Robots.txt Generator
Craft SEO-friendly robots.txt files with presets, validation, and live previews.
User-agent Groups
Allow rules
Disallow rules
No disallow rules. All paths are currently crawlable.
Custom directives
Add "Noindex: /path" or other vendor-specific directives.
Global settings
Optional and mostly used by Yandex. Leave blank for most sites.
Sitemaps
Add absolute URLs for each sitemap document.
Useful for vendor-specific fields like Clean-param or custom comments.
No global directives. Add one to append raw lines at the end.
Presets & automation
robots.txt preview
# robots.txt # Generated with the Utekar.com 's Robots.txt Generator Tool. Update contact and sitemap URLs. # Generated: 5/11/2025, 1:26:09 pm # Default rules for all user agents User-agent: * Allow: / Sitemap: https://example.com/sitemap.xml
Diagnostics
No issues detected. Your robots.txt looks healthy.
Groups
1
Allow rules
1
Disallow rules
0
Sitemaps
1
Custom directives
0
Total lines
7
Building an Effective robots.txt
Use this generator to produce high quality robots.txt files that respect search engine guidelines, protect private areas, and document your crawl strategy.
Core Concepts
- User-agent: Target specific crawlers with tailored rules. The wildcard
*applies to all bots. - Allow/Disallow: Grant or deny access to URL patterns. Keep rules specific to avoid unwanted blocking.
- Crawl-delay: Slow aggressive crawlers if your infrastructure needs breathing room.
- Sitemap: Always expose sitemap URLs so bots can discover content you want indexed.
Best Practices
- Keep critical resources such as CSS and JS accessible so search engines can render pages accurately.
- Pair robots.txt rules with on-page
metadirectives or HTTP headers for precise indexing control. - Document every rule with comments so future maintainers understand the rationale.
- Stage changes and verify the server actually returns HTTP 200 with
text/plaincontent.
How search engines interpret robots.txt
- The file is cached aggressively; announce changes by updating the modified timestamp or pinging search engines.
- Rules are evaluated top-to-bottom within a group, but the most specific match (longest path) wins.
Allowrules can reopen URLs blocked by broaderDisallowdirectives.- Google respects
Crawl-delayonly for select crawlers; Bing and Yandex use it more reliably.
Advanced Tips
- Create dedicated groups for specialised bots like Googlebot-News or AdsBot-Google so they can crawl at the right cadence.
- Use the global custom directives section for vendor-specific lines like
Clean-paramorRequest-rate. - Export and version-control your robots.txt so you can track historical changes.
- Remember that robots.txt is publicly accessible—do not place secrets here. Use authentication for true protection.
Deployment checklist
- Serve the file at the root:
https://example.com/robots.txt. - Ensure line endings are UNIX (LF) to avoid stray carriage returns.
- Validate absolute URLs for sitemaps and confirm they return 200 responses.
- Retest with Google Search Console → Robots Testing Tool before going live.
Common mistakes to avoid
- Blocking
/or/*unintentionally, which prevents all crawling. - Trying to hide sensitive files—robots.txt is public and only offers crawl guidance.
- Using relative sitemap URLs; most crawlers require absolute paths.
- Forgetting to update rules after migrating or renaming directories.
