Create a properly formatted robots.txt file to control how search engines crawl your website. Add rules for different user-agents, specify allowed and disallowed paths, and include your sitemap.
Optional: Include your XML sitemap location
Optional: Delay between crawler requests (not all bots honor this)
# Add rules above to generate robots.txt content
Save this content as robots.txt in your website's root directory.
The robots.txt file is a critical part of technical SEO. It tells search engine crawlers which pages they can and cannot access on your website, helping you control your crawl budget and protect sensitive areas.
User-agent: * Disallow: /
Blocks all bots from your entire site. Use for staging sites.
User-agent: * Disallow: /admin/ Disallow: /wp-admin/
Prevents indexing of administrative sections.
User-agent: * Allow: / Sitemap: https://example.com/sitemap.xml
Allows full access with sitemap reference.
User-agent: GPTBot Disallow: / User-agent: CCBot Disallow: /
Prevents AI training bots from scraping your content.
The robots.txt file must be placed in the root directory of your website and accessible at https://yourdomain.com/robots.txt. Search engines will only look for it at this exact location - placing it in a subdirectory won't work.
Start today and generate your first article within 15 minutes.