Robots.txt Generator

Create a robots.txt file for your website. Control how search engines crawl your site.

Site Info

Rule #1

Generated robots.txt

# robots.txt generated by Clarity (getclarityseo.com/robots)

User-agent: *
Allow: /

πŸ“– What is robots.txt?

A robots.txt file tells search engine crawlers which pages they can and can't access on your site. It lives at yourdomain.com/robots.txt.

Disallow blocks a path from being crawled. Allow explicitly permits crawling.

Sitemap tells crawlers where your XML sitemap lives for better indexing.

⚠️ robots.txt doesn't prevent pages from being indexed if they're linked elsewhere. Use noindex meta tags for that.

πŸ€– AI Crawler Blocking

Many AI companies crawl websites to train their models. You can block them by adding specific user-agent rules:

  • GPTBot β€” OpenAI's crawler
  • ChatGPT-User β€” ChatGPT browsing
  • CCBot β€” Common Crawl (used by many AI models)
  • Google-Extended β€” Google's AI training crawler
⚑ Clarity Pro β€” Now Available
Scheduled audits, rank tracking, competitor analysis, and white-label reports.

Technical SEO Guardrails in Pro

Keep robots directives, crawl controls, and technical checks aligned over time.

Related Tools