Free tool

AI Robots.txt Generator

Control which AI crawlers can access your site

Build your AI robots.txt

18 AI crawlers

Toggle which AI crawlers can access your site. Copy or download the generated robots.txt.

18 allowed
0 blocked
Output
# robots.txt — Generated by PingAura AI Robots.txt Generator
# Generated: 2026-03-26

# Default: allow all crawlers
User-agent: *
Allow: /

# Allowed AI crawlers
User-agent: GPTBot
Allow: /

User-agent: ChatGPT-User
Allow: /

User-agent: OAI-SearchBot
Allow: /

User-agent: ClaudeBot
Allow: /

User-agent: Claude-Web
Allow: /

User-agent: PerplexityBot
Allow: /

User-agent: Google-Extended
Allow: /

User-agent: Googlebot
Allow: /

User-agent: bingbot
Allow: /

User-agent: Bytespider
Allow: /

User-agent: CCBot
Allow: /

User-agent: Amazonbot
Allow: /

User-agent: Meta-ExternalAgent
Allow: /

User-agent: FacebookBot
Allow: /

User-agent: Applebot-Extended
Allow: /

User-agent: cohere-ai
Allow: /

User-agent: Diffbot
Allow: /

User-agent: YouBot
Allow: /

Your robots.txt file is the first thing AI crawlers check before indexing your site. This generator lets you toggle access for 18 major AI bots — including GPTBot, ClaudeBot, and PerplexityBot — and instantly produces a standards-compliant robots.txt file. Control exactly which AI systems can crawl your content.

Toggle access for 18 AI crawlers with a single click — no manual syntax required.
Generate a valid, standards-compliant robots.txt ready to deploy at your domain root.
Pair with LLMs.txt for complete AI access control — decide who crawls and what they discover.

How it works

Get started in 3 simple steps

1

Toggle each AI crawler on (allow) or off (block) based on your strategy.

2

Optionally add your sitemap URL for crawler discovery.

3

Copy or download the generated robots.txt and publish it at your domain root.

Best use cases

Built for teams that take AI visibility seriously

1

Brands that want to appear in ChatGPT and Perplexity but block data-harvesting bots.

2

Publishers managing AI crawler access across multiple domains and subdomains.

3

AEO teams building an intentional AI access strategy as part of their optimization workflow.

Want continuous monitoring instead of one-off checks?

Start free trial

FAQ

Frequently asked questions

Where do I put the robots.txt file?

Place it at the root of your domain (e.g. yoursite.com/robots.txt). AI crawlers check this path automatically before indexing your content.

Can I allow some AI bots and block others?

Yes. Each bot gets its own User-agent block with Allow or Disallow rules. You might allow GPTBot and ClaudeBot for AI citation while blocking data-harvesting bots like CCBot or Bytespider.

Does blocking a bot remove existing content from their index?

Not immediately. Blocking a bot in robots.txt prevents future crawling. To request removal of already-indexed content, you need to use each platform's specific content removal process.

Should I include a sitemap in robots.txt?

Yes. Adding a Sitemap directive helps both traditional search engines and AI crawlers discover your content structure more efficiently.

How does this relate to LLMs.txt?

robots.txt controls crawl access (who can visit). LLMs.txt controls content discovery (what to prioritise). Use both together — robots.txt manages the gate, LLMs.txt guides the crawlers once inside.

Start for free

Turn AI visibility insights into growth

Create your workspace to monitor AI visibility and activate optimization workflows.