OAI-SearchBot, GPTBot, and robots.txt: How to Control AI Access to Your Site
What OAI-SearchBot, GPTBot, and ChatGPT-User actually do, how to structure robots.txt rules without confusion, and how not to accidentally block ChatGPT search inclusion.
If you need to see whether this work is actually improving visibility in ChatGPT, Google AI, Perplexity, Alice, and other systems, GEO Scout helps track brand mentions, positions, and cited sources across live prompts.
There is now a lot of confusion around AI bots. Some teams allow everything without understanding the consequences. Others block both GPTBot and OAI-SearchBot in one line and then lose the chance to appear in ChatGPT search. In practice, the right approach is not “allow AI” versus “block AI,” but separating search, training, and user-triggered access as different policies.
Why one robots.txt file now affects several AI use cases
OpenAI uses different user agents for different jobs. That means robots.txt is no longer just a search setting. It is also a product-policy control: do you want to participate in ChatGPT search, do you allow training crawl, and how should user-triggered visits behave? The clearer that separation is, the lower the risk of shutting down the wrong channel by mistake.
Which scenarios to distinguish
- OAI-SearchBot controls appearance in ChatGPT search.
- GPTBot is tied to crawling that may be used for model training.
- ChatGPT-User is for user-triggered visits and actions.
- robots.txt rules should reflect real site sections, not a vague “allow all” policy.
- ChatGPT search behavior needs time to update after robots.txt changes.
How to structure access control
1. Separate search from training
If the business wants to appear in ChatGPT search but not allow training crawl, that logic should be split by user agent rather than handled through one blanket rule.
2. Check technical blocking layers
Even if robots.txt is open, the site can still be effectively closed by CDN rules, security plugins, rate limiting, or anti-bot settings at the hosting layer.
3. Expose only the valuable answer layer
There is no reason to open admin URLs, private sections, or noisy technical pages. Open the content you would actually want to surface in cited sources and search visits.
Implementation order
- Define policy first: do you want ChatGPT search inclusion, GPTBot access, and which sections should stay private?
- Set separate rules for OAI-SearchBot and GPTBot instead of relying on one global block.
- Check that key sections are not being blocked by CDN, WAF, anti-bot, or login requirements.
- Allow crawl on articles, FAQ, product pages, service pages, About pages, and other target URLs.
- Recheck search behavior after roughly 24 hours on branded and commercial prompts.
Common mistakes
- Adding a blanket block that accidentally removes ChatGPT search access.
- Confusing GPTBot and OAI-SearchBot, then diagnosing the wrong problem.
- Reviewing only robots.txt and ignoring CDN, WAF, and security-layer blocking.
- Allowing noisy utility pages while keeping the real answer layer weak or closed.
- Failing to document the policy and losing it on the next deployment.
Quick checklist
- The team knows whether ChatGPT search visibility is desired.
- OAI-SearchBot and GPTBot rules are separated and intentional.
- Important pages are not blocked by security infrastructure.
- Only high-value answer URLs are exposed.
- Everyone understands which rules affect search and which affect training.
- Behavior is rechecked after the policy change propagates.
Related reading
- Technical site checklist for AI systems
- Google AI Mode: the practical SEO guide
- How to configure CMS and hosting for IndexNow and AI bots
Частые вопросы
If you block GPTBot, will the site disappear from ChatGPT search?
What is ChatGPT-User for?
How quickly do robots.txt changes affect ChatGPT search?
Related Articles
Google AI Mode: The Practical SEO Guide for Websites and Brands
What actually matters for appearing in Google AI Mode. How to align classic SEO, page structure, Merchant Center, business profiles, and content that AI can safely use as a source.
Technical Checklist: How to Prepare Your Website for AI Indexing
Complete technical checklist for preparing your website for AI bot indexing: robots.txt, llms.txt, JSON-LD, Schema.org, sitemap.xml, loading speed, full table of all AI bots and user agents.