AI Crawler Access Control: Block Training, Allow Search

6 min read
AI CrawlersWordPressPrivacy

Not all AI crawlers want the same thing

A common mistake WordPress site owners make is treating all AI crawlers the same. Some block every AI bot out of privacy concerns. Others allow everything, including bots that harvest content for model training without any attribution.

The reality is nuanced. AI crawlers fall into two distinct categories, and treating them differently is not just possible — it is essential for a balanced AI strategy.

Training crawlers collect your content to build or fine-tune AI models. Once your content is absorbed into a model's training data, it becomes part of the model's general knowledge. There is no attribution, no link back to your site, and no traffic. Your content simply makes the AI smarter.

Search crawlers collect your content to provide real-time, cited answers. When a user asks Perplexity or ChatGPT a question and your content is relevant, these crawlers retrieve it and cite your site as a source. You get attribution and potentially traffic.

The distinction matters enormously for publishers, bloggers, and businesses. Blocking all AI crawlers means giving up AI search visibility. Allowing all AI crawlers means your content trains competing AI models for free. The correct approach is selective: block training, allow search.

Understanding the major AI crawlers

Here is a breakdown of the most active AI crawlers and their primary purposes:

CrawlerOperatorPrimary purpose
GPTBotOpenAISearch and training (configurable)
ChatGPT-UserOpenAIReal-time search only
ClaudeBotAnthropicTraining and search
PerplexityBotPerplexitySearch only
Google-ExtendedGoogleGemini training
CCBotCommon CrawlOpen dataset for training
BytespiderByteDanceTraining
Applebot-ExtendedAppleApple Intelligence training
meta-externalagentMetaTraining
AmazonbotAmazonAlexa/training

Some crawlers serve dual purposes. GPTBot, for example, crawls for both training data and ChatGPT's search feature. OpenAI allows you to configure access separately through robots.txt directives.

Configuring robots.txt for selective access

The primary mechanism for controlling AI crawler access is your robots.txt file. Here is a configuration that blocks training crawlers while preserving search visibility:

# Block training-only crawlers
User-agent: CCBot
Disallow: /

User-agent: Google-Extended
Disallow: /

User-agent: Bytespider
Disallow: /

User-agent: Applebot-Extended
Disallow: /

User-agent: meta-externalagent
Disallow: /

# Allow search crawlers
User-agent: ChatGPT-User
Allow: /

User-agent: PerplexityBot
Allow: /

User-agent: GPTBot
Allow: /

This configuration blocks known training-only crawlers while allowing the bots that power AI search features. Note that GPTBot serves both purposes, but blocking it would also block ChatGPT search — so most publishers choose to allow it.

Beyond robots.txt: HTTP headers and meta tags

robots.txt is a blunt instrument. It blocks entire pages or directories, and it operates on the honor system — crawlers are not required to obey it (though reputable ones do).

For more granular control, you can use HTTP headers and meta tags:

X-Robots-Tag HTTP header

X-Robots-Tag: noai
X-Robots-Tag: noimageai

These headers tell AI crawlers not to use the page content or images for training. They are more granular than robots.txt because they can be applied per-page or per-content-type.

Meta tags

<meta name="robots" content="noai, noimageai">

The meta tag equivalent, applied at the page level. Useful when you want to allow AI crawling for most of your site but protect specific pages.

AI-specific meta tags

Some AI platforms recognize proprietary meta tags:

<meta name="ai-training" content="disallow">
<meta name="ai-search" content="allow">

These explicitly separate training permission from search permission, giving you the most precise control available.

Implementing access control on WordPress

Manually editing robots.txt, adding HTTP headers, and inserting meta tags across a WordPress site is tedious and error-prone. Changes need to be maintained as you add new content, update existing pages, and adjust your AI strategy.

Arvo GEO simplifies this with a visual interface for AI crawler access control. You can:

  • Set global rules for each AI crawler (allow, block, or custom)
  • Override rules for specific post types, categories, or individual pages
  • Auto-generate appropriate robots.txt directives
  • Add AI-specific meta tags site-wide or selectively
  • Monitor which crawlers are actually respecting your rules

This means you can implement a nuanced access control strategy in minutes rather than hours, and adjust it as the AI crawler landscape evolves.

Common access control strategies

Different types of sites call for different approaches:

Strategy 1: Open search, closed training

Allow all search crawlers, block all training crawlers. Best for publishers and content creators who want AI search visibility but do not want their content used for model training.

Strategy 2: Selective content access

Allow all AI crawlers on marketing pages and blog posts, but block them from premium content, gated resources, or proprietary data. Best for sites with both public and private content.

Strategy 3: Full block with exceptions

Block all AI crawlers by default, then selectively allow specific bots on specific content. Best for sites with sensitive information that want minimal AI exposure.

Strategy 4: Full access

Allow all AI crawlers everywhere. Best for sites whose primary goal is maximum exposure, such as documentation sites, open-source projects, or public educational resources.

Monitoring compliance

Setting robots.txt rules is only half the battle. You also need to verify that crawlers are respecting your directives.

Track AI crawler activity on your site and compare it against your access control rules. If a crawler you have blocked is still appearing in your logs, you may need to escalate to server-level blocking via IP ranges or firewall rules.

Arvo GEO's crawler tracking dashboard makes this easy — you can see which bots visit, how often, and whether they are accessing pages they should not be.

The evolving legal landscape

AI crawler access control is not just a technical decision — it has legal dimensions. Copyright law around AI training is actively evolving, and several jurisdictions are considering or have passed regulations that give publishers more control over how their content is used by AI systems.

By implementing explicit access control now, you are documenting your preferences in a machine-readable format. This creates a clear record of consent (or non-consent) that may become legally relevant as regulations solidify.

Getting started today

If you do nothing else, take these three steps:

  1. Check your current robots.txt — Are you accidentally blocking search crawlers? Or accidentally allowing training crawlers?
  2. Install crawler monitoring — You cannot manage what you do not measure. Start tracking which AI bots visit your site.
  3. Implement selective access — Block training crawlers, allow search crawlers, and set up a system to maintain these rules as new crawlers emerge.

The AI crawler landscape is evolving rapidly. New bots appear regularly, existing bots change their behavior, and the distinction between training and search continues to sharpen. A proactive, well-maintained access control strategy keeps you in control of how your content is used.