How to Audit Your WordPress Site's AI Search Visibility
Most WordPress sites are invisible to AI search
Here is an uncomfortable truth: your WordPress site may be completely optimized for Google and completely invisible to AI search engines like ChatGPT, Perplexity, and Claude.
Traditional SEO audits do not check for AI visibility. They evaluate backlinks, keyword rankings, page speed, and meta tags — all important for Google, but insufficient for the AI search channel. A site can rank on page one of Google and still never be cited in a single AI-generated answer.
An AI search visibility audit closes this gap. It systematically evaluates every factor that determines whether AI models can discover, understand, and cite your content.
The complete AI visibility audit checklist
This audit covers seven areas. Work through each one systematically, documenting findings and prioritizing fixes.
Area 1: Crawler access
The most fundamental question: can AI crawlers actually reach your content?
Check your robots.txt file. Navigate to yourdomain.com/robots.txt and look for rules that affect AI crawlers:
- Is GPTBot blocked? If so, ChatGPT cannot access your content for search
- Is PerplexityBot blocked? This prevents Perplexity from citing your site
- Is ClaudeBot blocked? Claude will not be able to reference your content
- Are you blocking all bots with a wildcard rule that inadvertently catches AI crawlers?
Check for meta robots tags. Some WordPress themes or plugins add restrictive meta robots tags that prevent crawling. Look for noindex or nofollow tags on pages you want AI-visible.
Check server-level blocking. Firewalls, CDN rules, or security plugins may block AI crawlers by IP range or user-agent. This happens more often than you might expect, especially with aggressive security configurations.
What to fix: Allow search-oriented AI crawlers (GPTBot, ChatGPT-User, PerplexityBot, ClaudeBot) while optionally blocking training-only crawlers (CCBot, Google-Extended, Bytespider).
Area 2: llms.txt file
An llms.txt file is a machine-readable guide to your site for AI crawlers. Check whether you have one and whether it is properly configured.
Does yourdomain.com/llms.txt exist? If not, you are missing a key AI discovery signal.
Is it well-structured? A good llms.txt includes your site name, description, key content categories, and links to your most important pages.
Is it current? An outdated llms.txt that references deleted pages or missing content categories is worse than no llms.txt at all.
What to fix: Create or update your llms.txt file. On WordPress, Arvo GEO auto-generates this file from your published content and keeps it current as you add or modify posts.
Area 3: Content structure
AI models extract information most effectively from well-structured content. Audit your key pages for structural quality.
Heading hierarchy. Does each page have a single H1, with logically nested H2 and H3 headings? Skipped heading levels (jumping from H1 to H3) confuse AI parsers.
List and table usage. Are factual points presented in bulleted or numbered lists? Are comparisons presented in tables? These formats are far easier for AI to extract than prose paragraphs.
Paragraph length. AI models handle short, focused paragraphs better than long blocks of text. Paragraphs exceeding 5-6 sentences should be broken up.
Question-answer patterns. Does your content include explicit questions as headings with direct answers in the following paragraph? This pattern directly maps to how users query AI search engines.
What to fix: Restructure your top 20 pages by traffic, focusing on heading hierarchy, list formatting, and Q&A patterns.
Area 4: Schema markup
Structured data gives AI crawlers machine-readable context about your content type, structure, and key information.
Check for existing schema. Use Google's Rich Results Test or view page source to see what JSON-LD is present. Common findings:
- Many WordPress sites have basic Article or WebPage schema from their SEO plugin
- Few have FAQPage, HowTo, or Speakable schema
- Some have broken or invalid schema that provides no value
Key schemas for AI visibility:
| Schema type | Purpose | AI impact |
|---|---|---|
| Article | Identifies content type and metadata | Baseline — every post should have this |
| FAQPage | Marks up Q&A pairs | High — directly extractable by AI |
| HowTo | Structures step-by-step content | High — maps to procedural queries |
| Speakable | Identifies voice-friendly content | Medium — used by voice AI assistants |
| Organization | Establishes site identity | Medium — helps AI understand your brand |
What to fix: Add FAQPage schema to posts with FAQ sections. Add HowTo schema to tutorial content. Ensure Article schema is present and valid on every post.
Area 5: Content depth and freshness
AI models evaluate content quality through signals like depth, freshness, and factual density.
Content length. Flag pages under 300 words — they rarely provide enough depth for AI citation. Prioritize pages between 800-2,500 words of substantive content.
Publish and update dates. Check when your key content was last updated. Content older than six months may be deprioritized for time-sensitive topics.
Factual density. Does your content include specific data points, statistics, named sources, and concrete examples? Generic advice without supporting evidence is less citable.
What to fix: Update stale content with fresh data. Expand thin pages to substantive depth. Add specific facts and examples to generic content.
Area 6: Internal linking
AI crawlers use internal links to discover content and understand topical relationships between pages.
Orphaned pages. Identify pages with zero or very few internal links pointing to them. These are difficult for AI crawlers to discover.
Link depth. How many clicks from the homepage does it take to reach your key content? Content more than three clicks deep is less likely to be crawled.
Anchor text quality. Are internal links using descriptive anchor text or generic "click here" text? Descriptive anchors help AI understand what the linked page covers.
What to fix: Add internal links to orphaned content. Reduce link depth for important pages. Replace generic anchor text with descriptive phrases.
Area 7: AI crawler activity
The ultimate test of AI visibility: are AI crawlers actually visiting your site?
Check server logs. Look for user-agent strings containing GPTBot, ClaudeBot, PerplexityBot, ChatGPT-User, and other AI crawlers.
Measure crawl frequency. How often do AI bots visit? Daily crawls indicate healthy visibility. No crawls in 30+ days suggest a discovery problem.
Identify popular pages. Which pages do AI crawlers access most? This reveals what AI models consider your most valuable content.
What to fix: If AI crawlers are not visiting, focus on crawler access fixes first. If they are visiting but infrequently, improve content quality and llms.txt to encourage more frequent crawling.
Creating your audit report
Document your findings in a simple format:
- Current state — Summarize what you found in each area
- Critical issues — Problems that prevent AI visibility entirely (blocked crawlers, missing llms.txt)
- Optimization opportunities — Improvements that will increase citation likelihood
- Priority actions — Ranked list of fixes by impact and effort
Automating your audit with Arvo GEO
While you can perform this audit manually, Arvo GEO automates most of the assessment. It scores every page on your WordPress site for AI readiness, tracks crawler activity in real-time, manages your llms.txt file, and identifies structural issues that reduce AI visibility.
Start with the automated scan to get baseline scores, then use the manual checklist above to dig deeper into specific areas where your scores are lowest.
After the audit: building an optimization plan
An audit without action is just documentation. Take your prioritized list of fixes and schedule them:
- Week 1: Fix critical access issues (robots.txt, blocked crawlers)
- Week 2: Deploy llms.txt and schema markup
- Week 3-4: Restructure and refresh top content
- Ongoing: Monitor crawler activity and iterate
Revisit your audit quarterly to catch new issues and measure progress. AI search is evolving rapidly, and your optimization strategy needs to evolve with it.