What is the GEO Audit?
The GEO Audit evaluates each URL across three independent dimensions:
whether AI agents can reach the page,
whether they can parse its structure,
and whether the content itself is easy to process.
Each dimension produces a score from 0 to 100.
You can run the audit across a set of URLs and filter by score to identify which pages need attention first.
The three scores
Crawlability
Measures whether AI agents are technically allowed to access the page.
The audit checks your robots.txt file for each major AI crawler (GPTBot, ClaudeBot, Gemini Agent, PerplexityBot, and others), verifies your llm.txt configuration, and flags any sitemap errors. It also records page response time, since slow responses can cause crawlers to time out.
A low Crawlability score means some or all AI agents may be blocked from reading the page entirely — regardless of how well-structured the content is.
Score | What it means |
85–100 | All major AI agents can access the page |
60–84 | Some agents are restricted or the page is slow to respond |
0–59 | Significant access issues — content likely invisible to multiple AI agents |
Semantic
Measures how well the page is structured for AI parsing.
AI systems don't read pages the way humans do. They rely on HTML structure — headings, paragraphs, lists, tables — to understand what a page is about and how to extract information from it. A page that looks well-designed in a browser may have very little structure in its underlying HTML.
The Semantic score evaluates the presence and quality of these structural elements, adjusted for what is appropriate for the type of page being audited. A product page is expected to have a clear title and feature lists. A FAQ page is expected to have a Q&A structure. The score reflects whether the page meets those expectations.
The detected page type is shown as a badge next to the score. If the detection is incorrect, you can override it manually.
Score | What it means |
85–100 | Well-structured — AI agents can parse and extract content reliably |
65–84 | Acceptable structure — some improvements available |
40–64 | Structural gaps that may reduce how accurately AI agents interpret the page |
0–39 | Poor structure — content may be misread or ignored |
Readability
Measures whether the content is easy for an AI system to process and summarise.
This score looks at the language itself: sentence length, vocabulary complexity, reading grade level, and sentence structure. It also checks whether images have descriptive alt text, since AI agents cannot interpret images without it.
Content that is very dense, uses highly technical language, or relies on complex sentence constructions may be harder for an AI to accurately summarise or cite — even if it is perfectly accessible and well-structured.
Score | What it means |
75–100 | Clear, accessible content — easy for AI to process and summarise |
50–74 | Acceptable — some complexity may reduce summarisation accuracy |
0–49 | Dense or complex content — AI agents may struggle to extract key information |
AI Recommendations
Each audited URL includes an AI Recommendations panel that synthesizes the three scores into a plain-English summary and a prioritized action list.
The summary tells you how the page performs overall and where the most impactful opportunity lies. The recommendations below it are specific and ordered by impact — so if you can only fix one thing, you fix the right one first.
Reading the results
Each URL in the audit shows all three scores side by side. You can sort and filter by any score to prioritise fixes.
The panel below each score shows the specific issues found and what to do about them. Recommendations are ordered by impact — the most actionable fix is listed first.
A page can score well on Crawlability but poorly on Semantic or Readability. All three matter for AI citation: the agent needs to be able to reach the page, understand its structure, and process its content.
Common patterns
High Crawlability, low Semantic
The page is accessible but poorly structured. Common on pages built with heavy client-side rendering, where the visible content is generated by JavaScript and not present in the raw HTML. Improving server-side rendering or adding structured HTML markup typically resolves this.
High Semantic, low Readability
The page has good structure but the content itself is difficult to process — often found on legal, medical, or highly technical pages. Consider adding a plain-language summary or breaking long sentences into shorter ones.
Low Crawlability across all pages
Usually caused by an overly restrictive robots.txt that blocks AI crawlers as a group. Review your bot access policy and ensure the agents you want to allow are explicitly permitted.
