Skip to main content

Documentation Index

Fetch the complete documentation index at: https://docs.shareofmodel.ai/llms.txt

Use this file to discover all available pages before exploring further.

The GEO Audit evaluates each URL across three independent dimensions:
  • whether AI agents can reach the page,
  • whether they can parse its structure,
  • whether the content itself is easy to process.
Each dimension produces a score from 0 to 100. Run the audit on a set of URLs, then filter by score to identify the pages that need attention first.
GEO Audit

The three scores

Crawlability

Measures whether AI agents are technically allowed to access the page. The audit checks robots.txt for each major AI crawler (GPTBot, ClaudeBot, Gemini Agent, PerplexityBot and others), verifies your llm.txt configuration, and flags sitemap errors. It also records page response time, since slow responses can cause crawlers to time out. A low score means some or all AI agents may be blocked from reading the page entirely — regardless of how well-structured the content is.
ScoreWhat it means
85–100All major AI agents can access the page.
60–84Some agents are restricted or the page is slow to respond.
0–59Significant access issues — content likely invisible to multiple AI agents.

Semantic

Measures how well the page is structured for AI parsing. AI systems don’t read pages the way humans do. They rely on HTML structure — headings, paragraphs, lists, tables — to understand what a page is about and how to extract information. A page that looks well-designed in a browser may have very little structure in its underlying HTML. The Semantic score evaluates the presence and quality of these structural elements, adjusted for the type of page being audited. A product page is expected to have a clear title and feature lists; a FAQ page is expected to have a Q&A structure. The score reflects whether the page meets those expectations. The detected page type is shown as a badge next to the score. If the detection is incorrect, you can override it manually.
ScoreWhat it means
85–100Well-structured — AI agents can parse and extract content reliably.
65–84Acceptable structure — some improvements available.
40–64Structural gaps that may reduce how accurately AI agents interpret the page.
0–39Poor structure — content may be misread or ignored.

Readability

Measures whether the content is easy for an AI system to process and summarise. The score looks at the language itself: sentence length, vocabulary complexity, reading grade level, and sentence structure. It also checks whether images have descriptive alt text — AI agents cannot interpret images without it. Content that is dense, highly technical, or relies on complex sentence constructions may be hard to summarise or cite, even if it is perfectly accessible and well-structured.
ScoreWhat it means
75–100Clear, accessible content — easy for AI to process and summarise.
50–74Acceptable — some complexity may reduce summarisation accuracy.
0–49Dense or complex content — AI agents may struggle to extract key information.

AI Recommendations

Each audited URL includes an AI Recommendations panel that synthesises the three scores into a plain-English summary and a prioritised action list. The summary tells you how the page performs overall and where the most impactful opportunity lies. The recommendations below it are specific and ordered by impact — so if you can only fix one thing, you fix the right one first.

Reading the results

Each URL shows all three scores side by side. You can sort and filter by any score to prioritise fixes. The panel below each score lists the specific issues found and what to do about them — most actionable first.
A page can score well on Crawlability but poorly on Semantic or Readability. All three matter for AI citation: the agent needs to be able to reach the page, understand its structure, and process its content.

Common patterns

The page is accessible but poorly structured. Common on pages built with heavy client-side rendering, where visible content is generated by JavaScript and absent from the raw HTML. Improving server-side rendering or adding structured HTML markup typically resolves this.
The page has good structure but the content is difficult to process — common on legal, medical or highly technical pages. Add a plain-language summary or break long sentences into shorter ones.
Usually caused by an overly restrictive robots.txt that blocks AI crawlers as a group. Review your bot access policy and explicitly permit the agents you want to allow.

What’s next

Asset Evaluation

Score creative assets against brand attributes.

Search Visibility

Track domain and brand presence across engines.