Methodology
What we measure (and what we don't promise)
GEO is a young, fast-moving field. AI search algorithms (Claude, ChatGPT, Perplexity) are private and shift frequently. We're upfront about which parts of our product are technical ground truth, which are plausible heuristics, and which are generated suggestions you should treat as input โ not as ranking guarantees.
High confidence โ direct measurement
These checks are deterministic: we read what's there and report it. If the audit says you're blocking OAI-SearchBot in robots.txt, you are. If the bot access test gets a 403, the bot can't reach you.
- โ robots.txt parsing โ exact rules per user-agent.
- โ AI bot access (CDN/WAF) โ actual HTTP status returned to each bot's user-agent.
- โ SPA / SSR ratio โ bytes of content visible without JavaScript vs. with.
- โ Schema.org / JSON-LD presence โ what's in the HTML right now.
- โ Technical basics โ TTFB, HTTPS, sitemap, canonical, redirect chain.
- โ Citation snapshots โ when AI search returns sources, we report those URLs verbatim.
Medium confidence โ observed correlation
These are patterns we see, not laws we've proved. The data is real; the inference is reasonable but not validated.
- โ "Page-type playbook" in source maps โ descriptive of which URL patterns get cited today, not a prescription for ranking. Top domains may be cited because they have authority, not because they use
/blog/URLs. - โ Topic clustering (buying / how-to / comparison) โ heuristic regex on query keywords. Works for English, breaks on ambiguous or non-English queries.
- โ Domain rollup ranking โ accurate for the queries we ran. Rankings shift week to week.
- โ "Why am I (not) cited?" reasoning โ Claude's plausible explanation post-hoc. It can't see the actual retrieval algorithm; treat it as a brainstorm.
Low confidence โ heuristics from traditional SEO
We borrow these from established SEO practice and apply them to AI search. They might transfer cleanly. They might not. There's not enough public research yet either way.
- โ llms.txt โ spec is new (2024โ2025). Anthropic respects it; OpenAI partially; Perplexity unclear. We surface its presence but don't claim it lifts citations.
- โ E-E-A-T signals (author, date, Person/Org schema) โ comes from Google's quality guidelines. Reasonable to assume LLMs prefer authoritative content; not empirically validated for citation rates.
- โ "Recommended schemas" per page type โ informed guesses based on what makes sense for AI parsing. Adding
FAQPagemay or may not move the needle. - โ The 0โ100 score itself โ our weighting (retrieval 30, SPA 25, llms.txt 15, schema 15, technical 10, E-E-A-T 5) is sensible but arbitrary. The score is useful as a proxy, not as an industry benchmark.
Generated content โ input, not answer
Anything Claude or another LLM produces for you is an opinion based on what it observed. We never claim AI-generated output will itself be cited.
- โ Content briefs (title, outline, FAQs, recommended sources) โ Claude's synthesis from the search step. Good editorial input, not a guaranteed-to-rank specification.
- โ First-draft markdown โ Claude writing prose. Not publish-ready. Treat as a starting point for a writer to edit, fact-check, and add real examples.
- โ Competitive gap analysis โ Claude reading a top URL and listing what it covers / doesn't cover. Useful for differentiation; not a ranking factor.
- โ Pre-generated JSON-LD โ technically valid markup. Adding it doesn't guarantee anything beyond making your structured data parseable.
What we don't promise
- โ We don't promise that following our recommendations will increase your AI citation rate.
- โ We don't promise that a "Crawler-ready" score means you'll be cited by Claude / ChatGPT / Perplexity.
- โ We don't promise that AI rankings observed today will hold tomorrow.
- โ We don't promise that Claude's explanations of "why" are causally accurate โ they're plausible narratives, not algorithm introspection.
- โ We don't promise that briefs generated by us will outrank existing top-cited pages.
Where the real value sits
The strongest part of the product is the measurement loop: snapshot what AI cites today, change something on your site, snapshot again, see if it moved. That's defensible regardless of which heuristics turn out to be right. We're building toward better empirical grounding as the field matures.
We publish what we observe across all audited domains every month at /insights โ confidence-tiered, with full caveats. No other GEO tool does this.
If you have data โ public or private โ that strengthens or contradicts any of the above, we want to hear it. Email hello@citeai.io.
Want the short version? See how we frame this vs the rest of the GEO market โ
Run a free audit