Foundations

How accurate can LLM visibility tracking actually be?

Every "23.4% share of voice in ChatGPT" number you've ever seen is an estimate, not a fact. LLMs have memory. They generate stochastically. They personalize by account. The model itself silently changes underneath you. None of that means tracking is impossible — it just means anyone selling pixel-perfect AI visibility numbers is overselling. Here is what's actually knowable, what isn't, and what to demand from any tool that claims to measure your AI visibility.

By Gareth Hoyle Read time 12 min
TL;DR

You can only get clean LLM visibility data from cold-start sessions — fresh chats with no prior context. Once a model has any context, the signal degrades. On top of that, stochastic generation, account personalization, real-time retrieval shifts, and silent model updates each add noise. Best-in-class methodology controls for all of these and reports variance alongside means. The resulting number is a directional estimate accurate enough to make decisions with — but anyone claiming higher precision than that is selling a feeling, not a measurement.

The cold-start problem

Run an experiment. Open a new ChatGPT chat. Ask: "What's the best CRM for a 50-person SaaS company?" Note the answer. Then in the same conversation, ask: "What about for nonprofits?" The second answer will be biased by the first. The model remembers what you've already discussed. It carries context. It assumes continuity.

This is the foundational accuracy challenge in LLM visibility tracking, and it has a clean name: the cold-start problem. To get an unbiased measurement of how a brand surfaces in AI answers, every prompt must be asked in a fresh session, with no prior context, no carried history, no priming. The moment a session has been used for one query, it can't be used cleanly for another.

You can only really track first searches. Every subsequent prompt in the same conversation is contaminated.

This isn't a quirk. It's the design of modern conversational LLMs — they're built to maintain coherent dialogue, which means they're built to remember. The same property that makes them feel smart in conversation makes them unreliable as repeat measurement instruments.

What "cold-start" looks like in practice

For each prompt being measured, the tracking tool opens a brand-new session, asks the question once, captures the response, and closes the session. No follow-up questions in the same chat. No context retained. Each measurement is a clean room. The cost: this is slower and more expensive than running batches of prompts in long-running sessions. The benefit: it's the only way to get an unbiased reading.

Memory is just one source of noise

The cold-start problem is the cleanest example because it's the most binary — context either contaminates a measurement or it doesn't. But it's only one of seven significant sources of measurement noise that any honest tracking methodology has to address.

1. In-session memory (the cold-start problem)

Already covered. Solved by fresh sessions for every prompt.

2. Cross-session memory features

ChatGPT now has a Memories feature that persists context across separate conversations. Claude has Projects with persistent context. Gemini ties results to the user's Google account history. Even with a brand-new chat, a logged-in user gets answers shaped by past usage.

The control: track from logged-out states wherever possible, or use API access (which has cleaner isolation than the consumer apps), or use fresh accounts that haven't accumulated history. None of these fully replicate the experience of an everyday user — but they remove the user-specific personalization layer, which is what you want for an apples-to-apples brand measurement.

3. Stochastic generation

LLMs are not deterministic. Ask the same question twice and you'll get different answers — sometimes meaningfully different. The model samples tokens probabilistically; small variations propagate.

The control: run each prompt multiple times. Five is a minimum, nine is better, more is diminishing returns. Report the mean across runs and — critically — the variance. A brand that appears in 9 of 9 runs is in a different position from one that appears in 4 of 9, even if both produce a "44–100%" range. The variance is information, not noise.

4. Personalization by region, language, and account state

Search-augmented LLMs (Perplexity, ChatGPT with browsing, Google AI Overviews) consider user location for retrieval. The same prompt asked from London and from New York will retrieve different sources. Logged-in users get different inferences than logged-out ones. Models trained for international markets sometimes localize answers based on detected language even when the prompt is in English.

The control: declare your measurement context explicitly. "We measured from a logged-out US-based session in English at 14:00 UTC." Without that disclosure, two reports of "share of voice in ChatGPT" aren't comparable.

5. Real-time retrieval shifts

Engines that pull from the live web — Perplexity, AI Overviews, ChatGPT search — are reading a moving target. A new article published this morning can shift today's answers. A competitor's PR push that lands on Reuters at noon shows up in Perplexity by mid-afternoon. The "current" picture is genuinely a different picture every few hours.

The control: define your measurement window tightly. Don't average data collected across two weeks and call it a single point. For trend tracking, run measurements at the same time of day on a consistent cadence — daily or weekly, not "whenever we get to it."

6. Silent model updates

OpenAI, Anthropic, Google, and Perplexity all update their models without notice. GPT-4o gets a quiet refresh. Gemini's retrieval pipeline changes. Perplexity swaps the underlying generation model. Your visibility number can move five points overnight not because anything happened to your brand, but because the engine itself changed.

The control: track which model version was used (where the API exposes it) and flag step-changes in the data that align with public model release notes. Treat large unexplained shifts as suspect until you've ruled out a model update.

7. Prompt phrasing variance

"Best CRM for SMBs" and "what CRM should a small business use" and "recommend CRM software" return different answers. They activate different latent associations in the model. A tracking system that tests five prompts per topic captures less of the user-query distribution than one that tests fifty.

The control: large prompt panels (40–50 minimum for a meaningful brand measurement), drawn from real query data where possible, and grouped by intent rather than treated as interchangeable.

What "as accurate as possible" actually looks like

Stack the seven controls together and you get a methodology. None of these are exotic. The work is in doing all of them, every time.

Control What it looks like in practice
Cold-start sessions Every prompt asked in a brand-new chat with zero prior context
Account isolation Logged-out where possible, or API access, or fresh accounts
Multi-run sampling 5–9 runs per prompt, with variance reported alongside means
Context disclosure Declared region, language, time window, model version
Tight measurement window All data collected within hours, not weeks
Model version tracking Note which engine version produced each result
Wide prompt panel 40–50+ prompts grouped by intent, drawn from real queries

A measurement run that does all seven of these is roughly as accurate as anyone can ever tell you, given the technology. There's a hard ceiling here — set by the way LLMs work, not by the quality of the tooling — and the methodology above is what reaches it.

Anyone claiming above-ceiling precision is either using a different definition of accuracy or selling a story.

Triangulation: the only way past single-source uncertainty

One more layer matters. Even with all seven controls in place, a single-engine measurement is still a single-engine measurement. ChatGPT-only data tells you about ChatGPT — not about AI visibility broadly.

This is why methodology that's serious about accuracy triangulates across engines. Five engines — ChatGPT, Claude, Gemini, Perplexity, and Google's AI Overviews — give you a much wider view than any one. When four of five agree on your visibility ranking, you have a defensible cross-engine signal. When they disagree sharply, that disagreement is itself information: it tells you the brand is performing differently in retrieval-heavy engines versus training-heavy ones, or in real-time engines versus snapshot ones.

The same principle applies to the SEO authority side of the picture. AI visibility correlates strongly with SEO authority — at r = 0.89 in our reference dataset, meaning roughly 79% of the variance in AI citation rate is explained by underlying SEO authority. So serious measurement triangulates SEO authority too: across Ahrefs, Semrush, Moz, and DataForSEO. Where those four providers agree within a few points, you have a robust authority number. Where they diverge, the divergence flags something specific worth investigating.

Nine independent data sources — five AI engines, four SEO authority providers — give you a measurement defensible enough to take to a board. Anything less is a partial picture.

What you should demand from any tracking tool

Practical evaluation criteria. If you're considering an AI visibility tracking tool — SaaS dashboard, audit, monitoring service, anything — these are the questions that separate substantive methodology from marketing copy.

A tool that answers all seven cleanly is doing serious work. A tool that hand-waves any of them isn't selling measurement — it's selling a vibe.

The directional truth

Here's the resolution. None of the noise sources discussed here mean LLM visibility tracking is unreliable. They mean it's directional. With strong methodology, you get numbers that:

What they don't give you is a single perfect "your share of voice is exactly 23.4%" precision. Anyone presenting a number to that decimal place is conveying false confidence. The honest version is "23%, plus or minus a few points, in this engine, in this region, in this measurement window, with this confidence interval."

For the decisions marketing leaders need to make — where to invest content effort, which competitors are pulling ahead, which topic clusters need work — directional accuracy is plenty. Most decisions don't require pixel precision; they require a clear-enough signal of what's moving and where the gaps are.

The methodology test

Two reports could both say "your brand has 18% share of voice in ChatGPT." One was produced by a single uncontrolled session running 10 prompts back-to-back, with no run-to-run variance reporting, from a logged-in account. The other was produced by 50 prompts in cold-start sessions, run nine times each, from logged-out access, in a defined window, with variance reported per prompt.

Both numbers are 18%. Only one is a measurement.

That's the question to ask of any AI visibility data. Not "what's the number?" — but "how was the number produced?" The methodology is the work. The number is just the artifact.

Closing

LLM visibility tracking is achievable, defensible, and — with the right methodology — accurate enough to make real decisions on. It also has a hard ceiling on precision, set by how the engines work, that no amount of tooling can break through. Pretending otherwise is either an honest mistake or a sales tactic.

The brands that handle this well are the ones who treat their AI visibility numbers as directional signals — useful, reliable, but bounded. They invest in clean methodology, ask the right methodology questions of vendors, and use the resulting data for what it's actually good at: identifying where to focus, tracking whether work is moving the needle, and benchmarking against competitors with confidence.

That's as accurate as any tool can ever tell you. And for the work it needs to inform, that's plenty.

Methodology you can audit

Get a Search Visibility Audit.

Cold-start sessions. Nine runs per prompt. Five AI engines. Four SEO authority sources. Variance reported alongside every number. Full prompt log included. From $997, delivered in 48 hours.