Frameworks

The new search measurement stack: what to track when Google traffic isn't the only signal

The SEO measurement stack most teams use today was designed for a search world that no longer exists. One engine (Google), one outcome (the click), one funnel (rank → traffic → conversion). That model was unambiguous, defensible, and produced clear quarterly reports for two decades. It's now incomplete in ways that distort strategy. This is the full framework for what to measure when Google traffic isn't the only signal — the new stack, layer by layer, with the metrics, the tools, and the reporting cadence.

By Gareth Hoyle Read time 13 min
TL;DR

The old measurement stack collapsed three things into one funnel: visibility (do you appear?), traffic (do they click?), and conversion (do they buy?). When Google was the only meaningful search engine and clicks were the dominant outcome, the simplification worked. Now you need to measure across five layers: category presence, citation behaviour, retrieval performance, brand effect, and funnel conversion. Each layer has its own metrics, its own tools, its own cadence. Teams that adopt this stack early will look prescient in 12 months. Teams that wait will be reacting to traffic declines under panic conditions instead.

The old stack — and what it can't see

Most SEO measurement still runs on a model that was unchallenged from roughly 2005 to 2022:

LayerOld metricWhat it measured
PositionAverage ranking positionWhere you appeared in Google SERPs
TrafficOrganic sessionsHow many clicks you got from search
ConversionGoal completionsWhat share of visitors took the action

This stack worked because the connection between the layers was reliable. Better rankings produced more impressions; more impressions produced more clicks; more clicks produced more conversions. The math compounded predictably. Quarterly reports practically wrote themselves.

The model is breaking in ways that affect strategic decisions:

The old stack reports "traffic was up 3% MoM" and treats that as the whole story. The reality could be: visibility down 15%, citation share down 12%, branded search up 22% (because AI is driving recall), conversion rate up 8% (because AI-introduced traffic has higher intent). Five trends, all strategic, none of which the old stack captures.

The new stack — five layers

Layer 01

Category presence — where you appear

The first question of search measurement isn't "how much traffic did you get?" — it's "did the buyer encounter your brand at all?" Presence is the impression. It's the moment the buyer might learn about you, regardless of whether they click.

For Google: this is impression share, broken out by query type (commercial vs informational, branded vs unbranded). Available from Search Console, but most teams don't segment rigorously enough.

For AI engines: this is Share of AI Voice — what percentage of AI engine responses to category queries mention your brand. The closest analogue to "ranking" in the AI era. Measured by running representative query sets through ChatGPT, Claude, Perplexity, and Gemini and counting brand presence.

For AI Overviews: this is the AIO presence rate (do your priority queries trigger an AIO?) and AIO citation rate (when they trigger, does your URL appear in the sources strip?). SerpAPI and similar SERP-tracking tools can measure this systematically.

MetricSourceCadence
Impression share by query typeGoogle Search ConsoleWeekly trend, monthly review
Share of AI Voice (4-engine)Multi-engine query runsMonthly
AI Overview presence rateSerpAPI / SERP toolsMonthly
AI Overview citation rateSerpAPI / SERP toolsMonthly
Layer 02

Citation behaviour — how you're credited

The new layer that didn't exist five years ago. When AI engines answer queries, they often cite their sources — explicitly (Perplexity's numbered references, AI Overview source strips) or implicitly (the framing the AI uses, which reflects which sources it weighted heavily).

Citation is the bridge between presence and influence. Even when buyers don't click, being credited as the source shapes their perception. Repeated citation in a category builds the kind of authority that turns "a brand the AI mentions" into "the brand the AI defaults to."

Citation behaviour also has a structural dimension worth tracking: the difference between direct citation (your URLs cited) and proxy citation (third-party content about you cited). Both matter, differently. Direct citation indicates your content is structurally extractable. Proxy citation indicates your category framing is being shaped by sources you don't control — sometimes good (favourable third-party coverage), sometimes a warning (the framing you'd want isn't on your own pages).

MetricSourceCadence
Citation rate (% of AI responses citing your URL when mentioning you)Multi-engine query runsMonthly
Citation share vs top 3 competitorsMulti-engine query runsMonthly
Direct vs proxy citation ratioMulti-engine query runsQuarterly
Citation persistence (run-to-run consistency)Multi-engine query runsQuarterly
Layer 03

Retrieval performance — how well your content extracts

This layer measures whether your content is technically and structurally optimised for AI engines to use it well. It's the GEO equivalent of technical SEO — the foundational layer that determines whether the higher-leverage work pays off.

Retrieval performance combines several technical signals: whether AI crawlers can access your pages (robots.txt allows them), whether content renders in initial HTML (server-side rendering for AI extraction), whether key facts are positioned for clean extraction (factual claims at the start of paragraphs, not buried), whether structured data marks up the most-extractable content cleanly.

Most teams overlook this layer because it doesn't show up in headline metrics. But it's the layer that determines whether the work in layers 1 and 2 produces results. A brand with high category presence and strong potential authority can still under-perform on AI visibility if its retrieval performance is weak — the AI tries to use the content, fails to extract cleanly, falls back to other sources.

MetricSourceCadence
AI crawler accessibility auditDirect robots.txt and server log reviewQuarterly
Schema validity rate (priority page types)Google Rich Results TestQuarterly
Server-side rendering coverageCrawl + Inspect in Search ConsoleQuarterly
Content extractability score (manual sample)Audit checklistAnnual deep-dive, spot-check quarterly
Layer 04

Brand effect — downstream influence

The hardest layer to measure cleanly, and therefore the most important to attempt. Brand effect is what happens after the buyer has encountered you (in search, in an AI answer, anywhere) — does it shape their behaviour later?

Direct measurement of brand effect is impossible. Triangulation across proxies is necessary. Three signals correlate strongly with AI-driven brand effect:

Branded search volume. When AI starts mentioning your brand in category answers, buyers who encounter you that way often search your brand name later for verification. A rising branded search trend, when your overall organic traffic is flat or declining, is a strong signal that AI engines are doing brand-building work the standard funnel doesn't capture.

Direct traffic baseline. Visitors arriving with no referrer — typing the URL, using a saved bookmark, clicking from a chat or email — represents brand recall. AI-driven recall shows up here. The baseline shouldn't be flat; it should grow with your category presence.

Demand-gen self-reporting. For B2B with sales pipeline, asking "what AI engines did you use during your research?" on demo bookings or first sales calls. Anecdotal, but the directional pattern over months is informative.

MetricSourceCadence
Branded search trend (12mo)Search Console + GA4Monthly
Direct traffic trendGA4Monthly
AI-source self-report on demosSales processQuarterly review
Brand mention monitoring beyond AIMention/Brandwatch/TalkwalkerMonthly
Layer 05

Funnel conversion — the click and what comes after

Clicks haven't died — they've reallocated and changed in shape. The conversion layer measures what happens when buyers do click through, with appropriate segmentation for the new traffic sources.

The most important segmentation: separating AI-introduced visitors from search-introduced visitors. Their behaviour patterns differ. AI-introduced visitors often arrive more validated (they've already heard the AI's answer; they're verifying or buying) or more confused (they came for a specific extracted answer that doesn't match what's on the page). Both patterns deserve separate tracking.

Conversion rate by traffic source matters more than overall conversion rate. Time-to-conversion matters more than absolute conversions. Assisted conversions matter more than last-touch conversions. The old funnel still exists; the parameters describing it have shifted.

MetricSourceCadence
Conversion rate by source (organic, AI-referred, direct)GA4 with proper attributionMonthly
Time-to-conversion by sourceGA4 + CRMQuarterly
Assisted conversions including AI as touchpointMulti-touch attribution modelQuarterly
Pipeline by source (B2B)CRMMonthly

How the layers connect

The strategic value of the five-layer stack is in how the layers connect. Each layer is a leading indicator for the layers above it in funnel terms, and a lagging indicator for the layers below.

Retrieval performance leads citation behaviour. If your retrieval performance is weak, citation behaviour will erode in the months following. If you fix retrieval performance (cleaner extraction, better schema, allowed crawler access), citation behaviour improves over the following 30-90 days.

Citation behaviour leads category presence. Citations build the AI engine's confidence in your brand as a category authority. Sustained citation rates pull up Share of AI Voice over time.

Category presence leads brand effect. When you appear consistently in category queries, buyers internalise your brand as a category fixture. Branded search and direct traffic respond on a 60-90 day lag.

Brand effect leads funnel conversion. Strong brand recall produces visitors who arrive ready to convert. Direct traffic and branded search bring higher-quality buyers than non-branded organic search.

The implication: when you see something move in layer 5 (conversion changes), the cause is usually 60-90 days upstream in layers 3 and 2. By the time conversion shifts, the strategic decisions that produced it are already old. The leadership team that monitors only conversion is always reacting to outcomes; the team that monitors the upstream layers can act earlier.

The reporting cadence

One of the failure modes when teams adopt new measurement frameworks is reporting too much. The five-layer stack has 15+ metrics across all layers. Reporting all of them monthly is exhausting, generates noise, and obscures the strategic picture.

The recommended cadence:

Monthly — operational dashboard

Three numbers per layer, no more. Headline metric only, with trend arrow. Designed to be reviewed in 5 minutes by the marketing leader, with deeper investigation only triggered when a metric moves outside expected range. The dashboard exists to answer "is anything changing in a way I need to act on?" — not to provide comprehensive insight every month.

Quarterly — strategic review

Full five-layer breakdown with all metrics, plus competitive context (where do you sit vs the top 3 competitors on each layer?), plus the leading-vs-lagging analysis (which layers moved this quarter, which are about to move next quarter as a result?).

This is the report that drives strategic decisions and investment cases. It's also the report that produces the budget conversation with leadership — translated into language a CFO can act on.

Annual — methodology audit

Once a year, audit the methodology itself. Are the queries you're tracking still representative of your category? Have new AI engines emerged worth adding to the tracking set? Have your competitors changed enough that the comparative set needs updating? Are there new metrics worth incorporating that weren't established a year ago?

The methodology should evolve. Locking it permanently produces increasingly stale signal. Annual evolution, with discipline about not changing too much in any one cycle, keeps the measurement fresh while preserving year-over-year comparability.

What this looks like in practice

Imagine a quarterly review six months into the new stack. The marketing leader presents:

"Layer 1 — category presence. We're at 36% Share of AI Voice across the four major engines, up from 28% three months ago and 19% nine months ago. AI Overview presence rate is at 41% of priority commercial queries, with our citation rate when present at 23%. Both up sequentially.

Layer 2 — citation. Direct citation rate is 31%, slightly behind our top competitor at 35%. Proxy citation rate is 47% — most of which comes from three industry blogs we don't control. We need to bring more of that framing onto our own pages.

Layer 3 — retrieval. We've fixed the SSR gaps in the comparison pages, schema is now valid across all priority page types, robots.txt allows all major AI crawlers. Layer 3 health is high, which is why layers 1 and 2 are moving.

Layer 4 — brand effect. Branded search up 27% YoY, direct traffic up 19%. Demo self-reporting shows 38% of buyers mentioning AI engines as a research tool, up from 22% a year ago.

Layer 5 — conversion. AI-source visitors converting at 4.2% (vs 2.8% for unbranded organic). Time-to-conversion 23 days shorter on AI-introduced. Pipeline contribution from AI-influenced sources at $1.4M this quarter, vs $400K same quarter last year.

Strategic implication: the work in retrieval and citation is paying off in brand effect and pipeline. Recommended Q4 investment — extend to four more comparison page builds, scale Digital PR to grow citation share, address the proxy-citation framing gap."

Every number traces to a layer. Every layer connects to the next. Every recommendation is grounded in a specific metric movement. The CFO has a defensible model to evaluate. The marketing leader has a clear story for what's working and what's next. The agency or in-house team has clear targets for the next quarter.

This is what measurement-driven strategy looks like in the AI search era. It's more layers than the old stack — but each layer carries information the old stack hid.

The transition path

For teams currently running the old stack and wanting to move to the new one, the transition isn't all-or-nothing. A phased approach over 90 days:

90-day adoption plan

From old stack to five-layer stack

Days 1-30 — Add Layer 1 (category presence). Establish your Share of AI Voice baseline across the four major engines. Add AI Overview presence tracking. Continue reporting old stack alongside.

Days 31-60 — Add Layer 2 (citation) and Layer 4 (brand effect). Layer in citation tracking from your existing AI engine query runs. Wire up branded search and direct traffic trends from existing analytics. Begin self-reporting questions on sales calls.

Days 61-90 — Add Layer 3 (retrieval) and refine Layer 5 (conversion). Audit and document retrieval health. Segment conversion analytics by source properly. Build the integrated quarterly report.

Day 91+ — Sunset old-stack-only reporting. Five-layer monthly dashboard becomes the standard format. Old stack metrics still appear within the relevant layers but no longer dominate the headline.

By the end of one quarter, the new stack is live, the old stack is integrated into it, and leadership has been gradually re-educated on what the metrics mean. The transition is complete without a disruptive single switchover.

The strategic conclusion

Most marketing leaders we talk to know intuitively that the old stack isn't telling them the full story anymore. They feel the gap between what their dashboards show and what's actually happening in their market. They also know they can't make a budget case to leadership using vibes — they need numbers.

The five-layer stack is the bridge. It's enough additional measurement to capture what the old stack misses, but disciplined enough to avoid metric overload. It's defensible to a CFO because every metric ties to a specific decision. And it's leading-indicator-rich, which means it shows where the business is going, not just where it's been.

The teams adopting it now will look prescient in twelve months when the rest of the industry catches up. The teams waiting will be reacting to traffic declines and missed pipeline targets without the diagnostic data to explain what happened. The choice is whether to lead the change or be forced to react to it.

The measurement model, more than any single tactic, determines which side you're on.

Run the new stack across five engines

Get a Search Visibility Audit.

Layer 1 (category presence), Layer 2 (citation), and Layer 3 (retrieval) measured across five AI engines and four SEO data providers — in one report. The foundation of the new measurement stack, ready for your monthly dashboard. From $997, in 48 hours.