• Portfolio
  • About
  • Blog
  • Contact
Menu

Retina Media

  • Portfolio
  • About
  • Blog
  • Contact

The GEO White Paper v4.0: AI-Native Brand Visibility for B2B (March 2026)

March 6, 2026

Introduction

It’s an increasingly common scenario: your sales rep walks into a call and the buyer already has a shortlist. They know pricing bands, have integration concerns lined up, and raise a competitor's "known gap" unprompted. They even ask about a specific rate limit they "read about somewhere." 

The buyer built this list by asking ChatGPT, Perplexity, or Gemini for recommendations, and your brand either appeared in those responses or it didn’t. By the time sales enters the picture, the invisible committee member has already voted.

This is the reality that generative engine optimization (GEO) addresses. Not search engine rankings. Not website traffic. The question of whether your brand is visible, credible, and accurately represented in the AI-generated responses where your buyers are actually doing their research.

This document synthesizes nine months of accumulated research, over a hundred discrete findings across 40+ sources, and active client engagement data into the most comprehensive analysis of AI-native brand visibility available for B2B marketers. It is not a how-to guide. It is a state-of-the-field intelligence report designed to give marketing leaders the evidence, the strategic framework, and the honest assessment of what’s known and unknown that they need to make investment decisions in a category that didn’t exist two years ago.

Five key findings

  1. Nearly all B2B buyers now use AI tools during their purchase journey. The shortlist is forming before your sales team enters
    the room.

  2. The GEO tools market is growing fast, but dashboards show where you’re losing, not what to build. The gap between visibility diagnosis and actionable strategy is where most programs stall.

  3. Organic traffic explains 5% of AI citation behavior. Backlinks explain less than four percent. The skills transfer from SEO. The scoreboard is entirely different.

  4. 89% of citation opportunities are platform-specific. A strategy built for one AI platform fails on the others.

  5. If you don’t fill the information gap, someone else will, and the models will prefer their version because it’s more specific than yours.

Taken together, these findings point to a strategic conclusion: the current moment is the cheapest it will ever be to establish AI visibility. Every trajectory in Section H makes it harder and more expensive later.

How to use this document

This white paper is designed to be read in full or by section. Sections A through C establish the strategic argument. Sections D and E present the evidence base. Section F addresses risk. Section G describes what serious programs measure and prioritize. Sections H and I look forward.

Three supporting essays on retina.media (“The Committee Member Nobody Invited,” “GEO Tools Can’t Tell You What to Build,” and “The Trial of Traditional SEO”) develop individual arguments from this research in standalone form.

To discuss how these findings apply to your brand’s AI visibility strategy, contact shane@retina.media or visit retina.media.

A. The Pipeline Gate: AI as the Invisible Committee Member

The scenario described in the introduction isn’t an edge case. 6sense's 2025 B2B Buyer Experience Report found that 94% of buyers use LLMs during their buying process. G2’s August 2025 survey of over 1,000 B2B software buyers found that 87% say AI chatbots are changing how they research vendors, and half now start their research in AI rather than Google. Responsive’s data puts 25% of B2B buyers already using generative AI more than traditional search. TrustRadius reports 56% of tech buyers rely on AI chatbots as a top source for vendor discovery. 

These are not early-adopter numbers. This is the center of the market.

But the adoption data, by itself, tells a misleading story. It implies the challenge is visibility: that brands need to “show up” in AI responses the way they showed up in search results. The actual change is more structural than that.

The shortlist was formed, but not full

The 6sense 2025 B2B Buyer Experience Report referenced above studied nearly 4,000 B2B purchase decisions and found that 95% of winning vendors were on the buyer’s Day One shortlist: the initial consideration set formed before any sales engagement. On average, 3.4 of roughly 5 shortlist spots were filled on Day One through existing brand awareness and prior vendor relationships. 

These numbers predate the AI shift, and that's precisely the point: the shortlist was always the gate. AI didn't replace it. AI captured the remaining one to two seats. In competitive deals, that margin is often where outcomes are decided.

In the old model, the shortlist formed through a combination of brand awareness, peer recommendations, analyst reports, and organic search. The buyer controlled the process, drew on their professional network and their own research instincts. In the emerging model, an AI assistant synthesizes information from across the web, weighs sources the buyer will never see, and produces a ranked recommendation in seconds. The buyer doesn’t abandon their judgment, but they start from a position that’s already been shaped by a model’s interpretation of the market.

The defensibility mechanism

There’s a behavioral dynamic here that the adoption data doesn’t capture. When a buyer recommends a vendor to their internal stakeholders, they’re staking their professional credibility on that recommendation. AI-generated shortlists offer something that personal intuition doesn’t: air cover. “I ran this through our AI tools and these are the top three” is a safer sentence in a buying committee than “I think we should look at these three based on my experience.” The first frames the recommendation as data-driven. The second frames it as opinion. In a risk-averse enterprise buying culture, the AI recommendation becomes a shield against second-guessing. It doesn’t replace judgment, but rather insures it.

This means the AI’s influence is stickier than a simple recommendation engine would suggest. Once a buyer has used an AI-generated shortlist as the basis for an internal proposal, the cost of deviating from that list increases. Adding a vendor the AI didn’t surface requires the buyer to explain why they know better than the tool. Removing one the AI recommended requires explaining what they know that the model doesn’t. The path of least organizational resistance is to work from the AI’s list and refine from there.

Concentration, not chaos

SparkToro and Gumshoe’s January 2026 study (600 volunteers, 12 prompts, nearly 3,000 runs) found something that seems contradictory but isn’t. Full recommendation lists varied significantly between runs. Ask the same question twice and you’ll get different lists. But top brands in narrow categories appeared 70% to 90% of the time. The list is volatile. The consideration set is remarkably stable.

For B2B brands, the implication is direct (and likely understated: SparkToro’s study tested consumer categories with broad competitive sets, while narrow B2B software categories have even more constrained fields). In categories where the competitive set is small and well-defined (mid-market CRMs, cloud security platforms, performance management tools) the AI’s recommendations aren’t random. They’re concentrated. The same four or five vendors surface repeatedly. 

If you’re in that set, you have a structural advantage that compounds with every AI-assisted research session. If you’re not, you’re fighting for the attention of a buyer who has already been told, by a tool they trust, that your competitors are the answer.

B. The Measurement Trap: Visibility Dashboards Show Loss, Not the Fix

Give the GEO tools market its due: it solved the first problem. A year ago, most B2B companies had no idea whether they appeared in AI responses, let alone how often or in what context. That’s no longer true. An ecosystem of visibility platforms now exists, and it’s growing fast. Conductor’s January 2026 survey found 12% of enterprise digital budgets already allocated to GEO, with 94% of CMOs planning to increase spending. Separate data from GrowthUnhinged confirms the prioritization shift: 51% of marketers plan to increase AI search investment versus only 14% for traditional SEO.

The tools made the invisible visible. Brands can now see which AI platforms mention them, how often, for which queries, and in what competitive context. That’s genuinely valuable and was genuinely impossible eighteen months ago.

But knowing where you’re losing is not the same as knowing how to win.

The actionability gap

Every GEO dashboard on the market can tell you that you’re underrepresented on comparison queries in your category. The recommendations that follow are almost universally category-level advice: create more structured content, build topical authority, optimize for E-E-A-T signals, improve your technical SEO. This is not wrong. It’s also not strategy. It’s the marketing equivalent of telling someone who’s overweight to “eat better and exercise more”: true, unhelpful, and available to everyone including your competitors.

The distance between “you’re invisible on comparison queries” and “here is your specific content roadmap for these twelve high-value queries, prioritized by platform and buyer journey stage, with a competitive gap analysis that tells you exactly where to invest” is enormous. The first is a dashboard readout. The second is a strategy. The tools deliver the first. The second requires query-level specificity, competitive intelligence, platform-specific analysis, and a methodology for connecting visibility gaps to content gaps to business outcomes.

eMarketer captured this precisely when they identified scaling AI-optimized content as the top challenge marketers report. The challenge isn’t awareness of the problem. The challenge is that the problem, once identified, demands a kind of specificity that the tools themselves can’t provide.

Dashboards in a volatile landscape

The measurement problem is compounded by the nature of what’s being measured. AI responses are not like search rankings. They don’t hold still. Authoritas measured seventy percent turnover in AI Overview rankings within two to three months. Ahrefs found that between consecutive AI Overview responses, only 54.5% of cited URLs overlap: meaning the same query asked twice surfaced different cited URLs nearly half the time. Profound found 40% to 60% monthly domain drift.

A dashboard that shows your AI visibility at a single point in time is showing you a frame from a movie. The frame is real, but it’s not the movie. Any measurement approach that treats AI visibility as a static position to be tracked (the way SEO teams track keyword rankings) is fundamentally mismatched to the phenomenon it’s measuring. What matters is aggregate visibility across many queries and many runs, measured over time. That requires a different measurement philosophy, not just a different tool.

The market has over-invested in observation and under-invested in interpretation. Knowing you’re losing is the easy part. Understanding why you’re losing on specific queries across specific platforms (and translating that into a content and technical roadmap) is where the actual work begins. That work can’t be automated with a dashboard, and it can’t be generalized across categories. It requires the kind of query-level analytical depth that most teams aren’t staffed or structured to produce.

C. The Scoreboard Shift: Why SEO Mental Models Break in AI Surfaces

If Section A established the stakes and Section B identified the trap, this section presents the evidence. The case is straightforward: the signals that predict visibility in traditional search have almost no relationship to the signals that predict citation in AI responses. This isn’t a subtle difference. The correlations are near zero.

The prosecution’s case

Profound analyzed over fifty thousand prompts across five industries and measured how traditional SEO metrics correlated with AI citation. The results were definitive. Organic traffic explained 5% of citation behavior (r² = 0.05). Backlinks explained less than four percent (r² = 0.038). These are not weak correlations: they are effectively no correlation. A brand’s search ranking tells you almost nothing about whether an AI model will cite it.

The data gets more specific from there:

  • Only 40% of AI citations come from pages in Google’s top ten results. Semrush found that 90% of ChatGPT’s citations come from pages ranking position 21 or lower in Google.

  • 25% to 28% of top-cited pages have zero organic visibility: they don’t rank for anything in traditional search. AI models discover and cite content that SEO would consider invisible.

  • Citation distribution is flat. In Google, the difference between position one and position ten is a cliff in click-through rate. In AI responses, positions five through ten are all viable citation sources. The steep rank-CTR curve that defines SEO economics simply doesn’t exist.

  • Source divergence is extreme. Ahrefs found citation overlap between AI assistants and Google’s top ten averages just 11%. 89% of citation opportunities are platform-specific. A page that ChatGPT cites is unlikely to be cited by Perplexity for the same query, and vice versa.

Taken together, these findings do not support a nuanced “SEO and GEO are related” narrative. They support a stark one: the skills transfer, but the scoreboard is entirely different. The analytical thinking, the content strategy discipline, the technical competence that good SEO teams possess are all relevant to GEO. The specific signals those teams have spent years optimizing (keyword rankings, backlink profiles, organic traffic volume) are nearly meaningless as predictors of AI citation.

What does predict citation?

Seer Interactive’s analysis offers the clearest picture of what AI models actually respond to. Ahrefs confirmed that brand web mentions (r = 0.664) and brand anchor text (r = 0.527) far outperform backlinks (r = 0.218) and organic traffic (r = 0.274) as predictors of AI citation. The Princeton GEO study confirmed that content enriched with statistics, citations, and quotations boosts AI visibility by 30% to 40%. Domain authority matters at the domain level (the median Domain Rating for ChatGPT-cited sources is 90) but page-level authority is nearly irrelevant, with a median URL Rating of just 6.

The pattern is consistent: what matters is not whether your page ranks well, but whether your brand is well-represented across the web in the kinds of content AI models trust. Brand mentions in authoritative contexts. Specific, well-structured information. Freshness. Technical accessibility. These are the signals. Rankings are not.

The exception that proves the rule

There is one critical exception, and getting it right is what separates credible analysis from oversimplification.

Google’s AI Overviews do not work like other AI surfaces. Ahrefs analyzed 1.9 million AIO citations in mid-2025 and found that 76% come from pages already ranking in Google’s top ten, with a median organic search position of two. AI Overviews use retrieval-augmented generation from Google’s own search index. For this surface (and this surface only) traditional SEO is still the foundation for AI visibility.

Ahrefs’ March 2026 follow-up study with improved methodology found this figure has dropped to roughly 38%, suggesting AI Overviews are increasingly pulling from sources outside the top ten. The directional trend reinforces our argument: even Google’s own AI surface is relying less on the traditional SERP over time.

This matters because AI Overviews are often the first AI surface that SEO teams encounter, and it can create a dangerous false confidence. A team that optimizes for AI Overviews by doing better SEO will see results in that specific surface. They will then assume the same approach works for ChatGPT, Perplexity, Claude, and Gemini outside of Google Search. It doesn’t. The evidence is unambiguous: for every standalone AI platform, the old scoreboard is effectively irrelevant.

The practical implication is that GEO strategy must be platform-aware from the start. AI Overviews demand strong traditional SEO. Standalone chatbots demand something fundamentally different: brand authority, content specificity, technical accessibility, and a platform-by-platform approach to visibility. Treating “AI” as a single channel is as misguided as treating “social media” as a single channel was in 2010. The sooner that mental model breaks, the sooner effective strategy can begin.

D. What Has Changed Since v3.0

Version 3.0 of this white paper was published in August 2025. Since then, the evidence base for generative engine optimization has gone from suggestive to substantial. Not complete (the open questions in Section I are real), but the difference between mid-2025 and early 2026 is the difference between pattern recognition and confirmed findings. What follows are the six most consequential shifts, each sourced, each with its methodology noted, and each with a confidence level stated plainly.

The traffic inflection is no longer theoretical

In the summer of 2025, the most commonly cited figure for AI search traffic was “165x faster growth than organic.” That number was real, but it was a growth rate off a small base: the kind of statistic that sounds significant and means almost nothing operationally. By January 2026, the base caught up with the rate.

Akamai’s AI Pulse report, published in February 2026, measured 1.6 billion daily AI bot requests across its global CDN. Not estimates. Not survey responses. Actual request counts from one of the world’s largest content delivery networks. That number had plateaued around 1.1 to 1.2 billion through the fall of 2025 before jumping 33% in a single month: December to January. Over six months, the increase was 78%. This is an inflection point, not linear growth.

The distinction matters because these are bot requests, not human pageviews. They represent the raw material that feeds AI answer engines: the crawls that determine what models know, what they can cite, and what they can’t. Akamai’s earlier analysis had already flagged that AI and LLM bot management was becoming a business-critical issue for web operators. The January 2026 data proved it. Some publishers report AI crawlers visiting their sites more frequently than Googlebot, and unlike Googlebot, the AI crawlers don’t send traffic back.

This is Tier 1 data: first-party measurement from a neutral infrastructure provider with no competing GEO product to sell. When CMOs ask whether AI search is real enough to invest in, this is the number that answers it.

Measurement methodology matured, and exposed the old approach as flawed

Section B established that most GEO dashboards show frames, not movies. But the field didn’t just identify the problem: it began solving it. Three independent studies published between late 2025 and early 2026 converged on the same conclusion: single-response AI rankings are unreliable, but aggregate visibility across many queries and many runs is statistically meaningful.

SparkToro and Gumshoe’s January 2026 study showed that full recommendation lists vary wildly between runs. The probability of getting the exact same list twice is less than 1%. But top brands in narrow categories appeared 70% to 90% of the time. AirOps, analyzing over 45,000 citations, measured 30% consecutive persistence at the URL level: meaning 70% of cited links churned out of the response even when the model and prompt remained identical. And Ahrefs’ AI Overview change-rate research added the critical nuance: while the text of AI Overview responses changes 70% of the time and cited URLs change 46% of the time, cosine similarity between consecutive responses is 0.95. The underlying recommendation (the actual verdict) is remarkably stable even as the surface words and specific source links shuffle.

The implication for measurement is clear: any tool that shows you a single AI response and calls it your “ranking” is selling you noise. Valid measurement requires multi-run sampling across many queries, tracked over time, with aggregate visibility as the primary metric. The individual snapshot is meaningless. The pattern across hundreds of snapshots is not.

Citation mechanics went from vague to granular

Last summer, the best available guidance on what earned AI citations was “off-page signals matter more than on-page” and “domain authority correlates with visibility.” By early 2026, the picture is substantially more specific. Three findings in particular have reshaped how we think about what makes content citable.

First, structured content. Multiple analyses suggest that well-structured pages (with clear headings, defined data points, and organized information architecture) earn citations at meaningfully higher rates than unstructured prose. The Princeton GEO study first suggested that content enriched with statistics, citations, and quotations boosts AI visibility by 30% to 40%. The specific multiplier that circulates in practitioner communities is less well-documented, so we phrase this carefully: structured content is a demonstrated advantage. The exact magnitude depends on the competitive set and the platform.

Second, freshness. The “48-hour refresh” claim that circulated in mid-2025 was always poorly sourced, and Ahrefs’ rigorous analysis of 1.9 million AI Overview citations put a number on what actually matters: 76.4% of cited pages had been updated within 30 days, and 89.7% within the current year. Content from the past two to three months dominates citations. The defensible recommendation is a monthly refresh cadence. Not daily panic, not quarterly neglect.

Third, the distinction between being mentioned and being cited. Emerging evidence suggests that brands which are both mentioned in the response text and cited as a source are roughly forty percent more likely to resurface across repeated runs than brands that are only cited as a source link. This makes intuitive sense (a model that weaves your brand into its narrative answer has a stronger association than one that merely lists you in a footnote) but the finding hasn’t been independently replicated at scale. Treat it as a strong working hypothesis, not a confirmed law.

Taken together, these mechanics suggest a hierarchy. Seer Interactive’s analysis confirmed that brand web mentions and brand anchor text far outperform backlinks and organic traffic as predictors of AI citation, and Ahrefs’ brand visibility research validated the same pattern. The evidence supports a three-tier model: technical accessibility is a binary gate (if AI crawlers can’t see your content, nothing else matters). On-page quality (structure, freshness, intent match) is table stakes that won’t differentiate you from competitors who also have it, but its absence is penalized. Off-page signals (brand mentions, digital PR, third-party citations) are what actually differentiate between sites that are all technically sound.

The source composition picture came into focus

Ahrefs’ analysis of ChatGPT’s top 1,000 citations revealed that 67.7% are structurally uninfluenceable: Wikipedia at 29.7%, brand homepages at 23.8%, and app stores at 6.6%. These citations exist because the model draws on stable, high-authority reference material that no amount of content optimization will displace. The remaining roughly one-third is where GEO actually operates: the citations that are genuinely up for grabs based on content quality, brand authority, and strategic positioning.

Separately, Profound’s analysis of 27 million real answer engine prompts found that 97% of earned media citations come from non-Tier-1 sources: major outlets like AP, Bloomberg, and Forbes account for only 2.6% of earned media citations. (Profound’s data comes from consumer panel data; for B2B, Gartner, Forrester, and industry-specific publications carry more weight. The implication for B2B isn’t “skip Tier-1 press,” it’s “don’t only pursue Tier-1 press.” A Forrester Wave mention and three niche industry placements are probably worth more together than a generic Forbes feature.)

Manipulation vulnerabilities were proven, not just theorized

This is entirely new as of December 2025, and it represents something the field hadn’t grappled with: the question of whether AI models can be deliberately misled about brands.

Ahrefs’ misinformation experiment tested this directly. When an official brand FAQ stated “we don’t publish unit counts,” 5 of 8 major AI models chose fabricated concrete numbers from unofficial sources instead. The models preferred specific fiction over vague truth. ChatGPT-4 and ChatGPT-5 Thinking were resistant (93% to 96% accuracy) but Perplexity and Grok were heavily manipulated.

The most dangerous finding was the “partial debunk” attack vector. A Medium article that first corrected obvious misinformation earned the model’s trust, then successfully planted entirely new fabrications in the same post. The model treated the source as credible because it had demonstrated accuracy on one point, and then accepted its claims on a separate point without verification.

The practical implication is blunt: if you don’t fill your information gaps with specific, official content, someone else will fill them with fiction, and the models will prefer the fiction because it’s more specific. This creates an entirely new category of brand risk, and an entirely new service category in response. Section F explores the defensive implications in depth.

Search Engine Journal’s methodological critique is fair (the experiment used leading questions and the “official” site lacked typical brand authority signals), which means the real-world vulnerability may be lower for well-established brands. It doesn’t eliminate the risk.

The technical foundation became non-negotiable

In mid-2025, GEO advice was almost entirely about content strategy: what to write, how to structure it, where to publish. By late 2025, a set of technical prerequisites emerged that now sit upstream of everything else.

The most consequential: most LLM crawlers do not reliably render client-side JavaScript. Sites built on React, Vue, Angular, or any client-side rendering framework typically serve empty or near-empty pages to AI bots. If your site relies on JavaScript rendering and hasn’t implemented server-side rendering or pre-rendering for bot traffic, assume the crawl sees little or nothing. In practice, for most stacks, this behaves like a binary gate: either your content is accessible to AI crawlers, or it isn’t.

Compounding this, Cloudflare declared what it called “Content Independence Day” on July 1, 2025, changing its default to block AI crawlers unless they pay. Cloudflare’s network handles roughly 24% of all web traffic, and the default change made a significant share of the web effectively dark to AI crawlers overnight. Sites that didn’t explicitly opt in to allowing AI crawler access became invisible. 

Cloudflare’s own data illustrated the asymmetry driving this decision: OpenAI’s crawl-to-referral ratio was 1,700 to 1: for every 1,700 pages crawled, one referral visit was sent back. Anthropic’s ratio was 73,000 to 1. For comparison, Google’s ratio is roughly 14 to 1. Publishers were being crawled extensively and receiving almost nothing in return.

Additional technical findings have reinforced the non-negotiable status of this layer. Over thirteen percent of AI bots are ignoring robots.txt directives entirely. Core Web Vitals (specifically Largest Contentful Paint and Cumulative Layout Shift, not aggregate scores) show correlation with AI Mode citation rates. And the overall pattern is clear: a three-tier hierarchy where technical accessibility is the foundation, on-page content quality is the middle tier, and off-page brand signals are the differentiating top tier. Without the foundation, the upper tiers don’t matter.

Platform divergence deepened from statistical to behavioral

By late 2025, the platform divergence documented in Section C extends beyond source preferences to response behavior itself. Perplexity's responses are getting 2.7 times longer than ChatGPT's, with citation density jumping from an average of 5 to 9 sources per response. ChatGPT's responses are slightly contracting in length. The platforms are moving in opposite directions on the most basic behavioral dimensions.

Source preferences have gotten more specific. ChatGPT favors business listings at 48.7% of its citation mix; Gemini favors first-party websites at 52.1%. Manipulation resilience varies wildly: ChatGPT resists misinformation at 93% or higher, while Perplexity is heavily vulnerable. Different models now produce responses of different lengths, with different citation densities, from different source pools, with different susceptibility to external influence.

The operational consequence is that measurement, content strategy, and resource allocation all need to be platform-specific. Section G describes what that looks like in practice.

E. What Has Held: High-Confidence Findings

Section D covered what’s new. This section covers what didn’t change, and in a field this young, that’s the more important story. Anybody can identify a trend. The question is whether it holds. Several core findings from v3.0 of this document have now been confirmed and reconfirmed across multiple studies, timeframes, and methodologies. These are the highest-confidence foundations in GEO. If you’re building a program, build on these.

Three foundations confirmed

The three structural findings from Sections C and D have held without exception across nine months of additional measurement. The independence of AI citation from SEO signals has been retested by Profound, Semrush, and independent audits with no material change. The one exception remains Google’s AI Overviews, which directionally reinforces the decoupling even on Google’s own surface.

Platform divergence has been re-measured repeatedly and hasn’t materially changed. If anything, the platforms are getting more different, not less. And citation volatility remains structural: the words change between runs, but the underlying verdict holds.

AI-referred traffic converts at a premium for B2B

The conversion data has been remarkably consistent across independent studies: 4.4x (Semrush), 5x (Superprompt), 9x (Seer Interactive), 23x (Ahrefs) for B2B. The mechanism is clear: AI users arrive pre-qualified, having already evaluated alternatives and narrowed their consideration set before clicking through. They’re further down the funnel when they land.

The range is wide, and that matters. The highest estimate is more than five times the lowest. Not all of these studies have published whether their samples are purely B2B or mixed-category. The responsible framing is that AI-referred traffic converts at a meaningful premium for B2B (likely in the four to ten times range) but the precise multiplier depends on category, funnel stage, and attribution methodology.

One important caveat: the conversion premium is not universal. Ecommerce data shows AI traffic converting below Google organic on both conversion rate and revenue per session in some studies. The pattern appears to be category-dependent. For B2B, where the buyer arrives with research intent and high commercial motivation, the premium is real and persistent. For transactional ecommerce, it may not hold. Cite the B2B range, not a single number.

Domain authority matters; page authority doesn’t

Multiple studies have converged on the same finding: the median Domain Rating for ChatGPT-cited sources is 90, with 65.3% coming from domains rated 81 or higher. The average domain age of cited sources is 17 years. High-authority, established domains have a structural advantage in AI citation.

But page-level authority (URL Rating in Ahrefs’ taxonomy) is nearly irrelevant. The median URL Rating for cited pages is just 6. And 11.7% of top-cited pages come from domains rated 0 to 20. 25% to 28% have zero organic visibility. Domain authority is a strong signal but not a gatekeeper. The exception rate is high enough that content quality, specificity, and freshness can override a weak domain, and the page-level metrics that SEO teams obsess over simply don’t register.

For B2B brands on established domains, this is good news: your domain authority is an asset in AI surfaces, even though your specific page rankings don’t transfer. For challenger brands on newer domains, it means the path runs through content quality and third-party brand mentions, not through trying to build page-level authority signals that the models don’t use.

These five findings (the independence of AI citation from SEO signals, the persistence of platform divergence, the structural nature of citation volatility, the B2B conversion premium, and the domain-over-page authority pattern) are not new. They were present in v3.0. What’s new is the confidence level. Nine months of additional data, from independent sources using different methodologies, measuring different time periods, have confirmed every one of them. The foundation is solid. What you build on it is the question the rest of this document addresses.

F. Defensive GEO: Misinformation and
Narrative Hijacking

Most of this document is about offensive GEO: getting your brand into AI responses, earning citations, building visibility where buyers are looking. This section is about the other side: what happens when someone else shapes what AI says about you, and what it takes to prevent that.

The Ahrefs misinformation experiment detailed in Section D revealed something that deserves its own vocabulary. When a brand’s official content doesn’t answer a question with sufficient specificity, that’s a narrative gap, and it has a consequence: someone will fill that gap for you.

The mechanism: specificity beats authority

Traditional brand reputation assumed authoritative sources won: the brand’s own website, analyst reports, major press. AI models don’t follow this hierarchy reliably.

What the Ahrefs experiment demonstrated is that AI models are drawn to specificity as a proxy for credibility. A source that provides concrete numbers, even fabricated ones, gets prioritized over a source that says “we don’t disclose.” 

A blog post that names dates, figures, and details gets weighted over a corporate page that speaks in generalities. This isn’t a flaw the models will fix next quarter. It’s a reflection of how large language models process information: specific claims are more “useful” to the response the model is trying to construct, so they get incorporated. Vague claims don’t give the model anything to work with.

The practical implication is that the bar for defensive content isn’t “have an answer somewhere on your site.” It’s “have an answer that’s more specific than anything else the model can find.” If your pricing page says “contact sales,” and a comparison blog says “Company X starts at $49/month,” the model will use the blog’s number: whether or not it’s accurate.

This isn't limited to fabricated content. A founder recently documented his experience shopping for a PEO provider. He used Gusto for payroll and asked Google's AI Mode whether Gusto could handle PEO services. The AI pulled from a Rippling FAQ page, told him Gusto doesn't offer PEO, and injected Rippling into the conversation at the exact moment he was evaluating options. Rippling didn't lie. They published a factually accurate FAQ that answered a question their competitor's site didn't address. The invisible committee member used it.

The most dangerous attack: the partial debunk

Section D introduced this vector. Here’s how it works. 

A third-party article (a Medium post, a comparison blog, a Reddit thread) first corrects a piece of obvious misinformation about a brand. This earns the model’s trust. The source has demonstrated accuracy. Then, in the same piece, the author introduces new claims: claims that may be misleading, outdated, or entirely fabricated. The model has already classified the source as credible based on the debunking, so it accepts the new claims without independent verification.

This is not theoretical. The Ahrefs experiment tested this exact pattern and documented it working across multiple models. It’s the informational equivalent of a social engineering attack: establish trust first, then exploit it. And it’s particularly effective against brands because the “obvious misinformation” that gets corrected is usually something the brand could have proactively addressed but didn’t.

Not all platforms are equally vulnerable

Section D documented the stark range: ChatGPT-4 and ChatGPT-5 Thinking demonstrated 93% to 96% accuracy, while Perplexity and Grok were heavily manipulated.

This maps directly to the platform divergence discussed in Sections D and E. The platforms don’t just cite different sources and produce different response lengths: they have fundamentally different relationships with truth. A brand might be well-protected on ChatGPT and completely misrepresented on Perplexity, using the same set of available sources. Defensive GEO, like offensive GEO, has to be platform-aware.

The Reddit paradox: marginal citation source, material narrative risk

One of the more nuanced findings in our research is the disconnect between Reddit's role as a citation source and its role as a narrative shaper. For B2B enterprise queries, forums including Reddit account for roughly 2% of formal citations. Vendor sites, G2, Capterra, analyst content, and industry publications dominate the citation landscape for B2B buying queries. Reddit is not where your brand gets cited.

But Reddit is where narratives form. This operates through two channels. First, Reddit has signed data licensing agreements worth over $200 million annually with Google, OpenAI, and other AI companies, giving those platforms direct access to Reddit's content for model training. This means sentiment and claims from Reddit threads are baked into the model's baseline understanding of a brand, independent of whether Reddit appears as a cited source at query time. 

Second, AI models treat Reddit content as credible user testimony: firsthand experience from real practitioners. A disgruntled employee post, a misleading comparison thread, or a fabricated experience report can shape the narrative that models synthesize about a brand even when Reddit itself isn't formally cited as a source. The Ahrefs experiment confirmed the second channel directly: a fake Reddit AMA was treated as credible by multiple models.

For B2B brands, the operational implication is clear: monitor Reddit for narrative risk in category-relevant subreddits, but don't invest in Reddit content strategy at the expense of vendor site optimization, review platform presence, and earned media in industry publications. The citation value isn't there for B2B. The narrative risk is.

The defensive loop: back to the invisible committee member

Section A introduced the concept of AI as an invisible buying committee member: one that has already shaped the shortlist before the first sales call. The defensibility mechanism we described there works in both directions.

In the offensive direction, brands that appear consistently in AI responses benefit from a self-reinforcing cycle: buyers see the brand in AI recommendations, include it on their shortlist, and the brand’s presence in the market generates more signals that feed back into the model’s next response. Winning breeds winning.

In the defensive direction, misinformation follows the same loop. If a fabricated claim about your brand enters an AI model’s response, buyers encounter it during their research phase. Some of those buyers may repeat the claim in their own content (an internal Slack message, a LinkedIn post, a podcast discussion) which creates new training signals that reinforce the original fabrication. Misinformation compounds.

This is why defensive GEO isn’t optional for brands with significant AI visibility. The same mechanism that makes AI recommendations valuable (their role as career air cover for buying decisions) makes misinformation in those recommendations dangerous. If the invisible committee member is saying something wrong about your brand, that wrong thing is shaping pipeline before you even know it’s happening.

What defensive GEO requires

Defensive GEO is emerging as a distinct discipline within the broader GEO category. It’s not the same as reputation management, although it borrows from that field. And it’s not the same as offensive GEO, although the two share tooling. The core activities fall into three areas.

  1. Narrative gap identification. Every question a buyer naturally asks about your brand (pricing, implementation timelines, customer count, competitive differences, limitations) needs a specific, official answer that’s crawlable by AI models. The Ahrefs experiment proved that “we don’t publish that” is worse than not having a page at all, because it confirms the gap exists and signals to the model that it should look elsewhere. The audit process starts by mapping every buyer question to an existing content asset, then identifying where the answers are either missing, vague, or less specific than what third parties have published.

  2. Narrative monitoring. Brands need systematic, multi-platform tracking of what AI models actually say about them across different query types. Not just “are we cited?” but “what claims do the models make about us, and are those claims accurate?” This requires running prompts regularly across multiple platforms and analyzing the narrative content of responses, not just the citation lists. The methodology for doing this effectively is still maturing, but the need is clear and immediate.

  3. Rapid response content. When monitoring surfaces a factual error or misleading claim in an AI response, the fix isn’t to write an angry email to OpenAI. The fix is to publish content that’s more specific, more authoritative, and more crawlable than whatever source the model is currently using. This is where the freshness signal from Section D becomes directly relevant: 76.4 percent of cited pages were updated within thirty days. Models favor recent content. A brand that publishes a detailed, specific correction on a high-authority domain with strong crawl access has a reasonable chance of displacing the misinformation within a refresh cycle.

The brands that will fare best are not the ones that ignore it: they’re the ones that build the monitoring, gap analysis, and rapid response capabilities before they need them. The cost of prevention is a fraction of the cost of remediation.

G. What a Serious Program Looks Like

Sections A through F made the case. AI shortlists are forming before sales enters the picture. Dashboards show where you’re losing but not what to build. The old SEO scoreboard doesn’t apply. The evidence base is deeper than it’s ever been, and so are the risks. The natural next question is: what does a serious GEO program actually look like?

This section describes the model, not the playbook. It lays out the categories of work, the measurement architecture, and the strategic priorities that separate a real program from a dashboard subscription with a content calendar attached. It does not walk through step-by-step execution. That’s what engagements are for. What it does is give a VP of Marketing or a Head of Content enough structural clarity to evaluate whether their current approach is serious, or whether they’re running a checklist they found in a blog post.

What to measure (and what to stop measuring)

The first sign of a serious GEO program is what it measures. Or more precisely, what it stops treating as a proxy for AI visibility. Section C made the case that the metrics SEO teams have spent years optimizing have near-zero predictive value for AI citation. Continuing to measure them as a proxy for AI presence is measuring the wrong scoreboard.

A serious program tracks four things:

  • Presence: Is your brand appearing in AI responses for the queries that matter? Not all queries: the specific queries your buyers use at each stage of their journey. The metric that matters is aggregate presence across many queries and runs, not whether you appeared in any single response.

  • Citation share: Of the responses where your category is discussed, how often are you cited relative to competitors? As Section E established, the verdict is stable even as surface-level citations shuffle. Citation share, measured in aggregate, tells you whether you’re in the model’s settled view of your category.

  • Narrative accuracy: What do the models actually say about you, and is it correct? Section F made the case that this matters as much as presence. A brand that appears consistently but is described inaccurately (with outdated pricing, fabricated limitations, or competitor-planted framing) has a visibility problem dressed up as a visibility win. Narrative monitoring means running prompts regularly across multiple platforms and analyzing what the responses claim, not just whether your name shows up.

  • Platform-specific visibility: Not “are we visible in AI?” but “are we visible on ChatGPT, on Perplexity, on Google AI Overviews, on Claude?” The platform divergence documented throughout this paper means a single aggregate score hides more than it reveals. A brand might own eighty percent of relevant queries on ChatGPT and be completely absent on Perplexity. The measurement has to be platform-disaggregated, or it’s meaningless.

What to stop measuring, or at least stop using as a proxy for AI performance:

  • Page-level Google rankings (unless you’re specifically optimizing for
    AI Overviews)

  • Individual URL citation counts (too volatile to be actionable)

  • Any metric that treats “AI” as a single channel. The platforms are different. The measurement has to reflect that.

Platform divergence stays central

The platform divergence established throughout this paper should organize every GEO decision.

Section D documented the behavioral specifics: different response lengths, citation densities, source preferences, and manipulation resilience across platforms. The operational implication: even content structure and formatting decisions need to be platform-aware.

A serious program starts with platform prioritization: which platforms matter most for your buyers, and what does winning look like on each? For most B2B brands, the answer includes at least ChatGPT (83% of AI search traffic), Google AI Overviews (where SEO still matters, per Section E), and one or two additional platforms based on where your specific buyer persona is active. Platform-agnostic “optimize for AI” advice is increasingly useless.

From category-level advice to query-level specificity

This is the gap that separates programs from projects.

Section B described the actionability gap in the current GEO tools market: dashboards tell you where you’re invisible, then offer category-level recommendations: “add statistics,” “create FAQ pages,” “write listicles.” None of this is wrong. All of it is obvious. Emarketer’s January 2026 finding captures it precisely: AI visibility is the number-one goal for marketing leaders, but scaling AI-optimized content is their top challenge. The problem isn’t knowing what to do at the category level. It’s knowing what to build next, for whom, addressing what.

A serious program operates at query-level specificity. That means mapping specific buyer queries to specific personas and journey stages. It means understanding why competitors are winning particular queries (not just that they’re winning) by analyzing what the model is actually citing, what structure the cited content uses, and what claims it contains. It means producing content briefs specific enough to hand to a writer on Monday morning, not vague directives that require another round of interpretation.

The difference between “write a comparison page” and “write a comparison between your product and [specific competitor] for [specific persona], addressing [specific objections], structured for [specific platform’s citation patterns], targeting [specific stage of the buyer journey]” is the difference between a recommendation and a strategy. The first is available in any dashboard. The second requires research that most organizations are not yet equipped to do internally.

This is also where buyer journey mapping becomes operational rather than theoretical. A procurement manager asking “what are the best tools for X” is in a fundamentally different mental state than one asking “does [your company] integrate with Okta.” The content strategy for discovery queries, comparison queries, validation queries, and technical queries is different in structure, depth, and platform targeting. Programs that treat all AI queries as a single bucket produce content that’s adequate everywhere and compelling nowhere.

Evaluating what you’re reading

A serious GEO program evaluates its own evidence base. The best research in this field is commercially motivated: Conductor, Ahrefs, SparkToro, and others all sell tools adjacent to their studies. That’s not a reason to dismiss the research; convergence across independent, competing vendors makes findings more credible, not less. 

But look at sample size and methodology first, commercial interest second. A study of 50,000 prompts with documented methodology is more useful than a study of 50 prompts published by a company with no product to sell.

The technical foundation is non-negotiable

Section D covered these findings in detail: JavaScript-rendered sites are largely invisible to AI crawlers, and Cloudflare’s default AI crawler blocking means a significant share of the web is dark to models by infrastructure default. These aren’t optimization problems. They’re access problems.

A serious program starts with a technical audit: can AI crawlers actually reach and render your content? Do your robots.txt rules and CDN configurations allow crawl access? Is your site serving static HTML or relying on client-side rendering that the crawlers can’t process? In our experience, this is the single most common “quick win” in client engagements. Many enterprise sites run on JavaScript frameworks and don’t realize AI crawlers see little or none of their content. Fix this first, before any content strategy work.

Beyond crawl access, the technical foundation includes structured data implementation and content architecture designed for passage-level extraction. The freshness and structure signals documented in Section D aren’t nice-to-haves. They’re the mechanical prerequisites for citation.

Content, measurement, adjustment, repeat

A serious GEO program is not a project with a start and end date. It’s a loop.

The environment demands it. The citation volatility documented in Sections D and E is structural, not a temporary phase. A content strategy that doesn’t include continuous measurement and adjustment decays the month after you ship it.

The loop is straightforward to describe and difficult to execute: publish content built for specific queries, measure whether citation patterns shift, analyze what worked and what didn’t, adjust the strategy, repeat. 

The measurement cadence matters: weekly monitoring of high-priority queries, monthly analysis of aggregate trends, quarterly strategic reviews. But the harder part is the analytical layer between measurement and action. Knowing that your citation share dropped on comparison queries doesn’t tell you why it dropped or what to change. That diagnosis requires query-level analysis that most dashboards don’t provide.

This is also where defensive GEO integrates with offensive GEO. Section F described the narrative monitoring, gap identification, and rapid response capabilities that protect a brand from misinformation. In practice, those activities aren’t separate workstreams: they’re part of the same loop. You’re measuring what the models say about you, identifying gaps between what they say and what you want them to say, and publishing content to close those gaps. Whether the gap is “we’re not mentioned” or “we’re mentioned with incorrect pricing,” the response mechanism is the same.

H. Where It’s Heading

Everything in Sections A through G describes the current state: what’s happening now, what the evidence shows, and what serious programs are building. This section looks forward. Not with predictions, but with trajectories: forces already in motion, with observable momentum and directional clarity. The confidence levels are explicit.

The content source ecosystem is fracturing

The relationship between AI platforms and the publishers whose content they synthesize is breaking down. Google search referrals to publishers declined 33% globally in 2025 across more than 2,500 tracked sites. Zero-click searches rose from 56% to 69% between May 2024 and May 2025. Publishers are responding: 79% of top news sites now block at least one AI training bot, and Cloudflare flipped AI crawlers to blocked-by-default across roughly 20% of all websites in July 2025.

The economics driving this fracture (the crawl-to-referral asymmetry detailed in Section D) mean AI platforms consume content at massive scale and return almost nothing in traffic. The old social contract is broken, and publishers know it.

What this means for GEO: if publishers increasingly block AI crawlers or gate content behind licensing deals, the citation source pool changes. LLMs will rely more heavily on what remains accessible: owned brand content, open-access publications, review platforms, structured data. This could actually increase the value of GEO for brands that optimize their own properties. It also means today’s citation landscape data may not be stable. The source hierarchy documented in Sections D and E is a snapshot of the current access regime, not a permanent fixture.

AI advertising is arriving inside responses

OpenAI began testing ads inside ChatGPT in early 2026, initially limited to US users on Free and Go plans. Google is already placing ads in AI Overviews. OpenAI’s internal documents project $1 billion in free-user monetization in 2026, growing to $25 billion by 2029.

A paid layer is forming on top of organic AI visibility. This is the SEO-to-SEM split all over again, playing out on a compressed timeline. Organic GEO and paid GEO will become separate disciplines, separate budget lines, and separate optimization targets: just as organic search and paid search did two decades ago. For brands investing in GEO now, the implication is that the organic channel is still open and relatively uncluttered. That window has a shelf life.

AI Overviews are compressing organic visibility

The click-loss data has converged across independent studies. Ahrefs measured a 58% CTR reduction for position one across 300,000 keywords between December 2023 and December 2025. Seer Interactive found drops of 49.4% to 65.2% across 3,119 queries. Authoritas measured 47.5%. Pew Research found that users who encounter an AI summary click on a traditional result in just 8% of visits, compared to 15% without one.

But the picture is more nuanced than the headline numbers suggest. Seer’s data also shows that brands cited in AI Overviews earn 35% more organic clicks and 91% more paid clicks compared to brands that aren’t cited. Being in the AI response is becoming a prerequisite for being clicked at all. The old funnel where clicks came from rankings is collapsing into a new one where clicks come from citation.

These CTR studies are measured primarily on informational queries (AI Overviews trigger on 99.9% of informational keywords but only 5.5% of commercial and 1.2% of transactional queries), so B2B comparison-stage queries may be less exposed to AIO click loss (for now). The trajectory is toward broader coverage across query types.

The generational shift toward AI-first discovery

86% of Gen Z professionals now use AI daily at work. Gen Z trust in AI helpfulness jumped from 37% to 55% in a single year. 56% of tech buyers already rely on chatbots as a top source for vendor discovery. These are not projections: they are current behavioral data from people who are already in B2B buying roles and moving into more senior ones.

The tension: 6sense’s 2025 study of nearly 4,000 B2B buyers still shows LLM use peaking mid-journey, with 95% of winning vendors already on the buyer’s Day One shortlist. This is a constraint on how much AI currently reshapes discovery: most buyers still arrive with existing vendor awareness. But three mechanisms could change it: Gen Z professionals moving into senior purchasing roles, funnel compression as AI handles more of the early research, and agentic purchasing systems that don’t carry institutional memory at all.

The current position (that GEO’s primary value is mid-funnel influence and consideration-set reinforcement) is accurate for today. It should be understood as time-bound. If the shortlist-formation data starts shifting, discovery-stage GEO becomes the entire game, not just a supporting factor.

Agentic commerce changes who builds the shortlist

This is the trajectory with the biggest potential to reshape everything this document has discussed.

Gartner predicts that 90% of B2B buying will be AI-agent-intermediated by 2028, pushing over $15 trillion through agent exchanges. Those specific numbers may prove aggressive, but the direction is reinforced by nearer-term data: Forrester predicts that 20% of B2B sellers will face agent-led quote negotiations by the end of 2026. McKinsey’s 2025 survey shows 62% of organizations experimenting with AI agents, though only twenty-three percent have begun scaling.

When an AI agent builds the vendor shortlist, the “95% prior vendor experience” advantage that 6sense measured erodes: because agents don’t have institutional memory. They don’t remember which vendor the VP liked three years ago. They evaluate whatever they can access: structured data, API documentation, product feeds, published specifications. AI visibility becomes a literal prerequisite for consideration, not just an influence on it.

For GEO, the agentic shift means machine-readability, structured data, and API-accessible product information become as important as narrative content. The GEO program that Section G described (focused on AI responses to human queries) may need to expand into a broader discipline that includes agent-accessible data architecture. The playbook is still being written, but the direction is unmistakable.

The meta-implication

Every trajectory above points in the same direction: GEO’s value is increasing, not static. The content source ecosystem is fracturing in ways that favor optimized brand properties. Paid AI placement is arriving, which will raise the value of organic presence. AI Overviews are compressing organic clicks, making citation a prerequisite for visibility. A generational behavioral shift is making AI the default research tool. And agentic commerce is turning AI visibility into a literal gating function for vendor consideration.

Before paid layers arrive at scale, before source restrictions reshape the citation landscape, before agentic purchasing makes AI visibility a prerequisite rather than an advantage, the cost of entry is still low and the competitive landscape is still forming. Every trend above makes GEO harder and more expensive to do later. That’s not a sales pitch. It’s the structural logic of a market where first movers accumulate self-reinforcing advantages: the same defensibility mechanism Section A described, applied to the timeline of the field itself.

I. Open Questions the Field Hasn’t Answered

Sections A through H present what we know: with evidence, confidence levels, and sourcing. This section presents what nobody knows yet. These are not minor edge cases. They are structural gaps in the field’s understanding, and any organization claiming to have a complete GEO strategy is working around them, not through them.

We include this section because intellectual honesty is more useful than false certainty. Knowing where the evidence runs out is as important as knowing what it shows.

No rigorous longitudinal GEO intervention study has been published

This is the most significant gap in the entire field. Nobody has published a controlled study demonstrating: “We implemented these specific GEO interventions for this brand, and their AI visibility changed by this amount over this timeframe, controlling for other variables.”

What exists is cross-sectional: snapshots showing what kinds of content get cited, which platforms behave differently, how volatile citations are on any given day. What’s missing is the before-and-after. Did structured content improvements actually cause citation gains? Did technical access fixes lead to measurable visibility increases? Did query-specific content creation move the needle for a real brand in a real competitive landscape?

Without this data, GEO recommendations remain correlational. We know that structured content appears in citations at higher rates than unstructured content. We don’t have published evidence that restructuring existing content produces the same effect. The distinction matters, and it’s the kind of evidence that the field needs to move from plausible to proven.

AI visibility’s impact on pipeline and revenue is unproven externally

The conversion rate data is compelling. Multiple studies show AI referral traffic converting at 4 to 23 times the rate of traditional organic search, depending on the vertical. But conversion rate is not attribution. No independent study has documented the full causal chain: GEO investment → AI visibility improvement → pipeline growth → revenue increase.

The attribution problem is compounded by the measurement environment itself. Ahrefs demonstrated in January 2026 that clicks from Google AI Overviews appear in analytics as standard google/organic traffic or with no referrer at all. Google provides no native way to isolate AIO clicks in Search Console or GA4. The dark traffic problem from Section A isn’t just a conceptual framework, it’s an active measurement obstruction.

This means that even organizations seeing pipeline growth correlated with GEO investment can’t cleanly attribute it. The traffic looks like organic search. The referral source disappears. Building the analytical bridge between AI presence and business outcomes is one of the hardest problems in the field, and no one has published a rigorous solution.

Cross-platform citation mechanics beyond ChatGPT are thin

The vast majority of rigorous citation analysis has focused on ChatGPT, for understandable reasons: it commands 83% of AI search traffic. But Perplexity, Gemini, and Claude have much thinner data. Ahrefs’ platform overlap research tells us the platforms differ dramatically, but we don’t have platform-specific citation mechanics at the same granularity for any platform besides ChatGPT.

What does Perplexity’s source selection actually optimize for? How does Gemini’s citation behavior differ between its standalone product and Google AI Overviews? How does Claude weight training data against real-time retrieval? These aren’t theoretical questions. They’re operational ones. Any brand allocating resources across platforms is making bets without platform-specific evidence to inform the allocation.

How training data and real-time retrieval interact is poorly understood

Every major AI platform uses some combination of pre-trained knowledge (baked into the model during training) and real-time retrieval (pulled from the web at query time). But the weighting between these two sources (and how that weighting varies by model, query type, and topic) is not well documented.

This creates a practical ambiguity at the center of GEO strategy. If a brand has strong representation in training data (through years of authoritative content), does that persist even if the brand’s real-time content declines? Conversely, can a new entrant with excellent real-time content overcome a lack of training data presence? The answer almost certainly varies by platform and query type, but nobody has isolated these variables in a published study.

The implications are significant. If training data dominance is durable, incumbents have a structural advantage that real-time optimization can’t easily overcome. If real-time retrieval dominates, the field is more dynamic than it appears and the investment case for ongoing GEO programs is stronger. Both scenarios are plausible. Neither is proven.

The business listings question is unresolved

Yext’s analysis of 6.8 million AI citations found that 42% come from business listings: an enormous share if the finding generalizes. But the study was conducted by Yext, which sells business listings management. Their methodology is location-aware and intent-aware, which adds rigor. But the finding is heavily influenced by the four industries they tested: retail, financial services, healthcare, and food service, all of which have strong local and listings components.

No independent source has replicated this finding. For B2B technology companies (which don’t have storefronts, don’t appear on MapQuest, and aren’t listed on Vitals) the listings share is almost certainly much lower. But we don’t know by how much. And the broader question of how much citation behavior varies by industry vertical remains essentially unstudied outside of Yext’s consumer-facing sample.

International and non-English GEO dynamics are essentially unstudied

Virtually every finding cited in this document is English-language and US-centric. We have no substantive data on whether citation mechanics, platform preferences, volatility patterns, or source hierarchies hold for other languages or markets.

This is not a minor caveat. Regulatory environments differ: Europe’s AI Act imposes requirements that don’t exist in the US. Platform availability varies: some markets have limited access to ChatGPT or Perplexity. Language-specific training data differences likely create entirely different citation landscapes. Any organization operating globally is extrapolating from US English data to markets where the dynamics may be fundamentally different.

Agentic AI’s impact on citation mechanics is unknown

Section H described the trajectory toward agent-mediated commerce. But the specific question for GEO is narrower and more urgent: when an AI agent conducts research on behalf of a buyer, does it produce citations at all? Agent-mediated discovery may bypass the citation layer entirely, interacting directly with APIs, product feeds, and structured data instead of synthesizing web content into cited responses.

If that’s the case, the entire measurement framework this document describes (citation share, mention tracking, platform-specific visibility) may need to shift toward something closer to machine-readability scoring and data accessibility auditing. 

The current GEO paradigm assumes a human reading an AI-generated response with visible sources. Agentic commerce may produce a paradigm where the “audience” is another machine, and visibility means something entirely different.

Why this matters

These gaps don’t invalidate the evidence in the preceding sections. They contextualize it. The findings on platform divergence, citation volatility, structured content advantages, and technical prerequisites are all well-supported by the research available. But the research available has limits, and those limits should inform how confidently any organization acts on GEO strategy.

The field is roughly where SEO was in the mid-2000s: clearly important, directionally understood, but without the longitudinal evidence base that would make optimization a science rather than an informed craft. The organizations that invest in answering these questions (through rigorous measurement of their own GEO interventions, through cross-platform testing, through honest attribution modeling) will be the ones that eventually define best practices for everyone else.

Sources Cited

  1. 6sense, "B2B Buyer Experience Report 2025"

  2. G2, "How AI Chat is Rewriting B2B Software Buying" (Oct 2025)

  3. G2, "2025 Buyer Behavior Report" (May 2025)

  4. Responsive, "2025 State of Strategic Response Management" (2025)

  5. TrustRadius, "Bridging the Trust Gap: B2B Tech Buying in the Age of AI" (2025)

  6. Forrester, "The State of Business Buying, 2024" / Buyers’ Journey Survey (2024)

  7. Conductor, "The State of AEO/GEO in 2026: CMO Investment Report" (Jan 2026)

  8. Ahrefs, "AI Overviews Change Every 2 Days" (Nov 2025)

  9. GrowthUnhinged, "2025 State of B2B GTM Report" (2025)

  10. eMarketer, "Generative AI in Marketing 2025" (2025)

  11. Ahrefs, "76% of AI Overview Citations Pull From Top 10 Pages" (Aug 2025)

  12. Ahrefs, "Update: 38% of AI Overview Citations Pull From Top 10 Pages" (Mar 2026)

  13. Ahrefs, "Only 12% of AI Cited URLs Rank in Google’s Top 10" (Sep 2025)

  14. Aggarwal et al., "GEO: Generative Engine Optimization," KDD 2024, Princeton University

  15. Seer Interactive, "AIO Impact on Google CTR: September 2025 Update" (Nov 2025)

  16. Profound, "From SEO to GEO" (Comcast LIFT Labs, 2025)

  17. Seer Interactive, "What Drives Brand Mentions in AI Answers" (2025)

  18. Ahrefs, "What Correlates With AI Overview Brand Visibility" (Aug 2025)

  19. Semrush, "We Studied the Impact of AI Search on SEO Traffic" (Jul 2025)

  20. Akamai, "AI Pulse: How AI Bots and Agents Will Shape 2026" (Jan 2026)

  21. Akamai, "AI & LLM Bot Management Has Become a Business-Critical Issue" (Oct 2025)

  22. Ahrefs, "We Made Up a Brand to See if AI Would Recommend It" (Dec 2025)

  23. SparkToro & Gumshoe, "AI Recommendation Consistency Study" (Jan 2026)

  24. Cloudflare, "Content Independence Day: No AI Crawl Without Compensation" (July 2025)

  25. Cloudflare, "Control Content Use for AI Training" (July 2025)

  26. AirOps, "Staying Seen In AI Search: How Citations & Mentions Impact Brand Visibility" (Sep 2025)

  27. Authoritas, "SERP Organic and AI Overview Volatility Research" (2025)

  28. Ahrefs, "ChatGPT’s Top 1,000 Most-Cited Pages" (2025)

  29. eMarketer, "Marketers Invest More as GEO and AI Visibility Becomes Top Priority" (Jan 2026)

  30. Ahrefs, "How Google AI Overviews Impact Organic CTR" (Dec 2025)

  31. Forrester, "2026 B2B Marketing, Sales, and Product Predictions" (Oct 2025)

  32. Gartner, "Strategic Predictions for 2026 and Beyond" (Oct 2025)

  33. Seer Interactive, "Brand Mentions and AI Visibility" (2025)

  34. Chartbeat via Reuters Institute / Press Gazette, "Global Publisher Google Traffic Dropped by a Third in 2025" (Jan 2026)

  35. McKinsey, "The State of AI in 2025: Agents, Innovation, and Transformation" (Nov 2025)

  36. Pew Research Center, "Google Users Are Less Likely to Click on Links When an AI Summary Appears" (Jul 2025)

  37. Similarweb, "Zero-Click Searches and How They Impact Traffic" (May 2025)

  38. TrustRadius, "From Buzzword to Backbone: How AI Is Redefining B2B Search and Buying" (Oct 2025)

  39. Ahrefs, "Top AI Search Engines by Traffic" (2025)

  40. Yext, "AI Citations, User Locations, & Query Context" (Oct 2025)

  41. Columbia Journalism Review, "Reddit Is Winning the AI Game" (Oct 2025)

About Retina Media and Shane H. Tepper

Retina Media is a GEO consultancy that helps B2B brands become visible, credible, and accurately represented across AI-native discovery surfaces.

Shane Tepper is the architect of Retina’s GEO frameworks and one of the early practitioners in the field. Before launching Retina, he led content strategy at IDVerse (acquired by LexisNexis Risk Solutions in 2025), where he built AI-powered content workflows that contributed to the company’s growth through acquisition. His background spans fifteen-plus years across film, advertising, and B2B SaaS, with senior roles at Udacity, SoFi, BBDO, and agencies supporting brands including Wells Fargo, HP, Coca-Cola, and AT&T. Shane holds a Bachelor of Arts in Creative Writing and American History from the University of Pennsylvania.

He writes and speaks regularly on AI’s impact on brand discovery, B2B marketing strategy, and the broader implications of machine-mediated decision-making. 

Get in touch

To explore how Retina Media can make your brand visible where your buyers are actually researching:

Email: shane@retina.media

LinkedIn: linkedin.com/in/shanetepper

Web: retina.media

Phone: (404) 509-9910

Tags GEO, Generative Engine Optimization, AI Visibility, AI Citations, AI Measurement
The Committee Member Nobody Invited →

Latest

Featured
Mar 6, 2026
The GEO White Paper v4.0: AI-Native Brand Visibility for B2B (March 2026)
Mar 6, 2026
Mar 6, 2026
Feb 28, 2026
The Committee Member Nobody Invited
Feb 28, 2026
Feb 28, 2026
Feb 9, 2026
The Blind Automation Problem
Feb 9, 2026
Feb 9, 2026
Feb 1, 2026
The File That Makes AI Sound Like You
Feb 1, 2026
Feb 1, 2026
Jan 24, 2026
GEO Tools Can't Tell You What to Build
Jan 24, 2026
Jan 24, 2026
Jan 16, 2026
Why Proprietary Data Is the Last Content Advantage in AI Search
Jan 16, 2026
Jan 16, 2026
Jan 6, 2026
5 “Boring” AI Agents Every B2B GTM Team Should Deploy ASAP
Jan 6, 2026
Jan 6, 2026
Dec 20, 2025
11 Predictions for AI-Enabled GTM in 2026
Dec 20, 2025
Dec 20, 2025
Dec 15, 2025
The Trial of Traditional SEO
Dec 15, 2025
Dec 15, 2025
Dec 6, 2025
The Next Trillion-Dollar AI Opportunity is GTM. SaaS Won't Capture It.
Dec 6, 2025
Dec 6, 2025

© 2026  Shane H. Tepper