Publisher Traffic Is Collapsing. Here Is Why GEO Practitioners Should Care.
Small publishers lost 60% of search referrals. Google beat a publishers antitrust suit. For brands, the game is becoming a source inside the answer layer.
AI retrieval pipelines are shifting toward entity grounding and passage-level evidence. Structured signals now determine whether your content gets cited.
AI-generated answers in search are getting pickier about where they pull citations from. The retrieval pipelines behind systems like Google's AI Overviews are tightening—combining semantic similarity with entity grounding to keep citations stable and relevant. For content creators and SEO practitioners, this shift means structured signals and clearly anchored passages now carry more weight than ever in determining whether your content gets cited.
This is not a cosmetic change. It represents a fundamental evolution in how machines decide which sources to trust.
AI-generated summaries—whether from Google's AI Overviews, ChatGPT's search features, or Perplexity—operate through retrieval-augmented generation (RAG) systems. These pipelines don't scan the web in real time. They follow a structured sequence:
Google's own technical documentation on AI Overviews describes a "query fan-out" technique—where the system breaks complex queries into sub-queries, retrieves evidence for each, then assembles a coherent response. The retrieval layer works in tandem with Google's Knowledge Graph and existing ranking systems.
The key shift: passage-level evidence is increasingly tied to identifiable entities and structured context. This helps the model avoid citing the wrong source for a claim—a phenomenon researchers call citation drift.
Traditional search relied heavily on sparse retrieval—keyword frequency, TF-IDF scores, BM25 matching. Dense retrieval replaces this with learned embeddings that capture semantic meaning. The foundational Dense Passage Retrieval (DPR) paper from Facebook AI Research showed that learned dense embeddings outperform BM25 by 9–19% in passage retrieval accuracy.
What this means in practice: a page does not need to contain the exact words a user typed. It needs to contain passages that semantically match the intent behind the query—and those passages need to be clearly grounded in identifiable entities.
Recent work on interleaved reference-claim generation takes this further. Researchers have proposed systems that alternate between generating references and claims sentence-by-sentence, providing passage-level citation grounding. The goal is to ensure every factual claim in an AI-generated answer traces back to a specific, verifiable source passage.
In the old model, authority was measured by links. In the emerging model, authority is measured by entity clarity.
When an AI system retrieves a passage, it evaluates whether the entities mentioned in that passage are consistent with the entities the system has already identified as relevant to the query. If your content mentions "Google" but doesn't clearly establish whether you mean the company, the search engine, or the parent company Alphabet, the retrieval system has less confidence in your passage.
Schema.org structured data plays a direct role here. Structured metadata—Organization, Person, Article, Product types—provides machine-readable entity definitions that reinforce what your content is about and who is saying it. Google's developer documentation on AI features confirms that structured data and clear content organization help surfaces appear in AI-generated answers.
The pattern emerging across AI search systems—whether from Google, OpenAI, or Anthropic—is that machine readability is becoming as important as topical authority.
Pages that perform best in AI retrieval tend to share specific characteristics:
Google's Search Quality Rater Guidelines—updated in September 2025 to include explicit guidance on rating AI Overview responses—reinforce these signals. The guidelines emphasize experience, expertise, authoritativeness, and trustworthiness (E-E-A-T) as core quality dimensions, now applied to AI-generated content evaluation as well.
The web is slowly evolving from documents written for humans to knowledge modules optimized for retrieval systems. This is not speculation—it is the logical consequence of how RAG pipelines select and cite sources.
Search used to reward persuasion and keyword targeting. AI retrieval rewards clarity of knowledge structure. The machines are forcing the web to behave more like a library catalog than a marketing brochure.
For practitioners, the action items are concrete: define your entities explicitly, structure your content for passage-level extraction, deploy structured data that reinforces your identity, and write paragraphs that can stand alone as verifiable claims.
The retrieval pipeline does not care about your brand voice. It cares whether your passage answers the question, whether the entities match, and whether the structured context confirms the claim. Optimize for that.
James Calder is the editor of The Search Signal, covering AI-powered search, generative engine optimization, and the future of brand discovery.