Article imageLogo
Chatbots Behaving Badly™

From SEO to RAG - How the Web’s Attention Engine Is Being Rewired

By Markus Brinsa  |  August 27, 2025

Sources

What actually changed — and why it feels so different

For two decades, SEO taught writers to please two audiences at once: people and a crawler with a clipboard. The deal was transactional. Rank higher, earn clicks, capture attention. Then generative systems began answering questions directly. Retrieval-Augmented Generation (RAG) pulls relevant snippets from a knowledge store (often built from web content), hands them to a model, and synthesizes an answer on the fly. The user gets resolution; the publisher often loses the visit. Google’s own AI Overviews crystallized the shift: multiple independent analyses now show meaningful drops in organic clicks when AI answers occupy the “zero-click” zone at the top. Some publisher cohorts report average declines between roughly a quarter and a third for affected queries, and industry tallies note broad 1–25% referral dips since AI Overviews expanded. That’s not a theoretical trend line — it’s traffic leaving the building.

This is why people say “RAG is replacing SEO.” It isn’t that search is gone; it’s that the discovery surface has been skinned over by a synthetic explainer layer that preferentially absorbs your work before a human ever reaches you. In the blue-links era, the click was the prize. In the RAG era, inclusion is.

RAG in plain English

RAG doesn’t replace a large language model; it sobers it up. Models are eloquent but forgetful; retrieval grounding is how developers keep them factual. The pipeline is prosaic: chunk your page into passages, embed those into vectors (a map of meaning), retrieve the most relevant chunks when someone asks a question, and let the model weave them into an answer. The research literature has matured fast: whole surveys now dissect chunking, re-ranking, context filtering, and evaluation frameworks — in other words, the rules of what gets pulled and what gets ignored. If you want your ideas to show up in answers, you’re effectively negotiating with this pipeline.

The copyright trench warfare you can’t ignore

While technology marched forward, the law struggled to catch up. Major newsrooms have split their response: sue or license. The Associated Press took the licensing path in 2023. The Financial Times, News Corp, Axel Springer, Le Monde, and others followed with deals that permit training and attributed use inside products like ChatGPT. This is data diplomacy: pay for access, show attribution, reduce legal risk.

The other flank is litigation. The New York Times’ case against OpenAI and Microsoft survived a key motion to dismiss this spring, keeping core copyright claims alive and leading to sweeping preservation orders over user logs — a procedural turn with privacy and governance ripples far beyond one lawsuit. Expect the case (and similar actions globally) to shape norms for how training and real-time retrieval can use paywalled or licensed works.

Europe is adding its own pressure. The EU AI Act phases in transparency duties for general-purpose models, including disclosures about training data, while the earlier DSM Directive already created text-and-data-mining exceptions with a rightsholder opt-out — crucially, an opt-out that must be machine-readable. In short, if you reserve your rights properly, mining for training is not permitted in the EU. That’s a lever, and it’s starting to bite.

Opt-out theater, opt-in reality

“Just block the bots,” people say. Sometimes you can. OpenAI documents GPTBot and honors robots.txt disallow directives; Apple added Applebot-Extended so publishers can forbid use of their pages for training; Google offers Google-Extended as the training opt-out flag. Cloudflare now even manages these patterns for customers who don’t want to babysit their robots files. This is helpful, and it’s real.

But two uncomfortable truths follow. First, robots.txt is a courtesy, not a gate — even Google’s own documentation says it relies on the crawler's good faith. Second, opting out of training isn’t the same as opting out of generative use inside a search product. Reporting and testimony indicate that content opted out of model training may still be processed to power AI features like Overviews, because that’s categorized as “search,” not “training.” In practice, your blockade can be porous.

So is writing for humans over? The Goodhart trap

When a measure becomes a target, it ceases to be a good measure. That’s Goodhart’s law — and it ruined more than a few SEO playbooks. If creators pivot too hard toward “AI retrievability,” they risk collapsing their voice into the bland median that embeddings love. The research community is also warning about a parallel phenomenon on the model side: training repeatedly on generated content degrades models — the distribution tails become indistinguishable from one another. If we flood the open web with machine-optimized, machine-generated prose, we make both our content and the models worse. That’s the ouroboros nobody wants.

The counterweight is taste and specificity. Humans respond to narrative tension, memorable phrasing, original data, and earned authority. Machines respond to clarity, structure, and disambiguation. A professional strategy refuses the false choice and writes to both without turning into an instruction manual.

An AI-optimized content strategy that doesn’t sell your soul

The practical question isn’t “How do I beat the model?” It’s “How do I become a dependable node in the model’s knowledge graph without becoming invisible to people?” Here’s how that plays out when you build it into editorial practice.

Start with a structure that machines can parse without forcing your prose into bullet gibberish. Use real H2 and H3 headings, keep paragraphs coherent around a single claim, define acronyms, and introduce entities with their full names before using shorthands. RAG retrieval thrives on passages that answer a question cleanly. If a section genuinely is a question with a short answer, publish a companion Q&A or FAQ version with proper schema markup so the structure is explicit to crawlers. Schema.org’s QAPage and FAQPage types, expressed in JSON-LD, remain the clearest signals. Do not fake it; only mark up pages that truly are Q&A or FAQs — the search guidelines haven’t changed on that point.

Make your content retriever-friendly at the protocol layer. Serve clean, crawlable HTML; don’t hide core text inside images or render-blocking scripts. Publish article and FAQ sitemaps. Where appropriate, expose a machine-readable feed or API for high-value sections. The paradox of RAG is that the easier you make high-signal passages to extract, the more likely you are to be included in answers — and the more chances you have to carry your brand through.

Use licensing and attribution to your advantage. There’s an emerging premium on trustworthy, rights-cleared corpora. Many publishers have discovered that being part of a licensed set can increase the odds of name-on-screen presence in AI products, because the products can cite and quote with fewer legal headaches. The FT and News Corp deals explicitly mention attributed use inside ChatGPT; Stack Overflow’s partnership is framed around surfacing validated technical knowledge with attribution. If you’re a niche authority, don’t just “not block” — package, pitch, and license.

Keep humans on a higher rung. Use the open web for machine-readable clarity and guard the truly distinctive value — methodology, datasets, nuanced case studies, proprietary frameworks — for logged-in readers, newsletters, private podcasts, events, or subscribers. That’s not hoarding; it’s segmentation. In a world where top-of-funnel answers are commoditized, your brand grows by offering depth that a one-screen synopsis can’t compress.

Instrument what the machines are doing to you. Watch how your excerpts appear in AI answers. When you spot verbatim phrasing in generated results, lean into repeatable tag lines, author fingerprints, and signature formulations that are both quotable and identifiable. If you can trace your phrasing in the wild, you can demonstrate impact to partners — and to yourself.

Be realistic about blocking and provenance. If you intend to keep some work out of training sets, reserve rights in the EU using a machine-readable opt-out and backstop it with robots directives for the big crawlers (GPTBot, Applebot-Extended, Google-Extended). And if you publish media, begin adopting C2PA Content Credentials for images and video; it won’t stop scraping, but provenance is rapidly moving from “nice to have” to table stakes in tooling and policy. Pair that optimism with skepticism: robots files are voluntary; C2PA isn’t a magic shield; adoption is uneven and the standard has limits. But these are levers you control today.

The hard part: writing for two audiences without sounding like a bot

Here’s the editorial north star: make the first read work for a human and the second read work for a machine. That means your lead still needs to hook, your sections still need narrative logic, and you still need a voice. Inside that voice, practice explicitness. Use canonical names, dates, numbers, and definitions. Place the answer near the question. Avoid coy references that only make sense inside your personal canon. Humans forgive a little ambiguity; embeddings do not.

And then consider the macroeconomics of your own data exhaust. If you over-optimize every surface for retrieval, you contribute to the homogenization that research calls “model collapse” — the tails of language distributions get shaved off when models train on their own outputs and machine-optimized text. The antidote is not romantic Luddism. It’s originality. Keep generating signal that isn’t statistically convenient. The models need it, your readers crave it, and your brand depends on it.

Where this lands next

Expect three things. First, more licensing and syndication. It will feel like the early days of streaming: messy deals, uneven payouts, but a path for high-quality data to get compensated and cited. Second, increased European pressure on training transparency and opt-outs; the AI Act’s GPAI provisions and a new disclosure template are already transforming transparency from a blog-post promise into a compliance requirement. Third, more publisher pushback on AI Overviews-style features that divert clicks; some are already exploring competition and antitrust angles in Europe. The friction will continue — because for the first time since PageRank, the web’s attention engine is being meaningfully renegotiated.

Bottom line for your team

No, RAG hasn’t “killed” SEO. It has moved the goalposts. The click is no longer guaranteed even when you “win” the query. The pragmatic strategy is to be legible to machines without surrendering your voice to them, to license when it strengthens attribution, to reserve rights when it doesn’t, and to build owned channels where the relationship is direct and defensible. The web still rewards clarity and originality — and now, so do the models.

About the Author

Markus Brinsa is the Founder and CEO of SEIKOURI Inc., an international strategy consulting firm specializing in early-stage innovation discovery and AI Matchmaking. He is also the creator of Chatbots Behaving Badly, a platform and podcast that investigates the real-world failures, risks, and ethical challenges of artificial intelligence. With over 15 years of experience bridging technology, business strategy, and market expansion in the U.S. and Europe, Markus works with executives, investors, and developers to turn AI’s potential into sustainable, real-world impact.

©2025 Copyright by Markus Brinsa | Chatbots Behaving Badly™