Online search is changing radically: AI is no longer displaying classic results, but synthesizing direct answers, making SERP ranking irrelevant. This article explains the new 'search stack' built on vectors, embeddings, knowledge graphs, and LLMs, offering a practical guide for sites to stay visible in the AI era. Concrete optimization strategies and the advantages of adapting to new information retrieval flows are analyzed.
Think about this shift: today, when someone wants to learn about smart glasses, they’re no longer searching for “best smart sunglasses” and clicking through blue links.
Instead, they ask directly: “What’s up with the Meta Ray-Bans?” – and instantly get a complete answer, with specs, use cases, reviews – all without ever seeing a web page or even a SERP.
This is where the game changes completely: it’s no longer about your site ranking first. What matters is that it gets retrieved, understood, and used to build an answer.
We’re witnessing the collapse of a model we’ve known for years. Until now, it was straightforward: write a page, hope Google or Bing indexes it, pray your keywords match people’s searches, and hope you’re not outbid by someone with a bigger ads budget.
Generative AI systems don’t need your page to appear in a traditional list – they just need it to be well-structured, easy to read, and available when they have to answer. That’s where Generative Engine Optimization (GEO) and Answer Engine Optimization (AEO) come into play – strategies for making your content compatible with modern AI flows. In 2024, about 60% of Google searches no longer lead to a click on a website, but to a direct AI-generated answer (Zero-click searches rising, organic clicks dropping: report).
This is the new “search stack.” It’s not built on links, pages, or classic rankings – but on vectors, embeddings, ranking fusion, and LLMs that analyze semantically, not just alphabetically.
You’re no longer just optimizing a web page – you’re optimizing how your content can be fragmented, semantically scored, and assembled into a coherent answer for any AI query.
Once you understand this paradigm shift, classic SEO strategies start to look totally outdated. (The diagrams below are simplified for clarity.)
Behind every modern AI-powered retrieval system, there’s an invisible architecture for users – and it’s totally different from what we knew before.
Every phrase, paragraph, or whole document becomes a vector – basically, a complex snapshot of its meaning.
This is where embeddings are stored and retrieved at lightning speed. Think Pinecone, Weaviate, Qdrant, FAISS.
It may sound old-school, but it still works incredibly well!
BM25 evaluates content based on keyword frequency and rarity.
The graph below shows a conceptual comparison between BM25 and vector similarity ranking – based on hypothetical data to illustrate relevance differences. Notice how documents are ranked differently according to totally different principles.
This method combines results from multiple retrieval systems (for example, BM25 plus vector similarity) into one final ordered list.
RRF blends signals using the reciprocal of each document’s position in each system. The chart below shows how a document’s rank in one system contributes to its RRF score – favoring documents that perform well across several dimensions, even if they’re not #1 in any single one.
After the top results are retrieved, the LLM generates the final answer – either as a summary, a reformulation, or a direct quote.
And yes, indexing still exists – but it’s completely different now.
We’re not talking about slow crawling hoping for a good spot in the SERP. Information gets converted into vectors and stored in the DB to be retrieved by meaning, not by metadata or classic ranking.
The new model doesn’t eliminate classic search completely. It just outperforms it, especially where the old way never really worked well:
Key advantages:
It’s no surprise areas like enterprise search, customer support, and knowledge management are already adopting this paradigm. And, step by step, general search is heading the same way.
Vectors are powerful, but... a bit fuzzy. They get at meaning, but miss concrete “who, what, when” relationships that humans take for granted.
That’s where knowledge graphs come in.
They define concrete relationships between entities (people, products, brands) so the system can disambiguate and build deeper reasoning. Is it Apple the company, or apple the fruit? Who does “he/she” refer to in an ambiguous sentence?
Used together:
You don’t have to choose between a knowledge graph and the new search stack. The best generative AI systems use both – benefiting from hybrid advantages. Even in industries where context is everything (like legaltech), this integration can triple the accuracy and contextualization of AI results.
Let’s recap what used to matter for classic ranking.
This isn’t an exhaustive list, just setting the contrast for what follows. While traditional SEO was complex (speaking from direct Bing experience), what’s coming is way more dynamic.
For classic search, the focus was on:
Everything was a mix of technical hygiene, content relevance, and reputation – partly measured through external references.
But what should you actually do to stay visible in generative-AI search? Here are the practical recommendations, distilled from GEO/AEO experts and what’s actually cited in LLM answers:
Break info into easy-to-retrieve blocks. Use semantic HTML (<h2>
, <section>
, etc.), FAQ questions, modular formatting. This is the first layer the LLM processes. Implementing specific markups like FAQPage or Article on schema.org is essential for guiding AI crawlers.
Write to be understood, not admired. Avoid excessive jargon, complicated metaphors, or creative intros with no substance. Answer directly, just like a competent AI assistant would.
If GPTBot, Google-Extended, or CCBot can’t access your site, you basically don’t exist for AI-generated answers. Avoid content generated exclusively in JavaScript, make sure critical info is in HTML, and implement schema markup for structure and context. For detailed strategies, see The Proven AEO Guide: How to Do Answer Engine Optimization.
LLMs prefer credible sources. Clear bylines, publication dates, contact pages, external citations, and author bios help a lot. Statistically, pages with these signals are up to 30% more likely to be included in AI answers. Using concrete data or expert quotes increases your chances by over 20% (The future of building B2B authority in the AI search era).
Connect relevant pages, define concepts and clear links. Hub-and-spoke models, glossaries, and contextual links strengthen semantic coherence and boost retrievability.
Answer from multiple angles: what, why, how, comparisons, when, plus summaries, tables, checklists. This multiplies the odds that your fragments get synthesized in AI-generated snippets.
Use firm, declarative language. Avoid vague wording like “could,” “possibly,” “some believe.” The more categorical you are, the more likely AI will trust and cite you.
Restate the same idea in several ways to cover a wider range of possible queries. Expressive diversity amplifies your semantic footprint.
Each paragraph should contain only one clear, compact idea. This helps with embedding and precise retrieval.
Mention full names even if they seem obvious. For example: “the GPT-4 model from OpenAI,” not just “the new model.”
Back up claims with examples, stats, or analogies nearby. It strengthens coherence and helps the LLM reason better.
Think bullet points, short summaries, or “Key Takeaway” sections for each topic. This is perfect fuel for generative AI.
Create a dense “neighborhood”: glossaries, definitions, comparison pages, case studies. Interlink them to maximize recall in your core areas.
Test if your site is picked up by Perplexity, ChatGPT with browsing, etc. If not, review your structure, clarity, and authority signals, then adjust. Practical strategies for monitoring can be found in The Proven AEO Guide: How to Do Answer Engine Optimization.
Your site is no longer the final destination. It becomes raw material for answers.
In a world where generative AI rules, the best-case scenario is to be included – quoted, highlighted, extracted into an answer that someone finds useful. It’s a new imperative, especially since consumer access to information is diversifying fast – for example, the next generation of Meta Ray-Bans will search and get answers directly, no classic web interface required.
Web pages still matter. But increasingly, they become the skeleton, the basic infrastructure.
If you want to win in this new game, forget the old ranking obsession. Think like a credible source. It’s no longer about traffic volume, but about being included, cited, and intelligently reused.
This article was originally published on Duane Forrester Decodes, Substack (as Search Without a Webpage) and is republished with permission.