The advent of AI agents has fundamentally altered the paradigm of web search, moving beyond human-centric keyword matching to a demand for deep contextual understanding and comprehensive data retrieval. This profound shift was eloquently articulated by Will Bryk, CEO of Exa.ai, at the AI Engineer World's Fair in San Francisco, where he outlined how neural network RAG (Retrieval Augmented Generation) is rebuilding web search for the intelligence of machines, not just humans. Bryk demonstrated the stark contrast between traditional keyword-based search and Exa's neural RAG, emphasizing its critical role in empowering sophisticated AI agents.
Bryk traced the evolution of search, noting how Google in 1998 felt "magical" by simply finding documents containing specific keywords. However, by 2021, with the emergence of powerful language models like GPT-3, traditional search began to feel "ancient." He highlighted a common frustration: a Google search for "shirts without stripes" would still yield images of striped shirts. This disconnect underscored a core problem: keyword search algorithms lacked the semantic understanding inherent in advanced AI models. Exa's mission, born from this realization, was to create a search engine that truly understood complex queries at a deep level, leveraging the same transformer technology powering LLMs.
Exa’s core innovation lies in its transition from a keyword index to an embedding index. While a keyword index merely maps words to documents, an embedding index transforms entire documents into numerical vectors that capture their meaning, ideas, and even how they are referenced across the web. This "arbitrarily powerful representation" allows Exa to understand the nuances of a query, far beyond simple word matching.
The launch of ChatGPT in November 2022 initially presented an existential crisis for search companies, including Exa. The immediate thought was, "Is there even a role for search in this world?" However, the answer quickly became apparent: LLMs, despite their power, do not possess the entirety of the internet's knowledge. GPT-4, for instance, stores less than 10 terabytes of data, while the internet encompasses millions of terabytes, constantly updating. This fundamental information theory argument dictates that "LLMs always will need search."
The critical insight Will Bryk provided is that traditional search engines were built "for humans." Humans are "lazy" searchers, typing simple keywords and wanting to click a few links, prioritizing UI and quick, digestible results. AI agents, conversely, are profoundly different. They require "precise, controllable information" and "comprehensiveness," often submitting complex, multi-paragraph queries to find vast quantities of structured knowledge, which they can process instantly. A human receiving thousands of search results would be overwhelmed; an AI can digest them in seconds.
This divergence in needs creates a massive opportunity. AI agents demand a search API that offers granular control, allowing them to specify date ranges, domains, result counts (hundreds or thousands), and even exclude specific text. Exa is designed to provide this comprehensive, controllable access, effectively turning the web into a queryable database for AI. The "space of possible queries" has expanded dramatically, encompassing not just keyword and semantic searches, but also complex, multi-faceted inquiries previously unimaginable. Exa aims to be the single API that serves all these evolving needs, empowering AI systems with the precise, vast knowledge they require.

