The Rise of Agentic Retrieval Over Vector Indexing**
The successful stories of Claude Code have shown that you can skip heavyweight vector databases and let the model itself handle retrieval with simple tools: well-written llms.txt and grep calls. Surprisingly, this minimalist approach delivers more accurate and faster retrieval, proving that a reasoning-driven model can outperform embedding-based systems in both retrieval precision and latency.
We take that same spirit beyond code.
PageIndex is an LLM-native, vectorless index for PDFs and long-form documents — a hierarchical table-of-contents tree that lives inside the model’s context window, enabling the model to reason and navigate like a human reader. No vector DB required.
Classic RAG pipeline looks like this:
split content into chunks → embed → store in a vector DB → semantic search → (blend results with keyword search) → (rerank) → stuff the context → answer.
It works, but it’s complex to build, hard to maintain, and slow to iterate on, often more infrastructure than you really need.
In contrast, a new wave of code agents takes a refreshingly simple approach:
In real-world tests on developer docs, practitioners have found that a well-crafted llms.txt (URLs + succinct descriptions) plus simple tool calls like grep outperforms vector-DB pipelines for various coding tasks. It’s not only more accurate, but also dramatically easier to maintain and update.
Why does this minimalist approach work so well?
This is the essence of the new agentic retrieval paradigm: move retrieval intelligence into the model, keep external tools minimal, and represent the corpus in a way the LLM can navigate through reasoning rather than similarity. It’s about trusting and leveraging the model’s ability to reason, retrieve, and self-direct — not treating it as a passive consumer of externally retrieved chunks, but as an active agent capable of planning, retrieving, and deciding what to read next.