Corpus Architecture for RAG: The Indexing Decisions That Determine Quality Before Retrieval Starts
When a RAG system returns the wrong answer, the post-mortem almost always focuses on the same suspects: the retrieval query, the similarity threshold, the reranker, the prompt. Teams spend days tuning these components while the actual cause sits untouched in the indexing pipeline. The failure happened weeks ago when someone decided on a chunk size.
Most RAG quality problems are architectural, not operational. They stem from decisions made at index time that silently shape what the LLM will ever be allowed to see. By the time a user complains, the retrieval system is doing exactly what it was designed to do — it's just that the design was wrong.
This is a guide to the index-time decisions that matter most, what the 2024–2026 benchmarks actually show, and the patterns that consistently distinguish high-quality RAG systems from ones that look fine in demos and fall apart in production.
