What Does “Vectorless” Mean?
Traditional RAG systems require several vector-based components:Embedding Model
Converts text chunks into high-dimensional vectors
Vector Database
Stores and indexes embeddings for similarity search
Similarity Search
Computes distance metrics between query and document vectors
Chunk Management
Handles splitting, overlapping, and metadata for chunks
How PageIndex Works Without Vectors
PageIndex replaces the entire vector-based pipeline with two core components:1. Hierarchical Tree Index
Instead of embedding chunks, PageIndex creates a semantic tree structure:2. LLM Reasoning for Retrieval
Instead of similarity search, PageIndex uses the LLM to reason about which nodes are relevant:The LLM reads the tree structure (titles, summaries, page ranges) and makes intelligent decisions about where to look - no vector math required.
Why Go Vectorless?
1. Better Accuracy
PageIndex achieves 98.7% accuracy on FinanceBench without vectors. Why? Vectors lose information: Compressing text into fixed-dimensional embeddings loses nuance, structure, and context. Reasoning preserves understanding: LLMs can fully comprehend section titles, hierarchical relationships, and semantic meaning.Example: A section titled “Regulatory Developments” has clear semantic meaning that’s preserved in text but compressed away in a 1536-dimensional embedding.
2. True Explainability
Vector similarity is fundamentally opaque:3. No Infrastructure Overhead
Vector RAG requires significant infrastructure:- Vector databases: Pinecone, Weaviate, Qdrant, Milvus
- Embedding APIs: OpenAI, Cohere, or self-hosted models
- Index management: Building, updating, and maintaining vector indices
- Dimension tuning: Choosing embedding dimensions and distance metrics
- JSON storage: Any file system or database
- LLM API: For generation and reasoning (already needed for RAG)
4. No Chunking Problems
Vector RAG must split documents into chunks for embedding:- Lost boundaries: Important sections split mid-thought
- Lost hierarchy: Parent-child relationships destroyed
- Lost context: Chunks don’t know their position in the document
- Parameter sensitivity: Results vary wildly with chunk size
- Natural boundaries: Follows document structure
- Preserved hierarchy: Parent-child relationships maintained
- Full context: Exact page ranges and position
- No parameters: Structure is inherent to the document
5. Dynamic Context Windows
Vector RAG is limited by fixed chunk sizes:Technical Architecture Comparison
Traditional Vector RAG Pipeline
PageIndex Vectorless Pipeline
Notice how PageIndex eliminates the embedding and vector search steps entirely, simplifying the architecture while improving performance.
What About Scalability?
Common question: “Don’t vector databases scale better than reading JSON trees?” Answer: PageIndex scales differently, but effectively:Vector DB Scaling
- Millions of documents: Vector DBs excel at searching millions of embeddings
- Per-document cost: Constant time lookup (O(log n) or better with indexing)
- Use case: Finding similar documents across a large corpus
PageIndex Scaling
- Single document depth: PageIndex excels at deep understanding of individual documents
- Per-document cost: Proportional to tree depth (typically O(log n) nodes)
- Use case: Extracting precise information from long, complex documents
Real-world reality: Most enterprise RAG applications deal with 10-10,000 documents, not millions. At this scale, PageIndex’s JSON-based approach is more than sufficient and far simpler to maintain.
When Vectors Are Still Useful
The vectorless approach is optimal for: ✅ Long, structured documents (reports, manuals, textbooks) ✅ Professional/technical documents requiring precision ✅ Scenarios requiring explainability ✅ Complex, multi-step queries Vectors may still be better for:- Semantic search over large, unstructured text corpora
- Finding similar documents across millions of items
- Fuzzy matching where approximate results are acceptable
- Simple Q&A over homogeneous content
The AlphaGo Inspiration
PageIndex draws inspiration from AlphaGo’s success in mastering Go: AlphaGo: Used tree search (MCTS) + neural networks to evaluate board positions PageIndex: Uses tree search + LLM reasoning to evaluate document sections Both demonstrate that search + reasoning can outperform pure similarity-based approaches.Cost Implications
Removing vectors affects costs in several ways:Eliminated Costs
- ❌ Vector database hosting/licensing
- ❌ Embedding API calls (millions of chunks)
- ❌ Index maintenance and updates
New Costs
- ✅ Tree structure generation (one-time per document)
- ✅ LLM reasoning during retrieval (typically 2-5 tree search steps)
For most use cases, especially with long documents that would generate thousands of chunks, the vectorless approach is significantly more cost-effective.
Getting Started Without Vectors
Try PageIndex’s vectorless RAG approach:Next Steps
Tree Structure
Deep dive into PageIndex’s hierarchical tree format
Vectorless RAG Tutorial
Build a complete RAG system without vectors
API Quickstart
Start generating tree structures via API
Vision RAG
OCR-free, vectorless RAG with page images