Basic LLM Tree Search
A simple strategy is to use an LLM agent to conduct tree search. The LLM analyzes the document tree structure and identifies relevant nodes based on the query.Implementation
In our dashboard and retrieval API, we use a combination of LLM tree search and value function-based Monte Carlo Tree Search (MCTS). More details will be released soon.
Integrating User Preference or Expert Knowledge
Unlike vector-based RAG where integrating expert knowledge or user preference requires fine-tuning the embedding model, in PageIndex, you can incorporate user preferences or expert knowledge by simply adding knowledge to the LLM tree search prompt.Implementation Pipeline
Preference Retrieval
When a query is received, the system selects the most relevant user preference or expert knowledge snippets from a database or a set of domain-specific rules. This can be done using keyword matching, semantic similarity, or LLM-based relevance search.
Tree Search with Preference
Integrate the retrieved preference into the tree search prompt to guide the LLM’s node selection.
Example Expert Preference
If the query mentions EBITDA adjustments, prioritize Item 7 (MD&A) and footnotes in Item 8 (Financial Statements) in 10-K reports.By integrating user or expert preferences, node search becomes more targeted and effective, leveraging both the document structure and domain-specific insights.
Complete Example
Here’s a complete example that combines tree search with preference integration:Benefits of Tree Search with Preferences
- No Model Fine-tuning: Unlike vector-based RAG, preferences are integrated directly into prompts
- Dynamic Updates: Expert knowledge can be updated without retraining models
- Transparent Reasoning: LLM provides explicit reasoning for node selection
- Domain Expertise: Leverages document-specific knowledge and user preferences
- Flexible Integration: Supports multiple preference types (expert rules, user history, domain guidelines)