Installation
Setup
Basic Example
What Gets Traced
Phoenix automatically captures:- Queries: Query text, retrieved documents, synthesis
- Retrievals: Document chunks, similarity scores, ranking
- LLM Calls: Prompts, completions, tokens, model parameters
- Embeddings: Text being embedded, embedding model
- Agents: Tool selection, reasoning steps, execution
- Synthesis: Context assembly, prompt construction
Advanced Examples
RAG with Custom Retrieval
Agent with Tools
Sub-Question Query Engine
Streaming Responses
Chat Engine
Observability in Phoenix
Once instrumented, you can:- Visualize the RAG pipeline from query to response
- Inspect retrieved documents with similarity scores
- Monitor embedding calls and vector search performance
- Track LLM token usage and costs
- Debug agent reasoning and tool selection
- Analyze query latency across pipeline stages
Resources
Example Notebook
Complete tutorial
OpenInference Package
View source code
Working Examples
More code examples
LlamaIndex Docs
Official documentation
Legacy Versions
For LlamaIndex versions prior to 0.11.0, see the legacy integration guide in the Phoenix documentation.