Phoenix is an open-source AI observability platform designed for experimentation, evaluation, and troubleshooting. Whether you’re building chatbots, RAG applications, or AI agents, Phoenix gives you the visibility and tools you need to ship production-ready LLM applications.
What is Phoenix?
Phoenix provides a comprehensive suite of tools for understanding and improving your LLM applications:Tracing
Trace your LLM application’s runtime using OpenTelemetry-based instrumentation. See every LLM call, retrieval step, and tool usage.
Evaluation
Leverage LLMs to benchmark your application’s performance using response and retrieval evals.
Datasets
Create versioned datasets of examples for experimentation, evaluation, and fine-tuning.
Experiments
Track and evaluate changes to prompts, LLMs, and retrieval configurations.
Playground
Optimize prompts, compare models, adjust parameters, and replay traced LLM calls.
Prompt Management
Manage and test prompt changes systematically using version control, tagging, and experimentation.
Key Features
Open Source & Vendor Agnostic
Phoenix is fully open source and works with any LLM provider or framework. Out-of-the-box support includes: Frameworks: OpenAI Agents SDK, LangGraph, Vercel AI SDK, Mastra, CrewAI, LlamaIndex, DSPy, and more LLM Providers: OpenAI, Anthropic, Google GenAI, Google ADK, AWS Bedrock, LiteLLM, and moreFlexible Deployment
Phoenix runs practically anywhere:- Local development: Single command to launch (
python -m phoenix.server.main serve) - Docker: Pre-built container images available on Docker Hub
- Kubernetes: Helm charts for production deployments
- Cloud: Managed instances at app.phoenix.arize.com
Built on OpenTelemetry
Phoenix is built on OpenTelemetry, the industry standard for observability. This means:- Language and framework agnostic
- Compatible with existing observability infrastructure
- Future-proof as OpenTelemetry continues to evolve
- Integrates seamlessly with other OpenTelemetry-compatible tools
How Phoenix Works
Phoenix captures telemetry data from your LLM applications using the OpenInference specification, an extension of OpenTelemetry designed specifically for LLMs. This data includes:- LLM calls: Inputs, outputs, tokens, latency, and model parameters
- Retrievals: Documents fetched from vector stores and their relevance scores
- Tool usage: Function calls, arguments, and results
- Agent traces: Complete execution paths including reasoning steps
- Visualize traces in an intuitive UI
- Run evaluations to measure quality metrics
- Create datasets from production examples
- Run experiments to test improvements
- Optimize prompts in the playground
Getting Started
Quickstart
Get Phoenix running in 5 minutes with our quickstart guide.
Installation
Detailed installation instructions for pip, conda, Docker, and more.
Deployment
Learn about deployment options from local to production.
Tracing Integrations
Browse integrations for your framework and LLM provider.
Community & Support
Join thousands of AI builders using Phoenix:- Slack: Join our community Slack for support and discussions
- GitHub: Star us on GitHub and contribute
- Documentation: Explore our comprehensive docs
- X/Twitter: Follow @ArizePhoenix for updates
Phoenix is licensed under the Elastic License 2.0 (ELv2). It’s free to use and modify, with some restrictions on providing it as a managed service.