loading…
Search for a command to run...
loading…
A MCP server for Arize Phoenix
Phoenix is an open-source AI observability platform designed for experimentation, evaluation, and troubleshooting. It provides:
Phoenix is vendor and language agnostic with out-of-the-box support for popular frameworks (OpenAI Agents SDK, Claude Agent SDK, LangGraph, Vercel AI SDK, Mastra, CrewAI, LlamaIndex, DSPy) and LLM providers (OpenAI, Anthropic, Google GenAI, Google ADK, AWS Bedrock, OpenRouter, LiteLLM, and more). For details on auto-instrumentation, check out the OpenInference project.
Phoenix runs practically anywhere, including your local machine, a Jupyter notebook, a containerized deployment, or in the cloud.
Install Phoenix via pip or conda
pip install arize-phoenix
Phoenix container images are available via Docker Hub and can be deployed using Docker or Kubernetes. Arize AI also provides cloud instances at app.phoenix.arize.com.
The arize-phoenix package includes the entire Phoenix platform. However, if you have deployed the Phoenix platform, there are lightweight Python sub-packages and TypeScript packages that can be used in conjunction with the platform.
| Package | Version & Docs | Description |
|---|---|---|
| arize-phoenix-otel | PyPI Version Docs | Provides a lightweight wrapper around OpenTelemetry primitives with Phoenix-aware defaults |
| arize-phoenix-client | PyPI Version Docs | Lightweight client for interacting with the Phoenix server via its OpenAPI REST interface |
| arize-phoenix-evals | PyPI Version Docs | Tooling to evaluate LLM applications including RAG relevance, answer relevance, and more |
| Package | Version & Docs | Description |
|---|---|---|
| @arizeai/phoenix-otel | NPM Version Docs | Provides a lightweight wrapper around OpenTelemetry primitives with Phoenix-aware defaults |
| @arizeai/phoenix-client | NPM Version Docs | Client for the Arize Phoenix API |
| @arizeai/phoenix-evals | NPM Version Docs | TypeScript evaluation library for LLM applications (alpha release) |
| @arizeai/phoenix-mcp | NPM Version Docs | MCP server implementation for Arize Phoenix providing unified interface to Phoenix's capabilities |
| @arizeai/phoenix-cli | NPM Version Docs | CLI for fetching traces, datasets, and experiments for use with Claude Code, Cursor, and other coding agents |
Phoenix is built on top of OpenTelemetry and is vendor, language, and framework agnostic. For details about tracing integrations and example applications, see the OpenInference project.
Python Integrations
| Integration | Package | Version | |
|---|---|---|---|
| OpenAI | openinference-instrumentation-openai |
PyPI Version | |
| OpenAI Agents | openinference-instrumentation-openai-agents |
PyPI Version | |
| LlamaIndex | openinference-instrumentation-llama-index |
PyPI Version | |
| DSPy | openinference-instrumentation-dspy |
PyPI Version | |
| AWS Bedrock | openinference-instrumentation-bedrock |
PyPI Version | |
| LangChain | openinference-instrumentation-langchain |
PyPI Version | |
| MistralAI | openinference-instrumentation-mistralai |
PyPI Version | |
| Google GenAI | openinference-instrumentation-google-genai |
PyPI Version | |
| Google ADK | openinference-instrumentation-google-adk |
PyPI Version | |
| Guardrails | openinference-instrumentation-guardrails |
PyPI Version | |
| VertexAI | openinference-instrumentation-vertexai |
PyPI Version | |
| CrewAI | openinference-instrumentation-crewai |
PyPI Version | |
| Haystack | openinference-instrumentation-haystack |
PyPI Version | |
| LiteLLM | openinference-instrumentation-litellm |
PyPI Version | |
| Groq | openinference-instrumentation-groq |
PyPI Version | |
| Instructor | openinference-instrumentation-instructor |
PyPI Version | |
| Anthropic | openinference-instrumentation-anthropic |
PyPI Version | |
| Smolagents | openinference-instrumentation-smolagents |
PyPI Version | |
| Agno | openinference-instrumentation-agno |
PyPI Version | |
| MCP | openinference-instrumentation-mcp |
PyPI Version | |
| Pydantic AI | openinference-instrumentation-pydantic-ai |
PyPI Version | |
| Autogen AgentChat | openinference-instrumentation-autogen-agentchat |
PyPI Version | |
| Portkey | openinference-instrumentation-portkey |
PyPI Version | |
| Agent Spec | openinference-instrumentation-agentspec |
PyPI Version | |
| Claude Agent SDK | openinference-instrumentation-claude-agent-sdk |
PyPI Version |
Normalize and convert data across other instrumentation libraries by adding span processors that unify data.
| Package | Description | Version |
|---|---|---|
| openinference-instrumentation-openlit | OpenInference Span Processor for OpenLIT traces. | PyPI Version |
| openinference-instrumentation-openllmetry | OpenInference Span Processor for OpenLLMetry (Traceloop) traces. | PyPI Version |
| Integration | Package | Version | |
|---|---|---|---|
| OpenAI | @arizeai/openinference-instrumentation-openai |
NPM Version | |
| LangChain.js | @arizeai/openinference-instrumentation-langchain |
NPM Version | |
| Vercel AI SDK | @arizeai/openinference-vercel |
NPM Version | |
| BeeAI | @arizeai/openinference-instrumentation-beeai |
NPM Version | |
| Claude Agent SDK | @arizeai/openinference-instrumentation-claude-agent-sdk |
NPM Version | |
| Mastra | @mastra/arize |
NPM Version | |
| MCP | @arizeai/openinference-instrumentation-mcp |
NPM Version |
| Integration | Package | Version | |
|---|---|---|---|
| LangChain4j | openinference-instrumentation-langchain4j |
Maven Central | |
| SpringAI | openinference-instrumentation-springAI |
Maven Central | |
| Arconia | openinference-instrumentation-springAI |
Maven Central |
| Platform | Description | Docs | |
|---|---|---|---|
| BeeAI | AI agent framework with built-in observability | Integration Guide | |
| Dify | Open-source LLM app development platform | Integration Guide | |
| Envoy AI Gateway | AI Gateway built on Envoy Proxy for AI workloads | Integration Guide | |
| LangFlow | Visual framework for building multi-agent and RAG applications | Integration Guide | |
| LiteLLM Proxy | Proxy server for LLMs | Integration Guide | |
| Flowise | Visual framework for building LLM applications | Integration Guide | |
| Prompt Flow | Microsoft's prompt flow orchestration tool | Integration Guide | |
| NVIDIA NeMo | NVIDIA NeMo Agent Toolkit for enterprise agents | Integration Guide | |
| Graphite | Multi-agent LLM workflow framework with visual builder | Integration Guide |
This repository includes skills that teach coding agents how to work with Phoenix. They are located in .agents/skills/ and can be used with Claude Code, Cursor, and other compatible tools.
| Skill | Description |
|---|---|
| phoenix-cli | Debug LLM applications using the Phoenix CLI — fetch traces, analyze errors, review experiments, and query the GraphQL API |
| phoenix-evals | Build and run evaluators for AI/LLM applications using Phoenix |
| phoenix-tracing | OpenInference semantic conventions and instrumentation for tracing LLM applications |
We take data security and privacy very seriously. For more details, see our Security and Privacy documentation.
By default, Phoenix collects basic web analytics (e.g., page views, UI interactions) to help us understand how Phoenix is used and improve the product. None of your trace data, evaluation results, or any sensitive information is ever collected.
You can opt-out of telemetry by setting the environment variable: PHOENIX_TELEMETRY_ENABLED=false
Join our community to connect with thousands of AI builders.
See the migration guide for a list of breaking changes.
Copyright 2025 Arize AI, Inc. All Rights Reserved.
Portions of this code are patent protected by one or more U.S. Patents. See the IP_NOTICE.
This software is licensed under the terms of the Elastic License 2.0 (ELv2). See LICENSE.
Add this to claude_desktop_config.json and restart Claude Desktop.
{
"mcpServers": {
"phoenix-mcp": {
"command": "npx",
"args": [
"-y",
"@arizeai/phoenix-mcp"
]
}
}
}