Skip to content

Files

Latest commit

 

History

History

docs

description
AI Observability and Evaluation

Arize Phoenix

Phoenix is an open-source observability tool designed for experimentation, evaluation, and troubleshooting of AI and LLM applications. It allows AI engineers and data scientists to quickly visualize their data, evaluate performance, track down issues, and export data to improve.

Phoenix is built by Arize AI, the company behind the industry-leading AI observability platform, and a set of core contributors.

Phoenix works with OpenTelemetry and OpenInference instrumentation. See integrations-tracingfor details.

Features

{% tabs %} {% tab title="Prompt Engineering" %} {% embed url="https://storage.googleapis.com/arize-phoenix-assets/assets/gifs/prompt_playground.mp4" %} Phoenix Prompt Playground {% endembed %}

Phoenix offers tools to streamline your prompt engineering workflow.

  • Prompt Management - Create, store, modify, and deploy prompts for interacting with LLMs
  • Prompt Playground - Play with prompts, models, invocation parameters and track your progress via tracing and experiments
  • Span Replay - Replay the invocation of an LLM. Whether it's an LLM step in an LLM workflow or a router query, you can step into the LLM invocation and see if any modifications to the invocation would have yielded a better outcome.
  • Prompts in Code - Phoenix offers client SDKs to keep your prompts in sync across different applications and environments. {% endtab %}

{% tab title="Tracing" %} {% embed url="https://storage.googleapis.com/arize-phoenix-assets/assets/gifs/tracing.mp4" %} Tracing in Phoenix {% endembed %}

Tracing is a helpful tool for understanding how your LLM application works. Phoenix's open-source library offers comprehensive tracing capabilities that are not tied to any specific LLM vendor or framework.

Phoenix accepts traces over the OpenTelemetry protocol (OTLP) and supports first-class instrumentation for a variety of frameworks (LlamaIndex, LangChain, DSPy), SDKs (OpenAI, Bedrock, Mistral, Vertex), and Languages. (Python, Javascript, etc.) {% endtab %}

{% tab title="Evaluation" %} {% embed url="https://storage.googleapis.com/arize-phoenix-assets/assets/gifs/evals.mp4" %} Evals in the Phoenix UI {% endembed %}

Phoenix is built to help you evaluate your application and understand their true performance. To accomplish this, Phoenix includes:

{% tab title="Datasets & Experiments" %} {% embed url="https://storage.googleapis.com/arize-phoenix-assets/assets/gifs/experiments.mp4" %} Experiments in Phoenix {% endembed %}

Phoenix Datasets & Experiments let you test different versions of your application, store relevant traces for evaluation and analysis, and build robust evaluations into your development process.

Quickstarts

Running Phoenix for the first time? Select a quickstart below.

Tracingllm-traces-1.mdtracing-designed.png
Prompt Playgroundquickstart-promptsprompt-playground-designed.png
Datasets and Experimentsquickstart-datasets.mdexperiments_preview.png
Evaluationevals.mdevals-designed.png
Inferencesphoenix-inferences.mdScreenshot 2023-09-27 at 1.53.06 PM.png

Next Steps

Check out a comprehensive list of example notebooks for LLM Traces, Evals, RAG Analysis, and more.

Add instrumentation for popular packages and libraries such as OpenAI, LangGraph, Vercel AI SDK and more.

Join the Phoenix Slack community to ask questions, share findings, provide feedback, and connect with other developers.