[{"data":1,"prerenderedAt":-1},["ShallowReactive",2],{"tool-Arize-ai--phoenix":3,"similar-Arize-ai--phoenix":231},{"id":4,"github_repo":5,"name":6,"description_en":7,"description_zh":8,"ai_summary_zh":8,"readme_en":9,"readme_zh":10,"quickstart_zh":11,"use_case_zh":12,"hero_image_url":13,"owner_login":14,"owner_name":15,"owner_avatar_url":16,"owner_bio":17,"owner_company":18,"owner_location":18,"owner_email":19,"owner_twitter":20,"owner_website":21,"owner_url":22,"languages":23,"stars":59,"forks":60,"last_commit_at":61,"license":62,"difficulty_score":63,"env_os":64,"env_gpu":65,"env_ram":65,"env_deps":66,"category_tags":75,"github_topics":81,"view_count":98,"oss_zip_url":18,"oss_zip_packed_at":18,"status":99,"created_at":100,"updated_at":101,"faqs":102,"releases":130},8158,"Arize-ai\u002Fphoenix","phoenix","AI Observability & Evaluation","Phoenix 是一款专为人工智能应用打造的开源可观测性与评估平台。随着大模型和 RAG（检索增强生成）系统的普及，开发者常面临模型“黑盒”难题：难以追踪错误根源、无法量化回答质量或缺乏调试依据。Phoenix 正是为了解决这些痛点而生，它帮助团队深入洞察 AI 系统的内部运行状态，从追踪复杂的调用链到精准评估生成内容的准确性，让模型表现变得透明可控。\n\n这款工具主要面向 AI 工程师、数据科学家以及构建智能应用的开发团队。无论是正在微调大模型的算法研究员，还是致力于优化生产环境稳定性的后端开发者，都能通过 Phoenix 快速定位性能瓶颈或逻辑缺陷。其技术亮点在于对 OpenTelemetry 标准的原生支持，能够无缝集成主流 AI 框架，提供细粒度的分布式追踪能力；同时内置了先进的语义分析与可视化仪表盘，让用户无需编写复杂代码即可直观地分析模型行为、对比不同版本效果。通过 Phoenix，团队可以更高效地迭代模型，确保 AI 应用在真实场景中既聪明又可靠。","\u003Cp align=\"center\">\n    \u003Ca target=\"_blank\" href=\"https:\u002F\u002Fphoenix.arize.com\" style=\"background:none\">\n        \u003Cimg alt=\"phoenix banner\" src=\"https:\u002F\u002Foss.gittoolsai.com\u002Fimages\u002FArize-ai_phoenix_readme_f6d8660ba395.jpg\" width=\"auto\" height=\"auto\">\u003C\u002Fimg>\n    \u003C\u002Fa>\n    \u003Cbr\u002F>\n    \u003Cbr\u002F>\n    \u003Ca href=\"https:\u002F\u002Farize.com\u002Fdocs\u002Fphoenix\u002F\">\n        \u003Cimg src=\"https:\u002F\u002Fimg.shields.io\u002Fstatic\u002Fv1?message=Docs&logo=data:image\u002Fpng;base64,iVBORw0KGgoAAAANSUhEUgAAAIAAAACACAYAAADDPmHLAAAG4ElEQVR4nO2d4XHjNhCFcTf+b3ZgdWCmgmMqOKUC0xXYrsBOBVEqsFRB7ApCVRCygrMriFQBM7h5mNlwKBECARLg7jeDscamSQj7sFgsQfBL27ZK4MtXsT1vRADMEQEwRwTAHBEAc0QAzBEBMEcEwBwRAHNEAMwRATBnjAByFGE+MqVUMcYOY24GVUqpb\u002Fh8VErVKAf87QNFcEcbd4WSw+D6803njHscO5sATmGEURGBiCj6yUlv1uX2gv91FsDViArbcA2RUKF8QhAV8RQc0b15DcOt0VaTE1oAfWj3dYdCBfGGsmSM0XX5HsP3nEMAXbqCeCdiOERQPx9og5exGJ0S4zRQN9KrUupfpdQWjZciure\u002FYIj7K0bjqwTyAHdovA805iqCOg2xgnB1nZ97IvaoSCURdIPG\u002FIHGjTH\u002FYAz\u002FA8KdJai7lBQzgbpx\u002F0Hg6DT18UzWMXxSjMkDrElPNEmKfAbl6znwI3IMU\u002FOCa0\u002F1nfckwWaSbvWYYDnEsvCMJDNckhqu7GCMKWYOBXp9yPGd5kvqUAKf6rkAk7M2SY9QDXdEr9wEOr9x96EiejMFnixBNteDISsyNw7hHRqc22evWcP4vt39O85bzZH30AKg4+eo8cQRI4bHAJ7hyYM3CNHrG9RrimSXuZmUkZjN\u002FO6nAPpcwCcJNmipAle2QM\u002F1GU3vITCXhvY91u9geN\u002FjOY27VuTnYL1PCeAcRhwh7\u002FBl8Ai+IuxPiOCShtfX\u002FsPDtY8w+sZjby86dw6dBeoigD7obd\u002FKo6fI4BF8DA9HnGdrcU0fLt+n4dfE6H5jpjYcVdu2L23b5lpjHoo+18FDbcszddF1rUee\u002F4C6ZiO+80rHZmjDoIQUQLdRtm3brkcKIUPjjqVPBIUHgW1GGN4YfawAL2IqAVB8iEE31tvIelARlCPPVaFOLoIupzY6xVcM4MoRUyHXyHhslH6PaPl5RP1Lh4UsOeKR2e8dzC0Aiuvc2Nx3fwhfxf\u002FhknouUYbWUk5GTAIwmOh5e+H0cor8vEL91hfOdEqINLq1AV+RKImJ6869f9tFIBVc6y7gd3lHfWyNX0LEr7EuDElhRdAlQjig0e\u002FRU31xxDltM4pF7IY3pLIgxAhhgzF\u002FiC2M0Hi4dkOGlyGMd\u002Fg7dsMbUlsR9ICe9WhxbA3DjRkSdjiHzQzlBSKNJsCzIcUlYdfI0dcWS8LMkPDkcJ0n\u002FO+Qyy\u002FIAtDkSPnp4Fu4WpthQR\u002Fzm2VcoI\u002F51fI28iYld9\u002FHEh4Pf7D0Bm845pwIPnHMUJSf45pT5x68s5T9AW6INzhHDeP1BYcNMew5SghkinWOwVnaBhHGG5ybMn70zBDe8buh8X6DqV0Sa\u002F5tWOIOIbcWQ8KBiGBnMb\u002FP0OuTd\u002FlddCrY5jn\u002FVLm3nL+fY4X4YREuv8vS9wh6HSkAExMs0viKySZRd44iyOH2FzPe98Fll7A7GNMmjay4GF9BAKGXesfCN0sRsDG+YrhP4O2ACFgZXzHdKPL2RMJoxc34ivFOod3AMMNUj5XxFfOtYrUIXvB5MandS+G+V\u002FAzZ+MrEcBPlpoFtUIEwBwRAG+OIgDe1CIA5ogAmCMCYI4IgDkiAOaIAJgjAmCOCIA5IgDmiACYIwJgjgiAOSIA5ogAmCMCYI4IgDkiAOaIAJgjAmCOCIA5IgDmiACYIwJgjgiAOSIA5ogAmCMCYI4IgDkiAOaIAJgjAmDOVYBXvwvxQV8NWJOd0esvJ94babZaz7B5ovldxnlDpYhp0JFr\u002FKTlLKcEMMQKpcDPXIQxGXsYmhZnXAXQh\u002FEWBQrr3bc80mATyyrEvs4+BdBHgbdxFOIhrDkSg1\u002F6Iu2LCS0AyoqI4ftUF00EY\u002FQ3h1fRj2JKAVCMGErmnsH1lfnemEsAlByvgl0z2qx5B8OPCuB8EIMADBlEEOV79j1whNE3c\u002FX2PmISAGUNr7CEmUSUhjfEKgBDAY+QohCiNrwhdgEYzPv7UxkadvBg0RrekMrNoAozh3vLN4DPhc7S\u002FWL52vkoSO1u4BZC+DOCulC0KJ\u002FgqWaP7C8hlSGgjxyCmDuPsEePT\u002FKuasrrAcyr4H+f6fq01yd7Sz1lD0CZ2hs06PVJufs+lrIiyLwufjfBtXYpjvWnWIoHoJSYe4dIK\u002Ft4HX1ULFEACkPCm8e8wXFJvZ6y1EWhJkDcWxw7RINzLc74auGrgg8e4oIm9Sh\u002FCA7LwkvHqaIJ9pLI6Lmy1BigDy2EV8tjdzh+8XB6MGSLKH4INsZXDJ8MGhIBK+Mrpo+GnRIBO+MrZjFAFxoTNBwCvj6u4qvSZJiM3iNX4yvmHoA9Sh4PF0QAzBEBMEcEwBwRAHNEAMwRAXBGKfUfr5hKvglRfO4AAAAASUVORK5CYII=&labelColor=grey&color=blue&logoColor=white&label=%20\"\u002F>\n    \u003C\u002Fa>\n    \u003Ca target=\"_blank\" href=\"https:\u002F\u002Fjoin.slack.com\u002Ft\u002Farize-ai\u002Fshared_invite\u002Fzt-3r07iavnk-ammtATWSlF0pSrd1DsMW7g\">\n        \u003Cimg src=\"https:\u002F\u002Fimg.shields.io\u002Fstatic\u002Fv1?message=Community&logo=slack&labelColor=grey&color=blue&logoColor=white&label=%20\"\u002F>\n    \u003C\u002Fa>\n     \u003Ca target=\"_blank\" href=\"https:\u002F\u002Fbsky.app\u002Fprofile\u002Farize-phoenix.bsky.social\">\n        \u003Cimg src=\"https:\u002F\u002Fimg.shields.io\u002Fbadge\u002F-phoenix-blue.svg?color=blue&labelColor=gray&logo=bluesky\">\n    \u003C\u002Fa>\n    \u003Ca target=\"_blank\" href=\"https:\u002F\u002Fx.com\u002FArizePhoenix\">\n        \u003Cimg src=\"https:\u002F\u002Fimg.shields.io\u002Fbadge\u002F-ArizePhoenix-blue.svg?color=blue&labelColor=gray&logo=x\">\n    \u003C\u002Fa>\n    \u003Ca target=\"_blank\" href=\"https:\u002F\u002Fpypi.org\u002Fproject\u002Farize-phoenix\u002F\">\n        \u003Cimg src=\"https:\u002F\u002Fimg.shields.io\u002Fpypi\u002Fv\u002Farize-phoenix?color=blue\">\n    \u003C\u002Fa>\n    \u003Ca target=\"_blank\" href=\"https:\u002F\u002Fanaconda.org\u002Fconda-forge\u002Farize-phoenix\">\n        \u003Cimg src=\"https:\u002F\u002Fimg.shields.io\u002Fconda\u002Fvn\u002Fconda-forge\u002Farize-phoenix.svg?color=blue\">\n    \u003C\u002Fa>\n    \u003Ca target=\"_blank\" href=\"https:\u002F\u002Fpypi.org\u002Fproject\u002Farize-phoenix\u002F\">\n        \u003Cimg src=\"https:\u002F\u002Fimg.shields.io\u002Fpypi\u002Fpyversions\u002Farize-phoenix\">\n    \u003C\u002Fa>\n    \u003Ca target=\"_blank\" href=\"https:\u002F\u002Fhub.docker.com\u002Fr\u002Farizephoenix\u002Fphoenix\u002Ftags\">\n        \u003Cimg src=\"https:\u002F\u002Fimg.shields.io\u002Fdocker\u002Fv\u002Farizephoenix\u002Fphoenix?sort=semver&logo=docker&label=image&color=blue\">\n    \u003C\u002Fa>\n    \u003Ca target=\"_blank\" href=\"https:\u002F\u002Fhub.docker.com\u002Fr\u002Farizephoenix\u002Fphoenix-helm\">\n        \u003Cimg src=\"https:\u002F\u002Fimg.shields.io\u002Fbadge\u002FHelm-blue?style=flat&logo=helm&labelColor=grey\"\u002F>\n    \u003C\u002Fa>\n    \u003Ca target=\"_blank\" href=\"https:\u002F\u002Fgithub.com\u002FArize-ai\u002Fphoenix\u002Ftree\u002Fmain\u002Fjs\u002Fpackages\u002Fphoenix-mcp\">\n        \u003Cimg src=\"https:\u002F\u002Foss.gittoolsai.com\u002Fimages\u002FArize-ai_phoenix_readme_d4184d72d7bf.png\" title=\"MCP Enabled\"\u002F>\n    \u003C\u002Fa>\n    \u003Ca href=\"cursor:\u002F\u002Fanysphere.cursor-deeplink\u002Fmcp\u002Finstall?name=phoenix&config=eyJjb21tYW5kIjoibnB4IC15IEBhcml6ZWFpL3Bob2VuaXgtbWNwQGxhdGVzdCAtLWJhc2VVcmwgaHR0cHM6Ly9teS1waG9lbml4LmNvbSAtLWFwaUtleSB5b3VyLWFwaS1rZXkifQ%3D%3D\">\u003Cimg src=\"https:\u002F\u002Fcursor.com\u002Fdeeplink\u002Fmcp-install-dark.svg\" alt=\"Add Arize Phoenix MCP server to Cursor\" height=20 \u002F>\u003C\u002Fa>\n    \u003Cimg referrerpolicy=\"no-referrer-when-downgrade\" src=\"https:\u002F\u002Fstatic.scarf.sh\u002Fa.png?x-pxid=8e8e8b34-7900-43fa-a38f-1f070bd48c64&page=README.md\" \u002F>\n\u003C\u002Fp>\n\nPhoenix is an open-source AI observability platform designed for experimentation, evaluation, and troubleshooting. It provides:\n\n- [**_Tracing_**](https:\u002F\u002Farize.com\u002Fdocs\u002Fphoenix\u002Ftracing\u002Fllm-traces) - Trace your LLM application's runtime using OpenTelemetry-based instrumentation.\n- [**_Evaluation_**](https:\u002F\u002Farize.com\u002Fdocs\u002Fphoenix\u002Fevaluation\u002Fllm-evals) - Leverage LLMs to benchmark your application's performance using response and retrieval evals.\n- [**_Datasets_**](https:\u002F\u002Farize.com\u002Fdocs\u002Fphoenix\u002Fdatasets-and-experiments\u002Foverview-datasets) - Create versioned datasets of examples for experimentation, evaluation, and fine-tuning.\n- [**_Experiments_**](https:\u002F\u002Farize.com\u002Fdocs\u002Fphoenix\u002Fdatasets-and-experiments\u002Foverview-datasets#experiments) - Track and evaluate changes to prompts, LLMs, and retrieval.\n- [**_Playground_**](https:\u002F\u002Farize.com\u002Fdocs\u002Fphoenix\u002Fprompt-engineering\u002Foverview-prompts)- Optimize prompts, compare models, adjust parameters, and replay traced LLM calls.\n- [**_Prompt Management_**](https:\u002F\u002Farize.com\u002Fdocs\u002Fphoenix\u002Fprompt-engineering\u002Foverview-prompts\u002Fprompt-management)- Manage and test prompt changes systematically using version control, tagging, and experimentation.\n\nPhoenix is vendor and language agnostic with out-of-the-box support for popular frameworks ([OpenAI Agents SDK](https:\u002F\u002Farize.com\u002Fdocs\u002Fphoenix\u002Ftracing\u002Fintegrations-tracing\u002Fopenai-agents-sdk), [Claude Agent SDK](https:\u002F\u002Farize.com\u002Fdocs\u002Fphoenix\u002Fintegrations\u002Fpython\u002Fclaude-agent-sdk), [LangGraph](https:\u002F\u002Farize.com\u002Fdocs\u002Fphoenix\u002Ftracing\u002Fintegrations-tracing\u002Flangchain), [Vercel AI SDK](https:\u002F\u002Farize.com\u002Fdocs\u002Fphoenix\u002Ftracing\u002Fintegrations-tracing\u002Fvercel-ai-sdk), [Mastra](https:\u002F\u002Farize.com\u002Fdocs\u002Fphoenix\u002Fintegrations\u002Ftypescript\u002Fmastra), [CrewAI](https:\u002F\u002Farize.com\u002Fdocs\u002Fphoenix\u002Ftracing\u002Fintegrations-tracing\u002Fcrewai), [LlamaIndex](https:\u002F\u002Farize.com\u002Fdocs\u002Fphoenix\u002Ftracing\u002Fintegrations-tracing\u002Fllamaindex), [DSPy](https:\u002F\u002Farize.com\u002Fdocs\u002Fphoenix\u002Ftracing\u002Fintegrations-tracing\u002Fdspy)) and LLM providers ([OpenAI](https:\u002F\u002Farize.com\u002Fdocs\u002Fphoenix\u002Ftracing\u002Fintegrations-tracing\u002Fopenai), [Anthropic](https:\u002F\u002Farize.com\u002Fdocs\u002Fphoenix\u002Ftracing\u002Fintegrations-tracing\u002Fanthropic), [Google GenAI](https:\u002F\u002Farize.com\u002Fdocs\u002Fphoenix\u002Ftracing\u002Fintegrations-tracing\u002Fgoogle-genai), [Google ADK](https:\u002F\u002Farize.com\u002Fdocs\u002Fphoenix\u002Fintegrations\u002Fllm-providers\u002Fgoogle-gen-ai\u002Fgoogle-adk-tracing), [AWS Bedrock](https:\u002F\u002Farize.com\u002Fdocs\u002Fphoenix\u002Ftracing\u002Fintegrations-tracing\u002Fbedrock), [OpenRouter](https:\u002F\u002Farize.com\u002Fdocs\u002Fphoenix\u002Fintegrations\u002Fpython\u002Fopenrouter), [LiteLLM](https:\u002F\u002Farize.com\u002Fdocs\u002Fphoenix\u002Ftracing\u002Fintegrations-tracing\u002Flitellm), and more). For details on auto-instrumentation, check out the [OpenInference](https:\u002F\u002Fgithub.com\u002FArize-ai\u002Fopeninference) project.\n\nPhoenix runs practically anywhere, including your local machine, a Jupyter notebook, a containerized deployment, or in the cloud.\n\n## Installation\n\nInstall Phoenix via `pip` or `conda`\n\n```shell\npip install arize-phoenix\n```\n\nPhoenix container images are available via [Docker Hub](https:\u002F\u002Fhub.docker.com\u002Fr\u002Farizephoenix\u002Fphoenix) and can be deployed using Docker or Kubernetes. Arize AI also provides cloud instances at [app.phoenix.arize.com](https:\u002F\u002Fapp.phoenix.arize.com\u002F).\n\n## Packages\n\nThe `arize-phoenix` package includes the entire Phoenix platform. However, if you have deployed the Phoenix platform, there are lightweight Python sub-packages and TypeScript packages that can be used in conjunction with the platform.\n\n### Python Subpackages\n\n| Package                                                                                       | Version & Docs                                                                                                                                                                                                                                                                      | Description                                                                                |\n| --------------------------------------------------------------------------------------------- | ----------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------- | ------------------------------------------------------------------------------------------ |\n| [arize-phoenix-otel](https:\u002F\u002Fgithub.com\u002FArize-ai\u002Fphoenix\u002Ftree\u002Fmain\u002Fpackages\u002Fphoenix-otel)     | [![PyPI Version](https:\u002F\u002Fimg.shields.io\u002Fpypi\u002Fv\u002Farize-phoenix-otel)](https:\u002F\u002Fpypi.org\u002Fproject\u002Farize-phoenix-otel\u002F) [![Docs](https:\u002F\u002Fimg.shields.io\u002Fbadge\u002Fdocs-blue?logo=readthedocs&logoColor=white)](https:\u002F\u002Farize-phoenix.readthedocs.io\u002Fprojects\u002Fotel\u002Fen\u002Flatest\u002Findex.html)       | Provides a lightweight wrapper around OpenTelemetry primitives with Phoenix-aware defaults |\n| [arize-phoenix-client](https:\u002F\u002Fgithub.com\u002FArize-ai\u002Fphoenix\u002Ftree\u002Fmain\u002Fpackages\u002Fphoenix-client) | [![PyPI Version](https:\u002F\u002Fimg.shields.io\u002Fpypi\u002Fv\u002Farize-phoenix-client)](https:\u002F\u002Fpypi.org\u002Fproject\u002Farize-phoenix-client\u002F) [![Docs](https:\u002F\u002Fimg.shields.io\u002Fbadge\u002Fdocs-blue?logo=readthedocs&logoColor=white)](https:\u002F\u002Farize-phoenix.readthedocs.io\u002Fprojects\u002Fclient\u002Fen\u002Flatest\u002Findex.html) | Lightweight client for interacting with the Phoenix server via its OpenAPI REST interface  |\n| [arize-phoenix-evals](https:\u002F\u002Fgithub.com\u002FArize-ai\u002Fphoenix\u002Ftree\u002Fmain\u002Fpackages\u002Fphoenix-evals)   | [![PyPI Version](https:\u002F\u002Fimg.shields.io\u002Fpypi\u002Fv\u002Farize-phoenix-evals)](https:\u002F\u002Fpypi.org\u002Fproject\u002Farize-phoenix-evals\u002F) [![Docs](https:\u002F\u002Fimg.shields.io\u002Fbadge\u002Fdocs-blue?logo=readthedocs&logoColor=white)](https:\u002F\u002Farize-phoenix.readthedocs.io\u002Fprojects\u002Fevals\u002Fen\u002Flatest\u002Findex.html)    | Tooling to evaluate LLM applications including RAG relevance, answer relevance, and more   |\n\n### TypeScript Subpackages\n\n| Package                                                                                             | Version & Docs                                                                                                                                                                                                                                                                                 | Description                                                                                                  |\n| --------------------------------------------------------------------------------------------------- | ---------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------- | ------------------------------------------------------------------------------------------------------------ |\n| [@arizeai\u002Fphoenix-otel](https:\u002F\u002Fgithub.com\u002FArize-ai\u002Fphoenix\u002Ftree\u002Fmain\u002Fjs\u002Fpackages\u002Fphoenix-otel)     | [![NPM Version](https:\u002F\u002Fimg.shields.io\u002Fnpm\u002Fv\u002F%40arizeai%2Fphoenix-otel)](https:\u002F\u002Fwww.npmjs.com\u002Fpackage\u002F@arizeai\u002Fphoenix-otel) [![Docs](https:\u002F\u002Fimg.shields.io\u002Fbadge\u002Fdocs-blue?logo=typescript&logoColor=white)](https:\u002F\u002Farize-ai.github.io\u002Fphoenix\u002F)                                           | Provides a lightweight wrapper around OpenTelemetry primitives with Phoenix-aware defaults                   |\n| [@arizeai\u002Fphoenix-client](https:\u002F\u002Fgithub.com\u002FArize-ai\u002Fphoenix\u002Ftree\u002Fmain\u002Fjs\u002Fpackages\u002Fphoenix-client) | [![NPM Version](https:\u002F\u002Fimg.shields.io\u002Fnpm\u002Fv\u002F%40arizeai%2Fphoenix-client)](https:\u002F\u002Fwww.npmjs.com\u002Fpackage\u002F@arizeai\u002Fphoenix-client) [![Docs](https:\u002F\u002Fimg.shields.io\u002Fbadge\u002Fdocs-blue?logo=typescript&logoColor=white)](https:\u002F\u002Farize-ai.github.io\u002Fphoenix\u002F)                                       | Client for the Arize Phoenix API                                                                             |\n| [@arizeai\u002Fphoenix-evals](https:\u002F\u002Fgithub.com\u002FArize-ai\u002Fphoenix\u002Ftree\u002Fmain\u002Fjs\u002Fpackages\u002Fphoenix-evals)   | [![NPM Version](https:\u002F\u002Fimg.shields.io\u002Fnpm\u002Fv\u002F%40arizeai%2Fphoenix-evals)](https:\u002F\u002Fwww.npmjs.com\u002Fpackage\u002F@arizeai\u002Fphoenix-evals) [![Docs](https:\u002F\u002Fimg.shields.io\u002Fbadge\u002Fdocs-blue?logo=typescript&logoColor=white)](https:\u002F\u002Farize-ai.github.io\u002Fphoenix\u002F)                                         | TypeScript evaluation library for LLM applications (alpha release)                                           |\n| [@arizeai\u002Fphoenix-mcp](https:\u002F\u002Fgithub.com\u002FArize-ai\u002Fphoenix\u002Ftree\u002Fmain\u002Fjs\u002Fpackages\u002Fphoenix-mcp)       | [![NPM Version](https:\u002F\u002Fimg.shields.io\u002Fnpm\u002Fv\u002F%40arizeai%2Fphoenix-mcp)](https:\u002F\u002Fwww.npmjs.com\u002Fpackage\u002F@arizeai\u002Fphoenix-mcp) [![Docs](https:\u002F\u002Fimg.shields.io\u002Fbadge\u002Fdocs-blue?logo=markdown&logoColor=white)](.\u002Fjs\u002Fpackages\u002Fphoenix-mcp\u002FREADME.md)                                               | MCP server implementation for Arize Phoenix providing unified interface to Phoenix's capabilities            |\n| [@arizeai\u002Fphoenix-cli](https:\u002F\u002Fgithub.com\u002FArize-ai\u002Fphoenix\u002Ftree\u002Fmain\u002Fjs\u002Fpackages\u002Fphoenix-cli)       | [![NPM Version](https:\u002F\u002Fimg.shields.io\u002Fnpm\u002Fv\u002F%40arizeai%2Fphoenix-cli)](https:\u002F\u002Fwww.npmjs.com\u002Fpackage\u002F@arizeai\u002Fphoenix-cli) [![Docs](https:\u002F\u002Fimg.shields.io\u002Fbadge\u002Fdocs-blue?logo=typescript&logoColor=white)](https:\u002F\u002Farize.com\u002Fdocs\u002Fphoenix\u002Fsdk-api-reference\u002Ftypescript\u002Farizeai-phoenix-cli) | CLI for fetching traces, datasets, and experiments for use with Claude Code, Cursor, and other coding agents |\n\n## Tracing Integrations\n\nPhoenix is built on top of OpenTelemetry and is vendor, language, and framework agnostic. For details about tracing integrations and example applications, see the [OpenInference](https:\u002F\u002Fgithub.com\u002FArize-ai\u002Fopeninference) project.\n\n**Python Integrations**\n| | Integration | Package | Version |\n|:---:|---|---|---|\n| \u003Cpicture>\u003Csource media=\"(prefers-color-scheme: dark)\" srcset=\"https:\u002F\u002Funpkg.com\u002F@lobehub\u002Ficons-static-png@latest\u002Fdark\u002Fopenai.png\">\u003Cimg height=\"14\" src=\"https:\u002F\u002Foss.gittoolsai.com\u002Fimages\u002FArize-ai_phoenix_readme_fc1845c68da0.png\">\u003C\u002Fpicture> | [OpenAI](https:\u002F\u002Farize.com\u002Fdocs\u002Fphoenix\u002Ftracing\u002Fintegrations-tracing\u002Fopenai) | `openinference-instrumentation-openai` | [![PyPI Version](https:\u002F\u002Fimg.shields.io\u002Fpypi\u002Fv\u002Fopeninference-instrumentation-openai.svg)](https:\u002F\u002Fpypi.python.org\u002Fpypi\u002Fopeninference-instrumentation-openai) |\n| \u003Cpicture>\u003Csource media=\"(prefers-color-scheme: dark)\" srcset=\"https:\u002F\u002Funpkg.com\u002F@lobehub\u002Ficons-static-png@latest\u002Fdark\u002Fopenai.png\">\u003Cimg height=\"14\" src=\"https:\u002F\u002Foss.gittoolsai.com\u002Fimages\u002FArize-ai_phoenix_readme_fc1845c68da0.png\">\u003C\u002Fpicture> | [OpenAI Agents](https:\u002F\u002Farize.com\u002Fdocs\u002Fphoenix\u002Ftracing\u002Fintegrations-tracing\u002Fopenai-agents-sdk) | `openinference-instrumentation-openai-agents` | [![PyPI Version](https:\u002F\u002Fimg.shields.io\u002Fpypi\u002Fv\u002Fopeninference-instrumentation-openai-agents.svg)](https:\u002F\u002Fpypi.python.org\u002Fpypi\u002Fopeninference-instrumentation-openai-agents) |\n| \u003Cimg src=\"https:\u002F\u002Foss.gittoolsai.com\u002Fimages\u002FArize-ai_phoenix_readme_370c5501b150.png\" height=\"14\"> | [LlamaIndex](https:\u002F\u002Farize.com\u002Fdocs\u002Fphoenix\u002Ftracing\u002Fintegrations-tracing\u002Fllamaindex) | `openinference-instrumentation-llama-index` | [![PyPI Version](https:\u002F\u002Fimg.shields.io\u002Fpypi\u002Fv\u002Fopeninference-instrumentation-llama-index.svg)](https:\u002F\u002Fpypi.python.org\u002Fpypi\u002Fopeninference-instrumentation-llama-index) |\n| | [DSPy](https:\u002F\u002Farize.com\u002Fdocs\u002Fphoenix\u002Ftracing\u002Fintegrations-tracing\u002Fdspy) | `openinference-instrumentation-dspy` | [![PyPI Version](https:\u002F\u002Fimg.shields.io\u002Fpypi\u002Fv\u002Fopeninference-instrumentation-dspy.svg)](https:\u002F\u002Fpypi.python.org\u002Fpypi\u002Fopeninference-instrumentation-dspy) |\n| \u003Cimg src=\"https:\u002F\u002Foss.gittoolsai.com\u002Fimages\u002FArize-ai_phoenix_readme_7d660dac1f19.png\" height=\"14\"> | [AWS Bedrock](https:\u002F\u002Farize.com\u002Fdocs\u002Fphoenix\u002Ftracing\u002Fintegrations-tracing\u002Fbedrock) | `openinference-instrumentation-bedrock` | [![PyPI Version](https:\u002F\u002Fimg.shields.io\u002Fpypi\u002Fv\u002Fopeninference-instrumentation-bedrock.svg)](https:\u002F\u002Fpypi.python.org\u002Fpypi\u002Fopeninference-instrumentation-bedrock) |\n| \u003Cimg src=\"https:\u002F\u002Foss.gittoolsai.com\u002Fimages\u002FArize-ai_phoenix_readme_acb5f98fc60a.png\" height=\"14\"> | [LangChain](https:\u002F\u002Farize.com\u002Fdocs\u002Fphoenix\u002Ftracing\u002Fintegrations-tracing\u002Flangchain) | `openinference-instrumentation-langchain` | [![PyPI Version](https:\u002F\u002Fimg.shields.io\u002Fpypi\u002Fv\u002Fopeninference-instrumentation-langchain.svg)](https:\u002F\u002Fpypi.python.org\u002Fpypi\u002Fopeninference-instrumentation-langchain) |\n| \u003Cimg src=\"https:\u002F\u002Foss.gittoolsai.com\u002Fimages\u002FArize-ai_phoenix_readme_49ae725664be.png\" height=\"14\"> | [MistralAI](https:\u002F\u002Farize.com\u002Fdocs\u002Fphoenix\u002Ftracing\u002Fintegrations-tracing\u002Fmistralai) | `openinference-instrumentation-mistralai` | [![PyPI Version](https:\u002F\u002Fimg.shields.io\u002Fpypi\u002Fv\u002Fopeninference-instrumentation-mistralai.svg)](https:\u002F\u002Fpypi.python.org\u002Fpypi\u002Fopeninference-instrumentation-mistralai) |\n| \u003Cimg src=\"https:\u002F\u002Foss.gittoolsai.com\u002Fimages\u002FArize-ai_phoenix_readme_8d65b0abf33e.png\" height=\"14\"> | [Google GenAI](https:\u002F\u002Farize.com\u002Fdocs\u002Fphoenix\u002Ftracing\u002Fintegrations-tracing\u002Fgoogle-gen-ai) | `openinference-instrumentation-google-genai` | [![PyPI Version](https:\u002F\u002Fimg.shields.io\u002Fpypi\u002Fv\u002Fopeninference-instrumentation-google-genai.svg)](https:\u002F\u002Fpypi.python.org\u002Fpypi\u002Fopeninference-instrumentation-google-genai) |\n| \u003Cimg src=\"https:\u002F\u002Foss.gittoolsai.com\u002Fimages\u002FArize-ai_phoenix_readme_8d65b0abf33e.png\" height=\"14\"> | [Google ADK](https:\u002F\u002Farize.com\u002Fdocs\u002Fphoenix\u002Fintegrations\u002Fllm-providers\u002Fgoogle-gen-ai\u002Fgoogle-adk-tracing) | `openinference-instrumentation-google-adk` | [![PyPI Version](https:\u002F\u002Fimg.shields.io\u002Fpypi\u002Fv\u002Fopeninference-instrumentation-google-adk.svg)](https:\u002F\u002Fpypi.python.org\u002Fpypi\u002Fopeninference-instrumentation-google-adk) |\n| | [Guardrails](https:\u002F\u002Farize.com\u002Fdocs\u002Fphoenix\u002Ftracing\u002Fintegrations-tracing\u002Fguardrails) | `openinference-instrumentation-guardrails` | [![PyPI Version](https:\u002F\u002Fimg.shields.io\u002Fpypi\u002Fv\u002Fopeninference-instrumentation-guardrails.svg)](https:\u002F\u002Fpypi.python.org\u002Fpypi\u002Fopeninference-instrumentation-guardrails) |\n| \u003Cimg src=\"https:\u002F\u002Foss.gittoolsai.com\u002Fimages\u002FArize-ai_phoenix_readme_e8c58b26f28a.png\" height=\"14\"> | [VertexAI](https:\u002F\u002Farize.com\u002Fdocs\u002Fphoenix\u002Ftracing\u002Fintegrations-tracing\u002Fvertexai) | `openinference-instrumentation-vertexai` | [![PyPI Version](https:\u002F\u002Fimg.shields.io\u002Fpypi\u002Fv\u002Fopeninference-instrumentation-vertexai.svg)](https:\u002F\u002Fpypi.python.org\u002Fpypi\u002Fopeninference-instrumentation-vertexai) |\n| \u003Cimg src=\"https:\u002F\u002Foss.gittoolsai.com\u002Fimages\u002FArize-ai_phoenix_readme_6e375164f316.png\" height=\"14\"> | [CrewAI](https:\u002F\u002Farize.com\u002Fdocs\u002Fphoenix\u002Ftracing\u002Fintegrations-tracing\u002Fcrewai) | `openinference-instrumentation-crewai` | [![PyPI Version](https:\u002F\u002Fimg.shields.io\u002Fpypi\u002Fv\u002Fopeninference-instrumentation-crewai.svg)](https:\u002F\u002Fpypi.python.org\u002Fpypi\u002Fopeninference-instrumentation-crewai) |\n| | [Haystack](https:\u002F\u002Farize.com\u002Fdocs\u002Fphoenix\u002Ftracing\u002Fintegrations-tracing\u002Fhaystack) | `openinference-instrumentation-haystack` | [![PyPI Version](https:\u002F\u002Fimg.shields.io\u002Fpypi\u002Fv\u002Fopeninference-instrumentation-haystack.svg)](https:\u002F\u002Fpypi.python.org\u002Fpypi\u002Fopeninference-instrumentation-haystack) |\n| | [LiteLLM](https:\u002F\u002Farize.com\u002Fdocs\u002Fphoenix\u002Ftracing\u002Fintegrations-tracing\u002Flitellm) | `openinference-instrumentation-litellm` | [![PyPI Version](https:\u002F\u002Fimg.shields.io\u002Fpypi\u002Fv\u002Fopeninference-instrumentation-litellm.svg)](https:\u002F\u002Fpypi.python.org\u002Fpypi\u002Fopeninference-instrumentation-litellm) |\n| \u003Cpicture>\u003Csource media=\"(prefers-color-scheme: dark)\" srcset=\"https:\u002F\u002Funpkg.com\u002F@lobehub\u002Ficons-static-png@latest\u002Fdark\u002Fgroq.png\">\u003Cimg height=\"14\" src=\"https:\u002F\u002Foss.gittoolsai.com\u002Fimages\u002FArize-ai_phoenix_readme_8964c766bdfe.png\">\u003C\u002Fpicture> | [Groq](https:\u002F\u002Farize.com\u002Fdocs\u002Fphoenix\u002Ftracing\u002Fintegrations-tracing\u002Fgroq) | `openinference-instrumentation-groq` | [![PyPI Version](https:\u002F\u002Fimg.shields.io\u002Fpypi\u002Fv\u002Fopeninference-instrumentation-groq.svg)](https:\u002F\u002Fpypi.python.org\u002Fpypi\u002Fopeninference-instrumentation-groq) |\n| | [Instructor](https:\u002F\u002Farize.com\u002Fdocs\u002Fphoenix\u002Ftracing\u002Fintegrations-tracing\u002Finstructor) | `openinference-instrumentation-instructor` | [![PyPI Version](https:\u002F\u002Fimg.shields.io\u002Fpypi\u002Fv\u002Fopeninference-instrumentation-instructor.svg)](https:\u002F\u002Fpypi.python.org\u002Fpypi\u002Fopeninference-instrumentation-instructor) |\n| \u003Cpicture>\u003Csource media=\"(prefers-color-scheme: dark)\" srcset=\"https:\u002F\u002Funpkg.com\u002F@lobehub\u002Ficons-static-png@latest\u002Fdark\u002Fanthropic.png\">\u003Cimg height=\"14\" src=\"https:\u002F\u002Foss.gittoolsai.com\u002Fimages\u002FArize-ai_phoenix_readme_5d80b94217f0.png\">\u003C\u002Fpicture> | [Anthropic](https:\u002F\u002Farize.com\u002Fdocs\u002Fphoenix\u002Ftracing\u002Fintegrations-tracing\u002Fanthropic) | `openinference-instrumentation-anthropic` | [![PyPI Version](https:\u002F\u002Fimg.shields.io\u002Fpypi\u002Fv\u002Fopeninference-instrumentation-anthropic.svg)](https:\u002F\u002Fpypi.python.org\u002Fpypi\u002Fopeninference-instrumentation-anthropic) |\n| \u003Cimg src=\"https:\u002F\u002Foss.gittoolsai.com\u002Fimages\u002FArize-ai_phoenix_readme_d4b49bf027f0.png\" height=\"14\"> | [Smolagents](https:\u002F\u002Fhuggingface.co\u002Fdocs\u002Fsmolagents\u002Fen\u002Ftutorials\u002Finspect_runs) | `openinference-instrumentation-smolagents` | [![PyPI Version](https:\u002F\u002Fimg.shields.io\u002Fpypi\u002Fv\u002Fopeninference-instrumentation-smolagents.svg)](https:\u002F\u002Fpypi.python.org\u002Fpypi\u002Fopeninference-instrumentation-smolagents) |\n| | [Agno](https:\u002F\u002Farize.com\u002Fdocs\u002Fphoenix\u002Ftracing\u002Fintegrations-tracing\u002Fagno) | `openinference-instrumentation-agno` | [![PyPI Version](https:\u002F\u002Fimg.shields.io\u002Fpypi\u002Fv\u002Fopeninference-instrumentation-agno.svg)](https:\u002F\u002Fpypi.python.org\u002Fpypi\u002Fopeninference-instrumentation-agno) |\n| \u003Cpicture>\u003Csource media=\"(prefers-color-scheme: dark)\" srcset=\"https:\u002F\u002Funpkg.com\u002F@lobehub\u002Ficons-static-png@latest\u002Fdark\u002Fmcp.png\">\u003Cimg height=\"14\" src=\"https:\u002F\u002Foss.gittoolsai.com\u002Fimages\u002FArize-ai_phoenix_readme_53956c4cd629.png\">\u003C\u002Fpicture> | [MCP](https:\u002F\u002Farize.com\u002Fdocs\u002Fphoenix\u002Ftracing\u002Fintegrations-tracing\u002Fmodel-context-protocol-mcp) | `openinference-instrumentation-mcp` | [![PyPI Version](https:\u002F\u002Fimg.shields.io\u002Fpypi\u002Fv\u002Fopeninference-instrumentation-mcp.svg)](https:\u002F\u002Fpypi.python.org\u002Fpypi\u002Fopeninference-instrumentation-mcp) |\n| \u003Cimg src=\"https:\u002F\u002Foss.gittoolsai.com\u002Fimages\u002FArize-ai_phoenix_readme_f900f128bcfe.png\" height=\"14\"> | [Pydantic AI](https:\u002F\u002Farize.com\u002Fdocs\u002Fphoenix\u002Fintegrations\u002Fpython\u002Fpydantic) | `openinference-instrumentation-pydantic-ai` | [![PyPI Version](https:\u002F\u002Fimg.shields.io\u002Fpypi\u002Fv\u002Fopeninference-instrumentation-pydantic-ai.svg)](https:\u002F\u002Fpypi.python.org\u002Fpypi\u002Fopeninference-instrumentation-pydantic-ai) |\n| | [Autogen AgentChat](https:\u002F\u002Farize.com\u002Fdocs\u002Fphoenix\u002Fintegrations\u002Fframeworks\u002Fautogen\u002Fautogen-tracing) | `openinference-instrumentation-autogen-agentchat` | [![PyPI Version](https:\u002F\u002Fimg.shields.io\u002Fpypi\u002Fv\u002Fopeninference-instrumentation-autogen-agentchat.svg)](https:\u002F\u002Fpypi.python.org\u002Fpypi\u002Fopeninference-instrumentation-autogen-agentchat) |\n| | [Portkey](https:\u002F\u002Farize.com\u002Fdocs\u002Fphoenix\u002Fintegrations\u002Fportkey) | `openinference-instrumentation-portkey` | [![PyPI Version](https:\u002F\u002Fimg.shields.io\u002Fpypi\u002Fv\u002Fopeninference-instrumentation-portkey.svg)](https:\u002F\u002Fpypi.python.org\u002Fpypi\u002Fopeninference-instrumentation-portkey) |\n| | [Agent Spec](https:\u002F\u002Farize.com\u002Fdocs\u002Fphoenix\u002Ftracing\u002Fintegrations-tracing\u002Fagentspec) | `openinference-instrumentation-agentspec` | [![PyPI Version](https:\u002F\u002Fimg.shields.io\u002Fpypi\u002Fv\u002Fopeninference-instrumentation-agentspec.svg)](https:\u002F\u002Fpypi.python.org\u002Fpypi\u002Fopeninference-instrumentation-agentspec) |\n| \u003Cimg src=\"https:\u002F\u002Foss.gittoolsai.com\u002Fimages\u002FArize-ai_phoenix_readme_6ef465e6e627.png\" height=\"14\"> | [Claude Agent SDK](https:\u002F\u002Farize.com\u002Fdocs\u002Fphoenix\u002Fintegrations\u002Fpython\u002Fclaude-agent-sdk) | `openinference-instrumentation-claude-agent-sdk` | [![PyPI Version](https:\u002F\u002Fimg.shields.io\u002Fpypi\u002Fv\u002Fopeninference-instrumentation-claude-agent-sdk.svg)](https:\u002F\u002Fpypi.python.org\u002Fpypi\u002Fopeninference-instrumentation-claude-agent-sdk) |\n\n## Span Processors\n\nNormalize and convert data across other instrumentation libraries by adding span processors that unify data.\n\n| Package                                                                                                           | Description                                                      | Version                                                                                                                                                                |\n| ----------------------------------------------------------------------------------------------------------------- | ---------------------------------------------------------------- | ---------------------------------------------------------------------------------------------------------------------------------------------------------------------- |\n| [`openinference-instrumentation-openlit`](.\u002Fpython\u002Finstrumentation\u002Fopeninference-instrumentation-openlit)         | OpenInference Span Processor for OpenLIT traces.                 | [![PyPI Version](https:\u002F\u002Fimg.shields.io\u002Fpypi\u002Fv\u002Fopeninference-instrumentation-openlit.svg)](https:\u002F\u002Fpypi.python.org\u002Fpypi\u002Fopeninference-instrumentation-openlit)         |\n| [`openinference-instrumentation-openllmetry`](.\u002Fpython\u002Finstrumentation\u002Fopeninference-instrumentation-openllmetry) | OpenInference Span Processor for OpenLLMetry (Traceloop) traces. | [![PyPI Version](https:\u002F\u002Fimg.shields.io\u002Fpypi\u002Fv\u002Fopeninference-instrumentation-openllmetry.svg)](https:\u002F\u002Fpypi.python.org\u002Fpypi\u002Fopeninference-instrumentation-openllmetry) |\n\n### JavaScript Integrations\n\n| | Integration | Package | Version |\n|:---:|---|---|---|\n| \u003Cpicture>\u003Csource media=\"(prefers-color-scheme: dark)\" srcset=\"https:\u002F\u002Funpkg.com\u002F@lobehub\u002Ficons-static-png@latest\u002Fdark\u002Fopenai.png\">\u003Cimg height=\"14\" src=\"https:\u002F\u002Foss.gittoolsai.com\u002Fimages\u002FArize-ai_phoenix_readme_fc1845c68da0.png\">\u003C\u002Fpicture> | [OpenAI](https:\u002F\u002Farize.com\u002Fdocs\u002Fphoenix\u002Ftracing\u002Fintegrations-tracing\u002Fopenai-node-sdk) | `@arizeai\u002Fopeninference-instrumentation-openai` | [![NPM Version](https:\u002F\u002Fimg.shields.io\u002Fnpm\u002Fv\u002F@arizeai\u002Fopeninference-instrumentation-openai.svg)](https:\u002F\u002Fwww.npmjs.com\u002Fpackage\u002F@arizeai\u002Fopeninference-instrumentation-openai) |\n| \u003Cimg src=\"https:\u002F\u002Foss.gittoolsai.com\u002Fimages\u002FArize-ai_phoenix_readme_acb5f98fc60a.png\" height=\"14\"> | [LangChain.js](https:\u002F\u002Farize.com\u002Fdocs\u002Fphoenix\u002Ftracing\u002Fintegrations-tracing\u002Flangchain) | `@arizeai\u002Fopeninference-instrumentation-langchain` | [![NPM Version](https:\u002F\u002Fimg.shields.io\u002Fnpm\u002Fv\u002F@arizeai\u002Fopeninference-instrumentation-langchain.svg)](https:\u002F\u002Fwww.npmjs.com\u002Fpackage\u002F@arizeai\u002Fopeninference-instrumentation-langchain) |\n| \u003Cpicture>\u003Csource media=\"(prefers-color-scheme: dark)\" srcset=\"https:\u002F\u002Funpkg.com\u002F@lobehub\u002Ficons-static-png@latest\u002Fdark\u002Fvercel.png\">\u003Cimg height=\"14\" src=\"https:\u002F\u002Foss.gittoolsai.com\u002Fimages\u002FArize-ai_phoenix_readme_75660b5aa9b0.png\">\u003C\u002Fpicture> | [Vercel AI SDK](https:\u002F\u002Farize.com\u002Fdocs\u002Fphoenix\u002Ftracing\u002Fintegrations-tracing\u002Fvercel-ai-sdk) | `@arizeai\u002Fopeninference-vercel` | [![NPM Version](https:\u002F\u002Fimg.shields.io\u002Fnpm\u002Fv\u002F@arizeai\u002Fopeninference-vercel)](https:\u002F\u002Fwww.npmjs.com\u002Fpackage\u002F@arizeai\u002Fopeninference-vercel) |\n| | [BeeAI](https:\u002F\u002Farize.com\u002Fdocs\u002Fphoenix\u002Ftracing\u002Fintegrations-tracing\u002Fbeeai) | `@arizeai\u002Fopeninference-instrumentation-beeai` | [![NPM Version](https:\u002F\u002Fimg.shields.io\u002Fnpm\u002Fv\u002F@arizeai\u002Fopeninference-vercel)](https:\u002F\u002Fwww.npmjs.com\u002Fpackage\u002F@arizeai\u002Fopeninference-instrumentation-beeai) |\n| \u003Cimg src=\"https:\u002F\u002Foss.gittoolsai.com\u002Fimages\u002FArize-ai_phoenix_readme_6ef465e6e627.png\" height=\"14\"> | [Claude Agent SDK](https:\u002F\u002Farize.com\u002Fdocs\u002Fphoenix\u002Fintegrations\u002Ftypescript\u002Fclaude-agent-sdk) | `@arizeai\u002Fopeninference-instrumentation-claude-agent-sdk` | [![NPM Version](https:\u002F\u002Fimg.shields.io\u002Fnpm\u002Fv\u002F@arizeai%2Fopeninference-instrumentation-claude-agent-sdk.svg)](https:\u002F\u002Fwww.npmjs.com\u002Fpackage\u002F@arizeai\u002Fopeninference-instrumentation-claude-agent-sdk) |\n| \u003Cpicture>\u003Csource media=\"(prefers-color-scheme: dark)\" srcset=\"https:\u002F\u002Funpkg.com\u002F@lobehub\u002Ficons-static-png@latest\u002Fdark\u002Fmastra.png\">\u003Cimg height=\"14\" src=\"https:\u002F\u002Foss.gittoolsai.com\u002Fimages\u002FArize-ai_phoenix_readme_da5502535cec.png\">\u003C\u002Fpicture> | [Mastra](https:\u002F\u002Farize.com\u002Fdocs\u002Fphoenix\u002Fintegrations\u002Ftypescript\u002Fmastra) | `@mastra\u002Farize` | [![NPM Version](https:\u002F\u002Fimg.shields.io\u002Fnpm\u002Fv\u002F@mastra\u002Farize.svg)](https:\u002F\u002Fwww.npmjs.com\u002Fpackage\u002F@mastra\u002Farize) |\n| \u003Cpicture>\u003Csource media=\"(prefers-color-scheme: dark)\" srcset=\"https:\u002F\u002Funpkg.com\u002F@lobehub\u002Ficons-static-png@latest\u002Fdark\u002Fmcp.png\">\u003Cimg height=\"14\" src=\"https:\u002F\u002Foss.gittoolsai.com\u002Fimages\u002FArize-ai_phoenix_readme_53956c4cd629.png\">\u003C\u002Fpicture> | [MCP](https:\u002F\u002Farize.com\u002Fdocs\u002Fphoenix\u002Fintegrations\u002Ftypescript\u002Fmcp) | `@arizeai\u002Fopeninference-instrumentation-mcp` | [![NPM Version](https:\u002F\u002Fimg.shields.io\u002Fnpm\u002Fv\u002F@arizeai%2Fopeninference-instrumentation-mcp.svg)](https:\u002F\u002Fwww.npmjs.com\u002Fpackage\u002F@arizeai\u002Fopeninference-instrumentation-mcp) |\n\n### Java Integrations\n\n| | Integration | Package | Version |\n|:---:|---|---|---|\n| \u003Cimg src=\"https:\u002F\u002Foss.gittoolsai.com\u002Fimages\u002FArize-ai_phoenix_readme_acb5f98fc60a.png\" height=\"14\"> | [LangChain4j](https:\u002F\u002Fgithub.com\u002FArize-ai\u002Fopeninference\u002Ftree\u002Fmain\u002Fjava\u002Finstrumentation\u002Fopeninference-instrumentation-langchain4j) | `openinference-instrumentation-langchain4j` | [![Maven Central](https:\u002F\u002Fimg.shields.io\u002Fmaven-central\u002Fv\u002Fcom.arize\u002Fopeninference-instrumentation-langchain4j.svg)](https:\u002F\u002Fcentral.sonatype.com\u002Fartifact\u002Fcom.arize\u002Fopeninference-instrumentation-langchain4j) |\n| | SpringAI | `openinference-instrumentation-springAI` | [![Maven Central](https:\u002F\u002Fimg.shields.io\u002Fmaven-central\u002Fv\u002Fcom.arize\u002Fopeninference-instrumentation-springAI.svg)](https:\u002F\u002Fcentral.sonatype.com\u002Fartifact\u002Fcom.arize\u002Fopeninference-instrumentation-springAI) |\n| | [Arconia](https:\u002F\u002Farconia.io\u002Fdocs\u002Farconia\u002Flatest\u002Fobservability\u002Fgenerative-ai\u002F) | `openinference-instrumentation-springAI` | [![Maven Central](https:\u002F\u002Fimg.shields.io\u002Fmaven-central\u002Fv\u002Fcom.arize\u002Fopeninference-instrumentation-springAI.svg)](https:\u002F\u002Fcentral.sonatype.com\u002Fartifact\u002Fcom.arize\u002Fopeninference-instrumentation-springAI) |\n\n### Platforms\n\n| | Platform | Description | Docs |\n|:---:|---|---|---|\n| | [BeeAI](https:\u002F\u002Fdocs.beeai.dev\u002Fobservability\u002Fagents-traceability) | AI agent framework with built-in observability | [Integration Guide](https:\u002F\u002Fdocs.beeai.dev\u002Fobservability\u002Fagents-traceability) |\n| \u003Cimg src=\"https:\u002F\u002Foss.gittoolsai.com\u002Fimages\u002FArize-ai_phoenix_readme_8e9ca8ab7fe0.png\" height=\"14\"> | [Dify](https:\u002F\u002Fdocs.dify.ai\u002Fen\u002Fguides\u002Fmonitoring\u002Fintegrate-external-ops-tools\u002Fintegrate-phoenix) | Open-source LLM app development platform | [Integration Guide](https:\u002F\u002Fdocs.dify.ai\u002Fen\u002Fguides\u002Fmonitoring\u002Fintegrate-external-ops-tools\u002Fintegrate-phoenix) |\n| | [Envoy AI Gateway](https:\u002F\u002Fgithub.com\u002Fenvoyproxy\u002Fai-gateway) | AI Gateway built on Envoy Proxy for AI workloads | [Integration Guide](https:\u002F\u002Fgithub.com\u002Fenvoyproxy\u002Fai-gateway\u002Ftree\u002Fmain\u002Fcmd\u002Faigw#opentelemetry-setup-with-phoenix) |\n| | [LangFlow](https:\u002F\u002Farize.com\u002Fdocs\u002Fphoenix\u002Ftracing\u002Fintegrations-tracing\u002Flangflow) | Visual framework for building multi-agent and RAG applications | [Integration Guide](https:\u002F\u002Farize.com\u002Fdocs\u002Fphoenix\u002Ftracing\u002Fintegrations-tracing\u002Flangflow) |\n| | [LiteLLM Proxy](https:\u002F\u002Fdocs.litellm.ai\u002Fdocs\u002Fobservability\u002Fphoenix_integration#using-with-litellm-proxy) | Proxy server for LLMs | [Integration Guide](https:\u002F\u002Fdocs.litellm.ai\u002Fdocs\u002Fobservability\u002Fphoenix_integration#using-with-litellm-proxy) |\n| | [Flowise](https:\u002F\u002Farize.com\u002Fdocs\u002Fphoenix\u002Fintegrations\u002Fplatforms\u002Fflowise) | Visual framework for building LLM applications | [Integration Guide](https:\u002F\u002Farize.com\u002Fdocs\u002Fphoenix\u002Fintegrations\u002Fplatforms\u002Fflowise) |\n| | [Prompt Flow](https:\u002F\u002Farize.com\u002Fdocs\u002Fphoenix\u002Fintegrations\u002Fplatforms\u002Fprompt-flow) | Microsoft's prompt flow orchestration tool | [Integration Guide](https:\u002F\u002Farize.com\u002Fdocs\u002Fphoenix\u002Fintegrations\u002Fplatforms\u002Fprompt-flow) |\n| \u003Cimg src=\"https:\u002F\u002Foss.gittoolsai.com\u002Fimages\u002FArize-ai_phoenix_readme_6c74605c5569.png\" height=\"14\"> | [NVIDIA NeMo](https:\u002F\u002Farize.com\u002Fdocs\u002Fphoenix\u002Fintegrations\u002Fpython\u002Fnvidia) | NVIDIA NeMo Agent Toolkit for enterprise agents | [Integration Guide](https:\u002F\u002Farize.com\u002Fdocs\u002Fphoenix\u002Fintegrations\u002Fpython\u002Fnvidia) |\n| | [Graphite](https:\u002F\u002Farize.com\u002Fdocs\u002Fphoenix\u002Fintegrations\u002Fpython\u002Fgraphite) | Multi-agent LLM workflow framework with visual builder | [Integration Guide](https:\u002F\u002Farize.com\u002Fdocs\u002Fphoenix\u002Fintegrations\u002Fpython\u002Fgraphite) |\n\n## Coding Agent Skills\n\nThis repository includes [skills](https:\u002F\u002Fdocs.anthropic.com\u002Fen\u002Fdocs\u002Fclaude-code\u002Fskills) that teach coding agents how to work with Phoenix. They are located in [`.agents\u002Fskills\u002F`](.agents\u002Fskills\u002F) and can be used with Claude Code, Cursor, and other compatible tools.\n\n| Skill | Description |\n| ----- | ----------- |\n| [phoenix-cli](.agents\u002Fskills\u002Fphoenix-cli\u002F) | Debug LLM applications using the Phoenix CLI — fetch traces, analyze errors, review experiments, and query the GraphQL API |\n| [phoenix-evals](.agents\u002Fskills\u002Fphoenix-evals\u002F) | Build and run evaluators for AI\u002FLLM applications using Phoenix |\n| [phoenix-tracing](.agents\u002Fskills\u002Fphoenix-tracing\u002F) | OpenInference semantic conventions and instrumentation for tracing LLM applications |\n\n## Security & Privacy\n\nWe take data security and privacy very seriously. For more details, see our [Security and Privacy documentation](https:\u002F\u002Farize.com\u002Fdocs\u002Fphoenix\u002Fself-hosting\u002Fsecurity\u002Fprivacy).\n\n### Telemetry\n\nBy default, Phoenix collects basic web analytics (e.g., page views, UI interactions) to help us understand how Phoenix is used and improve the product. **None of your trace data, evaluation results, or any sensitive information is ever collected.**\n\nYou can opt-out of telemetry by setting the environment variable: `PHOENIX_TELEMETRY_ENABLED=false`\n\n## Community\n\nJoin our community to connect with thousands of AI builders.\n\n- 🌍 Join our [Slack community](https:\u002F\u002Fjoin.slack.com\u002Ft\u002Farize-ai\u002Fshared_invite\u002Fzt-3r07iavnk-ammtATWSlF0pSrd1DsMW7g).\n- 📚 Read our [documentation](https:\u002F\u002Farize.com\u002Fdocs\u002Fphoenix).\n- 💡 Ask questions and provide feedback in the _#phoenix-support_ channel.\n- 🌟 Leave a star on our [GitHub](https:\u002F\u002Fgithub.com\u002FArize-ai\u002Fphoenix).\n- 🐞 Report bugs with [GitHub Issues](https:\u002F\u002Fgithub.com\u002FArize-ai\u002Fphoenix\u002Fissues).\n- 𝕏 Follow us on [𝕏](https:\u002F\u002Ftwitter.com\u002FArizePhoenix).\n- 🗺️ Check out our [roadmap](https:\u002F\u002Fgithub.com\u002Forgs\u002FArize-ai\u002Fprojects\u002F45) to see where we're heading next.\n- 🧑‍🏫 Deep dive into everything [Agents](http:\u002F\u002Farize.com\u002Fai-agents\u002F) and [LLM Evaluations](https:\u002F\u002Farize.com\u002Fllm-evaluation) on Arize's Learning Hubs.\n\n## Breaking Changes\n\nSee the [migration guide](.\u002FMIGRATION.md) for a list of breaking changes.\n\n## Copyright, Patent, and License\n\nCopyright 2025 Arize AI, Inc. All Rights Reserved.\n\nPortions of this code are patent protected by one or more U.S. Patents. See the [IP_NOTICE](https:\u002F\u002Fgithub.com\u002FArize-ai\u002Fphoenix\u002Fblob\u002Fmain\u002FIP_NOTICE).\n\nThis software is licensed under the terms of the Elastic License 2.0 (ELv2). See [LICENSE](https:\u002F\u002Fgithub.com\u002FArize-ai\u002Fphoenix\u002Fblob\u002Fmain\u002FLICENSE).\n","\u003Cp align=\"center\">\n    \u003Ca target=\"_blank\" href=\"https:\u002F\u002Fphoenix.arize.com\" style=\"background:none\">\n        \u003Cimg alt=\"phoenix banner\" src=\"https:\u002F\u002Foss.gittoolsai.com\u002Fimages\u002FArize-ai_phoenix_readme_f6d8660ba395.jpg\" width=\"auto\" height=\"auto\">\u003C\u002Fimg>\n    \u003C\u002Fa>\n    \u003Cbr\u002F>\n    \u003Cbr\u002F>\n    \u003Ca href=\"https:\u002F\u002Farize.com\u002Fdocs\u002Fphoenix\u002F\">\n        \u003Cimg src=\"https:\u002F\u002Fimg.shields.io\u002Fstatic\u002Fv1?message=Docs&logo=data:image\u002Fpng;base64,iVBORw0KGgoAAAANSUhEUgAAAIAAAACACAYAAADDPmHLAAAG4ElEQVR4nO2d4XHjNhCFcTf+b3ZgdWCmgmMqOKUC0xXYrsBOBVEqsFRB7ApCVRCygrMriFQBM7h5mNlwKBECARLg7jeDscamSQj7sFgsQfBL27ZK4MtXsT1vRADMEQEwRwTAHBEAcEYdwBwRAHNEAMwRATBHBMAcEQBzRAAMEQAzRADAHBEAc0QAzBEBMEcEwBwRAHNEAMwRATBHBMAcEQBzRAAMEQAzRADAHBEAcEYdwBwRAHNEAMwRATBHBMAcEQBzRAAMEQAzRADAHBEAcEYdwBwRAHNEAMwRATBHBMAcEQBzRAAMEQAzRADAHBEAcEYdwBwRAHNEAMwRATBHBMAcEQBzRAAMEQAzRADAHBEAcEYdwBwRAHNEAMwRATBHBMAcEQBzRAAMEQAzRADAHBEAcEYdwBwRAHNEAMwRATBHBMAcEQBzRAAMEQAzRADAHBEAcEYdwBwRAHNEAMwRATBHBMAcEQBzRAAMEQAzRADAHBEAcEYdwBwRAHNEAMwRATBHBMAcEQBzRAAMEQAzRADAHBEAcEYdwBwRAHNEAMwRATBHBMAcEQBzRAAMEQAzRADAHBEAcEYdwBwRAHNEAMwRATBHBMAcEQBzRAAMEQAzRADAHBEAcEYdwBwRAHNEAMwRATBHBMAcEQBzRAAMEQAzRADAHBEAcEYdwBwRAHNEAMwRATBHBMAcEQBzRAAMEQAzRADAHBEAcEYdwBwRAHNEAMwRATBHBMAcEQBzRAAMEQAzRADAHBEAcEYdwBwRAHNEAMwRATBHBMAcEQBzRAAMEQAzRADAHBEAcEYdw......\u003Cp align=\"center\">\n    \u003Ca target=\"_blank\" href=\"https:\u002F\u002Fphoenix.arize.com\" style=\"background:none\">\n        \u003Cimg alt=\"phoenix banner\" src=\"https:\u002F\u002Foss.gittoolsai.com\u002Fimages\u002FArize-ai_phoenix_readme_f6d8660ba395.jpg\" width=\"auto\" height=\"auto\">\u003C\u002Fimg>\n    \u003C\u002Fa>\n    \u003Cbr\u002F>\n    \u003Cbr\u002F>\n    \u003Ca href=\"https:\u002F\u002Farize.com\u002Fdocs\u002Fphoenix\u002F\">\n        \u003Cimg src=\"https:\u002F\u002Fimg.shields.io\u002Fstatic\u002Fv1?message=Docs&logo=data:image\u002Fpng;base64,iVBORw0KGgoAAAANSUhEUgAAAIAAAACACAYAAADDPmHLAAAG4ElEQVR4nO2d4XHjNhCFcTf+b3ZgdWCmgmMqOKUC0xXYrsBOBVEqsFRB7ApCVRCygrMriFQBM7h5mNlwKBECARLg7jeDscamSQj7sFgsQfBL27ZK4MtXsT1vRADMEQEwRwTAHBEAcEcbd4WSw+D6803njHscO5sATmGEURGBiCj6yUlv1uX2gv91FsDViArbcA2RUKF8QhAV8RQc0b15DcOt0VaTE1oAfWj3dYdCBfGGsmSM0XX5HsP3nEMAXbqCeCdiOERQPx9og5exGJ0S4zRQN9KrUupfpdQWjZciure\u002FYIj7K0bjqwTyAHdovA805iqCOg2xgnB1nZ97IvaoSCURdIPGd5kvqUAKf6rkAk7M2SY9QDXdEr9wEOr9x96EiejMFnixBNteDISsyNw7hHRqc22evWcP4vt39O85bzZH30AKg4+eo8cQRI4bHAJ7hyYM3CNHrG9RrimSXuZmUkZjN\u002FO6nAPpcwCcJNmipAle2QM\u002F1GU3vITCXhvY91u9geN\u002FjOY27VuTnYL1PCeAcRhwh7\u002FBl8Ai+IuxPiOCShtfX\u002FsPDtY8w+sZjby86dw6dBeoigD7obd\u002FKo6fI4BF8DA9HnGdrcU0fLt+n4dfE6H5jpjYcVdu2L23b5lpjHoo+18FDbcszddF1rUee\u002F4C6ZiO+80rHZmjDoIQUQLdRtm3brkcKIUPjjqVPBIUPgW1GGN4YfawAL2IqAVB8iEE31tvIelARlCPPVaFOLoIupzY6xVcM4MoRUyHXyHhslH6PaPl5RP1Lh4UsOeKR2e8dzC0Aiuvc2Nx3fwhfxf\u002FhknouUYbWUk5GTAIwmOh5e+H0cor8vEL91hfOdEqINLq1AV+RKImJ6869f9tFIBVc6y7gd3lHfWyNX0LEr7EuDElhRdAlQjig0e\u002FRU31xxDltM4pF7IY3pLIgxAhhgzF\u002FiC2M0Hi4dkOGlyGMd\u002Fg7dsMbUlsR9ICe9WhxbA3DjRkSdjiHzQzlBSKNJsCzIcUlYdfI0dcWS8LMkPDkcJ0n\u002FO+Qyy\u002FIAtDkSPnp4Fu4WpthQR\u002Fzm2VcoI\u002F51fI28iYld9\u002FHEh4Pf7D0Bm845pwIPnHMUJSf45pT5x68s5T9AW6INzhHDeP1BYcNMew5SghkinWOwVnaBhHGG5ybMn70zBDe8buh8X6DqV0Sa\u002F5tWOIOIbcWQ8KBiGBnMb\u002FP0OuTd\u002FlddCrY5jn\u002FVLm3nL+fY4X4YREuv8vS9wh6HSkAExMs0viKySZRd44iyOH2FzPe98Fll7A7GNMmjay4GF9BAKGXesfCN0sRsDG+YrhP4O2ACFgZXzHdKPL2RMJoxc34ivFOod3AMMNUj5XxFfOtYrUIXvB5MandS+G+V\u002FAzZ+MrEcBPlpoFtUIEwBwRAG+OIgDe1CIA5ogAmCMCYI4IgDkiAOaIAJgjAmCOCIA5ogAmCMCYIwJgjgiAOSIA5ogAmCMCYI4IgDkiAOaIAJgjAmDOVYBXvwvxQV8NWJOd0esvJ94babZaz7B5ovldxnlDpYhp0JFr\u002FKTlLKcEMMQKpcDPXIQxGXsYmhZnXAXQh\u002FEWBQrr3bc80mATyyrEvs4+BdBHgbdxFOIhrDkSg1\u002F6Iu2LCS0AyoqI4ftUF00EY\u002FQ3h1fRj2JKAVCMGErmnsH1lfnemEsAlByvgl0z2qx5B8OPCuB8EIMADBlEEOV79j1whNE3c\u002FX2PmISAGUNr7CEmUSUhjfEKgBDAY+QohCiNrwhdgEYzPv7UxkadvBg0RrekAlJiM3iNX4yvmHoA9Sh4PF0QAzBEBMEcEwBwRAHNEAMwRAXBGKfUfr5hKvglRfO4AAAAASUVORK5CYII=&labelColor=grey&color=blue&logoColor=white&label=%20\"\u002F>\n    \u003C\u002Fa>\n    \u003Ca target=\"_blank\" href=\"https:\u002F\u002Fjoin.slack.com\u002Ft\u002Farize-ai\u002Fshared_invite\u002Fzt-3r07iavnk-ammtATWSlF0pSrd1DsMW7g\">\n        \u003Cimg src=\"https:\u002F\u002Fimg.shields.io\u002Fstatic\u002Fv1?message=Community&logo=slack&labelColor=grey&color=blue&logoColor=white&label=%20\"\u002F>\n    \u003C\u002Fa>\n     \u003Ca target=\"_blank\" href=\"https:\u002F\u002Fbsky.app\u002Fprofile\u002Farize-phoenix.bsky.social\">\n        \u003Cimg src=\"https:\u002F\u002Fimg.shields.io\u002Fbadge\u002F-phoenix-blue.svg?color=blue&labelColor=gray&logo=bluesky\">\n    \u003C\u002Fa>\n    \u003Ca target=\"_blank\" href=\"https:\u002F\u002Fx.com\u002FArizePhoenix\">\n        \u003Cimg src=\"https:\u002F\u002Fimg.shields.io\u002Fbadge\u002F-ArizePhoenix-blue.svg?color=blue&labelColor=gray&logo=x\">\n    \u003C\u002Fa>\n    \u003Ca target=\"_blank\" href=\"https:\u002F\u002Fpypi.org\u002Fproject\u002Farize-phoenix\u002F\">\n        \u003Cimg src=\"https:\u002F\u002Fimg.shields.io\u002Fpypi\u002Fv\u002Farize-phoenix?color=blue\">\n    \u003C\u002Fa>\n    \u003Ca target=\"_blank\" href=\"https:\u002F\u002Fanaconda.org\u002Fconda-forge\u002Farize-phoenix\">\n        \u003Cimg src=\"https:\u002F\u002Fimg.shields.io\u002Fconda\u002Fvn\u002Fconda-forge\u002Farize-phoenix.svg?color=blue\">\n    \u003C\u002Fa>\n    \u003Ca target=\"_blank\" href=\"https:\u002F\u002Fpypi.org\u002Fproject\u002Farize-phoenix\u002F\">\n        \u003Cimg src=\"https:\u002F\u002Fimg.shields.io\u002Fpypi\u002Fpyversions\u002Farize-phoenix\">\n    \u003C\u002Fa>\n    \u003Ca target=\"_blank\" href=\"https:\u002F\u002Fhub.docker.com\u002Fr\u002Farizephoenix\u002Fphoenix\u002Ftags\">\n        \u003Cimg src=\"https:\u002F\u002Fimg.shields.io\u002Fdocker\u002Fv\u002Farizephoenix\u002Fphoenix?sort=semver&logo=docker&label=image&color=blue\">\n    \u003C\u002Fa>\n    \u003Ca target=\"_blank\" href=\"https:\u002F\u002Fhub.docker.com\u002Fr\u002Farizephoenix\u002Fphoenix-helm\">\n        \u003Cimg src=\"https:\u002F\u002Fimg.shields.io\u002Fbadge\u002FHelm-blue?style=flat&logo=helm&labelColor=grey\"\u002F>\n    \u003C\u002Fa>\n    \u003Ca target=\"_blank\" href=\"https:\u002F\u002Fgithub.com\u002FArize-ai\u002Fphoenix\u002Ftree\u002Fmain\u002Fjs\u002Fpackages\u002Fphoenix-mcp\">\n        \u003Cimg src=\"https:\u002F\u002Foss.gittoolsai.com\u002Fimages\u002FArize-ai_phoenix_readme_d4184d72d7bf.png\" title=\"MCP Enabled\"\u002F>\n    \u003C\u002Fa>\n    \u003Ca href=\"cursor:\u002F\u002Fanysphere.cursor-deeplink\u002Fmcp\u002Finstall?name=phoenix&config=eyJjb21tYW5kIjoibnB4IC15IEBhcml6ZWFpL3Bob2VuaXgtbWNwQGxhdGVzdCAtLWJhc2VVcmwgaHR0cHM6Ly9teS1waG9lbml4LmNvbSAtLWFwaIKeyIBYXVyLWFwaIKeyifQ%3D%3D\">\u003Cimg src=\"https:\u002F\u002Fcursor.com\u002Fdeeplink\u002Fmcp-install-dark.svg\" alt=\"Add Arize Phoenix MCP server to Cursor\" height=20 \u002F>\u003C\u002Fa>\n    \u003Cimg referrerpolicy=\"no-referrer-when-downgrade\" src=\"https:\u002F\u002Fstatic.scarf.sh\u002Fa.png?x-pxid=8e8e8b34-7900-43fa-a38f-1f070bd48c64&page=README.md\" \u002F>\n\u003C\u002Fp>\n\nPhoenix 是一个开源的 AI 可观测性平台，专为实验、评估和故障排除而设计。它提供以下功能：\n\n- [**_追踪_**](https:\u002F\u002Farize.com\u002Fdocs\u002Fphoenix\u002Ftracing\u002Fllm-traces) - 使用基于 OpenTelemetry 的工具链追踪您的 LLM 应用程序运行时。\n- [**_评估_**](https:\u002F\u002Farize.com\u002Fdocs\u002Fphoenix\u002Fevaluation\u002Fllm-evals) - 利用 LLM 对您的应用程序性能进行基准测试，包括响应评估和检索评估。\n- [**_数据集_**](https:\u002F\u002Farize.com\u002Fdocs\u002Fphoenix\u002Fdatasets-and-experiments\u002Foverview-datasets) - 创建用于实验、评估和微调的版本化示例数据集。\n- [**_实验_**](https:\u002F\u002Farize.com\u002Fdocs\u002Fphoenix\u002Fdatasets-and-experiments\u002Foverview-datasets#experiments) - 跟踪并评估提示、LLM 和检索方面的变化。\n- [**_游乐场_**](https:\u002F\u002Farize.com\u002Fdocs\u002Fphoenix\u002Fprompt-engineering\u002Foverview-prompts)- 优化提示、比较模型、调整参数，并回放已追踪的 LLM 调用。\n- [**_提示管理_**](https:\u002F\u002Farize.com\u002Fdocs\u002Fphoenix\u002Fprompt-engineering\u002Foverview-prompts\u002Fprompt-management)- 使用版本控制、标记和实验系统地管理和测试提示变更。\n\nPhoenix 不依赖于特定的供应商或编程语言，开箱即用地支持多种流行框架（[OpenAI Agents SDK](https:\u002F\u002Farize.com\u002Fdocs\u002Fphoenix\u002Ftracing\u002Fintegrations-tracing\u002Fopenai-agents-sdk)、[Claude Agent SDK](https:\u002F\u002Farize.com\u002Fdocs\u002Fphoenix\u002Fintegrations\u002Fpython\u002Fclaude-agent-sdk)、[LangGraph](https:\u002F\u002Farize.com\u002Fdocs\u002Fphoenix\u002Ftracing\u002Fintegrations-tracing\u002Flangchain)、[Vercel AI SDK](https:\u002F\u002Farize.com\u002Fdocs\u002Fphoenix\u002Ftracing\u002Fintegrations-tracing\u002Fvercel-ai-sdk)、[Mastra](https:\u002F\u002Farize.com\u002Fdocs\u002Fphoenix\u002Fintegrations\u002Ftypescript\u002Fmastra)、[CrewAI](https:\u002F\u002Farize.com\u002Fdocs\u002Fphoenix\u002Ftracing\u002Fintegrations-tracing\u002Fcrewai)、[LlamaIndex](https:\u002F\u002Farize.com\u002Fdocs\u002Fphoenix\u002Ftracing\u002Fintegrations-tracing\u002Fllamaindex)、[DSPy](https:\u002F\u002Farize.com\u002Fdocs\u002Fphoenix\u002Ftracing\u002Fintegrations-tracing\u002Fdspy)）以及大模型服务提供商（[OpenAI](https:\u002F\u002Farize.com\u002Fdocs\u002Fphoenix\u002Ftracing\u002Fintegrations-tracing\u002Fopenai)、[Anthropic](https:\u002F\u002Farize.com\u002Fdocs\u002Fphoenix\u002Ftracing\u002Fintegrations-tracing\u002Fanthropic)、[Google GenAI](https:\u002F\u002Farize.com\u002Fdocs\u002Fphoenix\u002Ftracing\u002Fintegrations-tracing\u002Fgoogle-genai)、[Google ADK](https:\u002F\u002Farize.com\u002Fdocs\u002Fphoenix\u002Fintegrations\u002Fllm-providers\u002Fgoogle-gen-ai\u002Fgoogle-adk-tracing)、[AWS Bedrock](https:\u002F\u002Farize.com\u002Fdocs\u002Fphoenix\u002Ftracing\u002Fintegrations-tracing\u002Fbedrock)、[OpenRouter](https:\u002F\u002Farize.com\u002Fdocs\u002Fphoenix\u002Fintegrations\u002Fpython\u002Fopenrouter)、[LiteLLM](https:\u002F\u002Farize.com\u002Fdocs\u002Fphoenix\u002Ftracing\u002Fintegrations-tracing\u002Flitellm) 等）。有关自动埋点的详细信息，请参阅 [OpenInference](https:\u002F\u002Fgithub.com\u002FArize-ai\u002Fopeninference) 项目。\n\nPhoenix 几乎可以在任何环境中运行，包括您的本地机器、Jupyter Notebook、容器化部署或云端。\n\n\n\n## 安装\n\n您可以通过 `pip` 或 `conda` 安装 Phoenix：\n\n```shell\npip install arize-phoenix\n```\n\nPhoenix 的容器镜像可在 [Docker Hub](https:\u002F\u002Fhub.docker.com\u002Fr\u002Farizephoenix\u002Fphoenix) 上获取，可以使用 Docker 或 Kubernetes 进行部署。Arize AI 还在 [app.phoenix.arize.com](https:\u002F\u002Fapp.phoenix.arize.com\u002F) 提供云实例。\n\n## 软件包\n\n`arize-phoenix` 包含整个 Phoenix 平台。不过，如果您已经部署了 Phoenix 平台，则可以使用与平台配合使用的轻量级 Python 子包和 TypeScript 包。\n\n### Python 子包\n\n| 软件包                                                                                       | 版本与文档                                                                                                                                                                                                                                                                      | 描述                                                                                |\n| --------------------------------------------------------------------------------------------- | ----------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------- | ------------------------------------------------------------------------------------------ |\n| [arize-phoenix-otel](https:\u002F\u002Fgithub.com\u002FArize-ai\u002Fphoenix\u002Ftree\u002Fmain\u002Fpackages\u002Fphoenix-otel)     | [![PyPI 版本](https:\u002F\u002Fimg.shields.io\u002Fpypi\u002Fv\u002Farize-phoenix-otel)](https:\u002F\u002Fpypi.org\u002Fproject\u002Farize-phoenix-otel\u002F) [![文档](https:\u002F\u002Fimg.shields.io\u002Fbadge\u002Fdocs-blue?logo=readthedocs&logoColor=white)](https:\u002F\u002Farize-phoenix.readthedocs.io\u002Fprojects\u002Fotel\u002Fen\u002Flatest\u002Findex.html)       | 提供基于 OpenTelemetry 原语的轻量级封装，并带有 Phoenix 感知的默认配置 |\n| [arize-phoenix-client](https:\u002F\u002Fgithub.com\u002FArize-ai\u002Fphoenix\u002Ftree\u002Fmain\u002Fpackages\u002Fphoenix-client) | [![PyPI 版本](https:\u002F\u002Fimg.shields.io\u002Fpypi\u002Fv\u002Farize-phoenix-client)](https:\u002F\u002Fpypi.org\u002Fproject\u002Farize-phoenix-client\u002F) [![文档](https:\u002F\u002Fimg.shields.io\u002Fbadge\u002Fdocs-blue?logo=readthedocs&logoColor=white)](https:\u002F\u002Farize-phoenix.readthedocs.io\u002Fprojects\u002Fclient\u002Fen\u002Flatest\u002Findex.html) | 用于通过其 OpenAPI REST 接口与 Phoenix 服务器交互的轻量客户端  |\n| [arize-phoenix-evals](https:\u002F\u002Fgithub.com\u002FArize-ai\u002Fphoenix\u002Ftree\u002Fmain\u002Fpackages\u002Fphoenix-evals)   | [![PyPI 版本](https:\u002F\u002Fimg.shields.io\u002Fpypi\u002Fv\u002Farize-phoenix-evals)](https:\u002F\u002Fpypi.org\u002Fproject\u002Farize-phoenix-evals\u002F) [![文档](https:\u002F\u002Fimg.shields.io\u002Fbadge\u002Fdocs-blue?logo=readthedocs&logoColor=white)](https:\u002F\u002Farize-phoenix.readthedocs.io\u002Fprojects\u002Fevals\u002Fen\u002Flatest\u002Findex.html)    | 用于评估大模型应用的工具，包括 RAG 相关性、答案相关性等   |\n\n### TypeScript 子包\n\n| 包                                                                                             | 版本与文档                                                                                                                                                                                                                                                                                 | 描述                                                                                                  |\n| --------------------------------------------------------------------------------------------------- | ---------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------- | ------------------------------------------------------------------------------------------------------------ |\n| [@arizeai\u002Fphoenix-otel](https:\u002F\u002Fgithub.com\u002FArize-ai\u002Fphoenix\u002Ftree\u002Fmain\u002Fjs\u002Fpackages\u002Fphoenix-otel)     | [![NPM 版本](https:\u002F\u002Fimg.shields.io\u002Fnpm\u002Fv\u002F%40arizeai%2Fphoenix-otel)](https:\u002F\u002Fwww.npmjs.com\u002Fpackage\u002F@arizeai\u002Fphoenix-otel) [![文档](https:\u002F\u002Fimg.shields.io\u002Fbadge\u002Fdocs-blue?logo=typescript&logoColor=white)](https:\u002F\u002Farize-ai.github.io\u002Fphoenix\u002F)                                           | 提供基于 OpenTelemetry 原语的轻量级封装，并带有 Phoenix 特有的默认配置                   |\n| [@arizeai\u002Fphoenix-client](https:\u002F\u002Fgithub.com\u002FArize-ai\u002Fphoenix\u002Ftree\u002Fmain\u002Fjs\u002Fpackages\u002Fphoenix-client) | [![NPM 版本](https:\u002F\u002Fimg.shields.io\u002Fnpm\u002Fv\u002F%40arizeai%2Fphoenix-client)](https:\u002F\u002Fwww.npmjs.com\u002Fpackage\u002F@arizeai\u002Fphoenix-client) [![文档](https:\u002F\u002Fimg.shields.io\u002Fbadge\u002Fdocs-blue?logo=typescript&logoColor=white)](https:\u002F\u002Farize-ai.github.io\u002Fphoenix\u002F)                                       | Arize Phoenix API 的客户端                                                                             |\n| [@arizeai\u002Fphoenix-evals](https:\u002F\u002Fgithub.com\u002FArize-ai\u002Fphoenix\u002Ftree\u002Fmain\u002Fjs\u002Fpackages\u002Fphoenix-evals)   | [![NPM 版本](https:\u002F\u002Fimg.shields.io\u002Fnpm\u002Fv\u002F%40arizeai%2Fphoenix-evals)](https:\u002F\u002Fwww.npmjs.com\u002Fpackage\u002F@arizeai\u002Fphoenix-evals) [![文档](https:\u002F\u002Fimg.shields.io\u002Fbadge\u002Fdocs-blue?logo=typescript&logoColor=white)](https:\u002F\u002Farize-ai.github.io\u002Fphoenix\u002F)                                         | 用于 LLM 应用程序的 TypeScript 评估库（Alpha 版）                                           |\n| [@arizeai\u002Fphoenix-mcp](https:\u002F\u002Fgithub.com\u002FArize-ai\u002Fphoenix\u002Ftree\u002Fmain\u002Fjs\u002Fpackages\u002Fphoenix-mcp)       | [![NPM 版本](https:\u002F\u002Fimg.shields.io\u002Fnpm\u002Fv\u002F%40arizeai%2Fphoenix-mcp)](https:\u002F\u002Fwww.npmjs.com\u002Fpackage\u002F@arizeai\u002Fphoenix-mcp) [![文档](https:\u002F\u002Fimg.shields.io\u002Fbadge\u002Fdocs-blue?logo=markdown&logoColor=white)](.\u002Fjs\u002Fpackages\u002Fphoenix-mcp\u002FREADME.md)                                               | Arize Phoenix 的 MCP 服务器实现，为 Phoenix 的各项功能提供统一接口            |\n| [@arizeai\u002Fphoenix-cli](https:\u002F\u002Fgithub.com\u002FArize-ai\u002Fphoenix\u002Ftree\u002Fmain\u002Fjs\u002Fpackages\u002Fphoenix-cli)       | [![NPM 版本](https:\u002F\u002Fimg.shields.io\u002Fnpm\u002Fv\u002F%40arizeai%2Fphoenix-cli)](https:\u002F\u002Fwww.npmjs.com\u002Fpackage\u002F@arizeai\u002Fphoenix-cli) [![文档](https:\u002F\u002Fimg.shields.io\u002Fbadge\u002Fdocs-blue?logo=typescript&logoColor=white)](https:\u002F\u002Farize.com\u002Fdocs\u002Fphoenix\u002Fsdk-api-reference\u002Ftypescript\u002Farizeai-phoenix-cli) | 用于获取 Trace、数据集和实验的命令行工具，可与 Claude Code、Cursor 等编码助手配合使用 |\n\n## 跟踪集成\n\nPhoenix 基于 OpenTelemetry 构建，具有供应商、语言和框架无关性。有关跟踪集成及示例应用的详细信息，请参阅 [OpenInference](https:\u002F\u002Fgithub.com\u002FArize-ai\u002Fopeninference) 项目。\n\n**Python 集成**\n| | 集成 | 包 | 版本 |\n|:---:|---|---|---|\n| \u003Cpicture>\u003Csource media=\"(prefers-color-scheme: dark)\" srcset=\"https:\u002F\u002Funpkg.com\u002F@lobehub\u002Ficons-static-png@latest\u002Fdark\u002Fopenai.png\">\u003Cimg height=\"14\" src=\"https:\u002F\u002Foss.gittoolsai.com\u002Fimages\u002FArize-ai_phoenix_readme_fc1845c68da0.png\">\u003C\u002Fpicture> | [OpenAI](https:\u002F\u002Farize.com\u002Fdocs\u002Fphoenix\u002Ftracing\u002Fintegrations-tracing\u002Fopenai) | `openinference-instrumentation-openai` | [![PyPI 版本](https:\u002F\u002Fimg.shields.io\u002Fpypi\u002Fv\u002Fopeninference-instrumentation-openai.svg)](https:\u002F\u002Fpypi.python.org\u002Fpypi\u002Fopeninference-instrumentation-openai) |\n| \u003Cpicture>\u003Csource media=\"(prefers-color-scheme: dark)\" srcset=\"https:\u002F\u002Funpkg.com\u002F@lobehub\u002Ficons-static-png@latest\u002Fdark\u002Fopenai.png\">\u003Cimg height=\"14\" src=\"https:\u002F\u002Foss.gittoolsai.com\u002Fimages\u002FArize-ai_phoenix_readme_fc1845c68da0.png\">\u003C\u002Fpicture> | [OpenAI Agents](https:\u002F\u002Farize.com\u002Fdocs\u002Fphoenix\u002Ftracing\u002Fintegrations-tracing\u002Fopenai-agents-sdk) | `openinference-instrumentation-openai-agents` | [![PyPI 版本](https:\u002F\u002Fimg.shields.io\u002Fpypi\u002Fv\u002Fopeninference-instrumentation-openai-agents.svg)](https:\u002F\u002Fpypi.python.org\u002Fpypi\u002Fopeninference-instrumentation-openai-agents) |\n| \u003Cimg src=\"https:\u002F\u002Foss.gittoolsai.com\u002Fimages\u002FArize-ai_phoenix_readme_370c5501b150.png\" height=\"14\"> | [LlamaIndex](https:\u002F\u002Farize.com\u002Fdocs\u002Fphoenix\u002Ftracing\u002Fintegrations-tracing\u002Fllamaindex) | `openinference-instrumentation-llama-index` | [![PyPI 版本](https:\u002F\u002Fimg.shields.io\u002Fpypi\u002Fv\u002Fopeninference-instrumentation-llama-index.svg)](https:\u002F\u002Fpypi.python.org\u002Fpypi\u002Fopeninference-instrumentation-llama-index) |\n| | [DSPy](https:\u002F\u002Farize.com\u002Fdocs\u002Fphoenix\u002Ftracing\u002Fintegrations-tracing\u002Fdspy) | `openinference-instrumentation-dspy` | [![PyPI 版本](https:\u002F\u002Fimg.shields.io\u002Fpypi\u002Fv\u002Fopeninference-instrumentation-dspy.svg)](https:\u002F\u002Fpypi.python.org\u002Fpypi\u002Fopeninference-instrumentation-dspy) |\n| \u003Cimg src=\"https:\u002F\u002Foss.gittoolsai.com\u002Fimages\u002FArize-ai_phoenix_readme_7d660dac1f19.png\" height=\"14\"> | [AWS Bedrock](https:\u002F\u002Farize.com\u002Fdocs\u002Fphoenix\u002Ftracing\u002Fintegrations-tracing\u002Fbedrock) | `openinference-instrumentation-bedrock` | [![PyPI 版本](https:\u002F\u002Fimg.shields.io\u002Fpypi\u002Fv\u002Fopeninference-instrumentation-bedrock.svg)](https:\u002F\u002Fpypi.python.org\u002Fpypi\u002Fopeninference-instrumentation-bedrock) |\n| \u003Cimg src=\"https:\u002F\u002Foss.gittoolsai.com\u002Fimages\u002FArize-ai_phoenix_readme_acb5f98fc60a.png\" height=\"14\"> | [LangChain](https:\u002F\u002Farize.com\u002Fdocs\u002Fphoenix\u002Ftracing\u002Fintegrations-tracing\u002Flangchain) | `openinference-instrumentation-langchain` | [![PyPI 版本](https:\u002F\u002Fimg.shields.io\u002Fpypi\u002Fv\u002Fopeninference-instrumentation-langchain.svg)](https:\u002F\u002Fpypi.python.org\u002Fpypi\u002Fopeninference-instrumentation-langchain) |\n| \u003Cimg src=\"https:\u002F\u002Foss.gittoolsai.com\u002Fimages\u002FArize-ai_phoenix_readme_49ae725664be.png\" height=\"14\"> | [MistralAI](https:\u002F\u002Farize.com\u002Fdocs\u002Fphoenix\u002Ftracing\u002Fintegrations-tracing\u002Fmistralai) | `openinference-instrumentation-mistralai` | [![PyPI 版本](https:\u002F\u002Fimg.shields.io\u002Fpypi\u002Fv\u002Fopeninference-instrumentation-mistralai.svg)](https:\u002F\u002Fpypi.python.org\u002Fpypi\u002Fopeninference-instrumentation-mistralai) |\n| \u003Cimg src=\"https:\u002F\u002Foss.gittoolsai.com\u002Fimages\u002FArize-ai_phoenix_readme_8d65b0abf33e.png\" height=\"14\"> | [Google GenAI](https:\u002F\u002Farize.com\u002Fdocs\u002Fphoenix\u002Ftracing\u002Fintegrations-tracing\u002Fgoogle-gen-ai) | `openinference-instrumentation-google-genai` | [![PyPI 版本](https:\u002F\u002Fimg.shields.io\u002Fpypi\u002Fv\u002Fopeninference-instrumentation-google-genai.svg)](https:\u002F\u002Fpypi.python.org\u002Fpypi\u002Fopeninference-instrumentation-google-genai) |\n| \u003Cimg src=\"https:\u002F\u002Foss.gittoolsai.com\u002Fimages\u002FArize-ai_phoenix_readme_8d65b0abf33e.png\" height=\"14\"> | [Google ADK](https:\u002F\u002Farize.com\u002Fdocs\u002Fphoenix\u002Fintegrations\u002Fllm-providers\u002Fgoogle-gen-ai\u002Fgoogle-adk-tracing) | `openinference-instrumentation-google-adk` | [![PyPI 版本](https:\u002F\u002Fimg.shields.io\u002Fpypi\u002Fv\u002Fopeninference-instrumentation-google-adk.svg)](https:\u002F\u002Fpypi.python.org\u002Fpypi\u002Fopeninference-instrumentation-google-adk) |\n| | [Guardrails](https:\u002F\u002Farize.com\u002Fdocs\u002Fphoenix\u002Ftracing\u002Fintegrations-tracing\u002Fguardrails) | `openinference-instrumentation-guardrails` | [![PyPI 版本](https:\u002F\u002Fimg.shields.io\u002Fpypi\u002Fv\u002Fopeninference-instrumentation-guardrails.svg)](https:\u002F\u002Fpypi.python.org\u002Fpypi\u002Fopeninference-instrumentation-guardrails) |\n| \u003Cimg src=\"https:\u002F\u002Foss.gittoolsai.com\u002Fimages\u002FArize-ai_phoenix_readme_e8c58b26f28a.png\" height=\"14\"> | [VertexAI](https:\u002F\u002Farize.com\u002Fdocs\u002Fphoenix\u002Ftracing\u002Fintegrations-tracing\u002Fvertexai) | `openinference-instrumentation-vertexai` | [![PyPI 版本](https:\u002F\u002Fimg.shields.io\u002Fpypi\u002Fv\u002Fopeninference-instrumentation-vertexai.svg)](https:\u002F\u002Fpypi.python.org\u002Fpypi\u002Fopeninference-instrumentation-vertexai) |\n| \u003Cimg src=\"https:\u002F\u002Foss.gittoolsai.com\u002Fimages\u002FArize-ai_phoenix_readme_6e375164f316.png\" height=\"14\"> | [CrewAI](https:\u002F\u002Farize.com\u002Fdocs\u002Fphoenix\u002Ftracing\u002Fintegrations-tracing\u002Fcrewai) | `openinference-instrumentation-crewai` | [![PyPI 版本](https:\u002F\u002Fimg.shields.io\u002Fpypi\u002Fv\u002Fopeninference-instrumentation-crewai.svg)](https:\u002F\u002Fpypi.python.org\u002Fpypi\u002Fopeninference-instrumentation-crewai) |\n| | [Haystack](https:\u002F\u002Farize.com\u002Fdocs\u002Fphoenix\u002Ftracing\u002Fintegrations-tracing\u002Fhaystack) | `openinference-instrumentation-haystack` | [![PyPI 版本](https:\u002F\u002Fimg.shields.io\u002Fpypi\u002Fv\u002Fopeninference-instrumentation-haystack.svg)](https:\u002F\u002Fpypi.python.org\u002Fpypi\u002Fopeninference-instrumentation-haystack) |\n| | [LiteLLM](https:\u002F\u002Farize.com\u002Fdocs\u002Fphoenix\u002Ftracing\u002Fintegrations-tracing\u002Flitellm) | `openinference-instrumentation-litellm` | [![PyPI 版本](https:\u002F\u002Fimg.shields.io\u002Fpypi\u002Fv\u002Fopeninference-instrumentation-litellm.svg)](https:\u002F\u002Fpypi.python.org\u002Fpypi\u002Fopeninference-instrumentation-litellm) |\n| \u003Cpicture>\u003Csource media=\"(prefers-color-scheme: dark)\" srcset=\"https:\u002F\u002Funpkg.com\u002F@lobehub\u002Ficons-static-png@latest\u002Fdark\u002Fgroq.png\">\u003Cimg height=\"14\" src=\"https:\u002F\u002Foss.gittoolsai.com\u002Fimages\u002FArize-ai_phoenix_readme_8964c766bdfe.png\">\u003C\u002Fpicture> | [Groq](https:\u002F\u002Farize.com\u002Fdocs\u002Fphoenix\u002Ftracing\u002Fintegrations-tracing\u002Fgroq) | `openinference-instrumentation-groq` | [![PyPI 版本](https:\u002F\u002Fimg.shields.io\u002Fpypi\u002Fv\u002Fopeninference-instrumentation-groq.svg)](https:\u002F\u002Fpypi.python.org\u002Fpypi\u002Fopeninference-instrumentation-groq) |\n| | [Instructor](https:\u002F\u002Farize.com\u002Fdocs\u002Fphoenix\u002Ftracing\u002Fintegrations-tracing\u002Finstructor) | `openinference-instrumentation-instructor` | [![PyPI 版本](https:\u002F\u002Fimg.shields.io\u002Fpypi\u002Fv\u002Fopeninference-instrumentation-instructor.svg)](https:\u002F\u002Fpypi.python.org\u002Fpypi\u002Fopeninference-instrumentation-instructor) |\n| \u003Cpicture>\u003Csource media=\"(prefers-color-scheme: dark)\" srcset=\"https:\u002F\u002Funpkg.com\u002F@lobehub\u002Ficons-static-png@latest\u002Fdark\u002Fanthropic.png\">\u003Cimg height=\"14\" src=\"https:\u002F\u002Foss.gittoolsai.com\u002Fimages\u002FArize-ai_phoenix_readme_5d80b94217f0.png\">\u003C\u002Fpicture> | [Anthropic](https:\u002F\u002Farize.com\u002Fdocs\u002Fphoenix\u002Ftracing\u002Fintegrations-tracing\u002Fanthropic) | `openinference-instrumentation-anthropic` | [![PyPI 版本](https:\u002F\u002Fimg.shields.io\u002Fpypi\u002Fv\u002Fopeninference-instrumentation-anthropic.svg)](https:\u002F\u002Fpypi.python.org\u002Fpypi\u002Fopeninference-instrumentation-anthropic) |\n| \u003Cimg src=\"https:\u002F\u002Foss.gittoolsai.com\u002Fimages\u002FArize-ai_phoenix_readme_d4b49bf027f0.png\" height=\"14\"> | [Smolagents](https:\u002F\u002Fhuggingface.co\u002Fdocs\u002Fsmolagents\u002Fen\u002Ftutorials\u002Finspect_runs) | `openinference-instrumentation-smolagents` | [![PyPI 版本](https:\u002F\u002Fimg.shields.io\u002Fpypi\u002Fv\u002Fopeninference-instrumentation-smolagents.svg)](https:\u002F\u002Fpypi.python.org\u002Fpypi\u002Fopeninference-instrumentation-smolagents) |\n| | [Agno](https:\u002F\u002Farize.com\u002Fdocs\u002Fphoenix\u002Ftracing\u002Fintegrations-tracing\u002Fagno) | `openinference-instrumentation-agno` | [![PyPI 版本](https:\u002F\u002Fimg.shields.io\u002Fpypi\u002Fv\u002Fopeninference-instrumentation-agno.svg)](https:\u002F\u002Fpypi.python.org\u002Fpypi\u002Fopeninference-instrumentation-agno) |\n| \u003Cpicture>\u003Csource media=\"(prefers-color-scheme: dark)\" srcset=\"https:\u002F\u002Funpkg.com\u002F@lobehub\u002Ficons-static-png@latest\u002Fdark\u002Fmcp.png\">\u003Cimg height=\"14\" src=\"https:\u002F\u002Foss.gittoolsai.com\u002Fimages\u002FArize-ai_phoenix_readme_53956c4cd629.png\">\u003C\u002Fpicture> | [MCP](https:\u002F\u002Farize.com\u002Fdocs\u002Fphoenix\u002Ftracing\u002Fintegrations-tracing\u002Fmodel-context-protocol-mcp) | `openinference-instrumentation-mcp` | [![PyPI 版本](https:\u002F\u002Fimg.shields.io\u002Fpypi\u002Fv\u002Fopeninference-instrumentation-mcp.svg)](https:\u002F\u002Fpypi.python.org\u002Fpypi\u002Fopeninference-instrumentation-mcp) |\n| \u003Cimg src=\"https:\u002F\u002Foss.gittoolsai.com\u002Fimages\u002FArize-ai_phoenix_readme_f900f128bcfe.png\" height=\"14\"> | [Pydantic AI](https:\u002F\u002Farize.com\u002Fdocs\u002Fphoenix\u002Fintegrations\u002Fpython\u002Fpydantic) | `openinference-instrumentation-pydantic-ai` | [![PyPI 版本](https:\u002F\u002Fimg.shields.io\u002Fpypi\u002Fv\u002Fopeninference-instrumentation-pydantic-ai.svg)](https:\u002F\u002Fpypi.python.org\u002Fpypi\u002Fopeninference-instrumentation-pydantic-ai) |\n| | [Autogen AgentChat](https:\u002F\u002Farize.com\u002Fdocs\u002Fphoenix\u002Fintegrations\u002Fframeworks\u002Fautogen\u002Fautogen-tracing) | `openinference-instrumentation-autogen-agentchat` | [![PyPI 版本](https:\u002F\u002Fimg.shields.io\u002Fpypi\u002Fv\u002Fopeninference-instrumentation-autogen-agentchat.svg)](https:\u002F\u002Fpypi.python.org\u002Fpypi\u002Fopeninference-instrumentation-autogen-agentchat) |\n| | [Portkey](https:\u002F\u002Farize.com\u002Fdocs\u002Fphoenix\u002Fintegrations\u002Fportkey) | `openinference-instrumentation-portkey` | [![PyPI 版本](https:\u002F\u002Fimg.shields.io\u002Fpypi\u002Fv\u002Fopeninference-instrumentation-portkey.svg)](https:\u002F\u002Fpypi.python.org\u002Fpypi\u002Fopeninference-instrumentation-portkey) |\n| | [Agent Spec](https:\u002F\u002Farize.com\u002Fdocs\u002Fphoenix\u002Ftracing\u002Fintegrations-tracing\u002Fagentspec) | `openinference-instrumentation-agentspec` | [![PyPI 版本](https:\u002F\u002Fimg.shields.io\u002Fpypi\u002Fv\u002Fopeninference-instrumentation-agentspec.svg)](https:\u002F\u002Fpypi.python.org\u002Fpypi\u002Fopeninference-instrumentation-agentspec) |\n| \u003Cimg src=\"https:\u002F\u002Foss.gittoolsai.com\u002Fimages\u002FArize-ai_phoenix_readme_6ef465e6e627.png\" height=\"14\"> | [Claude Agent SDK](https:\u002F\u002Farize.com\u002Fdocs\u002Fphoenix\u002Fintegrations\u002Fpython\u002Fclaude-agent-sdk) | `openinference-instrumentation-claude-agent-sdk` | [![PyPI 版本](https:\u002F\u002Fimg.shields.io\u002Fpypi\u002Fv\u002Fopeninference-instrumentation-claude-agent-sdk.svg)](https:\u002F\u002Fpypi.python.org\u002Fpypi\u002Fopeninference-instrumentation-claude-agent-sdk) |\n\n## Span 处理器\n\n通过添加统一数据的 span 处理器，在其他观测库之间规范化和转换数据。\n\n| 软件包                                                                                                           | 描述                                                      | 版本                                                                                                                                                                |\n| ----------------------------------------------------------------------------------------------------------------- | ---------------------------------------------------------------- | ---------------------------------------------------------------------------------------------------------------------------------------------------------------------- |\n| [`openinference-instrumentation-openlit`](.\u002Fpython\u002Finstrumentation\u002Fopeninference-instrumentation-openlit)         | OpenInference Span 处理器，用于 OpenLIT 跟踪。                 | [![PyPI 版本](https:\u002F\u002Fimg.shields.io\u002Fpypi\u002Fv\u002Fopeninference-instrumentation-openlit.svg)](https:\u002F\u002Fpypi.python.org\u002Fpypi\u002Fopeninference-instrumentation-openlit)         |\n| [`openinference-instrumentation-openllmetry`](.\u002Fpython\u002Finstrumentation\u002Fopeninference-instrumentation-openllmetry) | OpenInference Span 处理器，用于 OpenLLMetry (Traceloop) 跟踪。 | [![PyPI 版本](https:\u002F\u002Fimg.shields.io\u002Fpypi\u002Fv\u002Fopeninference-instrumentation-openllmetry.svg)](https:\u002F\u002Fpypi.python.org\u002Fpypi\u002Fopeninference-instrumentation-openllmetry) |\n\n### JavaScript 集成\n\n| | 集成 | 包 | 版本 |\n|:---:|---|---|---|\n| \u003Cpicture>\u003Csource media=\"(prefers-color-scheme: dark)\" srcset=\"https:\u002F\u002Funpkg.com\u002F@lobehub\u002Ficons-static-png@latest\u002Fdark\u002Fopenai.png\">\u003Cimg height=\"14\" src=\"https:\u002F\u002Foss.gittoolsai.com\u002Fimages\u002FArize-ai_phoenix_readme_fc1845c68da0.png\">\u003C\u002Fpicture> | [OpenAI](https:\u002F\u002Farize.com\u002Fdocs\u002Fphoenix\u002Ftracing\u002Fintegrations-tracing\u002Fopenai-node-sdk) | `@arizeai\u002Fopeninference-instrumentation-openai` | [![NPM 版本](https:\u002F\u002Fimg.shields.io\u002Fnpm\u002Fv\u002F@arizeai\u002Fopeninference-instrumentation-openai.svg)](https:\u002F\u002Fwww.npmjs.com\u002Fpackage\u002F@arizeai\u002Fopeninference-instrumentation-openai) |\n| \u003Cimg src=\"https:\u002F\u002Foss.gittoolsai.com\u002Fimages\u002FArize-ai_phoenix_readme_acb5f98fc60a.png\" height=\"14\"> | [LangChain.js](https:\u002F\u002Farize.com\u002Fdocs\u002Fphoenix\u002Ftracing\u002Fintegrations-tracing\u002Flangchain) | `@arizeai\u002Fopeninference-instrumentation-langchain` | [![NPM 版本](https:\u002F\u002Fimg.shields.io\u002Fnpm\u002Fv\u002F@arizeai\u002Fopeninference-instrumentation-langchain.svg)](https:\u002F\u002Fwww.npmjs.com\u002Fpackage\u002F@arizeai\u002Fopeninference-instrumentation-langchain) |\n| \u003Cpicture>\u003Csource media=\"(prefers-color-scheme: dark)\" srcset=\"https:\u002F\u002Funpkg.com\u002F@lobehub\u002Ficons-static-png@latest\u002Fdark\u002Fvercel.png\">\u003Cimg height=\"14\" src=\"https:\u002F\u002Foss.gittoolsai.com\u002Fimages\u002FArize-ai_phoenix_readme_75660b5aa9b0.png\">\u003C\u002Fpicture> | [Vercel AI SDK](https:\u002F\u002Farize.com\u002Fdocs\u002Fphoenix\u002Ftracing\u002Fintegrations-tracing\u002Fvercel-ai-sdk) | `@arizeai\u002Fopeninference-vercel` | [![NPM 版本](https:\u002F\u002Fimg.shields.io\u002Fnpm\u002Fv\u002F@arizeai\u002Fopeninference-vercel.svg)](https:\u002F\u002Fwww.npmjs.com\u002Fpackage\u002F@arizeai\u002Fopeninference-vercel) |\n| | [BeeAI](https:\u002F\u002Farize.com\u002Fdocs\u002Fphoenix\u002Ftracing\u002Fintegrations-tracing\u002Fbeeai) | `@arizeai\u002Fopeninference-instrumentation-beeai` | [![NPM 版本](https:\u002F\u002Fimg.shields.io\u002Fnpm\u002Fv\u002F@arizeai\u002Fopeninference-vercel.svg)](https:\u002F\u002Fwww.npmjs.com\u002Fpackage\u002F@arizeai\u002Fopeninference-instrumentation-beeai) |\n| \u003Cimg src=\"https:\u002F\u002Foss.gittoolsai.com\u002Fimages\u002FArize-ai_phoenix_readme_6ef465e6e627.png\" height=\"14\"> | [Claude Agent SDK](https:\u002F\u002Farize.com\u002Fdocs\u002Fphoenix\u002Fintegrations\u002Ftypescript\u002Fclaude-agent-sdk) | `@arizeai\u002Fopeninference-instrumentation-claude-agent-sdk` | [![NPM 版本](https:\u002F\u002Fimg.shields.io\u002Fnpm\u002Fv\u002F@arizeai%2Fopeninference-instrumentation-claude-agent-sdk.svg)](https:\u002F\u002Fwww.npmjs.com\u002Fpackage\u002F@arizeai\u002Fopeninference-instrumentation-claude-agent-sdk) |\n| \u003Cpicture>\u003Csource media=\"(prefers-color-scheme: dark)\" srcset=\"https:\u002F\u002Funpkg.com\u002F@lobehub\u002Ficons-static-png@latest\u002Fdark\u002Fmastra.png\">\u003Cimg height=\"14\" src=\"https:\u002F\u002Foss.gittoolsai.com\u002Fimages\u002FArize-ai_phoenix_readme_da5502535cec.png\">\u003C\u002Fpicture> | [Mastra](https:\u002F\u002Farize.com\u002Fdocs\u002Fphoenix\u002Fintegrations\u002Ftypescript\u002Fmastra) | `@mastra\u002Farize` | [![NPM 版本](https:\u002F\u002Fimg.shields.io\u002Fnpm\u002Fv\u002F@mastra\u002Farize.svg)](https:\u002F\u002Fwww.npmjs.com\u002Fpackage\u002F@mastra\u002Farize) |\n| \u003Cpicture>\u003Csource media=\"(prefers-color-scheme: dark)\" srcset=\"https:\u002F\u002Funpkg.com\u002F@lobehub\u002Ficons-static-png@latest\u002Fdark\u002Fmcp.png\">\u003Cimg height=\"14\" src=\"https:\u002F\u002Foss.gittoolsai.com\u002Fimages\u002FArize-ai_phoenix_readme_53956c4cd629.png\">\u003C\u002Fpicture> | [MCP](https:\u002F\u002Farize.com\u002Fdocs\u002Fphoenix\u002Fintegrations\u002Ftypescript\u002Fmcp) | `@arizeai\u002Fopeninference-instrumentation-mcp` | [![NPM 版本](https:\u002F\u002Fimg.shields.io\u002Fnpm\u002Fv\u002F@arizeai%2Fopeninference-instrumentation-mcp.svg)](https:\u002F\u002Fwww.npmjs.com\u002Fpackage\u002F@arizeai\u002Fopeninference-instrumentation-mcp) |\n\n### Java 集成\n\n| | 集成 | 包 | 版本 |\n|:---:|---|---|---|\n| \u003Cimg src=\"https:\u002F\u002Foss.gittoolsai.com\u002Fimages\u002FArize-ai_phoenix_readme_acb5f98fc60a.png\" height=\"14\"> | [LangChain4j](https:\u002F\u002Fgithub.com\u002FArize-ai\u002Fopeninference\u002Ftree\u002Fmain\u002Fjava\u002Finstrumentation\u002Fopeninference-instrumentation-langchain4j) | `openinference-instrumentation-langchain4j` | [![Maven Central](https:\u002F\u002Fimg.shields.io\u002Fmaven-central\u002Fv\u002Fcom.arize\u002Fopeninference-instrumentation-langchain4j.svg)](https:\u002F\u002Fcentral.sonatype.com\u002Fartifact\u002Fcom.arize\u002Fopeninference-instrumentation-langchain4j) |\n| | SpringAI | `openinference-instrumentation-springAI` | [![Maven Central](https:\u002F\u002Fimg.shields.io\u002Fmaven-central\u002Fv\u002Fcom.arize\u002Fopeninference-instrumentation-springAI.svg)](https:\u002F\u002Fcentral.sonatype.com\u002Fartifact\u002Fcom.arize\u002Fopeninference-instrumentation-springAI) |\n| | [Arconia](https:\u002F\u002Farconia.io\u002Fdocs\u002Farconia\u002Flatest\u002Fobservability\u002Fgenerative-ai\u002F) | `openinference-instrumentation-springAI` | [![Maven Central](https:\u002F\u002Fimg.shields.io\u002Fmaven-central\u002Fv\u002Fcom.arize\u002Fopeninference-instrumentation-springAI.svg)](https:\u002F\u002Fcentral.sonatype.com\u002Fartifact\u002Fcom.arize\u002Fopeninference-instrumentation-springAI) |\n\n### 平台\n\n| | 平台 | 描述 | 文档 |\n|:---:|---|---|---|\n| | [BeeAI](https:\u002F\u002Fdocs.beeai.dev\u002Fobservability\u002Fagents-traceability) | 内置可观测性的 AI 代理框架 | [集成指南](https:\u002F\u002Fdocs.beeai.dev\u002Fobservability\u002Fagents-traceability) |\n| \u003Cimg src=\"https:\u002F\u002Foss.gittoolsai.com\u002Fimages\u002FArize-ai_phoenix_readme_8e9ca8ab7fe0.png\" height=\"14\"> | [Dify](https:\u002F\u002Fdocs.dify.ai\u002Fen\u002Fguides\u002Fmonitoring\u002Fintegrate-external-ops-tools\u002Fintegrate-phoenix) | 开源 LLM 应用开发平台 | [集成指南](https:\u002F\u002Fdocs.dify.ai\u002Fen\u002Fguides\u002Fmonitoring\u002Fintegrate-external-ops-tools\u002Fintegrate-phoenix) |\n| | [Envoy AI Gateway](https:\u002F\u002Fgithub.com\u002Fenvoyproxy\u002Fai-gateway) | 基于 Envoy Proxy 构建的用于 AI 工作负载的 AI 网关 | [集成指南](https:\u002F\u002Fgithub.com\u002Fenvoyproxy\u002Fai-gateway\u002Ftree\u002Fmain\u002Fcmd\u002Faigw#opentelemetry-setup-with-phoenix) |\n| | [LangFlow](https:\u002F\u002Farize.com\u002Fdocs\u002Fphoenix\u002Ftracing\u002Fintegrations-tracing\u002Flangflow) | 用于构建多代理和 RAG 应用程序的可视化框架 | [集成指南](https:\u002F\u002Farize.com\u002Fdocs\u002Fphoenix\u002Ftracing\u002Fintegrations-tracing\u002Flangflow) |\n| | [LiteLLM Proxy](https:\u002F\u002Fdocs.litellm.ai\u002Fdocs\u002Fobservability\u002Fphoenix_integration#using-with-litellm-proxy) | LLM 的代理服务器 | [集成指南](https:\u002F\u002Fdocs.litellm.ai\u002Fdocs\u002Fobservability\u002Fphoenix_integration#using-with-litellm-proxy) |\n| | [Flowise](https:\u002F\u002Farize.com\u002Fdocs\u002Fphoenix\u002Fintegrations\u002Fplatforms\u002Fflowise) | 用于构建 LLM 应用程序的可视化框架 | [集成指南](https:\u002F\u002Farize.com\u002Fdocs\u002Fphoenix\u002Fintegrations\u002Fplatforms\u002Fflowise) |\n| | [Prompt Flow](https:\u002F\u002Farize.com\u002Fdocs\u002Fphoenix\u002Fintegrations\u002Fplatforms\u002Fprompt-flow) | 微软的提示流程编排工具 | [集成指南](https:\u002F\u002Farize.com\u002Fdocs\u002Fphoenix\u002Fintegrations\u002Fplatforms\u002Fprompt-flow) |\n| \u003Cimg src=\"https:\u002F\u002Foss.gittoolsai.com\u002Fimages\u002FArize-ai_phoenix_readme_6c74605c5569.png\" height=\"14\"> | [NVIDIA NeMo](https:\u002F\u002Farize.com\u002Fdocs\u002Fphoenix\u002Fintegrations\u002Fpython\u002Fnvidia) | NVIDIA NeMo 企业级代理工具包 | [集成指南](https:\u002F\u002Farize.com\u002Fdocs\u002Fphoenix\u002Fintegrations\u002Fpython\u002Fnvidia) |\n| | [Graphite](https:\u002F\u002Farize.com\u002Fdocs\u002Fphoenix\u002Fintegrations\u002Fpython\u002Fgraphite) | 具有可视化构建器的多代理 LLM 工作流框架 | [集成指南](https:\u002F\u002Farize.com\u002Fdocs\u002Fphoenix\u002Fintegrations\u002Fpython\u002Fgraphite) |\n\n## 编码代理技能\n\n此仓库包含[技能](https:\u002F\u002Fdocs.anthropic.com\u002Fen\u002Fdocs\u002Fclaude-code\u002Fskills)，可教导编码代理如何使用 Phoenix。这些技能位于[`.agents\u002Fskills\u002F`](.agents\u002Fskills\u002F)，可与 Claude Code、Cursor 及其他兼容工具一起使用。\n\n| 技能 | 描述 |\n| ----- | ----------- |\n| [phoenix-cli](.agents\u002Fskills\u002Fphoenix-cli\u002F) | 使用 Phoenix CLI 调试 LLM 应用程序——获取跟踪记录、分析错误、查看实验并查询 GraphQL API |\n| [phoenix-evals](.agents\u002Fskills\u002Fphoenix-evals\u002F) | 使用 Phoenix 构建和运行 AI\u002FLLM 应用程序的评估工具 |\n| [phoenix-tracing](.agents\u002Fskills\u002Fphoenix-tracing\u002F) | OpenInference 语义规范及用于跟踪 LLM 应用程序的仪器化 |\n\n## 安全与隐私\n\n我们非常重视数据安全和隐私。有关详细信息，请参阅我们的[安全与隐私文档](https:\u002F\u002Farize.com\u002Fdocs\u002Fphoenix\u002Fself-hosting\u002Fsecurity\u002Fprivacy)。\n\n### 遥测\n\n默认情况下，Phoenix 会收集基本的 Web 分析数据（例如页面浏览量、UI 交互），以帮助我们了解 Phoenix 的使用情况并改进产品。**我们绝不会收集您的任何跟踪数据、评估结果或任何敏感信息。**\n\n您可以通过设置环境变量 `PHOENIX_TELEMETRY_ENABLED=false` 来选择退出遥测。\n\n## 社区\n\n加入我们的社区，与数千名 AI 构建者建立联系。\n\n- 🌍 加入我们的 [Slack 社区](https:\u002F\u002Fjoin.slack.com\u002Ft\u002Farize-ai\u002Fshared_invite\u002Fzt-3r07iavnk-ammtATWSlF0pSrd1DsMW7g)。\n- 📚 阅读我们的[文档](https:\u002F\u002Farize.com\u002Fdocs\u002Fphoenix)。\n- 💡 在 _#phoenix-support_ 频道中提问并提供反馈。\n- 🌟 在我们的 [GitHub](https:\u002F\u002Fgithub.com\u002FArize-ai\u002Fphoenix) 上留下一颗星。\n- 🐞 使用 [GitHub Issues](https:\u002F\u002Fgithub.com\u002FArize-ai\u002Fphoenix\u002Fissues) 报告 bug。\n- 𝕏 在 [𝕏](https:\u002F\u002Ftwitter.com\u002FArizePhoenix) 上关注我们。\n- 🗺️ 查看我们的[路线图](https:\u002F\u002Fgithub.com\u002Forgs\u002FArize-ai\u002Fprojects\u002F45)，了解我们接下来的发展方向。\n- 🧑‍🏫 在 Arize 的学习中心深入了解[代理](http:\u002F\u002Farize.com\u002Fai-agents\u002F)和[LLM 评估](https:\u002F\u002Farize.com\u002Fllm-evaluation)的相关内容。\n\n## 重大变更\n\n有关重大变更列表，请参阅[迁移指南](.\u002FMIGRATION.md)。\n\n## 版权、专利与许可\n\n版权所有 © 2025 Arize AI, Inc. 保留所有权利。\n\n本代码的部分内容受一项或多件美国专利保护。请参阅[IP_NOTICE](https:\u002F\u002Fgithub.com\u002FArize-ai\u002Fphoenix\u002Fblob\u002Fmain\u002FIP_NOTICE)。\n\n本软件根据 Elastic License 2.0 (ELv2) 的条款进行许可。请参阅[LICENSE](https:\u002F\u002Fgithub.com\u002FArize-ai\u002Fphoenix\u002Fblob\u002Fmain\u002FLICENSE)。","# Phoenix 快速上手指南\n\nPhoenix 是一个开源的 AI 可观测性平台，专为大语言模型（LLM）应用的实验、评估和故障排查而设计。它支持链路追踪（Tracing）、自动化评估（Evaluation）、数据集管理、提示词工程等功能，且与主流开发框架和模型提供商无关。\n\n## 环境准备\n\n在开始之前，请确保您的开发环境满足以下要求：\n\n*   **操作系统**：Linux、macOS 或 Windows（推荐 WSL2）。\n*   **Python 版本**：Python 3.9 或更高版本。\n*   **包管理器**：已安装 `pip` 或 `conda`。\n*   **可选依赖**：若需使用容器化部署，请确保已安装 Docker。\n\n> **提示**：国内开发者若遇到 PyPI 下载速度慢的问题，建议在安装命令中指定清华或阿里云镜像源。\n\n## 安装步骤\n\n您可以选择通过 `pip` 或 `conda` 进行安装。\n\n### 方式一：使用 pip 安装（推荐）\n\n**标准安装：**\n```bash\npip install arize-phoenix\n```\n\n**使用国内镜像加速安装（推荐国内用户）：**\n```bash\npip install arize-phoenix -i https:\u002F\u002Fpypi.tuna.tsinghua.edu.cn\u002Fsimple\n```\n\n### 方式二：使用 conda 安装\n\n```bash\nconda install -c conda-forge arize-phoenix\n```\n\n### 方式三：Docker 部署\n\n如果您希望独立部署服务而不占用本地 Python 环境，可以使用 Docker：\n\n```bash\ndocker run --rm -it -p 6006:6006 arizephoenix\u002Fphoenix\n```\n\n## 基本使用\n\n安装完成后，您可以通过以下几种最简单的方式启动 Phoenix 并开始观测您的 LLM 应用。\n\n### 1. 命令行启动（独立服务）\n\n在终端直接运行以下命令，Phoenix 将在本地启动一个 Web 界面（默认端口 6006）：\n\n```bash\nphoenix serve\n```\n\n启动后，浏览器访问 `http:\u002F\u002Flocalhost:6006` 即可查看仪表盘。\n\n### 2. 在 Jupyter Notebook 中使用\n\n如果您在 Notebook 中进行实验，可以直接导入并启动：\n\n```python\nimport phoenix as px\n\n# 启动 Phoenix 实例\nsession = px.launch_app()\n```\n\n### 3. 集成追踪示例\n\nPhoenix 基于 OpenTelemetry，支持自动追踪主流框架（如 OpenAI, LangChain, LlamaIndex 等）。以下是一个结合 OpenAI SDK 的最小化追踪示例：\n\n**前置准备：**\n确保已安装对应的集成包（以 OpenAI 为例）：\n```bash\npip install arize-phoenix-otel openai\n```\n\n**代码示例：**\n\n```python\nimport os\nfrom openai import OpenAI\nfrom phoenix.otel import register\n\n# 1. 注册 Phoenix 追踪器 (会自动配置 OpenTelemetry)\nregister(project_name=\"my-first-project\")\n\n# 2. 初始化您的 LLM 客户端\nclient = OpenAI(api_key=os.environ[\"OPENAI_API_KEY\"])\n\n# 3. 执行调用，Phoenix 将自动记录链路数据\nresponse = client.chat.completions.create(\n    model=\"gpt-3.5-turbo\",\n    messages=[{\"role\": \"user\", \"content\": \"Hello, Phoenix!\"}]\n)\n\nprint(response.choices[0].message.content)\n```\n\n运行上述代码后，返回 Phoenix Web 界面 (`http:\u002F\u002Flocalhost:6006`)，您即可看到完整的调用链路、延迟分析以及 Token 消耗情况。\n\n### 4. 连接云端服务（可选）\n\n除了本地运行，您也可以将数据发送到 Arize 云端实例：\n\n```python\nimport phoenix as px\n\n# 连接到云端实例\npx.connect_to_cloud(api_key=\"YOUR_API_KEY\", workspace_url=\"https:\u002F\u002Fapp.phoenix.arize.com\")\n```","某电商公司的算法团队正在优化其智能客服大模型，试图提升回答准确率并降低幻觉风险。\n\n### 没有 phoenix 时\n- **故障排查如“盲人摸象”**：当用户反馈模型胡编乱造时，开发人员只能翻阅海量原始日志，难以定位是哪一次检索增强生成（RAG）环节出现了数据污染。\n- **评估依赖人工抽样**：每次模型迭代后，测试人员需人工阅读数百条对话来打分，耗时数天且标准主观，无法量化细微的性能波动。\n- **缺乏细粒度归因**：发现回答质量下降时，无法区分是底层 Embedding 模型失效、检索片段不相关，还是提示词工程出了问题。\n- **线上问题响应滞后**：往往等到大量客诉爆发后才察觉异常，缺乏实时的质量监控仪表盘来预警潜在的分布漂移。\n\n### 使用 phoenix 后\n- **全链路追踪可视化**：通过 phoenix 的分布式追踪功能，一键还原问题对话的完整执行路径，精准定位到具体的检索片段或生成步骤。\n- **自动化批量评估**：利用内置的 LLM 评判器，几分钟内即可完成对数千条历史对话的自动评分，快速对比不同版本模型的优劣。\n- **多维根因分析**：借助嵌入空间（Embedding Space）可视化，直观发现聚类异常的数据点，迅速识别出是检索库噪声还是模型理解偏差。\n- **实时质量监控**：部署在线仪表盘实时监控延迟、令牌消耗及回答质量分数，一旦指标偏离阈值立即触发告警，将事故扼杀在萌芽期。\n\nphoenix 将原本黑盒的大模型运行过程转化为透明、可度量、可优化的工程闭环，显著提升了团队迭代效率与系统稳定性。","https:\u002F\u002Foss.gittoolsai.com\u002Fimages\u002FArize-ai_phoenix_f6d8660b.jpg","Arize-ai","Arize AI","https:\u002F\u002Foss.gittoolsai.com\u002Favatars\u002FArize-ai_f55f298b.jpg","Unified AI engineering and evaluation platform to accelerate development and improvement of AI apps and agents",null,"support@arize.com","arizeai","https:\u002F\u002Fwww.arize.com","https:\u002F\u002Fgithub.com\u002FArize-ai",[24,28,32,36,40,43,47,50,53,56],{"name":25,"color":26,"percentage":27},"Python","#3572A5",47.1,{"name":29,"color":30,"percentage":31},"TypeScript","#3178c6",35.1,{"name":33,"color":34,"percentage":35},"Jupyter Notebook","#DA5B0B",17,{"name":37,"color":38,"percentage":39},"Shell","#89e051",0.2,{"name":41,"color":42,"percentage":39},"JavaScript","#f1e05a",{"name":44,"color":45,"percentage":46},"PLpgSQL","#336790",0.1,{"name":48,"color":49,"percentage":46},"Makefile","#427819",{"name":51,"color":52,"percentage":46},"CSS","#663399",{"name":54,"color":55,"percentage":46},"HTML","#e34c26",{"name":57,"color":58,"percentage":46},"Go Template","#00ADD8",9303,821,"2026-04-16T07:35:33","NOASSERTION",1,"Linux, macOS, Windows","未说明",{"notes":67,"python":68,"dependencies":69},"该工具是厂商和语言无关的 AI 可观测性平台，支持本地机器、Jupyter Notebook、容器化部署（Docker\u002FKubernetes）及云端运行。可通过 pip 或 conda 安装。支持多种主流框架（如 LangChain, LlamaIndex, OpenAI Agents SDK 等）和 LLM 提供商的自动插桩。若需轻量级集成，可单独安装提供的 Python 子包或 TypeScript 包。","3.9+",[70,71,72,73,74],"arize-phoenix","OpenTelemetry","arize-phoenix-otel","arize-phoenix-client","arize-phoenix-evals",[76,77,78,79,80],"Agent","数据工具","其他","语言模型","开发框架",[82,83,84,85,86,87,88,89,90,91,92,93,94,95,96,97],"llmops","ai-monitoring","ai-observability","llm-eval","aiengineering","datasets","agents","llms","prompt-engineering","anthropic","evals","llm-evaluation","openai","langchain","llamaindex","smolagents",2,"ready","2026-03-27T02:49:30.150509","2026-04-17T09:54:36.061444",[103,108,113,118,122,126],{"id":104,"question_zh":105,"answer_zh":106,"source_url":107},36496,"为什么在使用 DSPy 连接器时，Phoenix UI 中的 Token 计数显示为 0？","这通常是因为 DSPy 的缓存功能被启用了，导致没有实际调用 LLM，因此 Token 数为 0。解决方法是禁用缓存：在导入 dspy 之前设置环境变量 `os.environ[\"DSP_CACHEBOOL\"] = \"False\"`。此外，对于 `dspy-ai>=2.5` 版本，DSPy 底层使用 LiteLLM，建议改用 `LiteLLMInstrumentor` 进行插桩以获取准确的 Token 计数。","https:\u002F\u002Fgithub.com\u002FArize-ai\u002Fphoenix\u002Fissues\u002F3119",{"id":109,"question_zh":110,"answer_zh":111,"source_url":112},36497,"遇到 ImportError: cannot import name '_ExtendedAttributes' 错误该如何解决？","该错误通常由 OpenTelemetry 版本不兼容引起（特别是版本 1.22.0 及以上）。解决方案是将 OpenTelemetry 相关包固定到 1.21.0 版本。请检查您的依赖项，确保所有 OpenTelemetry 包版本一致且兼容，例如运行 `pip install opentelemetry-api==1.21.0 opentelemetry-sdk==1.21.0` 等命令进行降级。","https:\u002F\u002Fgithub.com\u002FArize-ai\u002Fphoenix\u002Fissues\u002F8305",{"id":114,"question_zh":115,"answer_zh":116,"source_url":117},36498,"在 Span 详情视图中删除标注后，编辑器界面为何没有实时更新？","这是一个已知的前端刷新问题。当您在 \"annotation\" 表中删除标注（例如删除值为 \"yes\" 的 \"correct\" 标注）后，编辑器不会立即清除旧值。目前的临时解决方法是手动刷新页面，或者编辑另一个标注以触发界面的重新渲染。","https:\u002F\u002Fgithub.com\u002FArize-ai\u002Fphoenix\u002Fissues\u002F5917",{"id":119,"question_zh":120,"answer_zh":121,"source_url":117},36499,"Span 详情中的标注元数据（Metadata）列显示异常或过宽怎么办？","这是因为 \"annotation\" 表中的元数据列未对内容长度进行约束。当存储较大的 JSON 数据块时，会导致列宽异常扩展。目前建议在记录标注时尽量精简元数据内容，避免存入过大的 JSON 对象，直到后续版本对该列宽进行样式优化或限制。",{"id":123,"question_zh":124,"answer_zh":125,"source_url":107},36500,"如何正确配置 DSPy 和 Phoenix 以同时支持 OpenAI 和其他模型的 Token 追踪？","如果您使用的是 `dspy-ai>=2.5`，请通过 `LiteLLMInstrumentor` 进行插桩，因为新版 DSPy 底层使用 LiteLLM。配置代码示例：导入 `LiteLLMInstrumentor` 并调用 `.instrument()`。同时，确保设置了正确的 Phoenix 端点（默认为 `http:\u002F\u002Flocalhost:6006\u002Fv1\u002Ftraces` 或通过 `PHOENIX_COLLECTOR_ENDPOINT` 环境变量指定），并禁用了 DSPy 缓存以避免 Token 统计为零。",{"id":127,"question_zh":128,"answer_zh":129,"source_url":112},36501,"在容器化环境中部署 Phoenix 时遇到依赖冲突如何处理？","在 Linux 容器中使用 Phoenix 时，若遇到如 `_ExtendedAttributes` 导入错误，通常是由于基础镜像或环境中预装的 OpenTelemetry 版本过高。建议在 `requirements.txt` 或 Dockerfile 中明确锁定 OpenTelemetry 系列包的版本为 1.21.0，并确保 Python 版本（如 3.11）与所选的 Phoenix 版本兼容。",[131,136,141,146,151,156,161,166,171,176,181,186,191,196,201,206,211,216,221,226],{"id":132,"version":133,"summary_zh":134,"released_at":135},289320,"arize-phoenix-v14.6.0","## [14.6.0](https:\u002F\u002Fgithub.com\u002FArize-ai\u002Fphoenix\u002Fcompare\u002Farize-phoenix-v14.5.0...arize-phoenix-v14.6.0)（2026-04-15）\n\n\n### 功能特性\n\n* **agent:** 修复 PXI 系统提示中的文档 URL 格式，并添加评估框架（[#12677](https:\u002F\u002Fgithub.com\u002FArize-ai\u002Fphoenix\u002Fissues\u002F12677)）（[5dd794b](https:\u002F\u002Fgithub.com\u002FArize-ai\u002Fphoenix\u002Fcommit\u002F5dd794bbd6d4153b8a1ea0c6ead881e93fff95f0)）\n* **app:** 会话轮次的列表-详情布局（[#12674](https:\u002F\u002Fgithub.com\u002FArize-ai\u002Fphoenix\u002Fissues\u002F12674)）（[f5951fb](https:\u002F\u002Fgithub.com\u002FArize-ai\u002Fphoenix\u002Fcommit\u002Ff5951fbf15edaf79c65da37a363113a692a2882e)）\n\n\n### 错误修复\n\n* **ci:** 使 sync-models 自包含，添加 Slack 告警，并刷新模型列表（[#12679](https:\u002F\u002Fgithub.com\u002FArize-ai\u002Fphoenix\u002Fissues\u002F12679)）（[b96d21f](https:\u002F\u002Fgithub.com\u002FArize-ai\u002Fphoenix\u002Fcommit\u002Fb96d21f8a90dba890bca33864a509d8fcddbc0e3)）\n* **playground:** 将评估失败与完成结果分开统计（[#12675](https:\u002F\u002Fgithub.com\u002FArize-ai\u002Fphoenix\u002Fissues\u002F12675)）（[03172f3](https:\u002F\u002Fgithub.com\u002FArize-ai\u002Fphoenix\u002Fcommit\u002F03172f3fe957ba27b169dc10fef496777613e8f7)）","2026-04-15T16:12:59",{"id":137,"version":138,"summary_zh":139,"released_at":140},289321,"arize-phoenix-v14.5.0","## [14.5.0](https:\u002F\u002Fgithub.com\u002FArize-ai\u002Fphoenix\u002Fcompare\u002Farize-phoenix-v14.4.0...arize-phoenix-v14.5.0) (2026-04-14)\n\n\n### 功能特性\n\n* 为助手消息添加聊天元数据（[#12665](https:\u002F\u002Fgithub.com\u002FArize-ai\u002Fphoenix\u002Fissues\u002F12665)）（[a6f0f7f](https:\u002F\u002Fgithub.com\u002FArize-ai\u002Fphoenix\u002Fcommit\u002Fa6f0f7f7216ca675d9307f2afcfcee07dc537f28)）\n* 在 Phoenix CLI 中添加 Span 和 Trace 注解命令（[#12664](https:\u002F\u002Fgithub.com\u002FArize-ai\u002Fphoenix\u002Fissues\u002F12664)）（[c04a0c3](https:\u002F\u002Fgithub.com\u002FArize-ai\u002Fphoenix\u002Fcommit\u002Fc04a0c3ea1be246b8b91d40a64912f1879719da4)）\n* **agent:** 为助手消息添加反馈操作（[#12670](https:\u002F\u002Fgithub.com\u002FArize-ai\u002Fphoenix\u002Fissues\u002F12670)）（[b9af91a](https:\u002F\u002Fgithub.com\u002FArize-ai\u002Fphoenix\u002Fcommit\u002Fb9af91aeb37ca5648e20f92df6b6417d0b0d3dc6)）\n* **agent:** 从设置中切换助手代理的启用与禁用（[#12654](https:\u002F\u002Fgithub.com\u002FArize-ai\u002Fphoenix\u002Fissues\u002F12654)）（[4a6b1c7](https:\u002F\u002Fgithub.com\u002FArize-ai\u002Fphoenix\u002Fcommit\u002F4a6b1c79eb7f86bd4cce590ec50157c766e4f5cf)）\n\n\n### Bug 修复\n\n* pxi 技能 YAML 格式错误（[#12662](https:\u002F\u002Fgithub.com\u002FArize-ai\u002Fphoenix\u002Fissues\u002F12662)）（[fd82f2b](https:\u002F\u002Fgithub.com\u002FArize-ai\u002Fphoenix\u002Fcommit\u002Ffd82f2b2c753925534a3dfc2416db5f88ee24565)）\n\n\n### 文档更新\n\n* 添加 ATIF 轨迹导入文档（[#12631](https:\u002F\u002Fgithub.com\u002FArize-ai\u002Fphoenix\u002Fissues\u002F12631)）（[ef9e6cc](https:\u002F\u002Fgithub.com\u002FArize-ai\u002Fphoenix\u002Fcommit\u002Fef9e6cc12dfe54136f2f90b83718f7b98bb2597c)）\n* 在 AGENTS.md 中添加 Python 预发布依赖版本策略（[#12659](https:\u002F\u002Fgithub.com\u002FArize-ai\u002Fphoenix\u002Fissues\u002F12659)）（[4bdaf93](https:\u002F\u002Fgithub.com\u002FArize-ai\u002Fphoenix\u002Fcommit\u002F4bdaf93f8c118ad3e1e10dadfa120f1d3aeb31bd))","2026-04-14T20:49:38",{"id":142,"version":143,"summary_zh":144,"released_at":145},289322,"arize-phoenix-v14.4.0","## [14.4.0](https:\u002F\u002Fgithub.com\u002FArize-ai\u002Fphoenix\u002Fcompare\u002Farize-phoenix-v14.3.1...arize-phoenix-v14.4.0)（2026-04-14）\n\n\n### 功能特性\n\n* **agent:** 能力菜单 + 重构 PXI 前端工具，使其围绕能力注册表进行连接（[#12590](https:\u002F\u002Fgithub.com\u002FArize-ai\u002Fphoenix\u002Fissues\u002F12590)）（[0518f16](https:\u002F\u002Fgithub.com\u002FArize-ai\u002Fphoenix\u002Fcommit\u002F0518f1632dacd6df7e74e914a3bfdd4047a2f8a1)）\n\n\n### 错误修复\n\n* 移除 graphql-core 上的固定版本依赖（[#12657](https:\u002F\u002Fgithub.com\u002FArize-ai\u002Fphoenix\u002Fissues\u002F12657)）（[21b4c27](https:\u002F\u002Fgithub.com\u002FArize-ai\u002Fphoenix\u002Fcommit\u002F21b4c27c4de6755c4f72a7841a475403afad4517)）","2026-04-14T15:53:08",{"id":147,"version":148,"summary_zh":149,"released_at":150},289323,"arize-phoenix-v14.3.1","## [14.3.1](https:\u002F\u002Fgithub.com\u002FArize-ai\u002Fphoenix\u002Fcompare\u002Farize-phoenix-v14.3.0...arize-phoenix-v14.3.1) (2026-04-14)\n\n\n### 错误修复\n\n* 加强 authFetch 的安全性，并将其添加到数据集上传路径中 ([#12616](https:\u002F\u002Fgithub.com\u002FArize-ai\u002Fphoenix\u002Fissues\u002F12616)) ([62590d2](https:\u002F\u002Fgithub.com\u002FArize-ai\u002Fphoenix\u002Fcommit\u002F62590d25d1d8da90a0db2ed19d1291a8a3462220))","2026-04-14T01:28:14",{"id":152,"version":153,"summary_zh":154,"released_at":155},289324,"arize-phoenix-v14.3.0","## [14.3.0](https:\u002F\u002Fgithub.com\u002FArize-ai\u002Fphoenix\u002Fcompare\u002Farize-phoenix-v14.2.1...arize-phoenix-v14.3.0)（2026-04-14）\n\n\n### 功能特性\n\n* **agent:** 助理代理设置页面（[#12637](https:\u002F\u002Fgithub.com\u002FArize-ai\u002Fphoenix\u002Fissues\u002F12637)）（[bca2318](https:\u002F\u002Fgithub.com\u002FArize-ai\u002Fphoenix\u002Fcommit\u002Fbca23186770e6bf26489704e3bfe33bed42ce838)）\n* 通过文档重新导出 phoenix-otel 中的 openinference-core（[#12541](https:\u002F\u002Fgithub.com\u002FArize-ai\u002Fphoenix\u002Fissues\u002F12541)）（[1449f3d](https:\u002F\u002Fgithub.com\u002FArize-ai\u002Fphoenix\u002Fcommit\u002F1449f3db622915f9813bcc9b4f1d7d588204d708)）\n\n\n### 错误修复\n\n* 将自定义标量迁移到 strawberry-graphql 0.314.3 的 scalar_map 中（[#12568](https:\u002F\u002Fgithub.com\u002FArize-ai\u002Fphoenix\u002Fissues\u002F12568)）（[3488aa7](https:\u002F\u002Fgithub.com\u002FArize-ai\u002Fphoenix\u002Fcommit\u002F3488aa72f828f3ca72c4ddf5ec0d95f4f2035061)）\n\n\n### 文档更新\n\n* 每周文档缺失情况审计（2026-04-02 → 2026-04-09）（[#12606](https:\u002F\u002Fgithub.com\u002FArize-ai\u002Fphoenix\u002Fissues\u002F12606)）（[6372cfb](https:\u002F\u002Fgithub.com\u002FArize-ai\u002Fphoenix\u002Fcommit\u002F6372cfb59b38039fbad4dfd0c7deec7597793b66)）","2026-04-14T00:58:22",{"id":157,"version":158,"summary_zh":159,"released_at":160},289325,"arize-phoenix-v14.2.1","## [14.2.1](https:\u002F\u002Fgithub.com\u002FArize-ai\u002Fphoenix\u002Fcompare\u002Farize-phoenix-v14.2.0...arize-phoenix-v14.2.1)（2026-04-10）\n\n\n### 错误修复\n\n* 移除 GraphQL 订阅的 WebSocket 支持，改用 HTTP 多部分请求（[#12610](https:\u002F\u002Fgithub.com\u002FArize-ai\u002Fphoenix\u002Fissues\u002F12610)）（[ed77fea](https:\u002F\u002Fgithub.com\u002FArize-ai\u002Fphoenix\u002Fcommit\u002Fed77feafa4877ed01f89555002e4be372d3aa4ac)）","2026-04-10T15:07:29",{"id":162,"version":163,"summary_zh":164,"released_at":165},289326,"arize-phoenix-v14.2.0","## [14.2.0](https:\u002F\u002Fgithub.com\u002FArize-ai\u002Fphoenix\u002Fcompare\u002Farize-phoenix-v14.1.1...arize-phoenix-v14.2.0)（2026-04-10）\n\n\n### 功能特性\n\n* 添加 `\u002Fredirects\u002Fprojects\u002F:project_name`，用于基于名称的项目 URL（[#12593](https:\u002F\u002Fgithub.com\u002FArize-ai\u002Fphoenix\u002Fissues\u002F12593)）（[afd9584](https:\u002F\u002Fgithub.com\u002FArize-ai\u002Fphoenix\u002Fcommit\u002Fafd95840791251c2489ce2d83be6c50563148dfd)）\n\n\n### 文档\n\n* 添加 2026年4月2日至4月7日的 Phoenix 发布说明（[#12588](https:\u002F\u002Fgithub.com\u002FArize-ai\u002Fphoenix\u002Fissues\u002F12588)）（[3a55bae](https:\u002F\u002Fgithub.com\u002FArize-ai\u002Fphoenix\u002Fcommit\u002F3a55bae1c8c98882207bd632b9568c234ce125e8)）","2026-04-10T01:50:50",{"id":167,"version":168,"summary_zh":169,"released_at":170},289327,"arize-phoenix-v14.1.1","## [14.1.1](https:\u002F\u002Fgithub.com\u002FArize-ai\u002Fphoenix\u002Fcompare\u002Farize-phoenix-v14.1.0...arize-phoenix-v14.1.1) (2026-04-08)\n\n\n### Bug修复\n\n* 会话注释使用数据加载器，以防止连接耗尽（[#12585](https:\u002F\u002Fgithub.com\u002FArize-ai\u002Fphoenix\u002Fissues\u002F12585)）（[8549fde](https:\u002F\u002Fgithub.com\u002FArize-ai\u002Fphoenix\u002Fcommit\u002F8549fde5ba66a2e98d9f4d8bd5e7cc205212014d)）","2026-04-08T14:59:19",{"id":172,"version":173,"summary_zh":174,"released_at":175},289328,"arize-phoenix-v14.1.0","## [14.1.0](https:\u002F\u002Fgithub.com\u002FArize-ai\u002Fphoenix\u002Fcompare\u002Farize-phoenix-v14.0.0...arize-phoenix-v14.1.0)（2026-04-08）\n\n### 错误修复\n\n* 为创建跨度的 API 端点添加有效的 OpenAPI 示例 ([#12371](https:\u002F\u002Fgithub.com\u002FArize-ai\u002Fphoenix\u002Fissues\u002F12371)) ([d74be37](https:\u002F\u002Fgithub.com\u002FArize-ai\u002Fphoenix\u002Fcommit\u002Fd74be37cefff79313e216ce1ef80993e3c645345))\n* **agent：** 展示后端工具循环耗尽问题 ([#12569](https:\u002F\u002Fgithub.com\u002FArize-ai\u002Fphoenix\u002Fissues\u002F12569)) ([c887c93](https:\u002F\u002Fgithub.com\u002FArize-ai\u002Fphoenix\u002Fcommit\u002Fc887c9393243a5c1970ae3dffde9fe01b81743ad))\n* 将空的 PHOENIX_SQL_DATABASE_SCHEMA 规范化为 None ([#12582](https:\u002F\u002Fgithub.com\u002FArize-ai\u002Fphoenix\u002Fissues\u002F12582)) ([cd0dfaa](https:\u002F\u002Fgithub.com\u002FArize-ai\u002Fphoenix\u002Fcommit\u002Fcd0dfaa09e1c9d92927657fae10ad4ebc0a786cb))","2026-04-08T13:29:15",{"id":177,"version":178,"summary_zh":179,"released_at":180},289329,"arize-phoenix-client-v2.3.1","## [2.3.1](https:\u002F\u002Fgithub.com\u002FArize-ai\u002Fphoenix\u002Fcompare\u002Farize-phoenix-client-v2.3.0...arize-phoenix-client-v2.3.1)（2026-04-07）\n\n### 功能\n\n* **evals：** 弃用 evals 1.0 并移除旧版实验模块（[#12239](https:\u002F\u002Fgithub.com\u002FArize-ai\u002Fphoenix\u002Fissues\u002F12239)）（[15e6415](https:\u002F\u002Fgithub.com\u002FArize-ai\u002Fphoenix\u002Fcommit\u002F15e6415105ce4757de905172cbd0e07215d38990)）\n\n\n### 其他杂项\n\n* 发布 2.3.1（[#12572](https:\u002F\u002Fgithub.com\u002FArize-ai\u002Fphoenix\u002Fissues\u002F12572)）（[a33630d](https:\u002F\u002Fgithub.com\u002FArize-ai\u002Fphoenix\u002Fcommit\u002Fa33630d1913c2c43a6d2f3c39056ea11a2de8bee)）","2026-04-07T18:49:07",{"id":182,"version":183,"summary_zh":184,"released_at":185},289330,"arize-phoenix-evals-v3.0.0","## [3.0.0](https:\u002F\u002Fgithub.com\u002FArize-ai\u002Fphoenix\u002Fcompare\u002Farize-phoenix-evals-v2.13.0...arize-phoenix-evals-v3.0.0) (2026-04-07)\n\n\n### ⚠ BREAKING CHANGES\n\n* **evals:** deprecate evals 1.0 and remove legacy experiments module ([#12239](https:\u002F\u002Fgithub.com\u002FArize-ai\u002Fphoenix\u002Fissues\u002F12239))\n\n### Features\n\n* **evals:** deprecate evals 1.0 and remove legacy experiments module ([#12239](https:\u002F\u002Fgithub.com\u002FArize-ai\u002Fphoenix\u002Fissues\u002F12239)) ([15e6415](https:\u002F\u002Fgithub.com\u002FArize-ai\u002Fphoenix\u002Fcommit\u002F15e6415105ce4757de905172cbd0e07215d38990))","2026-04-07T17:12:24",{"id":187,"version":188,"summary_zh":189,"released_at":190},289331,"arize-phoenix-v14.0.0","See [MIGRATION.md](https:\u002F\u002Fgithub.com\u002FArize-ai\u002Fphoenix\u002Fblob\u002Fmain\u002FMIGRATION.md)\r\n\r\n## [14.0.0](https:\u002F\u002Fgithub.com\u002FArize-ai\u002Fphoenix\u002Fcompare\u002Farize-phoenix-v13.23.0...arize-phoenix-v14.0.0) (2026-04-07)\r\n\r\n\r\n### ⚠ BREAKING CHANGES\r\n\r\n* **server:** remove \u002Fv1\u002Fevaluations endpoint and Evaluations plumbing ([#12538](https:\u002F\u002Fgithub.com\u002FArize-ai\u002Fphoenix\u002Fissues\u002F12538))\r\n* **evals:** deprecate evals 1.0 and remove legacy experiments module ([#12239](https:\u002F\u002Fgithub.com\u002FArize-ai\u002Fphoenix\u002Fissues\u002F12239))\r\n* **cli:** CLI flags now follow the subcommand instead of preceding it.  For example:   Before: phoenix --dev serve   After:  phoenix serve --dev\r\n* **experiments:** define ephemeral experiments and add ExperimentSweeper daemon ([#11815](https:\u002F\u002Fgithub.com\u002FArize-ai\u002Fphoenix\u002Fissues\u002F11815))\r\n\r\n### Features\r\n\r\n* add [@defer](https:\u002F\u002Fgithub.com\u002Fdefer) support to Relay and ProjectPageHeader ([#12486](https:\u002F\u002Fgithub.com\u002FArize-ai\u002Fphoenix\u002Fissues\u002F12486)) ([a87aef5](https:\u002F\u002Fgithub.com\u002FArize-ai\u002Fphoenix\u002Fcommit\u002Fa87aef53922d589b3a76e0ea59dde22a50e6189d))\r\n* add read replica routing support for PostgreSQL ([#12532](https:\u002F\u002Fgithub.com\u002FArize-ai\u002Fphoenix\u002Fissues\u002F12532)) ([3163fae](https:\u002F\u002Fgithub.com\u002FArize-ai\u002Fphoenix\u002Fcommit\u002F3163faef6f9f6180e9db468424c4cacb7b42b0a6))\r\n* add Strands Agents integration docs and onboarding snippet ([#12406](https:\u002F\u002Fgithub.com\u002FArize-ai\u002Fphoenix\u002Fissues\u002F12406)) ([8864f9e](https:\u002F\u002Fgithub.com\u002FArize-ai\u002Fphoenix\u002Fcommit\u002F8864f9ec98a0333bdb604af2a05224efd0bd0efc))\r\n* **agent:** add ask_user elicitation tool with carousel UI ([#12517](https:\u002F\u002Fgithub.com\u002FArize-ai\u002Fphoenix\u002Fissues\u002F12517)) ([ced465e](https:\u002F\u002Fgithub.com\u002FArize-ai\u002Fphoenix\u002Fcommit\u002Fced465e56d57177c0fc55ce7189bba8c84dbee7b))\r\n* **agent:** backend MCP docs tool via Mintlify integration ([#12556](https:\u002F\u002Fgithub.com\u002FArize-ai\u002Fphoenix\u002Fissues\u002F12556)) ([b228134](https:\u002F\u002Fgithub.com\u002FArize-ai\u002Fphoenix\u002Fcommit\u002Fb228134a800d163f953134344ed529d706744058))\r\n* **agent:** keep chat streaming when panel closes ([#12564](https:\u002F\u002Fgithub.com\u002FArize-ai\u002Fphoenix\u002Fissues\u002F12564)) ([8168a8f](https:\u002F\u002Fgithub.com\u002FArize-ai\u002Fphoenix\u002Fcommit\u002F8168a8fde556aca661d9a57d346342b4f63124e7))\r\n* **agent:** PXI FAB, panel sizing, and stop button ([#12539](https:\u002F\u002Fgithub.com\u002FArize-ai\u002Fphoenix\u002Fissues\u002F12539)) ([d28c122](https:\u002F\u002Fgithub.com\u002FArize-ai\u002Fphoenix\u002Fcommit\u002Fd28c122f1271c336e377ac9fec415c35b3d549cf))\r\n* **agent:** trace PXI chat requests locally and remotely ([#12394](https:\u002F\u002Fgithub.com\u002FArize-ai\u002Fphoenix\u002Fissues\u002F12394)) ([cc64489](https:\u002F\u002Fgithub.com\u002FArize-ai\u002Fphoenix\u002Fcommit\u002Fcc644897c5e9deedf015528aacb2771064e0584a))\r\n* background experiment runner ([#11731](https:\u002F\u002Fgithub.com\u002FArize-ai\u002Fphoenix\u002Fissues\u002F11731)) ([a769a61](https:\u002F\u002Fgithub.com\u002FArize-ai\u002Fphoenix\u002Fcommit\u002Fa769a61f1e99aa7671b25b3a44b2ecf5f76fbe70))\r\n* **cli:** restructure CLI into subcommand-first architecture with serve\u002Fdb commands ([#12336](https:\u002F\u002Fgithub.com\u002FArize-ai\u002Fphoenix\u002Fissues\u002F12336)) ([b1c4da0](https:\u002F\u002Fgithub.com\u002FArize-ai\u002Fphoenix\u002Fcommit\u002Fb1c4da0266c8ee31c725adaf5363e9081f6d7637))\r\n* **evals:** deprecate evals 1.0 and remove legacy experiments module ([#12239](https:\u002F\u002Fgithub.com\u002FArize-ai\u002Fphoenix\u002Fissues\u002F12239)) ([15e6415](https:\u002F\u002Fgithub.com\u002FArize-ai\u002Fphoenix\u002Fcommit\u002F15e6415105ce4757de905172cbd0e07215d38990))\r\n* **experiments:** define ephemeral experiments and add ExperimentSweeper daemon ([#11815](https:\u002F\u002Fgithub.com\u002FArize-ai\u002Fphoenix\u002Fissues\u002F11815)) ([69d197e](https:\u002F\u002Fgithub.com\u002FArize-ai\u002Fphoenix\u002Fcommit\u002F69d197ed3f61d8be9381900a36c30dcab60f030d))\r\n* **experiments:** structured error table with sanitized messages ([#12558](https:\u002F\u002Fgithub.com\u002FArize-ai\u002Fphoenix\u002Fissues\u002F12558)) ([6211910](https:\u002F\u002Fgithub.com\u002FArize-ai\u002Fphoenix\u002Fcommit\u002F62119108caca5486fdd55fb0498c90d1f448287b))\r\n* **graphql:** require explicit first for forward pagination ([#12526](https:\u002F\u002Fgithub.com\u002FArize-ai\u002Fphoenix\u002Fissues\u002F12526)) ([ed559c4](https:\u002F\u002Fgithub.com\u002FArize-ai\u002Fphoenix\u002Fcommit\u002Fed559c46e58b22425fdc115a1f7c03d0ab6f0d3b))\r\n* include span and experiment run in error subscription payloads ([#12464](https:\u002F\u002Fgithub.com\u002FArize-ai\u002Fphoenix\u002Fissues\u002F12464)) ([33ec19a](https:\u002F\u002Fgithub.com\u002FArize-ai\u002Fphoenix\u002Fcommit\u002F33ec19a02f59129257525ffd2ecccd411e564490))\r\n* **onboarding:** add new provider integrations and env var support ([#12494](https:\u002F\u002Fgithub.com\u002FArize-ai\u002Fphoenix\u002Fissues\u002F12494)) ([81d296b](https:\u002F\u002Fgithub.com\u002FArize-ai\u002Fphoenix\u002Fcommit\u002F81d296bee7c3636514ccae6c06e78f3ec655b6bd))\r\n* **playground:** add experiment record toggle with ephemeral tracking ([#11944](https:\u002F\u002Fgithub.com\u002FArize-ai\u002Fphoenix\u002Fissues\u002F11944)) ([f14eaef](https:\u002F\u002Fgithub.com\u002FArize-ai\u002Fphoenix\u002Fcommit\u002Ff14eaef46cbaf8a34ebf87bf266be536a9d38739))\r\n* poll experiments table when jobs are running ([#12483](https:\u002F\u002Fgithub.com\u002FArize-ai\u002Fphoenix\u002Fissues\u002F12483)) ([7e7c204](https:\u002F\u002Fgithub.com\u002FArize-ai\u002Fphoenix\u002Fcommit\u002F7e7c20481787c1eb91d6373719a4c2d94ee5482d))\r\n* Remove legacy client ([#12164](https:\u002F\u002Fgithub.com\u002FArize-ai\u002Fphoenix\u002Fissues\u002F12164)) ([4c27a0b](https:\u002F\u002Fgithub.com\u002FArize-ai\u002Fphoenix\u002Fcommit\u002F4c27a0baa9472d7af873da5d5d441b46e2b9a5cb))\r\n* use non-streaming for LLM evaluato","2026-04-07T16:56:04",{"id":192,"version":193,"summary_zh":194,"released_at":195},289332,"arize-phoenix-client-v2.3.0","## [2.3.0](https:\u002F\u002Fgithub.com\u002FArize-ai\u002Fphoenix\u002Fcompare\u002Farize-phoenix-client-v2.2.0...arize-phoenix-client-v2.3.0) (2026-04-03)\n\n\n### Features\n\n* atif to trace trajectory conversion utility ([#12414](https:\u002F\u002Fgithub.com\u002FArize-ai\u002Fphoenix\u002Fissues\u002F12414)) ([28ecfe0](https:\u002F\u002Fgithub.com\u002FArize-ai\u002Fphoenix\u002Fcommit\u002F28ecfe023f3d94f97afa36b7161f80e1d316cc39))","2026-04-03T16:14:07",{"id":197,"version":198,"summary_zh":199,"released_at":200},289333,"arize-phoenix-v13.23.0","## [13.23.0](https:\u002F\u002Fgithub.com\u002FArize-ai\u002Fphoenix\u002Fcompare\u002Farize-phoenix-v13.22.2...arize-phoenix-v13.23.0) (2026-04-02)\n\n\n### Features\n\n* bundle TypeScript SDK docs into npm packages ([#12487](https:\u002F\u002Fgithub.com\u002FArize-ai\u002Fphoenix\u002Fissues\u002F12487)) ([1028be5](https:\u002F\u002Fgithub.com\u002FArize-ai\u002Fphoenix\u002Fcommit\u002F1028be5ea7400886a2abf64b2c227f66edb791e0))\n\n\n### Bug Fixes\n\n* **app:** fix broken Playground button in prompts table ([#12511](https:\u002F\u002Fgithub.com\u002FArize-ai\u002Fphoenix\u002Fissues\u002F12511)) ([4a4ddb3](https:\u002F\u002Fgithub.com\u002FArize-ai\u002Fphoenix\u002Fcommit\u002F4a4ddb3f80038578553760de2dab3561a7bf805d))\n* remove openinference dependency from migration ([#12510](https:\u002F\u002Fgithub.com\u002FArize-ai\u002Fphoenix\u002Fissues\u002F12510)) ([1059492](https:\u002F\u002Fgithub.com\u002FArize-ai\u002Fphoenix\u002Fcommit\u002F1059492a4bd90e6b98c96998e995794940168de3))\n\n\n### Documentation\n\n* Add Phoenix release notes for 03-26-2026 through 04-01-2026 ([#12489](https:\u002F\u002Fgithub.com\u002FArize-ai\u002Fphoenix\u002Fissues\u002F12489)) ([c414afc](https:\u002F\u002Fgithub.com\u002FArize-ai\u002Fphoenix\u002Fcommit\u002Fc414afc1d9008e55a60943cc37035a72c0fad613))","2026-04-02T06:27:43",{"id":202,"version":203,"summary_zh":204,"released_at":205},289334,"arize-phoenix-v13.22.2","## [13.22.2](https:\u002F\u002Fgithub.com\u002FArize-ai\u002Fphoenix\u002Fcompare\u002Farize-phoenix-v13.22.1...arize-phoenix-v13.22.2) (2026-04-01)\n\n\n### Bug Fixes\n\n* use DISTINCT instead of GROUP BY in ProjectHasTracesDataLoader ([#12495](https:\u002F\u002Fgithub.com\u002FArize-ai\u002Fphoenix\u002Fissues\u002F12495)) ([77bf1a3](https:\u002F\u002Fgithub.com\u002FArize-ai\u002Fphoenix\u002Fcommit\u002F77bf1a3e0e287b561e430b74f74bc0fa082ed7f9))","2026-04-01T17:39:29",{"id":207,"version":208,"summary_zh":209,"released_at":210},289335,"arize-phoenix-v13.22.1","## [13.22.1](https:\u002F\u002Fgithub.com\u002FArize-ai\u002Fphoenix\u002Fcompare\u002Farize-phoenix-v13.22.0...arize-phoenix-v13.22.1) (2026-04-01)\n\n\n### Bug Fixes\n\n* replace expensive count queries with hasTraces existence check ([#12490](https:\u002F\u002Fgithub.com\u002FArize-ai\u002Fphoenix\u002Fissues\u002F12490)) ([2d064fc](https:\u002F\u002Fgithub.com\u002FArize-ai\u002Fphoenix\u002Fcommit\u002F2d064fc263a70cc5705a535d0070632200d6d922))\n* skip redundant header stats refetch on initial mount ([#12491](https:\u002F\u002Fgithub.com\u002FArize-ai\u002Fphoenix\u002Fissues\u002F12491)) ([a10d8ed](https:\u002F\u002Fgithub.com\u002FArize-ai\u002Fphoenix\u002Fcommit\u002Fa10d8ed5664221aef53588ad7ec0c8c9b669e808))","2026-04-01T16:02:29",{"id":212,"version":213,"summary_zh":214,"released_at":215},289336,"arize-phoenix-client-v2.2.0","## [2.2.0](https:\u002F\u002Fgithub.com\u002FArize-ai\u002Fphoenix\u002Fcompare\u002Farize-phoenix-client-v2.1.0...arize-phoenix-client-v2.2.0) (2026-04-01)\n\n\n### Features\n\n* add GET \u002Fv1\u002Fuser endpoint and phoenix-rest-api skill ([#12274](https:\u002F\u002Fgithub.com\u002FArize-ai\u002Fphoenix\u002Fissues\u002F12274)) ([b5188ea](https:\u002F\u002Fgithub.com\u002FArize-ai\u002Fphoenix\u002Fcommit\u002Fb5188ea92eeb76ed0155e999b8fa0fc00519f32c))\n* add Python 3.14 support (except Windows) ([#12475](https:\u002F\u002Fgithub.com\u002FArize-ai\u002Fphoenix\u002Fissues\u002F12475)) ([50c08e3](https:\u002F\u002Fgithub.com\u002FArize-ai\u002Fphoenix\u002Fcommit\u002F50c08e3ff4e41c0693e92b7d3de7dbc41384fad1))\n* **client:** add get_traces method to Python and TypeScript clients ([#12245](https:\u002F\u002Fgithub.com\u002FArize-ai\u002Fphoenix\u002Fissues\u002F12245)) ([7d3f4ac](https:\u002F\u002Fgithub.com\u002FArize-ai\u002Fphoenix\u002Fcommit\u002F7d3f4aca8d5bc3e6df3e5a4f69332fa8ffcc274d))\n* **server:** REST endpoint for secrets management ([#12429](https:\u002F\u002Fgithub.com\u002FArize-ai\u002Fphoenix\u002Fissues\u002F12429)) ([39f68cf](https:\u002F\u002Fgithub.com\u002FArize-ai\u002Fphoenix\u002Fcommit\u002F39f68cf5449a5df04e4b8901975c01bbf76e93fc))\n\n\n### Documentation\n\n* **evals:** deprecate evals 1.0 docs ([#12123](https:\u002F\u002Fgithub.com\u002FArize-ai\u002Fphoenix\u002Fissues\u002F12123)) ([b0dc90f](https:\u002F\u002Fgithub.com\u002FArize-ai\u002Fphoenix\u002Fcommit\u002Fb0dc90f445a25ba74ddc86e0cc8bcdf721260344))\n* update all package READMEs with latest capabilities ([#12315](https:\u002F\u002Fgithub.com\u002FArize-ai\u002Fphoenix\u002Fissues\u002F12315)) ([b4ded15](https:\u002F\u002Fgithub.com\u002FArize-ai\u002Fphoenix\u002Fcommit\u002Fb4ded157987e0ce6fcbf749bf48225d9408313c2))\n* update package READMEs for recent features ([#12259](https:\u002F\u002Fgithub.com\u002FArize-ai\u002Fphoenix\u002Fissues\u002F12259)) ([d204898](https:\u002F\u002Fgithub.com\u002FArize-ai\u002Fphoenix\u002Fcommit\u002Fd204898cf63b48604e82403c1c6056fbb72fb860))","2026-04-01T01:35:20",{"id":217,"version":218,"summary_zh":219,"released_at":220},289337,"arize-phoenix-evals-v2.13.0","## [2.13.0](https:\u002F\u002Fgithub.com\u002FArize-ai\u002Fphoenix\u002Fcompare\u002Farize-phoenix-evals-v2.12.0...arize-phoenix-evals-v2.13.0) (2026-04-01)\n\n\n### Features\n\n* add Python 3.14 support (except Windows) ([#12475](https:\u002F\u002Fgithub.com\u002FArize-ai\u002Fphoenix\u002Fissues\u002F12475)) ([50c08e3](https:\u002F\u002Fgithub.com\u002FArize-ai\u002Fphoenix\u002Fcommit\u002F50c08e3ff4e41c0693e92b7d3de7dbc41384fad1))","2026-04-01T01:26:34",{"id":222,"version":223,"summary_zh":224,"released_at":225},289338,"arize-phoenix-v13.22.0","## [13.22.0](https:\u002F\u002Fgithub.com\u002FArize-ai\u002Fphoenix\u002Fcompare\u002Farize-phoenix-v13.21.0...arize-phoenix-v13.22.0) (2026-04-01)\n\n\n### Features\n\n* **ui:** updated provider and integration icons, logo icon agent skill ([#12470](https:\u002F\u002Fgithub.com\u002FArize-ai\u002Fphoenix\u002Fissues\u002F12470)) ([8668c43](https:\u002F\u002Fgithub.com\u002FArize-ai\u002Fphoenix\u002Fcommit\u002F8668c437d44969b6b52f67d3d4c9f0afce270ba7))\n\n\n### Bug Fixes\n\n* fail loudly on Windows + Python 3.14 install ([#12480](https:\u002F\u002Fgithub.com\u002FArize-ai\u002Fphoenix\u002Fissues\u002F12480)) ([50d66a3](https:\u002F\u002Fgithub.com\u002FArize-ai\u002Fphoenix\u002Fcommit\u002F50d66a3569be4eb5f4efda49890cdd4e30b3de42))","2026-04-01T01:11:46",{"id":227,"version":228,"summary_zh":229,"released_at":230},289339,"arize-phoenix-v13.21.0","## [13.21.0](https:\u002F\u002Fgithub.com\u002FArize-ai\u002Fphoenix\u002Fcompare\u002Farize-phoenix-v13.20.0...arize-phoenix-v13.21.0) (2026-04-01)\n\n\n### Features\n\n* add Python 3.14 support (except Windows) ([#12475](https:\u002F\u002Fgithub.com\u002FArize-ai\u002Fphoenix\u002Fissues\u002F12475)) ([50c08e3](https:\u002F\u002Fgithub.com\u002FArize-ai\u002Fphoenix\u002Fcommit\u002F50c08e3ff4e41c0693e92b7d3de7dbc41384fad1))\n* **agent:** Improve session summaries ([#12471](https:\u002F\u002Fgithub.com\u002FArize-ai\u002Fphoenix\u002Fissues\u002F12471)) ([f1750a9](https:\u002F\u002Fgithub.com\u002FArize-ai\u002Fphoenix\u002Fcommit\u002Ff1750a920a0d650314e30c7bf4ce5571163a9bd4))\n* **server:** REST endpoint for secrets management ([#12429](https:\u002F\u002Fgithub.com\u002FArize-ai\u002Fphoenix\u002Fissues\u002F12429)) ([39f68cf](https:\u002F\u002Fgithub.com\u002FArize-ai\u002Fphoenix\u002Fcommit\u002F39f68cf5449a5df04e4b8901975c01bbf76e93fc))","2026-04-01T00:19:24",[232,242,250,258,267,275],{"id":233,"name":234,"github_repo":235,"description_zh":236,"stars":237,"difficulty_score":238,"last_commit_at":239,"category_tags":240,"status":99},4358,"openclaw","openclaw\u002Fopenclaw","OpenClaw 是一款专为个人打造的本地化 AI 助手，旨在让你在自己的设备上拥有完全可控的智能伙伴。它打破了传统 AI 助手局限于特定网页或应用的束缚，能够直接接入你日常使用的各类通讯渠道，包括微信、WhatsApp、Telegram、Discord、iMessage 等数十种平台。无论你在哪个聊天软件中发送消息，OpenClaw 都能即时响应，甚至支持在 macOS、iOS 和 Android 设备上进行语音交互，并提供实时的画布渲染功能供你操控。\n\n这款工具主要解决了用户对数据隐私、响应速度以及“始终在线”体验的需求。通过将 AI 部署在本地，用户无需依赖云端服务即可享受快速、私密的智能辅助，真正实现了“你的数据，你做主”。其独特的技术亮点在于强大的网关架构，将控制平面与核心助手分离，确保跨平台通信的流畅性与扩展性。\n\nOpenClaw 非常适合希望构建个性化工作流的技术爱好者、开发者，以及注重隐私保护且不愿被单一生态绑定的普通用户。只要具备基础的终端操作能力（支持 macOS、Linux 及 Windows WSL2），即可通过简单的命令行引导完成部署。如果你渴望拥有一个懂你",349277,3,"2026-04-06T06:32:30",[76,80,241,77],"图像",{"id":243,"name":244,"github_repo":245,"description_zh":246,"stars":247,"difficulty_score":238,"last_commit_at":248,"category_tags":249,"status":99},3808,"stable-diffusion-webui","AUTOMATIC1111\u002Fstable-diffusion-webui","stable-diffusion-webui 是一个基于 Gradio 构建的网页版操作界面，旨在让用户能够轻松地在本地运行和使用强大的 Stable Diffusion 图像生成模型。它解决了原始模型依赖命令行、操作门槛高且功能分散的痛点，将复杂的 AI 绘图流程整合进一个直观易用的图形化平台。\n\n无论是希望快速上手的普通创作者、需要精细控制画面细节的设计师，还是想要深入探索模型潜力的开发者与研究人员，都能从中获益。其核心亮点在于极高的功能丰富度：不仅支持文生图、图生图、局部重绘（Inpainting）和外绘（Outpainting）等基础模式，还独创了注意力机制调整、提示词矩阵、负向提示词以及“高清修复”等高级功能。此外，它内置了 GFPGAN 和 CodeFormer 等人脸修复工具，支持多种神经网络放大算法，并允许用户通过插件系统无限扩展能力。即使是显存有限的设备，stable-diffusion-webui 也提供了相应的优化选项，让高质量的 AI 艺术创作变得触手可及。",162132,"2026-04-05T11:01:52",[80,241,76],{"id":251,"name":252,"github_repo":253,"description_zh":254,"stars":255,"difficulty_score":98,"last_commit_at":256,"category_tags":257,"status":99},1381,"everything-claude-code","affaan-m\u002Feverything-claude-code","everything-claude-code 是一套专为 AI 编程助手（如 Claude Code、Codex、Cursor 等）打造的高性能优化系统。它不仅仅是一组配置文件，而是一个经过长期实战打磨的完整框架，旨在解决 AI 代理在实际开发中面临的效率低下、记忆丢失、安全隐患及缺乏持续学习能力等核心痛点。\n\n通过引入技能模块化、直觉增强、记忆持久化机制以及内置的安全扫描功能，everything-claude-code 能显著提升 AI 在复杂任务中的表现，帮助开发者构建更稳定、更智能的生产级 AI 代理。其独特的“研究优先”开发理念和针对 Token 消耗的优化策略，使得模型响应更快、成本更低，同时有效防御潜在的攻击向量。\n\n这套工具特别适合软件开发者、AI 研究人员以及希望深度定制 AI 工作流的技术团队使用。无论您是在构建大型代码库，还是需要 AI 协助进行安全审计与自动化测试，everything-claude-code 都能提供强大的底层支持。作为一个曾荣获 Anthropic 黑客大奖的开源项目，它融合了多语言支持与丰富的实战钩子（hooks），让 AI 真正成长为懂上",158594,"2026-04-16T23:34:05",[80,76,79],{"id":259,"name":260,"github_repo":261,"description_zh":262,"stars":263,"difficulty_score":63,"last_commit_at":264,"category_tags":265,"status":99},8272,"opencode","anomalyco\u002Fopencode","OpenCode 是一款开源的 AI 编程助手（Coding Agent），旨在像一位智能搭档一样融入您的开发流程。它不仅仅是一个代码补全插件，而是一个能够理解项目上下文、自主规划任务并执行复杂编码操作的智能体。无论是生成全新功能、重构现有代码，还是排查难以定位的 Bug，OpenCode 都能通过自然语言交互高效完成，显著减少开发者在重复性劳动和上下文切换上的时间消耗。\n\n这款工具专为软件开发者、工程师及技术研究人员设计，特别适合希望利用大模型能力来提升编码效率、加速原型开发或处理遗留代码维护的专业人群。其核心亮点在于完全开源的架构，这意味着用户可以审查代码逻辑、自定义行为策略，甚至私有化部署以保障数据安全，彻底打破了传统闭源 AI 助手的“黑盒”限制。\n\n在技术体验上，OpenCode 提供了灵活的终端界面（Terminal UI）和正在测试中的桌面应用程序，支持 macOS、Windows 及 Linux 全平台。它兼容多种包管理工具，安装便捷，并能无缝集成到现有的开发环境中。无论您是追求极致控制权的资深极客，还是渴望提升产出的独立开发者，OpenCode 都提供了一个透明、可信",144296,"2026-04-16T14:50:03",[76,266],"插件",{"id":268,"name":269,"github_repo":270,"description_zh":271,"stars":272,"difficulty_score":98,"last_commit_at":273,"category_tags":274,"status":99},2271,"ComfyUI","Comfy-Org\u002FComfyUI","ComfyUI 是一款功能强大且高度模块化的视觉 AI 引擎，专为设计和执行复杂的 Stable Diffusion 图像生成流程而打造。它摒弃了传统的代码编写模式，采用直观的节点式流程图界面，让用户通过连接不同的功能模块即可构建个性化的生成管线。\n\n这一设计巧妙解决了高级 AI 绘图工作流配置复杂、灵活性不足的痛点。用户无需具备编程背景，也能自由组合模型、调整参数并实时预览效果，轻松实现从基础文生图到多步骤高清修复等各类复杂任务。ComfyUI 拥有极佳的兼容性，不仅支持 Windows、macOS 和 Linux 全平台，还广泛适配 NVIDIA、AMD、Intel 及苹果 Silicon 等多种硬件架构，并率先支持 SDXL、Flux、SD3 等前沿模型。\n\n无论是希望深入探索算法潜力的研究人员和开发者，还是追求极致创作自由度的设计师与资深 AI 绘画爱好者，ComfyUI 都能提供强大的支持。其独特的模块化架构允许社区不断扩展新功能，使其成为当前最灵活、生态最丰富的开源扩散模型工具之一，帮助用户将创意高效转化为现实。",108322,"2026-04-10T11:39:34",[80,241,76],{"id":276,"name":277,"github_repo":278,"description_zh":279,"stars":280,"difficulty_score":98,"last_commit_at":281,"category_tags":282,"status":99},6121,"gemini-cli","google-gemini\u002Fgemini-cli","gemini-cli 是一款由谷歌推出的开源 AI 命令行工具，它将强大的 Gemini 大模型能力直接集成到用户的终端环境中。对于习惯在命令行工作的开发者而言，它提供了一条从输入提示词到获取模型响应的最短路径，无需切换窗口即可享受智能辅助。\n\n这款工具主要解决了开发过程中频繁上下文切换的痛点，让用户能在熟悉的终端界面内直接完成代码理解、生成、调试以及自动化运维任务。无论是查询大型代码库、根据草图生成应用，还是执行复杂的 Git 操作，gemini-cli 都能通过自然语言指令高效处理。\n\n它特别适合广大软件工程师、DevOps 人员及技术研究人员使用。其核心亮点包括支持高达 100 万 token 的超长上下文窗口，具备出色的逻辑推理能力；内置 Google 搜索、文件操作及 Shell 命令执行等实用工具；更独特的是，它支持 MCP（模型上下文协议），允许用户灵活扩展自定义集成，连接如图像生成等外部能力。此外，个人谷歌账号即可享受免费的额度支持，且项目基于 Apache 2.0 协议完全开源，是提升终端工作效率的理想助手。",100752,"2026-04-10T01:20:03",[266,76,241,80]]