Phoenix does not provide a built-in model serving feature analogous to MLflow's model serving. Phoenix focuses on experimentation, evaluation, tracing, and observability for LLM and AI applications, but it does not offer a native mechanism to deploy or serve models or tasks as production endpoints. The recommended approach is to use your own serving infrastructure (e.g., Docker, ECS, Lambda) for deploying tasks or models, and then instrument those services with Phoenix for observability and evaluation purposes. There is no "Phoenix-native" serving or deployment abstraction鈥擯hoenix is designed to observe and evaluate, not to serve models directly or manage production inference endpoints.
You can continue wrapping your task logic in a container and deploying it as you currently do, then use Phoenix to trace, evaluate, and monitor the production service by instrumenting your code and sending traces to Phoenix. This workflow is consistent with Phoenix's intended usage and documentation; there are no additional serving features or deployment recommendations beyond this in the official docs or guides. For more, see <https://arize.com/docs/phoenix/datasets-and-experiments/how-to-experiments/run-experiments|Phoenix documentation on experiments and deployment> and <https://arize.com/docs/phoenix/self-hosting/deployment-options/docker|self-hosting/deployment options>.
Would you like more detail on how to instrument your production service for Phoenix observability?
Sources:
- <https://arize.com/docs/phoenix/datasets-and-experiments/how-to-experiments/run-experiments|Phoenix documentation on experiments and deployment>
- <https://arize.com/docs/phoenix/self-hosting/deployment-options/docker|self-hosting/deployment options>