Phoenix does not currently support a timeline view similar to AWS X-Ray. This feature is not available, but you can request it as an enhancement on the Phoenix GitHub repository. According to a recent Slack discussion, users are encouraged to open a feature request if they need this functionality. Would you like more detail on how to submit a feature request or alternative ways to visualize traces in Phoenix?Sources: - Slack discussion: timeline view not supported
馃挕聽Hint: Mention RunLLM in the thread for followups.
would a pip freeze answer your question?
openinference-instrumentation==0.1.32
openinference-instrumentation-smolagents==0.1.12
openinference-semantic-conventions==0.1.17this is my current environment.
if os.environ.get("APP_USE_TELEMETRY") == "true":
from phoenix.otel import register
register(project_name="zero", auto_instrument=True)
worker.run()
...
# in worker.run()
from openinference.instrumentation import using_session
with using_session(telemetry_session_id or f"{workspace_id}-{uuid()}"):
output = engine.query(task)this is what we run to setup everything.
(worker is a celery worker if it may be helpful) Then .query spawns some thread executors that run the smolagents (however the the pic i posted above should be executing directly in the .query thread if i recall corretly).
Okay that's great - trying to replicate this now!
Let me know if there is any other info i can share (we are mostly concerned about the fact that the innermost call takes half the time of the outermost call, so maybe we are calling things repeatdly without realising it)
Okay it doesn't look like there's a general problem with duplicate calls on those versions, I'm seeing the correct traces. Going to try and match your setup more closely. Are you using the InferenceClientModel or InferenceClient object within smolagents, and which model(s) are you using?
`model = LiteLLMModel("gemini-2.0-flash") model(messages)`
Ah okay think I found it - can you upgrade to openinference-instrumentation-smolagents==0.1.13 We added better support for the generate function and base model invocation approaches. I'm seeing that remove the repeat issue - though I only saw it repeat a total of 3 calls
Things got better (we also have a nice calls structure now), however it's still repeated twice per call sometimes (the second ZeroCodeAgent.run for example), and the total amount of tokens seems to not be adding up as one would expect?
