Hey All, need some directions on using phoenix evals on locally downloaded models like mistral/falcon etc
Curious if you're using a library such as HuggingFace as an interface to the models, or if you're running the model code directly.
Sounds like you're trying to run the model on bare metal as opposed to via a third-party API?
Right, so basically i want to use the falcon/mistral do hallucination evaluation instead of openai
We began looking into building a HuggingFace adapter a while back and can re-open the issue.
One thing to note is that the throughput for evals would be greatly constrained if you're running on a single node/ GPU. We are able to use concurrency to compute large volumes of evals quickly with third-party APIs.
Is your main motivation around using open-source models around data privacy/ not sending data to third-party APIs, or something else? If you are interested in using Mixtral or Falcon and are able to use a service like Anyscale Endpoints, it's straightforward to get set up.
I have got Huggingface to work through LLMlite
Oh nice, I did not know that.
