-
Notifications
You must be signed in to change notification settings - Fork 109
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
Add litellm inference #385
base: main
Are you sure you want to change the base?
Conversation
…ce#382) Co-authored-by: Clémentine Fourrier <[email protected]>
Fixes cache directory bug by using HF_HUB_CACHE instead of HF_HOME See documentation https://huggingface.co/docs/huggingface_hub/main/en/package_reference/environment_variables#hfhubcache
Co-authored-by: Nathan Habib <[email protected]> Co-authored-by: Clémentine Fourrier <[email protected]>
The docs for this PR live here. All of your documentation changes will be reflected on that endpoint. The docs are available until 30 days after the last update. |
Hi @JoelNiklaus ! Is this PR ready or do you need any hep with it ? |
From my side it is ready |
Nice ! I will merge main and solve possible conflict as we have made a big change on the way we call the CLI. Is it ok for you ? |
Sounds great, thanks! |
I'm trying to use it, can you provide me with the command you use ? |
For example like this:
|
Should be good to go ! Added a few logging fixes for convenience. @JoelNiklaus Tell me what you think, are you able to use it ? :) |
When testing it I receive this error:
|
Hey @JoelNiklaus modified the way we pass the chat-templated-messages to the litellm model. |
Ok cool, thanks! Would it make sense to add this the same way to the openai backend or remove that altogether? |
I don;t think we need the openai backend anymore if we have the litellm backend, i'm pretty against having 2 ways of doing the same thing |
Check the litellm docs in depth and I agree - I think we'll want to keep inference endpoints however (for leaderboards), even though it's slightly redundant |
Tests also need to be fixed ^^ |
…val into add_litellm_inference
…val into add_litellm_inference
Co-authored-by: Albert Villanova del Moral <[email protected]>
Co-authored-by: Albert Villanova del Moral <[email protected]>
This PR enables running inference using any model provider supported by litellm.