Chat with ollama/mistral-7b behind litellm returns strange answer #2259
francesco086
started this conversation in
LLM Usage | 语言模型研究
Replies: 2 comments
-
Thank you for raising an issue. We will investigate into the matter and get back to you as soon as possible. |
Beta Was this translation helpful? Give feedback.
0 replies
-
@francesco086 I think it might be the issue of parameter. Try to add a high value of |
Beta Was this translation helpful? Give feedback.
0 replies
Sign up for free
to join this conversation on GitHub.
Already have an account?
Sign in to comment
-
💻 Operating System
Other
📦 Environment
Docker
🌐 Browser
Other
🐛 Bug Description
I serve the mistral-7b model using ollama, and set a litellm proxy in front of it.
I am, for example, able to run the command:
and get the expected response
I setup lobechat to use several OpenAI models via litellm (gpt 3.5, 4, and dalle3), and everything works fine. However, with ollama/mistral-7b I get the following behaviour (I pressed the "Stop" button after a while because it was too slow):
🚦 Expected Behavior
No response
📷 Recurrence Steps
No response
📝 Additional Information
Services are running on Kubernetes, setup via ArgoCD.
Beta Was this translation helpful? Give feedback.
All reactions