Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

make it work with ollama and I will dockerise it for you #5

Open
fahadshery opened this issue Oct 4, 2024 · 4 comments
Open

make it work with ollama and I will dockerise it for you #5

fahadshery opened this issue Oct 4, 2024 · 4 comments

Comments

@fahadshery
Copy link

dont want to use openAI
already have ollama installed

@MR-444
Copy link
Contributor

MR-444 commented Oct 9, 2024

I don't know Ollama, but I guess it is a self-hosted server with an exposed API?
Edit: I looked on Ollama, it seems that this it is used to run a text-to-text LLM locally, but we need also an image-to-text system, if both are working in Ollama it would be possible, running a text-to-text LLM only would not work.

If these requirements are met, make it easy for us and give the API endpoints description(s) (+parameters) you need, then it would be easy to include it.

@fahadshery
Copy link
Author

Well ollama has a built-in openai compatibility. https://ollama.com/blog/openai-compatibility

image to text could be done via models like llava using ollama?

@MR-444
Copy link
Contributor

MR-444 commented Oct 10, 2024

Thank you, this is interesting, I did not see it yesterday.
This can really do it:
https://ollama.com/library?q=llava
We need to add another endpoint pointing to the oolama server api (I hope).

But I am not sure how good the image-to-text results would be, in my experience GPT-Vision is the best so far, maybe the Claude system is on par, followed by Gemini.

Well, I can't promise anything other than to try to find some time to investigate.

@MarkusR1805
Copy link

I use Ollama and love it!
You can quickly install a model, try it out, and if it's not good, you can uninstall it pretty quickly.
Here is a LLava model. I can't get along with it, but I installed it because I'm trying to do something with it in Python.
To generate prompts I use llama3.2:3b.

Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Labels
None yet
Projects
None yet
Development

No branches or pull requests

3 participants