Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

[broken-compatibility] chat completion breaks base64 standard / openAI spec #2820

Open
2 of 4 tasks
lucyknada opened this issue Dec 11, 2024 · 0 comments
Open
2 of 4 tasks

Comments

@lucyknada
Copy link

System Info

latest docker pull, --version says: text-generation-launcher 3.0.0
model used: https://huggingface.co/AI-Safeguard/Ivy-VL-llava

Information

  • Docker
  • The CLI directly

Tasks

  • An officially supported command
  • My own modifications

Reproduction

  1. host any vision model with tgi
  2. use any openAI / chat-completions compatible frontend or library, e.g. openwebUI to connect to it on /v1
  3. send an image and user-message
  4. it gets auto-converted into base64 and put into context as image-url (openAI spec)
  5. the response clearly indicates it sees the base64, but the model simply makes fun of that fact instead of tgi injecting it as image embeddings

model response:
image

payload sent by frontends and openAI-spec:
image

meanwhile tgi expects I assume base64 in an "inputs" key separate from messages or ignores base64 in messages entirely?

either way that breaks compatibility with all frontends and libraries that expect the backend to handle in-line base64.

thanks!

referencing original issue for cross-repo visibility: open-webui/open-webui#7760

Expected behavior

base64 image embedded into context is recognized as is with no markdown urls

Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Labels
None yet
Projects
None yet
Development

No branches or pull requests

1 participant