Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

Implement logit_bias correctly #2869

Open
juanwisz opened this issue Dec 28, 2024 · 0 comments
Open

Implement logit_bias correctly #2869

juanwisz opened this issue Dec 28, 2024 · 0 comments

Comments

@juanwisz
Copy link

juanwisz commented Dec 28, 2024

Feature request

Currently, logit_bias is labeled as unused https://github.com/huggingface/text-generation-inference/blob/main/router/src/lib.rs

Also, the documentation states that a JSON is needed with a mapping from token_ids to a number between -100 and 100. But this is misaligned with the code's typing, that asks for a vector of floats.

Motivation

Logit_bias is a very important parameter, the documentation in InferenceClient from huggingface_hub states that it can be used, but it does not work.
Also see:
huggingface/huggingface_hub#2720

Your contribution

I can definitely help building a PR but I will need details on what type of solution is expected.

Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Labels
None yet
Projects
None yet
Development

No branches or pull requests

1 participant