Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

Error occurs when loading the gemma model in bitsandbytes format. #2557

Merged
merged 3 commits into from
Dec 26, 2024

Conversation

upskyy
Copy link
Contributor

@upskyy upskyy commented Dec 23, 2024

Motivation

#2556

Modifications

I refer to vllm to solve the key error that occurs when loading in bitsandbytes format.

An error occurs when loading the gemma model with the command below.

python3 -m sglang.launch_server --model-path /models --tokenizer-path /models --port 30000 --tokenizer-mode auto --dtype bfloat16 --mem-fraction-static 0.5 --random-seed 0 --enable-torch-compile --disable-cuda-graph --schedule-conservativeness 1.3 --kv-cache-dtype fp8_e5m2 --load-format bitsandbytes --quantization bitsandbytes

Checklist

  • Format your code according to the Contributor Guide.
  • Add unit tests as outlined in the Contributor Guide.
  • Update documentation as needed, including docstrings or example tutorials.

@merrymercy merrymercy merged commit 08effbf into sgl-project:main Dec 26, 2024
15 checks passed
@merrymercy
Copy link
Contributor

@upskyy Thanks for the contribution. It is merged. Can you also help us fix this llama model? #2600

@upskyy
Copy link
Contributor Author

upskyy commented Dec 27, 2024

@merrymercy
Looking at the code, I think the issue will also be resolved with the PR #2557. This problem occurred when changing to qweight when doing bitsandbytes 4bit load, but in the PR #2557, it was corrected not only for the gemma model but also for loads of other models.

Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Labels
None yet
Projects
None yet
Development

Successfully merging this pull request may close these issues.

2 participants