Gliner quantization #145
Unanswered
carloronsi
asked this question in
Q&A
Replies: 2 comments
-
I'm also having a similar problem. When quantizing to 8bit I'm measuring a quite strong drop in performance. More than I usually see for this type of models. |
Beta Was this translation helpful? Give feedback.
0 replies
Sign up for free
to join this conversation on GitHub.
Already have an account?
Sign in to comment
-
Hi everyone! I've tried the quantization code present in this report:
https://github.com/urchade/GLiNER/blob/main/examples/convert_to_onnx.ipynb
Followed step-by-step, but the quantized model (last cell) is not returning any entity (if the threshold is reduced, texts are labeled incorrectly).
Am I missing something? Is there an additional step required to make the quantized model work?
Thank you all!
Beta Was this translation helpful? Give feedback.
All reactions