You signed in with another tab or window. Reload to refresh your session.You signed out in another tab or window. Reload to refresh your session.You switched accounts on another tab or window. Reload to refresh your session.Dismiss alert
It seem that the quantization is not applied to the model. How to solve this problem? How can I just quantize part of a model? Looking forward to your help!
The text was updated successfully, but these errors were encountered:
I want to quantize layers 12~23 of qwen0.5B. I used llmcompressor 0.3.1. Here is my script:
I use
ignore: ["lm_head", "re:model.layers.(0?[0-9]|1[01]).*$"]
to avoid quantizing layer 0~11. However, after the quantization, the model config isIt seem that the quantization is not applied to the model. How to solve this problem? How can I just quantize part of a model? Looking forward to your help!
The text was updated successfully, but these errors were encountered: