-
Notifications
You must be signed in to change notification settings - Fork 637
NotImplementedError when merge lora for qwen2 vl 7B #2344
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
Comments
是因为AWQ模型不能 merge lora 吗?那我要怎么推理呢 |
awq不支持merge-lora. 只能直接推理 |
单样本推理可以查看文档 |
可以把 AWQ 训练的 lora 和原模型合并吗? |
这样会有精度损失啊 不过 你额外指定--model_id_or_path 试试, 测测性能如何 |
我直接用原始模型训练了。但是我把 finetune 后的模型,做了一下 GPTQ 8bit 量化,对比量化前的模型,效果差不少。 |
效果差不少是 直接无法正常输出嘛,因为刚修复了一个gptq的bug |
gptq_quantizer.serialization_keys.append('block_name_to_quantize') |
感谢回复。我微调的 image_caption 任务,原模型能够正常描述人物属性,但是量化后描述不太准确。能够正常描述,但是准确率下降有点多。 |
Your hardware and system info
Write your system info like CUDA version/system/GPU/torch version here(在这里给出硬件信息和系统信息,如CUDA版本,系统,GPU型号和torch版本等)
system: ubuntu 22.04
python: 3.11
GPU: 4060ti 16G
torch: 2.4
Additional context
The text was updated successfully, but these errors were encountered: