-
Notifications
You must be signed in to change notification settings - Fork 636
Qwen2.5vl32B merge lora OOM问题 #4145
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
Comments
try |
非常感谢您的回复,merge lora使用这个确实可以了,但是推理的时候出现了新的bug NPROC_PER_NODE=1 \(此行我尝试过删除,和分别取1和8,删除和1是以下的bug报告,8是OOM)
|
删除应该没问题才对哇 |
方便测试一下 这个有效嘛 我这里没有NPU测试 |
是要修改infer的程序吗
|
Sign up for free
to join this conversation on GitHub.
Already have an account?
Sign in to comment
您好,我使用的是8卡910B的昇腾的机器
我在尝试微调并推理Qwen2.5vl32B的模型
直接训练的话,模型会在加载的时候出现OOM的情况,我才用了zero3,分布在八卡上,避免了这个问题
但现在,在推理阶段,我再次遇到了模型无法单卡加载的问题(OOM)
我尝试将模型分布到多卡上,受限于我匮乏的代码能力,一直没有成功,请问是否有参数可以完成这一点?
后续我还需要这个模型进行merge lora,也是相同的,无法单卡加载模型的问题,请问merge lora是否有参数可以解决?
非常期待您的回复,再次感谢您优秀的工作
The text was updated successfully, but these errors were encountered: