-
Notifications
You must be signed in to change notification settings - Fork 28
support svd quant #202
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
support svd quant #202
Conversation
Summary of ChangesHello @Glaceon-Hyy, I'm Gemini Code Assist1! I'm currently reviewing this pull request and will post my feedback shortly. In the meantime, here's a summary to help you and other reviewers quickly get up to speed! This pull request significantly enhances the Highlights
Using Gemini Code AssistThe full guide for Gemini Code Assist can be found on our documentation page, here are some quick tips. Invoking Gemini You can request assistance from Gemini at any point by creating a comment using either
Customization To customize Gemini Code Assist for GitHub experience, repository maintainers can create a configuration file and/or provide a custom code review style guide (such as PEP-8 for Python) by creating and adding files to a Limitations & Feedback Gemini Code Assist may make mistakes. Please leave feedback on any instances where its feedback is incorrect or counter productive. You can react with 👍 and 👎 on @gemini-code-assist comments. If you're interested in giving your feedback about your experience with Gemini Code Assist for Github and other Google products, sign up here. You can also get AI-powered code generation, chat, as well as code reviews directly in the IDE at no cost with the Gemini Code Assist IDE Extension. Footnotes
|
There was a problem hiding this comment.
Choose a reason for hiding this comment
The reason will be displayed to describe this comment to others. Learn more.
Code Review
This pull request introduces support for SVD quantization using the nunchaku library. It adds new configuration options, LoRA wrapper modules for quantized layers, and a quantized DiT model for Qwen Image. The pipeline logic is also updated to detect and load these quantized models. My review has identified a few critical bugs in the new LoRA handling logic for quantized layers that need to be addressed. Specifically, there are issues with rank management when fusing QKV LoRAs and a variable scope bug when loading them. There's also a minor bug in the clear method of a LoRA wrapper.
diffsynth_engine/configs/pipeline.py
Outdated
| # override OptimizationConfig | ||
| fbcache_relative_l1_threshold = 0.009 | ||
|
|
||
| # svd quant |
There was a problem hiding this comment.
Choose a reason for hiding this comment
The reason will be displayed to describe this comment to others. Learn more.
看后面逻辑这几个参数似乎不是给用户设置的?可以参考WanPipelineConfig里面boundary那几个参数设置成不能init的field
There was a problem hiding this comment.
Choose a reason for hiding this comment
The reason will be displayed to describe this comment to others. Learn more.
done
There was a problem hiding this comment.
Choose a reason for hiding this comment
The reason will be displayed to describe this comment to others. Learn more.
config里面还是保留一下字段比较好?用Field(init=False)的方式来写的话用户就不能初始化,但是还是有字段提示
| "4. Install it using pip, for example:\n" | ||
| " pip install nunchaku @ https://.../your_specific_nunchaku_file.whl\n" | ||
| ) | ||
| raise ImportError(error_message) |
There was a problem hiding this comment.
Choose a reason for hiding this comment
The reason will be displayed to describe this comment to others. Learn more.
这个error_message还是挺长的,感觉可以从flag里面引进来
No description provided.