Skip to content

Conversation

@ttompa
Copy link
Contributor

@ttompa ttompa commented Dec 1, 2025

This PR improves inference performance during validation and final model saving by merging LoRA weights back into the base model when gradients are not needed. This eliminates the inference overhead caused by additional torch calls in the LoRA implementation.
This is based off of and requires the E0 estimation branch.

@ttompa ttompa marked this pull request as ready for review December 1, 2025 21:34
@ttompa ttompa marked this pull request as draft December 1, 2025 21:35
Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment

Labels

None yet

Projects

None yet

Development

Successfully merging this pull request may close these issues.

1 participant