We read every piece of feedback, and take your input very seriously.
To see all available qualifiers, see our documentation.
There was an error while loading. Please reload this page.
1 parent c2a28c3 commit 4a3e574Copy full SHA for 4a3e574
examples/dreambooth/train_dreambooth_lora.py
@@ -871,7 +871,7 @@ def main(args):
871
lora_attn_processor_class = (
872
LoRAAttnProcessor2_0 if hasattr(F, "scaled_dot_product_attention") else LoRAAttnProcessor
873
)
874
-
+
875
module = lora_attn_processor_class(hidden_size=hidden_size, cross_attention_dim=cross_attention_dim)
876
unet_lora_attn_procs[name] = module
877
unet_lora_parameters.extend(module.parameters())
0 commit comments