@@ -391,8 +391,8 @@ def __call__(
391391 sigmas : List [float ] = None ,
392392 guidance_scale : float = 3.5 ,
393393 num_images_per_prompt : Optional [int ] = 1 ,
394- height : Optional [int ] = 512 ,
395- width : Optional [int ] = 512 ,
394+ height : Optional [int ] = 1024 ,
395+ width : Optional [int ] = 1024 ,
396396 generator : Optional [Union [torch .Generator , List [torch .Generator ]]] = None ,
397397 latents : Optional [torch .Tensor ] = None ,
398398 prompt_embeds : Optional [torch .Tensor ] = None ,
@@ -415,9 +415,9 @@ def __call__(
415415 `negative_prompt_embeds` instead. Ignored when not using guidance (i.e., ignored if `guidance_scale` is
416416 less than `1`).
417417 height (`int`, *optional*, defaults to self.transformer.config.sample_size * self.vae_scale_factor):
418- The height in pixels of the generated image. This is set to 512 by default.
418+ The height in pixels of the generated image. This is set to 1024 by default for best results .
419419 width (`int`, *optional*, defaults to self.transformer.config.sample_size * self.vae_scale_factor):
420- The width in pixels of the generated image. This is set to 512 by default.
420+ The width in pixels of the generated image. This is set to 1024 by default for best results .
421421 num_inference_steps (`int`, *optional*, defaults to 50):
422422 The number of denoising steps. More denoising steps usually lead to a higher quality image at the
423423 expense of slower inference.
0 commit comments