You signed in with another tab or window. Reload to refresh your session.You signed out in another tab or window. Reload to refresh your session.You switched accounts on another tab or window. Reload to refresh your session.Dismiss alert
Copy file name to clipboardExpand all lines: docs/source/en/using-diffusers/unconditional_image_generation.md
+18-31Lines changed: 18 additions & 31 deletions
Display the source diff
Display the rich diff
Original file line number
Diff line number
Diff line change
@@ -14,54 +14,41 @@ specific language governing permissions and limitations under the License.
14
14
15
15
[[open-in-colab]]
16
16
17
-
Unconditional image generation is a relatively straightforward task. The model only generates images - without any additional context like text or an image - resembling the training data it was trained on.
17
+
Unconditional image generation generates images that look like a random sample from the training data the model was trained on because the denoising process is not guided by any additional context like text or image.
18
18
19
-
The [`DiffusionPipeline`]is the easiest way to use a pre-trained diffusion system for inference.
19
+
To get started, use the [`DiffusionPipeline`]to load the [anton-l/ddpm-butterflies-128](https://huggingface.co/anton-l/ddpm-butterflies-128) checkpoint to generate images of butterflies. The [`DiffusionPipeline`] downloads and caches all the model components required to generate an image.
20
20
21
-
Start by creating an instance of [`DiffusionPipeline`] and specify which pipeline checkpoint you would like to download.
22
-
You can use any of the 🧨 Diffusers [checkpoints](https://huggingface.co/models?library=diffusers&sort=downloads) from the Hub (the checkpoint you'll use generates images of butterflies).
💡 Want to train your own unconditional image generation model? Take a look at the training [guide](../training/unconditional_training) to learn how to generate your own images.
31
+
Want to generate images of something else? Take a look at the training [guide](../training/unconditional_training) to learn how to train a model to generate your own images.
27
32
28
33
</Tip>
29
34
30
-
In this guide, you'll use [`DiffusionPipeline`] for unconditional image generation with [DDPM](https://arxiv.org/abs/2006.11239):
The output image is a [`PIL.Image`](https://pillow.readthedocs.io/en/stable/reference/Image.html?highlight=image#the-image-class) object that can be saved:
37
36
38
-
The [`DiffusionPipeline`] downloads and caches all modeling, tokenization, and scheduling components.
39
-
Because the model consists of roughly 1.4 billion parameters, we strongly recommend running it on a GPU.
40
-
You can move the generator object to a GPU, just like you would in PyTorch:
41
-
42
-
```python
43
-
generator.to("cuda")
37
+
```py
38
+
image.save("generated_image.png")
44
39
```
45
40
46
-
Now you can use the `generator`to generate an image:
41
+
You can also try experimenting with the `num_inference_steps` parameter, which controls the number of denoising steps. More denoising steps typically produce higher quality images, but it'll take longer to generate. Feel free to play around with this parameter to see how it affects the image quality.
The output is by default wrapped into a [`PIL.Image`](https://pillow.readthedocs.io/en/stable/reference/Image.html?highlight=image#the-image-class) object.
54
-
55
-
You can save the image by calling:
56
-
57
-
```python
58
-
image.save("generated_image.png")
59
-
```
60
-
61
-
Try out the Spaces below, and feel free to play around with the inference steps parameter to see how it affects the image quality!
48
+
Try out the Space below to generate an image of a butterfly!
0 commit comments