Skip to content

Commit af86b0c

Browse files
authored
Update fp16.mdx (huggingface#2746)
Fix typos
1 parent a9f28b6 commit af86b0c

File tree

1 file changed

+1
-1
lines changed

1 file changed

+1
-1
lines changed

docs/source/en/optimization/fp16.mdx

Lines changed: 1 addition & 1 deletion
Original file line numberDiff line numberDiff line change
@@ -221,7 +221,7 @@ image = pipe(prompt).images[0]
221221
Full-model offloading is an alternative that moves whole models to the GPU, instead of handling each model's constituent _modules_. This results in a negligible impact on inference time (compared with moving the pipeline to `cuda`), while still providing some memory savings.
222222

223223
In this scenario, only one of the main components of the pipeline (typically: text encoder, unet and vae)
224-
will be in the GPU while the others wait in the CPU. Compoments like the UNet that run for multiple iterations will stay on GPU until they are no longer needed.
224+
will be in the GPU while the others wait in the CPU. Components like the UNet that run for multiple iterations will stay on GPU until they are no longer needed.
225225

226226
This feature can be enabled by invoking `enable_model_cpu_offload()` on the pipeline, as shown below.
227227

0 commit comments

Comments
 (0)