Skip to content

Commit acd926f

Browse files
authored
[docs] Fix local links (huggingface#6440)
fix local links
1 parent 691d8d3 commit acd926f

File tree

1 file changed

+7
-7
lines changed

1 file changed

+7
-7
lines changed

docs/source/en/tutorials/fast_diffusion.md

Lines changed: 7 additions & 7 deletions
Original file line numberDiff line numberDiff line change
@@ -14,15 +14,15 @@ specific language governing permissions and limitations under the License.
1414

1515
Diffusion models are known to be slower than their counter parts, GANs, because of the iterative and sequential reverse diffusion process. Recent works try to address limitation with:
1616

17-
* progressive timestep distillation (such as [LCM LoRA](../using-diffusers/inference_with_lcm_lora.md))
17+
* progressive timestep distillation (such as [LCM LoRA](../using-diffusers/inference_with_lcm_lora))
1818
* model compression (such as [SSD-1B](https://huggingface.co/segmind/SSD-1B))
1919
* reusing adjacent features of the denoiser (such as [DeepCache](https://github.com/horseee/DeepCache))
2020

21-
In this tutorial, we focus on leveraging the power of PyTorch 2 to accelerate the inference latency of text-to-image diffusion pipeline, instead. We will use [Stable Diffusion XL (SDXL)](../using-diffusers/sdxl.md) as a case study, but the techniques we will discuss should extend to other text-to-image diffusion pipelines.
21+
In this tutorial, we focus on leveraging the power of PyTorch 2 to accelerate the inference latency of text-to-image diffusion pipeline, instead. We will use [Stable Diffusion XL (SDXL)](../using-diffusers/sdxl) as a case study, but the techniques we will discuss should extend to other text-to-image diffusion pipelines.
2222

2323
## Setup
2424

25-
Make sure you're on the latest version of `diffusers`:
25+
Make sure you're on the latest version of `diffusers`:
2626

2727
```bash
2828
pip install -U diffusers
@@ -42,7 +42,7 @@ _This tutorial doesn't present the benchmarking code and focuses on how to perfo
4242

4343
## Baseline
4444

45-
Let's start with a baseline. Disable the use of a reduced precision and [`scaled_dot_product_attention`](../optimization/torch2.0.md):
45+
Let's start with a baseline. Disable the use of a reduced precision and [`scaled_dot_product_attention`](../optimization/torch2.0):
4646

4747
```python
4848
from diffusers import StableDiffusionXLPipeline
@@ -104,11 +104,11 @@ _(We later ran the experiments in float16 and found out that the recent versions
104104
* The benefits of using the bfloat16 numerical precision as compared to float16 are hardware-dependent. Modern generations of GPUs tend to favor bfloat16.
105105
* Furthermore, in our experiments, we bfloat16 to be much more resilient when used with quantization in comparison to float16.
106106

107-
We have a [dedicated guide](../optimization/fp16.md) for running inference in a reduced precision.
107+
We have a [dedicated guide](../optimization/fp16) for running inference in a reduced precision.
108108

109109
## Running attention efficiently
110110

111-
Attention blocks are intensive to run. But with PyTorch's [`scaled_dot_product_attention`](../optimization/torch2.0.md), we can run them efficiently.
111+
Attention blocks are intensive to run. But with PyTorch's [`scaled_dot_product_attention`](../optimization/torch2.0), we can run them efficiently.
112112

113113
```python
114114
from diffusers import StableDiffusionXLPipeline
@@ -200,7 +200,7 @@ It provides a minor boost from 2.54 seconds to 2.52 seconds.
200200

201201
<Tip warning={true}>
202202

203-
Support for `fuse_qkv_projections()` is limited and experimental. As such, it's not available for many non-SD pipelines such as [Kandinsky](../using-diffusers/kandinsky.md). You can refer to [this PR](https://github.com/huggingface/diffusers/pull/6179) to get an idea about how to support this kind of computation.
203+
Support for `fuse_qkv_projections()` is limited and experimental. As such, it's not available for many non-SD pipelines such as [Kandinsky](../using-diffusers/kandinsky). You can refer to [this PR](https://github.com/huggingface/diffusers/pull/6179) to get an idea about how to support this kind of computation.
204204

205205
</Tip>
206206

0 commit comments

Comments
 (0)