Skip to content

Commit 07c46d9

Browse files
Add ToC
1 parent f23d95b commit 07c46d9

File tree

1 file changed

+12
-0
lines changed

1 file changed

+12
-0
lines changed

unit4/README.md

Lines changed: 12 additions & 0 deletions
Original file line numberDiff line numberDiff line change
@@ -13,6 +13,18 @@ Here are the steps for this unit:
1313

1414
:loudspeaker: Don't forget to join the [Discord](https://huggingface.co/join/discord), where you can discuss the material and share what you've made in the `#diffusion-models-class` channel.
1515

16+
## Table of Contents
17+
18+
- [Faster Sampling via Distillation](#faster-sampling-via-distillation)
19+
- [Training Improvements](#training-improvements)
20+
- [More Control for Generation and Editing](more-control-for-generation-and-editing)
21+
- [Video](#video)
22+
- [Audio](#audio)
23+
- [New Architectures and Approaches - Towards 'Iterative Refinement'](#new-architectures-and-approaches---towards-iterative-refinement)
24+
- [Hands-On Notebooks](#hands-on-notebooks)
25+
- [Where Next?](#where-next)
26+
27+
1628
## Faster Sampling via Distillation
1729

1830
Progressive distillation is a technique for taking an existing diffusion model and using it to train a new version of the model that requires fewer steps for inference. The 'student' model is initialized from the weights of the 'teacher' model. During training, the teacher model performs two sampling steps and the student model tries to match the resulting prediction in a single step. This process can be repeated mutiple times, with the previous iteration's student model becoming the teacher for the next stage. The end result is a model that can produce decent samples in much fewer steps (typically 4 or 8) than the original teacher model. The core mechanism is illustrated in this diagram from the [paper that introduced the idea](http://arxiv.org/abs/2202.00512):

0 commit comments

Comments
 (0)