Skip to content

Commit 716c3ee

Browse files
authored
Update document of orchestration (#1323)
1 parent bd17c49 commit 716c3ee

File tree

1 file changed

+9
-3
lines changed

1 file changed

+9
-3
lines changed

docs/orchestration.md

Lines changed: 9 additions & 3 deletions
Original file line numberDiff line numberDiff line change
@@ -9,15 +9,16 @@ pruning and then distillation and then quantization.
99
## Validated Orchestration Types
1010

1111
### One-shot
12-
12+
Since quantization-aware training, pruning and distillation all have training processes, we can achieve the goal of optimization through one shot training.
1313
- Pruning during quantization-aware training
1414
- Distillation with pattern lock pruning
1515
- Distillation with pattern lock pruning and quantization-aware training
1616

1717
### Multi-shot
18-
18+
Of course, besides one-shot, we also support separate execution of each optimization process.
1919
- Pruning and then post-training quantization
2020
- Distillation and then post-training quantization
21+
- Distillation, then pruning and post-training quantization
2122

2223
## Orchestration user facing API
2324

@@ -54,4 +55,9 @@ opt_model = scheduler.fit()
5455

5556
### Examples
5657

57-
For orchestration related examples, please refer to [Orchestration examples](../examples/README.md).
58+
For orchestration one-shot related examples, please refer to [One-shot examples](../examples/pytorch/nlp/huggingface_models/question-answering/optimization_pipeline/prune_once_for_all/fx/README.md).
59+
60+
For orchestration multi-shot related examples, please refer to [Multi-shot examples](../examples/pytorch/image_recognition/torchvision_models/optimization_pipeline/).
61+
62+
### Publications
63+
All the experiments from [Prune Once for ALL](https://arxiv.org/abs/2111.05754) can be reproduced using [Optimum-Intel](https://github.com/huggingface/optimum-intel) with Intel Neural Compressor.

0 commit comments

Comments
 (0)