@@ -17,29 +17,29 @@ Here is a basic example for exporting Qwen 3, although please refer to the Llama
17
17
Export 0.6b to XNNPack, quantized with 8da4w:
18
18
```
19
19
python -m extension.llm.export.export_llm \
20
- --config examples/models/qwen3/config/qwen3_xnnpack_q8da4w.yaml
20
+ --config examples/models/qwen3/config/qwen3_xnnpack_q8da4w.yaml \
21
21
+base.model_class="qwen3_0_6b" \
22
22
+base.params="examples/models/qwen3/config/0_6b_config.json" \
23
- +export.output_name="qwen3_0_6b.pte" \
23
+ +export.output_name="qwen3_0_6b.pte"
24
24
25
25
```
26
26
27
27
Export 1.7b to XNNPack, quantized with 8da4w:
28
28
```
29
29
python -m extension.llm.export.export_llm \
30
- --config examples/models/qwen3/config/qwen3_xnnpack_q8da4w.yaml
30
+ --config examples/models/qwen3/config/qwen3_xnnpack_q8da4w.yaml \
31
31
+base.model_class="qwen3_1_7b" \
32
32
+base.params="examples/models/qwen3/config/1_7b_config.json" \
33
- +export.output_name="qwen3_1_7b.pte" \
33
+ +export.output_name="qwen3_1_7b.pte"
34
34
```
35
35
36
36
Export 4b to XNNPack, quantized with 8da4w:
37
37
```
38
38
python -m extension.llm.export.export_llm \
39
- --config examples/models/qwen3/config/qwen3_xnnpack_q8da4w.yaml
39
+ --config examples/models/qwen3/config/qwen3_xnnpack_q8da4w.yaml \
40
40
+base.model_class="qwen3_4b" \
41
41
+base.params="examples/models/qwen3/config/4b_config.json" \
42
- +export.output_name="qwen3_4b.pte" \
42
+ +export.output_name="qwen3_4b.pte"
43
43
```
44
44
45
45
### Example run
0 commit comments