Skip to content

Commit 3a62889

Browse files
authored
[Feature] Add quant description file for new quant model generated by modelslim (#719)
### What this PR does / why we need it? After discussed with MindStudio about the quantization model format, we decide to support another quant format which may used in new modelslim tool, in which case, `quantization_config` may be removed from the `config.json` file and `quant_model_description.json` will be used for quantization configuration. ### Does this PR introduce _any_ user-facing change? Yes, using the latest quantization format ### How was this patch tested? Test locally Signed-off-by: ganyi <pleaplusone.gy@gmail.com>
1 parent affca6f commit 3a62889

File tree

1 file changed

+1
-1
lines changed

1 file changed

+1
-1
lines changed

vllm_ascend/quantization/quant_config.py

Lines changed: 1 addition & 1 deletion
Original file line numberDiff line numberDiff line change
@@ -67,7 +67,7 @@ def get_min_capability(cls) -> int:
6767

6868
@classmethod
6969
def get_config_filenames(cls) -> List[str]:
70-
return []
70+
return ["quant_model_description.json"]
7171

7272
@classmethod
7373
def from_config(cls, config: Dict[str, Any]) -> "AscendQuantConfig":

0 commit comments

Comments
 (0)