Skip to content

Commit 9d6808e

Browse files
[0.9.1]disable quantization in mindie_turbo (#1749)
### What this PR does / why we need it? since the interface in vllm-ascend has changed so quickly, the quantization function in mindie_turbo is no longer needed, so it needs to be discarded. ### Does this PR introduce _any_ user-facing change? no ### How was this patch tested? through ci Signed-off-by: zouyida <zouyida@huawei.com> Co-authored-by: zouyida <zouyida@huawei.com>
1 parent df18f1d commit 9d6808e

File tree

1 file changed

+2
-8
lines changed

1 file changed

+2
-8
lines changed

vllm_ascend/quantization/quantizer.py

Lines changed: 2 additions & 8 deletions
Original file line numberDiff line numberDiff line change
@@ -48,14 +48,8 @@ def get_quantizer(cls,
4848
if quantization_algorithm in CUSTOMIZED_QUANTIZER_TYPE:
4949
return
5050

51-
try:
52-
module = importlib.import_module("mindie_turbo")
53-
MindIETurboQuantizer = module.MindIETurboQuantizer
54-
return MindIETurboQuantizer.get_quantizer(quant_config, prefix,
55-
packed_modules_mapping)
56-
except ImportError:
57-
return VLLMAscendQuantizer.get_quantizer(quant_config, prefix,
58-
packed_modules_mapping)
51+
return VLLMAscendQuantizer.get_quantizer(quant_config, prefix,
52+
packed_modules_mapping)
5953

6054
def build_linear_method(self):
6155
raise NotImplementedError

0 commit comments

Comments
 (0)