We read every piece of feedback, and take your input very seriously.
To see all available qualifiers, see our documentation.
There was an error while loading. Please reload this page.
1 parent 7d5f7c9 commit cd28f40Copy full SHA for cd28f40
src/llmcompressor/transformers/compression/quantization_format.py
@@ -3,9 +3,7 @@
3
from compressed_tensors import CompressionFormat
4
from compressed_tensors.config import SparsityStructure
5
from compressed_tensors.quantization import QuantizationStrategy, QuantizationType
6
-from compressed_tensors.quantization.utils import (
7
- is_module_quantized,
8
-)
+from compressed_tensors.quantization.utils import is_module_quantized
9
10
__all__ = ["infer_quantization_format"]
11
@@ -50,8 +48,6 @@ def infer_quantization_format(
50
48
return quantization_format
51
49
52
weight_args, input_args = _get_unique_quant_args(model)
53
-
54
- # no quantization format if no weights are quantized
55
if len(weight_args) <= 0:
56
return None
57
0 commit comments