File tree Expand file tree Collapse file tree 2 files changed +4
-7
lines changed
src/compressed_tensors/quantization
tests/test_quantization/lifecycle Expand file tree Collapse file tree 2 files changed +4
-7
lines changed Original file line number Diff line number Diff line change @@ -230,10 +230,6 @@ def from_pretrained(
230
230
group_name = "group_" + str (idx )
231
231
config_groups [group_name ] = scheme
232
232
233
- # TODO: this is incorrect in compressed mode, since we are overwriting the
234
- # original weight we lose the uncompressed bit_depth indo
235
- compression_ratio = calculate_compression_ratio (model )
236
-
237
233
if format is None :
238
234
if quantization_status == QuantizationStatus .COMPRESSED :
239
235
format = CompressionFormat .int_quantized .value
@@ -244,7 +240,7 @@ def from_pretrained(
244
240
config_groups = config_groups ,
245
241
quantization_status = quantization_status ,
246
242
kv_cache_scheme = kv_cache_scheme ,
247
- global_compression_ratio = compression_ratio ,
243
+ global_compression_ratio = None ,
248
244
format = format ,
249
245
ignore = consolidated_ignore ,
250
246
)
Original file line number Diff line number Diff line change @@ -184,8 +184,9 @@ def test_serialize_config_tinyllama():
184
184
assert serialized_config .format == CompressionFormat .dense .value
185
185
assert serialized_config .quant_method == DEFAULT_QUANTIZATION_METHOD
186
186
assert serialized_config .ignore == ["model.layers.1.mlp.down_proj" ]
187
- assert serialized_config .global_compression_ratio > 1.0
188
- assert serialized_config .global_compression_ratio < 8.0
187
+ if serialized_config .global_compression_ratio is not None :
188
+ assert serialized_config .global_compression_ratio > 1.0
189
+ assert serialized_config .global_compression_ratio < 8.0
189
190
190
191
191
192
def _test_layer_quantization_status (
You can’t perform that action at this time.
0 commit comments