Skip to content

Commit 9ca6d85

Browse files
authored
Fixing assert bug (#89)
Summary: we support handling fp32, fp16 and bf16 tensors with quantization, asserting we only handle bf16 was breaking things. Test Plan: python test/test.py -k "test_int8_dynamic_quant_subclass" Reviewers: Subscribers: Tasks: Tags: ghstack-source-id: 6fd2e01 Pull Request resolved: #88
1 parent 56fb1b7 commit 9ca6d85

File tree

1 file changed

+1
-1
lines changed

1 file changed

+1
-1
lines changed

torchao/kernel/intmm.py

Lines changed: 1 addition & 1 deletion
Original file line numberDiff line numberDiff line change
@@ -64,7 +64,7 @@ def int_scaled_matmul(a, b, scales1):
6464
assert M == scales1.size(0)
6565
assert 1 == scales1.size(1)
6666
assert scales1.is_contiguous()
67-
assert scales1.dtype == torch.bfloat16
67+
6868
scales1 = scales1.expand((M, N))
6969
assert scales1.dim() == 2
7070
if intmm_triton is not None and AUTOTUNER_ENABLE:

0 commit comments

Comments
 (0)