Skip to content

Commit 0a39766

Browse files
committed
ruff format
Summary: Test Plan: Reviewers: Subscribers: Tasks: Tags:
1 parent 9fa36c5 commit 0a39766

File tree

1 file changed

+11
-11
lines changed

1 file changed

+11
-11
lines changed

test/quantization/test_moe_quant.py

Lines changed: 11 additions & 11 deletions
Original file line numberDiff line numberDiff line change
@@ -103,7 +103,7 @@ def test_int4wo_fake_dim(self, name, num_tokens, fullgraph):
103103
self.skipTest("Need CUDA available")
104104
if not TORCH_VERSION_AT_LEAST_2_5:
105105
self.skipTest("Test only enabled for 2.5+")
106-
106+
107107
config = MoEQuantConfig(Int4WeightOnlyConfig())
108108
tensor_impl_class = TensorCoreTiledAQTTensorImpl
109109

@@ -127,7 +127,7 @@ def test_int4wo_base(self, name, num_tokens, fullgraph):
127127
self.skipTest("Requires CUDA capability >= 9.0")
128128
if not TORCH_VERSION_AT_LEAST_2_5:
129129
self.skipTest("Test only enabled for 2.5+")
130-
130+
131131
config = Int4WeightOnlyConfig()
132132
tensor_impl_class = TensorCoreTiledAQTTensorImpl
133133

@@ -149,7 +149,7 @@ def test_int8wo_fake_dim(self, name, num_tokens, fullgraph):
149149
self.skipTest("Need CUDA available")
150150
if not TORCH_VERSION_AT_LEAST_2_5:
151151
self.skipTest("Test only enabled for 2.5+")
152-
152+
153153
config = MoEQuantConfig(Int8WeightOnlyConfig())
154154
tensor_impl_class = PlainAQTTensorImpl
155155

@@ -171,7 +171,7 @@ def test_int8wo_base(self, name, num_tokens, fullgraph):
171171
self.skipTest("Need CUDA available")
172172
if not TORCH_VERSION_AT_LEAST_2_5:
173173
self.skipTest("Test only enabled for 2.5+")
174-
174+
175175
config = Int8WeightOnlyConfig()
176176
tensor_impl_class = PlainAQTTensorImpl
177177

@@ -191,7 +191,7 @@ def test_int8wo_base(self, name, num_tokens, fullgraph):
191191
def test_int8wo_base_cpu(self, name, num_tokens, fullgraph):
192192
if not TORCH_VERSION_AT_LEAST_2_5:
193193
self.skipTest("Test only enabled for 2.5+")
194-
194+
195195
config = Int8WeightOnlyConfig()
196196
tensor_impl_class = PlainAQTTensorImpl
197197

@@ -213,7 +213,7 @@ def test_int8dq_fake_dim(self, name, num_tokens, fullgraph):
213213
self.skipTest("Need CUDA available")
214214
if not TORCH_VERSION_AT_LEAST_2_5:
215215
self.skipTest("Test only enabled for 2.5+")
216-
216+
217217
config = MoEQuantConfig(Int8DynamicActivationInt8WeightConfig())
218218
base_class = LinearActivationQuantizedTensor
219219

@@ -235,7 +235,7 @@ def test_int8dq_base(self, name, num_tokens, fullgraph):
235235
self.skipTest("Need CUDA available")
236236
if not TORCH_VERSION_AT_LEAST_2_5:
237237
self.skipTest("Test only enabled for 2.5+")
238-
238+
239239
config = Int8DynamicActivationInt8WeightConfig()
240240
base_class = LinearActivationQuantizedTensor
241241

@@ -258,7 +258,7 @@ def test_fp8wo_fake_dim(self, name, num_tokens, fullgraph):
258258
self.skipTest("Need CUDA available")
259259
if not is_sm_at_least_90():
260260
self.skipTest("Requires CUDA capability >= 9.0")
261-
261+
262262
config = MoEQuantConfig(Float8WeightOnlyConfig())
263263
tensor_impl_class = Float8AQTTensorImpl
264264

@@ -280,7 +280,7 @@ def test_fp8wo_base(self, name, num_tokens, fullgraph):
280280
self.skipTest("Need CUDA available")
281281
if not is_sm_at_least_90():
282282
self.skipTest("Requires CUDA capability >= 9.0")
283-
283+
284284
config = Float8WeightOnlyConfig()
285285
tensor_impl_class = Float8AQTTensorImpl
286286

@@ -302,7 +302,7 @@ def test_fp8dq_fake_dim(self, name, num_tokens, fullgraph):
302302
self.skipTest("Need CUDA available")
303303
if not is_sm_at_least_90():
304304
self.skipTest("Requires CUDA capability >= 9.0")
305-
305+
306306
config = MoEQuantConfig(Float8DynamicActivationFloat8WeightConfig())
307307
base_class = LinearActivationQuantizedTensor
308308

@@ -324,7 +324,7 @@ def test_fp8dq_base(self, name, num_tokens, fullgraph):
324324
self.skipTest("Need CUDA available")
325325
if not is_sm_at_least_90():
326326
self.skipTest("Requires CUDA capability >= 9.0")
327-
327+
328328
config = Float8DynamicActivationFloat8WeightConfig()
329329
base_class = LinearActivationQuantizedTensor
330330

0 commit comments

Comments
 (0)