We read every piece of feedback, and take your input very seriously.
To see all available qualifiers, see our documentation.
There was an error while loading. Please reload this page.
1 parent f460983 commit 76ddeffCopy full SHA for 76ddeff
vllm/model_executor/layers/quantization/utils/fp8_utils.py
@@ -378,8 +378,6 @@ def per_token_group_quant_fp8(
378
is supported for now.
379
column_major_scales: Outputs scales in column major.
380
out_q: Optional output tensor. If not provided, function will create.
381
- tuple[torch.Tensor, torch.Tensor]: The quantized tensor and the
382
- scaling factor for quantization.
383
Returns:
384
tuple[torch.Tensor, torch.Tensor]: The quantized tensor and the
385
scaling factor.
0 commit comments