Skip to content

Commit 8a8fc94

Browse files
authored
[Log] Debugging Log with more Information (#20770)
Signed-off-by: yewentao256 <zhyanwentao@126.com>
1 parent 4de7146 commit 8a8fc94

File tree

2 files changed

+37
-13
lines changed

2 files changed

+37
-13
lines changed

vllm/model_executor/layers/fused_moe/cutlass_moe.py

Lines changed: 17 additions & 9 deletions
Original file line numberDiff line numberDiff line change
@@ -571,34 +571,42 @@ def _valid_cutlass_block_scaled_grouped_gemm_shape(N: int, K: int):
571571

572572
_, K, N = w2.size()
573573
if not _valid_cutlass_block_scaled_grouped_gemm_shape(N, K):
574-
logger.debug(
575-
"CutlassBlockScaledGroupedGemm disabled: unalinged problem size.")
574+
logger.debug_once(
575+
"CutlassBlockScaledGroupedGemm disabled: unaligned problem size. "
576+
"N: %s, K: %s",
577+
N,
578+
K,
579+
)
576580
return False
577581

578582
if (w1.dtype != torch.float8_e4m3fn or w2.dtype != torch.float8_e4m3fn):
579-
logger.debug(
580-
"CutlassBlockScaledGroupedGemm disabled: invalid weight dtype(s).")
583+
logger.debug_once(
584+
"CutlassBlockScaledGroupedGemm disabled: invalid weight dtype(s). "
585+
"w1.dtype: %s, w2.dtype: %s",
586+
w1.dtype,
587+
w2.dtype,
588+
)
581589
return False
582590

583591
if expert_map is not None:
584-
logger.debug(
592+
logger.debug_once(
585593
"CutlassBlockScaledGroupedGemm disabled: expert_parallel is"
586594
" not supported.")
587595
return False
588596

589597
if activation != "silu":
590-
logger.debug(
598+
logger.debug_once(
591599
"CutlassBlockScaledGroupedGemm disabled: only activation silu is"
592600
" supported.")
593601
return False
594602

595603
if apply_router_weight_on_input:
596-
logger.debug("CutlassBlockScaledGroupedGemm disabled:"
597-
" apply_router_weight_on_input is not supported.")
604+
logger.debug_once("CutlassBlockScaledGroupedGemm disabled:"
605+
" apply_router_weight_on_input is not supported.")
598606
return False
599607

600608
if inplace:
601-
logger.debug(
609+
logger.debug_once(
602610
"CutlassBlockScaledGroupedGemm disabled: inplace is not supported."
603611
)
604612
return False

vllm/model_executor/layers/fused_moe/deep_gemm_moe.py

Lines changed: 20 additions & 4 deletions
Original file line numberDiff line numberDiff line change
@@ -50,17 +50,33 @@ def _valid_deep_gemm(hidden_states: torch.Tensor, w1: torch.Tensor,
5050
M = hidden_states.size(0)
5151
_, K, N = w2.size()
5252
if not _valid_deep_gemm_shape(M, N, K):
53-
logger.debug("DeepGemm disabled: unaligned problem size.")
53+
logger.debug_once(
54+
"DeepGemm disabled: unaligned problem size. M: %s, N: %s, K: %s",
55+
M,
56+
N,
57+
K,
58+
)
5459
return False
5560

5661
if (w1.dtype != torch.float8_e4m3fn or w2.dtype != torch.float8_e4m3fn):
57-
logger.debug("DeepGemm disabled: invalid weight dtype(s).")
62+
logger.debug_once(
63+
"DeepGemm disabled: invalid weight dtype(s). "
64+
"w1.dtype: %s, w2.dtype: %s",
65+
w1.dtype,
66+
w2.dtype,
67+
)
5868
return False
5969

6070
if (not hidden_states.is_contiguous() or not w1.is_contiguous()
6171
or not w2.is_contiguous()):
62-
logger.debug(
63-
"DeepGemm disabled: weights or activations not contiguous.")
72+
logger.debug_once(
73+
"DeepGemm disabled: weights or activations not contiguous. "
74+
"hidden_states.is_contiguous(): %s, w1.is_contiguous(): %s, "
75+
"w2.is_contiguous(): %s",
76+
hidden_states.is_contiguous(),
77+
w1.is_contiguous(),
78+
w2.is_contiguous(),
79+
)
6480
return False
6581

6682
return True

0 commit comments

Comments
 (0)