Skip to content

Commit 34dffa5

Browse files
handle out != None
1 parent cb1eae9 commit 34dffa5

File tree

2 files changed

+1
-8
lines changed

2 files changed

+1
-8
lines changed

torchao/prototype/moe_training/scaled_grouped_mm.py

Lines changed: 0 additions & 8 deletions
Original file line numberDiff line numberDiff line change
@@ -40,16 +40,8 @@ def _scaled_grouped_mm(
4040
offs (int32 torch.Tensor): The offsets to use to mark the starting index of each group along dim0 of the A tensor.
4141
out_dtype (Optional[torch.dtype]): The dtype of the output tensor. Currently only torch.bfloat16 is supported.
4242
"""
43-
<<<<<<< HEAD
44-
<<<<<<< HEAD
45-
# logger.info("Using scaled_grouped_mm")
46-
=======
47-
#logger.info("Using scaled_grouped_mm")
48-
>>>>>>> 6ca070de (handle out != None)
49-
=======
5043
# TODO: Remove once prototype is more mature. This is currently very useful for development and debugging.
5144
logger.info("Using scaled_grouped_mm")
52-
>>>>>>> 2f3bb137 (add tp support for fp8 moe training)
5345
return _Float8GroupedMM.apply(
5446
A,
5547
B_t,

torchao/prototype/moe_training/tensor.py

Lines changed: 1 addition & 0 deletions
Original file line numberDiff line numberDiff line change
@@ -183,6 +183,7 @@ def fsdp_post_all_gather(
183183
f"{out_data.dtype} {param_dtype}"
184184
)
185185
out_data.copy_(data)
186+
186187
return
187188

188189
# For training step 0, out=None, so we need to return a new ScaledGroupedMMTensor.

0 commit comments

Comments
 (0)