Skip to content

Commit 4821d26

Browse files
manman-renfacebook-github-bot
authored andcommitted
Add Gluon FA Blackwell for performance measurement (#268)
Summary: Copy python/tutorials/gluon/01-attention-forward.py from OSS Version from PR7238 Pull Request resolved: #268 Reviewed By: xuzhao9 Differential Revision: D77901791 Pulled By: manman-ren fbshipit-source-id: 4626816698fd55efc999f498660e02417b3b6b11
1 parent b519f89 commit 4821d26

File tree

5 files changed

+1232
-0
lines changed

5 files changed

+1232
-0
lines changed

test/test_gpu/skip_tests_h100_pytorch.yaml

Lines changed: 2 additions & 0 deletions
Original file line numberDiff line numberDiff line change
@@ -12,6 +12,8 @@ flash_attention:
1212
- triton_tutorial_flash_v2_ws
1313
- triton_tutorial_flash_v2_tma_ws
1414
- triton_tutorial_flash_v2_tma_ws_persistent
15+
blackwell_attentions:
16+
- gluon_blackwell_fwd
1517
# the two requires full fbgemm instead of genai flavor
1618
fp8_gemm_rowwise:
1719
fp8_gemm_rowwise_grouped:

test/test_gpu/skip_tests_h100_triton_main.yaml

Lines changed: 2 additions & 0 deletions
Original file line numberDiff line numberDiff line change
@@ -12,6 +12,8 @@ flash_attention:
1212
- triton_tutorial_flash_v2_ws
1313
- triton_tutorial_flash_v2_tma_ws
1414
- triton_tutorial_flash_v2_tma_ws_persistent
15+
blackwell_attentions:
16+
- gluon_blackwell_fwd
1517
# the two requires full fbgemm instead of genai flavor
1618
fp8_gemm_rowwise:
1719
fp8_gemm_rowwise_grouped:

tritonbench/kernels/attention_utils.py

Lines changed: 1 addition & 0 deletions
Original file line numberDiff line numberDiff line change
@@ -16,6 +16,7 @@
1616
PEEL_LAST = os.getenv("PEEL_LAST_ITER")
1717
WITH_TMA = os.getenv("WITH_TMA")
1818
HAS_EXPLICIT_WS = os.getenv("ENABLE_EXPLICIT_WS")
19+
SUPPORT_GLUON = os.getenv("WITH_GLUON")
1920

2021

2122
class TmaAutoTuneHelper:

0 commit comments

Comments
 (0)