Skip to content

Commit c34bbca

Browse files
committed
add module ut
Signed-off-by: Enwei Zhu <[email protected]>
1 parent e47b434 commit c34bbca

File tree

1 file changed

+6
-3
lines changed

1 file changed

+6
-3
lines changed

tests/unittest/_torch/modules/test_fused_moe.py

Lines changed: 6 additions & 3 deletions
Original file line numberDiff line numberDiff line change
@@ -1346,13 +1346,15 @@ def test_fused_moe_fp8_blockwise_cute_dsl_multi_gpu(ep_size, routing_method,
13461346

13471347
@skip_pre_blackwell
13481348
@pytest.mark.parametrize("dtype", [torch.float16, torch.bfloat16])
1349-
@pytest.mark.parametrize(
1350-
"moe_backend",
1351-
[pytest.param("TRTLLM", marks=skip_blackwell_geforce), "CUTLASS"])
1349+
@pytest.mark.parametrize("moe_backend", [
1350+
pytest.param("TRTLLM", marks=skip_blackwell_geforce), "CUTLASS", "CUTEDSL"
1351+
])
13521352
def test_fused_moe_nvfp4(dtype, moe_backend):
13531353

13541354
if moe_backend == "TRTLLM" and dtype == torch.float16:
13551355
pytest.skip("TRTLLM NVFP4 MoE backend does not support float16 yet")
1356+
if moe_backend == "CUTEDSL" and dtype == torch.float16:
1357+
pytest.skip("CUTEDSL NVFP4 MoE backend does not support float16 yet")
13561358

13571359
test_all_kernels = True
13581360
if get_sm_version() == 120:
@@ -1450,6 +1452,7 @@ def test_fused_moe_nvfp4(dtype, moe_backend):
14501452
moe_backend=moe_backend),
14511453
)
14521454
fused_moe.load_weights([weights])
1455+
fused_moe.post_load_weights()
14531456
fused_moe.cuda()
14541457

14551458
# Evaluate the outputs on a variant sequence length to cover all possible keys in Autotuner cache

0 commit comments

Comments
 (0)