@@ -34,7 +34,6 @@ def test_dense(rank, world_size, port, tp_size):
3434 mem_fraction_static = 0.4 ,
3535 # enable_torch_compile=True,
3636 enable_nccl_nvls = True ,
37- # enable_symm_mem=True,
3837 enable_symm_mem = False ,
3938 enable_torch_compile = True ,
4039 enable_dp_attention = False ,
@@ -73,9 +72,6 @@ def test_moe(rank, world_size, port, tp_size):
7372 mem_fraction_static = 0.4 ,
7473 enable_torch_compile = True ,
7574 enable_nccl_nvls = True ,
76- # enable_symm_mem=True,
77- # enable_dp_attention=True,
78- # enable_dp_lm_head=True,
7975 enable_symm_mem = False ,
8076 enable_dp_attention = False ,
8177 enable_dp_lm_head = False ,
@@ -213,11 +209,8 @@ def test_vlm(rank, world_size, port, tp_size):
213209 attention_backend = "fa3" ,
214210 mem_fraction_static = 0.75 ,
215211 enable_torch_compile = True ,
216- # enable_nccl_nvls=False,
217212 enable_nccl_nvls = True ,
218213 enable_symm_mem = False , # Disable to avoid nccl_allocator compilation issues
219- # enable_dp_attention=True,
220- # enable_dp_lm_head=True,
221214 enable_dp_attention = False ,
222215 enable_dp_lm_head = False ,
223216 enable_piecewise_cuda_graph = True ,
@@ -376,10 +369,7 @@ def test_vlm_multi_batch(rank, world_size, port, tp_size):
376369 mem_fraction_static = 0.4 ,
377370 enable_nccl_nvls = True ,
378371 enable_torch_compile = True ,
379- # enable_nccl_nvls=False,
380372 enable_symm_mem = False ,
381- # enable_dp_attention=True,
382- # enable_dp_lm_head=True,
383373 enable_dp_attention = False ,
384374 enable_dp_lm_head = False ,
385375 enable_piecewise_cuda_graph = True ,
0 commit comments