diff --git a/tests/comm/test_trtllm_mnnvl_allreduce_custom_comm.py b/tests/comm/test_trtllm_mnnvl_allreduce_custom_comm.py index 60933cf89b..772ceead0b 100644 --- a/tests/comm/test_trtllm_mnnvl_allreduce_custom_comm.py +++ b/tests/comm/test_trtllm_mnnvl_allreduce_custom_comm.py @@ -250,7 +250,7 @@ def test_mnnvl_allreduce_custom_communicator( available_gpus = torch.cuda.device_count() if world_size > available_gpus: - raise ValueError( + pytest.skip( f"world_size {world_size} is greater than available_gpus {available_gpus}" ) print(f"Running test for world_size={world_size}")