From c1fb5d7864ae65446f5c8664113e57fc75749fc7 Mon Sep 17 00:00:00 2001 From: Kirthi Shankar Sivamani Date: Wed, 4 Mar 2026 19:28:16 +0000 Subject: [PATCH] Skip test_nvfp4_partial_cast_matches_full test on hopper Signed-off-by: Kirthi Shankar Sivamani --- .../distributed/test_cast_master_weights_to_fp8.py | 9 +++++---- 1 file changed, 5 insertions(+), 4 deletions(-) diff --git a/tests/pytorch/distributed/test_cast_master_weights_to_fp8.py b/tests/pytorch/distributed/test_cast_master_weights_to_fp8.py index 373f27be2f..1606641b78 100644 --- a/tests/pytorch/distributed/test_cast_master_weights_to_fp8.py +++ b/tests/pytorch/distributed/test_cast_master_weights_to_fp8.py @@ -1101,10 +1101,6 @@ def _test_nvfp4_partial_cast_matches_full(dp_group) -> None: WORLD_RANK = dist.get_rank(dp_group) WORLD_SIZE = dist.get_world_size(dp_group) - available, reason = is_nvfp4_available(return_reason=True) - if not available: - pytest.skip(reason) - torch.manual_seed(1234) device = torch.device("cuda") # Shape must be divisible by WORLD_SIZE for even splitting @@ -1196,6 +1192,11 @@ def _test_nvfp4_partial_cast_matches_full(dp_group) -> None: @pytest.mark.parametrize("world_size", [2]) def test_nvfp4_partial_cast_matches_full(world_size: int) -> None: """Launch a distributed job for NVFP4 partial-cast equivalence test.""" + + available, reason = is_nvfp4_available(return_reason=True) + if not available: + pytest.skip(reason) + python_exe = pathlib.Path(sys.executable).resolve() current_file = pathlib.Path(__file__).resolve() command = [