diff --git a/torch/testing/_internal/distributed/distributed_test.py b/torch/testing/_internal/distributed/distributed_test.py index 353bd3b981e0..e684264c1265 100644 --- a/torch/testing/_internal/distributed/distributed_test.py +++ b/torch/testing/_internal/distributed/distributed_test.py @@ -5410,6 +5410,7 @@ def add(fut): f"The {BACKEND} backend does not support DistributedDataParallel", ) @skip_if_no_gpu + @skip_if_rocm def test_DistributedDataParallel(self): _group, _group_id, rank = self._init_global_test() rank_to_GPU = init_multigpu_helper(dist.get_world_size(), BACKEND)