From 6b6acac0c796b40ef200c8b24e16b01fdda572b5 Mon Sep 17 00:00:00 2001 From: ramcherukuri Date: Wed, 21 Aug 2024 04:27:15 +0000 Subject: [PATCH] skip test_DistributedDataParallel --- torch/testing/_internal/distributed/distributed_test.py | 1 + 1 file changed, 1 insertion(+) diff --git a/torch/testing/_internal/distributed/distributed_test.py b/torch/testing/_internal/distributed/distributed_test.py index 90de20eb9559..596f2c253583 100644 --- a/torch/testing/_internal/distributed/distributed_test.py +++ b/torch/testing/_internal/distributed/distributed_test.py @@ -5657,6 +5657,7 @@ def add(fut): f"The {BACKEND} backend does not support DistributedDataParallel", ) @skip_if_no_gpu + @skip_if_rocm def test_DistributedDataParallel(self): group, group_id, rank = self._init_global_test() rank_to_GPU = init_multigpu_helper(dist.get_world_size(), BACKEND)