File tree 1 file changed +2
-2
lines changed
torch/testing/_internal/distributed/rpc
1 file changed +2
-2
lines changed Original file line number Diff line number Diff line change 32
32
skip_if_lt_x_gpu ,
33
33
captured_output ,
34
34
tp_transports ,
35
- skip_if_rocm ,
36
35
)
37
36
from torch .testing ._internal .common_utils import (
38
37
IS_MACOS ,
39
38
load_tests ,
40
39
skip_but_pass_in_sandcastle_if ,
41
40
get_cycles_per_ms ,
41
+ skipIfRocm ,
42
42
)
43
43
44
44
from torch .testing ._internal .dist_utils import (
@@ -5055,7 +5055,7 @@ def test_dynamic_rpc_existing_rank_can_communicate_with_new_rank(self):
5055
5055
5056
5056
# Dynamic RPC existing ranks can communicate with new ranks using CUDA rpc
5057
5057
@skip_if_lt_x_gpu (2 )
5058
- @skip_if_rocm
5058
+ @skipIfRocm
5059
5059
@dist_init (setup_rpc = False )
5060
5060
def test_dynamic_rpc_existing_rank_can_communicate_with_new_rank_cuda (self ):
5061
5061
initialize_pg (self .file_init_method , self .rank , self .world_size )
You can’t perform that action at this time.
0 commit comments