From 5f6ff4b937f9594b55ebd64cfb8a95f1b1789c96 Mon Sep 17 00:00:00 2001 From: lcskrishna Date: Mon, 17 Sep 2018 14:41:30 -0700 Subject: [PATCH 1/2] re-enable tests in test_cuda after fixes --- test/test_cuda.py | 27 +++++++++++---------------- 1 file changed, 11 insertions(+), 16 deletions(-) diff --git a/test/test_cuda.py b/test/test_cuda.py index 1d898548def1f2..b390b2c5cb6699 100644 --- a/test/test_cuda.py +++ b/test/test_cuda.py @@ -253,8 +253,7 @@ def tmp(t): ('add', small_3d, lambda t: [number(0.2, 2, t), small_3d_positive(t)], 'scalar_tensor'), ('sub', small_3d, lambda t: [number(3.14, 3, t)]), ('sub', small_3d, lambda t: [small_3d_positive(t)], 'tensor'), - ('mul', small_3d, lambda t: [number(3.14, 3, t)], '', types, False, - "skipIfRocm:CharTensor,ShortTensor"), + ('mul', small_3d, lambda t: [number(3.14, 3, t)]), ('mul', small_3d, lambda t: [small_3d_positive(t)], 'tensor'), ('div', small_3d, lambda t: [number(3.14, 3, t)]), ('div', small_3d, lambda t: [small_3d_positive(t)], 'tensor'), @@ -337,13 +336,13 @@ def tmp(t): ('max', small_3d_unique, lambda t: [1], 'dim', types, False, skipIfRocm), ('max', small_3d_unique, lambda t: [-1], 'neg_dim', types, False, skipIfRocm), ('max', medium_2d, lambda t: [medium_2d(t)], 'elementwise'), - ('min', small_3d_unique, lambda t: [], '', types, False, "skipIfRocm:HalfTensor"), + ('min', small_3d_unique, lambda t: []), ('min', small_3d_unique, lambda t: [1], 'dim', types, False, skipIfRocm), ('min', small_3d_unique, lambda t: [-1], 'neg_dim', types, False, skipIfRocm), ('min', medium_2d, lambda t: [medium_2d(t)], 'elementwise'), ('mean', small_3d, lambda t: []), - ('mean', small_3d, lambda t: [-1], 'neg_dim', types, False, "skipIfRocm:DoubleTensor,FloatTensor"), - ('mean', small_3d, lambda t: [1], 'dim', types, False, "skipIfRocm:DoubleTensor,FloatTensor"), + ('mean', small_3d, lambda t: [-1], 'neg_dim'), + ('mean', small_3d, lambda t: [1], 'dim'), ('mode', small_3d, lambda t: [], '', types, False, skipIfRocm), ('mode', small_3d, lambda t: [1], 'dim', types, False, skipIfRocm), ('mode', small_3d, lambda t: [-1], 'neg_dim', types, False, skipIfRocm), @@ -375,16 +374,14 @@ def tmp(t): ('put_', new_t(2, 3), lambda t: [long_type(t)([]), t([])], 'empty'), ('put_', new_t(2, 2), lambda t: [long_type(t)([[1], [-3]]), t([[1], [2]]), True], 'accumulate'), ('prod', small_2d_oneish, lambda t: []), - ('prod', small_3d, lambda t: [1], 'dim', types, False, skipIfRocm), - ('prod', small_3d, lambda t: [-1], 'neg_dim', types, False, skipIfRocm), + ('prod', small_3d, lambda t: [1], 'dim'), + ('prod', small_3d, lambda t: [-1], 'neg_dim'), ('sum', small_2d, lambda t: []), - ('sum', small_3d, lambda t: [1], 'dim', types, False, skipIfRocm), - ('sum', small_3d, lambda t: [-1], 'neg_dim', types, False, skipIfRocm), - ('renorm', small_3d, lambda t: [2, 1, 1], '2_norm', types, False, "skipIfRocm:HalfTensor,DoubleTensor,FloatTensor"), - ('renorm', small_3d, lambda t: [2, -1, 1], '2_norm_neg_dim', types, - False, "skipIfRocm:HalfTensor,DoubleTensor,FloatTensor"), - ('renorm', small_3d, lambda t: [1.5, 1, 1], '1_5_norm', types, - False, "skipIfRocm:HalfTensor,DoubleTensor,FloatTensor"), + ('sum', small_3d, lambda t: [1], 'dim'), + ('sum', small_3d, lambda t: [-1], 'neg_dim'), + ('renorm', small_3d, lambda t: [2, 1, 1], '2_norm'), + ('renorm', small_3d, lambda t: [2, -1, 1], '2_norm_neg_dim'), + ('renorm', small_3d, lambda t: [1.5, 1, 1], '1_5_norm'), ('repeat', small_2d, lambda t: [2, 2, 2],), ('size', new_t(1, 2, 3, 4), lambda t: [],), ('size', new_t(1, 2, 3, 4), lambda t: [1], 'dim'), @@ -1110,7 +1107,6 @@ def test_gather(self): def test_gather_dim(self): self._test_gather(1) - @skipIfRocm def test_from_sequence(self): seq = [list(range(i * 4, i * 4 + 4)) for i in range(5)] reference = torch.arange(0, 20).resize_(5, 4) @@ -1846,7 +1842,6 @@ def test_nvtx(self): torch.cuda.nvtx.mark("bar") torch.cuda.nvtx.range_pop() - @skipIfRocm def test_randperm_cuda(self): cuda = torch.device('cuda:0') From ff6a7093fe0ee6d0f2f51a05f0fe184b084cc6ee Mon Sep 17 00:00:00 2001 From: lcskrishna Date: Tue, 18 Sep 2018 08:51:11 -0700 Subject: [PATCH 2/2] skipped test_scatter_gpu that failed in CI --- test/test_nn.py | 1 + 1 file changed, 1 insertion(+) diff --git a/test/test_nn.py b/test/test_nn.py index b6fb725edd211e..637adc4f1ca6cf 100644 --- a/test/test_nn.py +++ b/test/test_nn.py @@ -2695,6 +2695,7 @@ def test_scatter_cpu(self): self._test_scatter(torch.randn(4, 4)) @unittest.skipIf(not TEST_MULTIGPU, "multi-GPU not supported") + @skipIfRocm def test_scatter_gpu(self): self._test_scatter(torch.randn(4, 4).cuda())