Skip to content

Add tests for bfloat16 math functions on CUDA #57984

New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

Closed
wants to merge 17 commits into from

Conversation

zasdfgbnm
Copy link
Collaborator

@zasdfgbnm zasdfgbnm commented May 10, 2021

I copy paste cuda_half_test.cu to cuda_bfloat16_test.cu and change it to test bfloat16. It do find a few issues:

  • CUDA_VERSION sometimes doesn't work on c10 (I don't know the reason), I changed it to use __CUDACC_VER_MAJOR__ instead
  • The operator __nv_bfloat16() of c10::BFloat16 should not be explicit
  • PyTorch should be built with -D__CUDA_NO_BFLOAT16_OPERATORS__ to be consistent with half behavior
  • There is a assert(::abs(::atanh(Half(1.0)) - ::atanh(1.0f)) <= threshold); in the test, this doesn't make sense, because atanh(1) is inf.

@facebook-github-bot
Copy link
Contributor

facebook-github-bot commented May 10, 2021

🔗 Helpful links

❌ 5 New Failures

As of commit 0b0a0c7 (more details on the Dr. CI page):

Expand to see more
  • 5/5 failures introduced in this PR

🕵️ 5 new failures recognized by patterns

The following CI failures do not appear to be due to upstream breakages

See GitHub Actions build pull / linux-bionic-cuda11.3-py3.7-clang9 / build (1/5)

Step: "Build" (full log | diagnosis details | 🔁 rerun)

2022-05-21T21:28:56.5458551Z �[0m�[1m�[31mERROR...space/aten/src/ATen/test/cuda_half_test.cu\".\n" }
2022-05-21T21:28:46.5126996Z [ 94%] Built target cuda_complex_math_test
2022-05-21T21:28:55.9902251Z [ 94%] �[32m�[1mLinking CXX executable ../bin/cuda_bfloat16_test�[0m
2022-05-21T21:28:56.4827286Z [ 94%] Built target cuda_bfloat16_test
2022-05-21T21:28:56.4832963Z Makefile:145: recipe for target 'all' failed
2022-05-21T21:28:56.4833248Z make: *** [all] Error 2
2022-05-21T21:28:56.5222950Z + cleanup
2022-05-21T21:28:56.5223214Z + retcode=1
2022-05-21T21:28:56.5223398Z + set +x
2022-05-21T21:28:56.5224039Z ##[group]Sccache Compilation Log
2022-05-21T21:28:56.5224302Z =================== sccache compilation log ===================
2022-05-21T21:28:56.5458551Z �[0m�[1m�[31mERROR�[0m 2022-05-21T21:28:44Z: sccache::server: Compilation failed: Output { status: ExitStatus(unix_wait_status(256)), stdout: "", stderr: "/var/lib/jenkins/workspace/aten/src/ATen/test/cuda_half_test.cu(32): error: function \"kernel\" has already been defined\n\n/var/lib/jenkins/workspace/aten/src/ATen/test/cuda_half_test.cu(33): error: no instance of function template \"test\" matches the argument list\n\n2 errors detected in the compilation of \"/var/lib/jenkins/workspace/aten/src/ATen/test/cuda_half_test.cu\".\n" }
2022-05-21T21:28:56.5459157Z 
2022-05-21T21:28:56.5475338Z =========== If your build fails, please take a look at the log above for possible reasons ===========
2022-05-21T21:28:56.5499252Z Compile requests                    9325
2022-05-21T21:28:56.5499645Z Compile requests executed           6810
2022-05-21T21:28:56.5500009Z Cache hits                          6185
2022-05-21T21:28:56.5500322Z Cache hits (C/C++)                  6184
2022-05-21T21:28:56.5500608Z Cache hits (CUDA)                      1
2022-05-21T21:28:56.5500796Z Cache misses                         606
2022-05-21T21:28:56.5501004Z Cache misses (C/C++)                 446
2022-05-21T21:28:56.5501208Z Cache misses (CUDA)                  160

See GitHub Actions build pull / win-vs2019-cuda11.3-py3 / build (2/5)

Step: "Build" (full log | diagnosis details | 🔁 rerun)

2022-05-21T21:31:04.0718212Z FAILED: caffe2/CMa...st.dir/__/aten/src/ATen/test/cuda_half_test.cu.obj
2022-05-21T21:31:02.5138280Z 
2022-05-21T21:31:02.7652559Z [5927/6309] C:\actions-runner\_work\pytorch\pytorch\build\win_tmp\bin\sccache-cl.exe   /TP -DIDEEP_USE_MKL -DMINIZ_DISABLE_ZIP_READER_CRC32_CHECKS -DONNXIFI_ENABLE_EXT=1 -DONNX_ML=1 -DONNX_NAMESPACE=onnx_torch -DUSE_C10D_GLOO -DUSE_CUDA -DUSE_DISTRIBUTED -DUSE_EXTERNAL_MZCRC -DUSE_GTEST -DWIN32_LEAN_AND_MEAN -D_CRT_SECURE_NO_DEPRECATE=1 -D_OPENMP_NOFORCE_MANIFEST -IC:\actions-runner\_work\pytorch\pytorch\build\aten\src -IC:\actions-runner\_work\pytorch\pytorch\aten\src -IC:\actions-runner\_work\pytorch\pytorch\build -IC:\actions-runner\_work\pytorch\pytorch -IC:\actions-runner\_work\pytorch\pytorch\cmake\..\third_party\benchmark\include -IC:\actions-runner\_work\pytorch\pytorch\cmake\..\third_party\cudnn_frontend\include -IC:\actions-runner\_work\pytorch\pytorch\third_party\onnx -IC:\actions-runner\_work\pytorch\pytorch\build\third_party\onnx -IC:\actions-runner\_work\pytorch\pytorch\third_party\foxi -IC:\actions-runner\_work\pytorch\pytorch\build\third_party\foxi -IC:\actions-runner\_work\pytorch\pytorch\build\caffe2\..\aten\src -IC:\actions-runner\_work\pytorch\pytorch\torch\csrc\api -IC:\actions-runner\_work\pytorch\pytorch\torch\csrc\api\include -IC:\actions-runner\_work\pytorch\pytorch\c10\.. -IC:\actions-runner\_work\pytorch\pytorch\c10\cuda\..\.. -IC:\actions-runner\_work\pytorch\pytorch\third_party\pthreadpool\include -IC:\actions-runner\_work\pytorch\pytorch\build\third_party\gloo -IC:\actions-runner\_work\pytorch\pytorch\cmake\..\third_party\gloo -IC:\actions-runner\_work\pytorch\pytorch\cmake\..\third_party\googletest\googlemock\include -IC:\actions-runner\_work\pytorch\pytorch\cmake\..\third_party\googletest\googletest\include -IC:\actions-runner\_work\pytorch\pytorch\third_party\protobuf\src -IC:\actions-runner\_work\pytorch\pytorch\build\win_tmp\mkl\include -IC:\actions-runner\_work\pytorch\pytorch\third_party\XNNPACK\include -IC:\actions-runner\_work\pytorch\pytorch\cmake\..\third_party\eigen -I"C:\Program Files\NVIDIA GPU Computing Toolkit\CUDA\v11.3\include" -IC:\actions-runner\_work\pytorch\pytorch\third_party\ideep\mkl-dnn\third_party\oneDNN\include -IC:\actions-runner\_work\pytorch\pytorch\third_party\ideep\include -I"C:\Program Files\NVIDIA Corporation\NvToolsExt\include" -IC:\actions-runner\_work\pytorch\pytorch\third_party\googletest\googletest\include -IC:\actions-runner\_work\pytorch\pytorch\third_party\googletest\googletest /DWIN32 /D_WINDOWS /GR /EHsc /w /bigobj -DUSE_PTHREADPOOL -openmp:experimental -IC:/actions-runner/_work/pytorch/pytorch/build/win_tmp/mkl/include -DNDEBUG -DUSE_KINETO -DLIBKINETO_NOCUPTI -DUSE_FBGEMM -DUSE_XNNPACK -DSYMBOLICATE_MOBILE_DEBUG_HANDLE -DEDGE_PROFILER_USE_KINETO -DHAVE_AVX512_CPU_DEFINITION -DHAVE_AVX2_CPU_DEFINITION /MD /O2 /Ob2 /DNDEBUG /w /bigobj -DNDEBUG -DCAFFE2_USE_GLOO -DTH_HAVE_THREAD /EHsc /DNOMINMAX /wd4267 /wd4251 /wd4522 /wd4838 /wd4305 /wd4244 /wd4190 /wd4101 /wd4996 /wd4275 /bigobj -std:c++14 /showIncludes /Fotest_tensorexpr\CMakeFiles\test_tensorexpr.dir\test_type_specializations.cpp.obj /Fdtest_tensorexpr\CMakeFiles\test_tensorexpr.dir\ /FS -c C:\actions-runner\_work\pytorch\pytorch\test\cpp\tensorexpr\test_type_specializations.cpp
2022-05-21T21:31:02.7657972Z Microsoft (R) C/C++ Optimizing Compiler Version 19.28.29337 for x64
2022-05-21T21:31:02.7658283Z Copyright (C) Microsoft Corporation.  All rights reserved.
2022-05-21T21:31:02.7658470Z 
2022-05-21T21:31:03.4299654Z [5928/6309] C:\actions-runner\_work\pytorch\pytorch\build\win_tmp\bin\sccache-cl.exe   /TP -DIDEEP_USE_MKL -DMINIZ_DISABLE_ZIP_READER_CRC32_CHECKS -DONNXIFI_ENABLE_EXT=1 -DONNX_ML=1 -DONNX_NAMESPACE=onnx_torch -DUSE_C10D_GLOO -DUSE_CUDA -DUSE_DISTRIBUTED -DUSE_EXTERNAL_MZCRC -DUSE_GTEST -DWIN32_LEAN_AND_MEAN -D_CRT_SECURE_NO_DEPRECATE=1 -D_OPENMP_NOFORCE_MANIFEST -IC:\actions-runner\_work\pytorch\pytorch\build\aten\src -IC:\actions-runner\_work\pytorch\pytorch\aten\src -IC:\actions-runner\_work\pytorch\pytorch\build -IC:\actions-runner\_work\pytorch\pytorch -IC:\actions-runner\_work\pytorch\pytorch\cmake\..\third_party\benchmark\include -IC:\actions-runner\_work\pytorch\pytorch\cmake\..\third_party\cudnn_frontend\include -IC:\actions-runner\_work\pytorch\pytorch\third_party\onnx -IC:\actions-runner\_work\pytorch\pytorch\build\third_party\onnx -IC:\actions-runner\_work\pytorch\pytorch\third_party\foxi -IC:\actions-runner\_work\pytorch\pytorch\build\third_party\foxi -IC:\actions-runner\_work\pytorch\pytorch\build\caffe2\..\aten\src -IC:\actions-runner\_work\pytorch\pytorch\torch\csrc\api -IC:\actions-runner\_work\pytorch\pytorch\torch\csrc\api\include -IC:\actions-runner\_work\pytorch\pytorch\c10\.. -IC:\actions-runner\_work\pytorch\pytorch\c10\cuda\..\.. -IC:\actions-runner\_work\pytorch\pytorch\third_party\pthreadpool\include -IC:\actions-runner\_work\pytorch\pytorch\build\third_party\gloo -IC:\actions-runner\_work\pytorch\pytorch\cmake\..\third_party\gloo -IC:\actions-runner\_work\pytorch\pytorch\cmake\..\third_party\googletest\googlemock\include -IC:\actions-runner\_work\pytorch\pytorch\cmake\..\third_party\googletest\googletest\include -IC:\actions-runner\_work\pytorch\pytorch\third_party\protobuf\src -IC:\actions-runner\_work\pytorch\pytorch\build\win_tmp\mkl\include -IC:\actions-runner\_work\pytorch\pytorch\third_party\XNNPACK\include -IC:\actions-runner\_work\pytorch\pytorch\cmake\..\third_party\eigen -I"C:\Program Files\NVIDIA GPU Computing Toolkit\CUDA\v11.3\include" -IC:\actions-runner\_work\pytorch\pytorch\third_party\ideep\mkl-dnn\third_party\oneDNN\include -IC:\actions-runner\_work\pytorch\pytorch\third_party\ideep\include -I"C:\Program Files\NVIDIA Corporation\NvToolsExt\include" -IC:\actions-runner\_work\pytorch\pytorch\third_party\googletest\googletest\include -IC:\actions-runner\_work\pytorch\pytorch\third_party\googletest\googletest /DWIN32 /D_WINDOWS /GR /EHsc /w /bigobj -DUSE_PTHREADPOOL -openmp:experimental -IC:/actions-runner/_work/pytorch/pytorch/build/win_tmp/mkl/include -DNDEBUG -DUSE_KINETO -DLIBKINETO_NOCUPTI -DUSE_FBGEMM -DUSE_XNNPACK -DSYMBOLICATE_MOBILE_DEBUG_HANDLE -DEDGE_PROFILER_USE_KINETO -DHAVE_AVX512_CPU_DEFINITION -DHAVE_AVX2_CPU_DEFINITION /MD /O2 /Ob2 /DNDEBUG /w /bigobj -DNDEBUG -DCAFFE2_USE_GLOO -DTH_HAVE_THREAD /EHsc /DNOMINMAX /wd4267 /wd4251 /wd4522 /wd4838 /wd4305 /wd4244 /wd4190 /wd4101 /wd4996 /wd4275 /bigobj -std:c++14 /showIncludes /Fotest_tensorexpr\CMakeFiles\test_tensorexpr.dir\test_type.cpp.obj /Fdtest_tensorexpr\CMakeFiles\test_tensorexpr.dir\ /FS -c C:\actions-runner\_work\pytorch\pytorch\test\cpp\tensorexpr\test_type.cpp
2022-05-21T21:31:03.4304140Z Microsoft (R) C/C++ Optimizing Compiler Version 19.28.29337 for x64
2022-05-21T21:31:03.4304457Z Copyright (C) Microsoft Corporation.  All rights reserved.
2022-05-21T21:31:03.4304625Z 
2022-05-21T21:31:04.0708001Z [5929/6309] C:\actions-runner\_work\pytorch\pytorch\build\win_tmp\bin\randomtemp.exe C:/actions-runner/_work/pytorch/pytorch/build/win_tmp\bin\sccache.exe C:\PROGRA~1\NVIDIA~2\CUDA\v11.3\bin\nvcc.exe -forward-unknown-to-host-compiler -DIDEEP_USE_MKL -DMINIZ_DISABLE_ZIP_READER_CRC32_CHECKS -DONNXIFI_ENABLE_EXT=1 -DONNX_ML=1 -DONNX_NAMESPACE=onnx_torch -DUSE_C10D_GLOO -DUSE_DISTRIBUTED -DUSE_EXTERNAL_MZCRC -DWIN32_LEAN_AND_MEAN -D_CRT_SECURE_NO_DEPRECATE=1 -D_OPENMP_NOFORCE_MANIFEST -IC:\actions-runner\_work\pytorch\pytorch\build\aten\src -IC:\actions-runner\_work\pytorch\pytorch\aten\src -IC:\actions-runner\_work\pytorch\pytorch\build -IC:\actions-runner\_work\pytorch\pytorch -IC:\actions-runner\_work\pytorch\pytorch\cmake\..\third_party\benchmark\include -IC:\actions-runner\_work\pytorch\pytorch\cmake\..\third_party\cudnn_frontend\include -IC:\actions-runner\_work\pytorch\pytorch\third_party\onnx -IC:\actions-runner\_work\pytorch\pytorch\build\third_party\onnx -IC:\actions-runner\_work\pytorch\pytorch\third_party\foxi -IC:\actions-runner\_work\pytorch\pytorch\build\third_party\foxi -IC:\actions-runner\_work\pytorch\pytorch\build\caffe2\aten\src -IC:\actions-runner\_work\pytorch\pytorch\aten\..\third_party\catch\single_include -IC:\actions-runner\_work\pytorch\pytorch\aten\src\ATen\.. -IC:\actions-runner\_work\pytorch\pytorch\third_party\miniz-2.0.8 -IC:\actions-runner\_work\pytorch\pytorch\torch\csrc\api -IC:\actions-runner\_work\pytorch\pytorch\torch\csrc\api\include -IC:\actions-runner\_work\pytorch\pytorch\c10\.. -IC:\actions-runner\_work\pytorch\pytorch\c10\cuda\..\.. -isystem=C:\actions-runner\_work\pytorch\pytorch\build\third_party\gloo -isystem=C:\actions-runner\_work\pytorch\pytorch\cmake\..\third_party\gloo -isystem=C:\actions-runner\_work\pytorch\pytorch\cmake\..\third_party\googletest\googlemock\include -isystem=C:\actions-runner\_work\pytorch\pytorch\cmake\..\third_party\googletest\googletest\include -isystem=C:\actions-runner\_work\pytorch\pytorch\third_party\protobuf\src -isystem=C:\actions-runner\_work\pytorch\pytorch\build\win_tmp\mkl\include -isystem=C:\actions-runner\_work\pytorch\pytorch\third_party\XNNPACK\include -isystem=C:\actions-runner\_work\pytorch\pytorch\cmake\..\third_party\eigen -isystem="C:\Program Files\NVIDIA GPU Computing Toolkit\CUDA\v11.3\include" -isystem=C:\actions-runner\_work\pytorch\pytorch\third_party\ideep\mkl-dnn\third_party\oneDNN\include -isystem=C:\actions-runner\_work\pytorch\pytorch\third_party\ideep\include -isystem="C:\Program Files\NVIDIA Corporation\NvToolsExt\include" -isystem=C:\actions-runner\_work\pytorch\pytorch\third_party\googletest\googletest\include -isystem=C:\actions-runner\_work\pytorch\pytorch\third_party\googletest\googletest -Xcompiler /w -w -Xfatbin -compress-all -DONNX_NAMESPACE=onnx_torch --use-local-env -gencode arch=compute_70,code=sm_70 -Xcudafe --diag_suppress=cc_clobber_ignored,--diag_suppress=integer_sign_change,--diag_suppress=useless_using_declaration,--diag_suppress=set_but_not_used,--diag_suppress=field_without_dll_interface,--diag_suppress=base_class_has_different_dll_interface,--diag_suppress=dll_interface_conflict_none_assumed,--diag_suppress=dll_interface_conflict_dllexport_assumed,--diag_suppress=implicit_return_from_non_void_function,--diag_suppress=unsigned_compare_with_zero,--diag_suppress=declared_but_not_referenced,--diag_suppress=bad_friend_decl --Werror cross-execution-space-call --no-host-device-move-forward --expt-relaxed-constexpr --expt-extended-lambda  -Xcompiler=/wd4819,/wd4503,/wd4190,/wd4244,/wd4251,/wd4275,/wd4522 -Wno-deprecated-gpu-targets --expt-extended-lambda -DCUDA_HAS_FP16=1 -D__CUDA_NO_HALF_OPERATORS__ -D__CUDA_NO_HALF_CONVERSIONS__ -D__CUDA_NO_HALF2_OPERATORS__ -D__CUDA_NO_BFLOAT16_OPERATORS__ -D__CUDA_NO_BFLOAT16_CONVERSIONS__ -Xcompiler="-MD -O2 -Ob2" -DNDEBUG -Xcompiler /MD -DCAFFE2_USE_GLOO -DTH_HAVE_THREAD -std=c++14 -MD -MT caffe2\CMakeFiles\cuda_half_test.dir\__\aten\src\ATen\test\cuda_half_test.cu.obj -MF caffe2\CMakeFiles\cuda_half_test.dir\__\aten\src\ATen\test\cuda_half_test.cu.obj.d -x cu -c C:\actions-runner\_work\pytorch\pytorch\aten\src\ATen\test\cuda_half_test.cu -o caffe2\CMakeFiles\cuda_half_test.dir\__\aten\src\ATen\test\cuda_half_test.cu.obj -Xcompiler=-Fdcaffe2\CMakeFiles\cuda_half_test.dir\,-FS
2022-05-21T21:31:04.0718212Z FAILED: caffe2/CMakeFiles/cuda_half_test.dir/__/aten/src/ATen/test/cuda_half_test.cu.obj 
2022-05-21T21:31:04.0728818Z C:\actions-runner\_work\pytorch\pytorch\build\win_tmp\bin\randomtemp.exe C:/actions-runner/_work/pytorch/pytorch/build/win_tmp\bin\sccache.exe C:\PROGRA~1\NVIDIA~2\CUDA\v11.3\bin\nvcc.exe -forward-unknown-to-host-compiler -DIDEEP_USE_MKL -DMINIZ_DISABLE_ZIP_READER_CRC32_CHECKS -DONNXIFI_ENABLE_EXT=1 -DONNX_ML=1 -DONNX_NAMESPACE=onnx_torch -DUSE_C10D_GLOO -DUSE_DISTRIBUTED -DUSE_EXTERNAL_MZCRC -DWIN32_LEAN_AND_MEAN -D_CRT_SECURE_NO_DEPRECATE=1 -D_OPENMP_NOFORCE_MANIFEST -IC:\actions-runner\_work\pytorch\pytorch\build\aten\src -IC:\actions-runner\_work\pytorch\pytorch\aten\src -IC:\actions-runner\_work\pytorch\pytorch\build -IC:\actions-runner\_work\pytorch\pytorch -IC:\actions-runner\_work\pytorch\pytorch\cmake\..\third_party\benchmark\include -IC:\actions-runner\_work\pytorch\pytorch\cmake\..\third_party\cudnn_frontend\include -IC:\actions-runner\_work\pytorch\pytorch\third_party\onnx -IC:\actions-runner\_work\pytorch\pytorch\build\third_party\onnx -IC:\actions-runner\_work\pytorch\pytorch\third_party\foxi -IC:\actions-runner\_work\pytorch\pytorch\build\third_party\foxi -IC:\actions-runner\_work\pytorch\pytorch\build\caffe2\aten\src -IC:\actions-runner\_work\pytorch\pytorch\aten\..\third_party\catch\single_include -IC:\actions-runner\_work\pytorch\pytorch\aten\src\ATen\.. -IC:\actions-runner\_work\pytorch\pytorch\third_party\miniz-2.0.8 -IC:\actions-runner\_work\pytorch\pytorch\torch\csrc\api -IC:\actions-runner\_work\pytorch\pytorch\torch\csrc\api\include -IC:\actions-runner\_work\pytorch\pytorch\c10\.. -IC:\actions-runner\_work\pytorch\pytorch\c10\cuda\..\.. -isystem=C:\actions-runner\_work\pytorch\pytorch\build\third_party\gloo -isystem=C:\actions-runner\_work\pytorch\pytorch\cmake\..\third_party\gloo -isystem=C:\actions-runner\_work\pytorch\pytorch\cmake\..\third_party\googletest\googlemock\include -isystem=C:\actions-runner\_work\pytorch\pytorch\cmake\..\third_party\googletest\googletest\include -isystem=C:\actions-runner\_work\pytorch\pytorch\third_party\protobuf\src -isystem=C:\actions-runner\_work\pytorch\pytorch\build\win_tmp\mkl\include -isystem=C:\actions-runner\_work\pytorch\pytorch\third_party\XNNPACK\include -isystem=C:\actions-runner\_work\pytorch\pytorch\cmake\..\third_party\eigen -isystem="C:\Program Files\NVIDIA GPU Computing Toolkit\CUDA\v11.3\include" -isystem=C:\actions-runner\_work\pytorch\pytorch\third_party\ideep\mkl-dnn\third_party\oneDNN\include -isystem=C:\actions-runner\_work\pytorch\pytorch\third_party\ideep\include -isystem="C:\Program Files\NVIDIA Corporation\NvToolsExt\include" -isystem=C:\actions-runner\_work\pytorch\pytorch\third_party\googletest\googletest\include -isystem=C:\actions-runner\_work\pytorch\pytorch\third_party\googletest\googletest -Xcompiler /w -w -Xfatbin -compress-all -DONNX_NAMESPACE=onnx_torch --use-local-env -gencode arch=compute_70,code=sm_70 -Xcudafe --diag_suppress=cc_clobber_ignored,--diag_suppress=integer_sign_change,--diag_suppress=useless_using_declaration,--diag_suppress=set_but_not_used,--diag_suppress=field_without_dll_interface,--diag_suppress=base_class_has_different_dll_interface,--diag_suppress=dll_interface_conflict_none_assumed,--diag_suppress=dll_interface_conflict_dllexport_assumed,--diag_suppress=implicit_return_from_non_void_function,--diag_suppress=unsigned_compare_with_zero,--diag_suppress=declared_but_not_referenced,--diag_suppress=bad_friend_decl --Werror cross-execution-space-call --no-host-device-move-forward --expt-relaxed-constexpr --expt-extended-lambda  -Xcompiler=/wd4819,/wd4503,/wd4190,/wd4244,/wd4251,/wd4275,/wd4522 -Wno-deprecated-gpu-targets --expt-extended-lambda -DCUDA_HAS_FP16=1 -D__CUDA_NO_HALF_OPERATORS__ -D__CUDA_NO_HALF_CONVERSIONS__ -D__CUDA_NO_HALF2_OPERATORS__ -D__CUDA_NO_BFLOAT16_OPERATORS__ -D__CUDA_NO_BFLOAT16_CONVERSIONS__ -Xcompiler="-MD -O2 -Ob2" -DNDEBUG -Xcompiler /MD -DCAFFE2_USE_GLOO -DTH_HAVE_THREAD -std=c++14 -MD -MT caffe2\CMakeFiles\cuda_half_test.dir\__\aten\src\ATen\test\cuda_half_test.cu.obj -MF caffe2\CMakeFiles\cuda_half_test.dir\__\aten\src\ATen\test\cuda_half_test.cu.obj.d -x cu -c C:\actions-runner\_work\pytorch\pytorch\aten\src\ATen\test\cuda_half_test.cu -o caffe2\CMakeFiles\cuda_half_test.dir\__\aten\src\ATen\test\cuda_half_test.cu.obj -Xcompiler=-Fdcaffe2\CMakeFiles\cuda_half_test.dir\,-FS
2022-05-21T21:31:04.0738352Z C:\actions-runner\_work\pytorch\pytorch\aten\src\ATen\test\cuda_half_test.cu(32): error: function "kernel" has already been defined
2022-05-21T21:31:04.0738613Z 
2022-05-21T21:31:04.0738887Z C:\actions-runner\_work\pytorch\pytorch\aten\src\ATen\test\cuda_half_test.cu(33): error: no instance of function template "test" matches the argument list
2022-05-21T21:31:04.0739172Z 
2022-05-21T21:31:04.0739391Z 2 errors detected in the compilation of "C:/actions-runner/_work/pytorch/pytorch/aten/src/ATen/test/cuda_half_test.cu".
2022-05-21T21:31:04.0739693Z cuda_half_test.cu
2022-05-21T21:31:04.0739858Z Retry attempt: 1
2022-05-21T21:31:04.0740203Z C:\actions-runner\_work\pytorch\pytorch\aten\src\ATen\test\cuda_half_test.cu(32): error: function "kernel" has already been defined
2022-05-21T21:31:04.0740457Z 

See GitHub Actions build pull / deploy-linux-xenial-cuda11.3-py3.7-gcc7 / build (3/5)

Step: "Build" (full log | diagnosis details | 🔁 rerun)

2022-05-21T21:08:36.6695570Z ##[error]Process completed with exit code 1.
2022-05-21T21:08:36.6640984Z 
2022-05-21T21:08:36.6641083Z Non-cacheable reasons:
2022-05-21T21:08:36.6641311Z -E                                  936
2022-05-21T21:08:36.6641533Z -M                                   33
2022-05-21T21:08:36.6641717Z multiple input files                 15
2022-05-21T21:08:36.6641904Z @                                     1
2022-05-21T21:08:36.6642137Z unknown source language               1
2022-05-21T21:08:36.6642280Z 
2022-05-21T21:08:36.6648694Z Cache location                  S3, bucket: Bucket(name=ossci-compiler-cache-circleci-v2, base_url=http://ossci-compiler-cache-circleci-v2.s3.amazonaws.com/)
2022-05-21T21:08:36.6649688Z ##[endgroup]
2022-05-21T21:08:36.6695570Z ##[error]Process completed with exit code 1.
2022-05-21T21:08:36.6731727Z Prepare all required actions
2022-05-21T21:08:36.6751638Z ##[group]Run ./.github/actions/teardown-linux
2022-05-21T21:08:36.6751840Z with:
2022-05-21T21:08:36.6751976Z env:
2022-05-21T21:08:36.6752128Z   IN_CI: 1
2022-05-21T21:08:36.6752284Z   IS_GHA: 1
2022-05-21T21:08:36.6752431Z ##[endgroup]
2022-05-21T21:08:36.6779925Z ##[group]Run .github/scripts/wait_for_ssh_to_drain.sh
2022-05-21T21:08:36.6780187Z �[36;1m.github/scripts/wait_for_ssh_to_drain.sh�[0m
2022-05-21T21:08:36.6791854Z shell: /usr/bin/bash --noprofile --norc -e -o pipefail {0}

See GitHub Actions build pull / linux-xenial-cuda11.3-py3.7-gcc7 / build (4/5)

Step: "Build" (full log | diagnosis details | 🔁 rerun)

2022-05-21T21:32:07.8323337Z �[0m�[1m�[31mERROR...eof ((socklen_t)))\n ^\n" }
2022-05-21T21:32:07.8317424Z �[0m�[1m�[31mERROR�[0m 2022-05-21T21:06:14Z: sccache::server: Compilation failed: Output { status: ExitStatus(ExitStatus(256)), stdout: "", stderr: "conftest.c: In function \'main\':\nconftest.c:332:2: error: \'struct sockaddr\' has no member named \'sa_len\'\n x.sa_len = 0;\n  ^\n" }
2022-05-21T21:32:07.8317770Z 
2022-05-21T21:32:07.8318584Z �[0m�[1m�[31mERROR�[0m 2022-05-21T21:06:16Z: sccache::server: Compilation failed: Output { status: ExitStatus(ExitStatus(256)), stdout: "", stderr: "conftest.c: In function \'main\':\nconftest.c:366:10: error: \'RTLD_MEMBER\' undeclared (first use in this function); did you mean \'RTLD_NEXT\'?\n   (void) RTLD_MEMBER;\n          ^~~~~~~~~~~\n          RTLD_NEXT\nconftest.c:366:10: note: each undeclared identifier is reported only once for each function it appears in\n" }
2022-05-21T21:32:07.8319088Z 
2022-05-21T21:32:07.8319941Z �[0m�[1m�[31mERROR�[0m 2022-05-21T21:06:17Z: sccache::server: Compilation failed: Output { status: ExitStatus(ExitStatus(256)), stdout: "", stderr: "conftest.c:361:9: error: unknown type name \'not\'\n         not a universal capable compiler\n         ^~~\nconftest.c:361:15: error: expected \'=\', \',\', \';\', \'asm\' or \'__attribute__\' before \'universal\'\n         not a universal capable compiler\n               ^~~~~~~~~\nconftest.c:361:15: error: unknown type name \'universal\'\n" }
2022-05-21T21:32:07.8320469Z 
2022-05-21T21:32:07.8321246Z �[0m�[1m�[31mERROR�[0m 2022-05-21T21:06:17Z: sccache::server: Compilation failed: Output { status: ExitStatus(ExitStatus(256)), stdout: "", stderr: "conftest.c: In function \'main\':\nconftest.c:367:4: error: unknown type name \'not\'; did you mean \'ino_t\'?\n    not big endian\n    ^~~\n    ino_t\nconftest.c:367:12: error: expected \'=\', \',\', \';\', \'asm\' or \'__attribute__\' before \'endian\'\n    not big endian\n            ^~~~~~\n" }
2022-05-21T21:32:07.8321734Z 
2022-05-21T21:32:07.8322384Z �[0m�[1m�[31mERROR�[0m 2022-05-21T21:06:18Z: sccache::server: Compilation failed: Output { status: ExitStatus(ExitStatus(256)), stdout: "", stderr: "conftest.c: In function \'main\':\nconftest.c:378:4: error: \'struct stat\' has no member named \'st_mtimespec\'; did you mean \'st_mtim\'?\n st.st_mtimespec.tv_nsec = 1;\n    ^~~~~~~~~~~~\n    st_mtim\n" }
2022-05-21T21:32:07.8322773Z 
2022-05-21T21:32:07.8323337Z �[0m�[1m�[31mERROR�[0m 2022-05-21T21:06:19Z: sccache::server: Compilation failed: Output { status: ExitStatus(ExitStatus(256)), stdout: "", stderr: "conftest.c: In function \'main\':\nconftest.c:402:24: error: expected expression before \')\' token\n if (sizeof ((socklen_t)))\n                        ^\n" }
2022-05-21T21:32:07.8323729Z 
2022-05-21T21:32:07.8324661Z �[0m�[1m�[31mERROR�[0m 2022-05-21T21:31:53Z: sccache::server: Compilation failed: Output { status: ExitStatus(ExitStatus(256)), stdout: "", stderr: "/var/lib/jenkins/workspace/aten/src/ATen/test/cuda_half_test.cu(32): error: function \"kernel\" has already been defined\n\n/var/lib/jenkins/workspace/aten/src/ATen/test/cuda_half_test.cu(33): error: no instance of function template \"test\" matches the argument list\n\n2 errors detected in the compilation of \"/var/lib/jenkins/workspace/aten/src/ATen/test/cuda_half_test.cu\".\n" }
2022-05-21T21:32:07.8325220Z 
2022-05-21T21:32:07.8325398Z =========== If your build fails, please take a look at the log above for possible reasons ===========
2022-05-21T21:32:07.8336329Z Compile requests                   11123
2022-05-21T21:32:07.8336638Z Compile requests executed           8329
2022-05-21T21:32:07.8336873Z Cache hits                          7651
2022-05-21T21:32:07.8337077Z Cache hits (C/C++)                  7651
2022-05-21T21:32:07.8337282Z Cache misses                         604
2022-05-21T21:32:07.8337493Z Cache misses (C/C++)                 438

See GitHub Actions build pull / pytorch-xla-linux-bionic-py3.7-clang8 / test (xla, 1, 1, linux.2xlarge) (5/5)

Step: "Test" (full log | diagnosis details | 🔁 rerun)

2022-05-21T23:15:55.2769719Z RuntimeError: tens...OK() (UNKNOWN: Could not start gRPC server vs. OK)
2022-05-21T23:15:55.2764418Z   File "/opt/conda/lib/python3.7/site-packages/torch_xla-1.12-py3.7-linux-x86_64.egg/torch_xla/distributed/xla_multiprocessing.py", line 315, in _setup_replication
2022-05-21T23:15:55.2764922Z     device = xm.xla_device()
2022-05-21T23:15:55.2765689Z   File "/opt/conda/lib/python3.7/site-packages/torch_xla-1.12-py3.7-linux-x86_64.egg/torch_xla/core/xla_model.py", line 232, in xla_device
2022-05-21T23:15:55.2766112Z     devkind=devkind if devkind is not None else None)
2022-05-21T23:15:55.2766781Z   File "/opt/conda/lib/python3.7/site-packages/torch_xla-1.12-py3.7-linux-x86_64.egg/torch_xla/core/xla_model.py", line 137, in get_xla_supported_devices
2022-05-21T23:15:55.2767154Z     xla_devices = _DEVICES.value
2022-05-21T23:15:55.2767701Z   File "/opt/conda/lib/python3.7/site-packages/torch_xla-1.12-py3.7-linux-x86_64.egg/torch_xla/utils/utils.py", line 32, in value
2022-05-21T23:15:55.2768105Z     self._value = self._gen_fn()
2022-05-21T23:15:55.2768644Z   File "/opt/conda/lib/python3.7/site-packages/torch_xla-1.12-py3.7-linux-x86_64.egg/torch_xla/core/xla_model.py", line 19, in <lambda>
2022-05-21T23:15:55.2769119Z     _DEVICES = xu.LazyProperty(lambda: torch_xla._XLAC._xla_get_devices())
2022-05-21T23:15:55.2769719Z RuntimeError: tensorflow/compiler/xla/xla_client/xrt_local_service.cc:56 : Check failed: tensorflow::NewServer(server_def, &server_) == ::tensorflow::Status::OK() (UNKNOWN: Could not start gRPC server vs. OK)
2022-05-21T23:15:55.5290560Z Traceback (most recent call last):
2022-05-21T23:15:55.5291142Z   File "/var/lib/jenkins/workspace/xla/test/test_mp_all_to_all.py", line 34, in <module>
2022-05-21T23:15:55.5291646Z     xmp.spawn(_mp_fn, args=())
2022-05-21T23:15:55.5292225Z   File "/opt/conda/lib/python3.7/site-packages/torch_xla-1.12-py3.7-linux-x86_64.egg/torch_xla/distributed/xla_multiprocessing.py", line 395, in spawn
2022-05-21T23:15:55.5292575Z     start_method=start_method)
2022-05-21T23:15:55.5292948Z   File "/opt/conda/lib/python3.7/site-packages/torch/multiprocessing/spawn.py", line 198, in start_processes
2022-05-21T23:15:55.5293234Z     while not context.join():
2022-05-21T23:15:55.5293591Z   File "/opt/conda/lib/python3.7/site-packages/torch/multiprocessing/spawn.py", line 154, in join
2022-05-21T23:15:55.5293853Z     exit_code=exitcode
2022-05-21T23:15:55.5294164Z torch.multiprocessing.spawn.ProcessExitedException: process 3 terminated with exit code 17

This comment was automatically generated by Dr. CI (expand for details).

Please report bugs/suggestions to the (internal) Dr. CI Users group.

Click here to manually regenerate this comment.

@zasdfgbnm zasdfgbnm marked this pull request as ready for review May 11, 2021 06:01
@mruberry mruberry requested a review from ngimel May 12, 2021 03:47
@mruberry mruberry added the triaged This issue has been looked at a team member, and triaged and prioritized into an appropriate module label May 12, 2021
@@ -51,7 +51,7 @@ __device__ void test(){
assert(::abs(::acosh(Half(1.0)) - ::acosh(1.0f)) <= threshold);
assert(::abs(::acosh(Half(1.0)) - ::acosh(1.0f)) <= threshold);
assert(::abs(::asinh(Half(1.0)) - ::asinh(1.0f)) <= threshold);
assert(::abs(::atanh(Half(1.0)) - ::atanh(1.0f)) <= threshold);
assert(::abs(::atanh(Half(0.5)) - ::atanh(0.5f)) <= threshold);
Copy link
Collaborator

Choose a reason for hiding this comment

The reason will be displayed to describe this comment to others. Learn more.

how did this pass?

Copy link
Collaborator Author

Choose a reason for hiding this comment

The reason will be displayed to describe this comment to others. Learn more.

I don't know. Very strange.


using namespace at;

__device__ void test(){
Copy link
Collaborator

Choose a reason for hiding this comment

The reason will be displayed to describe this comment to others. Learn more.

should this be a templated test rather than copy-paste? You can have BFloat16, __nb_bfloat16, __float2bfloat16 and their half counterparts as template parameters, is there anything else?

Copy link
Collaborator Author

Choose a reason for hiding this comment

The reason will be displayed to describe this comment to others. Learn more.

I agree. Fixed.

@ezyang
Copy link
Contributor

ezyang commented May 12, 2021

This seems fine. @ngimel do you want to make the final call?

@facebook-github-bot
Copy link
Contributor

@ngimel has imported this pull request. If you are a Facebook employee, you can view this diff on Phabricator.

@pytorch-probot
Copy link

pytorch-probot bot commented Jan 6, 2022

CI Flow Status

⚛️ CI Flow

Ruleset - Version: v1
Ruleset - File: https://github.com/pytorch/pytorch/blob/6010e9c24b81bddcbbee79a8c9194ad760e4af46/.github/generated-ciflow-ruleset.json
PR ciflow labels: ciflow/default

Workflows Labels (bold enabled) Status
Triggered Workflows
linux-bionic-py3.7-clang9 ciflow/all, ciflow/cpu, ciflow/default, ciflow/linux, ciflow/noarch, ciflow/trunk ✅ triggered
linux-docs ciflow/all, ciflow/cpu, ciflow/default, ciflow/docs, ciflow/linux, ciflow/trunk ✅ triggered
linux-vulkan-bionic-py3.7-clang9 ciflow/all, ciflow/cpu, ciflow/default, ciflow/linux, ciflow/trunk, ciflow/vulkan ✅ triggered
linux-xenial-cuda11.3-py3.7-gcc7 ciflow/all, ciflow/cuda, ciflow/default, ciflow/linux, ciflow/trunk ✅ triggered
linux-xenial-cuda11.3-py3.7-gcc7-bazel-test ciflow/all, ciflow/bazel, ciflow/cpu, ciflow/default, ciflow/linux, ciflow/trunk ✅ triggered
linux-xenial-py3-clang5-mobile-build ciflow/all, ciflow/default, ciflow/linux, ciflow/mobile, ciflow/trunk ✅ triggered
linux-xenial-py3-clang5-mobile-custom-build-static ciflow/all, ciflow/default, ciflow/linux, ciflow/mobile, ciflow/trunk ✅ triggered
linux-xenial-py3.7-clang7-asan ciflow/all, ciflow/cpu, ciflow/default, ciflow/linux, ciflow/sanitizers, ciflow/trunk ✅ triggered
linux-xenial-py3.7-clang7-onnx ciflow/all, ciflow/cpu, ciflow/default, ciflow/linux, ciflow/onnx, ciflow/trunk ✅ triggered
linux-xenial-py3.7-gcc5.4 ciflow/all, ciflow/cpu, ciflow/default, ciflow/linux, ciflow/trunk ✅ triggered
linux-xenial-py3.7-gcc7 ciflow/all, ciflow/cpu, ciflow/default, ciflow/linux, ciflow/trunk ✅ triggered
linux-xenial-py3.7-gcc7-no-ops ciflow/all, ciflow/cpu, ciflow/default, ciflow/linux, ciflow/trunk ✅ triggered
pytorch-linux-xenial-py3-clang5-android-ndk-r19c-gradle-custom-build-single ciflow/all, ciflow/android, ciflow/cpu, ciflow/default, ciflow/linux, ciflow/trunk ✅ triggered
pytorch-linux-xenial-py3-clang5-android-ndk-r19c-gradle-custom-build-single-full-jit ciflow/all, ciflow/android, ciflow/cpu, ciflow/default, ciflow/linux, ciflow/trunk ✅ triggered
win-vs2019-cpu-py3 ciflow/all, ciflow/cpu, ciflow/default, ciflow/trunk, ciflow/win ✅ triggered
win-vs2019-cuda11.3-py3 ciflow/all, ciflow/cuda, ciflow/default, ciflow/trunk, ciflow/win ✅ triggered
Skipped Workflows
caffe2-linux-xenial-py3.7-gcc5.4 ciflow/all, ciflow/cpu, ciflow/linux, ciflow/trunk 🚫 skipped
docker-builds ciflow/all, ciflow/trunk 🚫 skipped
ios-12-5-1-arm64 ciflow/all, ciflow/ios, ciflow/macos, ciflow/trunk 🚫 skipped
ios-12-5-1-arm64-coreml ciflow/all, ciflow/ios, ciflow/macos, ciflow/trunk 🚫 skipped
ios-12-5-1-arm64-custom-ops ciflow/all, ciflow/ios, ciflow/macos, ciflow/trunk 🚫 skipped
ios-12-5-1-arm64-full-jit ciflow/all, ciflow/ios, ciflow/macos, ciflow/trunk 🚫 skipped
ios-12-5-1-arm64-metal ciflow/all, ciflow/ios, ciflow/macos, ciflow/trunk 🚫 skipped
ios-12-5-1-x86-64 ciflow/all, ciflow/ios, ciflow/macos, ciflow/trunk 🚫 skipped
ios-12-5-1-x86-64-coreml ciflow/all, ciflow/ios, ciflow/macos, ciflow/trunk 🚫 skipped
ios-12-5-1-x86-64-full-jit ciflow/all, ciflow/ios, ciflow/macos, ciflow/trunk 🚫 skipped
libtorch-linux-xenial-cuda10.2-py3.7-gcc7 ciflow/all, ciflow/cuda, ciflow/libtorch, ciflow/linux, ciflow/trunk 🚫 skipped
libtorch-linux-xenial-cuda11.3-py3.7-gcc7 ciflow/all, ciflow/cuda, ciflow/libtorch, ciflow/linux, ciflow/trunk 🚫 skipped
linux-bionic-cuda10.2-py3.9-gcc7 ciflow/all, ciflow/cuda, ciflow/linux, ciflow/slow, ciflow/trunk 🚫 skipped
linux-docs-push ciflow/all, ciflow/cpu, ciflow/linux, ciflow/scheduled 🚫 skipped
linux-xenial-cuda11.3-py3.7-gcc7-no-ops ciflow/all, ciflow/cuda, ciflow/linux, ciflow/trunk 🚫 skipped
macos-10-15-py3-arm64 ciflow/all, ciflow/macos, ciflow/trunk 🚫 skipped
macos-10-15-py3-lite-interpreter-x86-64 ciflow/all, ciflow/macos, ciflow/trunk 🚫 skipped
macos-11-py3-x86-64 ciflow/all, ciflow/macos, ciflow/trunk 🚫 skipped
parallelnative-linux-xenial-py3.7-gcc5.4 ciflow/all, ciflow/cpu, ciflow/linux, ciflow/trunk 🚫 skipped
periodic-libtorch-linux-bionic-cuda11.5-py3.7-gcc7 ciflow/all, ciflow/cuda, ciflow/libtorch, ciflow/linux, ciflow/scheduled 🚫 skipped
periodic-libtorch-linux-xenial-cuda11.1-py3.7-gcc7 ciflow/all, ciflow/cuda, ciflow/libtorch, ciflow/linux, ciflow/scheduled 🚫 skipped
periodic-linux-bionic-cuda11.5-py3.7-gcc7 ciflow/all, ciflow/cuda, ciflow/linux, ciflow/scheduled 🚫 skipped
periodic-linux-xenial-cuda10.2-py3-gcc7-slow-gradcheck ciflow/all, ciflow/cuda, ciflow/linux, ciflow/scheduled, ciflow/slow, ciflow/slow-gradcheck 🚫 skipped
periodic-linux-xenial-cuda11.1-py3.7-gcc7-debug ciflow/all, ciflow/cuda, ciflow/linux, ciflow/scheduled 🚫 skipped
periodic-win-vs2019-cuda11.1-py3 ciflow/all, ciflow/cuda, ciflow/scheduled, ciflow/win 🚫 skipped
periodic-win-vs2019-cuda11.5-py3 ciflow/all, ciflow/cuda, ciflow/scheduled, ciflow/win 🚫 skipped
pytorch-linux-xenial-py3-clang5-android-ndk-r19c-build ciflow/all, ciflow/android, ciflow/cpu, ciflow/linux, ciflow/trunk 🚫 skipped

You can add a comment to the PR and tag @pytorchbot with the following commands:
# ciflow rerun, "ciflow/default" will always be added automatically
@pytorchbot ciflow rerun

# ciflow rerun with additional labels "-l <ciflow/label_name>", which is equivalent to adding these labels manually and trigger the rerun
@pytorchbot ciflow rerun -l ciflow/scheduled -l ciflow/slow

For more information, please take a look at the CI Flow Wiki.

@zasdfgbnm
Copy link
Collaborator Author

@ngimel This was not merged because a CUDA 11.0 bug causes internal build failure. Is the internal build still using CUDA 11.0 today?

@github-actions
Copy link
Contributor

Looks like this PR hasn't been updated in a while so we're going to go ahead and mark this as Stale.
Feel free to remove the Stale label if you feel this was a mistake.
If you are unable to remove the Stale label please contact a maintainer in order to do so.
If you want the bot to never mark this PR stale again, add the no-stale label.
Stale pull requests will automatically be closed after 30 days of inactivity.

@github-actions github-actions bot added the Stale label May 21, 2022
@zasdfgbnm zasdfgbnm removed the Stale label May 21, 2022
@zasdfgbnm
Copy link
Collaborator Author

removing Stale label.
@ngimel Is meta internally still using CUDA 11.0, if not, I think this is ready to go.

@github-actions
Copy link
Contributor

Looks like this PR hasn't been updated in a while so we're going to go ahead and mark this as Stale.
Feel free to remove the Stale label if you feel this was a mistake.
If you are unable to remove the Stale label please contact a maintainer in order to do so.
If you want the bot to never mark this PR stale again, add the no-stale label.
Stale pull requests will automatically be closed after 30 days of inactivity.

@github-actions github-actions bot added the Stale label Jul 20, 2022
@zasdfgbnm
Copy link
Collaborator Author

@pytorchbot rebase

@zasdfgbnm zasdfgbnm removed the Stale label Jul 21, 2022
@pytorchmergebot
Copy link
Collaborator

@pytorchbot successfully started a rebase job. Check the current status here

@pytorchmergebot
Copy link
Collaborator

Rebase failed due to Command git -C /home/runner/work/pytorch/pytorch rebase refs/remotes/origin/master pull/57984/head returned non-zero exit code 1

Rebasing (1/12)
Rebasing (2/12)
Auto-merging c10/test/util/complex_test_common.h
CONFLICT (content): Merge conflict in c10/test/util/complex_test_common.h
Auto-merging c10/util/BFloat16-inl.h
Auto-merging c10/util/BFloat16.h
error: could not apply 9c5c6229dc... save
hint: Resolve all conflicts manually, mark them as resolved with
hint: "git add/rm <conflicted_files>", then run "git rebase --continue".
hint: You can instead skip this commit: run "git rebase --skip".
hint: To abort and get back to the state before "git rebase", run "git rebase --abort".
Could not apply 9c5c6229dc... save

Raised by https://github.com/pytorch/pytorch/actions/runs/2708457304

@pytorch-bot
Copy link

pytorch-bot bot commented Sep 10, 2022

🔗 Helpful Links

🧪 See artifacts and rendered test results at hud.pytorch.org/pr/57984

Note: Links to docs will display an error until the docs builds have been completed.

❌ 6 Failures, 1 Pending

As of commit 2a79341:

The following jobs have failed:

This comment was automatically generated by Dr. CI and updates every 15 minutes.

@facebook-github-bot
Copy link
Contributor

/easycla

As part of the transition to the PyTorch Foundation, this project now requires contributions be covered under the new CLA. See #85559 for additional details.

This comment will trigger a new check of this PR. If you are already covered, you will simply see a new "EasyCLA" check that passes. If you are not covered, a bot will leave a new comment with a link to sign.

@linux-foundation-easycla
Copy link

linux-foundation-easycla bot commented Oct 4, 2022

CLA Signed

The committers listed above are authorized under a signed CLA.

@kit1980
Copy link
Contributor

kit1980 commented Nov 22, 2022

@pytorchbot rebase

@pytorchmergebot
Copy link
Collaborator

@pytorchbot successfully started a rebase job. Check the current status here

@pytorchmergebot
Copy link
Collaborator

Rebase failed due to Command git -C /home/runner/work/pytorch/pytorch rebase refs/remotes/origin/viable/strict pull/57984/head returned non-zero exit code 1

Rebasing (1/12)
Auto-merging aten/src/ATen/test/CMakeLists.txt
CONFLICT (content): Merge conflict in aten/src/ATen/test/CMakeLists.txt
Auto-merging aten/src/ATen/test/cuda_half_test.cu
Auto-merging aten/tools/run_tests.sh
error: could not apply fbd1e1195b... Add tests for bfloat16
hint: Resolve all conflicts manually, mark them as resolved with
hint: "git add/rm <conflicted_files>", then run "git rebase --continue".
hint: You can instead skip this commit: run "git rebase --skip".
hint: To abort and get back to the state before "git rebase", run "git rebase --abort".
Could not apply fbd1e1195b... Add tests for bfloat16

Raised by https://github.com/pytorch/pytorch/actions/runs/3527515860

@kit1980
Copy link
Contributor

kit1980 commented Nov 22, 2022

@ngimel @zasdfgbnm Do we still want to merge this?

@github-actions
Copy link
Contributor

Looks like this PR hasn't been updated in a while so we're going to go ahead and mark this as Stale.
Feel free to remove the Stale label if you feel this was a mistake.
If you are unable to remove the Stale label please contact a maintainer in order to do so.
If you want the bot to never mark this PR stale again, add the no-stale label.
Stale pull requests will automatically be closed after 30 days of inactivity.

@github-actions github-actions bot added the Stale label Jan 21, 2023
@github-actions github-actions bot closed this Feb 20, 2023
@github-actions github-actions bot deleted the ci-all/bfloat16-cuda-tests branch March 11, 2024 01:50
Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Labels
cla signed open source Stale triaged This issue has been looked at a team member, and triaged and prioritized into an appropriate module
Projects
None yet
Development

Successfully merging this pull request may close these issues.

9 participants