You signed in with another tab or window. Reload to refresh your session.You signed out in another tab or window. Reload to refresh your session.You switched accounts on another tab or window. Reload to refresh your session.Dismiss alert
malfet
changed the title
Single GaussianBlur over 23x23 kernels fails on Windows
Single Channel GaussianBlur over 23x23 kernels fails on Windows
Feb 24, 2022
@malfet Thanks for your investigation at #5451 (comment). The gaussianblur uses conv2d so this bug has the potential of being quite significant. I saw that on your analysis you seemed to be familiar with similar previous bugs, could you link to some of them to get more context of what you suspect?
@NicolasHug@jdsgomes@vfdev-5 FYI Nikita disabled the test_gaussian_blur test for the specific configuration to restore the CI while the issue is being investigated. We should restore it once the issue is fixed.
@datumbox move just exposed the bug that existed for a while, but indeed the test code indicates that problem is somewhere in pytorch core, as test simply calls torch.nn.conv2d @ptrblck , @ngimel do you recall something of the nature?
Based on the compute-sanitizer output it's clearly an illegal memory access. I haven't seen one in this kernel so far, but will try to repro and forward to cuDNN.
Uh oh!
There was an error while loading. Please reload this page.
🐛 Describe the bug
See https://app.circleci.com/pipelines/github/pytorch/vision/14951/workflows/dd6ed737-f31d-4baa-8f42-9e69d55650aa/jobs/1205316
For some reason the same code works with CUDA-11.1, but fails with 11.3, which makes me highly suspicious of bug in cuDNN side
cuda-memcheck reports invalid memory access in `cudnnConvolutionForward ` call
cc @peterjc123 @nbcsm @guyang3532 @maxluk @gunandrose4u @mszhanyi @vfdev-5 @datumbox
The text was updated successfully, but these errors were encountered: