-
Notifications
You must be signed in to change notification settings - Fork 1.2k
RuntimeError with AttentionUNet #5422
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
Comments
thanks for reporting, that's a bug in the |
wyli
added a commit
that referenced
this issue
Oct 28, 2022
Signed-off-by: Wenqi Li <[email protected]> Fixes #5422 ### Description the kernel size and strides are hard-coded: https://github.com/Project-MONAI/MONAI/blob/a209b06438343830e561a0afd41b1025516a8977/monai/networks/nets/attentionunet.py#L151 this PR makes the values tunable. `level` parameter is not used and removed in this PR. ### Types of changes <!--- Put an `x` in all the boxes that apply, and remove the not applicable items --> - [x] Non-breaking change (fix or new feature that would not break existing functionality). - [ ] Breaking change (fix or new feature that would cause existing functionality to change). - [x] New tests added to cover the changes. - [ ] Integration tests passed locally by running `./runtests.sh -f -u --net --coverage`. - [ ] Quick tests passed locally by running `./runtests.sh --quick --unittests --disttests`. - [ ] In-line docstrings updated. - [ ] Documentation updated, tested `make html` command in the `docs/` folder. Signed-off-by: Wenqi Li <[email protected]>
KumoLiu
pushed a commit
that referenced
this issue
Nov 2, 2022
Signed-off-by: Wenqi Li <[email protected]> Fixes #5422 ### Description the kernel size and strides are hard-coded: https://github.com/Project-MONAI/MONAI/blob/a209b06438343830e561a0afd41b1025516a8977/monai/networks/nets/attentionunet.py#L151 this PR makes the values tunable. `level` parameter is not used and removed in this PR. ### Types of changes <!--- Put an `x` in all the boxes that apply, and remove the not applicable items --> - [x] Non-breaking change (fix or new feature that would not break existing functionality). - [ ] Breaking change (fix or new feature that would cause existing functionality to change). - [x] New tests added to cover the changes. - [ ] Integration tests passed locally by running `./runtests.sh -f -u --net --coverage`. - [ ] Quick tests passed locally by running `./runtests.sh --quick --unittests --disttests`. - [ ] In-line docstrings updated. - [ ] Documentation updated, tested `make html` command in the `docs/` folder. Signed-off-by: Wenqi Li <[email protected]> Signed-off-by: KumoLiu <[email protected]>
Sign up for free
to join this conversation on GitHub.
Already have an account?
Sign in to comment
Discussed in #5421
Originally posted by njneeteson October 27, 2022
I'm getting a RuntimeError when using the AttentionUnet in both 2D and 3D. This happens when training using a dataset and training script that works fine with a normal unet, a segan, and a unet-r (in both 2D and 3D) so there must be something specific that I am doing wrong or misunderstanding about how to use the attentionunet properly.
Here is a very minimal sample of code to reproduce the problem:
Here is the terminal output:
I tried to go in and look at the source code to trace the problem but I can't figure out what the problem would be, the code is a bit too complex for me to backtrace to where the issue might be.
Does anyone use the AttentionUnet successfully and can see what I am doing wrong? Or is this a bug?
The text was updated successfully, but these errors were encountered: