-
Notifications
You must be signed in to change notification settings - Fork 7.1k
Feature Pyramid Network code bug #2326
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
Comments
Hi @cl2227619761, what do you mean by
? Please post a minimal example where you encounter an issue with this line. |
That is a very good catch! @pmeier the issue is that A PR fixing this would be great, but it would also be great if we could measure the impact of this incorrect initialization in the model performance. @mthrok if you find the time, it would be great to assess how much this mistake affects performance. |
My bad, I misunderstood the comment above the loop: vision/torchvision/ops/feature_pyramid_network.py Lines 59 to 63 in c2e8a00
|
Can I help with this? If you point me to what model/scripts you used to train I can train with the correct init and the current one. |
@gan3sh500 You can find the training scripts here. |
I’ll have to train on single 2080ti. I’ve seen smaller batch size create worse convergence but is it fine if I compare with both the kaiming init and otherwise trained with same conditions? |
@gan3sh500 sorry for the delay in replying. If you change the number of GPUs, you'll need to adapt the learning rate to follow the linear scaling rule -- if you divide the global batch size by 8x, you should divide the learning rate by 8x |
vision/torchvision/ops/feature_pyramid_network.py
Line 60 in c2e8a00
this line does not work, I think it should be modified as self.modules() instead of self.children()
The text was updated successfully, but these errors were encountered: