-
Notifications
You must be signed in to change notification settings - Fork 7
Add short cut for recomputed tv #2134
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
Conversation
default: | ||
TORCH_INTERNAL_ASSERT(false, "Unknown tensor memory type."); |
There was a problem hiding this comment.
Choose a reason for hiding this comment
The reason will be displayed to describe this comment to others. Learn more.
Unrelated but I think it is a good idea to have it
There was a problem hiding this comment.
Choose a reason for hiding this comment
The reason will be displayed to describe this comment to others. Learn more.
LGTM
So, is this an alternative fix for the problem you previously proposed the "mirror" approach? How does having dummy outputs solve the problem? |
Yes, it is. Let's say we had
After projection, it becomes
Then neither the propagation path T2->T3->T4->T5 nor T2->T1->T0->T6->T4->T5 works because they both have missing root domain. But adding (edited) |
Oh, I see. Interesting approach. Please have your explanation as a code comment. I wonder how robust and generic this approach would be, but seems like good enough. |
If it was directly on iter domains instead of tensor views it would be very robust. The fact that it's on tensor views means that we can only build relationships across tensorviews that can go through a pointwise operation. Since this is replicated tensors exactly, it should be sufficient for all our use cases. Though same day in the future all replays should probably be IterDomain based, not TensorView based. |
Fixes #2094