-
Notifications
You must be signed in to change notification settings - Fork 24.4k
Fix the QuantizedAVX2 build issue #26854
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
Conversation
Any perf checks? |
There was a problem hiding this comment.
Choose a reason for hiding this comment
The reason will be displayed to describe this comment to others. Learn more.
@llyfacebook has imported this pull request. If you are a Facebook employee, you can view this diff on Phabricator.
Yes, I found it is even faster than the original one. (For bilinear2d interpolate case.)
|
I got some undeclared identifier errors when build locally: |
@llyfacebook yeah, QuantizeAVX2 operates on vectors of 32, but you were feeding it vectors of 8, so it was just running scalar code :p |
@llyfacebook I'd expect another 2x speedup if you switch to doing the float operations 4-wide and using QuantizeAVX2 again |
There was a problem hiding this comment.
Choose a reason for hiding this comment
The reason will be displayed to describe this comment to others. Learn more.
LGTM! My local build pass after rebasing.
There was a problem hiding this comment.
Choose a reason for hiding this comment
The reason will be displayed to describe this comment to others. Learn more.
@llyfacebook is landing this pull request. If you are a Facebook employee, you can view this diff on Phabricator.
@llyfacebook merged this pull request in 428204d. |
Summary: The QuantizedAVx2 does not support the int32 type. We switch to use at::quantize_vec function instead. Pull Request resolved: pytorch/pytorch#26854 Differential Revision: D17609872 Pulled By: llyfacebook fbshipit-source-id: b4a77d93ce0ebfef696506b5cdbe3e91fe44bb36
Summary: The QuantizedAVx2 does not support the int32 type. We switch to use at::quantize_vec function instead. Pull Request resolved: pytorch#26854 Differential Revision: D17609872 Pulled By: llyfacebook fbshipit-source-id: b4a77d93ce0ebfef696506b5cdbe3e91fe44bb36
The QuantizedAVx2 does not support the int32 type. We switch to use at::quantize_vec function instead.