-
Notifications
You must be signed in to change notification settings - Fork 26.3k
[quant][core][gpu] Implemented max pooling 2D using cudnn #74673
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
Conversation
[ghstack-poisoned]
🔗 Helpful links
💊 CI failures summary and remediationsAs of commit d22eda6 (more details on the Dr. CI page): 💚 💚 Looks good so far! There are no failures yet. 💚 💚 This comment was automatically generated by Dr. CI (expand for details).Please report bugs/suggestions to the (internal) Dr. CI Users group. |
[ghstack-poisoned]
Summary: Quantized 2D max pooling was implemented using cudnn. A corresponding test case was also added. The operator requires cudnn version 8.3.3 or higher. v7 APIs (https://docs.nvidia.com/deeplearning/cudnn/api/index.html#cudnnPoolingForward) are currently used as there are currently no v8 APIs for pooling. Note the current implementation does not support dilated pooling as it is not supported in cudnn. Test plan: In pytorch main dir, execute ``` python test/test_quantization.py TestQuantizedOps.test_max_pool2d_cudnn ``` [ghstack-poisoned]
|
@dzdang has imported this pull request. If you are a Facebook employee, you can view this diff on Phabricator. |
Summary: Quantized 2D max pooling was implemented using cudnn. A corresponding test case was also added. The operator requires cudnn version 8.3.3 or higher. v7 APIs (https://docs.nvidia.com/deeplearning/cudnn/api/index.html#cudnnPoolingForward) are currently used as there are currently no v8 APIs for pooling. Note the current implementation does not support dilated pooling as it is not supported in cudnn. Test plan: In pytorch main dir, execute ``` python test/test_quantization.py TestQuantizedOps.test_max_pool2d_cudnn ``` Differential Revision: [D35135200](https://our.internmc.facebook.com/intern/diff/D35135200) [ghstack-poisoned]
Summary: Quantized 2D max pooling was implemented using cudnn. A corresponding test case was also added. The operator requires cudnn version 8.3.3 or higher. v7 APIs (https://docs.nvidia.com/deeplearning/cudnn/api/index.html#cudnnPoolingForward) are currently used as there are currently no v8 APIs for pooling. Note the current implementation does not support dilated pooling as it is not supported in cudnn. Test plan: In pytorch main dir, execute ``` python test/test_quantization.py TestQuantizedOps.test_max_pool2d_cudnn ``` ghstack-source-id: ea8ef66 Pull Request resolved: #74673
|
@dzdang has imported this pull request. If you are a Facebook employee, you can view this diff on Phabricator. |
Summary: Quantized 2D max pooling was implemented using cudnn. A corresponding test case was also added. The operator requires cudnn version 8.3.3 or higher. v7 APIs (https://docs.nvidia.com/deeplearning/cudnn/api/index.html#cudnnPoolingForward) are currently used as there are currently no v8 APIs for pooling. Note the current implementation does not support dilated pooling as it is not supported in cudnn. Test plan: In pytorch main dir, execute ``` python test/test_quantization.py TestQuantizedOps.test_max_pool2d_cudnn ``` Differential Revision: [D35135200](https://our.internmc.facebook.com/intern/diff/D35135200) [ghstack-poisoned]
|
@dzdang has imported this pull request. If you are a Facebook employee, you can view this diff on Phabricator. |
Summary: Quantized 2D max pooling was implemented using cudnn. A corresponding test case was also added. The operator requires cudnn version 8.3.3 or higher. v7 APIs (https://docs.nvidia.com/deeplearning/cudnn/api/index.html#cudnnPoolingForward) are currently used as there are currently no v8 APIs for pooling. Note the current implementation does not support dilated pooling as it is not supported in cudnn. Test plan: In pytorch main dir, execute ``` python test/test_quantization.py TestQuantizedOps.test_max_pool2d_cudnn ``` Differential Revision: [D35135200](https://our.internmc.facebook.com/intern/diff/D35135200) [ghstack-poisoned]
Summary: Quantized 2D max pooling was implemented using cudnn. A corresponding test case was also added. The operator requires cudnn version 8.3.3 or higher. v7 APIs (https://docs.nvidia.com/deeplearning/cudnn/api/index.html#cudnnPoolingForward) are currently used as there are currently no v8 APIs for pooling. Note the current implementation does not support dilated pooling as it is not supported in cudnn. Test plan: In pytorch main dir, execute ``` python test/test_quantization.py TestQuantizedOps.test_max_pool2d_cudnn ``` ghstack-source-id: 4c9ae1a Pull Request resolved: #74673
|
@dzdang has imported this pull request. If you are a Facebook employee, you can view this diff on Phabricator. |
Summary: Quantized 2D max pooling was implemented using cudnn. A corresponding test case was also added. The operator requires cudnn version 8.3.3 or higher. v7 APIs (https://docs.nvidia.com/deeplearning/cudnn/api/index.html#cudnnPoolingForward) are currently used as there are currently no v8 APIs for pooling. Note the current implementation does not support dilated pooling as it is not supported in cudnn. Test plan: In pytorch main dir, execute ``` python test/test_quantization.py TestQuantizedOps.test_max_pool2d_cudnn ``` Differential Revision: [D35135200](https://our.internmc.facebook.com/intern/diff/D35135200) [ghstack-poisoned]
Summary: Quantized 2D max pooling was implemented using cudnn. A corresponding test case was also added. The operator requires cudnn version 8.3.3 or higher. v7 APIs (https://docs.nvidia.com/deeplearning/cudnn/api/index.html#cudnnPoolingForward) are currently used as there are currently no v8 APIs for pooling. Note the current implementation does not support dilated pooling as it is not supported in cudnn. Test plan: In pytorch main dir, execute ``` python test/test_quantization.py TestQuantizedOps.test_max_pool2d_cudnn ``` ghstack-source-id: 1bcda3f Pull Request resolved: #74673
|
@dzdang has imported this pull request. If you are a Facebook employee, you can view this diff on Phabricator. |
|
are we planning to register this under native_functions.yaml instead of a new quantized::maxpool function? |
I think that's what we're trending towards with the other max pool PRs. I'll make that change |
jerryzh168
left a comment
There was a problem hiding this comment.
Choose a reason for hiding this comment
The reason will be displayed to describe this comment to others. Learn more.
looks good!
Summary: Quantized 2D max pooling was implemented using cudnn. A corresponding test case was also added. The operator requires cudnn version 8.3.3 or higher. v7 APIs (https://docs.nvidia.com/deeplearning/cudnn/api/index.html#cudnnPoolingForward) are currently used as there are currently no v8 APIs for pooling. Note the current implementation does not support dilated pooling as it is not supported in cudnn. Test plan: In pytorch main dir, execute ``` python test/test_quantization.py TestQuantizedOps.test_max_pool2d_cudnn ``` Differential Revision: [D35135200](https://our.internmc.facebook.com/intern/diff/D35135200) [ghstack-poisoned]
Summary: Quantized 2D max pooling was implemented using cudnn. A corresponding test case was also added. The operator requires cudnn version 8.3.3 or higher. v7 APIs (https://docs.nvidia.com/deeplearning/cudnn/api/index.html#cudnnPoolingForward) are currently used as there are currently no v8 APIs for pooling. Note the current implementation does not support dilated pooling as it is not supported in cudnn. Test plan: In pytorch main dir, execute ``` python test/test_quantization.py TestQuantizedOps.test_max_pool2d_cudnn ``` ghstack-source-id: d96f94a Pull Request resolved: #74673
|
@dzdang has imported this pull request. If you are a Facebook employee, you can view this diff on Phabricator. |
Summary: Pull Request resolved: #74673 Quantized 2D max pooling was implemented using cudnn. A corresponding test case was also added. The operator requires cudnn version 8.3.3 or higher. v7 APIs (https://docs.nvidia.com/deeplearning/cudnn/api/index.html#cudnnPoolingForward) are currently used as there are currently no v8 APIs for pooling. Note the current implementation does not support dilated pooling as it is not supported in cudnn. Test Plan: In pytorch main dir, execute ``` python test/test_quantization.py TestQuantizedOps.test_max_pool2d_cudnn ``` In pytorch main dir, execute ``` python test/test_quantization.py TestQuantizedOps.test_max_pool2d_cudnn ``` Differential Revision: D35135200 D35135200 Reviewed By: jerryzh168 Pulled By: dzdang fbshipit-source-id: 199991031e0419e13578a1d4abbe17dd2ed98f66
Stack from ghstack (oldest at bottom):
Summary:
Quantized 2D max pooling was implemented using cudnn. A corresponding
test case was also added. The operator requires cudnn version 8.3.3
or higher. v7 APIs (https://docs.nvidia.com/deeplearning/cudnn/api/index.html#cudnnPoolingForward)
are currently used as there are currently no v8
APIs for pooling. Note the current implementation does not support dilated pooling
as it is not supported in cudnn.
Test plan:
In pytorch main dir, execute
Differential Revision: D35135200