-
Notifications
You must be signed in to change notification settings - Fork 26.3k
Batch norm layer for pseudo-fp16 support. #2388
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
Conversation
soumith
left a comment
There was a problem hiding this comment.
Choose a reason for hiding this comment
The reason will be displayed to describe this comment to others. Learn more.
thanks for introducing nn.contrib. Long-standing needed change.
|
It generates unused variable warnings, because now all cudnn routines call |
|
Removed BN layer in torch/nn left backend in place. |
.gitmodules
Outdated
This comment was marked as off-topic.
This comment was marked as off-topic.
Sorry, something went wrong.
This comment was marked as off-topic.
This comment was marked as off-topic.
Sorry, something went wrong.
|
I'm not sure what's going on in this change. It's changing the BatchNorm code path in C++, but stats will still be half on the Python side? Also, it's going to conflict with @ezyang's changes (to no longer use |
|
it's a rebased branch to get Zach going with Volta fp16. |
|
Yeah, but this is lacking Python changes. I can't see how will it not error out when you try to use BatchNorm |
|
@apaszke, It does not touch current batch norm in any way, just allows to call cudnn batch norm with fp16 inputs from a custom python module. It will have to be changed after Ed's changes go in. |
|
Replaced with #4021 due to cuDNN rewrite in aten. |
Batch norm layer that keeps parameters in 32-bit when using 16-bit input/output. This layer is necessary for successful pseudo fp16 training.