-
Notifications
You must be signed in to change notification settings - Fork 26.3k
[TensorExpr] Make KernelSumMultipleAxes much faster #43905
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
Conversation
Summary: Reduce input size, skip the dtype conversion. Test Plan: test_tensorexpr --gtest_filter=TensorExprTest.KernelSum* Reviewers: Subscribers: Tasks: Tags:
nickgg
left a comment
There was a problem hiding this comment.
Choose a reason for hiding this comment
The reason will be displayed to describe this comment to others. Learn more.
💨
facebook-github-bot
left a comment
There was a problem hiding this comment.
Choose a reason for hiding this comment
The reason will be displayed to describe this comment to others. Learn more.
@asuhan has imported this pull request. If you are a Facebook employee, you can view this diff on Phabricator.
Codecov Report
@@ Coverage Diff @@
## master #43905 +/- ##
==========================================
- Coverage 69.32% 69.32% -0.01%
==========================================
Files 378 378
Lines 46787 46787
==========================================
- Hits 32436 32435 -1
- Misses 14351 14352 +1
Continue to review full report at Codecov.
|
💊 CI failures summary and remediationsAs of commit 52f1f07 (more details on the Dr. CI page):
ci.pytorch.org: 1 failedThis comment was automatically generated by Dr. CI (expand for details).Follow this link to opt-out of these comments for your Pull Requests.Please report bugs/suggestions on the GitHub issue tracker or post in the (internal) Dr. CI Users group. This comment has been revised 1 time. |
Summary: Reduce input size, skip the dtype conversion.
Test Plan: test_tensorexpr --gtest_filter=TensorExprTest.KernelSum*