Skip to content

[Inductor] Make combo kernel MAX_NUM_ARGS configurable#166274

Closed
andyanwang wants to merge 1 commit intogh/andyanwang/40/basefrom
gh/andyanwang/40/head
Closed

[Inductor] Make combo kernel MAX_NUM_ARGS configurable#166274
andyanwang wants to merge 1 commit intogh/andyanwang/40/basefrom
gh/andyanwang/40/head

Conversation

@andyanwang
Copy link
Contributor

@andyanwang andyanwang commented Oct 26, 2025

@pytorch-bot
Copy link

pytorch-bot bot commented Oct 26, 2025

🔗 Helpful Links

🧪 See artifacts and rendered test results at hud.pytorch.org/pr/166274

Note: Links to docs will display an error until the docs builds have been completed.

✅ No Failures

As of commit 8e29b50 with merge base c7eee49 (image):
💚 Looks good so far! There are no failures yet. 💚

This comment was automatically generated by Dr. CI and updates every 15 minutes.

@pytorch-bot pytorch-bot bot added the ciflow/trunk Trigger trunk jobs on your pull request label Oct 27, 2025
@facebook-github-bot
Copy link
Contributor

@pytorchbot merge

(Initiating merge automatically since Phabricator Diff has merged)

@pytorchmergebot
Copy link
Collaborator

Merge started

Your change will be merged once all checks pass (ETA 0-4 Hours).

Learn more about merging in the wiki.

Questions? Feedback? Please reach out to the PyTorch DevX Team

Advanced Debugging
Check the merge workflow status
here

pytorchmergebot pushed a commit that referenced this pull request Oct 29, 2025
…uts (#166275)

MTIA triton currently has a limit that it can't support the cases when there are too many input/output buffers. This PR adds the limitation to prevent large fusion with many input/output buffer.

Differential Revision: [D85509351](https://our.internmc.facebook.com/intern/diff/D85509351/)

Pull Request resolved: #166275
Approved by: https://github.com/eellison
ghstack dependencies: #166274
tianrengao pushed a commit that referenced this pull request Oct 30, 2025
The MAX_NUM_ARGS of ComboKernel is currently a fixed number. We need to tune this number to avoid large fusion for MTIA, thus making it configurable.

Differential Revision: [D85509352](https://our.internmc.facebook.com/intern/diff/D85509352/)

Pull Request resolved: #166274
Approved by: https://github.com/eellison
BoyuanFeng pushed a commit that referenced this pull request Oct 31, 2025
…uts (#166275)

MTIA triton currently has a limit that it can't support the cases when there are too many input/output buffers. This PR adds the limitation to prevent large fusion with many input/output buffer.

Differential Revision: [D85509351](https://our.internmc.facebook.com/intern/diff/D85509351/)

Pull Request resolved: #166275
Approved by: https://github.com/eellison
ghstack dependencies: #166274
Khanaksahu pushed a commit to Khanaksahu/pytorch-fork that referenced this pull request Nov 17, 2025
The MAX_NUM_ARGS of ComboKernel is currently a fixed number. We need to tune this number to avoid large fusion for MTIA, thus making it configurable.

Pull Request resolved: pytorch/pytorch#166274

ghstack-source-id: 318804069
@exported-using-ghexport

Differential Revision: [D85509352](https://our.internmc.facebook.com/intern/diff/D85509352/)
Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment

Projects

None yet

Development

Successfully merging this pull request may close these issues.

4 participants