Skip to content

Conversation

@masnesral
Copy link
Contributor

@masnesral masnesral commented Jan 30, 2025

@pytorch-bot
Copy link

pytorch-bot bot commented Jan 30, 2025

🔗 Helpful Links

🧪 See artifacts and rendered test results at hud.pytorch.org/pr/146088

Note: Links to docs will display an error until the docs builds have been completed.

❗ 1 Active SEVs

There are 1 currently active SEVs. If your PR is affected, please view them below:

❌ 1 New Failure

As of commit f8d8bc3 with merge base 354fe48 (image):

NEW FAILURE - The following job has failed:

This comment was automatically generated by Dr. CI and updates every 15 minutes.

masnesral added a commit that referenced this pull request Jan 30, 2025
@masnesral masnesral added the topic: not user facing topic category label Jan 30, 2025
@masnesral masnesral marked this pull request as ready for review January 30, 2025 22:33
# This is mainly used to reduce test time in CI.
cutlass_max_profiling_configs: Optional[int] = None

# The L2 swizzle values to consider when profiling CUTLASS configs in max_autotune.
Copy link
Contributor

Choose a reason for hiding this comment

The reason will be displayed to describe this comment to others. Learn more.

probably also mention what are good values to put in here

Copy link
Collaborator

@Chillee Chillee left a comment

Choose a reason for hiding this comment

The reason will be displayed to describe this comment to others. Learn more.

I'm not sure I quite understand this PR? Is it just to reduce compilation time?

@masnesral
Copy link
Contributor Author

I'm not sure I quite understand this PR? Is it just to reduce compilation time?

@Chillee, yeah. For one data point, see: https://fburl.com/workplace/gx0zim0l
I'd think the default is not a big deal and the more important part is to make it configurable. WDYT? Would you make the default 1,2,3,4?

@henrylhtsang
Copy link
Contributor

I'm not sure I quite understand this PR? Is it just to reduce compilation time?

It will also 4x the number of configs, which cannot be controlled by cutlass_max_profiling_configs. For example, even if you set cutlass_max_profiling_configs = 10, you will still be autotuning 40 configs.

@henrylhtsang
Copy link
Contributor

I'm not sure I quite understand this PR? Is it just to reduce compilation time?

@Chillee, yeah. For one data point, see: https://fburl.com/workplace/gx0zim0l I'd think the default is not a big deal and the more important part is to make it configurable. WDYT? Would you make the default 1,2,3,4?

@Chillee @masnesral can you land this just to recover flaky test signal?
https://www.internalfb.com/intern/test/281475163758161?ref_report_id=0
https://www.internalfb.com/intern/test/562950140502597?ref_report_id=0

Even if we want to set default as [1, 2, 4, 8], as long as it is configurable, we can fix that in the test.

@masnesral
Copy link
Contributor Author

I'll land as is. I only chose '2' because @henrylhtsang suggest that in offline discussion. If someone can give insights on the "best" default, I'll gladly change it.

@masnesral
Copy link
Contributor Author

@pytorchbot merge

@pytorch-bot pytorch-bot bot added the ciflow/trunk Trigger trunk jobs on your pull request label Feb 4, 2025
@pytorchmergebot
Copy link
Collaborator

Merge started

Your change will be merged once all checks pass (ETA 0-4 Hours).

Learn more about merging in the wiki.

Questions? Feedback? Please reach out to the PyTorch DevX Team

Advanced Debugging
Check the merge workflow status
here

@pytorchmergebot
Copy link
Collaborator

Merge failed

Reason: 1 jobs have failed, first few of them are: trunk / linux-focal-rocm6.3-py3.10 / test (distributed, 1, 1, linux.rocm.gpu.4)

Details for Dev Infra team Raised by workflow job

@masnesral
Copy link
Contributor Author

@pytorchbot merge -i

@pytorchmergebot
Copy link
Collaborator

Merge started

Your change will be merged while ignoring the following 1 checks: trunk / linux-focal-rocm6.3-py3.10 / test (distributed, 1, 1, linux.rocm.gpu.4)

Learn more about merging in the wiki.

Questions? Feedback? Please reach out to the PyTorch DevX Team

Advanced Debugging
Check the merge workflow status
here

@github-actions github-actions bot deleted the gh/masnesral/174/head branch March 7, 2025 02:07
Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment

Labels

Projects

None yet

Development

Successfully merging this pull request may close these issues.

6 participants