Continue to build nightly CUDA 12.9 for internal#163029
Continue to build nightly CUDA 12.9 for internal#163029huydhn wants to merge 3 commits intopytorch:mainfrom
Conversation
🔗 Helpful Links🧪 See artifacts and rendered test results at hud.pytorch.org/pr/163029
Note: Links to docs will display an error until the docs builds have been completed. ❗ 1 Active SEVsThere are 1 currently active SEVs. If your PR is affected, please view them below: ✅ You can merge normally! (2 Unrelated Failures)As of commit 6295362 with merge base 12d7cc5 ( FLAKY - The following jobs failed but were likely due to flakiness present on trunk:
This comment was automatically generated by Dr. CI and updates every 15 minutes. |
malfet
left a comment
There was a problem hiding this comment.
Please mention an issue that answers a deadline on when to revert it, but sure, why not
I also doubt my sanity in doing this, so let's get this one ready, but not land it unless we really need it. Also add a note here that pytorch/test-infra#7074 needs to be reverted too to build domains on 12.9 |
|
@huydhn please provide some context on this. Supporting 4 CUDA versions across 3 platforms is quite expensive. Can we only build specific Python version and only Linux ? |
I'm keeping this around in case people ask for this internally (post). From the response so far, I don't think there is enough incentive to land this yet |
|
@pytorchbot rebase |
|
@pytorchbot started a rebase job onto refs/remotes/origin/viable/strict. Check the current status here |
Signed-off-by: Huy Do <huydhn@gmail.com>
Signed-off-by: Huy Do <huydhn@gmail.com>
|
Successfully rebased |
cc72948 to
b1f78ae
Compare
|
@pytorchbot drci |
|
@pytorchbot merge -f '12.9 build looks ok' |
Merge startedYour change will be merged immediately since you used the force (-f) flag, bypassing any CI checks (ETA: 1-5 minutes). Please use Learn more about merging in the wiki. Questions? Feedback? Please reach out to the PyTorch DevX Team |
|
@pytorchbot --help |
PyTorchBot HelpMergeRevertRebaseLabelDr CIcherry-pick |
|
@pytorchbot cherry-pick --onto release/2.9 --fixes 'vLLM CUDA 12.9 build' -c release |
Revert part of #161916 to continue building CUDA 12.9 nightly Pull Request resolved: #163029 Approved by: https://github.com/malfet (cherry picked from commit 4400c5d)
Cherry picking #163029The cherry pick PR is at #165466 and it is linked with issue vLLM CUDA 12.9 build. The following tracker issues are updated: Details for Dev Infra teamRaised by workflow job |
* Continue to build nightly CUDA 12.9 for internal (#163029) Revert part of #161916 to continue building CUDA 12.9 nightly Pull Request resolved: #163029 Approved by: https://github.com/malfet (cherry picked from commit 4400c5d) * Fix lint Signed-off-by: Huy Do <huydhn@gmail.com> --------- Signed-off-by: Huy Do <huydhn@gmail.com> Co-authored-by: Huy Do <huydhn@gmail.com>
When trying to bring cu129 back in #163029, I mainly looked at #163029 and missed another tweak coming from #162455 I discover this issue when testing aarch64+cu129 builds in https://github.com/pytorch/test-infra/actions/runs/18603342105/job/53046883322?pr=7373. Surprisingly, there is no test running for aarch64 CUDA build from what I see in https://hud.pytorch.org/pytorch/pytorch/commit/79a37055e790482c12bf32e69b28c8e473d0209d. Pull Request resolved: #165794 Approved by: https://github.com/malfet
When trying to bring cu129 back in #163029, I mainly looked at #163029 and missed another tweak coming from #162455 I discover this issue when testing aarch64+cu129 builds in https://github.com/pytorch/test-infra/actions/runs/18603342105/job/53046883322?pr=7373. Surprisingly, there is no test running for aarch64 CUDA build from what I see in https://hud.pytorch.org/pytorch/pytorch/commit/79a37055e790482c12bf32e69b28c8e473d0209d. Pull Request resolved: #165794 Approved by: https://github.com/malfet (cherry picked from commit 9095a9d)
[CD] Apply the fix from #162455 to aarch64+cu129 build (#165794) When trying to bring cu129 back in #163029, I mainly looked at #163029 and missed another tweak coming from #162455 I discover this issue when testing aarch64+cu129 builds in https://github.com/pytorch/test-infra/actions/runs/18603342105/job/53046883322?pr=7373. Surprisingly, there is no test running for aarch64 CUDA build from what I see in https://hud.pytorch.org/pytorch/pytorch/commit/79a37055e790482c12bf32e69b28c8e473d0209d. Pull Request resolved: #165794 Approved by: https://github.com/malfet (cherry picked from commit 9095a9d) Co-authored-by: Huy Do <huydhn@gmail.com>
Revert part of pytorch#161916 to continue building CUDA 12.9 nightly Pull Request resolved: pytorch#163029 Approved by: https://github.com/malfet
…h#165794) When trying to bring cu129 back in pytorch#163029, I mainly looked at pytorch#163029 and missed another tweak coming from pytorch#162455 I discover this issue when testing aarch64+cu129 builds in https://github.com/pytorch/test-infra/actions/runs/18603342105/job/53046883322?pr=7373. Surprisingly, there is no test running for aarch64 CUDA build from what I see in https://hud.pytorch.org/pytorch/pytorch/commit/79a37055e790482c12bf32e69b28c8e473d0209d. Pull Request resolved: pytorch#165794 Approved by: https://github.com/malfet
…h#165794) When trying to bring cu129 back in pytorch#163029, I mainly looked at pytorch#163029 and missed another tweak coming from pytorch#162455 I discover this issue when testing aarch64+cu129 builds in https://github.com/pytorch/test-infra/actions/runs/18603342105/job/53046883322?pr=7373. Surprisingly, there is no test running for aarch64 CUDA build from what I see in https://hud.pytorch.org/pytorch/pytorch/commit/79a37055e790482c12bf32e69b28c8e473d0209d. Pull Request resolved: pytorch#165794 Approved by: https://github.com/malfet
Revert part of #161916 to continue building CUDA 12.9 nightly
cc @albanD