Skip to content

Conversation

@eqy
Copy link
Collaborator

@eqy eqy commented Nov 23, 2022

@pytorch-bot pytorch-bot bot added the release notes: sparse release notes category label Nov 23, 2022
@pytorch-bot
Copy link

pytorch-bot bot commented Nov 23, 2022

🔗 Helpful Links

🧪 See artifacts and rendered test results at hud.pytorch.org/pr/89582

Note: Links to docs will display an error until the docs builds have been completed.

❌ 1 Failures

As of commit 39a85f4:

FLAKY - The following jobs failed but were likely due to flakiness present on master:

This comment was automatically generated by Dr. CI and updates every 15 minutes.

@yaox12
Copy link
Contributor

yaox12 commented Dec 5, 2022

There's some legacy code wrapped with the __CUDACC_VER_MAJOR__ macro, should we remove them too? For example,

#if __CUDACC_VER_MAJOR__ < 8 || defined(USE_ROCM)

})
#endif

// Workaround for C10_UNUSED because CUDA 10.2 and below fails to handle unused
Copy link
Collaborator

Choose a reason for hiding this comment

The reason will be displayed to describe this comment to others. Learn more.

remove this comment also

#if defined(__CUDACC__) && CUDA_VERSION < 11000
#define C10_UNUSED_DISPATCH_CUDA_WORKAROUND
#else
#define C10_UNUSED_DISPATCH_CUDA_WORKAROUND C10_UNUSED
Copy link
Collaborator

Choose a reason for hiding this comment

The reason will be displayed to describe this comment to others. Learn more.

probably all C10_UNUSED_DISPATCH... can be replaced by C10_UNUSED now?

Copy link
Collaborator

Choose a reason for hiding this comment

The reason will be displayed to describe this comment to others. Learn more.

Nit: can we just replace C10_UNUSED_DISPATCH_CUDA_WORKAROUND with C10_UNUSED everywhere?

template <typename T>
C10_HOST_DEVICE constexpr thrust::complex<T>
cuda101bug_cast_c10_complex_to_thrust_complex(const c10::complex<T>& x) {
#if defined(CUDA_VERSION) && (CUDA_VERSION < 10020)
Copy link
Collaborator

Choose a reason for hiding this comment

The reason will be displayed to describe this comment to others. Learn more.

similarly here, cuda101bug... uses should just be replaced with static_cast, as the comment suggests

@eqy eqy added the ciflow/trunk Trigger trunk jobs on your pull request label Dec 6, 2022
@eqy
Copy link
Collaborator Author

eqy commented Dec 6, 2022

Hmm, what macro should be used instead of USE_ROCM or TORCH_HIP_VERSION in c10/util/*? The former isn't defined there for hip builds and the latter seems to still break on the test CI runs (e.g., https://github.com/pytorch/pytorch/actions/runs/3625677594/jobs/6114427598#logs )

@eqy
Copy link
Collaborator Author

eqy commented Dec 6, 2022

CC @jithunnair-amd who might know more about the last issue

@eqy eqy changed the title [WIP][CUDA] Drop CUDA 10 support [CUDA] Drop CUDA 10 support Dec 30, 2022
@eqy
Copy link
Collaborator Author

eqy commented Dec 30, 2022

I've left the version guards in the c10/cuda files in for now as no one from AMD/ROCM has responded. Dropping the [WIP] tag.

@eqy eqy added the ciflow/periodic Trigger jobs ran periodically on master (periodic.yml) on the PR label Dec 30, 2022
@eqy eqy requested a review from ngimel December 30, 2022 04:39
#if defined(__CUDACC__) && CUDA_VERSION < 11000
#define C10_UNUSED_DISPATCH_CUDA_WORKAROUND
#else
#define C10_UNUSED_DISPATCH_CUDA_WORKAROUND C10_UNUSED
Copy link
Collaborator

Choose a reason for hiding this comment

The reason will be displayed to describe this comment to others. Learn more.

Nit: can we just replace C10_UNUSED_DISPATCH_CUDA_WORKAROUND with C10_UNUSED everywhere?

@eqy
Copy link
Collaborator Author

eqy commented Jan 5, 2023

@pytorchmergebot merge -f "ROCM failure appears unrelated"

@pytorchmergebot
Copy link
Collaborator

Merge started

Your change will be merged immediately since you used the force (-f) flag, bypassing any CI checks (ETA: 1-5 minutes).

Learn more about merging in the wiki.

Questions? Feedback? Please reach out to the PyTorch DevX Team

Advanced Debugging
Check the merge workflow status
here

pytorchmergebot pushed a commit that referenced this pull request Jan 24, 2023
Follow-up of #89582 to drop flags like `CUDA11OrLater` in tests. Note that in some places it appears that `TEST_WITH_ROCM` is _implicitly_ guarded against via the `CUDA11OrLater` version check, based on my best-guess of how `torch.version.cuda` would behave in ROCM builds, so I've added `not TEST_WITH_ROCM` in cases where ROCM wasn't previously explicitly allowed.

CC @ptrblck @malfet @ngimel
Pull Request resolved: #92605
Approved by: https://github.com/ngimel
Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment

Labels

ciflow/periodic Trigger jobs ran periodically on master (periodic.yml) on the PR ciflow/trunk Trigger trunk jobs on your pull request Merged open source release notes: sparse release notes category

Projects

None yet

Development

Successfully merging this pull request may close these issues.

5 participants