Skip to content

Conversation

@ngimel
Copy link
Collaborator

@ngimel ngimel commented May 22, 2018

Don't call cudaFree unconditionally, guard cudaFree call on cudaFreeMutex, submit kernel to current stream.

std::unique_lock<std::mutex> cudaFreeMutexLock(
*(THCCachingAllocator_getCudaFreeMutex()));
cudaFree(0);
cudaFreeMutexLock.unlock();

This comment was marked as off-topic.

TORCH_CU_CHECK(cuCtxGetCurrent(&pctx));
if (!pctx) {
std::unique_lock<std::mutex> cudaFreeMutexLock(
*(THCCachingAllocator_getCudaFreeMutex()));

This comment was marked as off-topic.

@soumith soumith merged commit 1feb1a9 into pytorch:master May 23, 2018
petrex pushed a commit to petrex/pytorch that referenced this pull request May 31, 2018
* origin:
  [Caffe2] Enabling AMD GPU Backend for Caffe2 (pytorch#7566)
  Call grad_mode.py context managers as decorators (pytorch#7737)
  catch CPU tensors in checkSameGPU (fixes pytorch#7689) (pytorch#7767)
  Mark stack as non-executable in NNPACK (pytorch#7752)
  small fixes in fusion_compiler (pytorch#7776)
  Run clang-format on c10d (pytorch#7791)
@ngimel ngimel deleted the jit_fixes branch June 6, 2018 05:02
weiyangfb pushed a commit to weiyangfb/pytorch that referenced this pull request Jun 11, 2018
* small fixes in fusion_compiler

* address review comments
@Oil3
Copy link

Oil3 commented Nov 12, 2023

NotImplementedError: The operator 'aten::upsample_bicubic2d.out' is not currently implemented for the MPS device. If **you want this op to be added in priority during the prototype phase of this feature, please comment on #77764. As a temporary fix, you can set the environment variable PYTORCH_ENABLE_MPS_FALLBACK=1 to use the CPU as a fallback for this op. WARNING: this will be slower than running natively on MPS

What's actually needed to implement that for MPS?

@rachelcenter
Copy link

just got this error:

Error occurred when executing AIO_Preprocessor:

The operator 'aten::upsample_bicubic2d.out' is not currently implemented for the MPS device. If you want this op to be added in priority during the prototype phase of this feature, please comment on #77764. As a temporary fix, you can set the environment variable PYTORCH_ENABLE_MPS_FALLBACK=1 to use the CPU as a fallback for this op. WARNING: this will be slower than running natively on MPS.

Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment

Projects

None yet

Development

Successfully merging this pull request may close these issues.

6 participants