We read every piece of feedback, and take your input very seriously.
To see all available qualifiers, see our documentation.
There was an error while loading. Please reload this page.
1 parent dc6c4a8 commit 8287fa8Copy full SHA for 8287fa8
aten/src/ATen/cuda/CUDABlas.cpp
@@ -193,7 +193,7 @@ static size_t _parseChosenWorkspaceSize() {
193
cudaDeviceProp* p = at::cuda::getDeviceProperties(c10::cuda::current_device());
194
// Keep workspace_size = 1024 for small Ampere GPUs
195
// See https://github.com/pytorch/pytorch/pull/120925#issuecomment-1977556485
196
- if (p->major == 8 && p->total_memory / 1073741824 >= 24) {
+ if (p->major == 8 && p->totalGlobalMem / 1073741824 >= 24) {
197
workspace_size = 4096;
198
} else if (p->major >= 9) {
199
workspace_size = 32768;
0 commit comments