Skip to content

Commit b39bc70

Browse files
Aidyn-Apytorchmergebot
authored andcommitted
fix totalGlobalMem
1 parent 853f7f0 commit b39bc70

File tree

1 file changed

+1
-1
lines changed

1 file changed

+1
-1
lines changed

aten/src/ATen/cuda/CUDABlas.cpp

Lines changed: 1 addition & 1 deletion
Original file line numberDiff line numberDiff line change
@@ -191,7 +191,7 @@ static size_t _parseChosenWorkspaceSize() {
191191
cudaDeviceProp* p = at::cuda::getDeviceProperties(c10::cuda::current_device());
192192
// Keep workspace_size = 1024 for small Ampere GPUs
193193
// See https://github.com/pytorch/pytorch/pull/120925#issuecomment-1977556485
194-
if (p->major == 8 && p->total_memory / 1073741824 >= 24) {
194+
if (p->major == 8 && p->totalGlobalMem / 1073741824 >= 24) {
195195
workspace_size = 4096;
196196
} else if (p->major >= 9) {
197197
workspace_size = 32768;

0 commit comments

Comments
 (0)