Spaces:
Running
Running
What are differences between Max Allocated Memory, Max Reserved Memory and Max Used Memory?
#21
by
zhiminy
- opened
Could anyone explain it? Thanks!
There are multiple ways to use CUDA memory, pytorch
for example will allocate some memory for tensors, but will also reserve some for its computation, so reserved = (allocated + cached). It's important to look at both because the performance we observe depends on that reserved memory. that's also why sometimes you can load a model but OOM when you run it.
Finally the used~= (resrved + non-releasable) is technically what you'll observe on nvidia-smi
, the most external view of memory usage.
More in https://pytorch.org/docs/stable/generated/torch.cuda.memory_stats.html
IlyasMoutawwakil
changed discussion status to
closed