About maximum GPU memory consumption

About maximum GPU memory consumption  

  By: ziyanhuang on July 9, 2022, 12:58 a.m.

Hi, It is observed that when inference first starts, there is usually a very short period of high memory usage in Pytorch.
For example, in the provided demo video for docker sanity check, the maximum GPU memory consumption is 2872 MiB when the inference starts. However, when the inference is stable, it keeps comsuming only 908 MiB GPU memory. So, I want to ask whether the definitin of maximum GPU memory consumption only consider the period when inference is stable and ignore the peak memory usage when the inference just starts?

Re: About maximum GPU memory consumption  

  By: junma on July 10, 2022, 6:15 p.m.

Yes! And the code is already availalbe on the evaluation page. You can find the details in the code.