'Looking for a solution for "CUDA out of memory" using pytorch

"CUDA out memory". I just want to know that is there any way using pytorch to not run out of CUDA memory without reducing any parameters like reducing batch size. btw, I'm working on a 6 GB NVIDIA RTX 3060.



Sources

This article follows the attribution requirements of Stack Overflow and is licensed under CC BY-SA 3.0.

Source: Stack Overflow

Solution Source