-
Notifications
You must be signed in to change notification settings - Fork 146
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
Met with cuda out of memory when I am testing the model #33
Comments
try reducing the batch size. |
i met the same error(cuda oom) when test rdn-liif. |
I have 4 GPU. But when I run code on GPU:2 ,I run this 'CUDA_VISIBLE_DEVICES=2 python demo.py --input 246.jpg --model rdn-liif.pth --resolution 2560,1440 --output output.png --gpu 0' and get 'cuda out of memory'. Than I try 'CUDA_VISIBLE_DEVICES=2 python demo.py --input 246.jpg --model rdn-liif.pth --resolution 1440,2560 --output output.png --gpu 2'. It work!!! |
I tried to reduce the batch size to 1, but it didn't work. |
I set the number of GPU to '1', but it still used GPU'0'. |
bash scripts/test-div2k.sh save/_train_edsr-baseline-liif/epoch-last.pth 0
then I met with RuntimeError: CUDA out of memory. Tried to allocate 676.00 MiB (GPU 0; 10.92 GiB total capacity; 827.98 MiB already allocated; 478.50 MiB free; 830.00 MiB reserved in total by PyTorch)
how to solve this problem?
The text was updated successfully, but these errors were encountered: