Skip to content
This repository has been archived by the owner on Oct 31, 2023. It is now read-only.

Learning rate value causes Infinite/nan loss value #26

Open
seo-95 opened this issue Jan 20, 2021 · 1 comment
Open

Learning rate value causes Infinite/nan loss value #26

seo-95 opened this issue Jan 20, 2021 · 1 comment

Comments

@seo-95
Copy link

seo-95 commented Jan 20, 2021

Hi,
first of all, thank you for sharing the code to reproduce your results. I was trying to train the model from scratch on VG using your configuration R-50-updn.yaml and I have found a potential problem. With the learning rate of 0.02 you have in your configs (both base grids and base bottom-up) the training script produces infinite/nan loss values causing the training to diverge.
The training works after decreasing the learning rate to 0.002.
I was wondering if there is something I am missing.
Thank you!

@endernewton
Copy link
Contributor

Have you retried after it failed? sometimes the randomness can cause that.

Sign up for free to subscribe to this conversation on GitHub. Already have an account? Sign in.
Labels
None yet
Projects
None yet
Development

No branches or pull requests

2 participants