-
Notifications
You must be signed in to change notification settings - Fork 56
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
what does the training loss curve look like #27
Comments
Sorry that I don't have any saved logs or plots for this. I might not find time soon to re-train to produce these plots. |
Would it be ok if I instead ask the following two questions?
|
If I remember correctly, Adam optimization results in a training curve with ups and downs. I do not remember how much ups and downs the loss is fluctuating. |
@coarsesand In my experiments, the position loss will increase along with iterations, and the reconstruction loss will decrease gradually. |
@CYang0515 does decrease in reconstruction loss eventually overpower the increase in position loss, resulting in a decrease in the overall combined loss with increasing number of iterations? |
I'm trying to train SSN via train_ssn.py, but after running ~40,000 iterations there seems to be a lot of jittering but no meaningful decrease in the training loss. I know from reading previous issues that convergence takes ~500,000K iterations, but with my computing resources it would take a few days to reach convergence.
So I was wondering whether the authors could kindly tell me / show me what the training loss curve looks like as a function of iteration number, starting from iteration 0 all the way to convergence.
Thank you in advance.
The text was updated successfully, but these errors were encountered: