-
Notifications
You must be signed in to change notification settings - Fork 1.9k
OnlineGradientDescent crash #4363
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
Comments
Does this still happen if you set DecreaseLearningRate to |
Constant learning rates work well in many cases. In AutoML, we sweep over both choices of true/false. My top guess is that I'd recommend building a meta-model to predict failures given a position in the hyperparameter space. Then AutoML can avoid areas that generally throw an error. It would be nice to attach a debugger and see what the trainer is doing. I'm guessing the model weights are oscillating and growing towards +/- Infinity. Though, it's possible there's a bug in the weight update code leading to the infinity. Stepping thru the trainer's weight update code will tell us. |
Hi @daholste , has your question been addressed by @justinormont 's comments ? If so, please feel free to close the issue. |
Closing this issue as the issue seems to be resolved. |
System information
Issue
I ran the script:
(I can provide the data as requested)
I get the exception
Successful training
The text was updated successfully, but these errors were encountered: