-
Notifications
You must be signed in to change notification settings - Fork 475
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
Update TrainConfig
optimizer hyperparameters
#82
Conversation
@herbiebradley since this is for your issue. |
There was a problem hiding this comment.
Choose a reason for hiding this comment
The reason will be displayed to describe this comment to others. Learn more.
Looks good! Can we change the default epsilon to 1.0e-7
(default in Tensorflow) since the ablation paper found it was significantly better?
Also, does anyone know where the weight decay of 1.0e-6
comes from, since it is significantly lower than the default?
|
There was a problem hiding this comment.
Choose a reason for hiding this comment
The reason will be displayed to describe this comment to others. Learn more.
lgtm!
@herbiebradley I intentionally left the default |
cc #76 for awareness |
Thanks for tagging the issue @jon-tow. |
This PR provides the following updates based on observations from issue #53:
weight_decay
TrainConfig parameter toAccelerateRLModel
optimizer initialization.opt_eps
TrainConfig parameter to theAccelerateRLModel
to allow users to specify an optimizer's epsilon value.learning_rate_init
andlearning_rate_target
tolr_init
andlr_target
TrainConfigs, respectively, for consistency with learning rate naming conventions intorch.optim
algorithms.lr_ramp_steps
andlr_decay_steps
TrainConfig parameters which look to be leftovers from the magicCARP lr scheduler.