Skip to content

Relate alpha, beta1, beta2 and epsilon to learning rate and momentum in adam_sgd optimizer

Chris Basoglu edited this page Dec 21, 2016 · 4 revisions

Here is a mapping : • Alpha is the learning_rate • Beta1 is momentum parameter • Beta2 is variance_momentum parameter

I don’t think you’re able to define the epsilon one.

Clone this wiki locally