You can use many optimizers based on various factors, such as the learning rate, performance metric, dropout, gradient, and more.
Following are some of the popular optimizers :
* Adam
* AdaDelta
* AdaGrad
* RMSprop
* Momentum
* Stochastic Gradient Descent