ADAM RMSPROP ADAGRAD.

for i in range(nb_epochs):
  params_grad = evaluate_gradient(loss_function, data, params)
  params = params - learning_rate * params_grad
Famous Fly