Awesome Optimizers Save

Neural Network optimizers implemented from scratch in numpy (Adam, Adadelta, RMSProp, SGD, etc.)

Project README

Awesome-optimizers - Deep Learning optimizers study and implementations in numpy

To see the detailed study, please open the notebook file: Report.ipynb

Introduction and objectives

The objective of this repository is to summarize the study I did for deeply understanding the most common neural networks oprimizers. In it, you will find implementations from scratch for the following gradient-based optimization algorithms.

  • Gradient Descent
  • Gradient Descent with Momentum
  • Gradient Descent with Nesterov Momentum
  • AdaGrad
  • RMSProp
  • AdaDelta
  • Adam
  • AdaMax
  • NAdam

These implementations have been done using numpy and autograd.

Getting started

If you are interested in reproducing the results of the study, please clone the repository, open the Report.ipynb notebook and follow it sequentially. I encourage you to try to experiment with a new objective function and to play with the parameters of the algorithms.

You will need to have installed the following libraries.

  • numpy
  • autograd
  • matplotlib

Results

In the following charts you can see how the implemented algorithms behave when applied in the log-Beale objective function.

error_evolution

animation_trajectories

Further work and contribution opportunities

There are several ways of continuing this study.

  • Research and implement state-of-the-art optimization algorithms
  • Implement more objective functions and show which are the potential benefits and disadvantages of each of the algorithms in each objective function
  • Try to understand better why algorithms like RMSProp show an oscilation when they converge

If you want to continue with it, please, consider sending a pull request, I'll be more than happy to merge your changes.

License

This project has been licensed under MIT agreement. Please, read the LICENSE file for further details. Copyright (c) 2018 Iván Vallés Pérez

Open Source Agenda is not affiliated with "Awesome Optimizers" Project. README Source: ivallesp/awesome-optimizers
Stars
45
Open Issues
0
Last Commit
4 years ago
License
MIT

Open Source Agenda Badge

Open Source Agenda Rating