My simple comparison of the Adam, Adadelta and SGD optimizers used in a CNN based MNIST classifier.
Visualizing Accuracy and Loss *:
- SGD OPTIMIZER
- ADAM OPTIMIZER
- ADADELTA OPTIMIZER
* Tuning the right hyperparams and altering the number of epochs and batch size could improve results vastly.





