Neural Network training with Adam optimizer from scratch

Full code for training and testing of a simple neural network on the MNIST data set for digit recognition.
296 Descargas
Actualizado 14 abr 2021

Ver licencia

Full code for training and testing of a simple neural network on the MNIST data set for recognition of single digits between 0 and 9 (Accuracy around 98 %). Everything is implemented from scratch, including the Adam optimizer. Make sure all the files are in your current folder and run "train.m".

Check out http://neuralnetworksanddeeplearning.com/index.html to learn about the theory of neural networks and https://arxiv.org/abs/1412.6980 to understand the Adam optimizer!

Citar como

Johannes Langelaar (2026). Neural Network training with Adam optimizer from scratch (https://es.mathworks.com/matlabcentral/fileexchange/90461-neural-network-training-with-adam-optimizer-from-scratch), MATLAB Central File Exchange. Recuperado .

Compatibilidad con la versión de MATLAB
Se creó con R2019a
Compatible con cualquier versión
Compatibilidad con las plataformas
Windows macOS Linux
Versión Publicado Notas de la versión
1.0.0