z-logo
open-access-imgOpen Access
Comparison of Effect of learning rate of Neural Network Performance in Deep learning neural networks using the stochastic gradient descent algorithm
Author(s) -
M. L. Sharma
Publication year - 2022
Publication title -
indian scientific journal of research in engineering and management
Language(s) - English
Resource type - Journals
ISSN - 2582-3930
DOI - 10.55041/ijsrem11651
Subject(s) - artificial intelligence , computer science , stochastic gradient descent , deep learning , artificial neural network , machine learning , gradient descent , test data , algorithm , programming language
The stochastic gradient descent optimization approach is used to train deep learning neural networks in this paper. Artificial neural networks are a subfield of deep learning that uses algorithms inspired by the structure and function of the brain. Deep learning systems are designed to learn feature hierarchies based on the composition of lower level characteristics at the top of the hierarchy. The various sorts of learning models are also discussed. We build a train and dataset with different samples (500, 2000, and 4000) and change the Epochs value in this research (100, 300 and 400).We also change the learning rate for different result. We Test the accuracy of learning rates. Classification accuracy on the training dataset is marked in blue, whereas accuracy on the test dataset is marked in orange. In This paper we find the best learning rate for good performance on the train and test sets. Keywords: Deep Learning, Epochs, Learning rate, Optimization

The content you want is available to Zendy users.

Already have an account? Click here to sign in.
Having issues? You can contact us here