Open Access iconOpen Access

ARTICLE

Performance Enhancement of Adaptive Neural Networks Based on Learning Rate

Swaleha Zubair1, Anjani Kumar Singha1, Nitish Pathak2, Neelam Sharma3, Shabana Urooj4,*, Samia Rabeh Larguech4

1 Department of Computer Science, Aligarh Muslim University, Aligarh, 202002, Uttar Pradesh, India
2 Department of Information Technology, Bhagwan Parshuram Institute of Technology (BPIT), GGSIPU, New Delhi, 110089, India
3 Department of Computer Science and Engineering, Maharaja Agrasen Institute of Technology (MAIT), GGSIPU, New Delhi, India
4 Department of Electrical Engineering, College of Engineering, Princess Nourah bint Abdulrahman University, P.O.Box 84428, Riyadh, 11671, Saudi Arabia

* Corresponding Author: Shabana Urooj. Email: email

Computers, Materials & Continua 2023, 74(1), 2005-2019. https://doi.org/10.32604/cmc.2023.031481

Abstract

Deep learning is the process of determining parameters that reduce the cost function derived from the dataset. The optimization in neural networks at the time is known as the optimal parameters. To solve optimization, it initialize the parameters during the optimization process. There should be no variation in the cost function parameters at the global minimum. The momentum technique is a parameters optimization approach; however, it has difficulties stopping the parameter when the cost function value fulfills the global minimum (non-stop problem). Moreover, existing approaches use techniques; the learning rate is reduced during the iteration period. These techniques are monotonically reducing at a steady rate over time; our goal is to make the learning rate parameters. We present a method for determining the best parameters that adjust the learning rate in response to the cost function value. As a result, after the cost function has been optimized, the process of the rate Schedule is complete. This approach is shown to ensure convergence to the optimal parameters. This indicates that our strategy minimizes the cost function (or effective learning). The momentum approach is used in the proposed method. To solve the Momentum approach non-stop problem, we use the cost function of the parameter in our proposed method. As a result, this learning technique reduces the quantity of the parameter due to the impact of the cost function parameter. To verify that the learning works to test the strategy, we employed proof of convergence and empirical tests using current methods and the results are obtained using Python.

Keywords


Cite This Article

APA Style
Zubair, S., Singha, A.K., Pathak, N., Sharma, N., Urooj, S. et al. (2023). Performance enhancement of adaptive neural networks based on learning rate. Computers, Materials & Continua, 74(1), 2005-2019. https://doi.org/10.32604/cmc.2023.031481
Vancouver Style
Zubair S, Singha AK, Pathak N, Sharma N, Urooj S, Larguech SR. Performance enhancement of adaptive neural networks based on learning rate. Comput Mater Contin. 2023;74(1):2005-2019 https://doi.org/10.32604/cmc.2023.031481
IEEE Style
S. Zubair, A.K. Singha, N. Pathak, N. Sharma, S. Urooj, and S.R. Larguech, “Performance Enhancement of Adaptive Neural Networks Based on Learning Rate,” Comput. Mater. Contin., vol. 74, no. 1, pp. 2005-2019, 2023. https://doi.org/10.32604/cmc.2023.031481



cc Copyright © 2023 The Author(s). Published by Tech Science Press.
This work is licensed under a Creative Commons Attribution 4.0 International License , which permits unrestricted use, distribution, and reproduction in any medium, provided the original work is properly cited.
  • 1132

    View

  • 556

    Download

  • 0

    Like

Share Link