International Journal For Multidisciplinary Research

E-ISSN: 2582-2160     Impact Factor: 9.24

A Widely Indexed Open Access Peer Reviewed Multidisciplinary Bi-monthly Scholarly International Journal

Call for Paper Volume 7, Issue 6 (November-December 2025) Submit your research before last 3 days of December to publish your research paper in the issue of November-December.

Addressing Neural Network Training Challenges through Optimizers and Activation Functions

Author(s) Ms. Madhuri Nitin Ghadge, Ms. Rohini Kapse
Country India
Abstract Training deep neural networks is often hindered by problems such as vanishing gradients (Sigmoid), neuronal instability (ReLU), slow convergence (SGDM), and poor generalization (Adam). This research systematically investigates how the choice of optimizers and activation functions mitigates these challenges. Using a Convolutional Neural Network (CNN) trained on the MNIST dataset, five activation functions (Sigmoid, ReLU, LeakyReLU, GeLU, SiLU) and four optimizers (SGDM, Adam, AdamW, RMSProp) were compared in 20 unique configurations. Performance was evaluated on the basis of validation accuracy, convergence speed, and gradient stability.
Keywords Deep Learning, Convolutional Neural Networks (CNN), Activation Functions, Optimizers, Gradient Vanishing Problem, Convergence Speed, MNIST Dataset
Field Computer > Artificial Intelligence / Simulation / Virtual Reality
Published In Volume 7, Issue 6, November-December 2025
Published On 2025-11-20
DOI https://doi.org/10.36948/ijfmr.2025.v07i06.61158
Short DOI https://doi.org/hbbz53

Share this