EconPapers    
Economics at your fingertips  
 

On the Synergy of Optimizers and Activation Functions: A CNN Benchmarking Study

Khuraman Aziz Sayın, Necla Kırcalı Gürsoy, Türkay Yolcu () and Arif Gürsoy
Additional contact information
Khuraman Aziz Sayın: Department of Mathematics, Ege University, Bornova, Izmir 35040, Türkiye
Necla Kırcalı Gürsoy: Department of Computer Programming, Ege University, Bornova, Izmir 35040, Türkiye
Türkay Yolcu: Department of Mathematics, Bradley University, Peoria, IL 61625, USA
Arif Gürsoy: Department of Mathematics, Ege University, Bornova, Izmir 35040, Türkiye

Mathematics, 2025, vol. 13, issue 13, 1-36

Abstract: In this study, we present a comparative analysis of gradient descent-based optimizers frequently used in Convolutional Neural Networks (CNNs), including SGD, mSGD, RMSprop, Adadelta, Nadam, Adamax, Adam, and the recent EVE optimizer. To explore the interaction between optimization strategies and activation functions, we systematically evaluate all combinations of these optimizers with four activation functions—ReLU, LeakyReLU, Tanh, and GELU—across three benchmark image classification datasets: CIFAR-10, Fashion-MNIST (F-MNIST), and Labeled Faces in the Wild (LFW). Each configuration was assessed using multiple evaluation metrics, including accuracy, precision, recall, F1-score, mean absolute error (MAE), and mean squared error (MSE). All experiments were performed using k -fold cross-validation to ensure statistical robustness. Additionally, two-way ANOVA was employed to validate the significance of differences across optimizer–activation combinations. This study aims to highlight the importance of jointly selecting optimizers and activation functions to enhance training dynamics and generalization in CNNs. We also consider the role of critical hyperparameters, such as learning rate and regularization methods, in influencing optimization stability. This work provides valuable insights into the optimizer–activation interplay and offers practical guidance for improving architectural and hyperparameter configurations in CNN-based deep learning models.

Keywords: Stochastic Gradient Descent; optimization; convolutional neural network; activation functions; image processing (search for similar items in EconPapers)
JEL-codes: C (search for similar items in EconPapers)
Date: 2025
References: Add references at CitEc
Citations:

Downloads: (external link)
https://www.mdpi.com/2227-7390/13/13/2088/pdf (application/pdf)
https://www.mdpi.com/2227-7390/13/13/2088/ (text/html)

Related works:
This item may be available elsewhere in EconPapers: Search for items with the same title.

Export reference: BibTeX RIS (EndNote, ProCite, RefMan) HTML/Text

Persistent link: https://EconPapers.repec.org/RePEc:gam:jmathe:v:13:y:2025:i:13:p:2088-:d:1687106

Access Statistics for this article

Mathematics is currently edited by Ms. Emma He

More articles in Mathematics from MDPI
Bibliographic data for series maintained by MDPI Indexing Manager ().

 
Page updated 2025-06-26
Handle: RePEc:gam:jmathe:v:13:y:2025:i:13:p:2088-:d:1687106