To read this content please select one of the options below:

Web-aided data set expansion in deep learning: evaluating trainable activation functions in ResNet for improved image classification

Zhiqiang Zhang (College of Science, Zhejiang University of Science and Technology, Hangzhou, China )
Xiaoming Li (School of International Business, Zhejiang Yuexiu University of Foreign Languages, Shaoxing, China)
Xinyi Xu (School of International Business, Zhejiang Yuexiu University of Foreign Languages, Shaoxing, China)
Chengjie Lu (School of International Business, Zhejiang Yuexiu University of Foreign Languages, Shaoxing, China)
Yihe Yang (College of Science, Zhejiang University of Science and Technology, Hangzhou, China )
Zhiyong Shi (College of Science, Zhejiang University of Science and Technology, Hangzhou, China )

International Journal of Web Information Systems

ISSN: 1744-0084

Article publication date: 12 July 2024

Issue publication date: 19 July 2024

30

Abstract

Purpose

The purpose of this study is to explore the potential of trainable activation functions to enhance the performance of deep neural networks, specifically ResNet architectures, in the task of image classification. By introducing activation functions that adapt during training, the authors aim to determine whether such flexibility can lead to improved learning outcomes and generalization capabilities compared to static activation functions like ReLU. This research seeks to provide insights into how dynamic nonlinearities might influence deep learning models' efficiency and accuracy in handling complex image data sets.

Design/methodology/approach

This research integrates three novel trainable activation functions – CosLU, DELU and ReLUN – into various ResNet-n architectures, where “n” denotes the number of convolutional layers. Using CIFAR-10 and CIFAR-100 data sets, the authors conducted a comparative study to assess the impact of these functions on image classification accuracy. The approach included modifying the traditional ResNet models by replacing their static activation functions with the trainable variants, allowing for dynamic adaptation during training. The performance was evaluated based on accuracy metrics and loss profiles across different network depths.

Findings

The findings indicate that trainable activation functions, particularly CosLU, can significantly enhance the performance of deep learning models, outperforming the traditional ReLU in deeper network configurations on the CIFAR-10 data set. CosLU showed the highest improvement in accuracy, whereas DELU and ReLUN offered varying levels of performance enhancements. These functions also demonstrated potential in reducing overfitting and improving model generalization across more complex data sets like CIFAR-100, suggesting that the adaptability of activation functions plays a crucial role in the training dynamics of deep neural networks.

Originality/value

This study contributes to the field of deep learning by introducing and evaluating the impact of three novel trainable activation functions within widely used ResNet architectures. Unlike previous works that primarily focused on static activation functions, this research demonstrates that incorporating trainable nonlinearities can lead to significant improvements in model performance and adaptability. The introduction of CosLU, DELU and ReLUN provides a new pathway for enhancing the flexibility and efficiency of neural networks, potentially setting a new standard for future deep learning applications in image classification and beyond.

Keywords

Acknowledgements

The authors sincerely thank the National Natural Science Foundation of China (Grant No. 62272311) for funding this research. The authors also appreciate the valuable feedback and suggestions from the peer reviewers, as well as the support from their colleagues at their institution during the research process. Lastly, the authors are grateful to their families and friends for their understanding and encouragement throughout the entire study.

Funding: This study was funded by the National Science Foundation of China (Grant No. 62272311).

Citation

Zhang, Z., Li, X., Xu, X., Lu, C., Yang, Y. and Shi, Z. (2024), "Web-aided data set expansion in deep learning: evaluating trainable activation functions in ResNet for improved image classification", International Journal of Web Information Systems, Vol. 20 No. 4, pp. 452-469. https://doi.org/10.1108/IJWIS-05-2024-0135

Publisher

:

Emerald Publishing Limited

Copyright © 2023, Emerald Publishing Limited

Related articles