Skip to main navigation Skip to search Skip to main content

Enhancing Deep Neural Network Convergence and Performance: A Hybrid Activation Function Approach by Combining ReLU and ELU Activation Function

  • Ritesh Maurya*
  • , Divyam Aggarwal
  • , T. Gopalakrishnan
  • , Nageshwar Nath Pandey
  • *Corresponding author for this work

Research output: Chapter in Book/Report/Conference proceedingConference contribution

Abstract

Activation functions play an important role in Deep Neural Networks. The activation function can learn nonlinearities present in the data; therefore, it can learn intricate patterns present in the data. Rectified Linear Unit (ReLU) is an activation function that helps in encountering the problem of vanishing gradient. However, it suffers from 'dying ReLU' problem for the negative values. Leaky ReLU can solve the problem of 'dying ReLU'; though it still suffers from a vanishing gradient problem due to the small gradient at for negative values, which results in slow convergence. Therefore, in this work, a combination of ReLU and Exponential Linear Unit (ELU) has been proposed considering the smoother convergence of the ELU activation function for the values on the negative side. Evaluating the effectiveness of the developed hybrid activation function compared to previous ReLU activation function versions such as SeLU, Leaky ReLU, ELU, etc. using a toy multi-layer perceptron and convolution neural network (CNN) model on the FashinMNIST and MNIST datasets. The improvement in the performance of these toy models when used with the proposed hybrid activation function on given datasets suggests the effectiveness of the proposed hybrid activation function.

Original languageEnglish
Title of host publicationProceedings of 2023 2nd International Conference on Informatics, ICI 2023
EditorsSandeep Kumar Singh, Vikas Saxena
PublisherInstitute of Electrical and Electronics Engineers Inc.
ISBN (Electronic)9798350343830
DOIs
Publication statusPublished - 2023
Event2nd International Conference on Informatics, ICI 2023 - Noida, India
Duration: 23-11-202325-11-2023

Publication series

NameProceedings of 2023 2nd International Conference on Informatics, ICI 2023

Conference

Conference2nd International Conference on Informatics, ICI 2023
Country/TerritoryIndia
CityNoida
Period23-11-2325-11-23

All Science Journal Classification (ASJC) codes

  • Artificial Intelligence
  • Computer Science Applications
  • Hardware and Architecture
  • Information Systems
  • Information Systems and Management

Fingerprint

Dive into the research topics of 'Enhancing Deep Neural Network Convergence and Performance: A Hybrid Activation Function Approach by Combining ReLU and ELU Activation Function'. Together they form a unique fingerprint.

Cite this