학술논문

Deeper Bidirectional Neural Networks with Generalized Non-Vanishing Hidden Neurons
Document Type
Conference
Source
2022 21st IEEE International Conference on Machine Learning and Applications (ICMLA) ICMLA Machine Learning and Applications (ICMLA), 2022 21st IEEE International Conference on. :69-76 Dec, 2022
Subject
Computing and Processing
Engineering Profession
Robotics and Control Systems
Signal Processing and Analysis
Backpropagation
Training
Neurons
Machine learning
Delays
Biological neural networks
Logistics
Bidirectional backpropagation
Logistic
ReLU
NoVa hidden neurons
vanishing gradient
Language
Abstract
The new NoVa hidden neurons have outperformed ReLU hidden neurons in deep classifiers on some large image test sets. The NoVa or nonvanishing logistic neuron additively perturbs the sigmoidal activation function so that its derivative is not zero. This helps avoid or delay the problem of vanishing gradients. We here extend the NoVa to the generalized perturbed logistic neuron and compare it to ReLU and several other hidden neurons on large image test sets that include CIFAR-100 and Caltech-256. Generalized NoVa classifiers allow deeper networks with better classification on the large datasets. This deep benefit holds for ordinary unidirectional backpropagation. It also holds for the more efficient bidirectional backpropagation that trains in both the forward and backward directions.