An efficient semi-sigmoidal non-linear activation function approach for deep neural networks
A non-linear activation function is one of the key contributing factors to the success of Deep Learning (DL). Since the revival of DL takes place in 2012, Rectified Linear Unit (ReLU) has been regarded as a de facto standard for many DL models by the community. Despite its popularity, however, Re...
Saved in:
Main Author: | Chieng, Hock Hung |
---|---|
Format: | Thesis |
Language: | English English English |
Published: |
2022
|
Subjects: | |
Online Access: | http://eprints.uthm.edu.my/8409/1/24p%20CHIENG%20HOCK%20HUNG.pdf http://eprints.uthm.edu.my/8409/2/CHIENG%20HOCK%20HUNG%20COPYRIGHT%20DECLARATION.pdf http://eprints.uthm.edu.my/8409/3/CHIENG%20HOCK%20HUNG%20WATERMARK.pdf http://eprints.uthm.edu.my/8409/ |
Tags: |
Add Tag
No Tags, Be the first to tag this record!
|
Similar Items
-
Review of the Sigmoid Function Implementation into Digital Hardware
by: Syahrulanuar, Ngah, et al.
Published: (2015) -
Sigmoid function implementation using the unequal segmentation of differential lookup table and second order nonlinear function
by: Syahrulanuar, Ngah, et al.
Published: (2017) -
Quality function model based on voice of customer using neural networks and statistical approaches
by: Siraj, Fadzilah, et al.
Published: (2011) -
Evaluation of knowledge encoded in deep neural network in the imputation of gene expression
by: Chow, Jenn Pang
Published: (2018) -
A Review of Deep Convolutional Neural Networks in Mobile Face Recognition
by: Jing, Chi, et al.
Published: (2023)