Bounded activation functions for enhanced training stability of deep neural networks on visual pattern recognition problems

This paper focuses on the enhancement of the generalization ability and training stability of deep neural networks (DNNs). New activation functions that we call bounded rectified linear unit (ReLU), bounded leaky ReLU, and bounded bi-firing are proposed. These activation functions are defined based...

Full description

Saved in:
Bibliographic Details
Main Authors: Liew, S. S., Khalil-Hani, M., Bakhteri, R.
Format: Article
Published: Elsevier B.V. 2016
Subjects:
Online Access:http://eprints.utm.my/id/eprint/71140/
https://www.scopus.com/inward/record.uri?eid=2-s2.0-84994477344&doi=10.1016%2fj.neucom.2016.08.037&partnerID=40&md5=5b940413f14332dd63cda37f4ebfbe4b
Tags: Add Tag
No Tags, Be the first to tag this record!
Be the first to leave a comment!
You must be logged in first