Bounded activation functions for enhanced training stability of deep neural networks on visual pattern recognition problems
This paper focuses on the enhancement of the generalization ability and training stability of deep neural networks (DNNs). New activation functions that we call bounded rectified linear unit (ReLU), bounded leaky ReLU, and bounded bi-firing are proposed. These activation functions are defined based...
Saved in:
Main Authors: | Liew, S. S., Khalil-Hani, M., Bakhteri, R. |
---|---|
Format: | Article |
Published: |
Elsevier B.V.
2016
|
Subjects: | |
Online Access: | http://eprints.utm.my/id/eprint/71140/ https://www.scopus.com/inward/record.uri?eid=2-s2.0-84994477344&doi=10.1016%2fj.neucom.2016.08.037&partnerID=40&md5=5b940413f14332dd63cda37f4ebfbe4b |
Tags: |
Add Tag
No Tags, Be the first to tag this record!
|
Similar Items
-
Bounded activation functions for enhanced training stability of deep neural networks on visual pattern recognition problems
by: Liew, S. S., et al.
Published: (2016) -
Distributed B-SDLM: accelerating the training convergence of deep neural networks through parallelism
by: Liew, S. S., et al.
Published: (2016) -
An optimized second order stochastic learning algorithm for neural network training
by: Liew, S. S., et al.
Published: (2016) -
An efficient and effective convolutional neural network for visual pattern recognition
by: Liew, Shan Sung
Published: (2016) -
Convolutional neural networks with fused layers applied to face recognition
by: Syafeeza, A. R., et al.
Published: (2015)