Cargando…
Deep Neural Networks with Multistate Activation Functions
We propose multistate activation functions (MSAFs) for deep neural networks (DNNs). These MSAFs are new kinds of activation functions which are capable of representing more than two states, including the N-order MSAFs and the symmetrical MSAF. DNNs with these MSAFs can be trained via conventional St...
Autores principales: | , , , |
---|---|
Formato: | Online Artículo Texto |
Lenguaje: | English |
Publicado: |
Hindawi Publishing Corporation
2015
|
Materias: | |
Acceso en línea: | https://www.ncbi.nlm.nih.gov/pmc/articles/PMC4581500/ https://www.ncbi.nlm.nih.gov/pubmed/26448739 http://dx.doi.org/10.1155/2015/721367 |
_version_ | 1782391575005888512 |
---|---|
author | Cai, Chenghao Xu, Yanyan Ke, Dengfeng Su, Kaile |
author_facet | Cai, Chenghao Xu, Yanyan Ke, Dengfeng Su, Kaile |
author_sort | Cai, Chenghao |
collection | PubMed |
description | We propose multistate activation functions (MSAFs) for deep neural networks (DNNs). These MSAFs are new kinds of activation functions which are capable of representing more than two states, including the N-order MSAFs and the symmetrical MSAF. DNNs with these MSAFs can be trained via conventional Stochastic Gradient Descent (SGD) as well as mean-normalised SGD. We also discuss how these MSAFs perform when used to resolve classification problems. Experimental results on the TIMIT corpus reveal that, on speech recognition tasks, DNNs with MSAFs perform better than the conventional DNNs, getting a relative improvement of 5.60% on phoneme error rates. Further experiments also reveal that mean-normalised SGD facilitates the training processes of DNNs with MSAFs, especially when being with large training sets. The models can also be directly trained without pretraining when the training set is sufficiently large, which results in a considerable relative improvement of 5.82% on word error rates. |
format | Online Article Text |
id | pubmed-4581500 |
institution | National Center for Biotechnology Information |
language | English |
publishDate | 2015 |
publisher | Hindawi Publishing Corporation |
record_format | MEDLINE/PubMed |
spelling | pubmed-45815002015-10-07 Deep Neural Networks with Multistate Activation Functions Cai, Chenghao Xu, Yanyan Ke, Dengfeng Su, Kaile Comput Intell Neurosci Research Article We propose multistate activation functions (MSAFs) for deep neural networks (DNNs). These MSAFs are new kinds of activation functions which are capable of representing more than two states, including the N-order MSAFs and the symmetrical MSAF. DNNs with these MSAFs can be trained via conventional Stochastic Gradient Descent (SGD) as well as mean-normalised SGD. We also discuss how these MSAFs perform when used to resolve classification problems. Experimental results on the TIMIT corpus reveal that, on speech recognition tasks, DNNs with MSAFs perform better than the conventional DNNs, getting a relative improvement of 5.60% on phoneme error rates. Further experiments also reveal that mean-normalised SGD facilitates the training processes of DNNs with MSAFs, especially when being with large training sets. The models can also be directly trained without pretraining when the training set is sufficiently large, which results in a considerable relative improvement of 5.82% on word error rates. Hindawi Publishing Corporation 2015 2015-09-10 /pmc/articles/PMC4581500/ /pubmed/26448739 http://dx.doi.org/10.1155/2015/721367 Text en Copyright © 2015 Chenghao Cai et al. https://creativecommons.org/licenses/by/3.0/ This is an open access article distributed under the Creative Commons Attribution License, which permits unrestricted use, distribution, and reproduction in any medium, provided the original work is properly cited. |
spellingShingle | Research Article Cai, Chenghao Xu, Yanyan Ke, Dengfeng Su, Kaile Deep Neural Networks with Multistate Activation Functions |
title | Deep Neural Networks with Multistate Activation Functions |
title_full | Deep Neural Networks with Multistate Activation Functions |
title_fullStr | Deep Neural Networks with Multistate Activation Functions |
title_full_unstemmed | Deep Neural Networks with Multistate Activation Functions |
title_short | Deep Neural Networks with Multistate Activation Functions |
title_sort | deep neural networks with multistate activation functions |
topic | Research Article |
url | https://www.ncbi.nlm.nih.gov/pmc/articles/PMC4581500/ https://www.ncbi.nlm.nih.gov/pubmed/26448739 http://dx.doi.org/10.1155/2015/721367 |
work_keys_str_mv | AT caichenghao deepneuralnetworkswithmultistateactivationfunctions AT xuyanyan deepneuralnetworkswithmultistateactivationfunctions AT kedengfeng deepneuralnetworkswithmultistateactivationfunctions AT sukaile deepneuralnetworkswithmultistateactivationfunctions |