TY - GEN
T1 - Use of a sparse structure to improve learning performance of recurrent neural networks
AU - Awano, Hiromitsu
AU - Nishide, Shun
AU - Arie, Hiroaki
AU - Tani, Jun
AU - Takahashi, Toru
AU - Okuno, Hiroshi G.
AU - Ogata, Tetsuya
PY - 2011
Y1 - 2011
N2 - The objective of our study is to find out how a sparse structure affects the performance of a recurrent neural network (RNN). Only a few existing studies have dealt with the sparse structure of RNN with learning like Back Propagation Through Time (BPTT). In this paper, we propose a RNN with sparse connection and BPTT called Multiple time scale RNN (MTRNN). Then, we investigated how sparse connection affects generalization performance and noise robustness. In the experiments using data composed of alphabetic sequences, the MTRNN showed the best generalization performance when the connection rate was 40%. We also measured sparseness of neural activity and found out that sparseness of neural activity corresponds to generalization performance. These results means that sparse connection improved learning performance and sparseness of neural activity would be used as metrics of generalization performance.
AB - The objective of our study is to find out how a sparse structure affects the performance of a recurrent neural network (RNN). Only a few existing studies have dealt with the sparse structure of RNN with learning like Back Propagation Through Time (BPTT). In this paper, we propose a RNN with sparse connection and BPTT called Multiple time scale RNN (MTRNN). Then, we investigated how sparse connection affects generalization performance and noise robustness. In the experiments using data composed of alphabetic sequences, the MTRNN showed the best generalization performance when the connection rate was 40%. We also measured sparseness of neural activity and found out that sparseness of neural activity corresponds to generalization performance. These results means that sparse connection improved learning performance and sparseness of neural activity would be used as metrics of generalization performance.
KW - Recurrent Neural Networks
KW - Sparse Coding
KW - Sparse Structure
UR - http://www.scopus.com/inward/record.url?scp=81855227141&partnerID=8YFLogxK
UR - http://www.scopus.com/inward/citedby.url?scp=81855227141&partnerID=8YFLogxK
U2 - 10.1007/978-3-642-24965-5_36
DO - 10.1007/978-3-642-24965-5_36
M3 - Conference contribution
AN - SCOPUS:81855227141
SN - 9783642249648
T3 - Lecture Notes in Computer Science (including subseries Lecture Notes in Artificial Intelligence and Lecture Notes in Bioinformatics)
SP - 323
EP - 331
BT - Neural Information Processing - 18th International Conference, ICONIP 2011, Proceedings
T2 - 18th International Conference on Neural Information Processing, ICONIP 2011
Y2 - 13 November 2011 through 17 November 2011
ER -