TY - JOUR
T1 - Emergence of hierarchical structure mirroring linguistic composition in a recurrent neural network
AU - Hinoshita, Wataru
AU - Arie, Hiroaki
AU - Tani, Jun
AU - Okuno, Hiroshi G.
AU - Ogata, Tetsuya
N1 - Funding Information:
This research was partially supported by a Grant-in-Aid for Scientific Research (B) 21300076 , Scientific Research (S) 19100003 , Creative Scientific Research 19GS0208 , and Global COE .
PY - 2011/5
Y1 - 2011/5
N2 - We show that a Multiple Timescale Recurrent Neural Network (MTRNN) can acquire the capabilities to recognize, generate, and correct sentences by self-organizing in a way that mirrors the hierarchical structure of sentences: characters grouped into words, and words into sentences. The model can control which sentence to generate depending on its initial states (generation phase) and the initial states can be calculated from the target sentence (recognition phase). In an experiment, we trained our model over a set of unannotated sentences from an artificial language, represented as sequences of characters. Once trained, the model could recognize and generate grammatical sentences, even if they were not learned. Moreover, we found that our model could correct a few substitution errors in a sentence, and the correction performance was improved by adding the errors to the training sentences in each training iteration with a certain probability. An analysis of the neural activations in our model revealed that the MTRNN had self-organized, reflecting the hierarchical linguistic structure by taking advantage of the differences in timescale among its neurons: in particular, neurons that change the fastest represented "characters", those that change more slowly, "words", and those that change the slowest, "sentences".
AB - We show that a Multiple Timescale Recurrent Neural Network (MTRNN) can acquire the capabilities to recognize, generate, and correct sentences by self-organizing in a way that mirrors the hierarchical structure of sentences: characters grouped into words, and words into sentences. The model can control which sentence to generate depending on its initial states (generation phase) and the initial states can be calculated from the target sentence (recognition phase). In an experiment, we trained our model over a set of unannotated sentences from an artificial language, represented as sequences of characters. Once trained, the model could recognize and generate grammatical sentences, even if they were not learned. Moreover, we found that our model could correct a few substitution errors in a sentence, and the correction performance was improved by adding the errors to the training sentences in each training iteration with a certain probability. An analysis of the neural activations in our model revealed that the MTRNN had self-organized, reflecting the hierarchical linguistic structure by taking advantage of the differences in timescale among its neurons: in particular, neurons that change the fastest represented "characters", those that change more slowly, "words", and those that change the slowest, "sentences".
KW - Hierarchical linguistic structure
KW - Language acquisition
KW - Multiple timescale recurrent neural network
KW - Self-organization
UR - http://www.scopus.com/inward/record.url?scp=79951944498&partnerID=8YFLogxK
UR - http://www.scopus.com/inward/citedby.url?scp=79951944498&partnerID=8YFLogxK
U2 - 10.1016/j.neunet.2010.12.006
DO - 10.1016/j.neunet.2010.12.006
M3 - Article
C2 - 21273043
AN - SCOPUS:79951944498
SN - 0893-6080
VL - 24
SP - 311
EP - 320
JO - Neural Networks
JF - Neural Networks
IS - 4
ER -