TY - GEN
T1 - Vocal imitation using physical vocal tract model
AU - Kanda, Hisashi
AU - Ogata, Tetsuya
AU - Komatani, Kazunori
AU - Okuno, Hiroshi G.
PY - 2007/12/1
Y1 - 2007/12/1
N2 - A vocal imitation system was developed using a computational model that supports the motor theory of speech perception. A critical problem in vocal imitation is how to generate speech sounds produced by adults, whose vocal tracts have physical properties (i.e., articulatory motions) differing from those of infants' vocal tracts. To solve this problem, a model based on the motor theory of speech perception, was constructed. This model suggests that infants simulate the speech generation by estimating their own articulatory motions in order to interpret the speech sounds of adults. Applying this model enables the vocal imitation system to estimate articulatory motions for unexperienced speech sounds that have not actually been generated by the system. The system was implemented by using Recurrent Neural Network with Parametric Bias (RNNPB) and a physical vocal tract model, called the Maeda model. Experimental results demonstrated that the system was sufficiently robust with respect to individual differences in speech sounds and could imitate unexperienced vowel sounds.
AB - A vocal imitation system was developed using a computational model that supports the motor theory of speech perception. A critical problem in vocal imitation is how to generate speech sounds produced by adults, whose vocal tracts have physical properties (i.e., articulatory motions) differing from those of infants' vocal tracts. To solve this problem, a model based on the motor theory of speech perception, was constructed. This model suggests that infants simulate the speech generation by estimating their own articulatory motions in order to interpret the speech sounds of adults. Applying this model enables the vocal imitation system to estimate articulatory motions for unexperienced speech sounds that have not actually been generated by the system. The system was implemented by using Recurrent Neural Network with Parametric Bias (RNNPB) and a physical vocal tract model, called the Maeda model. Experimental results demonstrated that the system was sufficiently robust with respect to individual differences in speech sounds and could imitate unexperienced vowel sounds.
UR - http://www.scopus.com/inward/record.url?scp=51349149903&partnerID=8YFLogxK
UR - http://www.scopus.com/inward/citedby.url?scp=51349149903&partnerID=8YFLogxK
U2 - 10.1109/IROS.2007.4399137
DO - 10.1109/IROS.2007.4399137
M3 - Conference contribution
AN - SCOPUS:51349149903
SN - 1424409128
SN - 9781424409129
T3 - IEEE International Conference on Intelligent Robots and Systems
SP - 1846
EP - 1851
BT - Proceedings of the 2007 IEEE/RSJ International Conference on Intelligent Robots and Systems, IROS 2007
T2 - 2007 IEEE/RSJ International Conference on Intelligent Robots and Systems, IROS 2007
Y2 - 29 October 2007 through 2 November 2007
ER -