Emotional speech synthesis by sensing affective information from text

Mostafa Al Masum Shaikh, Antonio Rui Ferreira Rebordao, Keikichi Hirose, Mitsuru Ishizuka

研究成果: Conference contribution

3 被引用数 (Scopus)

抄録

Speech can express subjective meanings and intents that, in order to be fully understood, rely heavily in its affective perception. Some Text-to-Speech (TTS) systems reveal weaknesses in their emotional expressivity but this situation can be improved by a better parametrization of the acoustic and prosodic parameters. This paper describes an approach for better emotional expressivity in a speech synthesizer. Our technique uses several linguistic resources that can recognize emotions in a text and assigns appropriate parameters to the synthesizer to carry out a suitable speech synthesis. For evaluation purposes we considered the MARY TTS system to readout "happy" and "sad" news. The preliminary perceptual test results are encouraging and human judges, by listening to the synthesized speech obtained with our approach, could perceive "happy" emotions much better than compared to when they listened nonaffective synthesized speech.

本文言語English
ホスト出版物のタイトルProceedings - 2009 3rd International Conference on Affective Computing and Intelligent Interaction and Workshops, ACII 2009
DOI
出版ステータスPublished - 2009
外部発表はい
イベント2009 3rd International Conference on Affective Computing and Intelligent Interaction and Workshops, ACII 2009 - Amsterdam
継続期間: 2009 9月 102009 9月 12

Other

Other2009 3rd International Conference on Affective Computing and Intelligent Interaction and Workshops, ACII 2009
CityAmsterdam
Period09/9/1009/9/12

ASJC Scopus subject areas

  • 人工知能
  • コンピュータ ビジョンおよびパターン認識
  • 人間とコンピュータの相互作用
  • ソフトウェア

フィンガープリント

「Emotional speech synthesis by sensing affective information from text」の研究トピックを掘り下げます。これらがまとまってユニークなフィンガープリントを構成します。

引用スタイル