A MULTIMODAL DATABASE OF GESTURES AND SPEECH

Satoru Hayamizu*, Shigeki Nagaya, Keiko Watanuki, Masayuki Nakazawa, Shuichi Nobe, Takashi Yoshimura

*この研究の対応する著者

研究成果: Paper査読

2 被引用数 (Scopus)

抄録

This paper describes a multimodal database which consists of image data of human gestures and corresponding speech data for the research on multimodal interaction systems. The purpose of this database is to provide an underlying foundation for research and development of multimodal interactive systems. Our primary concern in selecting utterances and gestures for inclusion in the database was to ascertain the kinds of expressions and gestures that artificial systems could produce and recognize. Total 25 kinds of gestures and speech were repeated four times for the recording of each subject. The speech and gestures for a total of 48 subjects were recorded, converted into files and in the first version, the files for 12 subjects were recorded on CD-ROMs.

本文言語English
ページ2247-2250
ページ数4
出版ステータスPublished - 1999
外部発表はい
イベント6th European Conference on Speech Communication and Technology, EUROSPEECH 1999 - Budapest, Hungary
継続期間: 1999 9月 51999 9月 9

Conference

Conference6th European Conference on Speech Communication and Technology, EUROSPEECH 1999
国/地域Hungary
CityBudapest
Period99/9/599/9/9

ASJC Scopus subject areas

  • コンピュータ サイエンスの応用
  • ソフトウェア
  • 言語学および言語
  • 通信

フィンガープリント

「A MULTIMODAL DATABASE OF GESTURES AND SPEECH」の研究トピックを掘り下げます。これらがまとまってユニークなフィンガープリントを構成します。

引用スタイル