Audio-visual interaction in model adaptation for multi-modal speech recognition

Satoshi Tamura*, Masanao Oonishi, Satoru Hayamizu

*この研究の対応する著者

研究成果査読

4 被引用数 (Scopus)

抄録

This paper investigates audio-visual interaction, i.e. inter-modal influences, in linear-regressive model adaptation for multi-modal speech recognition. In the multi-modal adaptation, inter-modal information may contribute the performance of speech recognition. Thus the influence and advantage of intermodal elements should be examined. Experiments were conducted to evaluate several transformation matrices including or excluding inter-modal and intra-modal elements, using noisy data in an audio-visual corpus. From the experimental results, the importance of effective use of audio-visual interaction is clarified.

本文言語English
ページ875-878
ページ数4
出版ステータスPublished - 2011
外部発表はい
イベントAsia-Pacific Signal and Information Processing Association Annual Summit and Conference 2011, APSIPA ASC 2011 - Xi'an, China
継続期間: 2011 10 182011 10 21

Conference

ConferenceAsia-Pacific Signal and Information Processing Association Annual Summit and Conference 2011, APSIPA ASC 2011
国/地域China
CityXi'an
Period11/10/1811/10/21

ASJC Scopus subject areas

  • 情報システム
  • 信号処理

フィンガープリント

「Audio-visual interaction in model adaptation for multi-modal speech recognition」の研究トピックを掘り下げます。これらがまとまってユニークなフィンガープリントを構成します。

引用スタイル