Intelligent facial image coding driven by speech and phoneme

Shigeo Morishima*, Kiyoharu Aizawa, Hiroshi Harashima

*この研究の対応する著者

研究成果: Conference article査読

43 被引用数 (Scopus)

抄録

The authors propose and compare two types of model-based facial motion coding schemes, i.e., synthesis by rules and synthesis by parameters. In synthesis by rules, facial motion images are synthesized on the basis of rules extracted by analysis of training image samples that include all of the phonemes and coarticulation. This system can be utilized as an automatic facial animation synthesizer from text input or as a man-machine interface using the facial motion image. In synthesis by parameters, facial motion images are synthesized on the basis of a code word index of speech parameters. Experimental results indicate good performance for both systems, which can create natural facial-motion images with very low transmission rate. Details of 3-D modeling, algorithm synthesis, and performance are discussed.

本文言語English
ページ(範囲)1795-1798
ページ数4
ジャーナルICASSP, IEEE International Conference on Acoustics, Speech and Signal Processing - Proceedings
3
出版ステータスPublished - 1989 12月 1
外部発表はい
イベント1989 International Conference on Acoustics, Speech, and Signal Processing - Glasgow, Scotland
継続期間: 1989 5月 231989 5月 26

ASJC Scopus subject areas

  • ソフトウェア
  • 信号処理
  • 電子工学および電気工学

フィンガープリント

「Intelligent facial image coding driven by speech and phoneme」の研究トピックを掘り下げます。これらがまとまってユニークなフィンガープリントを構成します。

引用スタイル