Estimation of Articulatory Movements From Speech Acoustics Using an HMM-Based Speech Production Model

Sadao Hiroya*, Masaaki Honda

*Corresponding author for this work

Research output: Contribution to journalArticlepeer-review

115 Citations (Scopus)

Abstract

We present a method that determines articulatory movements from speech acoustics using a Hidden Markov Model (HMM)-based speech production model. The model statistically generates speech spectrum and articulatory parameters from a given phonemic string. It consists of HMMs of articulatory parameters for each phoneme and an articulatory-to-acoustic mapping for each HMM state. For a given speech spectrum, maximum a posteriori estimation of the articulatory parameters of the statistical model is presented. The performance on sentences was evaluated by comparing the estimated articulatory parameters with the observed parameters. The average RMS errors of the estimated articulatory parameters were 1.50 mm from the speech acoustics and the phonemic information in an utterance and 1.73 mm from the speech acoustics only.

Original languageEnglish
Pages (from-to)175-185
Number of pages11
JournalIEEE Transactions on Speech and Audio Processing
Volume12
Issue number2
DOIs
Publication statusPublished - 2004 Mar
Externally publishedYes

Keywords

  • Articulatory HMM
  • Articulatory-to-acoustic mapping
  • HMM-based speech production model
  • Speech inversion

ASJC Scopus subject areas

  • Software
  • Acoustics and Ultrasonics
  • Computer Vision and Pattern Recognition
  • Electrical and Electronic Engineering

Fingerprint

Dive into the research topics of 'Estimation of Articulatory Movements From Speech Acoustics Using an HMM-Based Speech Production Model'. Together they form a unique fingerprint.

Cite this