Insertion-Based modeling for end-to-end automatic speech recognition

Yuya Fujita, Shinji Watanabe, Motoi Omachi, Xuankai Chang

Research output: Contribution to journalArticlepeer-review

Abstract

End-to-end (E2E) models have gained attention in the research field of automatic speech recognition (ASR). Many E2E models proposed so far assume left-to-right autoregressive generation of an output token sequence except for connectionist temporal classification (CTC) and its variants. However, left-to-right decoding cannot consider the future output context, and it is not always optimal for ASR. One of the non-left-to-right models is known as non-autoregressive Transformer (NAT) and has been intensively investigated in the area of neural machine translation (NMT) research. One NAT model, mask-predict, has been applied to ASR but the model needs some heuristics or additional component to estimate the length of the output token sequence. This paper proposes to apply another type of NAT called insertion-based models, that were originally proposed for NMT, to ASR tasks. Insertion-based models solve the above mask-predict issues and can generate an arbitrary generation order of an output sequence. In addition, we introduce a new formulation of joint training of the insertion-based models and CTC. This formulation reinforces CTC by making it dependent on insertion-based token generation in a non-autoregressive manner. We conducted experiments on three public benchmarks and achieved competitive performance to strong autoregressive Transformer with a similar decoding condition.

Original languageEnglish
JournalUnknown Journal
Publication statusPublished - 2020 May 27

Keywords

  • End-to-end
  • Non-autoregressive
  • Speech recognition
  • Transformer

ASJC Scopus subject areas

  • General

Fingerprint Dive into the research topics of 'Insertion-Based modeling for end-to-end automatic speech recognition'. Together they form a unique fingerprint.

Cite this