Publications

WORKSHOP (INTERNATIONAL) A Comparative Study on Non-Autoregressive Modelings for Speech-to-Text Generation

Yosuke Higuchi (Waseda University), Nanxin Chen (Johns Hopkins University), Yuya Fujita (Yahoo Japan Corporation), Hirofumi Inaguma (Kyoto University), Tatsuya Komatsu, Jaesong Lee (Naver Corporation), Jumon Nozaki (Kyoto University), Tianzi Wang (Johns Hopkins University), Shinji Watanabe (Carnegie Mellon University)

2021 IEEE Automatic Speech Recognition and Understanding Workshop (ASRU) (ASRU 2021)

December 13, 2021

Non-autoregressive (NAR) models simultaneously generate multiple outputs in a sequence, which significantly reduces the inference speed at the cost of accuracy drop compared to autoregressive baselines. Showing great potential for real-time applications, an increasing number of NAR models have been explored in different fields to mitigate the performance gap against AR models. In this work, we conduct a comparative study of various NAR modeling methods for end-to-end automatic speech recognition (ASR). Experiments are performed in the state-of-the-art setting using ESPnet. The results on various tasks provide interesting findings for developing an understanding of NAR ASR, such as the accuracy-speed trade-off and robustness against long-form utterances. We also show that the techniques can be combined for further improvement and applied to NAR end-to-end speech translation. All the implementations are publicly available to encourage further research in NAR speech processing.

Paper : A Comparative Study on Non-Autoregressive Modelings for Speech-to-Text Generationopen into new tab or window (external link)