Publications

CONFERENCE (INTERNATIONAL) InterAug: Augmenting Noisy Intermediate Predictions for CTC-based ASR

Yu Nakagome, Tatsuya Komatsu, Yusuke Fujita, Shuta Ichimura, Yusuke Kida

The 23rd Annual Conference of the International Speech Communication Association (INTERSPEECH 2022)

September 18, 2022

This paper proposes InterAug: a novel training method for CTC-based ASR using augmented intermediate representations for conditioning. The proposed method exploits the conditioning framework of self-conditioned CTC to train robust models by conditioning with ``noisy'' intermediate predictions. During the training, intermediate predictions are changed to incorrect intermediate predictions and fed into the next layer for conditioning. The subsequent layers are trained to correct the incorrect intermediate predictions with the intermediate losses. By repeating the augmentation and the correction, iterative refinements, which generally require a special decoder, can be realized only with the audio encoder. To produce noisy intermediate predictions, we also introduce new augmentation: intermediate feature space augmentation and intermediate token space augmentation that are designed to simulate typical errors. The combination of the proposed InterAug framework with new augmentation allows explicit training of the robust audio encoders. In experiments using augmentations simulating deletion, insertion, and substitution error, we confirmed that the trained model acquires robustness to each error, boosting the speech recognition performance of the strong self-conditioned CTC baseline.

Paper : InterAug: Augmenting Noisy Intermediate Predictions for CTC-based ASRopen into new tab or window (external link)