On the Comparison of Popular End-to-End Models for Large Scale Speech Recognition

Jinyu Li, Yu Wu, Yashesh Gaur, Chengyi Wang, Rui Zhao, Shujie Liu


Recently, there has been a strong push to transition from hybrid models to end-to-end (E2E) models for automatic speech recognition. Currently, there are three promising E2E methods: recurrent neural network transducer (RNN-T), RNN attention-based encoder-decoder (AED), and Transformer-AED. In this study, we conduct an empirical comparison of RNN-T, RNN-AED, and Transformer-AED models, in both non-streaming and streaming modes. We use 65 thousand hours of Microsoft anonymized training data to train these models. As E2E models are more data hungry, it is better to compare their effectiveness with large amount of training data. To the best of our knowledge, no such comprehensive study has been conducted yet. We show that although AED models are stronger than RNN-T in the non-streaming mode, RNN-T is very competitive in streaming mode if its encoder can be properly initialized. Among all three E2E models, transformer-AED achieved the best accuracy in both streaming and non-streaming mode. We show that both streaming RNN-T and transformer-AED models can obtain better accuracy than a highly-optimized hybrid model.


 DOI: 10.21437/Interspeech.2020-2846

Cite as: Li, J., Wu, Y., Gaur, Y., Wang, C., Zhao, R., Liu, S. (2020) On the Comparison of Popular End-to-End Models for Large Scale Speech Recognition. Proc. Interspeech 2020, 1-5, DOI: 10.21437/Interspeech.2020-2846.


@inproceedings{Li2020,
  author={Jinyu Li and Yu Wu and Yashesh Gaur and Chengyi Wang and Rui Zhao and Shujie Liu},
  title={{On the Comparison of Popular End-to-End Models for Large Scale Speech Recognition}},
  year=2020,
  booktitle={Proc. Interspeech 2020},
  pages={1--5},
  doi={10.21437/Interspeech.2020-2846},
  url={http://dx.doi.org/10.21437/Interspeech.2020-2846}
}