4th International Conference on Spoken Language Processing
Philadelphia, PA, USA
We study two key issues in task-independent training, namely selection of a universal set of subword units and modeling of the selected units. Since no a priori knowledge about the application vocabulary and syntax was used in the collection of the training corpus and the recognition task is frequently changing, the conventional strategy can no longer provide the best performance across many different tasks. We present a new approach that use the complete sets of right and left context-dependent units as the basis phone sets. Training of these models is accomplished by a new training criterion that maximizes phone separation between competing models. The proposed phone selection and modeling approach was evaluated across different tasks in American English. Good recognition results were obtained for both context-independent and context-dependent phone models even for unseen tasks. The same strategy has also been applied to two other languages, Mandarin Chinese and Spanish, with similar success.
Bibliographic reference. Lee, Chin-Hui / Juang, Biing-Hwang / Chou, Wu / Molina-Perez, J. J. (1996): "A study on task-independent subword selection and modeling for speech recognition", In ICSLP-1996, 1820-1823.