13th Annual Conference of the International Speech Communication Association

Portland, OR, USA
September 9-13, 2012

Posterior-Scaled MPE: Novel Discriminative Training Criteria

Markus Nussbaum-Thom (1), Zoltan Tuske (1), Georg Heigold (2), Ralf Schlüter (1), Hermann Ney (1)

(1) Computer Science Dept. 6, RWTH Aachen University, Aachen, Germany
(2) Google Research, Mountain View, CA, USA

We recently discovered novel discriminative training criteria following a principled approach. In this approach training criteria are developed from error bounds on the classification error for pattern classification tasks that depend on non-trivial loss functions. Automatic speech recognition (ASR) is a prominent example for such a task depending on the non-trivial Levenshtein loss. In this context, the posterior-scaled Minimum Phoneme Error (MPE) training criterion, which is the state-of-the-art discriminative training criterion in ASR, was shown to be an approximation to one of the novel criteria.
   Here, we describe the implementation of the posterior-scaled MPE criterion in a transducer-based framework, and compare this criterion to other discriminative training criteria on an ASR task. This comparison indicates that the posteriorscaled MPE criterion performs better than other discriminative criteria including MPE.

Index Terms: error bounds, discriminative training criteria, margin, MPE

Full Paper

Bibliographic reference.  Nussbaum-Thom, Markus / Tuske, Zoltan / Heigold, Georg / Schlüter, Ralf / Ney, Hermann (2012): "Posterior-scaled MPE: novel discriminative training criteria", In INTERSPEECH-2012, 2614-2617.