13th Annual Conference of the International Speech Communication Association

Portland, OR, USA
September 9-13, 2012

Constrained Maximum Mutual Information Dimensionality Reduction for Language Identification

Shuai Huang (1), Glen A. Coppersmith (1,2), Damianos Karakos (3)

(1) Center for Language and Speech Processing, Johns Hopkins University, Baltimore, MD, USA
(2) Human Language Technology Center of Excellence, Johns Hopkins University, Baltimore, MD, USA
(3) Raytheon BBN Technologies;, Cambridge, MA, USA

In this paper we propose Constrained Maximum Mutual Information dimensionality reduction (CMMI), an information-theoretic based dimensionality reduction technique. CMMI tries to maximize the mutual information between the class labels and the projected (lower dimensional) features, optimized via gradient ascent. Supervised and semi-supervised CMMI are introduced and compared with a state of the art dimensionality reduction technique (Minimum/Maximum Rényi's Mutual Information using the Stochastic Information Gradient; MRMISIG) for a language identification (LID) task using CallFriend corpus, with favorable results. CMMI also deals with higher dimensional data more gracefully than MRMI-SIG, permitting application to datasets for which MRMI-SIG is computationally prohibitive.

Full Paper

Bibliographic reference.  Huang, Shuai / Coppersmith, Glen A. / Karakos, Damianos (2012): "Constrained maximum mutual information dimensionality reduction for language identification", In INTERSPEECH-2012, 2037-2040.