4^{th} International Conference on Spoken Language ProcessingPhiladelphia, PA, USA |
This paper presents methods to improve the probability density estimation in hidden Markov models for phoneme recognition by exploiting the Self-Organizing Map (SOM) algorithm. The advantage of using the SOM is based on the created approximative topology between the mixture densities by training the Gaussian mean vectors used as the kernel centers by the SOM algorithm. The topology makes the neighboring mixtures to respond strongly for the same inputs and so most of the nearest mixtures used to approximate the current observation probability will be found in the topological neighborhood of the "winner" mixture. Also the knowledge about the previous winners are used to speed up the the search for the new winners. Tree-search SOMs and segmental SOM training are studied aiming at faster search and suitability for HMM training. The framework for the presented experiments includes melcepstrum features and phoneme-wise tied mixture density HMMs.
Bibliographic reference. Kurimo, Mikko / Somervuo, Panu (1996): "Using the self-organizing map to speed up the probability density estimation for speech recognition with mixture density HMMs", In ICSLP-1996, 358-361.