5th European Conference on Speech Communication and Technology

Rhodes, Greece
September 22-25, 1997

Driving Synthetic Mouth Gestures: Phonetic Recognition for FaceMe!

William Goldenthal, Keith Waters, Jean-Manuel Van Thong, Oren Glickman

Digital Equipment Corporation, Cambridge Research Laboratory Cambridge, Massachusetts, USA

The goal of this work is to use phonetic recognition to drive a synthetic image with speech. Phonetic units are identified by the phonetic recognition engine and mapped to mouth gestures, known as visemes, the visual counter-part of phonemes. The acoustic waveform and visemes are then sent to a synthetic image player, called FaceMe! where they are rendered synchronously. This paper provides background for the core technologies involved in this process and describes asynchronous and synchronous prototypes of a combined phonetic recognition/FaceMe! system which we use to render mouth gestures on an animated face.

Full Paper

Bibliographic reference.  Goldenthal, William / Waters, Keith / Thong, Jean-Manuel Van / Glickman, Oren (1997): "Driving synthetic mouth gestures: phonetic recognition for faceme!", In EUROSPEECH-1997, 1995-1998.