Prediction of Head Motion from Speech Waveforms with a Canonical-Correlation-Constrained Autoencoder

JinHong Lu, Hiroshi Shimodaira


This study investigates the direct use of speech waveforms to predict head motion for speech-driven head-motion synthesis, whereas the use of spectral features such as MFCC as basic input features together with additional features such as energy and F0 is common in the literature. We show that, rather than combining different features that originate from waveforms, it is more effective to use waveforms directly predicting corresponding head motion. The challenge with the waveform-based approach is that waveforms contain a large amount of information irrelevant to predict head motion, which hinders the training of neural networks. To overcome the problem, we propose a canonical-correlation-constrained autoencoder (CCCAE), where hidden layers are trained to not only minimise the error but also maximise the canonical correlation with head motion. Compared with an MFCC-based system, the proposed system shows comparable performance in objective evaluation, and better performance in subject evaluation.


 DOI: 10.21437/Interspeech.2020-1218

Cite as: Lu, J., Shimodaira, H. (2020) Prediction of Head Motion from Speech Waveforms with a Canonical-Correlation-Constrained Autoencoder. Proc. Interspeech 2020, 1301-1305, DOI: 10.21437/Interspeech.2020-1218.


@inproceedings{Lu2020,
  author={JinHong Lu and Hiroshi Shimodaira},
  title={{Prediction of Head Motion from Speech Waveforms with a Canonical-Correlation-Constrained Autoencoder}},
  year=2020,
  booktitle={Proc. Interspeech 2020},
  pages={1301--1305},
  doi={10.21437/Interspeech.2020-1218},
  url={http://dx.doi.org/10.21437/Interspeech.2020-1218}
}