Domain Adversarial Neural Networks for Dysarthric Speech Recognition

Dominika Woszczyk, Stavros Petridis, David Millard


Speech recognition systems have improved dramatically over the last few years, however, their performance is significantly degraded for the cases of accented or impaired speech. This work explores domain adversarial neural networks (DANN) for speaker-independent speech recognition on the UAS dataset of dysarthric speech. The classification task on 10 spoken digits is performed using an end-to-end CNN taking raw audio as input. The results are compared to a speaker-adaptive (SA) model as well as speaker-dependent (SD) and multi-task learning models (MTL). The experiments conducted in this paper show that DANN achieves an absolute recognition rate of 74.91% and outperforms the baseline by 12.18%. Additionally, the DANN model achieves comparable results to the SA model’s recognition rate of 77.65%. We also observe that when labelled dysarthric speech data is available DANN and MTL perform similarly, but when they are not DANN performs better than MTL.


 DOI: 10.21437/Interspeech.2020-2845

Cite as: Woszczyk, D., Petridis, S., Millard, D. (2020) Domain Adversarial Neural Networks for Dysarthric Speech Recognition. Proc. Interspeech 2020, 3875-3879, DOI: 10.21437/Interspeech.2020-2845.


@inproceedings{Woszczyk2020,
  author={Dominika Woszczyk and Stavros Petridis and David Millard},
  title={{Domain Adversarial Neural Networks for Dysarthric Speech Recognition}},
  year=2020,
  booktitle={Proc. Interspeech 2020},
  pages={3875--3879},
  doi={10.21437/Interspeech.2020-2845},
  url={http://dx.doi.org/10.21437/Interspeech.2020-2845}
}