The Zero Resource Speech Challenge 2020: Discovering Discrete Subword and Word Units

Ewan Dunbar, Julien Karadayi, Mathieu Bernard, Xuan-Nga Cao, Robin Algayres, Lucas Ondel, Laurent Besacier, Sakriani Sakti, Emmanuel Dupoux


We present the Zero Resource Speech Challenge 2020, which aims at learning speech representations from raw audio signals without any labels. It combines the data sets and metrics from two previous benchmarks (2017 and 2019) and features two tasks which tap into two levels of speech representation. The first task is to discover low bit-rate subword representations that optimize the quality of speech synthesis; the second one is to discover word-like units from unsegmented raw speech. We present the results of the twenty submitted models and discuss the implications of the main findings for unsupervised speech learning.


 DOI: 10.21437/Interspeech.2020-2743

Cite as: Dunbar, E., Karadayi, J., Bernard, M., Cao, X., Algayres, R., Ondel, L., Besacier, L., Sakti, S., Dupoux, E. (2020) The Zero Resource Speech Challenge 2020: Discovering Discrete Subword and Word Units. Proc. Interspeech 2020, 4831-4835, DOI: 10.21437/Interspeech.2020-2743.


@inproceedings{Dunbar2020,
  author={Ewan Dunbar and Julien Karadayi and Mathieu Bernard and Xuan-Nga Cao and Robin Algayres and Lucas Ondel and Laurent Besacier and Sakriani Sakti and Emmanuel Dupoux},
  title={{The Zero Resource Speech Challenge 2020: Discovering Discrete Subword and Word Units}},
  year=2020,
  booktitle={Proc. Interspeech 2020},
  pages={4831--4835},
  doi={10.21437/Interspeech.2020-2743},
  url={http://dx.doi.org/10.21437/Interspeech.2020-2743}
}