Domain Adaptation Using Class Similarity for Robust Speech Recognition

Han Zhu, Jiangjiang Zhao, Yuling Ren, Li Wang, Pengyuan Zhang


When only limited target domain data is available, domain adaptation could be used to promote performance of deep neural network (DNN) acoustic model by leveraging well-trained source model and target domain data. However, suffering from domain mismatch and data sparsity, domain adaptation is very challenging. This paper proposes a novel adaptation method for DNN acoustic model using class similarity. Since the output distribution of DNN model contains the knowledge of similarity among classes, which is applicable to both source and target domain, it could be transferred from source to target model for the performance improvement. In our approach, we first compute the frame level posterior probabilities of source samples using source model. Then, for each class, probabilities of this class are used to compute a mean vector, which we refer to as mean soft labels. During adaptation, these mean soft labels are used in a regularization term to train the target model. Experiments showed that our approach outperforms fine-tuning using one-hot labels on both accent and noise adaptation task, especially when source and target domain are highly mismatched.


 DOI: 10.21437/Interspeech.2020-3087

Cite as: Zhu, H., Zhao, J., Ren, Y., Wang, L., Zhang, P. (2020) Domain Adaptation Using Class Similarity for Robust Speech Recognition. Proc. Interspeech 2020, 4367-4371, DOI: 10.21437/Interspeech.2020-3087.


@inproceedings{Zhu2020,
  author={Han Zhu and Jiangjiang Zhao and Yuling Ren and Li Wang and Pengyuan Zhang},
  title={{Domain Adaptation Using Class Similarity for Robust Speech Recognition}},
  year=2020,
  booktitle={Proc. Interspeech 2020},
  pages={4367--4371},
  doi={10.21437/Interspeech.2020-3087},
  url={http://dx.doi.org/10.21437/Interspeech.2020-3087}
}