Dual-Path Transformer Network: Direct Context-Aware Modeling for End-to-End Monaural Speech Separation

Jingjing Chen, Qirong Mao, Dong Liu


The dominant speech separation models are based on complex recurrent or convolution neural network that model speech sequences indirectly conditioning on context, such as passing information through many intermediate states in recurrent neural network, leading to suboptimal separation performance. In this paper, we propose a dual-path transformer network (DPTNet) for end-to-end speech separation, which introduces direct context-awareness in the modeling for speech sequences. By introduces a improved transformer, elements in speech sequences can interact directly, which enables DPTNet can model for the speech sequences with direct context-awareness. The improved transformer in our approach learns the order information of the speech sequences without positional encodings by incorporating a recurrent neural network into the original transformer. In addition, the structure of dual paths makes our model efficient for extremely long speech sequence modeling. Extensive experiments on benchmark datasets show that our approach outperforms the current state-of-the-arts (20.6 dB SDR on the public WSj0-2mix data corpus).


 DOI: 10.21437/Interspeech.2020-2205

Cite as: Chen, J., Mao, Q., Liu, D. (2020) Dual-Path Transformer Network: Direct Context-Aware Modeling for End-to-End Monaural Speech Separation. Proc. Interspeech 2020, 2642-2646, DOI: 10.21437/Interspeech.2020-2205.


@inproceedings{Chen2020,
  author={Jingjing Chen and Qirong Mao and Dong Liu},
  title={{Dual-Path Transformer Network: Direct Context-Aware Modeling for End-to-End Monaural Speech Separation}},
  year=2020,
  booktitle={Proc. Interspeech 2020},
  pages={2642--2646},
  doi={10.21437/Interspeech.2020-2205},
  url={http://dx.doi.org/10.21437/Interspeech.2020-2205}
}