Home > Published Issues > 2023 > Volume 14, No. 4, 2023 >
JAIT 2023 Vol.14(4): 694-700
doi: 10.12720/jait.14.4.694-700

Multi-speaker Speech Separation under Reverberation Conditions Using Conv-Tasnet

Chunxi Wang, Maoshen Jia *, Yanyan Zhang, and Lu Li
Beijing Key Laboratory of Computational Intelligence and Intelligent System, Faculty of Information Technology Beijing University of Technology, Beijing, China; Email: chunxiwang@emails.bjut.edu.cn (C.W.), 13811321209@139.com (Y.Z.), lilubjut@163.com (L.L.)
*Correspondence: jiamaoshen@bjut.edu.cn (M.J.)

Manuscript received January 1, 2023; revised February 5, 2023; accepted March 22, 2023; published July 26, 2023.

Abstract—The goal of speech separation is to separate the target signal from the background interference. With the rapid development of artificial intelligence, speech separation technology combined with deep learning has received more attention as well as a lot of progress. However, in the “cocktail party problem”, it is still a challenge to achieve speech separation under reverberant conditions. In order to solve this problem, a model combining the Weighted Prediction Error (WPE) method and a fully-convolutional time-domain audio separation network (Conv-Tasnet) is proposed in this paper. The model target on separating multi-channel signals after dereverberation without prior knowledge of the second field environment. Subjective and objective evaluation results show that the proposed method outperforms existing methods in the speech separation tasks in reverberant and anechoic environments.
 
Keywords—speech separation, deep learning, dereverberation, speech enhancement

Cite: Chunxi Wang, Maoshen Jia, Yanyan Zhang, and Lu Li, "Multi-speaker Speech Separation under Reverberation Conditions Using Conv-Tasnet," Journal of Advances in Information Technology, Vol. 14, No. 4, pp. 694-700, 2023.

Copyright © 2023 by the authors. This is an open access article distributed under the Creative Commons Attribution License (CC BY-NC-ND 4.0), which permits use, distribution and reproduction in any medium, provided that the article is properly cited, the use is non-commercial and no modifications or adaptations are made.