A Target Speaker Separation Neural Network with Joint-Training

Wenjing Yang, Jing Wang, Hongfeng Li, Na Xu, Fei Xiang, Kai Qian, Shenghua Hu

科研成果: 书/报告/会议事项章节会议稿件同行评审

2 引用 (Scopus)

摘要

Target speaker separation aims to separate a target speech from multiple interference voices, which is promising for solving conventional difficulties in speech separation, such as arbitrary source permutation and unknown number of sources, and is useful for personal applications, like online meeting and personal phone calls. Recently, the application of deep-learning based models provided more alternatives for target speaker separation tasks. In this paper, we proposed a target speaker separation neural network with joint-training that separates the target voice in the spectrogram domain with the proposed combinative loss function. Experimental results show that compared with the baseline, our proposed method yields better performance on both test data and real data. Meanwhile, the proposed combinative loss function is more effective in addressing this issue.

源语言英语
主期刊名2021 Asia-Pacific Signal and Information Processing Association Annual Summit and Conference, APSIPA ASC 2021 - Proceedings
出版商Institute of Electrical and Electronics Engineers Inc.
614-618
页数5
ISBN(电子版)9789881476890
出版状态已出版 - 2021
活动2021 Asia-Pacific Signal and Information Processing Association Annual Summit and Conference, APSIPA ASC 2021 - Tokyo, 日本
期限: 14 12月 202117 12月 2021

出版系列

姓名2021 Asia-Pacific Signal and Information Processing Association Annual Summit and Conference, APSIPA ASC 2021 - Proceedings

会议

会议2021 Asia-Pacific Signal and Information Processing Association Annual Summit and Conference, APSIPA ASC 2021
国家/地区日本
Tokyo
时期14/12/2117/12/21

指纹

探究 'A Target Speaker Separation Neural Network with Joint-Training' 的科研主题。它们共同构成独一无二的指纹。

引用此