Integrating Spectrotemporal Context into Features Based on Auditory Perception for Classification-based Speech Separation

Xiang Li, Xihong Wu, Jing Chen

科研成果: 书/报告/会议事项章节会议稿件同行评审

摘要

Speech separation, which has been a challenging task for decades, especially at low signal-to-noise ratios (SNRs), can be cast as a classification problem. In such adverse acoustic environment, extracting robust features from noisy mixtures is crucial for successful classification. In the past studies, features representing temporal dynamics, known as delta features, have been widely used. Combining basic features with their deltas yields better speech separation results than using basic features alone. In this study, the commonly used delta feature was modified according to the characteristics of auditory perception, which included auditory processing on spectral change and spectral contrast. Therefore, we proposed a feature which integrated spectrotemporal context via replacing the commonly used delta feature by spectral change feature and spectral contrast feature. Experimental results showed that the proposed feature could produce better speech segregation performance than the common delta feature.

源语言英语
主期刊名2019 IEEE International Conference on Acoustics, Speech, and Signal Processing, ICASSP 2019 - Proceedings
出版商Institute of Electrical and Electronics Engineers Inc.
7165-7169
页数5
ISBN(电子版)9781479981311
DOI
出版状态已出版 - 5月 2019
已对外发布
活动44th IEEE International Conference on Acoustics, Speech, and Signal Processing, ICASSP 2019 - Brighton, 英国
期限: 12 5月 201917 5月 2019

出版系列

姓名ICASSP, IEEE International Conference on Acoustics, Speech and Signal Processing - Proceedings
2019-May
ISSN(印刷版)1520-6149

会议

会议44th IEEE International Conference on Acoustics, Speech, and Signal Processing, ICASSP 2019
国家/地区英国
Brighton
时期12/05/1917/05/19

指纹

探究 'Integrating Spectrotemporal Context into Features Based on Auditory Perception for Classification-based Speech Separation' 的科研主题。它们共同构成独一无二的指纹。

引用此