Robust speech recognition combining cepstral and articulatory features

Zhuan Ling Zha, Jin Hu, Qing Ran Zhan, Ya Hui Shan, Xiang Xie, Jing Wang, Hao Bo Cheng

科研成果: 书/报告/会议事项章节会议稿件同行评审

3 引用 (Scopus)

摘要

In this paper, a nonlinear relationship between pronunciation and auditory perception is introduced into speech recognition, and superior robustness is shown in the results. The Extreme Learning Machine mapping the relations was trained with Mocha-TIMIT database. Articulatory Features (AFs) were obtained by the network and MFCCs were fused for training acoustic model-DNN-HMM and GMM-HMM in this experiment. It has an 117.0% relative increment of WER with MFCCs-AFs-GMM-HMM while 125.6% with MFCCs-GMM-HMM And the performance of the model DNN-HMM is better than that of the model GMM-HMM, both with relative and absolute performance.

源语言英语
主期刊名2017 3rd IEEE International Conference on Computer and Communications, ICCC 2017
出版商Institute of Electrical and Electronics Engineers Inc.
1401-1405
页数5
ISBN(电子版)9781509063505
DOI
出版状态已出版 - 2 7月 2017
活动3rd IEEE International Conference on Computer and Communications, ICCC 2017 - Chengdu, 中国
期限: 13 12月 201716 12月 2017

出版系列

姓名2017 3rd IEEE International Conference on Computer and Communications, ICCC 2017
2018-January

会议

会议3rd IEEE International Conference on Computer and Communications, ICCC 2017
国家/地区中国
Chengdu
时期13/12/1716/12/17

指纹

探究 'Robust speech recognition combining cepstral and articulatory features' 的科研主题。它们共同构成独一无二的指纹。

引用此