Self-Supervised Speech Representation and Contextual Text Embedding for Match-Mismatch Classification with EEG Recording

Bo Wang, Xiran Xu, Zechen Zhang, Haolin Zhu, Yu Jie Yan, Xihong Wu, Jing Chen*

*此作品的通讯作者

科研成果: 书/报告/会议事项章节会议稿件同行评审

摘要

Relating speech to EEG holds considerable importance but is challenging. In this study, a deep convolutional network was employed to extract spatiotemporal features from EEG data. Self-supervised speech representation and contextual text embedding were used as speech features. Contrastive learning was used to relate EEG features to speech features. The experimental results demonstrate the benefits of using self-supervised speech representation and contextual text embedding. Through feature fusion and model ensemble, an accuracy of 60.29% was achieved, and the performance was ranked as No.2 in Task 1 of the Auditory EEG Challenge (ICASSP 2024). The code to implement our work is available on Github: https://github.com/bobwangPKU/EEG-Stimulus-Match-Mismatch.

源语言英语
主期刊名2024 IEEE International Conference on Acoustics, Speech, and Signal Processing Workshops, ICASSPW 2024 - Proceedings
出版商Institute of Electrical and Electronics Engineers Inc.
111-112
页数2
ISBN(电子版)9798350374513
DOI
出版状态已出版 - 2024
已对外发布
活动49th IEEE International Conference on Acoustics, Speech, and Signal Processing Workshops, ICASSPW 2024 - Seoul, 韩国
期限: 14 4月 202419 4月 2024

出版系列

姓名2024 IEEE International Conference on Acoustics, Speech, and Signal Processing Workshops, ICASSPW 2024 - Proceedings

会议

会议49th IEEE International Conference on Acoustics, Speech, and Signal Processing Workshops, ICASSPW 2024
国家/地区韩国
Seoul
时期14/04/2419/04/24

指纹

探究 'Self-Supervised Speech Representation and Contextual Text Embedding for Match-Mismatch Classification with EEG Recording' 的科研主题。它们共同构成独一无二的指纹。

引用此