TY - GEN
T1 - Comparing Human-Labeled and LLM-Generated Semantic Features via Cortical Neural Representation
AU - Xiao, Boda
AU - Wang, Bo
AU - Chen, Xuning
AU - Xu, Xiran
AU - Wu, Xihong
AU - Chen, Jing
N1 - Publisher Copyright:
©2024 IEEE.
PY - 2024
Y1 - 2024
N2 - Understanding how concepts are represented in the brain is a crucial yet unresolved question in language comprehension. Cognitive neuroscience research addresses this issue by linking the semantic features of concepts with neural responses. However, obtaining these semantic features from human raters is time-consuming and labor-intensive, limiting their application in conceptual representation studies. Recently, large language models (LLMs) have demonstrated near-human performance in semantic processing. We evaluated the feasibility of using ChatGPT to automatically generate semantic features by analyzing the correlation of semantic features for 80 concrete nouns obtained from both human raters and ChatGPT raters. Subsequently, we recorded MEG signals from seven participants as they listened to the 80 auditory nouns, and used multivariate pattern analysis (MVPA) to examine the spatiotemporal representations of ChatGPT-derived and human-derived semantic features in MEG source activity. Our results show that the two sets of semantic features are significantly correlated and exhibit high consistency across raters. Furthermore, MVPA results indicate that the semantic features derived from both methods display similar spatiotemporal patterns, which are left-lateralized and spatially overlap with the semantic network, emerging around 200 ms post-stimulus. These findings suggest that LLMs can effectively generate human-like semantic features for use in conceptual representation studies.
AB - Understanding how concepts are represented in the brain is a crucial yet unresolved question in language comprehension. Cognitive neuroscience research addresses this issue by linking the semantic features of concepts with neural responses. However, obtaining these semantic features from human raters is time-consuming and labor-intensive, limiting their application in conceptual representation studies. Recently, large language models (LLMs) have demonstrated near-human performance in semantic processing. We evaluated the feasibility of using ChatGPT to automatically generate semantic features by analyzing the correlation of semantic features for 80 concrete nouns obtained from both human raters and ChatGPT raters. Subsequently, we recorded MEG signals from seven participants as they listened to the 80 auditory nouns, and used multivariate pattern analysis (MVPA) to examine the spatiotemporal representations of ChatGPT-derived and human-derived semantic features in MEG source activity. Our results show that the two sets of semantic features are significantly correlated and exhibit high consistency across raters. Furthermore, MVPA results indicate that the semantic features derived from both methods display similar spatiotemporal patterns, which are left-lateralized and spatially overlap with the semantic network, emerging around 200 ms post-stimulus. These findings suggest that LLMs can effectively generate human-like semantic features for use in conceptual representation studies.
KW - LLM
KW - brain semantic representation
KW - semantic feature annotation
UR - https://www.scopus.com/pages/publications/85216406997
U2 - 10.1109/ISCSLP63861.2024.10800742
DO - 10.1109/ISCSLP63861.2024.10800742
M3 - Conference contribution
AN - SCOPUS:85216406997
T3 - 2024 14th International Symposium on Chinese Spoken Language Processing, ISCSLP 2024
SP - 666
EP - 670
BT - 2024 14th International Symposium on Chinese Spoken Language Processing, ISCSLP 2024
A2 - Qian, Yanmin
A2 - Jin, Qin
A2 - Ou, Zhijian
A2 - Ling, Zhenhua
A2 - Wu, Zhiyong
A2 - Li, Ya
A2 - Xie, Lei
A2 - Tao, Jianhua
PB - Institute of Electrical and Electronics Engineers Inc.
T2 - 14th International Symposium on Chinese Spoken Language Processing, ISCSLP 2024
Y2 - 7 November 2024 through 10 November 2024
ER -