TY - GEN
T1 - Implicit Discourse Relation Recognition by Scoring Special Tokens
AU - Cai, Mingyang
AU - Jian, Ping
AU - Tian, Yuhang
AU - Wang, Hai
N1 - Publisher Copyright:
© 2023 IEEE.
PY - 2023
Y1 - 2023
N2 - Implicit discourse relation recognition is one of the most difficult tasks in natural language understanding. Because of the lack of explicit connectives, the ability to extract logical information between the two discourse arguments is highly required. Previous studies mainly focus on designing complex neural network layers and various kinds of interactions between the two arguments, without further exploring the logical semantics in the pre-Trained language models. We propose a novel method that utilizes the power of the pre-Trained language model (RoBERTa) by introducing different kinds of extra special tokens, which can represent the relations between the discourse arguments respectively. On one hand, these special tokens learn to aggregate category features that profit classification; on the other hand, they can also be regarded as new 'words' that the prediction of them inspires the pre-Train language model. To effectively learn these special tokens, a scorer is then trained to give the discourse connected by corresponding special tokens higher scores than those with other special tokens. The experiments show the result of our approach exceeds our baseline by 4.24% F1, and the state-of-The-Art model by approximately 2.37% F1 on the four-way classification of the PDTB 2.0 dataset.
AB - Implicit discourse relation recognition is one of the most difficult tasks in natural language understanding. Because of the lack of explicit connectives, the ability to extract logical information between the two discourse arguments is highly required. Previous studies mainly focus on designing complex neural network layers and various kinds of interactions between the two arguments, without further exploring the logical semantics in the pre-Trained language models. We propose a novel method that utilizes the power of the pre-Trained language model (RoBERTa) by introducing different kinds of extra special tokens, which can represent the relations between the discourse arguments respectively. On one hand, these special tokens learn to aggregate category features that profit classification; on the other hand, they can also be regarded as new 'words' that the prediction of them inspires the pre-Train language model. To effectively learn these special tokens, a scorer is then trained to give the discourse connected by corresponding special tokens higher scores than those with other special tokens. The experiments show the result of our approach exceeds our baseline by 4.24% F1, and the state-of-The-Art model by approximately 2.37% F1 on the four-way classification of the PDTB 2.0 dataset.
KW - Implicit Discourse Relation Recognition
KW - Natural Language Processing
UR - http://www.scopus.com/inward/record.url?scp=85172155136&partnerID=8YFLogxK
U2 - 10.1109/ICFEICT59519.2023.00074
DO - 10.1109/ICFEICT59519.2023.00074
M3 - Conference contribution
AN - SCOPUS:85172155136
T3 - Proceedings - 2023 3rd International Conference on Frontiers of Electronics, Information and Computation Technologies, ICFEICT 2023
SP - 407
EP - 413
BT - Proceedings - 2023 3rd International Conference on Frontiers of Electronics, Information and Computation Technologies, ICFEICT 2023
A2 - Liu, Weijian
A2 - Wang, Zhuo Zheng
A2 - You, Peng
PB - Institute of Electrical and Electronics Engineers Inc.
T2 - 3rd International Conference on Frontiers of Electronics, Information and Computation Technologies, ICFEICT 2023
Y2 - 26 May 2023 through 29 May 2023
ER -