TY - GEN
T1 - Multi-view Contrastive Learning for Medical Question Summarization
AU - Wei, Sibo
AU - Peng, Xueping
AU - Guan, Hongjiao
AU - Geng, Lina
AU - Jian, Ping
AU - Wu, Hao
AU - Lu, Wenpeng
N1 - Publisher Copyright:
© 2024 IEEE.
PY - 2024
Y1 - 2024
N2 - Most Seq2Seq neural model-based medical question summarization (MQS) systems have a severe mismatch between training and inference, i.e., exposure bias. However, this problem remains unexplored in the MQS task. To bridge this research gap and alleviate the problem of exposure bias, we propose a novel re-ranking training framework for MQS called Multi-view Contrastive Learning (MvCL). MvCL simultaneously considers the similarity scores between medical questions and candidate summaries as well as the average similarity scores between candidate summaries and other candidates within the same group, and utilizes contrastive learning to optimize the model's ranking ability. Additionally, we propose a new multilevel inference approach to adapt to this training strategy. The approach first filters out candidate summaries that are dissimilar to the original medical question, and then selects the summary with the highest average similarity to other candidate summaries from the remaining candidates as the final output. We conducted extensive experiments, and the results demonstrate that our proposed MvCL framework achieves state-of-the-art results on the majority of evaluation metrics across four datasets.1
AB - Most Seq2Seq neural model-based medical question summarization (MQS) systems have a severe mismatch between training and inference, i.e., exposure bias. However, this problem remains unexplored in the MQS task. To bridge this research gap and alleviate the problem of exposure bias, we propose a novel re-ranking training framework for MQS called Multi-view Contrastive Learning (MvCL). MvCL simultaneously considers the similarity scores between medical questions and candidate summaries as well as the average similarity scores between candidate summaries and other candidates within the same group, and utilizes contrastive learning to optimize the model's ranking ability. Additionally, we propose a new multilevel inference approach to adapt to this training strategy. The approach first filters out candidate summaries that are dissimilar to the original medical question, and then selects the summary with the highest average similarity to other candidate summaries from the remaining candidates as the final output. We conducted extensive experiments, and the results demonstrate that our proposed MvCL framework achieves state-of-the-art results on the majority of evaluation metrics across four datasets.1
KW - Contrastive Learning
KW - Medical Question Summarization
KW - Re-ranking Framework
UR - http://www.scopus.com/inward/record.url?scp=85199080684&partnerID=8YFLogxK
U2 - 10.1109/CSCWD61410.2024.10580086
DO - 10.1109/CSCWD61410.2024.10580086
M3 - Conference contribution
AN - SCOPUS:85199080684
T3 - Proceedings of the 2024 27th International Conference on Computer Supported Cooperative Work in Design, CSCWD 2024
SP - 1826
EP - 1831
BT - Proceedings of the 2024 27th International Conference on Computer Supported Cooperative Work in Design, CSCWD 2024
A2 - Shen, Weiming
A2 - Shen, Weiming
A2 - Barthes, Jean-Paul
A2 - Luo, Junzhou
A2 - Qiu, Tie
A2 - Zhou, Xiaobo
A2 - Zhang, Jinghui
A2 - Zhu, Haibin
A2 - Peng, Kunkun
A2 - Xu, Tianyi
A2 - Chen, Ning
PB - Institute of Electrical and Electronics Engineers Inc.
T2 - 27th International Conference on Computer Supported Cooperative Work in Design, CSCWD 2024
Y2 - 8 May 2024 through 10 May 2024
ER -