TY - JOUR
T1 - DSciSum
T2 - Detailed summarization of long scientific documents
AU - Liu, Ran
AU - Mao, Xian Ling
AU - Huang, Heyan
N1 - Publisher Copyright:
© 2025 Elsevier B.V.
PY - 2025/5/23
Y1 - 2025/5/23
N2 - A summary is frequently considered by academics as a viable alternative to long scientific documents. Prior studies generally required well-annotated training datasets such as arXiv and PubMed, using abstracts from articles as supervised signals. However, these gold summaries merely provide a cursory overview of the subject matter, lacking crucial detailed information, such as datasets, evaluation metrics and model performance, which are essential for both academics and the general public. To address this problem, we propose DSciSum, an extract-then-generate framework that utilizes the zero-shot capabilities and a superior semantic understanding of large language models (LLMs). This approach focuses on previously overlooked details, thereby generating more human-related summaries. Moreover, an innovative LLM-based evaluation criterion is designed as a substitute for traditional metrics, providing a more meaningful and professional assessment for scientific summarization. Specifically, DSciSum first selects salient sentences containing both general and detailed information using a statistics-based heuristic approach. Thereafter, it pretrains and finetunes LLMs to acquire the generator tailored for scientific summarization. Finally, G-SciEval is designed to provide a human-related evaluation of scientific summarization from a deep semantic perspective. Experimental results show that DSciSum outperforms both the reference and state-of-the-art models on arXivCap.
AB - A summary is frequently considered by academics as a viable alternative to long scientific documents. Prior studies generally required well-annotated training datasets such as arXiv and PubMed, using abstracts from articles as supervised signals. However, these gold summaries merely provide a cursory overview of the subject matter, lacking crucial detailed information, such as datasets, evaluation metrics and model performance, which are essential for both academics and the general public. To address this problem, we propose DSciSum, an extract-then-generate framework that utilizes the zero-shot capabilities and a superior semantic understanding of large language models (LLMs). This approach focuses on previously overlooked details, thereby generating more human-related summaries. Moreover, an innovative LLM-based evaluation criterion is designed as a substitute for traditional metrics, providing a more meaningful and professional assessment for scientific summarization. Specifically, DSciSum first selects salient sentences containing both general and detailed information using a statistics-based heuristic approach. Thereafter, it pretrains and finetunes LLMs to acquire the generator tailored for scientific summarization. Finally, G-SciEval is designed to provide a human-related evaluation of scientific summarization from a deep semantic perspective. Experimental results show that DSciSum outperforms both the reference and state-of-the-art models on arXivCap.
KW - Automatic text summarization
KW - Evaluation
KW - Large language models
KW - Long document summarization
UR - http://www.scopus.com/inward/record.url?scp=105002224367&partnerID=8YFLogxK
U2 - 10.1016/j.knosys.2025.113409
DO - 10.1016/j.knosys.2025.113409
M3 - Article
AN - SCOPUS:105002224367
SN - 0950-7051
VL - 317
JO - Knowledge-Based Systems
JF - Knowledge-Based Systems
M1 - 113409
ER -