TY - JOUR
T1 - 基于细粒度可解释矩阵的摘要生成模型
AU - Wang, Haonan
AU - Gao, Yang
AU - Feng, Junlan
AU - Hu, Min
AU - Wang, Huixin
AU - Bai, Yu
N1 - Publisher Copyright:
© 2021 Peking University.
PY - 2021/1/20
Y1 - 2021/1/20
N2 - According to the great challenge of summarizing and interpreting the information of a long article in the summary model. A summary model (Fine-Grained Interpretable Matrix, FGIM), which is retracted and then generated, is proposed to improve the interpretability of the long text on the significance, update and relevance, and then guide to automatically generate a summary. The model uses a pair-wise extractor to compress the content of the article, capture the sentence with a high degree of centrality, and uses the compressed text to combine with the generator to achieve the process of generating the summary. At the same time, the interpretable mask matrix can be used to control the direction of digest generation at the generation end. The encoder uses two methods based on Transformer and BERT respectively. This method is better than the best baseline model on the benchmark text summary data set (CNN/DailyMail and NYT50). The experiment further builds two test data sets to verify the update and relevance of the abstract, and the proposed model achieves corresponding improvements in the controllable generation of the data set.
AB - According to the great challenge of summarizing and interpreting the information of a long article in the summary model. A summary model (Fine-Grained Interpretable Matrix, FGIM), which is retracted and then generated, is proposed to improve the interpretability of the long text on the significance, update and relevance, and then guide to automatically generate a summary. The model uses a pair-wise extractor to compress the content of the article, capture the sentence with a high degree of centrality, and uses the compressed text to combine with the generator to achieve the process of generating the summary. At the same time, the interpretable mask matrix can be used to control the direction of digest generation at the generation end. The encoder uses two methods based on Transformer and BERT respectively. This method is better than the best baseline model on the benchmark text summary data set (CNN/DailyMail and NYT50). The experiment further builds two test data sets to verify the update and relevance of the abstract, and the proposed model achieves corresponding improvements in the controllable generation of the data set.
KW - Abstractive summarization
KW - Centrality
KW - Controllable
KW - Interpretable extraction
KW - Mask matrix
UR - http://www.scopus.com/inward/record.url?scp=85101385191&partnerID=8YFLogxK
U2 - 10.13209/j.0479-8023.2020.082
DO - 10.13209/j.0479-8023.2020.082
M3 - 文章
AN - SCOPUS:85101385191
SN - 0479-8023
VL - 57
SP - 23
EP - 30
JO - Beijing Daxue Xuebao (Ziran Kexue Ban)/Acta Scientiarum Naturalium Universitatis Pekinensis
JF - Beijing Daxue Xuebao (Ziran Kexue Ban)/Acta Scientiarum Naturalium Universitatis Pekinensis
IS - 1
ER -