TY - JOUR
T1 - An approach to NMT re-ranking using sequence-labeling for grammatical error correction
AU - Wang, Bo
AU - Hirota, Kaoru
AU - Liu, Chang
AU - Dai, Yaping
AU - Jia, Zhiyang
N1 - Publisher Copyright:
© 2020 Fuji Technology Press. All rights reserved.
PY - 2020/7
Y1 - 2020/7
N2 - An approach to N-best hypotheses re-ranking using a sequence-labeling model is applied to resolve the data deficiency problem in Grammatical Error Correction (GEC). Multiple candidate sentences are generated using a Neural Machine Translation (NMT) model; thereafter, these sentences are re-ranked via a stacked Transformer following a Bidirectional Long Short-Term Memory (BiLSTM) with Conditional Random Field (CRF). Correlations within the sentences are extracted using the sequence-labeling model based on the Transformer, which is particularly suitable for long sentences. Meanwhile, the knowledge from a large amount of unlabeled data is acquired through the pre-trained structure. Thus, completely revised sentences are adopted instead of partially modified sentences. Compared with conventional NMT, experiments on the NUCLE and FCE datasets demonstrate that the model improves the F0.5 score by 8.22% and 2.09%, respectively. As an advantage, the proposed re-ranking method has the advantage of only requires a small set of easily computed features that do not need linguistic inputs.
AB - An approach to N-best hypotheses re-ranking using a sequence-labeling model is applied to resolve the data deficiency problem in Grammatical Error Correction (GEC). Multiple candidate sentences are generated using a Neural Machine Translation (NMT) model; thereafter, these sentences are re-ranked via a stacked Transformer following a Bidirectional Long Short-Term Memory (BiLSTM) with Conditional Random Field (CRF). Correlations within the sentences are extracted using the sequence-labeling model based on the Transformer, which is particularly suitable for long sentences. Meanwhile, the knowledge from a large amount of unlabeled data is acquired through the pre-trained structure. Thus, completely revised sentences are adopted instead of partially modified sentences. Compared with conventional NMT, experiments on the NUCLE and FCE datasets demonstrate that the model improves the F0.5 score by 8.22% and 2.09%, respectively. As an advantage, the proposed re-ranking method has the advantage of only requires a small set of easily computed features that do not need linguistic inputs.
KW - Grammatical error correction
KW - Neural machine translation
KW - Sequence-labeling
KW - Transformer
UR - http://www.scopus.com/inward/record.url?scp=85089576655&partnerID=8YFLogxK
U2 - 10.20965/jaciii.2020.p0557
DO - 10.20965/jaciii.2020.p0557
M3 - Article
AN - SCOPUS:85089576655
SN - 1343-0130
VL - 24
SP - 557
EP - 567
JO - Journal of Advanced Computational Intelligence and Intelligent Informatics
JF - Journal of Advanced Computational Intelligence and Intelligent Informatics
IS - 4
ER -