TY - GEN
T1 - Importance-Based Neuron Selective Distillation for Interference Mitigation in Multilingual Neural Machine Translation
AU - Zhang, Jiarui
AU - Huang, Heyan
AU - Hu, Yue
AU - Guo, Ping
AU - Xie, Yuqiang
N1 - Publisher Copyright:
© 2023, The Author(s), under exclusive license to Springer Nature Switzerland AG.
PY - 2023
Y1 - 2023
N2 - Multilingual neural machine translation employs a single model to translate multiple languages, enabling efficient cross-lingual transferability through shared parameters. However, multilingual training suffers from negative language interference, especially interference with high-resource languages. Existing approaches generally use language-specific modules to distinguish heterogeneous characteristics among different languages but suffer from the parameter explosion problem. In this paper, we propose a “divide and conquer” multilingual translation training method based on the importance of neurons that can mitigate negative language interference effectively without adding additional parameters. The key technologies can be summarized as estimation, pruning, distillation, and fine-tuning. Specifically, we estimate the importance of existing pre-trained model neurons, dividing them into the important ones representing general knowledge of each language and the unimportant ones representing individual knowledge of each low-resource language. Then, we prune the pre-trained model, retaining only the important neurons, and train the pruned model supervised by the original complete model via selective distillation to compensate for some performance loss due to unstructured pruning. Finally, we restore the pruned neurons and only fine-tune them. Experimental results on several language pairs demonstrate the effectiveness of the proposed method.
AB - Multilingual neural machine translation employs a single model to translate multiple languages, enabling efficient cross-lingual transferability through shared parameters. However, multilingual training suffers from negative language interference, especially interference with high-resource languages. Existing approaches generally use language-specific modules to distinguish heterogeneous characteristics among different languages but suffer from the parameter explosion problem. In this paper, we propose a “divide and conquer” multilingual translation training method based on the importance of neurons that can mitigate negative language interference effectively without adding additional parameters. The key technologies can be summarized as estimation, pruning, distillation, and fine-tuning. Specifically, we estimate the importance of existing pre-trained model neurons, dividing them into the important ones representing general knowledge of each language and the unimportant ones representing individual knowledge of each low-resource language. Then, we prune the pre-trained model, retaining only the important neurons, and train the pruned model supervised by the original complete model via selective distillation to compensate for some performance loss due to unstructured pruning. Finally, we restore the pruned neurons and only fine-tune them. Experimental results on several language pairs demonstrate the effectiveness of the proposed method.
KW - Importance estimation
KW - Multilingual translation
KW - Negative language interference
KW - Pruning
KW - Selective knowledge distillation
UR - http://www.scopus.com/inward/record.url?scp=85173063376&partnerID=8YFLogxK
U2 - 10.1007/978-3-031-40292-0_12
DO - 10.1007/978-3-031-40292-0_12
M3 - Conference contribution
AN - SCOPUS:85173063376
SN - 9783031402913
T3 - Lecture Notes in Computer Science (including subseries Lecture Notes in Artificial Intelligence and Lecture Notes in Bioinformatics)
SP - 140
EP - 150
BT - Knowledge Science, Engineering and Management - 16th International Conference, KSEM 2023, Proceedings
A2 - Jin, Zhi
A2 - Jiang, Yuncheng
A2 - Ma, Wenjun
A2 - Buchmann, Robert Andrei
A2 - Ghiran, Ana-Maria
A2 - Bi, Yaxin
PB - Springer Science and Business Media Deutschland GmbH
T2 - Knowledge Science, Engineering and Management - 16th International Conference, KSEM 2023, Proceedings
Y2 - 16 August 2023 through 18 August 2023
ER -