TY - GEN
T1 - Learning to Generalize Unseen Domains via Multi-source Meta Learning for Text Classification
AU - Hu, Yuxuan
AU - Zhang, Chenwei
AU - Yang, Min
AU - Liang, Xiaodan
AU - Li, Chengming
AU - Hu, Xiping
N1 - Publisher Copyright:
© The Author(s), under exclusive license to Springer Nature Switzerland AG 2025.
PY - 2025
Y1 - 2025
N2 - With the rapid development of deep learning methods, there have been many breakthroughs in the field of text classification. Models developed for this task have achievedhigh accuracy. However, most of these models are trained using labeled data from seen domains. It is difficult for these models to maintain high accuracy in a new challenging unseen domain, which is directly related to the generalization of the model. In this paper, we study the multi-source Domain Generalization for text classification and propose a framework to use multiple seen domains to train a model that can achieve high accuracy in an unseen domain. Specifically, we propose a multi-source meta-learning Domain Generalization framework to simulate the process of model generalization to an unseen domain, so as to extract sufficient domain-related features. We introduce a memory mechanism to store domain-specific features, which coordinate with the meta-learning framework. Besides, we adopt a novel “jury” mechanism that enables the model to learn sufficient domain-invariant features. Experiments demonstrate that our meta-learning framework can effectively enhance the ability of the model to generalize to an unseen domain and can outperform the state-of-the-art methods on multi-source text classification datasets.
AB - With the rapid development of deep learning methods, there have been many breakthroughs in the field of text classification. Models developed for this task have achievedhigh accuracy. However, most of these models are trained using labeled data from seen domains. It is difficult for these models to maintain high accuracy in a new challenging unseen domain, which is directly related to the generalization of the model. In this paper, we study the multi-source Domain Generalization for text classification and propose a framework to use multiple seen domains to train a model that can achieve high accuracy in an unseen domain. Specifically, we propose a multi-source meta-learning Domain Generalization framework to simulate the process of model generalization to an unseen domain, so as to extract sufficient domain-related features. We introduce a memory mechanism to store domain-specific features, which coordinate with the meta-learning framework. Besides, we adopt a novel “jury” mechanism that enables the model to learn sufficient domain-invariant features. Experiments demonstrate that our meta-learning framework can effectively enhance the ability of the model to generalize to an unseen domain and can outperform the state-of-the-art methods on multi-source text classification datasets.
KW - memory
KW - meta-learning
KW - multi sources
KW - Text classification
UR - http://www.scopus.com/inward/record.url?scp=85212299932&partnerID=8YFLogxK
U2 - 10.1007/978-3-031-78495-8_26
DO - 10.1007/978-3-031-78495-8_26
M3 - Conference contribution
AN - SCOPUS:85212299932
SN - 9783031784941
T3 - Lecture Notes in Computer Science (including subseries Lecture Notes in Artificial Intelligence and Lecture Notes in Bioinformatics)
SP - 412
EP - 428
BT - Pattern Recognition - 27th International Conference, ICPR 2024, Proceedings
A2 - Antonacopoulos, Apostolos
A2 - Chaudhuri, Subhasis
A2 - Chellappa, Rama
A2 - Liu, Cheng-Lin
A2 - Bhattacharya, Saumik
A2 - Pal, Umapada
PB - Springer Science and Business Media Deutschland GmbH
T2 - 27th International Conference on Pattern Recognition, ICPR 2024
Y2 - 1 December 2024 through 5 December 2024
ER -