TY - GEN
T1 - Neeko
T2 - 2024 Conference on Empirical Methods in Natural Language Processing, EMNLP 2024
AU - Yu, Xiaoyan
AU - Luo, Tongxu
AU - Wei, Yifan
AU - Lei, Fangyu
AU - Huang, Yiming
AU - Peng, Hao
AU - Zhu, Liehuang
N1 - Publisher Copyright:
© 2024 Association for Computational Linguistics.
PY - 2024
Y1 - 2024
N2 - Large Language Models (LLMs) have revolutionized open-domain dialogue agents but encounter challenges in multi-character role-playing (MCRP) scenarios. To address this issue, this work presents Neeko, an innovative framework designed for efficient multiple-character role-playing. The proposed framework breaks down the role-playing agent's training process into agent pre-tuning, multiple character playing, and character incremental learning, effectively handling both seen and unseen roles. Neeko employs a dynamic low-rank adapter (LoRA) strategy by training separate LoRA blocks independently for each character, alongside incorporating a gating network for role selection. This design allows Neeko to seamlessly adjust to a wide range of characters, thereby bolstering its adaptability to distinctive attributes, personalities, and speech patterns. As a result, Neeko demonstrates superior performance in MCRP over most existing methods, offering more engaging and versatile user interaction experiences. Code and data are available at https://github.com/weiyifan1023/Neeko.
AB - Large Language Models (LLMs) have revolutionized open-domain dialogue agents but encounter challenges in multi-character role-playing (MCRP) scenarios. To address this issue, this work presents Neeko, an innovative framework designed for efficient multiple-character role-playing. The proposed framework breaks down the role-playing agent's training process into agent pre-tuning, multiple character playing, and character incremental learning, effectively handling both seen and unseen roles. Neeko employs a dynamic low-rank adapter (LoRA) strategy by training separate LoRA blocks independently for each character, alongside incorporating a gating network for role selection. This design allows Neeko to seamlessly adjust to a wide range of characters, thereby bolstering its adaptability to distinctive attributes, personalities, and speech patterns. As a result, Neeko demonstrates superior performance in MCRP over most existing methods, offering more engaging and versatile user interaction experiences. Code and data are available at https://github.com/weiyifan1023/Neeko.
UR - http://www.scopus.com/inward/record.url?scp=85217805078&partnerID=8YFLogxK
U2 - 10.18653/v1/2024.emnlp-main.697
DO - 10.18653/v1/2024.emnlp-main.697
M3 - Conference contribution
AN - SCOPUS:85217805078
T3 - EMNLP 2024 - 2024 Conference on Empirical Methods in Natural Language Processing, Proceedings of the Conference
SP - 12540
EP - 12557
BT - EMNLP 2024 - 2024 Conference on Empirical Methods in Natural Language Processing, Proceedings of the Conference
A2 - Al-Onaizan, Yaser
A2 - Bansal, Mohit
A2 - Chen, Yun-Nung
PB - Association for Computational Linguistics (ACL)
Y2 - 12 November 2024 through 16 November 2024
ER -