Instruction Fine-Tuning Guidance: How PEFT Methods Impact the Generation of Language Model via Different Attributions

Linjia Jiang, Shumin Shi*, Cheng Yang

*Corresponding author for this work

Research output: Chapter in Book/Report/Conference proceedingConference contributionpeer-review

Abstract

In this paper, we investigate the impact of different Parameter Efficient Fine-tuning (PEFT) methods on the generation performance of large language models, focusing on how different methods affect the model through various attributes. We evaluate five well-known PEFT methods: LoRA, AdaLoRA, LoHA, IA3, and P-Tuning taking computational resource, scalability, efficiency, and other considerations into consideration. We use the Qwen-7B-Chat model as a baseline, instruction fine-tune it using the alpaca-cleaned dataset, and use the MT-Bench and HelpSteer datasets to evaluate the model by several metrics, including helpfulness, correctness, consistency, complexity and verbosity. Experimental results show that while all methods enhance generation performance of the model, they exhibit different advantages and trade-offs in terms of text quality, diversity, and computational efficiency. Ablation studies further explore the impact of the number of training parameters and target modules on model generation performance, and find that reducing the number of them can still lead to improved performance. Our analysis provides valuable guidance for selecting the most appropriate fine-tuning methods based on specific task requirements, and enlightening thoughts on how these methods affect language model generation after instruction fine-tuning.

Original languageEnglish
Title of host publicationIntelligent Multilingual Information Processing - 1st International Conference, IMLIP 2024, Proceedings
EditorsHuaping Zhang, Jianyun Shang, Jinsong Su
PublisherSpringer Science and Business Media Deutschland GmbH
Pages277-292
Number of pages16
ISBN (Print)9789819651221
DOIs
Publication statusPublished - 2025
Externally publishedYes
Event1st International Conference on Intelligent Multilingual Information Processing, IMLIP 2024 - Beijing, China
Duration: 16 Nov 202417 Nov 2024

Publication series

NameCommunications in Computer and Information Science
Volume2395 CCIS
ISSN (Print)1865-0929
ISSN (Electronic)1865-0937

Conference

Conference1st International Conference on Intelligent Multilingual Information Processing, IMLIP 2024
Country/TerritoryChina
CityBeijing
Period16/11/2417/11/24

Keywords

  • Instruction Tuning
  • Language Models
  • Parameter Efficient Fine-tuning

Fingerprint

Dive into the research topics of 'Instruction Fine-Tuning Guidance: How PEFT Methods Impact the Generation of Language Model via Different Attributions'. Together they form a unique fingerprint.

Cite this