Prompt-based and Fine-tuned GPT Models for Context-Dependent and -Independent Deductive Coding in Social Annotation

Chenyu Hou, Gaoxia Zhu, Juan Zheng, Lishan Zhang, Xiaoshan Huang, Tianlong Zhong, Shan Li, Hanxiang Du, Chin Lee Ker

Research output: Chapter in Book/Report/Conference proceedingConference contributionpeer-review

26 Citations (Scopus)

Abstract

GPT has demonstrated impressive capabilities in executing various natural language processing (NLP) and reasoning tasks, showcasing its potential for deductive coding in social annotations. This research explored the effectiveness of prompt engineering and fine-tuning approaches of GPT for deductive coding of context-dependent and context-independent dimensions. Coding context-dependent dimensions (i.e., Theorizing, Integration, Reflection) requires a contextualized understanding that connects the target comment with reading materials and previous comments, whereas coding context-independent dimensions (i.e., Appraisal, Questioning, Social, Curiosity, Surprise) relies more on the comment itself. Utilizing strategies such as prompt decomposition, multi-prompt learning, and a codebook-centered approach, we found that prompt engineering can achieve fair to substantial agreement with expert-labeled data across various coding dimensions. These results affirm GPT's potential for effective application in real-world coding tasks. Compared to context-independent coding, context-dependent dimensions had lower agreement with expert-labeled data. To enhance accuracy, GPT models were fine-tuned using 102 pieces of expert-labeled data, with an additional 102 cases used for validation. The fine-tuned models demonstrated substantial agreement with ground truth in context-independent dimensions and elevated the inter-rater reliability of context-dependent categories to moderate levels. This approach represents a promising path for significantly reducing human labor and time, especially with large unstructured datasets, without sacrificing the accuracy and reliability of deductive coding tasks in social annotation. The study marks a step toward optimizing and streamlining coding processes in social annotation. Our findings suggest the promise of using GPT to analyze qualitative data and provide detailed, immediate feedback for students to elicit deepening inquiries.

Original languageEnglish
Title of host publicationLAK 2024 Conference Proceedings - 14th International Conference on Learning Analytics and Knowledge
PublisherAssociation for Computing Machinery
Pages518-528
Number of pages11
ISBN (Electronic)9798400716188
DOIs
Publication statusPublished - 18 Mar 2024
Externally publishedYes
Event14th International Conference on Learning Analytics and Knowledge, LAK 2024 - Kyoto, Japan
Duration: 18 Mar 202422 Mar 2024

Publication series

NameACM International Conference Proceeding Series

Conference

Conference14th International Conference on Learning Analytics and Knowledge, LAK 2024
Country/TerritoryJapan
CityKyoto
Period18/03/2422/03/24

Keywords

  • Context-Dependent
  • Fine-tuning
  • GPT
  • Prompt Engineering
  • Social Annotation
  • deductive coding

Fingerprint

Dive into the research topics of 'Prompt-based and Fine-tuned GPT Models for Context-Dependent and -Independent Deductive Coding in Social Annotation'. Together they form a unique fingerprint.

Cite this