Exploring cultural commonsense in multilingual large language models: A survey

  • Geleta Negasa Binegde
  • , Huaping Zhang*
  • *Corresponding author for this work

Research output: Contribution to journalArticlepeer-review

Abstract

Large language models (LLMs) have demonstrated impressive proficiency in multilingual natural language processing (NLP), yet they frequently struggle with cultural commonsense—the implicit knowledge shaped by societal norms, traditions, and shared experiences. As these models are deployed in diverse linguistic and cultural settings, their ability to understand and apply cultural commonsense becomes crucial for ensuring fairness, inclusivity, and contextual accuracy. This paper presents a systematic review and a large-scale empirical benchmark for evaluating cultural commonsense in multilingual LLMs. Through a comprehensive evaluation of 15 models on the BLEnD dataset, our analysis reveals a critical performance gap of 64.2% between high-resource and low-resource cultures. The results demonstrate significant disparities across model architectures: encoder-only models show more consistent but lower overall performance compared to decoder-based models. We identify key limitations, including data scarcity, representational bias, and inadequate cross-lingual knowledge transfer. Finally, we propose future research directions, such as culturally diverse dataset curation, hybrid knowledge graph architectures, and fairness-aware fine-tuning. The primary contributions of this work are (1) a systematic review of challenges and mitigation strategies for cultural commonsense; (2) a large-scale empirical benchmark that evaluates 15 multilingual LLMs across 13 languages and 16 countries, revealing significant performance disparities; and (3) concrete findings on the effects of model architecture and the limitations of scale in cultural understanding. This research underscores the urgent need to advance cultural commonsense in multilingual LLMs to ensure the development of fair, inclusive, and contextually accurate AI systems globally.

Original languageEnglish
Article number102649
JournalInformation Systems
Volume138
DOIs
Publication statusPublished - Jun 2026
Externally publishedYes

Keywords

  • Commonsense knowledge
  • Cultural bias
  • Cultural understanding
  • Multilingual LLMs

Fingerprint

Dive into the research topics of 'Exploring cultural commonsense in multilingual large language models: A survey'. Together they form a unique fingerprint.

Cite this