Whose morality do they speak? Unraveling cultural bias in multilingual language models

Research output: Contribution to journalArticlepeer-review

1 Citation (Scopus)

Abstract

Large language models (LLMs) have become integral tools in diverse domains, yet their moral reasoning capabilities across cultural and linguistic contexts remain underexplored. This study investigates whether multilingual LLMs, such as GPT-3.5-Turbo, GPT-4o-mini, Llama 3.1, and MistralNeMo, reflect culturally specific moral values or impose dominant moral norms, particularly those rooted in English. Using the updated Moral Foundations Questionnaire (MFQ-2) in eight languages, Arabic, Farsi, English, Spanish, Japanese, Chinese, French, and Russian, the study analyzes the models’ adherence to six core moral foundations: care, equality, proportionality, loyalty, authority, and purity. The results reveal significant cultural and linguistic variability, challenging the assumption of universal moral consistency in LLMs. Although some models demonstrate adaptability to diverse contexts, others exhibit biases influenced by the composition of the training data. These findings underscore the need for culturally inclusive model development to improve fairness and trust in multi-lingual AI systems.

Original languageEnglish
Article number100172
JournalNatural Language Processing Journal
Volume12
DOIs
Publication statusPublished - Sept 2025
Externally publishedYes

Bibliographical note

Publisher Copyright:
© 2025 The Author(s)

Keywords

  • Bias
  • LLMs
  • MFQ
  • Morality
  • Multilingualism

Fingerprint

Dive into the research topics of 'Whose morality do they speak? Unraveling cultural bias in multilingual language models'. Together they form a unique fingerprint.

Cite this