Probing Pre-Trained Language Models for Cross-Cultural Differences in Values

Research output: Chapter in Book/Report/Conference proceedingArticle in proceedingsResearchpeer-review

Documents

  • Fulltext

    Final published version, 2.48 MB, PDF document

Language embeds information about social, cultural, and political values people hold. Prior work has explored potentially harmful social biases encoded in Pre-trained Language Models (PLMs). However, there has been no systematic study investigating how values embedded in these models vary across cultures. In this paper, we introduce probes to study which cross-cultural values are embedded in these models, and whether they align with existing theories and cross-cultural values surveys. We find that PLMs capture differences in values across cultures, but those only weakly align with established values surveys. We discuss implications of using mis-aligned models in cross-cultural settings, as well as ways of aligning PLMs with values surveys.
Original languageEnglish
Title of host publicationProceedings of the First Workshop on Cross-Cultural Considerations in NLP (C3NLP)
PublisherAssociation for Computational Linguistics (ACL)
Publication date2023
Pages114-130
DOIs
Publication statusPublished - 2023
Event1st Workshop on Cross-Cultural Considerations in NLP, C3NLP 2023 - Dubrovnik, Croatia
Duration: 5 May 2023 → …

Conference

Conference1st Workshop on Cross-Cultural Considerations in NLP, C3NLP 2023
LandCroatia
ByDubrovnik
Periode05/05/2023 → …

ID: 381220763