CIVICS: Building a Dataset for Examining Culturally-Informed Values in Large Language Models

Type: Preprint

Publication Date: 2024-05-22

Citations: 0

DOI: https://doi.org/10.48550/arxiv.2405.13974

Abstract

This paper introduces the "CIVICS: Culturally-Informed & Values-Inclusive Corpus for Societal impacts" dataset, designed to evaluate the social and cultural variation of Large Language Models (LLMs) across multiple languages and value-sensitive topics. We create a hand-crafted, multilingual dataset of value-laden prompts which address specific socially sensitive topics, including LGBTQI rights, social welfare, immigration, disability rights, and surrogacy. CIVICS is designed to generate responses showing LLMs' encoded and implicit values. Through our dynamic annotation processes, tailored prompt design, and experiments, we investigate how open-weight LLMs respond to value-sensitive issues, exploring their behavior across diverse linguistic and cultural contexts. Using two experimental set-ups based on log-probabilities and long-form responses, we show social and cultural variability across different LLMs. Specifically, experiments involving long-form responses demonstrate that refusals are triggered disparately across models, but consistently and more frequently in English or translated statements. Moreover, specific topics and sources lead to more pronounced differences across model answers, particularly on immigration, LGBTQI rights, and social welfare. As shown by our experiments, the CIVICS dataset aims to serve as a tool for future research, promoting reproducibility and transparency across broader linguistic settings, and furthering the development of AI technologies that respect and reflect global cultural diversities and value pluralism. The CIVICS dataset and tools will be made available upon publication under open licenses; an anonymized version is currently available at https://huggingface.co/CIVICS-dataset.

Locations

  • arXiv (Cornell University) - View - PDF

Similar Works

Action Title Year Authors
+ Assessing LLMs for Moral Value Pluralism 2023 Noam Benkler
Drisana Mosaphir
Scott Friedman
Andrew Smart
Sonja Schmer-Galunder
+ PDF Chat LocalValueBench: A Collaboratively Built and Extensible Benchmark for Evaluating Localized Value Alignment and Ethical Safety in Large Language Models 2024 Gwenyth Isobel Meadows
Nuno Lau
Eva Adelina Susanto
C. L. Yu
Aditya Paul
+ Use large language models to promote equity 2023 Emma Pierson
Divya Shanmugam
Rajiv Movva
Jon Kleinberg
Monica Agrawal
Mark Dredze
Kadija Ferryman
Judy Wawira Gichoya
Dan Jurafsky
Pang Wei Koh
+ PDF Chat WorldValuesBench: A Large-Scale Benchmark Dataset for Multi-Cultural Value Awareness of Language Models 2024 Wenlong Zhao
Debanjan Mondal
Niket Tandon
Danica Dillion
Kurt Gray
Yuling Gu
+ PDF Chat Towards "Differential AI Psychology" and in-context Value-driven Statement Alignment with Moral Foundations Theory 2024 Simon Münker
+ PDF Chat Bottom-Up and Top-Down Analysis of Values, Agendas, and Observations in Corpora and LLMs 2024 Scott Friedman
Noam Benkler
Drisana Mosaphir
Jeffrey M. Rye
Sonja Schmer-Galunder
Micah B. Goldwater
Matthew D. McLure
Ruta Wheelock
Jeremy Gottlieb
Robert P. Goldman
+ PDF Chat The PRISM Alignment Project: What Participatory, Representative and Individualised Human Feedback Reveals About the Subjective and Multicultural Alignment of Large Language Models 2024 Hannah Rose Kirk
Alexander Whitefield
Paul Röttger
Andrew Bean
Katerina Margatina
Juan Ciro
Rafael Mosquera
Max Bartolo
Adina Williams
He He
+ Value FULCRA: Mapping Large Language Models to the Multidimensional Spectrum of Basic Human Values 2023 Jing Yao
Xiaoyuan Yi
Xiting Wang
Yifan Gong
Xing Xie
+ PDF Chat The Generation Gap:Exploring Age Bias in Large Language Models 2024 Siyang Liu
Trish Maturi
Siqi Shen
Rada Mihalcea
+ PDF Chat The Potential and Challenges of Evaluating Attitudes, Opinions, and Values in Large Language Models 2024 Bolei Ma
Xinpeng Wang
Tiancheng Hu
Anna‐Carolina Haensch
Michael A. Hedderich
Barbara Plank
Frauke Kreuter
+ Evaluating and Mitigating Discrimination in Language Model Decisions 2023 Alex Tamkin
Amanda Askell
Liane Lovitt
Esin Durmus
Nicholas Joseph
Shauna Kravec
Karina Nguyen
Jared Kaplan
Deep Ganguli
+ The Ghost in the Machine has an American accent: value conflict in GPT-3 2022 Rebecca L. Johnson
Giada Pistilli
Natalia Menédez-González
Leslye Denisse Dias Duran
Enrico Panai
Julija Kalpokienė
Donald Jay Bertulfo
+ PDF Chat CLAVE: An Adaptive Framework for Evaluating Values of LLM Generated Responses 2024 Jing Yao
Xiaoyuan Yi
Xing Xie
+ The Unequal Opportunities of Large Language Models: Examining Demographic Biases in Job Recommendations by ChatGPT and LLaMA 2023 Abel Salinas
Parth Shah
Yuzhong Huang
Robert McCormack
Fred Morstatter
+ PDF Chat ValueBench: Towards Comprehensively Evaluating Value Orientations and Understanding of Large Language Models 2024 Yuanyi Ren
Haoran Ye
Hanjun Fang
Xin Zhang
Guojie Song
+ SurveyLM: A platform to explore emerging value perspectives in augmented language models' behaviors 2023 Steve J. Bickley
Ho Fai Chan
邦男 中島
Benno Torgler
Son N. Tran
+ PDF Chat ValuesRAG: Enhancing Cultural Alignment Through Retrieval-Augmented Contextual Learning 2025 Wonduk Seo
Zonghao Yuan
Yi Bu
+ Process for Adapting Language Models to Society (PALMS) with Values-Targeted Datasets 2021 Irene Solaiman
Christy Dennison
+ PDF Chat Political Compass or Spinning Arrow? Towards More Meaningful Evaluations for Values and Opinions in Large Language Models 2024 Paul Röttger
Valentin Hofmann
Valentina Pyatkin
Musashi Hinck
Hannah Rose Kirk
Hinrich Schütze
Dirk Hovy
+ PDF Chat Collective Constitutional AI: Aligning a Language Model with Public Input 2024 Saffron Huang
Divya Siddarth
Liane Lovitt
Thomas I. Liao
Esin Durmus
Alex Tamkin
Deep Ganguli

Works That Cite This (0)

Action Title Year Authors

Works Cited by This (0)

Action Title Year Authors