1. Measuring Spiritual Values and Bias of Large Language Models
- Author
-
Liu, Songyuan, Zhang, Ziyang, Yan, Runze, Wu, Wei, Yang, Carl, and Lu, Jiaying
- Subjects
Computer Science - Computation and Language - Abstract
Large language models (LLMs) have become integral tool for users from various backgrounds. LLMs, trained on vast corpora, reflect the linguistic and cultural nuances embedded in their pre-training data. However, the values and perspectives inherent in this data can influence the behavior of LLMs, leading to potential biases. As a result, the use of LLMs in contexts involving spiritual or moral values necessitates careful consideration of these underlying biases. Our work starts with verification of our hypothesis by testing the spiritual values of popular LLMs. Experimental results show that LLMs' spiritual values are quite diverse, as opposed to the stereotype of atheists or secularists. We then investigate how different spiritual values affect LLMs in social-fairness scenarios e.g., hate speech identification). Our findings reveal that different spiritual values indeed lead to different sensitivity to different hate target groups. Furthermore, we propose to continue pre-training LLMs on spiritual texts, and empirical results demonstrate the effectiveness of this approach in mitigating spiritual bias., Comment: 9 pages including appendix; 5 figures; 5 tables; submitted to ARR - Octobor 2024
- Published
- 2024