Decoding the Mind of Large Language Models: A Quantitative Evaluation of Ideology and Biases
By: Manari Hirose, Masato Uchida
Potential Business Impact:
Finds AI's hidden opinions and unfair ideas.
The widespread integration of Large Language Models (LLMs) across various sectors has highlighted the need for empirical research to understand their biases, thought patterns, and societal implications to ensure ethical and effective use. In this study, we propose a novel framework for evaluating LLMs, focusing on uncovering their ideological biases through a quantitative analysis of 436 binary-choice questions, many of which have no definitive answer. By applying our framework to ChatGPT and Gemini, findings revealed that while LLMs generally maintain consistent opinions on many topics, their ideologies differ across models and languages. Notably, ChatGPT exhibits a tendency to change their opinion to match the questioner's opinion. Both models also exhibited problematic biases, unethical or unfair claims, which might have negative societal impacts. These results underscore the importance of addressing both ideological and ethical considerations when evaluating LLMs. The proposed framework offers a flexible, quantitative method for assessing LLM behavior, providing valuable insights for the development of more socially aligned AI systems.
Similar Papers
Probing the Subtle Ideological Manipulation of Large Language Models
Computation and Language
Teaches computers to understand many political ideas.
Addressing Stereotypes in Large Language Models: A Critical Examination and Mitigation
Computation and Language
Fixes AI's unfairness and improves its understanding.
Echoes of Power: Investigating Geopolitical Bias in US and China Large Language Models
Computers and Society
AI models show country biases in political answers.