Large Language Models (LLMs) have become ubiquitous in today’s technological landscape, boasting a plethora of applications and even endangering human jobs in complex and creative fields. One such field is journalism: LLMs are being used for summarization, generation, and even fact-checking. However, in today’s political landscape, LLMs could accentuate tensions if they exhibit political bias. In this work, we evaluate the political bias of the 15 most-used multilingual LLMs via the Political Compass Test. We test different scenarios, where we vary the language of the prompt while also assigning a nationality to the model. We evaluate models on the 50 most populous countries and their official languages. Our results indicate that language has a strong influence on the political ideology displayed by a model. In addition, smaller models tend to display a more stable political ideology, i.e. ideology that is less affected by variations in the prompt.

doi.org/10.18653/v1/2025.findings-acl.883
The 63rd Annual Meeting of the Association for Computational Linguistics
creativecommons.org/licenses/by/4.0/
Human-Centered Data Analytics

Helwe, C., Balalau, O., & Ceolin, D. (2025). Navigating the political compass: Evaluating multilingual LLMs across languages and nationalities. In Findings of the Association for Computational Linguistics: ACL 2025 (pp. 17179–17204). doi:10.18653/v1/2025.findings-acl.883