Location Not Found: Exposing Implicit Local and Global Biases in Multilingual LLMs

📰 ArXiv cs.AI

arXiv:2604.19292v1 Announce Type: cross Abstract: Multilingual large language models (LLMs) have minimized the fluency gap between languages. This advancement, however, exposes models to the risk of biased behavior, as knowledge and norms may propagate across languages. In this work, we aim to quantify models' inter- and intra-lingual biases, via their ability to answer locale-ambiguous questions. To this end, we present LocQA, a test set containing 2,156 questions in 12 languages, referring to

Published 22 Apr 2026
Read full paper → ← Back to Reads