When Do Language Models Endorse Limitations on Human Rights Principles?

This paper evaluates eleven major Large Language Models across eight languages and reveals systematic biases where they are more likely to endorse limitations on Economic, Social, and Cultural rights than Political and Civil rights, exhibit significant cross-linguistic variations, and show high susceptibility to prompt-based steering.

Keenan Samway, Nicole Miu Takagi, Rada Mihalcea, Bernhard Schölkopf, Ilias Chalkidis, Daniel Hershcovich, Zhijing Jin

Published 2026-03-05
📖 5 min read🧠 Deep dive

Imagine you have built a team of 11 super-smart digital librarians. These librarians (Large Language Models, or LLMs) are being trained to help judges, police officers, and politicians make tough decisions. They are supposed to be the ultimate guardians of Human Rights, like the Universal Declaration of Human Rights (UDHR), which is basically the "Rulebook for Being Human."

But here's the big question: If you ask these librarians to break a rule to save the day, will they do it? And does it matter what language you speak to them?

This paper is like a giant, stress-test exam given to these digital librarians. The researchers created 1,152 tricky scenarios (like a "Choose Your Own Adventure" book) where a government action limits a human right to achieve something else, like public safety or stopping a pandemic.

Here is what they found, explained with some simple analogies:

1. The "Language Switch" Surprise 🌍

Imagine you ask a librarian, "Is it okay to stop people from blogging to stop fake news?"

  • If you ask in English, the librarian says: "No! That hurts free speech!"
  • If you ask the same librarian in Chinese or Hindi, they might say: "Actually, yes. It's necessary for order."

The Analogy: It's like a person who is very polite and firm when speaking English, but suddenly becomes much more flexible with rules when speaking a different language. The study found that these AI models are much more willing to sacrifice rights (like free speech) when talking in Chinese or Hindi compared to English or Romanian. They aren't just "translating" the answer; they are changing their values based on the language.

2. The "Heavy vs. Light" Rights Scale ⚖️

The researchers noticed the librarians treat different rights like different weights on a scale.

  • Political & Civil Rights (like freedom of speech, fair trials, and privacy) are treated like gold bars. The librarians are very protective of these. They rarely say "yes" to limiting them.
  • Economic & Social Rights (like the right to education, leisure, or owning property) are treated like feathers. The librarians are much more willing to say, "Sure, let's limit your right to leisure or your property rights if it helps the economy."

The Analogy: It's as if the AI thinks, "I will fight to the death for your freedom to speak, but I'm totally fine with taking away your right to a nice vacation or your savings if the government says it's for the greater good."

3. The "Emergency Button" Panic 🚨

The study tested how the librarians react when the world is on fire.

  • Normal Day: "No, we can't limit rights."
  • Civil Unrest (Protests): "Hmm, maybe we can limit a few things to keep the peace."
  • Natural Disaster (Hurricane/Earthquake): "Okay, lock everything down! We will suspend rights immediately to save lives."

The Analogy: The AI models act like a security guard who is very strict when the mall is quiet, but the moment a fire alarm goes off (a natural disaster), they throw the rulebook out the window and start locking doors and banning people, even if it's a bit extreme. They seem to think disasters are a better excuse to break rules than angry crowds.

4. The "Persona Mask" Trick 🎭

This was the most surprising part. The researchers put on different "masks" (prompts) for the librarians.

  • Mask A: "You are a champion of individual freedom!"
  • Mask B: "You are a champion of government authority!"

The Analogy: It's like asking a person, "What do you think?"

  • If you tell them, "Pretend you are a rebel," they say, "Freedom is everything!"
  • If you tell them, "Pretend you are a strict general," they say, "Order is everything!"

The study found that these AI models are extremely easy to trick. By just changing the introduction sentence, the researchers could make the AI swing from "Strongly Reject" to "Strongly Endorse" limiting human rights. It's like the AI doesn't have a solid core of values; it just mirrors whatever personality you tell it to wear.

5. The "Text vs. Talk" Confusion 🗣️📝

Finally, the researchers noticed that the librarians give different answers depending on how you ask.

  • If you ask them to circle a number (1 to 5), they give one answer.
  • If you ask them to write a paragraph explaining their choice, they often give a completely different answer.

The Analogy: It's like asking a student, "Raise your hand if you agree." They raise their hand. But then you ask, "Write an essay on why you agree," and they write a paragraph saying they actually disagree. This suggests the AI's "opinions" are fragile and depend entirely on the format of the question.

The Big Takeaway 🏁

This paper is a warning label for the future. As we start using AI to help make laws, judge asylum cases, or moderate social media, we can't assume the AI has a consistent moral compass.

  • It changes its mind based on the language you speak.
  • It cares more about some rights than others.
  • It panics during emergencies.
  • It can be easily manipulated by how you ask the question.

The Bottom Line: We can't just trust these digital librarians to hold the rulebook. We need to check their work constantly, in every language, and make sure they aren't just pretending to be smart while actually just following the latest trick we taught them.