Small Changes, Big Impact: Demographic Bias in LLM-Based Hiring Through Subtle Sociocultural Markers in Anonymised Resumes

This paper demonstrates that even when explicit personally identifiable information is removed, Large Language Models used in hiring can still exhibit significant demographic bias by inferring ethnicity and gender from subtle sociocultural markers in anonymized resumes, leading to systematic unfairness that is often amplified by explanation prompting.

Bryan Chen Zhengyu Tan, Shaun Khoo, Bich Ngoc Doan, Zhengyuan Liu, Nancy F. Chen, Roy Ka-Wei Lee

Published 2026-03-06
📖 5 min read🧠 Deep dive

Imagine you are hiring a new employee. To be fair, you decide to hide their name, photo, and address on their resume. You think, "Great! Now the computer (or AI) can't see if they are a man or a woman, or what their race is. It will just pick the best person based on their skills."

This paper says: You are wrong. Even with the names hidden, the AI can still guess who the person is, and it will treat them unfairly based on those guesses.

Here is the story of the paper, broken down with some simple analogies.

1. The "Invisible Ink" Problem

The researchers in Singapore decided to test this. They created 100 fake job descriptions (like "Staff Nurse" or "Software Engineer"). Then, they created 100 "perfectly neutral" resumes. These resumes had the same education and work experience, but they were missing the "personal" details.

Next, they played a game of "Resume Dress-Up." They took those neutral resumes and added tiny, seemingly innocent details to them, like:

  • Languages spoken: (e.g., "Mandarin" vs. "Tamil" vs. "Dutch").
  • Hobbies: (e.g., "Building custom PCs" vs. "Baking pastries").
  • Volunteering: (e.g., "Helping at a Mosque" vs. "Helping at a Temple").
  • School Clubs: (e.g., "Silat martial arts" vs. "Robotics Club").

They created 4,100 versions of these resumes, representing different combinations of ethnicity (Chinese, Malay, Indian, Caucasian) and gender (Male, Female).

The Analogy: Imagine you are trying to guess a person's favorite food without asking them. If you see they are wearing a red scarf, eating a specific type of bread, and listening to a specific genre of music, you might guess, "Ah, they probably like spicy curry." Even though they never said "I am Indian," their "accessories" gave it away. The researchers found that AI models are like super-sleuths; they can guess a person's background just by looking at these "accessories" on a resume.

2. The "Magic 8-Ball" Test

The researchers asked 18 different AI models (like the brains behind ChatGPT, Gemini, and others) to act as hiring managers. They gave them two types of tests:

  • Test A (The 1v1 Showdown): "Here are two resumes. One is the neutral one, one has the 'hobbies' added. Who do you pick?"
  • Test B (The Scoreboard): "Here are 41 resumes for one job. Rate them from 1 to 100. Who gets the top score?"

The Result: The AI didn't just pick the best candidate. It picked candidates based on their "accessories."

  • The Winners: Resumes with markers suggesting a Chinese Male or a Caucasian Male consistently got higher scores and were picked more often.
  • The Losers: Resumes suggesting a Malay Female or Indian Female were consistently ranked lower, even though their actual job skills were identical to the winners.

The Analogy: It's like a race where everyone starts at the same line with the same shoes. But the AI is the referee, and it secretly gives a head start to runners wearing red hats and a heavy backpack to runners wearing blue hats. The runners in blue hats are just as fast, but they never win because the referee is biased against their hat color.

3. The "Why" (The Clues)

The researchers wanted to know how the AI was guessing. They played "Remove the Clues" (Ablation study):

  • To guess Ethnicity: The AI mostly looked at Languages. If a resume said "Mandarin," the AI knew it was likely Chinese. If it said "Tamil," it knew it was likely Indian.
  • To guess Gender: The AI looked at Hobbies and Activities. "Building PCs" or "MMA" signaled "Male." "Baking" or "Yoga" signaled "Female."

The Analogy: It's like a detective solving a mystery. The "Language" clue is a giant neon sign saying "I am from this country." The "Hobby" clue is a smaller sign saying "I am likely a man or a woman." The AI reads both signs and makes a decision.

4. The "Explain Yourself" Trap

A common idea is: "If we ask the AI to explain why it picked someone, it will be more careful and fair."
The researchers tested this. They asked the AI: "Pick a winner, and tell us why."

The Result: This actually made things worse. When the AI had to explain its choice, it often doubled down on the stereotypes. It would say things like, "I picked the Chinese male because he has a strong technical background," when in reality, the "technical background" was just a hobby like "building PCs" that the AI associated with men.

The Analogy: It's like asking a biased judge to write down their reasoning. Instead of saying, "I picked him because he's better," the judge writes, "I picked him because he looks like the kind of person who usually wins." The explanation didn't fix the bias; it just gave the bias a voice.

5. The Big Takeaway

The paper concludes that hiding names is not enough.

As long as resumes contain "sociocultural markers" (languages, hobbies, volunteering, clubs), AI models will use them to guess who the person is and treat them unfairly.

  • The Systemic Issue: The AI has learned that "Chinese Male" = "Good Hire" and "Malay Female" = "Risky Hire," based on patterns in its training data, not on actual skill.
  • The Solution: Companies can't just rely on "anonymized" resumes. They need to:
    1. Test their AI tools specifically for these hidden biases before using them.
    2. Be careful about asking AI to "explain" its choices, as it might just be rationalizing its prejudice.
    3. Understand that "small changes" (like changing a hobby from "Chess" to "Knitting") can lead to "Big Impact" (getting the job or getting rejected).

In a nutshell: If you want a fair hiring process, you can't just cover the candidate's face; you have to blind the AI to the cultural "uniform" they are wearing, or else the AI will still know who they are and judge them for it.