Rethinking Role-Playing Evaluation: Anonymous Benchmarking and a Systematic Study of Personality Effects

This paper proposes an anonymous benchmarking method to expose the bias in current role-playing evaluations caused by character name exposure and demonstrates that incorporating self-generated personality traits effectively enhances model performance under these fairer conditions.

Ji-Lun Peng, Yun-Nung Chen

Published 2026-03-05
📖 4 min read☕ Coffee break read

Imagine you are hiring an actor to play a famous movie character, like Harry Potter, in a new play.

The Old Way (The "Name Drop" Problem)
In the past, researchers tested AI actors by simply telling them, "You are Harry Potter." The AI would then start acting like Harry. But here's the catch: The AI didn't actually learn how to be Harry from the script you gave it. It just remembered everything it had read about Harry Potter in its massive training data (like a student who memorized the entire encyclopedia).

If you asked the AI, "What is Harry's favorite color?" it didn't need to think; it just recalled a fact it already knew. This made the AI look great, but it was a bit of a cheat. It wasn't proving the AI could act; it was proving the AI had a good memory.

The New Experiment: The "Blind Audition"
The authors of this paper wanted to see if AI could really act, not just remember. So, they tried something called Anonymous Benchmarking.

Imagine the director walks into the audition room and says:

"You are an orphaned boy who discovers a magical world. You are brave, a bit clumsy, and love flying on a broomstick. But we cannot tell you your name."

Suddenly, the AI can't rely on its memory of "Harry Potter." It has to build the character from scratch using only the description provided.

The Result:
When they removed the name, the AI's performance dropped significantly. It stumbled, sounded less like the character, and made mistakes. This proved that the AI was previously "cheating" by using its memory of the name rather than truly understanding the role. The "Anonymous" test is like a fairer, stricter audition that shows what the AI can actually do.

The Solution: Giving the AI a "Personality Cheat Sheet"
Once the AI was struggling in the blind audition, the researchers asked: How do we help it act better without giving it the name?

They tried giving the AI a Personality Profile. Think of this like giving an actor a detailed character sheet that says:

  • "You are an INTJ (a logical, strategic, and private type of person)."
  • "You are Open to new ideas but Neurotic (easily stressed)."

They tested two ways to get this profile:

  1. The Human Expert: A real person reads the character's story and writes down their personality type (like a casting director).
  2. The AI Detective: The AI reads the story itself and guesses, "Hmm, this character seems like an INTJ," and writes down its own guess.

The Big Surprise:
The researchers found that giving the AI a personality profile made it act much better, even without knowing the character's name.

Even more surprisingly, the AI's own guess about the personality worked just as well as the Human Expert's notes. The AI didn't need a human to tell it how to be brave or shy; it could figure out the "vibe" of the character just by reading the description and then use that "vibe" to act more convincingly.

Why This Matters

  • Fairer Tests: We can now test AI on characters it has never heard of (like a real person or a brand-new fictional character) without it cheating by using its memory.
  • Better Actors: By adding a simple "personality tag" (like MBTI or Big Five traits), we can make AI actors sound much more human and consistent, even when they are playing a role they don't know well.

In a Nutshell:
The paper says, "Stop letting AI actors cheat by memorizing names. Instead, give them a personality map. And guess what? The AI can draw its own map just as well as a human can!"