Large language models show fragile cognitive reasoning about human emotions

This paper introduces CoRE, a benchmark based on cognitive appraisal theory, to reveal that while large language models capture systematic relations between cognitive appraisals and emotions, they exhibit misalignment with human judgments and instability across contexts, indicating fragile cognitive reasoning about human emotions.

Sree Bhattacharyya, Evgenii Kuriabov, Lucas Craig, Tharun Dilliraj, Reginald B. Adams, Jr., Jia Li, James Z. Wang

Published 2026-03-16
📖 5 min read🧠 Deep dive

The Big Idea: The "Emotion Detective" vs. The "Emotion Actor"

Imagine you hire a very smart robot detective to solve a mystery about why a person is crying.

  • The Old Way: You ask the robot, "Is this person sad?" The robot looks at its training manual, sees a picture of a crying face, and says, "Yes, that's Sadness." It's good at matching pictures to labels.
  • The New Way (This Paper): You ask the robot, "Why do you think they are sad?" You want it to explain the reasoning: "They are sad because they lost something they loved, they feel responsible for it, and they can't control the outcome."

This paper asks: Do Large Language Models (LLMs) actually understand the "Why" (the reasoning), or are they just really good at guessing the "What" (the label)?

The answer is a bit scary: They are great at guessing the label, but their internal reasoning is "fragile," inconsistent, and sometimes completely different from how humans think.


The Experiment: The "Emotion Gym"

The researchers built a massive gym called CoRE (Cognitive Reasoning for Emotions). They didn't just ask the AI, "What emotion is this?" Instead, they made the AI do a workout with 17 different mental muscles (called cognitive appraisals).

Think of these muscles as questions a human asks themselves when something happens:

  • Was this pleasant?
  • Did I cause this?
  • Was it fair?
  • How much effort did I have to use?
  • Was I certain about what would happen?

They tested 6 different AI models (like GPT, LLaMA, Gemini) on nearly 70,000 scenarios to see how they used these "muscles" to figure out emotions.

The Findings: Where the Robots Go Wrong

Here are the three main ways the AI's "emotional brain" is different from a human's:

1. The "Effort" Trap (The Over-Exercising Robot)

Humans have a balanced way of thinking. If we feel "Pride," we think, "I worked hard for this." If we feel "Happiness," we think, "This feels good."

  • The AI Glitch: The AI models treat Effort as the most important thing for almost every emotion. It's like a robot that thinks every problem in life is solved by "working harder."
  • The Metaphor: Imagine a human chef who thinks every dish needs extra salt to taste good. If the dish is sweet, salty, or spicy, the chef just dumps in more salt. The AI does this with "Effort." It overuses this one concept, even when humans wouldn't.

2. The "Secret Identity" Problem (The Lying Detective)

The researchers asked the AI two things:

  1. Implicitly: "Show me your work." (We looked at the numbers the AI generated).
  2. Explicitly: "Tell me what is most important." (We asked the AI to pick one factor).
  • The Glitch: The AI's "work" showed it was heavily relying on Effort and Problems. But when asked directly, it said, "Oh no, I rely on Responsibility and Control."
  • The Metaphor: It's like a student who studies the whole textbook (the implicit reasoning) but, when the teacher asks, "What was the most important thing you learned?", they confidently say, "The cover page." The AI doesn't even know what it's actually thinking.

3. The "Cultural Blindness" (The One-Size-Fits-All Suit)

Emotions are different depending on who you are and where you are from. A situation that makes a person in Japan feel "shame" might make a person in the US feel "anger."

  • The Glitch: When the researchers told the AI to "act like a person from Japan" or "act like a person from Mexico," the AI's emotional reasoning didn't change at all.
  • The Metaphor: Imagine a weather app that says "It's raining" no matter if you are in a desert or a rainforest. The AI has a "universal" emotional suit that fits everyone, but it's too loose to fit any specific culture.
  • The Exception: However, if you told the AI to "act like a grumpy person" vs. "act like a happy person" (personality), it did change its reasoning. It understands individual moods, but not cultural backgrounds.

The "Fragility" of the AI Mind

The paper concludes that AI emotional reasoning is fragile.

  • Simple is Easy: If you ask the AI to distinguish between "Happy" (good stuff) and "Sad" (bad stuff), it does great. It understands the basic "Good vs. Bad" split.
  • Complex is Broken: If you ask it to distinguish between "Shame" and "Guilt" (which are very similar but have different rules), the AI gets confused. It starts mixing them up or using the wrong logic.

Why Does This Matter?

If we use AI to help with mental health, teach kids, or run customer service, we need them to understand why we feel the way we do, not just what we feel.

  • The Risk: If an AI thinks "Anger" is always caused by "Unfairness" (which is true for humans), but it doesn't understand that in some cultures, anger is about "Loss of Face," it might give terrible advice.
  • The Future: We can't just train AI to say the right words. We need to train them to have a "human-like" internal structure, where they understand the complex web of reasons behind our feelings, not just the surface label.

Summary in One Sentence

Current AI models are like actors who have memorized the script for "Sadness" and "Anger" perfectly, but they don't actually understand the psychology behind the lines, and they often get confused when the scene changes to a different culture or a complex situation.

Get papers like this in your inbox

Personalized daily or weekly digests matching your interests. Gists or technical summaries, in your language.

Try Digest →