Behavioral characterization of dynamic facial expression perception in rhesus monkeys (Macaca mulatta) using naturalistic and synthetic stimuli

This study demonstrates that rhesus monkeys perceive facial expressions as context-dependent, functionally meaningful social signals shaped by both expression dynamics and signaler characteristics, rather than relying solely on fixed morphological categories or visual similarity.

Siebert, R., Taubert, N., Giese, M. A., Thier, P.

Published 2026-03-20
📖 4 min read☕ Coffee break read
⚕️

This is an AI-generated explanation of a preprint that has not been peer-reviewed. It is not medical advice. Do not make health decisions based on this content. Read full disclaimer

Imagine you are walking through a busy park. You see a dog barking, a cat hissing, and a squirrel chattering. Even without understanding the specific words they might be saying, you instantly know: "That dog is angry," "That cat is scared," and "That squirrel is just talking." You don't need a dictionary; you understand the vibe and the intent behind their faces.

This study asks a similar question about rhesus monkeys: Do they just see a bunch of moving muscles on a face, or do they actually understand what those faces mean?

Here is the story of how scientists figured it out, using some clever tricks and a lot of digital monkeys.

The Setup: A "Face-Off" Game

The researchers taught two real-life rhesus monkeys a game. They showed the monkeys videos of other monkeys making four specific faces:

  1. Neutral: Just chilling.
  2. Lip-Smacking: A friendly "hello" or "I'm happy."
  3. Silent Bared-Teeth (SBT): A "fear grin" (showing teeth but not biting), often meaning "I'm scared" or "I submit."
  4. Open-Mouth Threat: A scary "Back off!" face.

The monkeys had to press a button corresponding to the face they saw to get a juice reward. Once they got good at it, the scientists threw in a curveball: New monkeys they had never seen before.

The Big Discovery: It's Not Just About the Shape

If the monkeys were just matching shapes (like a child matching a red circle to another red circle), they would have gotten confused when the faces looked slightly different. But they didn't.

The "Aha!" Moment:
The scientists measured the monkeys' pupils. In humans and animals, pupils get bigger when we are excited, scared, or alert.

  • When the monkeys saw the Threat face, their pupils dilated (got huge). They were on high alert!
  • When they saw the Neutral face, their pupils stayed small. They were relaxed.

This proved that the monkeys weren't just memorizing shapes; they understood the meaning. They knew the threat face meant "Danger!" and the lip-smack meant "Friend!"

The Twist: Context is King

Here is where it gets really interesting. The monkeys didn't treat every "Fear Grin" (SBT) the same way. Sometimes they thought it was scary; other times they thought it was friendly. Why?

The scientists realized the monkeys were looking at the whole story, not just the mouth:

  • Body Size: If a big, heavy monkey made the "Fear Grin," the other monkeys thought, "Hmm, that big guy isn't actually scared; he's probably just being polite."
  • Eye Contact: If the monkey looked away while making the face, it changed the meaning.
  • Ears and Eyebrows: Just like humans raise their eyebrows when surprised, the monkeys used their ears and eyebrows to add nuance. A "Fear Grin" with lowered eyebrows meant "I'm terrified," but with raised eyebrows, it might mean something else.

The Analogy: Think of it like a text message. If you send "I'm fine," it's neutral. But if you send "I'm fine" with a crying emoji, it means "I'm sad." If you send it with a fire emoji, it means "I'm furious." The monkeys read the "emojis" (ears, eyes, body size) to understand the real message.

The Digital Lab: Testing with Avatars

To test exactly what parts of the face mattered, the scientists created digital avatars (computer-generated monkeys). They played with these avatars like a video game:

  • The "Uncanny Valley" Test: They removed the fur and color, leaving just a wireframe skeleton. The monkeys couldn't recognize the faces anymore. Lesson: You need to see a face that looks somewhat real (skin and fur) to understand the emotion.
  • The "Stop-Motion" Test: They scrambled the video frames so the movement was jerky and made no sense. Surprisingly, the monkeys still understood the faces! Lesson: The specific motion isn't as important as the final shape the face makes.
  • The "Human" Test: They put human faces (smiling, angry, scared) onto the monkey avatars. The monkeys were completely confused. They didn't know what to do. Lesson: Monkeys don't have a universal dictionary for all primates. They only understand their own species' "language." A human smile doesn't mean "friend" to a monkey; it's just a weird face.

The Takeaway

This study teaches us that rhesus monkeys are not just face-matching machines. They are social detectives.

They don't just look at a mouth; they look at the whole picture. They combine the mouth shape, the eye direction, the body size, and the ear position to figure out: Is this monkey a friend? Is it a threat? Is it scared?

It's like reading a book where the words are the mouth movements, but the tone of voice, the setting, and the character's size are the ears, eyes, and body. The monkeys read the whole story, not just the words.

Get papers like this in your inbox

Personalized daily or weekly digests matching your interests. Gists or technical summaries, in your language.

Try Digest →