Automatic pain face analysis in mice: Applied to a varied dataset with non-standardized conditions

This study introduces a deep learning model trained on a large, diverse dataset that outperforms human raters in automatically assessing Mouse Grimace Scale scores across varied strains and conditions, demonstrating that combining multiple subsets yields the most reliable performance for non-standardized pain assessment.

Andresen, N., Wöllhaf, M., Wilzopolski, J., Lang, A., Wolter, A., Howe-Wittek, L., Bekemeier, C., Pawlak, L.-I., Beyer, S., Cynis, H., Hietel, E., Rieckmann, V., Rieckmann, M., Thöne-Reineke, C., Lewejohann, L., Hellwich, O., Hohlbaum, K.

Published 2026-02-18
📖 5 min read🧠 Deep dive
⚕️

This is an AI-generated explanation of a preprint that has not been peer-reviewed. It is not medical advice. Do not make health decisions based on this content. Read full disclaimer

Imagine you are a detective trying to solve a mystery, but your suspect is a tiny mouse that can't speak. The mystery? Is the mouse in pain?

For decades, scientists have used a tool called the Mouse Grimace Scale (MGS) to solve this. It's like a "pain emoji" chart for mice. If a mouse is hurting, its face changes in five specific ways: its eyes squint tight, its nose and cheeks puff out, its ears stick out, and its whiskers straighten or curl back.

Usually, a human has to look at a photo of the mouse and manually score these changes. But humans get tired, they might be biased, and they can't watch the mouse 24 hours a day. Plus, if a human walks into the room, the mouse might get scared and hide its pain, just like a kid might stop crying when a parent walks in.

The Problem: The "One-Size-Fits-All" Failure
Scientists tried to build a computer program (an AI) to do this job automatically. But here's the catch: most of these AI programs were like students who only studied for one specific test.

  • If you trained an AI on black mice in a bright lab, it would fail miserably when shown a white mouse in a dim cage.
  • If it learned from mice that had surgery, it might get confused by mice that just had a blood test.
  • The real world is messy. Mice come in different colors (black, white, brown), live in different cages, and are filmed with different cameras. The AI was getting confused by all these variables, like a chef who only knows how to cook with a specific brand of stove and fails when given a different one.

The Solution: The "Super-Student" AI
This paper introduces a massive new project designed to create a "Super-Student" AI that can handle any situation.

  1. The Massive Library: Instead of looking at a few hundred photos, the researchers gathered a library of 35,000 images from five different laboratories. These images feature mice of different breeds, different fur colors, different ages, and different medical treatments. It's like training a student not just on one textbook, but on every book in the library, written in different languages and styles.
  2. The Training Method: They didn't just throw the photos at the computer. They used a clever "two-step" training method:
    • Step 1 (The Warm-up): First, they taught the AI a simple game: "Is this mouse in pain or not?" (Yes/No). This helped the AI learn to recognize the general "vibe" of a hurting mouse.
    • Step 2 (The Exam): Once the AI understood the basics, they taught it the hard math: "Give me a specific score from 0 to 2 for how much pain it's in."
  3. The Result: The AI became incredibly good. When tested on new, unseen mice, it made an average error of 0.26 on a scale of 0 to 2.
    • The Analogy: Imagine a human expert guessing the temperature of a room. They might be off by a degree or two. This AI was actually more accurate than the average human expert! It correlated with human experts at a level of 85%, which is like two best friends finishing each other's sentences.

The "Cross-Training" Surprise
The researchers also tested if the AI could handle a "cross-dataset" challenge. This is like taking a student who studied in New York and testing them in Tokyo.

  • The Bad News: If you trained the AI on just one type of mouse (e.g., only black mice), it struggled when shown a white mouse.
  • The Good News: The AI trained on the combined mix of all five groups was the most robust. It learned to ignore the "noise" (like the color of the cage or the fur) and focus only on the "signal" (the actual pain expression).
  • The Twist: Interestingly, the AI didn't get better by focusing only on the easiest feature (squinting eyes). It actually performed better when it looked at the whole face, just like a human does.

Why This Matters
This is a huge step forward for animal welfare.

  • 24/7 Monitoring: Instead of a human checking the mouse once a day, this AI could watch the mouse 24 hours a day, even while the mouse is sleeping or in the dark (when mice are most active).
  • No Stress: The mouse doesn't need to be taken out of its home cage and put in a scary box for a photo. The AI can analyze the mouse right in its home, meaning the mouse stays calm and shows its true feelings.
  • Better Science: If we know exactly when a mouse is in pain, we can treat it faster, and the scientific data we get from experiments is more accurate because pain doesn't skew the results.

In a Nutshell
The researchers built a "universal translator" for mouse faces. By feeding a computer a massive, diverse diet of images, they taught it to understand the universal language of pain, regardless of the mouse's color, size, or where it lives. This allows for a kinder, more scientific way to care for our tiny laboratory friends.

Get papers like this in your inbox

Personalized daily or weekly digests matching your interests. Gists or technical summaries, in your language.

Try Digest →