Apparent Age Estimation: Challenges and Outcomes

This paper reviews and evaluates distribution learning techniques for apparent age estimation, demonstrating that while methods like AMRL achieve high accuracy, significant demographic biases persist due to inconsistent feature focus, necessitating diverse datasets and rigorous fairness protocols beyond mere technical improvements.

Justin Rainier Go, Lorenz Bernard Marqueses, Mikaella Kaye Martinez, John Kevin Patrick Sarmiento, Abien Fred Agarap

Published 2026-04-07
📖 4 min read☕ Coffee break read

Imagine you walk into a room and instantly guess someone's age based on how they look. Maybe they have wrinkles, maybe their skin is glowing, or maybe they just have a "young at heart" vibe. That guess is what computer scientists call Apparent Age Estimation.

This paper is like a report card for a group of computer programs (AI models) trying to make that same guess. The researchers from De La Salle University wanted to see if these programs are good at guessing, but more importantly, if they are fair to everyone, regardless of their race or gender.

Here is the story of their findings, broken down into simple concepts:

1. The Problem: The AI is Biased

Think of the AI models as students taking a test. To study for the test, they were given a massive stack of flashcards (datasets) featuring mostly famous people from movies and Wikipedia.

  • The Issue: The flashcards were heavily skewed. They had way more pictures of white men than anyone else.
  • The Result: The AI became a "whiz kid" at guessing the age of white men but struggled terribly when looking at Asian or African American women. It was like a student who studied only for math but was suddenly asked to solve poetry problems—they just didn't have the right tools.

2. The Experiment: Trying New Study Methods

The researchers tried to fix this by teaching the AI three different ways to learn:

  • Method A (The Old Way): Just memorizing the answers (Cross-Entropy Loss).
  • Method B (The Statistical Way): Looking at the "average" and how much the answers usually vary (Mean-Variance Loss).
  • Method C (The Two-Step Way): First guessing a rough age, then making a small correction to get it right (Adaptive Mean-Residue Loss or AMRL).

The Winner: Method C (AMRL) was the smartest student. It got the most accurate guesses overall.

3. The Twist: Accuracy vs. Fairness

Here is where it gets tricky. Even though Method C was the most accurate on average, it still had blind spots.

  • The "Eye" Test: The researchers used a special tool called a "saliency map" (think of it as a heat map that shows where the AI is looking).
    • When looking at a white male, the AI correctly focused on the eyes and mouth.
    • When looking at an Asian or African American woman, the AI got confused. It started looking at the neck, the forehead, or even the background! It was like a detective looking at the wrong clues.
  • The Trade-off: They found that if they taught the AI using a more diverse set of photos (the FairFace dataset), the AI became much fairer. It stopped making huge mistakes for specific groups, even if its overall average score dropped slightly. It was like a teacher deciding to grade everyone fairly, even if it meant the top student's score went down a little.

4. Why Should You Care? (The Real World)

You might think, "So what if a computer guesses age wrong?" But this technology is already being used in the real world:

  • Cosmetics: Brands use it to recommend skincare. If the AI thinks a 30-year-old Asian woman is 50 because it's biased, it might sell her the wrong anti-aging cream.
  • Security: Banks use it to stop fraud. If the AI thinks a teenager looks like an adult (or vice versa) because of their race, it could deny a legitimate customer service or let a fraudster through.
  • The Philippines Context: The authors point out that most of these AI models are trained on Western data. If you use them in the Philippines, they are like a tourist trying to navigate Manila using a map of New York—they will get lost and make mistakes.

5. The Bottom Line

The paper concludes with a simple message: You can't just fix the math; you need to fix the data.

To make AI that is both smart and fair, we need to:

  1. Stop using only Western faces to train these models.
  2. Create local datasets (like photos of Filipino celebrities) so the AI learns what our faces look like as they age.
  3. Be careful with privacy, because facial data is sensitive.

In a nutshell: The AI is getting better at guessing age, but it's still a bit racist and sexist because it was taught by a biased teacher. To fix it, we need to give it a more diverse classroom and teach it to look at the whole picture, not just the parts it's used to seeing.

Get papers like this in your inbox

Personalized daily or weekly digests matching your interests. Gists or technical summaries, in your language.

Try Digest →