Fairness under Graph Uncertainty: Achieving Interventional Fairness with Partially Known Causal Graphs over Clusters of Variables

This paper proposes a learning framework that achieves interventional fairness under limited causal knowledge by leveraging a graph over variable clusters and minimizing the worst-case distributional discrepancy using a computationally efficient barycenter kernel maximum mean discrepancy (MMD).

Yoichi Chikahara

Published 2026-03-02
📖 5 min read🧠 Deep dive

Imagine you are a hiring manager at a company. You want to pick the best candidate for a job, but you also want to make sure you aren't discriminating against people based on their gender or race. This is the classic problem of Algorithmic Fairness.

Usually, computers make these decisions by looking at a person's resume (features) and predicting if they will be good at the job. But here's the catch: sometimes a resume contains hidden clues about a person's race or gender that lead to unfair bias, even if the computer doesn't "know" it's doing so.

To fix this, researchers use Causal Graphs. Think of a causal graph as a family tree of cause-and-effect. It maps out how one thing leads to another. For example:

  • Gender \rightarrow Education \rightarrow Job Score.
    If the computer sees "Education" and "Job Score," it might unfairly penalize someone because their gender influenced their education. To be truly fair, we need to "intervene" in this family tree—imagine magically changing a person's gender in a simulation to see if their job score would still be the same. If the score changes, the system is unfair.

The Problem: The Map is Missing

The big issue with current methods is that they require a perfect, detailed map of this family tree. They need to know exactly how every single variable connects to every other variable.

  • The Reality: In the real world, we rarely have this perfect map. We might know that "Education" and "Job Score" are related, but we don't know the exact path. Trying to draw the whole map from scratch is like trying to map every single street in a massive city just by looking at a few satellite photos. It's slow, expensive, and often leads to mistakes.

The Solution: The "Cluster" Shortcut

This paper proposes a clever workaround. Instead of trying to map every single street (variable), the authors suggest grouping related streets into neighborhoods (clusters).

  • The Old Way: Try to map every single house and street in the city. (Hard, slow, prone to errors).
  • The New Way: Group the city into neighborhoods. Map the roads between neighborhoods. (Much easier, faster, and more robust).

In this paper, the "neighborhoods" are groups of variables (like grouping "Education" and "Work Experience" into one "Background" cluster). The authors show that figuring out the connections between these clusters is much easier than figuring out connections between individual variables.

How It Works: The "Worst-Case" Safety Net

Since the map of the neighborhoods isn't perfect (we don't know exactly what's happening inside every neighborhood), the authors use a safety-first strategy.

  1. Guess the Paths: They look at their "cluster map" and list all the possible ways the bias could travel from the sensitive attribute (like gender) to the decision.
  2. The "What-If" Test: They simulate the hiring process for every possible path on their list.
  3. The Penalty: If the computer makes a decision that looks unfair in even one of these possible scenarios, the system gets a "penalty." It's like a teacher grading a student: if the student's answer is wrong in any possible interpretation of the question, they lose points.
  4. The Fix: The computer learns to adjust its decisions until it passes the fairness test for all possible scenarios simultaneously.

The Secret Sauce: The "Barycenter"

To make this math work fast, they invented a new way to measure "unfairness" called the Barycenter Kernel MMD.

  • The Analogy: Imagine you have 100 different groups of people (based on race, gender, etc.). Instead of comparing every single person to every other person (which takes forever), you find the "average person" (the barycenter) for each group.
  • Then, you just measure how far each group's "average person" is from the "grand average."
  • This is like comparing the average height of a basketball team to the average height of a gymnastics team, rather than measuring every single player against every other player. It's much faster and scales well even when you have many different groups to check.

The Results

The authors tested this on fake data and real-world datasets (like credit scoring and hiring).

  • Accuracy vs. Fairness: Usually, making a system fairer makes it less accurate (like a strict teacher who fails everyone to be "fair").
  • The Winner: Their method found the sweet spot. It was fairer than other methods that tried to guess the full map, and it was more accurate than methods that just ignored the sensitive data entirely.

In a Nutshell

This paper teaches us that when we don't have a perfect map of the world, we shouldn't try to draw one from scratch. Instead, we should group things together, map the big picture, and then use a safety net to ensure that no matter how the details turn out, our decisions remain fair. It's a smarter, faster, and more practical way to build AI that doesn't discriminate.

Get papers like this in your inbox

Personalized daily or weekly digests matching your interests. Gists or technical summaries, in your language.

Try Digest →