Missingness Bias Calibration in Feature Attribution Explanations

This paper introduces MCal, a lightweight post-hoc method that effectively corrects missingness bias in feature attribution explanations by fine-tuning a simple linear head on frozen models, outperforming or matching expensive retraining approaches across diverse medical benchmarks.

Shailesh Sridhar, Anton Xue, Eric Wong

Published 2026-03-06
📖 5 min read🧠 Deep dive

Imagine you have a brilliant, world-class doctor (the AI model) who can diagnose diseases by looking at X-rays or reading patient histories. This doctor is incredibly accurate when looking at a complete, clear picture.

But now, imagine you want to understand why the doctor made that diagnosis. You ask, "Which part of this X-ray made you think it's a tumor?" To find out, you try a simple experiment: you take a marker and scribble out (or "ablate") random parts of the X-ray to see if the doctor still gets the right answer.

The Problem: The "Blank Page" Panic
Here's the catch: When you scribble out parts of an X-ray, you aren't just removing information; you are creating a weird, distorted image that the doctor has never seen before. It's like showing a human a photo of a cat where half the face is a black square. The human might get confused, panic, and say, "I don't know, that looks like a healthy dog!"

In the world of AI, this is called Missingness Bias.

  • The AI gets confused by the "scribbles."
  • It starts guessing randomly or leaning toward a default answer (like "Healthy").
  • Because the AI is confused, the explanation it gives ("I think it's a tumor because of this spot") becomes a lie. It's not explaining its real logic; it's just reacting to the weird, scribbled mess you created.

This is dangerous. If a doctor (or an AI) gives a bad explanation because of a bad test, we can't trust their real diagnoses.

The Old Solutions: The Heavy Hammers
Previously, experts tried to fix this with massive, expensive solutions:

  1. Retraining: Teaching the AI from scratch how to handle scribbled images. (Like hiring a new teacher to re-educate the whole school).
  2. Architectural Changes: Rewiring the AI's brain to have special "scribble-proof" neurons. (Like rebuilding the school building to be earthquake-proof).
  3. Smart Filling: Trying to guess what the scribbled parts should have looked like and filling them in. (Like an artist trying to paint over the marker with a perfect forgery).

These methods are slow, expensive, and often impossible if you don't own the AI (like if you are using a service from a big tech company).

The New Solution: MCal (The "Translator" or "Tuning Knob")
The authors of this paper, Shailesh, Anton, and Eric, say: "Wait a minute. The AI isn't actually broken deep down. It's just that its output gets a little scrambled when it sees scribbles. We don't need to rebuild the brain; we just need to fix the translation."

They introduce MCal (Missingness Calibration).

The Analogy: The Radio Tuner
Imagine the AI is a radio station playing music.

  • Clean Input: The radio plays perfectly clear music.
  • Scribbled Input (Missingness): The radio starts picking up static and the volume gets weird. The music is still there, but it sounds distorted.
  • Old Solutions: You try to rebuild the radio tower or replace the entire radio.
  • MCal: You just turn a small tuning knob (a simple linear adjustment) on the radio. You don't change the music or the tower; you just correct the static so the music sounds right again.

How MCal Works (The Simple Version)

  1. Freeze the Brain: They take the original, powerful AI and lock it so it can't change.
  2. Add a Tiny Head: They attach a very small, simple "adapter" (a linear head) to the end of the AI. Think of this as a tiny translator that sits between the AI and the final answer.
  3. The Training: They show the AI some scribbled images and some clean images. They teach the tiny translator: "When the AI sees a scribbled image and gets confused, just adjust the final numbers slightly to match what it would have said if the image were clean."
  4. The Result: The translator learns to "calibrate" the confusion. Now, even when you scribble on the image, the translator fixes the AI's panic, and the explanation becomes accurate again.

Why This is a Big Deal

  • It's Cheap: You don't need supercomputers. It takes seconds to train this tiny "translator."
  • It's Universal: It works on images, text, and spreadsheets. It doesn't matter what kind of AI you have; you can just bolt this translator onto it.
  • It's Safe: The math guarantees that this little translator will find the best possible fix every time. No guessing, no trial-and-error.
  • It Beats the Giants: Surprisingly, this tiny, cheap fix often works better than the massive, expensive retraining methods.

In Summary
The paper argues that when AI explanations go wrong because we "scribble" on the input, we don't need to rebuild the AI. We just need a simple, lightweight "tuning knob" (MCal) to correct the AI's confusion. It's a cheap, fast, and reliable way to make sure our AI doctors are telling the truth about why they made their decisions.