GIFT: A Framework Towards Global Interpretable Faithful Textual Explanations of Vision Classifiers

The paper introduces GIFT, a post-hoc framework that generates global, interpretable, and faithful textual explanations for vision classifiers by aggregating natural-language descriptions of local visual counterfactuals and verifying their causal impact through image-based interventions.

Éloi Zablocki, Valentin Gerard, Amaia Cardiel, Eric Gaussier, Matthieu Cord, Eduardo Valle

Published 2026-02-23
📖 5 min read🧠 Deep dive

Imagine you have a super-smart robot that looks at pictures and decides what they are. Maybe it tells you if a face looks "Old" or "Young," or if a car in a video can "Turn Right" safely. But here's the problem: the robot is a black box. It gives you an answer, but it won't tell you why. It's like a chef who makes a delicious soup but refuses to tell you the recipe.

The paper introduces a new tool called GIFT (Global, Interpretable, Faithful, Textual explanations). Think of GIFT as a detective that interrogates the robot to find out its secret recipe, but it does so in a way that is honest, clear, and written in plain English.

Here is how GIFT works, broken down into four simple steps using a detective analogy:

The Detective's Four-Step Investigation

Step 1: The "What If?" Game (Counterfactuals)
Imagine you show the robot a picture of a person with glasses and it says, "This person is Old."
The detective (GIFT) asks the robot: "What if I took the glasses away? Would you still say they are Old?"
The robot tries to imagine the picture without glasses. If the robot suddenly changes its mind and says, "Actually, they look Young now," the detective knows: "Aha! The glasses are a clue!"

  • The Magic: GIFT does this thousands of times, changing tiny things in thousands of pictures (adding a red ball, removing a car, changing a color) to see what makes the robot change its mind. These are called counterfactuals.

Step 2: Translating the Clues (Vision-to-Text)
The robot's changes are just pixels. It's hard for humans to look at a slightly different picture and say exactly what changed.
So, GIFT brings in a translator (a Vision-Language Model). This translator looks at the "Before" and "After" pictures and writes a sentence.

  • Instead of: "Pixel values in the top-left quadrant shifted by 15%."
  • The Translator says: "The tiny red metal ball behind the brown block was removed."
    Now, the clues are in words we can understand.

Step 3: Finding the Pattern (The "Aha!" Moment)
The detective now has a huge pile of notes: "Removed glasses -> became Young," "Added wrinkles -> became Old," "Removed red ball -> changed class."
Individually, these notes are messy. But GIFT uses a super-smart brain (a Large Language Model, like the one behind this chat) to read all the notes and find the pattern.

  • It realizes: "Wait, every time the robot changes its mind, it's because of glasses or wrinkles."
  • It drafts a global rule: "This robot thinks people are Old if they have glasses or forehead wrinkles."

Step 4: The Lie Detector Test (Verification)
Here is the most important part. Sometimes, the super-smart brain might guess wrong or make things up. GIFT doesn't just trust the guess; it tests it.
The detective goes back to the robot and says: "I think your rule is 'Glasses = Old'. Let's test it."
GIFT uses an image editor to take a picture of a young person, add glasses, and show it to the robot.

  • If the robot says "Old": The rule is TRUE. The glasses really caused the change.
  • If the robot says "Young": The rule is FALSE. The robot was just confused, and the glasses didn't matter.
    This step ensures the explanation is Faithful—it's not just a guess; it's a proven fact.

Why is this a big deal?

Most other methods are like asking a human to guess the recipe by looking at the soup. They might say, "It probably has salt," but they could be wrong.

  • Old methods often give vague heatmaps (like a blurry red spot on a picture) that don't tell you what the object is.
  • GIFT gives you a clear sentence: "The robot is biased because it thinks cars on the left side of the road mean 'Don't Turn Right'."

Real-World Examples from the Paper

  1. The "Red Metal Ball" Test: In a toy world with blocks, GIFT figured out that a robot was trained to look for "Red Metal Objects." It didn't just say "Red"; it figured out it needed to be metal too.
  2. The "Old Face" Test: On a dataset of human faces, GIFT found that the robot was looking for wrinkles and glasses. But it also found a weird bias: the robot thought a "detailed background" made a person look old! (Maybe the training photos of old people were taken in busy rooms).
  3. The "Self-Driving Car" Test: This is the most critical one. The researchers tested a robot trained to decide if a car can turn right. They found the robot had a hidden bias: if there were cars in the left lane, the robot said "No Turn," even if the road was clear. GIFT found this hidden rule and proved it with the Lie Detector test. Without GIFT, this dangerous bias might have gone unnoticed.

The Bottom Line

GIFT is a framework that turns a black-box robot into a transparent partner.
It doesn't just guess why the robot made a decision; it plays "What If," translates the results into English, finds the big patterns, and then proves those patterns are true by editing the images and testing the robot again.

It's like having a detective who doesn't just tell you who the culprit is, but shows you the fingerprint, the motive, and the alibi, all written in plain English. This makes AI safer and more trustworthy for things like self-driving cars and medical diagnosis.

Get papers like this in your inbox

Personalized daily or weekly digests matching your interests. Gists or technical summaries, in your language.

Try Digest →