Axiomatic On-Manifold Shapley via Optimal Generative Flows

This paper proposes a novel Axiomatic On-Manifold Shapley framework that utilizes optimal generative flows and Wasserstein-2 geodesics to eliminate off-manifold artifacts, ensuring geometric efficiency, reparameterization invariance, and superior semantic alignment in model attribution.

Cenwei Zhang, Lin Zhu, Manxi Lin, Lei You

Published 2026-03-06
📖 5 min read🧠 Deep dive

The Big Problem: "The Hallucinating Detective"

Imagine you have a super-smart AI detective that can tell if a photo is of a Cat or a Dog. You show it a picture of a fluffy Golden Retriever, and it says, "That's a dog!"

Now, you want to know why. You ask the detective, "Which part of the image made you decide it's a dog? Was it the ears? The nose? The fur?"

This is what Explainable AI (XAI) tries to do. It assigns a "score" to every pixel to show how much it contributed to the decision.

The Problem:
Most current methods try to figure this out by asking, "What if we removed this part?"

  • Old Method: They take the dog picture and replace the dog's nose with a black square (a "baseline").
  • The Flaw: A black square doesn't exist in the real world. It's an "off-manifold" artifact. The AI gets confused because it's never seen a dog with a black square nose. It might start hallucinating, saying, "Oh, that black square looks like a hole in space, so the ears must be the most important thing!"

The explanation becomes unstable and misleading because the AI is being tested on things that don't make sense in its "universe" of training data.


The Solution: "The Perfect Road Trip"

This paper proposes a new way to explain the AI's decision. Instead of jumping to a fake black square, they imagine a smooth road trip from a "blank slate" to the actual dog photo.

1. The Concept: The "Coalition Formation"

Think of the AI's decision as a team building a house.

  • Old Way (Shapley Values): You try to build the house by randomly adding bricks one by one. Sometimes you add a roof before a wall. It's chaotic, and the order matters too much.
  • New Way (This Paper): You build the house in a perfectly logical, smooth order. You start with a foundation (a blank image) and slowly, smoothly, morph it into the final house (the dog photo).

2. The Innovation: "Optimal Generative Flows"

The authors realized that the "road" we take from the blank slate to the dog photo matters.

  • The Bad Road: If you take a winding, bumpy, zig-zag path through "imaginary land" (where pixels are random noise), the AI gets dizzy and gives a bad explanation.
  • The Good Road (The Paper's Idea): They use math (called Optimal Transport) to find the straightest, smoothest, most energy-efficient road that stays strictly on the "Highway of Real Data."

The Analogy:
Imagine you are a bird flying from a nest (the blank image) to a tree (the dog photo).

  • Old Methods: The bird flies in a chaotic spiral, sometimes going through a wall or a cloud that doesn't exist.
  • This Paper: The bird finds the perfect glide path. It stays in the sky (the "manifold" of real data) the whole time. It never touches the ground or flies through a wall. It takes the path of least resistance.

3. Why "Least Resistance" Matters

The authors proved a cool mathematical theorem: If you take the path that uses the least amount of "energy" (kinetic energy) to get from the blank image to the real image, the explanation you get is the only correct one.

It's like saying: "If you want to know how much effort it took to walk from your house to the store, you shouldn't walk in circles or jump over fences. You should walk the most direct, natural path. Only then is the measurement fair."

What Did They Actually Do?

  1. They built a "Time Machine" for images: They trained a model that knows exactly how to morph a random noise image into a real dog image without ever creating a "fake" or "impossible" intermediate image.
  2. They measured the journey: As the image morphs from noise to dog, they tracked how the AI's confidence changed at every tiny step.
  3. They summed it up: By adding up all those tiny changes along this perfect path, they got a score for every pixel.

The Results: "Crystal Clear Vision"

When they tested this on images (like birds, cars, and faces):

  • Old Methods (like Integrated Gradients): Produced "ghosting" effects. The explanation looked like a blurry mess with noise everywhere, because the AI was confused by the weird paths it was forced to take.
  • This New Method: Produced sharp, clean maps. If the AI decided it was a dog, the explanation highlighted the ears, nose, and tail perfectly. It didn't highlight random noise.

The Takeaway

This paper is like upgrading from a crumpled paper map to a GPS with real-time traffic.

  • Old XAI: "Here is a guess of what the AI saw, but it might be wrong because we asked it about things that don't exist."
  • New XAI: "We asked the AI to explain itself while walking the most natural, logical path through reality. The result is a trustworthy, stable, and mathematically proven explanation."

In short: To understand a black box, don't poke it with a stick (random baselines). Gently guide it along the smoothest, most natural path to the answer, and it will tell you the truth.