Assessing Model-Agnostic XAI Methods against EU AI Act Explainability Requirements

This paper addresses the gap between existing model-agnostic Explainable AI methods and the EU AI Act's regulatory requirements by proposing a qualitative-to-quantitative scoring framework that helps practitioners assess compliance and identify areas needing further research.

Original authors: Francesco Sovrano, Giulia Vilone, Michael Lognoul

Published 2026-04-14
📖 5 min read🧠 Deep dive

This is an AI-generated explanation of the paper below. It is not written or endorsed by the authors. For technical accuracy, refer to the original paper. Read full disclaimer

Imagine you have a magic black box that makes important decisions about your life—like whether you get a loan, a job, or medical treatment. You know the box gives an answer, but you have no idea how it decided that. It's like a chef who hands you a delicious meal but refuses to tell you the recipe or even what ingredients are in it.

In the past, if you asked, "Why did you do this?", the chef might say, "Because the magic box said so." But the EU AI Act (a new set of strict rules for Europe) is saying: "No more magic tricks. You must explain your recipe."

This paper is like a guidebook for chefs and inspectors trying to figure out which tools can help them explain the black box's secrets without breaking the law.

Here is the breakdown of what the authors did, using simple analogies:

1. The Problem: The "Translation Gap"

There is a huge misunderstanding between lawyers and computer scientists.

  • Lawyers want explanations that are fair, clear, and help people understand why a decision was made so they can fight it if it's wrong.
  • Computer Scientists have built many tools (called XAI methods) to explain AI, but they often speak a different language. They focus on math and code, not on human understanding or legal rights.

The authors realized there was a "gap" where companies didn't know which tool to use to stay legal. They asked: "Which of these computer tools actually satisfies the law's requirements?"

2. The Solution: A "Compliance Scorecard"

To fix this, the authors built a scoring system. Think of it like a restaurant health inspection, but instead of checking for dirty floors, they check if the AI explanation tool is "legal."

They looked at three main qualities (properties) that any good explanation needs:

  • Faithfulness (The "Truth" Test): Does the explanation actually match what the AI is thinking? Or is it just making up a story? (e.g., If the AI rejected your loan because of your age, the explanation shouldn't say it was because of your hair color).
  • Robustness (The "Stability" Test): If you change the input slightly (like changing your age by one day), does the explanation stay the same? If the explanation flips wildly with tiny changes, it's not reliable.
  • Complexity (The "Simplicity" Test): Is the explanation too complicated for a human to understand? A 100-page math proof isn't a good explanation for a regular person.

3. The Process: How They Scored the Tools

The authors took a list of popular AI explanation tools (like SHAP, LIME, Decision Trees, and Counterfactuals) and gave them grades from 1 to 5 on those three qualities.

  • SHAP was like the honest accountant: Very accurate and truthful, but sometimes a bit slow and complex.
  • Decision Trees were like the clear flowchart: Very easy to read, but if you change one line of data, the whole chart might collapse (unstable).
  • LIME was like a quick sketch: Good for a quick guess, but if you ask it twice, you might get two different sketches (unstable).

4. The Result: Matching Tools to Rules

The EU AI Act has different rules for different situations. The authors created a map to show which tool fits which rule:

  • Scenario A: "Why was my specific loan rejected?" (Art. 86)
    • The Law wants: A clear, specific reason for you.
    • The Best Tool: SHAP or RuleSHAP. They are the most honest about exactly which factors mattered for your specific case.
  • Scenario B: "How does this system work in general?" (Art. 11 & 13)
    • The Law wants: A big-picture view for the company's internal files.
    • The Best Tool: Decision Trees or RuleFit. They provide a simple, readable list of rules that humans can easily digest.

5. The Big Takeaway

The paper concludes that there is no single "perfect" tool. It's like asking, "What is the best vehicle?"

  • If you need to race, you need a Formula 1 car (SHAP: accurate but complex).
  • If you need to drive a family to the grocery store, you need a minivan (Decision Trees: simple and clear).

The authors' framework helps companies pick the right "vehicle" for the specific legal "road" they are driving on.

Why This Matters

Before this paper, companies were guessing which tools to use to comply with the new laws. This paper gives them a checklist and a score.

  • If a company uses a tool that scores low on "Truth" (Faithfulness), they might get fined for lying about how their AI works.
  • If they use a tool that scores low on "Simplicity" (Complexity), they might get fined for confusing the customer.

In short: This paper bridges the gap between the legal world (which wants fairness and clarity) and the tech world (which builds complex math). It tells engineers, "Here is the math tool that will keep you out of trouble with the law."

Drowning in papers in your field?

Get daily digests of the most novel papers matching your research keywords — with technical summaries, in your language.

Try Digest →