An interpretable prototype parts-based neural network for medical tabular data

This paper proposes an inherently interpretable, prototype-based neural network for medical tabular data that learns human-readable, discretized feature subsets to provide transparent, case-based clinical predictions while maintaining competitive accuracy.

Jacek Karolczak, Jerzy Stefanowski

Published 2026-03-06
📖 5 min read🧠 Deep dive

Imagine you are a doctor trying to diagnose a patient. You look at their blood work, their age, and their symptoms. You don't just look at one number in isolation; you look for patterns. You might think, "Ah, this patient has high sugar and low energy, which reminds me of a specific type of diabetes I've seen before."

For a long time, computers were terrible at this kind of "pattern matching" in a way humans could understand. They were like black boxes: you put data in, and a result came out, but no one knew why the computer made that decision. If a computer told a doctor, "This patient has a 90% chance of heart failure," the doctor would ask, "Why? What specific numbers made you say that?" The computer couldn't answer clearly.

This paper introduces a new computer model called MEDIC (Model for Explainable Diagnosis using Interpretable Concepts). Think of MEDIC not as a black box, but as a digital apprentice doctor who learns by studying a library of past cases and explaining its reasoning using simple, human language.

Here is how it works, broken down into simple analogies:

1. The "Fuzzy" to "Hard" Translation (The Ruler Analogy)

Medical data often comes as exact numbers (e.g., "Glucose is 142.7 mg/dL"). But doctors don't think in decimals; they think in ranges (e.g., "Normal," "High," or "Dangerous").

  • The Problem: Computers hate turning exact numbers into categories because it's hard to learn from them mathematically.
  • MEDIC's Solution: Imagine a ruler that can stretch and shrink. During training, MEDIC uses a "fuzzy" ruler that gently nudges numbers into categories so it can learn. Once it has learned the rules, it snaps the ruler into a "hard" position. Now, it speaks in clear categories: "This patient's sugar is in the High range." This makes the output instantly readable for a human.

2. The "Spotlight" on Clues (The Detective Analogy)

When a detective solves a crime, they don't look at every single piece of evidence equally. They focus on the key clues: "The muddy shoe print and the missing watch."

  • The Problem: Medical records have hundreds of features (blood pressure, cholesterol, age, etc.). Most are noise.
  • MEDIC's Solution: MEDIC uses "patching masks" which act like a spotlight. It learns to ignore the irrelevant background noise and shine a light only on the specific combination of clues that matter. It might say, "I am ignoring the patient's height, but I am focusing heavily on their high cholesterol combined with low albumin."

3. The "Case File" Library (The Memory Analogy)

This is the most important part. Instead of calculating a complex formula, MEDIC learns by building a library of Prototypes.

  • Think of it like this: Imagine a veteran doctor who has seen thousands of patients. They don't have a formula; they have a mental library of "archetypal cases."
    • Case A: "The patient with the 'High Fever + Rash' pattern."
    • Case B: "The patient with the 'Low Blood Pressure + Fast Heartbeat' pattern."
  • How MEDIC works: When a new patient walks in, MEDIC doesn't guess. It looks at the patient's "spotlighted" clues and asks: "Which case in my library does this patient look most like?"
    • If the new patient looks 90% like "Case A," MEDIC predicts "Case A's outcome" and says, "I think you have Condition A because your symptoms match this specific pattern I've seen before."

4. Why This Matters (The Trust Analogy)

In the past, if an AI gave a diagnosis, doctors were skeptical because they couldn't see the logic. It was like a GPS telling you to turn left without showing you the map.

  • MEDIC changes the game: It shows you the map. It says, "I am predicting 'High Risk' because your Bilirubin is in the 'High' range, your Platelets are 'Low', and this specific combination matches Prototype #4 in our database, which was a patient who unfortunately passed away."

The Results

The researchers tested MEDIC on real medical data (liver disease, kidney disease, and diabetes).

  • Accuracy: It performed just as well as the most powerful, complex AI models currently used.
  • Transparency: Unlike those complex models, MEDIC could explain its decisions in plain English, using ranges and specific feature combinations that real doctors recognize.

In a Nutshell

MEDIC is a new kind of AI that stops trying to be a "magic oracle" and starts acting like a collaborative partner. It learns the rules of medicine by finding patterns in past cases, translates complex numbers into simple ranges, and explains its decisions by saying, "I made this choice because your case looks just like this other real case I know."

It bridges the gap between the super-power of computers and the common sense of doctors, making AI something doctors can actually trust and use to save lives.

Get papers like this in your inbox

Personalized daily or weekly digests matching your interests. Gists or technical summaries, in your language.

Try Digest →