Weakly Supervised Concept Learning with Class-Level Priors for Interpretable Medical Diagnosis

This paper introduces Prior-guided Concept Predictor (PCP), a weakly supervised framework that leverages class-level concept priors and regularization to enable reliable, interpretable medical diagnosis without costly concept annotations, significantly outperforming zero-shot baselines while matching fully supervised models.

Md Nahiduzzaman, Steven Korevaar, Alireza Bab-Hadiashar, Ruwan Tennakoon

Published 2026-03-05
📖 5 min read🧠 Deep dive

Imagine you are a doctor trying to diagnose a patient. You look at an X-ray or a skin mole and see a bunch of tiny details: "irregular edges," "dark spots," "white patches." In the medical world, these details are called concepts.

For years, AI has been great at looking at these images and saying, "That's cancer!" or "That's healthy!" But it's a black box. It gives you the answer, but it can't tell you why. It's like a student who gets the right answer on a math test but can't show their work. Doctors don't trust students who can't show their work, so they don't trust the AI either.

To fix this, researchers built "interpretable" AI. These are like students who must list the steps they took before giving the answer. But there's a huge problem: to teach the AI these steps, you need a human expert to label every single image with every single concept.

  • "This mole has an irregular border."
  • "This blood cell has a weird shape."
  • "This X-ray shows fluid."

Doing this for thousands of images is like asking a librarian to read every book in the library and write a summary for every single page. It takes too long, costs too much money, and experts get tired.

The New Idea: The "Gut Feeling" AI (PCP)

This paper introduces a new method called PCP (Prior-guided Concept Predictor). Think of it as teaching the AI using general rules instead of specific homework.

Here is the analogy:

1. The Old Way (Full Supervision)

Imagine you are training a new chef. You give them a photo of a specific pizza and say: "This pizza has pepperoni, extra cheese, and burnt crust." You do this for 1,000 different pizzas. The chef learns perfectly but takes forever to train.

2. The "Zero-Shot" Way (The Current Trend)

Some researchers tried to skip the training entirely. They gave the AI a giant encyclopedia (a huge language model) and said, "Just guess what's on the pizza based on your general knowledge."

  • The Problem: The AI knows what a "pepperoni pizza" looks like in general, but it doesn't know the specific medical details. It might confuse a "burnt crust" with a "scab" on a skin image. It's too vague and makes mistakes.

3. The PCP Way (Weak Supervision with Priors)

This is the paper's breakthrough. Instead of labeling every single image, you give the AI a rulebook (the "Priors").

  • The Rulebook: "If a patient has Melanoma (skin cancer), there is a 90% chance they have an 'irregular border' and a 70% chance they have 'blue-white patches'."
  • The Training: You don't tell the AI exactly what is in this specific photo. You just say, "Based on the fact that this photo looks like a Melanoma case, check if it has those features."

The AI looks at the image, guesses the features, and then checks its own guess against the rulebook.

  • The "Refinement" (The Magic Sauce): The AI has a "self-correcting" mechanism.
    • KL Divergence: This is like a "Reality Check." If the AI thinks a cancer case never has a specific feature, but the rulebook says it usually does, the AI gets a gentle nudge to adjust its thinking.
    • Entropy Regularization: This is like a "Focus Filter." It stops the AI from saying, "Maybe it's a little bit of everything." It forces the AI to be decisive: "It's definitely this feature, and definitely not that one."

Why This is a Big Deal

  1. It's Cheaper and Faster: You don't need to hire experts to label thousands of images. You just need them to write down the general rules (the "Priors") once. It's like writing a recipe once instead of cooking the dish 1,000 times to prove it works.
  2. It's Trustworthy: The AI still explains its reasoning ("I think this is cancer because I see an irregular border"), but it learned to see those borders without being explicitly told to look for them in every single photo.
  3. It Works Better Than Guessing: The researchers tested this on skin images (dermoscopy) and blood cell images. The AI got the "concept" predictions right more than 33% better than the AI that just guessed based on general knowledge.

The Bottom Line

The authors built a system that teaches AI to be a medical detective without needing a human to point out every clue in every single case. Instead, they gave the AI a map of the territory (the class-level priors) and let it learn to find the clues on its own, while gently correcting itself to stay on track.

This means we can get AI that doctors can trust, explain, and use in real hospitals, without waiting years to collect millions of expensive, hand-labeled images. It's a smarter, faster way to teach machines how to "think" like a doctor.