Digging Deeper: Learning Multi-Level Concept Hierarchies

This paper introduces Multi-Level Concept Splitting (MLCS) and Deep-HiCEMs to overcome the limitations of shallow hierarchies in concept-based models by automatically discovering multi-level concept structures from coarse annotations and enabling effective interventions at various levels of abstraction.

Oscar Hill, Mateo Espinosa Zarlenga, Mateja Jamnik

Published 2026-03-12
📖 5 min read🧠 Deep dive

The Big Picture: Teaching AI to Think in Layers

Imagine you are trying to teach a robot to recognize a Red Apple.

  • The Old Way (Flat Thinking): You tell the robot, "Look for 'Red' and look for 'Apple'." The robot learns these two things are separate. It doesn't understand that "Red" is a type of color, or that "Apple" is a type of fruit. It's like giving the robot a list of 1,000 unrelated words and hoping it figures out the connections on its own.
  • The Previous Upgrade (Shallow Thinking): Researchers realized this was too simple. They taught the robot, "Apple" is a big category, and inside that, there are sub-categories like "Red Apple" and "Green Apple." This is better, but it stops there. It's a two-story building: the ground floor (Apple) and the first floor (Red Apple).
  • The New Breakthrough (Deep Thinking): This paper introduces a way to teach the robot to build a skyscraper of understanding. It can go from "Fruit" \rightarrow "Apple" \rightarrow "Red Apple" \rightarrow "Crunchy Red Apple."

The authors, Oscar Hill, Mateo Espinosa Zarlenga, and Mateja Jamnik, have created two new tools to make this happen: MLCS (the discovery tool) and Deep-HiCEM (the thinking machine).


1. The Problem: AI is Too "Flat"

Most AI models today are like a flat map. They know that "Dog" and "Cat" are animals, but they don't naturally understand that a "Golden Retriever" is a specific kind of dog.

To fix this, previous researchers tried to build a hierarchy (a family tree) for AI concepts. But they could only build one level deep.

  • Level 1: The big idea (e.g., "Vehicle").
  • Level 2: The sub-idea (e.g., "Car").
  • The Limit: They couldn't go deeper to "Sedan" or "Sports Car" without needing a human to manually label every single one. That takes forever and is expensive.

2. The Solution: MLCS (The "X-Ray" Machine)

The authors invented Multi-Level Concept Splitting (MLCS). Think of this as an X-ray machine for AI brains.

Usually, to teach an AI about "Red Apples," you need thousands of photos labeled "Red Apple." MLCS is magic because it doesn't need those labels.

  • How it works: You give the AI a picture of an apple and just say, "This is a fruit."
  • The Magic: MLCS looks inside the AI's "brain" (its internal math) and says, "Hey, I see a pattern here that looks like 'Red,' and inside that, I see a pattern that looks like 'Shiny'."
  • The Result: It automatically discovers the hidden layers of the hierarchy (Fruit \rightarrow Apple \rightarrow Red Apple) without anyone telling it what to look for. It's like finding a secret underground tunnel system in a building you thought was just a single floor.

3. The Solution: Deep-HiCEM (The "Skyscraper" Architect)

Once MLCS finds these hidden layers, you need a new type of building to house them. Enter Deep-HiCEM.

Think of a standard AI model as a bungalow (one floor). The old "Hierarchical" models were two-story houses. Deep-HiCEM is a skyscraper.

  • Arbitrary Depth: It can have as many floors as needed. It can handle "Animal" \rightarrow "Mammal" \rightarrow "Dog" \rightarrow "Poodle" \rightarrow "Fluffy Poodle."
  • Intervention (The "Human-in-the-Loop"): This is the coolest part. Because the AI understands the hierarchy, you can talk to it like a human.
    • Scenario: The AI thinks a picture is a "Dog," but you know it's actually a "Wolf."
    • Old AI: You have to retrain the whole thing or guess which specific feature is wrong.
    • Deep-HiCEM: You can just say, "No, that's not a dog, it's a wolf." Because the AI knows "Wolf" is a cousin of "Dog" but not a "Dog," it instantly updates its understanding of the whole picture. You can fix the AI's mistakes in real-time by correcting the concepts.

4. What Did They Prove? (The Results)

The team tested this on several datasets, including a made-up "PseudoKitchen" where they had ingredients (like "Apple") and sub-ingredients (like "Red Apple").

  • Did it find the hidden layers? Yes! The AI successfully discovered "sub-sub-concepts" (like specific colors or textures) that were never shown to it during training.
  • Did it get smarter? Yes. The AI was just as good at guessing the final answer (e.g., "Is this a fruit?") as the old models, but now it had a much richer understanding of why.
  • Can we fix it? Yes. When the researchers manually corrected the AI's concepts (e.g., "Actually, this is a red apple, not a green one"), the AI's final answer got better. It proved that the AI was listening to the hierarchy.

The Takeaway

This paper is about giving AI a deeper, more human-like way of thinking.

Instead of just memorizing a flat list of facts, the AI learns to organize knowledge into a family tree. It learns that a "Red Apple" isn't just a random word; it's a specific type of "Apple," which is a specific type of "Fruit."

Why does this matter?

  1. Less Work: We don't need humans to label every tiny detail. The AI can find the details itself.
  2. More Trust: If an AI makes a mistake, we can understand where in the hierarchy it went wrong and fix it easily.
  3. Better Explanations: Instead of saying "I think this is a cat because of pixels," the AI can say, "I think this is a cat because it has fur, whiskers, and pointy ears," and explain how those features fit together.

In short, they taught the AI to stop looking at the world in a flat line and start seeing the depth of reality.