On the definition and importance of interpretability in scientific machine learning

This paper argues that the scientific community often conflates mathematical sparsity with interpretability, proposing instead an operational definition for scientific machine learning that prioritizes mechanistic understanding over sparse expressions to better facilitate the discovery of fundamental physical principles.

Original authors: Conor Rowan, Alireza Doostan

Published 2026-04-23
📖 5 min read🧠 Deep dive

This is an AI-generated explanation of the paper below. It is not written or endorsed by the authors. For technical accuracy, refer to the original paper. Read full disclaimer

The Big Problem: The "Black Box" vs. The "Recipe Book"

Imagine you are a scientist trying to understand how the universe works. For centuries, scientists have been like master chefs writing down recipes. These recipes are short, simple lists of ingredients and steps (mathematical equations) that explain why a cake rises or why a bridge holds. If you read the recipe, you understand why the cake rises.

Now, imagine a new kind of chef: a Super-Computer AI. This AI has tasted millions of cakes and can predict exactly how a new cake will taste with 100% accuracy. But, the AI doesn't give you a recipe. Instead, it gives you a giant, tangled ball of yarn (a neural network) with millions of knots.

If you ask the AI, "Why did the cake rise?" it can't answer. It just says, "Because the yarn is tied this way."

This is the problem the paper addresses. In Scientific Machine Learning (SciML), we have these powerful "yarn ball" models that predict physical phenomena perfectly. But scientists are uneasy because they can't read the "recipe." They can't integrate these findings into their existing knowledge base. They want to know the mechanism, not just the prediction.

The Misunderstanding: "Simple" vs. "Understandable"

The paper argues that scientists in this field have been confused about what "interpretability" actually means.

The Current Mistake:
Many researchers think that if a model's output is mathematically simple (short and sparse), it is automatically understandable.

  • The Analogy: Imagine you find a note that says: x + y = z.
  • The Scientist's Thought: "Wow, that's short! It only has three letters. It must be easy to understand!"
  • The Reality: What do x, y, and z actually mean? Are they numbers? Are they emotions? Are they forces? Without knowing the context, a short equation is just as confusing as a long one.

The paper calls this the "Sparsity Trap." Just because an equation is short (sparse) doesn't mean it explains the physics behind the phenomenon.

The Real Definition: The "Story" Behind the Math

The authors propose a new, better definition of interpretability. They say:

A model is only "interpretable" if you can tell the story of how it connects to the fundamental laws of nature.

It's not about how short the sentence is; it's about whether the sentence makes sense in the context of the world.

The Analogy of the "Foreign Language":
Imagine you find a sentence written in a language you don't speak.

  • Scenario A (Sparse but Uninterpretable): The sentence is very short: "Gloop." It's short, but you have no idea what it means. It tells you nothing about the world.
  • Scenario B (Complex but Interpretable): The sentence is long and complicated: "The gravitational pull of the sun, combined with the inertia of the planet, creates an elliptical orbit." Even though it's long and complex, you understand the story. You know exactly what is happening.

The paper argues that Interpretability = Connection to Known Mechanisms.

  • If a model says, "This happens because of Advection (wind moving stuff) and Diffusion (spreading out)," that is interpretable because we know what those words mean.
  • If a model says, "This happens because of Term X," and we have no idea what "Term X" is, it is not interpretable, even if "Term X" is a very short mathematical term.

The Kepler Example: A Historical Lesson

The paper uses a famous historical example to prove their point: Johannes Kepler.

In the 1600s, Kepler discovered that planets move in ellipses. He wrote down simple, short mathematical laws to describe this.

  • Were they "Sparse"? Yes, very.
  • Were they "Interpretable" at the time? No.
    Kepler didn't know why the planets moved that way. He just knew that they did. It wasn't until 70 years later that Isaac Newton came along and said, "Ah! These laws are actually the result of Gravity and Inertia."

Only after Newton connected Kepler's short laws to the deeper "mechanism" of gravity did the laws become truly interpretable. Before that, they were just mysterious, short patterns.

The Solution: A New Way to Think

The authors suggest that in Scientific Machine Learning, we need to stop obsessing over making equations short (sparse) and start obsessing over making them connected to physical truth.

  1. Don't just look for short equations. A short equation with a mysterious term is useless for discovery.
  2. Look for the "Story." Can you explain the equation using known physics (like conservation of energy, force, or mass)?
  3. The Role of AI: The AI is great at finding patterns. But for it to be truly useful for science, we need to use those patterns to find the "missing link" to the fundamental laws of nature.

The Bottom Line

Think of Interpretability not as a "short summary," but as a bridge.

  • Old View: The bridge needs to be short and simple (Sparse).
  • New View: The bridge needs to connect two specific places: The Data we collected and the Fundamental Laws of Physics we already know.

If the AI gives us a short bridge that leads nowhere, it's not helpful. If it gives us a long, winding path that leads us to a deep understanding of how the universe works, that is true Interpretability.

In short: Scientists don't just want a model that predicts the future; they want a model that tells them why the future happens, using the language of physics, not the language of a black box.

Drowning in papers in your field?

Get daily digests of the most novel papers matching your research keywords — with technical summaries, in your language.

Try Digest →