Knowledge-Guided Machine Learning: Illustrating the use of Explainable Boosting Machines to Identify Overshooting Tops in Satellite Imagery

This paper demonstrates how to develop a fully interpretable machine learning model for detecting overshooting tops in satellite imagery by using Explainable Boosting Machines combined with knowledge-guided feature extraction and human-in-the-loop refinement to prioritize transparency and domain expertise over peak accuracy.

Nathan Mitchell, Lander Ver Hoef, Imme Ebert-Uphoff, Kristina Moen, Kyle Hilburn, Yoonjin Lee, Emily J. King

Published 2026-03-02
📖 5 min read🧠 Deep dive

Imagine you are trying to teach a computer to spot a very specific type of storm cloud, called an Overshooting Top (OT), in satellite photos. These are like giant, bubbling "heads" of clouds that punch through the sky, often signaling dangerous weather like tornadoes or hail.

For a long time, scientists have used "Black Box" AI (like deep neural networks) to do this. Think of a Black Box AI as a genius chef who can cook a perfect meal but refuses to tell you the recipe. You know the food tastes good, but if the chef suddenly decides to put salt in the dessert because they saw a picture of salt once, you have no idea why. In weather forecasting, this is dangerous. If the AI makes a mistake, we don't know why it happened, and we can't easily fix it.

This paper introduces a different kind of AI called an Explainable Boosting Machine (EBM). Instead of a mysterious genius chef, think of an EBM as a transparent, step-by-step recipe book.

Here is the story of how they used this new tool, explained simply:

1. The Problem: The "Clever Hans" Mistake

In the past, AI models sometimes learned "tricks" instead of real science. This is called the "Clever Hans" effect (named after a horse that seemed to do math but was actually just reading the trainer's body language).

  • The Trick: An AI might learn to spot a horse in a photo not by looking at the horse, but by looking for a "source tag" in the corner of the image that says "Horse."
  • The Risk: If you show the AI a horse without the tag, it fails. If you show it a picture of a pole with a tag, it thinks it's a horse.
  • The Goal: The authors wanted to build an AI that couldn't pull these tricks. They wanted a model that humans could look at and say, "Yes, that makes sense," or "No, that's wrong, let's fix it."

2. The Solution: The "Recipe Book" (EBM)

The authors used an Explainable Boosting Machine (EBM).

  • How it works: Instead of one giant, confusing brain, the EBM is like a team of small, simple experts.
    • Expert 1 looks at how bright the cloud is.
    • Expert 2 looks at how bumpy (textured) the cloud is.
    • Expert 3 looks at how cold the cloud is.
  • The Magic: Each expert writes down a simple rule on a piece of paper (a graph). You can literally see the graph: "If the cloud is this bright, add 5 points to the storm score. If it's that bright, subtract 2 points."
  • The Result: You can see exactly how the AI is thinking. If the AI thinks a dark shadow is a storm, you can look at the "Brightness Expert's" graph, see the mistake, and simply erase that part of the graph and draw a new line. You don't have to retrain the whole AI; you just edit the recipe.

3. The Process: From Raw Data to "Scalar" Clues

Satellite images are huge grids of pixels. EBMs can't read pictures directly; they need simple numbers (scalars). The authors had to translate the picture into three simple clues for their "experts":

  1. Brightness: How shiny is the cloud? (OTs are usually very bright in the sun).
  2. Texture: Is the cloud smooth like a pancake or bumpy like a popcorn kernel? (OTs are very bumpy). They used a math trick called a "GLCM" to measure this bumpiness.
  3. Temperature: How cold is the top of the cloud? (OTs are very high and very cold).

4. The Human Touch: "Editing the Recipe"

This is the most important part of the paper.

  • Step 1: They trained the EBM using data from radar (which tells them where rain is happening).
  • Step 2: They looked at the "Recipe Book" (the graphs). They noticed something weird: The AI was getting confused by shadows. It thought dark shadows were storms.
  • Step 3: The Edit. A human scientist looked at the graph, said, "Wait, shadows aren't storms," and manually flattened that part of the graph.
  • Step 4: They did this without retraining the whole model. They just changed the rules.

5. The Result: Good Enough, But Safe

The final model wasn't the most accurate AI in the world (it wasn't as "smart" as the Black Box models). However, it was honest.

  • It didn't make up fake reasons for its decisions.
  • It allowed humans to step in and fix its logic.
  • It successfully found the storm clouds, even if it sometimes got confused by other cold, bumpy clouds (like the "failure mode" they discussed).

The Big Takeaway

The authors are saying: "Don't just build a black box that works; build a glass box that we understand."

In high-stakes fields like weather forecasting, knowing why a computer made a prediction is just as important as the prediction itself. If a model says a tornado is coming, we need to know if it's looking at the right things (bumpy, cold clouds) or if it's just looking at a shadow. With EBMs, we can look under the hood, see the engine, and tune it ourselves.

In short: They built a weather AI that doesn't just guess; it explains its reasoning, lets humans correct its mistakes, and follows a recipe we can all read.

Get papers like this in your inbox

Personalized daily or weekly digests matching your interests. Gists or technical summaries, in your language.

Try Digest →