Data-Driven Global Sensitivity Analysis for Engineering Design Based on Individual Conditional Expectations

This paper proposes a novel global sensitivity analysis method based on Individual Conditional Expectation (ICE) curves that overcomes the limitations of traditional Partial Dependence Plots (PDPs) in capturing input interactions, offering a mathematically proven, more informative metric for explainable machine learning in engineering design.

Pramudita Satria Palar, Paul Saves, Rommel G. Regis, Koji Shimoyama, Shigeru Obayashi, Nicolas Verstaevel, Joseph Morlier

Published Mon, 09 Ma
📖 4 min read☕ Coffee break read

Imagine you are a chef trying to perfect a new, complex recipe. You have a "black box" machine (a machine learning model) that tells you exactly how delicious the dish will be based on your ingredients (input variables like salt, heat, cooking time).

Your goal is to understand which ingredients matter most and how they work together.

The Old Way: The "Average" Taste Test (PDP)

Traditionally, engineers used a method called Partial Dependence Plots (PDP). Think of this as asking the machine: "If I change the amount of salt, what happens to the taste, assuming everything else is just an average mix?"

The machine gives you a single line on a graph representing the average effect of salt.

  • The Problem: This is like averaging the taste of a dish where salt makes it salty in one region but bitter in another. The average might look like "no change at all" (a flat line). You might conclude, "Salt doesn't matter!" when in reality, salt is the most critical ingredient, but its effect depends entirely on how much pepper you added. The averaging process hides the drama.

The New Way: The "Individual" Taste Test (ICE)

The authors of this paper propose a better way using Individual Conditional Expectation (ICE). Instead of averaging everything immediately, they look at every single scenario individually.

Imagine you run the taste test 1,000 times, but each time you keep the other ingredients (pepper, heat, time) fixed at a specific setting.

  • Scenario A: Low pepper + High heat. (Salt makes it amazing).
  • Scenario B: High pepper + Low heat. (Salt makes it terrible).

When you plot all 1,000 lines, you see a messy, tangled web of curves. This is the ICE. It reveals that salt does matter, but its effect changes wildly depending on the other ingredients.

The Innovation: Measuring the "Messiness"

The paper's main contribution is turning this messy web of lines into a simple, useful score. They propose two new metrics:

  1. The "Average Impact" Score (μIice\mu_{Iice}):
    Instead of averaging the results (which hides the truth), they average the magnitude of the change.

    • Analogy: Imagine a rollercoaster. The old method (PDP) might say, "The average height of the ride is 50 feet," which sounds boring. The new method looks at the ups and downs of every single car. It says, "Even if the average height is 50, the thrill (the change) is huge!" This tells you the ingredient is important, even if the average effect cancels out.
  2. The "Chaos" Score (σIice\sigma_{Iice}):
    This measures how much the lines in your ICE web wiggle and disagree with each other.

    • Analogy: If all 1,000 lines look almost identical, the ingredient is predictable (low chaos). If the lines are all over the place—some going up, some down, some flat—that's high chaos. This "Chaos Score" tells you: "Hey, this ingredient is a troublemaker! Its effect depends entirely on what other ingredients are in the pot." This is a direct measure of interaction.

The "Correlation" Check

They also added a "Correlation" check. This asks: "Does the messy web of lines generally follow the same direction as the boring average line?"

  • If the answer is Yes, the average line was a good summary.
  • If the answer is No (the lines are zigzagging while the average is flat), it means the relationship is complex and the old method was lying to you.

Why This Matters for Engineers

The authors tested this on three real-world problems:

  1. A Math Puzzle: Proving the method works on paper.
  2. Wind Turbines: Figuring out why a turbine blade might break. They found that wind speed and wave height interact in complex ways that the old "average" method missed.
  3. Airplane Wings: Designing the shape of a wing to reduce drag. They discovered that the top and bottom curves of the wing interact in surprising ways.

The Takeaway

In the world of engineering design, we often rely on "black box" computers to make decisions.

  • Old Method: "On average, this variable doesn't do much." (Risk: Missing critical interactions).
  • New Method: "On average, it might look small, but look at the variability! It's actually a powerhouse that changes everything depending on the context."

This paper gives engineers a new set of glasses. Instead of seeing a blurry, averaged-out picture, they can now see the individual stories of how variables interact, ensuring that safety-critical designs (like planes and turbines) aren't built on misleading averages.