Enabling stratified sampling in high dimensions via nonlinear dimensionality reduction

This paper proposes a method to enable effective stratified sampling in high-dimensional spaces by using neural active manifolds to identify a one-dimensional latent space that captures model variability, allowing for the creation of input partitions that align with model level sets to significantly reduce variance in uncertainty propagation.

Gianluca Geraci, Daniele E. Schiavazzi, Andrea Zanoni

Published 2026-03-06
📖 4 min read🧠 Deep dive

Imagine you are trying to guess the average temperature of a giant, complex ocean. You can't measure every drop of water, so you have to take samples.

The Problem: The "Curse of Dimensions"
If the ocean were just a small bathtub (low dimensions), you could easily divide it into a grid of squares and take one sample from each square. This is called Stratified Sampling. It's like cutting a cake into equal slices; you know you've covered the whole cake, and your guess will be very accurate.

But what if the ocean isn't just a 2D surface, but a 3D volume, or even a 100-dimensional "hyper-ocean" (which happens in complex computer models for weather, finance, or engineering)?
If you try to cut a 100-dimensional cake into a grid, you would need more slices than there are atoms in the universe just to get a few samples per slice. This is the Curse of Dimensionality. Traditional methods fail because you run out of time and money before you can take enough samples.

The Old Way vs. The New Way

  • Standard Monte Carlo: This is like throwing darts blindly at the ocean map. You might hit the same spot twice and miss the cold currents entirely. It's slow and inaccurate.
  • The Paper's Solution: Instead of throwing darts blindly or trying to cut a 100D cake into a grid, the authors propose a clever trick: Find the "Spine" of the problem.

The Core Idea: The "NeurAM" (Neural Active Manifold)
Imagine the complex ocean has a hidden, winding river running through it. The temperature changes mostly along this river, while the water in the vast, empty spaces around the river stays mostly the same.

The authors use a special AI tool called NeurAM (Neural Active Manifold) to find this hidden river.

  1. The Detective (AI): The AI looks at the complex model and learns that, despite having 100 inputs, the output (the temperature) actually only cares about one specific combination of those inputs. It's like realizing that the weather in a city depends mostly on the wind direction, even though there are 50 other sensors measuring humidity, pressure, etc.
  2. The Map: The AI compresses the entire 100-dimensional ocean down into a single, 1-dimensional line (the river).
  3. The Slice: Now, instead of trying to slice a 100D cake, you just slice this 1D line into equal pieces. This is easy!
  4. The Back-Projection: You take those slices on the line and "project" them back onto the original 100D ocean. Because the AI found the river, these slices aren't random squares; they are smart, curved shapes that wrap around the interesting parts of the model (the "level sets").

The Analogy: The Mountain Hike
Think of the computer model as a giant, foggy mountain range. You want to know the average height of the terrain.

  • Old Method: You try to place a grid of 100x100x100... squares over the whole mountain. You get lost in the fog and can't place enough squares.
  • New Method: You use a drone (the AI) to fly over the mountain and realize that the height changes mostly along a single, winding ridge line. The rest of the mountain is just flat or repetitive.
    • You draw a line along that ridge.
    • You cut that line into 10 equal segments.
    • You send your hikers to sample those 10 specific segments.
    • Because you sampled the "spine" of the mountain where the action is, your average height calculation is incredibly accurate, even though you only took a few samples.

Why This is a Big Deal

  1. It Scales: It works for problems with 10, 100, or even 1,000 variables. The "1D line" trick bypasses the complexity of high dimensions.
  2. It's Smart: The slices aren't rigid boxes; they bend and twist to follow the shape of the data.
  3. It Saves Money: In engineering, running a simulation can cost thousands of dollars. This method gets you the same accuracy with far fewer (and cheaper) simulations.
  4. It Works with "Cheap" Models: The paper also shows you can mix this with "low-fidelity" (cheaper, less accurate) models to get even better results, like using a sketch to guide your search for the real painting.

In a Nutshell
The authors found a way to stop fighting the complexity of high-dimensional data. Instead of trying to cover every inch of a massive, multi-dimensional space, they use AI to find the "skeleton" of the problem, slice that skeleton, and use those slices to guide their sampling. It turns an impossible task into a manageable one, saving time, money, and computational power.