Hierarchical Inference and Closure Learning via Adaptive Surrogates for ODEs and PDEs

This paper proposes a hierarchical Bayesian framework that integrates adaptive surrogate models, such as Fourier Neural Operators and parametric PINNs, with ensemble MALA sampling to simultaneously infer individual system parameters and learn shared unknown dynamics via ML-based closure models for ODEs and PDEs.

Pengyu Zhang, Arnaud Vadeboncoeur, Alex Glyn-Davies, Mark Girolami

Published 2026-03-05
📖 5 min read🧠 Deep dive

Imagine you are a detective trying to solve a mystery, but you don't have the full picture. You have a bunch of similar crime scenes (physical systems), and you know the general rules of how the world works (the laws of physics), but you are missing two crucial pieces of the puzzle:

  1. The Specifics: Every crime scene has unique details (like the weight of a specific object or the initial speed of a car).
  2. The Missing Rule: There is a hidden, complicated law of nature that you don't know yet (like exactly how friction works in a specific type of engine).

This paper presents a clever new way for scientists and engineers to solve these mysteries simultaneously. They call it Hierarchical Inference and Closure Learning.

Here is a simple breakdown of how it works, using some creative analogies:

1. The "Classroom" Analogy (Hierarchical Inference)

Imagine a classroom of 20 students (the 20 different physical systems).

  • The Old Way: If you wanted to know how smart each student is, you would test them one by one, completely ignoring the others. If a student has bad test data (noise), you might get a wrong answer.
  • The New Way (Hierarchical): You realize these students are all in the same class. They share a common "population" intelligence. If Student A has messy data, you can look at Student B and Student C to help guess what Student A's true intelligence is.
  • The Result: By looking at the whole group together, you get a much more accurate estimate of each individual's specific traits, even if their personal data is a bit fuzzy.

2. The "Ghost in the Machine" (Closure Learning)

Now, imagine you know the equation for how a car moves, but you don't know exactly how the engine vibrates when it gets hot. That missing vibration rule is the "Closure."

  • Instead of trying to write a complex math formula for this vibration (which is hard), the researchers use a Neural Network (a type of AI).
  • Think of the Neural Network as a shape-shifting clay model. It starts as a lump of clay. As the researchers feed it data from the 20 students, the clay slowly molds itself into the exact shape of the missing vibration rule.
  • The AI learns this rule by trial and error, trying to make the math match the real-world observations.

3. The "Speeding Ticket" Problem (The Computational Bottleneck)

Here is the big problem: To figure out the missing rule and the specific details, the computer has to run the physics simulation millions of times.

  • The Analogy: Imagine you are trying to find the perfect recipe for a cake. Every time you change an ingredient, you have to bake the cake, wait 45 minutes for it to cool, taste it, and then write down the result. If you need to do this 10,000 times, you will never finish.
  • The Solution (Surrogates): The researchers build a Fast-Forward Simulator (a Surrogate Model).
    • Instead of baking the real cake (running the slow, expensive physics simulation), they use a Magic Crystal Ball (the Surrogate).
    • The Crystal Ball predicts what the cake will taste like in 0.01 seconds.
    • Crucially, they train this Crystal Ball while they are solving the mystery. It learns to mimic the real physics so well that it becomes a perfect stand-in, saving them years of computing time.

4. The "Two-Level Dance" (Bilevel Optimization)

The whole process is a synchronized dance between two goals:

  1. The Detective's Goal: "I need to find the specific details of these 20 students." (This uses a method called MALA, which is like a smart random walk that explores all possibilities).
  2. The Teacher's Goal: "I need to teach the Crystal Ball to be a better predictor."

They do this in a loop:

  • The Detective takes a step to guess the details.
  • The Teacher uses that guess to train the Crystal Ball to be more accurate.
  • The Detective uses the new Crystal Ball to take a better step.
  • They keep dancing back and forth until they have solved the mystery and built a perfect Crystal Ball.

Why is this a big deal?

  • It handles the unknown: It doesn't just guess numbers; it learns entire missing laws of physics.
  • It handles the messy: It works even when the data is noisy or incomplete, by using the "classroom" effect to help each other out.
  • It's fast: By using the "Magic Crystal Ball" (Surrogate), it solves problems that would usually take weeks in just hours.

In summary: This paper gives scientists a super-tool to figure out the hidden rules of nature and the specific details of complex systems, even when they don't have perfect data, by letting a group of systems teach each other and using a fast AI assistant to speed up the math.