Neural delay differential equations: learning non-Markovian closures for partially known dynamical systems

This paper introduces a constant-lag Neural Delay Differential Equations (NDDEs) framework, inspired by the Mori-Zwanzig formalism, to effectively learn non-Markovian dynamics from partially observed data by identifying memory effects through time delays, demonstrating superior performance over existing methods like LSTMs and ANODEs across synthetic, chaotic, and experimental datasets.

Thibault Monsel, Onofrio Semeraro, Lionel Mathelin, Guillaume Charpiat

Published Tue, 10 Ma
📖 5 min read🧠 Deep dive

Here is an explanation of the paper "Neural Delay Differential Equations: learning non-Markovian closures for partially known dynamical systems," translated into simple language with creative analogies.

The Big Problem: The "Blindfolded" System

Imagine you are trying to predict the weather, but you only have one tiny thermometer in your backyard. You don't know the wind speed, the humidity, the pressure, or what's happening in the clouds miles away. In the world of science, this is called partial observability.

Most modern AI models (like Neural ODEs) assume they can see the entire system at once. They think, "If I know the exact state of everything right now, I can predict the future." But in the real world, we rarely have that luxury. We only have a few sensors.

Furthermore, many systems aren't just about "what is happening now." They are about memory. A system might react to what happened 5 minutes ago, or 2 hours ago. If you ignore that history, your prediction will fail.

The Solution: The "Time-Traveling" AI

The authors propose a new type of AI called Neural Delay Differential Equations (NDDEs).

Think of a standard AI model as a driver who only looks through the windshield. They see the road right now and steer accordingly.

  • The Problem: If the car is on a winding road with a blind curve, looking only at the road right now isn't enough. You need to know where the road was 10 seconds ago to understand the curve you are currently entering.

NDDEs are like a driver who has a rear-view mirror that shows the road from the past. They don't just look at the current state; they explicitly look at the state of the system at specific times in the past (e.g., "What was the temperature 30 seconds ago?").

The Secret Sauce: Learning the "When"

Here is the clever part. In the past, scientists had to guess when to look back. They had to say, "Let's look at the data from exactly 5 seconds ago." If they guessed wrong, the model failed.

This paper introduces a method where the AI learns the time delays itself.

  • Analogy: Imagine you are trying to learn a song by listening to a recording, but you don't know the tempo. A standard model tries to guess the beat. This new model is like a musician who listens to the song and realizes, "Ah, the echo comes back exactly 0.4 seconds later." It learns the timing of the echo automatically.

The paper proves mathematically that if you have enough of these "echoes" (delays), you can perfectly reconstruct the hidden parts of the system, even if you can't see them directly. This is based on a famous math idea called Takens' Theorem, which basically says: "If you record a song at different times, you can reconstruct the whole melody."

The Physics Connection: The "Ghost" of the Past

The paper also leans on a physics concept called the Mori-Zwanzig formalism.

  • The Analogy: Imagine you are watching a billiard game, but you can only see the white ball. The other balls are hidden behind a curtain. When the white ball moves, it's because it was hit by a hidden ball.
  • The "hidden ball" is the unobserved variable.
  • The "hit" is the memory term.
  • The NDDE acts like a detective. It looks at the white ball's current path and its path from the past to deduce where the hidden balls must have been and how they are pushing the white ball around.

How They Tested It (The Lab Experiments)

The team tested this on three very different scenarios:

  1. Population Growth (The Rabbit Model): They modeled how a rabbit population grows. Rabbits don't reproduce instantly; there's a delay between mating and birth. The AI learned this delay perfectly, predicting the boom-and-bust cycles better than other models.
  2. Chemical Reactions (The Brusselator): This is a chemical system that oscillates (pulses) like a heartbeat. The AI had to predict the pulse using only partial data. The NDDE was the only model that stayed stable over a long time without going crazy.
  3. Fluid Dynamics (The Wind Tunnel): This was the "real world" test. They looked at air flowing over a cavity (like a hole in a car door). The air creates swirling vortices that bounce back and forth. This is a chaotic, noisy system.
    • The Result: The NDDE was the champion. It handled the noise better than the others. Why? Because by looking at the past, it could "average out" the random sensor noise and focus on the true physical rhythm of the wind.

Why This Matters

  1. It's Smarter: It doesn't just memorize data; it understands that the past influences the future.
  2. It's Efficient: Instead of needing a massive neural network with millions of parameters to "remember" everything (like a giant hard drive), it uses a few specific time delays (like a few key notes in a song) to capture the memory.
  3. It's Interpretable: Because the AI learns specific time delays (e.g., "The system reacts 2.5 seconds later"), scientists can actually look at the result and say, "Aha! The physics of this system has a 2.5-second lag." This gives us physical insight, not just a black-box prediction.

The Bottom Line

This paper gives us a new tool to predict complex systems when we don't have all the data. It teaches the AI to listen to the echoes of the past to understand the present, making it a powerful, efficient, and scientifically grounded way to model the messy, memory-filled real world.