Multilevel Picard approximations and deep neural networks with ReLU, leaky ReLU, and softplus activation overcome the curse of dimensionality when approximating semilinear parabolic partial differential equations in LpL^p-sense

This paper proves that multilevel Picard approximations and deep neural networks with ReLU, leaky ReLU, and softplus activation functions can approximate solutions to semilinear parabolic PDEs in the LpL^p-sense without suffering from the curse of dimensionality, as their computational cost and parameter count grow at most polynomially with respect to the dimension and the inverse of the desired accuracy.

Ariel Neufeld, Tuan Anh Nguyen

Published 2026-03-24
📖 5 min read🧠 Deep dive

Imagine you are trying to predict the weather, but instead of just looking at temperature and wind in your city, you have to track the weather for every single atom in the universe simultaneously.

In the world of mathematics and physics, this is the nightmare known as the "Curse of Dimensionality."

When scientists try to solve complex equations (called Partial Differential Equations or PDEs) that describe things like stock market crashes, quantum particles, or fluid dynamics, the number of variables (dimensions) can explode. Traditional computers hit a wall: every time you add one more variable, the time and money needed to solve the equation don't just go up a little; they go up exponentially. It's like trying to find a specific grain of sand on a beach, but every time you add a new beach, the number of grains doubles, then quadruples, until the universe runs out of sand.

This paper, written by Ariel Neufeld and Tuan Anh Nguyen, presents a breakthrough: They have proven that two modern tools can break this curse.

Here is the simple breakdown of their discovery:

1. The Two Heroes: MLPs and Deep Neural Networks

The authors focus on two specific methods that have been getting a lot of hype in the tech world:

  • Multilevel Picard (MLP) Approximations: Think of this as a "smart guess-and-check" team. Instead of one person trying to solve the whole puzzle alone, they break the problem into layers. They make a rough guess, then a slightly better guess, then an even better one, using a massive team of random simulations (like rolling dice millions of times) to average out the errors.
  • Deep Neural Networks (DNNs): These are the "brains" behind modern AI (like the chatbots you use). They are mathematical structures inspired by the human brain, made of layers of nodes that learn patterns. The authors specifically tested three types of "activation functions" (the switches that turn neurons on or off): ReLU, Leaky ReLU, and Softplus.

2. The Big Discovery: "Polynomial" vs. "Exponential"

The paper proves a mathematical fact that changes everything:

  • The Old Way (Curse): If you want to be twice as accurate, the old methods might need 21002^{100} (a number with 30 zeros) more computing power. This is impossible.
  • The New Way (The Breakthrough): With MLPs and these specific Neural Networks, if you want to be twice as accurate, the computing power only needs to go up by a polynomial amount (like 232^3 or 242^4).

The Analogy:
Imagine you are trying to paint a mural.

  • The Curse: If the mural gets 10% bigger, you need 10,000% more paint and time. You run out of paint before you finish.
  • The Solution: With the new method, if the mural gets 10% bigger, you only need a little bit more paint. The size of the mural doesn't matter anymore; the method scales efficiently.

3. Why "Lp-Sense" Matters

The paper mentions solving these equations in "Lp-sense."

  • Simple Translation: Most previous math proofs only guaranteed the method worked for "average" errors (L2). The authors proved it works for all kinds of errors, including the worst-case scenarios (L-infinity).
  • The Metaphor: Imagine a bridge. Previous proofs said, "This bridge is safe for the average car." This paper says, "This bridge is safe for every car, even the heaviest truck, and we can prove it mathematically." This makes the solution much more robust and reliable for real-world engineering.

4. The "Activation" Secret

The authors specifically looked at ReLU, Leaky ReLU, and Softplus.

  • ReLU is like a light switch: it's either ON or OFF.
  • Leaky ReLU is a dimmer switch that never fully turns off (it lets a tiny bit of light through).
  • Softplus is a smooth curve that mimics a switch but without the sharp edge.

The paper proves that all three of these "switches" work perfectly for solving these high-dimensional problems. This is great news because engineers can choose the switch that fits their specific hardware best without worrying about the math breaking.

5. The Real-World Impact

Why should you care?

  • Finance: Banks can better price complex options (like betting on a basket of 100 different stocks) without needing a supercomputer the size of a city.
  • Physics: Scientists can simulate how particles interact in complex systems with thousands of variables, which was previously impossible.
  • Engineering: Designing safer structures or optimizing energy grids becomes computationally feasible.

The Bottom Line

This paper is the mathematical receipt that proves the "magic" of AI and advanced Monte Carlo simulations isn't just luck. It proves that for a huge class of difficult equations, Deep Learning and Multilevel Picard methods are the keys to unlocking high-dimensional problems.

They have shown that the "Curse of Dimensionality" is not a law of nature, but a limitation of old methods. With these new tools, we can finally solve problems that were previously considered impossible.

Drowning in papers in your field?

Get daily digests of the most novel papers matching your research keywords — with technical summaries, in your language.

Try Digest →