Expressibility of neural quantum states: a Walsh-complexity perspective

This paper introduces Walsh complexity as a new measure of expressibility for neural quantum states, demonstrating that shallow additive networks struggle to represent certain short-range entangled states with maximal Walsh spectrum spread unless they reach a logarithmic depth or utilize activation saturation.

Taige Wang

Published 2026-04-07
📖 5 min read🧠 Deep dive

Imagine you are trying to teach a computer to mimic the behavior of a complex quantum system, like a swarm of interacting particles. In the world of physics, these systems are described by something called a "wavefunction." To do this, scientists use Neural Quantum States (NQS)—basically, artificial neural networks designed to guess what these particles are doing.

The big question this paper asks is: "How good are these neural networks at guessing?"

Specifically, the authors want to know: Can a simple, shallow neural network describe a complex quantum state, or do we need a very deep, complicated one?

Here is the breakdown of their discovery, using simple analogies.

1. The Problem: The "Hidden Complexity" Trap

Usually, physicists judge how hard a quantum state is to describe by looking at entanglement (how much the particles are "tangled" together).

  • The Old Belief: If particles are only tangled with their immediate neighbors (short-range entanglement), the state should be easy to describe.
  • The Surprise: The authors found a specific quantum state (a "dimerized state") that looks very simple. The particles are only tangled with their neighbors, and it can be described by a very short, simple formula.
  • The Twist: Despite looking simple, this state is a nightmare for certain types of neural networks (called "additive" networks). It's like a puzzle that looks like a picture of a cat, but to solve it, you have to rearrange every single piece in a way that defies logic.

2. The New Tool: "Walsh Complexity" (The Flavor Spectrum)

To measure this hidden difficulty, the authors invented a new ruler called Walsh Complexity.

The Analogy: The Cocktail Party
Imagine a huge party where everyone is shouting a different combination of words.

  • Simple State: Everyone is shouting the same phrase, or just a few distinct phrases. You can easily predict what the crowd sounds like.
  • Complex State (The Target): The crowd is shouting a chaotic mix of every possible combination of words, with equal volume. It's a "flat spectrum" of noise.

Walsh Complexity measures how "spread out" the noise is.

  • If the noise is concentrated in a few patterns, the complexity is low.
  • If the noise is spread evenly across all possible patterns (like our chaotic party), the complexity is maximal.

The authors found that their "simple" dimerized state has Maximal Walsh Complexity. It's a "flat spectrum" state.

3. The Limitation: The "Shallow Network" Bottleneck

The paper tests two types of neural networks:

  1. Multiplicative Networks: These build the answer by multiplying factors together (like stacking Lego bricks).
  2. Additive Networks: These build the answer by adding up signals (like mixing ingredients in a bowl). This is the standard way modern AI (like Transformers) works.

The Finding:

  • Multiplicative networks are great at handling this "flat spectrum" chaos. They can describe the simple dimer state easily.
  • Additive networks (the standard ones) struggle immensely.
    • If the network is shallow (few layers) and uses standard math functions (like polynomials), it physically cannot generate enough "Walsh Complexity" to match the target. It's like trying to paint a masterpiece using only a single drop of paint; you simply don't have enough "ink" to cover the canvas.
    • The network needs to get deeper (more layers) to build up enough complexity. The paper proves that for these networks to succeed, the depth must grow logarithmically with the size of the system (roughly, if you double the particles, you need a few more layers).

4. The "Saturation" Switch

The authors also looked at what happens when the neural network uses "saturated" activation functions (like the tanh function, which squashes numbers between -1 and 1).

  • The Analogy: Imagine a light switch. In the "tame" regime, the switch is dim and adjustable. In the "saturated" regime, the switch is either fully ON or fully OFF.
  • The Result: Once the network pushes its internal signals into this "ON/OFF" (threshold) mode, the rules change. The network suddenly becomes much more powerful, almost like a super-computer that can solve hard logic puzzles instantly.
  • The Catch: While this makes the network powerful, it also makes it impossible to prove mathematically that it can't solve a problem. It's like a magician who can pull a rabbit out of a hat; you know it's possible, but you can't easily explain how the trick works or prove a limit to what they can do.

Summary: What Does This Mean?

  1. Entanglement isn't everything: A quantum state can look simple (low entanglement) but be incredibly hard for standard AI to learn because of its "hidden spectrum" (Walsh complexity).
  2. Depth is a resource: For standard additive neural networks, depth (number of layers) is the key to unlocking complex states. You can't just throw more "width" (more neurons) at the problem; you need more layers to build up the necessary complexity.
  3. The "Tame" vs. "Wild" regimes:
    • In the "Tame" regime (standard math, no saturation), we can mathematically prove exactly what these networks cannot do.
    • In the "Wild" regime (saturated, threshold-like behavior), the networks become so expressive that it's very hard to predict their limits, which explains why modern AI seems to work so magically well on hard problems.

The Bottom Line:
This paper gives physicists a new "ruler" (Walsh Complexity) to measure how hard a quantum state is for AI to learn. It warns us that just because a state looks simple, it might be a "trick question" for standard neural networks, requiring deeper architectures to solve.

Get papers like this in your inbox

Personalized daily or weekly digests matching your interests. Gists or technical summaries, in your language.

Try Digest →