A Function-Centric Perspective on Flat and Sharp Minima

This paper challenges the conventional view that flat minima inherently ensure better generalization, arguing through extensive empirical studies that sharpness is a function-dependent property — sharper minima often correlate with improved performance, robustness, and calibration when models are properly regularized, though distinguishing task-driven sharpness from memorization-driven sharpness remains an open practical question.

Original authors: Israel Mason-Williams, Gabryel Mason-Williams, Helen Yannakoudakis

Published 2026-04-16✓ Author reviewed
📖 6 min read🧠 Deep dive

This is an AI-generated explanation of the paper below. It is not written by the authors. For technical accuracy, refer to the original paper. Read full disclaimer

The Big Idea: It's Not About the "Valley," It's About the "Map"

For years, scientists studying AI believed a simple rule: Flat is good, Sharp is bad.

Imagine training an AI is like a hiker trying to find the lowest point in a mountain range (the "loss landscape").

  • Flat Minima: A wide, gentle valley. If you take a small step in any direction, you stay at roughly the same low height. The old theory said this was great because the AI wouldn't get confused by tiny changes in data.
  • Sharp Minima: A narrow, jagged peak or a deep, thin canyon. If you take a tiny step, you fall up the side immediately. The old theory said this was bad because the AI was "memorizing" the training data too perfectly and would fail on new data.

This paper argues that this old rule is wrong.

The authors say: "Stop looking at the shape of the valley. Look at the map you are trying to draw."

They propose a Function-Centric Perspective. This means the shape of the solution (flat or sharp) isn't a sign of "good" or "bad" AI. Instead, it's a reflection of how complex the task is.


The Analogies

1. The "Perfectly Smooth" vs. "Roughly Detailed" Painting

Imagine you are an artist trying to paint a picture.

  • Task A: Paint a simple blue sky.
    • The Result: You can paint this with broad, sweeping, flat brushstrokes. The "valley" is wide and flat. This is easy.
  • Task B: Paint a hyper-realistic portrait of a face with every pore and hair visible.
    • The Result: To get this right, you need very precise, tight, and detailed brushstrokes. You can't just make broad, flat strokes; you have to be exact. The "valley" you end up in is narrow and sharp.

The Paper's Point: If you are trying to paint a complex face (a hard task), ending up in a "sharp" valley is actually a sign of success! It means your AI found a solution precise enough to handle the complexity. If you forced a "flat" solution on a complex task, you'd get a blurry, bad painting.

2. The "Rubber Band" vs. The "Steel Wire"

Think of the AI's decision boundary (the line it draws to separate cats from dogs) as a rubber band.

  • Flat Minima: A loose, floppy rubber band. It's easy to wiggle, but it might not fit the data tightly.
  • Sharp Minima: A tight, steel wire. It's rigid and hard to wiggle.

The paper shows that when we use Regularization (tools like "Data Augmentation" or "Weight Decay" that help the AI learn better), the AI often ends up with that tight steel wire.

  • Old View: "Oh no! The wire is too tight! It's sharp! It must be broken!"
  • New View: "Actually, the wire is tight because the task requires it to be precise. The AI is doing a better job because it learned to be precise, not because it got lucky." However, a crucial caveat remains: a tight wire can still sometimes indicate that the AI has simply memorized the data. The paper doesn't rule that out. The key takeaway is that sharpness alone is not a reliable signal either way—it doesn't automatically mean the model is broken, nor does it automatically mean it's perfect.

What Did They Actually Do?

The authors didn't just guess; they ran three types of experiments to prove their point:

1. The "Toy" Test (Single-Objective Optimization)
They gave the AI simple math problems to solve.

  • Some problems naturally had wide, flat solutions (like a sphere).
  • Some problems naturally had narrow, sharp solutions (like a twisted curve called "Rosenbrock").
  • Result: The AI found the "sharp" solution for the twisted problem and the "flat" solution for the sphere. The shape depended on the math problem, not on whether the AI was "smart" or "dumb."

2. The "Decision Boundary" Test
They made the AI separate two groups of dots.

  • When the dots were far apart, the AI made a wide, flat decision line.
  • When they forced the dots to be very close together (making the task harder), the AI made a tight, sharp decision line.
  • Result: Even though the line was "sharp," the AI still got 100% of the answers right. The sharpness was just a sign that the dots were close together, not a sign of failure.

3. The "Real World" Test (Images)
They trained big AI models (like ResNet and VGG) on image datasets (CIFAR, Tiny ImageNet) using standard tricks to make them better (Regularization).

  • The Surprise: The models that performed the best (most accurate, most robust, best at guessing correctly) were often the ones with the sharpest minima.
  • The "flat" models were actually the ones that performed the worst!

Why Does This Matter?

For a long time, researchers thought: "If I want a better AI, I must force it to find a flat valley." They built special tools (like SAM) to do exactly that.

This paper says: "Wait a minute. Sometimes, forcing the AI to be 'flat' makes it worse. If the task is complex, the AI needs to be sharp to get it right."

The Takeaway:

  • Sharpness is not always a bug — sometimes it's a feature. A sharp minimum often means the AI has learned a complex, precise function that fits the data perfectly.
  • However, sharpness can still coincide with memorization in some cases; it is simply not a guaranteed sign of it.
  • A flat minimum might just mean the AI is taking the "easy way out" and missing the details.

The "Goldilocks" Conclusion

The authors conclude that there is no single "perfect" shape for an AI solution.

  • If the task is simple, a flat solution is fine.
  • If the task is complex, a sharp solution is necessary.

Trying to force every AI to be "flat" is like trying to force a surgeon to use a butter knife because "knives are dangerous." Sometimes, you need the sharp edge to do the job right. The paper calls for us to stop judging AI by how "flat" its valley is, and start judging it by how well it actually solves the problem.

However, a critical practical question remains open: identifying exactly when sharpness reflects legitimate function complexity versus when it reflects memorization is still an unsolved problem. The paper successfully reframes the relationship between sharpness and generalization, showing that sharpness is not automatically a defect, but it does not yet provide a diagnostic tool to distinguish between a "good" sharp solution and a "bad" one in practice.

Therefore, sharpness should not be treated as an automatic defect to be eliminated. It can reflect complex, well-generalizing solutions, but it can also still reflect memorization in some cases. Distinguishing between these two scenarios in real-world applications remains an open challenge for the field.

Drowning in papers in your field?

Get daily digests of the most novel papers matching your research keywords — with technical summaries, in your language.

Try Digest →