Cutting Through the Noise: On-the-fly Outlier Detection for Robust Training of Machine Learning Interatomic Potentials

This paper introduces an automated, unsupervised on-the-fly outlier detection method that uses an exponential moving average of the loss distribution to down-weight noisy training data, enabling the robust and efficient training of machine learning interatomic potentials without the need for manual filtering or additional expensive calculations.

Original authors: Terry C. W. Lam, Niamh O'Neill, Christoph Schran, Lars L. Schaaf

Published 2026-02-10
📖 4 min read☕ Coffee break read

This is an AI-generated explanation of the paper below. It is not written or endorsed by the authors. For technical accuracy, refer to the original paper. Read full disclaimer

The Problem: The "Bad Apple" in the Digital Lab

Imagine you are a world-class chef trying to learn how to make the perfect soup. To learn, you study thousands of recipes written by other cooks. However, some of those recipes are terrible: one might say "add a cup of salt" instead of "a pinch," and another might have a typo that says "add 5 gallons of vinegar."

If you try to follow every recipe exactly as written, you will end up making a soup that is either undrinkable or wildly inconsistent. You’ll be "overfitting"—meaning you are working incredibly hard to master mistakes that aren't actually part of good cooking.

In the world of science, researchers use Machine Learning Interatomic Potentials (MLIPs). These are AI models that act like "digital chefs." They learn how atoms interact so they can simulate chemical reactions without needing a massive, expensive supercomputer to solve complex physics equations every single second.

The catch? The "recipes" (the data) used to train these AIs often come from digital simulations that didn't finish properly. They contain "numerical noise"—basically, digital typos that make the atoms behave in ways that are physically impossible.


The Solution: The "Smart Sieve" (On-the-fly Outlier Detection)

Usually, if a scientist finds bad data, they have to go back, fix the original math, and restart the entire training process from scratch. This is like a chef throwing away their entire cookbook and starting over because they found three bad recipes. It’s slow, expensive, and exhausting.

The authors of this paper created a way to fix this while the AI is learning. They call it "On-the-fly Outlier Detection."

Think of it like a Smart Sieve in a kitchen. As the chef (the AI) reads each recipe, they don't just blindly follow it. Instead, they keep a running mental note of what a "normal" recipe looks like.

  1. The Comparison: As the AI reads a new recipe, it asks: "Does this instruction look wildly different from the last 1,000 instructions I've read?"
  2. The Weighting: If the instruction says "add a pinch of salt," the AI gives it full attention. If the instruction says "add a bucket of salt," the AI recognizes this as an "outlier."
  3. The "Muting" Effect: Instead of throwing the recipe away (which would be a waste of time), the AI simply turns the volume down on that specific instruction. It says, "I see you, but I'm not going to let you influence my cooking too much."

This is done using a mathematical trick called an Exponential Moving Average. It’s like a chef who has a "gut feeling" that evolves over time; as they gain more experience, their ability to spot a "bad recipe" becomes sharper and sharper.


Why It Matters: Real-World Results

The researchers tested this "Smart Sieve" on three different levels, and the results were impressive:

  • Level 1: The Small Scale (The Soup Test). They took a standard dataset and intentionally "poisoned" 10% of it with bad data. The AI trained with the Smart Sieve ignored the poison and learned the correct physics, while the standard AI got confused and failed.
  • Level 2: The Liquid Test (The Water Test). They trained an AI on "messy" data about liquid water. A normal AI predicted that water wouldn't flow correctly (it messed up the "diffusion"). The Smart Sieve AI ignored the messy data and correctly predicted how water molecules actually move.
  • Level 3: The Giant Scale (The Foundation Model). They applied this to a massive "Foundation Model"—an AI trained on millions of different molecules (like a library of every recipe ever written). Even at this massive scale, the Smart Sieve helped the AI ignore "unphysical" structures (like atoms overlapping in impossible ways), making the AI three times more accurate in predicting energy.

The Bottom Line

This paper provides a way to train incredibly powerful scientific AIs using "imperfect" data. It allows scientists to build better digital tools for discovering new medicines and materials without having to spend years cleaning up their data by hand. It’s a way to cut through the noise and find the truth, even when the data is shouting lies.

Drowning in papers in your field?

Get daily digests of the most novel papers matching your research keywords — with technical summaries, in your language.

Try Digest →