Searching for Anomalies with Foundation Models

This paper investigates unexpected anomalies in CMS experiment data identified by the OmniLearned foundation model, finding that while background estimates perform well in validation regions, they fail to accurately model the signal region, prompting a call for further scrutiny.

Original authors: Vinicius Mikuni, Benjamin Nachman

Published 2026-03-26
📖 5 min read🧠 Deep dive

This is an AI-generated explanation of the paper below. It is not written or endorsed by the authors. For technical accuracy, refer to the original paper. Read full disclaimer

Imagine you are a detective trying to find a rare, mysterious criminal in a massive city. The city is full of ordinary people doing normal things (like going to work or shopping), but you suspect there's a hidden gang of spies doing something weird.

This paper is about two detectives (scientists) who tried to find these "spies" (new physics) in a giant dataset of particle collisions from the Large Hadron Collider (CMS experiment). They used a new, high-tech tool called a Foundation Model, which is like a super-smart AI that has read every book in the library to understand how the world usually works.

Here is the story of their investigation, broken down simply:

1. The Setup: The "Omni-Learned" Detective

The scientists used an AI called OmniLearned. Think of this AI as a detective who has studied millions of photos of "normal" traffic. It knows exactly what a normal car, a bicycle, or a pedestrian looks like.

  • The Small Detective: They first tried a smaller version of this AI. It worked great! It found the "Top Quark" (a known particle, like finding a specific type of car) exactly where physics said it should be.
  • The Big Detective: Then, they tried the Large version of the AI, which is much smarter and has seen way more data. They expected it to be even better.

2. The Glitch: The "Weird Noise"

When the Large Detective looked at the data, it found the Top Quark, but it also started screaming about something else. It pointed to a specific area in the data (a "mass sideband") where the numbers didn't look smooth.

  • The Analogy: Imagine you are listening to a radio station playing smooth jazz. Suddenly, the Large Detective hears a strange, rhythmic static in the background. The Small Detective didn't hear it; the Large Detective did.
  • The scientists thought, "Is this a new particle? Or is the AI just confused?"

3. The Investigation: Checking the Evidence

To be sure, the scientists had to do a full forensic audit. They couldn't just trust the AI; they had to prove the "static" wasn't just a glitch in the radio.

  • The Background Check: They used a method called ABCD. Imagine you have four rooms in a house. You know how many people are in three of them. If you know the rules of the house, you can mathematically guess how many people should be in the fourth room (the one with the "anomaly").
  • The Result: In most rooms, the math worked perfectly. The "noise" was just normal background static. But in the specific room the Large Detective pointed to, the math failed. The actual data didn't match the prediction. There was an unexpected "bump" in the data around a mass of 150 GeV.

4. The Suspect: The "Double Higgs" Theory

Since the background math didn't fit, the scientists asked: "What if there is a new signal here?"

  • They tested a hypothesis: Could this be Double Higgs Bosons (two Higgs particles created at once)?
  • The Fit: When they added this "Double Higgs" suspect to their model, the messy data suddenly looked much neater. It was like putting the missing puzzle piece in place.
  • The Catch: To make the math work, they had to assume there were 4,000 times more Double Higgs events than the Standard Model of physics predicts. That's like finding a needle in a haystack, but the needle is actually a giant golden statue. It's statistically unlikely to be real, but the pattern is suspicious.

5. The Twist: The "Substructure" Clue

The scientists dug deeper. They looked at the "substructure" of the particles (how the energy is packed inside the jets).

  • They found that the weird events had a specific signature: One jet was heavy (around 150 GeV), and the other was also heavy (over 100 GeV).
  • They also checked if these jets contained "bottom quarks" (a specific type of particle). When they filtered for those, the "weirdness" got even stronger.
  • The Confusion: They tried using a different, specialized tool designed specifically to find Double Higgs. Surprisingly, that tool didn't see the same weird events! Only the giant "OmniLearned" AI saw them. This suggests the AI is picking up on a very subtle, strange pattern that human-designed tools miss.

6. The Conclusion: "We Don't Know Yet"

The paper ends with a very honest conclusion:

  • The Good News: The Large AI found something the Small AI missed. The background models (our understanding of "normal") can't explain this specific bump in the data.
  • The Bad News: It's probably not a new particle (yet). The numbers are too extreme, and other tools don't see it.
  • The Mystery: The Large AI might be "hallucinating" (seeing patterns that aren't there), or it might be detecting a very subtle flaw in our current physics models.

The Takeaway:
The scientists are inviting the rest of the world to look at this "glitch." They are saying, "We have a strange signal that our best models can't explain. It might be a mistake in our math, or it might be the first hint of something new. Please come help us figure it out!"

It's a reminder that in science, sometimes the most interesting discoveries start with a weird noise that no one else can hear.

Drowning in papers in your field?

Get daily digests of the most novel papers matching your research keywords — with technical summaries, in your language.

Try Digest →