Transformer-Based Pulse Shape Discrimination in HPGe Detectors with Masked Autoencoder Pre-training

This paper demonstrates that transformer-based models, particularly when enhanced by masked autoencoder pre-training, outperform traditional gradient-boosted decision trees in pulse-shape discrimination and energy regression for HPGe detectors by leveraging full waveform data and significantly reducing the need for labeled training samples.

Marta Babicz, Saúl Alonso-Monsalve, Alain Fauquex, Laura Baudis

Published Mon, 09 Ma
📖 5 min read🧠 Deep dive

Imagine you are a detective trying to solve a mystery in a very quiet, dark room. The "mystery" is a rare event called neutrinoless double-beta decay, a process that, if found, would rewrite our understanding of the universe. But the room is full of "noise"—background radiation from rocks, cosmic rays, and other particles that look almost exactly like the signal you are looking for.

To find the real signal, you need to listen very carefully to the "footsteps" (electrical signals) left by particles as they hit a giant, super-cold germanium crystal detector.

This paper is about teaching a computer to listen to these footsteps better than any human or traditional method ever could. Here is the breakdown in simple terms:

1. The Problem: The "Summary Sheet" vs. The "Full Recording"

Traditionally, when physicists looked at these electrical signals (waveforms), they were like a musician listening to a symphony and only writing down three numbers: "How loud was the first note? How long did it last? How quiet was the end?"

They threw away the rest of the music. They thought, "These three numbers are enough to tell if it's a rock rolling by (background) or a ghost walking by (the signal)."

The Paper's Idea: Why throw away the rest of the recording? What if we let the computer listen to the entire song, from the first note to the last, to hear the subtle differences we missed?

2. The New Tool: The "Transformer"

The authors used a type of AI called a Transformer. You might know Transformers from chatbots that write essays or translate languages. They are amazing at understanding context and long sequences.

  • The Analogy: Imagine a traditional method is like a security guard who only checks your ID badge (a few summary numbers). The Transformer is like a detective who watches your entire walk through the building: your gait, your speed, how you look around, and your posture. It sees the whole picture.

In this paper, the Transformer looked directly at the raw electrical waves from the detector, without compressing them into summary numbers first.

3. The Secret Sauce: "Masked Autoencoders" (The "Fill-in-the-Blanks" Game)

Training a super-smart AI usually requires millions of labeled examples (e.g., "This wave is a ghost," "That wave is a rock"). But in physics, labeling data is hard, expensive, and slow. You have to be an expert to say, "Yes, that's definitely a background event."

The authors used a clever trick called Masked Autoencoder (MAE) pre-training.

  • The Analogy: Imagine you have a library of 1 million books, but only 10,000 have their endings written down (labeled).
    • Old Way: You try to learn the story using only the 10,000 books with endings.
    • The Paper's Way: You take the 1 million books, rip out random pages (masking them), and ask the AI to guess what the missing pages say based on the rest of the story.
    • The Result: The AI becomes an expert at understanding the structure of the language (the physics of the detector) just by reading the unlabelled books. Once it's an expert at "filling in the blanks," you only need a few labeled books to teach it the specific mystery you are solving.

The Benefit: This made the AI 2 to 4 times more efficient. It needed far fewer labeled examples to become just as good as the old methods.

4. The Results: Who Won the Race?

The authors compared three racers:

  1. The Old Guard (GBDT): A classic machine learning model that uses the "summary sheet" (hand-crafted numbers).
  2. The Newbie (Transformer from Scratch): The new AI, but it had to learn everything from zero using only the few labeled examples.
  3. The Pro (Transformer with Pre-training): The new AI that played the "fill-in-the-blanks" game first.

The Winner: The Pro (Transformer with Pre-training).

  • It beat the Old Guard in every category.
  • It was especially good at spotting the "tricky" background events that usually fool the Old Guard.
  • It was also better at measuring the energy of the event (like guessing the weight of a package just by looking at it).

5. Why Does This Matter?

In the search for neutrinoless double-beta decay, every bit of background noise you can remove brings you closer to finding the "Holy Grail" of physics.

  • Better Signal: By using the full waveform, the AI can reject more background noise without accidentally throwing away the real signal.
  • Faster Science: Because the AI learns faster (thanks to the "fill-in-the-blanks" trick), scientists don't have to wait years to collect enough labeled data to train a new model. They can adapt quickly to new detectors or new experimental setups.

Summary

Think of this paper as upgrading a security system. Instead of just checking a person's ID badge (the old way), the new system watches their entire body language and movement history (the Transformer). And to make sure the system is smart enough to do this without needing a million human trainers, it first plays a game of "guess the missing puzzle piece" using millions of unlabelled photos (Masked Autoencoding).

The result? A smarter, faster, and more accurate detector that brings us one step closer to solving one of the universe's biggest mysteries.