Selection and processing of calibration samples to measure the particle identification performance of the LHCb experiment in Run 2

This paper outlines the novel strategy developed by the LHCb experiment during Run 2 to select and process calibration samples using a dedicated online-offline computing model, enabling precise measurement of particle identification performance and data-quality monitoring across various decay channels.

Original authors: Roel Aaij, Lucio Anderlini, Sean Benson, Marco Cattaneo, Philippe Charpentier, Marco Clemencic, Antonio Falabella, Fabio Ferrari, Marianna Fontana, Vladimir Gligorov, Donal Hill, Thibaud Humair, Chris
Published 2018-03-02
📖 5 min read🧠 Deep dive

This is an AI-generated explanation of the paper below. It is not written or endorsed by the authors. For technical accuracy, refer to the original paper. Read full disclaimer

Imagine the LHCb experiment at CERN as a massive, high-speed supermarket that is constantly scanning millions of shoppers (particles) every second. The goal isn't to buy groceries, but to find very specific, rare shoppers (like "beauty" or "charm" particles) who might be carrying a secret message about new physics.

However, the supermarket is chaotic. There are millions of people, and many of them look very similar. A "pion" might look like a "kaon," or an "electron" might look like a "muon." If the security guards (the detectors) can't tell them apart, they might let the wrong people through or kick out the VIPs they are actually looking for.

This paper is about how the LHCb team built a super-smart ID system and a quality control lab to make sure their security guards are perfect at their jobs.

Here is the breakdown of their strategy using everyday analogies:

1. The Problem: The "Imposter" Crowd

The LHCb detector sees five main types of charged particles: electrons, muons, pions, kaons, and protons.

  • The Challenge: Imagine trying to spot a specific celebrity in a crowd of 100,000 people where everyone is wearing a similar hat. If your security camera (the detector) makes a mistake, you might think a random fan is the celebrity, or miss the celebrity entirely.
  • The Solution: They need to know exactly how good their cameras are at telling these people apart. But you can't just ask the cameras, "Are you sure?" because they might be biased. You need a Control Group.

2. The Control Group: The "Calibration Samples"

To test the cameras, the team needs a group of people whose identities are 100% guaranteed without looking at the cameras.

  • How they do it: They look for specific "family reunions" (particle decays) where the rules of physics make the identity obvious.
    • Example: If a particle decays into two muons, and we know the math perfectly, we know those two particles must be muons. We don't need the camera to tell us; the math tells us.
  • The Result: They have a "Gold Standard" dataset. They can run their ID software on these known muons and see: "Okay, the software correctly identified 99% of them, but it got confused 1% of the time." This tells them exactly how accurate their system is.

3. The New "Turbo" System: Speed vs. Detail

In the past (Run 1), the supermarket had a two-step process:

  1. Hardware Trigger: A quick glance to see if something interesting happened.
  2. Software Trigger: A deep dive to reconstruct the whole event.

In Run 2 (the focus of this paper), they upgraded to a hybrid system:

  • The "Turbo" Stream: For most events, they do the deep analysis instantly while the data is still flowing. They save the "receipt" (the result) but throw away the "raw video footage" to save space.
  • The "Full" Stream: For special cases, they keep the raw video footage.
  • The Innovation: They created a special "Calibration Stream" (TurboCalib). This is like a double-check station. They take the same event, process it once with the fast "online" method and once with the slow, detailed "offline" method. By comparing the two, they can see if the fast method is missing anything or if the slow method is changing its mind. This ensures that even if they only save the "receipt" later, they know the receipt is accurate.

4. The "Magic Mirror" (Correcting Simulations)

Scientists use computer simulations to predict what should happen. But computers aren't perfect; they are like a cartoon version of reality.

  • The Issue: The simulation might think a kaon looks like a pion 5% of the time, but in real life, it's 7%. If they use the simulation to plan their experiment, they will be wrong.
  • The Fix: They use their "Gold Standard" calibration data to create a Magic Mirror.
    • They take the cartoon simulation and "paint over" it with the real data.
    • They use a technique called sPlot (think of it as a statistical magic wand) to separate the signal from the background noise.
    • They then "resample" or "transform" the simulation data so that it behaves exactly like the real calibration samples. Now, their computer models are perfectly tuned to reality.

5. Why This Matters for the Future

The paper explains that this system is crucial for Run 3 of the LHC.

  • In the future, the data flow will be so fast that they won't be able to save the "raw video footage" for anything except the most interesting events. They will only save the "receipts."
  • Because they can't go back and re-analyze the raw footage later, the online analysis must be perfect.
  • The calibration samples described in this paper are the training wheels that ensure the online system is so good that they can eventually throw the training wheels away.

Summary

Think of this paper as the User Manual for a High-Tech Security System.

  1. We built a test lab using known particles (Calibration Samples) to measure how often our ID system makes mistakes.
  2. We built a double-processing machine that checks our work in real-time to ensure speed doesn't sacrifice accuracy.
  3. We created a "Magic Mirror" to fix our computer simulations so they match reality perfectly.
  4. We are ready for the future, ensuring that even when we have to make split-second decisions without saving all the raw data, we can still trust our results.

This allows physicists to search for the tiniest, rarest particles with the confidence that their "security guards" are telling the truth.

Drowning in papers in your field?

Get daily digests of the most novel papers matching your research keywords — with technical summaries, in your language.

Try Digest →