Energy efficiency of a GPU-based computing system for High Energy Physics experiments

This paper introduces energy efficiency as a new metric for evaluating GPU hardware and algorithm optimizations in High Energy Physics, presenting a model applied to the LHCb experiment's HLT1 trigger to relate throughput with hardware specifications and guide the development of sustainable computing ecosystems.

Original authors: Jiahui Zhuo, Arantza Oyanguren, Álvaro Fernández Casani, Luca Fiorini, Valerii Kholoimov

Published 2026-05-01
📖 5 min read🧠 Deep dive

This is an AI-generated explanation of the paper below. It is not written or endorsed by the authors. For technical accuracy, refer to the original paper. Read full disclaimer

Imagine you are running a massive, high-speed sorting factory. Every second, millions of tiny packages (data from particle collisions) arrive on a conveyor belt. Your job is to quickly inspect each package, decide if it's interesting, and sort it. This is what the LHCb experiment at CERN does with data from the Large Hadron Collider.

For a long time, this factory used standard "CPU" workers. But as the factory gets busier, these workers are getting tired and the electricity bill is skyrocketing. So, the team decided to hire a new kind of worker: GPUs (Graphics Processing Units). Think of GPUs as a team of thousands of super-fast, specialized robots that can work in parallel.

This paper is about figuring out which robots are the best to hire, not just by how fast they work, but by how much energy they waste.

The Problem: Speed vs. Energy

Usually, when you buy a new machine, you look at its speed. But in a giant factory, speed isn't everything. If a machine is super fast but guzzles electricity like a thirsty elephant, it costs too much to run and generates so much heat you need expensive air conditioning.

The authors wanted a new way to measure these robots: Energy Efficiency. This is simply: How many packages can this robot sort for every single drop of electricity it uses?

The Experiment: Testing the Robots

The team set up a test using 10 different models of NVIDIA GPUs (ranging from older models to the very newest, cutting-edge ones). They ran the exact same sorting task (called HLT1) on all of them.

They measured two things:

  1. Throughput: How many packages per second the robot sorted.
  2. Power: How much electricity the robot actually drank while doing the job.

The Surprising Discovery: The "Thirsty" vs. "Efficient" Robots

Here is the twist they found: Just because a robot is powerful doesn't mean it will run at its maximum power limit.

Think of a car. If you drive a Ferrari in heavy traffic, you might never reach its top speed, and you won't use all its fuel.

  • The "Power-Limited" Robots: Some older or specific workstation robots hit their "fuel cap" (TDP). They are working as hard as they can, but they are capped by their design. They are like a runner sprinting until they are out of breath.
  • The "Non-Power-Limited" Robots: Many of the newer, high-end robots were actually not using their full fuel capacity. Even though they were sorting packages at 100% speed, they weren't drinking as much electricity as their specs said they could. They were like a runner who could sprint faster but was only jogging because the task didn't require a full sprint.

The Magic Formula: Predicting the Future

The team didn't just measure these 10 robots; they built a predictive recipe (a mathematical model).

They realized that a robot's speed depends on two main things:

  1. How many hands it has (Number of Cores).
  2. How fast it can grab items (Memory Bandwidth).

However, they found that doubling the number of hands doesn't double the speed. Because the robots have to talk to each other and wait for instructions, the speed gains get smaller as you add more hands. It's like adding more chefs to a kitchen; eventually, they just get in each other's way.

Using this recipe, they can now look at the "spec sheet" of a brand-new robot that hasn't even been built yet. By plugging in its number of cores and memory speed, they can predict:

  • How fast it will sort packages.
  • How much electricity it will drink.
  • How energy-efficient it will be.

The Winner

When they ranked the robots by energy efficiency (packages per joule of electricity), the results were surprising:

  • The fastest robot (RTX PRO 6000) was not the most efficient. It was fast, but it drank a lot of power.
  • The most efficient robot (RTX PRO 4000) was actually slower, but it was so frugal with electricity that it sorted more packages per drop of energy than the giants.

Why This Matters

The LHCb experiment is planning to upgrade its factory soon. They can't afford to buy and test every single new robot model that comes out; it would take too long and cost too much.

Thanks to this paper, they can now look at the brochure of a future robot, run it through their "recipe," and know immediately if it's a good hire. They can choose the robot that gives them the best balance of speed and low energy bills, ensuring their massive data factory stays sustainable and affordable for years to come.

In short: They figured out how to predict exactly how much a new computer chip will cost to run and how fast it will work, just by reading its specifications, saving the scientists time, money, and electricity.

Drowning in papers in your field?

Get daily digests of the most novel papers matching your research keywords — with technical summaries, in your language.

Try Digest →