Physics-driven Comparative Analysis of Various Statistical Distance Metrics and Normalizing Functions

This paper presents a data-driven comparative analysis of various statistical distance metrics and normalizing functions using electron and photon events from a decaying Kr-83 isotope to evaluate the stability of a dimensionless Parameter of Interest under different conditions.

Original authors: Nafis Fuad (Center for Exploration of Energy,Matter, Indiana University, Bloomington, IN 47405, USA)

Published 2026-04-16
📖 5 min read🧠 Deep dive

This is an AI-generated explanation of the paper below. It is not written or endorsed by the authors. For technical accuracy, refer to the original paper. Read full disclaimer

Imagine you are a detective trying to tell the difference between two types of suspects: Electrons (charged particles) and Photons (light particles). In the world of physics, these two leave behind very specific "footprints" when they hit a detector. But sometimes, the footprints look a little blurry, or the camera isn't perfect.

This paper is essentially a tool test. The author, N. Fuad, is asking: "When I have two piles of footprints (data), which mathematical ruler is the best at measuring how different they really are?"

Here is the breakdown of the story, using simple analogies:

1. The Crime Scene: The Detector

The "detective" used a giant, super-cold camera made of Germanium (a semiconductor) to watch an isotope called Krypton-83 decay.

  • The Setup: Think of this camera as a high-speed microphone. When a particle hits it, it creates a sound wave (a signal).
  • The Difference:
    • Electrons are like a sprinter who stops abruptly. They hit the camera and stop almost instantly, creating a signal that rises very sharply.
    • Photons are like a marathon runner who slows down gradually. They travel further inside the camera before stopping, creating a signal that rises slowly.
  • The Goal: The team needed to separate these two groups perfectly. To do that, they created a "score" (a number between 0 and 1) for every event based on how sharp the signal was. This score is their Parameter of Interest (PoI).

2. The Problem: Too Many Rulers

In math and science, there are dozens of ways to measure "distance" or "difference" between two groups of data. Some are called Hellinger, some Wasserstein, some Kolmogorov-Smirnov, and so on.

Imagine you have two piles of sand.

  • Ruler A measures the difference in the height of the piles.
  • Ruler B measures how much sand you have to move to turn one pile into the other.
  • Ruler C measures the difference in the shape of the piles.

The problem is, some of these rulers are finicky.

  • If you change the size of the grains of sand (discretization), some rulers give wildly different answers.
  • If you only have a few grains of sand (low statistics), some rulers break or give nonsense numbers.
  • Some rulers are so sensitive that they say two piles are "completely different" even if they are just slightly different.

3. The Experiment: The "Normalization" Trick

The author realized that some of these rulers produce numbers that are too big or too small to compare fairly. So, they introduced Normalizing Functions.

Think of this like a translator or a compressor.

  • Imagine a ruler that measures distance in "light-years" (huge numbers) and another in "inches" (tiny numbers). It's hard to compare them.
  • The author invented special "squeeze functions" (like n(x)n(x)) that take any huge number and squash it down into a neat, tidy box between 0 and 1.
    • 0 means "Identical."
    • 1 means "Completely Different."

They tested four different "squeezers" to see which one made the rulers play nice together.

4. The Results: Who Won?

After running thousands of tests with their electron and photon data, here is what they found:

  • The Unreliable Ones:

    • Wasserstein-2 (W2W_2): This ruler is very sensitive to how you slice the data. If you change the grain size of your sand, this ruler panics and gives a different answer.
    • Fisher-Rao & LL_\infty: These are great if you have perfect data, but if you have a small sample size (few events), they become unstable and unreliable.
    • The "Saturators": Some rulers (like W2W_2 and LL_\infty) hit the "1.0" ceiling too easily. They say "These are totally different!" even when they are just mostly different. They lose the ability to tell the difference between "very different" and "maximally different."
  • The Winner: The JS\sqrt{JS} Distance

    • The Square Root of Jensen-Shannon (JS\sqrt{JS}) distance was the champion.
    • Why? It was the most stable. Whether they changed the size of the data grains, the number of events, or which "squeezer" they used, this ruler gave consistent, reliable answers.
    • It didn't get confused by small changes in the data, and it didn't break when the sample size was small.

5. The Takeaway

The paper concludes that if you are trying to compare two probability distributions (like electron vs. photon signals) in a noisy, real-world environment:

  1. Don't just pick a ruler at random; some are too sensitive to the "noise."
  2. Use the JS\sqrt{JS} distance. It's the most robust tool in the toolbox.
  3. If you need to squash big numbers into a 0-to-1 range, you can use simple mathematical "squeezers," but they all seem to work about the same for this specific job.

In short: The author tested a bunch of mathematical rulers to see which one is the best at telling apart electrons from photons. They found that the JS\sqrt{JS} ruler is the most reliable detective, while the others tend to get confused by the details of the experiment.

Drowning in papers in your field?

Get daily digests of the most novel papers matching your research keywords — with technical summaries, in your language.

Try Digest →