The mathematical landscape of partial information decomposition: A comprehensive review of properties and measures

This paper provides a comprehensive review of the Partial Information Decomposition (PID) framework by integrating diverse formalisms into a unified language, systematically evaluating their adherence to known properties, mapping theorems that reveal relationships and incompatibilities between these properties, and charting a path for future theoretical and empirical advancements.

Alberto Liardi, Keenan J. A. Down, George Blackburne, Matteo Neri, Pedro A. M. Mediano

Published Tue, 10 Ma
📖 5 min read🧠 Deep dive

Imagine you are trying to figure out how a group of friends (let's call them Sources) are helping you solve a mystery (the Target).

Sometimes, Friend A and Friend B tell you the exact same clue. That's Redundancy.
Sometimes, Friend A tells you one piece of the puzzle, and Friend B tells you a different piece, and you only get the full picture when you combine them. That's Synergy.
Sometimes, Friend A knows something Friend B doesn't, and that's Unique Information.

For decades, scientists had a great tool called Information Theory to measure how much information exists. But it had a blind spot: it couldn't tell the difference between "two friends repeating the same joke" (redundancy) and "two friends telling different parts of a story that only make sense together" (synergy).

Enter Partial Information Decomposition (PID). It's a mathematical framework designed to split that information into those three buckets. But here's the problem: since PID was invented in 2010, mathematicians have been arguing fiercely about how to do the splitting. They've built over 20 different "scales" (measures) to weigh the information, and they all give slightly different answers.

This paper is like a comprehensive map and rulebook for this chaotic landscape. The authors, a team of researchers from Imperial College London and others, have done three main things to help us navigate this mess:

1. The "Rulebook" (The Properties)

Imagine every PID measure is a different type of scale. Some scales are heavy, some are light, some are digital, some are analog. The authors listed all the "rules" (called axioms) that a good scale should follow.

  • Symmetry: It shouldn't matter if you swap Friend A and Friend B; the result should be the same.
  • Positivity: You shouldn't get a "negative amount of information" (though some modern scales argue that negative info is actually "misinformation," like a lie).
  • The "Copy" Test: If the mystery is just a copy of what the friends said, the scale should behave in a specific, logical way.

The paper lists about 20 of these rules. The big discovery? You can't satisfy all of them at once. It's like trying to build a car that is simultaneously the fastest, the most fuel-efficient, and the cheapest to build. You have to pick your priorities.

2. The "Scorecard" (The Comparison)

The authors took every single PID measure created so far (about 19 of them) and ran them through a checklist. They created a giant table (Table 5 in the paper) that acts like a scorecard.

  • The "Gold Standard" Measures: Some measures, like IBROJA, satisfy the most rules. They are the "Swiss Army Knives" of the field.
  • The "Specialists": Some measures, like ICCS, break the rule of "no negative numbers" because they think negative numbers are useful for spotting lies. They are like a specialized tool for a specific job.
  • The "Outliers": Some measures fail the basic "Copy Test," meaning they give weird results when the friends are just repeating each other.

By grouping these measures based on which rules they follow, the authors showed us that the field isn't just a random collection of ideas; it's actually organized into distinct "families" or "philosophies."

3. The "No-Go Zones" (The Theorems)

The paper maps out the "incompatibilities." Think of this as a traffic map showing which roads lead to dead ends.

They proved mathematically that if you want your scale to follow Rule A (e.g., "No negative numbers") and Rule B (e.g., "The Copy Test"), you cannot also follow Rule C (e.g., "The Chain Rule").

  • The Big Conflict: The biggest argument in the field is between Statistical Invariance (the idea that the labels of the data don't matter) and Mechanistic Redundancy (the idea that the way the data is generated matters). The paper shows you can't have both. You have to choose: do you care about the raw numbers, or the story behind how they were made?

Why Does This Matter?

Before this paper, if you were a biologist studying brain cells or a data scientist analyzing social networks, you wouldn't know which PID measure to use. You might pick one that gives you the answer you want to see, rather than the right one.

This paper acts as a guide for the traveler:

  • If you care about "mechanisms" (how things actually work physically), you should pick a measure that allows for "mechanistic redundancy" (like IBROJA or Ired).
  • If you care about "communication" (how much data is actually being sent), you should pick a measure that ensures all numbers are positive (like Imin or IMMI).
  • If you are dealing with "lies" or "misinformation" (where data confuses you), you might want a measure that allows negative numbers (like ICCS).

The Takeaway

The authors aren't saying "Here is the one true answer." Instead, they are saying: "Here is the map of the territory. We know the roads are full of dead ends and contradictions. Here is exactly which path leads to which destination, so you can choose the right tool for your specific job."

They have turned a confusing "multiverse" of math into a clear, organized landscape, helping scientists stop arguing about which measure is best and start using the right measure for the problem they are solving.