Imagine you are trying to figure out how two different people are thinking about the same thing. Maybe you want to know if a human brain and a computer program are "thinking" in the same way when they look at a picture of a cat.
For a long time, scientists have tried to answer this by using a single ruler to measure the distance between their thoughts. But this paper argues that one ruler isn't enough. Depending on how you measure, you might get completely different answers.
Here is the story of the paper, broken down into simple concepts and analogies.
1. The Problem: The "One-Ruler" Trap
Imagine you are trying to sort a pile of fruits.
- Ruler A measures weight.
- Ruler B measures color.
- Ruler C measures sweetness.
If you only use the Weight Ruler, a heavy watermelon and a heavy cantaloupe look very similar. But if you use the Color Ruler, they look totally different (green vs. orange). If you use the Sweetness Ruler, they might be similar again.
In the world of AI and brains, scientists have been using different "rulers" (mathematical formulas) to see if two systems are similar. Some rulers check if the systems organize data geometrically (like arranging shapes in a specific pattern). Others check if they can predict the same answers (like a teacher grading a test).
The problem? These rulers often disagree. One might say two AI models are twins, while another says they are strangers. This makes it hard to know what's actually going on.
2. The Experiment: Testing the Rulers
The authors decided to test all these different rulers on two groups:
- Artificial Brains: 35 different computer vision models (some trained to recognize cats, some trained to predict the next word, some built like human neurons, some built like modern Transformers).
- Real Brains: Scans of human brains looking at 1,000 natural images.
They asked two simple questions:
- For AI: Can the ruler tell the difference between a model trained with "Supervised Learning" (learning with a teacher) and "Self-Supervised Learning" (learning by guessing)?
- For Humans: Can the ruler tell the difference between the part of the brain that sees edges (V1) and the part that sees complex shapes (V4)?
The Result:
- The "Geometry" Rulers (like RSA and SoftMatch) were the best detectives. They looked at how the information was arranged and could easily tell the different families of models and brain regions apart.
- The "Prediction" Rulers (like Linear Predictivity) were weaker. They were too flexible. They could twist and turn the data to make anything look similar, so they missed the unique "fingerprints" that made each system special.
3. The Solution: The "Super-Blender" (Similarity Network Fusion)
If one ruler isn't enough, why not use all of them?
The authors tried a simple average (mixing all the rulers together), but that was like making a smoothie where the strong flavors cancel out the subtle ones. It didn't work well.
Instead, they used a clever technique called Similarity Network Fusion (SNF). Think of this not as a blender, but as a group of detectives meeting in a room.
- Detective A (Geometry) says: "These two models look alike because they arrange their data the same way."
- Detective B (Tuning) says: "I agree, and I also see they react to specific details similarly."
- Detective C (Prediction) says: "They both get the right answer, but their internal logic is different."
The SNF algorithm listens to all of them. It only draws a line between two models if most detectives agree they are similar. If one detective thinks they are similar but the others disagree, the line gets weaker. If they all agree, the line becomes a thick, solid highway.
The Magic:
When they used this "Group Detective" approach, the results were amazing.
- In AI: It perfectly grouped the models. It showed that all "Self-Supervised" models (regardless of whether they were old-school or new-school) formed one big family, while "Supervised" models formed another. It even showed that hybrid models (mixing old and new tech) were actually cousins to the self-supervised family.
- In Brains: It revealed the brain's structure perfectly. It showed the clear hierarchy of the visual cortex, from simple edge detectors to complex object recognizers, much clearer than any single ruler could.
4. The Big Takeaway
This paper teaches us two main lessons:
- There is no single "Truth" in similarity. Whether two brains or computers are "similar" depends entirely on how you look at them. Some similarities are about the shape of the thought, others are about the result of the thought.
- Combination is key. By fusing these different perspectives, we get a much clearer, more accurate map of how intelligence (both artificial and biological) is organized.
In short: Don't just look at a painting through one colored lens. Look through many, and then combine what you see to understand the masterpiece. That's what this paper does for the science of AI and the brain.
Get papers like this in your inbox
Personalized daily or weekly digests matching your interests. Gists or technical summaries, in your language.