GSVD for Geometry-Grounded Dataset Comparison: An Alignment Angle Is All You Need

This paper proposes a geometry-grounded framework for comparing datasets using the Generalized Singular Value Decomposition (GSVD) to derive an interpretable "angle score" that quantifies whether individual samples are better explained by one dataset, the other, or both.

Eduarda de Souza Marques, Arthur Sobrinho Ferreira da Rocha, Joao Paixao, Heudson Mirandola, Daniel Sadoc Menasche

Published 2026-03-12
📖 5 min read🧠 Deep dive

Here is an explanation of the paper "GSVD for Geometry-Grounded Dataset Comparison," translated into simple language with creative analogies.

The Big Idea: Comparing Datasets Without Losing the Plot

Imagine you have two huge libraries of books.

  • Library A is filled with mystery novels.
  • Library B is filled with sci-fi novels.

Usually, if you want to compare them, you might ask a librarian (a complex AI model) to read a few pages and tell you which library a new book belongs to. But this paper asks a different question: Can we compare the libraries themselves by looking at their "architecture" or "geometry"?

The authors propose a new way to look at data that treats every piece of information not just as a list of numbers, but as a direction in space. They want to know: Does this new book feel more like it belongs in the Mystery wing or the Sci-Fi wing?

The Problem: "Arbitrary Vectors" vs. "Geometry"

Most AI today treats data like a bag of random ingredients. It doesn't care that "red" and "blue" are related colors, or that "running" and "walking" are related actions. It just sees numbers.

This paper says: Stop treating data like a bag of marbles. Start treating it like a map.
If you have a map of the world, you can see that Paris and London are close, while Tokyo is far away. The authors want to build a map for their datasets so they can measure the "distance" and "angle" between them.

The Solution: The "Universal Translator" (GSVD)

To compare the two libraries (datasets), the authors use a mathematical tool called GSVD (Generalized Singular Value Decomposition).

The Analogy: The Shared Dance Floor
Imagine two groups of dancers: Group A (Mystery fans) and Group B (Sci-Fi fans). They are dancing in a huge room.

  • Sometimes, they dance in a way that is unique to them (Mystery fans do a specific spin; Sci-Fi fans do a specific jump).
  • Sometimes, they dance in a way that is the same (both groups clap their hands).

The GSVD is like a magic camera that finds a "Shared Dance Floor" (a common coordinate system). It separates the moves into three categories:

  1. The Mystery Moves: Unique to Group A.
  2. The Sci-Fi Moves: Unique to Group B.
  3. The Shared Moves: Moves both groups do.

This camera creates a "Joint Frame of Reference." Now, instead of looking at the messy original room, everyone is viewed through this clean, shared lens.

The Star of the Show: The "Alignment Angle" (θ\theta)

Once the data is on this shared dance floor, the authors introduce a simple score called the Alignment Angle. Think of this as a compass for a new piece of data (a new book, or a new image).

When a new item arrives, the compass points in a direction. The angle tells you everything you need to know:

  • Angle near 0°: The item is purely Mystery. It fits perfectly with Group A's unique moves.
  • Angle near 90°: The item is purely Sci-Fi. It fits perfectly with Group B's unique moves.
  • Angle near 45°: The item is ambiguous. It's doing a mix of both, or it's a "Shared Move" that fits neither group perfectly. It's like a book that is a "Sci-Fi Mystery."

Why is this cool?
Instead of a black-box AI saying "I'm 85% sure this is a Mystery," this method gives you a geometric reason: "This book is at a 10-degree angle from the Mystery direction, so it's definitely a Mystery." It's transparent and easy to understand.

How They Tested It: The MNIST Experiment

The authors tested this on MNIST, a famous dataset of handwritten digits (0 through 9).

  • They built a "Mystery Library" out of images of the number 4.
  • They built a "Sci-Fi Library" out of images of the number 9.

The Results:

  1. Clear Separation: When they tested images of 4s, the compass pointed almost straight to 0°. When they tested 9s, it pointed to 90°.
  2. The "Fuzzy" Ones: When they looked at the number 4 vs. 9, they found that some 4s looked a bit like 9s (maybe a curly tail). The compass for those specific images pointed to 45°.
  3. Visualizing the "Extreme" Directions: They could even generate "ghost images" of what a perfect 4 looks like according to their math, and what a perfect 9 looks like. These ghost images showed exactly why the computer thought they were different (e.g., the sharp angles of the 4 vs. the round loops of the 9).

Why Does This Matter?

  1. No More Black Boxes: Instead of guessing why an AI made a mistake, you can look at the angle and say, "Ah, this image was at 45 degrees, so the AI was confused because it looked like both classes."
  2. Better Data Cleaning: If you have a dataset full of "bad" data (like a photo of a cat labeled as a dog), this angle will be weird. It will point somewhere in the middle, flagging it for a human to check.
  3. Understanding Similarity: It helps us understand how two things are similar. Are they similar because they share a lot of features, or because they are just both "vague"?

Summary in One Sentence

This paper gives us a geometric compass that measures exactly how much a piece of data belongs to one group versus another, turning complex math into a simple angle that tells us if something is "Team A," "Team B," or "Confused."