Evaluating Graphical Perception Capabilities of Vision Transformers

This paper benchmarks Vision Transformers against CNNs and humans on elementary graphical perception tasks inspired by Cleveland and McGill, revealing that while ViTs excel in general vision, they exhibit significant limitations in aligning with human-like perceptual accuracy for interpreting visualizations.

Poonam Poonam, Pere-Pau Vázquez, Timo Ropinski

Published 2026-02-23
📖 5 min read🧠 Deep dive

Imagine you have two very different types of students trying to learn how to read a map.

Student A (The Human) is an experienced traveler. They look at a map and instantly understand that a long line means a long road, a big circle means a large city, and a steep angle means a sharp turn. They don't need to count every single pixel; they just "get" the shape and size intuitively.

Student B (The Old Robot - CNN) is a student who learned by looking at millions of tiny, blurry photos. They are very good at recognizing patterns like "this looks like a cat" or "that looks like a car" because they've seen them a thousand times. They are good at spotting details, but they sometimes struggle to see the whole picture at once.

Student C (The New Robot - Vision Transformer or ViT) is the star student of the modern era. They are famous for being able to look at a whole room and instantly understand how the furniture relates to the walls, the windows, and the door. They are incredibly smart at connecting distant dots and understanding complex scenes.

The Big Question

This paper asks a simple but crucial question: If we give these robots a basic math test involving shapes and sizes (like reading a bar chart), will the New Robot (ViT) act more like the Human traveler, or will it still act like a robot?

In the world of data visualization (charts, graphs, maps), we need computers to "see" things the way humans do. If a computer thinks a bar chart is easy to read but a human finds it confusing, the computer might give us bad advice.

The Experiment

The researchers set up a series of "elementary school" tests for these robots. They didn't ask them to write a poem or diagnose a disease. They asked them to do basic visual judgments, just like the famous Cleveland and McGill studies from the 1980s:

  • Length: Which bar is longer?
  • Angle: Which slice of the pie is bigger?
  • Position: Which dot is higher up?
  • Area: Which shape covers more space?
  • Counting: How many dots are in this cloud?

They tested three versions of the New Robot (ViT):

  1. The Pure Transformer: Just the raw attention mechanism.
  2. The Hybrid (CvT): A mix of the old robot's local vision and the new robot's global vision.
  3. The Swin: A version that looks at the image in small "windows" first, then zooms out.

The Results: The Plot Twist

Here is where it gets interesting. You might expect the New Robot (ViT) to be perfect because it's the "state-of-the-art" technology. But the results were surprising:

1. The Human is still the Gold Standard.
In almost every test, the Human traveler was the most accurate. They could compare lengths and positions with incredible precision.

2. The New Robot (ViT) is actually worse than the Old Robot (CNN) at basic math.
This is the biggest shock. While ViTs are amazing at recognizing complex scenes (like "is this a dog running in a park?"), they were struggling with the simple stuff.

  • When asked to compare the length of two bars, the ViT made more mistakes than the older CNN models.
  • When asked to count dots in a cloud, the ViT got very confused.
  • The ViT seemed to think some difficult tasks (like judging curved lines) were easy, and some easy tasks (like comparing lengths) were hard. It had a completely different "sense" of reality than humans.

3. The "Window" Robot (Swin) was the best of the bunch, but still not human.
The Swin Transformer performed the best among the robots, but it still couldn't match human accuracy. It was like a student who is great at writing essays but keeps failing basic arithmetic.

The Analogy: The "Zoom" Problem

Think of it this way:

  • Humans look at a chart and see the meaning. We know that a bar's height represents a number.
  • CNNs look at the chart and see local details. They see the edges of the bar very clearly.
  • ViTs look at the chart and try to see everything at once. They are so busy looking at the relationship between the title, the legend, and the background that they sometimes lose track of the exact length of the bar. They are "distracted" by the big picture and miss the small, precise details needed for accurate measurement.

Why Does This Matter?

We are building more and more tools that use AI to read charts, summarize data, and even design new graphs for us. If we use a Vision Transformer to build a dashboard for doctors or financial analysts, and that AI "sees" the data differently than a human does, it could lead to dangerous misunderstandings.

For example, if an AI thinks a small difference in a graph is huge (because it's bad at judging length), it might tell a doctor that a patient's condition has worsened when it hasn't.

The Takeaway

The paper concludes that while Vision Transformers are powerful and exciting, they are not yet "human-like" when it comes to basic visual perception.

They are like a genius who can write a symphony but can't tell you if two sticks are the same length. To use them safely in data visualization, we need to either:

  1. Train them specifically to be better at these basic tasks.
  2. Be very careful about where we use them, knowing they might "hallucinate" the size or shape of things.

The researchers are calling for more work to bridge the gap between how these super-smart robots see the world and how we, as humans, see it. Until then, we should probably keep a human in the loop to double-check the math.

Get papers like this in your inbox

Personalized daily or weekly digests matching your interests. Gists or technical summaries, in your language.

Try Digest →