Spatial-DISE: A Unified Benchmark for Evaluating Spatial Reasoning in Vision-Language Models

This paper introduces Spatial-DISE, a unified benchmark and large-scale dataset grounded in a four-quadrant taxonomy of spatial reasoning, which reveals significant gaps between current Vision-Language Models and human competence while providing a scalable framework for advancing human-like spatial intelligence.

Xinmiao Huang, Qisong He, Zhenglin Huang, Boxuan Wang, Zhuoyun Li, Guangliang Cheng, Yi Dong, Xiaowei Huang

Published 2026-02-25
📖 4 min read☕ Coffee break read

Imagine you are trying to teach a robot how to navigate a messy living room, fold a map, or assemble IKEA furniture without instructions. You'd need it to not just see the objects, but to understand how they fit together, how they move, and how they change shape.

This paper, Spatial-DISE, is a report card for the current generation of "Vision-Language Models" (AI that can see and talk). The authors from the University of Liverpool are saying: "These AI models are great at describing what they see, but they are terrible at figuring out how things work in 3D space."

Here is the breakdown using simple analogies:

1. The Problem: The "Flat-Earth" AI

Current AI models are like tourists who have only ever looked at postcards. They can tell you, "That's a red car next to a blue tree." But if you ask them, "If I fold this paper in half and poke a hole, what will the pattern look like when I unfold it?" they get confused.

Existing tests for these AIs were too easy. They mostly asked static questions like, "Is the cup to the left of the mug?" The paper argues this is like testing a pilot only on a calm, flat runway and never asking them to handle turbulence or a storm. We need to test their ability to mentally rotate, fold, and combine objects.

2. The Solution: The "Spatial-DISE" Gym

The authors built a new, tougher gym called Spatial-DISE. They created a massive set of puzzles (over 12,000 of them!) to test four specific types of "spatial muscles":

  • Intrinsic-Static (The Puzzle Piece): Looking at a single object and understanding its fixed parts. Example: "Which face is on the back of this cube?"
  • Intrinsic-Dynamic (The Mental Gymnast): Taking a single object and mentally twisting or folding it. Example: "If I rotate this block 90 degrees, what does it look like?"
  • Extrinsic-Static (The Photographer): Understanding how objects sit next to each other in a fixed scene. Example: "Is the cat under the table?"
  • Extrinsic-Dynamic (The Choreographer): Understanding how multiple objects move and interact. Example: "If I stack these blocks in this order, will the tower fall?"

The Cool Part: They didn't just hire humans to draw these puzzles. They built a robot factory (using a 3D software called Blender) that automatically generates thousands of unique, mathematically perfect puzzles. This ensures the answers are 100% correct and the AI can't cheat by memorizing previous answers.

3. The Results: The AI vs. Human Showdown

They tested 32 of the smartest AI models in the world (including GPT-4o, Gemini, and Claude) against this new gym.

  • The Score: The average AI scored about 28%. Since there are usually 4 options, random guessing would get you 25%. So, the AIs are barely doing better than a monkey throwing darts.
  • The Human Score: Humans scored around 77%.
  • The Gap: There is a massive canyon between human intelligence and current AI when it comes to "spatial reasoning."

The Surprise: Interestingly, some AIs got better at the hardest, most dynamic puzzles (like combining 3D shapes) than the static ones. It turns out, some AIs are like calculators: they are bad at "feeling" the shape, but if you force them to calculate the edges and angles mathematically, they can sometimes outperform humans who rely on intuition.

4. Why Do They Fail? (The Autopsy)

The authors looked at where the AI went wrong. It wasn't because the AI couldn't "see" the picture (it wasn't a vision problem). The problem was thinking.

  • Rule Breakers: The AI didn't know the basic laws of geometry. For example, it thought two opposite sides of a cube could be touching.
  • Bad Memory: In "Fold and Punch" tasks, the AI lost track of how many layers of paper existed after folding. It's like trying to juggle while forgetting how many balls you started with.
  • Superficial Glances: The AI would look at a shape and say, "It looks similar," without checking the tiny details that actually mattered.

5. The Takeaway

This paper is a wake-up call. If we want AI to drive cars, build robots, or help in surgery, they need to stop just "looking" and start "thinking" in 3D.

The authors have released their "gym" (the dataset and code) to the public. They hope that by training AI on these 12,000+ tough puzzles, we can finally teach machines to understand the physical world the way humans do: not just as a flat image, but as a dynamic, manipulatable space.

In short: We taught AI to read the menu, but we haven't taught it how to cook the meal yet. Spatial-DISE is the new cooking class.

Get papers like this in your inbox

Personalized daily or weekly digests matching your interests. Gists or technical summaries, in your language.

Try Digest →