From Perception to Action: An Interactive Benchmark for Vision Reasoning

This paper introduces CHAIN, an interactive 3D physics-driven benchmark that shifts Vision-Language Model evaluation from passive perception to active problem-solving, revealing that even top-performing models struggle to internalize physical constraints for reliable long-horizon planning and action execution.

Yuhao Wu, Maojia Song, Yihuai Lan, Lei Wang, Zhiqiang Hu, Yao Xiao, Heng Zhou, Weihua Zheng, Dylan Raharja, Soujanya Poria, Roy Ka-Wei Lee

Published 2026-02-25
📖 4 min read☕ Coffee break read

Imagine you are teaching a robot how to build a complex LEGO castle or solve a tricky wooden puzzle.

For a long time, we tested these robots by showing them a single, static photo of the finished castle and asking, "What is this?" or "How many blocks are there?" The robot would look at the picture, guess the answer, and get a score. This is like asking a chef to identify a cake just by looking at a photo of it, without ever letting them touch the ingredients or taste the batter.

The Problem:
Real life isn't a photo. Real life is messy, physical, and full of rules. If you try to put a big block on top of a tiny one, it falls. If you try to pull a piece out of a puzzle the wrong way, it gets stuck. The old tests didn't check if the robot understood these physical rules. They only checked if the robot could recognize things.

The Solution: CHAIN
The authors of this paper built a new test called CHAIN (Causal Hierarchy of Actions and Interactions). Think of CHAIN as a digital playground or a video game simulator where robots have to actually do things, not just talk about them.

Instead of a photo, the robot is dropped into a 3D world with physics (gravity, collisions, friction). It has to:

  1. Look at the scene.
  2. Think about what it can physically do next (e.g., "If I move this block, that one will fall").
  3. Act (move, rotate, stack).
  4. Watch what happens and adjust its plan if things go wrong.

The Two Main Games:
The benchmark uses two types of challenges to test the robots:

  1. The Interlocking Puzzles (The "Lockpick" Test):
    Imagine a complex wooden puzzle where pieces are locked together like a 3D jigsaw. To take it apart, you can't just pull randomly. You have to find the one specific piece that unlocks the rest, then move it in a specific direction, which unlocks the next piece.

    • The Challenge: It requires understanding hidden rules. If you pull the wrong piece, the whole thing jams.
    • The Result: Even the smartest AI models (like GPT-5 or Claude) got stuck almost immediately. They couldn't figure out the "secret handshake" needed to unlock the puzzle.
  2. The Stacking Game (The "Tetris" Test):
    Imagine you have a box and a pile of oddly shaped blocks. You have to pack them all in without them falling over or leaving gaps.

    • The Challenge: You can't just throw blocks in. You have to plan ahead. If you put a big block in the corner now, you might not have space for the last piece later.
    • The Result: The AIs were okay at easy levels but failed miserably at hard levels. They would greedily grab the "easy" blocks first, only to realize 10 steps later that they had trapped themselves with no way to finish the job.

The Big Surprise:
The researchers also tested "World Models" (AI that tries to predict what a video of an action would look like). They asked these AIs to generate a video of someone taking apart a puzzle.

  • The Disaster: The AIs hallucinated wildly. They made blocks pass through each other like ghosts, merged two blocks into one, or made pieces disappear. They looked cool visually, but they were physically impossible. It was like watching a movie where the laws of physics were broken.

The Takeaway:
The paper concludes that current AI is great at seeing (recognizing a cat in a photo) but terrible at doing (understanding that if you push the cat, it will move).

  • Old AI: "I see a puzzle. I know what a puzzle is."
  • Real-World AI (What we need): "I see a puzzle. I know that if I pull piece A, piece B will slide left, but only if I rotate piece C first. If I don't, the whole thing breaks."

Why This Matters:
If we want robots to help us in the real world—fixing cars, building houses, or helping in hospitals—they need to pass the CHAIN test. They need to understand that the world has rules, and that every action changes the future. Right now, our robots are like toddlers who can name objects but don't understand how the world works yet. This benchmark is the first step to teaching them the difference.

Get papers like this in your inbox

Personalized daily or weekly digests matching your interests. Gists or technical summaries, in your language.

Try Digest →