DexGrasp-Zero: A Morphology-Aligned Policy for Zero-Shot Cross-Embodiment Dexterous Grasping

The paper introduces DexGrasp-Zero, a novel policy utilizing a morphology-aligned graph representation and a Morphology-Aligned Graph Convolutional Network (MAGCN) with physical property injection to achieve robust zero-shot cross-embodiment dexterous grasping across diverse hardware without re-learning.

Yuliang Wu, Yanhan Lin, WengKit Lao, Yuhao Lin, Yi-Lin Wei, Wei-Shi Zheng, Ancong Wu

Published 2026-03-18
📖 5 min read🧠 Deep dive

Imagine you have a master chef who is an expert at cooking with a specific set of kitchen tools: a French knife, a wooden spoon, and a cast-iron skillet. If you suddenly hand them a pair of chopsticks, a ladle, and a wok, they might freeze. They know how to cook, but they don't know how to translate their muscle memory to these new, weirdly shaped tools.

This is the exact problem robots face today. We have built many different "robot hands" (dexterous hands), but they all look and move differently. If we train a robot brain to pick up a cup with a "Shadow Hand," that brain usually fails completely when we swap it for a "LEAP Hand" or a "Schunk Hand." It's like trying to drive a car by looking at the steering wheel of a boat.

DexGrasp-Zero is a new method that teaches a robot hand to be a "universal chef." It allows a single robot brain to learn how to grasp objects and then instantly use that skill on any new hand it sees, without needing to re-learn or practice.

Here is how they did it, using some simple analogies:

1. The Problem: The "Middleman" Trap

Previous methods tried to solve this by using a "middleman."

  • The Old Way: The robot brain would say, "I want my fingers to move to this specific spot in space." Then, a separate translator program would try to figure out how to move the specific robot hand to get there.
  • The Flaw: This is like telling a translator, "I want to go to the top of the mountain," without telling them which mountain you are standing on. The translator might try to walk up a cliff that doesn't exist for that specific hand, causing the robot to crash or try to bend a finger in a way that breaks it.

2. The Solution: Speaking a "Universal Language"

The authors realized that even though robot hands look different, they all share the same anatomy. They all have a wrist, a palm, a thumb, and fingers with joints that bend, spread, and twist.

Instead of teaching the robot to move "fingers," they taught it to speak a Universal Motion Language based on three simple movements:

  1. Flex: Bending the finger inward (like making a fist).
  2. Abduct: Spreading the finger out (like spreading your fingers wide).
  3. Rotate: Twisting the finger.

Think of this like LEGO bricks. Whether you have a small LEGO set or a giant one, the basic bricks (Flex, Abduct, Rotate) are the same. The robot brain learns to build a "grasp" using these universal bricks, rather than trying to memorize the specific instructions for every single hand model.

3. The Secret Sauce: The "Physical Cheat Sheet"

Just knowing the universal language isn't enough. A tiny robot hand can't lift a heavy rock, and a giant hand might crush a grape.

The researchers gave the robot brain a "Physical Cheat Sheet" (called a Morphology-Aligned Graph).

  • Before the robot tries to grab something, it looks at the "cheat sheet" for the specific hand it is currently using.
  • This sheet tells the brain: "Hey, this hand has short fingers, so don't try to wrap around that big ball," or "This hand has strong motors, so you can squeeze harder."
  • It's like a GPS that knows exactly what kind of car you are driving. If you are in a tiny Mini Cooper, the GPS won't tell you to take a route with a low bridge; if you are in a truck, it won't tell you to take a narrow alley.

4. The Result: Zero-Shot Transfer

Because the robot brain learns the concept of grasping (using the universal bricks) and checks the physical limits (using the cheat sheet) for every new hand, it works instantly.

  • Training: They trained the brain on four different types of robot hands.
  • The Test: They then gave it two completely new hands it had never seen before.
  • The Outcome: The robot didn't need to practice. It just looked at the new hand, checked the cheat sheet, and successfully grabbed objects.
    • Success Rate: It worked 85% of the time in simulation and 82% of the time in the real world.

Why This Matters

Imagine a future where a factory has 100 different robots, and a new one is added tomorrow. With old methods, you'd have to spend weeks teaching the new robot how to pick up a screw. With DexGrasp-Zero, you just plug the new robot in, and it already knows how to do the job because it understands the language of hands, not just the specific model.

It turns the robot from a "specialist" who only knows one tool into a "generalist" who can adapt to any tool in the shed.

Drowning in papers in your field?

Get daily digests of the most novel papers matching your research keywords — with technical summaries, in your language.

Try Digest →