Dance2Hesitate: A Multi-Modal Dataset of Dancer-Taught Hesitancy for Understandable Robot Motion

This paper introduces "Dance2Hesitate," an open-source multi-modal dataset comprising synchronized kinesthetic robot teaching and dancer motion capture data across three hesitancy levels, designed to facilitate the study and benchmarking of understandable, context-aware hesitancy in human-robot collaboration.

Srikrishna Bangalore Raghu, Anna Soukhovei, Divya Sai Sindhuja Vankineni, Alexandra Bacula, Alessandro Roncone

Published Thu, 12 Ma
📖 4 min read☕ Coffee break read

Imagine you are playing a game of Jenga with a robot. You reach out to pull a block, and the robot reaches out too. If the robot moves smoothly and confidently, you might feel safe. But what if the robot stops, wobbles, or moves very slowly? You might think, "Wait, is it unsure? Is it about to drop the block? Should I grab it before it falls?"

That "wobble" or "pause" is called hesitancy. In the world of human-robot teamwork, this hesitation is actually a superpower. It's a non-verbal way for the robot to say, "I'm not 100% sure I can do this safely," which helps humans react quickly to stay safe.

However, teaching a robot to hesitate just right is incredibly hard. It's like trying to teach a fish to dance; a fish doesn't have legs, and a robot arm doesn't have a whole body to express feelings. If a robot arm hesitates, it might look like a glitch. If a human hesitates, it looks like caution.

The Big Idea: "Dance2Hesitate"

To solve this, the researchers created a new dataset called Dance2Hesitate. Think of this dataset as a giant library of "hesitation recipes."

Instead of trying to program a robot to hesitate using math equations (which often looks robotic and weird), they asked professional dancers to teach them how to do it.

Here is how they did it, using some simple analogies:

1. The "Human Translator" (The Dancers)

Dancers are experts at using their bodies to tell stories without saying a word. They know exactly how to make a movement look "uncertain" or "cautious" without actually stopping.

  • The Experiment: The researchers set up a specific scene: A robot arm (or a human arm) needs to reach for a Jenga tower.
  • The Levels: They asked the dancers to perform this reach in three different "flavors" of hesitation:
    • Slight: Like a gentle "hmm, let me think."
    • Significant: Like a clear "whoa, I'm not sure about this."
    • Extreme: Like a dramatic "STOP! This is dangerous!"
  • The Magic: Because the dancers are so skilled, they could repeat these movements perfectly, giving the researchers clean, high-quality data to study.

2. The "Two-Way Mirror" (The Data)

The researchers recorded this in two ways to create a "Rosetta Stone" for robots:

  • The Robot Side: The dancers physically guided the robot arm (like holding a friend's hand to show them how to move) to reach the Jenga tower. This taught the robot exactly how a human wants it to move.
  • The Human Side: They filmed the dancers with special 3D cameras as they performed the same moves with their own arms and bodies.

This creates a bridge. Now, if a robot wants to hesitate, it can look at the dancer's movement, copy the "vibe," and translate it into its own mechanical language.

3. Why This Matters

Imagine a self-driving car approaching a pedestrian.

  • Without Hesitancy: The car stops abruptly. The pedestrian thinks, "Did the car break? Is it going to hit me?"
  • With Hesitancy: The car slows down, wobbles slightly, and pauses. The pedestrian understands, "Ah, the car sees me and is being careful. I can cross."

This dataset helps engineers build robots that don't just do tasks, but communicate their internal state. It turns a cold, mechanical machine into a partner that you can "read" like a human.

The Takeaway

The paper is essentially saying: "We asked dancers to teach robots how to be unsure, so that robots can stop looking like glitchy machines and start looking like thoughtful teammates."

They have now made all this data (videos, robot movement logs, and 3D motion files) free for anyone to use. This means researchers everywhere can now build robots that hesitate in a way that humans naturally understand, making our future interactions with robots safer and more intuitive.