OpenHEART: Opening Heterogeneous Articulated Objects with a Legged Manipulator

This paper presents OpenHEART, a robust and sample-efficient framework that enables legged manipulators to open diverse heterogeneous articulated objects by utilizing Sampling-based Abstracted Feature Extraction (SAFE) for compact geometric encoding and an Articulation Information Estimator (ArtIEst) for adaptive state estimation.

Seonghyeon Lim, Hyeonwoo Lee, Seunghyun Lee, I Made Aswin Nahrendra, Hyun Myung

Published 2026-03-09
📖 5 min read🧠 Deep dive

Imagine you have a robot dog with a robotic arm. It's a super-athlete, capable of walking over rough terrain and grabbing things. Now, imagine you ask this robot to open a bunch of different doors, drawers, and cabinets in a messy house.

Some doors swing on hinges (like a normal door), some drawers slide out (like a kitchen drawer), and some cabinets have weirdly shaped handles. The problem is that every single one of these objects is different. A robot trained to open a round doorknob might get confused by a long horizontal handle, or a sliding drawer might baffle a robot expecting a swinging door.

This paper, OpenHEART, introduces a new "brain" for this robot dog that lets it figure out how to open any of these weird, mixed-up objects without needing a specific manual for each one.

Here is how they did it, broken down into simple concepts:

1. The Problem: Too Much Noise, Not Enough Clarity

Most robots try to "see" the world using high-definition cameras or 3D scanners, creating a massive cloud of millions of dots (point clouds).

  • The Analogy: Imagine trying to learn how to drive a car by staring at a 4K video of the entire city, including every tree, cloud, and pedestrian. It's too much information! The robot gets overwhelmed and takes forever to learn the simple rule: "Turn the wheel to go left."
  • The Issue: Because legged robots (like the robot dog) are wobbly and complex, they need to learn fast. Looking at millions of dots is too slow and inefficient.

2. The Solution: The "Sketch" Method (SAFE)

The authors created a system called SAFE (Sampling-based Abstracted Feature Extraction). Instead of showing the robot the whole detailed object, they teach it to look at a simple "sketch."

  • The Analogy: Instead of giving the robot a photo of a complex door, they give it a simple box drawn around the handle and a box around the door panel. They ask the robot to ignore the wood grain, the color, and the scratches, and just focus on: "How long is the handle? Is the panel tall or wide?"
  • The Magic Trick: To make sure the robot doesn't just memorize the specific training doors, they randomly "shuffle" the dots inside these boxes. It's like giving the robot a puzzle where the pieces are always in a different order. This forces the robot to learn the shape of the object, not just the specific pattern of pixels. This makes the robot much better at handling objects it has never seen before.

3. The "Gut Feeling" Sensor (ArtIEst)

Sometimes, just looking at an object isn't enough. A cabinet might look like it opens left, right, or down, and the robot can't tell just by looking.

  • The Analogy: Imagine you are trying to open a jar. You look at it, but you aren't sure if the lid is stuck or if you're turning it the wrong way. So, you grab it and give it a little wiggle. Your hand (proprioception) tells you more than your eyes (exteroception) did.
  • How it Works: The robot has a smart system called ArtIEst.
    • Phase 1 (Eyes): Before touching the object, it guesses the direction based on the "sketch" (SAFE).
    • Phase 2 (Hands): Once it grabs the handle, it uses its own body sensors to feel the movement.
    • The Switch: A "Belief Gate" acts like a traffic cop. If the robot is just looking, it trusts its eyes. As soon as it grabs the handle and starts feeling resistance, it switches to trusting its hands. This helps it correct mistakes instantly.

4. The Result: A Versatile Robot

The team tested this on a real robot dog with an arm.

  • The Test: They threw it a mix of revolute doors (swinging) and prismatic drawers (sliding) with all sorts of handle shapes.
  • The Outcome: The robot didn't need a new program for each object. It used one single "brain" to figure out: "Ah, this handle is long and horizontal, so I need to pull straight out. That other one is a knob on the side, so I need to push and turn."
  • Real World Success: They even tested it on a real drawer in a real room. The robot missed the handle on the first try, but instead of giving up, it realized, "Oops, I missed," grabbed it again, and successfully opened it.

Summary

OpenHEART is like teaching a robot to be a handyman. Instead of memorizing the instructions for every single door in the world, the robot learns to look at the shape of the handle and the feeling of the movement. It ignores the clutter, focuses on what matters, and adapts on the fly, making it a true master of opening things in a chaotic, real-world environment.