Safe Navigation of Bipedal Robots via Koopman Operator-Based Model Predictive Control

This paper proposes a safe navigation framework for bipedal robots that combines deep reinforcement learning with Koopman operator-based model predictive control to linearize complex nonlinear dynamics in a lifted space, thereby achieving more accurate trajectory prediction and improved success rates in dense, narrow environments.

Jeonghwan Kim, Yunhai Han, Harish Ravichandar, Sehoon Ha

Published 2026-03-10
📖 5 min read🧠 Deep dive

Imagine you are trying to teach a very clumsy, two-legged robot (like a human) how to walk through a crowded, narrow hallway without bumping into walls or falling over.

This is a tough job for two reasons:

  1. The Robot is Chaotic: Walking is incredibly complex. Every step involves balancing, swinging legs, and hitting the ground. It's like trying to predict the path of a leaf blowing in a storm; it's full of "non-linear" chaos where small changes lead to big, unpredictable results.
  2. The Brain is Slow: To keep the robot safe, its "brain" needs to look ahead and plan every step. But if the brain tries to simulate the complex, chaotic physics of walking in real-time, it gets overwhelmed and crashes (or the robot falls).

This paper proposes a clever solution that acts like a super-smart translator and a crystal ball.

The Problem: The "Black Box" vs. The "Math Nightmare"

The researchers tried two main approaches before:

  • The "Black Box" (Deep Learning): They trained the robot to walk using trial and error (like teaching a dog tricks). It worked great for walking, but the robot's internal logic was a "black box." You couldn't easily ask, "What happens if I turn left here?" because the math was too messy to solve quickly.
  • The "Math Nightmare" (Standard Physics Models): They tried to write down exact physics equations. But bipedal robots are so complex that the equations are too hard to solve fast enough for real-time navigation.

The Solution: The "Koopman Translator"

The authors used a mathematical tool called the Koopman Operator. Here is the best way to understand it:

The Analogy: The Flat Map vs. The 3D Rollercoaster
Imagine the robot's movement is a rollercoaster track. It twists, turns, and loops in 3D space. Trying to predict the path on this twisted track is hard.

  • Standard Models try to calculate the twists and turns directly. It's slow and prone to errors.
  • The Koopman Approach is like taking a photo of that 3D rollercoaster and projecting it onto a flat 2D map. On this flat map, the crazy loops look like simple, straight lines.

The researchers trained a "translator" (using AI) that takes the robot's chaotic, 3D walking data and projects it into this "flat map" (a high-dimensional space). In this new space, the complex, chaotic movement suddenly looks linear (straight and predictable).

How It Works (Step-by-Step)

  1. Teach the Legs (The Low-Level Policy): First, they used Deep Reinforcement Learning to teach the robot how to walk. Think of this as teaching the robot's legs the muscle memory to balance and step.
  2. Build the Crystal Ball (The Koopman Model): They watched the robot walk and used the "Koopman Translator" to learn how the robot's overall position changes. They found a way to describe the robot's future path using simple, straight-line math (Linear Dynamics) instead of complex curves.
  3. The Safe Navigator (MPC): Now, they gave the robot a "planner" (Model Predictive Control). Because the math is now simple and straight (thanks to the translator), the planner can look 6 seconds into the future in a split second. It can say, "If I turn left now, I will hit the wall. If I turn right, I will make it."

The "Phase" Secret Sauce

The researchers added a special ingredient called Phase Augmentation.

  • The Metaphor: Walking is rhythmic, like a song with a beat. If you only look at where the robot is, you miss the beat.
  • The Fix: They told the model to also pay attention to where the robot is in its walking cycle (is the left foot down? is the right foot swinging?). By adding this "rhythm" to the math, the crystal ball became incredibly accurate at predicting turns and curves.

The Results: Why It Matters

They tested this in a virtual world and on a real robot (the Unitree G1) in narrow corridors and mazes.

  • Accuracy: The Koopman model predicted where the robot would be 6 seconds later with 50% less error than previous methods. It was like having a GPS that doesn't just guess your location, but knows exactly where you'll be.
  • Safety: In a maze with tight corners, the old methods (linear models) kept crashing into walls because they couldn't predict the turn well enough. The Koopman robot navigated the maze with a 96% success rate.
  • Speed: Because the math was simplified into straight lines, the robot could plan its path instantly, even on real hardware.

The Bottom Line

This paper is about taking a chaotic, difficult problem (walking bipedally) and using a mathematical "lens" to make it look simple and predictable. By doing this, they gave the robot a superpower: the ability to look far into the future, plan safe paths through crowded rooms, and not fall over, all while moving at real-time speeds.

It's the difference between trying to navigate a stormy sea by guessing (old methods) versus having a perfect, real-time map that turns the storm into a calm, straight road (Koopman MPC).