RVN-Bench: A Benchmark for Reactive Visual Navigation

The paper introduces RVN-Bench, a new collision-aware benchmark built on Habitat 2.0 and HM3D scenes that enables the training and evaluation of safe, robust indoor visual navigation policies for mobile robots in unseen, cluttered environments.

Jaewon Lee, Jaeseok Heo, Gunmin Lee, Howoong Jun, Jeongwoo Oh, Songhwai Oh

Published 2026-03-05
📖 4 min read☕ Coffee break read

Imagine you are teaching a robot dog to navigate a messy living room. The room is full of coffee tables, scattered toys, and fragile vases. Your goal is simple: tell the robot, "Go to the kitchen," and it needs to get there without knocking anything over.

This is the problem the paper RVN-Bench is trying to solve.

Here is the breakdown of the paper using simple analogies:

1. The Problem: The "Blind" Robot

Currently, most robot navigation tests are like playing a video game where the walls are invisible.

  • The Old Way: Existing benchmarks (tests) ask robots, "Did you reach the kitchen?" If the robot crashed through a wall or smashed a vase but still ended up in the kitchen, the old tests would say, "Great job! 100% success!"
  • The Reality: In the real world, crashing is bad. Robots need to be "collision-aware." They need to see the vase and stop, not just drive through it.
  • The Gap: Most tests are designed for cars driving on open highways (outdoors) or robots that ignore obstacles. There wasn't a good "driving school" for indoor robots that actually penalizes them for bumping into furniture.

2. The Solution: RVN-Bench (The "Obstacle Course")

The authors created RVN-Bench, which is like a high-tech, virtual obstacle course specifically for indoor robots.

  • The Simulator: They built this inside a video game engine called Habitat, using 3D scans of real houses (HM3D). It looks and feels like a real home.
  • The Rules: The robot is given a series of goals (e.g., "Go to the sofa," then "Go to the door"). It can only use its camera (eyes) to see. It has no map in its head.
  • The Twist: If the robot bumps into a wall, it gets a big "penalty." The test measures two things:
    1. Did it get there? (Success)
    2. Did it break anything? (Safety)

3. The Secret Weapon: The "Crash Dataset"

One of the coolest parts of this paper is how they teach the robot to avoid crashing.

  • The Problem with Real Life: If you want to teach a robot what a crash feels like, you have to actually crash it. In the real world, this is expensive (broken robots) and dangerous.
  • The RVN-Bench Trick: They created a special tool that generates "Negative Trajectories."
    • Imagine a video game where you intentionally drive the car into a wall on purpose, over and over, to record exactly what the camera sees right before the crash.
    • They call this the Negative Dataset. It's a library of "crash videos" that the robot can study without actually breaking anything.
    • They also have an Expert Dataset (videos of perfect, crash-free navigation).
    • By showing the robot both the "perfect path" and the "crash path," the robot learns much faster what not to do.

4. The Experiments: Who Won the Race?

The authors tested several different "brains" (algorithms) in this new obstacle course:

  • The Imitation Learners (The Copycats): These robots tried to learn by watching the "Expert Dataset" (like a student copying a teacher's homework). They were okay, but they struggled when the room looked slightly different.
  • The Reinforcement Learners (The Trial-and-Errorers): These robots learned by trying things, getting punished for crashing, and rewarded for moving forward. They were much better.
  • The Depth-Enhanced Robot: The best performer was a robot that didn't just use its camera (RGB) but also used a "depth sensor" (like a 3D eye) to understand how far away objects were. This robot was the champion, reaching goals more often and crashing less.

5. The Real-World Test: Does it Work Outside the Game?

Finally, they took the robot trained in the virtual house and put it in a real house.

  • The Result: The robot trained in the simulation (the video game) actually did better than robots trained only on real-world data!
  • Why? The simulation gave the robot thousands of hours of practice, including thousands of "crash lessons" that would have been too expensive to film in real life.
  • The Hybrid Winner: The absolute best robot was one trained on both real-world data and simulation data. It combined the "feel" of the real world with the "volume" of the simulation.

The Big Takeaway

This paper introduces a new standard for testing robots. It says: "Don't just ask if the robot can find the goal; ask if it can find the goal without breaking the house."

By creating a safe, virtual place to crash and learn, they are helping us build robots that are ready to safely roam our messy, cluttered homes. It's like moving from teaching a driver in an empty parking lot to teaching them in a busy city with traffic, pedestrians, and potholes.