PeRoI: A Pedestrian-Robot Interaction Dataset for Learning Avoidance, Neutrality, and Attraction Behaviors in Social Navigation

This paper introduces the PeRoI dataset, which captures diverse pedestrian reactions to robots in various contexts, and proposes the NeuRoSFM model to leverage this data for improved prediction of pedestrian-robot interactions in socially aware navigation.

Subham Agrawal, Nico Ostermann-Myrau, Nils Dengler, Maren Bennewitz

Published 2026-03-06
📖 4 min read☕ Coffee break read

Imagine you are walking down a busy sidewalk. Suddenly, a robot rolls into your path. What do you do?

Most people assume everyone would just step aside to avoid a collision. But in reality, human reactions are much more colorful. Some people might dodge the robot like it's a hot potato (Avoidance). Others might walk right past it without even glancing, as if it's invisible (Neutrality). And then, there are the curious ones who might actually slow down or turn toward the robot to get a better look (Attraction).

For a long time, robots trying to navigate our world have been taught to expect only the first reaction: "Everyone will run away." This paper, titled PeRoI, argues that this assumption is wrong and limits how well robots can interact with us.

Here is the breakdown of their solution, explained simply:

1. The Problem: Robots are "Blind" to Human Nuance

Think of current robot navigation systems like a driver who only knows how to drive in a straight line. They have maps of where people usually walk, but they don't understand why people move the way they do when a robot is around.

Existing data sets are like old photo albums: they show people walking, but they rarely show what happens when a robot enters the room. If a robot is trained only on data where people always avoid it, it will be confused and clumsy when someone actually stops to say "Hello" or just walks past it casually.

2. The Solution: The "PeRoI" Dataset (The Robot's Diary)

The researchers created a new database called PeRoI (Pedestrian-Robot Interaction). Imagine they set up a camera in two busy outdoor spots (like a university campus) and watched thousands of people walk by.

They introduced three different scenarios:

  • The Ghost: No robot is there (just people walking).
  • The Statue: A robot stands still in the middle of the path.
  • The Walker: A robot moves along a specific path.

They used three different types of robots to see if the look of the robot mattered:

  • A wheeled robot that looks like a friendly office assistant.
  • A four-legged robot that looks like a dog.
  • A boxy, industrial robot that looks like a delivery truck.

The Big Discovery: They found that people react differently based on the robot's shape and movement.

  • The "dog" robot (Unitree Go1) made people curious (Attraction).
  • The "truck" robot (MPO700) made people nervous and keep their distance (Avoidance).
  • Many people just walked by without caring (Neutrality).

They labeled every single person's reaction in the database, creating a "dictionary" of how humans actually behave, not just how we think they behave.

3. The New Brain: NeuRoSFM (The Robot's New Instincts)

Having the data is great, but the robot needs a way to use it. The authors built a new model called NeuRoSFM.

Think of the old way of programming robots as a recipe book. The recipe says: "If a person is 2 meters away, push them away with Force X." It's rigid and requires a human expert to tweak the numbers constantly.

The new NeuRoSFM is more like a muscle memory learned from experience.

  • Instead of hard-coded rules, the robot uses a "neural network" (a type of AI brain) to learn the forces.
  • It learns that a "dog robot" might pull people in, while a "truck robot" pushes them away.
  • It also learns about groups. If you are walking with friends, you might not move away from the robot even if you are scared, because you are sticking with your group. The old models ignored this; the new one accounts for it.

4. The Result: Smoother Dancing

The researchers tested this new "brain" on real-world data.

  • Old Model: Predicted that everyone would run away from the robot. It was wrong about 30% of the time.
  • New Model (NeuRoSFM): Predicted that some would run, some would ignore, and some would get curious. It was much more accurate.

The Takeaway

This paper is like teaching a robot to understand social etiquette.

Before, robots were like toddlers in a crowd: they knew they had to move, but they didn't understand that sometimes people want to say hi, sometimes they want to ignore you, and sometimes they just want to keep walking.

By collecting the PeRoI dataset and building the NeuRoSFM model, the authors have given robots the ability to "read the room." This means future robots won't just be safe; they will be polite, predictable, and comfortable to be around in our shopping malls, hospitals, and sidewalks.