DriverGaze360: OmniDirectional Driver Attention with Object-Level Guidance

This paper introduces DriverGaze360, a large-scale 360-degree driver attention dataset and a corresponding panoramic prediction network (DriverGaze360-Net) that leverages object-level guidance to overcome the limitations of existing frontal-view methods and achieve state-of-the-art performance in modeling omnidirectional driver gaze behavior.

Shreedhar Govil, Didier Stricker, Jason Rambach

Published 2026-03-06
📖 4 min read☕ Coffee break read

Imagine you are teaching a robot to drive a car. To do this safely, the robot needs to know not just where the car is going, but what the human driver is looking at. If the human glances at a child running near the curb, the robot should know that's important. If the human checks the rearview mirror before changing lanes, the robot needs to catch that too.

For a long time, scientists trying to teach robots this skill had a major blind spot: they were only looking through a narrow window.

Here is a simple breakdown of the new paper, "DriverGaze360," which changes the game.

1. The Problem: The "Tunnel Vision" Trap

Imagine trying to learn how to drive a car while wearing a blindfold that only lets you see a tiny rectangle directly in front of your nose. You wouldn't see the car merging from the left, the pedestrian stepping off the curb behind you, or the cyclist in your blind spot.

That's what previous research was like. They used cameras that only looked straight ahead. They missed the most critical moments of driving:

  • Checking the side mirrors before turning.
  • Looking back to merge onto a highway.
  • Watching a cyclist approach from the rear.

Because they couldn't see the whole picture, their AI models were "tunnel-visioned" and couldn't predict human behavior accurately in complex situations.

2. The Solution: A 360-Degree "Fishbowl"

The researchers built something new called DriverGaze360. Think of it as putting the driver inside a giant, transparent fishbowl where they can see everything around them—front, back, left, right, and up.

  • The Setup: They put 19 real human drivers in a high-tech driving simulator (like a video game, but it feels real).
  • The Gear: The drivers wore special glasses that tracked exactly where their eyes moved, 120 times per second.
  • The View: Instead of one camera, they used five cameras to stitch together a full 360-degree view.
  • The Data: They collected about 1 million snapshots of where drivers looked. This includes boring highway driving, scary near-miss accidents, and tricky city turns.

The Result: For the first time, we have a massive library of data showing exactly how humans look at the entire world while driving, not just the road ahead.

3. The Brain: The "Detective" AI

Just having the data isn't enough; you need a smart brain to understand it. The researchers created a new AI model called DriverGaze360-Net.

Here is the clever trick they used:

  • Old Way: The AI tried to guess where the driver was looking, but it was like guessing where a detective is looking in a dark room without any clues.
  • New Way: They gave the AI a second job. While it guesses where the driver is looking, it also has to identify what objects the driver is looking at (like a red car, a stop sign, or a pedestrian).

The Analogy: Imagine you are teaching a child to find a hidden toy.

  • Old Method: You just say, "Look here!" (The child guesses randomly).
  • New Method: You say, "Look at the red ball!" (The child knows exactly what to focus on).

By forcing the AI to identify the specific objects (the "red balls"), it becomes much better at predicting where the human's eyes will go. It learns that "drivers look at cars, not at the sky," or "drivers look at pedestrians, not at the clouds."

4. Why This Matters

This isn't just about better video games. This is about safer self-driving cars.

  • Explainable AI: When a self-driving car makes a decision, we want to know why. If the car brakes suddenly, we want it to say, "I stopped because I saw the human driver looking at a child crossing the street." This new system helps the car understand that logic.
  • Safety: It helps cars anticipate human mistakes. If the car knows the human is distracted or looking the wrong way, it can step in to prevent an accident.
  • Realism: Because this data covers the whole view (including the rear and sides), the AI won't get surprised when a car pulls out from a blind spot.

In a Nutshell

The researchers realized that to teach a robot to drive like a human, they had to stop looking through a keyhole and start looking through a fisheye lens. They built a massive dataset of 360-degree eye-tracking and created a smart AI that learns by identifying the specific things humans care about. This makes future autonomous vehicles more aware, safer, and easier to trust.