fMRI-Based Prediction of Eye Gaze During Naturalistic Movie Viewing Reveals Eye-Movement-Related Brain Activity

This study demonstrates that while a zero-shot DeepMReye model has limited accuracy for individual-level gaze prediction from fMRI data, group-averaged estimates effectively capture shared viewing behaviors and successfully reveal brain activation patterns associated with oculomotor control during naturalistic movie viewing.

Gao, L., Wei, Z., Biswal, B. B., Di, X.

Published 2026-04-12
📖 5 min read🧠 Deep dive
⚕️

This is an AI-generated explanation of a preprint that has not been peer-reviewed. It is not medical advice. Do not make health decisions based on this content. Read full disclaimer

Imagine you are trying to understand how people's brains work while they watch a movie. Usually, scientists put a camera on the person's face to track exactly where their eyes are looking. This is like having a spotlight that shows you exactly what the person is paying attention to.

But here's the problem: Most old movie-watching brain scans don't have these eye cameras. They were taken years ago, or in hospitals where the cameras didn't fit. It's like having a recording of a concert but missing the audio of the singer's voice—you can see the crowd, but you don't know who they were looking at.

This paper asks a big question: Can we use Artificial Intelligence (AI) to "guess" where people were looking, just by looking at their brain scan data?

Here is the simple breakdown of what the researchers did and what they found, using some everyday analogies.

1. The Magic Trick: Reading Minds from Brain Scans

The researchers used a pre-trained AI model called DeepMReye. Think of this AI as a "mind-reading detective" that has been trained on a huge library of brain scans and eye-tracking data.

  • The Job: The AI looks at the tiny signals coming from the eyeball area inside the brain scan. Even though the brain scan is blurry and slow (like a low-frame-rate video), the movement of the eye creates tiny ripples in the signal.
  • The Challenge: The researchers wanted to see if this AI could work on new movies and new people without being re-trained first. This is called a "zero-shot" approach. It's like hiring a translator who speaks 10 languages and asking them to translate a new language they've never heard of, just by guessing based on context.

2. The Results: The "Crowd" vs. The "Individual"

The results were a mix of "Wow!" and "Not so fast."

The Group Level: The "Chorus" Effect

When the researchers averaged the eye-gaze predictions of everyone in the room together, the AI was amazing.

  • The Analogy: Imagine a choir singing. If you listen to just one singer, they might be slightly off-key or out of rhythm. But if you listen to the whole choir, the sound is perfect and powerful.
  • The Finding: When they looked at the average gaze of the group, the AI's guess matched the real eye-tracking data almost perfectly (about 80% accuracy). It successfully predicted that when a character in the movie moved left, the whole group looked left.

The Individual Level: The "Soloist" Problem

When they tried to guess the eye movements of just one person, the AI struggled.

  • The Analogy: Trying to hear one specific singer in a noisy room is hard. Everyone's eyes are shaped slightly differently, they move their heads differently, and the brain scanner creates different amounts of "static" for each person.
  • The Finding: For a single person, the AI was only about 25-35% accurate. It was too noisy to tell exactly where one specific person was looking at any given second.

3. What Did They Learn About the Brain?

Even though the AI wasn't perfect for individuals, the "group average" was good enough to map the brain.

  • The Map: They used the AI's "group guess" of where people were looking to see which parts of the brain lit up.
  • The Discovery: They found the brain's "Eye Control Center." This included the Frontal Eye Fields (the boss that tells the eyes where to go) and the Visual Cortex (the screen where the movie plays).
  • The Takeaway: Even without a real eye camera, the AI could reconstruct the brain's "eye-movement map" well enough to show us how our brains control where we look.

4. The Age Factor: Growing Up Changes How We Look

The researchers also looked at how age changes things, comparing children to adults.

  • The Finding: As people get older, their eye movements become more synchronized with the group.
  • The Analogy: Think of a toddler watching a cartoon. They might look at the dog, then the cat, then the ceiling, then the dog again—very randomly. An adult, however, tends to look at the main character, just like everyone else.
  • The Twist: This "growing up" of eye habits wasn't a straight line. It was like a hill: kids' brains get better at following the action as they enter their teenage years, but then the pattern shifts again as they become young adults. It's a complex journey, not a simple straight line.

The Bottom Line

Can we use AI to guess where people looked in old brain scans?

  • For a single person? Not really yet. It's too fuzzy. You still need a real camera for that.
  • For a group of people? Yes! It works surprisingly well.

Why does this matter?
There are thousands of old brain scans sitting in databases that we can't fully use because we don't know where the people were looking. This study shows that we can use AI to "fill in the blanks" for the group. It allows scientists to study how our brains handle attention and eye movements in movies, even for studies that were done years ago without eye-tracking cameras.

It's like finding a way to hear the melody of a song even if the original recording was missing the vocals, as long as you listen to the whole orchestra together.

Get papers like this in your inbox

Personalized daily or weekly digests matching your interests. Gists or technical summaries, in your language.

Try Digest →