EgoCross: Benchmarking Multimodal Large Language Models for Cross-Domain Egocentric Video Question Answering

This paper introduces EgoCross, a comprehensive benchmark comprising 1,000 QA pairs across four challenging domains (surgery, industry, extreme sports, and animal perspective) to evaluate and expose the poor cross-domain generalization capabilities of current Multimodal Large Language Models in egocentric video question answering.

Yanjun Li, Yuqian Fu, Tianwen Qian, Qi'ao Xu, Silong Dai, Danda Pani Paudel, Luc Van Gool, Xiaoling Wang

Published Wed, 11 Ma
📖 5 min read🧠 Deep dive

Imagine you've trained a brilliant student to be the ultimate "House Helper." This student has watched thousands of hours of videos about cooking, cleaning, and gardening. They are a wizard at answering questions like, "What ingredient is the chef adding?" or "When does the soup start boiling?"

Now, imagine you hire this same student to work in a surgical theater, a factory floor, a ski jump, or even strapped to the back of a cat.

Suddenly, the student freezes. They can't tell a scalpel from a screwdriver. They get dizzy trying to predict where a skier will land next. They have no idea what a cat sees when it chases a laser pointer.

This is exactly the problem the paper EgoCross is solving.

The Problem: The "Daily Life" Bubble

For a long time, AI researchers have been testing video-understanding AI (called Multimodal Large Language Models, or MLLMs) only on "daily life" videos. It's like only testing a car on a smooth, flat parking lot. The AI gets an A+ because it's never seen a pothole, a mountain road, or a muddy trail.

But in the real world, AI needs to work everywhere. If a robot is going to help a surgeon, it can't just know how to chop vegetables; it needs to know how to hold a surgical tool. If a drone is going to help a firefighter, it needs to understand smoke and fire, not just a sunny picnic.

The current AI models are like that "House Helper" student: they are great at home, but they panic when they leave the house.

The Solution: EgoCross (The "Field Trip" Exam)

The authors created a new test called EgoCross. Think of it as a "Field Trip Exam" designed to see if these AI students can handle totally different environments.

Instead of just cooking videos, EgoCross tests the AI in four very different, high-stakes worlds:

  1. Surgery: Watching a surgeon's hands from their own eyes. The tools are tiny, the movements are precise, and the vocabulary is medical.
  2. Industry: Watching someone repair complex circuit boards. It's about logic, tools, and assembly lines.
  3. Extreme Sports: First-person views of skiing or skydiving. The camera is shaking, moving fast, and the scenery is wild.
  4. Animal Perspective: Videos from the point of view of a dog, cat, or turtle. The world looks different (lower to the ground, different motion), and the "human" logic doesn't apply.

The Test: "What Happens Next?"

The exam isn't just "What do you see?" It's much harder. The AI has to answer questions like:

  • Prediction: "The surgeon just cut this artery; what tool will they use next?"
  • Localization: "Where exactly was the screwdriver when the hand touched it?"
  • Counting: "How many different tools did the mechanic use in this 10-second clip?"
  • Recognition: "Is that a 'grasper' or a 'scissors'?" (Even though they look similar).

The test includes about 1,000 questions, split into two types:

  • Multiple Choice (CloseQA): Like a standard quiz.
  • Free Response (OpenQA): Like a short essay where the AI has to explain its reasoning.

The Results: The AI Got Stuck

The authors ran the test on the smartest AI models available today (including the big ones from Google and OpenAI).

The verdict? Most of them failed miserably.

  • On the "daily life" tests, they scored high.
  • On the EgoCross test, their scores dropped dramatically. They were barely better than guessing randomly.

It turns out, these AIs are "specialists" in daily life, not "generalists." They haven't learned the deep rules of how the world works; they just memorized patterns of cooking and cleaning. When the visual style and the logic change (like switching from a kitchen to an operating room), they get lost.

The Silver Lining: Trying to Fix It

The paper didn't just point out the problem; they tried to fix it. They tested three methods to help the AI "study" for the new subjects:

  1. Prompting: Giving the AI a "cheat sheet" of context (e.g., "Remember, you are in a hospital"). Result: A little help, but not enough.
  2. Fine-Tuning: Letting the AI study a small amount of the new videos. Result: Better, but it only learned the specific examples it saw.
  3. Reinforcement Learning: Letting the AI try, fail, get feedback, and try again (like training a dog with treats). Result: This was the winner! It helped the AI adapt much better to the new worlds.

The Big Takeaway

EgoCross is a wake-up call. It tells us that while AI is getting very good at understanding our daily lives, it is still very fragile when it steps out of its comfort zone.

If we want AI to truly help us in hospitals, factories, or rescue missions, we can't just train it on YouTube cooking videos. We need to build models that are robust enough to handle the messy, weird, and specialized parts of the real world. This paper provides the map and the exam to help us get there.