Can Vision-Language Models Answer Face to Face Questions in the Real-World?

This paper introduces the Qualcomm Interactive Video Dataset (IVD) to benchmark the ability of vision-language models to answer real-time questions about live video and audio, revealing that while current models significantly lag behind humans, targeted fine-tuning can substantially improve their performance on these interactive tasks.

Reza Pourreza, Rishit Dagli, Apratim Bhattacharyya, Sunny Panchal, Guillaume Berger, Roland Memisevic

Published 2026-02-24
📖 4 min read☕ Coffee break read

Imagine you have a robot friend named "VisionBot." You've trained VisionBot to look at a photo of a living room and tell you, "That's a red couch," or "There's a cat sleeping on the rug." VisionBot is great at this. It's like a very smart librarian who can instantly find facts in a static book.

But now, imagine you want VisionBot to be a real-time conversation partner. You are holding your phone, walking around your house, and you ask VisionBot, "Hey, am I holding this cup the right way?" or "How many times did I just clap?" while you are doing it.

This is where VisionBot starts to stumble. It's like asking a librarian to join you in a live dance party and keep up with the rhythm, but the librarian is used to reading books in a quiet library. They know the facts, but they don't know when to speak or how to react to things happening right now.

The Problem: The "Time Travel" Gap

The paper argues that current AI models are like time travelers who only know the future.

  • Old AI: You give it the entire video of a person clapping 10 times, and then you ask, "How many times did they clap?" The AI watches the whole thing, counts, and answers. It's like watching a movie and then taking a quiz.
  • Real Life: In the real world, you ask the question while the person is clapping. The AI has to listen to the question, watch the video unfold, realize, "Oh, they are still clapping, I need to wait," and then answer at the exact right moment.

Current AI models are terrible at this. They either answer too early (before the action is finished) or they get confused because they can't combine what they see (video) with what they hear (audio) in real-time.

The Solution: The "QIVD" Playground

To fix this, the researchers at Qualcomm created a new playground called QIVD (Qualcomm Interactive Video Dataset).

Think of QIVD as a gym for AI robots.

  • The Workout: They recorded 2,900 short videos of real people doing random things (clapping, pointing, holding objects) and asking questions like, "Is this my nose or my eye?" or "Did I just throw the ball?"
  • The Twist: The dataset includes a special "stopwatch" for every question. It tells the AI exactly when enough information has appeared to answer correctly.
    • Example: If someone asks, "How many times did I clap?" the stopwatch says, "Don't answer yet! Wait until the clapping stops."
  • The Goal: To train AI to learn not just what to say, but when to say it.

What They Discovered

The researchers put the smartest AI models (like GPT-4o and others) through this gym, and the results were surprising:

  1. The "Blind Spot": Even the smartest AIs failed miserably at "face-to-face" questions. They often answered before the action was done or missed the audio cues. It's like a student who knows the math formula but forgets to wait for the teacher to finish reading the question before shouting out the answer.
  2. The "Audio Blindness": Many models ignored the sound. If you asked, "Am I speaking loudly?" and the model only looked at the video, it couldn't answer. They are like people trying to have a conversation in a noisy room while wearing noise-canceling headphones.
  3. The "Magic of Fine-Tuning": Here is the good news. When they took these clumsy robots and gave them extra practice specifically on this "gym" data (fine-tuning), they got much better. They learned to wait, to listen, and to combine sight and sound. It's like giving a new driver a few hours of practice in a parking lot; suddenly, they aren't crashing into everything.

The Big Picture

This paper is a wake-up call. We are building AI that is amazing at describing static pictures, but we are far from having AI that can be a helpful, real-time companion in our daily lives.

The Analogy:

  • Current AI is like a photographer who takes a perfect picture and writes a caption later.
  • What we need is a cameraman who can walk alongside you, listen to your questions, watch what you're doing, and give you helpful advice while you are doing it.

The QIVD dataset is the first step toward teaching our AI cameramen how to stop, look, listen, and speak at the right time. It's the bridge between "smart computers" and "helpful robots."

Get papers like this in your inbox

Personalized daily or weekly digests matching your interests. Gists or technical summaries, in your language.

Try Digest →