Benchmarking Deflection and Hallucination in Large Vision-Language Models

This paper introduces VLM-DeflectionBench, a dynamic benchmark and evaluation protocol designed to assess Large Vision-Language Models' ability to handle conflicting or insufficient multimodal evidence by generating appropriate deflections rather than hallucinating answers.

Nicholas Moratelli, Christopher Davis, Leonardo F. R. Ribeiro, Bill Byrne, Gonzalo Iglesias

Published 2026-04-15
📖 5 min read🧠 Deep dive

Imagine you have a brilliant, super-smart robot assistant named "Visionary." Visionary can look at a picture, read a book, and answer almost any question you ask. But there's a catch: sometimes, Visionary gets too confident and makes things up when it doesn't actually know the answer. This is called hallucination.

On the other hand, sometimes Visionary is so scared of making a mistake that it refuses to answer even when it does have the right information. This is called deflection (or "abstaining").

The paper you shared introduces a new test called VLM-DeflectionBench. Think of it as a "stress test" or a "driving exam" for these AI robots to see if they can tell the difference between "I know this" and "I have no idea."

Here is the breakdown of what the researchers did, using simple analogies:

1. The Problem: The "Know-It-All" vs. The "Shy Kid"

In the past, researchers tested these robots by asking them questions where the answer was hidden in a pile of documents.

  • The Old Way: They just checked if the robot got the answer right.
  • The Flaw: If the robot didn't know the answer, it would either:
    • Hallucinate: Guess wildly and make up a story (like a student who guesses "42" on a math test just to get points).
    • Deflect: Say "I don't know" even when the answer was right in front of them (like a shy student who stays silent even though they raised their hand).

The old tests didn't care how the robot failed, only if it failed. But in the real world, we need robots that know when to speak up and when to stay quiet.

2. The Solution: A Dynamic "Trap" Test

The researchers built a new testing ground called VLM-DeflectionBench. Imagine this as a high-tech escape room with four different rooms:

  • Room 1: The Memory Room (Parametric). No books, no pictures, just the question. The robot must rely only on what it learned in school.
    • Goal: If the robot doesn't know it, it should say, "I don't know."
  • Room 2: The Perfect Library (Oracle). The robot is given the exact right book with the answer highlighted.
    • Goal: It should answer correctly. If it says "I don't know" here, it's being too shy.
  • Room 3: The Messy Desk (Realistic). The robot gets the right book, but it's buried under 10 other books with wrong information.
    • Goal: It needs to find the right book and ignore the noise.
  • Room 4: The Trap (Adversarial). The robot is given only the wrong books (distractors).
    • Goal: It should realize the books are lying and say, "I cannot answer this." If it tries to guess based on the wrong books, it fails.

3. The "Filter" Machine

One of the biggest problems with old tests is that as robots get smarter, they memorize the answers. A question that used to require looking up a book might now be answered from memory.

The researchers created a dynamic pipeline (a smart filter). Before a question becomes part of the test, they run it through a panel of other super-smart robots.

  • If any robot can answer it from memory, the question is thrown out.
  • Only the questions that truly require looking up new information are kept.
  • This ensures the test stays hard and relevant, even as AI gets smarter in the future.

4. What They Discovered

They tested 20 of the smartest robots on the market (both open-source and expensive commercial ones). The results were surprising:

  • The "Confident Liar" Problem: Even the best robots often hallucinate. When they were given misleading information (Room 4), they didn't say "I don't know." Instead, they confidently made up answers based on the lies.
  • The "Text Bias": The robots are weirdly biased toward text. If they see a picture that says "A" but a piece of paper says "B," they often ignore the picture and believe the paper, even if the picture is the truth.
  • The "Shy Kid" Problem: When the researchers told the robots to be very strict ("Only answer if you are 100% sure!"), the robots became too cautious. They stopped answering even when they had the right information, causing their accuracy to crash.

5. The Big Takeaway

The paper concludes that we need to stop just asking AI "Are you smart?" and start asking "Are you reliable?"

A truly reliable AI isn't just one that knows facts; it's one that knows when it doesn't know. It needs to be brave enough to guess when it's right, but humble enough to say "I don't know" when the evidence is missing or confusing.

In short: The researchers built a better "driver's license test" for AI. They found that while these robots are great at driving on empty roads, they tend to crash when the road is foggy (noisy data) or when they are tricked by a fake sign (distractors). They need to learn to pull over and ask for help instead of crashing the car.

Get papers like this in your inbox

Personalized daily or weekly digests matching your interests. Gists or technical summaries, in your language.

Try Digest →