InfoGatherer: Principled Information Seeking via Evidence Retrieval and Strategic Questioning

InfoGatherer is a principled framework for high-stakes document-grounded QA that combines retrieved evidence with strategic user questioning, utilizing Dempster-Shafer belief assignments to model uncertainty and improve decision reliability while reducing interaction turns.

Maksym Taranukhin, Shuyue Stella Li, Evangelos Milios, Geoff Pleiss, Yulia Tsvetkov, Vered Shwartz

Published 2026-03-09
📖 4 min read☕ Coffee break read

Imagine you are a detective trying to solve a mystery, but instead of a crime scene, you are dealing with a patient who has a sore throat or a client with a legal problem. The person comes to you with a vague story: "I have a cough and a sore throat."

In the past, if you asked an AI (a Large Language Model) to solve this, it would often act like a confident but overzealous detective. It would look at the limited clues, guess the culprit (maybe "Flu"), and give you a very loud, very sure answer, even if it was wrong. It hates saying, "I don't know enough yet."

INFOGATHERER is a new, smarter way to build these AI detectives. It changes the game from "guessing fast" to "investigating carefully."

Here is how it works, using some simple analogies:

1. The Two-Pronged Investigation

Most AI detectives rely only on what they memorized in school (their internal training data). If that data is old or missing a detail, they fail.

INFOGATHERER has two tools:

  • The Library (Retrieved Documents): Before asking the user anything, it runs to a library of trusted, up-to-date rulebooks (like medical textbooks or legal codes) to see what the rules say about the symptoms.
  • The Interviewer (Strategic Questioning): It doesn't just guess; it asks the user specific questions to fill in the blanks.

2. The "Fuzzy Map" vs. The "Sharp Guess"

This is the most important part.

  • Old AI: Uses a Sharp Guess. It says, "There is a 90% chance it's the Flu." If it's wrong, it's still 90% sure it's right. This is dangerous in medicine or law.
  • INFOGATHERER: Uses a Fuzzy Map (based on something called Dempster-Shafer Theory). Instead of forcing a single percentage, it admits, "I have some evidence for Flu, some for Allergies, and a big chunk of 'I don't know yet'."

The Analogy: Imagine you are trying to find a lost dog.

  • Old AI points at a bush and says, "It's definitely in that bush!" (Even if the dog is actually in the garage).
  • INFOGATHERER draws a circle around the neighborhood and says, "The dog is likely in the neighborhood, but I'm not sure which house. I need to check the garage next." It keeps the "I don't know" part visible until it finds the dog.

3. The Detective's Strategy (The "Stop" Button)

How does the AI know when to stop asking questions?

  • Old AI often stops too early because it feels "confident" (even if that confidence is fake).
  • INFOGATHERER uses a Confidence Meter that only moves when it finds new, solid evidence. It keeps asking questions until the "I don't know" part of its Fuzzy Map shrinks enough that it can point to one specific answer with high certainty.

It's like a game of 20 Questions, but the AI is playing perfectly. It doesn't ask, "Is it an animal?" if it already knows it's a disease. It asks, "Do you have a fever?" because that specific question will cut the list of possibilities in half.

4. Why This Matters

The paper tested this on Medical (diagnosing diseases) and Legal (solving court cases) problems.

  • The Result: INFOGATHERER got the right answer more often than other AI methods, but it asked fewer questions.
  • The Reason: It didn't waste time asking useless questions, and it didn't guess until it was sure.

The Big Picture

Think of INFOGATHERER as a wise, cautious expert rather than a fast, confident guesser.

  • It reads the rulebook first.
  • It admits when it is confused.
  • It asks the right questions to clear up the confusion.
  • It only gives an answer when it has built a solid case.

In high-stakes fields like healthcare and law, where a wrong guess can hurt someone, this shift from "confident guessing" to "principled investigation" is a huge step forward for making AI trustworthy.