DRBench: A Realistic Benchmark for Enterprise Deep Research

This paper introduces DRBench, a realistic benchmark comprising 100 human-verified, multi-step deep research tasks across 10 enterprise domains that evaluate AI agents on their ability to synthesize information from both public web and private company data to generate accurate, structured reports.

Amirhossein Abaskohi, Tianyi Chen, Miguel Muñoz-Mármol, Curtis Fox, Amrutha Varshini Ramesh, Étienne Marcotte, Xing Han Lù, Nicolas Chapados, Spandana Gella, Peter West, Giuseppe Carenini, Christopher Pal, Alexandre Drouin, Issam H. Laradji

Published Wed, 11 Ma
📖 3 min read☕ Coffee break read

Imagine you've hired a super-smart, tireless research assistant named "AI." You ask it a simple question like, "What's the weather in Tokyo?" It checks the internet and gives you the answer instantly. That's easy.

But now, imagine you give that same assistant a much harder, real-world job: "Look at our company's private emails, check our internal sales spreadsheets, scan our chat logs, and then compare all that with the latest news on the open web. Based on everything you find, write a report telling us how to change our product plan so we don't break new safety laws."

This is exactly what the paper DRBench is all about. Here is the breakdown in simple terms:

1. The Problem: The "Google Test" vs. The "Real World"

Previous tests for AI were like a multiple-choice quiz. They asked simple questions where the answer was just one click away on Google. But in the real business world, answers aren't sitting on a single webpage. They are hidden in a messy pile of:

  • Private company documents (like a locked filing cabinet).
  • Old emails and Slack messages (like a chaotic group chat).
  • Public news and websites (like the open internet).

Old AI tests were like asking a student to solve a math problem on a clean whiteboard. DRBench is like throwing that student into a busy library, a noisy office, and a construction site all at once and asking them to build a house.

2. The Solution: DRBench (The "Deep Research" Gym)

The authors created DRBench, which is essentially a gym for AI agents. Instead of lifting light weights (simple questions), they are lifting heavy, complex tasks.

  • The Workouts: They created 100 realistic scenarios across 10 different departments (like Sales, Cybersecurity, and Compliance).
  • The Rules: The AI has to be a detective. It can't just guess; it has to dig through the "private vault" (company data) and the "public square" (the web) to find the clues.
  • The Goal: The AI isn't just graded on finding the answer; it's graded on how well it connects the dots, tells the truth, and writes a clear, organized report that a human boss could actually use.

3. How They Built It

They didn't just make up fake questions. They used a special recipe:

  1. Synthesis: They used computers to generate realistic "personas" (like a stressed Sales Manager or a strict Compliance Officer).
  2. Human Check: Real humans stepped in to verify that the tasks made sense and were actually hard enough.
  3. The Result: A massive dataset of 100 "deep research" missions that mimic the chaos and complexity of a real office.

4. Why It Matters

The authors tested many different AI models (like GPT, Llama, and Qwen) on this new gym. They found that while some AIs are great at simple trivia, they often get lost when the task requires juggling private data and public info simultaneously.

In a nutshell:
Think of DRBench as the Olympics for AI researchers. Before this, AI was competing in the 100-meter dash (simple, fast answers). Now, they are competing in the Decathlon, where they have to run, jump, throw, and navigate complex obstacles (private data, public web, and logical reasoning) all at once.

This benchmark helps companies figure out which AI assistants are actually ready to handle the messy, complicated work of running a business, rather than just answering simple questions.

Where to find it: If you want to see the "gym" yourself, the code and data are open for everyone to use on GitHub.