Succeeding at Scale: Automated Dataset Construction and Query-Side Adaptation for Multi-Tenant Search

This paper introduces DevRev-Search, an automated benchmark for technical support retrieval, and proposes an Index-Preserving Adaptation strategy that fine-tunes only the query encoder to achieve scalable, high-performance multi-tenant search without the prohibitive cost of re-indexing.

Prateek Jain, Shabari S Nair, Ritesh Goru, Prakhar Agarwal, Ajay Yadav, Yoga Sri Varshan Varadharajan, Constantine Caramanis

Published 2026-03-05
📖 4 min read☕ Coffee break read

Imagine you run a massive, bustling library that serves thousands of different neighborhoods (tenants). Each neighborhood has its own unique set of books, rules, and slang. Your goal is to build a super-smart librarian robot that can instantly find the right book for any question a visitor asks.

However, you face two massive problems:

  1. The "Dark Data" Problem: You have millions of books, but no one has written "review cards" saying which books are actually good answers to specific questions. It's like having a library where the books are there, but the catalog is blank. You can't train your robot because you don't know what "good" looks like.
  2. The "Re-Shelving" Tax: Every time you want to teach your robot a new trick, you usually have to take every single book off the shelf, re-read it, and re-shelve it in a new order. If you have 1,000 neighborhoods, doing this for every single update would take forever and cost a fortune.

This paper, "Succeeding at Scale," introduces a new way to solve both problems. Here is how they did it, explained simply:

1. Building the Training Manual Without Humans (The "AI Detective" Pipeline)

Usually, you need human experts to read questions and mark the correct answers to train a search engine. But that's slow and expensive.

The authors built an automated factory to create this training data:

  • The Scavenger Hunt: Instead of relying on one search tool, they sent out a team of seven different "scouts" (some look for exact word matches, others look for meaning). They gathered every possible answer these scouts could find.
  • The Judge: They then used a super-smart AI (an LLM) as a "Judge." This Judge looked at the pile of answers and asked: "Does this actually answer the question, or is it just a fancy-looking distraction?"
  • The Result: The AI filtered out the junk and kept only the gold. They created a massive, high-quality training dataset (called DevRev-Search) without a single human having to manually label a single document.

2. The "One-Sided" Makeover (Index-Preserving Adaptation)

This is the paper's biggest breakthrough.

In the old way, to make the librarian smarter, you had to reorganize the entire library (the documents) every time.

  • The Old Way: Imagine you want to teach the librarian how to understand a new neighborhood's slang. You have to re-shelve every book in the entire building to match the new slang. Impossible for a huge library.
  • The New Way (Query-Only Adaptation): The authors realized they only needed to change the Librarian's brain (the query encoder), not the books themselves.
    • They kept the library shelves exactly as they were (frozen document index).
    • They only gave the librarian a "brain upgrade" to understand the specific questions from that neighborhood.
    • The Analogy: It's like giving your librarian a pair of specialized glasses for a specific customer. You don't need to move the books; you just change how the librarian looks at the question. This makes updates instant and cheap.

3. The "Lightweight" Upgrade (Parameter-Efficient Fine-Tuning)

Even upgrading the librarian's whole brain is heavy. So, they used a technique called PEFT (Parameter-Efficient Fine-Tuning).

  • The Analogy: Instead of rebuilding the librarian's entire brain (which has billions of neurons), they just added a few smart sticky notes or a small cheat sheet to the librarian's desk.
  • They found that using a method called LoRA (Low-Rank Adaptation) is like giving the librarian a tiny, highly efficient notebook.
  • The Magic: This tiny notebook allows the librarian to learn the new neighborhood's needs almost as well as if they had rebuilt their whole brain, but it uses 99% less computing power and memory.

The Bottom Line

The authors proved that:

  1. You can build a perfect training dataset using AI judges instead of humans.
  2. You can make a search engine smarter for specific customers without ever touching the massive database of documents.
  3. You can do this with a tiny, efficient "upgrade" that saves massive amounts of money and time.

In short: They figured out how to teach a giant, multi-tenant search engine to be a genius for every specific customer, without ever having to move a single book on the shelf. It's a win for speed, cost, and quality.