Benchmarking Legal RAG: The Promise and Limits of AI Statutory Surveys

This paper benchmarks emerging legal RAG tools against the LaborBench dataset, revealing that while commercial AI platforms underperform standard RAG, a custom tool (STARA) significantly improves accuracy, though further analysis suggests its true performance is even higher due to previously unaccounted omissions in the ground truth data.

Mohamed Afane, Emaan Hariri, Derek Ouyang, Daniel E. Ho

Published 2026-03-05
📖 5 min read🧠 Deep dive

Imagine you are a lawyer trying to answer a very specific question: "Does the state of California allow a worker to keep their unemployment benefits while starting a small business?"

To get the right answer, you can't just ask one person. You have to check the rulebooks of all 50 states. Each state has its own thick, complex book of laws, written in confusing legal jargon, with rules that change depending on the year, the type of job, or even the size of the company.

Doing this manually is like trying to find a specific needle in 50 haystacks, where the needles are made of glass and the haystacks are on fire. It takes a team of experts months to do it.

This paper is about testing Artificial Intelligence (AI) to see if it can do this job faster and better. The researchers set up a "test drive" (called a benchmark) to see how well different AI tools can read these 50 state rulebooks and answer questions accurately.

Here is the breakdown of what they found, using simple analogies:

1. The Three Contenders

The researchers put three different "AI students" to the test:

  • The Custom Student (STARA): A specialized tool built by the researchers specifically designed to read legal codes. It's like a student who has memorized the entire library and knows exactly how to find a book by its spine.
  • The Big Brand Student A (Westlaw AI): A famous, expensive legal tool marketed to lawyers. It's like a general encyclopedia that claims to know everything but is trying to do a very specific math problem.
  • The Big Brand Student B (Lexis+ AI): Another famous legal tool, similar to Student A, claiming to be a "game-changer" for legal research.

2. The Results: Who Passed the Test?

The test was a series of True/False questions about unemployment laws across the US.

  • The Custom Student (STARA) Aced It: It got 83% of the answers right. When the researchers double-checked the "wrong" answers, they realized the student was actually right, but the official answer key (made by human experts) was wrong. Once they fixed the answer key, this student's score jumped to 92%.

    • Analogy: This student didn't just guess; they actually read the fine print. They even found rules that the human experts had missed!
  • The Big Brand Students Struggled:

    • Westlaw AI got about 58% right. It was so eager to say "Yes" that it hallucinated rules that didn't exist.
    • Lexis+ AI got about 64% right. It was very careful and rarely said "Yes," but that meant it missed a lot of valid rules (it said "No" when the answer was "Yes").
    • Analogy: Westlaw AI was like a student who guesses "True" for everything to get points, even when they don't know the answer. Lexis+ AI was like a student who is so afraid of being wrong that they leave most questions blank.

3. The Big Surprise: The "Answer Key" Was Wrong

The most shocking discovery wasn't about the AI; it was about the humans.

The test was graded against an official report created by the U.S. Department of Labor (DOL). This report was made by teams of human lawyers who spent six months manually reading the laws. The researchers assumed this report was the "perfect truth."

But when they checked the AI's "wrong" answers, they found that the AI was often right, and the human experts were wrong.

  • The human experts had missed valid laws in several states.
  • The AI found these missing laws because it read the raw text more thoroughly than the tired human team.
  • Analogy: Imagine a teacher grading a test using an old, outdated answer key. The student (AI) writes the correct answer, but the teacher marks it wrong because the key is missing the update. The researchers realized the "teacher" (the DOL) needed to update their key!

4. Why Did the Big Brands Fail?

The paper explains that the commercial tools (Westlaw and Lexis) have some major design flaws for this specific job:

  • The "Character Limit" Trap: Westlaw AI has a strict limit on how much text you can type into it. It's like trying to explain a complex legal question using only a tweet. You have to cut out all the important details, so the AI gets confused.
  • The "Keyword" Trap: These tools often look for words that sound similar rather than understanding the meaning. If you ask about "unemployment," they might pull up a law about "employment discrimination" just because they both have the word "employment."
  • The "Black Box" Problem: We don't know exactly how these commercial tools work. They are like magic 8-balls. You ask a question, they give an answer, but they won't show you the specific page of the law they used to decide. This makes it hard to trust them.

5. The Takeaway

The paper concludes that AI has huge potential for legal research, but we need to build it differently.

  • Don't just use generic AI: You can't just plug a general chatbot into a legal database and expect it to work. You need tools designed specifically to understand the structure of laws (like STARA).
  • Human experts make mistakes too: Even the best human teams miss things. AI can actually help humans by double-checking their work and finding the rules they overlooked.
  • Transparency is key: Legal AI needs to show its work. It shouldn't just say "Yes" or "No"; it needs to say "Yes, and here is the exact paragraph in the law that proves it."

In short: The specialized AI student (STARA) outperformed the famous commercial brands and even found mistakes in the human experts' work. It turns out, when it comes to reading the fine print of 50 different state laws, a smart, specialized robot might just be better than a tired team of humans or a flashy, generic commercial tool.