Legal interpretation and AI: from expert systems to argumentation and LLMs

This paper traces the evolution of AI and Law research on legal interpretation, highlighting the shift from expert systems focused on knowledge engineering, to argumentation frameworks analyzing dialectical interactions, and finally to machine learning models generating interpretive suggestions.

Václav Janeček, Giovanni Sartor

Published 2026-03-06
📖 5 min read🧠 Deep dive

The Big Picture: Who Decides What the Law Means?

Imagine the law is a giant, ancient rulebook written in a language that is sometimes vague, sometimes contradictory, and often open to debate. The central question of this paper is: How can we teach computers to figure out what these rules actually mean?

The authors, Václav Janeček and Giovanni Sartor, take us on a journey through three different eras of trying to solve this puzzle, using the classic example of a park sign that says: "No vehicles in the park."


Era 1: The "Rule-Maker" (Expert Systems)

The Analogy: The Strict Librarian.

In the 1970s and 90s, researchers tried to build AI like a super-strict librarian. They believed that if humans could write down every single rule perfectly, the computer could just follow them like a recipe.

  • How it worked: A human expert (the librarian) had to decide beforehand exactly what a "vehicle" is. They would write a rule in the computer: "If it has wheels and an engine, it's a vehicle. If it's a bicycle, it's a vehicle."
  • The Problem: This is like trying to pack a suitcase for a trip to every possible climate in the world before you leave. What if a child rides a tiny tricycle? What if it's a toy car?
  • The Flaw: The computer couldn't handle the "gray areas." If the human expert made a mistake in the rules (e.g., deciding all bikes are vehicles), the computer would blindly ban a child's bike, even if the park guard would have let it slide. The computer had no common sense; it just followed the manual.

Era 2: The "Debate Club" (Argumentation)

The Analogy: The Courtroom Drama.

Researchers realized that law isn't just about following a list; it's about arguing. So, they changed the AI from a rule-follower to a debate moderator.

  • How it works: Instead of hard-coding the answer, the AI builds two teams of arguments.
    • Team A (The Guard): "Bikes are vehicles! The sign says 'No vehicles,' so the kid can't ride." (Argument based on Ordinary Language).
    • Team B (The Kid): "But the point of the rule is to stop loud, dangerous motorbikes. A quiet kid's bike doesn't hurt anyone. So, 'vehicle' shouldn't include this." (Argument based on Purpose/Teleology).
  • The Magic: The AI doesn't just pick one; it weighs the arguments. It asks: "Which argument is stronger?" In this case, the "Purpose" argument might win, allowing the kid to ride.
  • The Result: This approach mimics how real lawyers think. It acknowledges that the law is a conversation, not a computer program.

Era 3: The "Super-Reader" (Large Language Models / LLMs)

The Analogy: The Over-Confident Intern.

Today, we have Generative AI (like ChatGPT). These aren't rule-followers or debate moderators; they are pattern matchers. They have read almost every book, law, and article on the internet.

  • How it works: You ask the AI, "Can a kid ride a bike in the park?" The AI doesn't "know" the answer. Instead, it looks at billions of similar sentences it has read before and predicts what a human would say. It's incredibly good at sounding smart.
  • The Good News: It can explain the rule in plain English, list different ways to interpret it, and even find the "purpose" of the law just like a lawyer. It's like having a brilliant intern who has read the entire library in 5 seconds.
  • The Bad News (The "Hallucination"): Because the AI is guessing based on patterns, it can lie confidently. It might invent a fake law or cite a court case that doesn't exist. It has no understanding of truth, only of what looks like truth.
    • The Danger: If you let this "Intern" make the final decision, it might fine the kid because it "thinks" bikes are vehicles, even if the real human judge would say no. It lacks the moral compass and the ability to understand the spirit of the law.

The Verdict: What Should We Do?

The authors conclude with a very important warning: Don't let the AI be the Judge.

  • AI is a Tool, Not a Boss: Think of the AI as a "research assistant." It can find the arguments, summarize the laws, and suggest ideas.
  • The Human is the Captain: A human lawyer or judge must look at what the AI says, check if it's true, and make the final call. The human brings the "soul" of the law—the ability to understand fairness, context, and human values.
  • The Future: The best path forward is to combine the strengths. Use the AI's speed to find information and the human's wisdom to weigh the arguments.

Summary in One Sentence

We moved from trying to teach computers rigid rules (which failed), to teaching them how to argue (which worked better), to giving them a super-memory (which is powerful but dangerous), and the lesson is: Let the computer do the research, but let the human make the judgment.

Get papers like this in your inbox

Personalized daily or weekly digests matching your interests. Gists or technical summaries, in your language.

Try Digest →