Elenchus: Generating Knowledge Bases from Prover-Skeptic Dialogues

This paper introduces Elenchus, a dialogue system that leverages prover-skeptic interactions between a human expert and an LLM to construct knowledge bases grounded in inferentialist semantics, mapping dialectical states to formal logic to explicitly capture and verify the inferential relationships and design rationales of complex ontologies like PROV-O.

Bradley P. Allen

Published Tue, 10 Ma
📖 5 min read🧠 Deep dive

Imagine you are trying to write a rulebook for a complex game, like chess or a new video game. Traditionally, a "Knowledge Engineer" would sit down with a Grandmaster, ask them questions like, "What happens if a pawn reaches the end?" and then try to write down the rules based on the Grandmaster's answers.

The problem is that experts often don't have a perfect, pre-written rulebook in their heads. They know how to play by feeling and practice, but explaining the "why" and "what if" is hard. They might give a vague answer, or forget a weird edge case. This is the "Knowledge Acquisition Bottleneck."

Elenchus is a new way to build these rulebooks. Instead of just asking questions, it turns the process into a debate between a human expert and an AI.

Here is how it works, using some simple analogies:

1. The Setup: The "Socratic Sparring Partner"

Think of the AI not as a smart encyclopedia that knows everything, but as a Socratic sparring partner (like a debate coach).

  • The Human (The Expert): Has the deep knowledge of the game. They make a claim, like "If a pawn moves forward, it captures diagonally."
  • The AI (The Skeptic): Its job isn't to know the answer, but to poke holes in the expert's logic. It says, "Wait, if the pawn moves forward, but there's a wall there, does it still capture? Or what if two pawns are in the same spot?"

The AI acts as a "Defeasible Oracle." This is a fancy way of saying: "I'm going to guess where your logic might be broken. You are the boss. If I'm wrong, you tell me. If I'm right, you fix your rule."

2. The Process: Finding the "Tensions"

As they talk, the AI points out Tensions.

  • Analogy: Imagine you are building a house of cards. You say, "I put a card here." The AI says, "If you put a card there, and you also put one there, the whole thing collapses."
  • The expert has to fix the collapse. They can:
    • Retract: "Okay, I take that back. That rule doesn't work."
    • Refine: "Ah, I see. It only works if the card is red."
    • Contest: "No, actually, my rule is fine, and your guess about the collapse was wrong."

Every time the expert agrees that a tension is real, they are essentially saying, "Okay, I admit these two rules cannot exist together." This agreement becomes a permanent part of the new rulebook.

3. The Result: A "Living Map" of Logic

At the end of the conversation, you don't just have a list of rules. You have a Material Base.

  • The Analogy: Think of a standard rulebook as a flat map. Elenchus creates a 3D topographic map. It doesn't just show you where the roads are; it shows you the relationships between them. It knows that "Road A leads to Road B," but also that "Road A breaks if you add Road C."
  • This map is special because it is non-monotonic. In normal logic, if A leads to B, adding more facts usually doesn't change that. But in the real world (and in Elenchus), adding a new fact (like "It's raining") might change the outcome (the road is now closed). Elenchus captures these real-world "exceptions" naturally.

4. The Proof: The "W3C PROV-O" Test

The authors tested this on a real-world standard called PROV-O, which is like the "Constitution" for tracking where data comes from on the internet.

  • They fed the AI a short, 350-word paragraph describing the rules.
  • The AI and the human expert had a debate.
  • The Magic: The final "rulebook" the AI helped build matched the actual design decisions made by the original team of experts years ago.
  • Even better, the original team had to sift through 8,820 emails and 152 meetings to figure out their own design choices. Elenchus did it in a single dialogue session, and the resulting logic was so precise that a computer could verify it was mathematically correct.

Why This Matters

  • No More "Black Boxes": In many AI systems, you don't know why the AI made a rule. In Elenchus, every single rule in the final database can be traced back to a specific moment in the conversation where the expert said, "Yes, I agree that is a problem."
  • The Human is Still King: The AI can hallucinate (make things up), but because the human gets to say "No, that's wrong," the AI's mistakes are filtered out. The human remains the authority.
  • Making the Implicit Explicit: Experts often know things they can't easily explain. Elenchus forces them to explain by challenging them, turning their "gut feeling" into a formal, logical structure.

In short: Elenchus is a tool that turns a messy, natural conversation between a human and an AI into a perfectly structured, mathematically verified rulebook, by treating the conversation as a game of "find the contradiction" rather than "extract the answer."