BIS Reasoning 1.0: The First Large-Scale Japanese Benchmark for Belief-Inconsistent Syllogistic Reasoning

This paper introduces BIS Reasoning 1.0, the first large-scale Japanese benchmark for evaluating belief-inconsistent syllogistic reasoning in LLMs, demonstrating that robustness against belief bias is driven primarily by explicit reasoning optimization rather than language specialization or model scale alone.

Ha-Thanh Nguyen, Hideyuki Tachibana, Chaoran Liu, Qianying Liu, Su Myat Noe, Koichi Takeda, Sadao Kurohashi

Published 2026-03-17
📖 5 min read🧠 Deep dive

🧠 The Big Idea: Can AI Think, or Just Guess?

Imagine you are taking a logic test. The teacher gives you two rules:

  1. Rule: All cats are dogs.
  2. Rule: All dogs are purple.
  3. Question: Therefore, are all cats purple?

If you are a human who knows cats aren't dogs, your brain screams, "Wait, that's wrong!" You might get confused and say "No" because your real-world knowledge is fighting the logic.

But if you are a perfect logic machine, you ignore the fact that cats aren't actually dogs. You follow the rules strictly: If A is B, and B is C, then A must be C. So, you correctly answer "Yes."

This paper is about building a giant test to see if AI (Large Language Models) can be that perfect logic machine, or if they get tripped up by their own "common sense."


🏗️ What Did They Build? (The "BIS" Dataset)

The researchers created a new dataset called BIS Reasoning 1.0. Think of this as a giant obstacle course for AI, specifically designed to trick them.

  • The Trap: The test contains 5,000 puzzles in Japanese. In every puzzle, the answer is logically correct, but it sounds absurd or wrong based on real life.
    • Example: "All birds can fly. Penguins are birds. Therefore, penguins can fly." (Logically valid based on the rules, but factually false in the real world).
  • The Goal: They wanted to see if the AI would say "No" because it knows penguins can't fly (believing bias), or "Yes" because it followed the math of the sentence (logical reasoning).

They call this "Belief-Inconsistent Reasoning." It's like asking the AI to wear "logic goggles" and ignore its "common sense glasses."


🏆 The Race: Who Passed the Test?

The researchers put many different AI models through this obstacle course. Here is how they fared:

1. The "Super-Reasoners" (The Gold Medalists) 🥇

  • Who: Newer, specialized models like GPT-5 and Qwen.
  • Performance: They got almost 100% correct.
  • Analogy: These are like students who have been trained specifically to ignore distractions. When they see the puzzle, they put on their "logic goggles," ignore the fact that penguins can't fly, and solve the math perfectly.

2. The "Old Guard" (The Strugglers) 🥉

  • Who: Older models like GPT-4o and early Japanese models.
  • Performance: They scored around 60–80%.
  • Analogy: These models are like smart people who are too easily distracted. When they see "Penguins can fly," their brain says, "Wait, that's impossible!" and they get confused, failing the logic test even though the answer was right based on the rules.

3. The "Japanese Specialists" (The Up-and-Comers) 🇯🇵

  • Who: Models built specifically for the Japanese language (like llm-jp).
  • Performance: The older versions were terrible (scoring below 50%, sometimes even guessing wrong on purpose!). But the newest version (llm-jp-3.1) jumped up to the 80s.
  • Analogy: The old Japanese models were like chefs who knew how to cook delicious Japanese food but couldn't follow a recipe if the ingredients were weird. The new version learned to follow the recipe strictly, even if the ingredients were strange.

🔍 Key Discoveries (The "Aha!" Moments)

1. Size Doesn't Matter, Training Does 📏

You might think a bigger, smarter AI would automatically be better at logic. Not true.

  • Analogy: A giant library full of books (a big model) doesn't mean the librarian knows how to solve a riddle. It's about how the librarian was trained. If you train them to prioritize logic over facts, they get better. If you just train them to sound polite or fluent, they get worse at logic.

2. The "Prompt" is the Secret Sauce 📝

The researchers found that how you ask the question changes the answer.

  • The "Casual" Ask: "Hey, does this make sense?" -> The AI gets lazy and uses common sense.
  • The "Strict" Ask: "Ignore real life. Follow the rules step-by-step." -> The AI wakes up and solves it correctly.
  • Analogy: It's like asking a friend, "What's 2+2?" vs. "Pretend 2+2 equals 5. Now, what is 2+2?" The second question forces them to switch gears.

3. The "Thinking Time" Matters ⏳

Newer models have a setting called "Reasoning Effort."

  • Low Effort: The AI guesses quickly. (Score: Low)
  • High Effort: The AI takes a moment to "think" through the steps. (Score: High)
  • Analogy: It's the difference between a student who guesses the answer on a multiple-choice test in 2 seconds vs. one who actually works out the problem on scratch paper.

🚨 Why Should We Care? (The Real World Impact)

Why do we need to test AI on "nonsense" logic? Because real life is full of traps.

Imagine an AI lawyer or a doctor:

  • The Lawyer: If a client says, "All people who wear red are guilty. John wears red. Therefore, John is guilty," a biased AI might say, "Well, John looks innocent, so no." But a logical AI must say, "Based on the rule you gave me, yes, he is guilty."
  • The Doctor: If a new study says, "All patients with this symptom have a rare disease," the AI shouldn't say, "That can't be true, I've never heard of it." It needs to follow the evidence, even if it feels weird.

The Conclusion:
To make AI safe for law, medicine, and science, we can't just rely on them being "fluent" or "polite." We need to train them to be stubbornly logical, even when the answer feels wrong. This new test (BIS Reasoning 1.0) is the first step in making sure Japanese AI can do that.

Drowning in papers in your field?

Get daily digests of the most novel papers matching your research keywords — with technical summaries, in your language.

Try Digest →