Social Teaching: Being Informative vs. Being Right in Sequential Decision Making

This paper demonstrates that in sequential Bayesian decision-making, agents can minimize the error risk of the final decision by intentionally adopting inaccurate initial beliefs that are systematically biased toward the unlikely hypothesis, thereby prioritizing the informativeness of their actions over immediate correctness.

Joong Bum Rhim, Vivek K Goyal

Published 2026-03-12
📖 6 min read🧠 Deep dive

Here is an explanation of the paper "Social Teaching: Being Informative vs. Being Right in Sequential Decision Making," translated into simple, everyday language with creative analogies.

The Big Idea: Sometimes, Being "Wrong" Helps Everyone Else

Imagine a group of friends trying to guess the answer to a tricky riddle. They are sitting in a line, one behind the other.

  • Friend #1 gets a hint (a private signal) and makes a guess.
  • Friend #2 hears Friend #1's guess, gets their own hint, and makes a guess.
  • Friend #3 hears both previous guesses, gets their own hint, and guesses.
  • And so on, until the Last Friend makes the final call.

The paper asks a surprising question: Should the first friend try their hardest to be 100% correct based on the facts they know?

The answer, surprisingly, is no.

The authors (Rhim and Goyal) discovered that if the first few people in the line are too confident in their own "facts" (their prior beliefs), they might actually make the final decision worse. Instead, to help the last person make the best possible choice, the early people should be a little bit "open-minded"—even if it means they are technically "wrong" about the odds.

The Metaphor: The Jury and the Glasses

To understand why, let's look at the paper's analogy of a Jury Trial.

Imagine a defendant is on trial. The jury has to decide: Guilty or Not Guilty.

  • The Evidence: The lawyers present clues. Sometimes the clues are clear; sometimes they are blurry (like a noisy phone call).
  • The Prior Belief: Before the trial starts, the jurors have a gut feeling about how likely it is that the defendant is guilty. Maybe they think, "Most people who wear glasses look smart, so this guy is probably innocent," or "This guy looks shady, so he's probably guilty."

In the real world, jurors often let their gut feelings (biases) influence how they weigh the evidence. The paper argues that in a sequential setting (where one juror speaks, then the next, then the next), having a "perfectly accurate" gut feeling isn't always the best strategy for the group.

The Problem: The "Echo Chamber" of Confidence

Let's say the first person in line (Alexis) is 100% sure the answer is "No."

  • She sees a tiny bit of evidence that suggests "Yes," but because she is so confident in her "No" belief, she ignores it and says "No."
  • The second person (Blake) hears "No." He thinks, "Well, Alexis is smart. She probably saw something I didn't." Even if Blake's own hint suggests "Yes," he might just follow Alexis.
  • The third person hears "No, No." They follow suit.

This is called herding. The group gets stuck on the wrong answer because the first person was too rigid.

The Solution: The "Open-Minded" Adviser

The paper suggests a counter-intuitive strategy called Social Teaching.

If you are the first person in line, your job isn't just to be right for yourself. Your job is to be informative for the people behind you.

  • The Scenario: Imagine the truth is that "Guilty" is very rare (only 10% chance).
  • The "Right" Way: You see a hint. You think, "It's only 10% likely, so I'll bet on 'Not Guilty'." You are technically correct.
  • The "Open-Minded" Way (The Paper's Recommendation): You should act as if the chance of "Guilty" is actually higher (say, 30%).
    • Why? Because if you act as if the rare event is more likely, you will be more willing to say "Guilty" if you see a hint that suggests it.
    • If you do say "Guilty," the next person will think, "Wow! Even though the odds were low, Alexis said 'Guilty'. She must have seen a really strong hint!"
    • This makes your decision more valuable information for the next person.

The Analogy of the Lighthouse:
Imagine you are a lighthouse keeper.

  • If you are rigid, you only flash your light when you are 100% sure a ship is there. If the fog is thick, you stay dark. The ships behind you (the next agents) get no signal and crash.
  • If you are open-minded, you flash your light even when you are only 60% sure. You are saying, "Hey, there might be a ship here!"
  • The next lighthouse keeper sees your flash. They think, "Okay, the first guy flashed. He's usually cautious, so he must have seen something." They combine your signal with their own view and make a much better decision.

The "Gaussian" Twist (The Math Part Made Simple)

The paper uses math (specifically "Gaussian likelihoods," which is just a fancy way of saying "noise that looks like a bell curve") to prove this.

They found a specific pattern:

  1. Early Agents (The Advisers): If the true odds of an event are low, they should pretend the odds are higher. If the true odds are high, they should pretend they are lower. They need to be "open-minded" toward the unlikely.
  2. The Last Agent (The Decision Maker): The final person should do the opposite. They should be slightly more conservative to balance out the "open-mindedness" of the people before them.

Why Does This Happen?

It comes down to a trade-off:

  • Being Right: Optimizing your own decision based on what you know.
  • Being Informative: Optimizing your decision so that it sends the clearest possible message to the next person.

Sometimes, to send the clearest message, you have to take a risk that you might be wrong yourself. By being slightly "wrong" about the odds, you force yourself to react more strongly to new evidence, which creates a stronger signal for the next person.

The Takeaway

In a team where people make decisions one after another:

  • Don't just be a robot following the data.
  • Be a "Social Teacher."
  • If you are the first to speak, don't be too rigid. Be open to the unlikely possibilities. This might make your decision slightly less accurate, but it will make the entire team's final decision much smarter.

In short: The best adviser isn't the one who is always right; it's the one who is open-minded enough to show the rest of the team the way.