Robust Testing Of the Allais Paradox By Paired Choices vs. Paired Valuations

This paper refutes the claim that valuation tests are a superior alternative to paired choices for testing the Allais paradox by demonstrating that valuation tests are inherently biased under standard stochastic choice models, whereas a "strong" paired choice test remains robust and confirms the continued prevalence of the common ratio effect.

Federico Echenique, Gerelt Tserenjigmid

Published 2026-04-08
📖 6 min read🧠 Deep dive

Imagine you are a detective trying to solve a mystery about human behavior. The mystery is the Allais Paradox, a famous puzzle where people often make choices that seem to break the rules of "rational" math.

Specifically, there's a pattern called the Common Ratio Effect. It's like this:

  • Scenario A: You can take $100 for sure, or gamble for a chance at $150. Most people pick the sure $100.
  • Scenario B: Now, imagine the odds are cut in half. You can take $50 for sure, or gamble for a chance at $150. Suddenly, many people switch and pick the gamble.

Mathematically, if you liked the sure thing in the first scenario, you should like it in the second. But people flip-flop. This suggests our brains don't work like simple calculators.

For decades, economists have used Choice Tests (asking people to pick Option A or B) to prove this happens. But recently, a new group of researchers (MNOSS) argued: "Wait a minute! Maybe people aren't actually flipping their preferences. Maybe they just make random mistakes when they choose. If we account for those mistakes, the 'paradox' disappears."

To prove their point, MNOSS stopped asking people to choose and started asking them to value things (e.g., "What is the lowest amount of cash you would accept instead of this gamble?"). Using these "Valuation Tests," they claimed the paradox was an illusion and that people are actually rational after all.

Enter Echenique and Tserenjigmid (the authors of this paper). They are the detectives who say: "Hold on. You changed the tool, but you might have broken the tool you used to measure it."

Here is the breakdown of their argument using simple analogies:

1. The "Noisy Scale" Problem (Why the old tests were criticized)

Imagine you are weighing two bags of apples.

  • The Weak Test: You look at the scale. If Bag A is heavier than Bag B, you pick A.
  • The Criticism: MNOSS argued that if the scale is "noisy" (shaky), sometimes it might show Bag A is heavier, and sometimes Bag B, even if they are the same weight. They claimed that if you just look at the frequency of choices, the noise makes it look like people are flipping their preferences (the paradox), even if they aren't.

2. The "Valuation" Trap (Why the new tests are flawed)

MNOSS said, "Let's stop weighing the bags and just ask people, 'How much money is this bag worth to you?'"

  • The Authors' Counter: This is like asking someone to guess the weight of a bag of apples while they are drunk.
    • The "Mean" Problem: If you ask for an average value, the answer depends heavily on how "risk-averse" (scared of losing) the person is. It's like asking a tightrope walker and a trapeze artist how much a rope is worth; their answers will be wildly different based on their personality, not the rope's actual weight. The math shows you can get any answer you want just by tweaking the person's risk personality.
    • The "Sign" Problem: If you just ask, "Is it worth more than $10?", it only works if the person's "drunkenness" (random errors) is perfectly symmetrical. If their mistakes lean one way, the test is broken.

The Analogy: MNOSS tried to fix a shaky camera (the choice test) by switching to a blurry mirror (the valuation test). The authors argue the mirror is actually more distorted than the camera.

3. The "Strong" Solution (The Robust Test)

The authors propose a better way to look at the "noisy" choice data. Instead of asking, "Did they pick A more often than B?" (which is sensitive to noise), they suggest a Strong Test:

  • The Rule: If a person picks Option A more than 50% of the time, we assume they prefer A. If they pick it less than 50%, they prefer B.
  • Why it works: This is like looking at a crowd of people voting. If 60% vote for Candidate A, we know A is the winner. It doesn't matter if 10 people made a mistake and voted for B. The "majority rule" cuts through the noise.
  • The Result: The authors prove mathematically that this "Strong Test" is immune to the "shaky scale" problem. It works whether the noise is random, correlated, or weirdly distributed.

4. The Verdict: The Paradox is Real!

When the authors applied this "Strong Test" to the data from the previous studies (including the new data from MNOSS), the results were shocking:

  • MNOSS's Conclusion: "We found no evidence of the paradox."
  • Authors' Conclusion: "We found the paradox in 41% of the studies (and 10% in MNOSS's own data)."

The "Arbitrary Parameters" Twist:
The authors also found a sneaky trick in how the experiments were designed. The "Common Ratio Effect" only shows up if you pick very specific numbers for the money and probabilities (like $100 vs. $50). If you pick random numbers (like $12 vs. $30), the effect often disappears.

  • Analogy: It's like trying to find a specific type of fish. If you only cast your net in the exact spot where that fish lives, you'll find it. If you cast your net randomly in the whole ocean, you won't. MNOSS cast their net randomly and said, "No fish here!" The authors say, "You just didn't look in the right spots."

Summary for the Everyday Reader

  • The Conflict: Some researchers said the famous "Allais Paradox" (where people act irrationally) was just a statistical illusion caused by random mistakes. They used "Valuation" (asking for prices) to prove it.
  • The Defense: The authors say the Valuation method is actually more broken and biased than the old Choice method.
  • The Fix: They introduced a "Strong Test" (Majority Rule) that ignores the noise and looks at the clear preference.
  • The Outcome: When you use the Strong Test, the "irrational" behavior is still there. People really do flip-flop their choices in predictable ways. The Allais Paradox is real, and the "Valuation" method was a red herring.

In short: The authors saved the day by showing that the "new" method used to debunk the paradox was actually the one that was broken, and the "old" paradox is still very much alive and kicking.

Get papers like this in your inbox

Personalized daily or weekly digests matching your interests. Gists or technical summaries, in your language.

Try Digest →