Assessing Sensitivity to IV Exclusion and Exogeneity without First Stage Monotonicity

This paper introduces new sensitivity analysis methods for instrumental variable studies that derive identified sets for treatment effects under relaxed exclusion and exogeneity assumptions without requiring first-stage monotonicity, utilizing computationally tractable linear programming techniques demonstrated through an empirical application to peer effects in movie viewership.

Paul Diegert, Matthew A. Masten, Alexandre Poirier

Published 2026-04-10
📖 5 min read🧠 Deep dive

Imagine you are a detective trying to solve a mystery: Does watching a movie with friends actually make you want to watch it again later?

To solve this, you can't just ask people what they did, because maybe they just really liked the movie to begin with. You need a clever trick. You decide to use the weather as your clue.

  • The Logic: If it's a beautiful, sunny Saturday, people stay outside and don't go to the movies. If it's rainy, they stay inside and do go.
  • The Assumption: You assume the weather is a "random" event that only affects whether people go to the movies, and nothing else. It doesn't change the quality of the movie, and it doesn't change how much people like talking about it.

This is the standard way economists use Instrumental Variables (IV). But here's the problem: What if your clue is flawed?

Maybe on sunny days, people also go to outdoor festivals where they talk about movies, which changes their opinion. Or maybe movie studios release better movies on sunny days to compete with the weather. If the weather affects the outcome directly (not just through the movie ticket sales), your whole investigation is ruined.

The Paper's Big Idea: "What If We're Wrong?"

This paper by Diegert, Masten, and Poirier is like a stress test for your detective work.

Most previous methods for checking these clues relied on a strict rule called "Monotonicity." Think of this as a rule saying: "The weather must always push people in the same direction. It can never make a sunny day more likely to result in a movie night for some people and less likely for others."

The authors say: "That's too strict! Real life is messy."

They developed a new toolkit that says: *"We don't need to assume the weather always pushes people the same way. We just want to know: How much can our clue be slightly wrong before our whole conclusion falls apart?"*

The New Toolkit: The "Sensitivity Slider"

Imagine you have a slider on a control panel labeled "How much do we trust the weather?"

  1. Slider at 0 (Perfect Trust): You assume the weather is a perfect, random clue. The paper calculates the answer based on this.
  2. Slider at 1 (No Trust): You assume the weather has nothing to do with the outcome. The answer becomes a huge, useless range (like "The effect is between -100% and +100%").
  3. The Middle Ground: The authors created a way to slide the knob to 0.1, 0.2, 0.5, etc. At each step, they ask: "If the weather is slightly 'contaminated' (e.g., it affects the movie choice by 5% more than we thought), what happens to our answer?"

They use a mathematical method called Linear Programming (think of it as a super-fast, high-tech calculator) to find the "best case" and "worst case" scenarios for every position of that slider.

The Real-World Test: Movie Tickets

The authors tested their new toolkit on a famous study about movie peer effects.

  • The Original Study: Found that if a movie does poorly on its opening weekend (because of bad weather), it actually does better later because people talk about it. They concluded: "Yes, peer pressure works!"
  • The New Stress Test: The authors took that study and ran their "Sensitivity Slider."

The Shocking Result:
When they assumed the weather was perfect, the result held up. But the moment they allowed for a tiny amount of doubt (just a 1.5% chance that the weather was influencing things in a weird way), the result collapsed.

The "positive effect" disappeared. The answer became: "We don't know if peer pressure works or not; the data is consistent with zero effect."

Why This Matters

Think of this paper as a safety net for science.

In the past, if a study said "X causes Y," we often had to take it on faith that their assumptions were perfect. If the assumptions were slightly wrong, the whole conclusion could be a lie, and nobody would know until years later.

This paper gives researchers a way to say:

"Here is our conclusion. But look at this graph. If our assumption is wrong by just a tiny bit, the conclusion changes. If it's wrong by a medium bit, the conclusion vanishes. So, you should be very careful trusting this result."

The Takeaway

This paper doesn't just give you an answer; it gives you a map of uncertainty.

It tells us that in the world of data, perfection is rare. Instead of pretending our clues are perfect, we should measure exactly how fragile our conclusions are. If a conclusion breaks with the slightest nudge, it's not a solid fact—it's a house of cards.

In short: The authors built a machine that tells you how much "wiggle room" you have in your assumptions before your scientific discovery turns into a guess. And in the case of movie peer effects, it turns out the wiggle room was almost non-existent.

Get papers like this in your inbox

Personalized daily or weekly digests matching your interests. Gists or technical summaries, in your language.

Try Digest →