Instrumental and Proximal Causal Inference with Gaussian Processes

This paper introduces a Deconditional Gaussian Process framework that unifies instrumental variable and proximal causal inference methods to provide reliable, well-calibrated epistemic uncertainty quantification and principled model selection while maintaining predictive precision.

Yuqi Zhang, Krikamol Muandet, Dino Sejdinovic, Edwin Fong, Siu Lun Chau

Published 2026-03-03
📖 4 min read☕ Coffee break read

Imagine you are a doctor trying to figure out if a new medicine actually cures a disease. You look at your patient records and see that people who took the medicine got better. But wait! Maybe those people were also healthier to begin with, or maybe they ate better food. You can't tell if the medicine worked or if it was just their healthy lifestyle. In statistics, this hidden factor (like lifestyle) is called a "confounder," and it makes it very hard to prove cause and effect.

This paper introduces a new, smarter way to solve this problem, especially when you can't see those hidden factors. The authors call their method GPIV and GPProxy.

Here is the breakdown using simple analogies:

1. The Problem: The "Hidden Puppeteer"

In many real-world situations (like economics or medicine), we can't run a perfect experiment where we control everything. We only have observational data.

  • The Confounder: Imagine a hidden puppeteer pulling strings on both the "Treatment" (the medicine) and the "Outcome" (getting better). If you just look at the data, you think the medicine caused the recovery, but really, the puppeteer did.
  • The Old Tools: Scientists have used two main tricks to find the puppeteer:
    • Instrumental Variables (IV): Using a "randomizer" (like a lottery for who gets the medicine) that isn't influenced by the puppeteer.
    • Proxies: Using "surrogate" clues (like a patient's mood or a side effect) that hint at what the puppeteer is doing.

2. The Flaw in Old Tools: "Guessing without a Safety Net"

The existing methods are good at giving you a single number (e.g., "The medicine improves health by 10%"). But they are terrible at telling you how sure they are.

  • It's like a weather forecaster saying, "It will rain tomorrow," but refusing to say if there's a 10% chance or a 90% chance.
  • If you are a doctor, you need to know: Is this 10% improvement a solid fact, or just a lucky guess? If it's a guess, you shouldn't prescribe the medicine to everyone.

3. The Solution: The "Gaussian Process" (The Flexible Rubber Sheet)

The authors propose using a Gaussian Process (GP).

  • The Analogy: Imagine a giant, stretchy rubber sheet stretched over a landscape of data points.
  • How it works: When you feed the data into this sheet, it doesn't just snap to a single line. It stretches and bends to fit the data, but it also "knows" how much it is stretching.
  • The Magic: The sheet gives you two things at once:
    1. The Prediction: The height of the sheet at any point (the estimated effect of the medicine).
    2. The Uncertainty: How wobbly or shaky the sheet is at that point. If the sheet is very wobbly, it means "I'm not sure here; I need more data." If it's flat and steady, it means "I'm very confident."

4. The Secret Sauce: "Deconditioning"

The paper uses a clever mathematical trick called Deconditioning.

  • The Analogy: Imagine you are trying to hear a whisper (the true effect) in a noisy room (the confounders).
  • The Trick: Instead of trying to shout over the noise, the authors use a "noise-canceling headphone" algorithm. They mathematically reverse the way the noise (the confounders) mixes with the signal.
  • The Result: Their method recovers the exact same "best guess" as the old, popular methods (so it's just as accurate), but it also keeps the "wobble" information (the uncertainty) that the old methods threw away.

5. Why This Matters: "Knowing When to Say 'I Don't Know'"

The paper shows that their new method is a game-changer for two reasons:

  • Better Decisions: Because it tells you how confident it is, you can make safer decisions. If the "wobble" is high, you can choose not to make a decision yet (like not prescribing a risky drug) until you have more evidence. This is called "selective prediction."
  • Self-Correction: The method can automatically tune itself to find the best settings without needing a human to guess, making it easier to use.

Summary

Think of the old methods as a crystal ball that gives you a single, blurry number and says, "Trust me."
The new method is like a smart GPS that gives you the route and tells you, "I'm 95% sure this is the right way, but this next turn is foggy, so drive carefully."

By combining the accuracy of old tools with the "confidence meter" of modern AI, this paper provides a safer, more reliable way to figure out cause and effect in a messy, confusing world.

Get papers like this in your inbox

Personalized daily or weekly digests matching your interests. Gists or technical summaries, in your language.

Try Digest →