Causal Survival Analysis in Platform Trials with Non-Concurrent Controls

This paper develops a causal survival framework for platform trials that demonstrates while pooling non-concurrent controls can improve precision under strict assumptions, the most robust approach to avoid bias and maintain efficiency is to target concurrent causal estimands using covariate-adjusted doubly robust estimators that rely solely on concurrent controls.

Antonio D'Alessandro, Samrachana Adhikari, Michele Santacatterina

Published Thu, 12 Ma
📖 5 min read🧠 Deep dive

Imagine you are running a massive, high-stakes cooking competition to find the best new recipe for a soup.

The Setup: The "Platform" Kitchen

In a traditional cooking show, you have two teams: Team A makes the old recipe, and Team B makes the new one. They cook side-by-side for the whole season.

But in a Platform Trial (like the one described in this paper), the kitchen is dynamic.

  • Team A (The Control): They are always there, making the "standard" soup. They are the shared reference point.
  • Team B, C, D (The Treatments): New teams enter the kitchen at different times. Team B arrives in January, Team C in March, and Team D in June.
  • The Twist: When Team B is cooking, they compare their soup to Team A. When Team C arrives, they also compare their soup to Team A.

This is efficient! You don't need a new "Team A" for every new recipe. You just keep the same standard team.

The Problem: The "Time Drift"

Here is the catch: The kitchen changes over time.

  • In January, the ingredients were fresh, the chefs were well-rested, and the weather was cold (people eat more soup).
  • In June, the ingredients are different, the chefs are tired, and it's hot (people eat less soup).

The paper calls this "Time Drift."

If you simply mix the data from Team B's January cooks with Team C's June cooks to compare against Team A, you might get a false result. Maybe Team C's soup looks "worse" not because the recipe is bad, but because it was summer and no one wanted soup.

  • Concurrent Controls: Team A cooks at the same time as Team B. (Fair comparison).
  • Non-Concurrent Controls (NCC): Team A cooked in January, but Team C is cooking in June. (Unfair comparison if you just mash the data together).

The Big Question

The researchers asked: Can we use the "old" Team A data (Non-Concurrent) to help us get a more precise answer for the "new" Team C, or will it mess things up?

Many statisticians wanted to say, "Yes! More data is better data!" They wanted to pool all the Team A cooks together to get a super-accurate baseline.

The Solution: A "Causal" Lens

The authors of this paper built a new mathematical framework to answer this. They used a concept called "Causal Survival Analysis."

Think of it like this: Instead of just looking at the average soup taste, they are trying to answer a specific "What If?" question:

"If a patient entered the trial today (concurrent), how long would they survive if they got the new treatment vs. the old treatment?"

They realized that to answer this specific question, you have to be very careful about which "Team A" cooks you use.

The Key Findings (The "Secret Sauce")

1. The "Double-Robust" Chef is the Safest Bet
The paper tested different ways to analyze the data. They found that the most reliable method is a technique called "Doubly Robust Estimation."

  • Analogy: Imagine you have two ways to guess the soup's quality:
    1. Taste Test (Outcome Regression): You guess based on the ingredients used.
    2. Chef's Reputation (Propensity): You guess based on who cooked it and when.
  • The Magic: The "Doubly Robust" method uses both. If your guess about the ingredients is slightly wrong, the "Chef's Reputation" part saves you. If your guess about the chef is wrong, the "Taste Test" saves you.
  • The Result: This method works best when you only use the "Concurrent" Team A cooks (those who cooked at the exact same time as the new treatment). It gives you a reliable answer without needing to risk mixing in the "old" data.

2. When Pooling Data Backfires
The paper showed that if you try to mix in the "Non-Concurrent" (old) Team A data:

  • If your math model is perfect: You might get a slightly more precise answer (a sharper picture).
  • If your math model is even slightly wrong (which happens often in real life): You introduce bias. You might conclude the new soup is great when it's actually terrible, or vice versa, just because you mixed in data from a different season.
  • The Verdict: It's like trying to fix a blurry photo by adding pixels from a completely different photo. If the lighting is different, you just make the picture worse.

3. The "Time Drift" is Real
The study confirmed that in real-world scenarios (like the COVID-19 trials they analyzed), the "Non-Concurrent" data often carries hidden biases. The "Concurrent" controls are the only ones that truly represent the current reality.

The Bottom Line for Everyone

If you are running a complex experiment where new treatments are added over time:

  1. Don't be greedy with data. Just because you have old control data doesn't mean you should use it.
  2. Focus on the "Now." Compare new treatments only against controls that existed at the same time.
  3. Use the "Double-Check" method. Use advanced statistical tools (Doubly Robust estimators) that protect you if your assumptions about the data are slightly off.

In short: The paper argues that in the race to find better treatments, accuracy is more important than speed. It is better to have a slightly less precise but correct answer using current data, than a very precise but wrong answer by mixing in old, irrelevant data.