A Bayesian adaptive enrichment design using aggregate historical data to inform individualized treatment recommendations

This paper proposes a Bayesian adaptive enrichment design that leverages aggregate historical data via a normalized power prior to inform individualized treatment recommendations, demonstrating through simulations and a motivating obstructive sleep apnea trial that this approach improves statistical power and efficiency compared to non-borrowing designs.

Lara Maleyeff, Shirin Golchi, Erica E. M. Moodie

Published Wed, 11 Ma
📖 6 min read🧠 Deep dive

Imagine you are a chef trying to create the perfect recipe for a new dish. You want to know exactly which ingredients work best for which type of person. Maybe spicy food makes some people happy, but gives others a stomach ache.

In the world of medicine, this is called Precision Medicine. Doctors want to know: "Does this drug work for you specifically, based on your unique biology?"

However, running a clinical trial to test this is expensive, slow, and difficult. You can't test every single person in the world. Usually, scientists run a big trial on a mixed group of people and just look at the "average" result. But as the paper explains, the average often hides the truth. The drug might work wonders for 50% of people and do nothing for the other 50%, but the average looks like "meh, it's okay."

Here is the problem: To find the "50% who benefit," you need a lot of data. But you don't always have time or money to collect it all from scratch.

The Solution: Borrowing Wisdom from the Past

The authors of this paper propose a clever way to speed things up. They suggest borrowing information from old studies (historical data) to help design the new trial.

Think of it like this:

  • The Old Way (No Borrowing): You are building a new house from scratch. You buy every single brick, nail, and beam yourself, even though there are warehouses full of perfectly good bricks from old houses nearby. It's safe, but slow and expensive.
  • The New Way (Adaptive Enrichment with Borrowing): You look at the blueprints and materials from those old houses. If the old house was built with similar materials and in a similar climate, you use those bricks to speed up your construction. But, you keep a safety inspector (the math) to make sure those old bricks aren't rotten or from a completely different style of house that would make your new house collapse.

The "Magic" Tool: The Normalized Power Prior

The paper introduces a specific mathematical tool called a Normalized Power Prior (NPP). Let's break down what that means in plain English:

  1. The Summary Problem: Often, old studies don't give you the raw data (like a spreadsheet of every patient). They only give you a summary, like "The average blood pressure dropped by 5 points."
  2. The Mismatch: Your new trial is looking for something specific, like "Did blood pressure drop for people with high heart rates?" The old study didn't tell you that.
  3. The Bridge: The authors built a mathematical bridge. They take that old "average" summary and translate it into a guess about your specific group.
    • Analogy: Imagine the old study says, "The average height of people in this city is 5'10"." Your new study is looking for the average height of basketball players in that city. The math takes the city average and, based on what you know about basketball players, estimates what the basketball players' height might be, without needing to measure every single player in the city first.

How the Trial Changes While It Runs (Adaptive Enrichment)

This isn't just about using old data at the start. The trial changes its mind as it goes along.

Imagine a detective investigating a crime.

  • Start: The detective questions everyone in the neighborhood (broad recruitment).
  • Mid-Trial (Interim Analysis): After talking to 200 people, the detective notices a pattern: "Wait, all the clues point to people who live on the North Side."
  • The Switch: Instead of wasting time questioning the South Side, the detective stops there and focuses all remaining resources on the North Side.

In the paper's medical trial:

  1. They start with all patients.
  2. They check the data halfway through.
  3. If the data shows the drug is working great for a specific group (e.g., people with high "hypoxic burden" in sleep apnea), they stop recruiting people who don't fit that profile.
  4. They focus only on the people who are likely to benefit. This saves money, time, and spares people who won't benefit from taking a useless drug.

The Safety Net: What if the Old Data is Wrong?

The biggest fear is: "What if the old study was wrong or biased?"

The authors' method has a built-in safety valve.

  • If the new data looks very different from the old data (e.g., the old study said the drug works, but your new trial shows it does nothing), the math automatically turns down the volume on the old data. It says, "Okay, we'll ignore the old study and just trust our new data."
  • If the new data matches the old data, the math turns up the volume, letting the old study help confirm the results.

The Results: Why This Matters

The authors ran thousands of computer simulations to test this idea. Here is what they found:

  • Faster: They could stop the trial earlier because they were more confident in the results.
  • Smaller: They needed fewer patients to get a clear answer.
  • Smarter: They were better at finding the specific group of people who actually needed the medicine.
  • Safe: Even with the "borrowing," they didn't accidentally declare a fake cure (they kept the error rate low).

The Real-World Example: Sleep Apnea

To prove it works, they applied this to a real-world problem: Obstructive Sleep Apnea (OSA).

  • Some people with sleep apnea have high "hypoxic burden" (their bodies struggle to get oxygen).
  • Some have low burden.
  • Old studies showed that CPAP machines (the masks people wear to sleep) didn't help everyone.
  • Using their new method, they showed that if you borrow data from past studies, you can quickly identify that the CPAP machine works amazingly for the high-burden group, but is useless for the low-burden group. This helps doctors prescribe the right treatment to the right person much faster.

The Bottom Line

This paper is about being efficient and smart in medical research. Instead of reinventing the wheel for every new drug trial, we can use the wheels from previous trials to get to the finish line faster. But, we use a special "suspension system" (the math) to make sure that if the old wheels are broken, we don't crash.

It's a way to bring history into future medicine, ensuring that treatments are tailored to the individual, not just the average.