Data Science x Public Health
This podcast discusses the concepts of data science and public health, and then delves into their intersection, exploring the connection between the two fields in greater detail.
Data Science x Public Health
Bayesian Borrowing Explained: The FDA’s 2026 Clinical Trial Shift
Use Left/Right to seek, Home/End to jump to start or end. Hold shift to jump forward or backward.
Clinical trials traditionally rely only on data from newly enrolled patients. But Bayesian borrowing allows researchers to incorporate external data from past studies to strengthen new trials.
In January 2026, the FDA released draft guidance expanding how sponsors can use external controls and Bayesian methods in clinical trials. This episode explains how Bayesian borrowing works, why it can make trials faster and smaller, and the risks regulators are trying to control.
👉 Enjoyed the episode? Follow the show to get new episodes automatically.
If you found the content helpful, consider leaving a rating or review — it helps support the podcast.
For business and sponsorship inquiries, email us at:
📧 contact@bjanalytics.com
Youtube: https://www.youtube.com/@BJANALYTICS
Instagram: https://www.instagram.com/bjanalyticsconsulting/
Twitter/X: https://x.com/BJANALYTICS
Welcome to today's deep dive. Today we are looking at the FDA's January 2026 draft guidance and a recent February 2026 analysis on Bayesian borrowing.
SPEAKER_00Yeah, Bayesian borrowing. It's quite a heavy term, but honestly so important right now.
SPEAKER_01Exactly. And the mission for you today, as you're listening, is to understand how reusing historical patient data could completely revolutionize clinical trials. I mean, making them smaller, faster, and sparing patients from unnecessary placebos.
SPEAKER_00Which is a huge deal for the patients, obviously.
SPEAKER_01Right. So to really wrap our heads around this, imagine building a brand new house. But instead of spending months, you know, digging a massive hole and pouring fresh concrete, you get to build right on top of a perfectly good foundation left over from a past project.
SPEAKER_00Oh, I like that analogy. You save immense time and money that way.
SPEAKER_01But well, that only works if the old foundation actually matches your new floor plan.
SPEAKER_00And if it doesn't match, well, the entire structure is compromised. So that is the fundamental challenge statisticians face right now. We have mountains of historical clinical data, sure. But safely merging that old data with incoming trial data, it requires a highly specific mathematical engine. Basically, we are combining a prior distribution, which is the existing knowledge, with new trial data.
SPEAKER_01And that generates the updated posterior distribution.
SPEAKER_00Exactly. You've got it.
SPEAKER_01So the guidance outlines two ways to build that posterior, static and dynamic borrowing. And static borrowing feels a bit like, I don't know, locking into a fixed-rate mortgage while totally ignoring the current housing market.
SPEAKER_00Yeah, that is a very fair way to put it. You decide ahead of time that, say, each historical patient is worth exactly half of a new patient.
SPEAKER_01Right. And you never deviate from that weight, regardless of what the new data actually shows, which seems incredibly dangerous to me. Wait, isn't locking in a fixed rate a huge risk if the old data turns out to be wildly divergent from our new patients?
SPEAKER_00Aaron Powell It is a massive vulnerability. You'd be blindly dragging your results toward a false conclusion. And that is exactly why the FDA mandates rigorous simulations before trial even begins.
SPEAKER_01Wow, so they have to prove it beforehand.
SPEAKER_00Yeah, sponsors have to computationally stress test their mathematical models across thousands of worst-case scenarios. They essentially run virtual trials where the drug actually fails.
SPEAKER_01Oh, just to prove their algorithm won't trigger a type I error under pressure. That makes sense.
SPEAKER_00Type I error is that catastrophic risk of falsely concluding a treatment works when it doesn't. Because static borrowing is so rigid, the industry is really pivoting toward dynamic borrowing.
SPEAKER_01Which sounds much smarter, but mechanically, how does it actually adjust? Like you hear terms like the modified power prior thrown around a lot.
SPEAKER_00Right. So think of the modified power prior as an algorithmic trading bot. It constantly compares the variance between the historical data curve and the incoming new data curve. If those data sets overlap cleanly, meaning the old patients and new patients are responding identically, the algorithm automatically dials the borrowing weight up.
SPEAKER_01Sometimes to a full 100% right.
SPEAKER_00Yeah, exactly. But if the curves start to diverge, it mathematically throttles the historical data's weight down.
SPEAKER_01And if they clash completely.
SPEAKER_00It drops the borrowing straight to zero.
SPEAKER_01So because the math acts as its own fail-safe, the FDA seems pretty willing to entertain this in scenarios where running a massive traditional trial is basically impossible. Pediatric extrapolation is a great example of this.
SPEAKER_00Yes, absolutely. Using adult efficacy data to get a children's drug approved without making kids go through a redundant trial.
SPEAKER_01Or building synthetic comparators for severe rare diseases so you don't have to randomize desperate patients to a placebo.
SPEAKER_00Those are definitely the absolute ideal use cases. But relying heavily on historical data introduces a major structural flaw. If we pull patient records from a trial conducted, say 10 years ago, the baseline standard of care was likely very different.
SPEAKER_01Right. The disease progression might have been worse, or the diagnostic criteria might have been looser.
SPEAKER_00Exactly.
SPEAKER_01So that completely skews the baseline. The old foundation literally doesn't match the new floor plan anymore.
SPEAKER_00Aaron Powell And in statistics, we call that prior data conflict. When you blend mismatched historical data with new data, the algorithm doesn't just introduce bias. It actually creates something called inflated confidence.
SPEAKER_01Inflated confidence. What does that mean exactly in this context?
SPEAKER_00Aaron Powell Well, the whole point of borrowing data is to reduce uncertainty. But if the borrowed data is fundamentally flawed, you just end up generating incredibly narrow, precise confidence intervals around a completely wrong answer.
SPEAKER_01Aaron Powell Oh, wow. Being precisely wrong instead of vaguely right.
SPEAKER_00Aaron Powell Exactly.
SPEAKER_01Aaron Powell So the FDA must be scrutinizing the inputs mercilessly, then. Like if a sponsor tries to construct an external control group from old databases, what kind of pushback are they seeing?
SPEAKER_00Aaron Powell The February 2026 analysis really highlights this. The agency's tolerance is narrowing drastically right now. They are demanding granular proof that the historical data was collected under nearly identical clinical conditions.
SPEAKER_01Meaning perfectly comparable endpoints in patient population. Well, for you listening, this delicate balance of speed versus statistical rigor directly impacts how quickly life-saving treatments reach you.
SPEAKER_00Aaron Powell It really does. It's a quiet revolution in biostatistics that respects the data we already have.
SPEAKER_01Right. So the next time you hear about a breakthrough drug for a rare disease getting approved in record time, you now know the hidden dynamic math that likely helped fast track it leaves us with a pretty provocative question to ponder.
SPEAKER_00Oh, what's that?
SPEAKER_01Well, if dynamic borrowing algorithms become sophisticated enough to perfectly synthesize vast global databases of historical medical records, will the traditional standalone clinical trial eventually become obsolete?
SPEAKER_00Wow, that is a fascinating thought.
SPEAKER_01Right. I mean, if we have enough perfect foundations already poured out there, maybe one day we'll never have to dig a new one from scratch again.