The Epstein Files
The Epstein Files is the first AI-native documentary podcast to systematically analyze the Jeffrey Epstein case at scale. With over 3 million pages of DOJ documents, court records, flight logs, and public resources now available, traditional journalism simply cannot process this volume of information. AI can.
This series leverages artificial intelligence at every layer of production. From custom-built architecture that ingests and cross-references millions of pages of evidence, to AI-generated audio that delivers findings in a consistent, accessible format, this project represents a new model for investigative journalism. What would take a newsroom years to analyze, AI can process in days, surfacing connections, patterns, and details that would otherwise remain buried in the sheer volume of data.
Each episode draws directly from primary sources: unsealed court documents, FBI files, the black book, flight logs, victim depositions, and the DOJ's ongoing document releases. The AI architecture identifies relevant passages, cross-references names and dates across thousands of files, and synthesizes findings into episodes that make this information digestible for the public.
The series covers Epstein's mysterious rise to wealth, his network of enablers, the properties where crimes occurred, the 2008 sweetheart deal, his death in federal custody, the Maxwell trial, and the unanswered questions that remain.
This is not sensationalized content. It is documented fact, processed at scale, and presented with journalistic rigor. The goal is simple: make the public record accessible to the public.
New episodes release as additional documents become available, with AI enabling rapid analysis and production that keeps pace with ongoing revelations. Our Standards AI enables scale, but journalistic standards guide the output. Every claim is tied to specific documents. The series clearly distinguishes between proven facts and allegations. Victim testimony is handled with dignity. Names that appear in documents are not accused of wrongdoing unless documents support such claims.
This is documented fact, processed at scale, presented for the public.
The Epstein Files
File 44 - 3.5 Million Pages the DOJ Released on Epstein
Use Left/Right to seek, Home/End to jump to start or end. Hold shift to jump forward or backward.
The Epstein Files Transparency Act compelled the release of thousands of previously classified government documents related to the case. FBI interview summaries, DOJ correspondence, and intelligence agency records emerged for the first time. This episode examines the most significant revelations and the files that remain sealed.
Sources for this episode are available at: https://epsteinfiles.fm/?episode=ep44
About The Epstein Files
The Epstein Files is an AI-generated podcast analyzing the 3.5 million pages released under the Epstein Files Transparency Act (EFTA). All claims are grounded in primary source documents.
Produced by Island Investigation
3 million pages of evidence. Thousands of unsealed flight logs. Millions of data points, names, themes and timelines connected. You are listening to the Epstein Files, the world's first AI native investigation into the case that traditional journalism simply could not handle. Welcome back to the Epstein Files. Last time we walked through ISBE versus JP Morgan. Today, we are following the 2024 document releases through the documentary record, so that timeline decisions and institutional failures are clear. As always, every document and source we reference is available at epsteinfiles fm. So start with court unsealing process. That is where the paper trail becomes specific and testable. We need to set the parameters immediately. This is not a speculative discussion. This is a forensic audit of the 2024 document releases. We are analyzing the metadata, the file integrity and the chain of custody regarding the Department of Justice's release pattern. So we're looking at the digital footprint, the bureaucracy itself. Exactly. Not just what was released, but how. And more importantly, what was retracted after the fact. Agreed. The constraint for today is strict. What do the documents actually prove? What remains unproven? And where are the specific institutional decision points? Let's start with the release mechanism. Okay, so this wasn't standard document dump. It wasn't, you know, just a box being overturned. We have to look at the pinpoint database of U.S. justice Department's Epstein files, which is managed by the Courier Newsroom. And that's a critical detail. It's the whole ball game. The Courier newsroom archive, it acts as our control group. In any forensic review, you need a baseline, something to measure against. Without their archive, we'd only have the DOJ's current. Let's call the official version of the repository Sanitized version. That's one word for it. The record shows a very specific pattern that Courier identified and crucially captured. They call it a pattern of release and delete. Okay, let's break that down. Release and delete. You're talking about the sequence of events recorded in August of 2024? I am. The source documentation confirms it. The U.S. justice Department began uploading documents and multimedia files. These were official uploads to their public facing server. Not leaks? No, absolutely not. They appeared on the server live, accessible, and then shortly after they were removed, deleted. But not before they were archived. Right. Courier Newsroom retained the files during that. That window of availability, which lets us perform a delta analysis, a comparison. We can see what the DOJ initially released, or maybe accidentally made public, and compare it against what they decided retroactively should not be public. We have specific serial numbers for these Files, too. I'm looking at a document here. EFTA 023-35898. PDF. And another one, EFTA 027-30741. PDF. And look at the acquisition date listed in The Pinpoint database. 2024. 08. 14. August 14th. August 14th, 2024. That date is. Well, it's very significant. It's well past the official closing of the case, which is a whole other decision point we need to get to. We will, but just looking at those serials for a second of eight, fda. That prefix usually denotes a specific agency archives a task force. The fact that these files exist in the courier archive, but are absent from the current DOJ site, that's what we call a documented discrepancy. It's a mismatch between the master file and the public record. It is a testable, provable mismatch. And the content of these deleted files isn't random. The source documentation notes that the items that were pulled back specifically connected Epstein to President Donald Trump. And we have to state that very neutrally, exactly as it appears in the record. The files that were selected for removal contained those specific keywords and associations. Now, the logs don't state why, right? It could be a legal privilege review, a privacy redaction, an error. It could be a simple error. We don't have documentation for the motive. However, the selection of the material for deletion, that's a data point in itself. It suggests the filter is active. It's active and it's reactive. This isn't a static release of historical records from a dusty box. It's an ongoing curation process happening in 2024. Which really changes the whole nature of this unsealing process. It implies administrative friction. If the mandate was total, simple transparency, the files would remain online. The removal indicates that someone somewhere inside the DOJ flagged these specific serial numbers for retraction after the initial upload. And that's the decision point. That is the decision point. We are tracking who made that call and based on what criteria. The documents don't tell us that, but they prove the call was made. Okay, so let's move to the structure of the data that did remain. The data that was cataloged. The pinpoint database organizes all of this into what they call data data sets. And this is where we start to see the scale of what might be withheld. The records. They verify the existence of data sets one through eight. Okay, I see that in the log, one through eight are accounted for. And we also have verification for data set 12. So it jumps. It jumps. Which immediately forces the question about the sequence gap. We do not have documentation for data sets 9, 10, and 11. I said they're just unaccounted for. Completely missing from the upload logs as of the audit date. We have January 31, 2025. Yes, they are missing from the sequence. And in any kind of forensic accounting, or federal discovery for that matter, a gap in sequential numbering, that's a major red flag. You don't just skip numbers. You don't skip numbers. Not in invoices, not in check numbers, and certainly not in evidence batches. It represents a volume of information that was either processed and then withheld entirely or is processed to move to the different and non public archive. And we have no idea the size of those missing sets. We don't. But if we assume they're comparable in size to, say, data sets one through eight, we could be talking about thousands of pages. A substantial void. A substantial void in the evidentiary sequence. It's a black box sitting right in the middle of the timeline. Let's look inside the boxes. We do have the metadata tags in the pinpoint database. They give us a kind of architecture of Epstein's digital life. I'm seeing entity tags for the Federal Bureau of Investigation, obviously, Apple, Bradford Limited. Bradford limited is a key one. It's a corporate entity that shows up again and again in the financial webs. But I want to focus on the hardware and software tags. For me, that's where we see the real age of this evidence. I see Compaq and another one, book 4U. Compaq, right there, that's a hardware tag. That single word dates the origin of the forensic image. We're looking at Compaq as a major independent company, effectively stopped existing after the HP merger, which was what, early 2000s? 2002. Exactly. So when you see a compact tag in a 2024 document release, it confirms you're looking at digital forensics from the very earliest era of his operations, the late 90s, the very early 2000s. This isn't data from the cloud. This is data pulled from a physical hard drive, physical media that was seized, imaged, and is only now being processed for public release decades later. And what about book 4U? It's likely a proprietary software directory or maybe a legacy contact management system, something custom. It indicates the specific digital tools Epstein was using to manage his network. This level of metadata is unusual. It's very rare. Usually we just get a PDF of a scanned printed document to see the system tag's compact book. For you, that tells us the DOJ is processing the native files, the original digital material, which means they have the original directory structures, the folder names, everything. They have the machine state. They. They know not just what was written, but when it was created, what machine it was created on, using what software they possess. The entire digital environment, not just the text within it and the locations tagged in the metadata. United States, of course, but also Austin Cranston. Right. And these geographic tags, they correlate to flight patterns and property records we've seen in other document tranches. But their appearance here in the digital file metadata, that's different. It links the physical location to the digital act. It anchors it. It means someone was sitting at a computer in Austin or in Cranston. Creating these files on an old compact machine. It connects the digital evidence to the physical world in a way a simple document can't. Okay, let's shift to evidence block 3. The names, the headline grabbing part of these releases, the 2024 releases, combined with the earlier 2015 tranches, the. They confirm a lot of the individuals we already knew about. The established list. Yes. Prince Andrew, Alan Dershowitz, Bill Clinton. Those names are confirmed in multiple sources. They're part of the broad documentary record. Now, the headline names, they are, but for this audit, we need to look past them, at the new names that are appearing in the 2024 metadata. I see them here. Alfredo Rodriguez and Carlos Perez. Exactly. These names are showing up in the pinpoint database metadata. These are not headline names. These are. Well, they're likely support staff, pilots, household facilitators, could be drivers, schedulers, IT maintenance. And their significance is enormous because an operation like this requires logistics. It's pure logistics. A trafficking network is an infrastructure project. The appearance of names like Rodriguez and Perez in the metadata suggests the core investigation files contain testimony or employment records from that operational tier, not just the guests on the plane. The infrastructure of the crimes. Yes. And that's often where the most damning, verifiable evidence is found, in the mundane records of the people who made the machine run. And yet, while we're getting these new lower level names unsealed, we still have pseudonyms like Jane Doe 3 appearing. And this is the core tension in all of these unsealing orders. The public narrative is one of transparency, of naming names, but the legal reality is much more complex. The protective orders are still active. For victims? For certain witnesses, absolutely. Jane Doe 3 appearing in the 2024 metadata confirms that the unsealing is not a blanket process. I'm looking at another document serial here. EFT A02730486.PDF. The status on this one is listed as Access denied you. Access denied you. Now, the U in government markings typically stands for unclassified. So this is a specific category of withholding. It's not a national security issue. No, it's not a classification or clearance issue. It's a privacy hold or a relevance objection. Something is being shielded for reasons other than state secrets. And we actually have the justification text in another document. EFTA 023-35898. PDF. It's quoted right here. What does it say? It says, production not necessary. Documents are not relevant or you document does not exhibit the mental impressions of. Stop there. Mental impressions. That's the key. That is work product. Legal terminology. It's a very specific legal doctrine. So this suggests that some of these withheld documents, they're attorney's notes or internal prosecution strategy memos, deliberations, the thought process of the case itself. So they aren't direct evidence of the original crime. They're evidence of how the case was built or not built. Precisely. The government is shielding the internal deliberations of its own prosecutors. Or perhaps the defense. The phrase production not necessary is an administrative decision, not a legal one handed down by a judge. It means someone within the DOJ has decided that the public's interest in seeing that specific page does not outweigh the institutional privilege of the author. But that fundamentally contradicts the public narrative of unsealing, doesn't it? If the headline is Epstein Files unsealed, but the audit trail shows potentially thousands of pages are being held back as not relevant or as mental impressions, then what we are seeing is not the full picture. We are seeing a curated selection from the archive. The unsealing is the legal procedure, but the redaction and withholding, that is the administrative reality. And the audit trail proves significant Content remains shielded. Let's reconstruct the timeline. Now, to understand how these 2024 digital decisions fit, we have to look at how the physical evidence has been handled. Starting from the moment of Epstein's death. We have to go back to 2019 and early 2020. There's contemporaneous reporting from January 17, 2020, that details the immediate aftermath. This is where the dispute over the pathology comes from. Yes. Dr. Michael Baden. He was the forensic pathologist hired by the family to observe the autopsy. His testimony is what created the conflict. He cited the brust capillaries in the eyes, petechial hemorrhaging. He claimed on the record that this finding was more consistent with homicidal strangulation than with a drop hanging suicide, which is the official cause. Day one. We have a fork in the road from day one. The documents show conflicting pathology reports. We have the official ruling of suicide from the city's medical examiner. And we have a contemporaneous contradictory claim from an expert witness who was physically present at the autopsy. It's a fundamental disagreement on the core fact of the event. And what about the physical evidence from the jail itself? The surveillance cameras? This is the critical custodial failure. The reporting from that same period confirms it. And it's never been credibly refuted. Surveillance video from Epstein's first suicide attempt erased. And to be clear, this was the video from the attempt that happened before his death. Correct. The incident a few weeks prior where he was found semi conscious in his cell. The video of that event which would have shown who entered or left his cell area around that time was reported as erased. And the explanation was a clerical error, a technical glitch. The justifications have varied, but the result is the same. The evidence is gone. Now connect that specific event in 2019 to our findings about the 2024 document release at the Continuum of Behavior. In 2019, we had the confirmed erasure of custodial video evidence from a high security federal facility. A physical asset. In 2024, we had the confirmed release and delete pattern of digital documents by the Department of Justice. Which was caught by Courier newsroom. A digital asset. It's the same pattern, just different media. It establishes a repeatable pattern of data loss regarding key custodial evidence. Whether it's a videotape from the MCC or a PDF of a flight manifest on a DOJ server. The documented pattern is that the government possesses data that subsequently vanishes from the public record. The timeline then shifts. In July 2020, right. Ghislaine Maxwell is arrested in New Hampshire after being in hiding for about a year. And this marks a critical transition in the entire case. The investigation pivots. It is no longer the Epstein case. It is now the Maxwell prosecution. And that allows for a compartmentalization of the evidence. Of course, once Epstein is dead and Maxwell becomes the sole prosecutorial target, Evidence is filtered through a new lens. Is it relevant to convicting her? Evidence that was relevant only to him or to his vast network of other associates that can be deprioritized. He can be deprioritized. Or as we see with those production not necessary tactics tags, it can be formally deemed irrelevant to the act of prosecution. The focus narrows and a lot of material gets left on the cutting room floor. Let's talk about the decision to close the case entirely. There's reporting from June 28, 2023. Yes, the headline was Definitive Gov closes case on Epstein. So that's mid 2023. The book is closed. Officially June 2023. That's the public statement that the pinpoint database, the one we've been auditing, shows the acquisition of new document serials In August of 2024, a full year later. This is inconsistent, say the least. If the case is closed, why is the system still processing new serials? Why is it acquiring documents? The metadata is unambiguous. Acquired on 2024. 0814. Now, there are a couple of possible explanations. One is that these are archival processing dates, meaning they're just now getting around to scanning and logging old files for the unsealing process. Okay, so the other explanation is that the closure of the case was a public relations or administrative status, and it did not actually stop the internal movement and processing of data related to the investigation. So the case might be closed to the public, but it's not closed inside the DOJ servers. The government is still actively handling this material, generating new acquisition tags in its own database. More than a year after telling the public that the book was shut. It's a direct contradiction. Now, let's look at another institutional decision point, this one outside of law enforcement. The institutional funding. We have to look at the MIT Media Lab. This is a critical piece of the puzzle because it demonstrates the complicity, the broader institutional failure. The reporting details MIT's decision to continue accepting funding from Epstein and the justification they offered. The Founder claimed that taking the money was justified. And this is the crucial part. This was happening after Epstein's 2008 conviction in Florida. This wasn't before he was a known sex offender. This was after. So we have a timeline where federal video evidence is being erased. Yes. The DOJ is actively deleting its own released files in 2024. Correct. And one of the world's premier scientific institutions is simultaneously justifying taking millions of dollars from him, knowing his status as a convicted offender. It places the DOJ's failures in a much broader context. The rot wasn't just in the jail or in the prosecutor's office. The financial permeability of these elite institutions is what allowed Epstein to operate with impunity for so long. The audit of institutional failures have to include entities like mit. Because the money was accepted, knowing the source. It was a conscious decision. Okay, let's Move to evidence block 6. The unresolved gaps, the things we know are missing from the record. We have to address the specific claims of missing evidence. There's an interview from January 12, 2024. It's with the Jeffrey Epstein's brother, Mark Epstein. Right. And in that interview, Mark Epstein lists very specific pieces of evidence that he claims the feds are covering up specifically related to the death scene. He talks about photos, a hyoid bone. He's very specific. He is. Now, we have to be extremely precise here in our audit. We don't have documentation for the specific inventory that Mark Estein lists. We cannot independently verify his claims about those items. But his central claim, that there is a cover up is happening in a context that we can document. His claim is corroborated by the proven erasure of the jailhouse surveillance video and the proven existence of conflicting autopsy notes from Dr. Baden. So while we can't verify his specific list of missing items, the environment of his claim is supported by the forensic record we've already established. Exactly. His claim is not being made in a vacuum. It is being made in a documented environment of systemic data loss and conflicting official records that gives his claim only a weight it would not otherwise have. And that brings us to the flight logs. The flight logs. There's a source from August 12, 2019, just after Epstein's death, that frames the question perfectly. The topic was the logs, but the specific question was, what vanished? What vanished? And that's still the question. The logs themselves exist, at least in part. We have lists of names. We know many of the powerful people who flew on his planes. We know who flew. We still don't know why. That's it. The context is what vanished. The 2024 release gives us metadata. It gives us these fascinating tags like book4u and compaq, but it does not provide the investigative narratives, the FBI 302s, the witness interviews, that would explain the purpose of these flights. We have thousands of data points, but the Department of Justice has not released the connective tissue that would turn that data into a story. So let's synthesize this. Let's bring the entire audit together. We've walked through the documentary record. Wow. We have a clear timeline. It starts with erased video evidence from the jail in 2019 and 2020. A key piece of physical evidence is gone. Gone with an official explanation that doesn't hold up to scrutiny. Then in mid 2023, we have a public declaration from the government that the case is officially closed. A line is drawn in the sand. And then a year later, in 2024, we see this subsequent erratic release of documents. A process where the Department of Justice itself is caught actively deleting files that it just made public. The pattern is not one of transparency. It's a pattern of what can only be described as administrative friction. Friction? Yes. Resistance and institutional resistance to full disclosure. The documents prove without a doubt that the DOJ possesses more data than it retains in the public record. We know this for a fact because Courier newsroom caught them deleting it. We also know it because of the gaps. We know it because data sets 9, 10 and 11 are missing from the official log. A huge unexplained hole in the middle of the evidence. So this whole unsealing process is what, a performance. It's a selective process. That's the most neutral way to put it. It is not a total dump of the archive. It is a highly curated release, subject to production, not necessary tags, access denied flags, and as we've seen, active deletion of politically sensitive connections after the fact. The documents do confirm some things. They confirm new names on the operational level, like Rodriguez and Perez. They confirm the existence of these vast data sets. But the gaps, the deletions, that seems to be the real story. The gaps are where the institutional liability resides. From the physical jail cell in 2019 to the digital server room in 2024. The documented continuity is one of evidence destruction or suppression. Next time. DOJ. 3.5 million pages. You have just heard an analysis of the official record. Every claim, name and date mentioned in this episode is backed by PR source documents. You can view the original files for yourself at Epsteinfiles fm. If you value this data first approach to journalism. Please leave a five star review wherever you're listening right now. It helps keep this investigation visible. We'll see you in the next file.