Build by AI
Build by AI is your daily briefing on everything happening in the world of artificial intelligence, delivered straight to your ears every single day.
Whether you're a founder trying to stay ahead of the curve, a professional figuring out how AI fits into your work, or simply someone who wants to understand what's actually going on in one of the fastest-moving industries on the planet, Build by AI cuts through the noise and brings you what matters, in plain English, in under ten minutes.
Every episode covers the latest AI news, model releases, industry shifts, and research breakthroughs, so you never have to spend hours scrolling to stay informed. Think of it as your morning coffee briefing for the AI age.
Build by AI is produced by artificial intelligence, from research to script to publish, with every episode reviewed and verified by a human editor before it reaches your ears. So you get the speed and consistency of automation, without sacrificing accuracy or trust. Which also raises the question we're quietly exploring with every episode: how good can AI-generated content actually get? You be the judge.
New episodes drop daily.
Subscribe wherever you get your podcasts and wake up smarter every morning.
Collaboration requests: wiktoria@womenlead.ai
Topics covered: artificial intelligence news, large language models, generative AI, AI tools, ChatGPT, Claude, Gemini, AI regulation, machine learning research, tech industry news, AI startups, and the future of work.
Build by AI
Space Data Centers and the $830M Infrastructure Arms Race I 31st March
Use Left/Right to seek, Home/End to jump to start or end. Hold shift to jump forward or backward.
Okay, so let me get this straight. In one day, we've got companies raising over a billion dollars combined to build data centers, both on Earth and literally in space. And I'm genuinely not sure which one sounds more realistic at this point.
SPEAKER_01Dude, right? Like, when space data centers start sounding more feasible than some of these Earth-based infrastructure plays, we've officially entered the twilight zone of uh AI funding.
SPEAKER_00And that's just the beginning. We're also seeing a$400 million bet on toppling NVIDIA's chip dominance, plus this fascinating problem where AI is writing so much code that we need other AI just to figure out if the first AI's code actually works.
SPEAKER_01It's like we're building the digital equivalent of the Tower of Babel, except with venture capital and orbital mechanics involved.
SPEAKER_00And the crazy part is, all of this infrastructure build out is happening because companies are convinced we're still in the early days of AI adoption. These aren't defensive moves, these are massive offensive plays.
SPEAKER_01Right. When you're talking about putting data centers in orbit, you're basically saying the current approach to computing infrastructure is fundamentally broken at scale. That's either visionary or completely delusional.
SPEAKER_00You're listening to Build by AI. I'm Alex Shannon. And yeah, March 31st, 2026 is shaping up to be one of those days where the future feels like it's arriving faster than we can keep up with.
SPEAKER_01And I'm Sam Hinton. And honestly, today's stories read like someone fed a sci-fi novel into a funding announcement generator. We've got space data centers, AI chip wars, and digital human twins. And somehow it's all connected to this massive infrastructure arms race that's reshaping how we think about computing.
SPEAKER_00Alright, let's dive in because there's a lot to unpack here, and some of these moves are going to fundamentally change how AI gets built and deployed. So let's start with what might be the wildest story of the day. StarCloud just closed a$170 million Series A to build data centers in space. And get this, they've become the fastest Y Combinator startup ever to hit unicorn status. Just 17 months after demo day.
SPEAKER_01Wait, 17 months? That's insane. But okay, let's talk about the elephant in the room. Are we seriously at the point where launching computers into orbit makes economic sense?
SPEAKER_00Right, because on the surface it sounds completely ridiculous. I mean the cost of getting anything to space is still astronomical, no pun intended. What's the actual value proposition here that convinced investors to drop 170 million dollars?
SPEAKER_01Well, think about it this way. Space has some unique advantages that are becoming more relevant in the AI era. You've got basically unlimited cooling because space is really, really cold. You've got no physical security concerns once you're up there. And here's the big one: latency. If you're serving global applications, being in low Earth orbit might actually give you better average latency to users worldwide than any single ground-based data center.
SPEAKER_00But hold on, I'm still skeptical about the economics. Even if the operational advantages are real, the upfront costs have to be enormous. You're talking about space hardened hardware, launch costs, maintenance. How do you ever make that pencil out compared to just building more data centers on Earth?
SPEAKER_01That's where I think the timing is everything. SpaceX and other companies have driven launch costs down by like 90% over the past decade. Plus, with AI workloads, you're dealing with such high value computations that the premium might actually be worth it. If you're running a global AI service and you can reduce latency by 50 milliseconds for every user, that could be worth hundreds of millions in improved performance.
SPEAKER_00And I guess there's also the angle that as AI models get bigger and more complex, maybe the infrastructure requirements become so demanding that you need to think outside the box, literally outside Earth's atmosphere.
SPEAKER_01Exactly. This feels like one of those things that sounds crazy until it doesn't. Remember when people thought cloud computing was a fad? Now we're talking about orbital computing. The fact that they hit unicorn status so fast suggests the market sees something real here.
SPEAKER_00But let's get practical for a second. What happens when something breaks? Like if you have a hardware failure in a traditional data center, you call a technician. If you have a hardware failure in orbit, what do you do? Send up a SpaceX mission.
SPEAKER_01That's actually a fascinating question. And I think it completely changes how you design these systems. You probably need to over-engineer everything for redundancy in a way that ground-based data centers don't. Which brings the costs up even more, but also potentially makes the whole system more robust.
SPEAKER_00And there's the regulatory aspect too. Who regulates space-based data centers? Is this a NASA thing? FCC.
SPEAKER_01Yeah, we're we're basically an uncharted territory there. But you know what? Maybe that's actually an advantage. If you can figure out the regulatory framework first, you might have a huge moat against competitors who come later.
SPEAKER_00And think about data sovereignty. If your data is literally an international space, whose laws apply? That could be either a massive advantage or a massive headache for enterprise customers.
SPEAKER_01Right. And for companies that are paranoid about data security and government surveillance, having your data literally out of reach of any earthbound authority might be worth paying a premium for.
SPEAKER_00Keep an eye on this because if StarCloud actually pulls this off and demonstrates viable space-based AI infrastructure, it's going to completely change how we think about global computing architecture.
SPEAKER_01And honestly, the speed of their growth trajectory suggests they might have some serious technical breakthroughs or partnerships that we don't know about yet. Y Combinator doesn't usually produce unicorns in 17 months unless there's something really special happening.
SPEAKER_00Now speaking of infrastructure plays, early reports suggest Mistral AI just secured$830 million in debt financing to build a data center near Paris, with operations planned to start by Q2 2026. So while StarCloud is going to space, he Mistrel is making a massive bet on Earth-based infrastructure.
SPEAKER_01Okay,$803 million in debt, that's a huge number. And the fact that it's debt rather than equity tells you something important. They're confident enough in their business model to take on that kind of obligation, which suggests they see very predictable revenue streams ahead.
SPEAKER_00Right. And timing-wise, if they're aiming for Q2 2026 operations, that's basically tomorrow in data center construction terms. This feels like a response to immediate capacity constraints rather than a long-term strategic play.
SPEAKER_01Absolutely. And um and think about what this means for the European AI landscape. Mistral has been positioning itself as the European answer to open AI and Anthropic, and now they're building the infrastructure to back that up. This isn't just about having more compute, it's about data sovereignty and reducing dependence on US cloud providers.
SPEAKER_00But I'm curious about the economics here too.$830 million buys a lot of GPUs, but with the current chip shortage and the crazy prices NVIDIA is charging, are they going to get enough compute power to really compete with the big US players?
SPEAKER_01That's the million-dollar question, or I guess the$830 million question. But here's what's interesting. Mistral has been really focused on efficiency. Their models punch above their weight in terms of performance per parameter, so maybe they don't need to match open AI's compute dollar for dollar if they're more efficient with what they've got.
SPEAKER_00And there's also the geographic angle. Having a major AI infrastructure hub in Europe could attract a lot of European companies who want to keep their data local for regulatory reasons. GDPR compliance alone could drive significant demand.
SPEAKER_01Yeah. This feels like Mistrol is making a bet that AI infrastructure is going to regionalize rather than centralize. Instead of everyone depending on a few massive US-based cloud providers, you'll have regional champions building out local capacity.
SPEAKER_00But let's talk about the competitive dynamics here.$830 million sounds like a lot. But OpenAI and Microsoft are throwing around numbers that make this look small. Can a single European player really compete at the scale needed?
SPEAKER_01That's where I think the strategy might be different. Maybe they're not trying to beat OpenAI at their own game. Maybe they're building for European enterprise customers who prioritize data, residency, regulatory compliance, and cultural alignment over raw scale.
SPEAKER_00And the debt financing structure is really interesting too. It suggests they have concrete business commitments, like signed contracts or letters of intent that justify taking on that level of financial obligation.
SPEAKER_01Exactly. You don't get banks to lend you$830 million for speculative AI infrastructure unless you can show them a clear path to revenue. This feels like Mistral has locked in some major enterprise customers already.
SPEAKER_00And if they can get this facility operational by Q2 2026, that timing could be perfect. A lot of European companies are probably getting frustrated with relying on US cloud providers and are looking for alternatives.
SPEAKER_01Plus, there's probably some government support behind the scenes here. European governments are definitely interested in reducing technological dependence on the US, especially for critical AI infrastructure.
SPEAKER_00If confirmed, this could be the beginning of a much broader trend, where AI infrastructure becomes more distributed geographically, driven by a combination of regulatory requirements, latency concerns, and good old-fashioned national competitiveness. Alright, so staying on this infrastructure theme, early reports suggest AI chip startup rebellions just raised$400 million at a$2.3 billion valuation in what they're calling a pre-IPO round. They're designing specialized chips for AI inference and positioning themselves as a challenger to NVIDIA's market dominance.
SPEAKER_01Okay,$2.3 billion valuation for a chip company that's challenging NVIDIA. That's either brilliant or completely insane. And honestly, it might be both. Everyone and their grandmother has been trying to build the NVIDIA killer for years.
SPEAKER_00Right. And the graveyard of NVIDIA competitors is pretty extensive at this point. But what's interesting here is they're specifically focusing on inference rather than training. That might actually be a smarter play than going head to head with NVIDIA's training dominance.
SPEAKER_01That's a really good point. Training is where NVIDIA has this massive moat with CUDA and their ecosystem, but inference is a different game. And there's definitely room for specialized silicon that can beat general-purpose GPUs on those metrics.
SPEAKER_00And think about the market timing. We're seeing this explosion in AI applications that need to run inference at scale. Chatbots, image generation, code completion. The total addressable market for inference chips is growing exponentially. So maybe there's room for multiple winners.
SPEAKER_01But here's what I'm skeptical about. It's not just about having better hardware. Nvidia's real advantage is the software ecosystem. Developers know CUDA, you know, their tools are mature, the libraries all work together. How do you break into that without spending a decade building ecosystem?
SPEAKER_00That's the billion-dollar question, literally. Maybe the answer is you don't try to replicate the NVIDIA ecosystem. You build something completely different that's so much better for specific use cases that developers are willing to learn new tools. Which makes me wonder, who are those customers? Are we talking about cloud providers who want to reduce their dependence on Nvidia? Enterprise companies building their own AI infrastructure, startups looking for cost-effective inference.
SPEAKER_01Companies like AWS, Google Cloud, Azure, that they're all paying massive premiums to NVIDIA, and they desperately want alternatives. If Rebellions can offer 80% of the performance at 50% of the cost for inference workloads, that's a huge win.
SPEAKER_00And there's also the geopolitical angle here. With all the chip export restrictions and trade tensions, having non-US chip alternatives become strategically important for a lot of companies and countries.
SPEAKER_01Yeah, especially in Asia and Europe, where there's growing concern about technological dependence. A successful NVIDIA alternative could capture a lot of that demand.
SPEAKER_00But let's be real about the technical challenges. Nvidia has spent decades optimizing their chips and software stack. Can a startup really match that performance and reliability in their first generation of products?
SPEAKER_01That's the big risk. But you know what? Sometimes it takes fresh thinking and modern architecture to leapfrog incumbent technology. Nvidia's chips are incredibly powerful, but they're also designed to be general purpose. If you can build something that's specifically optimized for the inference workloads that most companies actually run, you might be able to beat them on the metrics that matter.
SPEAKER_00If this plays out, it could be huge for anyone building AI applications. More competition in the chip space means better performance and lower costs, which makes AI more accessible across the board. That's definitely worth watching as they move toward that IPO. Now here's a story that really gets to the heart of where AI development is heading. Early reports suggest Codo just raised$70 million to focus on code verification as AI generated code becomes more prevalent. Essentially, they're building tools to make sure AI generated code actually works properly.
SPEAKER_01Oh man, this is such an important problem that nobody's talking about enough. We're in this phase where AI can write code that looks reasonable, passes basic tests, but then has these subtle bugs or security vulnerabilities that only show up in production.
unknownRight.
SPEAKER_00Developers are becoming more productive at generating code, but potentially less good at understanding what that code actually does under the hood.
SPEAKER_01As AI coding tools get better, junior developers especially are going to rely on them more heavily. But if you don't deeply understand the code you're shipping, you can't really verify whether it's correct or secure.
SPEAKER_00So in some ways, AI coding tools might be making the overall quality problem worse. Even as they make developers more productive. It's like having a really fast typist who might not understand what they're typing.
SPEAKER_01Exactly, and that's where a company like Kodo comes in. If they can build tools that automatically verify AI generated code for correctness, security, performance, that could be incredibly valuable. You get the productivity benefits of AI coding without the quality risks.
SPEAKER_00But I wonder how technically feasible this really is. Code verification is a notoriously hard problem even for human written code. Can you really build automated tools that are smart enough to catch the subtle issues that AI coding introduces?
SPEAKER_01That's the multi-million dollar question. My guess is it's going to be about building specialized verification tools for different types of AA-generated code patterns. Like if you know that the code was generated by GPT-4 for a specific type of task, you can probably predict the most likely failure modes and test for those specifically.
SPEAKER_00And from a business perspective, this makes total sense. Every company using AI coding tools is going to need some way to ensure code quality. And 70 million million suggests investors think this market is going to be huge.
SPEAKER_01Yeah, and and think about the liability issues. If you're a you're a company shipping software that was partially written by AI and that software has a security vulnerability that causes a data breach, who's responsible? Legal departments are going to demand these kinds of verification tools.
SPEAKER_00That's a really good point. It's not just about code quality, it's about legal and regulatory compliance. Companies need to be able to demonstrate that they've done due diligence on AI generated code.
SPEAKER_01AI-generated code might work, but is it efficient? Is it maintainable? Does it follow best practices? These are all things that a verification platform would need to check.
SPEAKER_00Plus, as AI coding tools get more sophisticated, the verification needs are going to get more complex too. Today's AI might generate simple functions, but tomorrow's AI might be architecting entire microservices. The verification challenge scales with the capability.
SPEAKER_01And here's something else. As more code gets generated by AI, human developers are going to lose some of their intuitive ability to spot problems. We're going to become more dependent on automated verification, whether we want to or not.
SPEAKER_00Keep an eye on this space because I think we're going to see a whole ecosystem of tools emerge around making AI-generated code production ready. Verification is just the beginning. You'll probably need specialized testing, monitoring, and debugging tools too.
SPEAKER_01Absolutely. But someone needs to provide the tools to make sure that code is actually good.
SPEAKER_00Alright, let's hit some rapid fire stories. Early reports suggest ScaleOps just secured$130 million to address GPU shortages and high AI cloud costs through real-time infrastructure automation.
SPEAKER_01This is basically the make AI cheaper to run play, which is smart because cloud costs are becoming a real barrier to AI adoption. If you can automate infrastructure to be more efficient, that's a huge value prop.
SPEAKER_00And with GPU shortages still being a major issue, anything that can squeeze more performance out of existing hardware is going to be in high demand.
SPEAKER_01Exactly. This feels like infrastructure tooling that could become essential as AI workloads scale up.
SPEAKER_00What I like about this approach is that it's solving a problem that affects everyone running AI workloads, not just the big cloud providers. Even smaller companies could benefit from better infrastructure efficiency.
SPEAKER_01And$130 million suggests they're seeing serious demand already. Companies are clearly willing to pay to optimize their AI infrastructure costs.
SPEAKER_00Next up, according to reports, Lite LLM terminated its relationship with security compliance partner Delve following a credential stealing malware attack. Lite LLM had previously gotten security certifications through Delve.
SPEAKER_01Yikes, that's a messy situation. When your security compliance partner gets compromised, it kind of defeats the whole purpose. This is going to make enterprises even more paranoid about vetting their AI tool vendors. Right. The stakes are just way higher when you're routing AI traffic for major companies. One security incident can tank your credibility overnight.
SPEAKER_00What's particularly concerning is that Light LLM had obtained two security certifications through Delve. So enterprise customers probably thought they were covered from a compliance perspective.
SPEAKER_01Yeah, and now they have to figure out how to maintain those certifications and rebuild trust with customers. It's a reminder that security is only as strong as your weakest link.
SPEAKER_00This whole situation is going to make enterprise customers much more demanding about security practices from their AI vendors. Expect to see a lot more direct audits and certifications.
SPEAKER_01And honestly, that's probably a good thing. The potential for medical research is is huge if you can create realistic synthetic patient data. But the implications are kind of mind-bending.
SPEAKER_00Right. Imagine being able to test treatments on thousands of virtual patients before ever running a real clinical trial. It could dramatically speed up medical research.
SPEAKER_01But it also raises all these questions about how accurate these digital twins really are and whether synthetic data can truly replace real patient data for research purposes.
SPEAKER_00And they're apparently aggregating disparate data sources to represent anatomy, physiology, and behavior, which sounds incredibly complex from a technical standpoint.
SPEAKER_01The data availability problem in medicine is real, though. Patient privacy regulations make it really hard to get large data sets for research. If synthetic data can solve that while preserving privacy, it's a huge win.
SPEAKER_00Plus, you could potentially create digital twins of rare conditions where you don't have enough real patient data to do meaningful research.
SPEAKER_01Yeah, this could democratize medical research in ways we haven't seen before. Though I'd want to see a lot of validation that these synthetic patients actually behave like real ones. But there's still surprisingly little rigorous evaluation of whether they actually improve patient outcomes. Exactly. The gap between this AI tool can detect something and this AI tool improves patient care is enormous, and we're just starting to close that gap.
SPEAKER_00And there's this tension between innovation and safety in healthcare that doesn't exist in other industries. You can't just ship an MVP and iterate based on user feedback when patient lives are on the line.
SPEAKER_01Right, but the regulatory approval process is so slow that by the time an AI health tool gets approved, the underlying technology might be completely outdated. It's a really challenging problem to solve. Yeah, it's like we're seeing the build-out of the physical and digital infrastructure that's going to power the next phase of AI development. Space data centers, massive European facilities, specialized chips, code verification tools. It's all connected.
SPEAKER_00And what strikes me is how much money is flowing into these infrastructure plays. We're talking about well over a billion dollars just in today's stories. That suggests investors think we're still in the early innings of AI adoption.
SPEAKER_01But here's what I find most interesting. A lot of these bets are about solving problems that AI itself has created. AI coding creates a need for code verification. AI model scaling creates chip shortages and infrastructure bottlenecks. It's like we're building solutions to problems that didn't exist five years ago.
SPEAKER_00That's a really good point. We're not just scaling AI, we're having to completely rethink computing infrastructure, software development practices, even data center location strategies to support AI workloads.
SPEAKER_01And I think what we're seeing is just the beginning. As AI models get bigger and more capable, the infrastructure requirements are going to get even more demanding. Space data centers might sound crazy today, but they might be necessary tomorrow.
SPEAKER_00There's also this interesting geographic competition happening. Mistral building infrastructure in Europe, rebellions challenging Nvidia's dominance. Companies looking at space as a way to transcend geographic limitations entirely.
SPEAKER_01Right. Maybe the future of AI infrastructure isn't three big cloud providers. Maybe it's hundreds of specialized providers serving different needs.
SPEAKER_00And the quality and verification angle is huge too. As AI becomes more capable and autonomous, we need better tools to ensure that what it produces is actually reliable and safe. That's not just a nice to have, it's becoming mission critical.
SPEAKER_01Especially in regulated industries like healthcare, where we saw those digital twins and effectiveness questions. The stakes keep getting higher as AI tools become more sophisticated and widely deployed.
SPEAKER_00What's really fascinating is how all these pieces connect. Better infrastructure enables more sophisticated AI, which creates new verification and quality challenges, which drives demand for specialized tools, which requires even more infrastructure.
SPEAKER_01It's this virtuous cycle, or maybe vicious cycle, depending on your perspective, where each advancement in AI capability creates new infrastructure needs and business opportunities.
SPEAKER_00And the speed is just incredible. StarCloud went from Y Combinator Demo Day to Unicorn in 17 months. Rebellions is already planning an IPO. These aren't 10-year infrastructure build-outs, they're sprint speed developments.
SPEAKER_01Which suggests that either the market opportunity is so massive that everyone's rushing to capture it, or we're in some kind of bubble where reality hasn't caught up with valuations yet. Probably a bit of both.