Build by AI

The Enterprise AI Wars Heat Up I 9th April

Use Left/Right to seek, Home/End to jump to start or end. Hold shift to jump forward or backward.

0:00 | 26:44
OpenAI is making bold economic proposals to Washington while AWS plays both sides by investing billions in OpenAI AND Anthropic. Meanwhile, Anthropic just dropped a game-changing tool for building AI agents, and early reports suggest Meta's new Muse Spark model might finally put Zuckerberg in the same league as the AI giants. Plus, we dive into OpenAI's new child safety blueprint and what the next phase of enterprise AI adoption really looks like. Buckle up - the AI landscape is shifting fast.
SPEAKER_00

So let me get this straight. AWS is investing billions of dollars in both OpenAI and Anthropic, essentially funding two companies that are in direct competition with each other and they're saying this isn't a conflict of interest.

SPEAKER_01

Dude, it's like being married to two people and telling them both it's totally fine because you have experience managing complicated relationships. Like what?

SPEAKER_00

Right? And this is happening at the exact same time that OpenAI is basically lobbying Washington with economic proposals while releasing safety blueprints. The timing feels strategic.

SPEAKER_01

Oh, it's absolutely strategic. And wait until you hear what Anthropic just dropped and what Meta might be cooking up. The enterprise AI wars are getting wild.

SPEAKER_00

This is either brilliant business maneuvering or we're watching the tech industry completely lose its mind. Maybe both. You're listening to Build by AI, the daily show where we decode what's actually happening in artificial intelligence. I'm Alex Shannon.

SPEAKER_01

And I'm Sam Hinton. Today we're talking about some major power plays in the AI world, from Washington lobbying to billion-dollar investment strategies that make zero sense on the surface.

SPEAKER_00

Plus, we've got some potentially huge news from Meta that could shake up everything. And Atlassian is making some interesting moves in the visual AI space.

SPEAKER_01

It's Wednesday, April 9th, 2026, and the AI landscape is shifting under our feet. Let's dive in.

SPEAKER_00

AWS is investing billions with a B on uh in both OpenAI and Anthropics simultaneously. These are two companies that are directly competing with each other in the large language model space. And when people started asking, hey, isn't this a conflict of interest? AWS leadership basically said, nah, we're good at managing competition with our partners. Their argument is that they have a culture of handling situations where they compete with their own partners. But Sam, help me understand this. Is this actually normal business practice or are we in uncharted territory here?

SPEAKER_01

Okay, so here's the thing. Yes, AWS does have experience competing with partners. They've done it with companies like Salesforce and Netflix for years. But this feels different because of the scale and the stakes involved. We're talking about billions of dollars going to two companies that are basically in an arms race to build the most powerful AI models. It's like if if during the space race, NASA had funded both the US and Soviet programs.

SPEAKER_00

But wait, let me play devil's advocate here. Isn't this actually smart diversification? I mean, nobody knows which AI approach is going to win long term. By betting on both horses, AWS ensures they have a relationship with whoever comes out on top.

SPEAKER_01

That's a fair point. But here's what worries me. What happens when OpenAI and Anthropic start competing for the same enterprise contracts? Does AWS have to choose sides? Um do they share information between the two? The potential for conflicts is huge.

SPEAKER_00

And there's another layer to this. Both of these companies need massive amounts of compute power to train their models. Guess who provides that? AWS. So they're essentially landlords to both competitors.

SPEAKER_01

Exactly. You know, it's like owning the racetrack and betting on multiple horses in the same race. You know, sure, you might say you're neutral, but you literally control the conditions of the competition.

SPEAKER_00

So what does this mean for businesses that are trying to choose between these AI platforms? Should they be concerned about AWS's dual allegiances?

SPEAKER_01

I think companies need to ask hard questions about data handling, preferential treatment, and long-term commitments. AWS says they can manage it, but trust needs to be earned, not just declared.

SPEAKER_00

Aaron Ross Powell But here's what I'm wondering. Could this actually benefit customers? If AWS has deep relationships with both companies, maybe they can push both to improve their offerings more aggressively.

SPEAKER_01

Hmm. That's interesting. Like they become this neutral party that can influence both sides to innovate faster? I could see that, but it requires a level of transparency and ethical behavior that's hard to enforce.

SPEAKER_00

Right. And what's the accountability mechanism here? If AWS makes a decision that benefits one AI company over another, who's watching? Who's making sure they're being fair?

SPEAKER_01

That's the crux of it. You know, in traditional industries, you'd have regulators or or industry bodies overseeing these kinds of arrangements. But AI is moving so fast that governance is way behind.

SPEAKER_00

And let's be real about the power dynamics here. AWS isn't just an investor. They're providing critical infrastructure. That gives them enormous leverage over both companies' operations and strategic decisions.

SPEAKER_01

Which brings up another question. Because from their perspective, they're essentially sharing a sugar daddy who's also funding their biggest rival.

SPEAKER_00

I imagine they don't love it, but they need the compute power and the investment. It's like being in a relationship you're not thrilled about because you need the apartment and the Netflix password.

unknown

Ha.

SPEAKER_01

Exactly. But that dependency could become a real problem if AWS starts making demands or if the competitive landscape shifts. These companies might find themselves in a very uncomfortable position.

SPEAKER_00

For people building AI applications, I think the takeaway is to be aware of these interconnections. The AI ecosystem is more intertwined than it appears on the surface. And that affects everything from pricing to availability to strategic direction.

SPEAKER_01

And watch for signs of preferential treatment. If you notice one AI platform getting better AWS integration, better pricing, or faster performance, that might not be coincidental.

SPEAKER_00

Keep an eye on this because as these AI investments get bigger and the competition gets fiercer, these kinds of conflicts are only going to become more common and more complicated.

SPEAKER_01

And honestly, this might be a preview of what happens when big tech companies start consolidating I assets. We could see a lot more of these awkward, multi-sided relationships in the future.

SPEAKER_00

Speaking of OpenAI, they've been busy in Washington lately. The company has made some economic proposals to DC policymakers, and from what we're seeing, the political reception has been, let's call it mixed. Now we don't have all the details of what exactly these proposals contain, but the fact that OpenAI is actively lobbying in Washington tells us they're thinking about regulation and policy at the highest levels. This comes at a time when there's growing scrutiny about AI safety, market concentration, and the role these companies should play in society. Sam, what's your read on OpenAI's DC strategy?

SPEAKER_01

This is classic tech company playbook, right? Get ahead of regulation by trying to shape it yourself. But here's what's interesting. OpenAI is doing this while they're still relatively early in their corporate evolution. Usually companies wait until they're facing serious regulatory pressure before they start heavy lobbying. OpenAI seems to be taking a proactive approach, which could be really smart or could backfire spectacularly.

SPEAKER_00

That's a good point. And timing-wise, this is happening right as they're releasing safety blueprints and making other public commitments. It feels coordinated, like they're trying to position themselves as the responsible AI company.

SPEAKER_01

Yeah, but here's my concern. When tech companies start making economic proposals to Washington, it's usually because they want something specific. Tax breaks, regulatory frameworks that favor them, protection from competitors. The question is whether these proposals actually serve the public interest or just OpenAI's business interests. And frankly, DC's track record with understanding tech issues isn't great.

SPEAKER_00

Right, and there's this broader question about whether AI companies should be writing their own rules. I mean, we've seen how that worked out with social media platforms over the past decade.

SPEAKER_01

Exactly. Facebook basically wrote the playbook on move fast and break things, apologize later. Do we really want the same approach with AI, which could have much bigger consequences?

SPEAKER_00

But on the flip side, who else has the technical expertise to craft meaningful AI policy? Congress can barely handle basic tech issues, let alone something as complex as artificial intelligence.

SPEAKER_01

That's the catch 22. We need people who understand the technology to write good policy, but the people who understand it best also have the biggest financial stakes in the outcome.

SPEAKER_00

And you know what's interesting? The fact that DC has formed opinions on these proposals suggests there's actually some substantive engagement happening. That's not always the case with tech policy.

SPEAKER_01

True, but I'm curious about what those opinions actually are. Are lawmakers pushing back on certain aspects? Are they buying into OpenAI's vision wholesale? The devil's in those details.

SPEAKER_00

Right. And there's a political dimension here too. AI policy is becoming a bipartisan issue, but for different reasons. Republicans worry about economic competitiveness, Democrats worry about worker displacement and safety.

SPEAKER_01

So OpenAI has to thread a really narrow needle, appeal to both sides without alienating either. That's why these economic proposals are smart. They speak to the competitiveness concerns while the safety blueprints address the regulatory worries.

SPEAKER_00

But here's what I'm watching for. Are other AI companies going to follow OpenAI's lead with their own economic proposals? Because if everyone starts lobbying with different visions, things could get messy fast.

SPEAKER_01

Oh, they absolutely will. Google, Microsoft, Meta, they're all watching this closely. If OpenAI gains regulatory advantage through these proposals, everyone else will be scrambling to catch up.

SPEAKER_00

Which could actually be good for the policy process, right? If multiple companies are proposing different frameworks, maybe policymakers get a more complete picture of the issues and trade-offs.

SPEAKER_01

Maybe. Or maybe they just get confused by competing corporate interests dressed up as public policy recommendations. It depends on whether DC has the expertise to separate good ideas from corporate spin.

SPEAKER_00

So for people watching this space, I'd say pay attention to what these economic proposals actually contain when more details emerge. The devil's always in the details with this stuff.

SPEAKER_01

And watch how other AI companies respond. If OpenAI is getting cozy with Washington, you can bet Google, Microsoft, and others are going to ramp up their own lobbying efforts. This could get messy fast.

SPEAKER_00

Plus, keep an eye on which lawmakers are engaging with these proposals and how. The political coalition around AI policy is still forming, and these early interactions could shape it for years to come, Yi To.

SPEAKER_01

And honestly, this is one of those moments where public engagement matters. If citizens don't weigh in on AI policy, these companies will fill the vacuum by default. That might not be the outcome we want.

SPEAKER_00

Alright, let's shift gears and talk about something that could be a real game changer. Anthropic just launched a new product that's designed to simplify the process of building AI agents using Clawed. The big selling point here is that it's supposed to lower the barrier to entry for businesses. Right now, building AI agents requires a lot of technical expertise, custom coding, and frankly a lot of trial and error. If Anthropic can actually make this accessible to regular businesses without big technical teams, that could accelerate enterprise AI adoption significantly. Sam, how big a deal is this?

SPEAKER_01

This could be huge, and here's why. Right now, most businesses are stuck at the cool demo stage with AI. They see the potential, but actually implementing useful AI agents feels like climbing Mount Everest. It's like the difference between seeing a beautiful website and actually knowing how to build one. There's been this massive gap between AI capability and AI usability for regular businesses.

SPEAKER_00

And the timing is interesting because we're seeing enterprise AI adoption growing rapidly across industries. But most of that growth has been concentrated among tech-savvy companies with big budgets for custom development.

SPEAKER_01

Right. This could be Anthropic's play to democratize AI agents. Think about it. If a small marketing agency or a local law firm can suddenly deploy sophisticated AI agents without hiring a team of developers, that changes everything.

SPEAKER_00

But let me ask you this. Are businesses actually ready for this? Because making AI agents easier to build is one thing. But do most companies have the processes and understanding to use them effectively?

SPEAKER_01

That's the million-dollar question. It reminds me of when WordPress made website building accessible to everyone. Suddenly everyone could build a website, but that didn't mean everyone built good websites. We might see a wave of poorly designed AI agents that don't actually solve business problems. Just because the technology became available doesn't mean the strategy became clearer.

SPEAKER_00

Although maybe that's okay. Like maybe businesses need to go through that experimental phase where they build some clunky AI agents before they figure out what actually works.

SPEAKER_01

Yeah, that's fair. And Anthropic has been pretty thoughtful about AI safety and responsible deployment, so hopefully they're building in guardrails and best practices from the start.

SPEAKER_00

Plus, this puts competitive pressure on OpenAI, Google, and others to make their tools more accessible too. Competition in the ease of use space is great for everyone.

SPEAKER_01

Absolutely. And for businesses listening, this is worth keeping an eye on because if Anthropic delivers on this promise, it could be your entry point into practical AI implementation.

SPEAKER_00

The key thing to watch is not just whether the tool works, but whether Anthropic provides the education and support that businesses need to use it effectively. Building the tool is only half the battle.

SPEAKER_01

Right, because here's what I'm wondering. What happens when thousands of businesses suddenly have access to AI agents but don't understand the implications? Are we prepared for that kind of rapid adoption?

SPEAKER_00

That's a great point. There are ethical considerations, privacy implications, job displacement concerns, all the stuff that gets glossed over in the excitement of easy AI agent building.

SPEAKER_01

And what about quality control? Yeah, you know, if building AI agents becomes as easy as creating a PowerPoint presentation, how do we ensure these agents are actually helpful and not just generating digital busy worth?

SPEAKER_00

I think that's where Anthropic's approach to AI safety and their focus on helpful, harmless, and honest AI could actually be a competitive advantage. They're not just making it easier, they're hopefully making it better.

SPEAKER_01

True. But there's also the question of vendor lock-in. If Anthropic makes it really easy to build agents with Claude, are businesses going to find themselves dependent on that ecosystem? That's a strategic consideration.

SPEAKER_00

Good point. Although if the alternative is spending months and thousands of dollars on custom development, a little vendor dependence might be worth a trade-off for smaller businesses.

SPEAKER_01

Fair enough. And honestly, if this works well, it could be the moment when AI stops being a tech company thing and becomes an every business thing. That's a pretty big shift.

SPEAKER_00

Which brings us back to that acceleration of enterprise AI adoption. If Anthropics succeeds here, we might look back at this as the moment AI went mainstream in business operations.

SPEAKER_01

Absolutely. You know, the question is whether the business world is ready for that acceleration, or if we're about to see a lot of trial and error in real time across entire industries.

SPEAKER_00

Now we need to talk about something much more serious. OpenAI has released what they're calling a child safety blueprint, and this is in response to what they describe as an alarming rise in child sexual exploitation that's been linked to eye advancements. This is obviously a deeply concerning issue, and it highlights some of the darker potential uses of AI technology that we don't always talk about, but absolutely need to address. The blueprint is designed to combat these problems, though we don't have all the specific details about what measures they're implementing. Sam, this feels like a critical moment for i safety discussions.

SPEAKER_01

Yeah, this is exactly the kind of issue that shows why AI safety isn't just about preventing artificial general intelligence from going rogue. There are immediate real-world harms happening right now. The fact that OpenAI is releasing a specific blueprint for this suggests they're seeing enough concerning activity that they felt compelled to take action. That's both good that they're responding and troubling that it's necessary.

SPEAKER_00

And this ties into broader concerns about AI-generat, deep fakes, synthetic media, and the ways these technologies can be misused. The same capabilities that can create amazing art can also create harmful content.

SPEAKER_01

Exactly. And here's what's challenging about this. These protections need to be built into the foundation of AI systems, which means thinking about potential misuse from day one.

SPEAKER_00

It also raises questions about industry-wide standards. OpenAI releasing their own blueprint is good, but shouldn't there be coordinated efforts across all AI companies to address these issues?

SPEAKER_01

That's a great point. Child safety shouldn't be a competitive advantage. It should be a baseline requirement. Maybe this is where we actually need government regulation to ensure consistent protections across all AI platforms.

SPEAKER_00

And for parents, educators, and anyone working with young people, this is a reminder that as AI becomes more prevalent, we need to be more vigilant about digital safety and education.

SPEAKER_01

The technology is advancing faster than our social systems can adapt. We need better education, better reporting mechanisms, and frankly, better accountability from AI companies.

SPEAKER_00

This is one of those areas where the AI industry's reputation and social license to operate is really on the line. They have to get this right. Not just for ethical reasons, but for their own long-term viability.

SPEAKER_01

And we'll be watching to see if other companies follow OpenAI's lead with their own safety blueprints, or if this becomes another area where industry coordination falls short.

SPEAKER_00

But here's what's tricky about this. How do you balance safety measures with legitimate uses of AI technology? Overly aggressive content filtering could limit beneficial applications.

SPEAKER_01

That's the eternal challenge with content moderation, right? You want to catch the bad stuff without throwing out the good stuff. With AI, the stakes are higher and the volume is much larger.

SPEAKER_00

And there's a detection arms race happening. As AI gets better at creating synthetic content, it also needs to get better at identifying synthetic content. It's like a cat and mouse game.

SPEAKER_01

Which is why I think the blueprint approach makes sense. It's not just about technology solutions, it's about processes, governance, partnerships with law enforcement, education initiatives.

SPEAKER_00

Right. This is a multifaceted problem that requires multifaceted solutions. Technology alone isn't going to solve child exploitation, but it can be part of a broader strategy.

SPEAKER_01

And timing-wise, releasing this alongside their DC economic proposals and enterprise AI initiatives, it feels like OpenAI is trying to demonstrate they can be both innovative and responsible.

SPEAKER_00

That's cynical, but probably accurate. Public trust is crucial for AI companies right now, and safety initiatives like this are part of building and maintaining that trust.

SPEAKER_01

Whether it's cynical or genuine doesn't matter as much as whether it's effective. If this blueprint actually reduces harm to children, then the motivations are secondary.

SPEAKER_00

Alright, let's move into some rapid-fire coverage. OpenAI is also talking about what they call the next phase of enterprise AI, with new products including something called Frontier, ChatGPT Enterprise Codex, and company wide AI agents.

SPEAKER_01

This feels like OpenAI's big push to own the enterprise market. Company-wide AI agents is particularly interesting because that suggests AI that can work across different departments and workflows, not just isolated use cases.

SPEAKER_00

Right. And this is happening while they're making those economic proposals to DC we talked about earlier. It's like they're trying to establish market dominance while also positioning themselves as policy leaders.

SPEAKER_01

Smart strategy, but risky. If they move too aggressively on market capture, they might invite more regulatory scrutiny. It's a delicate balance between growth and maintaining that responsible AI company image.

SPEAKER_00

And notice how AI adoption is accelerating across industries according to their messaging. That creates urgency for businesses, like get on board now or get left behind.

SPEAKER_01

Yeah, but I wonder if the market is actually ready for company-wide AI agents. That's a massive change in how businesses operate. Are most organizations prepared for that level of AI integration?

SPEAKER_00

Probably not, but maybe that's the point. If OpenI can help them get there faster than competitors, that's a huge competitive advantage. First mover advantage in the enterprise space is powerful.

SPEAKER_01

True. And with Frontier, ChatGPT Enterprise, and Codex, they're covering the full spectrum from cutting-edge research to practical business applications. That's a comprehensive approach to enterprise domination.

SPEAKER_00

Atlassian is making moves too. They've launched visual AI tools in Confluence and integrated third-party agents from companies like Lovable, Replit, and Gamma. Users can now create visual assets directly within Confluence.

SPEAKER_01

This is actually really smart positioning. Instead of trying to build everything in-house, they're becoming the platform where other AI tools can plug in. It's like the App Store model, but for AI agents in the workplace.

SPEAKER_00

And Confluence is already where a lot of teams do their documentation and collaboration. So adding AI capabilities there feels natural. It's meeting people where they already work instead of asking them to adopt new tools.

SPEAKER_01

Plus, visual AI tools for documentation could be huge for productivity. And if you can automatically generate diagrams, charts, and visual explanations, that saves tons of time for teams trying to communicate complex ideas.

SPEAKER_00

The partnerships with Lovable, Replit, and Gamma are interesting too. Those are all companies with specific AI specialties. So Atlassian is curating best-in-class tools rather than trying to do everything themselves.

SPEAKER_01

Which could be the winning strategy in the long run. Instead of building mediocre AI features across the board, they're offering excellent AI features from specialists. That's probably better for users.

SPEAKER_00

And it positions Confluence as the central hub for AI-powered collaboration. If teams can access multiple AI agents from one familiar interface, that reduces friction significantly.

SPEAKER_01

The question is whether these integrations are deep enough to be useful, or just surface-level connections that create more complexity than value. Integration quality matters more than integration quantity.

SPEAKER_00

Now this is interesting. Early reports suggest that Meta has released something called Muse Spark, which is apparently their first AI model following what's being described as a strategic AI reboot.

SPEAKER_01

Okay. If confirmed, this could be big news. The report mentions benchmark results showing formidable performance that puts Meta in competition with the top AI companies. Zuckerberg has been pretty quiet on the AI front lately.

SPEAKER_00

Right. And the framing of getting a seat at the big kids' table suggests Meta might have been falling behind in AI. And this is their attempt to catch up to OpenAI, Google, and others. The timing is curious though. Right when everyone else is making big enterprise plays, Meta drops a model that could shake up the competitive landscape. That's either great timing or terrible timing, depending on your perspective.

SPEAKER_01

And remember, Meta has massive amounts of data from their social platforms, plus serious compute infrastructure. If they can leverage those assets effectively, MuseSpark could be formidable indeed.

SPEAKER_00

The question is what their go-to-market strategy looks like. Are they going after enterprise customers like everyone else, or do they have a different approach given their social media expertise?

SPEAKER_01

That'll be fascinating to watch. Meta could integrate AI deeply into their existing platforms in ways that other companies can't match. That's potentially a huge advantage if they execute well.

SPEAKER_00

And finally, there's this company called Poke that's taking a completely different approach. They're making AI agents accessible through simple text messaging, no complex setup required.

SPEAKER_01

Because it flips the whole paradigm. Instead of making people learn new interfaces, they're using the interface everyone already knows, texting. It's like the ultimate in user experience simplification.

SPEAKER_00

According to reports, the platform can handle tasks and automations all through text messages. It's almost like having a personal assistant you can reach by SMS.

SPEAKER_01

If this works well, it could be huge for adoption among less tech savvy users. Sometimes the best innovation isn't making technology more complex. It's making it feel invisible and natural.

SPEAKER_00

And it removes all the barriers that usually prevent people from trying AI agents. No app downloads, no account creation, no learning new commands, just text like you normally would.

SPEAKER_01

The challenge will be handling complex tasks through such a simple interface. There's a reason most AI platforms have elaborate UIs. Sometimes you need more than text to communicate effectively with AI.

SPEAKER_00

True. But maybe that constraint is actually beneficial. If the AI agents have to work through text messaging, they're forced to be more conversational and intuitive. Less feature bloats, more focused functionality.

SPEAKER_01

That's a really good point. Polk might have found the sweet spot between powerful AI capabilities and human-friendly interaction. That could be exactly what mainstream adoption needs.

SPEAKER_00

If you zoom out and look at everything we covered today, there's a clear theme emerging. This is all about the battle for enterprise AI dominance.

SPEAKER_01

Absolutely. You've got uh OpenAI lobbying Washington while releasing safety blueprints, AWS hedging their bets by investing in multiple AI companies, Anthropic simplifying agent development, and potentially Meta making a major comeback play.

SPEAKER_00

And what's interesting is how different their strategies are. OpenAI is going the policy route, AWS is playing venture capitalist, anthropic is focusing on usability, and companies like Atlassian are becoming platforms for AI integration.

SPEAKER_01

What I'm watching for is whether any of these approaches prove to be clearly superior, or if we end up with a fractured market where different strategies work for different segments. The enterprise market is huge and diverse enough to support multiple winners.

SPEAKER_00

But here's my prediction. And that might not be the companies with the most advanced models.

SPEAKER_01

That's a great point. Sometimes the best technology doesn't win. The most practical and accessible technology does. We might be at a turning point where ease of use becomes more important than raw capability.

SPEAKER_00

And think about the interconnections here. AWS is funding both OpenAI and Anthropic. OpenAI is courting policymakers while pushing enterprise products. Anthropic is democratizing AI agent development. These aren't isolated strategies.

SPEAKER_01

Right. And meanwhile, you have companies like Polk and Atlassian taking completely different approaches. One through radical simplification, the other through platform integration. The diversity of approaches is actually really healthy.

SPEAKER_00

Plus, we can't ignore the safety angle. OpenAI's Child Safety Blueprint isn't just about doing the right thing, it's about maintaining the social license to operate. Companies that get safety wrong will face backlash.

SPEAKER_01

And that that creates interesting dynamics. Companies have to balance innovation speed with responsibility, competitive advantage with industry collaboration, growth with regulatory compliance. It's a complex optimization problem.

SPEAKER_00

The meta situation is particularly intriguing because if Muse Spark is as competitive as early reports suggest, it could scramble all these careful strategies. Sudden disruption changes everything.

SPEAKER_01

Which is why AWS's multi-investment approach might actually be brilliant. They're not betting on any single winner. They're positioning themselves to benefit no matter who comes out on top.

SPEAKER_00

Although that creates its own risks, as we discussed. Playing all sides works until the sides realize you're playing all sides. Trust and exclusivity have value too.

SPEAKER_01

True. And for businesses watching all this unfold, I think the key insight is that the enterprise AI landscape is still very much in flux. Early decisions about platforms and vendors could have long-term consequences.

SPEAKER_00

But also the barriers to entry are dropping rapidly. Whether it's Anthropic's simplified agent building, Polk's text-based interface, or at Lassian's platform approach, AI is becoming more accessible to regular businesses.

SPEAKER_01

Which creates both opportunity and risk. More businesses can benefit from AI, but more businesses can also make mistakes with AI. The democratization of powerful technology is always a double-edged sword.

SPEAKER_00

That's a wrap for today's Build by AI. As always, the AI world is moving fast and these enterprise battles are just heating up.

SPEAKER_01

If you're getting value from these daily deep dives, definitely subscribe wherever you get your podcasts. And hit us up on social if you've got thoughts on any of these stories. We love the discussion.

SPEAKER_00

We'll be back tomorrow with more AI news and analysis. I'm Alex Shannon.

SPEAKER_01

And I'm Sam Hinton. Until next time, keep building.