The Macro AI Podcast

Model Context Protocol (MCP) Explained: The Economics of Scaling Enterprise AI Without Exploding Costs

The AI Guides - Gary Sloper & Scott Bryan Season 2 Episode 65

Use Left/Right to seek, Home/End to jump to start or end. Hold shift to jump forward or backward.

0:00 | 18:22

In this episode of The Macro AI Podcast, Gary Sloper and Scott Bryan revisit the Model Context Protocol (MCP)—a topic that continues to generate strong listener interest and real-world enterprise questions. 

As organizations move beyond AI pilots and demos, many are discovering that AI isn’t failing because of the models—it’s failing because of integration, governance, and cost. This episode explores why enterprise AI so often hits scaling walls and how MCP is emerging as a critical piece of infrastructure to remove them. 

The conversation breaks down MCP at a practical, executive level—explaining how it standardizes the way AI systems discover, understand, and safely interact with enterprise tools and data. Gary and Scott walk through why traditional API-based integrations struggle in AI-driven environments, how MCP changes the N-by-M integration problem, and why this matters for CIOs, CFOs, and CEOs planning long-term AI strategies. 

A major focus of the episode is AI economics, including a deep dive into token costs—one of the most misunderstood and underestimated drivers of enterprise AI spend. Using clear, real-world examples, the discussion shows how MCP can dramatically reduce token usage, improve performance, and turn unpredictable inference costs into a controllable operating expense. 

The episode also covers: 

  • Why MCP fundamentally changes the economics of scaling enterprise AI 
  • How token efficiency directly impacts ROI, latency, and adoption 
  • The infrastructure and total cost of ownership tradeoffs leaders need to understand 
  • Governance risks, including the rise of “shadow MCP,” and why centralized oversight matters 
  • How MCP complements—not replaces—RAG in modern enterprise AI architectures 

Bottom line: MCP is not a feature or a framework—it’s becoming core infrastructure for serious enterprise AI. If you’re responsible for AI strategy, governance, or budgets, this episode explains why MCP belongs on your radar now. 

Send a Text to the AI Guides on the show!


About your AI Guides

Gary Sloper

https://www.linkedin.com/in/gsloper/


Scott Bryan

https://www.linkedin.com/in/scottjbryan/

Macro AI Website:

https://www.macroaipodcast.com/

Macro AI LinkedIn Page:

https://www.linkedin.com/company/macro-ai-podcast/


Gary's Free AI Readiness Assessment:

https://macronetservices.com/events/the-comprehensive-guide-to-ai-readiness


Scott's Content & Blog

https://www.macronomics.ai/blog





00:00
Welcome to the Macro AI Podcast,  where your expert guides Gary Sloper and Scott Bryan navigate the ever-evolving world of artificial intelligence.  Step into the future with us  as we uncover how AI is revolutionizing the global business landscape  from nimble startups to Fortune 500 giants.  Whether you're a seasoned executive,  an ambitious entrepreneur,

00:27
or simply eager to harness AI's potential,  we've got you covered.  Expect actionable insights,  conversations with industry trailblazers  and service providers,  and proven strategies to keep you ahead in a world being shaped rapidly by innovation.  Gary and Scott are here to decode the complexities of AI  and to bring forward ideas that can transform cutting-edge technology  into real-world business success.

00:57
So join us,  let's explore, learn  and lead together.  Welcome back to the Macro AI Podcast, the show where we break down what's happening in artificial intelligence and more importantly, what it means for business leaders who are trying to deploy AI responsibly, securely and at scale. I'm Gary Slover. And I'm Scott Bryan. In today's episode, we are revisiting Model Context Protocol, MCP, a topic that we covered in season one, episode 14.

01:27
And we have been continuing to get a lot of downloads  on the episode and queries around MCP. We're going to talk about why enterprise AI in many cases keep hitting walls, how MCP is designed to remove those walls and what that means financially and strategically for CIOs, CFOs and CEOs. If you're responsible for AI strategy or the budget behind it, this episode is for you.

01:58
So Scott, let's start  with the core problem MCP is addressing. Most organizations  really underestimate how fundamentally different artificial intelligence systems are from, say, traditional software, SaaS and other  legacy software that they're accustomed to. Traditional applications are deterministic. You might give them an input, and they produce a predictable output.

02:24
AI systems are probabilistic and stateful. They reason, they adapt and change behavior based on context. That difference becomes critical the moment AI needs to interact with real enterprise systems. Yeah. And I think that's, you know, because now AI isn't just answering questions. It's reading from CRMs, it's querying data warehouses, triggering workflows.

02:52
and writing back into operational systems.  And in reality, enterprise environments were never designed for that kind of interaction. They were really more designed in silos. So each system has its own API, its own authentication model, its own permissioning scheme and its own data structures. So now when AI arrives,  organizations do what they've always done. They just integrate because enterprise systems of today were not built for AI from the ground up. Greenfield.

03:22
Yeah, that's a point. mean, the problem is scale. Instead of one application talking to one system, really you now have multiple models, multiple agents, multiple workflows and multiple tools all trying to interact at once. This creates the end by end integration problem. And that's where AI velocity quietly collapses. And many of you have probably seen this. The real cost isn't the first integration. So when you start

03:50
your AI journey. It's really the second, the third, or the tenth. Every new AI initiative reopens security reviews.  It  has data access debates and custom glue code conversations. uh Really over time, progress slows,  risk increases, and leadership confidence erodes in what is being built because you're having those issues, like I said, the second or tenth time afterwards. m

04:20
Yeah. So, so in those environments, you know, how does MCP, you know, assist or, kind of apply here. And, and so, you know, I think the question that you're probably asking, you know, what is MCP really solving and at its core, you know, model context protocol standardizes  how AI systems understand interact with those external capabilities. This isn't just about

04:46
calling APIs, many of us have done that for years where you have a SaaS environment and you have APIs into others. ah It's really about discovery, context, and control for your artificial intelligence environment. Yeah, I totally agree. And I think that distinction really matters because uh APIs  assume the caller already knows what exists,  how to call it, and what inputs are valid. And AI just doesn't work that way. AI needs to discover what it's allowed to do.

05:16
and then understand how to do it safely and operate within the defined guardrails. Right, right, exactly. And  MCP introduces a formal contract, right? So, you know, as you're familiar with it, clearly defines what tools are available in that moment, what each tool does,  how it can be invoked and really, you know, an important one, what permissions apply  in that environment. So architecturally, this decouples

05:43
reasoning from execution. The model decides what should happen. The MCP server controls how it happens in practice.  And I think that separation is what makes enterprise grade governance, security and cost control possible.  critically MCP is stateful, as I mentioned before. So this is not a stateless chat. The system maintains awareness of available tools and permissions and uh active context, which oh

06:12
is essential for long running enterprise grade workflows.  And so that is something that, you know, you really should just kind of stop and think about versus what you may be missing today in your environment.  And to understand why MCP scales, you need to understand its primitives. Resources are, you know, intentionally read only. ah They allow artificial intelligence to see the business, the data, document, records, et cetera.

06:42
without immediately introducing risk. And I think that's an important concept as well. Yep, I totally agree. And tools are where the action happens. these are executable capabilities that can update systems. They can trigger complete workflows,  perform operations. And what's critical is that every tool is  schema described. So the model isn't  guessing how to use it. It's operating against a clear machine readable contract.

07:12
Yeah. mean, prompts are their most underappreciated piece here and in MCP, know, prompts become reusable workflows over time. They encode how you, how really your organization solves problems,  executes processes, uh, and makes decisions  for the business. And that's how artificial intelligence stops being individual productivity, which I think a lot of people assume and are accustomed to. And now it starts to become more institutional.

07:41
from a capability standpoint. Yeah, and I think on that note,  let's talk about why MCP kind of changes enterprise AI economics. Sure, yes. ah I think this is where MCP becomes a board level discussion. ah Most enterprise AI budgets spiral for, I would say, three reasons. Integration labor, uh ongoing maintenance, and unpredictable inference costs.

08:13
MCP directly addresses all three. And I think that's something that as a business leader, you really need to take note of as well. Yeah, I agree. Costs are hugely important. Just to break that down. on the, on integration labor, MCP turns bespoke connector work into reusable infrastructure. So instead of building the same integration repeatedly for different models and different workflows, organizations, can

08:41
build it once with MCP and reuse it everywhere. Yeah, that's a, that you're spot on there. That's a point. And, and so then if you were to look at, you know, regarding maintenance, the curve fundamentally changes without MCP integration, complexity really grows exponentially with MCP. grows linearly that difference compounds year, year after year.

09:05
And the difference between AI as a perpetual cost center and AI as a scalable platform investment is an important note.  And that kind of really pivots into, I'd say token economics. I think we need to probably slow down and really talk about token economics because this is where enterprise, artificial intelligence costs quietly get out of control. ah

09:32
Many of you are probably still burned from the quiet cost of cloud many years ago. And most business leaders  really don't understand and realize it until the bill shows up. know, very similar It can be that budget killer. Yeah, yeah. And so you don't want to have a surprise. So every interaction with a large language model costs money based on tokens in and tokens out. And most of you are probably familiar with that.

10:00
tokens aren't just the words and the prompt or response. include things like system instructions, tool definitions, schema, conversation history, and really any of the data loaded into context. And I think this is where things go wrong, but it's also an awareness component because people that are using the environment are not aware of, you know, all of the backend information that I just talked about, you know, tool definitions, the schemas, the conversation history, for example.

10:30
So a lot of people just think, hey, it's just the input that I'm seeing on the screen. Exactly. Yeah. Then you come up with a misaligned budget. And I think in many, enterprise implementations today, the system loads everything into the model's context upfront. So dozens of tools, thousands of lines of schema, big chunks of data, whether they're relevant or not. So before the model even starts reasoning, you've already paid for a massive prompt. Yeah.

10:59
To your point. mean,  it's simple. Mat law  can be sometimes simple. Um, if  to really think about this, if one AI interaction uses,  say 20,000 tokens and it costs only a few cents. That doesn't sound like a large bill, a big deal, not a problem. But if the, if that interaction happens a hundred thousand times a month across employees.

11:22
your workflows, and maybe even some customer interactions, you're suddenly spending tens of thousands of dollars every month just on inference. And that's even before you even scale as a business within your artificial intelligence environment. Right. Yeah. So now MCP allows you to change that equation. So instead of loading everything into context, MCP enables selected discovery. Instead of 20,000 tokens, maybe you send

11:51
3000 or 4000 tokens. That's a, you know, an 80 % reduction in the context size for the same outcome. And then you can multiply that across your workflows. And to the listeners, when I said simple math, Scott just addressed it. Cause that, that 80 % matters. If your AI spend is $25,000 a month, reducing token usage by 80 % could take that down to $5,000 a month. I think I did the math, right?

12:18
That's in it. so a $240,000 annual savings, that's a huge swing. That's a quarter million dollars driven purely by infrastructure. Yeah. And then on the technical side, I think there's also latency to consider. larger prompts that you're loading through workloads, uh, take longer to process. So when responses slow down, adoption drops and when adoption drops, the ROI can, can become suddenly in question. So token efficiency isn't just about costs.

12:47
It's about making AI also feel fast and reliable. Yeah. Yeah. I can, I can just  imagine somebody saying this AI environment stinks. It's too slow, right? Um, it's kind of like bandwidth. So  it's a, it's a really good point. Um, I'd say another key point is where the work happens without MCP. We often ask the model to do things. It's not optimized for like reasoning over, you know, uh, thousands of database rows with MCP.

13:13
The model delegates that work into a tool and gets back a small clean result  of  an output, maybe 500 tokens instead of 50,000. So that's not just an optimization  scenario. That's a structural shift for the business.  And this is where token economics become an issue that leaders really need to understand.  Token costs will  scale with success. So the more you're

13:40
organization uses AI, the more you automate, the more value you create. And of course, then the more tokens you burn. So without a structural approach, leaders are forced to kind of slow down adoption just to control the costs and figure out what happened. Yeah. MCP flips that dynamic. makes AI usage predictable, controllable, and aligned with value creation.

14:08
you know, again, for business leaders, I think the boardroom takeaway that you could respond with is the simplest way to think about MCP is that it prevents AI success from becoming a cost problem by turning token spend from an unpredictable tax in the business into a controlled operating expense. And that's ultimately what changes the economics, but also provides awareness back into the board. Right. Yeah. No, that's a good point.

14:37
So I think let's just consider a little bit about uh the infrastructure reality and total cost of ownership.  when considering TCO, total cost of ownership, MCP adds infrastructure.  there's orchestration, state management, monitoring, governance. So it's not  simple, but the comparison isn't MCP versus nothing. It's MCP versus integration teams,  middleware platformers, manual workflows, and operational drag to try to

15:07
to try to make AI work.  When framed correctly, MCP often lowers the total cost while dramatically increasing  reliability and speed like we've already talked about.  Yep. Good point. And then if we were to kind of shift towards governance and risk and shadow MCP we talked about, I think this is where leadership discipline matters.  MCP gives  artificial intelligence the ability to act, not just answer.

15:36
action without governance introduces real risk. And those of you that have been in a governance role or have responsibility for it, you know what I'm talking about, because it's applied throughout your business tech for many years. And now we're hearing that shadow MCP is already starting to show up. So that is something that you want to get out in front of. Yep. Yep. We're already hearing about that. So teams will spin up MCP servers so that they can

16:04
You know, they can move fast and work on those projects, but without centralized oversight, organizations are going to lose  or they're already starting to lose visibility, access control and audit ability. So leaders need to be fully aware of what their teams are working on, including MCP and MCP servers uh and inspect everything accordingly. Yeah, that's a good one. Especially the audit trail to your point.  And enterprise MCP.

16:31
requires identity first design, role-based access control, logging, and centralized policy enforcement.  And this is where AI governance stops being theoretical that  I think a lot of organizations talk about and kind of mull around. And this is where it really becomes  operational and  has to be operational if it's not already.  So uh the final takeaways for business leaders and you all listening today, I think

16:59
You know, here's the bottom line, the competitive advantage in enterprise AI will not be the chosen model or models. In addition to selecting the right use cases and managing them, it will be how efficiently you integrate artificial intelligence in the business, how well you govern it, and how predictably you control the cost to prevent things like shadow costs we just talked about. Right. Yeah. I think, um, model context protocol, MCP kind of sits at the center of all three of those.

17:29
Yeah, that's, that's why MCP is something that business leaders should understand if they are planning  really for a serious enterprise AI environment. It is a component of the infrastructure for sure. Yeah. I think this is probably a good, point to wrap it up. So if this  episode gave you something to think about with respect to enterprise AI, please share it with your network. And as always, we'll keep rolling out the topics that are on the cutting edge of enterprise AI.

17:58
Thanks for listening to the Macri podcast. Please join us on our LinkedIn page, share the podcast with your peers. We're on all the major players such as Apple, Spotify, Amazon, iHeartRadio. Really appreciate your support and look forward to seeing you in the future. Thank you.