The Digital Transformation Playbook

Beyond Workflows: The Rise of AI Agents

โ€ข Kieran Gilmurray

Use Left/Right to seek, Home/End to jump to start or end. Hold shift to jump forward or backward.

0:00 | 16:20

Curious about what AI agents really are and how they're reshaping automation? This deep dive cuts through the jargon to deliver precisely what you need to know about these powerful systems built on large language models. Well listen to AI explain AI.

TLDR:

  • Three key components make an AI agent: an LLM as its brain, tools to interact with the world, and guardrails to ensure appropriate behaviour
  • AI agents excel where traditional rule-based systems fail: complex decision-making, overly complicated rules, and processing unstructured data
  • Start building with the most capable model to prove your concept, then optimize later with smaller, faster, cheaper options if needed
  • Tools come in three types: data tools for fetching information, action tools for doing things, and orchestration tools for calling other agents
  • Clear instructions are vital - leverage existing SOPs, prompt the agent to break down tasks, and anticipate edge cases
  • Begin with simple single-agent systems before moving to multi-agent approaches like the manager pattern or decentralized pattern
  • Implement layered guardrails including relevance checks, safety classifiers, PII filters, moderation tools, and risk-based controls
  • Human intervention remains critical, especially for high-risk actions or when the agent struggles with certain tasks

We explore how AI agents fundamentally differ from traditional software by independently accomplishing tasks through their LLM "brain," specialized tools, and carefully designed guardrails. Rather than just following rigid rules, these systems can reason through complex problems, adapt on the fly, and make nuanced judgment calls โ€“ like having tiny specialized workers available 24/7.

You'll discover the three key scenarios where AI agents truly shine: handling complex decisions requiring judgment, replacing brittle rule systems that have become maintenance nightmares, and processing mountains of unstructured data. We break down the building blocks of effective agent design, from choosing the right model to crafting clear instructions and implementing proper safety mechanisms.

The conversation moves from simple single-agent systems to sophisticated multi-agent architectures, explaining when to use manager patterns versus decentralized approaches. We emphasize the critical importance of layered safety measures โ€“ from privacy protections to content moderation โ€“ and the continuing role of human oversight, especially for high-risk actions.

Whether you're just exploring the concept or actively looking to implement AI agents in your organization, this episode provides the clear, practical understanding you need to evaluate their potential and approach their development responsibly. The future of work is changing โ€“ are you ready to rethink what automation can accomplish?

Support the show


๐—–๐—ผ๐—ป๐˜๐—ฎ๐—ฐ๐˜ my team and I to get business results, not excuses.

โ˜Ž๏ธ https://calendly.com/kierangilmurray/results-not-excuses
โœ‰๏ธ kieran@gilmurray.co.uk
๐ŸŒ www.KieranGilmurray.com
๐Ÿ“˜ Kieran Gilmurray | LinkedIn
๐Ÿฆ‰ X / Twitter: https://twitter.com/KieranGilmurray
๐Ÿ“ฝ YouTube: https://www.youtube.com/@KieranGilmurray

๐Ÿ“• Want to learn more about agentic AI then read my new book on Agentic AI and the Future of Work https://tinyurl.com/MyBooksOnAmazonUK


Introduction to AI Agents

AI Voice One

Hi there. You're probably here because, like us, you want to get a real handle on something important you know, without wading through endless articles and jargon.

AI Voice Two

Yeah, cut right to the chase.

AI Voice One

Exactly, and today we're diving deep into AI agents. We've gathered a bunch of info that really paints a picture of these systems, the ones powered by large language models, or LLMs.

AI Voice Two

Right.

AI Voice One

And they seem like more than just you know clever software. This could genuinely be a fundamental shift in automation.

AI Voice Two

It really could. A move beyond the workflows you have to manually kick off.

AI Voice One

But something that can, what act independently on your behalf.

AI Voice Two

Pretty much. Our mission today, then, is to pull out the core knowledge you really need to understand this potential game changer.

AI Voice One

Okay, so let's start there. An AI agent How's it different from software that helps me do something?

AI Voice Two

Ah, good question. Well, the sources we looked at define an agent as a system specifically designed to independently accomplish tasks. It's about delegating entire processes, not just like individual steps.

AI Voice One

Independently accomplishing tasks, yeah, okay. So what makes it an agent, then, rather than just a really fancy program that happens to use an LLM? What are the essential ingredients?

AI Voice Two

The material consistently points to three key things. First, the agent uses an LLM as its core. Like its brain, its reasoning engine, it actively manages a workflow and makes decisions as it goes. So it's not just spitting out text, it's reasoning engine, it actively manages a workflow and makes decisions as it goes.

AI Voice One

So it's not just spitting out text, it's directing things.

AI Voice Two

Exactly Orchestrating actions. Second, it needs access to what are called tools.

AI Voice One

Tools Like software tools.

AI Voice Two

Sort of yeah, Think of them like extensions or plugins. They let the agent interact with the outside world, query a database, send an email, search the web, that kind of thing.

AI Voice One

Gotcha, so it can actually do stuff.

Core Components of AI Agents

AI Voice Two

Right. And third, and this is crucial, its operation is governed by defined guardrails, instructions and boundaries to make sure it behaves acceptably.

AI Voice One

Okay, LLM brain tools for action and guardrails for safety Makes sense, but when would you actually go through the effort of building one? It sounds like a bigger deal than standard automation.

AI Voice Two

It definitely can be, and that's a really important question. The sources address Agents truly shine where traditional, like rule-based automation starts hitting its limits. Limits like what Well take payment fraud analysis, for instance. A standard system might just flag transactions matching very specific preset rules, bang rule triggered. But an AI agent, it can reason through the context. It can look at subtle indicators, things that don't fit a neat rule, and make a more nuanced judgment. It's almost like having a tiny fraud investigator working 24-7.

AI Voice One

Ah, I see. So it's less about rigid if this than that and more about understanding the bigger picture.

AI Voice Two

Exactly, it moves beyond those brittle rules towards something more flexible, almost intuitive, you could say.

AI Voice One

So are there specific areas where this really pays off, signs that an agent might be the way to go?

When to Use AI Agents

AI Voice Two

Yeah. The material highlights three main value areas. First is complex decision making. You know workflows needing judgment calls, handling weird exceptions, adapting on the fly, like approving a tricky customer refund.

AI Voice One

Right where it's not just black and white.

AI Voice Two

Precisely. Second, situations where your rules have become insanely complicated and a nightmare to maintain Think vendor security reviews with thousands of branching rules.

AI Voice One

Oh yeah, I can imagine.

AI Voice Two

And third is when you're drowning in unstructured data, like sifting through thousands of customer emails written in natural language or pulling key facts from messy insurance claim documents.

AI Voice One

Okay, complex decisions, hard to maintain rules or lots of unstructured data.

AI Voice Two

If your problem ticks one or more of those boxes, an agent is definitely worth considering.

AI Voice One

Right. So okay, let's say you've identified a good use case. Where do you start designing one? What are those core building blocks?

AI Voice Two

again, so back to those three core components. We mentioned First the model, the LLM itself.

AI Voice One

The brain.

AI Voice Two

The brain. Yeah, and different models have different strengths, right. Some are better at complex reasoning, some are faster, some are cheaper.

AI Voice One

So how do you choose?

AI Voice Two

Well, the common advice seems to be start prototyping with the most capable model you can get access to. Really push the boundaries, see what's possible.

AI Voice One

Prove the concept first.

AI Voice Two

Exactly. Then, once you've got something working, you can experiment, try smaller, faster, cheaper models and see if the performance is still good enough for your specific needs. Optimization comes later.

AI Voice One

Smart, Prove it, then refine it. Component one the model. What was number two?

AI Voice Two

The tools. These are those external functions or APIs application programming interfaces that let the agent interact with the world outside the LLM.

AI Voice One

The hands, basically the hands.

AI Voice Two

yeah, that's a good way to put it. The sources break them down into roughly three types. You've got data tools for fetching info, querying databases, reading files, searching the web.

AI Voice One

Okay.

AI Voice Two

Then action tools for doing things sending emails, updating Salesforce records, creating support tickets.

AI Voice One

Makes sense.

AI Voice Two

And interestingly, there are also orchestration tools where one agent can actually call another agent as one of its tools to handle a subtask.

AI Voice One

Whoa agents using other agents? Okay, meta.

AI Voice Two

It can get pretty sophisticated. The point is equipping the agent with exactly the capabilities it needs for its job.

Building Blocks: Models and Tools

AI Voice One

Got it Model tools and the third piece was instructions.

AI Voice Two

Instructions yes, these are the explicit guidelines and the guardrails that define how the agent should behave. Think of it as the agent's rulebook or standard operating procedure.

AI Voice One

And getting these right sounds critical.

AI Voice Two

Absolutely vital. Clear instructions reduce ambiguity, improve the quality of the agent's decisions and prevent it from going off the rails.

AI Voice One

So how do you write good instructions for an AI? It can be quite like writing an email to a colleague, right?

AI Voice Two

Not quite. No, the sources suggest starting with what you already have existing standard operating procedures, maybe customer support scripts, internal wikis.

AI Voice One

Leverage existing knowledge Exactly.

AI Voice Two

It's also really helpful to prompt the agent itself to break down big tasks into smaller steps Like okay, outline the steps you'd take to resolve this issue.

AI Voice One

Ah, make it think about its own process.

AI Voice Two

Yes, and for each step you need to define a really clear action or outcome. Minimize wiggle room and this is key. Anticipate the weird stuff, the edge cases. What happens if the database is down? What if the customer gives contradictory information? You need instructions for that.

AI Voice One

Plan for the unexpected.

AI Voice Two

You have to. Interestingly, the sources even mention using other advanced LLMs to help generate the initial set of instructions by feeding them your existing documents. There is even an example prompt for doing that.

Instructions and Orchestration

AI Voice One

Using AI to bootstrap the instructions for another AI. That's efficient, I guess. It's a potential accelerator for sure. Okay, so you've got your model, your tools, your carefully crafted instructions. How do you actually make the agent you know run? How does it execute a workflow? This is orchestration, right.

AI Voice Two

Precisely. Orchestration is all about the patterns and strategies that let the agent follow those instruction and use its tools effectively to reach the goal and where do you start?

AI Voice One

seems like it could get complicated fast it can.

AI Voice Two

The advice is generally to start simple, usually with what's called a single agent system meaning, just one agent does everything well, one primary agent manages the whole process.

AI Voice Two

It might have lots of tools, but it's one central brain coordinating things. It runs in a loop. Basically, yeah, think of it as read the instructions, figure out the next step, maybe use a tool, get the result, figure out the next step, and so on. This run keeps going until a specific condition is met, like what maybe the agent calls a specific task, complete tool, or it generates the final output you wanted, or maybe it hits an error it can't resolve, or, importantly, it might hit a maximum number of turns or steps to prevent it from just running forever a safety mechanism.

AI Voice Two

Definitely the material actually mentioned a function like runner dot run from something called the agents SDK, a software development kit for building these. Think of that as the go button for the agent's loop.

AI Voice One

Okay, and if that single agent have like dozens of tools and complex logic, how to keep that manageable?

AI Voice Two

Ah, good point. Prompt templates are apparently very useful here. Instead of writing unique instructions for every tiny variation, you create a template with placeholders, variables.

AI Voice One

Like a fill in the blanks prompt.

AI Voice Two

Exactly so for a call center agent. You might have variables for customer name accountage issue type. You fill those in based on the current situation. It makes the core instructions much easier to manage and scale.

AI Voice One

Makes sense. Reuse the core logic.

AI Voice Two

Yeah, and the sources generally advise pushing that single agent approach as far as you can before jumping to multiple agents.

AI Voice One

Why is that?

AI Voice Two

Because coordinating multiple agents just adds another layer of complexity. You'd only really move to multi-agent systems if the logic gets super tangled or if the single agent has so many tools it keeps picking the wrong one, you know.

AI Voice One

Okay, so only add complexity when you really have to yeah, but if you do need more than one agent, what then? That's multi-agent systems right.

Multi-Agent Systems and Patterns

AI Voice Two

This is where you break down the workflow and have several agents collaborating. The sources focus on two main patterns here okay, pattern one is manager pattern. Imagine a central manager agent acting like a project lead. It doesn't do all the work itself, instead, it directs traffic. It calls on specialized worker agents using tools. Hey translation agent, translate this to Spanish. Hey database agent fetch this customer record.

AI Voice One

So the worker agents were basically tools for the manager agent Pretty much the manager assigns tasks, collects the results from the workers agent.

AI Voice Two

Pretty much the manager assigns tasks, collects the results from the workers and then synthesizes the final output or decides the next overall step. The example given was that translation scenario a manager using separate Spanish, french, italian agents.

AI Voice One

Got it Like an orchestra conductor, making sure everyone plays their part.

AI Voice Two

It's a perfect analogy the manager keeps control. The sources did mention a contrast here with some visual flowchart style builders saying that, while those look clear, a code first approach, like with the agent SDK, might offer more flexibility for these complex interactions.

AI Voice One

Interesting trade-off. Okay, so manager pattern is one. What's the other? Big one?

AI Voice Two

The other is the decentralized pattern. Here agents act more like peers on a team. They hand off tasks directly to each other, based on specialization.

AI Voice One

So no central manager.

AI Voice Two

Not really. No, it's more like an assembly line or a relay race. An agent finishes its part and then uses a specific tool or function to pass the whole task onto the next appropriate specialist agent.

AI Voice One

And it's usually a one-way handoff.

AI Voice Two

Typically yeah. Once Agent A hands off to Agent B, Agent B takes over. The example used was a customer service flow.

AI Voice One

How did that work?

AI Voice Two

Well, you might have a general triage agent that first talks to the customer, Based on the issue. It might hand off to a technical support agent or sales agent or an order management agent.

AI Voice One

Ah. Routing based on need.

AI Voice Two

Exactly. Each specialist handles their piece. This pattern is apparently really good for that kind of conversation routing or task triage.

Implementing Strong Guardrails

AI Voice One

I guess you're building a team of specialists. Okay, but with all those power agents making decisions, taking actions, potentially using other agents, how do you keep them from messing up or doing things they shouldn't? Guardrails right.

AI Voice Two

Absolutely critical. Guardrails are your safety net. You're managing risks like exposing private data, saying something off-brand or just making bad decisions. Think of them like safety features on heavy machinery.

AI Voice One

And it's not just one big stop button.

AI Voice Two

No, the sources really emphasize a layered defense, multiple types of guardrails working together.

AI Voice One

Okay, like what? Give me some examples.

AI Voice Two

Sure, you might have a relevance classifier that flags if a user asks the agent something totally unrelated to its job.

AI Voice One

Keep it on topic.

AI Voice Two

Right, A safety classifier to detect harmful inputs. People trying to jailbreak the agent or feed it malicious instructions.

AI Voice One

I'm taking the agent itself.

AI Voice Two

Exactly. Then things like a PII filter to stop the agent from unnecessarily asking for or revealing personal info like credit card numbers.

AI Voice One

Privacy protection Crucial.

AI Voice Two

Very. Also moderation tools to check the agent's output for harmful or inappropriate content before it reaches the user.

AI Voice One

So checking both input and output, yes, you can also have tool safeguards.

AI Voice Two

Maybe certain tools are riskier, like delete customer account. You could rate that tool as high risk, triggering extra checks or even needing human approval before the agent can use it. Smart.

AI Voice One

Risk-based controls.

AI Voice Two

And then there are more traditional things too Simple, rules-based protections like block lists for certain words, limits on input length, using rejects patterns to validate formats and, finally, output validation, just to ensure the agent's tone and style match your brand voice.

AI Voice One

Wow, that's quite a few layers. How do you decide where to focus? You can't build all of that on day one, surely?

AI Voice Two

No, probably not. The guidance suggests this pragmatic approach. Start by focusing on the big risks privacy and basic safety. Get those fundamentals in place, then add more specific guardrails reactively, based on actual failures or near misses. You see, when testing or deploying the agent, learn from experience.

AI Voice One

Let reality guide the hardening process.

AI Voice Two

Pretty much it's a continuous balancing act between security and making sure the agent is still useful and not annoying to interact with. The material showed a code snippet using the agent's SDK for an input guardrail, specifically detecting if a customer seems likely to churn.

AI Voice One

And how did that work?

AI Voice Two

It used an optimistic execution approach. The main agent process would continue, but in the background, this guardrail would analyze the input for churn signals. If detected, it could trigger a specific action, like alerting a human retention specialist.

AI Voice One

So the guardrail runs in parallel, potentially.

AI Voice Two

In that example. Yes, it avoids blocking the main flow unless necessary.

AI Voice One

Okay, but even with all these automated checks, is there still a place for a human in the loop?

AI Voice Two

Oh, absolutely. Human intervention is highlighted as a critical safeguard, especially early on.

AI Voice One

Why especially early on?

AI Voice Two

Well, it helps you catch those unforeseen issues, discover edge cases you didn't anticipate in your instructions and just generally build confidence in the agent's performance before you let it run completely free.

AI Voice One

Makes sense Train it with supervision first.

AI Voice Two

Right, and the sources point to two main triggers for pulling a human in. First, if the agent starts failing too often, maybe it exceeds a certain threshold for errors or retries on a task.

AI Voice One

Too many mistakes Call for help.

AI Voice Two

Exactly. And second, when the agent is about to perform a particularly high-risk action we mentioned deleting an account, maybe issuing a large refund or sending a critical communication. For those kinds of things, having a human review and give the final okay is often the safest bet.

Conclusion and Future Implications

AI Voice One

Better safe than sorry, especially with high stakes of things, having a human review and give the final OK is often the safest bet. Better safe than sorry, especially with high stakes. Ok, so let's try and wrap this up. If we boil it all down, what's the main thing people should take away about AI agents from this deep dive?

AI Voice Two

I think the core message is that AI agents are a significant step up in automation. They're not just about making existing processes faster. They enable automation of complex, multi-step tasks that require judgment and interaction with the world in ways that, frankly, older software just couldn't handle.

AI Voice One

And they're especially good for.

AI Voice Two

For those really tricky workflows, the ones involving complex decisions, messy, unstructured data or those brittle, hard-to-maintain rule systems we talked about. That's where they can be transformative.

AI Voice One

And building them reliably means.

AI Voice Two

It means focusing on those foundations the right model, the right tools and crystal-clear instructions. Then choosing the right orchestration patterns. Start simple, scale up carefully and, crucially, layering in those robust guardrails to manage the risks Safety, privacy, reliability they're paramount.

AI Voice One

Right, so for you listening. Hopefully that gives you a much clearer picture of what AI agents are, where they might fit and what it takes to build them effectively and responsibly.

AI Voice Two

Yeah, the potential is definitely there.

AI Voice One

It really is. And it leads to a final thought, I suppose as these agents become more common, more integrated, how is that going to change our basic ideas about what work even means or what assistance looks like?

AI Voice Two

That's a big question.

AI Voice One

It is Definitely something to chew on. Well, thanks for joining us for this deep dive.

AI Voice Two

My pleasure.