
Digital Transformation Playbook
Kieran Gilmurray is a globally recognised authority on Artificial Intelligence, cloud, intelligent automation, data analytics, agentic AI, and digital transformation. He has authored three influential books and hundreds of articles that have shaped industry perspectives on digital transformation, data analytics, intelligent automation, agentic AI and artificial intelligence.
๐ช๐ต๐ฎ๐ does Kieran doโ
When I'm not chairing international conferences, serving as a fractional CTO or Chief AI Officer, Iโm delivering AI, leadership, and strategy masterclasses to governments and industry leaders.
My team and I help global businesses drive AI, agentic ai, digital transformation and innovation programs that deliver tangible business results.
๐ ๐๐ฐ๐๐ซ๐๐ฌ:
๐นTop 25 Thought Leader Generative AI 2025
๐นTop 50 Global Thought Leaders and Influencers on Agentic AI 2025
๐นTop 100 Thought Leader Agentic AI 2025
๐นTop 100 Thought Leader Legal AI 2025
๐นTeam of the Year at the UK IT Industry Awards
๐นTop 50 Global Thought Leaders and Influencers on Generative AI 2024
๐นTop 50 Global Thought Leaders and Influencers on Manufacturing 2024
๐นBest LinkedIn Influencers Artificial Intelligence and Marketing 2024
๐นSeven-time LinkedIn Top Voice.
๐นTop 14 people to follow in data in 2023.
๐นWorld's Top 200 Business and Technology Innovators.
๐นTop 50 Intelligent Automation Influencers.
๐นTop 50 Brand Ambassadors.
๐นGlobal Intelligent Automation Award Winner.
๐นTop 20 Data Pros you NEED to follow.
๐๐ผ๐ป๐๐ฎ๐ฐ๐ my team and I to get business results, not excuses.
โ๏ธ https://calendly.com/kierangilmurray/30min
โ๏ธ kieran@gilmurray.co.uk
๐ www.KieranGilmurray.com
๐ Kieran Gilmurray | LinkedIn
Digital Transformation Playbook
Beyond Workflows: The Rise of AI Agents
Curious about what AI agents really are and how they're reshaping automation? This deep dive cuts through the jargon to deliver precisely what you need to know about these powerful systems built on large language models. Well listen to AI explain AI.
TLDR:
- Three key components make an AI agent: an LLM as its brain, tools to interact with the world, and guardrails to ensure appropriate behaviour
- AI agents excel where traditional rule-based systems fail: complex decision-making, overly complicated rules, and processing unstructured data
- Start building with the most capable model to prove your concept, then optimize later with smaller, faster, cheaper options if needed
- Tools come in three types: data tools for fetching information, action tools for doing things, and orchestration tools for calling other agents
- Clear instructions are vital - leverage existing SOPs, prompt the agent to break down tasks, and anticipate edge cases
- Begin with simple single-agent systems before moving to multi-agent approaches like the manager pattern or decentralized pattern
- Implement layered guardrails including relevance checks, safety classifiers, PII filters, moderation tools, and risk-based controls
- Human intervention remains critical, especially for high-risk actions or when the agent struggles with certain tasks
We explore how AI agents fundamentally differ from traditional software by independently accomplishing tasks through their LLM "brain," specialized tools, and carefully designed guardrails. Rather than just following rigid rules, these systems can reason through complex problems, adapt on the fly, and make nuanced judgment calls โ like having tiny specialized workers available 24/7.
You'll discover the three key scenarios where AI agents truly shine: handling complex decisions requiring judgment, replacing brittle rule systems that have become maintenance nightmares, and processing mountains of unstructured data. We break down the building blocks of effective agent design, from choosing the right model to crafting clear instructions and implementing proper safety mechanisms.
The conversation moves from simple single-agent systems to sophisticated multi-agent architectures, explaining when to use manager patterns versus decentralized approaches. We emphasize the critical importance of layered safety measures โ from privacy protections to content moderation โ and the continuing role of human oversight, especially for high-risk actions.
Whether you're just exploring the concept or actively looking to implement AI agents in your organization, this episode provides the clear, practical understanding you need to evaluate their potential and approach their development responsibly. The future of work is changing โ are you ready to rethink what automation can accomplish?
๐๐ผ๐ป๐๐ฎ๐ฐ๐ my team and I to get business results, not excuses.
โ๏ธ https://calendly.com/kierangilmurray/results-not-excuses
โ๏ธ kieran@gilmurray.co.uk
๐ www.KieranGilmurray.com
๐ Kieran Gilmurray | LinkedIn
๐ฆ X / Twitter: https://twitter.com/KieranGilmurray
๐ฝ YouTube: https://www.youtube.com/@KieranGilmurray
Hi there. You're probably here because, like us, you want to get a real handle on something important you know, without wading through endless articles and jargon.
AI Voice Two:Yeah, cut right to the chase.
AI Voice One:Exactly, and today we're diving deep into AI agents. We've gathered a bunch of info that really paints a picture of these systems, the ones powered by large language models, or LLMs.
AI Voice Two:Right.
AI Voice One:And they seem like more than just you know clever software. This could genuinely be a fundamental shift in automation.
AI Voice Two:It really could. A move beyond the workflows you have to manually kick off.
AI Voice One:But something that can, what act independently on your behalf.
AI Voice Two:Pretty much. Our mission today, then, is to pull out the core knowledge you really need to understand this potential game changer.
AI Voice One:Okay, so let's start there. An AI agent How's it different from software that helps me do something?
AI Voice Two:Ah, good question. Well, the sources we looked at define an agent as a system specifically designed to independently accomplish tasks. It's about delegating entire processes, not just like individual steps.
AI Voice One:Independently accomplishing tasks, yeah, okay. So what makes it an agent, then, rather than just a really fancy program that happens to use an LLM? What are the essential ingredients?
AI Voice Two:The material consistently points to three key things. First, the agent uses an LLM as its core. Like its brain, its reasoning engine, it actively manages a workflow and makes decisions as it goes. So it's not just spitting out text, it's reasoning engine, it actively manages a workflow and makes decisions as it goes.
AI Voice One:So it's not just spitting out text, it's directing things.
AI Voice Two:Exactly Orchestrating actions. Second, it needs access to what are called tools.
AI Voice One:Tools Like software tools.
AI Voice Two:Sort of yeah, Think of them like extensions or plugins. They let the agent interact with the outside world, query a database, send an email, search the web, that kind of thing.
AI Voice One:Gotcha, so it can actually do stuff.
AI Voice Two:Right. And third, and this is crucial, its operation is governed by defined guardrails, instructions and boundaries to make sure it behaves acceptably.
AI Voice One:Okay, LLM brain tools for action and guardrails for safety Makes sense, but when would you actually go through the effort of building one? It sounds like a bigger deal than standard automation.
AI Voice Two:It definitely can be, and that's a really important question. The sources address Agents truly shine where traditional, like rule-based automation starts hitting its limits. Limits like what Well take payment fraud analysis, for instance. A standard system might just flag transactions matching very specific preset rules, bang rule triggered. But an AI agent, it can reason through the context. It can look at subtle indicators, things that don't fit a neat rule, and make a more nuanced judgment. It's almost like having a tiny fraud investigator working 24-7.
AI Voice One:Ah, I see. So it's less about rigid if this than that and more about understanding the bigger picture.
AI Voice Two:Exactly, it moves beyond those brittle rules towards something more flexible, almost intuitive, you could say.
AI Voice One:So are there specific areas where this really pays off, signs that an agent might be the way to go?
AI Voice Two:Yeah. The material highlights three main value areas. First is complex decision making. You know workflows needing judgment calls, handling weird exceptions, adapting on the fly, like approving a tricky customer refund.
AI Voice One:Right where it's not just black and white.
AI Voice Two:Precisely. Second, situations where your rules have become insanely complicated and a nightmare to maintain Think vendor security reviews with thousands of branching rules.
AI Voice One:Oh yeah, I can imagine.
AI Voice Two:And third is when you're drowning in unstructured data, like sifting through thousands of customer emails written in natural language or pulling key facts from messy insurance claim documents.
AI Voice One:Okay, complex decisions, hard to maintain rules or lots of unstructured data.
AI Voice Two:If your problem ticks one or more of those boxes, an agent is definitely worth considering.
AI Voice One:Right. So okay, let's say you've identified a good use case. Where do you start designing one? What are those core building blocks?
AI Voice Two:again, so back to those three core components. We mentioned First the model, the LLM itself.
AI Voice One:The brain.
AI Voice Two:The brain. Yeah, and different models have different strengths, right. Some are better at complex reasoning, some are faster, some are cheaper.
AI Voice One:So how do you choose?
AI Voice Two:Well, the common advice seems to be start prototyping with the most capable model you can get access to. Really push the boundaries, see what's possible.
AI Voice One:Prove the concept first.
AI Voice Two:Exactly. Then, once you've got something working, you can experiment, try smaller, faster, cheaper models and see if the performance is still good enough for your specific needs. Optimization comes later.
AI Voice One:Smart, Prove it, then refine it. Component one the model. What was number two?
AI Voice Two:The tools. These are those external functions or APIs application programming interfaces that let the agent interact with the world outside the LLM.
AI Voice One:The hands, basically the hands.
AI Voice Two:yeah, that's a good way to put it. The sources break them down into roughly three types. You've got data tools for fetching info, querying databases, reading files, searching the web.
AI Voice One:Okay.
AI Voice Two:Then action tools for doing things sending emails, updating Salesforce records, creating support tickets.
AI Voice One:Makes sense.
AI Voice Two:And interestingly, there are also orchestration tools where one agent can actually call another agent as one of its tools to handle a subtask.
AI Voice One:Whoa agents using other agents? Okay, meta.
AI Voice Two:It can get pretty sophisticated. The point is equipping the agent with exactly the capabilities it needs for its job.
AI Voice One:Got it Model tools and the third piece was instructions.
AI Voice Two:Instructions yes, these are the explicit guidelines and the guardrails that define how the agent should behave. Think of it as the agent's rulebook or standard operating procedure.
AI Voice One:And getting these right sounds critical.
AI Voice Two:Absolutely vital. Clear instructions reduce ambiguity, improve the quality of the agent's decisions and prevent it from going off the rails.
AI Voice One:So how do you write good instructions for an AI? It can be quite like writing an email to a colleague, right?
AI Voice Two:Not quite. No, the sources suggest starting with what you already have existing standard operating procedures, maybe customer support scripts, internal wikis.
AI Voice One:Leverage existing knowledge Exactly.
AI Voice Two:It's also really helpful to prompt the agent itself to break down big tasks into smaller steps Like okay, outline the steps you'd take to resolve this issue.
AI Voice One:Ah, make it think about its own process.
AI Voice Two:Yes, and for each step you need to define a really clear action or outcome. Minimize wiggle room and this is key. Anticipate the weird stuff, the edge cases. What happens if the database is down? What if the customer gives contradictory information? You need instructions for that.
AI Voice One:Plan for the unexpected.
AI Voice Two:You have to. Interestingly, the sources even mention using other advanced LLMs to help generate the initial set of instructions by feeding them your existing documents. There is even an example prompt for doing that.
AI Voice One:Using AI to bootstrap the instructions for another AI. That's efficient, I guess. It's a potential accelerator for sure. Okay, so you've got your model, your tools, your carefully crafted instructions. How do you actually make the agent you know run? How does it execute a workflow? This is orchestration, right.
AI Voice Two:Precisely. Orchestration is all about the patterns and strategies that let the agent follow those instruction and use its tools effectively to reach the goal and where do you start?
AI Voice One:seems like it could get complicated fast it can.
AI Voice Two:The advice is generally to start simple, usually with what's called a single agent system meaning, just one agent does everything well, one primary agent manages the whole process.
AI Voice Two:It might have lots of tools, but it's one central brain coordinating things. It runs in a loop. Basically, yeah, think of it as read the instructions, figure out the next step, maybe use a tool, get the result, figure out the next step, and so on. This run keeps going until a specific condition is met, like what maybe the agent calls a specific task, complete tool, or it generates the final output you wanted, or maybe it hits an error it can't resolve, or, importantly, it might hit a maximum number of turns or steps to prevent it from just running forever a safety mechanism.
AI Voice Two:Definitely the material actually mentioned a function like runner dot run from something called the agents SDK, a software development kit for building these. Think of that as the go button for the agent's loop.
AI Voice One:Okay, and if that single agent have like dozens of tools and complex logic, how to keep that manageable?
AI Voice Two:Ah, good point. Prompt templates are apparently very useful here. Instead of writing unique instructions for every tiny variation, you create a template with placeholders, variables.
AI Voice One:Like a fill in the blanks prompt.
AI Voice Two:Exactly so for a call center agent. You might have variables for customer name accountage issue type. You fill those in based on the current situation. It makes the core instructions much easier to manage and scale.
AI Voice One:Makes sense. Reuse the core logic.
AI Voice Two:Yeah, and the sources generally advise pushing that single agent approach as far as you can before jumping to multiple agents.
AI Voice One:Why is that?
AI Voice Two:Because coordinating multiple agents just adds another layer of complexity. You'd only really move to multi-agent systems if the logic gets super tangled or if the single agent has so many tools it keeps picking the wrong one, you know.
AI Voice One:Okay, so only add complexity when you really have to yeah, but if you do need more than one agent, what then? That's multi-agent systems right.
AI Voice Two:This is where you break down the workflow and have several agents collaborating. The sources focus on two main patterns here okay, pattern one is manager pattern. Imagine a central manager agent acting like a project lead. It doesn't do all the work itself, instead, it directs traffic. It calls on specialized worker agents using tools. Hey translation agent, translate this to Spanish. Hey database agent fetch this customer record.
AI Voice One:So the worker agents were basically tools for the manager agent Pretty much the manager assigns tasks, collects the results from the workers agent.
AI Voice Two:Pretty much the manager assigns tasks, collects the results from the workers and then synthesizes the final output or decides the next overall step. The example given was that translation scenario a manager using separate Spanish, french, italian agents.
AI Voice One:Got it Like an orchestra conductor, making sure everyone plays their part.
AI Voice Two:It's a perfect analogy the manager keeps control. The sources did mention a contrast here with some visual flowchart style builders saying that, while those look clear, a code first approach, like with the agent SDK, might offer more flexibility for these complex interactions.
AI Voice One:Interesting trade-off. Okay, so manager pattern is one. What's the other? Big one?
AI Voice Two:The other is the decentralized pattern. Here agents act more like peers on a team. They hand off tasks directly to each other, based on specialization.
AI Voice One:So no central manager.
AI Voice Two:Not really. No, it's more like an assembly line or a relay race. An agent finishes its part and then uses a specific tool or function to pass the whole task onto the next appropriate specialist agent.
AI Voice One:And it's usually a one-way handoff.
AI Voice Two:Typically yeah. Once Agent A hands off to Agent B, Agent B takes over. The example used was a customer service flow.
AI Voice One:How did that work?
AI Voice Two:Well, you might have a general triage agent that first talks to the customer, Based on the issue. It might hand off to a technical support agent or sales agent or an order management agent.
AI Voice One:Ah. Routing based on need.
AI Voice Two:Exactly. Each specialist handles their piece. This pattern is apparently really good for that kind of conversation routing or task triage.
AI Voice One:I guess you're building a team of specialists. Okay, but with all those power agents making decisions, taking actions, potentially using other agents, how do you keep them from messing up or doing things they shouldn't? Guardrails right.
AI Voice Two:Absolutely critical. Guardrails are your safety net. You're managing risks like exposing private data, saying something off-brand or just making bad decisions. Think of them like safety features on heavy machinery.
AI Voice One:And it's not just one big stop button.
AI Voice Two:No, the sources really emphasize a layered defense, multiple types of guardrails working together.
AI Voice One:Okay, like what? Give me some examples.
AI Voice Two:Sure, you might have a relevance classifier that flags if a user asks the agent something totally unrelated to its job.
AI Voice One:Keep it on topic.
AI Voice Two:Right, A safety classifier to detect harmful inputs. People trying to jailbreak the agent or feed it malicious instructions.
AI Voice One:I'm taking the agent itself.
AI Voice Two:Exactly. Then things like a PII filter to stop the agent from unnecessarily asking for or revealing personal info like credit card numbers.
AI Voice One:Privacy protection Crucial.
AI Voice Two:Very. Also moderation tools to check the agent's output for harmful or inappropriate content before it reaches the user.
AI Voice One:So checking both input and output, yes, you can also have tool safeguards.
AI Voice Two:Maybe certain tools are riskier, like delete customer account. You could rate that tool as high risk, triggering extra checks or even needing human approval before the agent can use it. Smart.
AI Voice One:Risk-based controls.
AI Voice Two:And then there are more traditional things too Simple, rules-based protections like block lists for certain words, limits on input length, using rejects patterns to validate formats and, finally, output validation, just to ensure the agent's tone and style match your brand voice.
AI Voice One:Wow, that's quite a few layers. How do you decide where to focus? You can't build all of that on day one, surely?
AI Voice Two:No, probably not. The guidance suggests this pragmatic approach. Start by focusing on the big risks privacy and basic safety. Get those fundamentals in place, then add more specific guardrails reactively, based on actual failures or near misses. You see, when testing or deploying the agent, learn from experience.
AI Voice One:Let reality guide the hardening process.
AI Voice Two:Pretty much it's a continuous balancing act between security and making sure the agent is still useful and not annoying to interact with. The material showed a code snippet using the agent's SDK for an input guardrail, specifically detecting if a customer seems likely to churn.
AI Voice One:And how did that work?
AI Voice Two:It used an optimistic execution approach. The main agent process would continue, but in the background, this guardrail would analyze the input for churn signals. If detected, it could trigger a specific action, like alerting a human retention specialist.
AI Voice One:So the guardrail runs in parallel, potentially.
AI Voice Two:In that example. Yes, it avoids blocking the main flow unless necessary.
AI Voice One:Okay, but even with all these automated checks, is there still a place for a human in the loop?
AI Voice Two:Oh, absolutely. Human intervention is highlighted as a critical safeguard, especially early on.
AI Voice One:Why especially early on?
AI Voice Two:Well, it helps you catch those unforeseen issues, discover edge cases you didn't anticipate in your instructions and just generally build confidence in the agent's performance before you let it run completely free.
AI Voice One:Makes sense Train it with supervision first.
AI Voice Two:Right, and the sources point to two main triggers for pulling a human in. First, if the agent starts failing too often, maybe it exceeds a certain threshold for errors or retries on a task.
AI Voice One:Too many mistakes Call for help.
AI Voice Two:Exactly. And second, when the agent is about to perform a particularly high-risk action we mentioned deleting an account, maybe issuing a large refund or sending a critical communication. For those kinds of things, having a human review and give the final okay is often the safest bet.
AI Voice One:Better safe than sorry, especially with high stakes of things, having a human review and give the final OK is often the safest bet. Better safe than sorry, especially with high stakes. Ok, so let's try and wrap this up. If we boil it all down, what's the main thing people should take away about AI agents from this deep dive?
AI Voice Two:I think the core message is that AI agents are a significant step up in automation. They're not just about making existing processes faster. They enable automation of complex, multi-step tasks that require judgment and interaction with the world in ways that, frankly, older software just couldn't handle.
AI Voice One:And they're especially good for.
AI Voice Two:For those really tricky workflows, the ones involving complex decisions, messy, unstructured data or those brittle, hard-to-maintain rule systems we talked about. That's where they can be transformative.
AI Voice One:And building them reliably means.
AI Voice Two:It means focusing on those foundations the right model, the right tools and crystal-clear instructions. Then choosing the right orchestration patterns. Start simple, scale up carefully and, crucially, layering in those robust guardrails to manage the risks Safety, privacy, reliability they're paramount.
AI Voice One:Right, so for you listening. Hopefully that gives you a much clearer picture of what AI agents are, where they might fit and what it takes to build them effectively and responsibly.
AI Voice Two:Yeah, the potential is definitely there.
AI Voice One:It really is. And it leads to a final thought, I suppose as these agents become more common, more integrated, how is that going to change our basic ideas about what work even means or what assistance looks like?
AI Voice Two:That's a big question.
AI Voice One:It is Definitely something to chew on. Well, thanks for joining us for this deep dive.
AI Voice Two:My pleasure.