The AI Argument

AI Agents Under Fire, LLM Bias Runs Deep, and a Wizard of Oz Fail: The AI Argument EP68

Frank Prendergast and Justin Collery

AI agents crumble faster than wet cardboard when under attack. A recent study proved it. Every single agent tested failed against prompt injections. That’s a 100% failure rate. 

Justin sees this as a fixable engineering problem with smart design and strict access controls. 

Frank isn’t convinced. Real-world complexity means isolation isn’t that simple. 

And while Justin rails against regulation, Frank points to the EU’s looming rules as a possible safety net.

The bigger takeaway? Businesses racing to deploy open-ended agents could be building ticking time bombs. The safer bet might be narrow, well-scoped agents that automate specific tasks. But will hype win over common sense?

From there, the debate shifts to a study exposing bias in LLMs. It found they recommend lower salaries for women and minority groups. Can removing personal details fix the problem, or is the bias baked in? 

Then it takes a technical turn with Chinese researchers using LLMs to design stronger models, before veering into the unexpected: a football club handing legal contracts to AI and a Wizard of Oz remake that left Vegas audiences unimpressed.

02:12 Can any AI agent survive a prompt attack?
14:51 Is AI quietly spreading bias everywhere?
25:19 Are LLMs now designing better LLMs?
29:32 Did United just make AI their star player?
31:13 Did AI butcher the Wizard of Oz in Vegas?

► LINKS TO CONTENT WE DISCUSSED


► CONNECT WITH US
For more in-depth discussions, connect Justin and Frank on LinkedIn.
Justin: https://www.linkedin.com/in/justincollery/
Frank: https://www.linkedin.com/in/frankprendergast/