The Macro AI Podcast

Taylor Swift, AI Clones, and the Future of Human Identity

Season 2 Episode 78

Use Left/Right to seek, Home/End to jump to start or end. Hold shift to jump forward or backward.

0:00 | 15:27

Fresh in the headlines, Taylor Swift is reportedly taking aggressive legal steps to protect her voice, likeness, and digital identity from AI replication. But is this really just a celebrity story—or is it the beginning of a much larger transformation in business, law, and society? 

In this episode of the Macro AI Podcast, we explore an important emerging issue of the AI era: the rise of synthetic identity

As generative AI rapidly advances, businesses are entering a world where voices can be cloned, faces can be synthesized, personalities can be modeled, and human authenticity itself becomes programmable. The discussion goes far beyond entertainment and dives into what executives across every industry need to understand right now. 

The episode examines: 

  • Why AI-generated identity replication is becoming a major enterprise risk  
  • How deepfakes and synthetic media are already impacting trust and cybersecurity  
  • Why current copyright and intellectual property laws are not prepared for this shift  
  • The growing importance of digital provenance, authentication, and AI governance  
  • How organizations may eventually manage AI “digital twins” of executives and employees  
  • Why trust may become one of the most valuable assets in the AI economy  
  • The enormous opportunities around scalable AI personas and trusted digital interaction  

We also explore the broader macro implications of a world where identity itself becomes software—and what that means for brands, leadership, customer experience, security, and the future of human authenticity. 

This is a thoughtful and highly relevant conversation for CEOs, CIOs, legal leaders, marketers, cybersecurity professionals, and anyone trying to understand where AI is truly heading next. 

Send a Text to the AI Guides on the show!


About your AI Guides

Gary Sloper

https://www.linkedin.com/in/gsloper/


Scott Bryan

https://www.linkedin.com/in/scottjbryan/

 

Macro AI Website

https://www.macroaipodcast.com/

Macro AI LinkedIn Page:  

https://www.linkedin.com/company/macro-ai-podcast/


Gary's Free AI Readiness Assessment:

https://macronetservices.com/events/the-comprehensive-guide-to-ai-readiness


Scott's Content & Blog

https://www.macronomics.ai/blog





Welcome to the Macro AI podcast, where we explore the biggest macro trends in artificial intelligence and what they mean for business, technology and society. I'm Gary Sloper. And I'm Scott Bryan. And today we're talking about something that on the surface kind of sounds like a celebrity story, but, really it represents an important business and legal shift in this new era of AI. So

01:25
Fresh in the news cycles, Taylor Swift is reportedly taking aggressive steps to protect her voice, image,  likeness, and broader digital identity from AI replication. And while it's easy to dismiss that as something that only affects entertainers,  we actually think that it's an early warning signal for pretty much every industry, because this is no longer just about content creation. It's really about ownership of identity itself. Yeah, that's a point. And I think business leaders need to realize

01:55
how quickly this is evolving. If you look back for decades, enterprises focused on protecting intellectual property in the traditional sense. So think of software code, patents,  your trademarks, trade secrets, and ultimately customer data. That was really important for organizations. But generative AI is introducing an entirely new category of enterprise risks and enterprise value. So now your voice can be cloned.

02:23
Your face can be synthesized. Your writing style can be replicated. Your personality can be replicated. And eventually your digital presence  can persist independently of you. So if you're not there, it's moving on without you. And that's a profound shift that's happened in a very short window in this AI era. Right. Yeah. And Taylor Swift is really just the most visible example of a much bigger phenomenon. uh Her economic value is tied

02:53
closely to identity, obviously,  not just her music, but the emotional connection people feel with her voice or image, her personality and her authenticity. And AI now has the ability to reproduce pretty much all of those things at scale. And once that capability exists, it doesn't stop with just celebrities. So think about how this will impact uh executives, sales professionals,  financial advisors, teachers, educators, attorneys, politicians.

03:23
pretty much anybody, really anyone whose value depends on trust and recognizable identity. Yeah, it's a good point there, Scott. mean, AI has crossed an important threshold, you know, now in this time of 2026. It can now imitate the subtle signals humans use to establish credibility and familiarity. So if you look backwards, historically, if you heard someone's voice or saw their face, your brain assumed authenticity.

03:50
Now we're entering a world where that assumption no longer holds. And I think a lot of organizations still underestimate how disruptive this can become or will become. Cause once your identity itself becomes, you know, reproducible, companies have to rethink a lot of areas and they need to think of security, branding, uh governance, and even customer trust models in this, this, this time that we're seeing change right in front of us. Yeah, it's definitely getting

04:20
hard to tell what's real, what's not very, very quick. think one of the reasons that the story matters so much is because it exposes how unprepared the existing legal frameworks really are. ah Most of the current AI debate still revolves around copyright law, but copyright was designed for protecting content. And what we're now dealing with is much deeper than that. We're moving into a world where courts and regulators will increasingly have to answer questions about

04:50
ownership of actual human identity.  Who owns your voice?  Who owns a digital version of your likeness?  Who controls an AI generated personality that behaves like you, but not be you?  And those are totally different questions than traditional intellectual property law was really  built over time to handle. Yeah, that's a point. And AI blurs  all the old categories together. You can...

05:18
have an AI-generated song using synthetic voice that sounds like a real artist. uh You could even generate a video, and you've probably seen it online,  with a realistic digital likeness of that person. You can mimic someone's communication style almost perfectly. I I've seen this  on videos on Instagram, for example.  The old frameworks start breaking down because the machine is synthesizing parts of the identity itself. And this is why

05:45
You know, someone like Taylor Swift matters so much in this discussion. When someone with that level of influence starts aggressively defending identity rights, you know, what they feel is important to them and, and outliving them industries respond very quickly. You know, law firms respond, media companies respond, legislators respond, technology companies respond. So eventually enterprise legal departments will respond as well.  Um, because this, this can be something that will be impactful to them. Yeah.

06:15
Exactly. That's probably a good way to pivot into enterprise risk. ah So the business leaders  shouldn't think of this as a future problem that's years and years away. The enterprise implications are definitely already here. So we're already seeing  AI generated executive voices used in social engineering attacks. ah Obviously we've seen deep fake videos influence public perception, particularly in politics.  And we've seen synthetic media

06:45
actually move markets and create confusion, whether it's temporary or lasting confusion  online, ah obviously has an impact on markets.  And the underlying models are improving at an extraordinary pace. So voice cloning now requires very little training data and video generation quality is already accelerating at a rapid pace  and multimodal AI systems are getting

07:11
dramatically better at simulating believable human interaction. So it's really hard to tell in some cases what's real and what's not. And the realism curve is moving faster than most organizations out there are prepared for. And a lot of them just aren't even thinking about it. Yeah, it's a good point. I don't think a lot of them are thinking about it. And the real challenge is that trust infrastructure on the internet was never designed for this world. you you and I have been in the

07:39
global internet space for a long time. ah We know how it was built and it was not built for this. And most communication systems implicitly assume that audio and video are authentic. If you think about it, that assumption is no longer holding weight.  So, which means enterprises are eventually going to need entirely new verification mechanisms around identity, authenticity, to really prevent a nefarious scenario. I think we'll see much greater emphasis on

08:08
digital provenance, uh cryptographic authentication, watermarking, identity verification systems, ah trusted communication frameworks. Because- ahead. A lot of opportunity for uh startups out there, obviously, around  all these angles. Completely wide open. And I think because in the future, hearing a CEO's voice may not actually mean you heard the CEO.

08:37
And that's something that  I think a lot of business leaders will start to need to  really think about.  Yeah. And I think an interesting angle is that this isn't only about risk, it's also about enterprise opportunity. So organizations are going to realize that trusted human identity can become a scalable digital asset. So you get the digital asset angle. So  imagine a highly effective financial advisor.

09:04
creating an AI version of themselves  that they can interact with thousands of clients simultaneously, or something like a world-class position training on an AI uh assistant modeled on their exact communication style or expertise.  Or another one would be like a uh CEO deploying an AI generated internal uh update, team update that sounds and behaves consistently with their real personality.

09:34
there's a lot of opportunity there and things start kind of changing the economics of  expertise really directly. mean, am I really speaking with Scott right now? Or am I speaking with  Scott's cloned twin that's  5X across a bunch of others?  Yeah, you need to verify that. Right, and that's where things become philosophically interesting, I think.

09:59
Companies will eventually need policies around digital twins, AI personas,  and synthetic representations of  employees and executives. But who owns those systems? Can they continue to operate after someone leaves the company? So Scott, if you left the company ah and you're still representing the company digitally, is that allowable? ah What if you decided to retire? So can that persist after your retirement? ah

10:27
What rights does an employee retain over an AI trained version of themselves? And so these questions sound futuristic, but they're going to become real business discussions surprisingly quickly. And I think  many enterprises, yeah, haven't, haven't started considering that reality just yet. Yeah. Yeah. Good stuff. And I think kind of at the, at the deepest level, this, this entire episode, you know, everything that we've been talking about is, is kind of about that topic of trust.

10:55
So just, you know, going back, know, human history, authenticity was, was relatively easy to establish,  uh, physical presence mattered,  um, and, human interaction had all different types of friction and replication was difficult. But even with the AI tools that are out there as of right now, you know, early mid, almost mid 2026 replication is actually pretty easy and identity can be synthesized,  scaled,  automated.

11:25
and then distributed globally almost instantly. And that definitely fundamentally changes how humans evaluate credibility.  Credibility is already being questioned  all over the place. Yeah. And the business implications are enormous. Brands that can establish trusted authenticity. ah They may gain major strategic advantages here. Enterprises that fail to build trust frameworks may struggle in environments

11:51
you know, flooded with synthetic media and AI generated interaction or interactions.  Um, this is why we believe identity protection will eventually become  as important as cybersecurity itself, not because every company is trying to defend against celebrity impersonation, but because enterprises are increasingly operating in digital environments where, know, that, that human authenticity becomes harder to verify. Yeah, definitely. And I think executive advice for present.

12:21
is to start thinking about this in  three areas  simultaneously. So first, understand where identity creates enterprise value inside your organization. That includes the executives, customer-facing employees,  subject matter experts, advisors, brand representatives. Brand representatives, obviously huge.  And then second, begin developing governance policies  and assess tools that are specific to synthetic media.

12:49
AI generated likeness, uh voice cloning and authentication standards. Obviously that's important.  And then a third one would be,  recognize that this is not just a technology issue. uh It's a strategic leadership issue and organizations that adapt early will kind of help define the trust standards for their industries  and avoid that first costly issue. ah And then not only that, they may be able to take advantage of some opportunity like we talked about in the opportunity uh segment.

13:18
Yeah, those are great points. And I'd add one more thing. think leaders should avoid viewing this purely defensively. This will absolutely be enormous opportunities here because  some companies will create an extraordinary AI powered customer experience using trusted digital personas. Like that will happen.  Some  will scale education and advisory services in ways that  previously were impossible.

13:45
And I think some will create entirely new business models around licensed digital identity. You talked about this a little while ago, you know, there's opportunity for a slew of startups to pop up. But the organizations that succeed will be the ones that balance innovation with authenticity. So because you, know, in this AI era, trust may become a hugely valuable asset for the enterprise.

14:11
And thus could be a huge catalyst for a lot of organizations, whether you're servicing companies for identity protection or you're  building it internally. So there's a huge opportunity here. Yeah. Yeah. Good stuff.  That's probably a good place to wrap it. uh So Taylor Swift protecting her likeness from AI may kind of sound like entertainment news, like how we open this up, but we think it's actually one of the

14:38
probably one of the most clear signals yet about where society is headed. Human identity is becoming programmable. And once  identity becomes  software, the implications extend into pretty much every corner of business and society. Yeah. And the companies that recognize this early won't just protect themselves from risk. They'll help define how trust, authenticity, and digital interaction function move forward for them in their business ecosystem. Yep.

15:08
Yeah, so thanks for joining us on the Macro AI podcast. If you enjoyed this episode, share it with another executive or a technology leader who's trying to understand where AI is really headed.  So thank you and we'll talk next time.  See you next time.