
EDGE AI POD
Discover the cutting-edge world of energy-efficient machine learning, edge AI, hardware accelerators, software algorithms, and real-world use cases with this podcast feed from all things in the world's largest EDGE AI community.
These are shows like EDGE AI TALKS, EDGE AI BLUEPRINTS as well as EDGE AI FOUNDATION event talks on a range of research, product and business topics.
Join us to stay informed and inspired!
EDGE AI POD
Recap of the "Beyond Chatbots - The Journey of Generative AI to the Edge"
Get ready for the recap of the Beyond Chatbots - The Journey of Generative AI to the Edge - exploring the frontier of Generative AI and edge technology with Davis Sawyer, Danilo Pau and Pete Bernard as they discuss the dynamic innovations transforming these fields. We uncover the fascinating shift from cloud-based proof-of-concepts to the compelling reality of edge solutions. This conversation dives into the ever-evolving landscape of generative AI, particularly in edge environments where cost and power efficiency reign supreme. From automotive breakthroughs to advanced memory optimization techniques, discover how these innovations are redefining the role of AI in our world.
Listen in as we dissect the gravitational pull towards edge solutions with insights from recent industry research and captivating discussions. Dave shares intriguing observations on AI as crucial infrastructure, likening it to essential utilities that promise lasting impacts across sectors. As we look ahead, the conversation turns to the future of AI in manufacturing and the exciting potential for AGI-capable equipment. With Foundation partners like STMicroelectronics and NXP at the helm, the potential for edge AI advancements is limitless. Join us for an engaging exploration of the trends shaping the future of AI and edge computing.
Learn more about the EDGE AI FOUNDATION - edgeaifoundation.org
And now we are back to the recap with Davis and Pete.
Speaker 3:It was a great day, pete, and the years man, my head is exploding right yeah, your endurance is impressive and I think the engagement from the common side is impressive. It's hard, this, I think this is the hardest part of the day. It's like where to begin to recap. It's got to sink in. You need to sleep on it.
Speaker 2:Yeah, yeah, I think there were really diversified contributions which express somehow the richness of the people we have in the community.
Speaker 1:Yeah, agreed.
Speaker 3:Agreed, some things stay the same and some things change. There's always great to see the continued interest in pruning and quantization and some of these familiar topics, but they take different forms in terms of the methodologies applied to you, but also the diversity of technical details plus some applications, like we saw with ALOC's automotive presentation was a good day to dive on a specific domain.
Speaker 1:I thought yeah, that made sense. I like also the memory optimization discussion Kind of went to math camp for a little bit there. For a lot of these systems they're very kind of memory bound and that's cost and power when you get into these edge platforms. So the ability to sort of get more creative with memory usage is really going to help a lot of the generative AI deployments out there on the edge. Not really an issue when you're in the cloud, you just kind of throw in sort of infinite memory there, but out in the real world you need to sort of be careful about how much you add in. So I thought that was interesting.
Speaker 2:I was also touched by some comments from Dave at the beginning when he was sharing his view about this layer representation of the edge. Edge is not a monolithic thing. It really is composed by layers and really these layers are the opportunity for us and for the community to contribute. And the other point was really the concept of the infrastructure. Ai is an infrastructure. That is, I think is really a very important comment from Dave Infrastructure like a train, like a water pipe, like energy distribution, which is amazing, and it means that it's here to stay for long.
Speaker 2:And it's a continuum between the cloud and the edge.
Speaker 3:Yeah, dave's onto something.
Speaker 1:It was interesting to hear, I mean, the way they do the real research with Fortune 500s and kind of trying to understand where their head's at.
Speaker 1:And it was interesting that he was pointing out too, I think reinforcing the fact that there's always a gravitational pull toward the edge. Things start as a POC on the cloud, but then, when they get real and get commercialized, everyone's like do I need to pay all that cloud stuff and do I need to actually, can I just have this in a box on my factory floor and do that? So I think that's just always the gravity, always the gravity.
Speaker 3:Once you get that first AWS bill and you think but the one data point from Dave's slides I think he was talking about the percent increase, the uptick in evaluation, so people at the evaluation stage of AI, the one data point that stuck out to me was the percent not interested in using Gen AI. It was very small. I think the global average was like 5%. I think Europe EU is highest with 11%. So, yeah, it's here to stay and manufacturing intelligence.
Speaker 1:Yeah, I was at Embedded World last week. I did the rounds. I mean, it's still early, so I asked people some of the box makers that make gateways. I'm like what do you have for me to run like a Gen AI model in your actual gateway? What do you have that's outfitted with that and it's still pretty slim pickings. It's some Warren-based stuff with 32 gigs of RAM stuff, but that's that's like 5 000 plus dollars.
Speaker 1:So I think that's the next thing is like now we'll start to see some of the box makers, the device builders, you know, building out platforms that are agi capable. I mean, you know, qualcomm makes a pretty good case with their snapdragons on what they can really do there. Um, I mean, I think we'll see more and more, and I'm sure with nxp as well, but that's the next thing is to see actual equipment out there. So when Dave talks about evaluating, how do people actually evaluate this stuff in a cost-effective way? I think that's the next step.
Speaker 3:I think SST and Qualcomm, you guys just partnered up or announced some partnership along these lines of combining computation plus connectivity.
Speaker 1:Yeah, yeah. So it's going gonna get real, I'm sure, by the next, next embedded world. I'm sure we'll see all kinds of stuff going on. But you know, even six months from now it's inevitable if there's uh demand. So, yeah, I thought the, the, yeah, the range of uh stuff was really good. The execute torch stuff I learned a lot on that. Um, from meta, from chen lai, that was cool too. Um Mentioned the Bosch stuff. Yeah, it was just like so many different ways. I mean, you know, there's an old phrase there's never a shortage of work to do. There's like a ton of work to do here, but a lot of opportunity for people out there that are building out their careers and want to really be on the cutting edge. No pun intended. You know, edge edge AI gener want to really be on the cutting edge.
Speaker 3:No, pun intended edge AI generative AI, tiny ML.
Speaker 1:This is the frontier for folks to invest in, and you know.
Speaker 2:Pete and Davis automotive. You said already Davis, but it still surprised me Meta. Maybe I'm wrong, maybe it was a personal feeling. It still surprised me Meta. Maybe I'm wrong, maybe it was a personal feeling, but back to 2018, 2019, I didn't feel that Meta was really interested at the edge. But what I saw today from Che, really Meta is committed at the edge and is bringing innovation at the edge, starting with generative AI.
Speaker 3:I think it's a very bold message.
Speaker 3:Yeah, well, if you kind of tie a couple of threads together, pytorch is a common framework to develop and train these models, but the footprint of PyTorch is the framework itself and the dependent libraries is pretty large. So when you're trying to port to a memory-constrained board it kind of begets this natural evolution from the team. So it's been great to see Executorch and I think Chad alluded to it in the apps like Instagram and WhatsApp and Meta's apps. They are edge-native. You know mobile at least mobile native applications, right when where data is created, the user interaction points. So while a lot of these big models live, live and train on servers, still, yeah, they, they've been kind of catapulted into the future, with, with the success of Lama and Lama 3.2 being a very popular model. I mean, we certainly see that a lot too. So no, I'm glad you brought it up to me. It kind of it's good to see.
Speaker 1:Yeah, no, it's cool For future reference. And also, if you haven't been to the Discord channel, it's dscgg forward slash edge gen AI. There you go. Event channel on Discord. A bunch of people joined today, so and including the more detailed agenda and we'll post some decks up there and some other slides and stuff.
Speaker 2:So those are two good good locations.
Speaker 1:Yes, uh, everyone in in in europe.
Speaker 2:Good night in other time zones.
Speaker 1:Have a good afternoon and we'll see you tomorrow. Thank you.