
EDGE AI POD
Discover the cutting-edge world of energy-efficient machine learning, edge AI, hardware accelerators, software algorithms, and real-world use cases with this podcast feed from all things in the world's largest EDGE AI community.
These are shows like EDGE AI TALKS, EDGE AI BLUEPRINTS as well as EDGE AI FOUNDATION event talks on a range of research, product and business topics.
Join us to stay informed and inspired!
EDGE AI POD
Revolutionizing AI Deployment with Open-Source Edge Technologies with Odin Chen of Arm
What if you could deploy AI solutions seamlessly across every platform with just a single stroke? Join us as we explore this intriguing possibility alongside our guest, Odin Chen of Arm, who brings a wealth of experience from his journey from Taiwan to Cambridge. Together, we embark on a journey through the vibrant landscape of Edge AI, tackling its transformative impact on both enterprises and startups. We shine a light on the mantra "write once, deploy everywhere," unraveling its significance in a world where AI's reach is rapidly expanding. Along the way, we confront the challenges of AI deployment, from soaring costs to energy demands, and celebrate the critical role of open-source projects. Discover how technologies spanning image classification to generative AI are revolutionizing industries, while open-source contributions, particularly in Arm architecture and Python optimizations, are paving the way for a more collaborative future.
In the second segment, we traverse the vast potential of the AI ecosystem, shedding light on AI frameworks like MediaPipe and ESSEC torch that guide industries toward optimized AI solutions. With Google's AI offerings on the horizon, we introduce a cutting-edge context model for ESSEC torch and delve into key partnerships with Arm technology pioneers such as Alif and HiMAS. These collaborations aim to turbocharge scalable computing solutions and high-efficiency image processing. Excitingly, we unveil our developer program and open software stack, inviting listeners to engage actively and collaborate. Keep your calendars marked for upcoming live demos and panel discussions, promising a firsthand look at the immense potential of AI deployment across diverse sectors.
Learn more about the EDGE AI FOUNDATION - edgeaifoundation.org
Tanya Cushman Reviewer. Reviewer. Without further ado, odin, it's always good to see the old friend here, especially in Taiwan. Hi everyone, it's Odin. I was the Taiwan's employee, but now I'm moving to Cambridge.
Speaker 1:The topic today is Unleash the Edge AI Potential. The one of the terms I want to share is the right ones and the deploy everywhere. So why is it so important? In today's morning session we talk a lot about how the possibility AI can did for, like an age device, for the research, from the security, for a lot of different segments. So, as I am very happy to share our view and what the World1 provider ecosystem looks like, so yeah, as always, any questions, just raise your hand or let me know so we can have more interactive discussion. So I think everyone believes, especially in the AI segment, ai is everywhere right. We have a lot of enterprise companies. They commit, they are studying, doing the AI, no matter what kind of purpose, for to enhance the productivity, for building some of the AI future, for their product, or even that's just safety in the cost and also from the finance point of view. We have a lot of investment for the AI enterprise and even the startup, for the AI enterprise and even the startup. So in the booth we already saw a lot of AI innovations is using the AI. Also, it's some of the buzzy word called AI server, right. So for the compute tasks for a lot of companies they have moved the standard server to the AI server. So those are the things that is very convincing for everyone.
Speaker 1:Say that AI is everywhere, but AI also is very expensive. So we talk about the power consumption, we talk about how difficult it is to deploy the software. We're talking about how to maintain the model, the security. So in the very beginning, the AI we're talking about AI is called like an image classification, so we can detect the cats, dog, car, bus kind of a thing. So in that time this is more like oh yeah, we do. Aik can help the people to do some of the image regulation or maybe the object detection kind of work. But for the genetic AI, they start doing some of the tax generation, they can generate the question.
Speaker 1:So maybe you can give us some of the quickest survey here. Have anyone used the chat GPT kind of? Oh yeah, yes, a lot of people use that. That's good. So for me this is like a kind of a different type of a net phrase, right. So we use a net for your home entertainment, but for us we also use the generative AI kind of a service for your daily work. No matter it's translation, no matter it's helping your kid to do the homework, right, okay, so test generation is kind of a new AI use case and furthermore it's the image. We have a lot of image generation, so mid-generation kind of a thing.
Speaker 1:So different type of AI is more and more costed. They consume more energy in a different way. So we realize AI is very expensive. So it's become a challenge when we try to have more AI for edge. For in general terms, so use case could be like your personal content or there will be some of your AI energy so they help you to improve the productivity, to help you to assess your work. So that's a use case and we see that a lot of people are asking how can I have a lot of AI experience not only in the server but also in your device like a smartphone, your AG server, even the notebook?
Speaker 1:So the challenge is that it's so many different kind of computing platform and the power efficiency, as I say, also some of the different type of storage. That's caused the challenge for when we try to adopt AI on the AG, for when we try to adopt AI on the edge. So that is the main challenge for the software developer. So if we were talking about the hardware, it's more like how to do the hardware implementation, how to manufacture the hardware. But for software a lot of challenges from reform is so many different kind of hardware. So how can you optimize your model, your software, your application in different type of hardware? Some hardware that have a GPU, some hardware that have DSP, some hardware just CPU. Another one is for software. So because it's a very diverse, different kind of platform, so the software should be very different and we saw that a lot of developer they are keeping up how they gross the different type of AI application. They will also cause the front challenge for the software developer as an.
Speaker 1:I want to share what on to help the open source investment. The goal is that if we have make it the open source, they have a more more to optimize the architecture. They will benefit the whole ecosystem. The graph here is the Linaro and the Microsoft Windows. They are both of the majority of the operation systems around the world. We have a lot of open source projects working to enhance, enable that. On the ARM architecture, give you some example. So it's over the 2020, the kind of a different project, open source project around ARM for the open source space. So if you are like working with check, with linear code, if you're tracking with the Windows, some of the utility, most of that is contributed by an employee, and then not only for the OS perspective. Like a GPU, we have the Mali GPU. They also have the DDK. They have the open source, the DDL driver. This also contribute to the open source.
Speaker 1:As a software developer, probably you will not only use the C and the C++. You may use Java, you will use R, you can use Go, so they use a different kind of language. Arm also contribute a lot of optimization through lots of projects Python for short. So every AI scientist, some of the machine learning engineers, they use heavy Python. So we do a lot of Python implementation optimization as well. If you are using Cortex-N, so like the demo in outside the demo booths, so they have a lot of Cortex-N product. They are leveraging the CNSYS. It's an open source library. They optimize for the DSP or some machine learning feature, not only like a Cortex-A and Cortex-A. If you're doing some of the 5G, like open RANs and also contribute the RAN. It's kind of a isolation library to help the open RAN easier to adopt, help the open RAN easier to adopt.
Speaker 1:So I want you to the page to show all the ecosystem partners To make the industrial success. It's not only for the hardware part, the software is the matter. So, as ARM, we're not only developing the IP part, we also develop another software, so Windows R1. So this is another exciting part I want to share with everyone. So, because Windows R1 is a lot of different types of software, so we work with lots of partners and and we integrate lots of application tools into the on-active ecosystem. So Claudia is another good technology I want to share with everyone.
Speaker 1:So, claudia, the mini is a key. I think that was a key to enable the LLN into the device. So you can see that this is some of the demo we're running on the in general, the CPU smartphone. You can use that to do some of the chatbot without the internet. So you asking the question, what is Q, what is stack? The robot gave you the answer. We have very convinced that the ecosystem, say the cloud, ai, they can generate almost double the time to first tokens performance and that we in the end device, in the smartphone, we can provide the 25% of leave and also you can reduce the size. So the good news is all the software is open source. So you can bring that check out from the check, check out from the GitHub and download in your device.
Speaker 1:So let me skip this one. So I think the best part is that we are not only tell the industrial, say you can do that, but we will guide you over the software device how to do that. So here is a sample. So MediaPipe, smpack this is one of Google's providers some of the LLM, the framework, the Cloud AI. We have a learning path. We have some tutorial to show you how to do that. So you have a phone, you just apply the software to follow the learning path. You can get it. Here's the results for the ESSEC torch. Essec torch is another good AI framework for deploy to the end device. So recently we just announced we have a new context model for the ESSEC torch. Here is the link. So PyTorch they have a blog. He mentioned about how good the Lama 3.0 could be optimized through the Cloud AI. Next time I hope I have a chance to give you some more data, maybe a live demo to show the whole ecosystem part of how to make it. I think everyone knows, so maybe I can skip this one.
Speaker 1:So we have a lot of computing platform. We have a lot of partners. We have a three major partners not only three, but I want to show three of that. The beautiful part is we have three different MCU and C partner. They are using the ARM technology but they can do some differentiation by their target market Alif. They are doing the scalable computing solution. They have the Cortex-A, cortex-n and the MPU. Also we are HiMAS. They have a very high low power camera sensor camera. They can do a lot of image pipeline optimization and they're in the very low power mode. We also have a general purpose MCU, like a new model. They can do the connectivity, security and the machine learning. So I want to show the beautiful part. So through the ARM ecosystem, through the ARM technology, you still can build a lot of different kinds of business potential in different areas.
Speaker 1:So I want to highlight that it's all about software, right? So we have the best-in-class, the IP. We have a very open software stack and we care about the software development. So we have a program called the developer program. So that is a lot of members and we have staff, some of the ambassadors, so actually we have one of our ambassadors in here. So Jack will bring some of the panel discussion later. So the call for action is that I want everyone. If you have a chance, if you're interested about, if you're interested in how to deploy the AI, just look at the link, just tell us what you want. So we are open. We are very interested in how we can help you see the every ecosystem part.