エピソード

  • Recap of the "Beyond Chatbots - The Journey of Generative AI to the Edge"
    2024/12/26

    Send us a text

    Get ready for the recap of the Beyond Chatbots - The Journey of Generative AI to the Edge - exploring the frontier of Generative AI and edge technology with Davis Sawyer, Danilo Pau and Pete Bernard as they discuss the dynamic innovations transforming these fields. We uncover the fascinating shift from cloud-based proof-of-concepts to the compelling reality of edge solutions. This conversation dives into the ever-evolving landscape of generative AI, particularly in edge environments where cost and power efficiency reign supreme. From automotive breakthroughs to advanced memory optimization techniques, discover how these innovations are redefining the role of AI in our world.

    Listen in as we dissect the gravitational pull towards edge solutions with insights from recent industry research and captivating discussions. Dave shares intriguing observations on AI as crucial infrastructure, likening it to essential utilities that promise lasting impacts across sectors. As we look ahead, the conversation turns to the future of AI in manufacturing and the exciting potential for AGI-capable equipment. With Foundation partners like STMicroelectronics and NXP at the helm, the potential for edge AI advancements is limitless. Join us for an engaging exploration of the trends shaping the future of AI and edge computing.

    Support the show

    Learn more about the EDGE AI FOUNDATION - edgeaifoundation.org

    続きを読む 一部表示
    9 分
  • Crafting Artistic Images with Embedded AI with Alberto Ancilotto of FBK
    2024/12/19

    Send us a text

    Unlock the secrets of neural style transfer on microcontrollers with our special guest, Alberto Ancilotto of FBK, as he explores a groundbreaking approach to image generation on low-power devices. Discover how this innovative technique allows us to combine the content of one image with the artistic style of another, transforming simple visuals into unique masterpieces—like turning a regular cat photo into a Van Gogh-inspired work of art. Alberto introduces Xinet, a cutting-edge convolutional neural network designed to perform these creative tasks efficiently on embedded platforms. Gain insight into the process of optimizing performance by evaluating CNN operators for energy efficiency and adapting networks for a variety of devices, from the smallest microcontrollers to advanced TPUs and accelerators.

    We dive deep into the collaboration between Clip and style transfer networks, enhancing the precision of semantic representation in generated images. Witness the impressive capabilities of this technology through real-world examples, such as generating images in just 60 milliseconds on the STM32N6 microcontroller. Experience the advanced applications in video anonymization, where style transfer provides a superior alternative to traditional blurring methods, altering appearances while maintaining action consistency. Alberto also addresses the broader implications of anonymization technology in public spaces, including privacy protection and GDPR compliance, while maintaining artistic integrity. Join us as we tackle audience questions about model parameters, deployment flexibility, and the exciting potential of this technology across various sectors.

    Support the show

    Learn more about the EDGE AI FOUNDATION - edgeaifoundation.org

    続きを読む 一部表示
    32 分
  • Revolutionizing Software Development with GenAI-Powered Edge Solutions with Anirban Bhattacharjee of Wipro
    2024/12/12

    Send us a text

    What if AI could transform the landscape of software development, making traditional methods seem like relics of the past? Join us as we sit down with a visionary guest from Wipro Engineering to uncover how GenAI-based custom code systems are revolutionizing the future of AI PCs. We promise a deep dive into how these intelligent tools are not just enhancing productivity but are redefining the very essence of software creation. Discover why the shift from cloud-based to edge-based solutions is imperative for enterprise developers dealing with proprietary codebases, and explore the power of open-source models like Code Llama, integrated through advanced frameworks such as LanChain and Lama Index. This episode promises to unravel the intricacies of an innovative web-based application designed to boost performance and user efficiency.

    Our conversation also shines a light on the tangible productivity gains brought by cutting-edge code assistance tools, effectively reducing weeks of work into days. We explore the vital role of hardware—like RAM and discrete GPUs—in maximizing these tools' potential. Discussions reveal varying levels of acceptance among developers, with younger professionals more readily embracing the shift. To bridge this gap, we recommend intensive training and boot camps as pathways to wider adoption. Furthermore, the potential of generative AI in deciphering and documenting legacy code is highlighted, offering a glimpse into how these advancements are reshaping the programming landscape. Engage with us to understand the profound impact of these innovations on the future of software development.

    Support the show

    Learn more about the EDGE AI FOUNDATION - edgeaifoundation.org

    続きを読む 一部表示
    29 分
  • Tomorrow's Edge AI: Cutting-Edge Memory Optimization for Large Language Models with Seonyeong Heo of Kyung Hee University
    2024/12/05

    Send us a text

    Discover the cutting-edge techniques behind memory optimization for large language models with our guest, Seonyeong Heo from Kyung-Hee University. Join us as we promise to unlock the secrets of deploying 7-billion-parameter models on small devices with limited memory. This episode delves into the intricacies of key-value caching in decoder-only transformers, a crucial innovation that reduces computational overhead by efficiently storing and reusing outputs. Seon-young shares insightful strategies that tackle the high demands of memory management, offering a glimpse into how these models can be more feasible and energy-efficient.

    Our conversation also ventures into the world of dynamic compression methods essential for optimizing memory usage. We unpack the challenges of compressing key-value arrays and explore the merits of techniques like quantization, pruning, and dimensionality reduction with autoencoders. Weighted quantization is highlighted as a standout method for achieving remarkable compression rates with minimal errors, provided it's fine-tuned effectively. This episode is a must-listen for those interested in the future of on-device LLMs, as we underscore the significance of efficient memory management in enhancing their performance, especially in resource-constrained settings. Tune in for this enlightening discussion paving the way for innovative advancements in the field.

    Support the show

    Learn more about the EDGE AI FOUNDATION - edgeaifoundation.org

    続きを読む 一部表示
    30 分
  • Revolutionizing Automotive AI with Small Language Models with Alok Ranjan of BOSCH
    2024/11/28

    Send us a text

    Unlock the future of automotive AI with insights from Ashutosh and Guru Prashad of BOSS India, as they unravel the transformative potential of Small Language Models (SLMs) in vehicles. Discover how these compact powerhouses are reshaping the landscape of vehicular edge systems by offering customization and performance reliability, all while keeping data privacy at the forefront. Dive into the captivating interplay between SLMs and their larger counterparts, Large Language Models (LLMs), and learn how they together address domain-specific tasks and complex computations in the cloud. This episode promises to equip you with a deeper understanding of why SLMs are the secret sauce for advancing automotive AI in this data-sensitive era.

    We also spotlight the remarkable optimization journey of the Tiny Llama 1.1 model. Learn about the fine-tuning process that brought about an astounding 700% improvement in throughput and a drastic reduction in model size. Uncover the fascinating world of edge deployment using devices like Pi 4B and Jetson Oren Nano, and explore the audio and chat functionalities that are setting new standards in vehicular AI. Finally, imagine the future of personalized interactions in cars, where generative AI transforms the way we communicate and engage with our vehicles and surroundings. This episode is a treasure trove of forward-thinking solutions and innovative ideas, perfect for anyone eager to explore the cutting edge of automotive AI.

    Support the show

    Learn more about the EDGE AI FOUNDATION - edgeaifoundation.org

    続きを読む 一部表示
    32 分
  • From tinyML to the edge of AI: Introducing the EDGE AI FOUNDATION
    2024/11/25

    Send us a text

    Discover how the EDGE AI FOUNDATION is evolving from its roots in the tinyML Foundation to becoming a central hub for innovation and collaboration. Learn how initiatives like EDGE AI Labs and the EDGE AIP program are bridging the gap between academia and industry, training future AI leaders while tackling the ethical challenges of responsible AI development.

    Explore the transformative potential of generative AI on edge devices, from providing vital healthcare diagnostics in remote areas to enabling adaptive robotics in factories. We'll highlight compelling reasons for companies to engage with the Edge AI Foundation, offering unparalleled access to cutting-edge research, top talent, and a voice in shaping the industry's future. As we navigate through real-life scenarios and ethical considerations, you’ll see why the urgency and opportunity surrounding Edge AI is something you don't want to miss.

    Join us on this journey to ensure the benefits of AI are shared widely and responsibly by visiting edgeaifoundation.org.

    Support the show

    Learn more about the EDGE AI FOUNDATION - edgeaifoundation.org

    続きを読む 一部表示
    14 分
  • Unveiling the Technological Breakthroughs of ExecuTorch with Meta's Chen Lai
    2024/11/21

    Send us a text

    Unlock the secrets to deploying machine learning models on edge devices with Chen Lai from the PyTorch Edge team at Meta. Discover how XTorch, a brainchild of the PyTorch team, is transforming edge deployment by addressing challenges like memory constraints and hardware diversity. Get an insider's view on the technical collaborations with tech giants like Apple, Arm, Qualcomm, and MediaTek, which are revolutionizing the deployment of advanced language models like LLAMA on platforms such as iOS and Android. With Chen's expert insights, explore the fascinating process of converting PyTorch models into executable programs optimized for performance, stability, and broad hardware compatibility, ensuring seamless integration from server to edge environments.

    Immerse yourself in the world of XTorch within the Red Bull Ecosystem, where deploying machine learning models becomes effortless even without extensive hardware knowledge. Learn how key components like Torchexport and Torchio capture compute graphs and support quantization, elevating edge deployment capabilities. Discover how Torchchat facilitates large language model inference on various devices, ensuring compatibility with popular models from Hugging Face. As we wrap up, hear about the community impact of Meta's Executorch initiative, showcasing a commitment to innovation and collaboration. Chen shares his passion and dedication to advancing edge computing, leaving a lasting impression on listeners eager for the next wave of technological breakthroughs.

    Support the show

    Learn more about the EDGE AI FOUNDATION - edgeaifoundation.org

    続きを読む 一部表示
    31 分
  • Revolutionizing TinyML: Integrating Large Language Models for Enhanced Efficiency
    2024/11/14

    Send us a text

    Unlock the future of TinyML by learning how to harness the power of large language models, as we sit down with Roberto Morabito to dissect this intriguing technological convergence. Discover how the collaborative efforts with Eurocom and the University of Helsinki are shaping a groundbreaking framework designed to elevate TinyML's lifecycle management. We promise to unravel the complexities and opportunities that stem from integrating these technologies, focusing on the essential role of prompt templates and the dynamic challenges posed by hardware constraints. Through a proof-of-concept demonstration, we bring you invaluable insights into resource consumption, potential bottlenecks, and the exciting prospect of automating lifecycle stages.

    Our conversation ventures into optimizing language models for end devices, delving into the transformative potential of Arduinos and single-board computers in enhancing efficiency and slashing costs. Roberto shares his expertise on the nuances of model conversion across varying hardware capabilities, revealing the impact this has on success rates. The episode crescendos with a compelling discussion on automating industrial time series forecasting, underscoring the critical need for adaptive solutions to maintain accuracy and efficiency. Through Roberto's expert insights, listeners are invited to explore the forefront of technology that is poised to revolutionize industrial applications.

    Support the show

    Learn more about the EDGE AI FOUNDATION - edgeaifoundation.org

    続きを読む 一部表示
    27 分