Episodes

  • Revolutionizing TinyML: Integrating Large Language Models for Enhanced Efficiency
    Nov 14 2024

    Send us a text

    Unlock the future of TinyML by learning how to harness the power of large language models, as we sit down with Roberto Morabito to dissect this intriguing technological convergence. Discover how the collaborative efforts with Eurocom and the University of Helsinki are shaping a groundbreaking framework designed to elevate TinyML's lifecycle management. We promise to unravel the complexities and opportunities that stem from integrating these technologies, focusing on the essential role of prompt templates and the dynamic challenges posed by hardware constraints. Through a proof-of-concept demonstration, we bring you invaluable insights into resource consumption, potential bottlenecks, and the exciting prospect of automating lifecycle stages.

    Our conversation ventures into optimizing language models for end devices, delving into the transformative potential of Arduinos and single-board computers in enhancing efficiency and slashing costs. Roberto shares his expertise on the nuances of model conversion across varying hardware capabilities, revealing the impact this has on success rates. The episode crescendos with a compelling discussion on automating industrial time series forecasting, underscoring the critical need for adaptive solutions to maintain accuracy and efficiency. Through Roberto's expert insights, listeners are invited to explore the forefront of technology that is poised to revolutionize industrial applications.

    Learn more about the tinyML Foundation - tinyml.org

    Show More Show Less
    27 mins
  • Harnessing Edge AI: Transforming Industries with Advanced Transformer Models with Dave McCarthy of IDC and Pete Bernard of tinyML Foundation
    Nov 7 2024

    Send us a text

    Unlock the transformative potential of edge computing with the insights of industry experts Dave McCarthy from IDC and Pete Bernard. Ever wondered how advanced transformer models are catalyzing technological leaps at the edge? This episode promises to enlighten you on the nuances of AI-ready infrastructure, pushing the boundaries of autonomous operations in multi-cloud and multi-edge environments. With an emphasis on trust, security, and sustainability, our guests illuminate the strategic importance of optimizing edge designs and the benefits of hybrid and multi-cloud strategies.

    Explore the dynamic world of Edge AI as we dissect the complexities of heavy and light edge scenarios, particularly within industrial contexts. Dave and Pete help navigate the shift from centralized systems to the cutting-edge distributed frameworks necessary for processing the explosion of data generated outside traditional data centers. Discover how Edge AI and TinyML are reshaping industries by empowering smarter devices and solutions, pushing AI workloads from the cloud to resource-constrained environments for improved efficiency and real-time data processing.

    Dive into the fascinating migration of AI workloads from the cloud to the edge, driven by the demands of smart cities and critical infrastructure. Our experts share insights from global surveys, examining how inference is increasingly shifting to the edge, while training remains cloud-based. Listen in as we explore the evolving edge AI hardware landscape, cost-effective solutions, and the burgeoning interest in specialized models. Uncover emerging generative AI use cases poised to revolutionize various sectors, and gain a glimpse into the future opportunities and challenges in the ever-evolving landscape of edge AI. Join us for a riveting discussion that promises to leave you informed and inspired.

    Learn more about the tinyML Foundation - tinyml.org

    Show More Show Less
    34 mins
  • Transforming the Edge with Generative AI: Unraveling Innovations Beyond Chatbots with Danilo Pau, IEEE Fellow from STMicroelectronics
    Oct 31 2024

    Send us a text

    Generative AI is poised to transform the edge, but what does this mean for technology, innovation, and everyday life? Join us for an enlightening discussion led by Danilo Pau, featuring a distinguished panel of experts including Dave, Roberto Chen, Alok, Seung-Yang, Arniban, and Alberto. They promise to unravel the mysteries beyond large language models and chatbots, offering fresh insights into the interplay of algorithms, software, chips, and methodologies from an EDA perspective. The conversation aims to democratize generative AI, making its groundbreaking potential accessible to all, and sparking inspiration throughout the tiny ML community.

    With gratitude to the TinyML Foundation for their invaluable support, this episode builds on the momentum of previous forums. Reflecting on the foundation laid by Davis Sawyer's inspiring March session, we explore how generative AI is not just a cloud-based innovation but a technology set to revolutionize the edge. Expect to hear how these developments could impact everyone's work and life, and gain a glimpse into the collective vision for the future from our esteemed speakers. Don’t miss the chance to be part of this vibrant exchange, where innovation is not just discussed but propelled forward.

    Learn more about the tinyML Foundation - tinyml.org

    Show More Show Less
    7 mins
  • Revolutionizing Weather Forecasting with Acoustic Smart Technology
    Oct 27 2024

    Send us a text

    Unlock the secrets to revolutionary weather forecasting with our latest episode featuring Jonah Beysens, the brilliant mind behind the Aurora project. Imagine a world where remote and resource-limited areas can access reliable weather data without the hassle of maintaining traditional mechanical stations. Learn how the Aurora, an acoustic smart weather station, is making this possible. Born from the TinyML Challenge 2022, Aurora employs a TinyML board with a microphone to classify wind and rain intensity, ensuring robustness and ease of deployment. Jonah walks us through the journey of crafting a real-world dataset and shares the collaborative spirit behind making this data open for community-driven innovation.

    Discover how the shift towards low-cost, acoustics-based devices is reshaping weather forecasting, offering enhanced spatial resolution with multiple ground stations. Jonah sheds light on the collaborative efforts to refine prediction models with open datasets, emphasizing the profound global impact, especially in developing nations where agriculture depends heavily on accurate forecasts. As we discuss the ongoing work to make weather data more accessible worldwide, we highlight the role of community and open access in driving forward weather-related technologies. Join us in exploring how these innovative solutions promise timely, dependable forecasts, paving the way for a future where environmental data is a shared resource for all.

    Learn more about the tinyML Foundation - tinyml.org

    Show More Show Less
    25 mins
  • Revolutionizing Nano-UAV Technology with Cutting-Edge On-Device Learning Strategies
    Oct 20 2024

    Send us a text

    Unlock the secrets of edge computing and on-device learning with Elias Cereda, a trailblazing PhD student at the Dalle Molle Institute for Artificial Intelligence in Italy, as we explore the transformative potential of AI in nano-UAV technology. Discover how Elias is pioneering solutions to the challenges of domain shifts in AI models, particularly those trained in simulations and deployed in real-world scenarios. From enhancing privacy and security through on-device learning to mastering the intricacies of human pose estimation for autonomous navigation, this episode is a treasure trove of insights. We dive into the practicalities of deploying systems using the Crazyflie 2.1 drone alongside GreenWaves' sophisticated RISC-V GAP8 and GAP9 chips, all while managing tight power constraints.

    Take a closer look at the complexities of on-device and online learning with Elias as we delve into the nuances of continual learning. Learn how maintaining a balance between retaining previous knowledge and integrating new data is crucial in the field. Elias introduces us to a self-supervised metric that operates without initial labels, cleverly utilizing odometry data to refine model predictions. By considering human movement as measurement noise and refining data through post-processing techniques, the episode unveils strategies to enhance model performance without explicit tracking.

    For those eager to delve deeper into the methodologies discussed, Elias points to a comprehensive paper that promises to enrich your understanding of this cutting-edge technology.

    Learn more about the tinyML Foundation - tinyml.org

    Show More Show Less
    22 mins
  • Deploying TinyML Models at Scale: Insights on Monitoring and Automation with Alessandro Grande of Edge Impulse
    Oct 13 2024

    Send us a text

    Unlock the secrets of deploying TinyML models in real-world scenarios with Alessandro Grande, Head of Product at Edge Impulse. Curious about how TinyML has evolved since its early days? Alessandro takes us through a journey from his initial demos at Arm to the sophisticated, scalable deployments we see today. Learn why continuous model monitoring is not just important but essential for the reliability and functionality of machine learning applications, especially in large-scale IoT deployments. Alessandro shares actionable insights on how to maintain a continuous lifecycle for ML models to handle unpredictable changes and ensure sustained success.

    Delve into the intricacies of health-related use cases with a spotlight on the HIFE AI cough monitoring system. Discover best practices for data collection and preparation, including identifying outliers and leveraging Generative AI like ChatGPT 4.0 for efficient data labeling. We also emphasize the importance of building scalable infrastructure for automated ML development. Learn how continuous integration and continuous deployment (CI/CD) pipelines can enhance the lifecycle management of ML models, ensuring security and scalability from day one. This episode is a treasure trove of practical advice for anyone tackling the challenges of deploying ML models in diverse environments.

    Learn more about the tinyML Foundation - tinyml.org

    Show More Show Less
    21 mins
  • Revolutionizing Smart Glasses with Ultra-Efficient AI Algorithms
    Oct 6 2024

    Send us a text

    Unlock the secrets of ultra-efficient AI with Julian Musman from ETH Zurich, as we explore the cutting-edge world of smart glasses. Discover how Julian and his team have mastered the art of deploying powerful AI algorithms on the tiniest of devices, tackling the monumental challenge of performing object detection on microcontrollers. Tainissimo YOLO, a breakthrough network, is not just a marvel of engineering but a testament to the balance between computational power, network size, and battery life, revolutionizing the way smart glasses operate without relying on the cloud.

    Join us as we unpack the hardware innovations behind these intelligent glasses, featuring components like the NRF53 microcontroller and the GAP9 neural engine, meticulously designed for energy efficiency and high-speed performance. With an astonishing end-to-end latency of 18 frames per second, these glasses promise all-day functionality on a single charge. Julian sheds light on the intricate harmony of PCB design and neural network architecture and explores the potential of advanced convolutional changes, all while emphasizing the strategic choice to remain loyal to traditional networks for unmatched performance with hardware accelerators. Get ready for an insightful journey into the future of AI-powered wearable technology!

    Learn more about the tinyML Foundation - tinyml.org

    Show More Show Less
    31 mins
  • Cloud to the Edge: Future of LLMs w/ Mahesh Yadav of Google
    Sep 24 2024

    Send us a text

    Curious about how you can run a colossal 405 billion parameter model on a device with a mere 2 billion footprint? Join us with Mahesh Yadav from Google, as he shares his journey from developing small devices to working with massive language models. Mahesh reveals the groundbreaking possibilities of operating large models on minimal hardware, making internet-free, edge AI a reality even on devices as small as a smartwatch. This eye-opening discussion is packed with insights into the future of AI and edge computing that you don't want to miss.

    Explore the strategic shifts by tech giants in the language model arena with Mahesh and our hosts. We dissect Microsoft's investment in OpenAI’s Phi model and Google's development of Gamma, exploring how increasing the parameters in large language models leads to emergent behaviors like logical reasoning and translation. Delving into the technical and financial implications of these advancements, we also address privacy concerns and the critical need for cost-effective model optimization in enterprise environments handling sensitive data.

    Advancements in edge AI training take center stage as Mahesh unpacks the latest techniques for model size reduction. Learn about synthetic data generation and the use of quantization, pruning, and distillation to shrink models without losing accuracy. Mahesh also highlights practical applications of small language models in enterprise settings, from contract management to sentiment analysis, and discusses the challenges of deploying these models on edge devices. Tune in to discover cutting-edge strategies for model compression and adaptation, and how startups are leveraging base models with specialized adapters to revolutionize the AI landscape.

    Learn more about the tinyML Foundation - tinyml.org

    Show More Show Less
    1 hr