EDGE AI POD cover art

EDGE AI POD

EDGE AI POD

By: EDGE AI FOUNDATION
Listen for free

About this listen

Discover the cutting-edge world of energy-efficient machine learning, edge AI, hardware accelerators, software algorithms, and real-world use cases with this podcast feed from all things in the world's largest EDGE AI community.

These are shows like EDGE AI Talks, EDGE AI Blueprints as well as EDGE AI FOUNDATION event talks on a range of research, product and business topics.

Join us to stay informed and inspired!

© 2025 EDGE AI FOUNDATION
Episodes
  • Bridging the Digital Divide by Generative AI through the Edge
    Dec 23 2025

    The technological revolution sparked by generative AI threatens to create the deepest digital divide we've ever seen. In this illuminating talk, Danilo Pau from STMicroelectronics reveals how only a handful of companies worldwide possess the resources to fully harness large-scale generative AI, while the rest of humanity risks being left behind.

    Pau takes us through the sobering reality of today's AI landscape: hyperparameterized models requiring nuclear power plants for training, hundreds of millions in costs, and worrying environmental impacts. But rather than accept this centralized future, he presents a compelling alternative path – bringing generative AI to edge devices.

    Through a comprehensive survey of recent research, Pau demonstrates that generative AI is already running on edge devices ranging from smartphones to microcontrollers. His team's work with STMicroelectronics processors showcases practical implementations including style transfer, language models, and perhaps most impressively, an intelligent thermostat capable of natural language interaction with reasoning capabilities.

    What emerges is a vision for AI not as another backend classifier but as a transformative interface between humans and machines. "GenAI is not for another detector," Pau explains. "We need to offer new added value" through natural interactions that understand context and can reason about the world.

    For researchers and developers, this talk provides concrete pathways to explore: from audio processing as a "low-hanging fruit" to visual question answering systems that run on minimal hardware. The future of AI isn't just in massive data centers – it's in the devices all around us, waiting to be unleashed through energy-efficient processing and innovative approaches to model optimization.

    Ready to join the movement bringing AI capabilities to everyone? Explore how edge-based generative AI could transform your products and help bridge the growing digital divide.

    Send us a text

    Support the show

    Learn more about the EDGE AI FOUNDATION - edgeaifoundation.org

    Show More Show Less
    32 mins
  • Networked AI Agents Decentralized Architecture
    Dec 16 2025

    What happens when trillions of AI agents can discover, communicate, and collaborate across organizational boundaries? Pradyumna Shari from MIT Media Lab unveils NANDA (Networked AI Agents in a Decentralized Architecture), a groundbreaking open protocol that could fundamentally transform how we interact with artificial intelligence.

    Drawing a fascinating parallel between computing history and our AI trajectory, Pradyumna explains how we've evolved from isolated large language models to action-capable agents that can reason and act in the world. Yet despite this progress, we're still missing the crucial infrastructure that would allow these agents to find and collaborate with each other across organizational boundaries – essentially, an "Internet of AI Agents."

    Using a relatable birthday party planning scenario, Pradyumna demonstrates how interconnected agents could effortlessly coordinate calendars, groceries, and bakery orders without human micromanagement. But enabling this vision requires solving complex challenges around agent discovery, authentication, verifiability, and privacy that differ significantly from traditional web architecture.

    At the heart of NANDA's approach is a three-layer registry system designed specifically for dynamic, peer-to-peer agent interactions. The demonstration showcases how this architecture enables diverse communications – from personal agents that adapt messages between family members to commercial interactions between customers and businesses, all while supporting different communication protocols like Google's A2A and Anthropic's MCP.

    What makes NANDA particularly exciting is its commitment to democratic, open-source development. Rather than dictating standards, the project invites collaboration from academic and industry partners to build this agent ecosystem together, ensuring it remains transparent, trustworthy, and accessible to all.

    Visit nanda.mit.edu to learn more about how you can contribute to this vision of a decentralized, collaborative future for artificial intelligence.

    Send us a text

    Support the show

    Learn more about the EDGE AI FOUNDATION - edgeaifoundation.org

    Show More Show Less
    38 mins
  • Generative AI on NXP Microprocessors
    Dec 9 2025

    Stepping into a future where AI doesn't require the cloud, NXP is revolutionizing edge computing by bringing generative AI directly to microprocessors. Alberto Alvarez offers an illuminating journey through NXP's approach to private, secure, and efficient AI inference that operates entirely at the edge.

    The heart of NXP's innovation is their EAQ GenAI Flow, a comprehensive software pipeline designed for iMX SoCs that enables both fine-tuning and optimization of AI models. This dual capability allows developers to adapt openly available Large Language Models for specific use cases without compromising data privacy, while also tackling the challenge of memory footprint through quantization techniques that maintain model accuracy. The conversational AI implementation creates a seamless experience by combining wake word detection, speech recognition, language processing with retrieval-augmented generation, and natural speech synthesis—all accelerated by NXP's Neutron NPU.

    Most striking is NXP's partnership with Kinara, which introduces truly groundbreaking multimodal AI capabilities running entirely at the edge. Their demonstration of the LAVA model—combining LLAMA3's 8 billion parameters with CLIP vision encoding—showcases the ability to process both images and language queries without any cloud connectivity. Imagine industrial systems analyzing visual scenes, detecting subtle anomalies like water spills, and providing spoken reports—all while keeping sensitive data completely private. With quantization reducing these massive models to manageable 4-bit and 8-bit precision, NXP is making previously impossible edge AI applications practical reality.

    Ready to experience the future of edge intelligence? Explore NXP's application code hub to start building with EIQ GenAI resources on compatible hardware and discover how your next project can harness the power of generative AI without surrendering privacy or security to the cloud.

    Send us a text

    Support the show

    Learn more about the EDGE AI FOUNDATION - edgeaifoundation.org

    Show More Show Less
    29 mins
No reviews yet
In the spirit of reconciliation, Audible acknowledges the Traditional Custodians of country throughout Australia and their connections to land, sea and community. We pay our respect to their elders past and present and extend that respect to all Aboriginal and Torres Strait Islander peoples today.