Lightning Talk: Flexible Deployment of PyTorch Models on MCU-Class... Robert Kalmar & Martin Pavella
Skills:
Model Deployment90%
Lightning Talk: Flexible Deployment of PyTorch Models on MCU-Class Devices Using ExecuTorch - Robert Kalmar & Martin Pavella, NXP
ExecuTorch has recently matured into a production ready framework designed specifically for efficient edge deployment of PyTorch models. Its architecture supports a broad spectrum of hardware targets—from low power, bare metal or RTOS based microcontrollers (MCU) to higher performance Linux or Android based microprocessor platforms—while meeting the demanding constraints of memory, compute, and power typically found in real world embedded applications.
This talk focuses on the deployment flexibility ExecuTorch offers for MCU class devices, highlighting how different backends enable efficient execution across heterogeneous compute units. We will explore CPU, DSP, and NPU acceleration paths using the Cortex-M, Cadence, Ethos-U, and eIQ Neutron backends, and discuss how these integrate into typical ML model deployment workflows.
To make the session practical and application oriented, we will present an optimization journey aimed at reducing power consumption—an essential requirement for ML workloads in energy constrained environments. Attendees will gain insights into backend selection, performance trade offs, and best practices for suitable deploying PyTorch models on edge devices.
Watch on YouTube ↗
(saves to browser)
Sign in to unlock AI tutor explanation · ⚡30
More on: Model Deployment
View skill →Related AI Lessons
⚡
⚡
⚡
⚡
`setTimeout()` Is NOT Part of JavaScript
Dev.to · CodeWithIshwar
Installing Node.js and npm on Ubuntu 26.04
Dev.to · Sanskriti Harmukh
How to Modernize a Node.js Backend Without Rewriting It (Using Zuplo)
Dev.to · Chidera Humphrey
Firebase for Startups: When to Switch to Enterprise Solutions
Dev.to · Horizon Dev
🎓
Tutor Explanation
DeepCamp AI