Principal Investigator: Song Han
Modern deep learning requires a massive amount of computational resource, carbon footprint, and engineering efforts. On mobile devices, the hardware resource and power budget are very limited, and on-device machine learning is challenging; retraining the model on-device is even more difficult. We make machine learning efficient and fit tiny devices (TinyML). We holistically developed a stack of techniques to (1) automatically synthesize efficient and tiny neural architectures given hardware constraints, outperforming the human design while requiring less engineer efforts; (2) train the model with tiny amount of data without losing accuracy, with better data efficiency; (3) compress the model and co-design the inference framework to fit tiny edge devices; (4) retrain the model locally on edge devices and customize for new data with tiny memory resources. We uncover new design insights under resource constraints. Our research is highlighted by full-stack optimizations, including the neural network topology, inference library, and the hardware architecture, which allows a larger design space to unearth the underlying principles.
Partially funded by the MIT-IBM Watson AI Lab, and IBM Research.
Related Publications
TinyTL: Reduce Activations, Not Trainable Parameters for Efficient On-Device Learning, NeurIPS 2020
Delayed Gradient Averaging: Tolerate the Communication Latency in Federated Learning, NeurIPS 2021
Network Augmentation for Tiny Deep Learning, ICLR 2022
Explore
This Compact, Low-Power Receiver could Give a Boost to 5G Smart Devices
Adam Zewe | MIT News
Researchers designed a tiny receiver chip that is more resilient to interference, which could enable smaller 5G “internet of things” devices with longer battery lives.
Photonic Processor could Streamline 6G Wireless Signal Processing
Adam Zewe | MIT News
By performing deep learning at the speed of light, this chip could give edge devices new capabilities for real-time data analysis.
AI Tool Generates High-Quality Images Faster Than State-of-the-Art Approaches
Adam Zewe | MIT News
Researchers fuse the best of two popular methods to create an image generator that uses less energy and can run locally on a laptop or smartphone.