학술논문

Tiny Machine Learning: Progress and Futures
Document Type
Working Paper
Source
IEEE Circuits and Systems Magazine, 23(3), pp. 8-34, October 2023
Subject
Computer Science - Machine Learning
Computer Science - Artificial Intelligence
Computer Science - Computer Vision and Pattern Recognition
Language
Abstract
Tiny Machine Learning (TinyML) is a new frontier of machine learning. By squeezing deep learning models into billions of IoT devices and microcontrollers (MCUs), we expand the scope of AI applications and enable ubiquitous intelligence. However, TinyML is challenging due to hardware constraints: the tiny memory resource makes it difficult to hold deep learning models designed for cloud and mobile platforms. There is also limited compiler and inference engine support for bare-metal devices. Therefore, we need to co-design the algorithm and system stack to enable TinyML. In this review, we will first discuss the definition, challenges, and applications of TinyML. We then survey the recent progress in TinyML and deep learning on MCUs. Next, we will introduce MCUNet, showing how we can achieve ImageNet-scale AI applications on IoT devices with system-algorithm co-design. We will further extend the solution from inference to training and introduce tiny on-device training techniques. Finally, we present future directions in this area. Today's large model might be tomorrow's tiny model. The scope of TinyML should evolve and adapt over time.
Comment: arXiv admin note: text overlap with arXiv:2206.15472