Tinyml meets iot: A comprehensive survey

L Dutta, S Bharali - Internet of Things, 2021 - Elsevier
The rapid growth in miniaturization of low-power embedded devices and advancement in
the optimization of machine learning (ML) algorithms have opened up a new prospect of the …

A review of convolutional neural network architectures and their optimizations

S Cong, Y Zhou - Artificial Intelligence Review, 2023 - Springer
The research advances concerning the typical architectures of convolutional neural
networks (CNNs) as well as their optimizations are analyzed and elaborated in detail in this …

Spvit: Enabling faster vision transformers via latency-aware soft token pruning

Z Kong, P Dong, X Ma, X Meng, W Niu, M Sun… - European conference on …, 2022 - Springer
Abstract Recently, Vision Transformer (ViT) has continuously established new milestones in
the computer vision field, while the high computation and memory cost makes its …

Efficient acceleration of deep learning inference on resource-constrained edge devices: A review

MMH Shuvo, SK Islam, J Cheng… - Proceedings of the …, 2022 - ieeexplore.ieee.org
Successful integration of deep neural networks (DNNs) or deep learning (DL) has resulted
in breakthroughs in many areas. However, deploying these highly accurate models for data …

Tinyml-enabled frugal smart objects: Challenges and opportunities

R Sanchez-Iborra, AF Skarmeta - IEEE Circuits and Systems …, 2020 - ieeexplore.ieee.org
The TinyML paradigm proposes to integrate Machine Learning (ML)-based mechanisms
within small objects powered by Microcontroller Units (MCUs). This paves the way for the …

TinyML for ultra-low power AI and large scale IoT deployments: A systematic review

N Schizas, A Karras, C Karras, S Sioutas - Future Internet, 2022 - mdpi.com
The rapid emergence of low-power embedded devices and modern machine learning (ML)
algorithms has created a new Internet of Things (IoT) era where lightweight ML frameworks …

Enable deep learning on mobile devices: Methods, systems, and applications

H Cai, J Lin, Y Lin, Z Liu, H Tang, H Wang… - ACM Transactions on …, 2022 - dl.acm.org
Deep neural networks (DNNs) have achieved unprecedented success in the field of artificial
intelligence (AI), including computer vision, natural language processing, and speech …

Sparcl: Sparse continual learning on the edge

Z Wang, Z Zhan, Y Gong, G Yuan… - Advances in …, 2022 - proceedings.neurips.cc
Existing work in continual learning (CL) focuses on mitigating catastrophic forgetting, ie,
model performance deterioration on past tasks when learning a new task. However, the …

Mest: Accurate and fast memory-economic sparse training framework on the edge

G Yuan, X Ma, W Niu, Z Li, Z Kong… - Advances in …, 2021 - proceedings.neurips.cc
Recently, a new trend of exploring sparsity for accelerating neural network training has
emerged, embracing the paradigm of training on the edge. This paper proposes a novel …

Dnnfusion: accelerating deep neural networks execution with advanced operator fusion

W Niu, J Guan, Y Wang, G Agrawal, B Ren - Proceedings of the 42nd …, 2021 - dl.acm.org
Deep Neural Networks (DNNs) have emerged as the core enabler of many major
applications on mobile devices. To achieve high accuracy, DNN models have become …