A survey on deep neural network pruning: Taxonomy, comparison, analysis, and recommendations

H Cheng, M Zhang, JQ Shi - IEEE Transactions on Pattern …, 2024 - ieeexplore.ieee.org
Modern deep neural networks, particularly recent large language models, come with
massive model sizes that require significant computational and storage resources. To …

Structured pruning for deep convolutional neural networks: A survey

Y He, L Xiao - IEEE transactions on pattern analysis and …, 2023 - ieeexplore.ieee.org
The remarkable performance of deep Convolutional neural networks (CNNs) is generally
attributed to their deeper and wider architectures, which can come with significant …

Advancing neuromorphic computing with loihi: A survey of results and outlook

M Davies, A Wild, G Orchard… - Proceedings of the …, 2021 - ieeexplore.ieee.org
Deep artificial neural networks apply principles of the brain's information processing that led
to breakthroughs in machine learning spanning many problem domains. Neuromorphic …

Sparsity in deep learning: Pruning and growth for efficient inference and training in neural networks

T Hoefler, D Alistarh, T Ben-Nun, N Dryden… - Journal of Machine …, 2021 - jmlr.org
The growing energy and performance costs of deep learning have driven the community to
reduce the size of neural networks by selectively pruning components. Similarly to their …

Machine learning for microcontroller-class hardware: A review

SS Saha, SS Sandha, M Srivastava - IEEE Sensors Journal, 2022 - ieeexplore.ieee.org
The advancements in machine learning (ML) opened a new opportunity to bring intelligence
to the low-end Internet-of-Things (IoT) nodes, such as microcontrollers. Conventional ML …

Zero-cost proxies for lightweight nas

MS Abdelfattah, A Mehrotra, Ł Dudziak… - arXiv preprint arXiv …, 2021 - arxiv.org
Neural Architecture Search (NAS) is quickly becoming the standard methodology to design
neural network models. However, NAS is typically compute-intensive because multiple …

Neural architecture search without training

J Mellor, J Turner, A Storkey… - … conference on machine …, 2021 - proceedings.mlr.press
The time and effort involved in hand-designing deep neural networks is immense. This has
prompted the development of Neural Architecture Search (NAS) techniques to automate this …

Neural architecture search: Insights from 1000 papers

C White, M Safari, R Sukthanker, B Ru, T Elsken… - arXiv preprint arXiv …, 2023 - arxiv.org
In the past decade, advances in deep learning have resulted in breakthroughs in a variety of
areas, including computer vision, natural language understanding, speech recognition, and …

A fast post-training pruning framework for transformers

W Kwon, S Kim, MW Mahoney… - Advances in …, 2022 - proceedings.neurips.cc
Pruning is an effective way to reduce the huge inference cost of Transformer models.
However, prior work on pruning Transformers requires retraining the models. This can add …

Chip: Channel independence-based pruning for compact neural networks

Y Sui, M Yin, Y Xie, H Phan… - Advances in Neural …, 2021 - proceedings.neurips.cc
Filter pruning has been widely used for neural network compression because of its enabled
practical acceleration. To date, most of the existing filter pruning works explore the …