A survey on deep neural network pruning: Taxonomy, comparison, analysis, and recommendations

H Cheng, M Zhang, JQ Shi - IEEE Transactions on Pattern …, 2024 - ieeexplore.ieee.org
Modern deep neural networks, particularly recent large language models, come with
massive model sizes that require significant computational and storage resources. To …

Sparsity in deep learning: Pruning and growth for efficient inference and training in neural networks

T Hoefler, D Alistarh, T Ben-Nun, N Dryden… - Journal of Machine …, 2021 - jmlr.org
The growing energy and performance costs of deep learning have driven the community to
reduce the size of neural networks by selectively pruning components. Similarly to their …

Pruning neural networks without any data by iteratively conserving synaptic flow

H Tanaka, D Kunin, DL Yamins… - Advances in neural …, 2020 - proceedings.neurips.cc
Pruning the parameters of deep neural networks has generated intense interest due to
potential savings in time, memory and energy both during training and at test time. Recent …

Sparse training via boosting pruning plasticity with neuroregeneration

S Liu, T Chen, X Chen, Z Atashgahi… - Advances in …, 2021 - proceedings.neurips.cc
Works on lottery ticket hypothesis (LTH) and single-shot network pruning (SNIP) have raised
a lot of attention currently on post-training pruning (iterative magnitude pruning), and before …

The emergence of essential sparsity in large pre-trained models: The weights that matter

A Jaiswal, S Liu, T Chen… - Advances in Neural …, 2024 - proceedings.neurips.cc
Large pre-trained transformers are $\textit {show-stealer} $ in modern-day deep learning,
and it becomes crucial to comprehend the parsimonious patterns that exist within them as …

Do we actually need dense over-parameterization? in-time over-parameterization in sparse training

S Liu, L Yin, DC Mocanu… - … on Machine Learning, 2021 - proceedings.mlr.press
In this paper, we introduce a new perspective on training deep neural networks capable of
state-of-the-art performance without the need for the expensive over-parameterization by …

Structural pruning via latency-saliency knapsack

M Shen, H Yin, P Molchanov, L Mao… - Advances in Neural …, 2022 - proceedings.neurips.cc
Structural pruning can simplify network architecture and improve inference speed. We
propose Hardware-Aware Latency Pruning (HALP) that formulates structural pruning as a …

Efficient joint optimization of layer-adaptive weight pruning in deep neural networks

K Xu, Z Wang, X Geng, M Wu, X Li… - Proceedings of the …, 2023 - openaccess.thecvf.com
In this paper, we propose a novel layer-adaptive weight-pruning approach for Deep Neural
Networks (DNNs) that addresses the challenge of optimizing the output distortion …

When to prune? a policy towards early structural pruning

M Shen, P Molchanov, H Yin… - Proceedings of the …, 2022 - openaccess.thecvf.com
Pruning enables appealing reductions in network memory footprint and time complexity.
Conventional post-training pruning techniques lean towards efficient inference while …

Winning the lottery ahead of time: Efficient early network pruning

J Rachwan, D Zügner, B Charpentier… - International …, 2022 - proceedings.mlr.press
Pruning, the task of sparsifying deep neural networks, received increasing attention recently.
Although state-of-the-art pruning methods extract highly sparse models, they neglect two …