Structured pruning for deep convolutional neural networks: A survey

Y He, L Xiao - IEEE transactions on pattern analysis and …, 2023 - ieeexplore.ieee.org
The remarkable performance of deep Convolutional neural networks (CNNs) is generally
attributed to their deeper and wider architectures, which can come with significant …

Model compression for deep neural networks: A survey

Z Li, H Li, L Meng - Computers, 2023 - mdpi.com
Currently, with the rapid development of deep learning, deep neural networks (DNNs) have
been widely applied in various computer vision tasks. However, in the pursuit of …

Pruning's effect on generalization through the lens of training and regularization

T Jin, M Carbin, D Roy, J Frankle… - Advances in Neural …, 2022 - proceedings.neurips.cc
Practitioners frequently observe that pruning improves model generalization. A long-
standing hypothesis based on bias-variance trade-off attributes this generalization …

Efficient joint optimization of layer-adaptive weight pruning in deep neural networks

K Xu, Z Wang, X Geng, M Wu, X Li… - Proceedings of the …, 2023 - openaccess.thecvf.com
In this paper, we propose a novel layer-adaptive weight-pruning approach for Deep Neural
Networks (DNNs) that addresses the challenge of optimizing the output distortion …

The combinatorial brain surgeon: pruning weights that cancel one another in neural networks

X Yu, T Serra, S Ramalingam… - … Conference on Machine …, 2022 - proceedings.mlr.press
Neural networks tend to achieve better accuracy with training if they are larger {—} even if
the resulting models are overparameterized. Nevertheless, carefully removing such excess …

Otov2: Automatic, generic, user-friendly

T Chen, L Liang, T Ding, Z Zhu, I Zharkov - arXiv preprint arXiv …, 2023 - arxiv.org
The existing model compression methods via structured pruning typically require
complicated multi-stage procedures. Each individual stage necessitates numerous …

An architecture-level analysis on deep learning models for low-impact computations

H Li, Z Wang, X Yue, W Wang, H Tomiyama… - Artificial Intelligence …, 2023 - Springer
Deep neural networks (DNNs) have made significant achievements in a wide variety of
domains. For the deep learning tasks, multiple excellent hardware platforms provide efficient …

Topology-aware network pruning using multi-stage graph embedding and reinforcement learning

S Yu, A Mazaheri, A Jannesari - International conference on …, 2022 - proceedings.mlr.press
Abstract Model compression is an essential technique for deploying deep neural networks
(DNNs) on power and memory-constrained resources. However, existing model …

Auto-Train-Once: Controller Network Guided Automatic Network Pruning from Scratch

X Wu, S Gao, Z Zhang, Z Li, R Bao… - Proceedings of the …, 2024 - openaccess.thecvf.com
Current techniques for deep neural network (DNN) pruning often involve intricate multi-step
processes that require domain-specific expertise making their widespread adoption …

Towards data-agnostic pruning at initialization: what makes a good sparse mask?

H Pham, S Liu, L Xiang, D Le, H Wen… - Advances in Neural …, 2024 - proceedings.neurips.cc
Pruning at initialization (PaI) aims to remove weights of neural networks before training in
pursuit of training efficiency besides the inference. While off-the-shelf PaI methods manage …