Structured pruning for deep convolutional neural networks: A survey
The remarkable performance of deep Convolutional neural networks (CNNs) is generally
attributed to their deeper and wider architectures, which can come with significant …
attributed to their deeper and wider architectures, which can come with significant …
Pruning's effect on generalization through the lens of training and regularization
Practitioners frequently observe that pruning improves model generalization. A long-
standing hypothesis based on bias-variance trade-off attributes this generalization …
standing hypothesis based on bias-variance trade-off attributes this generalization …
Efficient joint optimization of layer-adaptive weight pruning in deep neural networks
In this paper, we propose a novel layer-adaptive weight-pruning approach for Deep Neural
Networks (DNNs) that addresses the challenge of optimizing the output distortion …
Networks (DNNs) that addresses the challenge of optimizing the output distortion …
The combinatorial brain surgeon: pruning weights that cancel one another in neural networks
Neural networks tend to achieve better accuracy with training if they are larger {—} even if
the resulting models are overparameterized. Nevertheless, carefully removing such excess …
the resulting models are overparameterized. Nevertheless, carefully removing such excess …
Otov2: Automatic, generic, user-friendly
The existing model compression methods via structured pruning typically require
complicated multi-stage procedures. Each individual stage necessitates numerous …
complicated multi-stage procedures. Each individual stage necessitates numerous …
An architecture-level analysis on deep learning models for low-impact computations
Deep neural networks (DNNs) have made significant achievements in a wide variety of
domains. For the deep learning tasks, multiple excellent hardware platforms provide efficient …
domains. For the deep learning tasks, multiple excellent hardware platforms provide efficient …
Topology-aware network pruning using multi-stage graph embedding and reinforcement learning
Abstract Model compression is an essential technique for deploying deep neural networks
(DNNs) on power and memory-constrained resources. However, existing model …
(DNNs) on power and memory-constrained resources. However, existing model …
Auto-Train-Once: Controller Network Guided Automatic Network Pruning from Scratch
Current techniques for deep neural network (DNN) pruning often involve intricate multi-step
processes that require domain-specific expertise making their widespread adoption …
processes that require domain-specific expertise making their widespread adoption …
Towards data-agnostic pruning at initialization: what makes a good sparse mask?
Pruning at initialization (PaI) aims to remove weights of neural networks before training in
pursuit of training efficiency besides the inference. While off-the-shelf PaI methods manage …
pursuit of training efficiency besides the inference. While off-the-shelf PaI methods manage …