Structured pruning for deep convolutional neural networks: A survey

Y He, L Xiao - IEEE transactions on pattern analysis and …, 2023 - ieeexplore.ieee.org
The remarkable performance of deep Convolutional neural networks (CNNs) is generally
attributed to their deeper and wider architectures, which can come with significant …

Learning best combination for efficient n: M sparsity

Y Zhang, M Lin, Z Lin, Y Luo, K Li… - Advances in Neural …, 2022 - proceedings.neurips.cc
By forcing N out of M consecutive weights to be non-zero, the recent N: M fine-grained
network sparsity has received increasing attention with its two attractive advantages over …

Rgp: Neural network pruning through regular graph with edges swapping

Z Chen, J Xiang, Y Lu, Q Xuan, Z Wang… - … on Neural Networks …, 2023 - ieeexplore.ieee.org
Deep learning technology has found a promising application in lightweight model design, for
which pruning is an effective means of achieving a large reduction in both model parameters …

Model compression of deep neural network architectures for visual pattern recognition: Current status and future directions

S Bhalgaonkar, M Munot - Computers and Electrical Engineering, 2024 - Elsevier
Abstract Visual Pattern Recognition Networks (VPRNs) are widely used in various visual
data based applications such as computer vision and edge AI. VPRNs help to enhance a …

Multidimensional pruning and its extension: A unified framework for model compression

J Guo, D Xu, W Ouyang - IEEE Transactions on Neural …, 2023 - ieeexplore.ieee.org
Observing that the existing model compression approaches only focus on reducing the
redundancies in convolutional neural networks (CNNs) along one particular dimension (eg …

EACP: An effective automatic channel pruning for neural networks

Y Liu, D Wu, W Zhou, K Fan, Z Zhou - Neurocomputing, 2023 - Elsevier
The large data scale and computational resources required by Convolutional Neural
Networks (CNNs) hinder the practical application on mobile devices. However, channel …

Reaf: Remembering enhancement and entropy-based asymptotic forgetting for filter pruning

X Zhang, W Xie, Y Li, K Jiang… - IEEE Transactions on …, 2023 - ieeexplore.ieee.org
Neurologically, filter pruning is a procedure of forgetting and remembering recovering.
Prevailing methods directly forget less important information from an unrobust baseline at …

FPFS: Filter-level pruning via distance weight measuring filter similarity

W Zhang, Z Wang - Neurocomputing, 2022 - Elsevier
Abstract Deep Neural Networks (DNNs) enjoy the welfare of convolution, while also bearing
huge computational pressure. Therefore, model compression techniques are used to …

An accelerating convolutional neural networks via a 2D entropy based-adaptive filter search method for image recognition

C Li, H Li, G Gao, Z Liu, P Liu - Applied Soft Computing, 2023 - Elsevier
The success of CNNs for various vision tasks has been accompanied by a significant
increase in required FLOPs and parameter quantities, which has impeded the deployment of …

Co-exploring structured sparsification and low-rank tensor decomposition for compact dnns

Y Sui, M Yin, Y Gong, B Yuan - IEEE Transactions on Neural …, 2024 - ieeexplore.ieee.org
Sparsification and low-rank decomposition are two important techniques to compress deep
neural network (DNN) models. To date, these two popular yet distinct approaches are …