Structured pruning for deep convolutional neural networks: A survey

Y He, L Xiao - IEEE transactions on pattern analysis and …, 2023 - ieeexplore.ieee.org
The remarkable performance of deep Convolutional neural networks (CNNs) is generally
attributed to their deeper and wider architectures, which can come with significant …

Text-visual prompting for efficient 2d temporal video grounding

Y Zhang, X Chen, J Jia, S Liu… - Proceedings of the IEEE …, 2023 - openaccess.thecvf.com
In this paper, we study the problem of temporal video grounding (TVG), which aims to predict
the starting/ending time points of moments described by a text sentence within a long …

Convolutional neural network pruning based on multi-objective feature map selection for image classification

P Jiang, Y Xue, F Neri - Applied soft computing, 2023 - Elsevier
Deep convolutional neural networks (CNNs) are widely used for image classification. Deep
CNNs often require a large memory and abundant computation resources, limiting their …

Performance-aware approximation of global channel pruning for multitask cnns

H Ye, B Zhang, T Chen, J Fan… - IEEE Transactions on …, 2023 - ieeexplore.ieee.org
Global channel pruning (GCP) aims to remove a subset of channels (filters) across different
layers from a deep model without hurting the performance. Previous works focus on either …

Towards deploying DNN models on edge for predictive maintenance applications

R Pandey, S Uziel, T Hutschenreuther, S Krug - Electronics, 2023 - mdpi.com
Almost all rotating machinery in the industry has bearings as their key building block and
most of these machines run 24× 7. This makes bearing health prediction an active research …

Pruning foundation models for high accuracy without retraining

P Zhao, F Sun, X Shen, P Yu, Z Kong, Y Wang… - arXiv preprint arXiv …, 2024 - arxiv.org
Despite the superior performance, it is challenging to deploy foundation models or large
language models (LLMs) due to their massive parameters and computations. While pruning …

CoAxNN: Optimizing on-device deep learning with conditional approximate neural networks

G Li, X Ma, Q Yu, L Liu, H Liu, X Wang - Journal of Systems Architecture, 2023 - Elsevier
While deep neural networks have achieved superior performance in a variety of intelligent
applications, the increasing computational complexity makes them difficult to be deployed …

End-to-End Neural Network Compression via l1/l2 Regularized Latency Surrogates

A Nasery, H Shah, AS Suggala… - Proceedings of the IEEE …, 2024 - openaccess.thecvf.com
Neural network (NN) compression via techniques such as pruning quantization requires
setting compression hyperparameters (eg number of channels to be pruned bitwidths for …

UPSCALE: unconstrained channel pruning

A Wan, H Hao, K Patnaik, Y Xu… - International …, 2023 - proceedings.mlr.press
As neural networks grow in size and complexity, inference speeds decline. To combat this,
one of the most effective compression techniques–channel pruning–removes channels from …

Equivariance-aware architectural optimization of neural networks

K Maile, DG Wilson, P Forré - arXiv preprint arXiv:2210.05484, 2022 - arxiv.org
Incorporating equivariance to symmetry groups as a constraint during neural network
training can improve performance and generalization for tasks exhibiting those symmetries …