Structured pruning for deep convolutional neural networks: A survey
The remarkable performance of deep Convolutional neural networks (CNNs) is generally
attributed to their deeper and wider architectures, which can come with significant …
attributed to their deeper and wider architectures, which can come with significant …
Text-visual prompting for efficient 2d temporal video grounding
In this paper, we study the problem of temporal video grounding (TVG), which aims to predict
the starting/ending time points of moments described by a text sentence within a long …
the starting/ending time points of moments described by a text sentence within a long …
Convolutional neural network pruning based on multi-objective feature map selection for image classification
Deep convolutional neural networks (CNNs) are widely used for image classification. Deep
CNNs often require a large memory and abundant computation resources, limiting their …
CNNs often require a large memory and abundant computation resources, limiting their …
Performance-aware approximation of global channel pruning for multitask cnns
Global channel pruning (GCP) aims to remove a subset of channels (filters) across different
layers from a deep model without hurting the performance. Previous works focus on either …
layers from a deep model without hurting the performance. Previous works focus on either …
Towards deploying DNN models on edge for predictive maintenance applications
R Pandey, S Uziel, T Hutschenreuther, S Krug - Electronics, 2023 - mdpi.com
Almost all rotating machinery in the industry has bearings as their key building block and
most of these machines run 24× 7. This makes bearing health prediction an active research …
most of these machines run 24× 7. This makes bearing health prediction an active research …
Pruning foundation models for high accuracy without retraining
Despite the superior performance, it is challenging to deploy foundation models or large
language models (LLMs) due to their massive parameters and computations. While pruning …
language models (LLMs) due to their massive parameters and computations. While pruning …
CoAxNN: Optimizing on-device deep learning with conditional approximate neural networks
While deep neural networks have achieved superior performance in a variety of intelligent
applications, the increasing computational complexity makes them difficult to be deployed …
applications, the increasing computational complexity makes them difficult to be deployed …
End-to-End Neural Network Compression via l1/l2 Regularized Latency Surrogates
Neural network (NN) compression via techniques such as pruning quantization requires
setting compression hyperparameters (eg number of channels to be pruned bitwidths for …
setting compression hyperparameters (eg number of channels to be pruned bitwidths for …
UPSCALE: unconstrained channel pruning
As neural networks grow in size and complexity, inference speeds decline. To combat this,
one of the most effective compression techniques–channel pruning–removes channels from …
one of the most effective compression techniques–channel pruning–removes channels from …
Equivariance-aware architectural optimization of neural networks
Incorporating equivariance to symmetry groups as a constraint during neural network
training can improve performance and generalization for tasks exhibiting those symmetries …
training can improve performance and generalization for tasks exhibiting those symmetries …