Structured pruning for deep convolutional neural networks: A survey

Y He, L Xiao - IEEE transactions on pattern analysis and …, 2023 - ieeexplore.ieee.org
The remarkable performance of deep Convolutional neural networks (CNNs) is generally
attributed to their deeper and wider architectures, which can come with significant …

[HTML][HTML] Applications and techniques for fast machine learning in science

AMC Deiana, N Tran, J Agar, M Blott… - Frontiers in big …, 2022 - frontiersin.org
In this community review report, we discuss applications and techniques for fast machine
learning (ML) in science—the concept of integrating powerful ML methods into the real-time …

A survey of quantization methods for efficient neural network inference

A Gholami, S Kim, Z Dong, Z Yao… - Low-Power Computer …, 2022 - taylorfrancis.com
This chapter provides approaches to the problem of quantizing the numerical values in deep
Neural Network computations, covering the advantages/disadvantages of current methods …

Hrank: Filter pruning using high-rank feature map

M Lin, R Ji, Y Wang, Y Zhang… - Proceedings of the …, 2020 - openaccess.thecvf.com
Neural network pruning offers a promising prospect to facilitate deploying deep neural
networks on resource-limited devices. However, existing methods are still challenged by the …

Group sparsity: The hinge between filter pruning and decomposition for network compression

Y Li, S Gu, C Mayer, LV Gool… - Proceedings of the …, 2020 - openaccess.thecvf.com
In this paper, we analyze two popular network compression techniques, ie filter pruning and
low-rank decomposition, in a unified sense. By simply changing the way the sparsity …

Operation-aware soft channel pruning using differentiable masks

M Kang, B Han - International conference on machine …, 2020 - proceedings.mlr.press
We propose a simple but effective data-driven channel pruning algorithm, which
compresses deep neural networks in a differentiable way by exploiting the characteristics of …

Self-regulated feature learning via teacher-free feature distillation

L Li - European Conference on Computer Vision, 2022 - Springer
Abstract Knowledge distillation conditioned on intermediate feature representations always
leads to significant performance improvements. Conventional feature distillation framework …

Towards efficient model compression via learned global ranking

TW Chin, R Ding, C Zhang… - Proceedings of the …, 2020 - openaccess.thecvf.com
Pruning convolutional filters has demonstrated its effectiveness in compressing ConvNets.
Prior art in filter pruning requires users to specify a target model complexity (eg, model size …

Yolobile: Real-time object detection on mobile devices via compression-compilation co-design

Y Cai, H Li, G Yuan, W Niu, Y Li, X Tang… - Proceedings of the …, 2021 - ojs.aaai.org
The rapid development and wide utilization of object detection techniques have aroused
attention on both accuracy and speed of object detectors. However, the current state-of-the …

Methods for pruning deep neural networks

S Vadera, S Ameen - IEEE Access, 2022 - ieeexplore.ieee.org
This paper presents a survey of methods for pruning deep neural networks. It begins by
categorising over 150 studies based on the underlying approach used and then focuses on …