Structured pruning for deep convolutional neural networks: A survey

Y He, L Xiao - IEEE transactions on pattern analysis and …, 2023 - ieeexplore.ieee.org
The remarkable performance of deep Convolutional neural networks (CNNs) is generally
attributed to their deeper and wider architectures, which can come with significant …

An overview of neural network compression

JO Neill - arXiv preprint arXiv:2006.03669, 2020 - arxiv.org
Overparameterized networks trained to convergence have shown impressive performance
in domains such as computer vision and natural language processing. Pushing state of the …

Beyond transmitting bits: Context, semantics, and task-oriented communications

D Gündüz, Z Qin, IE Aguerri, HS Dhillon… - IEEE Journal on …, 2022 - ieeexplore.ieee.org
Communication systems to date primarily aim at reliably communicating bit sequences.
Such an approach provides efficient engineering designs that are agnostic to the meanings …

Sparsity in deep learning: Pruning and growth for efficient inference and training in neural networks

T Hoefler, D Alistarh, T Ben-Nun, N Dryden… - Journal of Machine …, 2021 - jmlr.org
The growing energy and performance costs of deep learning have driven the community to
reduce the size of neural networks by selectively pruning components. Similarly to their …

Meta-learning in neural networks: A survey

T Hospedales, A Antoniou, P Micaelli… - IEEE transactions on …, 2021 - ieeexplore.ieee.org
The field of meta-learning, or learning-to-learn, has seen a dramatic rise in interest in recent
years. Contrary to conventional approaches to AI where tasks are solved from scratch using …

Rigging the lottery: Making all tickets winners

U Evci, T Gale, J Menick, PS Castro… - … on machine learning, 2020 - proceedings.mlr.press
Many applications require sparse neural networks due to space or inference time
restrictions. There is a large body of work on training dense networks to yield sparse …

The state of sparsity in deep neural networks

T Gale, E Elsen, S Hooker - arXiv preprint arXiv:1902.09574, 2019 - arxiv.org
We rigorously evaluate three state-of-the-art techniques for inducing sparsity in deep neural
networks on two large-scale learning tasks: Transformer trained on WMT 2014 English-to …

Mest: Accurate and fast memory-economic sparse training framework on the edge

G Yuan, X Ma, W Niu, Z Li, Z Kong… - Advances in …, 2021 - proceedings.neurips.cc
Recently, a new trend of exploring sparsity for accelerating neural network training has
emerged, embracing the paradigm of training on the edge. This paper proposes a novel …

Do we actually need dense over-parameterization? in-time over-parameterization in sparse training

S Liu, L Yin, DC Mocanu… - … on Machine Learning, 2021 - proceedings.mlr.press
In this paper, we introduce a new perspective on training deep neural networks capable of
state-of-the-art performance without the need for the expensive over-parameterization by …

Layer-adaptive sparsity for the magnitude-based pruning

J Lee, S Park, S Mo, S Ahn, J Shin - arXiv preprint arXiv:2010.07611, 2020 - arxiv.org
Recent discoveries on neural network pruning reveal that, with a carefully chosen layerwise
sparsity, a simple magnitude-based pruning achieves state-of-the-art tradeoff between …