What is the state of neural network pruning?
D Blalock, JJ Gonzalez Ortiz… - … of machine learning …, 2020 - proceedings.mlsys.org
Neural network pruning---the task of reducing the size of a network by removing parameters--
-has been the subject of a great deal of work in recent years. We provide a meta-analysis of …
-has been the subject of a great deal of work in recent years. We provide a meta-analysis of …
Patient knowledge distillation for bert model compression
S Sun, Y Cheng, Z Gan, J Liu - arXiv preprint arXiv:1908.09355, 2019 - arxiv.org
Pre-trained language models such as BERT have proven to be highly effective for natural
language processing (NLP) tasks. However, the high demand for computing resources in …
language processing (NLP) tasks. However, the high demand for computing resources in …
Mobilenetv2: Inverted residuals and linear bottlenecks
M Sandler, A Howard, M Zhu… - Proceedings of the …, 2018 - openaccess.thecvf.com
In this paper we describe a new mobile architecture, mbox {MobileNetV2}, that improves the
state of the art performance of mobile models on multiple tasks and benchmarks as well as …
state of the art performance of mobile models on multiple tasks and benchmarks as well as …
Inverted residuals and linear bottlenecks: Mobile networks for classification, detection and segmentation
A Howard, A Zhmoginov, LC Chen, M Sandler… - Proc. CVPR, 2018 - research.google
In this paper we describe a new mobile architecture MobileNetV2 that improves the state of
the art performance of mobile models on multiple benchmarks across a spectrum of different …
the art performance of mobile models on multiple benchmarks across a spectrum of different …
To prune, or not to prune: exploring the efficacy of pruning for model compression
M Zhu, S Gupta - arXiv preprint arXiv:1710.01878, 2017 - arxiv.org
Model pruning seeks to induce sparsity in a deep neural network's various connection
matrices, thereby reducing the number of nonzero-valued parameters in the model. Recent …
matrices, thereby reducing the number of nonzero-valued parameters in the model. Recent …
Learning efficient convolutional networks through network slimming
Z Liu, J Li, Z Shen, G Huang, S Yan… - Proceedings of the …, 2017 - openaccess.thecvf.com
The deployment of deep convolutional neural networks (CNNs) in many real world
applications is largely hindered by their high computational cost. In this paper, we propose a …
applications is largely hindered by their high computational cost. In this paper, we propose a …
Variational dropout sparsifies deep neural networks
D Molchanov, A Ashukha… - … conference on machine …, 2017 - proceedings.mlr.press
We explore a recently proposed Variational Dropout technique that provided an elegant
Bayesian interpretation to Gaussian Dropout. We extend Variational Dropout to the case …
Bayesian interpretation to Gaussian Dropout. We extend Variational Dropout to the case …
Scalable training of artificial neural networks with adaptive sparse connectivity inspired by network science
DC Mocanu, E Mocanu, P Stone, PH Nguyen… - Nature …, 2018 - nature.com
Through the success of deep learning in various domains, artificial neural networks are
currently among the most used artificial intelligence methods. Taking inspiration from the …
currently among the most used artificial intelligence methods. Taking inspiration from the …
Interleaved group convolutions
T Zhang, GJ Qi, B Xiao, J Wang - Proceedings of the IEEE …, 2017 - openaccess.thecvf.com
In this paper, we present a simple and modularized neural network architecture, named
interleaved group convolutional neural networks (IGCNets). The main point lies in a novel …
interleaved group convolutional neural networks (IGCNets). The main point lies in a novel …
A comprehensive guide to bayesian convolutional neural network with variational inference
K Shridhar, F Laumann, M Liwicki - arXiv preprint arXiv:1901.02731, 2019 - arxiv.org
Artificial Neural Networks are connectionist systems that perform a given task by learning on
examples without having prior knowledge about the task. This is done by finding an optimal …
examples without having prior knowledge about the task. This is done by finding an optimal …