Dynamic neural networks: A survey
Dynamic neural network is an emerging research topic in deep learning. Compared to static
models which have fixed computational graphs and parameters at the inference stage …
models which have fixed computational graphs and parameters at the inference stage …
Ghostnet: More features from cheap operations
Deploying convolutional neural networks (CNNs) on embedded devices is difficult due to the
limited memory and computation resources. The redundancy in feature maps is an important …
limited memory and computation resources. The redundancy in feature maps is an important …
GhostNets on heterogeneous devices via cheap operations
Deploying convolutional neural networks (CNNs) on mobile devices is difficult due to the
limited memory and computation resources. We aim to design efficient neural networks for …
limited memory and computation resources. We aim to design efficient neural networks for …
Manifold regularized dynamic network pruning
Neural network pruning is an essential approach for reducing the computational complexity
of deep models so that they can be well deployed on resource-limited devices. Compared …
of deep models so that they can be well deployed on resource-limited devices. Compared …
Adapting Neural Networks at Runtime: Current Trends in At-Runtime Optimizations for Deep Learning
Adaptive optimization methods for deep learning adjust the inference task to the current
circumstances at runtime to improve the resource footprint while maintaining the model's …
circumstances at runtime to improve the resource footprint while maintaining the model's …
Dynamic neural network structure: A review for its theories and applications
The dynamic neural network (DNN), in contrast to the static counterpart, offers numerous
advantages, such as improved accuracy, efficiency, and interpretability. These benefits stem …
advantages, such as improved accuracy, efficiency, and interpretability. These benefits stem …
Towards performance-maximizing neural network pruning via global channel attention
Network pruning has attracted increasing attention recently for its capability of transferring
large-scale neural networks (eg, CNNs) into resource-constrained devices. Such a transfer …
large-scale neural networks (eg, CNNs) into resource-constrained devices. Such a transfer …
Kernel based progressive distillation for adder neural networks
Abstract Adder Neural Networks (ANNs) which only contain additions bring us a new way of
developing deep neural networks with low energy consumption. Unfortunately, there is an …
developing deep neural networks with low energy consumption. Unfortunately, there is an …
Prior gradient mask guided pruning-aware fine-tuning
Abstract We proposed a Prior Gradient Mask Guided Pruning-aware Fine-Tuning (PGMPF)
framework to accelerate deep Convolutional Neural Networks (CNNs). In detail, the …
framework to accelerate deep Convolutional Neural Networks (CNNs). In detail, the …
[HTML][HTML] Zero time waste in pre-trained early exit neural networks
The problem of reducing processing time of large deep learning models is a fundamental
challenge in many real-world applications. Early exit methods strive towards this goal by …
challenge in many real-world applications. Early exit methods strive towards this goal by …