Communication-efficient distributed deep learning: A comprehensive survey

Z Tang, S Shi, W Wang, B Li, X Chu - arXiv preprint arXiv:2003.06307, 2020 - arxiv.org
Distributed deep learning (DL) has become prevalent in recent years to reduce training time
by leveraging multiple computing devices (eg, GPUs/TPUs) due to larger models and …

Sparsity in deep learning: Pruning and growth for efficient inference and training in neural networks

T Hoefler, D Alistarh, T Ben-Nun, N Dryden… - Journal of Machine …, 2021 - jmlr.org
The growing energy and performance costs of deep learning have driven the community to
reduce the size of neural networks by selectively pruning components. Similarly to their …

Model pruning enables efficient federated learning on edge devices

Y Jiang, S Wang, V Valls, BJ Ko… - … on Neural Networks …, 2022 - ieeexplore.ieee.org
Federated learning (FL) allows model training from local data collected by edge/mobile
devices while preserving data privacy, which has wide applicability to image and vision …

A comprehensive survey on training acceleration for large machine learning models in IoT

H Wang, Z Qu, Q Zhou, H Zhang, B Luo… - IEEE Internet of …, 2021 - ieeexplore.ieee.org
The ever-growing artificial intelligence (AI) applications have greatly reshaped our world in
many areas, eg, smart home, computer vision, natural language processing, etc. Behind …

Gossipfl: A decentralized federated learning framework with sparsified and adaptive communication

Z Tang, S Shi, B Li, X Chu - IEEE Transactions on Parallel and …, 2022 - ieeexplore.ieee.org
Recently, federated learning (FL) techniques have enabled multiple users to train machine
learning models collaboratively without data sharing. However, existing FL algorithms suffer …

Adaptive gradient sparsification for efficient federated learning: An online learning approach

P Han, S Wang, KK Leung - 2020 IEEE 40th international …, 2020 - ieeexplore.ieee.org
Federated learning (FL) is an emerging technique for training machine learning models
using geographically dispersed data collected by local entities. It includes local computation …

Joint model pruning and device selection for communication-efficient federated edge learning

S Liu, G Yu, R Yin, J Yuan, L Shen… - IEEE Transactions on …, 2021 - ieeexplore.ieee.org
In recent years, wireless federated learning (FL) has been proposed to support the mobile
intelligent applications over the wireless network, which protects the data privacy and …

Near-optimal sparse allreduce for distributed deep learning

S Li, T Hoefler - Proceedings of the 27th ACM SIGPLAN Symposium on …, 2022 - dl.acm.org
Communication overhead is one of the major obstacles to train large deep learning models
at scale. Gradient sparsification is a promising technique to reduce the communication …

Compressed-vfl: Communication-efficient learning with vertically partitioned data

TJ Castiglia, A Das, S Wang… - … on Machine Learning, 2022 - proceedings.mlr.press
Abstract We propose Compressed Vertical Federated Learning (C-VFL) for communication-
efficient training on vertically partitioned data. In C-VFL, a server and multiple parties …

Towards scalable distributed training of deep learning on public cloud clusters

S Shi, X Zhou, S Song, X Wang, Z Zhu… - Proceedings of …, 2021 - proceedings.mlsys.org
Distributed training techniques have been widely deployed in large-scale deep models
training on dense-GPU clusters. However, on public cloud clusters, due to the moderate …