A comprehensive survey on training acceleration for large machine learning models in IoT

H Wang, Z Qu, Q Zhou, H Zhang, B Luo… - IEEE Internet of …, 2021 - ieeexplore.ieee.org
The ever-growing artificial intelligence (AI) applications have greatly reshaped our world in
many areas, eg, smart home, computer vision, natural language processing, etc. Behind …

Communication compression techniques in distributed deep learning: A survey

Z Wang, M Wen, Y Xu, Y Zhou, JH Wang… - Journal of Systems …, 2023 - Elsevier
Nowadays, the training data and neural network models are getting increasingly large. The
training time of deep learning will become unbearably long on a single machine. To reduce …

Federated learning over wireless device-to-device networks: Algorithms and convergence analysis

H Xing, O Simeone, S Bi - IEEE Journal on Selected Areas in …, 2021 - ieeexplore.ieee.org
The proliferation of Internet-of-Things (IoT) devices and cloud-computing applications over
siloed data centers is motivating renewed interest in the collaborative training of a shared …

Quasi-global momentum: Accelerating decentralized deep learning on heterogeneous data

T Lin, SP Karimireddy, SU Stich, M Jaggi - arXiv preprint arXiv:2102.04761, 2021 - arxiv.org
Decentralized training of deep learning models is a key element for enabling data privacy
and on-device learning over networks. In realistic learning scenarios, the presence of …

1% vs 100%: Parameter-efficient low rank adapter for dense predictions

D Yin, Y Yang, Z Wang, H Yu… - Proceedings of the …, 2023 - openaccess.thecvf.com
Fine-tuning large-scale pre-trained vision models to downstream tasks is a standard
technique for achieving state-of-the-art performance on computer vision benchmarks …

Cross-gradient aggregation for decentralized learning from non-iid data

Y Esfandiari, SY Tan, Z Jiang, A Balu… - International …, 2021 - proceedings.mlr.press
Decentralized learning enables a group of collaborative agents to learn models using a
distributed dataset without the need for a central parameter server. Recently, decentralized …

Datalens: Scalable privacy preserving training via gradient compression and aggregation

B Wang, F Wu, Y Long, L Rimanic, C Zhang… - Proceedings of the 2021 …, 2021 - dl.acm.org
Recent success of deep neural networks (DNNs) hinges on the availability of large-scale
dataset; however, training on such dataset often poses privacy risks for sensitive training …

Rank diminishing in deep neural networks

R Feng, K Zheng, Y Huang, D Zhao… - Advances in Neural …, 2022 - proceedings.neurips.cc
The rank of neural networks measures information flowing across layers. It is an instance of
a key structural condition that applies across broad domains of machine learning. In …

Layer-wise adaptive model aggregation for scalable federated learning

S Lee, T Zhang, AS Avestimehr - … of the AAAI Conference on Artificial …, 2023 - ojs.aaai.org
Abstract In Federated Learning (FL), a common approach for aggregating local solutions
across clients is periodic full model averaging. It is, however, known that different layers of …

Secure decentralized image classification with multiparty homomorphic encryption

G Xu, G Li, S Guo, T Zhang, H Li - IEEE Transactions on …, 2023 - ieeexplore.ieee.org
Decentralized image classification plays a key role in various scenarios due to its attractive
properties, including tolerating high network latency and less prone to single-point failures …