Grace: A compressed communication framework for distributed machine learning

H Xu, CY Ho, AM Abdelmoniem, A Dutta… - 2021 IEEE 41st …, 2021 - ieeexplore.ieee.org
Powerful computer clusters are used nowadays to train complex deep neural networks
(DNN) on large datasets. Distributed training increasingly becomes communication bound …

An efficient statistical-based gradient compression technique for distributed training systems

AM Abdelmoniem, A Elzanaty… - Proceedings of …, 2021 - proceedings.mlsys.org
The recent many-fold increase in the size of deep neural networks makes efficient
distributed training challenging. Many proposals exploit the compressibility of the gradients …

Empirical analysis of federated learning in heterogeneous environments

AM Abdelmoniem, CY Ho, P Papageorgiou… - Proceedings of the 2nd …, 2022 - dl.acm.org
Federated learning (FL) is becoming a popular paradigm for collaborative learning over
distributed, private datasets owned by non-trusting entities. FL has seen successful …

Learned gradient compression for distributed deep learning

L Abrahamyan, Y Chen, G Bekoulis… - IEEE Transactions on …, 2021 - ieeexplore.ieee.org
Training deep neural networks on large datasets containing high-dimensional data requires
a large amount of computation. A solution to this problem is data-parallel distributed training …

Deepreduce: A sparse-tensor communication framework for federated deep learning

H Xu, K Kostopoulou, A Dutta, X Li… - Advances in …, 2021 - proceedings.neurips.cc
Sparse tensors appear frequently in federated deep learning, either as a direct artifact of the
deep neural network's gradients, or as a result of an explicit sparsification process. Existing …

Towards energy-aware federated learning on battery-powered clients

A Arouj, AM Abdelmoniem - Proceedings of the 1st ACM Workshop on …, 2022 - dl.acm.org
Federated learning (FL) is a newly emerged branch of AI that facilitates edge devices to
collaboratively train a global machine learning model without centralizing data and with …

Personalized federated learning with communication compression

EH Bergou, K Burlachenko, A Dutta… - arXiv preprint arXiv …, 2022 - arxiv.org
In contrast to training traditional machine learning (ML) models in data centers, federated
learning (FL) trains ML models over local datasets contained on resource-constrained …

Joint online optimization of model training and analog aggregation for wireless edge learning

J Wang, B Liang, M Dong, G Boudreau… - IEEE/ACM …, 2023 - ieeexplore.ieee.org
We consider federated learning in a wireless edge network, where multiple power-limited
mobile devices collaboratively train a global model, using their local data with the assistance …

Adaptive synchronous strategy for distributed machine learning

M Tan, WX Liu, J Luo, H Chen… - International Journal of …, 2022 - Wiley Online Library
In distributed machine learning training, bulk synchronous parallel (BSP) and asynchronous
parallel (ASP) are two main synchronization methods to help achieve gradient aggregation …

Leveraging cloud-native microservices architecture for high performance real-time intra-day trading: A tutorial

M Hota, AM Abdelmoniem, M Xu, SS Gill - 6G Enabled Fog Computing in …, 2023 - Springer
Day trading has been gaining attention from prospective investors over the past decades,
even more so in the last decade due to a plethora of factors such as instantaneous …