Grace: A compressed communication framework for distributed machine learning
Powerful computer clusters are used nowadays to train complex deep neural networks
(DNN) on large datasets. Distributed training increasingly becomes communication bound …
(DNN) on large datasets. Distributed training increasingly becomes communication bound …
An efficient statistical-based gradient compression technique for distributed training systems
AM Abdelmoniem, A Elzanaty… - Proceedings of …, 2021 - proceedings.mlsys.org
The recent many-fold increase in the size of deep neural networks makes efficient
distributed training challenging. Many proposals exploit the compressibility of the gradients …
distributed training challenging. Many proposals exploit the compressibility of the gradients …
Empirical analysis of federated learning in heterogeneous environments
Federated learning (FL) is becoming a popular paradigm for collaborative learning over
distributed, private datasets owned by non-trusting entities. FL has seen successful …
distributed, private datasets owned by non-trusting entities. FL has seen successful …
Learned gradient compression for distributed deep learning
L Abrahamyan, Y Chen, G Bekoulis… - IEEE Transactions on …, 2021 - ieeexplore.ieee.org
Training deep neural networks on large datasets containing high-dimensional data requires
a large amount of computation. A solution to this problem is data-parallel distributed training …
a large amount of computation. A solution to this problem is data-parallel distributed training …
Deepreduce: A sparse-tensor communication framework for federated deep learning
Sparse tensors appear frequently in federated deep learning, either as a direct artifact of the
deep neural network's gradients, or as a result of an explicit sparsification process. Existing …
deep neural network's gradients, or as a result of an explicit sparsification process. Existing …
Towards energy-aware federated learning on battery-powered clients
A Arouj, AM Abdelmoniem - Proceedings of the 1st ACM Workshop on …, 2022 - dl.acm.org
Federated learning (FL) is a newly emerged branch of AI that facilitates edge devices to
collaboratively train a global machine learning model without centralizing data and with …
collaboratively train a global machine learning model without centralizing data and with …
Personalized federated learning with communication compression
In contrast to training traditional machine learning (ML) models in data centers, federated
learning (FL) trains ML models over local datasets contained on resource-constrained …
learning (FL) trains ML models over local datasets contained on resource-constrained …
Joint online optimization of model training and analog aggregation for wireless edge learning
We consider federated learning in a wireless edge network, where multiple power-limited
mobile devices collaboratively train a global model, using their local data with the assistance …
mobile devices collaboratively train a global model, using their local data with the assistance …
Adaptive synchronous strategy for distributed machine learning
M Tan, WX Liu, J Luo, H Chen… - International Journal of …, 2022 - Wiley Online Library
In distributed machine learning training, bulk synchronous parallel (BSP) and asynchronous
parallel (ASP) are two main synchronization methods to help achieve gradient aggregation …
parallel (ASP) are two main synchronization methods to help achieve gradient aggregation …
Leveraging cloud-native microservices architecture for high performance real-time intra-day trading: A tutorial
Day trading has been gaining attention from prospective investors over the past decades,
even more so in the last decade due to a plethora of factors such as instantaneous …
even more so in the last decade due to a plethora of factors such as instantaneous …