Speeding up distributed gradient descent by utilizing non-persistent stragglers

E Ozfatura, D Gündüz, S Ulukus - 2019 IEEE International …, 2019 - ieeexplore.ieee.org
When gradient descent (GD) is scaled to many parallel computing servers (workers) for
large scale machine learning problems, its per-iteration computation time is limited by the …

Etalumis: Bringing probabilistic programming to scientific simulators at scale

AG Baydin, L Shao, W Bhimji, L Heinrich… - Proceedings of the …, 2019 - dl.acm.org
Probabilistic programming languages (PPLs) are receiving widespread attention for
performing Bayesian inference in complex generative models. However, applications to …

[HTML][HTML] Privacy-preserved learning from non-iid data in fog-assisted IoT: A federated learning approach

M Abdel-Basset, H Hawash, N Moustafa… - Digital Communications …, 2022 - Elsevier
With the prevalence of the Internet of Things (IoT) systems, smart cities comprise complex
networks, including sensors, actuators, appliances, and cyber services. The complexity and …

Backdoor attacks in peer-to-peer federated learning

G Syros, G Yar, S Boboila, C Nita-Rotaru… - ACM Transactions on …, 2023 - dl.acm.org
Most machine learning applications rely on centralized learning processes, opening up the
risk of exposure of their training datasets. While federated learning (FL) mitigates to some …

Neighborhood-correction algorithm for classification of normal and malignant cells

Y Pan, M Liu, Y Xia, D Shen - ISBI 2019 C-NMC Challenge: Classification …, 2019 - Springer
Classification of normal and malignant cells observed under a microscope is an essential
and challenging step in the development of a cost-effective computer-aided diagnosis tool …

Totoro: A Scalable Federated Learning Engine for the Edge

CW Ching, X Chen, T Kim, B Ji, Q Wang… - Proceedings of the …, 2024 - dl.acm.org
Federated Learning (FL) is an emerging distributed machine learning (ML) technique that
enables in-situ model training and inference on decentralized edge devices. We propose …

Straggler-resilient distributed machine learning with dynamic backup workers

G Xiong, G Yan, R Singh, J Li - arXiv preprint arXiv:2102.06280, 2021 - arxiv.org
With the increasing demand for large-scale training of machine learning models, consensus-
based distributed optimization methods have recently been advocated as alternatives to the …

Straggler-resilient decentralized learning via adaptive asynchronous updates

G Xiong, G Yan, S Wang, J Li - … and Protocol Design for Mobile Networks …, 2024 - dl.acm.org
With the increasing demand for large-scale training of machine learning models, fully
decentralized optimization methods have recently been advocated as alternatives to the …

Distributed dual coordinate ascent with imbalanced data on a general tree network

M Cho, L Lai, W Xu - … Workshop on Machine Learning for Signal …, 2023 - ieeexplore.ieee.org
In this paper, we investigate the impact of imbalanced data on the convergence of
distributed dual coordinate ascent in a tree network for solving an empirical loss …

Weighted aggregating stochastic gradient descent for parallel deep learning

P Guo, Z Ye, K Xiao, W Zhu - IEEE Transactions on Knowledge …, 2020 - ieeexplore.ieee.org
This paper investigates the stochastic optimization problem focusing on developing scalable
parallel algorithms for deep learning tasks. Our solution involves a reformation of the …