Variance reduced proxskip: Algorithm, theory and application to federated learning

G Malinovsky, K Yi, P Richtárik - Advances in Neural …, 2022 - proceedings.neurips.cc
We study distributed optimization methods based on the {\em local training (LT)} paradigm,
ie, methods which achieve communication efficiency by performing richer local gradient …

Towards out-of-distribution generalization: A survey

J Liu, Z Shen, Y He, X Zhang, R Xu, H Yu… - arXiv preprint arXiv …, 2021 - arxiv.org
Traditional machine learning paradigms are based on the assumption that both training and
test data follow the same statistical pattern, which is mathematically referred to as …

Proxskip: Yes! local gradient steps provably lead to communication acceleration! finally!

K Mishchenko, G Malinovsky, S Stich… - International …, 2022 - proceedings.mlr.press
We introduce ProxSkip—a surprisingly simple and provably efficient method for minimizing
the sum of a smooth ($ f $) and an expensive nonsmooth proximable ($\psi $) function. The …

Linear convergence in federated learning: Tackling client heterogeneity and sparse gradients

A Mitra, R Jaafar, GJ Pappas… - Advances in Neural …, 2021 - proceedings.neurips.cc
We consider a standard federated learning (FL) setup where a group of clients periodically
coordinate with a central server to train a statistical model. We develop a general algorithmic …

Communication-efficient and distributed learning over wireless networks: Principles and applications

J Park, S Samarakoon, A Elgabli, J Kim… - Proceedings of the …, 2021 - ieeexplore.ieee.org
Machine learning (ML) is a promising enabler for the fifth-generation (5G) communication
systems and beyond. By imbuing intelligence into the network edge, edge nodes can …

Fedavg with fine tuning: Local updates lead to representation learning

L Collins, H Hassani, A Mokhtari… - Advances in Neural …, 2022 - proceedings.neurips.cc
Abstract The Federated Averaging (FedAvg) algorithm, which consists of alternating
between a few local stochastic gradient updates at client nodes, followed by a model …

A unified analysis of federated learning with arbitrary client participation

S Wang, M Ji - Advances in Neural Information Processing …, 2022 - proceedings.neurips.cc
Federated learning (FL) faces challenges of intermittent client availability and computation/
communication efficiency. As a result, only a small subset of clients can participate in FL at a …

Fedmsplit: Correlation-adaptive federated multi-task learning across multimodal split networks

J Chen, A Zhang - Proceedings of the 28th ACM SIGKDD conference on …, 2022 - dl.acm.org
With the advancement of data collection techniques, end users are interested in how
different types of data can collaborate to improve our life experiences. Multimodal Federated …

FedNL: Making Newton-type methods applicable to federated learning

M Safaryan, R Islamov, X Qian, P Richtárik - arXiv preprint arXiv …, 2021 - arxiv.org
Inspired by recent work of Islamov et al (2021), we propose a family of Federated Newton
Learn (FedNL) methods, which we believe is a marked step in the direction of making …

SoteriaFL: A unified framework for private federated learning with communication compression

Z Li, H Zhao, B Li, Y Chi - Advances in Neural Information …, 2022 - proceedings.neurips.cc
To enable large-scale machine learning in bandwidth-hungry environments such as
wireless networks, significant progress has been made recently in designing communication …