The min-max complexity of distributed stochastic convex optimization with intermittent communication

BE Woodworth, B Bullins, O Shamir… - … on Learning Theory, 2021 - proceedings.mlr.press
We resolve the min-max complexity of distributed stochastic convex optimization (up to a log
factor) in the intermittent communication setting, where $ M $ machines work in parallel over …

Towards optimal communication complexity in distributed non-convex optimization

KK Patel, L Wang, BE Woodworth… - Advances in Neural …, 2022 - proceedings.neurips.cc
We study the problem of distributed stochastic non-convex optimization with intermittent
communication. We consider the full participation setting where $ M $ machines work in …

Federated online and bandit convex optimization

KK Patel, L Wang, A Saha… - … Conference on Machine …, 2023 - proceedings.mlr.press
We study the problems of distributed online and bandit convex optimization against an
adaptive adversary. We aim to minimize the average regret on $ M $ machines working in …

The limits and potentials of local sgd for distributed heterogeneous learning with intermittent communication

KK Patel, M Glasgow, A Zindari, L Wang… - arXiv preprint arXiv …, 2024 - arxiv.org
Local SGD is a popular optimization method in distributed learning, often outperforming
other algorithms in practice, including mini-batch SGD. Despite this success, theoretically …

Flecs: A federated learning second-order framework via compression and sketching

A Agafonov, D Kamzolov, R Tappenden… - arXiv preprint arXiv …, 2022 - arxiv.org
Inspired by the recent work FedNL (Safaryan et al, FedNL: Making Newton-Type Methods
Applicable to Federated Learning), we propose a new communication efficient second-order …

Distributed online and bandit convex optimization

KK Patel, A Saha, L Wang, N Srebro - OPT 2022: Optimization for …, 2022 - openreview.net
We study the problems of distributed online and bandit convex optimization against an
adaptive adversary. Our goal is to minimize the average regret on M machines working in …

Exploiting higher-order derivatives in convex optimization methods

D Kamzolov, A Gasnikov, P Dvurechensky… - arXiv preprint arXiv …, 2022 - arxiv.org
Exploiting higher-order derivatives in convex optimization is known at least since 1970's. In
each iteration higher-order (also called tensor) methods minimize a regularized Taylor …

Natural Policy Gradient and Actor Critic Methods for Constrained Multi-Task Reinforcement Learning

S Zeng, TT Doan, J Romberg - arXiv preprint arXiv:2405.02456, 2024 - arxiv.org
Multi-task reinforcement learning (RL) aims to find a single policy that effectively solves
multiple tasks at the same time. This paper presents a constrained formulation for multi-task …

FLECS-CGD: A Federated Learning Second-Order Framework via Compression and Sketching with Compressed Gradient Differences

A Agafonov, B Erraji, M Takáč - arXiv preprint arXiv:2210.09626, 2022 - arxiv.org
In the recent paper FLECS (Agafonov et al, FLECS: A Federated Learning Second-Order
Framework via Compression and Sketching), the second-order framework FLECS was …

[HTML][HTML] Privacy-Preserving Distributed Learning via Newton Algorithm

Z Cao, X Guo, H Zhang - Mathematics, 2023 - mdpi.com
Federated learning (FL) is a prominent distributed learning framework. The main barriers of
FL include communication cost and privacy breaches. In this work, we propose a novel …