Randomized numerical linear algebra: Foundations and algorithms

PG Martinsson, JA Tropp - Acta Numerica, 2020 - cambridge.org
This survey describes probabilistic algorithms for linear algebraic computations, such as
factorizing matrices and solving linear systems. It focuses on techniques that have a proven …

Asynchronous parallel stochastic gradient for nonconvex optimization

X Lian, Y Huang, Y Li, J Liu - Advances in neural …, 2015 - proceedings.neurips.cc
The asynchronous parallel implementations of stochastic gradient (SG) have been broadly
used in solving deep neural network and received many successes in practice recently …

Asynchronous stochastic gradient descent with delay compensation

S Zheng, Q Meng, T Wang, W Chen… - International …, 2017 - proceedings.mlr.press
With the fast development of deep learning, it has become common to learn big neural
networks using massive training data. Asynchronous Stochastic Gradient Descent (ASGD) is …

An asynchronous parallel stochastic coordinate descent algorithm

J Liu, S Wright, C Ré, V Bittorf… - … on Machine Learning, 2014 - proceedings.mlr.press
We describe an asynchronous parallel stochastic coordinate descent algorithm for
minimizing smooth unconstrained or separably constrained functions. The method achieves …

Arock: an algorithmic framework for asynchronous parallel coordinate updates

Z Peng, Y Xu, M Yan, W Yin - SIAM Journal on Scientific Computing, 2016 - SIAM
Finding a fixed point to a nonexpansive operator, ie, x^*=Tx^*, abstracts many problems in
numerical linear algebra, optimization, and other areas of data science. To solve fixed-point …

Perturbed iterate analysis for asynchronous stochastic optimization

H Mania, X Pan, D Papailiopoulos, B Recht… - SIAM Journal on …, 2017 - SIAM
We introduce and analyze stochastic optimization methods where the input to each update
is perturbed by bounded noise. We show that this framework forms the basis of a unified …

Asynchronous stochastic coordinate descent: Parallelism and convergence properties

J Liu, SJ Wright - SIAM Journal on Optimization, 2015 - SIAM
We describe an asynchronous parallel stochastic proximal coordinate descent algorithm for
minimizing a composite objective function, which consists of a smooth convex function …

From server-based to client-based machine learning: A comprehensive survey

R Gu, C Niu, F Wu, G Chen, C Hu, C Lyu… - ACM Computing Surveys …, 2021 - dl.acm.org
In recent years, mobile devices have gained increasing development with stronger
computation capability and larger storage space. Some of the computation-intensive …

A comprehensive linear speedup analysis for asynchronous stochastic parallel optimization from zeroth-order to first-order

X Lian, H Zhang, CJ Hsieh… - Advances in Neural …, 2016 - proceedings.neurips.cc
Asynchronous parallel optimization received substantial successes and extensive attention
recently. One of core theoretical questions is how much speedup (or benefit) the …

Stochastic reformulations of linear systems: algorithms and convergence theory

P Richtárik, M Takác - SIAM Journal on Matrix Analysis and Applications, 2020 - SIAM
We develop a family of reformulations of an arbitrary consistent linear system into a
stochastic problem. The reformulations are governed by two user-defined parameters: a …