A single-timescale method for stochastic bilevel optimization

T Chen, Y Sun, Q Xiao, W Yin - International Conference on …, 2022 - proceedings.mlr.press
Stochastic bilevel optimization generalizes the classic stochastic optimization from the
minimization of a single objective to the minimization of an objective function that depends …

A Single-Timescale Method for Stochastic Bilevel Optimization

T Chen, Y Sun, Q Xiao, W Yin - arXiv preprint arXiv:2102.04671, 2021 - arxiv.org
Stochastic bilevel optimization generalizes the classic stochastic optimization from the
minimization of a single objective to the minimization of an objective function that depends …

Conditional gradient method for stochastic submodular maximization: Closing the gap

A Mokhtari, H Hassani… - … Conference on Artificial …, 2018 - proceedings.mlr.press
In this paper, we study the problem of constrained and stochastic continuous submodular
maximization. Even though the objective function is not concave (nor convex) and is defined …

Adaptive quadratically regularized Newton method for Riemannian optimization

J Hu, A Milzarek, Z Wen, Y Yuan - SIAM Journal on Matrix Analysis and …, 2018 - SIAM
Optimization on Riemannian manifolds widely arises in eigenvalue computation, density
functional theory, Bose--Einstein condensates, low rank nearest correlation, image …

Escaping saddle points in constrained optimization

A Mokhtari, A Ozdaglar… - Advances in Neural …, 2018 - proceedings.neurips.cc
In this paper, we study the problem of escaping from saddle points in smooth nonconvex
optimization problems subject to a convex set $\mathcal {C} $. We propose a generic …

Exploiting negative curvature in deterministic and stochastic optimization

FE Curtis, DP Robinson - Mathematical Programming, 2019 - Springer
This paper addresses the question of whether it can be beneficial for an optimization
algorithm to follow directions of negative curvature. Although prior work has established …

Doubly adaptive scaled algorithm for machine learning using second-order information

M Jahani, S Rusakov, Z Shi, P Richtárik… - arXiv preprint arXiv …, 2021 - arxiv.org
We present a novel adaptive optimization algorithm for large-scale machine learning
problems. Equipped with a low-cost estimate of local curvature and Lipschitz smoothness …

Finding second-order stationary points efficiently in smooth nonconvex linearly constrained optimization problems

S Lu, M Razaviyayn, B Yang… - Advances in Neural …, 2020 - proceedings.neurips.cc
This paper proposes two efficient algorithms for computing approximate second-order
stationary points (SOSPs) of problems with generic smooth non-convex objective functions …

Flecs: A federated learning second-order framework via compression and sketching

A Agafonov, D Kamzolov, R Tappenden… - arXiv preprint arXiv …, 2022 - arxiv.org
Inspired by the recent work FedNL (Safaryan et al, FedNL: Making Newton-Type Methods
Applicable to Federated Learning), we propose a new communication efficient second-order …

A scalable second order method for ill-conditioned matrix completion from few samples

C Kümmerle, CM Verdun - International Conference on …, 2021 - proceedings.mlr.press
We propose an iterative algorithm for low-rank matrix completion with that can be interpreted
as an iteratively reweighted least squares (IRLS) algorithm, a saddle-escaping smoothing …