A single-timescale method for stochastic bilevel optimization
Stochastic bilevel optimization generalizes the classic stochastic optimization from the
minimization of a single objective to the minimization of an objective function that depends …
minimization of a single objective to the minimization of an objective function that depends …
A Single-Timescale Method for Stochastic Bilevel Optimization
Stochastic bilevel optimization generalizes the classic stochastic optimization from the
minimization of a single objective to the minimization of an objective function that depends …
minimization of a single objective to the minimization of an objective function that depends …
Conditional gradient method for stochastic submodular maximization: Closing the gap
A Mokhtari, H Hassani… - … Conference on Artificial …, 2018 - proceedings.mlr.press
In this paper, we study the problem of constrained and stochastic continuous submodular
maximization. Even though the objective function is not concave (nor convex) and is defined …
maximization. Even though the objective function is not concave (nor convex) and is defined …
Adaptive quadratically regularized Newton method for Riemannian optimization
Optimization on Riemannian manifolds widely arises in eigenvalue computation, density
functional theory, Bose--Einstein condensates, low rank nearest correlation, image …
functional theory, Bose--Einstein condensates, low rank nearest correlation, image …
Escaping saddle points in constrained optimization
A Mokhtari, A Ozdaglar… - Advances in Neural …, 2018 - proceedings.neurips.cc
In this paper, we study the problem of escaping from saddle points in smooth nonconvex
optimization problems subject to a convex set $\mathcal {C} $. We propose a generic …
optimization problems subject to a convex set $\mathcal {C} $. We propose a generic …
Exploiting negative curvature in deterministic and stochastic optimization
FE Curtis, DP Robinson - Mathematical Programming, 2019 - Springer
This paper addresses the question of whether it can be beneficial for an optimization
algorithm to follow directions of negative curvature. Although prior work has established …
algorithm to follow directions of negative curvature. Although prior work has established …
Doubly adaptive scaled algorithm for machine learning using second-order information
We present a novel adaptive optimization algorithm for large-scale machine learning
problems. Equipped with a low-cost estimate of local curvature and Lipschitz smoothness …
problems. Equipped with a low-cost estimate of local curvature and Lipschitz smoothness …
Finding second-order stationary points efficiently in smooth nonconvex linearly constrained optimization problems
This paper proposes two efficient algorithms for computing approximate second-order
stationary points (SOSPs) of problems with generic smooth non-convex objective functions …
stationary points (SOSPs) of problems with generic smooth non-convex objective functions …
Flecs: A federated learning second-order framework via compression and sketching
Inspired by the recent work FedNL (Safaryan et al, FedNL: Making Newton-Type Methods
Applicable to Federated Learning), we propose a new communication efficient second-order …
Applicable to Federated Learning), we propose a new communication efficient second-order …
A scalable second order method for ill-conditioned matrix completion from few samples
C Kümmerle, CM Verdun - International Conference on …, 2021 - proceedings.mlr.press
We propose an iterative algorithm for low-rank matrix completion with that can be interpreted
as an iteratively reweighted least squares (IRLS) algorithm, a saddle-escaping smoothing …
as an iteratively reweighted least squares (IRLS) algorithm, a saddle-escaping smoothing …