Online nonconvex optimization with limited instantaneous oracle feedback

Z Guan, Y Zhou, Y Liang - The Thirty Sixth Annual …, 2023 - proceedings.mlr.press
We investigate online nonconvex optimization from a local regret minimization perspective.
Previous studies along this line implicitly required the access to sufficient gradient oracles at …

Regret minimization in stochastic non-convex learning via a proximal-gradient approach

N Hallak, P Mertikopoulos… - … Conference on Machine …, 2021 - proceedings.mlr.press
This paper develops a methodology for regret minimization with stochastic first-order oracle
feedback in online, constrained, non-smooth, non-convex problems. In this setting, the …

Online bilevel optimization: Regret analysis of online alternating gradient methods

DA Tarzanagh, P Nazari, B Hou… - International …, 2024 - proceedings.mlr.press
This paper introduces\textit {online bilevel optimization} in which a sequence of time-varying
bilevel problems is revealed one after the other. We extend the known regret bounds for …

Adaptive first-order methods revisited: Convex minimization without lipschitz requirements

K Antonakopoulos… - Advances in Neural …, 2021 - proceedings.neurips.cc
We propose a new family of adaptive first-order methods for a class of convex minimization
problems that may fail to be Lipschitz continuous or smooth in the standard sense …

Non-convex bilevel optimization with time-varying objective functions

S Lin, D Sow, K Ji, Y Liang… - Advances in Neural …, 2024 - proceedings.neurips.cc
Bilevel optimization has become a powerful tool in a wide variety of machine learning
problems. However, the current nonconvex bilevel optimization considers an offline dataset …

On the Hardness of Online Nonconvex Optimization with Single Oracle Feedback

Z Guan, Y Zhou, Y Liang - The Twelfth International Conference on …, 2024 - openreview.net
Online nonconvex optimization has been an active area of research recently. Previous
studies either considered the global regret with full information about the objective functions …

Nested bandits

M Martin, P Mertikopoulos, T Rahier… - … on Machine Learning, 2022 - proceedings.mlr.press
In many online decision processes, the optimizing agent is called to choose between large
numbers of alternatives with many inherent similarities; in turn, these similarities imply …

Distributed stochastic nash equilibrium learning in locally coupled network games with unknown parameters

Y Huang, J Hu - Learning for Dynamics and Control …, 2022 - proceedings.mlr.press
In stochastic Nash equilibrium problems (SNEPs), it is natural for players to be uncertain
about their complex environments and have multi-dimensional unknown parameters in their …

Gradient and projection free distributed online min-max resource optimization

J Wang, B Liang - Learning for Dynamics and Control …, 2022 - proceedings.mlr.press
We consider distributed online min-max resource allocation with a set of parallel agents and
a parameter server. Our goal is to minimize the pointwise maximum over a set of time …

Unlocking TriLevel Learning with Level-Wise Zeroth Order Constraints: Distributed Algorithms and Provable Non-Asymptotic Convergence

Y Jiao, K Yang, C Jian - arXiv preprint arXiv:2412.07138, 2024 - arxiv.org
Trilevel learning (TLL) found diverse applications in numerous machine learning
applications, ranging from robust hyperparameter optimization to domain adaptation …