Recent advances in stochastic gradient descent in deep learning

Y Tian, Y Zhang, H Zhang - Mathematics, 2023 - mdpi.com
In the age of artificial intelligence, the best approach to handling huge amounts of data is a
tremendously motivating and hard problem. Among machine learning models, stochastic …

随机梯度下降算法研究进展

史加荣, 王丹, 尚凡华, 张鹤于 - 自动化学报, 2021 - aas.net.cn
在机器学习领域中, 梯度下降算法是求解最优化问题最重要, 最基础的方法. 随着数据规模的不断
扩大, 传统的梯度下降算法已不能有效地解决大规模机器学习问题. 随机梯度下降算法在迭代 …

Dash: Semi-supervised learning with dynamic thresholding

Y Xu, L Shang, J Ye, Q Qian, YF Li… - International …, 2021 - proceedings.mlr.press
While semi-supervised learning (SSL) has received tremendous attentions in many machine
learning tasks due to its successful use of unlabeled data, existing SSL algorithms use either …

PAGE: A simple and optimal probabilistic gradient estimator for nonconvex optimization

Z Li, H Bao, X Zhang… - … conference on machine …, 2021 - proceedings.mlr.press
In this paper, we propose a novel stochastic gradient estimator—ProbAbilistic Gradient
Estimator (PAGE)—for nonconvex optimization. PAGE is easy to implement as it is designed …

Stochastic recursive gradient descent ascent for stochastic nonconvex-strongly-concave minimax problems

L Luo, H Ye, Z Huang, T Zhang - Advances in Neural …, 2020 - proceedings.neurips.cc
We consider nonconvex-concave minimax optimization problems of the form $\min_ {\bf
x}\max_ {\bf y\in {\mathcal Y}} f ({\bf x},{\bf y}) $, where $ f $ is strongly-concave in $\bf y $ but …

Sample efficient policy gradient methods with recursive variance reduction

P Xu, F Gao, Q Gu - arXiv preprint arXiv:1909.08610, 2019 - arxiv.org
Improving the sample efficiency in reinforcement learning has been a long-standing
research problem. In this work, we aim to reduce the sample complexity of existing policy …

An improved convergence analysis of stochastic variance-reduced policy gradient

P Xu, F Gao, Q Gu - Uncertainty in Artificial Intelligence, 2020 - proceedings.mlr.press
We revisit the stochastic variance-reduced policy gradient (SVRPG) method proposed
by\citet {papini2018stochastic} for reinforcement learning. We provide an improved …

Stochastic auc maximization with deep neural networks

M Liu, Z Yuan, Y Ying, T Yang - arXiv preprint arXiv:1908.10831, 2019 - arxiv.org
Stochastic AUC maximization has garnered an increasing interest due to better fit to
imbalanced data classification. However, existing works are limited to stochastic AUC …

EF21 with bells & whistles: Practical algorithmic extensions of modern error feedback

I Fatkhullin, I Sokolov, E Gorbunov, Z Li… - arXiv preprint arXiv …, 2021 - arxiv.org
First proposed by Seide (2014) as a heuristic, error feedback (EF) is a very popular
mechanism for enforcing convergence of distributed gradient-based optimization methods …

Improved fine-tuning by better leveraging pre-training data

Z Liu, Y Xu, Y Xu, Q Qian, H Li, X Ji… - Advances in Neural …, 2022 - proceedings.neurips.cc
As a dominant paradigm, fine-tuning a pre-trained model on the target data is widely used in
many deep learning applications, especially for small data sets. However, recent studies …