Optimization for deep learning: An overview
RY Sun - Journal of the Operations Research Society of China, 2020 - Springer
Optimization is a critical component in deep learning. We think optimization for neural
networks is an interesting topic for theoretical research due to various reasons. First, its …
networks is an interesting topic for theoretical research due to various reasons. First, its …
The global landscape of neural networks: An overview
One of the major concerns for neural network training is that the nonconvexity of the
associated loss functions may cause a bad landscape. The recent success of neural …
associated loss functions may cause a bad landscape. The recent success of neural …
Optimization for deep learning: theory and algorithms
R Sun - arXiv preprint arXiv:1912.08957, 2019 - arxiv.org
When and why can a neural network be successfully trained? This article provides an
overview of optimization algorithms and theory for training neural networks. First, we discuss …
overview of optimization algorithms and theory for training neural networks. First, we discuss …
Deep learning-based channel estimation for doubly selective fading channels
In this paper, online deep learning (DL)-based channel estimation algorithm for doubly
selective fading channels is proposed by employing the deep neural network (DNN). With …
selective fading channels is proposed by employing the deep neural network (DNN). With …
What Happens after SGD Reaches Zero Loss?--A Mathematical Framework
Understanding the implicit bias of Stochastic Gradient Descent (SGD) is one of the key
challenges in deep learning, especially for overparametrized models, where the local …
challenges in deep learning, especially for overparametrized models, where the local …
Deep transfer learning-based downlink channel prediction for FDD massive MIMO systems
Artificial intelligence (AI) based downlink channel state information (CSI) prediction for
frequency division duplexing (FDD) massive multiple-input multiple-output (MIMO) systems …
frequency division duplexing (FDD) massive multiple-input multiple-output (MIMO) systems …
Going beyond linear mode connectivity: The layerwise linear feature connectivity
Recent work has revealed many intriguing empirical phenomena in neural network training,
despite the poorly understood and highly complex loss landscapes and training dynamics …
despite the poorly understood and highly complex loss landscapes and training dynamics …
Recent theoretical advances in non-convex optimization
Motivated by recent increased interest in optimization algorithms for non-convex
optimization in application to training deep neural networks and other optimization problems …
optimization in application to training deep neural networks and other optimization problems …
On connected sublevel sets in deep learning
Q Nguyen - International conference on machine learning, 2019 - proceedings.mlr.press
This paper shows that every sublevel set of the loss function of a class of deep over-
parameterized neural nets with piecewise linear activation functions is connected and …
parameterized neural nets with piecewise linear activation functions is connected and …
Explaining landscape connectivity of low-cost solutions for multilayer nets
Mode connectivity is a surprising phenomenon in the loss landscape of deep nets. Optima---
at least those discovered by gradient-based optimization---turn out to be connected by …
at least those discovered by gradient-based optimization---turn out to be connected by …