Solving the Kolmogorov PDE by means of deep learning
Stochastic differential equations (SDEs) and the Kolmogorov partial differential equations
(PDEs) associated to them have been widely used in models from engineering, finance, and …
(PDEs) associated to them have been widely used in models from engineering, finance, and …
On the bias-variance-cost tradeoff of stochastic optimization
We consider stochastic optimization when one only has access to biased stochastic oracles
of the objective, and obtaining stochastic gradients with low biases comes at high costs. This …
of the objective, and obtaining stochastic gradients with low biases comes at high costs. This …
Full error analysis for the training of deep neural networks
Deep learning algorithms have been applied very successfully in recent years to a range of
problems out of reach for classical solution paradigms. Nevertheless, there is no completely …
problems out of reach for classical solution paradigms. Nevertheless, there is no completely …
Convergence of stochastic gradient descent schemes for Lojasiewicz-landscapes
In this article, we consider convergence of stochastic gradient descent schemes (SGD)
under weak assumptions on the underlying landscape. More explicitly, we show that on the …
under weak assumptions on the underlying landscape. More explicitly, we show that on the …
A proof of convergence for the gradient descent optimization method with random initializations in the training of neural networks with ReLU activation for piecewise …
Gradient descent (GD) type optimization methods are the standard instrument to train
artificial neural networks (ANNs) with recti_ed linear unit (ReLU) activation. Despite the …
artificial neural networks (ANNs) with recti_ed linear unit (ReLU) activation. Despite the …
Strong error analysis for stochastic gradient descent optimization algorithms
Stochastic gradient descent (SGD) optimization algorithms are key ingredients in a series of
machine learning applications. In this article we perform a rigorous strong error analysis for …
machine learning applications. In this article we perform a rigorous strong error analysis for …
Blow up phenomena for gradient descent optimization methods in the training of artificial neural networks
In this article we investigate blow up phenomena for gradient descent optimization methods
in the training of artificial neural networks (ANNs). Our theoretical analysis is focused on …
in the training of artificial neural networks (ANNs). Our theoretical analysis is focused on …
A proof of convergence for stochastic gradient descent in the training of artificial neural networks with ReLU activation for constant target functions
In this article we study the stochastic gradient descent (SGD) optimization method in the
training of fully connected feedforward artificial neural networks with ReLU activation. The …
training of fully connected feedforward artificial neural networks with ReLU activation. The …
Multi-level monte-carlo gradient methods for stochastic optimization with biased oracles
We consider stochastic optimization when one only has access to biased stochastic oracles
of the objective and the gradient, and obtaining stochastic gradients with low biases comes …
of the objective and the gradient, and obtaining stochastic gradients with low biases comes …
Convergence proof for stochastic gradient descent in the training of deep neural networks with ReLU activation for constant target functions
In many numerical simulations stochastic gradient descent (SGD) type optimization methods
perform very effectively in the training of deep neural networks (DNNs) but till this day it …
perform very effectively in the training of deep neural networks (DNNs) but till this day it …