A review of nonlinear FFT-based computational homogenization methods
M Schneider - Acta Mechanica, 2021 - Springer
Since their inception, computational homogenization methods based on the fast Fourier
transform (FFT) have grown in popularity, establishing themselves as a powerful tool …
transform (FFT) have grown in popularity, establishing themselves as a powerful tool …
Optimization for deep learning: An overview
RY Sun - Journal of the Operations Research Society of China, 2020 - Springer
Optimization is a critical component in deep learning. We think optimization for neural
networks is an interesting topic for theoretical research due to various reasons. First, its …
networks is an interesting topic for theoretical research due to various reasons. First, its …
Derivative-free optimization methods
In many optimization problems arising from scientific, engineering and artificial intelligence
applications, objective and constraint functions are available only as the output of a black …
applications, objective and constraint functions are available only as the output of a black …
Ego-planner: An esdf-free gradient-based local planner for quadrotors
Gradient-based planners are widely used for quadrotor local planning, in which a Euclidean
Signed Distance Field (ESDF) is crucial for evaluating gradient magnitude and direction …
Signed Distance Field (ESDF) is crucial for evaluating gradient magnitude and direction …
Machine learning a general-purpose interatomic potential for silicon
The success of first-principles electronic-structure calculation for predictive modeling in
chemistry, solid-state physics, and materials science is constrained by the limitations on …
chemistry, solid-state physics, and materials science is constrained by the limitations on …
Input convex neural networks
This paper presents the input convex neural network architecture. These are scalar-valued
(potentially deep) neural networks with constraints on the network parameters such that the …
(potentially deep) neural networks with constraints on the network parameters such that the …
Performant implementation of the atomic cluster expansion (PACE) and application to copper and silicon
The atomic cluster expansion is a general polynomial expansion of the atomic energy in
multi-atom basis functions. Here we implement the atomic cluster expansion in the …
multi-atom basis functions. Here we implement the atomic cluster expansion in the …
[图书][B] Tensor analysis: spectral theory and special tensors
L Qi, Z Luo - 2017 - SIAM
Matrix theory is one of the most fundamental tools of mathematics and science, and a
number of classical books on matrix analysis have been written to explore this theory. As a …
number of classical books on matrix analysis have been written to explore this theory. As a …
Adagrad stepsizes: Sharp convergence over nonconvex landscapes
Adaptive gradient methods such as AdaGrad and its variants update the stepsize in
stochastic gradient descent on the fly according to the gradients received along the way; …
stochastic gradient descent on the fly according to the gradients received along the way; …
Optimization for deep learning: theory and algorithms
R Sun - arXiv preprint arXiv:1912.08957, 2019 - arxiv.org
When and why can a neural network be successfully trained? This article provides an
overview of optimization algorithms and theory for training neural networks. First, we discuss …
overview of optimization algorithms and theory for training neural networks. First, we discuss …