Model complexity of deep learning: A survey

X Hu, L Chu, J Pei, W Liu, J Bian - Knowledge and Information Systems, 2021 - Springer
Abstract Model complexity is a fundamental problem in deep learning. In this paper, we
conduct a systematic overview of the latest studies on model complexity in deep learning …

Mish: A self regularized non-monotonic activation function

D Misra - arXiv preprint arXiv:1908.08681, 2019 - arxiv.org
We propose $\textit {Mish} $, a novel self-regularized non-monotonic activation function
which can be mathematically defined as: $ f (x)= x\tanh (softplus (x)) $. As activation …

A deep collocation method for the bending analysis of Kirchhoff plate

H Guo, X Zhuang, T Rabczuk - arXiv preprint arXiv:2102.02617, 2021 - arxiv.org
In this paper, a deep collocation method (DCM) for thin plate bending problems is proposed.
This method takes advantage of computational graphs and backpropagation algorithms …

How good is the Bayes posterior in deep neural networks really?

F Wenzel, K Roth, BS Veeling, J Świątkowski… - arXiv preprint arXiv …, 2020 - arxiv.org
During the past five years the Bayesian deep learning community has developed
increasingly accurate and efficient approximate inference procedures that allow for …

A novel PCA–whale optimization-based deep neural network model for classification of tomato plant diseases using GPU

TR Gadekallu, DS Rajput, MPK Reddy… - Journal of Real-Time …, 2021 - Springer
The human population is growing at a very rapid scale. With this progressive growth, it is
extremely important to ensure that healthy food is available for the survival of the inhabitants …

Finite versus infinite neural networks: an empirical study

J Lee, S Schoenholz, J Pennington… - Advances in …, 2020 - proceedings.neurips.cc
We perform a careful, thorough, and large scale empirical study of the correspondence
between wide neural networks and kernel methods. By doing so, we resolve a variety of …

Neural tangents: Fast and easy infinite neural networks in python

R Novak, L Xiao, J Hron, J Lee, AA Alemi… - arXiv preprint arXiv …, 2019 - arxiv.org
Neural Tangents is a library designed to enable research into infinite-width neural networks.
It provides a high-level API for specifying complex and hierarchical neural network …

Dynamical isometry and a mean field theory of cnns: How to train 10,000-layer vanilla convolutional neural networks

L Xiao, Y Bahri, J Sohl-Dickstein… - International …, 2018 - proceedings.mlr.press
In recent years, state-of-the-art methods in computer vision have utilized increasingly deep
convolutional neural network architectures (CNNs), with some of the most successful models …

Tensor programs ii: Neural tangent kernel for any architecture

G Yang - arXiv preprint arXiv:2006.14548, 2020 - arxiv.org
We prove that a randomly initialized neural network of* any architecture* has its Tangent
Kernel (NTK) converge to a deterministic limit, as the network widths tend to infinity. We …

Kanqas: Kolmogorov-arnold network for quantum architecture search

A Kundu, A Sarkar, A Sadhu - EPJ Quantum Technology, 2024 - epjqt.epj.org
Quantum architecture Search (QAS) is a promising direction for optimization and automated
design of quantum circuits towards quantum advantage. Recent techniques in QAS …