A survey on statistical theory of deep learning: Approximation, training dynamics, and generative models

N Suh, G Cheng - Annual Review of Statistics and Its Application, 2024 - annualreviews.org
In this article, we review the literature on statistical theories of neural networks from three
perspectives: approximation, training dynamics, and generative models. In the first part …

Classification with deep neural networks and logistic loss

Z Zhang, L Shi, DX Zhou - Journal of Machine Learning Research, 2024 - jmlr.org
Deep neural networks (DNNs) trained with the logistic loss (also known as the cross entropy
loss) have made impressive advancements in various binary classification tasks. Despite the …

Generalization analysis of deep CNNs under maximum correntropy criterion

Y Zhang, Z Fang, J Fan - Neural Networks, 2024 - Elsevier
Convolutional neural networks (CNNs) have gained immense popularity in recent years,
finding their utility in diverse fields such as image recognition, natural language processing …

Optimal Rates of Approximation by Shallow ReLU Neural Networks and Applications to Nonparametric Regression

Y Yang, DX Zhou - Constructive Approximation, 2024 - Springer
We study the approximation capacity of some variation spaces corresponding to shallow
ReLU k neural networks. It is shown that sufficiently smooth functions are contained in these …

Deeper or wider: A perspective from optimal generalization error with sobolev loss

Y Yang, J He - arxiv preprint arxiv:2402.00152, 2024 - arxiv.org
Constructing the architecture of a neural network is a challenging pursuit for the machine
learning community, and the dilemma of whether to go deeper or wider remains a persistent …

Approximation with cnns in sobolev space: with applications to classification

G Shen, Y Jiao, Y Lin, J Huang - Advances in neural …, 2022 - proceedings.neurips.cc
We derive a novel approximation error bound with explicit prefactor for Sobolev-regular
functions using deep convolutional neural networks (CNNs). The bound is non-asymptotic in …

Approximation of nonlinear functionals using deep ReLU networks

L Song, J Fan, DR Chen, DX Zhou - Journal of Fourier Analysis and …, 2023 - Springer
In recent years, functional neural networks have been proposed and studied in order to
approximate nonlinear continuous functionals defined on L p ([-1, 1] s) for integers s≥ 1 and …

CNN models for readability of Chinese texts.

H Feng, S Hou, LY Wei… - … Foundations of Computing, 2022 - search.ebscohost.com
Readability of Chinese texts considered in this paper is a multi-class classification problem
with 12 grade classes corresponding to 6 grades in primary schools, 3 grades in middle …

On the rates of convergence for learning with convolutional neural networks

Y Yang, H Feng, DX Zhou - arxiv preprint arxiv:2403.16459, 2024 - arxiv.org
We study approximation and learning capacities of convolutional neural networks (CNNs)
with one-side zero-padding and multiple channels. Our first result proves a new …

Learning rates of deep nets for geometrically strongly mixing sequence

Y Men, L Li, Z Hu, Y Xu - IEEE Transactions on Neural …, 2024 - ieeexplore.ieee.org
The great success of deep learning poses an urgent challenge to establish the theoretical
basis for its working mechanism. Recently, research on the convergence of deep neural …