Deep learning: a statistical viewpoint

PL Bartlett, A Montanari, A Rakhlin - Acta numerica, 2021 - cambridge.org
The remarkable practical success of deep learning has revealed some major surprises from
a theoretical perspective. In particular, simple gradient methods easily find near-optimal …

Understanding deep neural networks with rectified linear units

R Arora, A Basu, P Mianjy, A Mukherjee - arxiv preprint arxiv:1611.01491, 2016 - arxiv.org
In this paper we investigate the family of functions representable by deep neural networks
(DNN) with rectified linear units (ReLU). We give an algorithm to train a ReLU DNN with one …

[책][B] Neural network learning: Theoretical foundations

M Anthony, PL Bartlett - 2009 - dl.acm.org
This important work describes recent theoretical advances in the study of artificial neural
networks. It explores probabilistic models of supervised learning problems, and addresses …

[책][B] Feedforward neural network methodology

TL Fine - 2006 - books.google.com
The decade prior to publication has seen an explosive growth in com-tational speed and
memory and a rapid enrichment in our understa-ing of arti? cial neural networks. These two …

[책][B] Handbook of approximation algorithms and metaheuristics

TF Gonzalez - 2007 - taylorfrancis.com
Delineating the tremendous growth in this area, the Handbook of Approximation Algorithms
and Metaheuristics covers fundamental, theoretical topics as well as advanced, practical …

[책][B] A theory of learning and generalization

M Vidyasagar - 2002 - dl.acm.org
From the Publisher: How does it differ from first edition__ __ Includes new material on:*
support vector machines (SVM's),* fat shattering dimensions* applications to neural network …

Global optimality beyond two layers: Training deep relu networks via convex programs

T Ergen, M Pilanci - International Conference on Machine …, 2021 - proceedings.mlr.press
Understanding the fundamental mechanism behind the success of deep neural networks is
one of the key challenges in the modern machine learning literature. Despite numerous …

[HTML][HTML] On the difficulty of approximately maximizing agreements

S Ben-David, N Eiron, PM Long - Journal of Computer and System …, 2003 - Elsevier
We address the computational complexity of learning in the agnostic framework. For a
variety of common concept classes we prove that, unless P= NP, there is no polynomial time …

Deep neural networks with multi-branch architectures are intrinsically less non-convex

H Zhang, J Shao… - The 22nd International …, 2019 - proceedings.mlr.press
Several recently proposed architectures of neural networks such as ResNeXt, Inception,
Xception, SqueezeNet and Wide ResNet are based on the designing idea of having multiple …

Path regularization: A convexity and sparsity inducing regularization for parallel relu networks

T Ergen, M Pilanci - Advances in Neural Information …, 2024 - proceedings.neurips.cc
Understanding the fundamental principles behind the success of deep neural networks is
one of the most important open questions in the current literature. To this end, we study the …