Deep learning: a statistical viewpoint
The remarkable practical success of deep learning has revealed some major surprises from
a theoretical perspective. In particular, simple gradient methods easily find near-optimal …
a theoretical perspective. In particular, simple gradient methods easily find near-optimal …
Understanding deep neural networks with rectified linear units
In this paper we investigate the family of functions representable by deep neural networks
(DNN) with rectified linear units (ReLU). We give an algorithm to train a ReLU DNN with one …
(DNN) with rectified linear units (ReLU). We give an algorithm to train a ReLU DNN with one …
[책][B] Neural network learning: Theoretical foundations
M Anthony, PL Bartlett - 2009 - dl.acm.org
This important work describes recent theoretical advances in the study of artificial neural
networks. It explores probabilistic models of supervised learning problems, and addresses …
networks. It explores probabilistic models of supervised learning problems, and addresses …
[책][B] Handbook of approximation algorithms and metaheuristics
TF Gonzalez - 2007 - taylorfrancis.com
Delineating the tremendous growth in this area, the Handbook of Approximation Algorithms
and Metaheuristics covers fundamental, theoretical topics as well as advanced, practical …
and Metaheuristics covers fundamental, theoretical topics as well as advanced, practical …
[책][B] A theory of learning and generalization
M Vidyasagar - 2002 - dl.acm.org
From the Publisher: How does it differ from first edition__ __ Includes new material on:*
support vector machines (SVM's),* fat shattering dimensions* applications to neural network …
support vector machines (SVM's),* fat shattering dimensions* applications to neural network …
Global optimality beyond two layers: Training deep relu networks via convex programs
Understanding the fundamental mechanism behind the success of deep neural networks is
one of the key challenges in the modern machine learning literature. Despite numerous …
one of the key challenges in the modern machine learning literature. Despite numerous …
[HTML][HTML] On the difficulty of approximately maximizing agreements
S Ben-David, N Eiron, PM Long - Journal of Computer and System …, 2003 - Elsevier
We address the computational complexity of learning in the agnostic framework. For a
variety of common concept classes we prove that, unless P= NP, there is no polynomial time …
variety of common concept classes we prove that, unless P= NP, there is no polynomial time …
Deep neural networks with multi-branch architectures are intrinsically less non-convex
Several recently proposed architectures of neural networks such as ResNeXt, Inception,
Xception, SqueezeNet and Wide ResNet are based on the designing idea of having multiple …
Xception, SqueezeNet and Wide ResNet are based on the designing idea of having multiple …
Path regularization: A convexity and sparsity inducing regularization for parallel relu networks
Understanding the fundamental principles behind the success of deep neural networks is
one of the most important open questions in the current literature. To this end, we study the …
one of the most important open questions in the current literature. To this end, we study the …