Learning single-index models with shallow neural networks

A Bietti, J Bruna, C Sanford… - Advances in Neural …, 2022 - proceedings.neurips.cc
Single-index models are a class of functions given by an unknown univariate``link''function
applied to an unknown one-dimensional projection of the input. These models are …

Statistically meaningful approximation: a case study on approximating turing machines with transformers

C Wei, Y Chen, T Ma - Advances in Neural Information …, 2022 - proceedings.neurips.cc
A common lens to theoretically study neural net architectures is to analyze the functions they
can approximate. However, the constructions from approximation theory often have …

Provable guarantees for nonlinear feature learning in three-layer neural networks

E Nichani, A Damian, JD Lee - Advances in Neural …, 2024 - proceedings.neurips.cc
One of the central questions in the theory of deep learning is to understand how neural
networks learn hierarchical features. The ability of deep networks to extract salient features …

Optimizing solution-samplers for combinatorial problems: The landscape of policy-gradient method

C Caramanis, D Fotakis, A Kalavasis… - Advances in …, 2023 - proceedings.neurips.cc
Abstract Deep Neural Networks and Reinforcement Learning methods have empirically
shown great promise in tackling challenging combinatorial problems. In those methods a …

Designing Universally-Approximating Deep Neural Networks: A First-Order Optimization Approach

Z Wu, M **ao, C Fang, Z Lin - IEEE Transactions on Pattern …, 2024 - ieeexplore.ieee.org
Universal approximation capability, also referred to as universality, is an important property
of deep neural networks, endowing them with the potency to accurately represent the …

Optimization-based separations for neural networks

I Safran, J Lee - Conference on Learning Theory, 2022 - proceedings.mlr.press
Depth separation results propose a possible theoretical explanation for the benefits of deep
neural networks over shallower architectures, establishing that the former possess superior …

Width is less important than depth in relu neural networks

G Vardi, G Yehudai, O Shamir - Conference on learning …, 2022 - proceedings.mlr.press
We solve an open question from Lu et al.(2017), by showing that any target network with
inputs in $\mathbb {R}^ d $ can be approximated by a width $ O (d) $ network (independent …

Size and depth separation in approximating benign functions with neural networks

G Vardi, D Reichman, T Pitassi… - … on Learning Theory, 2021 - proceedings.mlr.press
When studying the expressive power of neural networks, a main challenge is to understand
how the size and depth of the network affect its ability to approximate real functions …

Depth separation beyond radial functions

L Venturi, S Jelassi, T Ozuch, J Bruna - Journal of machine learning …, 2022 - jmlr.org
High-dimensional depth separation results for neural networks show that certain functions
can be efficiently approximated by two-hidden-layer networks but not by one-hidden-layer …

Regret guarantees for online deep control

X Chen, E Minasyan, JD Lee… - Learning for Dynamics …, 2023 - proceedings.mlr.press
Despite the immense success of deep learning in reinforcement learning and control, few
theoretical guarantees for neural networks exist for these problems. Deriving performance …