A review of off-policy evaluation in reinforcement learning

M Uehara, C Shi, N Kallus - arxiv preprint arxiv:2212.06355, 2022 - arxiv.org
Reinforcement learning (RL) is one of the most vibrant research frontiers in machine
learning and has been recently applied to solve a number of challenging problems. In this …

Score approximation, estimation and distribution recovery of diffusion models on low-dimensional data

M Chen, K Huang, T Zhao… - … Conference on Machine …, 2023 - proceedings.mlr.press
Diffusion models achieve state-of-the-art performance in various generation tasks. However,
their theoretical foundations fall far behind. This paper studies score approximation …

Diffusion models are minimax optimal distribution estimators

K Oko, S Akiyama, T Suzuki - International Conference on …, 2023 - proceedings.mlr.press
While efficient distribution learning is no doubt behind the groundbreaking success of
diffusion modeling, its theoretical guarantees are quite limited. In this paper, we provide the …

High-dimensional asymptotics of feature learning: How one gradient step improves the representation

J Ba, MA Erdogdu, T Suzuki, Z Wang… - Advances in Neural …, 2022 - proceedings.neurips.cc
We study the first gradient descent step on the first-layer parameters $\boldsymbol {W} $ in a
two-layer neural network: $ f (\boldsymbol {x})=\frac {1}{\sqrt {N}}\boldsymbol {a}^\top\sigma …

A primer on Bayesian neural networks: review and debates

J Arbel, K Pitas, M Vladimirova, V Fortuin - arxiv preprint arxiv:2309.16314, 2023 - arxiv.org
Neural networks have achieved remarkable performance across various problem domains,
but their widespread applicability is hindered by inherent limitations such as overconfidence …

A theoretical analysis of deep Q-learning

J Fan, Z Wang, Y **e, Z Yang - Learning for dynamics and …, 2020 - proceedings.mlr.press
Despite the great empirical success of deep reinforcement learning, its theoretical
foundation is less well understood. In this work, we make the first attempt to theoretically …

Variational physics-informed neural networks for solving partial differential equations

E Kharazmi, Z Zhang, GE Karniadakis - arxiv preprint arxiv:1912.00873, 2019 - arxiv.org
Physics-informed neural networks (PINNs)[31] use automatic differentiation to solve partial
differential equations (PDEs) by penalizing the PDE in the loss function at a random set of …

On the rate of convergence of fully connected deep neural network regression estimates

M Kohler, S Langer - The Annals of Statistics, 2021 - JSTOR
Recent results in nonparametric regression show that deep learning, that is, neural network
estimates with many hidden layers, are able to circumvent the so-called curse of …

Deep network approximation for smooth functions

J Lu, Z Shen, H Yang, S Zhang - SIAM Journal on Mathematical Analysis, 2021 - SIAM
This paper establishes the optimal approximation error characterization of deep rectified
linear unit (ReLU) networks for smooth functions in terms of both width and depth …

A tale of tails: Model collapse as a change of scaling laws

E Dohmatob, Y Feng, P Yang, F Charton… - arxiv preprint arxiv …, 2024 - arxiv.org
As AI model size grows, neural scaling laws have become a crucial tool to predict the
improvements of large models when increasing capacity and the size of original (human or …