Sampling is as easy as learning the score: theory for diffusion models with minimal data assumptions
We provide theoretical convergence guarantees for score-based generative models (SGMs)
such as denoising diffusion probabilistic models (DDPMs), which constitute the backbone of …
such as denoising diffusion probabilistic models (DDPMs), which constitute the backbone of …
NP-hardness of learning programs and partial MCSP
S Hirahara - 2022 ieee 63rd annual symposium on foundations …, 2022 - ieeexplore.ieee.org
A long-standing open question in computational learning theory is to prove NP-hardness of
learning efficient programs, the setting of which is in between proper learning and improper …
learning efficient programs, the setting of which is in between proper learning and improper …
Hardness of noise-free learning for two-hidden-layer neural networks
We give superpolynomial statistical query (SQ) lower bounds for learning two-hidden-layer
ReLU networks with respect to Gaussian inputs in the standard (noise-free) model. No …
ReLU networks with respect to Gaussian inputs in the standard (noise-free) model. No …
Learning deep relu networks is fixed-parameter tractable
We consider the problem of learning an unknown ReLU network with respect to Gaussian
inputs and obtain the first nontrivial results for networks of depth more than two. We give an …
inputs and obtain the first nontrivial results for networks of depth more than two. We give an …
Provably learning a multi-head attention layer
The multi-head attention layer is one of the key components of the transformer architecture
that sets it apart from traditional feed-forward models. Given a sequence length $ k …
that sets it apart from traditional feed-forward models. Given a sequence length $ k …
On the cryptographic hardness of learning single periodic neurons
We show a simple reduction which demonstrates the cryptographic hardness of learning a
single periodic neuron over isotropic Gaussian distributions in the presence of noise. More …
single periodic neuron over isotropic Gaussian distributions in the presence of noise. More …
Efficient algorithms for learning depth-2 neural networks with general relu activations
We present polynomial time and sample efficient algorithms for learning an unknown depth-
2 feedforward neural network with general ReLU activations, under mild non-degeneracy …
2 feedforward neural network with general ReLU activations, under mild non-degeneracy …
Computational complexity of learning neural networks: Smoothness and degeneracy
Understanding when neural networks can be learned efficientlyis a fundamental question in
learning theory. Existing hardness results suggest that assumptions on both the input …
learning theory. Existing hardness results suggest that assumptions on both the input …
Hardness of agnostically learning halfspaces from worst-case lattice problems
S Tiegel - The Thirty Sixth Annual Conference on Learning …, 2023 - proceedings.mlr.press
We show hardness of improperly learning halfspaces in the agnostic model, both in the
distribution-independent as well as the distribution-specific setting, based on the assumption …
distribution-independent as well as the distribution-specific setting, based on the assumption …
Learning (very) simple generative models is hard
Motivated by the recent empirical successes of deep generative models, we study the
computational complexity of the following unsupervised learning problem. For an unknown …
computational complexity of the following unsupervised learning problem. For an unknown …