Turnitin
降AI改写
早检测系统
早降重系统
Turnitin-UK版
万方检测-期刊版
维普编辑部版
Grammarly检测
Paperpass检测
checkpass检测
PaperYY检测
Quantum variational algorithms are swamped with traps
One of the most important properties of classical neural networks is how surprisingly
trainable they are, though their training algorithms typically rely on optimizing complicated …
trainable they are, though their training algorithms typically rely on optimizing complicated …
Fl-ntk: A neural tangent kernel-based framework for federated learning analysis
Federated Learning (FL) is an emerging learning scheme that allows different distributed
clients to train deep neural networks together without data sharing. Neural networks have …
clients to train deep neural networks together without data sharing. Neural networks have …
When deep learning meets polyhedral theory: A survey
In the past decade, deep learning became the prevalent methodology for predictive
modeling thanks to the remarkable accuracy of deep neural networks in tasks such as …
modeling thanks to the remarkable accuracy of deep neural networks in tasks such as …
Provably learning a multi-head attention layer
The multi-head attention layer is one of the key components of the transformer architecture
that sets it apart from traditional feed-forward models. Given a sequence length $ k …
that sets it apart from traditional feed-forward models. Given a sequence length $ k …
Towards lower bounds on the depth of ReLU neural networks
We contribute to a better understanding of the class of functions that is represented by a
neural network with ReLU activations and a given architecture. Using techniques from mixed …
neural network with ReLU activations and a given architecture. Using techniques from mixed …
Bounding the width of neural networks via coupled initialization a worst case analysis
A common method in training neural networks is to initialize all the weights to be
independent Gaussian vectors. We observe that by instead initializing the weights into …
independent Gaussian vectors. We observe that by instead initializing the weights into …
Hardness of noise-free learning for two-hidden-layer neural networks
We give superpolynomial statistical query (SQ) lower bounds for learning two-hidden-layer
ReLU networks with respect to Gaussian inputs in the standard (noise-free) model. No …
ReLU networks with respect to Gaussian inputs in the standard (noise-free) model. No …
Training Fully Connected Neural Networks is -Complete
We consider the algorithmic problem of finding the optimal weights and biases for a two-
layer fully connected neural network to fit a given set of data points, also known as empirical …
layer fully connected neural network to fit a given set of data points, also known as empirical …
Learning narrow one-hidden-layer relu networks
We consider the well-studied problem of learning a linear combination of $ k $ ReLU
activations with respect to a Gaussian distribution on inputs in $ d $ dimensions. We give the …
activations with respect to a Gaussian distribution on inputs in $ d $ dimensions. We give the …
Agnostically learning multi-index models with queries
We study the power of query access for the fundamental task of agnostic learning under the
Gaussian distribution. In the agnostic model, no assumptions are made on the labels of the …
Gaussian distribution. In the agnostic model, no assumptions are made on the labels of the …