Turnitin
降AI改写
早检测系统
早降重系统
Turnitin-UK版
万方检测-期刊版
维普编辑部版
Grammarly检测
Paperpass检测
checkpass检测
PaperYY检测
To compress or not to compress—self-supervised learning and information theory: A review
Deep neural networks excel in supervised learning tasks but are constrained by the need for
extensive labeled data. Self-supervised learning emerges as a promising alternative …
extensive labeled data. Self-supervised learning emerges as a promising alternative …
Generalization bounds: Perspectives from information theory and PAC-Bayes
A fundamental question in theoretical machine learning is generalization. Over the past
decades, the PAC-Bayesian approach has been established as a flexible framework to …
decades, the PAC-Bayesian approach has been established as a flexible framework to …
Control batch size and learning rate to generalize well: Theoretical and empirical evidence
Deep neural networks have received dramatic success based on the optimization method of
stochastic gradient descent (SGD). However, it is still not clear how to tune hyper …
stochastic gradient descent (SGD). However, it is still not clear how to tune hyper …
Recent advances in deep learning theory
Deep learning is usually described as an experiment-driven field under continuous criticizes
of lacking theoretical foundations. This problem has been partially fixed by a large volume of …
of lacking theoretical foundations. This problem has been partially fixed by a large volume of …
On the power of over-parametrization in neural networks with quadratic activation
We provide new theoretical insights on why over-parametrization is effective in learning
neural networks. For a $ k $ hidden node shallow network with quadratic activation and $ n …
neural networks. For a $ k $ hidden node shallow network with quadratic activation and $ n …
Tightening mutual information-based bounds on generalization error
An information-theoretic upper bound on the generalization error of supervised learning
algorithms is derived. The bound is constructed in terms of the mutual information between …
algorithms is derived. The bound is constructed in terms of the mutual information between …
Information-theoretic generalization bounds for SGLD via data-dependent estimates
In this work, we improve upon the stepwise analysis of noisy iterative learning algorithms
initiated by Pensia, Jog, and Loh (2018) and recently extended by Bu, Zou, and Veeravalli …
initiated by Pensia, Jog, and Loh (2018) and recently extended by Bu, Zou, and Veeravalli …
Sharpened generalization bounds based on conditional mutual information and an application to noisy, iterative algorithms
The information-theoretic framework of Russo and Zou (2016) and Xu and Raginsky (2017)
provides bounds on the generalization error of a learning algorithm in terms of the mutual …
provides bounds on the generalization error of a learning algorithm in terms of the mutual …
Information-theoretic generalization bounds for stochastic gradient descent
We study the generalization properties of the popular stochastic optimization method known
as stochastic gradient descent (SGD) for optimizing general non-convex loss functions. Our …
as stochastic gradient descent (SGD) for optimizing general non-convex loss functions. Our …
Topological generalization bounds for discrete-time stochastic optimization algorithms
We present a novel set of rigorous and computationally efficient topology-based complexity
notions that exhibit a strong correlation with the generalization gap in modern deep neural …
notions that exhibit a strong correlation with the generalization gap in modern deep neural …