Turnitin
降AI改写
早检测系统
早降重系统
Turnitin-UK版
万方检测-期刊版
维普编辑部版
Grammarly检测
Paperpass检测
checkpass检测
PaperYY检测
Revisiting knowledge distillation: An inheritance and exploration framework
Abstract Knowledge Distillation (KD) is a popular technique to transfer knowledge from a
teacher model or ensemble to a student model. Its success is generally attributed to the …
teacher model or ensemble to a student model. Its success is generally attributed to the …
Model-independent detection of new physics signals using interpretable semisupervised classifier tests
The supplementary material contains the proof of Theorem 4.1, some of the proposed
algorithms from Section 3.2, and details about the exploratory data analysis of the Higgs …
algorithms from Section 3.2, and details about the exploratory data analysis of the Higgs …
Robust tensor decomposition via orientation invariant tubal nuclear norms
Aiming at recovering an unknown tensor (ie, multi-way array) corrupted by both sparse
outliers and dense noises, robust tensor decomposition (RTD) serves as a powerful pre …
outliers and dense noises, robust tensor decomposition (RTD) serves as a powerful pre …
On the deep active-subspace method
W Edeling - SIAM/ASA Journal on Uncertainty Quantification, 2023 - SIAM
The deep active-subspace method is a neural-network based tool for the propagation of
uncertainty through computational models with high-dimensional input spaces. Unlike the …
uncertainty through computational models with high-dimensional input spaces. Unlike the …
A supervised learning approach involving active subspaces for an efficient genetic algorithm in high-dimensional optimization problems
In this work, we present an extension of genetic algorithm (GA) which exploits the
supervised learning technique called active subspaces (AS) to evolve the individuals on a …
supervised learning technique called active subspaces (AS) to evolve the individuals on a …
Low-rank gradient descent
Several recent empirical studies demonstrate that important machine learning tasks such as
training deep neural networks, exhibit a low-rank structure, where most of the variation in the …
training deep neural networks, exhibit a low-rank structure, where most of the variation in the …
Learning the subspace of variation for global optimization of functions with low effective dimension
We propose an algorithmic framework, that employs active subspace techniques, for
scalable global optimization of functions with low effective dimension (also referred to as low …
scalable global optimization of functions with low effective dimension (also referred to as low …
Learning active subspaces for effective and scalable uncertainty quantification in deep neural networks
Bayesian inference for neural networks, or Bayesian deep learning, has the potential to
provide well-calibrated predictions with quantified uncertainty and robustness. However, the …
provide well-calibrated predictions with quantified uncertainty and robustness. However, the …
A dimensionality reduction approach for convolutional neural networks
The focus of this work is on the application of classical Model Order Reduction techniques,
such as Active Subspaces and Proper Orthogonal Decomposition, to Deep Neural …
such as Active Subspaces and Proper Orthogonal Decomposition, to Deep Neural …
Few-bit backward: Quantized gradients of activation functions for memory footprint reduction
Memory footprint is one of the main limiting factors for large neural network training. In
backpropagation, one needs to store the input to each operation in the computational graph …
backpropagation, one needs to store the input to each operation in the computational graph …