Turnitin
降AI改写
早检测系统
早降重系统
Turnitin-UK版
万方检测-期刊版
维普编辑部版
Grammarly检测
Paperpass检测
checkpass检测
PaperYY检测
Rethinking data distillation: Do not overlook calibration
Neural networks trained on distilled data often produce over-confident output and require
correction by calibration methods. Existing calibration methods such as temperature scaling …
correction by calibration methods. Existing calibration methods such as temperature scaling …
Advancing neural network calibration: The role of gradient decay in large-margin Softmax optimization
S Zhang, L **e - Neural Networks, 2024 - Elsevier
This study introduces a novel hyperparameter in the Softmax function to regulate the rate of
gradient decay, which is dependent on sample probability. Our theoretical and empirical …
gradient decay, which is dependent on sample probability. Our theoretical and empirical …
Distributionally robust ensemble of lottery tickets towards calibrated sparse network training
The recently developed sparse network training methods, such as Lottery Ticket Hypothesis
(LTH) and its variants, have shown impressive learning capacity by finding sparse sub …
(LTH) and its variants, have shown impressive learning capacity by finding sparse sub …
Quantifying lottery tickets under label noise: accuracy, calibration, and complexity
Pruning deep neural networks is a widely used strategy to alleviate the computational
burden in machine learning. Overwhelming empirical evidence suggests that pruned …
burden in machine learning. Overwhelming empirical evidence suggests that pruned …
Uncovering the Hidden Cost of Model Compression
In an age dominated by resource-intensive foundation models the ability to efficiently adapt
to downstream tasks is crucial. Visual Prompting (VP) drawing inspiration from the prompting …
to downstream tasks is crucial. Visual Prompting (VP) drawing inspiration from the prompting …
Balance is essence: Accelerating sparse training via adaptive gradient correction
Despite impressive performance, deep neural networks require significant memory and
computation costs, prohibiting their application in resource-constrained scenarios. Sparse …
computation costs, prohibiting their application in resource-constrained scenarios. Sparse …
Parametric -Norm Scaling Calibration
S Zhang, L **e - arxiv preprint arxiv:2412.15301, 2024 - arxiv.org
Output uncertainty indicates whether the probabilistic properties reflect objective
characteristics of the model output. Unlike most loss functions and metrics in machine …
characteristics of the model output. Unlike most loss functions and metrics in machine …
Embracing Unknown Step by Step: Towards Reliable Sparse Training in Real World
Sparse training has emerged as a promising method for resource-efficient deep neural
networks (DNNs) in real-world applications. However, the reliability of sparse models …
networks (DNNs) in real-world applications. However, the reliability of sparse models …
A PID Controller Approach for Adaptive Probability-dependent Gradient Decay in Model Calibration
S Zhang, L **e - The Thirty-eighth Annual Conference on Neural … - openreview.net
Modern deep learning models often exhibit overconfident predictions, inadequately
capturing uncertainty. During model optimization, the expected calibration error tends to …
capturing uncertainty. During model optimization, the expected calibration error tends to …
Learning Under Implicit Bias and Data Bias
J Li - 2023 - search.proquest.com
Modern machine learning tasks often involve the training of over-parameterized models and
the challenge of addressing data bias. However, despite recent advances, there remains a …
the challenge of addressing data bias. However, despite recent advances, there remains a …