Rethinking data distillation: Do not overlook calibration

D Zhu, B Lei, J Zhang, Y Fang, Y **e… - Proceedings of the …, 2023 - openaccess.thecvf.com
Neural networks trained on distilled data often produce over-confident output and require
correction by calibration methods. Existing calibration methods such as temperature scaling …

Advancing neural network calibration: The role of gradient decay in large-margin Softmax optimization

S Zhang, L **e - Neural Networks, 2024 - Elsevier
This study introduces a novel hyperparameter in the Softmax function to regulate the rate of
gradient decay, which is dependent on sample probability. Our theoretical and empirical …

Distributionally robust ensemble of lottery tickets towards calibrated sparse network training

H Sapkota, D Wang, Z Tao… - Advances in Neural …, 2023 - proceedings.neurips.cc
The recently developed sparse network training methods, such as Lottery Ticket Hypothesis
(LTH) and its variants, have shown impressive learning capacity by finding sparse sub …

Quantifying lottery tickets under label noise: accuracy, calibration, and complexity

V Arora, D Irto, S Goldt… - Uncertainty in Artificial …, 2023 - proceedings.mlr.press
Pruning deep neural networks is a widely used strategy to alleviate the computational
burden in machine learning. Overwhelming empirical evidence suggests that pruned …

Uncovering the Hidden Cost of Model Compression

D Misra, M Chaudhary, A Goyal… - Proceedings of the …, 2024 - openaccess.thecvf.com
In an age dominated by resource-intensive foundation models the ability to efficiently adapt
to downstream tasks is crucial. Visual Prompting (VP) drawing inspiration from the prompting …

Balance is essence: Accelerating sparse training via adaptive gradient correction

B Lei, D Xu, R Zhang, S He… - … on Parsimony and …, 2024 - proceedings.mlr.press
Despite impressive performance, deep neural networks require significant memory and
computation costs, prohibiting their application in resource-constrained scenarios. Sparse …

Parametric -Norm Scaling Calibration

S Zhang, L **e - arxiv preprint arxiv:2412.15301, 2024 - arxiv.org
Output uncertainty indicates whether the probabilistic properties reflect objective
characteristics of the model output. Unlike most loss functions and metrics in machine …

Embracing Unknown Step by Step: Towards Reliable Sparse Training in Real World

B Lei, D Xu, R Zhang, B Mallick - arxiv preprint arxiv:2403.20047, 2024 - arxiv.org
Sparse training has emerged as a promising method for resource-efficient deep neural
networks (DNNs) in real-world applications. However, the reliability of sparse models …

A PID Controller Approach for Adaptive Probability-dependent Gradient Decay in Model Calibration

S Zhang, L **e - The Thirty-eighth Annual Conference on Neural … - openreview.net
Modern deep learning models often exhibit overconfident predictions, inadequately
capturing uncertainty. During model optimization, the expected calibration error tends to …

Learning Under Implicit Bias and Data Bias

J Li - 2023 - search.proquest.com
Modern machine learning tasks often involve the training of over-parameterized models and
the challenge of addressing data bias. However, despite recent advances, there remains a …