Spanning training progress: Temporal dual-depth scoring (tdds) for enhanced dataset pruning

X Zhang, J Du, Y Li, W **e… - Proceedings of the IEEE …, 2024 - openaccess.thecvf.com
Dataset pruning aims to construct a coreset capable of achieving performance comparable
to the original full dataset. Most existing dataset pruning methods rely on snapshot-based …

Refined Coreset Selection: Towards Minimal Coreset Size under Model Performance Constraints

X **a, J Liu, S Zhang, Q Wu, H Wei… - Forty-first International …, 2024 - openreview.net
Coreset selection is powerful in reducing computational costs and accelerating data
processing for deep learning algorithms. It strives to identify a small subset from large-scale …

Mitigating dialogue hallucination for large multi-modal models via adversarial instruction tuning

D Park, Z Qian, G Han, SN Lim - arxiv preprint arxiv:2403.10492, 2024 - arxiv.org
Mitigating hallucinations of Large Multi-modal Models (LMMs) is crucial to enhance their
reliability for general-purpose assistants. This paper shows that such hallucinations of LMMs …

Coreset selection with prioritized multiple objectives

X **a, J Liu, S Zhang, Q Wu, T Liu - arxiv preprint arxiv:2311.08675, 2023 - arxiv.org
Coreset selection is powerful in reducing computational costs and accelerating data
processing for deep learning algorithms. It strives to identify a small subset from large-scale …

Learning with noisy labels for robust fatigue detection

M Wang, R Hu, X Zhu, D Zhu, X Wang - Knowledge-Based Systems, 2024 - Elsevier
Fatigue is a significant safety concern across various domains, and accurate detection is
vital. However, the commonly employed fine-grained labels (seconds-based) frequently …

Efficient quantization-aware training with adaptive coreset selection

X Huang, Z Liu, SY Liu, KT Cheng - 2023 - openreview.net
The expanding model size and computation of deep neural networks (DNNs) have
increased the demand for efficient model deployment methods. Quantization-aware training …

Clipcleaner: Cleaning noisy labels with clip

C Feng, G Tzimiropoulos, I Patras - Proceedings of the 32nd ACM …, 2024 - dl.acm.org
Learning with Noisy labels (LNL) poses a significant challenge for the Machine Learning
community. Some of the most widely used approaches that select as clean samples for …

Lightweight spatial-channel feature disentanglement modeling with confidence evaluation for uncertain industrial image

L Lei, HX Li, HD Yang - Applied Mathematical Modelling, 2025 - Elsevier
Process uncertainty has a significant impact on industrial image processing. Existing deep
learning methods were established on high-quality datasets without considering the …

GDeR: Safeguarding Efficiency, Balancing, and Robustness via Prototypical Graph Pruning

G Zhang, H Dong, Y Zhang, Z Li, D Chen… - arxiv preprint arxiv …, 2024 - arxiv.org
Training high-quality deep models necessitates vast amounts of data, resulting in
overwhelming computational and memory demands. Recently, data pruning, distillation, and …

DynImpt: A Dynamic Data Selection Method for Improving Model Training Efficiency

W Huang, Y Zhang, S Guo, Y Shang… - IEEE Transactions on …, 2024 - ieeexplore.ieee.org
Selecting key data subsets for model training is an effective way to improve training
efficiency. Existing methods generally utilize a well-trained model to evaluate samples and …