Spanning training progress: Temporal dual-depth scoring (tdds) for enhanced dataset pruning
Dataset pruning aims to construct a coreset capable of achieving performance comparable
to the original full dataset. Most existing dataset pruning methods rely on snapshot-based …
to the original full dataset. Most existing dataset pruning methods rely on snapshot-based …
Refined Coreset Selection: Towards Minimal Coreset Size under Model Performance Constraints
Coreset selection is powerful in reducing computational costs and accelerating data
processing for deep learning algorithms. It strives to identify a small subset from large-scale …
processing for deep learning algorithms. It strives to identify a small subset from large-scale …
Mitigating dialogue hallucination for large multi-modal models via adversarial instruction tuning
Mitigating hallucinations of Large Multi-modal Models (LMMs) is crucial to enhance their
reliability for general-purpose assistants. This paper shows that such hallucinations of LMMs …
reliability for general-purpose assistants. This paper shows that such hallucinations of LMMs …
Coreset selection with prioritized multiple objectives
Coreset selection is powerful in reducing computational costs and accelerating data
processing for deep learning algorithms. It strives to identify a small subset from large-scale …
processing for deep learning algorithms. It strives to identify a small subset from large-scale …
Learning with noisy labels for robust fatigue detection
M Wang, R Hu, X Zhu, D Zhu, X Wang - Knowledge-Based Systems, 2024 - Elsevier
Fatigue is a significant safety concern across various domains, and accurate detection is
vital. However, the commonly employed fine-grained labels (seconds-based) frequently …
vital. However, the commonly employed fine-grained labels (seconds-based) frequently …
Efficient quantization-aware training with adaptive coreset selection
The expanding model size and computation of deep neural networks (DNNs) have
increased the demand for efficient model deployment methods. Quantization-aware training …
increased the demand for efficient model deployment methods. Quantization-aware training …
Clipcleaner: Cleaning noisy labels with clip
Learning with Noisy labels (LNL) poses a significant challenge for the Machine Learning
community. Some of the most widely used approaches that select as clean samples for …
community. Some of the most widely used approaches that select as clean samples for …
Lightweight spatial-channel feature disentanglement modeling with confidence evaluation for uncertain industrial image
L Lei, HX Li, HD Yang - Applied Mathematical Modelling, 2025 - Elsevier
Process uncertainty has a significant impact on industrial image processing. Existing deep
learning methods were established on high-quality datasets without considering the …
learning methods were established on high-quality datasets without considering the …
GDeR: Safeguarding Efficiency, Balancing, and Robustness via Prototypical Graph Pruning
Training high-quality deep models necessitates vast amounts of data, resulting in
overwhelming computational and memory demands. Recently, data pruning, distillation, and …
overwhelming computational and memory demands. Recently, data pruning, distillation, and …
DynImpt: A Dynamic Data Selection Method for Improving Model Training Efficiency
Selecting key data subsets for model training is an effective way to improve training
efficiency. Existing methods generally utilize a well-trained model to evaluate samples and …
efficiency. Existing methods generally utilize a well-trained model to evaluate samples and …