Towards lossless dataset distillation via difficulty-aligned trajectory matching
The ultimate goal of Dataset Distillation is to synthesize a small synthetic dataset such that a
model trained on this synthetic set will perform equally well as a model trained on the full …
model trained on this synthetic set will perform equally well as a model trained on the full …
Distill gold from massive ores: Bi-level data pruning towards efficient dataset distillation
Data-efficient learning has garnered significant attention, especially given the current trend
of large multi-modal models. Recently, dataset distillation has become an effective approach …
of large multi-modal models. Recently, dataset distillation has become an effective approach …
Prioritize Alignment in Dataset Distillation
Dataset Distillation aims to compress a large dataset into a significantly more compact,
synthetic one without compromising the performance of the trained models. To achieve this …
synthetic one without compromising the performance of the trained models. To achieve this …
ATOM: Attention Mixer for Efficient Dataset Distillation
Recent works in dataset distillation seek to minimize training expenses by generating a
condensed synthetic dataset that encapsulates the information present in a larger real …
condensed synthetic dataset that encapsulates the information present in a larger real …
Emphasizing discriminative features for dataset distillation in complex scenarios
Dataset distillation has demonstrated strong performance on simple datasets like CIFAR,
MNIST, and TinyImageNet but struggles to achieve similar results in more complex …
MNIST, and TinyImageNet but struggles to achieve similar results in more complex …
Going Beyond Feature Similarity: Effective Dataset distillation based on Class-aware Conditional Mutual Information
Dataset distillation (DD) aims to minimize the time and memory consumption needed for
training deep neural networks on large datasets, by creating a smaller synthetic dataset that …
training deep neural networks on large datasets, by creating a smaller synthetic dataset that …
Diffusion-Augmented Coreset Expansion for Scalable Dataset Distillation
With the rapid scaling of neural networks, data storage and communication demands have
intensified. Dataset distillation has emerged as a promising solution, condensing information …
intensified. Dataset distillation has emerged as a promising solution, condensing information …
DELT: A Simple Diversity-driven EarlyLate Training for Dataset Distillation
Recent advances in dataset distillation have led to solutions in two main directions. The
conventional batch-to-batch matching mechanism is ideal for small-scale datasets and …
conventional batch-to-batch matching mechanism is ideal for small-scale datasets and …
FairDD: Fair Dataset Distillation via Synchronized Matching
Condensing large datasets into smaller synthetic counterparts has demonstrated its promise
for image classification. However, previous research has overlooked a crucial concern in …
for image classification. However, previous research has overlooked a crucial concern in …
Dataset Distillers Are Good Label Denoisers In the Wild
Learning from noisy data has become essential for adapting deep learning models to real-
world applications. Traditional methods often involve first evaluating the noise and then …
world applications. Traditional methods often involve first evaluating the noise and then …