Dataset distillation: A comprehensive review
Recent success of deep learning is largely attributed to the sheer amount of data used for
training deep neural networks. Despite the unprecedented success, the massive data …
training deep neural networks. Despite the unprecedented success, the massive data …
A comprehensive survey of dataset distillation
Deep learning technology has developed unprecedentedly in the last decade and has
become the primary choice in many application domains. This progress is mainly attributed …
become the primary choice in many application domains. This progress is mainly attributed …
Dataset distillation by matching training trajectories
Dataset distillation is the task of synthesizing a small dataset such that a model trained on
the synthetic set will match the test accuracy of the model trained on the full dataset. In this …
the synthetic set will match the test accuracy of the model trained on the full dataset. In this …
Dataset distillation via factorization
In this paper, we study dataset distillation (DD), from a novel perspective and introduce
a\emph {dataset factorization} approach, termed\emph {HaBa}, which is a plug-and-play …
a\emph {dataset factorization} approach, termed\emph {HaBa}, which is a plug-and-play …
Cafe: Learning to condense dataset by aligning features
Dataset condensation aims at reducing the network training effort through condensing a
cumbersome training set into a compact synthetic one. State-of-the-art approaches largely …
cumbersome training set into a compact synthetic one. State-of-the-art approaches largely …
Generalizing dataset distillation via deep generative prior
Dataset Distillation aims to distill an entire dataset's knowledge into a few synthetic images.
The idea is to synthesize a small number of synthetic data points that, when given to a …
The idea is to synthesize a small number of synthetic data points that, when given to a …
Slimmable dataset condensation
Dataset distillation, also known as dataset condensation, aims to compress a large dataset
into a compact synthetic one. Existing methods perform dataset condensation by assuming a …
into a compact synthetic one. Existing methods perform dataset condensation by assuming a …
Dataset condensation with distribution matching
Computational cost of training state-of-the-art deep models in many learning problems is
rapidly increasing due to more sophisticated models and larger datasets. A recent promising …
rapidly increasing due to more sophisticated models and larger datasets. A recent promising …
Dataset condensation via efficient synthetic-data parameterization
The great success of machine learning with massive amounts of data comes at a price of
huge computation costs and storage for training and tuning. Recent studies on dataset …
huge computation costs and storage for training and tuning. Recent studies on dataset …
Dataset distillation using neural feature regression
Dataset distillation aims to learn a small synthetic dataset that preserves most of the
information from the original dataset. Dataset distillation can be formulated as a bi-level …
information from the original dataset. Dataset distillation can be formulated as a bi-level …