Data-to-Model Distillation: Data-Efficient Learning Framework
Dataset distillation aims to distill the knowledge of a large-scale real dataset into small yet informative synthetic data such that a model trained on it performs as well as a model trained on the full dataset. Despite recent progress, existing dataset distillation methods often struggle with compu...
Gespeichert in:
Hauptverfasser: | , , , , , |
---|---|
Format: | Artikel |
Sprache: | eng |
Schlagworte: | |
Online-Zugang: | Volltext bestellen |
Tags: |
Tag hinzufügen
Keine Tags, Fügen Sie den ersten Tag hinzu!
|
Zusammenfassung: | Dataset distillation aims to distill the knowledge of a large-scale real
dataset into small yet informative synthetic data such that a model trained on
it performs as well as a model trained on the full dataset. Despite recent
progress, existing dataset distillation methods often struggle with
computational efficiency, scalability to complex high-resolution datasets, and
generalizability to deep architectures. These approaches typically require
retraining when the distillation ratio changes, as knowledge is embedded in raw
pixels. In this paper, we propose a novel framework called Data-to-Model
Distillation (D2M) to distill the real dataset's knowledge into the learnable
parameters of a pre-trained generative model by aligning rich representations
extracted from real and generated images. The learned generative model can then
produce informative training images for different distillation ratios and deep
architectures. Extensive experiments on 15 datasets of varying resolutions show
D2M's superior performance, re-distillation efficiency, and cross-architecture
generalizability. Our method effectively scales up to high-resolution 128x128
ImageNet-1K. Furthermore, we verify D2M's practical benefits for downstream
applications in neural architecture search. |
---|---|
DOI: | 10.48550/arxiv.2411.12841 |