Home

Awesome

🌟 Diversity-Driven Synthesis: Enhancing Dataset Distillation through Directed Weight Adjustment

πŸ”₯ NeurIPS 2024 Spotlight πŸ”₯

Diversity-Driven Synthesis: Enhancing Dataset Distillation through Directed Weight Adjustment.<br> Jiawei Du, Xin Zhang, Juncheng Hu, Wenxin Huang, Joey Tianyi Zhou <br> A*Star, XiDian University, National University of Singapore, Hubei University

πŸ“– Introduction

<p align="justify"> The sharp increase in data-related expenses has motivated research into condensing datasets while retaining the most informative features. Dataset distillation has thus recently come to the fore. This paradigm generates synthetic datasets that are representative enough to replace the original dataset in training a neural network. To avoid redundancy in these synthetic datasets, it is crucial that each element contains unique features and remains diverse from others during the synthesis stage. In this paper, we provide a thorough theoretical and empirical analysis of diversity within synthesized datasets. We argue that enhancing diversity can improve the parallelizable yet isolated synthesizing approach. Specifically, we introduce a novel method that employs dynamic and directed weight adjustment techniques to modulate the synthesis process, thereby maximizing the representativeness and diversity of each synthetic instance. Our method ensures that each batch of synthetic data mirrors the characteristics of a large, varying subset of the original dataset. Extensive experiments across multiple datasets, including CIFAR, Tiny-ImageNet, and ImageNet-1K, demonstrate the superior performance of our method, highlighting its effectiveness in producing diverse and representative synthetic datasets with minimal computational expense.</p>

βš™οΈ Installation

To get started, follow these instructions to set up the environment and install dependencies.

  1. Clone this repository:

    git clone https://github.com/AngusDujw/Diversity-Driven-Synthesis.git
    cd Diversity-Driven-Synthesis
    
  2. Install required packages: You don’t need to create a new environment; simply ensure that you have compatible versions of CUDA and PyTorch installed.


πŸš€ Usage

Here’s how to use this code for distillation and evaluation:

we also provide the .sh script in the scripts directory.


πŸ“Š Results

Our experiments demonstrate the effectiveness of the proposed approach across various benchmarks.

Results

For detailed experimental results and further analysis, please refer to the full paper.


πŸ“‘ Citation

If you find this code useful in your research, please consider citing our work:

@inproceedings{dwa2024neurips,
    title={Diversity-Driven Synthesis: Enhancing Dataset Distillation through Directed Weight Adjustment},
    author={Du, Jiawei and Zhang, Xin and Hu, Juncheng and Huang, Wenxin and Zhou, Joey Tianyi},
    booktitle={Adv. Neural Inf. Process. Syst. (NeurIPS)},
    year={2024}
}

πŸŽ‰ Reference

Our code has referred to previous work:

Squeeze, Recover and Relabel: Dataset Condensation at ImageNet Scale From A New Perspective