Home

Awesome

MEND: Meta dEmonstratioN Distillation for Efficient and Effective In-Context Learning

This repository focuses on tools and scripts for data distillation in the context of efficient in-context learning. Our work builds upon the MetaICL codebase.

Dependencies

Data Preprocessing

Pretrain C4 dataset

We utilize the validation set of C4 dataset, select "en" subset of validation split. You can also check our preprocessed data on Huggingface datasets.

Meta-train and Meta-test dataset

For details on downloading and preprocessing, kindly refer to the MetaICL documentation. You can also check our preprocessed data on Huggingface datasets.

Model Checkpoint

The model checkpoint is available in Google Drive.

Data Distillation Training

Inside src directory, you will find:

Pre-training:

cd scripts
sh c4_pretrain.sh

FineTuning

cd scripts
sh finetune.sh

License

MetaICL is CC-BY-NC 4.0 licensed.

Citation

If you use this code for your research, please cite our paper:

@inproceedings{
li2024mend,
title={{MEND}: Meta Demonstration Distillation for Efficient and Effective In-Context Learning},
author={Yichuan Li and Xiyao Ma and Sixing Lu and Kyumin Lee and Xiaohu Liu and Chenlei Guo},
booktitle={The Twelfth International Conference on Learning Representations},
year={2024},
url={https://openreview.net/forum?id=2Y5kBPtU0o}
}