Home

Awesome

<div align="center"> <h1>Sparse Low-rank Adaptation of Pre-trained Language Models</h1> </div>

🎉 This is the implementation of EMNLP 2023 paper:Sparse Low-rank Adaptation of Pre-trained Language Models

Requirements

To run our code, please install all the dependency packages by using the following command:

pip install -r requirements.txt

Preparation

Prepare the Data and Modify the Data Path

In the paper/code, we use the GLUE datasets, you can download the data from Huggingface or from our Google Drive

After download the data, please replace the following data path definition with your data path:

Prepare the model

You can download the base model and the corresponding tokenizer from Huggingface. And after that, do not forget to modify the model_name_or_path and tokenizer_name in script file (.sh).

Baseline

We provide the implementation of LoRA, Adapter, BitFit and Full-parameter Fine-Tune. You can apply these baselines by running the following codes:

cd scripts
# LoRA
bash run_glue_lora.sh
# Adapter
bash run_glue_adapter.sh
# BitFit
bash run_glue_bitfit.sh
# Full-parameter Fine-Tune
bash run_glue_finetune.sh

SoRA

You can apply SoRA by running the following codes:

cd scripts
# without the sparsifying scheduler
bash run_glue_sora_no_schedule.sh
# with the sparsifying scheduler (Algorithm 1)
bash run_glue_sora_schedule_dense.sh

We explain some of the arguments as follows:

Bugs or questions?

If you have any questions related to the codes or the paper, please contact Ning Ding (dn97@mail.tsinghua.edu.cn), Xingtai Lv (lvxt20@mails.tsinghua.edu.cn) or open an issue.

Citation

If you find our work useful, please use the following citation:

@article{ding2023sparse,
  title={Sparse Low-rank Adaptation of Pre-trained Language Models},
  author={Ding, Ning and Lv, Xingtai and Wang, Qiaosen and Chen, Yulin and Zhou, Bowen and Liu, Zhiyuan and Sun, Maosong},
  journal={arXiv preprint arXiv:2311.11696},
  year={2023}
}