Home

Awesome

Increasing Model Capacity for Free: A Simple Strategy for Parameter Efficient Fine-Tuning [ICLR 2024]

<a href="https://www.linkedin.com/in/haobo-song-28a969167/?originalSubdomain=ch">Haobo Song*</a>, <a href="https://marcelluszhao.github.io/">Hao Zhao*</a>, <a href="https://scholar.google.com/citations?user=QSUfCpQAAAAJ&hl=en">Soumajit Majumder</a>, <a href="https://tlin-taolin.github.io/">Tao Lin</a>

Paper: http://arxiv.org/abs/2407.01320 (accepted at ICLR 2024)

TL;DR: We pursue the effective use of parameter-efficient modules (e.g., LoRA and Adapter) from the perspective of boosting model capacity.

This dicrectory contains code to run all the RoBERTa experiments in our paper. The code is based on adapterhub.

Environments

We use Pytorch 1.11.0+cu113 and Nvidia RTX 4090. Before running the code, please install the requirements and the propetl package by

python install -r requirements
python install .

How to run the models

To reproduce the experiment in the paper Table 1, you can simply run the following 3 shell (Adapter, LoRA and prefix).

ModelCoLASST-2MRPCQQPSTS-BMNLIQNLIRTEAvg
ProAdapter65.4394.1588.24/91.4189.40/86.0491.34/90.9586.5392.5876.5086.6
ProLoRA61.8194.0087.42/91.0088.85/85.2290.48/90.4785.7391.0563.7984.53
ProPrefix62.1693.6288.73/91.8087.59/83.7190.92/90.8385.3091.7572.6685.37
# propetl adapter
bash scripts/run_adapter.sh
# propetl LoRA
bash scripts/run_lora.sh
# propetl prefix tuning
bash scripts/run_prefix.sh

Citation

If you find this useful in your research, please consider citing:

@inproceedings{haobo2023increasing,
  title={Increasing Model Capacity for Free: A Simple Strategy for Parameter Efficient Fine-tuning},
  author={Haobo, SONG and Zhao, Hao and Majumder, Soumajit and Lin, Tao},
  booktitle={The Twelfth International Conference on Learning Representations},
  year={2024}
}