Home

Awesome

PeFoMed

This is the official implementation of PeFoMed: Parameter Efficient Fine-tuning of Multimodal Large Language Models for Medical Imaging.

<div align=center> <img src="fig/model.png" style="zoom:75%;"> </div> <center>Figure 1: Overview of the PeFoMed. </center>

Datasets

The configuration of all datasets needs to be set in the corresponding dataset configuration file in the pefomed/configs/datasets/medical

Stage 1 finetune datasets: ROCO, CLEF2022, MEDICAT, and MIMIC-CXR

Stage 2 finetune medical VQA datasets: VQA-RAD, PathVQA and Slake.

Stage 2 finetune MRG dataset: IU-Xray

Acknowledgement

If you're using PeFoMed in your research or applications, please cite using this BibTeX:

@misc{liu2024pefomedparameterefficientfinetuning,
      title={PeFoMed: Parameter Efficient Fine-tuning of Multimodal Large Language Models for Medical Imaging}, 
      author={Gang Liu and Jinlong He and Pengfei Li and Genrong He and Zhaolin Chen and Shenjun Zhong},
      year={2024},
      eprint={2401.02797},
      archivePrefix={arXiv},
      primaryClass={cs.CL},
      url={https://arxiv.org/abs/2401.02797}, 
}

License

This repository is under BSD 3-Clause License.

Many codes are based on Lavis and MiniGPT-v2