Home

Awesome

ECAMP

The official implementation of "ECAMP: Entity-centered Context-aware Medical Vision Language Pre-training".
Our paper can be found here

framework Some code is borrowed from MAE, huggingface and MRM

Installation

Clone this repository:

git clone https://github.com/ToniChopp/ECAMP.git

Install Python dependencies:

conda env create -f environment.yml

Resource fetching

As of now, we exclusively offer pre-training code, focusing solely on illustrating the process of retrieving MIMIC-CXR data

You can download ViTB/16 checkpoint here for pretraining.
Our pre-trained model can be found here.

New: Our distilled reports by LLM have been released. You can fetch them here

Pre-training

The distilled report and attention weights will be released as soon as our paper is accepted, but you can still use the original radiographs and report for pre-training.
We pre-train ECAMP on MIMIC-CXR using this command:

cd ECAMP/ECAMP/Pre-training
chmod a+x run.sh
./run.sh

Note that it is flexible to develop other pre-training models under this framework.

Reference

If you have found our work valuable for your research, we kindly suggest that you acknowledge and cite our contribution(s) by referencing:

@misc{wang2023ecamp,
      title={ECAMP: Entity-centered Context-aware Medical Vision Language Pre-training}, 
      author={Rongsheng Wang and Qingsong Yao and Haoran Lai and Zhiyang He and Xiaodong Tao and Zihang Jiang and S. Kevin Zhou},
      year={2023},
      eprint={2312.13316},
      archivePrefix={arXiv},
      primaryClass={cs.CV}
}

Hope you enjoy!