Home

Awesome

VQAMix: Conditional Triplet Mixup for Medical Visual Question Answering paper

IEEE Transaction on Medical Imaging

This repository is the official implementation of VQAMix for the visual question answering task in medical domain. In this paper, we propose a simple yet effective data augmentation method, VQAMix, to mitigate the data limitation problem. Specifically, VQAMix generates more labeled training samples by linearly combining a pair of VQA samples, which can be easily embedded into any visual-language model to boost performance.

This repository is based on and inspired by @Jin-Hwa Kim's work and @Aizo-ai's work. We sincerely thank for their sharing of the codes.

Citation

Please cite this paper in your publications if it helps your research

@article{gong2022vqamix,
  title={VQAMix: Conditional Triplet Mixup for Medical Visual Question Answering},
  author={Haifan Gong and Guanqi Chen and Mingzhi Mao and Zhen Li and Guanbin Li},
  journal={IEEE Trans. on Medical Imaging},
  year={2022}
}

Overview of the vqamix framework

In VQAMix, two image-question pairs {Vi, Qi} and {Vj, Qj} are mixed. When the mixed sample is sent to the VQA model, the linguistic feature extracted from Qi will interact with the visual feature extracted from Vj, which constructs a new connection {Vj, Qi}. So is {Vi, Qj}. Thus, the label for the mixed image-question pair consists of four answer labels (Yi for {Vi, Qi}, Yj for {Vj, Qj}, Yk for {Vi, Qj} and Yl for {Vj, Qi}). And the weights of those answer labels are the probabilities of occurrence of those imagequestion pairs. The answer A is encoded as a one-hot vector Y.

Details of the vqamix framework

An overview of our proposed VQAMix enhanced by Learning with Missing Labels (LML) and Learning with Conditional-mixed Labels (LCL) strategies. Two VQA samples are combined linearly in the training phase. To ensure that the mixed label can be used to supervise the learning of VQA models, both LML and LCL scheme discards those two unspecified labels to solve the missing labels issue. Moreover, to avoid meaningless answers, the LCL scheme further utilizes the category of the question to avoid the model suffering from meaningless mixed labels.

Prerequisites

torch 1.6.0+ torchvision 0.6+

Dataset and Pre-trained Models

The processed data should be downloaded via Baidu Drive with the extract code: ioz1. Or you can download the data from the previous work MMQlink. The downloaded file should be extracted to data_RAD/ and data_PATH directory.

The trained models are available at Baidu Drive with extract code: i800.

Training and Testing

Just run the run_rad.sh run_path.sh for training and evaluation. The result json file can be found in the directory results/.

Comaprison with the sota

Comparison

License

MIT License

More information

If you have any problem, no hesitate contact us at haifangong@outlook.com