Home

Awesome

MA-AGIQA:Large Multi-modality Model Assiste AI-Generated Image Quality Assessment

Platform Python Pytorch License arXiv

This repository is the official PyTorch implementation of MA-AGIQA:Large Multi-modality Model Assisted AI-Generated Image Quality Assessment (ACM MM'24 oral).

Network Architecture

image.png

Requirements

You can use your exsiting conda environment.

If you want to create an new environment, please refer to mPLUG-Owl2 and follow their installation instruction.

After that, you need

conda activate mplug_owl2
pip install omegaconf, opencv-python

Usage

To use our code, firstly you should extract fine-grained semantic features. After that, you can run the train or test steps. Here, I use AGIQA-3k as an example. You can easily change to your own datasets with slightly changes to the config files.

Semantic Feature Extraction

We use official mPLUG-Owl2 to extract semantic features. The feature extraction codes are based on mPLUG-Owl2, great thanks to them!

You can get semantic feature by ( you should run for question_type 1 and 2)

python getFeature.py --config configs/AGIQA_3k/MA_AGIQA.yaml

if you have error when connect to Hugging Face, we recommand you use

HF_ENDPOINT=https://hf-mirror.com python getFeature.py --config configs/AGIQA_3k/MA_AGIQA.yaml

Train and Test

Train and test procedure are integrated in the main.py file. The process will run test step when training is done.

If you only want to run test step, please set "epoch" in yaml file to 0. Train and Test

python main.py --config configs/AGIQA_3k/MA_AGIQA.yaml

Performance

image.png

Model Checkpoints

You could download the checkpoints (30 epochs) through Google Drive.

<!-- - AIGCQA-20k [Pre-trained](https://drive.google.com/file/d/1nKVcmBw-K9nS4tplZhMwHySpyy8ToAzK/view?usp=sharing) -->

TODO

Citation

If you find our code or model useful for your research, please cite:

@misc{wang2024large,
      title={Large Multi-modality Model Assisted AI-Generated Image Quality Assessment}, 
      author={Puyi Wang and Wei Sun and Zicheng Zhang and Jun Jia and Yanwei Jiang and Zhichao Zhang and Xiongkuo Min and Guangtao Zhai},
      year={2024},
      eprint={2404.17762},
      archivePrefix={arXiv},
      primaryClass={cs.CV}
}

Acknowledgement

Part of our codes are based on MANIQA and mPLUG-Owl2. Thanks for their awesome work!