Home

Awesome

<div align="center"> <h2 class="papername"> <img src="./assets/logo3.png" style="vertical-align: -10px;" :height="70px" width="70px"> Peacock: A Family of Arabic Multimodal Large Language Models and Benchmarks </h2> <div> <div> <a href="https://dlnlp.ai/index.html#Team" target="_blank">Fakhraddin Alwajih</a>, <a href="https://dlnlp.ai/index.html#Team" target="_blank">El Moatez Billah Nagoudi</a>, <a href="https://dlnlp.ai/index.html#Team" target="_blank">Gagan Bhatia</a>, <a href="https://dlnlp.ai/index.html#Team" target="_blank">Abdelrahman Mohamed</a>, <a href="https://dlnlp.ai/index.html#Team" target="_blank">Muhammad Abdul-Mageed</a> </div>

The University of British Columbia, Invertible AI<br>

<p align="center"> <a href='https://github.com/UBC-NLP/peacock'><img src='https://img.shields.io/badge/Project-Page-Green'></a> <a href='https://arxiv.org/abs/2403.01031'><img src='https://img.shields.io/badge/Paper-Arxiv-red'></a> <a href='https://huggingface.co/datasets/UBC-NLP/Henna-Bench-Eval'><img src='https://img.shields.io/badge/Dataset-HuggingFace-blue'></a> <a href='https://aclanthology.org/2024.acl-long.689/'><img src='https://img.shields.io/badge/Publication-ACL2024-orange'></a> </p>

:fire: Details will be released. Stay tuned :beers: :+1:

</div> </div>

If you find this work useful for your research, please kindly cite our paper and star our repository.

Updates

Abstract

Multimodal large language models (MLLMs) have proven effective in a wide range of tasks requiring complex reasoning and linguistic comprehension. However, due to a lack of high-quality multimodal resources in languages other than English, success of MLLMs remains relatively limited to English-based settings. This poses significant challenges in developing comparable models for other languages, including even those with large speaker populations such as Arabic. To alleviate this challenge, we introduce a comprehensive family of Arabic MLLMs, dubbed Peacock, with strong vision and language capabilities. Through comprehensive qualitative and quantitative analysis, we demonstrate the solid performance of our models on various visual reasoning tasks and further show their emerging dialectal potential. Additionally, we introduce Henna, a new benchmark specifically designed for assessing MLLMs on aspects related to Arabic culture, setting the first stone for culturally-aware Arabic MLLMs.

Henna Benchmark

This collection of images showcases a curated subset selected from Henna dataset, representing 11 Arab countries, and capturing the essence of traditional food, local customs, historical monuments, everyday activities, and distinctive architecture that characterize the diverse and rich heritage of each region. Henna Samples

Henna Dataset Generation

Dataset Generation Example using GPT-4V. This figure demonstrates the process of generating a question-answer dataset for an attraction in Yemen as an example. For each site, an image and its corresponding Wikipedia article were used to provide GPT-4V with rich contextual information. The model then generated ten contextually relevant questions and answers per image. Henna Pipeline

Evaluation results

example Comparison between the performance of Peacock models on SEED-Benchmark dimensions.

<p align="center"> <img src="assets/seed.png" alt="Seed-Bench" width="50%"/> </p>

Examples

example example

Citation

If you find this work useful for your research, please kindly cite our paper:

@inproceedings{alwajih-etal-2024-peacock,
    title = "Peacock: A Family of {A}rabic Multimodal Large Language Models and Benchmarks",
    author = "Alwajih, Fakhraddin  and
      Nagoudi, El Moatez Billah  and
      Bhatia, Gagan  and
      Mohamed, Abdelrahman  and
      Abdul-Mageed, Muhammad",
    editor = "Ku, Lun-Wei  and
      Martins, Andre  and
      Srikumar, Vivek",
    booktitle = "Proceedings of the 62nd Annual Meeting of the Association for Computational Linguistics (Volume 1: Long Papers)",
    month = aug,
    year = "2024",
    address = "Bangkok, Thailand",
    publisher = "Association for Computational Linguistics",
    url = "https://aclanthology.org/2024.acl-long.689",
    doi = "10.18653/v1/2024.acl-long.689",
    pages = "12753--12776"