Home

Awesome

Uni-NLX: Unifying Textual Explanations for Vision and Vision-Language Tasks

<br> <br> <p align="center"> <img src="demo_uninlx.png" width="784"/> </p>

[arXiv]<br> [video presentation at ICCV]

Requirements

Images Download

Data

The trianing and test data (combined for all datasets) can be found here

Annotations

The annotations in the format that cococaption expects can be found here. Please place them inside the cococaption folder.

Code

train_nlx.py: script for training only<br> test_datasets.py: script for validation/testing for all epochs on all 7 NLE tasks<br> clip_model.py: script for vision backbone we use (CLIP visual encoder)<br>

Models

Results