Home

Awesome

:houses: Airbert: In-domain Pretraining for Vision-and-Language Navigation :houses:

MIT arXiv R2R 1st ICCV 2021 website

This repository stores some models trained in our experiments.

:keyboard: Downloading from the command line

We stored our models on Google Drive, as the service provides a storage of 15 GB for free.

You can use the great gdown script for downloading the models:

pip install gdown
gdown [link to Google Drive]

We also provide a Makefile to help you:

# Download everything
make all 
# Download a specific model
make airbert-r2rRSA
# Get all commands
make help

:houses: Model pretrained on the BNB dataset

ModelDescription
airbertAirbert model pretrained on the BNB dataset

:alien: External models used in our scripts

ModelDescription
vilbertViLBERT model pretrained on Conceptual Captions
vlnbertVLN-BERT: vilbert fine-tuned on R2R

:robot: Finetuned models in discriminative setting

ModelDescription
airbert-r2rRSairbert fine-tuned on R2R with the shuffling loss
airbert-r2rRSAairbert fine-tuned on R2R with the shuffling loss + speaker data

:robot: Finetuned models in generative setting

ModelDescription
REVERIERecurrent VLN-BERT for remote referring expression with pretrained Airbert as backbone
R2RRecurrent VLN-BERT for vision-and-language navigation with pretrained Airbert as backbone