Home

Awesome

LEXA Benchmark

Codebase for the self-supervised goal reaching benchmark introduced in the LEXA paper (Discovering and Achieving Goals via World Models, NeurIPS 2021).

<img src="https://russellmendonca.github.io/data/lexa_benchmark.png" width="1000">

Setup

Create the conda environment by running : conda env create -f environment.yml

Alternatively, you can update an existing conda environment by running : conda env update -f environment.yml

Modify the python path
export PYTHONPATH=<path to lexa-benchmark>

Export the following variables for rendering
export MUJOCO_RENDERER=egl; export MUJOCO_GL=egl

Please follow these instructions to install mujoco

Usage

WARNING! The success criteria defined in this benchmark should not be used as a reward for the agent. The criteria were tuned to roughly match human intuition, but produce some false positives, which the agent can exploit if it observes the success variable.

Bibtex

If you find this code useful, please cite:

@misc{lexa2021,
    title={Discovering and Achieving Goals via World Models},
    author={Mendonca, Russell and Rybkin, Oleh and
    Daniilidis, Kostas and Hafner, Danijar and Pathak, Deepak},
    year={2021},
    Booktitle={NeurIPS}
}

Acknowledgements

This benchmark is built on top of the following environments: Adept, MetaWorld, and DeepMind Control Suite.