Home

Awesome

<img src="./images/tasks.png" width="50%">

Benchmarks for Deep Off-Policy Evaluation

Contact: offline-rl@google.com

In D4RL and RL Unplugged: Benchmarks for Offline Reinforcement Learning, we released a suite of benchmarks for offline reinforcement learning. They are designed to facilitate ease of use, so we provided the datasets with a unified API which makes it easy for the practitioner to work with all data in the suite once a general pipeline has been established.

Here, we release policies which can be used in conjunction with the RL Unplugged and D4RL datasets to facilitate off-policy evaluation and offline model selection benchmarking.

In this release, we provide:

The policies are available under gs://gresearch/deep-ope, with the RL Unplugged policies in the subdirectory gs://gresearch/deep-ope/rlunplugged and the D4RL policies in the subdirectory gs://gresearch/deep-ope/d4rl.

Task Descriptions

DeepMind Locomotion Dataset

These tasks are made up of the corridor locomotion tasks involving the CMU Humanoid, for which prior efforts have either used motion capture data (Merel et al., 2019a, Merel et al., 2019b) or training from scratch (Song et al., 2020). In addition, the DM Locomotion repository contains a set of tasks adapted to be suited to a virtual rodent (Merel et al., 2020). We emphasize that the DM Locomotion tasks feature the combination of challenging high-DoF continuous control along with perception from rich egocentric observations. For details on how the dataset was generated, please refer to RL Unplugged: Benchmarks for Offline Reinforcement Learning.

DeepMind Control Suite Dataset

DeepMind Control Suite (Tassa et al., 2018) is a set of control tasks implemented in MuJoCo (Todorov et al., 2012). We consider a subset of the tasks provided in the suite that cover a wide range of difficulties.

Most of the datasets in this domain are generated using D4PG. For the environments Manipulator insert ball and Manipulator insert peg we use V-MPO (Song et al., 2020) to generate the data as D4PG is unable to solve these tasks. We release datasets for 9 control suite tasks. For details on how the dataset was generated, please refer to RL Unplugged: Benchmarks for Offline Reinforcement Learning.

D4RL Dataset

A subset of the tasks within the D4RL (Fu et. al. 2020) for offline reinforcement learning is included. These tasks include maze navigation with different robot morphologies, hand manipulation tasks (Rajeswaran et. al. 2017), and tasks from the OpenAI Gym bechmark (Brockman et. al. 2016).

Each task includes a variety of datasets in order to study the interaction between dataset distributions and policies. For further information on what datasets are available, please refer to D4RL: Datasets for Deep Data-Driven Reinforcement Learning.

Using the policies

The rlunplugged_policies.json file provides metadata about the policies in this dataset. It is structured as a list of dictionaries, one for each policy, where the keys contain metadata including:

The 'd4rl_policies.json' file contains metadata in a similar format:

Requirements:

Policy loading example

RLUnplugged policies are stored as TensorFlow SavedModels. Calling the policy on an observation returns an action sample. See load_rlunplugged_policy_example.py for an example of loading a policy.

D4RL policies are stored as pickle files containing weights. See load_d4rl_policy_example.py for an example of loading a policy.

Compute evaluation metrics

TODO Fill in example computing groundtruth and evaluation metrics.

Dataset Metadata

The following table is necessary for this dataset to be indexed by search engines such as <a href="https://g.co/datasetsearch">Google Dataset Search</a>.

<div itemscope itemtype="http://schema.org/Dataset"> <table> <tr> <th>property</th> <th>value</th> </tr> <tr> <td>name</td> <td><code itemprop="name">Benchmarks for Deep Off-Policy Evaluation</code></td> </tr> <tr> <td>url</td> <td><code itemprop="url">https://github.com/google-research/deep_ope</code></td> </tr> <tr> <td>sameAs</td> <td><code itemprop="sameAs">https://github.com/google-research/deep_ope</code></td> </tr> <tr> <td>description</td> <td><code itemprop="description"> Data accompanying [Benchmarks for Deep Off-Policy Evaluation](). </code></td> </tr> <tr> <td>provider</td> <td> <div itemscope itemtype="http://schema.org/Organization" itemprop="provider"> <table> <tr> <th>property</th> <th>value</th> </tr> <tr> <td>name</td> <td><code itemprop="name">Google</code></td> </tr> <tr> <td>sameAs</td> <td><code itemprop="sameAs">https://en.wikipedia.org/wiki/Google</code></td> </tr> </table> </div> </td> </tr> </table> </div>