Home

Awesome

Awesome Model-Based Reinforcement Learning

Awesome docs GitHub stars GitHub forks GitHub license

This is a collection of research papers for model-based reinforcement learning (mbrl). And the repository will be continuously updated to track the frontier of model-based rl.

Welcome to follow and star!

<pre name="code" class="html"> <font color="red">[2024.10.27] <b>New: We update the NeurIPS 2024 paper list of model-based rl!</b></font> [2024.05.20] We update the ICML 2024 paper list of model-based rl. [2023.11.29] We update the ICLR 2024 paper list of model-based rl. [2023.09.29] We update the NeurIPS 2023 paper list of model-based rl. [2023.06.15] We update the ICML 2023 paper list of model-based rl. [2023.02.05] We update the ICLR 2023 paper list of model-based rl. [2022.11.03] We update the NeurIPS 2022 paper list of model-based rl. [2022.07.06] We update the ICML 2022 paper list of model-based rl. [2022.02.13] We update the ICLR 2022 paper list of model-based rl. [2021.12.28] We release the awesome model-based rl. </pre>

Table of Contents

A Taxonomy of Model-Based RL Algorithms

We’ll start this section with a disclaimer: it’s really quite hard to draw an accurate, all-encompassing taxonomy of algorithms in the Model-Based RL space, because the modularity of algorithms is not well-represented by a tree structure. So we will publish a series of related blogs to explain more Model-Based RL algorithms.

<p align="center"> <img style="border-radius: 0.3125em; box-shadow: 0 2px 4px 0 rgba(34,36,38,.12),0 2px 10px 0 rgba(34,36,38,.08);" src="./assets/mbrl-taxonomy.png"> <br> <em style="display: inline-block;">A non-exhaustive, but useful taxonomy of algorithms in modern Model-Based RL.</em> </p>

We simply divide Model-Based RL into two categories: Learn the Model and Given the Model.

And we give some examples as shown in the figure above. There are links to algorithms in taxonomy.

[1] World Models: Ha and Schmidhuber, 2018
[2] I2A (Imagination-Augmented Agents): Weber et al, 2017
[3] MBMF (Model-Based RL with Model-Free Fine-Tuning): Nagabandi et al, 2017
[4] MBVE (Model-Based Value Expansion): Feinberg et al, 2018
[5] ExIt (Expert Iteration): Anthony et al, 2017
[6] AlphaZero: Silver et al, 2017
[7] POPLIN (Model-Based Policy Planning): Wang et al, 2019
[8] M2AC (Masked Model-based Actor-Critic): Pan et al, 2020

Papers

format:
- [title](paper link) [links]
  - author1, author2, and author3
  - Key: key problems and insights
  - OpenReview: optional
  - ExpEnv: experiment environments

Classic Model-Based RL Papers

<details open> <summary>Toggle</summary> </details>

NeurIPS 2024

<details open> <summary>Toggle</summary> <!--- [Parallelizing Model-based Reinforcement Learning Over the Sequence Length]() - Zirui Wang, Yue DENG, Junfeng Long, Yin Zhang - Key: - ExpEnv: - [Constrained Latent Action Policies for Model-Based Offline Reinforcement Learning]() - Marvin Alles, Philip Becker-Ehmck, Patrick van der Smagt, Maximilian Karl - Key: - ExpEnv: - [Policy-shaped prediction: avoiding distractions in model-based RL]() - Miles Hutson, Isaac Kauvar, Nick Haber - Key: - ExpEnv: --> </details>

ICML 2024

<details open> <summary>Toggle</summary> <!-- - [Trust the Model Where It Trusts Itself - Model-Based Actor-Critic with Uncertainty-Aware Rollout Adaption]() - Bernd Frauenknecht, Artur Eisele, Devdutt Subhasish, Friedrich Solowjow, Sebastian Trimpe - Key: - ExpEnv: - [Efficient World Models with Time-Aware and Context-Augmented Tokenization]() - Vincent Micheli, Eloi Alonso, François Fleuret - Key: - ExpEnv: - [Coprocessor Actor Critic: A Model-Based Reinforcement Learning Approach For Adaptive Deep Brain Stimulation]() - Michelle Pan, Mariah Schrum, Vivek Myers, Erdem Biyik, Anca Dragan - Key: - ExpEnv: --> </details>

ICLR 2024

<details open> <summary>Toggle</summary> </details>

NeurIPS 2023

<details open> <summary>Toggle</summary> </details>

ICML 2023

<details open> <summary>Toggle</summary> </details>

ICLR 2023

<details open> <summary>Toggle</summary> <!-- - [The Benefits of Model-Based Generalization in Reinforcement Learning](https://openreview.net/forum?id=w1w4dGJ4qV) - Kenny Young, Aditya Ramesh, Louis Kirsch, JĂĽrgen Schmidhuber - Key: model generalization can be considered more useful than value function generalization - OpenReview: 8, 6, 5, 5 - ExpEnv: [ProcMaze, ButtonGrid, PanFlute]() --> </details>

NeurIPS 2022

<details open> <summary>Toggle</summary> </details>

ICML 2022

<details open> <summary>Toggle</summary> </details>

ICLR 2022

<details open> <summary>Toggle</summary> </details>

NeurIPS 2021

<details open> <summary>Toggle</summary> </details>

ICLR 2021

<details open> <summary>Toggle</summary> </details>

ICML 2021

<details open> <summary>Toggle</summary> </details>

Other

Tutorial

Codebase

Contributing

Our purpose is to make this repo even better. If you are interested in contributing, please refer to HERE for instructions in contribution.

License

Awesome Model-Based RL is released under the Apache 2.0 license.

<p align="right">(<a href="#top">Back to top</a>)</p>