Home

Awesome

RANS_LOGO

About this Repository

This repository serves as an extension to the OmniIsaacGymEnvs framework, enhancing its capabilities with additional robots and advanced features. The primary goal is to provide a more flexible environment for deploying various code-generated robotic and performing complex navigation tasks.

Key Features

<head> <style> .responsive-img { max-height: 100px; /* Adjust this value to your desired height */ width: 200px; } </style> </head>

Robots Included

2D Satellite3D SatelliteHeron USVTurtle-botsHusky
<img src="data/tasks_imgs/3dof_gotoxy.png" alt="2D Satellite" class="responsive-img"><img src="omniisaacgymenvs/images/6DofGoToPose.png" alt="3D Satellite" class="responsive-img"/><img src="data/robots_imgs/usv_boat.png" alt="Heron USV" class="responsive-img"/><img src="data/robots_imgs/turtlebot2e.png" alt="Turtle-bots" width=130/><img src="data/robots_imgs/husky_car.png" alt="Husky" class="responsive-img"/>

Navigation Tasks Available

This library provides a set of predefined navigation tasks for robotic control and reinforcement learning. It allows for easy extensions to add new tasks or modify existing ones to suit different requirements.

Tasks list

<details> <summary><h3 style="display: inline;">Click to Open Task Descriptions</h3></summary>
TaskObjective and ChallengeImage
GoToPositionNavigate to a specific target position. Accurately reach position within a given distance.<img src="data/tasks_imgs/3dof_gotoxy.png" alt="GoToPosition Image" class="responsive-img"/>
GoToPoseReach a specific target position and orientation. Manage both position and heading errors, within a given distance.<img src="data/tasks_imgs/3dof_pose.png" alt="GoToPose Image" class="responsive-img"/>
Track Linear VelocityAchieve and maintain a specific linear velocity. Precisely track the target velocity.<img src="path_to_image.png" alt="TrackLinearVelocity Image" class="responsive-img"/>
Track Linear & Angular VelocityAchieve and maintain specific linear and angular velocities. Manage both linear and angular velocity tracking.<img src="path_to_image.png" alt="TrackLinearAngularVelocity Image" class="responsive-img"/>
Track Linear Velocity & HeadingAchieve and maintain a specific linear velocity while following a heading. Track linear velocity and heading simultaneously.<img src="path_to_image.png" alt="TrackLinearVelocityHeading Image" class="responsive-img"/>
GoThroughPosition / SequencePass through a specific position or sequence while continuing the trajectory. Ensure precise traversal within a given distance (0.1cm default).<img src="data/tasks_imgs/position_seq.png" alt="GoThroughPosition Image" class="responsive-img"/>
GoThroughPose / SequencePass through a target position and orientation or sequence. Achieve accuracy in both position and heading within a given distance (0.1cm default).<img src="data/tasks_imgs/pose_seq.png" alt="GoThroughPose Image" class="responsive-img"/>
GoThroughGate / SequencePass through a gate or a series of gates in a specific direction without touching them. Maintain accuracy with a tolerance of 1m (default) and avoid crossing in the wrong direction.<img src="data/tasks_imgs/gate_seq.png" alt="GoThroughGate Image" class="responsive-img"/>
</details>

Installation

Native

Follow the Isaac Sim documentation to install the latest Isaac Sim release.

Examples in this repository rely on features from the most recent Isaac Sim release. Please make sure to update any existing Isaac Sim build to the latest release version, 2023.1.1, to ensure examples work as expected.

[!IMPORTANT] Make sure Isaac sim was installed localy. Locate it's python executable path, it can usually be found here: ~/.local/share/ov/pkg/isaac-sim-2023.1.1/python.sh or here: ~/.local/share/ov/pkg/isaac-sim-2023.1.1/python.sh

Clone this repository:

git clone https://github.com/elharirymatteo/RANS.git
cd RANS

[!CAUTION] In the following we refer to python.sh as the full path to the python executable of isaac sim. Make sure you use the full path.

python.sh -m pip install -e .

We use a modified version of the rl-games library to train our agents.

To install it clone this repository INSIDE RANS:

git clone https://github.com/AntoineRichard/rl_games

Then install the module:

cd rl_games
python.sh -m pip install --upgrade pip
pythons.sh -m pip install -e .

With these steps done you should be all set! Refer to the Getting Started section to learn how to launch your first trainings.

Docker

Before we install the simulation, please follow the procedure here to install all the required components to install IsaacSim in a docker container.

[!TIP] You will need an nvcr.io account.

Once you're all set, clone this repository as well as our own version of RL_games.

git clone https://github.com/elharirymatteo/RANS.git
cd RANS
git clone https://github.com/AntoineRichard/rl_games

[!IMPORTANT] RL games must be cloned inside the RANS repository.

One this is done, the docker image can be built by calling the build script.

./docker/build_docker.sh

With this out of the way you should be off to the races! Check the Getting Started section to start your first training.

Getting Started

[!IMPORTANT] If you are using docker you can start a new docker container by running the docker/run_docker.sh or docker/run_docker_viewer.sh. The viewer version allows the user to visualize the environments.

[!Note] All commands should be executed from RANS/omniisaacgymenvs.* python.sh referes to the python.sh inside Isaac's sim folder. In docker this will be at /isaac-sim/python.sh.

Training new agents

To train your first policy, (example for the USV robot) run:

python.sh scripts/rlgames_train_RANS.py task=ASV/GoToPose train=RANS/PPOcontinuous_MLP headless=True num_envs=1024

Modify num_envs appropriately to scale with your current machine capabilities. Turn headless to False if you want to visualize the envs while training occurs.

You should see an Isaac Sim window pop up. Once Isaac Sim initialization completes, the scene for the selected robot will be constructed and simulation will start running automatically. The process will terminate once training finishes.

Here's another example - GoToPose for the Satellite robot (MFP - modular floating platform) - using the multi-threaded training script:

python.sh scripts/rlgames_train_RANS.py task=MFP2D/GoToPose train=RANS/PPOmulti_discrete_MLP

Note that by default, we show a Viewport window with rendering, which slows down training. You can choose to close the Viewport window during training for better performance. The Viewport window can be re-enabled by selecting Window > Viewport from the top menu bar.

To achieve maximum performance, launch training in headless mode as follows:

python.sh scripts/rlgames_train_RANS.py task=MFP2D/GoToPose train=PPOmulti_discrete_MLP headless=True

[!NOTE] Some of the examples could take a few minutes to load because the startup time scales based on the number of environments. The startup time will continually be optimized in future releases.

Loading trained models (or checkpoints)

Checkpoints are saved in the folder runs/EXPERIMENT_NAME/nn where EXPERIMENT_NAME defaults to the task name, but can also be overridden via the experiment argument.

To load a trained checkpoint and continue training, use the checkpoint argument:

python.sh scripts/rlgames_train_RANS.py task=MFP2D/GoToPose train=RANS/PPOmulti_discrete_MLP checkpoint=runs/MFP2D_GoToPose/nn/MFP2D_GoToPose.pth

To load a trained checkpoint and only perform inference (no training), pass test=True as an argument, along with the checkpoint name. To avoid rendering overhead, you may also want to run with fewer environments using num_envs=64:

python.sh scripts/rlgames_train_RANS.py task=MFP2D/GoToPose train=RANS/PPOmulti_discrete_MLP checkpoint=runs/MFP2D_GoToPose/nn/MFP2D_GoToPose.pth test=True num_envs=64

Note that if there are special characters such as [ or = in the checkpoint names, you will need to escape them and put quotes around the string. For example, checkpoint="runs/Ant/nn/last_Antep\=501rew\[5981.31\].pth"

Going further

All scripts provided in omniisaacgymenvs/scripts can be launched directly with PYTHON_PATH.

<details> <summary><span style="font-size: 1.3em; font-weight: bold;">Random policy</span></summary> To test out a task without RL in the loop, run the random policy script with:
python.sh scripts/random_policy.py task=MFP2D/GoToPose

This script will sample random actions from the action space and apply these actions to your task without running any RL policies. Simulation should start automatically after launching the script, and will run indefinitely until terminated.

</details> <details> <summary><span style="font-size: 1.3em; font-weight: bold;">Train on single GPU</span></summary> To run a simple form of PPO from `rl_games`, use the single-threaded training script:
python.sh scripts/rlgames_train_RANS.py task=MFP2D/GoToPosition

This script creates an instance of the PPO runner in rl_games and automatically launches training and simulation. Once training completes (the total number of iterations have been reached), the script will exit. If running inference with test=True checkpoint=<path/to/checkpoint>, the script will run indefinitely until terminated. Note that this script will have limitations on interaction with the UI.

</details> <details> <summary><span style="font-size: 1.3em; font-weight: bold;">Train on multiple GPUs</span></summary> TBD </details> <details> <summary><span style="font-size: 1.3em; font-weight: bold;">Configuration and command line arguments</span></summary>

We use Hydra to manage the config.

Common arguments for the training scripts are:

Hydra also allows setting variables inside config files directly as command line arguments. For example, to set the minibatch size for an rl_games training run, you can use train.params.config.minibatch_size=64. Similarly, variables in task configs can also be set, such as task.env.episodeLength=100.

Hydra Notes

Default values for each of these are found in the omniisaacgymenvs/cfg/config.yaml file.

The way that the task and train portions of the config works are through the use of config groups. You can learn more about how these work here The actual configs for task are in omniisaacgymenvs/cfg/task/<TASK>.yaml and for train in omniisaacgymenvs/cfg/train/<TASK>PPO.yaml.

In some places in the config you will find other variables referenced (for example, num_actors: ${....task.env.numEnvs}). Each . represents going one level up in the config hierarchy. This is documented fully here.

</details>

Tensorboard

Tensorboard can be launched during training via the following command:

python.sh -m tensorboard.main --logdir runs/EXPERIMENT_NAME/summaries

WandB support

You can run (WandB)[https://wandb.ai/] with OmniIsaacGymEnvs by setting wandb_activate=True flag from the command line. You can set the group, name, entity, and project for the run by setting the wandb_group, wandb_name, wandb_entity and wandb_project arguments. Make sure you have WandB installed in the Isaac Sim Python executable with PYTHON_PATH -m pip install wandb before activating.

Citation

If you use the current repository in your work, we suggest citing the following papers:

@article{el2023drift,
  title={DRIFT: Deep Reinforcement Learning for Intelligent Floating Platforms Trajectories},
  author={El-Hariry, Matteo and Richard, Antoine and Muralidharan, Vivek and Yalcin, Baris Can and Geist, Matthieu and Olivares-Mendez, Miguel},
  journal={arXiv preprint arXiv:2310.04266},
  year={2023}
}

@article{el2023rans,
  title={RANS: Highly-Parallelised Simulator for Reinforcement Learning based Autonomous Navigating Spacecrafts},
  author={El-Hariry, Matteo and Richard, Antoine and Olivares-Mendez, Miguel},
  journal={arXiv preprint arXiv:2310.07393},
  year={2023}
}

Directory Structure (most relevant folders shown)

.
├── cfg                                 # Configuration files
│   ├── controller                      # Controller configurations
│   └──  hl_task                         # High-level task configurations convergence task
│   └── train                           # Training configurations
│       └── MFP                         # Training configurations for Modular Floating Platform
├── demos                               # Demonstration files (e.g., gifs, videos)
├── doc                                 # Documentation files
│   ├── curriculum.md                   # Documentation for curriculum
│   ├── domain_randomization.md         # Documentation for domain randomization
│   ├── figures                         # Figures used in documentation
│   │   └── ...                         # Other figure files
│   └── penalties.md                    # Documentation for penalties
├── envs                                # Environment scripts
│   ├── vec_env_rlgames_mfp.py          # Vectorized environment for rlgames with MFP
│   ├── vec_env_rlgames_mt.py           # Vectorized environment for rlgames with multiple tasks
│   └── vec_env_rlgames.py              # General vectorized environment for rlgames
├── extension.py                        # Extension script
├── images                              # Image files
│   ├── 3dof_gotoxy.png                 # Image for 3DOF GoToXY task
│   └── ...                             # Other image files
├── __init__.py                         # Initialization script for the package
├── lab_tests                           # Lab test scripts and data
├── mj_runs                             # Mujoco run scripts and data
├── models                              # Model files
├── robots                              # Robot related files
│   ├── articulations                   # Articulation files for robots
│   ├── sensors                         # Sensor files for robots
│   └── usd                             # USD files for robots
├── ros                                 # ROS related files
├── scripts                             # Utility scripts
├── tasks                               # Task implementations
│   └── MFP                             # Task implementations for Modular Floating Platform
│       ├── curriculum_helpers.py       # Helper functions for curriculum
│       └── unit_tests                  # Unit tests for MFP tasks
├── utils                               # Utility functions and scripts
│   ├── aggregate_and_eval_mujoco_batch_data.py  # Script to aggregate and evaluate Mujoco batch data
│   ├── rlgames                         # RL games related utilities
│   │   ├── __pycache__                 # Compiled Python files
│   │   ├── rlgames_train_mt.py         # Training script for RL games with multiple tasks
│   │   └── rlgames_utils.py            # Utility functions for RL games
├── videos                              # Video files
└── wandb                               # Weights and Biases integration files