Home

Awesome

Domain Randomization-Enhanced Depth Simulation and Restoration for Perceiving and Grasping Specular and Transparent Objects (ECCV 2022)

This is the official repository of Domain Randomization-Enhanced Depth Simulation and Restoration for Perceiving and Grasping Specular and Transparent Objects.

For more information, please visit our project page.

Introduction

teaser

This paper investigates the problem of specular and transparent object depth simulation and restoration. We propose a system composed of a RGBD fusion network SwinDRNet for depth restoration, along with a synthetic data generation pipeline, Domain Randomization-Enhanced Depth Simulation, to generate the large-scale synthetic RGBD dataset, DREDS, that contains 130k photorealistic RGB images and simulated depths with realistic sensor noise. We also curate a real-world dataset, STD, that captures 30 cluttered scenes composed of 50 objects with various materials from specular, transparent, to diffuse. Experiments demonstrate that training on our simulated data, SwinDRNet can directly generalize to real RGBD images and significantly boosts the performance of perception and interaction tasks (e.g. category-level pose estimation, object grasping)

Overview

This repository provides:

Dataset

<!-- Please download the compressed files and store them from -->

DREDS dataset (simulated)

STD dataset (real)

CAD models

We provide the CAD models of our DREDS and STD dataset, including: 1,801 of DREDS-CatKnown (syn_train & syn_test), 42 of STD-CatKnown (real_cat_known), and 8 of STD-CatNovel (real_cat_known).

NOTE: The data is only for non-commercial use.

Citation

If you find our work useful in your research, please consider citing:

@inproceedings{dai2022dreds,
	title={Domain Randomization-Enhanced Depth Simulation and Restoration for Perceiving and Grasping Specular and Transparent Objects},
	author={Dai, Qiyu and Zhang, Jiyao and Li, Qiwei and Wu, Tianhao and Dong, Hao and Liu, Ziyuan and Tan, Ping and Wang, He},
	booktitle={European Conference on Computer Vision (ECCV)},
	year={2022}
    }

License

This work and the dataset are licensed under CC BY-NC 4.0.

CC BY-NC 4.0

Contact

If you have any questions, please open a github issue or contact us:

Qiyu Dai: qiyudai@pku.edu.cn, Jiyao Zhang: zhangjiyao@stu.xjtu.edu.cn, Qiwei Li: lqw@pku.edu.cn, He Wang: hewang@pku.edu.cn