Home

Awesome

ODMS Dataset

ODMS is the first dataset for learning Object Depth via Motion and Segmentation. ODMS training data are configurable and extensible, with each training example consisting of a series of object segmentation masks, camera movement distances, and ground truth object depth. As a benchmark evaluation, we also provide four ODMS validation and test sets with 15,650 examples in multiple domains, including robotics and driving. In our paper, we use an ODMS-trained network to perform object depth estimation in real-time robot grasping experiments, demonstrating how ODMS is a viable tool for 3D perception from a single RGB camera.

(New) An object detection-based version of the ODMS benchmark is now available here!

Contact: Brent Griffin (griffb at umich dot edu)

Quick Introduction: https://youtu.be/c90Fg_whjpI

IMAGE ALT TEXT HERE

Using ODMS

Run ./demo/demo_datagen.py to generate random ODMS data to train your model. <br /> Example training data configurations are provided in the ./config/ folder. Has the option to save a static dataset. <br /> [native Python, has scipy dependency]

Run ./demo/demo_dataset_eval.py to evaluate your model on the ODMS validation and test sets. <br /> Provides an example evaluation for the VOS-DE baseline. Results are saved in the ./results/ folder. <br /> [native Python, VOS-DE baseline has skimage dependency]

Benchmark

MethodRobotDrivingNormalPerturbAll
DBox11.524.811.820.317.1
ODN<sub>lr</sub>13.131.78.617.917.8
Box<sub>LS</sub>17.633.313.736.625.3
VOS-DE32.636.07.933.627.5

Is your technique missing although it's published and the code is public? Let us know and we'll add it.

Using ODN Method

Run ./demo/demo_odn_train.py to train your own ODN model using ODMS. <br /> Run ./demo/demo_odn_eval.py after training to evaluate your ODN model. <br /> Example training and ODN model configurations are provided in the ./config/ folder. Models are saved in the ./results/model/ folder. <br /> [native Python, has Torch dependency]

Publication

Please cite our paper if you find it useful for your research.

@inproceedings{GrCoECCV20,
  author = {Griffin, Brent A. and Corso, Jason J.},
  booktitle={The European Conference on Computer Vision (ECCV)},
  title = {Learning Object Depth from Camera Motion and Video Object Segmentation},
  year = {2020}
}

ECCV 2020 Presentation: https://youtu.be/ZD4Y4oQbdks

IMAGE ALT TEXT HERE

Use

This code is available for non-commercial research purposes only.