Home

Awesome

An Adaptive Framework for Learning Unsupervised Depth Completion

PyTorch implementation of An Adaptive Framework for Learning Unsupervised Depth Completion

Project AdaFrame: Ada(ptive) Frame(work) for Depth Completion

Published in RA-L January 2021 and ICRA 2021

[publication]

Model have been tested on Ubuntu 16.04, 20.04 using Python 3.5, 3.6, PyTorch 1.2.0

Authors: Alex Wong, Xiaohan Fei

If this work is useful to you, please cite our paper:

@article{wong2021adaptive,
    title={An Adaptive Framework for Learning Unsupervised Depth Completion},
    author={Wong, Alex and Fei, Xiaohan and Hong, Byung-Woo and Soatto, Stefano},
    journal={IEEE Robotics and Automation Letters},
    volume={6},
    number={2},
    pages={3120--3127},
    year={2021},
    publisher={IEEE}
}

Table of Contents

  1. About sparse-to-dense depth completion
  2. About AdaFrame
  3. Related projects
  4. License and disclaimer

About sparse-to-dense depth completion <a name="about-sparse-to-dense"></a>

In the sparse-to-dense depth completion problem, we seek to infer the dense depth map of a 3-D scene using an RGB image and its associated sparse depth measurements in the form of a sparse depth map, obtained either from computational methods such as SfM (Strcuture-from-Motion) or active sensors such as lidar or structured light sensors.

RGB image from the VOID datasetOur densified depth map -- colored and backprojected to 3D
<img src="figures/void_teaser_image.png" width="400"><img src="figures/void_teaser_pointcloud.gif" width="400">
RGB image from the KITTI datasetOur densified depth map -- colored and backprojected to 3D
<img src="figures/kitti_teaser_image.png" width="400"><img src="figures/kitti_teaser_pointcloud.gif" width="400">

To follow the literature and benchmarks for this task, you may visit: Awesome State of Depth Completion

About AdaFrame <a name="about-adaframe"></a>

A number of computer vision problems can be formulated as an energy function which consists of the linear combination of a data fidelity (fitness to data) term and a regularizer (bias or prior). The data fidelity is weighted uniformly by a scalar α and the regularizer by γ that determine their relative significance.

However, uniform static α does not account for visibility phenomenon (occlusions) and uniform static γ does may impose too much or too little regularization. We propose an adaptive framework (α and γ) that consists of weighting schemes that vary spatially (image domain) and temporally (over training time) based on the residual or fitness of model to data.

α starts by weighting all pixel locations approximately uniformly and gradually downweights regions with high residual over time. α is conditioned on the mean or global residual, as the model become better fitted to the data, we become more confident that the high residual regions be results of occlusions yielding a sharper curve over time. Here is a visualization of α:

<p align="center"> <img align="center" src="figures/alpha_over_time.gif" width="400"> </p>

γ starts by imposing a low degree of regularization and only increase regularization where appropriate based on the fitness of the model to the data. Here is a visualization of γ:

<p align="center"> <img align="center" src="figures/gamma_over_time.gif" width="400"> </p>

We note that alpha and gamma are complementary. Ill-posed regions such as occlusions cannot be uniquely determined by the data and hence we need regularization. So, gamma increases regularization around ill-posed regions allowing neighboring point estimates that fit the data well to fill them in the gaps.

Source code coming soon!

We apologize for the delay, we want to release the PyTorch implementation in conjunction to an update to VOICED.

Related projects <a name="related-projects"></a>

You may also find the following projects useful:

We also have works in adversarial attacks on depth estimation methods:

License and disclaimer <a name="license-disclaimer"></a>

This software is property of the UC Regents, and is provided free of charge for research purposes only. It comes with no warranties, expressed or implied, according to these terms and conditions. For commercial use, please contact UCLA TDG.