scispace - formally typeset
Open AccessProceedings ArticleDOI

Sparse and Noisy LiDAR Completion with RGB Guidance and Uncertainty

TLDR
In this article, the authors proposed a new depth completion framework which extracts both global and local information in order to produce proper depth maps, and further proposed a fusion method with RGB guidance from a monocular camera to leverage object information and to correct mistakes in the sparse input.
Abstract
This work proposes a new method to accurately complete sparse LiDAR maps guided by RGB images. For autonomous vehicles and robotics the use of LiDAR is indispensable in order to achieve precise depth predictions. A multitude of applications depend on the awareness of their surroundings, and use depth cues to reason and react accordingly. On the one hand, monocular depth prediction methods fail to generate absolute and precise depth maps. On the other hand, stereoscopic approaches are still significantly outperformed by LiDAR based approaches. The goal of the depth completion task is to generate dense depth predictions from sparse and irregular point clouds which are mapped to a 2D plane. We propose a new framework which extracts both global and local information in order to produce proper depth maps. We argue that simple depth completion does not require a deep network. However, we additionally propose a fusion method with RGB guidance from a monocular camera in order to leverage object information and to correct mistakes in the sparse input. This improves the accuracy significantly. Moreover, confidence masks are exploited in order to take into account the uncertainty in the depth predictions from each modality. This fusion method outperforms the state-of-the-art and ranks first on the KITTI depth completion benchmark [21]. Our code with visualizations is available at https://github.com/wvangansbeke/Sparse-Depth-Completion.

read more

Citations
More filters
Journal ArticleDOI

Deep Learning for Image and Point Cloud Fusion in Autonomous Driving: A Review

TL;DR: A review of recent deep-learning-based data fusion approaches that leverage both image and point cloud data processing and identifies gaps and over-looked challenges between current academic researches and real-world applications.
Journal ArticleDOI

Learning Guided Convolutional Network for Depth Completion

TL;DR: Inspired by the guided image filtering, a novel guided network is designed to predict kernel weights from the guidance image and these predicted kernels are then applied to extract the depth image features.
Proceedings ArticleDOI

PENet: Towards Precise and Efficient Image Guided Depth Completion

TL;DR: PENet-ICRA2021 as mentioned in this paper proposes a two-branch backbone that consists of a color-dominant branch to exploit and fuse two modalities thoroughly, and a simple geometric convolutional layer to encode 3D geometric cues.
Proceedings ArticleDOI

Learning Joint 2D-3D Representations for Depth Completion

TL;DR: This paper designs a simple yet effective neural network block that learns to extract joint 2D and 3D features from RGBD data and shows that it outperforms the state-of-the-art on the challenging KITTI depth completion benchmark.
References
More filters
Proceedings ArticleDOI

Deep Residual Learning for Image Recognition

TL;DR: In this article, the authors proposed a residual learning framework to ease the training of networks that are substantially deeper than those used previously, which won the 1st place on the ILSVRC 2015 classification task.
Proceedings Article

Adam: A Method for Stochastic Optimization

TL;DR: This work introduces Adam, an algorithm for first-order gradient-based optimization of stochastic objective functions, based on adaptive estimates of lower-order moments, and provides a regret bound on the convergence rate that is comparable to the best known results under the online convex optimization framework.
Proceedings Article

Batch Normalization: Accelerating Deep Network Training by Reducing Internal Covariate Shift

TL;DR: Applied to a state-of-the-art image classification model, Batch Normalization achieves the same accuracy with 14 times fewer training steps, and beats the original model by a significant margin.
Proceedings ArticleDOI

Focal Loss for Dense Object Detection

TL;DR: This paper proposes to address the extreme foreground-background class imbalance encountered during training of dense detectors by reshaping the standard cross entropy loss such that it down-weights the loss assigned to well-classified examples, and develops a novel Focal Loss, which focuses training on a sparse set of hard examples and prevents the vast number of easy negatives from overwhelming the detector during training.
Related Papers (5)