scispace - formally typeset
Journal ArticleDOI

BundleFusion: real-time globally consistent 3D reconstruction using on-the-fly surface re-integration

Reads0
Chats0
TLDR
In this paper, a robust pose estimation strategy is proposed for real-time, high-quality, 3D scanning of large-scale scenes using RGB-D input with an efficient hierarchical approach, which removes heavy reliance on temporal tracking and continually localizes to the globally optimized frames instead.
Abstract
Real-time, high-quality, 3D scanning of large-scale scenes is key to mixed reality and robotic applications. However, scalability brings challenges of drift in pose estimation, introducing significant errors in the accumulated model. Approaches often require hours of offline processing to globally correct model errors. Recent online methods demonstrate compelling results but suffer from (1) needing minutes to perform online correction, preventing true real-time use; (2) brittle frame-to-frame (or frame-to-model) pose estimation, resulting in many tracking failures; or (3) supporting only unstructured point-based representations, which limit scan quality and applicability. We systematically address these issues with a novel, real-time, end-to-end reconstruction framework. At its core is a robust pose estimation strategy, optimizing per frame for a global set of camera poses by considering the complete history of RGB-D input with an efficient hierarchical approach. We remove the heavy reliance on temporal tracking and continually localize to the globally optimized frames instead. We contribute a parallelizable optimization framework, which employs correspondences based on sparse features and dense geometric and photometric matching. Our approach estimates globally optimized (i.e., bundle adjusted) poses in real time, supports robust tracking with recovery from gross tracking failures (i.e., relocalization), and re-estimates the 3D model in real time to ensure global consistency, all within a single framework. Our approach outperforms state-of-the-art online systems with quality on par to offline methods, but with unprecedented speed and scan completeness. Our framework leads to a comprehensive online scanning solution for large indoor environments, enabling ease of use and high-quality results.1

read more

Citations
More filters
Proceedings ArticleDOI

Rendering and Tracking the Directional TSDF: Modeling Surface Orientation for Coherent Maps

TL;DR: In this paper, the authors presented methods for rendering depth and color maps from the Directional Truncated Signed Distance Function (DTSDF), making it a true drop-in replacement for the regular TSDF.
Journal ArticleDOI

OwlFusion: Depth-Only Onboard Real-Time 3D Reconstruction of Scalable Scenes for Fast-Moving MAV

TL;DR: In this article , a dense SLAM system that combines surface hierarchical sparse representation and particle swarm pose optimization is proposed for onboard, depth-only, real-time 3D reconstruction capable of accommodating fast-moving vehicles.
Journal ArticleDOI

Self-distillation framework for indoor and outdoor monocular depth estimation

TL;DR: This work designs a student encoder that extracts features from two datasets of indoor and outdoor scenes, respectively, and introduces a dissimilarity loss to pull apart encoded features of different scenes in the feature space, and proposes a self-distillation MDE framework to improve the generalization ability across scenes.
Journal ArticleDOI

Local Homography Estimation on User-Specified Textureless Regions

TL;DR: This paper presents a novel deep neural network for designated point tracking (DPT) in a monocular RGB video, and shows that the algorithm outperforms the state-of-the-art approaches on ScanDPT.
Journal Article

Large-Scale 3D Semantic Reconstruction for Automated Driving Vehicles with Adaptive Truncated Signed Distance Function

TL;DR: A novel 3D reconstruction, texturing and semantic mapping system using LiDAR and camera sensors using an Adaptive Truncated Signed Distance Function and a Markov Random Field-based data fusion approach to estimate the optimal semantic class for each triangle mesh.
References
More filters
Journal ArticleDOI

Distinctive Image Features from Scale-Invariant Keypoints

TL;DR: This paper presents a method for extracting distinctive invariant features from images that can be used to perform reliable matching between different views of an object or scene and can robustly identify objects among clutter and occlusion while achieving near real-time performance.
Journal ArticleDOI

A method for registration of 3-D shapes

TL;DR: In this paper, the authors describe a general-purpose representation-independent method for the accurate and computationally efficient registration of 3D shapes including free-form curves and surfaces, based on the iterative closest point (ICP) algorithm, which requires only a procedure to find the closest point on a geometric entity to a given point.
Book

A Mathematical Introduction to Robotic Manipulation

TL;DR: In this paper, the authors present a detailed overview of the history of multifingered hands and dextrous manipulation, and present a mathematical model for steerable and non-driveable hands.
Book ChapterDOI

Indoor segmentation and support inference from RGBD images

TL;DR: The goal is to parse typical, often messy, indoor scenes into floor, walls, supporting surfaces, and object regions, and to recover support relationships, to better understand how 3D cues can best inform a structured 3D interpretation.
Proceedings ArticleDOI

KinectFusion: Real-time dense surface mapping and tracking

TL;DR: A system for accurate real-time mapping of complex and arbitrary indoor scenes in variable lighting conditions, using only a moving low-cost depth camera and commodity graphics hardware, which fuse all of the depth data streamed from a Kinect sensor into a single global implicit surface model of the observed scene in real- time.
Related Papers (5)