scispace - formally typeset
Open AccessProceedings ArticleDOI

DPPTAM: Dense piecewise planar tracking and mapping from a monocular sequence

Alejo Concha, +1 more
- pp 5686-5693
Reads0
Chats0
TLDR
This paper proposes a direct monocular SLAM algorithm that estimates a dense reconstruction of a scene in real-time on a CPU, based on the information of a superpixel segmentation and the semidense map from highly textured areas.
Abstract
This paper proposes a direct monocular SLAM algorithm that estimates a dense reconstruction of a scene in real-time on a CPU. Highly textured image areas are mapped using standard direct mapping techniques [1], that minimize the photometric error across different views. We make the assumption that homogeneous-color regions belong to approximately planar areas. Our contribution is a new algorithm for the estimation of such planar areas, based on the information of a superpixel segmentation and the semidense map from highly textured areas. We compare our approach against several alternatives using the public TUM dataset [2] and additional live experiments with a hand-held camera. We demonstrate that our proposal for piecewise planar monocular SLAM is faster, more accurate and more robust than the piecewise planar baseline [3]. In addition, our experimental results show how the depth regularization of monocular maps can damage its accuracy, being the piecewise planar assumption a reasonable option in indoor scenarios.

read more

Content maybe subject to copyright    Report

DPPTAM: Dense Piecewise Planar Tracking and Mapping from a
Monocular Sequence
Alejo Concha and Javier Civera
I3A, Universidad de Zaragoza
{alejocb,jcivera}@unizar.es
(a) Semidense map
(b) Piecewise planar low-gradient regions
(c) Dense map
Fig. 1: Illustrative results of our demo. We estimate a semidense 3D map from a monocular sequence and reconstruct
low-gradient areas assuming they are piecewise planar.
Abstract Our demo is a direct monocular SLAM algorithm
that estimates a dense reconstruction of a scene in real-time on
a CPU. Highly textured image areas are mapped using standard
direct mapping techniques [1], that minimizes the photometric
error across different views. We make the assumption that
homogeneous-color regions belong to approximately planar
areas. Our contribution is a new algorithm for the estimation
of such planar areas, based on the information of a superpixel
segmentation and the semidense map from highly textured
areas.
I. INTRODUCTION
One of the key pieces of any virtual or augmented reality
system is the 3D estimation of the surrounding scene and the
pose of the device from sensing data, sequentially and in real-
time. This is also an essential component of an autonomous
robots and has been usually denoted with the acronym SLAM
–Simultaneous Localization and Mapping. The monocular
camera stands out as one of the most convenient sensors for
several reasons.
One of the hardest challenges in monocular SLAM is the
estimation of a fully dense map of the imaged scene. Pixels
in textureless areas cannot be reliably matched across views
and standard 3D reconstructions from monocular SLAM are
limited to areas of high photometric gradients.
Our research starts in [2], [3] modelling the environment
with 3D points for high-gradient areas and 3D planes for low-
gradient areas. The assumption made is that image areas with
low color gradients are mostly planar; which is met in most
indoors and man-made scenes. Low-gradient image areas are
segmented using superpixels.
II. OVERVIEW
In our approach, the camera is tracked in real time at video
frequency by minimizing the photometric error between the
high-gradient pixels of the current frame and the reprojection
of the corresponding map points.
A semidense map is estimated from a sparse set of selected
keyframes. This map is used to register the current camera
in a global reference frame; and hence it should be estimated
at a high rate.
Finally, a dense map is estimated from the same set of
keyframes but at a slower rate. This dense map can be
used for realistic augmentation or robotic navigation. The
regularization that produces fully dense maps can be very
demanding and a GPU is needed to do it in real-time, limiting
its use to high-end devices. Our proposal is to leverage scene
priors, specifically the Manhattan and piecewise planar struc-
tures in man-made scenes, to reduce the complexity of the
map estimation. Some illustrative results of our algorithms
can be seen in figure 1. The maps in this figure have been
estimated in real-time in a CPU. The results can be better
appreciated in the video of the footnote link
1
.
ACKNOWLEDGMENT
This research was funded by the Spanish government with
the projects IPT-2012-1309-430000 and DPI2012-32168
REFERENCES
[1] J. Engel, T. Sch
¨
ops, and D. Cremers, “LSD-SLAM: Large-scale direct
monocular slam, in Computer Vision–ECCV 2014. Springer, 2014,
pp. 834–849.
[2] A. Concha and J. Civera, “Using superpixels in monocular SLAM,
in IEEE International Conference on Robotics and Automation, Hong
Kong, June 2014.
[3] A. Concha, W. Hussain, L. Montano, and J. Civera, “Manhattan
and piecewise-planar constraints for dense monocular mapping, in
Robotics:Science and Systems, 2014.
1
http://webdiis.unizar.es/
˜
jcivera/videos/
iros15submission.mp4
Citations
More filters
Journal ArticleDOI

DynaSLAM: Tracking, Mapping, and Inpainting in Dynamic Scenes

TL;DR: DynaSLAM is a visual SLAM system that, building on ORB-SLAM2, adds the capabilities of dynamic object detection and background inpainting, and outperforms the accuracy of standard visualSLAM baselines in highly dynamic scenarios.
Journal ArticleDOI

DynaSLAM: Tracking, Mapping and Inpainting in Dynamic Scenes

TL;DR: DynaSLAM as mentioned in this paper is a visual SLAM system that, building over ORB-SLAM2, adds the capabilities of dynamic object detection and background inpainting, and it is robust in dynamic scenarios for monocular, stereo and RGB-D configurations.
Journal ArticleDOI

Keyframe-based monocular SLAM

TL;DR: This paper presents a survey that covers the various keyframe-based monocular SLAM systems in the literature, detailing the components of their implementation, and critically assessing the specific strategies made in each proposed solution.
Proceedings ArticleDOI

Pop-up SLAM: Semantic monocular plane SLAM for low-texture environments

TL;DR: In this article, a real-time monocular plane SLAM algorithm was proposed to improve the robustness of low-texture environments by combining planes with point-based SLAM.
Proceedings ArticleDOI

Comparison of Various SLAM Systems for Mobile Robot in an Indoor Environment

TL;DR: A comparative analysis of a mobile robot trajectories computed by various ROS-based SLAM systems is presented, demonstrating encouraging results for lidar-based Cartographer SLAM, Monocular ORB SLAM and Stereo RTAB Map methods.
References
More filters
Journal ArticleDOI

Efficient Graph-Based Image Segmentation

TL;DR: An efficient segmentation algorithm is developed based on a predicate for measuring the evidence for a boundary between two regions using a graph-based representation of the image and it is shown that although this algorithm makes greedy decisions it produces segmentations that satisfy global properties.
Proceedings ArticleDOI

Parallel Tracking and Mapping for Small AR Workspaces

TL;DR: A system specifically designed to track a hand-held camera in a small AR workspace, processed in parallel threads on a dual-core computer, that produces detailed maps with thousands of landmarks which can be tracked at frame-rate with accuracy and robustness rivalling that of state-of-the-art model-based systems.
Journal ArticleDOI

MonoSLAM: Real-Time Single Camera SLAM

TL;DR: The first successful application of the SLAM methodology from mobile robotics to the "pure vision" domain of a single uncontrolled camera, achieving real time but drift-free performance inaccessible to structure from motion approaches is presented.
Book ChapterDOI

LSD-SLAM: Large-Scale Direct Monocular SLAM

TL;DR: A novel direct tracking method which operates on \(\mathfrak{sim}(3)\), thereby explicitly detecting scale-drift, and an elegant probabilistic solution to include the effect of noisy depth values into tracking are introduced.
Journal ArticleDOI

Lucas-Kanade 20 Years On: A Unifying Framework

TL;DR: In this paper, a wide variety of extensions have been made to the original formulation of the Lucas-Kanade algorithm and their extensions can be used with the inverse compositional algorithm without any significant loss of efficiency.
Related Papers (5)
Frequently Asked Questions (4)
Q1. What are the contributions in "Dpptam: dense piecewise planar tracking and mapping from a monocular sequence" ?

In this paper, a direct monocular SLAM algorithm that estimates a dense reconstruction of a scene in real-time on a CPU is presented. 

Their proposal is to leverage scene priors, specifically the Manhattan and piecewise planar structures in man-made scenes, to reduce the complexity of the map estimation. 

In their approach, the camera is tracked in real time at video frequency by minimizing the photometric error between thehigh-gradient pixels of the current frame and the reprojection of the corresponding map points. 

The regularization that produces fully dense maps can be very demanding and a GPU is needed to do it in real-time, limiting its use to high-end devices.