scispace - formally typeset
Open AccessProceedings ArticleDOI

Detection-based object labeling in 3D scenes

TLDR
This work utilizes sliding window detectors trained from object views to assign class probabilities to pixels in every RGB-D frame, and performs efficient inference on a Markov Random Field over the voxels, combining cues from view-based detection and 3D shape, to label the scene.
Abstract
We propose a view-based approach for labeling objects in 3D scenes reconstructed from RGB-D (color+depth) videos. We utilize sliding window detectors trained from object views to assign class probabilities to pixels in every RGB-D frame. These probabilities are projected into the reconstructed 3D scene and integrated using a voxel representation. We perform efficient inference on a Markov Random Field over the voxels, combining cues from view-based detection and 3D shape, to label the scene. Our detection-based approach produces accurate scene labeling on the RGB-D Scenes Dataset and improves the robustness of object detection.

read more

Content maybe subject to copyright    Report

Citations
More filters
Proceedings ArticleDOI

Real-time grasp detection using convolutional neural networks

TL;DR: An accurate, real-time approach to robotic grasp detection based on convolutional neural networks that outperforms state-of-the-art approaches by 14 percentage points and runs at 13 frames per second on a GPU.
Proceedings ArticleDOI

Multimodal deep learning for robust RGB-D object recognition

TL;DR: This paper leverages recent progress on Convolutional Neural Networks (CNNs) and proposes a novel RGB-D architecture for object recognition that is composed of two separate CNN processing streams - one for each modality - which are consecutively combined with a late fusion network.
Book ChapterDOI

Sliding Shapes for 3D Object Detection in Depth Images

TL;DR: This paper proposes to use depth maps for object detection and design a 3D detector to overcome the major difficulties for recognition, namely the variations of texture, illumination, shape, viewpoint, clutter, occlusion, self-occlusion and sensor noises.
Posted Content

Real-Time Grasp Detection Using Convolutional Neural Networks

TL;DR: In this paper, a convolutional neural network (CNN) is used for robotic grasp detection, which performs single-stage regression to graspable bounding boxes without using standard sliding window or region proposal techniques.
Proceedings ArticleDOI

Unsupervised feature learning for 3D scene labeling

TL;DR: This paper presents an approach for labeling objects in 3D scenes that combines features learned from raw RGB-D images and 3D point clouds directly, without any hand-designed features, to assign an object label to every3D point in the scene.
References
More filters
Proceedings ArticleDOI

ImageNet: A large-scale hierarchical image database

TL;DR: A new database called “ImageNet” is introduced, a large-scale ontology of images built upon the backbone of the WordNet structure, much larger in scale and diversity and much more accurate than the current image datasets.
Proceedings ArticleDOI

Histograms of oriented gradients for human detection

TL;DR: It is shown experimentally that grids of histograms of oriented gradient (HOG) descriptors significantly outperform existing feature sets for human detection, and the influence of each stage of the computation on performance is studied.
Journal ArticleDOI

Object Detection with Discriminatively Trained Part-Based Models

TL;DR: An object detection system based on mixtures of multiscale deformable part models that is able to represent highly variable object classes and achieves state-of-the-art results in the PASCAL object detection challenges is described.
Journal ArticleDOI

Fast approximate energy minimization via graph cuts

TL;DR: This work presents two algorithms based on graph cuts that efficiently find a local minimum with respect to two types of large moves, namely expansion moves and swap moves that allow important cases of discontinuity preserving energies.
Book

Probabilistic Robotics

TL;DR: This research presents a novel approach to planning and navigation algorithms that exploit statistics gleaned from uncertain, imperfect real-world environments to guide robots toward their goals and around obstacles.
Related Papers (5)