scispace - formally typeset
Proceedings ArticleDOI

Towards shape-based visual object categorization for humanoid robots

Reads0
Chats0
TLDR
A shape model-based approach using stereo vision and machine learning for object categorization is introduced allowing proper categorization of unknown objects even when object appearance and shape substantially differ from the training set.
Abstract
Humanoid robots should be able to grasp and handle objects in the environment, even if the objects are seen for the first time. A plausible solution to this problem is to categorize these objects into existing classes with associated actions and functional knowledge. So far, efforts on visual object categorization using humanoid robots have either been focused on appearance-based methods or have been restricted to object recognition without generalization capabilities. In this work, a shape model-based approach using stereo vision and machine learning for object categorization is introduced. The state-of-the-art features for shape matching and shape retrieval were evaluated and selectively transfered into the visual categorization. Visual sensing from different vantage points allows the reconstruction of 3D mesh models of the objects found in the scene by exploiting knowledge about the environment for model-based segmentation and registration. These reconstructed 3D mesh models were used for shape feature extraction for categorization and provide sufficient information for grasping and manipulation. Finally, the visual categorization was successfully performed with a variety of features and classifiers allowing proper categorization of unknown objects even when object appearance and shape substantially differ from the training set. Experimental evaluation with the humanoid robot ARMAR-IIIa is presented.

read more

Content maybe subject to copyright    Report

Citations
More filters
Journal ArticleDOI

Task-based Grasp Adaptation on a Humanoid Robot

TL;DR: This paper presents an approach towards autonomous grasping of objects according to their category and a given task, which allows to transfer task-specific grasp experience between objects of the same category.
Journal ArticleDOI

Robot semantic mapping through human activity recognition

TL;DR: This work presents a new robot semantic mapping approach through human activity recognition in a human-robot coexisting environment that is able to create a semantic map of an indoor environment effectively.
Proceedings ArticleDOI

Object categorization and grasping by parts from range scan data

TL;DR: This work proposes a novel approach for object categorization and grasping that is focused on topological shape segmentation that allows generation of watertight triangulated models of the objects and theirshape segmentation into parts.
Proceedings ArticleDOI

Shape distributions on voxel surfaces for 3D object classification from depth images

TL;DR: This work addresses the problem of 3D shape based object class recognition directly from point cloud data obtained from RGB-D cameras like the Kinect sensor from Microsoft, and presents a novel shape descriptor, capable of classifying 'never before seen objects' at their first occurrence in a single view in a fast and robust manner.
Proceedings ArticleDOI

Improving robot vision models for object detection through interaction

TL;DR: In this article, a method for learning specific object representations that can be applied (and reused) in visual detection and identification tasks is proposed. But this method relies on the use of Cartesian Genetic Programming (CGP).
References
More filters
Journal ArticleDOI

Distinctive Image Features from Scale-Invariant Keypoints

TL;DR: This paper presents a method for extracting distinctive invariant features from images that can be used to perform reliable matching between different views of an object or scene and can robustly identify objects among clutter and occlusion while achieving near real-time performance.

Distinctive Image Features from Scale-Invariant Keypoints

TL;DR: The Scale-Invariant Feature Transform (or SIFT) algorithm is a highly robust method to extract and consequently match distinctive invariant features from images that can then be used to reliably match objects in diering images.
Journal ArticleDOI

Robust Real-Time Face Detection

TL;DR: In this paper, a face detection framework that is capable of processing images extremely rapidly while achieving high detection rates is described. But the detection performance is limited to 15 frames per second.
Proceedings ArticleDOI

Robust real-time face detection

TL;DR: A new image representation called the “Integral Image” is introduced which allows the features used by the detector to be computed very quickly and a method for combining classifiers in a “cascade” which allows background regions of the image to be quickly discarded while spending more computation on promising face-like regions.
Proceedings Article

Visual categorization with bags of keypoints

TL;DR: This bag of keypoints method is based on vector quantization of affine invariant descriptors of image patches and shows that it is simple, computationally efficient and intrinsically invariant.
Related Papers (5)