Open Access
Distinctive Image Features from Scale-Invariant Keypoints
TLDR
The Scale-Invariant Feature Transform (or SIFT) algorithm is a highly robust method to extract and consequently match distinctive invariant features from images that can then be used to reliably match objects in diering images.Abstract:
The Scale-Invariant Feature Transform (or SIFT) algorithm is a highly robust method to extract and consequently match distinctive invariant features from images. These features can then be used to reliably match objects in diering images. The algorithm was rst proposed by Lowe [12] and further developed to increase performance resulting in the classic paper [13] that served as foundation for SIFT which has played an important role in robotic and machine vision in the past decade.read more
Citations
More filters
Proceedings ArticleDOI
Clustering of static-adaptive correspondences for deformable object tracking
Georg Nebehay,Roman Pflugfelder +1 more
TL;DR: This work proposes a novel method for establishing correspondences on deformable objects for single-target object tracking that outperforms the state of the art on a dataset of 77 sequences and builds a keypoint-based tracker that outputs rotated bounding boxes.
Proceedings ArticleDOI
A learned feature descriptor for object recognition in RGB-D data
TL;DR: A new, learned, local feature descriptor for RGB-D images, the convolutional k-means descriptor, which automatically learns feature responses in the neighborhood of detected interest points and is able to combine all available information, such as color and depth into one, concise representation.
Journal ArticleDOI
Real-Time Fire Detection for Video-Surveillance Applications Using a Combination of Experts Based on Color, Shape, and Motion
TL;DR: A method that is able to detect fires by analyzing videos acquired by surveillance cameras is proposed, and a novel descriptor based on a bag-of-words approach has been proposed for representing motion.
Proceedings ArticleDOI
Co-localization in Real-World Images
TL;DR: An extensive evaluation of the method compared to previous state-of-the-art approaches on the challenging PASCAL VOC 2007 and Object Discovery datasets and a large-scale study of co-localization on ImageNet, involving ground-truth annotations for 3, 624 classes and approximately 1 million images.
Journal ArticleDOI
A survey on Visual-Based Localization: On the benefit of heterogeneous data
TL;DR: A survey about recent methods that localize a visual acquisition system according to a known environment by categorizing VBL methods into two distinct families: indirect and direct localization systems.
References
More filters
Journal ArticleDOI
Distinctive Image Features from Scale-Invariant Keypoints
TL;DR: This paper presents a method for extracting distinctive invariant features from images that can be used to perform reliable matching between different views of an object or scene and can robustly identify objects among clutter and occlusion while achieving near real-time performance.
Proceedings ArticleDOI
Object recognition from local scale-invariant features
TL;DR: Experimental results show that robust object recognition can be achieved in cluttered partially occluded images with a computation time of under 2 seconds.
Proceedings ArticleDOI
A Combined Corner and Edge Detector
Chris Harris,Mike Stephens +1 more
TL;DR: The problem the authors are addressing in Alvey Project MMI149 is that of using computer vision to understand the unconstrained 3D world, in which the viewed scenes will in general contain too wide a diversity of objects for topdown recognition techniques to work.
Journal ArticleDOI
A performance evaluation of local descriptors
TL;DR: It is observed that the ranking of the descriptors is mostly independent of the interest region detector and that the SIFT-based descriptors perform best and Moments and steerable filters show the best performance among the low dimensional descriptors.
Journal ArticleDOI
Robust wide-baseline stereo from maximally stable extremal regions
TL;DR: The high utility of MSERs, multiple measurement regions and the robust metric is demonstrated in wide-baseline experiments on image pairs from both indoor and outdoor scenes.