scispace - formally typeset
Search or ask a question
Topic

Scale-invariant feature transform

About: Scale-invariant feature transform is a research topic. Over the lifetime, 6779 publications have been published within this topic receiving 197891 citations. The topic is also known as: SIFT.


Papers
More filters
Journal ArticleDOI
TL;DR: This work reduces the size of the descriptors by representing them as short binary strings and learn descriptor invariance from examples, and shows extensive experimental validation, demonstrating the advantage of the proposed approach.
Abstract: SIFT-like local feature descriptors are ubiquitously employed in computer vision applications such as content-based retrieval, video analysis, copy detection, object recognition, photo tourism, and 3D reconstruction. Feature descriptors can be designed to be invariant to certain classes of photometric and geometric transformations, in particular, affine and intensity scale transformations. However, real transformations that an image can undergo can only be approximately modeled in this way, and thus most descriptors are only approximately invariant in practice. Second, descriptors are usually high dimensional (e.g., SIFT is represented as a 128-dimensional vector). In large-scale retrieval and matching problems, this can pose challenges in storing and retrieving descriptor data. We map the descriptor vectors into the Hamming space in which the Hamming metric is used to compare the resulting representations. This way, we reduce the size of the descriptors by representing them as short binary strings and learn descriptor invariance from examples. We show extensive experimental validation, demonstrating the advantage of the proposed approach.

654 citations

Journal ArticleDOI
TL;DR: A variety of circle detection methods which are based on variations of the Hough Transform are investigated, and the accuracy, reliability, computational efficiency and storage requirements of each of the methods are presented.

652 citations

Posted Content
TL;DR: This paper presents a self-supervised framework for training interest point detectors and descriptors suitable for a large number of multiple-view geometry problems in computer vision and introduces Homographic Adaptation, a multi-scale, multi-homography approach for boosting interest point detection repeatability and performing cross-domain adaptation.
Abstract: This paper presents a self-supervised framework for training interest point detectors and descriptors suitable for a large number of multiple-view geometry problems in computer vision. As opposed to patch-based neural networks, our fully-convolutional model operates on full-sized images and jointly computes pixel-level interest point locations and associated descriptors in one forward pass. We introduce Homographic Adaptation, a multi-scale, multi-homography approach for boosting interest point detection repeatability and performing cross-domain adaptation (e.g., synthetic-to-real). Our model, when trained on the MS-COCO generic image dataset using Homographic Adaptation, is able to repeatedly detect a much richer set of interest points than the initial pre-adapted deep model and any other traditional corner detector. The final system gives rise to state-of-the-art homography estimation results on HPatches when compared to LIFT, SIFT and ORB.

641 citations

Journal ArticleDOI
TL;DR: It is shown that if just a small subset of the edge points in the image, selected at random, is used as input for the Hough Transform, the performance is often only slightly impaired, thus the execution time can be considerably shortened.

640 citations

Journal ArticleDOI
TL;DR: Hough forests can be regarded as task-adapted codebooks of local appearance that allow fast supervised training and fast matching at test time that improve the performance of the generalized Hough transform for object detection on a categorical level and extend to new domains such as object tracking and action recognition.
Abstract: The paper introduces Hough forests, which are random forests adapted to perform a generalized Hough transform in an efficient way. Compared to previous Hough-based systems such as implicit shape models, Hough forests improve the performance of the generalized Hough transform for object detection on a categorical level. At the same time, their flexibility permits extensions of the Hough transform to new domains such as object tracking and action recognition. Hough forests can be regarded as task-adapted codebooks of local appearance that allow fast supervised training and fast matching at test time. They achieve high detection accuracy since the entries of such codebooks are optimized to cast Hough votes with small variance and since their efficiency permits dense sampling of local image patches or video cuboids during detection. The efficacy of Hough forests for a set of computer vision tasks is validated through experiments on a large set of publicly available benchmark data sets and comparisons with the state-of-the-art.

629 citations


Network Information
Related Topics (5)
Image segmentation
79.6K papers, 1.8M citations
93% related
Feature extraction
111.8K papers, 2.1M citations
93% related
Convolutional neural network
74.7K papers, 2M citations
91% related
Feature (computer vision)
128.2K papers, 1.7M citations
91% related
Image processing
229.9K papers, 3.5M citations
90% related
Performance
Metrics
No. of papers in the topic in previous years
YearPapers
2023190
2022442
2021165
2020276
2019353
2018445