scispace - formally typeset
Search or ask a question
Journal Articleā€¢DOIā€¢

Distinctive Image Features from Scale-Invariant Keypoints

01 Nov 2004-International Journal of Computer Vision (Kluwer Academic Publishers)-Vol. 60, Iss: 2, pp 91-110
TL;DR: This paper presents a method for extracting distinctive invariant features from images that can be used to perform reliable matching between different views of an object or scene and can robustly identify objects among clutter and occlusion while achieving near real-time performance.
Abstract: This paper presents a method for extracting distinctive invariant features from images that can be used to perform reliable matching between different views of an object or scene. The features are invariant to image scale and rotation, and are shown to provide robust matching across a substantial range of affine distortion, change in 3D viewpoint, addition of noise, and change in illumination. The features are highly distinctive, in the sense that a single feature can be correctly matched with high probability against a large database of features from many images. This paper also describes an approach to using these features for object recognition. The recognition proceeds by matching individual features to a database of features from known objects using a fast nearest-neighbor algorithm, followed by a Hough transform to identify clusters belonging to a single object, and finally performing verification through least-squares solution for consistent pose parameters. This approach to recognition can robustly identify objects among clutter and occlusion while achieving near real-time performance.

Content maybe subject toĀ copyrightĀ Ā Ā  Report

Citations
More filters
Journal Articleā€¢DOIā€¢
TL;DR: The inherent difficulties in PIFR are discussed and a comprehensive review of established techniques are presented, that is, pose-robust feature extraction approaches, multiview subspace learning approaches, face synthesis approaches, and hybrid approaches.
Abstract: The capacity to recognize faces under varied poses is a fundamental human ability that presents a unique challenge for computer vision systems. Compared to frontal face recognition, which has been intensively studied and has gradually matured in the past few decades, Pose-Invariant Face Recognition (PIFR) remains a largely unsolved problem. However, PIFR is crucial to realizing the full potential of face recognition for real-world applications, since face recognition is intrinsically a passive biometric technology for recognizing uncooperative subjects. In this article, we discuss the inherent difficulties in PIFR and present a comprehensive review of established techniques. Existing PIFR methods can be grouped into four categories, that is, pose-robust feature extraction approaches, multiview subspace learning approaches, face synthesis approaches, and hybrid approaches. The motivations, strategies, pros/cons, and performance of representative approaches are described and compared. Moreover, promising directions for future research are discussed.

269Ā citations


Cites methods from "Distinctive Image Features from Sca..."

  • ...Similarly, Biswas et al (2013) described each landmark with SIFT features (Lowe 2004) and concatenated the SIFT features of all landmarks as the face representation. More recent engineered features benefit from the rapid progress in face alignment (Wang et al 2014a), which makes dense landmark detection more reliable. For example, Chen et al (2013) extracted multi-scale Local Binary Patterns (LBP) features from patches around 27 landmarks. LBP features for all patches are concatenated to become a high-dimensional feature vector as the pose-robust feature. A similar idea is adopted for feature extraction in (Prince et al 2008; Zhang et al 2013b). Intuitively, the larger the number of landmarks employed, the tighter semantic correspondence that can be achieved. Li et al (2009) proposed the detection of a number of landmarks with the help of a generic 3D face model. In comparison, Yi et al (2013) proposed a more accurate approach by employing a deformable 3D face model with 352 pre-labeled landmarks. Similar to (Li et al 2009), the 2D face image is aligned to the deformable 3D face model using the weak perspective projection model, after which the dense landmarks on the 3D model are projected to the 2D image. Lastly, Gabor magnitude coefficients at all landmarks are extracted and concatenated as the pose-robust feature. Concatenating the features of all landmarks across the face brings about highly non-linear intra-personal variation. To relieve this problem, Ding et al (2014) combined the component-level and landmark-level methods. In their approach, the Dual-Cross Patterns (DCP) (Ding et al 2014) features of landmarks belonging to the same facial component are concatenated as the description of the component. The pose-robust face representation incorporates a set of features of facial components. While the above methods crop patches centered around facial landmarks, Fischer et al (2012) found that the location of the patches for non-frontal faces has a noticeable impact on the recognition results. For example, the positions of patches around some landmarks, e.g., the nose tip and mouth corners, for face images of extreme pose should be adjusted so that fewer background pixels are included. The accuracy and reliability of dense landmark detection are critical for building semantic correspondence. However, accurate landmark detection in unconstrained images is still challenging. To handle this problem, Zhao and Gao (2009); Liao et al (2013b); Weng et al (2013) proposed alignment-free approaches to extract features around the so-called facial keypoints....

    [...]

  • ...Similarly, Biswas et al (2013) described each landmark with SIFT features (Lowe 2004) and concatenated the SIFT features of all landmarks as the face representation. More recent engineered features benefit from the rapid progress in face alignment (Wang et al 2014a), which makes dense landmark detection more reliable. For example, Chen et al (2013) extracted multi-scale Local Binary Patterns (LBP) features from patches around 27 landmarks. LBP features for all patches are concatenated to become a high-dimensional feature vector as the pose-robust feature. A similar idea is adopted for feature extraction in (Prince et al 2008; Zhang et al 2013b). Intuitively, the larger the number of landmarks employed, the tighter semantic correspondence that can be achieved. Li et al (2009) proposed the detection of a number of landmarks with the help of a generic 3D face model. In comparison, Yi et al (2013) proposed a more accurate approach by employing a deformable 3D face model with 352 pre-labeled landmarks. Similar to (Li et al 2009), the 2D face image is aligned to the deformable 3D face model using the weak perspective projection model, after which the dense landmarks on the 3D model are projected to the 2D image. Lastly, Gabor magnitude coefficients at all landmarks are extracted and concatenated as the pose-robust feature. Concatenating the features of all landmarks across the face brings about highly non-linear intra-personal variation. To relieve this problem, Ding et al (2014) combined the component-level and landmark-level methods....

    [...]

  • ...Similarly, Biswas et al (2013) described each landmark with SIFT features (Lowe 2004) and concatenated the SIFT features of all landmarks as the face representation. More recent engineered features benefit from the rapid progress in face alignment (Wang et al 2014a), which makes dense landmark detection more reliable. For example, Chen et al (2013) extracted multi-scale Local Binary Patterns (LBP) features from patches around 27 landmarks. LBP features for all patches are concatenated to become a high-dimensional feature vector as the pose-robust feature. A similar idea is adopted for feature extraction in (Prince et al 2008; Zhang et al 2013b). Intuitively, the larger the number of landmarks employed, the tighter semantic correspondence that can be achieved. Li et al (2009) proposed the detection of a number of landmarks with the help of a generic 3D face model. In comparison, Yi et al (2013) proposed a more accurate approach by employing a deformable 3D face model with 352 pre-labeled landmarks. Similar to (Li et al 2009), the 2D face image is aligned to the deformable 3D face model using the weak perspective projection model, after which the dense landmarks on the 3D model are projected to the 2D image. Lastly, Gabor magnitude coefficients at all landmarks are extracted and concatenated as the pose-robust feature. Concatenating the features of all landmarks across the face brings about highly non-linear intra-personal variation. To relieve this problem, Ding et al (2014) combined the component-level and landmark-level methods. In their approach, the Dual-Cross Patterns (DCP) (Ding et al 2014) features of landmarks belonging to the same facial component are concatenated as the description of the component. The pose-robust face representation incorporates a set of features of facial components. While the above methods crop patches centered around facial landmarks, Fischer et al (2012) found that the location of the patches for non-frontal faces has a noticeable impact on the recognition results. For example, the positions of patches around some landmarks, e.g., the nose tip and mouth corners, for face images of extreme pose should be adjusted so that fewer background pixels are included. The accuracy and reliability of dense landmark detection are critical for building semantic correspondence. However, accurate landmark detection in unconstrained images is still challenging. To handle this problem, Zhao and Gao (2009); Liao et al (2013b); Weng et al (2013) proposed alignment-free approaches to extract features around the so-called facial keypoints. For example, Liao et al (2013b) proposed the extraction of Multi-Keypoint Descriptors (MKD) around keypoints detected by SIFT-like detectors....

    [...]

  • ...Similarly, Biswas et al. (2013) described each landmark with SIFT features Lowe (2004) and concatenated the SIFT features of all landmarks as the face representation....

    [...]

  • ...Similarly, Biswas et al (2013) described each landmark with SIFT features (Lowe 2004) and concatenated the SIFT features of all landmarks as the face representation. More recent engineered features benefit from the rapid progress in face alignment (Wang et al 2014a), which makes dense landmark detection more reliable. For example, Chen et al (2013) extracted multi-scale Local Binary Patterns (LBP) features from patches around 27 landmarks. LBP features for all patches are concatenated to become a high-dimensional feature vector as the pose-robust feature. A similar idea is adopted for feature extraction in (Prince et al 2008; Zhang et al 2013b). Intuitively, the larger the number of landmarks employed, the tighter semantic correspondence that can be achieved. Li et al (2009) proposed the detection of a number of landmarks with the help of a generic 3D face model....

    [...]

Proceedings Articleā€¢DOIā€¢
17 Jun 2007
TL;DR: A fast and principled solution to the discovery of significant spatial co-occurrent patterns using frequent itemset mining; a pattern summarization method that deals with the compositional uncertainties in visual phrases; and a top-down refinement scheme of the visual word lexicon by feeding back discovered phrases to tune the similarity measure through metric learning.
Abstract: A visual word lexicon can be constructed by clustering primitive visual features, and a visual object can be described by a set of visual words. Such a "bag-of-words" representation has led to many significant results in various vision tasks including object recognition and categorization. However, in practice, the clustering of primitive visual features tends to result in synonymous visual words that over-represent visual patterns, as well as polysemous visual words that bring large uncertainties and ambiguities in the representation. This paper aims at generating a higher-level lexicon, i.e. visual phrase lexicon, where a visual phrase is a meaningful spatially co-occurrent pattern of visual words. This higher-level lexicon is much less ambiguous than the lower-level one. The contributions of this paper include: (1) a fast and principled solution to the discovery of significant spatial co-occurrent patterns using frequent itemset mining; (2) a pattern summarization method that deals with the compositional uncertainties in visual phrases; and (3) a top-down refinement scheme of the visual word lexicon by feeding back discovered phrases to tune the similarity measure through metric learning.

269Ā citations

Journal Articleā€¢DOIā€¢
TL;DR: This work comprehensively review the existing small object detection methods based on deep learning from five aspects, including multi-scale feature learning, data augmentation, training strategy, context-based detection and GAN- based detection.

269Ā citations

Journal Articleā€¢DOIā€¢
TL;DR: The design and implementation of a deep neural network model referred to as ElemNet is presented; it automatically captures the physical and chemical interactions and similarities between different elements using artificial intelligence which allows it to predict the materials properties with better accuracy and speed.
Abstract: Conventional machine learning approaches for predicting material properties from elemental compositions have emphasized the importance of leveraging domain knowledge when designing model inputs. Here, we demonstrate that by using a deep learning approach, we can bypass such manual feature engineering requiring domain knowledge and achieve much better results, even with only a few thousand training samples. We present the design and implementation of a deep neural network model referred to as ElemNet; it automatically captures the physical and chemical interactions and similarities between different elements using artificial intelligence which allows it to predict the materials properties with better accuracy and speed. The speed and best-in-class accuracy of ElemNet enable us to perform a fast and robust screening for new material candidates in a huge combinatorial space; where we predict hundreds of thousands of chemical systems that could contain yet-undiscovered compounds.

268Ā citations

Proceedings Articleā€¢DOIā€¢
28 Sep 2009
TL;DR: The feasibility of using periocular images of an individual as a biometric trait using texture and point operators resulting in a feature set that can be used for matching is studied.
Abstract: Periocular biometric refers to the facial region in the immediate vicinity of the eye. Acquisition of the periocular biometric does not require high user cooperation and close capture distance unlike other ocular biometrics (e.g., iris, retina, and sclera). We study the feasibility of using periocular images of an individual as a biometric trait. Global and local information are extracted from the periocular region using texture and point operators resulting in a feature set that can be used for matching. The effect of fusing these feature sets is also studied. The experimental results show a 77% rank-1 recognition accuracy using 958 images captured from 30 different subjects.

267Ā citations

References
More filters
Proceedings Articleā€¢DOIā€¢
20 Sep 1999
TL;DR: Experimental results show that robust object recognition can be achieved in cluttered partially occluded images with a computation time of under 2 seconds.
Abstract: An object recognition system has been developed that uses a new class of local image features. The features are invariant to image scaling, translation, and rotation, and partially invariant to illumination changes and affine or 3D projection. These features share similar properties with neurons in inferior temporal cortex that are used for object recognition in primate vision. Features are efficiently detected through a staged filtering approach that identifies stable points in scale space. Image keys are created that allow for local geometric deformations by representing blurred image gradients in multiple orientation planes and at multiple scales. The keys are used as input to a nearest neighbor indexing method that identifies candidate object matches. Final verification of each match is achieved by finding a low residual least squares solution for the unknown model parameters. Experimental results show that robust object recognition can be achieved in cluttered partially occluded images with a computation time of under 2 seconds.

16,989Ā citations


"Distinctive Image Features from Sca..." refers background or methods in this paper

  • ...The initial implementation of this approach (Lowe, 1999) simply located keypoints at the location and scale of the central sample point....

    [...]

  • ...Earlier work by the author (Lowe, 1999) extended the local feature approach to achieve scale invariance....

    [...]

  • ...More details on applications of these features to recognition are available in other pape rs (Lowe, 1999; Lowe, 2001; Se, Lowe and Little, 2002)....

    [...]

  • ...To efficiently detect stable keypoint locations in scale space, we have proposed (Lowe, 1999) using scalespace extrema in the difference-of-Gaussian function convolved with the image, D(x, y, Ļƒ ), which can be computed from the difference of two nearby scales separated by a constant multiplicativeā€¦...

    [...]

  • ...More details on applications of these features to recognition are available in other papers (Lowe, 1999, 2001; Se et al., 2002)....

    [...]

Bookā€¢
01 Jan 2000
TL;DR: In this article, the authors provide comprehensive background material and explain how to apply the methods and implement the algorithms directly in a unified framework, including geometric principles and how to represent objects algebraically so they can be computed and applied.
Abstract: From the Publisher: A basic problem in computer vision is to understand the structure of a real world scene given several images of it. Recent major developments in the theory and practice of scene reconstruction are described in detail in a unified framework. The book covers the geometric principles and how to represent objects algebraically so they can be computed and applied. The authors provide comprehensive background material and explain how to apply the methods and implement the algorithms directly.

15,558Ā citations

01 Jan 2001
TL;DR: This book is referred to read because it is an inspiring book to give you more chance to get experiences and also thoughts and it will show the best book collections and completed collections.
Abstract: Downloading the book in this website lists can give you more advantages. It will show you the best book collections and completed collections. So many books can be found in this website. So, this is not only this multiple view geometry in computer vision. However, this book is referred to read because it is an inspiring book to give you more chance to get experiences and also thoughts. This is simple, read the soft file of the book and you get it.

14,282Ā citations


"Distinctive Image Features from Sca..." refers background in this paper

  • ...A more general solution would be to solve for the fundamental matrix (Luong and Faugeras, 1996; Hartley and Zisserman, 2000)....

    [...]

Proceedings Articleā€¢DOIā€¢
01 Jan 1988
TL;DR: The problem the authors are addressing in Alvey Project MMI149 is that of using computer vision to understand the unconstrained 3D world, in which the viewed scenes will in general contain too wide a diversity of objects for topdown recognition techniques to work.
Abstract: The problem we are addressing in Alvey Project MMI149 is that of using computer vision to understand the unconstrained 3D world, in which the viewed scenes will in general contain too wide a diversity of objects for topdown recognition techniques to work. For example, we desire to obtain an understanding of natural scenes, containing roads, buildings, trees, bushes, etc., as typified by the two frames from a sequence illustrated in Figure 1. The solution to this problem that we are pursuing is to use a computer vision system based upon motion analysis of a monocular image sequence from a mobile camera. By extraction and tracking of image features, representations of the 3D analogues of these features can be constructed.

13,993Ā citations

Journal Articleā€¢DOIā€¢
TL;DR: The high utility of MSERs, multiple measurement regions and the robust metric is demonstrated in wide-baseline experiments on image pairs from both indoor and outdoor scenes.

3,422Ā citations

Trending Questions (1)
How can distinctive features theory be applied to elision?

The provided information does not mention anything about the application of distinctive features theory to elision.