Open Access
Distinctive Image Features from Scale-Invariant Keypoints
Reads0
Chats0
TLDR
The Scale-Invariant Feature Transform (or SIFT) algorithm is a highly robust method to extract and consequently match distinctive invariant features from images that can then be used to reliably match objects in diering images.Abstract:
The Scale-Invariant Feature Transform (or SIFT) algorithm is a highly robust method to extract and consequently match distinctive invariant features from images. These features can then be used to reliably match objects in diering images. The algorithm was rst proposed by Lowe [12] and further developed to increase performance resulting in the classic paper [13] that served as foundation for SIFT which has played an important role in robotic and machine vision in the past decade.read more
Citations
More filters
Journal ArticleDOI
Residual enhanced visual vector as a compact signature for mobile visual search
David Chen,Sam S. Tsai,Vijay Chandrasekhar,Gabriel Takacs,Ramakrishna Vedantham,Radek Grzeszczuk,Bernd Girod +6 more
TL;DR: This paper presents a new architecture for searching a large database directly on a mobile device, which can provide numerous benefits for network-independent, low-latency, and privacy-protected image retrieval.
Proceedings ArticleDOI
Taking a Deeper Look at Co-Salient Object Detection
TL;DR: A new high-quality dataset, named CoSOD3k, is collected, which contains 3,316 images divided into 160 groups with multiple level annotations, i.e., category, bounding box, object, and instance levels, and makes a significant leap in terms of diversity, difficulty and scalability.
Journal ArticleDOI
Goal-Oriented Rectification of Camera-Based Document Images
TL;DR: Experimental results demonstrate the robustness and effectiveness of the proposed rectification methodology using a consistent evaluation methodology that encounters OCR accuracy and a newly introduced measure using a semi-automatic procedure.
Journal ArticleDOI
Fusion of block and keypoints based approaches for effective copy-move image forgery detection
TL;DR: Experimental results show that the proposed fusion scheme outperforms the keypoint- based method in reliability of detection and the block-based method in efficiency.
Content based leaf image retrieval (cblir) using shape, color and texture features
TL;DR: This paper proposes an efficient computer-aided Plant Image Retrieval method based on plant leaf images using Shape, Color and Texture features intended mainly for medical industry, botanical gardening and cosmetic industry, which outperforms the recently developed methods.
References
More filters
Journal ArticleDOI
Distinctive Image Features from Scale-Invariant Keypoints
TL;DR: This paper presents a method for extracting distinctive invariant features from images that can be used to perform reliable matching between different views of an object or scene and can robustly identify objects among clutter and occlusion while achieving near real-time performance.
Proceedings ArticleDOI
Object recognition from local scale-invariant features
TL;DR: Experimental results show that robust object recognition can be achieved in cluttered partially occluded images with a computation time of under 2 seconds.
Proceedings ArticleDOI
A Combined Corner and Edge Detector
Chris Harris,Mike Stephens +1 more
TL;DR: The problem the authors are addressing in Alvey Project MMI149 is that of using computer vision to understand the unconstrained 3D world, in which the viewed scenes will in general contain too wide a diversity of objects for topdown recognition techniques to work.
Journal ArticleDOI
A performance evaluation of local descriptors
TL;DR: It is observed that the ranking of the descriptors is mostly independent of the interest region detector and that the SIFT-based descriptors perform best and Moments and steerable filters show the best performance among the low dimensional descriptors.
Journal ArticleDOI
Robust wide-baseline stereo from maximally stable extremal regions
TL;DR: The high utility of MSERs, multiple measurement regions and the robust metric is demonstrated in wide-baseline experiments on image pairs from both indoor and outdoor scenes.