Pruning SIFT & SURF for Efficient Clustering of Near-duplicate Images
09 May 2019-pp 3132-3136
...read more
Citations
More filters
[...]
TL;DR: This thesis proposes approaches for efficient clustering, fast direction oriented motion estimation algorithm, and an image reordering scheme with minimum predictive costs for better compression of near-duplicate image collection.
Abstract: The explosion of digital photos has posed a challenge for storage and transmission bandwidth. The thesis briefly discusses my work on efficient image set compression. The lossless compression of near-duplicate image collection is carried out using multiple steps, where each step is computationally demanding. My task is to make them work faster without compromising on compression efficiency. In this pursuit, we propose approaches for efficient clustering, fast direction oriented motion estimation algorithm, and an image reordering scheme with minimum predictive costs for better compression. The preliminary results for the proposed approach are promising. We also aim to extend our approach to hyperspectral and medical image set compression.
2 citations
Cites background from "Pruning SIFT & SURF for Efficient C..."
[...]
[...]
TL;DR: The proposed Bag-of-Visual Word Modelling in which Image Feature Extraction is achieved using Deep Feature Learning via Stacked-Autoencoder is tested and the ability of deep feature learning to yield optimum image categorisation performance is confirmed.
Abstract: The Bag-of-Visual Words has been recognised as an effective mean of representing images for image classification. However, its reliance on hand crafted image feature extraction algorithms often results in significant computational overhead, and poor classification accuracies. Therefore, this paper presents a Bag-of-Visual Word Modelling in which Image Feature Extraction is achieved using Deep Feature Learning via Stacked-Autoencoder. The proposed method is tested using three image collections constituted from the Caltech 101 image collection and the results confirm the ability of deep feature learning to yield optimum image categorisation performance.
2 citations
[...]
TL;DR: This study presents an adaptive BOVW modelling, in which image feature extraction is achieved using deep feature learning and the amount of computation required for the development of visual codebook is minimised using a batch implementation of particle swarm optimisation.
Abstract: The bag-of-visual words (BOVWs) have been recognised as an effective mean of representing images for image classification. However, its reliance on a visual codebook developed using handcrafted image feature extraction algorithms and vector quantisation via k
-means clustering often results in significant computational overhead, and poor classification accuracies. Therefore, this study presents an adaptive BOVW modelling, in which image feature extraction is achieved using deep feature learning and the amount of computation required for the development of visual codebook is minimised using a batch implementation of particle swarm optimisation. The proposed method is tested using Caltech-101 image dataset, and the results confirm the suitability of the proposed method in improving the categorisation performance while reducing the computational load.
1 citations
Cites background from "Pruning SIFT & SURF for Efficient C..."
[...]
[...]
[...]
[...]
References
More filters
[...]
TL;DR: Experimental results show that robust object recognition can be achieved in cluttered partially occluded images with a computation time of under 2 seconds.
Abstract: An object recognition system has been developed that uses a new class of local image features. The features are invariant to image scaling, translation, and rotation, and partially invariant to illumination changes and affine or 3D projection. These features share similar properties with neurons in inferior temporal cortex that are used for object recognition in primate vision. Features are efficiently detected through a staged filtering approach that identifies stable points in scale space. Image keys are created that allow for local geometric deformations by representing blurred image gradients in multiple orientation planes and at multiple scales. The keys are used as input to a nearest neighbor indexing method that identifies candidate object matches. Final verification of each match is achieved by finding a low residual least squares solution for the unknown model parameters. Experimental results show that robust object recognition can be achieved in cluttered partially occluded images with a computation time of under 2 seconds.
15,597 citations
"Pruning SIFT & SURF for Efficient C..." refers methods in this paper
[...]
[...]
TL;DR: A novel scale- and rotation-invariant detector and descriptor, coined SURF (Speeded-Up Robust Features), which approximates or even outperforms previously proposed schemes with respect to repeatability, distinctiveness, and robustness, yet can be computed and compared much faster.
Abstract: This article presents a novel scale- and rotation-invariant detector and descriptor, coined SURF (Speeded-Up Robust Features). SURF approximates or even outperforms previously proposed schemes with respect to repeatability, distinctiveness, and robustness, yet can be computed and compared much faster. This is achieved by relying on integral images for image convolutions; by building on the strengths of the leading existing detectors and descriptors (specifically, using a Hessian matrix-based measure for the detector, and a distribution-based descriptor); and by simplifying these methods to the essential. This leads to a combination of novel detection, description, and matching steps. The paper encompasses a detailed description of the detector and descriptor and then explores the effects of the most important parameters. We conclude the article with SURF's application to two challenging, yet converse goals: camera calibration as a special case of image registration, and object recognition. Our experiments underline SURF's usefulness in a broad range of topics in computer vision.
11,276 citations
"Pruning SIFT & SURF for Efficient C..." refers methods in this paper
[...]
[...]
[...]
TL;DR: In this article, the authors derived necessary conditions for any finite number of quanta and associated quantization intervals of an optimum finite quantization scheme to achieve minimum average quantization noise power.
Abstract: It has long been realized that in pulse-code modulation (PCM), with a given ensemble of signals to handle, the quantum values should be spaced more closely in the voltage regions where the signal amplitude is more likely to fall. It has been shown by Panter and Dite that, in the limit as the number of quanta becomes infinite, the asymptotic fractional density of quanta per unit voltage should vary as the one-third power of the probability density per unit voltage of signal amplitudes. In this paper the corresponding result for any finite number of quanta is derived; that is, necessary conditions are found that the quanta and associated quantization intervals of an optimum finite quantization scheme must satisfy. The optimization criterion used is that the average quantization noise power be a minimum. It is shown that the result obtained here goes over into the Panter and Dite result as the number of quanta become large. The optimum quautization schemes for 2^{b} quanta, b=1,2, \cdots, 7 , are given numerically for Gaussian and for Laplacian distribution of signal amplitudes.
9,657 citations
[...]
TL;DR: The corresponding result for any finite number of quanta is derived; that is, necessary conditions are found that the quanta and associated quantization intervals of an optimum finite quantization scheme must satisfy.
Abstract: It has long been realized that in pulse-code modulation (PCM), with a given ensemble of signals to handle, the quantum values should be spaced more closely in the voltage regions where the signal amplitude is more likely to fall. It has been shown by Panter and Dite that, in the limit as the number of quanta becomes infinite, the asymptotic fractional density of quanta per unit voltage should vary as the one-third power of the probability density per unit voltage of signal amplitudes. In this paper the corresponding result for any finite number of quanta is derived; that is, necessary conditions are found that the quanta and associated quantization intervals of an optimum finite quantization scheme must satisfy. The optimization criterion used is that the average quantization noise power be a minimum. It is shown that the result obtained here goes over into the Panter and Dite result as the number of quanta become large. The optimum quautization schemes for 2^{b} quanta, b=1,2, \cdots, 7 , are given numerically for Gaussian and for Laplacian distribution of signal amplitudes.
9,602 citations
"Pruning SIFT & SURF for Efficient C..." refers methods in this paper
[...]
[...]
17 Jun 2006
TL;DR: This paper presents a method for recognizing scene categories based on approximate global geometric correspondence that exceeds the state of the art on the Caltech-101 database and achieves high accuracy on a large database of fifteen natural scene categories.
Abstract: This paper presents a method for recognizing scene categories based on approximate global geometric correspondence. This technique works by partitioning the image into increasingly fine sub-regions and computing histograms of local features found inside each sub-region. The resulting "spatial pyramid" is a simple and computationally efficient extension of an orderless bag-of-features image representation, and it shows significantly improved performance on challenging scene categorization tasks. Specifically, our proposed method exceeds the state of the art on the Caltech-101 database and achieves high accuracy on a large database of fifteen natural scene categories. The spatial pyramid framework also offers insights into the success of several recently proposed image descriptions, including Torralbas "gist" and Lowes SIFT descriptors.
8,415 citations
"Pruning SIFT & SURF for Efficient C..." refers background in this paper
[...]
Related Papers (5)
[...]
[...]
[...]