scispace - formally typeset
Search or ask a question
Topic

Histogram of oriented gradients

About: Histogram of oriented gradients is a research topic. Over the lifetime, 2037 publications have been published within this topic receiving 55881 citations. The topic is also known as: HOG.


Papers
More filters
Proceedings ArticleDOI
04 Jul 2016
TL;DR: A novel method to learn rotation-invariant HOG (RIHOG) features for object detection in optical remote sensing images is proposed via optimizing a new objective function, which constrains the training samples before and after rotation to share the similar features to achieve rotation- Invariance.
Abstract: Object detection in very high resolution (VHR) optical remote sensing images is one of the most fundamental but challenging problems in the field of remote sensing image analysis. As object detection is usually carried out in feature space, effective feature representation is very important to construct a high-performance object detection system. During the last decades, a great deal of effort has been made to develop various feature representations for the detection of different types of objects. Among various features developed for visual object detection, the histogram of oriented gradients (HOG) feature is maybe one of the most popular features that has been successfully applied to computer vision community. However, although the HOG feature has achieved great success in nature scene images, it is problematic to directly use it for object detection in optical remote sensing images because it is difficult to effectively handle the problem of object rotation variations. To explore a possible solution to the problem, this paper proposes a novel method to learn rotation-invariant HOG (RIHOG) features for object detection in optical remote sensing images. This is achieved by learning a rotation-invariant transformation model via optimizing a new objective function, which constrains the training samples before and after rotation to share the similar features to achieve rotation-invariance. In the experiments, we evaluate the proposed method on a publicly available 10-class VHR geospatial object detection dataset and comprehensive comparisons with state-of-the-arts demonstrate the effectiveness the proposed method.

29 citations

Journal ArticleDOI
TL;DR: The proposed method consistently improves the detection rate by 4.5% in detection accuracy, compared with the original HOG, and is evaluated with a Histogram Intersection Kernel SVM (HIKSVM) on the public “INRIA” pedestrian detection benchmark dataset.
Abstract: The outstanding Histogram-of-Oriented-Gradients (HOG) feature proposed by Dalal and Triggs is a state-of-art technique for pedestrian detection, and it is usually applied with a linear support vector machine (SVM) in a sliding-window framework. Most other algorithms for pedestrian detection use HOG as the basic feature, and combine other features with HOG to form the feature set. Hence, the HOG feature is actually the most efficient and fundamental feature for pedestrian detection. However, the HOG feature cannot adequately handle scale variation of pedestrians. In addition, simply downsampling an image into a different scale, or decomposing via wavelet into multi-resolution subimages, calculating their HOG feature and combining them cannot enhance performance. Therefore, in this paper, based on the idea of multi-resolution feature descriptors, we propose a new robust edge feature referred to as Enhanced HOG (eHOG). It is a complementary descriptor for the histograms-of-oriented-gradients feature. Though the extraction process of the eHOG descriptor is derived only from HOG itself, similar to the process of extracting edge information from the downscaling image, it retains much more information for the edge gradient than that of the original HOG, without significantly increasing the complexity of computation. In the INRIA pedestrian dataset, many experiments have been conducted with eHOG and HOG, and the results show that the proposed new feature consistently improves the detection rate more than the original HOG feature detector. Particularly, eHOG with a Histogram Intersection Kernel SVM (HIKSVM) classifier has greatly improved performance. These results suggest that eHOG may be a better substitute for HOG for pedestrian detection in many applications.

29 citations

Journal ArticleDOI
TL;DR: Experimental results show that the proposed modified feature descriptors can effectively recognize emotions on CK+ dataset and JAFFE dataset.

29 citations

Journal ArticleDOI
TL;DR: Comparative analysis on these databases using various descriptors shows the superiority of BSIF with Cosine, Chi square and Cityblock distance measures using 1-NN as classifier over other descriptors and distance measures and even some of the current state-of-art benchmark database results.

29 citations

Journal ArticleDOI
TL;DR: A hybrid approach of facial expression based sentiment analysis has been presented combining local and global features, boosting performance of the proposed technique over face images containing noise and occlusions.
Abstract: Facial sentiment analysis has been an enthusiastic research area for the last two decades A fair amount of work has been done by researchers in this field due to its utility in numerous applications such as facial expression driven knowledge discovery However, developing an accurate and efficient facial expression recognition system is still a challenging problem Although many efficient recognition systems have been introduced in the past, the recognition rate is not satisfactory in general due to inherent limitations including light, pose variations, noise, and occlusion In this paper, a hybrid approach of facial expression based sentiment analysis has been presented combining local and global features Feature extraction is performed fusing the histogram of oriented gradients (HOG) descriptor with the uniform local ternary pattern (U-LTP) descriptor These features are extracted from the entire face image rather than from individual components of faces like eyes, nose, and mouth The most suitable set of HOG parameters are selected after analyzing them experimentally along with the ULTP descriptor, boosting performance of the proposed technique over face images containing noise and occlusions Face sentiments are analyzed classifying them into seven universal emotional expressions: Happy, Angry, Fear, Disgust, Sad, Surprise, and Neutral Extracted features via HOG and ULTP are fused into a single feature vector and this feature vector is fed into a Multi-class Support Vector Machine classifier for emotion classification Three types of experiments are conducted over three public facial image databases including JAFFE, MMI, and CK+ to evaluate the recognition rate of the proposed technique during experimental evaluation; recognition accuracy in percent, ie, 9571, 9820, and 9968 are achieved for JAFFE, MMI, and CK+, respectively

29 citations


Network Information
Related Topics (5)
Feature extraction
111.8K papers, 2.1M citations
89% related
Convolutional neural network
74.7K papers, 2M citations
87% related
Deep learning
79.8K papers, 2.1M citations
87% related
Image segmentation
79.6K papers, 1.8M citations
87% related
Feature (computer vision)
128.2K papers, 1.7M citations
86% related
Performance
Metrics
No. of papers in the topic in previous years
YearPapers
202356
2022181
2021116
2020189
2019179
2018240