scispace - formally typeset
Search or ask a question
Topic

Histogram of oriented gradients

About: Histogram of oriented gradients is a research topic. Over the lifetime, 2037 publications have been published within this topic receiving 55881 citations. The topic is also known as: HOG.


Papers
More filters
Posted Content
TL;DR: In this article, a novel unsupervised deep neural network architecture of a feature embedding for visual loop closure is proposed, which is built upon the autoencoder architecture tailored specifically to the problem at hand.
Abstract: Robust efficient loop closure detection is essential for large-scale real-time SLAM. In this paper, we propose a novel unsupervised deep neural network architecture of a feature embedding for visual loop closure that is both reliable and compact. Our model is built upon the autoencoder architecture, tailored specifically to the problem at hand. To train our network, we inflict random noise on our input data as the denoising autoencoder does, but, instead of applying random dropout, we warp images with randomized projective transformations to emulate natural viewpoint changes due to robot motion. Moreover, we utilize the geometric information and illumination invariance provided by histogram of oriented gradients (HOG), forcing the encoder to reconstruct a HOG descriptor instead of the original image. As a result, our trained model extracts features robust to extreme variations in appearance directly from raw images, without the need for labeled training data or environment-specific training. We perform extensive experiments on various challenging datasets, showing that the proposed deep loop-closure model consistently outperforms the state-of-the-art methods in terms of effectiveness and efficiency. Our model is fast and reliable enough to close loops in real time with no dimensionality reduction, and capable of replacing generic off-the-shelf networks in state-of-the-art ConvNet-based loop closure systems.

56 citations

Proceedings ArticleDOI
11 Jul 2011
TL;DR: A novel approach for detecting highlights in sports videos based on an unsupervised event discovery and detection framework based on easy-to-extract low-level visual features such as color histogram (CH) or histogram of oriented gradients (HOG), which can potentially be generalized to different sports.
Abstract: In this paper, we propose a novel approach for detecting highlights in sports videos. The videos are temporally decomposed into a series of events based on an unsupervised event discovery and detection framework. The framework solely depends on easy-to-extract low-level visual features such as color histogram (CH) or histogram of oriented gradients (HOG), which can potentially be generalized to different sports. The unigram and bigram statistics of the detected events are then used to provide a compact representation of the video. The effectiveness of the proposed representation is demonstrated on cricket video classification: Highlight vs. Non-Highlight for individual video clips (7000 training and 7000 test instances). We achieve a low equal error rate of 12.1% using event statistics based on CH and HOG features.

56 citations

Journal ArticleDOI
TL;DR: The experimental results demonstrate that the proposed vehicle detection method not only improves detection performance but also reduces computation time.
Abstract: In this paper, a new on-road vehicle detection method is presented. First, a new feature named the Position and Intensity-included Histogram of Oriented Gradients (PIHOG or $\pi$ HOG) is proposed. Unlike the conventional HOG, $\pi$ HOG compensates the information loss involved in the construction of a histogram with position information, and it improves the discriminative power using intensity information. Second, a new search space reduction (SSR) method is proposed to speed up the detection and reduce the computational load. The SSR additionally decreases the false positive rate. A variety of classifiers, including support vector machine, extreme learning machine, and $k$ -nearest neighbor, are used to train and classify vehicles using $\pi$ HOG. The validity of the proposed method is demonstrated by its application to Caltech, IR, Pittsburgh, and Kitti datasets. The experimental results demonstrate that the proposed vehicle detection method not only improves detection performance but also reduces computation time.

56 citations

Journal ArticleDOI
TL;DR: A method for localizing and labeling the lumbar vertebrae and intervertebral discs in mid-sagittal MR image slices based on a Markov-chain-like graphical model that can scale-invariantly localize discs and vertebra at the same time even in the existence of missing structures is presented.
Abstract: This paper presents a method for localizing and labeling the lumbar vertebrae and intervertebral discs in mid-sagittal MR image slices. The approach is based on a Markov-chain-like graphical model of the ordered discs and vertebrae in the lumbar spine. The graphical model is formulated by combining local image features and semiglobal geometrical information. The local image features are extracted from the image by employing pyramidal histogram of oriented gradients (PHOG) and a novel descriptor that we call image projection descriptor (IPD). These features are trained with support vector machines (SVM) and each pixel in the target image is locally assigned a score. These local scores are combined with the semiglobal geometrical information like the distance ratio and angle between the neighboring structures under the Markov random field (MRF) framework. An exact localization of discs and vertebrae is inferred from the MRF by finding a maximum a posteriori solution efficiently using dynamic programming. As a result of the novel features introduced, our system can scale-invariantly localize discs and vertebra at the same time even in the existence of missing structures. The proposed system is tested and validated on a clinical lumbar spine MR image dataset containing 80 subjects of which 64 have disc- and vertebra-related diseases and abnormalities. The experiments show that our system is successful even in abnormal cases and our results are comparable to the state of the art.

55 citations

Proceedings ArticleDOI
01 Dec 2016
TL;DR: A facial expression recognition framework which infers the emotional states in real-time, thereby enabling the computers to interact more intelligently with people.
Abstract: This paper presents a facial expression recognition framework which infers the emotional states in real-time, thereby enabling the computers to interact more intelligently with people. The proposed method determines the face as well as the facial landmark points, extracts discriminating features from suitable facial regions, and classifies the expressions in real-time from live webcam feed. The speed of the system is improved by the appropriate combination of the detection and tracking algorithms. Further, instead of the whole face, histogram of oriented gradients (HOG) features are extracted from the active facial patches which makes the system robust against the scale and pose variations. The feature vectors are further fed to a support vector machine (SVM) classifier to classify into neutral or six universal expressions. Experimental results show an accuracy of 95% with 5 folds cross-validation in extended Cohn-Kanade (CK+) dataset.

55 citations


Network Information
Related Topics (5)
Feature extraction
111.8K papers, 2.1M citations
89% related
Convolutional neural network
74.7K papers, 2M citations
87% related
Deep learning
79.8K papers, 2.1M citations
87% related
Image segmentation
79.6K papers, 1.8M citations
87% related
Feature (computer vision)
128.2K papers, 1.7M citations
86% related
Performance
Metrics
No. of papers in the topic in previous years
YearPapers
202356
2022181
2021116
2020189
2019179
2018240