scispace - formally typeset
Search or ask a question
Author

Haibo Luo

Bio: Haibo Luo is an academic researcher from Chinese Academy of Sciences. The author has contributed to research in topics: Computer science & Artificial intelligence. The author has an hindex of 12, co-authored 49 publications receiving 440 citations. Previous affiliations of Haibo Luo include Shenyang Institute of Automation.


Papers
More filters
Journal ArticleDOI
03 Oct 2018-PLOS ONE
TL;DR: This work proposes a novel method to detect violence sequences by extracting two kinds of low-level features to represent the appearance and dynamics for violent behaviors in video surveillance scenarios.
Abstract: It is very important to automatically detect violent behaviors in video surveillance scenarios, for instance, railway stations, gymnasiums and psychiatric centers. However, the previous detection methods usually extract descriptors around the spatiotemporal interesting points or extract statistic features in the motion regions, leading to limited abilities to effectively detect video-based violence activities. To address this issue, we propose a novel method to detect violence sequences. Firstly, the motion regions are segmented according to the distribution of optical flow fields. Secondly, in the motion regions, we propose to extract two kinds of low-level features to represent the appearance and dynamics for violent behaviors. The proposed low-level features are the Local Histogram of Oriented Gradient (LHOG) descriptor extracted from RGB images and the Local Histogram of Optical Flow (LHOF) descriptor extracted from optical flow images. Thirdly, the extracted features are coded using Bag of Words (BoW) model to eliminate redundant information and a specific-length vector is obtained for each video clip. At last, the video-level vectors are classified by Support Vector Machine (SVM). Experimental results on three challenging benchmark datasets demonstrate that the proposed detection approach is superior to the previous methods.

87 citations

Journal ArticleDOI
01 Jun 2017
TL;DR: A new input modality, image acceleration field is proposed to better extract the motion attributes and experimental results demonstrate that the proposed model for violent interaction detection shows higher accuracy and better robustness.
Abstract: Violent interaction detection is of vital importance in some video surveillance scenarios like railway stations, prisons or psychiatric centres. Existing vision-based methods are mainly based on hand-crafted features such as statistic features between motion regions, leading to a poor adaptability to another dataset. En lightened by the development of convolutional networks on common activity recognition, we construct a FightNet to represent the complicated visual violence interaction. In this paper, a new input modality, image acceleration field is proposed to better extract the motion attributes. Firstly, each video is framed as RGB images. Secondly, optical flow field is computed using the consecutive frames and acceleration field is obtained according to the optical flow field. Thirdly, the FightNet is trained with three kinds of input modalities, i.e., RGB images for spatial networks, optical flow images and acceleration images for temporal networks. By fusing results from different inputs, we conclude whether a video tells a violent event or not. To provide researchers a common ground for comparison, we have collected a violent interaction dataset (VID), containing 2314 videos with 1077 fight ones and 1237 no-fight ones. By comparison with other algorithms, experimental results demonstrate that the proposed model for violent interaction detection shows higher accuracy and better robustness.

86 citations

Journal ArticleDOI
TL;DR: A new interpolation method for DoFP polarimeters is presented by using intensity correlation to detect edges and then implement interpolation along edges, which can achieve better visual effects and a lower RMSE than other methods.
Abstract: Division of focal plane (DoFP) polarimeters operate by integrating micro-polarizer elements with a focal plane. These polarization imaging sensors reduce spatial resolution output and each pixel has a varying instantaneous field of view (IFoV). These drawbacks can be mitigated by applying proper interpolation methods. In this paper, we present a new interpolation method for DoFP polarimeters by using intensity correlation. We employ the correlation of intensity measurements in different orientations to detect edges and then implement interpolation along edges. The performance of the proposed method is compared with several previous methods by using root mean square error (RMSE) comparison and visual comparison. Experimental results showed that our proposed method can achieve better visual effects and a lower RMSE than other methods.

81 citations

Journal ArticleDOI
TL;DR: A new approach is proposed, where the task of phase unwrapping is transferred into a multi-class classification problem and an efficient segmentation network is introduced to identify classes and a noise-to-noise denoised network is integrated to preprocess noisy wrapped phase.
Abstract: The interferometry technique is commonly used to obtain the phase information of an object in optical metrology. The obtained wrapped phase is subject to a 2π ambiguity. To remove the ambiguity and obtain the correct phase, phase unwrapping is essential. Conventional phase unwrapping approaches are time-consuming and noise sensitive. To address those issues, we propose a new approach, where we transfer the task of phase unwrapping into a multi-class classification problem and introduce an efficient segmentation network to identify classes. Moreover, a noise-to-noise denoised network is integrated to preprocess noisy wrapped phase. We have demonstrated the proposed method with simulated data and in a real interferometric system.

78 citations

Journal ArticleDOI
TL;DR: A mathematical derivation method based on Intersection over Union (IOU) was proposed to select the number and the aspect ratio dimensions of the candidate anchor boxes for each scale of the improved YOLO V3 to improve the detection performance of the network.
Abstract: Target detection is one of the most important research directions in computer vision. Recently, a variety of target detection algorithms have been proposed. Since the targets have varying sizes in a scene, it is essential to be able to detect the targets at different scales. To improve the detection performance of targets with different sizes, a multi-scale target detection algorithm was proposed involving improved YOLO (You Only Look Once) V3. The main contributions of our work include: (1) a mathematical derivation method based on Intersection over Union (IOU) was proposed to select the number and the aspect ratio dimensions of the candidate anchor boxes for each scale of the improved YOLO V3; (2) To further improve the detection performance of the network, the detection scales of YOLO V3 have been extended from 3 to 4 and the feature fusion target detection layer downsampled by 4× is established to detect the small targets; (3) To avoid gradient fading and enhance the reuse of the features, the six convolutional layers in front of the output detection layer are transformed into two residual units. The experimental results upon PASCAL VOC dataset and KITTI dataset show that the proposed method has obtained better performance than other state-of-the-art target detection algorithms.

75 citations


Cited by
More filters
01 Jan 2004
TL;DR: Comprehensive and up-to-date, this book includes essential topics that either reflect practical significance or are of theoretical importance and describes numerous important application areas such as image based rendering and digital libraries.
Abstract: From the Publisher: The accessible presentation of this book gives both a general view of the entire computer vision enterprise and also offers sufficient detail to be able to build useful applications. Users learn techniques that have proven to be useful by first-hand experience and a wide range of mathematical methods. A CD-ROM with every copy of the text contains source code for programming practice, color images, and illustrative movies. Comprehensive and up-to-date, this book includes essential topics that either reflect practical significance or are of theoretical importance. Topics are discussed in substantial and increasing depth. Application surveys describe numerous important application areas such as image based rendering and digital libraries. Many important algorithms broken down and illustrated in pseudo code. Appropriate for use by engineers as a comprehensive reference to the computer vision enterprise.

3,627 citations

Reference EntryDOI
15 Oct 2004

2,118 citations

Journal Article
TL;DR: A new approach to visual navigation under changing conditions dubbed SeqSLAM, which removes the need for global matching performance by the vision front-end - instead it must only pick the best match within any short sequence of images.
Abstract: Learning and then recognizing a route, whether travelled during the day or at night, in clear or inclement weather, and in summer or winter is a challenging task for state of the art algorithms in computer vision and robotics. In this paper, we present a new approach to visual navigation under changing conditions dubbed SeqSLAM. Instead of calculating the single location most likely given a current image, our approach calculates the best candidate matching location within every local navigation sequence. Localization is then achieved by recognizing coherent sequences of these “local best matches”. This approach removes the need for global matching performance by the vision front-end - instead it must only pick the best match within any short sequence of images. The approach is applicable over environment changes that render traditional feature-based techniques ineffective. Using two car-mounted camera datasets we demonstrate the effectiveness of the algorithm and compare it to one of the most successful feature-based SLAM algorithms, FAB-MAP. The perceptual change in the datasets is extreme; repeated traverses through environments during the day and then in the middle of the night, at times separated by months or years and in opposite seasons, and in clear weather and extremely heavy rain. While the feature-based method fails, the sequence-based algorithm is able to match trajectory segments at 100% precision with recall rates of up to 60%.

686 citations

01 Jan 2016
TL;DR: The the radon transform and some of its applications is universally compatible with any devices to read and an online access to it is set as public so you can get it instantly.
Abstract: the radon transform and some of its applications is available in our book collection an online access to it is set as public so you can get it instantly. Our digital library spans in multiple locations, allowing you to get the most less latency time to download any of our books like this one. Kindly say, the the radon transform and some of its applications is universally compatible with any devices to read.

212 citations

Journal ArticleDOI
Sungmin Lee1, Seok Min Yun1, Ju-Hun Nam, Chee Sun Won1, Seung-Won Jung1 
TL;DR: A detailed survey and experimental analysis on DCP-based methods will help readers understand the effectiveness of the individual step of the dehazing process and will facilitate development of advanced dehazed algorithms.
Abstract: The presence of haze in the atmosphere degrades the quality of images captured by visible camera sensors. The removal of haze, called dehazing, is typically performed under the physical degradation model, which necessitates a solution of an ill-posed inverse problem. To relieve the difficulty of the inverse problem, a novel prior called dark channel prior (DCP) was recently proposed and has received a great deal of attention. The DCP is derived from the characteristic of natural outdoor images that the intensity value of at least one color channel within a local window is close to zero. Based on the DCP, the dehazing is accomplished through four major steps: atmospheric light estimation, transmission map estimation, transmission map refinement, and image reconstruction. This four-step dehazing process makes it possible to provide a step-by-step approach to the complex solution of the ill-posed inverse problem. This also enables us to shed light on the systematic contributions of recent researches related to the DCP for each step of the dehazing process. Our detailed survey and experimental analysis on DCP-based methods will help readers understand the effectiveness of the individual step of the dehazing process and will facilitate development of advanced dehazing algorithms.

174 citations