Other affiliations: Tseung Kwan O Hospital, University of Sydney, University of Technology, Sydney ...read more
Bio: Kin-Man Lam is an academic researcher from Hong Kong Polytechnic University. The author has contributed to research in topics: Facial recognition system & Feature extraction. The author has an hindex of 44, co-authored 316 publications receiving 6226 citations. Previous affiliations of Kin-Man Lam include Tseung Kwan O Hospital & University of Sydney.
Papers published on a yearly basis
••08 Sep 2018
TL;DR: This paper proposes a fast video salient object detection model, based on a novel recurrent network architecture, named Pyramid Dilated Bidirectional ConvLSTM (PDB-ConvL STM), which achieves state-of-the-art results on two popular benchmarks, well demonstrating its superior performance and high applicability.
Abstract: This paper proposes a fast video salient object detection model, based on a novel recurrent network architecture, named Pyramid Dilated Bidirectional ConvLSTM (PDB-ConvLSTM). A Pyramid Dilated Convolution (PDC) module is first designed for simultaneously extracting spatial features at multiple scales. These spatial features are then concatenated and fed into an extended Deeper Bidirectional ConvLSTM (DB-ConvLSTM) to learn spatiotemporal information. Forward and backward ConvLSTM units are placed in two layers and connected in a cascaded way, encouraging information flow between the bi-directional streams and leading to deeper feature extraction. We further augment DB-ConvLSTM with a PDC-like structure, by adopting several dilated DB-ConvLSTMs to extract multi-scale spatiotemporal information. Extensive experimental results show that our method outperforms previous video saliency models in a large margin, with a real-time speed of 20 fps on a single GPU. With unsupervised video object segmentation as an example application, the proposed model (with a CRF-based post-process) achieves state-of-the-art results on two popular benchmarks, well demonstrating its superior performance and high applicability.
TL;DR: The proposed level set method can be directly applied to simultaneous segmentation and bias correction for 3 and 7T magnetic resonance images and demonstrates the superiority of the proposed method over other representative algorithms.
Abstract: It is often a difficult task to accurately segment images with intensity inhomogeneity, because most of representative algorithms are region-based that depend on intensity homogeneity of the interested object. In this paper, we present a novel level set method for image segmentation in the presence of intensity inhomogeneity. The inhomogeneous objects are modeled as Gaussian distributions of different means and variances in which a sliding window is used to map the original image into another domain, where the intensity distribution of each object is still Gaussian but better separated. The means of the Gaussian distributions in the transformed domain can be adaptively estimated by multiplying a bias field with the original signal within the window. A maximum likelihood energy functional is then defined on the whole image region, which combines the bias field, the level set function, and the piecewise constant function approximating the true image signal. The proposed level set method can be directly applied to simultaneous segmentation and bias correction for 3 and 7T magnetic resonance images. Extensive evaluation on synthetic and real-images demonstrate the superiority of the proposed method over other representative algorithms.
TL;DR: The corner detection scheme introduced in this paper can provide accurate information about the corners and accurately locate the templates in relation to the eye images and greatly reduce the processing time for the templates.
Abstract: Facial feature extraction is an important step in automated visual interpretation and human face recognition. Among the facial features, the eye plays the most important part in the recognition process. The deformable template can be used in extracting the eye boundaries. However, the weaknesses of the deformable template are that the processing time is lengthy and that its success relies on the initial position of the template. In this paper, the head boundary is first located in a head-and-shoulders image. The approximate positions of the eyes are estimated by means of average anthropometric measures. Corners, the salient features of the eyes, are detected and used to set the initial parameters of the eye templates. The corner detection scheme introduced in this paper can provide accurate information about the corners. Based on the corner positions, we can accurately locate the templates in relation to the eye images and greatly reduce the processing time for the templates. The performance of the deformable template is assessed with and without using the information on corner positions. Experiments show that a saving in execution time of about 40% on average and a better eye boundary representation can be achieved by using the corner information.
TL;DR: An analytic-to-holistic approach which can identify faces at different perspective variations is proposed, and it is shown that this approach can achieve a similar level of performance from different viewing directions of a face.
Abstract: We propose an analytic-to-holistic approach which can identify faces at different perspective variations. The database for the test consists of 40 frontal-view faces. The first step is to locate 15 feature points on a face. A head model is proposed, and the rotation of the face can be estimated using geometrical measurements. The positions of the feature points are adjusted so that their corresponding positions for the frontal view are approximated. These feature points are then compared with the feature points of the faces in a database using a similarity transform. In the second step, we set up windows for the eyes, nose, and mouth. These feature windows are compared with those in the database by correlation. Results show that this approach can achieve a similar level of performance from different viewing directions of a face. Under different perspective variations, the overall recognition rates are over 84 percent and 96 percent for the first and the first three likely matched faces, respectively.
TL;DR: An efficient algorithm for human face detection and facial feature extraction is devised using the genetic algorithm and the eigenface technique, and the lighting effect and orientation of the faces are considered and solved.
Abstract: In this paper, an efficient algorithm for human face detection and facial feature extraction is devised. Firstly, the location of the face regions is detected using the genetic algorithm and the eigenface technique. The genetic algorithm is applied to search for possible face regions in an image, while the eigenface technique is used to determine the fitness of the regions. As the genetic algorithm is computationally intensive, the searching space is reduced and limited to the eye regions so that the required timing is greatly reduced. Possible face candidates are then further verified by measuring their symmetries and determining the existence of the different facial features. Furthermore, in order to improve the level of detection reliability in our approach, the lighting effect and orientation of the faces are considered and solved.
01 Jan 2015
TL;DR: In this article, the authors categorize and evaluate face detection algorithms and discuss relevant issues such as data collection, evaluation metrics and benchmarking, and conclude with several promising directions for future research.
Abstract: Images containing faces are essential to intelligent vision-based human-computer interaction, and research efforts in face processing include face recognition, face tracking, pose estimation and expression recognition. However, many reported methods assume that the faces in an image or an image sequence have been identified and localized. To build fully automated systems that analyze the information contained in face images, robust and efficient face detection algorithms are required. Given a single image, the goal of face detection is to identify all image regions which contain a face, regardless of its 3D position, orientation and lighting conditions. Such a problem is challenging because faces are non-rigid and have a high degree of variability in size, shape, color and texture. Numerous techniques have been developed to detect faces in a single image, and the purpose of this paper is to categorize and evaluate these algorithms. We also discuss relevant issues such as data collection, evaluation metrics and benchmarking. After analyzing these algorithms and identifying their limitations, we conclude with several promising directions for future research.
01 Jan 2006