scispace - formally typeset
Search or ask a question
Proceedings ArticleDOI

Gender and ethnicity classification of Iris images using deep class-encoder

TL;DR: Deep Class Encoder as mentioned in this paper uses class labels to learn discriminative representation for the given sample by mapping the learned feature vector to its label, which leads to improved recognition performance, reduced computation time and faster processing of test samples.
Abstract: Soft biometric modalities have shown their utility in different applications including reducing the search space significantly. This leads to improved recognition performance, reduced computation time, and faster processing of test samples. Some common soft biometric modalities are ethnicity, gender, age, hair color, iris color, presence of facial hair or moles, and markers. This research focuses on performing ethnicity and gender classification on iris images. We present a novel supervised auto-encoder based approach, Deep Class-Encoder, which uses class labels to learn discriminative representation for the given sample by mapping the learned feature vector to its label. The proposed model is evaluated on two datasets each for ethnicity and gender classification. The results obtained using the proposed Deep Class-Encoder demonstrate its effectiveness in comparison to existing approaches and state-of-the-art methods.
Citations
More filters
Journal ArticleDOI
TL;DR: Taxonomy for conceptual categorisation of methods with the aim of reducing or optimising the computational workload, and thereby speeding-up the identification transactions, in biometric identification systems are presented.
Abstract: The computational workload is one of the key challenges in biometric identification systems. The naive retrieval method based on an exhaustive search becomes impractical with the growth of the number of the enrolled data subjects. Consequently, in recent years, many methods with the aim of reducing or optimising the computational workload, and thereby speeding-up the identification transactions, in biometric identification systems have been developed. In this article, taxonomy for conceptual categorisation of such methods is presented, followed by a comprehensive survey of the relevant academic publications, including computational workload reduction and software/hardware-based acceleration. Lastly, the pertinent technical considerations and trade-offs of the surveyed methods are discussed, along with an industry perspective, and open issues/challenges in the field.

34 citations

Journal ArticleDOI
TL;DR: An overview of the existing publicly available datasets and their popularity in the research community using a bibliometric approach is provided to help investigators conducting research in the domain of iris recognition to identify relevant datasets.

28 citations

Journal ArticleDOI
TL;DR: This paper presents compelling experimental evidences in terms of correct classification accuracy (CCR) on CASIA-IrisV4 Twins’ dataset that manifest the effectiveness of ocular biometrics in identifying twins and quantified the cross-domain capability of the proposed subnetworks (i.e. CNN and MCNN) on ND-GFI dataset that outperforms the state-of-the-art methods.
Abstract: In this paper, we tackle the twins identification problem; a challenging task in biometric authentication. The experiments are carried out using deep learning techniques and for that we have proposed a Convolutional Siamese Network (CNN-Siamese) and a multiscale Convolutional Siamese network (MCNN-Siamese). We have also explored and evaluated three pre-trained CNNs namely, ResNet-50, VGG-16 and NASNet-Large, and utilized them in Siamese network after appropriate modifications. In addition, a simple 5-layer neural network (sNN) is also utilised as Siamese subnetwork in the experiments. We have presented compelling experimental evidences in terms of correct classification accuracy (CCR) on CASIA-IrisV4 Twins’ dataset that manifest the effectiveness of ocular biometrics in identifying twins. The results achieve the existing state-of-the-art human level accuracy. Also, the proximity of CCRs of all the models asserts that the ocular regions in twins hold a significant correlation to label them as twins. We have also quantified the cross-domain capability of the proposed subnetworks (i.e. CNN and MCNN) on ND-GFI dataset that outperforms the state-of-the-art methods. Notably, for the ocular biometrics, to the best of our knowledge, there is currently no literature available as of date that explores the association between twins and subsequently unriddles the classification using Deep Learning.

18 citations

Journal ArticleDOI
TL;DR: A comprehensive study of the effects of different pre-training over the employed architectures is carried out, showing that, in most cases, a better accuracy is obtained after the networks have been fine-tuned for face recognition.
Abstract: We address the use of selfie ocular images captured with smartphones to estimate age and gender. Partial face occlusion has become an issue due to the mandatory use of face masks. Also, the use of mobile devices has exploded, with the pandemic further accelerating the migration to digital services. However, state‐of‐the‐art solutions in related tasks such as identity or expression recognition employ large Convolutional Neural Networks, whose use in mobile devices is infeasible due to hardware limitations and size restrictions of downloadable applications. To counteract this, we adapt two existing lightweight CNNs proposed in the context of the ImageNet Challenge, and two additional architectures proposed for mobile face recognition. Since datasets for soft‐biometrics prediction using selfie images are limited, we counteract over‐fitting by using networks pre‐trained on ImageNet. Furthermore, some networks are further pre‐trained for face recognition, for which very large training databases are available. Since both tasks employ similar input data, we hypothesise that such strategy can be beneficial for soft‐biometrics estimation. A comprehensive study of the effects of different pre‐training over the employed architectures is carried out, showing that, in most cases, a better accuracy is obtained after the networks have been fine‐tuned for face recognition. [ABSTRACT FROM AUTHOR] Copyright of IET Biometrics (Wiley-Blackwell) is the property of Wiley-Blackwell and its content may not be copied or emailed to multiple sites or posted to a listserv without the copyright holder's express written permission. However, users may print, download, or email articles for individual use. This abstract may be abridged. No warranty is given about the accuracy of the copy. Users should refer to the original published version of the material for the full abstract. (Copyright applies to all Abstracts.)

16 citations

Proceedings ArticleDOI
01 Jan 2019
TL;DR: In this paper, the authors used probabilistic occlusion masking to gain insight on the discriminative power of the iris texture for gender prediction, and found that the gender related information is primarily in the periocular region.
Abstract: Predicting gender from iris images has been reported by several researchers as an application of machine learning in biometrics. Recent works on this topic have suggested that the preponderance of the gender cues is located in the periocular region rather than in the iris texture itself. This paper focuses on teasing out whether the information for gender prediction is in the texture of the iris stroma, the periocular region, or both. We present a larger dataset for gender from iris, and evaluate gender prediction accuracy using linear SVM and CNN, comparing hand-crafted and deep features. We use probabilistic occlusion masking to gain insight on the problem. Results suggest the discriminative power of the iris texture for gender is weaker than previously thought, and that the gender-related information is primarily in the periocular region.

14 citations

References
More filters
Proceedings Article
03 Dec 2012
TL;DR: The state-of-the-art performance of CNNs was achieved by Deep Convolutional Neural Networks (DCNNs) as discussed by the authors, which consists of five convolutional layers, some of which are followed by max-pooling layers, and three fully-connected layers with a final 1000-way softmax.
Abstract: We trained a large, deep convolutional neural network to classify the 1.2 million high-resolution images in the ImageNet LSVRC-2010 contest into the 1000 different classes. On the test data, we achieved top-1 and top-5 error rates of 37.5% and 17.0% which is considerably better than the previous state-of-the-art. The neural network, which has 60 million parameters and 650,000 neurons, consists of five convolutional layers, some of which are followed by max-pooling layers, and three fully-connected layers with a final 1000-way softmax. To make training faster, we used non-saturating neurons and a very efficient GPU implementation of the convolution operation. To reduce overriding in the fully-connected layers we employed a recently-developed regularization method called "dropout" that proved to be very effective. We also entered a variant of this model in the ILSVRC-2012 competition and achieved a winning top-5 test error rate of 15.3%, compared to 26.2% achieved by the second-best entry.

73,978 citations

Journal ArticleDOI
28 May 2015-Nature
TL;DR: Deep learning is making major advances in solving problems that have resisted the best attempts of the artificial intelligence community for many years, and will have many more successes in the near future because it requires very little engineering by hand and can easily take advantage of increases in the amount of available computation and data.
Abstract: Deep learning allows computational models that are composed of multiple processing layers to learn representations of data with multiple levels of abstraction. These methods have dramatically improved the state-of-the-art in speech recognition, visual object recognition, object detection and many other domains such as drug discovery and genomics. Deep learning discovers intricate structure in large data sets by using the backpropagation algorithm to indicate how a machine should change its internal parameters that are used to compute the representation in each layer from the representation in the previous layer. Deep convolutional nets have brought about breakthroughs in processing images, video, speech and audio, whereas recurrent nets have shone light on sequential data such as text and speech.

46,982 citations

Book
23 May 2011
TL;DR: It is argued that the alternating direction method of multipliers is well suited to distributed convex optimization, and in particular to large-scale problems arising in statistics, machine learning, and related areas.
Abstract: Many problems of recent interest in statistics and machine learning can be posed in the framework of convex optimization. Due to the explosion in size and complexity of modern datasets, it is increasingly important to be able to solve problems with a very large number of features or training examples. As a result, both the decentralized collection or storage of these datasets as well as accompanying distributed solution methods are either necessary or at least highly desirable. In this review, we argue that the alternating direction method of multipliers is well suited to distributed convex optimization, and in particular to large-scale problems arising in statistics, machine learning, and related areas. The method was developed in the 1970s, with roots in the 1950s, and is equivalent or closely related to many other algorithms, such as dual decomposition, the method of multipliers, Douglas–Rachford splitting, Spingarn's method of partial inverses, Dykstra's alternating projections, Bregman iterative algorithms for l1 problems, proximal methods, and others. After briefly surveying the theory and history of the algorithm, we discuss applications to a wide variety of statistical and machine learning problems of recent interest, including the lasso, sparse logistic regression, basis pursuit, covariance selection, support vector machines, and many others. We also discuss general distributed optimization, extensions to the nonconvex setting, and efficient implementation, including some details on distributed MPI and Hadoop MapReduce implementations.

17,433 citations

Journal ArticleDOI
28 Jul 2006-Science
TL;DR: In this article, an effective way of initializing the weights that allows deep autoencoder networks to learn low-dimensional codes that work much better than principal components analysis as a tool to reduce the dimensionality of data is described.
Abstract: High-dimensional data can be converted to low-dimensional codes by training a multilayer neural network with a small central layer to reconstruct high-dimensional input vectors. Gradient descent can be used for fine-tuning the weights in such "autoencoder" networks, but this works well only if the initial weights are close to a good solution. We describe an effective way of initializing the weights that allows deep autoencoder networks to learn low-dimensional codes that work much better than principal components analysis as a tool to reduce the dimensionality of data.

16,717 citations

Journal ArticleDOI
TL;DR: A fast, greedy algorithm is derived that can learn deep, directed belief networks one layer at a time, provided the top two layers form an undirected associative memory.
Abstract: We show how to use "complementary priors" to eliminate the explaining-away effects that make inference difficult in densely connected belief nets that have many hidden layers. Using complementary priors, we derive a fast, greedy algorithm that can learn deep, directed belief networks one layer at a time, provided the top two layers form an undirected associative memory. The fast, greedy algorithm is used to initialize a slower learning procedure that fine-tunes the weights using a contrastive version of the wake-sleep algorithm. After fine-tuning, a network with three hidden layers forms a very good generative model of the joint distribution of handwritten digit images and their labels. This generative model gives better digit classification than the best discriminative learning algorithms. The low-dimensional manifolds on which the digits lie are modeled by long ravines in the free-energy landscape of the top-level associative memory, and it is easy to explore these ravines by using the directed connections to display what the associative memory has in mind.

15,055 citations