scispace - formally typeset
Search or ask a question

Showing papers on "Facial recognition system published in 2022"


Journal ArticleDOI
TL;DR: A novel Multi-feature Fusion and Decomposition framework for age-invariant face recognition, which learns more discriminative and robust features and reduces the intra-class variants and improves the AIFR performance than previous state-of-the-art methods.
Abstract: Different from general face recognition, age-invariant face recognition (AIFR) aims at matching faces with a big age gap. Previous discriminative methods usually focus on decomposing facial feature into age-related and age-invariant components, which suffer from the loss of facial identity information. In this article, we propose a novel Multi-feature Fusion and Decomposition (MFD) framework for age-invariant face recognition, which learns more discriminative and robust features and reduces the intra-class variants. Specifically, we first sample multiple face images of different ages with the same identity as a face time sequence. Then, the multi-head attention is employed to capture contextual information from facial feature series, extracted by the backbone network. Next, we combine feature decomposition with fusion based on the face time sequence to ensure that the final age-independent features effectively represent the identity information of the face and have stronger robustness against the aging process. Besides, we also mitigate imbalanced age distribution in the training data by a re-weighted age loss. We experimented with the proposed MFD over the popular CACD and CACD-VS datasets, where we show that our approach improves the AIFR performance than previous state-of-the-art methods. We simultaneously show the performance of MFD on LFW dataset.

57 citations


Journal ArticleDOI
TL;DR: In this article , an additive angular margin loss (ArcFace) was proposed to enhance the discriminative power of the face recognition model by encouraging one dominant sub-class that contains the majority of clean faces and non-dominant sub-classes that include hard or noisy faces.
Abstract: Recently, a popular line of research in face recognition is adopting margins in the well-established softmax loss function to maximize class separability. In this paper, we first introduce an Additive Angular Margin Loss (ArcFace), which not only has a clear geometric interpretation but also significantly enhances the discriminative power. Since ArcFace is susceptible to the massive label noise, we further propose sub-center ArcFace, in which each class contains K sub-centers and training samples only need to be close to any of the K positive sub-centers. Sub-center ArcFace encourages one dominant sub-class that contains the majority of clean faces and non-dominant sub-classes that include hard or noisy faces. Based on this self-propelled isolation, we boost the performance through automatically purifying raw web faces under massive real-world noise. Besides discriminative feature embedding, we also explore the inverse problem, mapping feature vectors to face images. Without training any additional generator or discriminator, the pre-trained ArcFace model can generate identity-preserved face images for both subjects inside and outside the training data only by using the network gradient and Batch Normalization (BN) priors. Extensive experiments demonstrate that ArcFace can enhance the discriminative feature embedding as well as strengthen the generative face synthesis.

43 citations


Journal ArticleDOI
TL;DR: Zhang et al. as mentioned in this paper proposed a pseudo RGB-D face recognition framework and provided data-driven ways to generate the depth maps from 2D face images, which can teach computers to generate correct answers by training.
Abstract: In the last decade, advances and popularity of low-cost RGB-D sensors have enabled us to acquire depth information of objects. Consequently, researchers began to solve face recognition problems by capturing RGB-D face images using these sensors. Until now, it is not easy to acquire the depth of human faces because of limitations imposed by privacy policies, and RGB face images are still more common. Therefore, obtaining the depth map directly from the corresponding RGB image could be helpful to improve the performance of subsequent face processing tasks, such as face recognition. Intelligent creatures can use a large amount of experience to obtain 3D spatial information only from 2D plane scenes. It is machine learning methodology, which is to solve such problems, that can teach computers to generate correct answers by training. To replace the depth sensors by generated pseudo-depth maps, in this article, we propose a pseudo RGB-D face recognition framework and provide data-driven ways to generate the depth maps from 2D face images. Specially we design and implement a generative adversarial network model named “D+GAN” to perform the multiconditional image-to-image translation with face attributes. By this means, we validate the pseudo RGB-D face recognition with experiments on various datasets. With the cooperation of image fusion technologies, especially non-subsampled shearlet transform (NSST), the accuracy of face recognition has been significantly improved.

42 citations


Journal ArticleDOI
TL;DR: In this article , the authors proposed an IDL-ERCFI technique, which is based on intelligent DL, to distinguish and classify ethnicity based on facial photos using face landmarks to align photos before sending them to the network.

39 citations


Journal ArticleDOI
TL;DR: Wang et al. as discussed by the authors proposed the Embedding Unmasking Model (EUM) operated on top of existing face recognition models, which enabled the EUM to produce embeddings similar to these of unmasked faces of the same identities.

39 citations


Journal ArticleDOI
TL;DR: Wang et al. as mentioned in this paper proposed a novel DeepMasknet framework capable of both the face mask detection and masked facial recognition, which can detect the people not wearing the face masks and recognizing different persons while wearing the mask.

37 citations


Proceedings ArticleDOI
01 Jun 2022
TL;DR: ElasticFace as mentioned in this paper relaxes the fixed penalty margin constrain by using random margin values drawn from a normal distribution in each training iteration to give the decision boundary chances to extract and retract to allow space for flexible class separability learning.
Abstract: Learning discriminative face features plays a major role in building high-performing face recognition models. The recent state-of-the-art face recognition solutions proposed to incorporate a fixed penalty margin on commonly used classification loss function, softmax loss, in the normalized hypersphere to increase the discriminative power of face recognition models, by minimizing the intra-class variation and maximizing the inter-class variation. Marginal penalty softmax losses, such as ArcFace and CosFace, assume that the geodesic distance between and within the different identities can be equally learned using a fixed penalty margin. However, such a learning objective is not realistic for real data with inconsistent inter-and intra-class variation, which might limit the discriminative and generalizability of the face recognition model. In this paper, we relax the fixed penalty margin constrain by proposing elastic penalty margin loss (ElasticFace) that allows flexibility in the push for class separability. The main idea is to utilize random margin values drawn from a normal distribution in each training iteration. This aims at giving the decision boundary chances to extract and retract to allow space for flexible class separability learning. We demonstrate the superiority of our ElasticFace loss over ArcFace and CosFace losses, using the same geometric transformation, on a large set of mainstream benchmarks. From a wider perspective, our ElasticFace has advanced the state-of-the-art face recognition performance on seven out of nine mainstream benchmarks. All training codes, pre-trained models, training logs will be publicly released 1.

35 citations


Journal ArticleDOI
TL;DR: In this paper , the authors proposed an occluded expression recognition model based on the generated countermeasure network, which is divided into two modules, namely, occlussed face image restoration and face recognition.

33 citations


Journal ArticleDOI
TL;DR: In this paper, a series of experiments comparing the masked face recognition performances of CNN architectures available in literature and exploring possible alterations in loss functions, architectures, and training methods that can enable existing methods to fully extract and leverage the limited facial information available in a masked face.

28 citations


Journal ArticleDOI
11 Feb 2022-PLOS ONE
TL;DR: In this paper , the authors investigated whether facial emotion recognition is impaired for faces wearing a mask compared to uncovered faces, in a sample of 790 participants between 18 and 89 years (condition mask vs. original).
Abstract: Facial emotion recognition is crucial for social interaction. However, in times of a global pandemic, where wearing a face mask covering mouth and nose is widely encouraged to prevent the spread of disease, successful emotion recognition may be challenging. In the current study, we investigated whether emotion recognition, assessed by a validated emotion recognition task, is impaired for faces wearing a mask compared to uncovered faces, in a sample of 790 participants between 18 and 89 years (condition mask vs. original). In two more samples of 395 and 388 participants between 18 and 70 years, we assessed emotion recognition performance for faces that are occluded by something other than a mask, i.e., a bubble as well as only showing the upper part of the faces (condition half vs. bubble). Additionally, perception of threat for faces with and without occlusion was assessed. We found impaired emotion recognition for faces wearing a mask compared to faces without mask, for all emotions tested (anger, fear, happiness, sadness, disgust, neutral). Further, we observed that perception of threat was altered for faces wearing a mask. Upon comparison of the different types of occlusion, we found that, for most emotions and especially for disgust, there seems to be an effect that can be ascribed to the face mask specifically, both for emotion recognition performance and perception of threat. Methodological constraints as well as the importance of wearing a mask despite temporarily compromised social interaction are discussed.

28 citations


Journal ArticleDOI
TL;DR: In this paper , a series of experiments comparing the masked face recognition performances of CNN architectures were conducted and exploring possible alterations in loss functions, architectures, and training methods that can enable existing methods to fully extract and leverage the limited facial information available in a masked face.

Journal ArticleDOI
TL;DR: In this paper , a lightweight A-MobileNet model is proposed to enhance the local feature extraction of facial expressions, where the center loss and softmax loss are combined to optimize the model parameters to reduce intra-class distance and increase interclass distance.
Abstract: Facial expression recognition (FER) is to separate the specific expression state from the given static image or video to determine the psychological emotions of the recognized object, the realization of the computer's understanding and recognition of facial expressions have fundamentally changed the relationship between human and computer, to achieve better human computer interaction (HCI). In recent years, FER has attracted widespread attention in the fields of HCI, security, communications and driving, and has become one of the research hotspots. In the mobile Internet era, the need for lightweight networking and real-time performance is growing. In this paper, a lightweight A-MobileNet model is proposed. First, the attention module is introduced into the MobileNetV1 model to enhance the local feature extraction of facial expressions. Then, the center loss and softmax loss are combined to optimize the model parameters to reduce intra-class distance and increase inter-class distance. Compared with the original MobileNet series models, our method significantly improves recognition accuracy without increasing the number of model parameters. Compared with others, A- MobileNet model achieves better results on the FERPlus and RAF-DB datasets.

Journal ArticleDOI
TL;DR: Zhang et al. as mentioned in this paper proposed a novel face recognition method that is robust to occlusions based on a single end-to-end deep neural network, which learns to discover the corrupted features from the deep convolutional neural networks, and clean them by the dynamically learned masks.
Abstract: With the recent advancement of deep convolutional neural networks, significant progress has been made in general face recognition. However, the state-of-the-art general face recognition models do not generalize well to occluded face images, which are exactly the common cases in real-world scenarios. The potential reasons are the absences of large-scale occluded face data for training and specific designs for tackling corrupted features brought by occlusions. This article presents a novel face recognition method that is robust to occlusions based on a single end-to-end deep neural network. Our approach, named FROM (Face Recognition with Occlusion Masks), learns to discover the corrupted features from the deep convolutional neural networks, and clean them by the dynamically learned masks. In addition, we construct massive occluded face images to train FROM effectively and efficiently. FROM is simple yet powerful compared to the existing methods that either rely on external detectors to discover the occlusions or employ shallow models which are less discriminative. Experimental results on the LFW, Megaface challenge 1, RMF2, AR dataset and other simulated occluded/masked datasets confirm that FROM dramatically improves the accuracy under occlusions, and generalizes well on general face recognition.

Journal ArticleDOI
TL;DR: A CCTV image-based human face recognition system using different techniques for feature extraction and face recognition, which recognized faces with a minimum computing time and an accuracy of more than 90%.
Abstract: This paper aims to develop a machine learning and deep learning-based real-time framework for detecting and recognizing human faces in closed-circuit television (CCTV) images. The traditional CCTV system needs a human for 24/7 monitoring, which is costly and insufficient. The automatic recognition system of faces in CCTV images with minimum human intervention and reduced cost can help many organizations, such as law enforcement, identifying the suspects, missing people, and people entering a restricted territory. However, image-based recognition has many issues, such as scaling, rotation, cluttered backgrounds, and variation in light intensity. This paper aims to develop a CCTV image-based human face recognition system using different techniques for feature extraction and face recognition. The proposed system includes image acquisition from CCTV, image preprocessing, face detection, localization, extraction from the acquired images, and recognition. We use two feature extraction algorithms, principal component analysis (PCA) and convolutional neural network (CNN). We use and compare the performance of the algorithms K-nearest neighbor (KNN), decision tree, random forest, and CNN. The recognition is done by applying these techniques to the dataset with more than 40K acquired real-time images at different settings such as light level, rotation, and scaling for simulation and performance evaluation. Finally, we recognized faces with a minimum computing time and an accuracy of more than 90%.

Journal ArticleDOI
TL;DR: In this paper , the influence of 47 attributes on the verification performance of two popular face recognition models was investigated on the publicly available MAAD-Face attribute database with over 120M high-quality attribute annotations.
Abstract: Face recognition (FR) systems have a growing effect on critical decision-making processes. Recent works have shown that FR solutions show strong performance differences based on the user’s demographics. However, to enable a trustworthy FR technology, it is essential to know the influence of an extended range of facial attributes on FR beyond demographics. Therefore, in this work, we analyze FR bias over a wide range of attributes. We investigate the influence of 47 attributes on the verification performance of two popular FR models. The experiments were performed on the publicly available MAAD-Face attribute database with over 120M high-quality attribute annotations. To prevent misleading statements about biased performances, we introduced control group-based validity values to decide if unbalanced test data causes the performance differences. The results demonstrate that also many nondemographic attributes strongly affect recognition performance, such as accessories, hairstyles and colors, face shapes, or facial anomalies. The observations of this work show the strong need for further advances in making the FR system more robust, explainable, and fair. Moreover, our findings might help to a better understanding of how FR networks work, enhance the robustness of these networks, and develop more generalized bias-mitigating FR solutions.

Proceedings ArticleDOI
Anchun Cheng1
21 Jan 2022
TL;DR: Wang et al. as discussed by the authors introduced three main learning models in deep learning: convolutional neural networks, recurrent neural networks and generative adversarial networks, and provided a comparative analysis of these three learning models.
Abstract: Deep learning is a technical tool with broad application prospects and has an important role in the field of image recognition. In view of the theoretical value and practical significance of image recognition technology in promoting the development of computer vision and artificial intelligence, this paper will review and study the application of deep learning in image recognition. This paper first outlines the development of icon recognition technology, and then introduces three main learning models in deep learning: convolutional neural networks, recurrent neural networks, and generative adversarial networks, and provides a comparative analysis of these three learning models. Finally, the research results of deep learning image recognition application fields, such as face recognition, medical image recognition, and remote sensing image classification, are analyzed and discussed. This paper also analyze the development trend of deep learning in the field of image recognition, and conclude that the future development direction is the effective recognition of video images and the theoretical strengthening of models.

Journal ArticleDOI
TL;DR: Wang et al. as mentioned in this paper proposed a model inversion attack (MIA) which aims to reveal the identity of a targeted user by generating the most proper datapoint input to the system with maximum corresponding confidence score at the output.
Abstract: Cybersecurity in front of attacks to a face recognition system is an emerging issue in the cloud era, especially due to its strong bonds with the privacy of the users registered to the system. A possible attack is the model inversion attack (MIA) which aims to reveal the identity of a targeted user by generating the most proper datapoint input to the system with maximum corresponding confidence score at the output. The generated data of a registered user can be maliciously used as a serious invasion of the user privacy. In literature, MIA processes are categorized into white-box and black-box scenarios which are respectively with and without information about the system structure, parameters, and partially about the users. This research work assumes the MIA under semi-white box scenario of availability of system model structure and parameters but not any user data information, and verifies it as a severe threat even for a deep-learning-based face recognition system despite its complex structure and the diversity of registered user data. The alert state is promoted by Deep MIA which is the integration of deep generative models in MIA, and $\alpha $ -GAN integrated MIA-initilized by a face based seed ( $\alpha $ -GAN-MIA-FS) is proposed. As a novel MIA search strategy, a pre-trained deep generative model with capability of generating a face image from a random feature vector is used for narrowing down the image search space to the feature vectors space, which has much lower dimensions. This allows the MIA process to efficiently search for a low-dimensional feature vector whose corresponding face image maximizes the confidence score. We have experimentally evaluated the proposed method by two objective criteria and three subjective criteria in comparison to $\alpha $ -GAN-integrated MIA initialized with a random seed ( $\alpha $ -GAN-MIA-RS), DCGAN-integrated MIA (DCGAN-MIA), and the conventional MIA. The evaluation results approve the efficiency and superiority of the proposed technique in generating natural looking face clones with high recognizability as the targeted users.

Journal ArticleDOI
TL;DR: Wang et al. as discussed by the authors proposed an efficient face recognition system based on Gabor filter bank and a deep learning method known as Sparse AutoEncoder (SAE), the main aim of the proposed system is to improve the features extracted by Gabor Filter Bank using SAE method.
Abstract: These days, face recognition systems are widely being employed in various daily applications such as smart phone unlocking, tracking school attendance, and secure online bank transactions, smarter border control, to name a few. In spite of the remarkable progress, face recognition systems still suffer from occlusions, light variations, camera types and their resolutions. Face recognition is still a dynamic research field. In this paper, we propose an efficient face recognition system based on Gabor filter bank and a deep learning method known as Sparse AutoEncoder (SAE). The main aim of the proposed system is to improve the features extracted by Gabor filter bank using SAE method. Then, these enhanced features are subjected to features reduction using principal component analysis and linear discriminant analysis (PCA + LDA) technique. Finally, the matching stage is accomplished via cosine Mahalanobis distance. Experiments on seven publicly available databases (i.e., JAFFE, AT&T, Yale, Georgia Tech, CASIA, Extended Yale, Essex) show that the proposed system can achieve promising results with the combination of Gabor and SAE, as well as outperform previously proposed methods.

Journal ArticleDOI
TL;DR: The statistical analysis is used to extract and select the statistical features, whereas, the SVM algorithm is employed to merge and classify the different features in order to increase the quality of the information and to obtain an optimal Human face recognition.

Journal ArticleDOI
TL;DR: In this paper , a dual-branch training strategy was proposed to guide the model to focus on the upper half of the face to extract robust features for Masked face recognition.

Proceedings ArticleDOI
13 Oct 2022
TL;DR: In this article , a framework for secure face recognition (SoF) is proposed, which allows verified customers to enter a private residence using a TensorFlow-based classifier.
Abstract: Internet-based innovation has advanced significantly during the past decade. As a result, advances in security technology have become a crucial resource for protecting our daily lives. In this research, we suggest a framework for secure face recognition (SoF). In particular, we tailor this infrastructure to allowing verified customers entry into a private residence. In order to get the classifier ready, another flexible learning technique is used. For starters, we get our preliminary data from social groups and institutions. As the client continues to make use of the framework, the classifier's accuracy increases. The classifier model has been improved by employing an epic technique that makes use of human interaction and online community. Using the powerful learning framework TensorFlow, the system may be easily repurposed to work with a wide variety of devices and applications.

Journal ArticleDOI
TL;DR: The novelty of the approach in this work is the inclusion of hyperparameter tuning using a nature-inspired optimization algorithm, which is an important and essential step in discovering the optimal hyperparameters for training the model which in turn increases the accuracy.
Abstract: The Intelligent Transportation System (ITS) is said to revolutionize the travel experience by making it safe, secure, and comfortable for the people. Although vehicles have been automated up to a certain extent, it still has critical security issues that require thorough study and advanced solutions. The security vulnerabilities of ITS allows the attacker to steal the vehicle. Therefore, the identification of drivers is required in order to develop a safe and secure system so that the vehicles can be protected from theft. There are two ways in which a driver can be identified: 1) face recognition of the driver, and 2) based on driving behavior. Face recognition includes image processing of 2-D images and learning of the features, which require high computational power. Drivers are known to have unique driving styles, whose data can be captured by the sensors. Therefore, the second method identifies drivers based on the analysis of the sensor data and it requires comparatively lesser computational power. In this paper, an optimized deep learning model is trained on the sensor data to correctly identify the drivers. The Long Short-Term Memory (LSTM) deep learning model is optimized for better performance. The novelty of the approach in this work is the inclusion of hyperparameter tuning using a nature-inspired optimization algorithm, which is an important and essential step in discovering the optimal hyperparameters for training the model which in turn increases the accuracy. The CAN-BUS dataset is used for experimentation and evaluation of the training model. Evaluation parameters such as accuracy, precision score, F1 score, and ROC AUC curve are considered to evaluate the performance of the model.

Journal ArticleDOI
TL;DR: Wang et al. as mentioned in this paper designed a feature set-based representation learning (FSRL) scheme to collaboratively represent the hierarchical features for more accurate recognition, and they used the primitive form of feature maps to keep the latent structural information.
Abstract: Cross-resolution face recognition (CRFR), which is important in intelligent surveillance and biometric forensics, refers to the problem of matching a low-resolution (LR) probe face image against high-resolution (HR) gallery face images. Existing shallow learning-based and deep learning-based methods focus on mapping the HR-LR face pairs into a joint feature space where the resolution discrepancy is mitigated. However, little works consider how to extract and utilize the intermediate discriminative features from the noisy LR query faces to further mitigate the resolution discrepancy due to the resolution limitations. In this study, we desire to fully exploit the multi-level deep convolutional neural network (CNN) feature set for robust CRFR. In particular, our contributions are threefold. (i) To learn more robust and discriminative features, we desire to adaptively fuse the contextual features from different layers. (ii) To fully exploit these contextual features, we design a feature set-based representation learning (FSRL) scheme to collaboratively represent the hierarchical features for more accurate recognition. Moreover, FSRL utilizes the primitive form of feature maps to keep the latent structural information, especially in noisy cases. (iii) To further promote the recognition performance, we desire to fuse the hierarchical recognition outputs from different stages. Meanwhile, the discriminability from different scales can also be fully integrated. By exploiting these advantages, the efficiency of the proposed method can be delivered. Experimental results on several face datasets have verified the superiority of the presented algorithm to the other competitive CRFR approaches.

Journal ArticleDOI
TL;DR: The experimental results established that the GWOECN-FR technology outperformed more contemporary approaches in real-time face recognition, and was primarily concerned with reliably and rapidly recognizing faces in input photos.
Abstract: Face recognition (FR) is a technique for recognizing individuals through the use of face photographs. The FR technology is widely applicable in a variety of fields, including security, biometrics, authentication, law enforcement, smart cards, and surveillance. Recent advances in deep learning (DL) models, particularly convolutional neural networks (CNNs), have demonstrated promising results in the field of FR. CNN models that have been pretrained can be utilized to extract characteristics for effective FR. In this regard, this research introduces the GWOECN-FR approach, a unique grey wolf optimization with an enhanced capsule network-based deep transfer learning model for real-time face recognition. The proposed GWOECN-FR approach is primarily concerned with reliably and rapidly recognizing faces in input photos. Additionally, the GWOECN-FR approach is preprocessed in two steps, namely, data augmentation and noise reduction by bilateral filtering (BF). Additionally, for feature vector extraction, an expanded capsule network (ECN) model can be used. Additionally, grey wolf optimization (GWO) combined with a stacked autoencoder (SAE) model is used to identify and classify faces in images. The GWO algorithm is used to optimize the SAE model’s weight and bias settings. The GWOECN-FR technique’s performance is validated using a benchmark dataset, and the results are analyzed in a variety of aspects. The GWOECN-FR approach achieved a TST of 0.03 s on the FEI dataset, whereas the AlexNet-SVM, ResNet-SVM, and AlexNet models achieved TSTs of 0.125 s, 0.0051 s, and 0.0062 s, respectively. The experimental results established that the GWOECN-FR technology outperformed more contemporary approaches.

Journal ArticleDOI
TL;DR: This work presents significant literacy calculations used in facial protestation for exact distinctive verification and acknowledgment that can effectively and capably see sentiments from the vibes of the client.
Abstract: This work proposes the implementation of the idea of real-time human emotion recognition through digital image processing techniques using CNN. This work presents significant literacy calculations used in facial protestation for exact distinctive verification and acknowledgment that can effectively and capably see sentiments from the vibes of the client. The proposed model gives six probability values based on six different expressions. Large datasets are explored and investigated for training facial emotion recognition model. In support of this work, CNN using Deep learning model, OpenCV, Tensorflow, Keras, Pandas, and Numpy is used for digital computer vision procedures involved, and an lite experiment is conducted for various men and women of different age, race, and colour to descry their feelings and variations for different faces are found. This work is improved in 3 targets as face location, acknowledgment and feeling arrangement. Open CV library, and facial expression images dataset are used in this proposed work. Also python writing computer programs is utilized for computer vision (using webcam) procedures. To demonstrate ongoing adequacy, an investigation is directed for a very long time to distinguish their internal feelings and track down physiological changes for each face. The consequences of the examinations exhibit the idealizations in face investigation framework. At long last, the exhibition of programmed face detection and recognition are measured with very high accuracy and in real-time. This method can be implemented and is widely useful in various domains such as security, schools, colleges and universities, military, airlines, banking etc.

Book ChapterDOI
01 Jan 2022
TL;DR: In this article, the authors proposed a hybrid method for face recognition using Principal Component Analysis (PCA), Discrete Wavelet Transform (DWT), Independent Component Analysis and Support Vector Machine (SVM).
Abstract: Face recognition need is fine assured as enormous industrial relevance use them to implement one or another objective. As the programmes move closer to everyday usage to hold a database of actual events, an individual’s identification primarily demanded as an instance of consistency. As facial recognition has beating advantages over other industrial applications and human eyes can quickly evaluate performance, improved algorithms and smaller computing costs are continuously improving this methodology. This research takes the conventional algorithms of recognition in the first stage and uses hybrid approaches to counter their limitations. The study starts with basic computation of global face features using Principal Component Analysis (PCA), Discrete Wavelet Transform (DWT) and Independent Component Analysis (ICA), with some standard classifiers like Neural Network (NN) and Support Vector Machine (SVM). As the learning rate is high in machine learning, then the system’s accuracy goes high, but increases the area and cost overhead. Fusion-based methods have been proposed in further work to overcome that training limitation, based on Harris corner, Speed up Robust Features (SURF) and DWT + PCA system model where only 10% training sample has been taken on Essex database, and 99.45% accuracy is achieved. Creating the Fusion rule requires some hit and trial methods that may not be Universal in every database. To overcome this limitation further an efficient Hybrid method proposed which elaborates the local features Linear Binary Pattern (LBP), Histogram Oriented Gradients (HOG), Gabor wavelet and global features (DWT, PCA) of the face. Further, this feature trained with Neural Network classifier to obtained better accuracy nearly 99.40% with single image training from each class.

Journal ArticleDOI
TL;DR: Li et al. as mentioned in this paper proposed a cattle face recognition model based on a two-branch convolutional neural network (TB-CNN), which collected two cattle face images from different angles are input to the convolution network of different channels for feature extraction, the features of the two channels are feature fused, and the global average pooling layer is combined with the classifier to identify the individual cattle.

Journal ArticleDOI
TL;DR: In this paper , a discriminative learning method based on triplet loss function and a sensitive triplet generator is proposed to improve both the accuracy and fairness of biased face recognition algorithms.
Abstract: We propose a discrimination-aware learning method to improve both the accuracy and fairness of biased face recognition algorithms. The most popular face recognition benchmarks assume a distribution of subjects without paying much attention to their demographic attributes. In this work, we perform a comprehensive discrimination-aware experimentation of deep learning-based face recognition. We also propose a notational framework for algorithmic discrimination with application to face biometrics. The experiments include three popular face recognition models and three public databases composed of 64,000 identities from different demographic groups characterized by sex and ethnicity. We experimentally show that learning processes based on the most used face databases have led to popular pre-trained deep face models that present evidence of strong algorithmic discrimination. Finally, we propose a discrimination-aware learning method, Sensitive Loss, based on the popular triplet loss function and a sensitive triplet generator. Our approach works as an add-on to pre-trained networks and is used to improve their performance in terms of average accuracy and fairness. The method shows results comparable to state-of-the-art de-biasing networks and represents a step forward to prevent discriminatory automatic systems.

Proceedings ArticleDOI
01 Jun 2022
TL;DR: AdaFace as discussed by the authors proposes a new loss function that emphasizes samples of different difficulties based on their image quality, which achieves this in the form of an adaptive margin function by approximating the image quality with feature norms.
Abstract: Recognition in low quality face datasets is challenging because facial attributes are obscured and degraded. Advances in margin-based loss functions have resulted in enhanced discriminability of faces in the embedding space. Further, previous studies have studied the effect of adaptive losses to assign more importance to misclassified (hard) examples. In this work, we introduce another aspect of adaptiveness in the loss function, namely the image quality. We argue that the strategy to emphasize misclassified samples should be adjusted according to their image quality. Specifically, the relative importance of easy or hard samples should be based on the sample's image quality. We propose a new loss function that emphasizes samples of different difficulties based on their image quality. Our method achieves this in the form of an adaptive margin function by approximating the image quality with feature norms. Extensive experiments show that our method, AdaFace, improves the face recognition performance over the state-of-the-art (SoTA) on four datasets (IJB-B, IJB-C, IJB-S and TinyFace). Code and models are released in Supp.

Journal ArticleDOI
TL;DR: Zhang et al. as mentioned in this paper proposed a deep Age-Invariant Model (AIM) for face recognition in the wild with three distinct novelties: first, a unified deep architecture jointly performs cross-age face synthesis and recognition in a mutual boosting way.
Abstract: Despite the remarkable progress in face recognition related technologies, reliably recognizing faces across ages remains a big challenge. The appearance of a human face changes substantially over time, resulting in significant intra-class variations. As opposed to current techniques for age-invariant face recognition, which either directly extract age-invariant features for recognition, or first synthesize a face that matches target age before feature extraction, we argue that it is more desirable to perform both tasks jointly so that they can leverage each other. To this end, we propose a deep Age-Invariant Model (AIM) for face recognition in the wild with three distinct novelties. First, AIM presents a novel unified deep architecture jointly performing cross-age face synthesis and recognition in a mutual boosting way. Second, AIM achieves continuous face rejuvenation/aging with remarkable photorealistic and identity-preserving properties, avoiding the requirement of paired data and the true age of testing samples. Third, effective and novel training strategies are developed for end-to-end learning of the whole deep architecture, which generates powerful age-invariant face representations explicitly disentangled from the age variation. Moreover, we construct a new large-scale Cross-Age Face Recognition (CAFR) benchmark dataset to facilitate existing efforts and push the frontiers of age-invariant face recognition research. Extensive experiments on both our CAFR dataset and several other cross-age datasets (MORPH, CACD, and FG-NET) demonstrate the superiority of the proposed AIM model over the state-of-the-arts. Benchmarking our model on the popular unconstrained face recognition datasets YTF and IJB-C additionally verifies its promising generalization ability in recognizing faces in the wild.