scholarly journals Adaptive Deep Supervised Autoencoder Based Image Reconstruction for Face Recognition

2016 ◽  
Vol 2016 ◽  
pp. 1-14 ◽  
Author(s):  
Rongbing Huang ◽  
Chang Liu ◽  
Guoqi Li ◽  
Jiliu Zhou

Based on a special type of denoising autoencoder (DAE) and image reconstruction, we present a novel supervised deep learning framework for face recognition (FR). Unlike existing deep autoencoder which is unsupervised face recognition method, the proposed method takes class label information from training samples into account in the deep learning procedure and can automatically discover the underlying nonlinear manifold structures. Specifically, we define an Adaptive Deep Supervised Network Template (ADSNT) with the supervised autoencoder which is trained to extract characteristic features from corrupted/clean facial images and reconstruct the corresponding similar facial images. The reconstruction is realized by a so-called “bottleneck” neural network that learns to map face images into a low-dimensional vector and reconstruct the respective corresponding face images from the mapping vectors. Having trained the ADSNT, a new face image can then be recognized by comparing its reconstruction image with individual gallery images, respectively. Extensive experiments on three databases including AR, PubFig, and Extended Yale B demonstrate that the proposed method can significantly improve the accuracy of face recognition under enormous illumination, pose change, and a fraction of occlusion.

Author(s):  
Amal A. Moustafa ◽  
Ahmed Elnakib ◽  
Nihal F. F. Areed

This paper presents a methodology for Age-Invariant Face Recognition (AIFR), based on the optimization of deep learning features. The proposed method extracts deep learning features using transfer deep learning, extracted from the unprocessed face images. To optimize the extracted features, a Genetic Algorithm (GA) procedure is designed in order to select the most relevant features to the problem of identifying a person based on his/her facial images over different ages. For classification, K-Nearest Neighbor (KNN) classifiers with different distance metrics are investigated, i.e., Correlation, Euclidian, Cosine, and Manhattan distance metrics. Experimental results using a Manhattan distance KNN classifier achieves the best Rank-1 recognition rate of 86.2% and 96% on the standard FGNET and MORPH datasets, respectively. Compared to the state-of-the-art methods, our proposed method needs no preprocessing stages. In addition, the experiments show its privilege over other related methods.


2018 ◽  
Vol 2018 ◽  
pp. 1-8 ◽  
Author(s):  
Hae-Min Moon ◽  
Min-Gu Kim ◽  
Ju-Hyun Shin ◽  
Sung Bum Pan

In recent years, various studies have been conducted to provide a real-time service based on face recognition in Internet of things environments such as in a smart home environment. In particular, face recognition in a network-based surveillance camera environment can significantly change the performance or utilization of face recognition technology because the size of image information to be transmitted varies depending on the communication capabilities. In this paper, we propose a multiresolution face recognition method that uses virtual facial images by distance as learning to solve the problem of low recognition rate caused by communication, camera, and distance change. Face images for each virtual distance are generated through clarity and image degradation for each resolution, using a single high-resolution face image. The proposed method achieved a performance that was 5.9% more accurate than methods using MPCA and SVM, when LDA and the Euclidean distance were employed for a DB that was configured using faces that were acquired from the real environments of five different streets.


2019 ◽  
Vol 8 (3) ◽  
pp. 1932-1938

In this work, deep learning methods are used to classify the facial images. ORL Database is used for the purpose of training the models and for testing. Three kinds of models are developed and their performances are measured. Convolutional Neural Networks (CNN), Convolutional Neural Network Based Inception Model with single training image per class (CNN-INC) and Convolutional Neural Network Based Inception Model with several training images per class (CNN-INC-MEAN) are developed. The ORL database has ten facial images for each person. Five images are used for training purpose and remaining 5 images are used for testing. The five images for the training are chosen randomly so that two sets of training and testing data is generated. The models are trained and tested on the two sets that are drawn from the same population. The results are presented for accuracy of face recognition


Author(s):  
PEI CHEN ◽  
DAVID SUTER

Illumination effects, including shadows and varying lighting, make the problem of face recognition challenging. Experimental and theoretical results show that the face images under different illumination conditions approximately lie in a low-dimensional subspace, hence principal component analysis (PCA) or low-dimensional subspace techniques have been used. Following this spirit, we propose new techniques for the face recognition problem, including an outlier detection strategy (mainly for those points not following the Lambertian reflectance model), and a new error criterion for the recognition algorithm. Experiments using the Yale-B face database show the effectiveness of the new strategies.


2019 ◽  
Vol 16 (10) ◽  
pp. 4309-4312
Author(s):  
Rajeshwar Moghekar ◽  
Sachin Ahuja

Face recognition from videos is challenging problem as the face image captured has variations in terms of pose, Occlusion, blur and resolution. It has many applications including security monitoring and authentication. A subset of Indian Movies Face database (IMFDB) which has collection of face images retrieved from movie/video of actors which vary in terms of blur, pose, noise and illumination is used in our work. Our work focuses on the use of pre-trained deep learning models and applies transfer learning to the features extracted from the CNN layers. Later we compare it Fine tuned model. The results show that the accuracy is 99.89 using CNN as feature extractor and 96.3 when we fine tune the VGG-Face. The Fine tuned network of VGG-Face learnt more generic features when compared with its counterpart transfer learning. When applied on VGG16 transfer learning achieved 93.9.


Author(s):  
Wencan Zhong ◽  
Vijayalakshmi G. V. Mahesh ◽  
Alex Noel Joseph Raj ◽  
Nersisson Ruban

Finding faces in the clutter scenes is a challenging task in automatic face recognition systems as facial images are subjected to changes in the illumination, facial expression, orientation, and occlusions. Also, in the cluttered scenes, faces are not completely visible and detecting them is essential as it is significant in surveillance applications to study the mood of the crowd. This chapter utilizes the deep learning methods to understand the cluttered scenes to find the faces and discriminate them into partial and full faces. The work proves that MTCNN used for detecting the faces and Zernike moments-based kernels employed in CNN for classifying the faces into partial and full takes advantage in delivering a notable performance as compared to the other techniques. Considering the limitation of recognition on partial face emotions, only the full faces are preserved, and further, the KDEF dataset is modified by MTCNN to detect only faces and classify them into four emotions. PatternNet is utilized to train and test the modified dataset to improve the accuracy of the results.


Symmetry ◽  
2019 ◽  
Vol 11 (10) ◽  
pp. 1234 ◽  
Author(s):  
Jo ◽  
Kim

Face recognition using a near-infrared (NIR) sensor is widely applied to practical applications such as mobile unlocking or access control. However, unlike RGB sensors, few deep learning approaches have studied NIR face recognition. We conducted comparative experiments for the application of deep learning to NIR face recognition. To accomplish this, we gathered five public databases and trained two deep learning architectures. In our experiments, we found that simple architecture could have a competitive performance on the NIR face databases that are mostly composed of frontal face images. Furthermore, we propose a data augmentation method to train the architectures to improve recognition of users who wear glasses. With this augmented training set, the recognition rate for users who wear glasses increased by up to 16%. This result implies that the recognition of those who wear glasses can be overcome using this simple method without constructing an additional training set. Furthermore, the model that uses augmented data has symmetry with those trained with real glasses-wearing data regarding the recognition of people who wear glasses.


2016 ◽  
Vol 13 (10) ◽  
pp. 7696-7700
Author(s):  
Xuansheng Wang ◽  
Zheqi Lin ◽  
Zhen Chen ◽  
Zhongming Teng

Flexible Manifold Embedding (FME) is a semi-supervised manifold learning framework with good applicability. FME can effectively exploit label information from labeled data as well as a manifold structure from both labeled and unlabeled data to reduce the dimension of data. FME employs a simple but effective linear regression function to map new data points and it aims to find optimal prediction labels F for all training samples X, the linear regression h(X) and the regression residue F0 = F− h(X). In this paper, we proposed a novel idea to improve the accuracy of FME. We produce the mirror image of the face and then integrate the original face image and its mirror image to extend training samples X. The mirror image shares the same class label with the original face image, so to simultaneously use the mirror image and original face image can partially eliminate the side-effect on face recognition of the variation of poses and illuminations of original face images of the same person. The proposed method combines the advantages of the FME and extended training samples for face recognition. Moreover, the mirror image used in the proposed method can also enhance the robustness of the method. Experiments carried out on public face databases show that the proposed method can obtain notable accuracy improvement.


Author(s):  
Yunke Li ◽  
Hongyuan Shi ◽  
Liang Chen ◽  
Fan Jiang

Convolutional neural networks (CNN) are widely used deep learning frameworks and are applied in the field of face recognition, achieving outstanding results. The Macropixel comparison approach is a shallow mathematical approach that recognizes faces by comparing the original pixel blocks of face images. In this article, the authors are inspired by ideas of the currently popular deep neural network framework and introduce two features into the mathematical approach: deep overlap and weighted filter. The aim is to explore if the idea of deep learning could benefit mathematical recognition method, which might extend the scope of face recognition research. Results from the experiments show that the new proposed approach achieves markedly better recognition rates than the original macropixel methods.


Deep learning has attracted several researchers in the field of computer vision due to its ability to perform face and object recognition tasks with high accuracy than the traditional shallow learning systems. The convolutional layers present in the deep learning systems help to successfully capture the distinctive features of the face. For biometric authentication, face recognition (FR) has been preferred due to its passive nature. Processing face images are accompanied by a series of complexities, like variation of pose, light, face expression, and make up. Although all aspects are important, the one that impacts the most face-related computer vision applications is pose. In face recognition, it has been long desired to have a method capable of bringing faces to the same pose, usually a frontal view, in order to ease recognition. Synthesizing different views of a face is still a great challenge, mostly because in nonfrontal face images there are loss of information when one side of the face occludes the other. Most solutions for FR fail to perform well in cases involving extreme pose variations as in such scenarios, the convolutional layers of the deep models are unable to find discriminative parts of the face for extracting information. Most of the architectures proposed earlier deal with the scenarios where the face images used for training as well as testing the deep learning models are frontal and nearfrontal. On the contrary, here a limited number of face images at different poses is used to train the model, where a number of separate generator models learn to map a single face image at any arbitrary pose to specific poses and the discriminator performs the task of face recognition along with discriminating a synthetic face from a realworld sample. To this end, this paper proposes a representation learning by rotating the face. Here an encoderdecoder structure of the generator enables to learn a representation that is both generative and discriminative, which can be used for face image synthesis and pose-invariant face recognition. This representation is explicitly disentangled from other face variations such as pose, through the pose code provided to the decoder and pose estimation in the discriminator.


Sign in / Sign up

Export Citation Format

Share Document