Dissertations / Theses on the topic 'Face Recognition Across Pose'
Create a spot-on reference in APA, MLA, Chicago, Harvard, and other styles
Consult the top 50 dissertations / theses for your research on the topic 'Face Recognition Across Pose.'
Next to every source in the list of references, there is an 'Add to bibliography' button. Press on it, and we will generate automatically the bibliographic reference to the chosen work in the citation style you need: APA, MLA, Harvard, Chicago, Vancouver, etc.
You can also download the full text of the academic publication as pdf and read online its abstract whenever available in the metadata.
Browse dissertations / theses on a wide variety of disciplines and organise your bibliography correctly.
Graham, Daniel B. "Pose-varying face recognition." Thesis, University of Manchester, 2000. http://ethos.bl.uk/OrderDetails.do?uin=uk.bl.ethos.488288.
Full textAbi, Antoun Ramzi. "Pose-Tolerant Face Recognition." Research Showcase @ CMU, 2013. http://repository.cmu.edu/dissertations/244.
Full textLincoln, Michael C. "Pose-independent face recognition." Thesis, University of Essex, 2003. http://ethos.bl.uk/OrderDetails.do?uin=uk.bl.ethos.250063.
Full textGodzich, Elliot J. "Automated Pose Correction for Face Recognition." Scholarship @ Claremont, 2012. http://scholarship.claremont.edu/cmc_theses/376.
Full textZhang, Xiaozheng. "Pose-invariant Face Recognition through 3D Reconstructions." Thesis, Griffith University, 2008. http://hdl.handle.net/10072/366373.
Full textThesis (PhD Doctorate)
Doctor of Philosophy (PhD)
School of Engineering
Science, Environment, Engineering and Technology
Full Text
Wibowo, Moh Edi. "Towards pose-robust face recognition on video." Thesis, Queensland University of Technology, 2014. https://eprints.qut.edu.au/77836/1/Moh%20Edi_Wibowo_Thesis.pdf.
Full textKumar, Sooraj. "Face recognition with variation in pose angle using face graphs /." Online version of thesis, 2009. http://hdl.handle.net/1850/9482.
Full textKing, Steve. "Robust face recognition under varying illumination and pose." Thesis, University of Huddersfield, 2005. http://ethos.bl.uk/OrderDetails.do?uin=uk.bl.ethos.417305.
Full textBeymer, David James. "Pose-invariant face recognition using real and virtual views." Thesis, Massachusetts Institute of Technology, 1995. http://hdl.handle.net/1721.1/38101.
Full textIncludes bibliographical references (p. 173-184).
by David James Beymer.
Ph.D.
Du, Shan. "Image-based face recognition under varying pose and illuminations conditions." Thesis, University of British Columbia, 2008. http://hdl.handle.net/2429/2814.
Full textRajwade, Ajit. "Facial pose estimation and face recognition from three-dimensional data." Thesis, McGill University, 2004. http://digitool.Library.McGill.CA:80/R/?func=dbin-jump-full&object_id=82410.
Full textFirstly, the thesis proposes a generic learning strategy using support vector regression [11] to estimate the approximate pose of a 3D scan. The support vector machine (SVM) is trained on range images in several poses, belonging to a small set of individuals. This thesis also examines the relationship between size of the range image and the accuracy of the pose prediction from the scan.
Secondly, a hierarchical two-step strategy is proposed to normalize a facial scan to a nearly frontal pose before performing recognition. The first step consists of a coarse normalization making use of either the spatial relationships between salient facial features or the generic learning algorithm using the SVM. This is followed by an iterative technique to refine the alignment to the frontal pose, which is basically an improved form of the Iterated Closest Point Algorithm [17]. The latter step produces a residual error value, which can be used as a metric to gauge the similarity between two faces. Our two-step approach is experimentally shown to outdo both the individual normalization methods in terms of recognition rates, over a very wide range of facial poses. Our strategy has been tested on a large database of 3D facial scans in which the training and test images of each individual were acquired at significantly different times, unlike several existing 3D face recognition methods.
Arashloo, Shervin Rahimzadeh. "Pose-invariant 2D face recognition by matching using graphical models." Thesis, University of Surrey, 2010. http://ethos.bl.uk/OrderDetails.do?uin=uk.bl.ethos.527013.
Full textEl, Seuofi Sherif M. "Performance Evaluation of Face Recognition Using Frames of Ten Pose Angles." Youngstown State University / OhioLINK, 2007. http://rave.ohiolink.edu/etdc/view?acc_num=ysu1198184813.
Full textZhao, Sanqiang. "On Sparse Point Representation for Face Localisation and Recognition." Thesis, Griffith University, 2009. http://hdl.handle.net/10072/366629.
Full textThesis (PhD Doctorate)
Doctor of Philosophy (PhD)
Griffith School of Engineering
Science, Environment, Engineering and Technology
Full Text
Lucey, Patrick Joseph. "Lipreading across multiple views." Thesis, Queensland University of Technology, 2007. https://eprints.qut.edu.au/16676/1/Patrick_Joseph_Lucey_Thesis.pdf.
Full textLucey, Patrick Joseph. "Lipreading across multiple views." Queensland University of Technology, 2007. http://eprints.qut.edu.au/16676/.
Full textHwang, June Youn. "Fast pose and automatic matching using hybrid method for the three dimensional face recognition." Thesis, University of Newcastle Upon Tyne, 2009. http://ethos.bl.uk/OrderDetails.do?uin=uk.bl.ethos.514469.
Full textZeni, Luis Felipe de Araujo. "Reconhecimento facial tolerante à variação de pose utilizando uma câmera RGB-D de baixo custo." reponame:Biblioteca Digital de Teses e Dissertações da UFRGS, 2014. http://hdl.handle.net/10183/101659.
Full textRecognizing the identity of human beings from recorded digital images of their faces is important for a variety of applications, namely, security access, human computer interation, digital entertainment, etc. This dissertation proposes a new method for automatic face recognition that uses both 2D and 3D information of an RGB-D(Kinect) camera. The method uses the color information of the 2D image to locate faces in the scene, once a face is properly located it is cut and normalized to a standard size and color. Afterwards, using depth information the method estimates the pose of the head relative to the camera. With the normalized faces and their respective pose information, the proposed method trains a model of faces that is robust to pose and expressions using a new automatic technique that separates different poses in different models of faces. With the trained model, the method is able to identify whether people used to train the model are present or not in new acquired images, which the model had no access during the training phase. The experiments demonstrate that the proposed method considerably improves the result of classification in real images with varying pose and expression.
Derkach, Dmytro. "Spectrum analysis methods for 3D facial expression recognition and head pose estimation." Doctoral thesis, Universitat Pompeu Fabra, 2018. http://hdl.handle.net/10803/664578.
Full textFacial analysis has attracted considerable research efforts over the last decades, with a growing interest in improving the interaction and cooperation between people and computers. This makes it necessary that automatic systems are able to react to things such as the head movements of a user or his/her emotions. Further, this should be done accurately and in unconstrained environments, which highlights the need for algorithms that can take full advantage of 3D data. These systems could be useful in multiple domains such as human-computer interaction, tutoring, interviewing, health-care, marketing etc. In this thesis, we focus on two aspects of facial analysis: expression recognition and head pose estimation. In both cases, we specifically target the use of 3D data and present contributions that aim to identify meaningful representations of the facial geometry based on spectral decomposition methods: 1. We propose a spectral representation framework for facial expression recognition using exclusively 3D geometry, which allows a complete description of the underlying surface that can be further tuned to the desired level of detail. It is based on the decomposition of local surface patches in their spatial frequency components, much like a Fourier transform, which are related to intrinsic characteristics of the surface. We propose the use of Graph Laplacian Features (GLFs), which result from the projection of local surface patches into a common basis obtained from the Graph Laplacian eigenspace. The proposed approach is tested in terms of expression and Action Unit recognition and results confirm that the proposed GLFs produce state-of-the-art recognition rates. 2. We propose an approach for head pose estimation that allows modeling the underlying manifold that results from general rotations in 3D. We start by building a fully-automatic system based on the combination of landmark detection and dictionary-based features, which obtained the best results in the FG2017 Head Pose Estimation Challenge. Then, we use tensor representation and higher order singular value decomposition to separate the subspaces that correspond to each rotation factor and show that each of them has a clear structure that can be modeled with trigonometric functions. Such representation provides a deep understanding of data behavior, and can be used to further improve the estimation of the head pose angles.
Cament, Riveros Leonardo. "Enhancements by weighted feature fusion, selection and active shape model for frontal and pose variation face recognition." Tesis, Universidad de Chile, 2015. http://repositorio.uchile.cl/handle/2250/132854.
Full textFace recognition is one of the most active areas of research in computer vision because of its wide range of possible applications in person identification, access control, human computer interfaces, and video search, among many others. Face identification is a one-to-n matching problem where a captured face is compared to n samples in a database. In this work a new method for robust face recognition is proposed. The methodology is divided in two parts, the first one focuses in face recognition robust to illumination, expression and small age variation and the second part focuses in pose variation. The proposed algorithm is based on Gabor features; which have been widely studied in face identification because of their good results and robustness. In the first part, a new method for face identification is proposed that combines local normalization for an illumination compensation stage, entropy-like weighted Gabor features for a feature extraction stage, and improvements in the Borda count classification through a threshold to eliminate low-score Gabor jets from the voting process. The FERET, AR, and FRGC 2.0 databases were used to test and compare the proposed method results with those previously published. Results on these databases show significant improvements relative to previously published results, reaching the best performance on the FERET and AR databases. Our proposed method also showed significant robustness to slight pose variations. The method was tested assuming noisy eye detection to check its robustness to inexact face alignment. Results show that the proposed method is robust to errors of up to three pixels in eye detection. However, face identification is strongly affected when the test images are very different from those of the gallery, as is the case in varying face pose. The second part of this work proposes a new 2D Gabor-based method which modifies the grid from which the Gabor features are extracted using a mesh to model face deformations produced by varying pose. Also, a statistical model of the Borda count scores computed by using the Gabor features is used to improve recognition performance across pose. The method was tested on the FERET and CMU-PIE databases, and the performance improvement provided by each block was assessed. The proposed method achieved the highest classification accuracy ever published on the FERET database with 2D face recognition methods. The performance obtained in the CMU-PIE database is among those obtained by the best published methods. Extensive experimental results are provided for different combinations of the proposed method, including results with two poses enrolled as a gallery.
Brown, Dane. "Faster upper body pose recognition and estimation using compute unified device architecture." Thesis, University of Western Cape, 2013. http://hdl.handle.net/11394/3455.
Full textThe SASL project is in the process of developing a machine translation system that can translate fully-fledged phrases between SASL and English in real-time. To-date, several systems have been developed by the project focusing on facial expression, hand shape, hand motion, hand orientation and hand location recognition and estimation. Achmed developed a highly accurate upper body pose recognition and estimation system. The system is capable of recognizing and estimating the location of the arms from a twodimensional video captured from a monocular view at an accuracy of 88%. The system operates at well below real-time speeds. This research aims to investigate the use of optimizations and parallel processing techniques using the CUDA framework on Achmed’s algorithm to achieve real-time upper body pose recognition and estimation. A detailed analysis of Achmed’s algorithm identified potential improvements to the algorithm. Are- implementation of Achmed’s algorithm on the CUDA framework, coupled with these improvements culminated in an enhanced upper body pose recognition and estimation system that operates in real-time with an increased accuracy.
Chu, Baptiste. "Neutralisation des expressions faciales pour améliorer la reconnaissance du visage." Thesis, Ecully, Ecole centrale de Lyon, 2015. http://www.theses.fr/2015ECDL0005/document.
Full textExpression and pose variations are major challenges for reliable face recognition (FR) in 2D. In this thesis, we aim to endow state of the art face recognition SDKs with robustness to simultaneous facial expression variations and pose changes by using an extended 3D Morphable Model (3DMM) which isolates identity variations from those due to facial expressions. Specifically, given a probe with expression, a novel view of the face is generated where the pose is rectified and the expression neutralized. We present two methods of expression neutralization. The first one uses prior knowledge to infer the neutral expression from an input image. The second method, specifically designed for verification, is based on the transfer of the gallery face expression to the probe. Experiments using rectified and neutralized view with a standard commercial FR SDK on two 2D face databases show significant performance improvement and demonstrates the effectiveness of the proposed approach. Then, we aim to endow the state of the art FR SDKs with the capabilities to recognize faces in videos. Finally, we present different methods for improving biometric performances for specific cases
Kramer, Annika. "Model based methods for locating, enhancing and recognising low resolution objects in video." Thesis, Curtin University, 2009. http://hdl.handle.net/20.500.11937/585.
Full textEmir, Alkazhami. "Facial Identity Embeddings for Deepfake Detection in Videos." Thesis, Linköpings universitet, Datorseende, 2020. http://urn.kb.se/resolve?urn=urn:nbn:se:liu:diva-170587.
Full textFiche, Cécile. "Repousser les limites de l'identification faciale en contexte de vidéo-surveillance." Thesis, Grenoble, 2012. http://www.theses.fr/2012GRENT005/document.
Full textThe person identification systems based on face recognition are becoming increasingly widespread and are being used in very diverse applications, particularly in the field of video surveillance. In this context, the performance of the facial recognition algorithms largely depends on the image acquisition context, especially because the pose can vary, but also because the acquisition methods themselves can introduce artifacts. The main issues are focus imprecision, which can lead to blurred images, or the errors related to compression, which can introduce the block artifact. The work done during the thesis focuses on facial recognition in images taken by video surveillance cameras, in cases where the images contain blur or block artifacts or show various poses. First, we are proposing a new approach that allows to significantly improve facial recognition in images with high blur levels or with strong block artifacts. The method, which makes use of specific noreference metrics, starts with the evaluation of the quality level of the input image and then adapts the training database of the recognition algorithms accordingly. Second, we have focused on the facial pose estimation. Normally, it is very difficult to recognize a face in an image taken from another viewpoint than the frontal one and the majority of facial identification algorithms which are robust to pose variation need to know the pose in order to achieve a satisfying recognition rate in a relatively short time. We have therefore developed a fast and satisfying pose estimation method based on recent recognition techniques
Zhang, Yuyao. "Non-linear dimensionality reduction and sparse representation models for facial analysis." Thesis, Lyon, INSA, 2014. http://www.theses.fr/2014ISAL0019/document.
Full textFace analysis techniques commonly require a proper representation of images by means of dimensionality reduction leading to embedded manifolds, which aims at capturing relevant characteristics of the signals. In this thesis, we first provide a comprehensive survey on the state of the art of embedded manifold models. Then, we introduce a novel non-linear embedding method, the Kernel Similarity Principal Component Analysis (KS-PCA), into Active Appearance Models, in order to model face appearances under variable illumination. The proposed algorithm successfully outperforms the traditional linear PCA transform to capture the salient features generated by different illuminations, and reconstruct the illuminated faces with high accuracy. We also consider the problem of automatically classifying human face poses from face views with varying illumination, as well as occlusion and noise. Based on the sparse representation methods, we propose two dictionary-learning frameworks for this pose classification problem. The first framework is the Adaptive Sparse Representation pose Classification (ASRC). It trains the dictionary via a linear model called Incremental Principal Component Analysis (Incremental PCA), tending to decrease the intra-class redundancy which may affect the classification performance, while keeping the extra-class redundancy which is critical for sparse representation. The other proposed work is the Dictionary-Learning Sparse Representation model (DLSR) that learns the dictionary with the aim of coinciding with the classification criterion. This training goal is achieved by the K-SVD algorithm. In a series of experiments, we show the performance of the two dictionary-learning methods which are respectively based on a linear transform and a sparse representation model. Besides, we propose a novel Dictionary Learning framework for Illumination Normalization (DL-IN). DL-IN based on sparse representation in terms of coupled dictionaries. The dictionary pairs are jointly optimized from normally illuminated and irregularly illuminated face image pairs. We further utilize a Gaussian Mixture Model (GMM) to enhance the framework's capability of modeling data under complex distribution. The GMM adapt each model to a part of the samples and then fuse them together. Experimental results demonstrate the effectiveness of the sparsity as a prior for patch-based illumination normalization for face images
Peng, Hsiao-Chia, and 彭小佳. "3D Face Reconstruction on RGB and RGB-D Images for Recognition Across Pose." Thesis, 2015. http://ndltd.ncl.edu.tw/handle/88142215912683274078.
Full text國立臺灣科技大學
機械工程系
103
Face recognition across pose is a challenging problem in computer vision. Two scenarios are considered in this thesis. One is the common setup with one single frontal facial image of each subject in the gallery set and the images of other poses in the probe set. The other considers a RGB-D image of the frontal face for each subject in the gallery, but the probe set is the same as in the previous case that only contains RGB images of other poses. The second scenario simulates the case that RGB-D camera can be available for user registration only and recognition can be performed on regular RGB images without the depth channel. Two approaches are proposed for handling the first scenario, one is holistic and the other is component-based. The former is extended from a face reconstruction approach and improved with different sets of landmarks for alignment and multiple reference models considered in the reconstruction phase. The latter focuses on the reconstruction of facial components obtained by the pose-invariant landmarks, and the recognition with different components considered at different poses. Such a component-based reconstruction for handling cross-pose recognition is rarely seen in the literature. Although the approach for handling the second scenario, i.e., the RGB-D based recognition, is partially similar to the approach for handling the first scenario, the novelty is on the handling of the depth readings corrupted by quantization noise, which are often encountered when the face is not close enough to the RGB-D camera at registration. An approach is proposed to resurface the corrupted depth map and substantially improve the recognition performance. All of the proposed approaches are evaluated on benchmark databases and proven comparable to state-of-the-art approaches.
Sanyal, Soubhik. "Discriminative Descriptors for Unconstrained Face and Object Recognition." Thesis, 2017. http://etd.iisc.ac.in/handle/2005/4177.
Full textLing-ying, Lee, and 李玲瑩. "Face Recognition Across Poses Using A Single Reference Model." Thesis, 2012. http://ndltd.ncl.edu.tw/handle/thxax6.
Full text國立臺灣科技大學
機械工程系
100
Given a frontal facial image as a gallery sample, a scheme is developed to generate novel views of the face for recognition across poses. The core part of the scheme is a recently published 3D face reconstruction which exploits a single reference 3D face model to build a 3D shape model for each face in the gallery set. The 3D shape model combined with the texture of each facial image in the gallery allows novel poses of the face to be generated. The LBP features are then extracted from these generated poses to train an SVM classifier for recognition. Assuming Lambertian surface with a reflectance function approximated by spherical harmonics, the 3D reference model would be made to deform so that the 2D projection of the deformed model can approximate the facial image in the gallery. The problem is cast as an image irradiance equation with unknown lighting, albedo, and surface normals. Using the reference model to estimate lighting, and providing an initial estimate of albedo, the reflectance function becomes only a function of the unknown surface normals, and the irradiance equation becomes a partial differential equation which is then solved for depth. A 3D face from the FRGC database is used as the reference model in the experiments, and the performance is evaluated on the PIE database. It is shown that the developed scheme gives a satisfactory performance, and can be further improved if the alignment between the reference model and the gallery image can be enhanced.
Beymer, David J. "Face Recognition Under Varying Pose." 1993. http://hdl.handle.net/1721.1/6621.
Full textYang, Feng. "Face recognition under significant pose variation." Thesis, 2007. http://spectrum.library.concordia.ca/975540/1/MR28958.pdf.
Full textChiu, Kuo-Yu, and 邱國育. "Face recognition system and its applications by using face pose estimation and face pose synthesis." Thesis, 2011. http://ndltd.ncl.edu.tw/handle/85747296243675764247.
Full text國立交通大學
電控工程研究所
99
In this dissertation, an improved face pose estimation algorithm to increase face recognition rate is proposed. There are three main stages. The first stage is chin curve estimation by using active contour model. The active contour model is auto-initialized to approach the chin curve under various face poses according to statistical experimental results. The second stage is the face pose estimation and synthesis. Using the chin contour information along with other facial features, simulated annealing algorithm is adopted to estimate various face poses. Using the face pose information, input face image with arbitrary face poses can be synthesized to be frontal. The last stage is face recognition. The synthesized frontal face pose image is utilized to solve the problem that face recognition rate is dramatically reduced when non-frontal face pose images are presented. From experimental results, it can be seen that the face recognition rates of traditional algorithms are only about 40%, while the proposed method greatly improves the recognition rate to about 80%. When face recognition system is applied in surveillance system, the recognition rates are 23% and 70% for traditional algorithm and the proposed system respectively.
WANG, HSIANG-JUNG, and 王湘蓉. "Multi-pose Face Recognition Using an Enhanced 3D Face Modeling." Thesis, 2016. http://ndltd.ncl.edu.tw/handle/36599347215052254848.
Full text開南大學
資訊學院碩士班
104
With the development of information technology, face recognition has reached a near perfect positive identification. However, there is recognition on the side faces some difficulty, because of the similarity of the structure of the face, in the case only in profile, lack of information can give more to deepen the side of face recognition difficulties, so in recent years began to study side face recognition based. In only a single front to capture an image, use the point cloud 3D modeling, in turn flank angle over the General Assembly to produce due to lack of depth of information generated holes, can reduce the degree of identification. In this study, improvement of traditional point cloud 3D face modeling, in addition to the hole in the side of the angle is too large to improve the result, and with reference to active shape model selection feature points 3D modeling training samples, and then SVM identity and face angle. Identification of the correct angle of classification results was 80.64%; facial recognition part, the false acceptance rate is 0%, the false rejection rate of 100%.
Beymer, David. "Pose-Invariant Face Recognition Using Real and Virtual Views." 1996. http://hdl.handle.net/1721.1/6772.
Full textJu-ChinChen and 陳洳瑾. "Subspace Learning for Face Detection, Recognition and Pose Estimation." Thesis, 2010. http://ndltd.ncl.edu.tw/handle/03701616334187369722.
Full text國立成功大學
資訊工程學系碩博士班
98
This thesis concerns the subspace learning methods on performing dimensionality reduction and extracting discriminant features for face detection, recognition and pose estimation. We examine the subspace learning methods and then the novel subspace learning methods are derived according to the data distribution of each application. The first application is based on analyzing the manifold in eigenspace to develop a statistic-based multi-view face detection and the pose estimation. In the eigenspace, not only the simple cascaded rejecter module can be developed to exclude 85% of the non-face images to enhance the overall system performance but also the manifold of face data can be applied to develop coarse pose estimation. In addition, to improve tolerance toward different partial occlusions and lighting conditions, the five-module detection system is based on significant local facial features (or subregions) rather than the entire face. In order to extract the low- and high-frequency feature information of each subregion of the facial image, the eigenspace and residual independent basis space are constructed. In addition, either projection weight vectors or coefficient vectors in the PCA (principal component analysis) or ICA (independent component analysis) space have divergent distributions and are therefore modeled by using the weighted Gaussian mixture model (GMM) with parameters estimated by Expectation-Maximization (EM) algorithm. Face detection is then performed by conducting a likelihood evaluation process based on the estimated joint probability of the weight and coefficient vectors and the corresponding geometric positions of the subregions. The use of subregion position information can reduce the risk of false acceptances. Following the use of PCA+ICA to model the face images, in the second application the kernel discriminant transformation (KDT) algorithm is proposed by extending the idea of canonical correlation analysis (CCA) of comparing facial image sets for face recognition. The recognition performance is rendered more robust by utilizing a set of test facial images characterized by arbitrary head poses, facial expressions and lighting conditions. Since the manifolds of the image sets in the training database are highly-overlapped and non-linearly distributed, each facial image set is non-linearly mapped into a high-dimensional space and a corresponding kernel subspace is then constructed using kernel principal component analysis (KPCA). To extract the discriminant features for recognition, a KDT matrix is proposed that maximizes the similarities of within-kernel subspaces and simultaneously minimizes those of between-kernel subspaces. While the KDT matrix cannot be computed explicitly in the high-dimensional feature space, an iterative kernel discriminant transformation algorithm is developed to solve the matrix in an implicit way. The proposed face recognition system is demonstrated to outperform existing still-image-based as well as image set-based recognition systems using the Yale face database B.
Hsu, Heng-Wei, and 許恆瑋. "Face Recognition Using Metric Learning with Head Pose Information." Thesis, 2018. http://ndltd.ncl.edu.tw/handle/69788s.
Full text國立交通大學
電子研究所
107
Face recognition has gained much interest recently and is widely used in daily applications such as video surveillance, applications for smart phones and airport security. Nevertheless, recognizing faces in large profile views still remains a hard problem since important features start to be obscured as a person’s head turns. This problem can be divided into two sub-problems: first, an accurate head pose estimation model is required to predict the angle given a face image. Second, a face recognition model that leverages the angle information is also needed to discriminate different people in different angles. In this dissertation, we aim to fulfill this gap. Instead of estimating the angle of head poses through a commonly used two-step process, a set of landmarks are first detected from faces then angles are estimated through these detected landmarks, we propose to directly predict angles from face images by training a deep convolutional neural network model. We further provide a metric learning based face recognition framework to leverage the angle information and improve the overall performance. Our contribution can be mainly divided into three parts: first, we propose a novel geometric loss for face recognition that explores the area relations within quadruplets of samples, which inherently considers the geometric characteristics of each sample set. The sampled quadruplet includes three positive samples and one negative sample which form a tetrahedron in the embedding space. The area of the triangular face formed by positive samples is minimized to reduce intraclass variations, whereas the areas of the triangular faces including the negative sample are maximized to increase interclass distances. With our area based objective function, the gradient of each sample considers its neighboring samples and adapts to local geometry which leads to improved performance. Second, we conduct an in-depth study of head pose estimation and present a multi-regression loss function, a L2 regression loss combined with an ordinal regression loss, to train a convolutional neural network (CNN) that is dedicated to estimating head poses from RGB images without depth information. The ordinal regression loss is utilized to address the non-stationary property observed as the facial features change with respect to different head pose angles and learn robust features. The L2 regression loss leverages these features to provide precise angle predictions for input images. To avoid the ambiguity problem in the commonly used Euler angle representation, we further formulate the head pose estimation problem in quaternions. Our quaternion-based multi-regression loss method achieves state-of-the-art performance on several public benchmark datasets. Third, we designed a sophisticated face recognition training framework. We start from data cleaning, an automatic method to deal with the labeling noise issue which most recent large datasets suffer. We then designed a data augmentation method that randomly augments the input image under various condition, such as adjusting the contrast, saturation, and the lighting condition of an image. Sharpening, blurring and noises are also applied to the images to simulate cases from different camera sources. The boundary values of the parameters for each image processing method are designed such that the resulting images are reasonable. Experiment results demonstrate that models trained with this kind of data augmentation show robust performance to unseen images. When training with large datasets, the size of the last fully connected layer for the classification loss are often large since the datasets consist of large number of identities. This makes the training process hard to converge, as the weights are randomly initialized. Thus we propose an iterative training and finetuning process that makes the training loss converge smoothly. Furthermore, to leverage the angle information for improving face recognition performance, we provide a detailed analysis of a metric learning based method that learns to minimize the distance between a person’s frontal and profile images. Qualitative and quantitative results are shown to demonstrate the effectiveness of our proposed training methodology. The following publications form the foundation of this thesis Heng-Wei Hsu, Tung-Yu Wu, Sheng Wan, Wing Hung Wong, and Chen-Yi Lee, “QuatNet: Quaternion-Based Head Pose Estimation With Multiregression Loss,” IEEE Transactions on Multimedia, Aug 2018. • Heng-Wei Hsu, Tung-Yu Wu, Wing Hung Wong, and Chen-Yi Lee, “Correlation-based Face Detection for Recognizing Faces in Videos,” in IEEE International Conference on Acoustics, Speech and Signal Processing (ICASSP), pp. 3101–3105, Apr 2018. • Heng-Wei Hsu, Tung-Yu Wu, Sheng Wan, Wing Hung Wong, and Chen-Yi Lee, “Deep Metric Learning with Geometric Loss,” under review. • Sheng Wan, Tung-Yu Wu, Heng-Wei Hsu, Yi-Wei Chen, Wing H. Wong, and Chen-Yi Lee, “Model-based JPEG for Convolutional Neural Network Classifications,” under review.
Chu, Tsu-ying, and 朱姿穎. "Correlation Filter for Face Recognition Across Illumination." Thesis, 2012. http://ndltd.ncl.edu.tw/handle/96q2k7.
Full text國立臺灣科技大學
機械工程系
100
Face recognition across illumination variation involves illumination normalization, feature extraction and classification. This research compares a few state-of-the-art illumination normalization methods, and selects the most potential one. We also investigate the impacts made by different facial regions on the recognition performance. Many believe that the facial region considered for face recognition is better bounded within the facial contour to minimize the degradation due to background and hair. However, we have found that the inclusion of the boundary of the forehead, contours of the cheeks, and the contour of the chin can effectively improve the performance. Minimum average correlation energy filter (MACE) combined with kernel class-dependence feature analysis (KCFA) is proven an effective solution, and therefore is adopted in this study with some minor modification. Following the protocol FGRC 2.0, the recognition rate can be improved from 72.91% to 84.83% using the recommended illumination normalization, and further improved to 88.17% with the recommended facial region.
Wang, Ding-En, and 王鼎恩. "Features Selection and GMM Classifier for Multi-Pose Face Recognition." Thesis, 2015. http://ndltd.ncl.edu.tw/handle/03408850317662581389.
Full text國立東華大學
資訊工程學系
103
Face recognition is widely used in security application, such as homeland security, video surveillance, law enforcement, and identity management. However, there are still some problems in face recognition system. The main problems include the light changes, facial expression changes, pose variations and partial occlusion. Although many face recognition approaches reported satisfactory performance, their successes are limited to the conditions of controlled environment. In fact, pose variation has been identified as one of the most current problems in the real world. Therefore, many algorithms focusing on how to handle pose variation have received much attention. To solve the pose variations problem, in this thesis, we propose a multi-pose face recognition system based on an effective design of classifier using SURF feature. In training phase, the proposed method utilizes SURF features to calculate similarity between two images from different poses of the same face. Face recognition model (GMM) is trained using the robust SURF features from different poses. In testing phase, feature vectors corresponding to the test images are input to all trained models for the decision of the recognized face. Experiment results show that the performance of the proposed method is better than other existing methods.
Zhuo, You-Lin, and 卓佑霖. "A Comparative Study on Face Recognition Across Illumination." Thesis, 2011. http://ndltd.ncl.edu.tw/handle/mp2v5w.
Full text國立臺灣科技大學
機械工程系
99
Face recognition across illumination is one of the most challenging problems in image-based face analysis. Most research focus on the methods for illumination normalization, illumination-invariant feature extraction, or classifier design, but few compare the performance of different approaches. This research evaluates and compares the performance of a few competitive approaches for illumination normalization and several methods for local feature extraction, aiming at determining an effective approach for face recognition across illumination. Because the other issue of the central concern of this research is the appropriateness of the determined approach in making a real-time system, the methods with high computational cost are excluded, although some may result in high recognition rates. The approach recommended by this comparison study can attain 85.19% in recognition rate on FRGC 2.0 database. With its relatively low computational cost, the approach is experimentally proven appropriate for making a real-time system.
Wu, Wei-Ting, and 吳韋霆. "Face Recognition Across Illumination Using Local DCT Features." Thesis, 2010. http://ndltd.ncl.edu.tw/handle/31101436077802528442.
Full text國立臺灣科技大學
機械工程系
98
Based on the holistic Log-DCT features, which are proven effective for face recognition across illumination conditions, this research considers the same features combined with local square patches for face recognition across illumination and with imprecise face localization. The objectives of this research include the following: (1) define the performance upper bound attainable by the combination of holistic Log-DCT features and local patches, (2) investigate the impacts on the performance from imprecise face localization contributed by a face detector. Satisfactory results are shown from the experiments on the CMU PIE database which offers faces with almost perfect localization, revealing that the combination of Log-DCT features and local patches can be an effective solution for recognizing perfectly localized faces across illumination. However, the performance degrades substantially when evaluating the combination on the FRGC 2.0 database, which offer faces with imprecise localization and variations on pose and expression, reflecting the fact that an actual face recognition system cannot leave alone these parameters. A local alignment and masking scheme is proposed to tackle the problems caused by these parameters, and is proven effective in an extensive experimental study.
Lin, Chiunhsiun, and 林群雄. "Face Detection, Pose Classification, and Face Recognition Based on Triangle Geometry and Color Features." Thesis, 2001. http://ndltd.ncl.edu.tw/handle/00914481948416413620.
Full text國立中央大學
資訊工程研究所
89
ABSTRACT In this dissertation, the problems of face detection, pose classification, and face recognition are studied and solved completely. The applications of face detection, pose classification, and face recognition are extended to various topics. The applications include: computer vision, security system, authentication for remove banking and access-control application. In the past, the problems of face detection, pose classification, and face recognition were introduced by numerous researches. However, experimental results reveal the practicability and competence of our proposed approaches in finding human face, pose classification, and face recognition. The feasibility and efficiency of our approaches is confirmed by experimental results. In this thesis, the relationship between two eyes and one mouth is shown clearly based on the geometrical structure of an isosceles triangle. The first proposed face detection system consists of two primary parts. The first part is to search for the potential face regions. The second part is to perform face verification. This system can conquer different size, different lighting condition, varying pose and expression, and noise and defocus problems. In addition to overcome the problem of partial occlusion of mouth and sunglasses, the system can also detect faces from the side view. Experimental results demonstrate that an approximately 98 % success rate is achieved. In addition, a new method of extracting the human-skin-like colors is proposed for reduction of the total computation effort in complicated surroundings. In this approach, skin-color-segmentation is used to remove the complex backgrounds according to the values of R, G, and B directly. This partition method reveals the skin-color-segmentation, which results in the saving of the total computation effort nearly by 80% in complicated backgrounds. The third chapter presents another novel face detection algorithm that is presented to locate multiple faces in color scenery images. A binary skin color map is first obtained in the color analysis stage. Then, color regions corresponding to the facial and non-facial areas in the color map are separated with a clustering-based splitting algorithm. Thereafter, an elliptic face model is devised to crop the real human faces through the shape location procedure. Last, local thresholding technique and a statistic-based verification procedure are utilized to confirm the human faces. The proposed detection algorithm combines both the color and shape properties of faces. In this work, the color span of human face can be expanded as wilder as possible to cover different faces by using the clustering-based splitting algorithm. Experimental results also reveal the feasibility of our proposed approach in solving face detection problem. The fourth chapter presents a method for automatic estimation of the poses/degrees of human faces. The proposed system consists of two primary parts. The first part is to search the potential face regions that are gotten from the isosceles-triangle criteria based on the rules of "the combination of two eyes and one mouth". The second part of the proposed system is to perform the task of pose verification by utilizing face weighting mask function, direction weighting mask function, and pose weighting mask function. The proposed face poses/degrees classification system can determine the poses of multiple faces. Experimental results demonstrate that an approximately 99 % success rate is achieved and the relative false estimation rate is very low. The fifth chapter presented a robust and efficient feature-based classification to recognize human faces embedded in photographs. The proposed system consists of two main parts. The first part is to detect the face regions. The second part of the proposed system is to perform the face recognition task. The proposed face recognition system can handle different size and different brightness conditions problems. Experimental results demonstrate that we can succeed overcome the various brightness conditions. Finally, conclusions and future works are given in Chapter 6.
Goren, Deborah. "Quantifying facial expression recognition across viewing conditions /." 2004. http://wwwlib.umi.com/cr/yorku/fullcit?pMQ99314.
Full textTypescript. Includes bibliographical references (leaves 59-66). Also available on the Internet. MODE OF ACCESS via web browser by entering the following URL: http://wwwlib.umi.com/cr/yorku/fullcit?pMQ99314
Wang, Shih Chieh, and 王仕傑. "Pose-Variant Face Recognition and Its Application to Human-Robot Interaction." Thesis, 2008. http://ndltd.ncl.edu.tw/handle/63299410428657032826.
Full text國立交通大學
電機與控制工程系所
97
In this thesis, a pose-variant face recognition system has been developed for human-robot interaction. In order to extract the facial feature points from different poses, active appearance model (AAM) is employed to find the position of feature point. The improved Lucas-Kanade algorithm is used to solve the image alignment. After obtaining the location of feature points, the eigenspace of texture model is reduced the dimension and sent to the back propagation neural network (BPNN). By using the BPNN, the proposed recognizes that which family-member is the user. The proposed pose-variant face recognition system has been implemented on an embedded image system of a pet robot. In order to test our method, UMIST and self-built database are both used to evaluate the performance of the proposed algorithm. Experimental results show that the average recognition rate of the UMIST database and self-built database in our lab are 91% and 95.56% respectively. The proposed pose-variant face recognition system is suitable for applying to human-robot interaction.
Hsieh, Chao-Kuei, and 謝兆魁. "Research on Robust 2D Face Recognition under Expression and Pose Variations." Thesis, 2009. http://ndltd.ncl.edu.tw/handle/29837810055342825192.
Full text國立清華大學
電機工程學系
97
Face recognition is one of the most intensively studied topics in computer vision and pattern recognition. There are three essential issues to be dealt with in the research of face recognition; namely, pose, illumination, and expression variations. The recognition rate will drop considerably when the head pose or illumination variation is too large, or when there is expression on the face. Although many researches were focused on overcoming these challenges, few were focused on how to robustly recognize expressional faces with one single training sample per class. In this thesis, we modify the regularization-based optical flow algorithm by imposing constraints on some given point correspondences to compute precise pixel displacements and intensity variations. The constrained optical flow computation can be efficiently computed by applying the modified ICPCG algorithm. By using the optical flow computed from the input expression-variant face image with respect to a reference neutral face image, we can remove the expression from the face image by elastic image warping to recognize the subject with facial expression. On the other hand, the optical flow can be computed in the opposite direction, which is from the neutral face image to the input expression-variant face image. By combining information from the computed intra-person optical flow and the synthesized face image in a probabilistic framework, an integrated face recognition system is proposed, which can be robust against facial expressions with a limited size of training database. Experimental validation on the Binghamton University 3D Face Expression (BU-3DFE) Database is given to show that the proposed expression normalization algorithm significantly improves the accuracy of face recognition on expression variant faces. A possible solution for overcoming the pose variation problem in face recognition is also presented in this thesis. The ideal solution is to reconstruct a 3D model from the input images and synthesize the virtual image with the corresponding pose, which might be too complex to be implemented in a real-time application. By formulating this kind of solution as a nonlinear pose normalization problem, we propose an algorithm that integrates the nonlinear kernel function and the linear regression method, which makes the solution resemble to the ideal one. Some discussions and experiments on CMU PIE database are carried out, and the experimental results show that the proposed method is robust against pose variations.
Huang, Jia-ji, and 黃嘉吉. "Automatic Face Recognition based on Head Pose Estimation and SIFT Features." Thesis, 2009. http://ndltd.ncl.edu.tw/handle/50374859509522309664.
Full text國立中正大學
資訊工程所
97
Video based face recognition can be divided three categories: (1)still-to-still, (2)multip stills-to-stills and (3)multiple-still-to-multiple-still. In this thesis, we are interested in video based recognition. Our system is divided into several parts. The first part is face detection procedure which detects face region in a frame from video. The detected face could contain pose variant. Therefore, we empoly head pose estimation method as ltering procedure which select frontal faces from video. In third part, the matching procedure is included. we extract facial feature using Scale-Invariant Feature Transform (SIFT). While features is extracted, each feature of frames in the probe set is matched with gallery set. We also exploit spacial information to eliminate false matching. Our performance will be evaluated in FRGC, MBGC, and IDIAP dataset.
Chen, Hong-Wen, and 陳宏文. "Pose and Expression Invariant Face Recognition with One Training Image P4er Person." Thesis, 2008. http://ndltd.ncl.edu.tw/handle/42556758739964511010.
Full text亞洲大學
資訊工程學系碩士班
96
Face recognition security system has becoming important for many applications such as automatic access control and video surveillance. Most face recognition security systems today require proper frontal view of a person, and these systems will fail if the person to be recognized does not face the camera correctly. In this paper, we present a method for pose and expression invariant face recognition using only a single sample image per person. The method utilizes the similarities of a face image against a set of faces from a prototype set taken at the same pose and expression to establish pose and expression invariant similarity vector which can be used for comparing face images of a person taken in different poses and facial expressions. Experimental results indicate that the proposed method achieves high recognition rate even for large pose and expression variations.
Lazarus, Toby Ellen. "Changes in face processing across ages : the role of experience /." 2002. http://gateway.proquest.com/openurl?url_ver=Z39.88-2004&res_dat=xri:pqdiss&rft_val_fmt=info:ofi/fmt:kev:mtx:dissertation&rft_dat=xri:pqdiss:3048398.
Full textChien, Ming-yen, and 簡名彥. "LBP-Based On-line Multi-Pose Face Model Learning and the Application in Real-time Face Tracking and Recognition." Thesis, 2012. http://ndltd.ncl.edu.tw/handle/u9wznr.
Full text國立臺灣科技大學
自動化及控制研究所
100
Because human face is not a rigid object, the changes of face expression or poses would cause huge variation in the image. Furthermore, there are other disturbances such as the varying of illumination and partial occlusions. Therefore, it is necessary to have a robust multi-pose measurement model to achieve stable tracking. In addition, tracked face could be lost when the occluded region is too large. To recover the tracking, specific features are needed to learn previously. However, how to overcome the disturbance and construct the multi-pose specific feature model is a challenge. Regarding to on-line face recognition, the more personal information we collect, the more accurate result we’ll get. The collection of such information would also be affected by the instability of tracking. How to obtain correct information for on-line face recognition is a problem. In this thesis, we propose an integrated tracking algorithm combining generic face model and specific face model. We use the color kernel histogram of human face to assistant the integration of combining two models, and use generic face model to help construct multi-pose specific face model. Via the specific face model, even if the tracking target is lost in some situations, the model can help to find the losing target. Because the specific face model is constructed by LBP texture feature, it can achieve robust tracking, including partial occlusion. And the learned information of specific face model can be used in face recognition. In our experiment, the purposed method can achieve good tracking result in the condition of complex background, varying of illumination, partial occlusion, changes of poses and so on. The target losing during tracking can be recovered correctly as well. For face recognition, the multi-pose specific face model can provide sufficient information to achieve acceptable accuracy rate. Through the experiment the accuracy rate is above 70%.
Lin, Zong-xian, and 林宗賢. "Fast Semi-3D Vertical Pose Recovery for Face Recognition and Its OMAP Embedded System Implementation." Thesis, 2008. http://ndltd.ncl.edu.tw/handle/87878543624234482851.
Full text雲林科技大學
電機工程系碩士班
96
Face recognition is the key part in biometric field because it can provide noninvasive and convenient features. Most of conventional face recognition systems just focus on the frontal face cases, but the fact is that the face data captured by the camera are often filled up with the pose variation, whether horizontal or vertical pose variations. It decreases the recognition accuracy and reliability. Based on semi-3D face model, this paper proposes a simple but practical preprocessing method to recover the vertical pose variation simply from a single 2-D model view. The proposed method evaluates the angle of the vertical pose variation and thereby recovers the flanked face to the frontal face. Consequently, the recovered face data can be processed by the original face recognition system accurately and efficiently. In the experiment, we adopt the Gabor Wavelet transform for the feature extraction core of the face recognition system. The experimental result shows the proposed Fast Semi-3D Vertical Pose Recovery method can significantly help to raise both similarity and precision of the face recognition system.
Yan, Yi-wei, and 嚴逸緯. "Integration of Human Feature Detection and Geometry Analysis for Real-time Face Pose Estimation and Gesture Recognition." Thesis, 2009. http://ndltd.ncl.edu.tw/handle/63197581381191989404.
Full text國立成功大學
電機工程學系碩博士班
97
In recent years, the digital products have become more accessible to people. The requirement of the different levels of intelligent human-machine interface is increased gradually and conduces to grow more and more related research of human face technique. Face detection and face recognition technology applications such as identifications system, access control monitoring systems. Human-Computer Interactions (HCI) is more and more common in real life. The survey of human face pose estimation, a human face-related research field is a popular topic in HCI. In this paper, we divide the human faces into several viewpoint categories according to their poses in 3D and propose a system to estimate human face pose based on object detection and geometry analysis. The system architecture includes two components: 1) Face detection, 2) Face Pose estimation. It is not only considered about performance, but also the extension of the system by using the modular structure design. We define 9-posed in this system by the human features detection such as eyes, head and shoulders, frontal face and profile face and we defined a detect array for these detectors. Because of the fast object detection algorithm, the features can be detected and get good detect rate in low resolution 320*240. To improve the detect array of this system, we design a cascade detector array which detect only the interested region in image and can detect 9 face poses in real-time. We can speed up the detection system by using the cascade detector array. We have proposed a gesture detection system based on Paul and Viola’s object detection, and combine it with image processing to recognize the defined gesture. We define two gestures in gesture detection to control the appliance. In final chapter of this thesis, we will show the experimental results using the test videos we took. Then we combine the pose estimation system and gesture detection system and apply it to the appliance control in NCKU Aspire Home. The proposed system can not only detect the human pose’s position and pose effectively in image, but also order the appliance. In this research, we proposed a human face pose estimation system. If we add the face model with a pre-training mode, we can increase the system detect rate and it will be a complete detection approach.