Pose-robust face recognition

With the development of deep learning technology, smart photo gallery system which supports text-based searching becomes deployable on personal smart phones. In order to retrieve images containing specific person, it is necessary to find out who is contained inside each image. Above feature is a typ...

Full description

Saved in:
Bibliographic Details
Main Author: Xie, Zhuofan
Other Authors: Yap Kim Hui
Format: Final Year Project
Language:English
Published: Nanyang Technological University 2020
Subjects:
Online Access:https://hdl.handle.net/10356/139988
Tags: Add Tag
No Tags, Be the first to tag this record!
Institution: Nanyang Technological University
Language: English
id sg-ntu-dr.10356-139988
record_format dspace
spelling sg-ntu-dr.10356-1399882023-07-07T18:36:36Z Pose-robust face recognition Xie, Zhuofan Yap Kim Hui School of Electrical and Electronic Engineering EKHYap@ntu.edu.sg Engineering::Electrical and electronic engineering With the development of deep learning technology, smart photo gallery system which supports text-based searching becomes deployable on personal smart phones. In order to retrieve images containing specific person, it is necessary to find out who is contained inside each image. Above feature is a typical face recognition task which usually contains detection stage and identification stage. Our work focus on the identification stage. At this stage, face detected is encoded into a vector for componence with those in the database to pick out the label of closest one as the identity. The problem is, encoded vectors of one’s frontal face image may differ a lot from those generated from profile face image although they are from the same person. This problem occurs mainly because it often contains much more front face images than profile face images in the training data. To solve this problem, we need to find a way to map profile vector into front vector of the same person . Specifically, it consists of two parts. First, a head rotation estimator is developed to get the yaw angle as the weight parameter, representing how much modification needed. Second, a light-weight CNN network is trained to learn the profile-front mapping. Images taken from different angles for groups of people are used to train the network. The goal is to minimize the Euclid distance between encoded profile face vectors and encoded front face vectors for each person. Our model achieves an accuracy of 97.4% tested on LFW dataset compared with 95.6% achieved by original FaceNet. Bachelor of Engineering (Electrical and Electronic Engineering) 2020-05-26T02:35:58Z 2020-05-26T02:35:58Z 2020 Final Year Project (FYP) https://hdl.handle.net/10356/139988 en A3286-191 application/pdf Nanyang Technological University
institution Nanyang Technological University
building NTU Library
continent Asia
country Singapore
Singapore
content_provider NTU Library
collection DR-NTU
language English
topic Engineering::Electrical and electronic engineering
spellingShingle Engineering::Electrical and electronic engineering
Xie, Zhuofan
Pose-robust face recognition
description With the development of deep learning technology, smart photo gallery system which supports text-based searching becomes deployable on personal smart phones. In order to retrieve images containing specific person, it is necessary to find out who is contained inside each image. Above feature is a typical face recognition task which usually contains detection stage and identification stage. Our work focus on the identification stage. At this stage, face detected is encoded into a vector for componence with those in the database to pick out the label of closest one as the identity. The problem is, encoded vectors of one’s frontal face image may differ a lot from those generated from profile face image although they are from the same person. This problem occurs mainly because it often contains much more front face images than profile face images in the training data. To solve this problem, we need to find a way to map profile vector into front vector of the same person . Specifically, it consists of two parts. First, a head rotation estimator is developed to get the yaw angle as the weight parameter, representing how much modification needed. Second, a light-weight CNN network is trained to learn the profile-front mapping. Images taken from different angles for groups of people are used to train the network. The goal is to minimize the Euclid distance between encoded profile face vectors and encoded front face vectors for each person. Our model achieves an accuracy of 97.4% tested on LFW dataset compared with 95.6% achieved by original FaceNet.
author2 Yap Kim Hui
author_facet Yap Kim Hui
Xie, Zhuofan
format Final Year Project
author Xie, Zhuofan
author_sort Xie, Zhuofan
title Pose-robust face recognition
title_short Pose-robust face recognition
title_full Pose-robust face recognition
title_fullStr Pose-robust face recognition
title_full_unstemmed Pose-robust face recognition
title_sort pose-robust face recognition
publisher Nanyang Technological University
publishDate 2020
url https://hdl.handle.net/10356/139988
_version_ 1772825162443390976