Pose-robust face recognition
With the development of deep learning technology, smart photo gallery system which supports text-based searching becomes deployable on personal smart phones. In order to retrieve images containing specific person, it is necessary to find out who is contained inside each image. Above feature is a typ...
Saved in:
Main Author: | |
---|---|
Other Authors: | |
Format: | Final Year Project |
Language: | English |
Published: |
Nanyang Technological University
2020
|
Subjects: | |
Online Access: | https://hdl.handle.net/10356/139988 |
Tags: |
Add Tag
No Tags, Be the first to tag this record!
|
Institution: | Nanyang Technological University |
Language: | English |
id |
sg-ntu-dr.10356-139988 |
---|---|
record_format |
dspace |
spelling |
sg-ntu-dr.10356-1399882023-07-07T18:36:36Z Pose-robust face recognition Xie, Zhuofan Yap Kim Hui School of Electrical and Electronic Engineering EKHYap@ntu.edu.sg Engineering::Electrical and electronic engineering With the development of deep learning technology, smart photo gallery system which supports text-based searching becomes deployable on personal smart phones. In order to retrieve images containing specific person, it is necessary to find out who is contained inside each image. Above feature is a typical face recognition task which usually contains detection stage and identification stage. Our work focus on the identification stage. At this stage, face detected is encoded into a vector for componence with those in the database to pick out the label of closest one as the identity. The problem is, encoded vectors of one’s frontal face image may differ a lot from those generated from profile face image although they are from the same person. This problem occurs mainly because it often contains much more front face images than profile face images in the training data. To solve this problem, we need to find a way to map profile vector into front vector of the same person . Specifically, it consists of two parts. First, a head rotation estimator is developed to get the yaw angle as the weight parameter, representing how much modification needed. Second, a light-weight CNN network is trained to learn the profile-front mapping. Images taken from different angles for groups of people are used to train the network. The goal is to minimize the Euclid distance between encoded profile face vectors and encoded front face vectors for each person. Our model achieves an accuracy of 97.4% tested on LFW dataset compared with 95.6% achieved by original FaceNet. Bachelor of Engineering (Electrical and Electronic Engineering) 2020-05-26T02:35:58Z 2020-05-26T02:35:58Z 2020 Final Year Project (FYP) https://hdl.handle.net/10356/139988 en A3286-191 application/pdf Nanyang Technological University |
institution |
Nanyang Technological University |
building |
NTU Library |
continent |
Asia |
country |
Singapore Singapore |
content_provider |
NTU Library |
collection |
DR-NTU |
language |
English |
topic |
Engineering::Electrical and electronic engineering |
spellingShingle |
Engineering::Electrical and electronic engineering Xie, Zhuofan Pose-robust face recognition |
description |
With the development of deep learning technology, smart photo gallery system which supports text-based searching becomes deployable on personal smart phones. In order to retrieve images containing specific person, it is necessary to find out who is contained inside each image. Above feature is a typical face recognition task which usually contains detection stage and identification stage. Our work focus on the identification stage. At this stage, face detected is encoded into a vector for componence with those in the database to pick out the label of closest one as the identity. The problem is, encoded vectors of one’s frontal face image may differ a lot from those generated from profile face image although they are from the same person. This problem occurs mainly because it often contains much more front face images than profile face images in the training data. To solve this problem, we need to find a way to map profile vector into front vector of the same person . Specifically, it consists of two parts. First, a head rotation estimator is developed to get the yaw angle as the weight parameter, representing how much modification needed. Second, a light-weight CNN network is trained to learn the profile-front mapping. Images taken from different angles for groups of people are used to train the network. The goal is to minimize the Euclid distance between encoded profile face vectors and encoded front face vectors for each person. Our model achieves an accuracy of 97.4% tested on LFW dataset compared with 95.6% achieved by original FaceNet. |
author2 |
Yap Kim Hui |
author_facet |
Yap Kim Hui Xie, Zhuofan |
format |
Final Year Project |
author |
Xie, Zhuofan |
author_sort |
Xie, Zhuofan |
title |
Pose-robust face recognition |
title_short |
Pose-robust face recognition |
title_full |
Pose-robust face recognition |
title_fullStr |
Pose-robust face recognition |
title_full_unstemmed |
Pose-robust face recognition |
title_sort |
pose-robust face recognition |
publisher |
Nanyang Technological University |
publishDate |
2020 |
url |
https://hdl.handle.net/10356/139988 |
_version_ |
1772825162443390976 |