Facial image synthesis

With the ability to carry out challenging tasks such as photo generation, Generative Adversarial Networks have attracted increasing attention and achieved impressive progress in recent years [6]. Researchers are also exploring its possible applications for more complex tasks such as facial image syn...

Full description

Saved in:
Bibliographic Details
Main Author: Shen, Guangxu
Other Authors: Lu Shijian
Format: Final Year Project
Language:English
Published: Nanyang Technological University 2022
Subjects:
Online Access:https://hdl.handle.net/10356/156571
Tags: Add Tag
No Tags, Be the first to tag this record!
Institution: Nanyang Technological University
Language: English
id sg-ntu-dr.10356-156571
record_format dspace
spelling sg-ntu-dr.10356-1565712022-04-20T07:31:17Z Facial image synthesis Shen, Guangxu Lu Shijian School of Computer Science and Engineering Shijian.Lu@ntu.edu.sg Engineering::Computer science and engineering::Computing methodologies::Image processing and computer vision With the ability to carry out challenging tasks such as photo generation, Generative Adversarial Networks have attracted increasing attention and achieved impressive progress in recent years [6]. Researchers are also exploring its possible applications for more complex tasks such as facial image synthesis. GAN has proven to have an outstanding performance in carrying out facial expression and attribute editing. A well-trained model could easily transform a facial image with one specific attribute/expression to another while preserving the identity information [7]. In this project, we will first discuss, in a very brief manner, the general problems that are faced by researchers in facial image synthesis. Subsequently, we will evaluate the common practices to solve those problems and their respective limitations. We will carry out an analysis on two advanced approaches, StarGan[2] and STGan[10], and discuss their respective ways to carry out facial image generation. . We will also explore the possibility of combining the best parts of these two models so that our designed facial expression GAN, CombineGAN, will be able to address both image feature transfer and quality issues. One possible way is to utilize STGan’s generator, built from a selective transfer perspective where Selective Transfer Units (STU) are built in the encoder-decoder generator architecture for it to adaptively choose and modify the encoder feature for an improved facial image synthesis. We will adopt evaluation metrics such as Inception Score (IS) and Frechet Inception Distance (FID) [18] to quantitively evaluate the model’s performance. We will also use qualitative method such as Amazon Mechanical Turk (AMT) [14] to evaluate the model performance from a human’s perspective. Lastly, our model will be applied to translate real life images for us to better understand its performance in a different context. Bachelor of Engineering (Computer Science) 2022-04-20T07:31:17Z 2022-04-20T07:31:17Z 2022 Final Year Project (FYP) Shen, G. (2022). Facial image synthesis. Final Year Project (FYP), Nanyang Technological University, Singapore. https://hdl.handle.net/10356/156571 https://hdl.handle.net/10356/156571 en SCSE21-0070 application/pdf Nanyang Technological University
institution Nanyang Technological University
building NTU Library
continent Asia
country Singapore
Singapore
content_provider NTU Library
collection DR-NTU
language English
topic Engineering::Computer science and engineering::Computing methodologies::Image processing and computer vision
spellingShingle Engineering::Computer science and engineering::Computing methodologies::Image processing and computer vision
Shen, Guangxu
Facial image synthesis
description With the ability to carry out challenging tasks such as photo generation, Generative Adversarial Networks have attracted increasing attention and achieved impressive progress in recent years [6]. Researchers are also exploring its possible applications for more complex tasks such as facial image synthesis. GAN has proven to have an outstanding performance in carrying out facial expression and attribute editing. A well-trained model could easily transform a facial image with one specific attribute/expression to another while preserving the identity information [7]. In this project, we will first discuss, in a very brief manner, the general problems that are faced by researchers in facial image synthesis. Subsequently, we will evaluate the common practices to solve those problems and their respective limitations. We will carry out an analysis on two advanced approaches, StarGan[2] and STGan[10], and discuss their respective ways to carry out facial image generation. . We will also explore the possibility of combining the best parts of these two models so that our designed facial expression GAN, CombineGAN, will be able to address both image feature transfer and quality issues. One possible way is to utilize STGan’s generator, built from a selective transfer perspective where Selective Transfer Units (STU) are built in the encoder-decoder generator architecture for it to adaptively choose and modify the encoder feature for an improved facial image synthesis. We will adopt evaluation metrics such as Inception Score (IS) and Frechet Inception Distance (FID) [18] to quantitively evaluate the model’s performance. We will also use qualitative method such as Amazon Mechanical Turk (AMT) [14] to evaluate the model performance from a human’s perspective. Lastly, our model will be applied to translate real life images for us to better understand its performance in a different context.
author2 Lu Shijian
author_facet Lu Shijian
Shen, Guangxu
format Final Year Project
author Shen, Guangxu
author_sort Shen, Guangxu
title Facial image synthesis
title_short Facial image synthesis
title_full Facial image synthesis
title_fullStr Facial image synthesis
title_full_unstemmed Facial image synthesis
title_sort facial image synthesis
publisher Nanyang Technological University
publishDate 2022
url https://hdl.handle.net/10356/156571
_version_ 1731235724692291584