Generating human faces by generative adversarial networks

Over the years, computer vision improves significantly. From recognising and understanding what lies underneath an image, we can now generate images by modelling training distribution using generative adversarial network(GAN). Since then, researchers come out with various variants of GAN and ways to...

Full description

Saved in:
Bibliographic Details
Main Author: Quek, Chin Wei
Other Authors: Chen Change Loy
Format: Final Year Project
Language:English
Published: Nanyang Technological University 2020
Subjects:
Online Access:https://hdl.handle.net/10356/139259
Tags: Add Tag
No Tags, Be the first to tag this record!
Institution: Nanyang Technological University
Language: English
Description
Summary:Over the years, computer vision improves significantly. From recognising and understanding what lies underneath an image, we can now generate images by modelling training distribution using generative adversarial network(GAN). Since then, researchers come out with various variants of GAN and ways to stabalize GAN training. This results in improved quality of generated image. The application of GAN has sparked the interest of many people. In this project, we first analyse the use of StarGAN, a unified generative adversarial network for multi-domain image-to-image translation task to generate human facial expressions. We also explore the possible use of StarGAN in cartoon character facial expression generation and video generation.