Anime characters creation using generative adversarial networks with user inputs
Generative Adversarial Network (GAN) is a framework that has been used to generate realistic images of faces, objects, and even landscapes. With its increasing popularity, it can be used to generate anime facial images. Diffusion models have also recently been on the rise with models like Stable Dif...
Saved in:
Main Author: | |
---|---|
Other Authors: | |
Format: | Final Year Project |
Language: | English |
Published: |
Nanyang Technological University
2024
|
Subjects: | |
Online Access: | https://hdl.handle.net/10356/175300 |
Tags: |
Add Tag
No Tags, Be the first to tag this record!
|
Institution: | Nanyang Technological University |
Language: | English |
id |
sg-ntu-dr.10356-175300 |
---|---|
record_format |
dspace |
spelling |
sg-ntu-dr.10356-1753002024-04-26T15:44:16Z Anime characters creation using generative adversarial networks with user inputs Ang, Himari Lixin Seah Hock Soon School of Computer Science and Engineering ASHSSEAH@ntu.edu.sg Computer and Information Science Generative adversarial networks Anime Diffusion Generative Adversarial Network (GAN) is a framework that has been used to generate realistic images of faces, objects, and even landscapes. With its increasing popularity, it can be used to generate anime facial images. Diffusion models have also recently been on the rise with models like Stable Diffusion taking the center stage due to how realistic the images generated are. These models has potential applications in the entertainment industry for creating virtual worlds. The final year project aims to develop a GAN-based system for generating anime char- acters with user-defined attributes, allowing users to input desired characteristics such as hair color, eye shape, clothing style, and more to create customizable and unique anime character designs. In this project, GAN and diffusion models are proposed. For the GAN model, we explored an Auxiliary Classifier GAN (ACGAN) architecture to constraint the model such that users will be able to request specific attributes to be generated with their image. Similarly, for the diffusion model, we adopted the Denoising Diffusion Probabilistic Models (DDPM) framework with a UNet base model. To allow for attributes, an attribute mapper was designed to learn to map user-input attributes to the random noise that diffu- sion models use for generation. To tie the models together for the ease of use, a backend was created that allows the model to be deployed as a Model-as-a-Service (MaaS) with a NextJS frontend that interacts with it. Instead of having to work with the model directly, users will only have to interact with the Web Application. In the report, we show how we are able to generate anime character faces using both the GAN and diffusion model. At the same time, we present potential future works that can help further improve the model such as improving on the dataset for both the image and the attribute tags. Finally, we touch on the lifecycle of the project: architecture of our models, system architecture, implementation and deployment of the project. Bachelor's degree 2024-04-23T10:52:42Z 2024-04-23T10:52:42Z 2024 Final Year Project (FYP) Ang, H. L. (2024). Anime characters creation using generative adversarial networks with user inputs. Final Year Project (FYP), Nanyang Technological University, Singapore. https://hdl.handle.net/10356/175300 https://hdl.handle.net/10356/175300 en SCSE23-0318 application/pdf Nanyang Technological University |
institution |
Nanyang Technological University |
building |
NTU Library |
continent |
Asia |
country |
Singapore Singapore |
content_provider |
NTU Library |
collection |
DR-NTU |
language |
English |
topic |
Computer and Information Science Generative adversarial networks Anime Diffusion |
spellingShingle |
Computer and Information Science Generative adversarial networks Anime Diffusion Ang, Himari Lixin Anime characters creation using generative adversarial networks with user inputs |
description |
Generative Adversarial Network (GAN) is a framework that has been used to generate realistic images of faces, objects, and even landscapes. With its increasing popularity, it can be used to generate anime facial images. Diffusion models have also recently been on the rise with models like Stable Diffusion taking the center stage due to how realistic the images generated are. These models has potential applications in the entertainment industry for creating virtual worlds.
The final year project aims to develop a GAN-based system for generating anime char- acters with user-defined attributes, allowing users to input desired characteristics such as hair color, eye shape, clothing style, and more to create customizable and unique anime character designs. In this project, GAN and diffusion models are proposed. For the GAN model, we explored an Auxiliary Classifier GAN (ACGAN) architecture to constraint the model such that users will be able to request specific attributes to be generated with their image. Similarly, for the diffusion model, we adopted the Denoising Diffusion Probabilistic Models (DDPM) framework with a UNet base model. To allow for attributes, an attribute mapper was designed to learn to map user-input attributes to the random noise that diffu- sion models use for generation. To tie the models together for the ease of use, a backend was created that allows the model to be deployed as a Model-as-a-Service (MaaS) with a NextJS frontend that interacts with it. Instead of having to work with the model directly, users will only have to interact with the Web Application.
In the report, we show how we are able to generate anime character faces using both the GAN and diffusion model. At the same time, we present potential future works that can help further improve the model such as improving on the dataset for both the image and the attribute tags. Finally, we touch on the lifecycle of the project: architecture of our models, system architecture, implementation and deployment of the project. |
author2 |
Seah Hock Soon |
author_facet |
Seah Hock Soon Ang, Himari Lixin |
format |
Final Year Project |
author |
Ang, Himari Lixin |
author_sort |
Ang, Himari Lixin |
title |
Anime characters creation using generative adversarial networks with user inputs |
title_short |
Anime characters creation using generative adversarial networks with user inputs |
title_full |
Anime characters creation using generative adversarial networks with user inputs |
title_fullStr |
Anime characters creation using generative adversarial networks with user inputs |
title_full_unstemmed |
Anime characters creation using generative adversarial networks with user inputs |
title_sort |
anime characters creation using generative adversarial networks with user inputs |
publisher |
Nanyang Technological University |
publishDate |
2024 |
url |
https://hdl.handle.net/10356/175300 |
_version_ |
1800916243973144576 |