Stuff segmentation and non-photorealistic rendering with generative adversarial networks
Generative Adversarial Networks (GANs) have shown impressive results in a variety of image generation tasks in recent years, including rendering photorealistic images with artistic styles. However, current work in transforming images have mostly focused on either transforming the whole image, or on...
Saved in:
Main Author: | |
---|---|
Other Authors: | |
Format: | Thesis-Master by Research |
Language: | English |
Published: |
Nanyang Technological University
2020
|
Subjects: | |
Online Access: | https://hdl.handle.net/10356/137213 |
Tags: |
Add Tag
No Tags, Be the first to tag this record!
|
Institution: | Nanyang Technological University |
Language: | English |
id |
sg-ntu-dr.10356-137213 |
---|---|
record_format |
dspace |
spelling |
sg-ntu-dr.10356-1372132020-10-28T08:29:22Z Stuff segmentation and non-photorealistic rendering with generative adversarial networks Choy, Jin Xiang Ong Yew Soon School of Computer Science and Engineering ASYSOng@ntu.edu.sg Engineering::Computer science and engineering Generative Adversarial Networks (GANs) have shown impressive results in a variety of image generation tasks in recent years, including rendering photorealistic images with artistic styles. However, current work in transforming images have mostly focused on either transforming the whole image, or on the thing classes. There have been little attention on the artistic rendering of only the stuff classes of images. Current possible methods of performing painting of specific image regions also result in unnatural boundaries between painted and non-painted regions. Therefore, we aim to develop an end-to-end model for the novel task of Non-Photorealistically Rendering the stuff of images. In order to train a model capable of doing so, we first require images with partially painted stuff classes for training. However, due to a lack of such images, we propose a flexible and extensible partially painted image generation pipeline that uses an image segmentation dataset to generate partially painted image datasets. We use these datasets with a GAN framework based on the Pix2Pix model for training. We find that the trained model can perform image stuff painting acceptably well, and it generated results that have more natural boundaries between painted and non-painted image regions. We then provide an analysis of the effects of the Pix2Pix architecture on the training task. We find that satisfactory results can also be obtained. We then discuss the limitations for this context and compare the performance to that of image stuff painting. We then introduce the Stuff-Painting GAN(SPGAN) for reducing errors in identifying image regions for painting by introducing segmentation masks into the training process. An additional discriminator that takes in segmentation masks as input is introduced to the architecture. We then show that it performs on par or better than the baseline GAN framework in identifying image regions for painting. To improve the segmentation learning we also introduce a Gaussian error correction in each training iteration. Master of Engineering 2020-03-06T06:09:43Z 2020-03-06T06:09:43Z 2019 Thesis-Master by Research Choy, J. X. (2019). Stuff segmentation and non-photorealistic rendering with generative adversarial networks. Master's thesis, Nanyang Technological University, Singapore. https://hdl.handle.net/10356/137213 10.32657/10356/137213 en This work is licensed under a Creative Commons Attribution-NonCommercial 4.0 International License (CC BY-NC 4.0). application/pdf Nanyang Technological University |
institution |
Nanyang Technological University |
building |
NTU Library |
continent |
Asia |
country |
Singapore Singapore |
content_provider |
NTU Library |
collection |
DR-NTU |
language |
English |
topic |
Engineering::Computer science and engineering |
spellingShingle |
Engineering::Computer science and engineering Choy, Jin Xiang Stuff segmentation and non-photorealistic rendering with generative adversarial networks |
description |
Generative Adversarial Networks (GANs) have shown impressive results in a variety of image generation tasks in recent years, including rendering photorealistic images with artistic styles. However, current work in transforming images have mostly focused on either transforming the whole image, or on the thing classes. There have been little attention on the artistic rendering of only the stuff classes of images. Current possible methods of performing painting of specific image regions also result in unnatural boundaries between painted and non-painted regions. Therefore, we aim to develop an end-to-end model for the novel task of Non-Photorealistically Rendering the stuff of images.
In order to train a model capable of doing so, we first require images with partially painted stuff classes for training. However, due to a lack of such images, we propose a flexible and extensible partially painted image generation pipeline that uses an image segmentation dataset to generate partially painted image datasets.
We use these datasets with a GAN framework based on the Pix2Pix model for training. We find that the trained model can perform image stuff painting acceptably well, and it generated results that have more natural boundaries between painted and non-painted image regions. We then provide an analysis of the effects of the Pix2Pix architecture on the training task. We find that satisfactory results can also be obtained. We then discuss the limitations for this context and compare the performance to that of image stuff painting.
We then introduce the Stuff-Painting GAN(SPGAN) for reducing errors in identifying image regions for painting by introducing segmentation masks into the training process. An additional discriminator that takes in segmentation masks as input is introduced to the architecture. We then show that it performs on par or better than the baseline GAN framework in identifying image regions for painting. To improve the segmentation learning we also introduce a Gaussian error correction in each training iteration. |
author2 |
Ong Yew Soon |
author_facet |
Ong Yew Soon Choy, Jin Xiang |
format |
Thesis-Master by Research |
author |
Choy, Jin Xiang |
author_sort |
Choy, Jin Xiang |
title |
Stuff segmentation and non-photorealistic rendering with generative adversarial networks |
title_short |
Stuff segmentation and non-photorealistic rendering with generative adversarial networks |
title_full |
Stuff segmentation and non-photorealistic rendering with generative adversarial networks |
title_fullStr |
Stuff segmentation and non-photorealistic rendering with generative adversarial networks |
title_full_unstemmed |
Stuff segmentation and non-photorealistic rendering with generative adversarial networks |
title_sort |
stuff segmentation and non-photorealistic rendering with generative adversarial networks |
publisher |
Nanyang Technological University |
publishDate |
2020 |
url |
https://hdl.handle.net/10356/137213 |
_version_ |
1683493126894780416 |