Stuff segmentation and non-photorealistic rendering with generative adversarial networks

Generative Adversarial Networks (GANs) have shown impressive results in a variety of image generation tasks in recent years, including rendering photorealistic images with artistic styles. However, current work in transforming images have mostly focused on either transforming the whole image, or on...

Full description

Saved in:
Bibliographic Details
Main Author: Choy, Jin Xiang
Other Authors: Ong Yew Soon
Format: Thesis-Master by Research
Language:English
Published: Nanyang Technological University 2020
Subjects:
Online Access:https://hdl.handle.net/10356/137213
Tags: Add Tag
No Tags, Be the first to tag this record!
Institution: Nanyang Technological University
Language: English
id sg-ntu-dr.10356-137213
record_format dspace
spelling sg-ntu-dr.10356-1372132020-10-28T08:29:22Z Stuff segmentation and non-photorealistic rendering with generative adversarial networks Choy, Jin Xiang Ong Yew Soon School of Computer Science and Engineering ASYSOng@ntu.edu.sg Engineering::Computer science and engineering Generative Adversarial Networks (GANs) have shown impressive results in a variety of image generation tasks in recent years, including rendering photorealistic images with artistic styles. However, current work in transforming images have mostly focused on either transforming the whole image, or on the thing classes. There have been little attention on the artistic rendering of only the stuff classes of images. Current possible methods of performing painting of specific image regions also result in unnatural boundaries between painted and non-painted regions. Therefore, we aim to develop an end-to-end model for the novel task of Non-Photorealistically Rendering the stuff of images. In order to train a model capable of doing so, we first require images with partially painted stuff classes for training. However, due to a lack of such images, we propose a flexible and extensible partially painted image generation pipeline that uses an image segmentation dataset to generate partially painted image datasets. We use these datasets with a GAN framework based on the Pix2Pix model for training. We find that the trained model can perform image stuff painting acceptably well, and it generated results that have more natural boundaries between painted and non-painted image regions. We then provide an analysis of the effects of the Pix2Pix architecture on the training task. We find that satisfactory results can also be obtained. We then discuss the limitations for this context and compare the performance to that of image stuff painting. We then introduce the Stuff-Painting GAN(SPGAN) for reducing errors in identifying image regions for painting by introducing segmentation masks into the training process. An additional discriminator that takes in segmentation masks as input is introduced to the architecture. We then show that it performs on par or better than the baseline GAN framework in identifying image regions for painting. To improve the segmentation learning we also introduce a Gaussian error correction in each training iteration. Master of Engineering 2020-03-06T06:09:43Z 2020-03-06T06:09:43Z 2019 Thesis-Master by Research Choy, J. X. (2019). Stuff segmentation and non-photorealistic rendering with generative adversarial networks. Master's thesis, Nanyang Technological University, Singapore. https://hdl.handle.net/10356/137213 10.32657/10356/137213 en This work is licensed under a Creative Commons Attribution-NonCommercial 4.0 International License (CC BY-NC 4.0). application/pdf Nanyang Technological University
institution Nanyang Technological University
building NTU Library
continent Asia
country Singapore
Singapore
content_provider NTU Library
collection DR-NTU
language English
topic Engineering::Computer science and engineering
spellingShingle Engineering::Computer science and engineering
Choy, Jin Xiang
Stuff segmentation and non-photorealistic rendering with generative adversarial networks
description Generative Adversarial Networks (GANs) have shown impressive results in a variety of image generation tasks in recent years, including rendering photorealistic images with artistic styles. However, current work in transforming images have mostly focused on either transforming the whole image, or on the thing classes. There have been little attention on the artistic rendering of only the stuff classes of images. Current possible methods of performing painting of specific image regions also result in unnatural boundaries between painted and non-painted regions. Therefore, we aim to develop an end-to-end model for the novel task of Non-Photorealistically Rendering the stuff of images. In order to train a model capable of doing so, we first require images with partially painted stuff classes for training. However, due to a lack of such images, we propose a flexible and extensible partially painted image generation pipeline that uses an image segmentation dataset to generate partially painted image datasets. We use these datasets with a GAN framework based on the Pix2Pix model for training. We find that the trained model can perform image stuff painting acceptably well, and it generated results that have more natural boundaries between painted and non-painted image regions. We then provide an analysis of the effects of the Pix2Pix architecture on the training task. We find that satisfactory results can also be obtained. We then discuss the limitations for this context and compare the performance to that of image stuff painting. We then introduce the Stuff-Painting GAN(SPGAN) for reducing errors in identifying image regions for painting by introducing segmentation masks into the training process. An additional discriminator that takes in segmentation masks as input is introduced to the architecture. We then show that it performs on par or better than the baseline GAN framework in identifying image regions for painting. To improve the segmentation learning we also introduce a Gaussian error correction in each training iteration.
author2 Ong Yew Soon
author_facet Ong Yew Soon
Choy, Jin Xiang
format Thesis-Master by Research
author Choy, Jin Xiang
author_sort Choy, Jin Xiang
title Stuff segmentation and non-photorealistic rendering with generative adversarial networks
title_short Stuff segmentation and non-photorealistic rendering with generative adversarial networks
title_full Stuff segmentation and non-photorealistic rendering with generative adversarial networks
title_fullStr Stuff segmentation and non-photorealistic rendering with generative adversarial networks
title_full_unstemmed Stuff segmentation and non-photorealistic rendering with generative adversarial networks
title_sort stuff segmentation and non-photorealistic rendering with generative adversarial networks
publisher Nanyang Technological University
publishDate 2020
url https://hdl.handle.net/10356/137213
_version_ 1683493126894780416