Real-time arbitrary style transfer via deep learning

Neural style transfer is the process of merging the content of one image with the style of another to create a new image. Many applications have recently exploited style transfer to create highly popular content on social media. Existing methods typically face limitations such as a small number of t...

Full description

Saved in:
Bibliographic Details
Main Author: Wang, Zijian
Other Authors: Chen Change Loy
Format: Final Year Project
Language:English
Published: Nanyang Technological University 2021
Subjects:
Online Access:https://hdl.handle.net/10356/147930
Tags: Add Tag
No Tags, Be the first to tag this record!
Institution: Nanyang Technological University
Language: English
id sg-ntu-dr.10356-147930
record_format dspace
spelling sg-ntu-dr.10356-1479302021-04-16T06:49:47Z Real-time arbitrary style transfer via deep learning Wang, Zijian Chen Change Loy School of Computer Science and Engineering ccloy@ntu.edu.sg Engineering::Computer science and engineering::Computing methodologies::Image processing and computer vision Neural style transfer is the process of merging the content of one image with the style of another to create a new image. Many applications have recently exploited style transfer to create highly popular content on social media. Existing methods typically face limitations such as a small number of transferable styles and a sluggish image generation speed. In this work, we discuss two approaches, AdaIN and MUNIT, to achieve real-time arbitrary style transfer and apply it to videos. The AdaIN method can produce aesthetically pleasing stylized images by changing the content-style weight ratio. It is found that the AdaIN method can be sped up by eliminating convolutional layers from the decoder. The refined decoder of AdaIN achieves a large speed boost without compromising image quality of style transfer. The MUNIT method has advantages when training on a small dataset that style and content samples are from two specific domains. We analyze these two methods and derive possible theoretical reasons behind them. Since the refined AdaIN method only needs to be trained once and can produce stylized images at real-time speed, its application can be extended to perform real-time arbitrary video style transfer. Finally, we conclude with more discussions about several future improvement directions. Bachelor of Engineering (Computer Science) 2021-04-16T06:49:47Z 2021-04-16T06:49:47Z 2021 Final Year Project (FYP) Wang, Z. (2021). Real-time arbitrary style transfer via deep learning. Final Year Project (FYP), Nanyang Technological University, Singapore. https://hdl.handle.net/10356/147930 https://hdl.handle.net/10356/147930 en SCSE20-0408 application/pdf Nanyang Technological University
institution Nanyang Technological University
building NTU Library
continent Asia
country Singapore
Singapore
content_provider NTU Library
collection DR-NTU
language English
topic Engineering::Computer science and engineering::Computing methodologies::Image processing and computer vision
spellingShingle Engineering::Computer science and engineering::Computing methodologies::Image processing and computer vision
Wang, Zijian
Real-time arbitrary style transfer via deep learning
description Neural style transfer is the process of merging the content of one image with the style of another to create a new image. Many applications have recently exploited style transfer to create highly popular content on social media. Existing methods typically face limitations such as a small number of transferable styles and a sluggish image generation speed. In this work, we discuss two approaches, AdaIN and MUNIT, to achieve real-time arbitrary style transfer and apply it to videos. The AdaIN method can produce aesthetically pleasing stylized images by changing the content-style weight ratio. It is found that the AdaIN method can be sped up by eliminating convolutional layers from the decoder. The refined decoder of AdaIN achieves a large speed boost without compromising image quality of style transfer. The MUNIT method has advantages when training on a small dataset that style and content samples are from two specific domains. We analyze these two methods and derive possible theoretical reasons behind them. Since the refined AdaIN method only needs to be trained once and can produce stylized images at real-time speed, its application can be extended to perform real-time arbitrary video style transfer. Finally, we conclude with more discussions about several future improvement directions.
author2 Chen Change Loy
author_facet Chen Change Loy
Wang, Zijian
format Final Year Project
author Wang, Zijian
author_sort Wang, Zijian
title Real-time arbitrary style transfer via deep learning
title_short Real-time arbitrary style transfer via deep learning
title_full Real-time arbitrary style transfer via deep learning
title_fullStr Real-time arbitrary style transfer via deep learning
title_full_unstemmed Real-time arbitrary style transfer via deep learning
title_sort real-time arbitrary style transfer via deep learning
publisher Nanyang Technological University
publishDate 2021
url https://hdl.handle.net/10356/147930
_version_ 1698713636391878656