Parsing-conditioned anime translation: A new dataset and method

Anime is an abstract art form that is substantially different from the human portrait, leading to a challenging misaligned image translation problem that is beyond the capability of existing methods. This can be boiled down to a highly ambiguous unconstrained translation between two domains. To this...

Full description

Saved in:
Bibliographic Details
Main Authors: LI, Zhansheng, XU, Yangyang, ZHAO, Nanxuan, ZHOU, Yang, LIU, Yongtuo, LIN, Dahua, HE, Shengfeng
Format: text
Language:English
Published: Institutional Knowledge at Singapore Management University 2023
Subjects:
Online Access:https://ink.library.smu.edu.sg/sis_research/8358
https://ink.library.smu.edu.sg/context/sis_research/article/9361/viewcontent/Parsing.pdf
Tags: Add Tag
No Tags, Be the first to tag this record!
Institution: Singapore Management University
Language: English
id sg-smu-ink.sis_research-9361
record_format dspace
spelling sg-smu-ink.sis_research-93612023-12-13T03:15:44Z Parsing-conditioned anime translation: A new dataset and method LI, Zhansheng XU, Yangyang ZHAO, Nanxuan ZHOU, Yang LIU, Yongtuo LIN, Dahua HE, Shengfeng Anime is an abstract art form that is substantially different from the human portrait, leading to a challenging misaligned image translation problem that is beyond the capability of existing methods. This can be boiled down to a highly ambiguous unconstrained translation between two domains. To this end, we design a new anime translation framework by deriving the prior knowledge of a pre-Trained StyleGAN model. We introduce disentangled encoders to separately embed structure and appearance information into the same latent code, governed by four tailored losses. Moreover, we develop a FaceBank aggregation method that leverages the generated data of the StyleGAN, anchoring the prediction to produce in-domain animes. To empower our model and promote the research of anime translation, we propose the first anime portrait parsing dataset, Danbooru-Parsing, containing 4,921 densely labeled images across 17 classes. This dataset connects the face semantics with appearances, enabling our new constrained translation setting. We further show the editability of our results, and extend our method to manga images, by generating the first manga parsing pseudo data. Extensive experiments demonstrate the values of our new dataset and method, resulting in the first feasible solution on anime translation. 2023-04-01T07:00:00Z text application/pdf https://ink.library.smu.edu.sg/sis_research/8358 info:doi/10.1145/3585002 https://ink.library.smu.edu.sg/context/sis_research/article/9361/viewcontent/Parsing.pdf http://creativecommons.org/licenses/by-nc-nd/4.0/ Research Collection School Of Computing and Information Systems eng Institutional Knowledge at Singapore Management University Abstract arts Aggregation methods Anchorings Feasible solution Image editing Image translation Image-to-image translation Labeled images Prior-knowledge Two domains Databases and Information Systems
institution Singapore Management University
building SMU Libraries
continent Asia
country Singapore
Singapore
content_provider SMU Libraries
collection InK@SMU
language English
topic Abstract arts
Aggregation methods
Anchorings
Feasible solution
Image editing
Image translation
Image-to-image translation
Labeled images
Prior-knowledge
Two domains
Databases and Information Systems
spellingShingle Abstract arts
Aggregation methods
Anchorings
Feasible solution
Image editing
Image translation
Image-to-image translation
Labeled images
Prior-knowledge
Two domains
Databases and Information Systems
LI, Zhansheng
XU, Yangyang
ZHAO, Nanxuan
ZHOU, Yang
LIU, Yongtuo
LIN, Dahua
HE, Shengfeng
Parsing-conditioned anime translation: A new dataset and method
description Anime is an abstract art form that is substantially different from the human portrait, leading to a challenging misaligned image translation problem that is beyond the capability of existing methods. This can be boiled down to a highly ambiguous unconstrained translation between two domains. To this end, we design a new anime translation framework by deriving the prior knowledge of a pre-Trained StyleGAN model. We introduce disentangled encoders to separately embed structure and appearance information into the same latent code, governed by four tailored losses. Moreover, we develop a FaceBank aggregation method that leverages the generated data of the StyleGAN, anchoring the prediction to produce in-domain animes. To empower our model and promote the research of anime translation, we propose the first anime portrait parsing dataset, Danbooru-Parsing, containing 4,921 densely labeled images across 17 classes. This dataset connects the face semantics with appearances, enabling our new constrained translation setting. We further show the editability of our results, and extend our method to manga images, by generating the first manga parsing pseudo data. Extensive experiments demonstrate the values of our new dataset and method, resulting in the first feasible solution on anime translation.
format text
author LI, Zhansheng
XU, Yangyang
ZHAO, Nanxuan
ZHOU, Yang
LIU, Yongtuo
LIN, Dahua
HE, Shengfeng
author_facet LI, Zhansheng
XU, Yangyang
ZHAO, Nanxuan
ZHOU, Yang
LIU, Yongtuo
LIN, Dahua
HE, Shengfeng
author_sort LI, Zhansheng
title Parsing-conditioned anime translation: A new dataset and method
title_short Parsing-conditioned anime translation: A new dataset and method
title_full Parsing-conditioned anime translation: A new dataset and method
title_fullStr Parsing-conditioned anime translation: A new dataset and method
title_full_unstemmed Parsing-conditioned anime translation: A new dataset and method
title_sort parsing-conditioned anime translation: a new dataset and method
publisher Institutional Knowledge at Singapore Management University
publishDate 2023
url https://ink.library.smu.edu.sg/sis_research/8358
https://ink.library.smu.edu.sg/context/sis_research/article/9361/viewcontent/Parsing.pdf
_version_ 1787136840334049280