Editing out-of-domain GAN inversion via differential activations

Despite the demonstrated editing capacity in the latent space of a pretrained GAN model, inverting real-world images is stuck in a dilemma that the reconstruction cannot be faithful to the original input. The main reason for this is that the distributions between training and real-world data are mis...

Full description

Saved in:
Bibliographic Details
Main Authors: SONG, Haorui, DU, Yong, XIANG, Tianyi, DONG, Junyu, QIN, Jing, HE, Shengfeng
Format: text
Language:English
Published: Institutional Knowledge at Singapore Management University 2022
Subjects:
Online Access:https://ink.library.smu.edu.sg/sis_research/8426
https://ink.library.smu.edu.sg/context/sis_research/article/9429/viewcontent/2207.08134.pdf
Tags: Add Tag
No Tags, Be the first to tag this record!
Institution: Singapore Management University
Language: English
id sg-smu-ink.sis_research-9429
record_format dspace
spelling sg-smu-ink.sis_research-94292024-01-09T03:29:20Z Editing out-of-domain GAN inversion via differential activations SONG, Haorui DU, Yong XIANG, Tianyi DONG, Junyu QIN, Jing HE, Shengfeng Despite the demonstrated editing capacity in the latent space of a pretrained GAN model, inverting real-world images is stuck in a dilemma that the reconstruction cannot be faithful to the original input. The main reason for this is that the distributions between training and real-world data are misaligned, and because of that, it is unstable of GAN inversion for real image editing. In this paper, we propose a novel GAN prior based editing framework to tackle the out-of-domain inversion problem with a composition-decomposition paradigm. In particular, during the phase of composition, we introduce a differential activation module for detecting semantic changes from a global perspective, i.e., the relative gap between the features of edited and unedited images. With the aid of the generated Diff-CAM mask, a coarse reconstruction can intuitively be composited by the paired original and edited images. In this way, the attribute-irrelevant regions can be survived in almost whole, while the quality of such an intermediate result is still limited by an unavoidable ghosting effect. Consequently, in the decomposition phase, we further present a GAN prior based deghosting network for separating the final fine edited image from the coarse reconstruction. Extensive experiments exhibit superiorities over the state-of-the-art methods, in terms of qualitative and quantitative evaluations. The robustness and flexibility of our method is also validated on both scenarios of single attribute and multi-attribute manipulations. Code is available at https://github.com/HaoruiSong622/Editing-Out-of-Domain. 2022-10-01T07:00:00Z text application/pdf https://ink.library.smu.edu.sg/sis_research/8426 info:doi/10.1007/978-3-031-19790-1_1 https://ink.library.smu.edu.sg/context/sis_research/article/9429/viewcontent/2207.08134.pdf http://creativecommons.org/licenses/by-nc-nd/4.0/ Research Collection School Of Computing and Information Systems eng Institutional Knowledge at Singapore Management University Image reconstruction Semantics Artificial Intelligence and Robotics
institution Singapore Management University
building SMU Libraries
continent Asia
country Singapore
Singapore
content_provider SMU Libraries
collection InK@SMU
language English
topic Image reconstruction
Semantics
Artificial Intelligence and Robotics
spellingShingle Image reconstruction
Semantics
Artificial Intelligence and Robotics
SONG, Haorui
DU, Yong
XIANG, Tianyi
DONG, Junyu
QIN, Jing
HE, Shengfeng
Editing out-of-domain GAN inversion via differential activations
description Despite the demonstrated editing capacity in the latent space of a pretrained GAN model, inverting real-world images is stuck in a dilemma that the reconstruction cannot be faithful to the original input. The main reason for this is that the distributions between training and real-world data are misaligned, and because of that, it is unstable of GAN inversion for real image editing. In this paper, we propose a novel GAN prior based editing framework to tackle the out-of-domain inversion problem with a composition-decomposition paradigm. In particular, during the phase of composition, we introduce a differential activation module for detecting semantic changes from a global perspective, i.e., the relative gap between the features of edited and unedited images. With the aid of the generated Diff-CAM mask, a coarse reconstruction can intuitively be composited by the paired original and edited images. In this way, the attribute-irrelevant regions can be survived in almost whole, while the quality of such an intermediate result is still limited by an unavoidable ghosting effect. Consequently, in the decomposition phase, we further present a GAN prior based deghosting network for separating the final fine edited image from the coarse reconstruction. Extensive experiments exhibit superiorities over the state-of-the-art methods, in terms of qualitative and quantitative evaluations. The robustness and flexibility of our method is also validated on both scenarios of single attribute and multi-attribute manipulations. Code is available at https://github.com/HaoruiSong622/Editing-Out-of-Domain.
format text
author SONG, Haorui
DU, Yong
XIANG, Tianyi
DONG, Junyu
QIN, Jing
HE, Shengfeng
author_facet SONG, Haorui
DU, Yong
XIANG, Tianyi
DONG, Junyu
QIN, Jing
HE, Shengfeng
author_sort SONG, Haorui
title Editing out-of-domain GAN inversion via differential activations
title_short Editing out-of-domain GAN inversion via differential activations
title_full Editing out-of-domain GAN inversion via differential activations
title_fullStr Editing out-of-domain GAN inversion via differential activations
title_full_unstemmed Editing out-of-domain GAN inversion via differential activations
title_sort editing out-of-domain gan inversion via differential activations
publisher Institutional Knowledge at Singapore Management University
publishDate 2022
url https://ink.library.smu.edu.sg/sis_research/8426
https://ink.library.smu.edu.sg/context/sis_research/article/9429/viewcontent/2207.08134.pdf
_version_ 1787590773547466752