GDFace: Gated deformation for multi-view face image synthesis

Photorealistic multi-view face synthesis from a single image is an important but challenging problem. Existing methods mainly learn a texture mapping model from the source face to the target face. However, they fail to consider the internal deformation caused by the change of poses, leading to the u...

Full description

Saved in:
Bibliographic Details
Main Authors: XU, Xuemiao, LI, Keke, XU, Cheng, HE, Shengfeng
Format: text
Language:English
Published: Institutional Knowledge at Singapore Management University 2020
Subjects:
Online Access:https://ink.library.smu.edu.sg/sis_research/8520
https://ink.library.smu.edu.sg/context/sis_research/article/9523/viewcontent/123500205__1_.pdf
Tags: Add Tag
No Tags, Be the first to tag this record!
Institution: Singapore Management University
Language: English
id sg-smu-ink.sis_research-9523
record_format dspace
spelling sg-smu-ink.sis_research-95232024-01-22T15:07:03Z GDFace: Gated deformation for multi-view face image synthesis XU, Xuemiao LI, Keke XU, Cheng HE, Shengfeng Photorealistic multi-view face synthesis from a single image is an important but challenging problem. Existing methods mainly learn a texture mapping model from the source face to the target face. However, they fail to consider the internal deformation caused by the change of poses, leading to the unsatisfactory synthesized results for large pose variations. In this paper, we propose a Gated Deformable Face Synthesis Network to model the deformation of faces that aids the synthesis of the target face image. Specifically, we propose a dual network that consists of two modules. The first module estimates the deformation of two views in the form of convolution offsets according to the input and target poses. The second one, on the other hand, leverages the predicted deformation offsets to create the target face image. In this way, pose changes are explicitly modeled in the face generator to cope with geometric transformation, by adaptively focusing on pertinent regions of the source image. To compensate offset estimation errors, we introduce a soft-gating mechanism that enables adaptive fusion between deformable features and primitive features. Extensive experimental results on five widely-used benchmarks show that our approach performs favorably against the state-of-the-arts on multi-view face synthesis, especially for large pose changes. 2020-02-01T08:00:00Z text application/pdf https://ink.library.smu.edu.sg/sis_research/8520 info:doi/10.1609/aaai.v34i07.6942 https://ink.library.smu.edu.sg/context/sis_research/article/9523/viewcontent/123500205__1_.pdf http://creativecommons.org/licenses/by-nc-nd/4.0/ Research Collection School Of Computing and Information Systems eng Institutional Knowledge at Singapore Management University Artificial Intelligence and Robotics Graphics and Human Computer Interfaces
institution Singapore Management University
building SMU Libraries
continent Asia
country Singapore
Singapore
content_provider SMU Libraries
collection InK@SMU
language English
topic Artificial Intelligence and Robotics
Graphics and Human Computer Interfaces
spellingShingle Artificial Intelligence and Robotics
Graphics and Human Computer Interfaces
XU, Xuemiao
LI, Keke
XU, Cheng
HE, Shengfeng
GDFace: Gated deformation for multi-view face image synthesis
description Photorealistic multi-view face synthesis from a single image is an important but challenging problem. Existing methods mainly learn a texture mapping model from the source face to the target face. However, they fail to consider the internal deformation caused by the change of poses, leading to the unsatisfactory synthesized results for large pose variations. In this paper, we propose a Gated Deformable Face Synthesis Network to model the deformation of faces that aids the synthesis of the target face image. Specifically, we propose a dual network that consists of two modules. The first module estimates the deformation of two views in the form of convolution offsets according to the input and target poses. The second one, on the other hand, leverages the predicted deformation offsets to create the target face image. In this way, pose changes are explicitly modeled in the face generator to cope with geometric transformation, by adaptively focusing on pertinent regions of the source image. To compensate offset estimation errors, we introduce a soft-gating mechanism that enables adaptive fusion between deformable features and primitive features. Extensive experimental results on five widely-used benchmarks show that our approach performs favorably against the state-of-the-arts on multi-view face synthesis, especially for large pose changes.
format text
author XU, Xuemiao
LI, Keke
XU, Cheng
HE, Shengfeng
author_facet XU, Xuemiao
LI, Keke
XU, Cheng
HE, Shengfeng
author_sort XU, Xuemiao
title GDFace: Gated deformation for multi-view face image synthesis
title_short GDFace: Gated deformation for multi-view face image synthesis
title_full GDFace: Gated deformation for multi-view face image synthesis
title_fullStr GDFace: Gated deformation for multi-view face image synthesis
title_full_unstemmed GDFace: Gated deformation for multi-view face image synthesis
title_sort gdface: gated deformation for multi-view face image synthesis
publisher Institutional Knowledge at Singapore Management University
publishDate 2020
url https://ink.library.smu.edu.sg/sis_research/8520
https://ink.library.smu.edu.sg/context/sis_research/article/9523/viewcontent/123500205__1_.pdf
_version_ 1789483258035044352