GDFace: Gated deformation for multi-view face image synthesis
Photorealistic multi-view face synthesis from a single image is an important but challenging problem. Existing methods mainly learn a texture mapping model from the source face to the target face. However, they fail to consider the internal deformation caused by the change of poses, leading to the u...
Saved in:
Main Authors: | , , , |
---|---|
Format: | text |
Language: | English |
Published: |
Institutional Knowledge at Singapore Management University
2020
|
Subjects: | |
Online Access: | https://ink.library.smu.edu.sg/sis_research/8520 https://ink.library.smu.edu.sg/context/sis_research/article/9523/viewcontent/123500205__1_.pdf |
Tags: |
Add Tag
No Tags, Be the first to tag this record!
|
Institution: | Singapore Management University |
Language: | English |
id |
sg-smu-ink.sis_research-9523 |
---|---|
record_format |
dspace |
spelling |
sg-smu-ink.sis_research-95232024-01-22T15:07:03Z GDFace: Gated deformation for multi-view face image synthesis XU, Xuemiao LI, Keke XU, Cheng HE, Shengfeng Photorealistic multi-view face synthesis from a single image is an important but challenging problem. Existing methods mainly learn a texture mapping model from the source face to the target face. However, they fail to consider the internal deformation caused by the change of poses, leading to the unsatisfactory synthesized results for large pose variations. In this paper, we propose a Gated Deformable Face Synthesis Network to model the deformation of faces that aids the synthesis of the target face image. Specifically, we propose a dual network that consists of two modules. The first module estimates the deformation of two views in the form of convolution offsets according to the input and target poses. The second one, on the other hand, leverages the predicted deformation offsets to create the target face image. In this way, pose changes are explicitly modeled in the face generator to cope with geometric transformation, by adaptively focusing on pertinent regions of the source image. To compensate offset estimation errors, we introduce a soft-gating mechanism that enables adaptive fusion between deformable features and primitive features. Extensive experimental results on five widely-used benchmarks show that our approach performs favorably against the state-of-the-arts on multi-view face synthesis, especially for large pose changes. 2020-02-01T08:00:00Z text application/pdf https://ink.library.smu.edu.sg/sis_research/8520 info:doi/10.1609/aaai.v34i07.6942 https://ink.library.smu.edu.sg/context/sis_research/article/9523/viewcontent/123500205__1_.pdf http://creativecommons.org/licenses/by-nc-nd/4.0/ Research Collection School Of Computing and Information Systems eng Institutional Knowledge at Singapore Management University Artificial Intelligence and Robotics Graphics and Human Computer Interfaces |
institution |
Singapore Management University |
building |
SMU Libraries |
continent |
Asia |
country |
Singapore Singapore |
content_provider |
SMU Libraries |
collection |
InK@SMU |
language |
English |
topic |
Artificial Intelligence and Robotics Graphics and Human Computer Interfaces |
spellingShingle |
Artificial Intelligence and Robotics Graphics and Human Computer Interfaces XU, Xuemiao LI, Keke XU, Cheng HE, Shengfeng GDFace: Gated deformation for multi-view face image synthesis |
description |
Photorealistic multi-view face synthesis from a single image is an important but challenging problem. Existing methods mainly learn a texture mapping model from the source face to the target face. However, they fail to consider the internal deformation caused by the change of poses, leading to the unsatisfactory synthesized results for large pose variations. In this paper, we propose a Gated Deformable Face Synthesis Network to model the deformation of faces that aids the synthesis of the target face image. Specifically, we propose a dual network that consists of two modules. The first module estimates the deformation of two views in the form of convolution offsets according to the input and target poses. The second one, on the other hand, leverages the predicted deformation offsets to create the target face image. In this way, pose changes are explicitly modeled in the face generator to cope with geometric transformation, by adaptively focusing on pertinent regions of the source image. To compensate offset estimation errors, we introduce a soft-gating mechanism that enables adaptive fusion between deformable features and primitive features. Extensive experimental results on five widely-used benchmarks show that our approach performs favorably against the state-of-the-arts on multi-view face synthesis, especially for large pose changes. |
format |
text |
author |
XU, Xuemiao LI, Keke XU, Cheng HE, Shengfeng |
author_facet |
XU, Xuemiao LI, Keke XU, Cheng HE, Shengfeng |
author_sort |
XU, Xuemiao |
title |
GDFace: Gated deformation for multi-view face image synthesis |
title_short |
GDFace: Gated deformation for multi-view face image synthesis |
title_full |
GDFace: Gated deformation for multi-view face image synthesis |
title_fullStr |
GDFace: Gated deformation for multi-view face image synthesis |
title_full_unstemmed |
GDFace: Gated deformation for multi-view face image synthesis |
title_sort |
gdface: gated deformation for multi-view face image synthesis |
publisher |
Institutional Knowledge at Singapore Management University |
publishDate |
2020 |
url |
https://ink.library.smu.edu.sg/sis_research/8520 https://ink.library.smu.edu.sg/context/sis_research/article/9523/viewcontent/123500205__1_.pdf |
_version_ |
1789483258035044352 |