Fully deformable network for multiview face image synthesis

Photorealistic multiview face synthesis from a single image is a challenging problem. Existing works mainly learn a texture mapping model from the source to the target faces. However, they rarely consider the geometric constraints on the internal deformation arising from pose variations, which cause...

Full description

Saved in:
Bibliographic Details
Main Authors: XU, Cheng, LI, Keke, LUO, Xuandi, XU, Xuemiao, HE, Shengfeng, ZHANG, Kun
Format: text
Language:English
Published: Institutional Knowledge at Singapore Management University 2022
Subjects:
Online Access:https://ink.library.smu.edu.sg/sis_research/7861
Tags: Add Tag
No Tags, Be the first to tag this record!
Institution: Singapore Management University
Language: English
id sg-smu-ink.sis_research-8864
record_format dspace
spelling sg-smu-ink.sis_research-88642023-06-15T09:00:05Z Fully deformable network for multiview face image synthesis XU, Cheng LI, Keke LUO, Xuandi XU, Xuemiao HE, Shengfeng ZHANG, Kun Photorealistic multiview face synthesis from a single image is a challenging problem. Existing works mainly learn a texture mapping model from the source to the target faces. However, they rarely consider the geometric constraints on the internal deformation arising from pose variations, which causes a high level of uncertainty in face pose modeling, and hence, produces inferior results for large pose variations. Moreover, current methods typically suffer from undesired facial details loss due to the adoption of the de-facto standard encoder-decoder architecture without any skip connections (SCs). In this article, we directly learn and exploit geometric constraints and propose a fully deformable network to simultaneously model the deformations of both landmarks and faces for face synthesis. Specifically, our model consists of two parts: a deformable landmark learning network (DLLN) and a gated deformable face synthesis network (GDFSN). The DLLN converts an initial reference landmark to an individual-specific target landmark as delicate pose guidance for face rotation. The GDFSN adopts a dual-stream structure, with one stream estimating the deformation of two views in the form of convolution offsets according to the source pose and the converted target pose, and the other leveraging the predicted deformation offsets to create the target face. In this way, individual-aware pose changes are explicitly modeled in the face generator to cope with geometric transformation, by adaptively focusing on pertinent regions of the source face. To compensate for offset estimation errors, we introduce a soft-gating mechanism for adaptive fusion between deformable features and primitive features. Additionally, a pose-aligned SC (PASC) is tailored to propagate low-level input features to the appropriate positions in the output features for further enhancing the facial details and identity preservation. Extensive experiments on six benchmarks show that our approach performs favorably against the state-of-the-arts, especially with large pose changes. Code is available at https://github.com/cschengxu/FDFace. 2022-11-07T08:00:00Z text https://ink.library.smu.edu.sg/sis_research/7861 info:doi/10.1109/TNNLS.2022.3216018 Research Collection School Of Computing and Information Systems eng Institutional Knowledge at Singapore Management University Faces Face recognition Strain Deformable models Electronic mail Convolution Standards Deformable convolution gating multiview face synthesis pose-invariant face recognition Information Security
institution Singapore Management University
building SMU Libraries
continent Asia
country Singapore
Singapore
content_provider SMU Libraries
collection InK@SMU
language English
topic Faces
Face recognition
Strain
Deformable models
Electronic mail
Convolution
Standards
Deformable convolution
gating
multiview face synthesis
pose-invariant face recognition
Information Security
spellingShingle Faces
Face recognition
Strain
Deformable models
Electronic mail
Convolution
Standards
Deformable convolution
gating
multiview face synthesis
pose-invariant face recognition
Information Security
XU, Cheng
LI, Keke
LUO, Xuandi
XU, Xuemiao
HE, Shengfeng
ZHANG, Kun
Fully deformable network for multiview face image synthesis
description Photorealistic multiview face synthesis from a single image is a challenging problem. Existing works mainly learn a texture mapping model from the source to the target faces. However, they rarely consider the geometric constraints on the internal deformation arising from pose variations, which causes a high level of uncertainty in face pose modeling, and hence, produces inferior results for large pose variations. Moreover, current methods typically suffer from undesired facial details loss due to the adoption of the de-facto standard encoder-decoder architecture without any skip connections (SCs). In this article, we directly learn and exploit geometric constraints and propose a fully deformable network to simultaneously model the deformations of both landmarks and faces for face synthesis. Specifically, our model consists of two parts: a deformable landmark learning network (DLLN) and a gated deformable face synthesis network (GDFSN). The DLLN converts an initial reference landmark to an individual-specific target landmark as delicate pose guidance for face rotation. The GDFSN adopts a dual-stream structure, with one stream estimating the deformation of two views in the form of convolution offsets according to the source pose and the converted target pose, and the other leveraging the predicted deformation offsets to create the target face. In this way, individual-aware pose changes are explicitly modeled in the face generator to cope with geometric transformation, by adaptively focusing on pertinent regions of the source face. To compensate for offset estimation errors, we introduce a soft-gating mechanism for adaptive fusion between deformable features and primitive features. Additionally, a pose-aligned SC (PASC) is tailored to propagate low-level input features to the appropriate positions in the output features for further enhancing the facial details and identity preservation. Extensive experiments on six benchmarks show that our approach performs favorably against the state-of-the-arts, especially with large pose changes. Code is available at https://github.com/cschengxu/FDFace.
format text
author XU, Cheng
LI, Keke
LUO, Xuandi
XU, Xuemiao
HE, Shengfeng
ZHANG, Kun
author_facet XU, Cheng
LI, Keke
LUO, Xuandi
XU, Xuemiao
HE, Shengfeng
ZHANG, Kun
author_sort XU, Cheng
title Fully deformable network for multiview face image synthesis
title_short Fully deformable network for multiview face image synthesis
title_full Fully deformable network for multiview face image synthesis
title_fullStr Fully deformable network for multiview face image synthesis
title_full_unstemmed Fully deformable network for multiview face image synthesis
title_sort fully deformable network for multiview face image synthesis
publisher Institutional Knowledge at Singapore Management University
publishDate 2022
url https://ink.library.smu.edu.sg/sis_research/7861
_version_ 1770576571290091520