Disentangled person image generation

Generating novel, yet realistic, images of persons is a challenging task due to the complex interplay between the different image factors, such as the foreground, background and pose information. In this work, we aim at generating such images based on a novel, two-stage reconstruction pipeline that...

Full description

Saved in:
Bibliographic Details
Main Authors: MA, Liqian, SUN, Qianru, GEORGOULIS, Stamatios, VAN GOOL, Luc, SCHIELE, Bernt, FRITZ, Mario
Format: text
Language:English
Published: Institutional Knowledge at Singapore Management University 2018
Subjects:
Online Access:https://ink.library.smu.edu.sg/sis_research/4456
https://ink.library.smu.edu.sg/context/sis_research/article/5459/viewcontent/Ma_Disentangled_Person_Image_CVPR_2018_paper.pdf
Tags: Add Tag
No Tags, Be the first to tag this record!
Institution: Singapore Management University
Language: English
id sg-smu-ink.sis_research-5459
record_format dspace
spelling sg-smu-ink.sis_research-54592021-02-19T02:25:03Z Disentangled person image generation MA, Liqian SUN, Qianru GEORGOULIS, Stamatios VAN GOOL, Luc SCHIELE, Bernt FRITZ, Mario Generating novel, yet realistic, images of persons is a challenging task due to the complex interplay between the different image factors, such as the foreground, background and pose information. In this work, we aim at generating such images based on a novel, two-stage reconstruction pipeline that learns a disentangled representation of the aforementioned image factors and generates novel person images at the same time. First, a multi-branched reconstruction network is proposed to disentangle and encode the three factors into embedding features, which are then combined to re-compose the input image itself. Second, three corresponding mapping functions are learned in an adversarial manner in order to map Gaussian noise to the learned embedding feature space, for each factor, respectively. Using the proposed framework, we can manipulate the foreground, background and pose of the input image, and also sample new embedding features to generate such targeted manipulations, that provide more control over the generation process. Experiments on the Market-1501 and Deepfashion datasets show that our model does not only generate realistic person images with new foregrounds, backgrounds and poses, but also manipulates the generated factors and interpolates the in-between states. Another set of experiments on Market-1501 shows that our model can also be beneficial for the person re-identification task. 2018-06-01T07:00:00Z text application/pdf https://ink.library.smu.edu.sg/sis_research/4456 info:doi/10.1109/CVPR.2018.00018 https://ink.library.smu.edu.sg/context/sis_research/article/5459/viewcontent/Ma_Disentangled_Person_Image_CVPR_2018_paper.pdf http://creativecommons.org/licenses/by-nc-nd/4.0/ Research Collection School Of Computing and Information Systems eng Institutional Knowledge at Singapore Management University Image generation generative adversarial networks pose estimation person re-identification Artificial Intelligence and Robotics Numerical Analysis and Scientific Computing
institution Singapore Management University
building SMU Libraries
continent Asia
country Singapore
Singapore
content_provider SMU Libraries
collection InK@SMU
language English
topic Image generation
generative adversarial networks
pose estimation
person re-identification
Artificial Intelligence and Robotics
Numerical Analysis and Scientific Computing
spellingShingle Image generation
generative adversarial networks
pose estimation
person re-identification
Artificial Intelligence and Robotics
Numerical Analysis and Scientific Computing
MA, Liqian
SUN, Qianru
GEORGOULIS, Stamatios
VAN GOOL, Luc
SCHIELE, Bernt
FRITZ, Mario
Disentangled person image generation
description Generating novel, yet realistic, images of persons is a challenging task due to the complex interplay between the different image factors, such as the foreground, background and pose information. In this work, we aim at generating such images based on a novel, two-stage reconstruction pipeline that learns a disentangled representation of the aforementioned image factors and generates novel person images at the same time. First, a multi-branched reconstruction network is proposed to disentangle and encode the three factors into embedding features, which are then combined to re-compose the input image itself. Second, three corresponding mapping functions are learned in an adversarial manner in order to map Gaussian noise to the learned embedding feature space, for each factor, respectively. Using the proposed framework, we can manipulate the foreground, background and pose of the input image, and also sample new embedding features to generate such targeted manipulations, that provide more control over the generation process. Experiments on the Market-1501 and Deepfashion datasets show that our model does not only generate realistic person images with new foregrounds, backgrounds and poses, but also manipulates the generated factors and interpolates the in-between states. Another set of experiments on Market-1501 shows that our model can also be beneficial for the person re-identification task.
format text
author MA, Liqian
SUN, Qianru
GEORGOULIS, Stamatios
VAN GOOL, Luc
SCHIELE, Bernt
FRITZ, Mario
author_facet MA, Liqian
SUN, Qianru
GEORGOULIS, Stamatios
VAN GOOL, Luc
SCHIELE, Bernt
FRITZ, Mario
author_sort MA, Liqian
title Disentangled person image generation
title_short Disentangled person image generation
title_full Disentangled person image generation
title_fullStr Disentangled person image generation
title_full_unstemmed Disentangled person image generation
title_sort disentangled person image generation
publisher Institutional Knowledge at Singapore Management University
publishDate 2018
url https://ink.library.smu.edu.sg/sis_research/4456
https://ink.library.smu.edu.sg/context/sis_research/article/5459/viewcontent/Ma_Disentangled_Person_Image_CVPR_2018_paper.pdf
_version_ 1770574844431171584