EAD-GAN: a generative adversarial network for disentangling affine transforms in images

This article proposes a generative adversarial network called explicit affine disentangled generative adversarial network (EAD-GAN), which explicitly disentangles affine transform in a self-supervised manner. We propose an affine transform regularizer to force the InfoGAN to have explicit properties...

Full description

Saved in:
Bibliographic Details
Main Authors: Liu, Letao, Jiang, Xudong, Saerbeck, Martin, Dauwels, Justin
Other Authors: School of Electrical and Electronic Engineering
Format: Article
Language:English
Published: 2023
Subjects:
Online Access:https://hdl.handle.net/10356/164532
Tags: Add Tag
No Tags, Be the first to tag this record!
Institution: Nanyang Technological University
Language: English
id sg-ntu-dr.10356-164532
record_format dspace
spelling sg-ntu-dr.10356-1645322023-01-31T02:53:50Z EAD-GAN: a generative adversarial network for disentangling affine transforms in images Liu, Letao Jiang, Xudong Saerbeck, Martin Dauwels, Justin School of Electrical and Electronic Engineering Engineering::Electrical and electronic engineering Affine Transform Disentanglement This article proposes a generative adversarial network called explicit affine disentangled generative adversarial network (EAD-GAN), which explicitly disentangles affine transform in a self-supervised manner. We propose an affine transform regularizer to force the InfoGAN to have explicit properties of affine transform. To facilitate training an affine transform encoder, we decompose the affine matrix into two separate matrices and infer the explicit transform parameters by the least-squares method. Unlike the existing approaches, representations learned by the proposed EAD-GAN have clear physical meaning, where transforms, such as rotation, horizontal and vertical zooms, skews, and translations, are explicitly learned from training data. Thus, we set different values of each transform parameter individually to generate specifically affine transformed data by the learned network. We show that the proposed EAD-GAN successfully disentangles these attributes on the MNIST, CelebA, and dSprites datasets. EAD-GAN achieves higher disentanglement scores with a large margin compared to the state-of-the-art methods on the dSprites dataset. For example, on the dSprites dataset, EAD-GAN achieves the MIG and DCI score of 0.59 and 0.96 respectively, compared to 0.37 and 0.71, respectively, for the state-of-the-art methods. Economic Development Board (EDB) Published version This work was supported in part by the Singapore Economic Development Board Industrial Postgraduate Program under Grant S17-1298-IPP-II. 2023-01-31T02:53:50Z 2023-01-31T02:53:50Z 2022 Journal Article Liu, L., Jiang, X., Saerbeck, M. & Dauwels, J. (2022). EAD-GAN: a generative adversarial network for disentangling affine transforms in images. IEEE Transactions On Neural Networks and Learning Systems, PP, 1-11. https://dx.doi.org/10.1109/TNNLS.2022.3195533 2162-237X https://hdl.handle.net/10356/164532 10.1109/TNNLS.2022.3195533 35939476 2-s2.0-85136140446 PP 1 11 en S17-1298-IPP-II IEEE transactions on neural networks and learning systems © 2022 The authors. This work is licensed under a Creative Commons Attribution 4.0 License. For more information, see https://creativecommons.org/licenses/by/4.0/ application/pdf
institution Nanyang Technological University
building NTU Library
continent Asia
country Singapore
Singapore
content_provider NTU Library
collection DR-NTU
language English
topic Engineering::Electrical and electronic engineering
Affine Transform
Disentanglement
spellingShingle Engineering::Electrical and electronic engineering
Affine Transform
Disentanglement
Liu, Letao
Jiang, Xudong
Saerbeck, Martin
Dauwels, Justin
EAD-GAN: a generative adversarial network for disentangling affine transforms in images
description This article proposes a generative adversarial network called explicit affine disentangled generative adversarial network (EAD-GAN), which explicitly disentangles affine transform in a self-supervised manner. We propose an affine transform regularizer to force the InfoGAN to have explicit properties of affine transform. To facilitate training an affine transform encoder, we decompose the affine matrix into two separate matrices and infer the explicit transform parameters by the least-squares method. Unlike the existing approaches, representations learned by the proposed EAD-GAN have clear physical meaning, where transforms, such as rotation, horizontal and vertical zooms, skews, and translations, are explicitly learned from training data. Thus, we set different values of each transform parameter individually to generate specifically affine transformed data by the learned network. We show that the proposed EAD-GAN successfully disentangles these attributes on the MNIST, CelebA, and dSprites datasets. EAD-GAN achieves higher disentanglement scores with a large margin compared to the state-of-the-art methods on the dSprites dataset. For example, on the dSprites dataset, EAD-GAN achieves the MIG and DCI score of 0.59 and 0.96 respectively, compared to 0.37 and 0.71, respectively, for the state-of-the-art methods.
author2 School of Electrical and Electronic Engineering
author_facet School of Electrical and Electronic Engineering
Liu, Letao
Jiang, Xudong
Saerbeck, Martin
Dauwels, Justin
format Article
author Liu, Letao
Jiang, Xudong
Saerbeck, Martin
Dauwels, Justin
author_sort Liu, Letao
title EAD-GAN: a generative adversarial network for disentangling affine transforms in images
title_short EAD-GAN: a generative adversarial network for disentangling affine transforms in images
title_full EAD-GAN: a generative adversarial network for disentangling affine transforms in images
title_fullStr EAD-GAN: a generative adversarial network for disentangling affine transforms in images
title_full_unstemmed EAD-GAN: a generative adversarial network for disentangling affine transforms in images
title_sort ead-gan: a generative adversarial network for disentangling affine transforms in images
publishDate 2023
url https://hdl.handle.net/10356/164532
_version_ 1757048199951941632