The spatially-correlative loss for various image translation tasks
We propose a novel spatially-correlative loss that is simple, efficient and yet effective for preserving scene structure consistency while supporting large appearance changes during unpaired image-to-image (I2I) translation. Previous methods attempt this by using pixel-level cycle-consistency or fea...
Saved in:
Main Authors: | , , |
---|---|
Other Authors: | |
Format: | Conference or Workshop Item |
Language: | English |
Published: |
2021
|
Subjects: | |
Online Access: | https://hdl.handle.net/10356/151225 |
Tags: |
Add Tag
No Tags, Be the first to tag this record!
|
Institution: | Nanyang Technological University |
Language: | English |
id |
sg-ntu-dr.10356-151225 |
---|---|
record_format |
dspace |
spelling |
sg-ntu-dr.10356-1512252021-06-11T11:14:41Z The spatially-correlative loss for various image translation tasks Zheng, Chuanxia Cham, Tat-Jen Cai, Jianfei School of Computer Science and Engineering IEEE Conference on Computer Vision and Pattern Recognition Engineering::Computer science and engineering::Computing methodologies::Pattern recognition Spatially-correlative Loss Translation Tasks We propose a novel spatially-correlative loss that is simple, efficient and yet effective for preserving scene structure consistency while supporting large appearance changes during unpaired image-to-image (I2I) translation. Previous methods attempt this by using pixel-level cycle-consistency or feature-level matching losses, but the domain-specific nature of these losses hinder translation across large domain gaps. To address this, we exploit the spatial patterns of self-similarity as a means of defining scene structure. Our spatially-correlative loss is geared towards only capturing spatial relationships within an image rather than domain appearance. We also introduce a new self-supervised learning method to explicitly learn spatially-correlative maps for each specific translation task. We show distinct improvement over baseline models in all three modes of unpaired I2I translation: single-modal, multi-modal, and even single-image translation. This new loss can easily be integrated into existing network architectures and thus allows wide applicability. 2021-06-11T11:14:40Z 2021-06-11T11:14:40Z 2021 Conference Paper Zheng, C., Cham, T. & Cai, J. (2021). The spatially-correlative loss for various image translation tasks. IEEE Conference on Computer Vision and Pattern Recognition. https://hdl.handle.net/10356/151225 en © 2021 Institute of Electrical and Electronics Engineers (IEEE). All rights reserved. |
institution |
Nanyang Technological University |
building |
NTU Library |
continent |
Asia |
country |
Singapore Singapore |
content_provider |
NTU Library |
collection |
DR-NTU |
language |
English |
topic |
Engineering::Computer science and engineering::Computing methodologies::Pattern recognition Spatially-correlative Loss Translation Tasks |
spellingShingle |
Engineering::Computer science and engineering::Computing methodologies::Pattern recognition Spatially-correlative Loss Translation Tasks Zheng, Chuanxia Cham, Tat-Jen Cai, Jianfei The spatially-correlative loss for various image translation tasks |
description |
We propose a novel spatially-correlative loss that is simple, efficient and yet effective for preserving scene structure consistency while supporting large appearance changes during unpaired image-to-image (I2I) translation. Previous methods attempt this by using pixel-level cycle-consistency or feature-level matching losses, but the domain-specific nature of these losses hinder translation across large domain gaps. To address this, we exploit the spatial patterns of self-similarity as a means of defining scene structure. Our spatially-correlative loss is geared towards only capturing spatial relationships within an image rather than domain appearance. We also introduce a new self-supervised learning method to explicitly learn spatially-correlative maps for each specific translation task. We show distinct improvement over baseline models in all three modes of unpaired I2I translation: single-modal, multi-modal, and even single-image translation. This new loss can easily be integrated into existing network architectures and thus allows wide applicability. |
author2 |
School of Computer Science and Engineering |
author_facet |
School of Computer Science and Engineering Zheng, Chuanxia Cham, Tat-Jen Cai, Jianfei |
format |
Conference or Workshop Item |
author |
Zheng, Chuanxia Cham, Tat-Jen Cai, Jianfei |
author_sort |
Zheng, Chuanxia |
title |
The spatially-correlative loss for various image translation tasks |
title_short |
The spatially-correlative loss for various image translation tasks |
title_full |
The spatially-correlative loss for various image translation tasks |
title_fullStr |
The spatially-correlative loss for various image translation tasks |
title_full_unstemmed |
The spatially-correlative loss for various image translation tasks |
title_sort |
spatially-correlative loss for various image translation tasks |
publishDate |
2021 |
url |
https://hdl.handle.net/10356/151225 |
_version_ |
1702431235846963200 |