Back to Search Start Over

Unsupervised Object Transfiguration with Attention.

Authors :
Ye, Zihan
Lyu, Fan
Li, Linyan
Sun, Yu
Fu, Qiming
Hu, Fuyuan
Source :
Cognitive Computation; Dec2019, Vol. 11 Issue 6, p869-878, 10p
Publication Year :
2019

Abstract

Object transfiguration is a subtask of the image-to-image translation, which translates two independent image sets and has a wide range of applications. Recently, some studies based on Generative Adversarial Network (GAN) have achieved impressive results in the image-to-image translation. However, the object transfiguration task only translates regions containing target objects instead of whole images; most of the existing methods never consider this issue, which results in mistranslation on the backgrounds of images. To address this problem, we present a novel pipeline called Deep Attention Unit Generative Adversarial Networks (DAU-GAN). During the translating process, the DAU computes attention masks that point out where the target objects are. DAU makes GAN concentrate on translating target objects while ignoring meaningless backgrounds. Additionally, we construct an attention-consistent loss and a background-consistent loss to compel our model to translate intently target objects and preserve backgrounds further effectively. We have comparison experiments on three popular related datasets, demonstrating that the DAU-GAN achieves superior performance to the state-of-the-art. We also export attention masks in different stages to confirm its effect during the object transfiguration task. The proposed DAU-GAN can translate object effectively as well as preserve backgrounds information at the same time. In our model, DAU learns to focus on the most important information by producing attention masks. These masks compel DAU-GAN to effectively distinguish target objects and backgrounds during the translation process and to achieve impressive translation results in two subsets of ImageNet and CelebA. Moreover, the results show that we cannot only investigate the model from the image itself but also research from other modal information. [ABSTRACT FROM AUTHOR]

Details

Language :
English
ISSN :
18669956
Volume :
11
Issue :
6
Database :
Complementary Index
Journal :
Cognitive Computation
Publication Type :
Academic Journal
Accession number :
139721647
Full Text :
https://doi.org/10.1007/s12559-019-09633-3