TY - GEN
T1 - CSAGAN
T2 - 2021 IEEE International Conference on Systems, Man, and Cybernetics, SMC 2021
AU - Yang, Rui
AU - Peng, Chao
AU - Wang, Chenchao
AU - Wang, Mengdan
AU - Chen, Yao
AU - Zheng, Peng
AU - Xiong, Neal N.
N1 - Publisher Copyright:
© 2021 IEEE.
PY - 2021
Y1 - 2021
N2 - Unsupervised image-to-image translation is to learn a mapping function from one image domain to another with unpaired samples, which is an important task of computer vision. However, current unsupervised image-to-image translation methods only perform well on certain datasets. To handle the limitation, this paper proposes a novel framework termed as CSAGAN which contains a new discriminator structure, a novel attention module, and a new normalized function. The discriminator is an attention-guided feature pyramid discriminator which makes use of low-level and high-level features to determine an image's realness. The new attention module integrating channel attention and spatial attention can guide generators focus on the most discriminative regions of feature maps to generate high-quality translated images. Moreover, our attention module embedded into generators requires less computation compared with other self-attention methods. In addition, the new normalized function helps generators limberly control the variation of shape, color, and texture through learning parameters. Experimental results indicate that our approach performs better than the current state-of-the-art methods.
AB - Unsupervised image-to-image translation is to learn a mapping function from one image domain to another with unpaired samples, which is an important task of computer vision. However, current unsupervised image-to-image translation methods only perform well on certain datasets. To handle the limitation, this paper proposes a novel framework termed as CSAGAN which contains a new discriminator structure, a novel attention module, and a new normalized function. The discriminator is an attention-guided feature pyramid discriminator which makes use of low-level and high-level features to determine an image's realness. The new attention module integrating channel attention and spatial attention can guide generators focus on the most discriminative regions of feature maps to generate high-quality translated images. Moreover, our attention module embedded into generators requires less computation compared with other self-attention methods. In addition, the new normalized function helps generators limberly control the variation of shape, color, and texture through learning parameters. Experimental results indicate that our approach performs better than the current state-of-the-art methods.
KW - Attention
KW - Generative Adversarial Networks
KW - Unsupervised Image-to-Image Translation
UR - https://www.scopus.com/pages/publications/85124325422
U2 - 10.1109/SMC52423.2021.9658979
DO - 10.1109/SMC52423.2021.9658979
M3 - 会议稿件
AN - SCOPUS:85124325422
T3 - Conference Proceedings - IEEE International Conference on Systems, Man and Cybernetics
SP - 3258
EP - 3265
BT - 2021 IEEE International Conference on Systems, Man, and Cybernetics, SMC 2021
PB - Institute of Electrical and Electronics Engineers Inc.
Y2 - 17 October 2021 through 20 October 2021
ER -