Proceedings of the International Workshop on Advances in Deep Learning for Image Analysis and Computer Vision (IWADIC 2025)

Attention-Enhanced CycleGAN for Unpaired Image-to-Image Translation

Authors
Wenqi Zheng1, *
1Department of Mathematics, University College London, London, UK
*Corresponding author. Email: wenqi.zheng.23@ucl.ac.uk
Corresponding Author
Wenqi Zheng
Available Online 24 April 2026.
DOI
10.2991/978-94-6239-648-7_100How to use a DOI?
Keywords
Unpaired translation; CycleGAN; Frozen attention; Mask-guided discriminator; Foreground-aware translation
Abstract

Cycle-consistent GANs are widely used for unpaired image-to-image translation, but they often over-translate textures in regions that should remain largely unchanged (e.g., background grass or sky in horse ↔ zebra). This failure mode is encouraged by discriminators that score the full image uniformly, which can reward the generator for spreading domain-specific cues beyond the intended foreground. The paper proposes a lightweight, plug-in attention head attached to the CycleGAN generator to predict a soft foreground attention mask. During training the paper uses this mask to (i) gate the discriminator’s input and (ii) weight the adversarial loss so that discriminator feedback is concentrated on attended regions. A key stability issue in attention-guided adversarial training is that attention maps can drift late in training; to prevent this, the paper monitor attention-map changes across epochs and freeze the attention parameters once the maps stabilize. On the Horse ↔ Zebra benchmark, our method improves Kernel Inception Distance (KID; lower is better) compared with a vanilla CycleGAN baseline and yields visibly cleaner backgrounds with reduced texture bleeding.

Copyright
© 2026 The Author(s)
Open Access
Open Access This chapter is licensed under the terms of the Creative Commons Attribution-NonCommercial 4.0 International License (http://creativecommons.org/licenses/by-nc/4.0/), which permits any noncommercial use, sharing, adaptation, distribution and reproduction in any medium or format, as long as you give appropriate credit to the original author(s) and the source, provide a link to the Creative Commons license and indicate if changes were made.

Download article (PDF)

Volume Title
Proceedings of the International Workshop on Advances in Deep Learning for Image Analysis and Computer Vision (IWADIC 2025)
Series
Advances in Computer Science Research
Publication Date
24 April 2026
ISBN
978-94-6239-648-7
ISSN
2352-538X
DOI
10.2991/978-94-6239-648-7_100How to use a DOI?
Copyright
© 2026 The Author(s)
Open Access
Open Access This chapter is licensed under the terms of the Creative Commons Attribution-NonCommercial 4.0 International License (http://creativecommons.org/licenses/by-nc/4.0/), which permits any noncommercial use, sharing, adaptation, distribution and reproduction in any medium or format, as long as you give appropriate credit to the original author(s) and the source, provide a link to the Creative Commons license and indicate if changes were made.

Cite this article

TY  - CONF
AU  - Wenqi Zheng
PY  - 2026
DA  - 2026/04/24
TI  - Attention-Enhanced CycleGAN for Unpaired Image-to-Image Translation
BT  - Proceedings of the International Workshop on Advances in Deep Learning for Image Analysis and Computer Vision (IWADIC 2025)
PB  - Atlantis Press
SP  - 934
EP  - 942
SN  - 2352-538X
UR  - https://doi.org/10.2991/978-94-6239-648-7_100
DO  - 10.2991/978-94-6239-648-7_100
ID  - Zheng2026
ER  -