摘要
This paper proposes a novel task for saliency-guided image translation, with the goal of image-to-image translation conditioned on the user specified saliency map. To address this problem, we develop a novel generative adversarial network (GAN) method -based model, called SalG-GAN method. Given the original image and target saliency map, proposed method can generate a translated image that satisfies the target saliency map. In proposed method, a disentangled representation framework is proposed to encourage the model to learn diverse translations for the same target saliency condition. A saliency-based attention module is introduced as a special attention mechanism to facilitate the developed structures of saliency-guided generator, saliency cue encoder, and saliency-guided global and local discriminators. Furthermore, we build a synthetic dataset and a real-world dataset with labeled visual attention for training and evaluating proposed method. The experimental results on both datasets verify the effectiveness of our model for saliency-guided image translation.
| 投稿的翻译标题 | Saliency-guided image translation |
|---|---|
| 源语言 | 繁体中文 |
| 页(从-至) | 2689-2698 |
| 页数 | 10 |
| 期刊 | Beijing Hangkong Hangtian Daxue Xuebao/Journal of Beijing University of Aeronautics and Astronautics |
| 卷 | 49 |
| 期 | 10 |
| DOI | |
| 出版状态 | 已出版 - 1 10月 2023 |
关键词
- attention mechanism
- dataset
- generative adversarial network
- image translation
- saliency
指纹
探究 '显著性指导下图像迁移' 的科研主题。它们共同构成独一无二的指纹。引用此
- APA
- Author
- BIBTEX
- Harvard
- Standard
- RIS
- Vancouver