Multilevel contrast strategy for unpaired image-to-image translation

被引:1
作者
Han, Minggui [1 ]
Shao, Mingwen [1 ]
Meng, Lingzhuang [1 ]
Liu, Yuexian [1 ]
Qiao, Yuanjian [1 ]
机构
[1] China Univ Petr, Qingdao Inst Software, Coll Comp Sci & Technol, Qingdao, Peoples R China
基金
中国国家自然科学基金;
关键词
image-to-image translation; contrastive learning; multilevel contrast strategy; GENERATIVE ADVERSARIAL NETWORK; REPRESENTATION;
D O I
10.1117/1.JEI.32.6.063030
中图分类号
TM [电工技术]; TN [电子技术、通信技术];
学科分类号
0808 ; 0809 ;
摘要
Contrastive learning for unpaired image-to-image translation utilizes adversarial loss to ensure the realism of generated images in the target domain and incorporates pixel-wise contrastive loss to maximize the correlation between them. However, existing methods only contrast pixel-wise features, ignoring higher-level features, and the pixel-wise contrast is imperfect, which leads to poorer perceptual and visual results. In order to alleviate these problems, we propose an effective multilevel contrast strategy for unpaired image-to-image translation (MLCUT), which contrasts features at three levels to generate more harmonious and realistic images. Specifically, we strengthen the pixel-wise level contrast and introduce the contrasts of plane and voxel-wise levels. On the one hand, MLCUT enhances training effectiveness by picking over hard negative keys for each query at the pixel-wise level. On the other hand, we strengthen the learning preferences of generators on features of objects rather than backgrounds by contrasting the plane-wise discriminative matrices in adversarial loss. Furthermore, by contrasting voxel-wise global semantic vectors, MLCUT effectively improves the realism of generated images and avoids mode collapse. Qualitative and quantitative experiments demonstrate that our method effectively improves performance in perception and vision.
引用
收藏
页数:18
相关论文
共 50 条
  • [41] Unsupervised Image-to-Image Translation with Style Consistency
    Lai, Binxin
    Wang, Yuan-Gen
    PATTERN RECOGNITION AND COMPUTER VISION, PRCV 2023, PT VI, 2024, 14430 : 322 - 334
  • [42] Research on Image-to-Image Translation with Capsule Network
    Ye, Jian
    Chang, Qing
    Jia, Xiaotian
    ARTIFICIAL NEURAL NETWORKS AND MACHINE LEARNING - ICANN 2019: THEORETICAL NEURAL COMPUTATION, PT I, 2019, 11727 : 141 - 151
  • [43] Edge Sensitive Unsupervised Image-to-Image Translation
    Akkaya, Ibrahim Batuhan
    Halici, Ugur
    2020 28TH SIGNAL PROCESSING AND COMMUNICATIONS APPLICATIONS CONFERENCE (SIU), 2020,
  • [44] Equivariant Adversarial Network for Image-to-image Translation
    Zareapoor, Masoumeh
    Yang, Jie
    ACM TRANSACTIONS ON MULTIMEDIA COMPUTING COMMUNICATIONS AND APPLICATIONS, 2021, 17 (02)
  • [45] Consistent Embedded GAN for Image-to-Image Translation
    Xiong, Feng
    Wang, Qianqian
    Gao, Quanxue
    IEEE ACCESS, 2019, 7 : 126651 - 126661
  • [46] Robotic Instrument Segmentation With Image-to-Image Translation
    Colleoni, Emanuele
    Stoyanov, Danail
    IEEE ROBOTICS AND AUTOMATION LETTERS, 2021, 6 (02) : 935 - 942
  • [47] Edge-guided Adversarial Network Based on Contrastive Learning for Image-to-Image Translation
    Zhu, Chen
    Lai, Ru
    Bi, Luzheng
    Wang, Xuyang
    Du, Jiarong
    2021 PROCEEDINGS OF THE 40TH CHINESE CONTROL CONFERENCE (CCC), 2021, : 7949 - 7954
  • [48] Image-to-Image Translation for Near-Infrared Image Colorization
    Kim, Hyeongyu
    Kim, Jonghyun
    Kim, Joongkyu
    2022 INTERNATIONAL CONFERENCE ON ELECTRONICS, INFORMATION, AND COMMUNICATION (ICEIC), 2022,
  • [49] TriGAN: image-to-image translation for multi-source domain adaptation
    Roy, Subhankar
    Siarohin, Aliaksandr
    Sangineto, Enver
    Sebe, Nicu
    Ricci, Elisa
    MACHINE VISION AND APPLICATIONS, 2021, 32 (01)
  • [50] Image-to-Image Translation Using Identical-Pair Adversarial Networks
    Sung, Thai Leang
    Lee, Hyo Jong
    APPLIED SCIENCES-BASEL, 2019, 9 (13):