JurassicWorld Remake: Bringing Ancient Fossils Back to Life via Zero-Shot Long Image-to-Image Translation

被引:0
|
作者
Martin, Alexander [1 ]
Zheng, Haitian [1 ]
An, Jie [1 ]
Luo, Jiebo [1 ]
机构
[1] Univ Rochester, Rochester, NY 14627 USA
关键词
image-to-image translation; large domain gap; stable diffusion;
D O I
10.1145/3581783.3612708
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
With a strong understanding of the target domain from natural language, we produce promising results in translating across large domain gaps and bringing skeletons back to life. In thiswork, we use text-guided latent diffusion models for zero-shot image-to-image translation (I2I) across large domain gaps (longI2I), where large amounts of new visual features and new geometry need to be generated to enter the target domain. Being able to perform translations across large domain gaps has a wide variety of real-world applications in criminology, astrology, environmental conservation, and paleontology. In this work, we introduce a new task Skull2Animal for translating between skulls and living animals. On this task, we find that unguided Generative Adversarial Networks (GANs) are not capable of translating across large domain gaps. Instead of these traditional I2I methods, we explore the use of guided diffusion and image editing models and provide a new benchmark model, Revive2I, capable of performing zero-shot I2I via text-prompting latent diffusion models. We find that guidance is necessary for longI2I because, to bridge the large domain gap, prior knowledge about the target domain is needed. In addition, we find that prompting provides the best and most scalable information about the target domain as classifier-guided diffusion models require retraining for specific use cases and lack stronger constraints on the target domain because of the wide variety of images they are trained on.
引用
收藏
页码:9320 / 9328
页数:9
相关论文
共 39 条
  • [31] Zero-Shot Low-Light Image Enhancement via Joint Frequency Domain Priors Guided Diffusion
    He, Jinhong
    Palaiahnakote, Shivakumara
    Ning, Aoxiang
    Xue, Minglong
    IEEE SIGNAL PROCESSING LETTERS, 2025, 32 : 1091 - 1095
  • [32] Zero-Shot Self-Supervised Joint Temporal Image and Sensitivity Map Reconstruction via Linear Latent Space
    Zhang, Molin
    Xu, Junshen
    Arefeen, Yamin
    Adalsteinsson, Elfar
    MEDICAL IMAGING WITH DEEP LEARNING, VOL 227, 2023, 227 : 1713 - 1725
  • [33] Mining Fine-Grained Image-Text Alignment for Zero-Shot Captioning via Text-Only Training
    Qiu, Longtian
    Ning, Shan
    He, Xuming
    THIRTY-EIGHTH AAAI CONFERENCE ON ARTIFICIAL INTELLIGENCE, VOL 38 NO 5, 2024, : 4605 - 4613
  • [34] CGUN-2A: Deep Graph Convolutional Network via Contrastive Learning for Large-Scale Zero-Shot Image Classification
    Li, Liangwei
    Liu, Lin
    Du, Xiaohui
    Wang, Xiangzhou
    Zhang, Ziruo
    Zhang, Jing
    Zhang, Ping
    Liu, Juanxiu
    SENSORS, 2022, 22 (24)
  • [35] Zero-Shot Low-Dose CT Image Denoising via Patch-Based Content-Guided Diffusion Models
    Su, Bo
    Hu, Xiangyun
    Zha, Yunfei
    Wu, Zijun
    Ma, Yuncheng
    Xu, Jiabo
    Zhang, Baochang
    IEEE TRANSACTIONS ON INSTRUMENTATION AND MEASUREMENT, 2025, 74
  • [36] Uieanything: zero-shot underwater image enhancement via advanced depth estimation, white balance models, and improved sea-thru
    Shao, Jinxin
    Zhang, Haosu
    Miao, Jianming
    PATTERN ANALYSIS AND APPLICATIONS, 2025, 28 (02)
  • [37] ZEPI-Net: Light Field Super Resolution via Internal Cross-Scale Epipolar Plane Image Zero-Shot Learning
    Zhaolin Xiao
    Yinhai Liu
    Haiyan Jin
    Christine Guillemot
    Neural Processing Letters, 2023, 55 : 1649 - 1662
  • [38] ZEPI-Net: Light Field Super Resolution via Internal Cross-Scale Epipolar Plane Image Zero-Shot Learning
    Xiao, Zhaolin
    Liu, Yinhai
    Jin, Haiyan
    Guillemot, Christine
    NEURAL PROCESSING LETTERS, 2023, 55 (02) : 1649 - 1662
  • [39] Significantly improving zero-shot X-ray pathology classification via fine-tuning pre-trained image-text encoders
    Jang, Jongseong
    Kyung, Daeun
    Kim, Seung Hwan
    Lee, Honglak
    Bae, Kyunghoon
    Choi, Edward
    SCIENTIFIC REPORTS, 2024, 14 (01):