Face Inpainting with Pre-trained Image Transformers

被引:0
|
作者
Gonc, Kaan [1 ]
Saglam, Baturay [2 ]
Kozat, Suleyman S. [2 ]
Dibeklioglu, Hamdi [1 ]
机构
[1] Bilkent Univ, Bilgisayar Muhendisligi Bolumu, Ankara, Turkey
[2] Bilkent Univ, Elekt & Elekt Muhendisligi Bolumu, Ankara, Turkey
来源
2022 30TH SIGNAL PROCESSING AND COMMUNICATIONS APPLICATIONS CONFERENCE, SIU | 2022年
关键词
image inpainting; transformers; deep generative models;
D O I
10.1109/SIU55565.2022.9864676
中图分类号
TP39 [计算机的应用];
学科分类号
081203 ; 0835 ;
摘要
Image inpainting is an underdetermined inverse problem that allows various contents to fill in the missing or damaged regions realistically. Convolutional neural networks (CNNs) are commonly used to create aesthetically pleasing content, yet CNNs have restricted perception fields for collecting global characteristics. Transformers enable long-range relationships to be modeled and different content generated with autoregressive modeling of pixel-sequence distributions using image-level attention mechanism. However, the current approaches to inpainting with transformers are limited to task-specific datasets and require larger-scale data. We introduce an approach to image inpainting by leveraging pre-trained vision transformers to remedy this issue. Experiments show that our approach can outperform CNN-based approaches and have a remarkable performance closer to the task-specific transformer methods.
引用
收藏
页数:4
相关论文
共 50 条
  • [11] DRVMon-VM: Distracted driver recognition using large pre-trained video transformers
    Pizarro, Ricardo
    Bergasa, Luis M.
    Baumela, Luis
    Buenaposada, Jose M.
    Barea, Rafael
    2024 35TH IEEE INTELLIGENT VEHICLES SYMPOSIUM, IEEE IV 2024, 2024, : 1901 - 1906
  • [12] Can pre-trained Transformers be used in detecting complex sensitive sentences? - A Monsanto case study
    Timmer, Roelien C.
    Liebowitz, David
    Nepal, Surya
    Kanhere, Salil S.
    2021 THIRD IEEE INTERNATIONAL CONFERENCE ON TRUST, PRIVACY AND SECURITY IN INTELLIGENT SYSTEMS AND APPLICATIONS (TPS-ISA 2021), 2021, : 90 - 97
  • [13] SPIQ: A Self-Supervised Pre-Trained Model for Image Quality Assessment
    Chen, Pengfei
    Li, Leida
    Wu, Qingbo
    Wu, Jinjian
    IEEE SIGNAL PROCESSING LETTERS, 2022, 29 : 513 - 517
  • [14] Diverse Image Inpainting with Bidirectional and Autoregressive Transformers
    Yu, Yingchen
    Zhan, Fangneng
    Wu, Rongliang
    Pan, Jianxiong
    Cui, Kaiwen
    Lu, Shijian
    Ma, Feiying
    Xie, Xuansong
    Miao, Chunyan
    PROCEEDINGS OF THE 29TH ACM INTERNATIONAL CONFERENCE ON MULTIMEDIA, MM 2021, 2021, : 69 - 78
  • [15] Deep Learning-based POS Tagger and Chunker for Odia Language Using Pre-trained Transformers
    Dalai, Tusarkanta
    Kumarmishra, Tapas
    Sa, Andpankaj K.
    ACM TRANSACTIONS ON ASIAN AND LOW-RESOURCE LANGUAGE INFORMATION PROCESSING, 2024, 23 (02)
  • [16] PART: Pre-trained Authorship Representation Transformer
    Huertas-Tato, Javier
    Martin, Alejandro
    Camacho, David
    HUMAN-CENTRIC COMPUTING AND INFORMATION SCIENCES, 2024, 14
  • [17] A Survey on Time-Series Pre-Trained Models
    Ma, Qianli
    Liu, Zhen
    Zheng, Zhenjing
    Huang, Ziyang
    Zhu, Siying
    Yu, Zhongzhong
    Kwok, James T.
    IEEE TRANSACTIONS ON KNOWLEDGE AND DATA ENGINEERING, 2024, 36 (12) : 7536 - 7555
  • [18] Face Image Inpainting via Variational Autoencoder
    Zhang X.
    Cheng L.
    Bai S.
    Zhang F.
    Sun N.
    Wang Z.
    Jisuanji Fuzhu Sheji Yu Tuxingxue Xuebao/Journal of Computer-Aided Design and Computer Graphics, 2020, 32 (03): : 401 - 409
  • [19] CoLaNet: Adaptive Context and Latent Information Blending for Face Image Inpainting
    Park, Joonkyu
    Hong, Cheeun
    Baik, Sungyong
    Lee, Kyoung Mu
    IEEE SIGNAL PROCESSING LETTERS, 2024, 31 : 91 - 95
  • [20] A deep connection to Khasi language through pre-trained embedding
    Thabah, N. Donald Jefferson
    Mitri, Aiom Minnette
    Saha, Goutam
    Maji, Arnab Kumar
    Purkayastha, Bipul Shyam
    INNOVATIONS IN SYSTEMS AND SOFTWARE ENGINEERING, 2025, 21 (01) : 179 - 193