Blur2Sharp: A GAN-Based Model for Document Image Deblurring

被引:5
作者
Neji, Hala [1 ,2 ,3 ]
Ben Halima, Mohamed [2 ]
Hamdani, Tarek. M. [2 ]
Nogueras-Iso, Javier [3 ]
Alimi, Adel M. [2 ,4 ]
机构
[1] Univ Gabes, Natl Engn Sch Gabes ENIG, Gabes, Tunisia
[2] Univ Sfax, Natl Engn Sch Sfax ENIS, REGIM Lab Res Groups Intelligent Machines, Sfax, Tunisia
[3] Univ Zaragoza, Aragon Inst Engn Res I3A, Zaragoza, Spain
[4] Univ Johannesburg, Fac Engn & Built Environm, Dept Elect & Elect Engn Sci, Johannesburg, South Africa
关键词
Generative adversarial network (GAN); Cycle-consistent generative adversarial network (CycleGAN); Document deblurring; Blind deconvolution; Motion blur; Out-of-focus blur;
D O I
10.2991/ijcis.d.210407.001
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
The advances in mobile technology and portable cameras have facilitated enormously the acquisition of text images. However, the blur caused by camera shake or out-of-focus problems may affect the quality of acquired images and their use as input for optical character recognition (OCR) or other types of document processing. This work proposes an end-to-end model for document deblurring using cycle-consistent adversarial networks. The main novelty of this work is to achieve blind document deblurring, i.e., deblurring without knowledge of the blur kernel. Our method, named "Blur2Sharp CycleGAN," generates a sharp image from a blurry one and shows how cycle-consistent generative adversarial networks (CycleGAN) can be used in document deblurring. Using only a blurred image as input, we try to generate the sharp image. Thus, no information about the blur kernel is required. In the evaluation part, we use peak signal to noise ratio (PSNR) and structural similarity index (SSIM) to compare the deblurring images. The experiments demonstrate a clear improvement in visual quality with respect to the state-of-the-art using a dataset of text images. (C) 2021 The Authors. Published by Atlantis Press B.V.
引用
收藏
页码:1315 / 1321
页数:7
相关论文
共 21 条
  • [1] [Anonymous], 2015, P BMVC, DOI DOI 10.5244/C.29.6
  • [2] Chen XG, 2011, PROC CVPR IEEE, P369, DOI 10.1109/CVPR.2011.5995568
  • [3] Cho HJ, 2012, LECT NOTES COMPUT SC, V7576, P524, DOI 10.1007/978-3-642-33715-4_38
  • [4] Fu B., INT J COMPUT INT SYS, V14, P84
  • [5] Google Open Source,, 2021, TESSERACT OCR
  • [6] Exploring the Effects of Blur and Deblurring to Visual Object Tracking
    Guo, Qing
    Feng, Wei
    Gao, Ruijun
    Liu, Yang
    Wang, Song
    [J]. IEEE TRANSACTIONS ON IMAGE PROCESSING, 2021, 30 : 1812 - 1824
  • [7] Image-to-Image Translation with Conditional Adversarial Networks
    Isola, Phillip
    Zhu, Jun-Yan
    Zhou, Tinghui
    Efros, Alexei A.
    [J]. 30TH IEEE CONFERENCE ON COMPUTER VISION AND PATTERN RECOGNITION (CVPR 2017), 2017, : 5967 - 5976
  • [8] Text image deblurring via two-tone prior
    Jiang, Xiaolei
    Yao, Hongxun
    Zhao, Sicheng
    [J]. NEUROCOMPUTING, 2017, 242 : 1 - 14
  • [9] Progressive Semantic Face Deblurring
    Lee, Tae Bok
    Jung, Soo Hyun
    Heo, Yong Seok
    [J]. IEEE ACCESS, 2020, 8 (08): : 223548 - 223561
  • [10] Compressed Sensing Image Reconstruction Based on Convolutional Neural Network
    Liu, Yuhong
    Liu, Shuying
    Li, Cuiran
    Yang, Danfeng
    [J]. INTERNATIONAL JOURNAL OF COMPUTATIONAL INTELLIGENCE SYSTEMS, 2019, 12 (02) : 873 - 880