Subgraph-Induced Extraction Technique for Information (SETI) from Administrative Documents

被引:0
作者
Kafle, Dipendra Sharma [1 ]
Thomas, Eliott [1 ]
Coustaty, Mickael [1 ]
Joseph, Aurelie [2 ]
Doucet, Antoine [1 ]
DAndecy, Vincent Poulain [2 ]
机构
[1] Univ La Rochelle, L3i Ave Michel Crepeau, F-17042 La Rochelle, France
[2] Yooz, Aimargues, France
来源
DOCUMENT ANALYSIS AND RECOGNITION - ICDAR 2023 WORKSHOPS, PT II | 2023年 / 14194卷
关键词
Information Extraction; Invoice; Graph Neural Network;
D O I
10.1007/978-3-031-41501-2_8
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
Information Extraction plays a key role in the automation of auditing processes in administrative documents. However, variety in layout and language is always a challenging task. On the other hand, large volumes of public training datasets related to administrative documents such as invoices are rare to find. In this work, we use Graph Attention Network model for information extraction. This type of model makes it easier to understand the mechanism as compared to classical neural networks due to the visualization of link between entities in the graph. Moreover, it maximizes the layout and structure retrieval which is a crucial advantage in administrative documents. From the same graph, our model learns at different graph levels to encapsulate dynamic and more enriched knowledge in each batch, thus maximizing the generalization on smaller dataset. We present how the model learns in each graph level and compare the results with baselines on private as well as public datasets. Our model succeeds in improving recall and precision scores for some classes in our private dataset and produces comparable results for public datasets designed for Form Understanding and Information Extraction.
引用
收藏
页码:108 / 122
页数:15
相关论文
共 19 条
[1]  
Brody S, 2022, Arxiv, DOI [arXiv:2105.14491, DOI 10.48550/ARXIV.2105.14491]
[2]   Named Entity Recognition and Relation Extraction with Graph Neural Networks in Semi Structured Documents [J].
Carbonell, Manuel ;
Riba, Pau ;
Villegas, Mauricio ;
Fornes, Alicia ;
Llados, Josep .
2020 25TH INTERNATIONAL CONFERENCE ON PATTERN RECOGNITION (ICPR), 2021, :9622-9627
[3]  
Devlin J, 2019, Arxiv, DOI [arXiv:1810.04805, 10.48550/arXiv.1810.04805]
[4]  
Finkel J.R., 2005, ACL
[5]   Information Extraction from Invoices [J].
Hamdi, Ahmed ;
Carel, Elodie ;
Joseph, Aurelie ;
Coustaty, Mickael ;
Doucet, Antoine .
DOCUMENT ANALYSIS AND RECOGNITION - ICDAR 2021, PT II, 2021, 12822 :699-714
[6]   LayoutLMv3: Pre-training for Document AI with Unified Text and Image Masking [J].
Huang, Yupan ;
Lv, Tengchao ;
Cui, Lei ;
Lu, Yutong ;
Wei, Furu .
PROCEEDINGS OF THE 30TH ACM INTERNATIONAL CONFERENCE ON MULTIMEDIA, MM 2022, 2022, :4083-4091
[7]  
Denk TI, 2019, Arxiv, DOI arXiv:1909.04948
[8]   FUNSD: A Dataset for Form Understanding in Noisy Scanned Documents [J].
Jaume, Guillaume ;
Ekenel, Hazim Kemal ;
Thiran, Jean-Philippe .
2019 INTERNATIONAL CONFERENCE ON DOCUMENT ANALYSIS AND RECOGNITION WORKSHOPS (ICDARW) AND 2ND INTERNATIONAL WORKSHOP ON OPEN SERVICES AND TOOLS FOR DOCUMENT ANALYSIS (OST), VOL 2, 2019, :1-6
[9]  
Lee CY, 2021, Arxiv, DOI arXiv:2106.10786
[10]   Focal Loss for Dense Object Detection [J].
Lin, Tsung-Yi ;
Goyal, Priya ;
Girshick, Ross ;
He, Kaiming ;
Dollar, Piotr .
2017 IEEE INTERNATIONAL CONFERENCE ON COMPUTER VISION (ICCV), 2017, :2999-3007