GetPt: Graph-enhanced General Table Pre-training with Alternate Attention Network

被引:1
|
作者
Jia, Ran [1 ]
Guo, Haoming [2 ,6 ]
Jin, Xiaoyuan [3 ,6 ]
Yan, Chao [4 ,6 ]
Du, Lun [1 ]
Ma, Xiaojun [1 ]
Stankovic, Tamara [5 ]
Lozajic, Marko [5 ]
Zoranovic, Goran [5 ]
Ilic, Igor [5 ]
Han, Shi [1 ]
Zhang, Dongmei [1 ]
机构
[1] Microsoft, Beijing, Peoples R China
[2] Univ Calif Berkeley, Berkeley, CA USA
[3] Swiss Fed Inst Technol, Zurich, Switzerland
[4] Peking Univ, Beijing, Peoples R China
[5] Microsoft, Belgrade, Serbia
[6] Microsoft Res Asia, Beijing, Peoples R China
来源
PROCEEDINGS OF THE 29TH ACM SIGKDD CONFERENCE ON KNOWLEDGE DISCOVERY AND DATA MINING, KDD 2023 | 2023年
关键词
table pre-training; graph transformer; table understanding;
D O I
10.1145/3580305.3599366
中图分类号
TP [自动化技术、计算机技术];
学科分类号
0812 ;
摘要
Tables are widely used for data storage and presentation due to their high flexibility in layout. The importance of tables as information carriers and the complexity of tabular data understanding attract a great deal of research on large-scale pre-training for tabular data. However, most of the works design models for specific types of tables, such as relational tables and tables withwell-structured headers, neglecting tables with complex layouts. In real-world scenarios, there are many such tables beyond the target scope of previous research and are thus not well supported. In this paper, we propose GetPt, a unified pre-training architecture for general table representation applicable even to tables with complex structures and layouts. First, we convert a table to a heterogeneous graph to represent the layout of the table. Based on the graph, a specially designed transformer is applied to jointly model the semantics and structure of the table. Second, we devise an Alternate Attention Network (AAN) to better model the contextual information across multiple granularities of a table including the tokens, cells, and table. To better support a wide range of downstream tasks, we further employ three pre-training objectives and pre-train the model on a large table dataset. We fine-tune and evaluate GetPt model on two representative tasks, table type classification, and table structure recognition. Experiments show that GetPt outperforms existing state-of-the-art methods on these tasks.
引用
收藏
页码:941 / 950
页数:10
相关论文
共 3 条
  • [1] Schema dependency-enhanced curriculum pre-training for table semantic parsing
    Qin, Bowen
    Hui, Binyuan
    Wang, Lihan
    Yang, Min
    Li, Binhua
    Huang, Fei
    Si, Luo
    Jiang, Qingshan
    Li, Yongbin
    KNOWLEDGE-BASED SYSTEMS, 2023, 262
  • [2] Pre-training Graph Transformer with Multimodal Side Information for Recommendation
    Liu, Yong
    Yang, Susen
    Lei, Chenyi
    Wang, Guoxin
    Tang, Haihong
    Zhang, Juyong
    Sun, Aixin
    Miao, Chunyan
    PROCEEDINGS OF THE 29TH ACM INTERNATIONAL CONFERENCE ON MULTIMEDIA, MM 2021, 2021, : 2853 - 2861
  • [3] KPGT: Knowledge-Guided Pre-training of Graph Transformer for Molecular Property Prediction
    Li, Han
    Zhao, Dan
    Zeng, Jianyang
    PROCEEDINGS OF THE 28TH ACM SIGKDD CONFERENCE ON KNOWLEDGE DISCOVERY AND DATA MINING, KDD 2022, 2022, : 857 - 867