Unifying Structure Reasoning and Language Pre-Training for Complex Reasoning Tasks
被引:2
作者:
Wang, Siyuan
论文数: 0引用数: 0
h-index: 0
机构:
Fudan Univ, Sch Data Sci, Shanghai 200433, Peoples R ChinaFudan Univ, Sch Data Sci, Shanghai 200433, Peoples R China
Wang, Siyuan
[1
]
Wei, Zhongyu
论文数: 0引用数: 0
h-index: 0
机构:
Fudan Univ, Sch Data Sci, Shanghai 200433, Peoples R China
Fudan Univ, Res Inst Intelligent & Complex Syst, Shanghai 200433, Peoples R ChinaFudan Univ, Sch Data Sci, Shanghai 200433, Peoples R China
Wei, Zhongyu
[1
,2
]
Xu, Jiarong
论文数: 0引用数: 0
h-index: 0
机构:
Fudan Univ, Sch Management, Shanghai 200433, Peoples R ChinaFudan Univ, Sch Data Sci, Shanghai 200433, Peoples R China
Xu, Jiarong
[3
]
Li, Taishan
论文数: 0引用数: 0
h-index: 0
机构:
ShanghaiTech Univ, Sch Informat Sci & Technol, Shanghai 201210, Peoples R ChinaFudan Univ, Sch Data Sci, Shanghai 200433, Peoples R China
Li, Taishan
[4
]
Fan, Zhihao
论文数: 0引用数: 0
h-index: 0
机构:
Fudan Univ, Sch Data Sci, Shanghai 200433, Peoples R ChinaFudan Univ, Sch Data Sci, Shanghai 200433, Peoples R China
Fan, Zhihao
[1
]
机构:
[1] Fudan Univ, Sch Data Sci, Shanghai 200433, Peoples R China
[2] Fudan Univ, Res Inst Intelligent & Complex Syst, Shanghai 200433, Peoples R China
[3] Fudan Univ, Sch Management, Shanghai 200433, Peoples R China
[4] ShanghaiTech Univ, Sch Informat Sci & Technol, Shanghai 201210, Peoples R China
Cognition;
Task analysis;
Semantics;
Films;
Speech processing;
Context modeling;
Data models;
Structure reasoning skill;
language model pre-training;
complex reasoning;
D O I:
10.1109/TASLP.2023.3325973
中图分类号:
O42 [声学];
学科分类号:
070206 ;
082403 ;
摘要:
Recent pre-trained language models (PLMs) equipped with foundation reasoning skills have shown remarkable performance on downstream complex tasks. However, the significant structure reasoning skill has been rarely studied, which involves modeling implicit structure information within the text and performing explicit logical reasoning over them to deduce the conclusion. This paper proposes a unified learning framework that combines explicit structure reasoning and language pre-training to endow PLMs with the structure reasoning skill. It first identifies several elementary structures within contexts to construct structured queries and performs step-by-step reasoning along the queries to identify the answer entity. The fusion of textual semantics and structure reasoning is achieved by using contextual representations learned by PLMs to initialize the representation space of structures, and performing stepwise reasoning on this semantic representation space. Experimental results on four datasets demonstrate that the proposed model achieves significant improvements in complex reasoning tasks involving diverse structures, and shows transferability to downstream tasks with limited training data and effectiveness for complex reasoning of KGs modality.