Judicial Text Relation Extraction Based on Prompt Tuning

被引:0
作者
Chen, Xue [1 ]
Li, Yi [1 ]
Fan, Shuhuan [1 ]
Hou, Mengshu [1 ]
机构
[1] Univ Elect Sci & Technol China, Chengdu, Peoples R China
来源
2024 2ND ASIA CONFERENCE ON COMPUTER VISION, IMAGE PROCESSING AND PATTERN RECOGNITION, CVIPPR 2024 | 2024年
关键词
relation extraction; pre-training model; prompt-tuning; judicial text processing;
D O I
10.1145/3663976.3664029
中图分类号
TM [电工技术]; TN [电子技术、通信技术];
学科分类号
0808 ; 0809 ;
摘要
With the acceleration of the digital transformation in the judicial field, the need for automated processing and analysis of voluminous legal documents becomes increasingly imperative. Relationship extraction, a pivotal step in comprehending these documents, is crucial for automated case analysis, discovery of legal facts, and prediction of judgments. However, the idiosyncrasies of judicial texts, such as their intricate linguistic structures, specialized terminology, and implicit semantic relationships, present significant challenges to the efficacy of traditional text analysis methods. Traditional relation extraction techniques often fail to adapt well to professional terms, intricate linguistic structures and implicit semantics in judicial texts. Moreover, the general lack of training data tailored to the specific context of the legal domain hampers the ability of models to learn effectively and accurately extract relationships from judicial texts. This paper introduces a novel approach utilizing pre-trained models for prompt-based fine-tuning to achieve relationship extraction in judicial texts. Leveraging the robust language comprehension capabilities of large-scale pre-trained models, we have designed domain-specific prompt templates to guide the model in more effectively capturing and understanding key information within legal texts. This method not only enhances the accuracy of relationship extraction but also significantly reduces reliance on extensive annotated data. The effectiveness and superiority of our approach are corroborated by test results on a judicial text dataset.
引用
收藏
页数:6
相关论文
共 17 条
[1]  
Alt C, 2020, Arxiv, DOI arXiv:2004.14855
[2]  
.cailabs, ABOUT US
[3]  
Chalkidis I, 2020, Arxiv, DOI [arXiv:2010.02559, DOI 10.48550/ARXIV.2010.02559]
[4]  
Chen X, 2023, Arxiv, DOI arXiv:2104.07650
[5]  
Chen Yanguang, 2020, P 28 INT C COMP LING, P1561, DOI DOI 10.18653/V1/2020.COLING-MAIN.137
[6]  
Han X, 2019, Arxiv, DOI [arXiv:1909.13078, DOI 10.48550/ARXIV.1909.13078]
[7]  
Hendrickx I., 2019, arXiv
[8]   SpanBERT: Improving Pre-training by Representing and Predicting Spans [J].
Joshi, Mandar ;
Chen, Danqi ;
Liu, Yinhan ;
Weld, Daniel S. ;
Zettlemoyer, Luke ;
Levy, Omer .
TRANSACTIONS OF THE ASSOCIATION FOR COMPUTATIONAL LINGUISTICS, 2020, 8 :64-77
[9]  
Liu YH, 2019, Arxiv, DOI [arXiv:1907.11692, DOI 10.48550/ARXIV.1907.11692]
[10]  
Shen Y., 2016, P COLING 2016 26 INT, P2526