NAS-CRE: Neural Architecture Search for Context-Based Relation Extraction

被引:0
作者
Yan, Rongen [1 ,2 ]
Li, Dongmei [1 ,2 ]
Wu, Yan [3 ,4 ]
Dang, Depeng [5 ]
Tao, Ye [6 ]
Wang, Shaofei [7 ]
机构
[1] Beijing Forestry Univ, Sch Informat Sci & Technol, Beijing 100083, Peoples R China
[2] Natl Forestry & Grassland Adm, Engn Res Ctr Forestry Oriented Intelligent Informa, Beijing 100083, Peoples R China
[3] CSSC Syst Engn Res Inst, Beijing 100094, Peoples R China
[4] CSSC Intelligent Innovat Res Inst, Beijing 100094, Peoples R China
[5] Beijing Normal Univ, Sch Artificial Intelligence, Beijing 100875, Peoples R China
[6] Zhongguancun Lab, Beijing 100094, Peoples R China
[7] Capital Normal Univ, Coll Informat Engn, Beijing 100048, Peoples R China
来源
APPLIED SCIENCES-BASEL | 2024年 / 14卷 / 23期
基金
中央高校基本科研业务费专项资金资助;
关键词
natural language processing; relation extraction; neural architecture search; INFORMATION EXTRACTION;
D O I
10.3390/app142310960
中图分类号
O6 [化学];
学科分类号
0703 ;
摘要
Relation extraction, a crucial task in natural language processing (NLP) for constructing knowledge graphs, entails extracting relational semantics between pairs of entities within a sentence. Given the intricacy of language, a single sentence often encompasses multiple entities that mutually influence one another. Recently, various iterations of recurrent neural networks (RNNs) have been introduced into relation extraction tasks, where the efficacy of neural network structures directly influences task performance. However, many neural networks necessitate manual determination of optimal parameters and network architectures, resulting in limited generalization capabilities for specific tasks. In this paper, we formally define the context-based relation extraction problem and propose a solution utilizing neural architecture search (NAS) to optimize RNN. Specifically, NAS employs an RNN controller to delineate an RNN cell, yielding an optimal structure to represent all relationships, thereby aiding in extracting relationships between target entities. Additionally, to enhance relation extraction performance, we leverage the XLNet pretrained model to comprehensively capture the semantic features of the sentence. Extensive experiments conducted on a real-world dataset containing words with multiple relationships demonstrate that our proposed method significantly enhances micro-F1 scores compared to state-of-the-art baselines.
引用
收藏
页数:16
相关论文
共 45 条
  • [1] Wang Y., Lipka N., Rossi R.A., Siu A., Zhang R., Derr T., Knowledge graph prompting for multi-document question answering, Proceedings of the AAAI Conference on Artificial Intelligence, 38, pp. 19206-19214
  • [2] Tong X., Yu L., Deacon S.H., A Meta-Analysis of the Relation Between Syntactic Skills and Reading Comprehension: A Cross-Linguistic and Developmental Investigation, Rev. Educ. Res, 2024, (2024)
  • [3] Gui S., Shao C., Ma Z., Chen Y., Feng Y., Non-autoregressive Machine Translation with Probabilistic Context-free Grammar, Adv. Neural Inf. Process. Syst, 36, pp. 5598-5615, (2024)
  • [4] Chen H., Wang Y., Guo J., Tao D., Vanillanet: The power of minimalism in deep learning, Adv. Neural Inf. Process. Syst, 36, pp. 7050-7064, (2024)
  • [5] Li X., Li Y., Yang J., Liu H., Hu P., A relation aware embedding mechanism for relation extraction, Appl. Intell, 52, pp. 10022-10031, (2022)
  • [6] Zhang M., Qian T., Liu B., Exploit Feature and Relation Hierarchy for Relation Extraction, IEEE/ACM Trans. Audio Speech Lang. Process, 30, pp. 917-930, (2022)
  • [7] Luo F.M., Xu T., Lai H., Chen X.H., Zhang W., Yu Y., A survey on model-based reinforcement learning, Sci. China Inf. Sci, 67, (2024)
  • [8] Li Z., Hu Z., Luo W., Hu X., SaberNet: Self-attention based effective relation network for few-shot learning, Pattern Recognit, 133, (2023)
  • [9] Sui D., Zeng X., Chen Y., Liu K., Zhao J., Joint entity and relation extraction with set prediction networks, IEEE Trans. Neural Netw. Learn. Syst, 35, pp. 12784-12795, (2023)
  • [10] Sorokin D., Gurevych I., Context-aware representations for knowledge base relation extraction, Proceedings of the 2017 Conference on Empirical Methods in Natural Language Processing, pp. 1784-1789