A distant supervision method based on paradigmatic relations for learning word embeddings

被引:3
|
作者
Li, Jianquan [1 ]
Hu, Renfen [2 ]
Liu, Xiaokang [1 ]
Tiwari, Prayag [4 ]
Pandey, Hari Mohan [3 ]
Chen, Wei [1 ]
Wang, Benyou [4 ]
Jin, Yaohong [1 ]
Yang, Kaicheng [1 ]
机构
[1] Beijing Ultrapower Software Co Ltd, Beijing, Peoples R China
[2] Beijing Normal Univ, Beijing, Peoples R China
[3] Edge Hill Univ, Dept Comp Sci, Ormskirk, England
[4] Univ Padua, Dept Informat Engn, Padua, Italy
来源
NEURAL COMPUTING & APPLICATIONS | 2020年 / 32卷 / 12期
基金
欧盟地平线“2020”;
关键词
Neural network; Word embedding; Text classification; Sentence matching;
D O I
10.1007/s00521-019-04071-6
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
Word embeddings learned on external resources have succeeded in improving many NLP tasks. However, existing embedding models still face challenges in situations where fine-gained semantic information is required, e.g., distinguishing antonyms from synonyms. In this paper, a distant supervision method is proposed to guide the training process by introducing semantic knowledge in a thesaurus. Specifically, the proposed model shortens the distance between target word and its synonyms by controlling the movements of them in both unidirectional and bidirectional, yielding three different models, namelyUnidirectional Movement of Target Model(UMT),Unidirectional Movement of Synonyms Model(UMS) andBidirectional Movement of Target and Synonyms Model(BMTS). Extensive computational experiments have been conducted, and results are collected for analysis purpose. The results show that the proposed models not only efficiently capture semantic information of antonyms but also achieve significant improvements in both intrinsic and extrinsic evaluation tasks. To validate the performance of the proposed models (UMT, UMS and BMTS), results are compared against well-known models, namelySkip-gram,JointRCM,WE-TDanddict2vec. The performances of the proposed models are evaluated on four tasks (benchmarks):word analogy(intrinsic),synonym-antonym detection(intrinsic),sentence matching(extrinsic) andtext classification(extrinsic). A case study is provided to illustrate the working of the proposed models in an effective manner. Overall, a distant supervision method based on paradigmatic relations is proposed for learning word embeddings and it outperformed when compared against other existing models.
引用
收藏
页码:7759 / 7768
页数:10
相关论文
共 50 条
  • [1] A distant supervision method based on paradigmatic relations for learning word embeddings
    Jianquan Li
    Renfen Hu
    Xiaokang Liu
    Prayag Tiwari
    Hari Mohan Pandey
    Wei Chen
    Benyou Wang
    Yaohong Jin
    Kaicheng Yang
    Neural Computing and Applications, 2020, 32 : 7759 - 7768
  • [2] Distant Supervision Learning of DBPedia Relations
    Zajac, Marcin
    Przepiorkowski, Adam
    TEXT, SPEECH, AND DIALOGUE, TSD 2013, 2013, 8082 : 193 - 200
  • [3] Learning Word Embeddings from Dependency Relations
    Zhao, Yinggong
    Huang, Shujian
    Dai, Xinyu
    Zhang, Jianbing
    Chen, Jiajun
    PROCEEDINGS OF THE 2014 INTERNATIONAL CONFERENCE ON ASIAN LANGUAGE PROCESSING (IALP 2014), 2014, : 123 - 127
  • [4] Learning Word Representations by Jointly Modeling Syntagmatic and Paradigmatic Relations
    Sun, Fei
    Guo, Jiafeng
    Lan, Yanyan
    Xu, Jun
    Cheng, Xueqi
    PROCEEDINGS OF THE 53RD ANNUAL MEETING OF THE ASSOCIATION FOR COMPUTATIONAL LINGUISTICS AND THE 7TH INTERNATIONAL JOINT CONFERENCE ON NATURAL LANGUAGE PROCESSING, VOL 1, 2015, : 136 - 145
  • [5] Evaluating Word Embeddings based on Hypernymy Relations
    Han, Xiao
    Wang, Qi
    PROCEEDINGS OF 2019 IEEE 3RD INFORMATION TECHNOLOGY, NETWORKING, ELECTRONIC AND AUTOMATION CONTROL CONFERENCE (ITNEC 2019), 2019, : 954 - 958
  • [6] A survey of word embeddings based on deep learning
    Shirui Wang
    Wenan Zhou
    Chao Jiang
    Computing, 2020, 102 : 717 - 740
  • [7] A survey of word embeddings based on deep learning
    Wang, Shirui
    Zhou, Wenan
    Jiang, Chao
    COMPUTING, 2020, 102 (03) : 717 - 740
  • [8] Derivational Morphological Relations in Word Embeddings
    Musil, Tomas
    Vidra, Jonas
    Marecek, David
    BLACKBOXNLP WORKSHOP ON ANALYZING AND INTERPRETING NEURAL NETWORKS FOR NLP AT ACL 2019, 2019, : 173 - 180
  • [9] Adaptive entity extraction method based on distant supervision
    Ge L.
    Zhang Y.
    Li W.
    Harbin Gongcheng Daxue Xuebao/Journal of Harbin Engineering University, 2022, 43 (04): : 564 - 571
  • [10] Learning Phrase Representations Based on Word and Character Embeddings
    Huang, Jiangping
    Ji, Donghong
    Yao, Shuxin
    Huang, Wenzhi
    Chen, Bo
    NEURAL INFORMATION PROCESSING, ICONIP 2016, PT IV, 2016, 9950 : 547 - 554