Word Embedding with Neural Probabilistic Prior

被引:0
作者
Ren, Shaogang [1 ]
Li, Dingcheng [1 ]
Li, Ping [1 ]
机构
[1] Baidu Res, Cognit Comp Lab, 10900 NE 8th St, Bellevue, WA 98004 USA
来源
PROCEEDINGS OF THE 2024 SIAM INTERNATIONAL CONFERENCE ON DATA MINING, SDM | 2024年
关键词
NONLINEAR ICA;
D O I
暂无
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
To improve word representation learning, we propose a probabilistic prior which can be seamlessly integrated with word embedding models. Different from previous methods, word embedding is taken as a probabilistic generative model, and it enables us to impose a prior regularizing word representation learning. The proposed prior not only enhances the representation of embedding vectors but also improves the model's robustness and stability. The structure of the proposed prior is simple and effective, and it can be easily implemented and flexibly plugged in most existing word embedding models. Extensive experiments show the proposed method improves word representation on various tasks.
引用
收藏
页码:896 / 904
页数:9
相关论文
共 43 条
  • [21] Lee Kenton, 2018, P 2018 C THE N AM, V2, P687
  • [22] Levy O, 2014, ADV NEUR IN, V27
  • [23] Dependency-Based Word Embeddings
    Levy, Omer
    Goldberg, Yoav
    [J]. PROCEEDINGS OF THE 52ND ANNUAL MEETING OF THE ASSOCIATION FOR COMPUTATIONAL LINGUISTICS, VOL 2, 2014, : 302 - 308
  • [24] Liao KT, 2020, 1ST CONFERENCE OF THE ASIA-PACIFIC CHAPTER OF THE ASSOCIATION FOR COMPUTATIONAL LINGUISTICS AND THE 10TH INTERNATIONAL JOINT CONFERENCE ON NATURAL LANGUAGE PROCESSING (AACL-IJCNLP 2020), P720
  • [25] Luong T., 2013, P CONLL
  • [26] Manaal, 2015, P NA ACL HUM LANG TE, P1606, DOI DOI 10.3115/V1/N15-1184
  • [27] Marcheggiani D., 2017, P 2017 C EMP METH NA, P1506, DOI DOI 10.18653/V1/D17-1159
  • [28] Marcus M., 1994, HUMAN LANGUAGE TECHN
  • [29] Mikolov T., 2013, P 2013 C N AM CHAPT, P746
  • [30] Mikolov T., 2013, Advances in neural information processing systems, DOI DOI 10.5555/2999792.2999959