A survey of word embeddings based on deep learning

被引:112
|
作者
Wang, Shirui [1 ]
Zhou, Wenan [1 ]
Jiang, Chao [1 ]
机构
[1] Beijing Univ Posts & Telecommun, Dept Comp Sci, Beijing 100876, Peoples R China
关键词
Word embeddings; Neural networks; Distributed hypothesis; Multi-source data; PERFORMANCE;
D O I
10.1007/s00607-019-00768-7
中图分类号
TP301 [理论、方法];
学科分类号
081202 ;
摘要
The representational basis for downstream natural language processing tasks is word embeddings, which capture lexical semantics in numerical form to handle the abstract semantic concept of words. Recently, the word embeddings approaches, represented by deep learning, has attracted extensive attention and widely used in many tasks, such as text classification, knowledge mining, question-answering, smart Internet of Things systems and so on. These neural networks-based models are based on the distributed hypothesis while the semantic association between words can be efficiently calculated in low-dimensional space. However, the expressed semantics of most models are constrained by the context distribution of each word in the corpus while the logic and common knowledge are not better utilized. Therefore, how to use the massive multi-source data to better represent natural language and world knowledge still need to be explored. In this paper, we introduce the recent advances of neural networks-based word embeddings with their technical features, summarizing the key challenges and existing solutions, and further give a future outlook on the research and application.
引用
收藏
页码:717 / 740
页数:24
相关论文
共 50 条
  • [11] Using Deep Learning Word Embeddings for Citations Similarity in Academic Papers
    Hourrane, Oumaima
    Mifrah, Sara
    Benlahmar, El Habib
    Bouhriz, Nadia
    Rachdi, Mohamed
    BIG DATA, CLOUD AND APPLICATIONS, BDCA 2018, 2018, 872 : 185 - 196
  • [12] Learning Phrase Representations Based on Word and Character Embeddings
    Huang, Jiangping
    Ji, Donghong
    Yao, Shuxin
    Huang, Wenzhi
    Chen, Bo
    NEURAL INFORMATION PROCESSING, ICONIP 2016, PT IV, 2016, 9950 : 547 - 554
  • [13] Using deep learning and word embeddings for predicting human agreeableness behavior
    Alsini, Raed
    Naz, Anam
    Khan, Hikmat Ullah
    Bukhari, Amal
    Daud, Ali
    Ramzan, Muhammad
    SCIENTIFIC REPORTS, 2024, 14 (01):
  • [14] Arabic Quran Verses Authentication Using Deep Learning and Word Embeddings
    Touati-Hamad, Zineb
    Laouar, Mohamed Ridda
    Bendib, Issam
    Hakak, Saqib
    INTERNATIONAL ARAB JOURNAL OF INFORMATION TECHNOLOGY, 2022, 19 (04) : 681 - 688
  • [15] Deep Fake Recognition in Tweets Using Text Augmentation, Word Embeddings and Deep Learning
    Tesfagergish, Senait G.
    Damasevicius, Robertas
    Kapociute-Dzikiene, Jurgita
    COMPUTATIONAL SCIENCE AND ITS APPLICATIONS, ICCSA 2021, PT VI, 2021, 12954 : 523 - 538
  • [16] Complementary Learning of Word Embeddings
    Song, Yan
    Shi, Shuming
    PROCEEDINGS OF THE TWENTY-SEVENTH INTERNATIONAL JOINT CONFERENCE ON ARTIFICIAL INTELLIGENCE, 2018, : 4368 - 4374
  • [17] A survey on training and evaluation of word embeddings
    François Torregrossa
    Robin Allesiardo
    Vincent Claveau
    Nihel Kooli
    Guillaume Gravier
    International Journal of Data Science and Analytics, 2021, 11 : 85 - 103
  • [18] A survey on training and evaluation of word embeddings
    Torregrossa, Francois
    Allesiardo, Robin
    Claveau, Vincent
    Kooli, Nihel
    Gravier, Guillaume
    INTERNATIONAL JOURNAL OF DATA SCIENCE AND ANALYTICS, 2021, 11 (02) : 85 - 103
  • [19] Deep-STP: a deep learning-based approach to predict snake toxin proteins by using word embeddings
    Zulfiqar, Hasan
    Guo, Zhiling
    Ahmad, Ramala Masood
    Ahmed, Zahoor
    Cai, Peiling
    Chen, Xiang
    Zhang, Yang
    Lin, Hao
    Shi, Zheng
    FRONTIERS IN MEDICINE, 2024, 10
  • [20] eXplainable AI for Word Embeddings: A Survey
    Boselli, Roberto
    D'Amico, Simone
    Nobani, Navid
    COGNITIVE COMPUTATION, 2025, 17 (01)