Large Margin Nearest Neighbor Embedding for Knowledge Representation

被引:1
作者
Fan, Miao [1 ,2 ]
Zhou, Qiang [1 ]
Zheng, Thomas Fang [1 ]
Grishman, Ralph [2 ]
机构
[1] Tsinghua Univ, Tsinghua Natl Lab Informat Sci & Technol, Div Tech Innovat & Dev, CSLT, Beijing 100084, Peoples R China
[2] NYU, Proteus Grp, New York, NY 10003 USA
来源
2015 IEEE/WIC/ACM INTERNATIONAL CONFERENCE ON WEB INTELLIGENCE AND INTELLIGENT AGENT TECHNOLOGY (WI-IAT), VOL 1 | 2015年
基金
美国国家科学基金会;
关键词
D O I
10.1109/WI-IAT.2015.125
中图分类号
TP301 [理论、方法];
学科分类号
081202 ;
摘要
Traditional way of storing facts in triplets (head entity, relation, tail_entity), abbreviated as (h, r, t), allows the knowledge to be intuitively displayed and easily acquired by human beings, but hardly computed or even reasoned about by AI machines. Inspired by the success in applying Distributed Representations to AI-related fields, recent studies expect to represent each entity and relation with a unique low dimensional embedding, which is different from the symbolic and atomic framework of displaying knowledge in triplets. In this way, the knowledge computing and reasoning can be essentially facilitated by means of a simple vector calculation, i.e. h r approximate to t. We thus contribute an effective model to learn better embeddings satisfying the formula by pulling the positive tail entities t(+) together and close to h + r (Nearest Neighbor), and simultaneously pushing the negatives t(-) away from the positives t(+) via keeping a Large Margin. We also design a corresponding learning algorithm to efficiently find the optimal solution based on Stochastic Gradient Descent in iterative fashion. Quantitative experiments illustrate that our approach can achieve the state-of-the-art performance, compared with several recent methods on some benchmark datasets for two classical applications, i.e. Link prediction and Triplet classification. Moreover, we analyze the parameter complexities among all the evaluated models, and analytical results indicate that our model needs fewer computational resources while outperforming the other methods.
引用
收藏
页码:53 / 59
页数:7
相关论文
共 20 条
  • [1] Hate Speech Detection with Comment Embeddings
    Djuric, Nemanja
    Zhou, Jing
    Morris, Robin
    Grbovic, Mihajlo
    Radosavljevic, Vladan
    Bhamidipati, Narayan
    [J]. WWW'15 COMPANION: PROCEEDINGS OF THE 24TH INTERNATIONAL CONFERENCE ON WORLD WIDE WEB, 2015, : 29 - 30
  • [2] [Anonymous], 2011, P 25 AAAI C ARTIFICI
  • [3] [Anonymous], 2013, P 2013 C N AM CHAPTE
  • [4] [Anonymous], 2007, AAAI
  • [5] [Anonymous], 2012, NIPS
  • [6] [Anonymous], 2013, P NIPS
  • [7] [Anonymous], 2013, P 26 INT C NEUR INF
  • [8] [Anonymous], 2009, P 22 INT C NEURAL IN
  • [9] [Anonymous], 2007, WWW
  • [10] A neural probabilistic language model
    Bengio, Y
    Ducharme, R
    Vincent, P
    Jauvin, C
    [J]. JOURNAL OF MACHINE LEARNING RESEARCH, 2003, 3 (06) : 1137 - 1155