Quantum-inspired semantic matching based on neural networks with the duality of density matrices

被引:0
作者
Zhang, Chenchen [1 ]
Li, Qiuchi [3 ]
Song, Dawei [1 ,2 ]
Tiwari, Prayag [4 ]
机构
[1] Beijing Inst Technol, Sch Comp Sci & Technol, Beijing, Peoples R China
[2] Open Univ, Sch Comp & Commun, Milton Keynes, England
[3] Univ Copenhagen, Dept Comp Sci, Copenhagen, Denmark
[4] Halmstad Univ, Sch Informat Technol, Halmstad, Sweden
关键词
Neural network; Complex-valued neural network; State-probability duality; Density matrix; Quantum theory;
D O I
10.1016/j.engappai.2024.109667
中图分类号
TP [自动化技术、计算机技术];
学科分类号
0812 ;
摘要
Social media text can be semantically matched in different ways, viz paraphrase identification, answer selection, community question answering, and soon. The performance of the above semantic matching tasks depends largely on the ability of language modeling. Neural network based language models and probabilistic language models are two main streams of language modeling approaches. However, few prior work has managed to unify them in a single framework on the premise of preserving probabilistic features during the neural network learning process. Motivated by recent advances of quantum-inspired neural networks for text representation learning, we fill the gap by resorting to density matrices, a key concept describing a quantum state as well as a quantum probability distribution. The state and probability views of density matrices are mapped respectively to the neural and probabilistic aspects of language models. Concretizing this state-probability duality to the semantic matching task, we build a unified neural-probabilistic language model through a quantum-inspired neural network. Specifically, we take the state view to construct a density matrix representation of sentence, and exploit its probabilistic nature by extracting its main semantics, which form the basis of a legitimate quantum measurement. When matching two sentences, each sentence is measured against the main semantics of the other. Such a process is implemented in a neural structure, facilitating an end-to-end learning of parameters. The learned density matrix representation reflects an authentic probability distribution over the semantic space throughout the training process. Experiments show that our model significantly outperforms a wide range of prominent classical and quantum-inspired baselines.
引用
收藏
页数:15
相关论文
共 64 条
  • [1] Ai Q., 2016, P 2016 ACM INT C THE, P133, DOI DOI 10.1145/2970398.2970409
  • [2] Embedding-based Query Expansion for Weighted Sequential Dependence Retrieval Model
    Balaneshin-kordan, Saeid
    Kotov, Alexander
    [J]. SIGIR'17: PROCEEDINGS OF THE 40TH INTERNATIONAL ACM SIGIR CONFERENCE ON RESEARCH AND DEVELOPMENT IN INFORMATION RETRIEVAL, 2017, : 1213 - 1216
  • [3] Busemeyer J.R., 2012, Quantum Models of Cognition and Decision, DOI [10.1017/CBO9780511997716, DOI 10.1017/CBO9780511997716]
  • [4] Quantum Language Model With Entanglement Embedding for Question Answering
    Chen, Yiwei
    Pan, Yu
    Dong, Daoyi
    [J]. IEEE TRANSACTIONS ON CYBERNETICS, 2023, 53 (06) : 3467 - 3478
  • [5] ELECTRA: PRE-TRAINING TEXT ENCODERS AS DISCRIMINATORS RATHER THAN GENERATORS
    Clark, Kevin
    Luong, Minh-Thang
    Le, Quoc V.
    Manning, Christopher D.
    [J]. INFORMATION SYSTEMS RESEARCH, 2020,
  • [6] Conneau A., 2017, P 2017 C EMP METH NA, P670, DOI [DOI 10.18653/V1/D17-1070, 10.18653/v1/D17-1070]
  • [7] Convolutional Neural Networks for Soft-Matching N-Grams in Ad-hoc Search
    Dai, Zhuyun
    Xiong, Chenyan
    Callan, Jamie
    Liu, Zhiyuan
    [J]. WSDM'18: PROCEEDINGS OF THE ELEVENTH ACM INTERNATIONAL CONFERENCE ON WEB SEARCH AND DATA MINING, 2018, : 126 - 134
  • [8] Devlin J, 2019, 2019 CONFERENCE OF THE NORTH AMERICAN CHAPTER OF THE ASSOCIATION FOR COMPUTATIONAL LINGUISTICS: HUMAN LANGUAGE TECHNOLOGIES (NAACL HLT 2019), VOL. 1, P4171
  • [9] Dolan William B., 2005, P 3 INT WORKSHOP PAR
  • [10] dos Santos C, 2016, Arxiv, DOI arXiv:1602.03609