Building siamese attention-augmented recurrent convolutional neural networks for document similarity scoring

被引:9
|
作者
Han, Sifei [1 ]
Shi, Lingyun [1 ]
Richie, Russell [1 ]
Tsui, Fuchiang R. Rich [1 ,2 ]
机构
[1] Childrens Hosp Philadelphia, Dept Biomed & Hlth Informat, Tsui Lab, 2716 South St, Philadelphia, PA 19104 USA
[2] Univ Penn, Perelman Sch Med, 3400 Spruce St,Suite 680 Dulles, Philadelphia, PA USA
基金
美国国家科学基金会;
关键词
Attention neural network; Deep learning; Machine learning; Natural language processing; Information retrieval; Text similarity;
D O I
10.1016/j.ins.2022.10.032
中图分类号
TP [自动化技术、计算机技术];
学科分类号
0812 ;
摘要
Automatically measuring document similarity is imperative in natural language process-ing, with applications ranging from recommendation to duplicate document detection. State-of-the-art approach in document similarity commonly involves deep neural net-works, yet there is little study on how different architectures may be combined. Thus, we introduce the Siamese Attention-augmented Recurrent Convolutional Neural Network (S-ARCNN) that combines multiple neural network architectures. In each subnet-work of S-ARCNN, a document passes through a bidirectional Long Short-Term Memory (bi-LSTM) layer, which sends representations to local and global document modules. A local document module uses convolution, pooling, and attention layers, whereas a global document module uses last states of the bi-LSTM. Both local and global features are con-catenated to form a single document representation. Using the Quora Question Pairs data -set, we evaluated S-ARCNN, Siamese convolutional neural networks (S-CNNs), Siamese LSTM, and two BERT models. While S-CNNs (82.02% F1) outperformed S-ARCNN (79.83% F1) overall, S-ARCNN slightly outperformed S-CNN on duplicate question pairs with more than 50 words (39.96% vs. 39.42% accuracy). With the potential advantage of S-ARCNN for processing longer documents, S-ARCNN may help researchers identify collaborators with similar research interests, help editors find potential reviewers, or match resumes with job descriptions.(c) 2022 Published by Elsevier Inc.
引用
收藏
页码:90 / 102
页数:13
相关论文
共 50 条
  • [41] Pruning convolutional neural networks via filter similarity analysis
    Geng, Lili
    Niu, Baoning
    MACHINE LEARNING, 2022, 111 (09) : 3161 - 3180
  • [42] Sentence Semantic Similarity Model Using Convolutional Neural Networks
    Karthiga M.
    Sountharrajan S.
    Suganya E.
    Sankarananth S.
    EAI Endorsed Transactions on Energy Web, 2021, 8 (35) : 1 - 6
  • [43] Learning visual similarity for product design with convolutional neural networks
    Bell, Sean
    Bala, Kavita
    ACM TRANSACTIONS ON GRAPHICS, 2015, 34 (04):
  • [44] Combining Very Deep Convolutional Neural Networks and Recurrent Neural Networks for Video Classification
    Kiziltepe, Rukiye Savran
    Gan, John Q.
    Escobar, Juan Jose
    ADVANCES IN COMPUTATIONAL INTELLIGENCE, IWANN 2019, PT II, 2019, 11507 : 811 - 822
  • [45] Convolutional and Recurrent Neural Networks for Face Image Analysis
    Yuksel, Kivanc
    Skarbek, Wladyslaw
    FOUNDATIONS OF COMPUTING AND DECISION SCIENCES, 2019, 44 (03) : 331 - 347
  • [46] Image Captioning using Convolutional Neural Networks and Recurrent Neural Network
    Calvin, Rachel
    Suresh, Shravya
    2021 6TH INTERNATIONAL CONFERENCE FOR CONVERGENCE IN TECHNOLOGY (I2CT), 2021,
  • [47] Attention Augmented Convolutional Neural Network for acoustics based machine state estimation
    Tan, Jiannan
    Oyekan, John
    APPLIED SOFT COMPUTING, 2021, 110
  • [48] Embedding Graph Convolutional Networks in Recurrent Neural Networks for Predictive Monitoring
    Rama-Maneiro, Efren
    Vidal, Juan C.
    Lama, Manuel
    IEEE TRANSACTIONS ON KNOWLEDGE AND DATA ENGINEERING, 2024, 36 (01) : 137 - 151
  • [49] A Comparative Review of Convolutional Neural Networks, Long Short-Term Memory, and Recurrent Neural Networks in Recommendation Systems
    Tyagi, Geetanjali
    Ray, Susmita
    ARTIFICIAL INTELLIGENCE: THEORY AND APPLICATIONS, VOL 1, AITA 2023, 2024, 843 : 395 - 408
  • [50] Attention Based Recurrent Neural Networks for Online Advertising
    Zhai, Shuangfei
    Chang, Keng-Hao
    Zhang, Ruofei
    Zhang, Zhongfei
    PROCEEDINGS OF THE 25TH INTERNATIONAL CONFERENCE ON WORLD WIDE WEB (WWW'16 COMPANION), 2016, : 141 - 142