Exploiting covariate embeddings for classification using Gaussian processes

被引:1
|
作者
Andrade, Daniel [1 ]
Tamura, Akihiro [2 ]
Tsuchida, Masaaki [3 ]
机构
[1] NEC Corp Ltd, Secur Res Labs, Tokyo, Japan
[2] Ehime Univ, Grad Sch Sci & Engn, Matsuyama, Ehime, Japan
[3] DeNA Co Ltd, Tokyo, Japan
关键词
Logistic regression; Auxiliary information of covariates; Gaussian process; Text classification; TEXT CLASSIFICATION;
D O I
10.1016/j.patrec.2018.01.011
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
In many logistic regression tasks, auxiliary information about the covariates is available. For example, a user might be able to specify a similarity measure between the covariates, or an embedding (feature vector) for each covariate, which is created from unlabeled data. In particular for text classification, the covariates (words) can be described by word embeddings or similarity measures from lexical resources like WordNet. We propose a new method to use such embeddings of covariates for logistic regression. Our method consists of two main components. The first component is a Gaussian process (GP) with a covariance function that models the correlations between covariates, and returns a noise-free estimate of the covariates. The second component is a logistic regression model that uses these noise-free estimates. One advantage of our model is that the covariance function can be adjusted to the training data using maximum likelihood. Another advantage is that new covariates that never occurred in the training data can be incorporated at test time, while run-time increases only linearly in the number of new covariates. Our experiments demonstrate the usefulness of our method in situations when only small training data is available. (c) 2018 Elsevier B.V. All rights reserved.
引用
收藏
页码:8 / 14
页数:7
相关论文
共 50 条
  • [21] Accurate posterior probability estimates for channel equalization using Gaussian processes for classification
    Perez-Cruz, Fernando
    Martinez-Olmos, Pablo
    Jose Murillo-Fuentest, Juan
    2007 IEEE 8TH WORKSHOP ON SIGNAL PROCESSING ADVANCES IN WIRELESS COMMUNICATIONS, VOLS 1 AND 2, 2007, : 380 - +
  • [22] Divergence-Based One-Class Classification Using Gaussian Processes
    Bodesheim, Paul
    Rodner, Erik
    Freytag, Alexander
    Denzler, Joachim
    PROCEEDINGS OF THE BRITISH MACHINE VISION CONFERENCE 2012, 2012,
  • [23] Gaussian Copula Embeddings
    Lu, Chien
    Peltonen, Jaakko
    ADVANCES IN NEURAL INFORMATION PROCESSING SYSTEMS 35 (NEURIPS 2022), 2022,
  • [24] Adversarial Robustness Guarantees for Classification with Gaussian Processes
    Blaas, Arno
    Patane, Andrea
    Laurenti, Luca
    Cardelli, Luca
    Kwiatkowska, Marta
    Roberts, Stephen
    INTERNATIONAL CONFERENCE ON ARTIFICIAL INTELLIGENCE AND STATISTICS, VOL 108, 2020, 108 : 3372 - 3381
  • [25] Extrinsic Gaussian Processes for Regression and Classification on Manifolds
    Lin, Lizhen
    Mu, Niu
    Cheung, Pokman
    Dunson, David
    BAYESIAN ANALYSIS, 2019, 14 (03): : 887 - 906
  • [26] Mean field methods for classification with Gaussian processes
    Opper, M
    Winther, O
    ADVANCES IN NEURAL INFORMATION PROCESSING SYSTEMS 11, 1999, 11 : 309 - 315
  • [27] One-Class Classification with Gaussian Processes
    Kemmler, Michael
    Rodner, Erik
    Denzler, Joachim
    COMPUTER VISION - ACCV 2010, PT II, 2011, 6493 : 489 - 500
  • [28] One-class classification with Gaussian processes
    Kemmler, Michael
    Rodner, Erik
    Wacker, Esther-Sabrina
    Denzler, Joachim
    PATTERN RECOGNITION, 2013, 46 (12) : 3507 - 3518
  • [29] Evaluation of Deep Gaussian Processes for Text Classification
    Jayashree, P.
    Srijith, P. K.
    PROCEEDINGS OF THE 12TH INTERNATIONAL CONFERENCE ON LANGUAGE RESOURCES AND EVALUATION (LREC 2020), 2020, : 1485 - 1491
  • [30] Semantic Mapping with Confidence Scores through Metric Embeddings and Gaussian Process Classification
    Hong, Jungseok
    Garg, Suveer
    Isler, Volkan
    2023 IEEE INTERNATIONAL CONFERENCE ON ROBOTICS AND AUTOMATION, ICRA, 2023, : 1723 - 1730