Sparse Topical Coding with Sparse Groups

被引:2
作者
Peng, Min [1 ]
Xie, Qianqian [1 ]
Huang, Jiajia [1 ]
Zhu, Jiahui [1 ]
Ouyang, Shuang [1 ]
Huang, Jimin [1 ]
Tian, Gang [1 ]
机构
[1] Wuhan Univ, Sch Comp, Wuhan, Peoples R China
来源
WEB-AGE INFORMATION MANAGEMENT, PT I | 2016年 / 9658卷
关键词
Document representation; Topic model; Sparse coding; Sparse group lasso; REGRESSION; SELECTION;
D O I
10.1007/978-3-319-39937-9_32
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
Learning a latent semantic representing from a large number of short text corpora makes a profound practical significance in research and engineering. However, it is difficult to use standard topic models in microblogging environments since microblogs have short length, large amount, snarled noise and irregular modality characters, which prevent topic models from using full information of microblogs. In this paper, we propose a novel non-probabilistic topic model called sparse topical coding with sparse groups (STCSG), which is capable of discovering sparse latent semantic representations of large short text corpora. STCSG relaxes the normalization constraint of the inferred representations with sparse group lasso, a sparsity-inducing regularizer, which is convenient to directly control the sparsity of document, topic and word codes. Furthermore, the relaxed non-probabilistic STCSG can be effectively learned with alternating direction method of multipliers (ADMM). Our experimental results on Twitter dataset demonstrate that STCSG performs well in finding meaningful latent representations of short documents. Therefore, it can substantially improve the accuracy and efficiency of document classification.
引用
收藏
页码:415 / 426
页数:12
相关论文
共 50 条
  • [41] Local and global regularized sparse coding for data representation
    Shu, Zhenqiu
    Zhou, Jun
    Huang, Pu
    Yu, Xun
    Yang, Zhangjing
    Zhao, Chunxia
    [J]. NEUROCOMPUTING, 2016, 175 : 188 - 197
  • [42] Sparse Coding for Transform Domain-Based Sparse OFDM Channel Estimation
    Nazzal, Mahmoud
    Aygul, Mehmet Ali
    Gorcin, Ali
    Arslan, Huseyin
    [J]. 2019 27TH SIGNAL PROCESSING AND COMMUNICATIONS APPLICATIONS CONFERENCE (SIU), 2019,
  • [43] Selectivity and robustness of sparse coding networks
    Paiton, Dylan M.
    Frye, Charles G.
    Lundquist, Sheng Y.
    Bowen, Joel D.
    Zarcone, Ryan
    Olshausen, Bruno A.
    [J]. JOURNAL OF VISION, 2020, 20 (12): : 1 - 28
  • [44] Toward Sparse Coding on Cosine Distance
    Choi, Jonghyun
    Cho, Hyunjong
    Kwac, Jungsuk
    Davis, Larry S.
    [J]. 2014 22ND INTERNATIONAL CONFERENCE ON PATTERN RECOGNITION (ICPR), 2014, : 4423 - 4428
  • [45] Neural associative memories and sparse coding
    Palm, Guenther
    [J]. NEURAL NETWORKS, 2013, 37 : 163 - 169
  • [46] Continuous speech recognition with sparse coding
    Smit, W. J.
    Barnard, E.
    [J]. COMPUTER SPEECH AND LANGUAGE, 2009, 23 (02) : 200 - 219
  • [47] Maximum-margin sparse coding
    Liu, Chien-Liang
    Hsaio, Wen-Hoar
    Xiao, Bin
    Chen, Chun-Yu
    Wu, Wei-Liang
    [J]. NEUROCOMPUTING, 2017, 238 : 340 - 350
  • [48] Sparse and silent coding in neural circuits
    Lorincz, Andras
    Palotai, Zsolt
    Szirtes, Gabor
    [J]. NEUROCOMPUTING, 2012, 79 : 115 - 124
  • [49] Robust sparse coding for subspace learning
    Dai, Xiangguang
    Tao, Yingyin
    Xiong, Jiang
    Feng, Yuming
    [J]. ITALIAN JOURNAL OF PURE AND APPLIED MATHEMATICS, 2020, (44): : 986 - 994
  • [50] Classifying chart images with sparse coding
    Gao, Jinglun
    Zhou, Yin
    Barner, Kenneth E.
    [J]. COMPRESSIVE SENSING, 2012, 8365