A parameter-free text classification method based on dual compressors

被引:0
|
作者
Mao, Yanxu [1 ]
Ding, Ying [2 ]
Cui, Tiehan [1 ]
机构
[1] Henan Univ, Sch Software, Kaifeng, Peoples R China
[2] Henan Inst Sci & Technol, Sch Comp Sci & Technol, Xinxiang, Peoples R China
关键词
Text classification; Compressor; Few-shot learning; LSTM;
D O I
10.1007/s10115-024-02335-9
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
With the development of the information age, the emergence of massive amounts of text data has made effective text classification a critical challenge. Traditional classification methods often underperform or incur high computational costs when dealing with heterogeneous data, limited labeled data, or domain-specific data. To address these challenges, this paper proposes a novel text classification model, GZclassifier, designed to improve both accuracy and efficiency. GZclassifier employs two distinct compressors to handle information data and calculates distances in parallel to facilitate classification. This dual-compressor approach enhances the model's ability to manage diverse and sparse data effectively. We conducted extensive experimental evaluations on a range of public datasets, including those with few-shot learning scenarios, to assess the proposed method's performance. The results demonstrate that our model significantly outperforms traditional methods in terms of classification accuracy, robustness, and computational efficiency. The GZclassifier's ability to handle limited labeled data and domain-specific contexts highlights its potential as an efficient solution for real-world text classification tasks. This study not only advances the field of text classification but also showcases the model's practical applicability and benefits in various text processing scenarios.
引用
收藏
页码:3737 / 3767
页数:31
相关论文
共 50 条
  • [1] "Low-Resource" Text Classification: A Parameter-Free Classification Method with Compressors
    Jiang, Zhiying
    Yang, Matthew Y. R.
    Tsirlin, Mikhail
    Tang, Raphael
    Dai, Yiqin
    Lin, Jimmy
    FINDINGS OF THE ASSOCIATION FOR COMPUTATIONAL LINGUISTICS, ACL 2023, 2023, : 6810 - 6828
  • [2] A parameter-free associative classification method
    Cerf, Loic
    Gay, Dominique
    Selmaoui, Nazha
    Boulicaut, Jean-Francois
    DATA WAREHOUSING AND KNOWLEDGE DISCOVERY, PROCEEDINGS, 2008, 5182 : 293 - +
  • [3] A Parameter-Free Classification Method for Large Scale Learning
    Boulle, Marc
    JOURNAL OF MACHINE LEARNING RESEARCH, 2009, 10 : 1367 - 1385
  • [4] A Parameter-Free Cleaning Method for SMOTE in Imbalanced Classification
    Yan, Yuanting
    Liu, Ruiqing
    Ding, Zihan
    Du, Xiuquan
    Chen, Jie
    Zhang, Yanping
    IEEE ACCESS, 2019, 7 : 23537 - 23548
  • [5] An Improved Audio Classification Method Based on Parameter-Free Attention Combined with Self-Supervision
    Gong X.
    Li Z.
    Jisuanji Fuzhu Sheji Yu Tuxingxue Xuebao/Journal of Computer-Aided Design and Computer Graphics, 2023, 35 (03): : 434 - 440
  • [6] A Parameter-Free Linear Sampling Method
    Liu, Lei
    Hu, Guanzhong
    IEEE ACCESS, 2019, 7 : 17935 - 17940
  • [7] Parameter-Free Extreme Learning Machine for Imbalanced Classification
    Li, Li
    Zhao, Kaiyi
    Sun, Ruizhi
    Gan, Jiangzhang
    Yuan, Gang
    Liu, Tong
    NEURAL PROCESSING LETTERS, 2020, 52 (03) : 1927 - 1944
  • [8] A parameter-free affinity based clustering
    Mukhoty, Bhaskar
    Gupta, Ruchir
    Lakshmanan, K.
    Kumar, Mayank
    APPLIED INTELLIGENCE, 2020, 50 (12) : 4543 - 4556
  • [9] Parameter-Free Extreme Learning Machine for Imbalanced Classification
    Li Li
    Kaiyi Zhao
    Ruizhi Sun
    Jiangzhang Gan
    Gang Yuan
    Tong Liu
    Neural Processing Letters, 2020, 52 : 1927 - 1944
  • [10] A parameter-free affinity based clustering
    Bhaskar Mukhoty
    Ruchir Gupta
    Lakshmanan K.
    Mayank Kumar
    Applied Intelligence, 2020, 50 : 4543 - 4556