A Parameter-Free Classification Method for Large Scale Learning

被引:0
|
作者
Boulle, Marc [1 ]
机构
[1] Orange Labs, F-22300 Lannion, France
关键词
large scale learning; naive Bayes; Bayesianism; model selection; model averaging; BAYES;
D O I
暂无
中图分类号
TP [自动化技术、计算机技术];
学科分类号
0812 ;
摘要
With the rapid growth of computer storage capacities, available data and demand for scoring models both follow an increasing trend, sharper than that of the processing power. However, the main limitation to a wide spread of data mining solutions is the non-increasing availability of skilled data analysts, which play a key role in data preparation and model selection. In this paper, we present a parameter-free scalable classification method, which is a step towards fully automatic data mining. The method is based on Bayes optimal univariate conditional density estimators, naive Bayes classification enhanced with a Bayesian variable selection scheme, and averaging of models using a logarithmic smoothing of the posterior distribution. We focus on the complexity of the algorithms and show how they can cope with data sets that are far larger than the available central memory. We finally report results on the Large Scale Learning challenge, where our method obtains state of the art performance within practicable computation time.
引用
收藏
页码:1367 / 1385
页数:19
相关论文
共 50 条
  • [1] A parameter-free associative classification method
    Cerf, Loic
    Gay, Dominique
    Selmaoui, Nazha
    Boulicaut, Jean-Francois
    DATA WAREHOUSING AND KNOWLEDGE DISCOVERY, PROCEEDINGS, 2008, 5182 : 293 - +
  • [2] Parameter-Free Extreme Learning Machine for Imbalanced Classification
    Li, Li
    Zhao, Kaiyi
    Sun, Ruizhi
    Gan, Jiangzhang
    Yuan, Gang
    Liu, Tong
    NEURAL PROCESSING LETTERS, 2020, 52 (03) : 1927 - 1944
  • [3] Parameter-Free Extreme Learning Machine for Imbalanced Classification
    Li Li
    Kaiyi Zhao
    Ruizhi Sun
    Jiangzhang Gan
    Gang Yuan
    Tong Liu
    Neural Processing Letters, 2020, 52 : 1927 - 1944
  • [4] A Parameter-Free Cleaning Method for SMOTE in Imbalanced Classification
    Yan, Yuanting
    Liu, Ruiqing
    Ding, Zihan
    Du, Xiuquan
    Chen, Jie
    Zhang, Yanping
    IEEE ACCESS, 2019, 7 : 23537 - 23548
  • [5] A parameter-free text classification method based on dual compressors
    Mao, Yanxu
    Ding, Ying
    Cui, Tiehan
    KNOWLEDGE AND INFORMATION SYSTEMS, 2025, : 3737 - 3767
  • [6] "Low-Resource" Text Classification: A Parameter-Free Classification Method with Compressors
    Jiang, Zhiying
    Yang, Matthew Y. R.
    Tsirlin, Mikhail
    Tang, Raphael
    Dai, Yiqin
    Lin, Jimmy
    FINDINGS OF THE ASSOCIATION FOR COMPUTATIONAL LINGUISTICS, ACL 2023, 2023, : 6810 - 6828
  • [7] A parameter-free learning automaton scheme
    Ren, Xudie
    Li, Shenghong
    Ge, Hao
    FRONTIERS IN NEUROROBOTICS, 2022, 16
  • [8] Gene Classification Using Parameter-Free Semi-Supervised Manifold Learning
    Huang, Hong
    Feng, Hailiang
    IEEE-ACM TRANSACTIONS ON COMPUTATIONAL BIOLOGY AND BIOINFORMATICS, 2012, 9 (03) : 818 - 827
  • [9] A Parameter-Free Linear Sampling Method
    Liu, Lei
    Hu, Guanzhong
    IEEE ACCESS, 2019, 7 : 17935 - 17940
  • [10] Parameter-Free Loss for Class-Imbalanced Deep Learning in Image Classification
    Du, Jie
    Zhou, Yanhong
    Liu, Peng
    Vong, Chi-Man
    Wang, Tianfu
    IEEE TRANSACTIONS ON NEURAL NETWORKS AND LEARNING SYSTEMS, 2023, 34 (06) : 3234 - 3240