Communication-efficient distributed large-scale sparse multinomial logistic regression

被引:0
|
作者
Lei, Dajiang [1 ]
Huang, Jie [1 ]
Chen, Hao [1 ]
Li, Jie [1 ]
Wu, Yu [2 ]
机构
[1] Chongqing Univ Posts & Telecommun, Coll Comp, Chongqing, Peoples R China
[2] Chongqing Univ Posts & Telecommun, Inst Web Intelligence, Chongqing, Peoples R China
来源
CONCURRENCY AND COMPUTATION-PRACTICE & EXPERIENCE | 2023年 / 35卷 / 18期
关键词
alternating direction method of multipliers; big data; distributed parallel; sparse multinomial logistic regression; PARALLEL;
D O I
10.1002/cpe.6148
中图分类号
TP31 [计算机软件];
学科分类号
081202 ; 0835 ;
摘要
Sparse multinomial logistic regression (SMLR) is widely used in image classification and text classification due to its feature selection and probabilistic output. However, the traditional SMLR algorithm cannot satisfy the memory and time needs of big data, which makes it necessary to propose a new distributed solution algorithm. The existing distributed SMLR algorithm has some shortcomings in network strategy and cannot make full use of the computing resources of the current high-performance cluster. Therefore, we propose communication-efficient sparse multinomial logistic regression (CESMLR), which adopts the efficient network strategy of each node to solve the SMLR subproblem and achieve a large number of data partitions, taking full advantage of the computing resources of the cluster to achieve an efficient SMLR solution. The big data experimental results show that the performance of our algorithm exceeds those of state-of-the-art algorithms. CESMLR is suitable for processing tasks with high-dimensional features and consumes less running time while maintaining high classification accuracy.
引用
收藏
页数:12
相关论文
共 50 条
  • [1] Communication-efficient distributed estimation for high-dimensional large-scale linear regression
    Liu, Zhan
    Zhao, Xiaoluo
    Pan, Yingli
    METRIKA, 2023, 86 (04) : 455 - 485
  • [2] Communication-efficient distributed estimation for high-dimensional large-scale linear regression
    Zhan Liu
    Xiaoluo Zhao
    Yingli Pan
    Metrika, 2023, 86 : 455 - 485
  • [3] Large-Scale Sparse Logistic Regression
    Liu, Jun
    Chen, Jianhui
    Ye, Jieping
    KDD-09: 15TH ACM SIGKDD CONFERENCE ON KNOWLEDGE DISCOVERY AND DATA MINING, 2009, : 547 - 555
  • [4] Communication-Efficient Distributed Multiple Testing for Large-Scale Inference
    Pournaderi, Mehrdad
    Xiang, Yu
    arXiv, 2022,
  • [5] Distributed Parallel Sparse Multinomial Logistic Regression
    Lei, Dajiang
    Du, Meng
    Chen, Hao
    Li, Zhixing
    Wu, Yu
    IEEE ACCESS, 2019, 7 : 55496 - 55508
  • [6] Communication-efficient sparse regression
    Lee, Jason D.
    Liu, Qiang
    Sun, Yuekai
    Taylor, Jonathan E.
    Journal of Machine Learning Research, 2017, 18 : 1 - 30
  • [7] Communication-efficient Sparse Regression
    Lee, Jason D.
    Liu, Qiang
    Sun, Yuekai
    Taylor, Jonathan E.
    JOURNAL OF MACHINE LEARNING RESEARCH, 2017, 18
  • [8] Communication-efficient sparse composite quantile regression for distributed data
    Yaohong Yang
    Lei Wang
    Metrika, 2023, 86 : 261 - 283
  • [9] Communication-efficient sparse composite quantile regression for distributed data
    Yang, Yaohong
    Wang, Lei
    METRIKA, 2023, 86 (03) : 261 - 283
  • [10] Exploring Fast and Communication-Efficient Algorithms in Large-scale Distributed Networks
    Yu, Yue
    Wu, Jiaxiang
    Huang, Junzhou
    22ND INTERNATIONAL CONFERENCE ON ARTIFICIAL INTELLIGENCE AND STATISTICS, VOL 89, 2019, 89 : 674 - 683