A Hierarchical Mixture-Of-Experts Framework for Few Labeled Node Classification

被引:0
|
作者
Wang, Yimeng [1 ,2 ]
Yang, Zhiyao [1 ,2 ]
Che, Xiangjiu [1 ,2 ]
机构
[1] Jilin Univ, Coll Comp Sci & Technol, Changchun 130012, Jilin, Peoples R China
[2] Jilin Univ, Key Lab Symbol Computat & Knowledge Engn MOE, Changchun 130012, Jilin, Peoples R China
基金
中国国家自然科学基金;
关键词
Node classification; Mixture of experts; Data augmentation; Few labeled graph;
D O I
10.1016/j.neunet.2025.107285
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
Few Labeled Node Classification (FLNC) is a challenging subtask of node classification, where training nodes are extremely limited, often with only one or two labels per class. While Graph Neural Networks (GNNs) show promise, they often suffer from feature convergence. A common method to address this challenge is multi-perspective feature extraction, with the Mixture of Experts (MoE) model being a popular approach. However, directly applying MoE to FLNC frequently results in overfitting. To address these issues, we propose the Hierarchical Mixture-of-Experts (HMoE) framework. First, we mitigate overfitting by applying three data augmentation techniques to enrich input features. Next, we design a novel hierarchical mixture-of-experts encoder to achieve diversified feature representations, where the first layer extracts unique feature information, and the second layer refines shared information. Additionally, we design an auxiliary task to distinguish between original and augmented data, using a gradient reversal mechanism to enhance the feature representation ability of graph data. The experimental results show that HMoE outperforms the baseline methods, achieving an average 1.2% performance improvement across six datasets.
引用
收藏
页数:12
相关论文
共 50 条
  • [21] TabMoE: A General Framework for Diverse Table-Based Reasoning with Mixture-of-Experts
    Wu, Jie
    Hou, Mengshu
    MATHEMATICS, 2024, 12 (19)
  • [22] MoDE: A Mixture-of-Experts Model with Mutual Distillation among the Experts
    Xie, Zhitian
    Zhang, Yinger
    Zhuang, Chenyi
    Shi, Qitao
    Liu, Zhining
    Gu, Jinjie
    Zhang, Guannan
    THIRTY-EIGHTH AAAI CONFERENCE ON ARTIFICIAL INTELLIGENCE, VOL 38 NO 14, 2024, : 16067 - 16075
  • [23] Improving risk classification and ratemaking using mixture-of-experts models with random effects
    Tseung, Spark C.
    Chan, Ian Weng
    Fung, Tsz Chai
    Badescu, Andrei L.
    Lin, X. Sheldon
    JOURNAL OF RISK AND INSURANCE, 2023, 90 (03) : 789 - 820
  • [24] A Universal Approximation Theorem for Mixture-of-Experts Models
    Nguyen, Hien D.
    Lloyd-Jones, Luke R.
    McLachlan, Geoffrey J.
    NEURAL COMPUTATION, 2016, 28 (12) : 2585 - 2593
  • [25] On the Benefits of Learning to Route in Mixture-of-Experts Models
    Dikkala, Nishanth
    Ghosh, Nikhil
    Meka, Raghu
    Panigrahy, Rina
    Vyas, Nikhil
    Wang, Xin
    2023 CONFERENCE ON EMPIRICAL METHODS IN NATURAL LANGUAGE PROCESSING (EMNLP 2023), 2023, : 9376 - 9396
  • [26] Measurement of the probability of insolvency with mixture-of-experts networks
    Baetge, J
    Jerschensky, A
    CLASSIFICATION IN THE INFORMATION AGE, 1999, : 421 - 429
  • [27] Advances in using hierarchical mixture of experts for signal classification
    Ramamurti, V
    Ghosh, J
    1996 IEEE INTERNATIONAL CONFERENCE ON ACOUSTICS, SPEECH, AND SIGNAL PROCESSING, CONFERENCE PROCEEDINGS, VOLS 1-6, 1996, : 3569 - 3572
  • [28] Breaking the gridlock in Mixture-of-Experts: Consistent and Efficient Algorithms
    Makkuva, Ashok Vardhan
    Oh, Sewoong
    Kannan, Sreeram
    Viswanath, Pramod
    INTERNATIONAL CONFERENCE ON MACHINE LEARNING, VOL 97, 2019, 97
  • [29] A Mixture-of-Experts Model for Antonym-Synonym Discrimination
    Xie, Zhipeng
    Zeng, Nan
    ACL-IJCNLP 2021: THE 59TH ANNUAL MEETING OF THE ASSOCIATION FOR COMPUTATIONAL LINGUISTICS AND THE 11TH INTERNATIONAL JOINT CONFERENCE ON NATURAL LANGUAGE PROCESSING, VOL 2, 2021, : 558 - 564
  • [30] A similarity-based Bayesian mixture-of-experts model
    Tianfang Zhang
    Rasmus Bokrantz
    Jimmy Olsson
    Statistics and Computing, 2023, 33