Data Pruning-enabled High Performance and Reliable Graph Neural Network Training on ReRAM-based Processing-in-Memory Accelerators

被引:1
|
作者
Ogbogu, Chukwufumnanya [1 ]
Joardar, Biresh [2 ]
Chakrabarty, Krishnendu [3 ]
Doppa, Jana [1 ]
Pande, Partha Pratim [1 ]
机构
[1] Washington State Univ, Pullman, WA 99164 USA
[2] Univ Houston Syst, Houston, TX USA
[3] Arizona State Univ, Tempe, AZ USA
基金
美国国家科学基金会;
关键词
Performance; reliability; non-volatile memory; endurance;
D O I
10.1145/3656171
中图分类号
TP3 [计算技术、计算机技术];
学科分类号
0812 ;
摘要
Graph Neural Networks (GNNs) have achieved remarkable accuracy in cognitive tasks such as predictive analytics on graph-structured data. Hence, they have become very popular in diverse real-world applications. However, GNN training with large real-world graph datasets in edge-computing scenarios is both memory- and compute-intensive. Traditional computing platforms such as CPUs and GPUs do not provide the energy efficiency and low latency required in edge intelligence applications due to their limited memory bandwidth. Resistive random-access memory (ReRAM)-based processing-in-memory (PIM) architectures have been proposed as suitable candidates for accelerating AI applications at the edge, including GNN training. However, ReRAM-based PIM architectures suffer from low reliability due to their limited endurance, and low performance when they are used for GNN training in real-world scenarios with large graphs. In this work, we propose a learning-for-data-pruning framework, which leverages a trained Binary Graph Classifier (BGC) to reduce the size of the input data graph by pruning subgraphs early in the training process to accelerate the GNN training process on ReRAM-based architectures. The proposed light-weight BGC model reduces the amount of redundant information in input graph(s) to speed up the overall training process, improves the reliability of the ReRAM-based PIM accelerator, and reduces the overall training cost. This enables fast, energy-efficient, and reliable GNN training on ReRAM-based architectures. Our experimental results demonstrate that using this learning for data pruning framework, we can accelerate GNN training and improve the reliability of ReRAM-based PIM architectures by up to 1.6 x, and reduce the overall training cost by 100 x compared to state-of-the-art data pruning techniques. CCS Concepts: center dot Hardware -> Emerging technologies ; Analysis and design of emerging devices and systems ; Emerging architectures ;
引用
收藏
页数:29
相关论文
共 17 条
  • [1] Runtime Row/Column Activation Pruning for ReRAM-based Processing-in-Memory DNN Accelerators
    Jiang, Xikun
    Shen, Zhaoyan
    Sun, Siqing
    Yin, Ping
    Jia, Zhiping
    Ju, Lei
    Zhang, Zhiyong
    Yu, Dongxiao
    2023 IEEE/ACM INTERNATIONAL CONFERENCE ON COMPUTER AIDED DESIGN, ICCAD, 2023,
  • [2] ReRAM-Based Processing-in-Memory Architecture for Recurrent Neural Network Acceleration
    Long, Yun
    Na, Taesik
    Mukhopadhyay, Saibal
    IEEE TRANSACTIONS ON VERY LARGE SCALE INTEGRATION (VLSI) SYSTEMS, 2018, 26 (12) : 2781 - 2794
  • [3] A Novel ReRAM-based Processing-in-Memory Architecture for Graph Computing
    Han, Lei
    Shen, Zhaoyan
    Shao, Zili
    Huang, H. Howie
    Li, Tao
    2017 IEEE 6TH NON-VOLATILE MEMORY SYSTEMS AND APPLICATIONS SYMPOSIUM (NVMSA 2017), 2017,
  • [4] A Novel ReRAM-Based Processing-in-Memory Architecture for Graph Traversal
    Han, Lei
    Shen, Zhaoyan
    Liu, Duo
    Shao, Zili
    Huang, H. Howie
    Li, Tao
    ACM TRANSACTIONS ON STORAGE, 2018, 14 (01)
  • [5] A Survey of ReRAM-Based Architectures for Processing-In-Memory and Neural Networks
    Mittal, Sparsh
    MACHINE LEARNING AND KNOWLEDGE EXTRACTION, 2019, 1 (01): : 75 - 114
  • [6] PRIME: A Novel Processing-in-memory Architecture for Neural Network Computation in ReRAM-based Main Memory
    Chi, Ping
    Li, Shuangchen
    Xu, Cong
    Zhang, Tao
    Zhao, Jishen
    Liu, Yongpan
    Wang, Yu
    Xie, Yuan
    2016 ACM/IEEE 43RD ANNUAL INTERNATIONAL SYMPOSIUM ON COMPUTER ARCHITECTURE (ISCA), 2016, : 27 - 39
  • [7] ReaDy: A ReRAM-Based Processing-in-Memory Accelerator for Dynamic Graph Convolutional Networks
    Huang, Yu
    Zheng, Long
    Yao, Pengcheng
    Wang, Qinggang
    Liu, Haifeng
    Liao, Xiaofei
    Jin, Hai
    Xue, Jingling
    IEEE TRANSACTIONS ON COMPUTER-AIDED DESIGN OF INTEGRATED CIRCUITS AND SYSTEMS, 2022, 41 (11) : 3567 - 3578
  • [8] Accelerating Graph Neural Network Training on ReRAM-Based PIM Architectures via Graph and Model Pruning
    Ogbogu, Chukwufumnanya O.
    Arka, Aqeeb Iqbal
    Pfromm, Lukas
    Joardar, Biresh Kumar
    Doppa, Janardhan Rao
    Chakrabarty, Krishnendu
    Pande, Partha Pratim
    IEEE TRANSACTIONS ON COMPUTER-AIDED DESIGN OF INTEGRATED CIRCUITS AND SYSTEMS, 2023, 42 (08) : 2703 - 2716
  • [9] CRPIM: An efficient compute-reuse scheme for ReRAM-based Processing-in-Memory DNN accelerators
    Hong, Shihao
    Chung, Yeh-Ching
    JOURNAL OF SYSTEMS ARCHITECTURE, 2024, 153
  • [10] A Quantized Training Framework for Robust and Accurate ReRAM-based Neural Network Accelerators
    Zhang, Chenguang
    Zhou, Pingqiang
    2021 26TH ASIA AND SOUTH PACIFIC DESIGN AUTOMATION CONFERENCE (ASP-DAC), 2021, : 43 - 48