Triangle Counting Accelerations: From Algorithm to In-Memory Computing Architecture

被引:13
作者
Wang, Xueyan [1 ]
Yang, Jianlei [2 ]
Zhao, Yinglin [1 ]
Jia, Xiaotao [1 ]
Yin, Rong [3 ]
Chen, Xuhang [1 ]
Qu, Gang [4 ,5 ]
Zhao, Weisheng [1 ]
机构
[1] Beihang Univ, Sch Integrated Circuit Sci & Engn, MIIT Key Lab Spintron, Beijing 100191, Peoples R China
[2] Beihang Univ, Sch Comp Sci & Engn, State Key Lab Software Dev Environm NLSDE, BDBC, Beijing 100191, Peoples R China
[3] Chinese Acad Sci, Inst Informat Engn, Beijing 100049, Peoples R China
[4] Univ Maryland, Dept Elect & Comp Engn, College Pk, MD 20742 USA
[5] Univ Maryland, Inst Syst Res, College Pk, MD 20742 USA
基金
中国国家自然科学基金;
关键词
Triangle counting acceleration; processing-in-memory; algorithm-architecture co-design; graph computing; NONVOLATILE MEMORY; ENERGY;
D O I
10.1109/TC.2021.3131049
中图分类号
TP3 [计算技术、计算机技术];
学科分类号
0812 ;
摘要
Triangles are the basic substructure of networks and triangle counting (TC) has been a fundamental graph computing problem in numerous fields such as social network analysis. Nevertheless, like other graph computing problems, due to the high memory-computation ratio and random memory access pattern, TC involves a large amount of data transfers thus suffers from the bandwidth bottleneck in the traditional Von-Neumann architecture. To overcome this challenge, in this paper, we propose to accelerate TC with the emerging processingin-memory (PIM) architecture through an algorithm-architecture co-optimization manner. To enable the efficient in-memory implementations, we come up to reformulate TC with bitwise logic operations (such as AND), and develop customized graph compression and mapping techniques for efficient data flow management. With the emerging computational Spin-Transfer Torque Magnetic RAM(STT-MRAM) array, which is one of the most promising PIM enabling techniques, the device-to-architecture co-simulation results demonstrate that the proposed TC in-memory accelerator outperforms the state-of-the-art GPU and FPGA accelerations by 12.2 x and 31.8 x, respectively, and achieves a 34 x energy efficiency improvement over the FPGA accelerator.
引用
收藏
页码:2462 / 2472
页数:11
相关论文
共 50 条
  • [21] Optimization of Projected Phase Change Memory for Analog In-Memory Computing Inference
    Li, Ning
    Mackin, Charles
    Chen, An
    Brew, Kevin
    Philip, Timothy
    Simon, Andrew
    Saraf, Iqbal
    Han, Jin-Ping
    Sarwat, Syed Ghazi
    Burr, Geoffrey W.
    Rasch, Malte
    Sebastian, Abu
    Narayanan, Vijay
    Saulnier, Nicole
    ADVANCED ELECTRONIC MATERIALS, 2023, 9 (06)
  • [22] STT-RAM-Based Hierarchical in-Memory Computing
    Gajaria, Dhruv
    Gomez, Kevin Antony
    Adegbija, Tosiron
    IEEE TRANSACTIONS ON PARALLEL AND DISTRIBUTED SYSTEMS, 2024, 35 (09) : 1615 - 1629
  • [23] Cryogenic Hyperdimensional In-Memory Computing Using Ferroelectric TCAM
    Singh Parihar, Shivendra
    Kumar, Shubham
    Chatterjee, Swetaki
    Pahwa, Girish
    Singh Chauhan, Yogesh
    Amrouch, Hussam
    IEEE JOURNAL ON EXPLORATORY SOLID-STATE COMPUTATIONAL DEVICES AND CIRCUITS, 2025, 11 : 34 - 41
  • [24] NNgine: Ultra-Efficient Nearest Neighbor Accelerator Based on In-Memory Computing
    Imani, Mohsen
    Kim, Yeseong
    Rosing, Tajana
    2017 IEEE INTERNATIONAL CONFERENCE ON REBOOTING COMPUTING (ICRC), 2017, : 228 - 235
  • [25] Domain-Specific STT-MRAM-Based In-Memory Computing: A Survey
    Yusuf, Alaba
    Adegbija, Tosiron
    Gajaria, Dhruv
    IEEE ACCESS, 2024, 12 : 28036 - 28056
  • [26] Impact of Random Phase Distribution in 3D Vertical NAND Architecture of Ferroelectric Transistors on In-Memory Computing
    Choe, Gihun
    Shim, Wonbo
    Hur, Jae
    Khan, Asif Islam
    Yu, Shimeng
    2020 INTERNATIONAL CONFERENCE ON SIMULATION OF SEMICONDUCTOR PROCESSES AND DEVICES (SISPAD 2020), 2020, : 165 - 168
  • [27] A Scalable and Efficient In-Memory Interconnect Architecture for Automata Processing
    Sadredini, Elaheh
    Rahimi, Reza
    Verma, Vaibhav
    Stan, Mircea
    Skadron, Kevin
    IEEE COMPUTER ARCHITECTURE LETTERS, 2019, 18 (02) : 87 - 90
  • [28] Resistive Memory-Based In-Memory Computing: From Device and Large-Scale Integration System Perspectives
    Yan, Bonan
    Li, Bing
    Qiao, Ximing
    Xue, Cheng-Xin
    Chang, Meng-Fan
    Chen, Yiran
    Li, Hai
    ADVANCED INTELLIGENT SYSTEMS, 2019, 1 (07)
  • [29] Optimization of OLAP In-Memory Database Management Systems with Processing-In-Memory Architecture
    Hosseinzadeh, Shima
    Parvaresh, Amirhossein
    Fey, Dietmar
    ARCHITECTURE OF COMPUTING SYSTEMS, ARCS 2023, 2023, 13949 : 264 - 278
  • [30] MOL-Based In-Memory Computing of Binary Neural Networks
    Ali, Khaled Alhaj
    Baghdadi, Amer
    Dupraz, Elsa
    Leonardon, Mathieu
    Rizk, Mostafa
    Diguet, Jean-Philippe
    IEEE TRANSACTIONS ON VERY LARGE SCALE INTEGRATION (VLSI) SYSTEMS, 2022, 30 (07) : 869 - 880