Triangle Counting Accelerations: From Algorithm to In-Memory Computing Architecture

被引:13
作者
Wang, Xueyan [1 ]
Yang, Jianlei [2 ]
Zhao, Yinglin [1 ]
Jia, Xiaotao [1 ]
Yin, Rong [3 ]
Chen, Xuhang [1 ]
Qu, Gang [4 ,5 ]
Zhao, Weisheng [1 ]
机构
[1] Beihang Univ, Sch Integrated Circuit Sci & Engn, MIIT Key Lab Spintron, Beijing 100191, Peoples R China
[2] Beihang Univ, Sch Comp Sci & Engn, State Key Lab Software Dev Environm NLSDE, BDBC, Beijing 100191, Peoples R China
[3] Chinese Acad Sci, Inst Informat Engn, Beijing 100049, Peoples R China
[4] Univ Maryland, Dept Elect & Comp Engn, College Pk, MD 20742 USA
[5] Univ Maryland, Inst Syst Res, College Pk, MD 20742 USA
基金
中国国家自然科学基金;
关键词
Triangle counting acceleration; processing-in-memory; algorithm-architecture co-design; graph computing; NONVOLATILE MEMORY; ENERGY;
D O I
10.1109/TC.2021.3131049
中图分类号
TP3 [计算技术、计算机技术];
学科分类号
0812 ;
摘要
Triangles are the basic substructure of networks and triangle counting (TC) has been a fundamental graph computing problem in numerous fields such as social network analysis. Nevertheless, like other graph computing problems, due to the high memory-computation ratio and random memory access pattern, TC involves a large amount of data transfers thus suffers from the bandwidth bottleneck in the traditional Von-Neumann architecture. To overcome this challenge, in this paper, we propose to accelerate TC with the emerging processingin-memory (PIM) architecture through an algorithm-architecture co-optimization manner. To enable the efficient in-memory implementations, we come up to reformulate TC with bitwise logic operations (such as AND), and develop customized graph compression and mapping techniques for efficient data flow management. With the emerging computational Spin-Transfer Torque Magnetic RAM(STT-MRAM) array, which is one of the most promising PIM enabling techniques, the device-to-architecture co-simulation results demonstrate that the proposed TC in-memory accelerator outperforms the state-of-the-art GPU and FPGA accelerations by 12.2 x and 31.8 x, respectively, and achieves a 34 x energy efficiency improvement over the FPGA accelerator.
引用
收藏
页码:2462 / 2472
页数:11
相关论文
共 50 条
  • [31] Interplay Bitwise Operation in Emerging MRAM for Efficient In-memory Computing
    Cai, Hao
    Jiang, Honglan
    Zhou, Yongliang
    Han, Menglin
    Liu, Bo
    [J]. CCF TRANSACTIONS ON HIGH PERFORMANCE COMPUTING, 2020, 2 (03) : 282 - 296
  • [32] In-Memory Data Rearrangement for Irregular, Data-Intensive Computing
    Lloyd, Scott
    Gokhale, Maya
    [J]. COMPUTER, 2015, 48 (08) : 18 - 25
  • [33] In-Memory Computing based Machine Learning Accelerators: Opportunities and Challenges
    Roy, Kaushik
    [J]. PROCEEDINGS OF THE 32ND GREAT LAKES SYMPOSIUM ON VLSI 2022, GLSVLSI 2022, 2022, : 203 - 204
  • [34] Impact of Random Phase Distribution in Ferroelectric Transistors-Based 3-D NAND Architecture on In-Memory Computing
    Choe, Gihun
    Shim, Wonbo
    Wang, Panni
    Hur, Jae
    Khan, Asif Islam
    Yu, Shimeng
    [J]. IEEE TRANSACTIONS ON ELECTRON DEVICES, 2021, 68 (05) : 2543 - 2548
  • [35] An Energy-Efficient Nonvolatile In-Memory Computing Architecture for Extreme Learning Machine by Domain-Wall Nanowire Devices
    Wang, Yuhao
    Yu, Hao
    Ni, Leibin
    Huang, Guang-Bin
    Yan, Mei
    Weng, Chuliang
    Yang, Wei
    Zhao, Junfeng
    [J]. IEEE TRANSACTIONS ON NANOTECHNOLOGY, 2015, 14 (06) : 998 - 1012
  • [36] Processing In-Memory Architecture with On-Chip Transfer Learning Function for Compensating Characteristic Variation
    Hayashikoshi, Masanori
    Suzuki, Junichi
    Watanabe, Yasuaki
    Furuya, Akiko
    Yabuta, Toshio
    Ido, Masamichi
    Kiyota, Yukiyoshi
    Kono, Takashi
    [J]. 2020 IEEE INTERNATIONAL MEMORY WORKSHOP (IMW 2020), 2020, : 95 - 98
  • [37] A Communication-Aware DNN Accelerator on ImageNet Using In-Memory Entry-Counting Based Algorithm-Circuit-Architecture Co-Design in 65-nm CMOS
    Zhu, Haozhe
    Chen, Chixiao
    Liu, Shiwei
    Zou, Qiaosha
    Wang, Mingyu
    Zhang, Lihua
    Zeng, Xiaoyang
    Richard Shi, C. -J.
    [J]. IEEE JOURNAL ON EMERGING AND SELECTED TOPICS IN CIRCUITS AND SYSTEMS, 2020, 10 (03) : 283 - 294
  • [38] In-Memory Stateful Logic Computing Using Memristors: Gate, Calculation, and Application
    Xu, Nuo
    Park, Taegyun
    Yoon, Kyung Jean
    Hwang, Cheol Seong
    [J]. PHYSICA STATUS SOLIDI-RAPID RESEARCH LETTERS, 2021, 15 (09):
  • [39] Achieving software-equivalent accuracy for hyperdimensional computing with ferroelectric-based in-memory computing
    Kazemi, Arman
    Mueller, Franz
    Sharifi, Mohammad Mehdi
    Errahmouni, Hamza
    Gerlach, Gerald
    Kaempfe, Thomas
    Imani, Mohsen
    Hu, Xiaobo Sharon
    Niemier, Michael
    [J]. SCIENTIFIC REPORTS, 2022, 12 (01)
  • [40] Resistive-RAM-Based In-Memory Computing for Neural Network: A Review
    Chen, Weijian
    Qi, Zhi
    Akhtar, Zahid
    Siddique, Kamran
    [J]. ELECTRONICS, 2022, 11 (22)