Machine Learning Computers With Fractal von Neumann Architecture

被引:10
|
作者
Zhao, Yongwei [1 ,2 ,3 ]
Fan, Zhe [1 ,2 ,3 ]
Du, Zidong [1 ,3 ]
Zhi, Tian [1 ,3 ]
Li, Ling [4 ]
Guo, Qi [1 ]
Liu, Shaoli [1 ,3 ]
Xu, Zhiwei [1 ,2 ]
Chen, Tianshi [1 ,3 ]
Chen, Yunji [1 ,2 ,5 ]
机构
[1] Chinese Acad Sci, Inst Comp Technol, State Key Lab Comp Architecture, Beijing 100190, Peoples R China
[2] Univ Chinese Acad Sci, Beijing 100049, Peoples R China
[3] Cambricon Technol, Beijing, Peoples R China
[4] Chinese Acad Sci, Inst Software, Beijing 100190, Peoples R China
[5] CAS Ctr Excellence Brain Sci & Intelligence Techn, Shanghai Res Ctr Brian Sci & Brain Inspired Intel, Inst Brain Intelligence Technol, Zhangjiang Lab BIT,ZfLab, Beijing, Peoples R China
基金
北京市自然科学基金;
关键词
Machine learning; Computers; Fractals; Programming; Computer architecture; Graphics processing units; Matrix decomposition; architecture; neural networks; programming efficiency;
D O I
10.1109/TC.2020.2982159
中图分类号
TP3 [计算技术、计算机技术];
学科分类号
0812 ;
摘要
Machine learning techniques are pervasive tools for emerging commercial applications and many dedicated machine learning computers on different scales have been deployed in embedded devices, servers, and data centers. Currently, most machine learning computer architectures still focus on optimizing performance and energy efficiency instead of programming productivity. However, with the fast development in silicon technology, programming productivity, including programming itself and software stack development, becomes the vital reason instead of performance and power efficiency that hinders the application of machine learning computers. In this article, we propose Cambricon-F, which is a series of homogeneous, sequential, multi-layer, layer-similar, and machine learning computers with same ISA. A Cambricon-F machine has a fractal von Neumann architecture to iteratively manage its components: it is with von Neumann architecture and its processing components (sub-nodes) are still Cambricon-F machines with von Neumann architecture and the same ISA. Since different Cambricon-F instances with different scales can share the same software stack on their common ISA, Cambricon-Fs can significantly improve the programming productivity. Moreover, we address four major challenges in Cambricon-F architecture design, which allow Cambricon-F to achieve a high efficiency. We implement two Cambricon-F instances at different scales, i.e., Cambricon-F100 and Cambricon-F1. Compared to GPU based machines (DGX-1 and 1080Ti), Cambricon-F instances achieve 2.82x, 5.14x better performance, 8.37x, 11.39x better efficiency on average, with 74.5, 93.8 percent smaller area costs, respectively. We further propose Cambricon-FR, which enhances the Cambricon-F machine learning computers to flexibly and efficiently support all the fractal operations with a reconfigurable fractal instruction set architecture. Compared to the Cambricon-F instances, Cambricon-FR machines achieve 1.96x, 2.49x better performance on average. Most importantly, Cambricon-FR computers are able to save the code length with a factor of 5.83, thus significantly improving the programming productivity.
引用
收藏
页码:998 / 1014
页数:17
相关论文
共 50 条
  • [1] A Heterogeneous Parallel Non-von Neumann Architecture System for Accurate and Efficient Machine Learning Molecular Dynamics
    Zhao, Zhuoying
    Tan, Ziling
    Mo, Pinghui
    Wang, Xiaonan
    Zhao, Dan
    Zhang, Xin
    Tao, Ming
    Liu, Jie
    IEEE TRANSACTIONS ON CIRCUITS AND SYSTEMS I-REGULAR PAPERS, 2023, 70 (06) : 2439 - 2449
  • [2] The Heterogeneous Deep Neural Network Processor With a Non-von Neumann Architecture
    Shin, Dongjoo
    Yoo, Hoi-Jun
    PROCEEDINGS OF THE IEEE, 2020, 108 (08) : 1245 - 1260
  • [3] A Study on Teaching and Learning the von Neumann Machine in a 3D Learning Environment
    Cecilia, Maria Rosita
    De Gasperis, Giovanni
    METHODOLOGIES AND INTELLIGENT SYSTEMS FOR TECHNOLOGY ENHANCED LEARNING (MIS4TEL), 2016, 478 : 91 - 101
  • [4] IASSim: A Programmable Emulator for the Princeton IAS/Von Neumann Machine
    Fagin, Barry
    Skrien, Dale
    SIGCSE 11: PROCEEDINGS OF THE 42ND ACM TECHNICAL SYMPOSIUM ON COMPUTER SCIENCE EDUCATION, 2011, : 359 - 364
  • [5] Molecular Cellular Networks: A Non von Neumann Architecture for Molecular Electronics
    Lent, Craig S.
    Henderson, Kenneth W.
    Kandel, S. Alex
    Corcelli, Steven A.
    Snider, Gregory L.
    Orlov, Alexei O.
    Kogge, Peter M.
    Niemier, Michael T.
    Brown, Ryan C.
    Christie, John A.
    Wasio, Natalie A.
    Quardokus, Rebecca C.
    Forrest, Ryan P.
    Peterson, Jacob P.
    Silski, Angela
    Turner, David A.
    Blair, Enrique P.
    Lu, Yuhui
    2016 IEEE INTERNATIONAL CONFERENCE ON REBOOTING COMPUTING (ICRC), 2016,
  • [6] A Study on Decoupling Von Neumann Architecture from Rasterization in Neural Networks
    Bin, Miao
    2014 IEEE WORKSHOP ON ELECTRONICS, COMPUTER AND APPLICATIONS, 2014, : 307 - 309
  • [7] Leveraging million-scale Non von Neumann computations for accelerated Machine Learning and High Performance Computing
    Daudet, Laurent
    2021 PHOTONICS NORTH (PN), 2021,
  • [8] SEVERAL SPECIAL FUNCTIONS IN FRACTALS AND APPLICATIONS OF THE FRACTAL IN MACHINE LEARNING
    Wang, Jun
    Cao, Lei
    Chen, Xiliang
    Tang, Wei
    Xu, Zhixiong
    FRACTALS-COMPLEX GEOMETRY PATTERNS AND SCALING IN NATURE AND SOCIETY, 2022, 30 (01)
  • [9] Reliable Binarized Neural Networks on Unreliable Beyond Von-Neumann Architecture
    Yayla, Mikail
    Thomann, Simon
    Buschjaeger, Sebastian
    Morik, Katharina
    Chen, Jian-Jia
    Amrouch, Hussam
    IEEE TRANSACTIONS ON CIRCUITS AND SYSTEMS I-REGULAR PAPERS, 2022, 69 (06) : 2516 - 2528
  • [10] Learning Machine Learning as an Architect, How to? Presenting and evaluating a Grasshopper based platform to teach architecture students machine learning
    Khean, Nariddh
    Fabbri, Alessandra
    Haeusler, M. Hank
    ECAADE 2018: COMPUTING FOR A BETTER TOMORROW, VO 1, 2018, : 95 - 102