A High-Performance and Energy-Efficient Photonic Architecture for Multi-DNN Acceleration

被引:0
|
作者
Li, Yuan [1 ]
Louri, Ahmed [1 ]
Karanth, Avinash [2 ]
机构
[1] George Washington Univ, Dept Elect & Comp Engn, Washington, DC 20052 USA
[2] Ohio Univ, Sch Elect Engn & Comp Sci, Athens, OH 45701 USA
基金
美国国家科学基金会;
关键词
Accelerator; dataflow; deep neural network; silicon photonics;
D O I
10.1109/TPDS.2023.3327535
中图分类号
TP301 [理论、方法];
学科分类号
081202 ;
摘要
Large-scale deep neural network (DNN) accelerators are poised to facilitate the concurrent processing of diverse DNNs, imposing demanding challenges on the interconnection fabric. These challenges encompass overcoming performance degradation and energy increase associated with system scaling while also necessitating flexibility to support dynamic partitioning and adaptable organization of compute resources. Nevertheless, conventional metallic-based interconnects frequently confront inherent limitations in scalability and flexibility. In this paper, we leverage silicon photonic interconnects and adopt an algorithm-architecture co-design approach to develop MDA, a DNN accelerator meticulously crafted to empower high-performance and energy-efficient concurrent processing of diverse DNNs. Specifically, MDA consists of three novel components: 1) a resource allocation algorithm that assigns compute resources to concurrent DNNs based on their computational demands and priorities; 2) a dataflow selection algorithm that determines off-chip and on-chip dataflows for each DNN, with the objectives of minimizing off-chip and on-chip memory accesses, respectively; 3) a flexible silicon photonic network that can be dynamically segmented into sub-networks, each interconnecting the assigned compute resources of a certain DNN while adapting to the communication patterns dictated by the selected on-chip dataflow. Simulation results show that the proposed MDA accelerator outperforms other state-of-the-art multi-DNN accelerators, including PREMA, AI-MT, Planaria, and HDA. MDA accelerator achieves a speedup of 3.6, accompanied by substantial improvements of 7.3x, 12.7x, and 9.2x in energy efficiency, service-level agreement (SLA) satisfaction rate, and fairness, respectively.
引用
收藏
页码:46 / 58
页数:13
相关论文
共 50 条
  • [1] A Silicon Photonic Multi-DNN Accelerator
    Li, Yuan
    Louri, Ahmed
    Karanth, Avinash
    2023 32ND INTERNATIONAL CONFERENCE ON PARALLEL ARCHITECTURES AND COMPILATION TECHNIQUES, PACT, 2023, : 238 - 249
  • [2] Serving Multi-DNN Workloads on FPGAs: A Coordinated Architecture, Scheduling, and Mapping Perspective
    Zeng, Shulin
    Dai, Guohao
    Zhang, Niansong
    Yang, Xinhao
    Zhang, Haoyu
    Zhu, Zhenhua
    Yang, Huazhong
    Wang, Yu
    IEEE TRANSACTIONS ON COMPUTERS, 2023, 72 (05) : 1314 - 1328
  • [3] Energy-Efficient High-Performance Photonic Backplane Network for Rack-Scale Computing Systems
    Feng, Jun
    Chen, Shixi
    Zhang, Jiaxu
    Fu, Yuxiang
    Xu, Jiang
    2022 IEEE COMPUTER SOCIETY ANNUAL SYMPOSIUM ON VLSI (ISVLSI 2022), 2022, : 122 - 127
  • [4] MuDBN: An Energy-Efficient and High-Performance Multi-FPGA Accelerator for Deep Belief Networks
    Cheng, Yuming
    Wang, Chao
    Zhao, Yangyang
    Chen, Xianglan
    Zhou, Xuehai
    Li, Xi
    PROCEEDINGS OF THE 2018 GREAT LAKES SYMPOSIUM ON VLSI (GLSVLSI'18), 2018, : 435 - 438
  • [5] SPRINT: A High-Performance, Energy-Efficient, and Scalable Chiplet-Based Accelerator With Photonic Interconnects for CNN Inference
    Li, Yuan
    Louri, Ahmed
    Karanth, Avinash
    IEEE TRANSACTIONS ON PARALLEL AND DISTRIBUTED SYSTEMS, 2022, 33 (10) : 2332 - 2345
  • [6] MAICC : A Lightweight Many-core Architecture with In-Cache Computing for Multi-DNN Parallel Inference
    Fan, Renhao
    Cui, Yikai
    Chen, Qilin
    Wang, Mingyu
    Zhang, Youhui
    Zheng, Weimin
    Li, Zhaolin
    56TH IEEE/ACM INTERNATIONAL SYMPOSIUM ON MICROARCHITECTURE, MICRO 2023, 2023, : 411 - 423
  • [7] Ascend: A Scalable and Energy-Efficient Deep Neural Network Accelerator With Photonic Interconnects
    Li, Yuan
    Wang, Ke
    Zheng, Hao
    Louri, Ahmed
    Karanth, Avinash
    IEEE TRANSACTIONS ON CIRCUITS AND SYSTEMS I-REGULAR PAPERS, 2022, 69 (07) : 2730 - 2741
  • [8] A 5.99 TFLOPS/W Heterogeneous CIM-NPU Architecture for an Energy Efficient Floating-Point DNN Acceleration
    Park, Wonhoon
    Ryu, Junha
    Kim, Sangjin
    Um, Soyeon
    Jo, Wooyoung
    Kim, Sangyoeb
    Yoo, Hoi-Jun
    2023 IEEE INTERNATIONAL SYMPOSIUM ON CIRCUITS AND SYSTEMS, ISCAS, 2023,
  • [9] Energy-Efficient DNN Training Processorson Micro-AI Systems
    Han, Donghyeon
    Kang, Sanghoon
    Kim, Sangyeob
    Lee, Juhyoung
    Yoo, Hoi-Jun
    IEEE OPEN JOURNAL OF THE SOLID-STATE CIRCUITS SOCIETY, 2022, 2 : 259 - 275
  • [10] An Embedded Architecture for Energy-Efficient Stream Computing
    Panda, Amrit
    Chatha, Karam S.
    IEEE EMBEDDED SYSTEMS LETTERS, 2014, 6 (03) : 57 - 60