A High-Performance and Energy-Efficient Photonic Architecture for Multi-DNN Acceleration

被引:0
|
作者
Li, Yuan [1 ]
Louri, Ahmed [1 ]
Karanth, Avinash [2 ]
机构
[1] George Washington Univ, Dept Elect & Comp Engn, Washington, DC 20052 USA
[2] Ohio Univ, Sch Elect Engn & Comp Sci, Athens, OH 45701 USA
基金
美国国家科学基金会;
关键词
Accelerator; dataflow; deep neural network; silicon photonics;
D O I
10.1109/TPDS.2023.3327535
中图分类号
TP301 [理论、方法];
学科分类号
081202 ;
摘要
Large-scale deep neural network (DNN) accelerators are poised to facilitate the concurrent processing of diverse DNNs, imposing demanding challenges on the interconnection fabric. These challenges encompass overcoming performance degradation and energy increase associated with system scaling while also necessitating flexibility to support dynamic partitioning and adaptable organization of compute resources. Nevertheless, conventional metallic-based interconnects frequently confront inherent limitations in scalability and flexibility. In this paper, we leverage silicon photonic interconnects and adopt an algorithm-architecture co-design approach to develop MDA, a DNN accelerator meticulously crafted to empower high-performance and energy-efficient concurrent processing of diverse DNNs. Specifically, MDA consists of three novel components: 1) a resource allocation algorithm that assigns compute resources to concurrent DNNs based on their computational demands and priorities; 2) a dataflow selection algorithm that determines off-chip and on-chip dataflows for each DNN, with the objectives of minimizing off-chip and on-chip memory accesses, respectively; 3) a flexible silicon photonic network that can be dynamically segmented into sub-networks, each interconnecting the assigned compute resources of a certain DNN while adapting to the communication patterns dictated by the selected on-chip dataflow. Simulation results show that the proposed MDA accelerator outperforms other state-of-the-art multi-DNN accelerators, including PREMA, AI-MT, Planaria, and HDA. MDA accelerator achieves a speedup of 3.6, accompanied by substantial improvements of 7.3x, 12.7x, and 9.2x in energy efficiency, service-level agreement (SLA) satisfaction rate, and fairness, respectively.
引用
收藏
页码:46 / 58
页数:13
相关论文
共 50 条
  • [21] High-Performance Photonic Integrated Circuits on Silicon
    Helkey, Roger
    Saleh, Adel A. M.
    Buckwalter, Jim
    Bowers, John E.
    IEEE JOURNAL OF SELECTED TOPICS IN QUANTUM ELECTRONICS, 2019, 25 (05)
  • [22] Photonic Architectures for High-Performance Data Centers
    Beausoleil, Raymond G.
    McLaren, Moray
    Jouppi, Norman P.
    IEEE JOURNAL OF SELECTED TOPICS IN QUANTUM ELECTRONICS, 2013, 19 (02)
  • [23] Hierarchical DNN with Heterogeneous Computing Enabled High-Performance DNA Sequencing
    Luo, Shaobo
    Xie, Zhiyuan
    Chen, Gengxin
    Cui, Lei
    Yan, Mei
    Huang, Xiwei
    Lie, Shuwei
    Man, Changhai
    Mao, Wei
    Yu, Hao
    2022 IEEE ASIA PACIFIC CONFERENCE ON CIRCUITS AND SYSTEMS, APCCAS, 2022, : 35 - 40
  • [24] Albireo: Energy-Efficient Acceleration of Convolutional Neural Networks via Silicon Photonics
    Shiflett, Kyle
    Karanth, Avinash
    Bunescu, Razvan
    Louri, Ahmed
    2021 ACM/IEEE 48TH ANNUAL INTERNATIONAL SYMPOSIUM ON COMPUTER ARCHITECTURE (ISCA 2021), 2021, : 860 - 873
  • [25] A Flexible and Energy-Efficient Convolutional Neural Network Acceleration With Dedicated ISA and Accelerator
    Chen, Xiaobai
    Yu, Zhiyi
    IEEE TRANSACTIONS ON VERY LARGE SCALE INTEGRATION (VLSI) SYSTEMS, 2018, 26 (07) : 1408 - 1412
  • [26] An Energy-Efficient Optical Interconnect Architecture for Bandwidth-Balanced Systems
    Schwetman, Herb
    Patel, Avadh
    Robinson, Leick
    Zheng, Xuezhe
    Wood, Alan
    Krishnamoorthy, Ashok V.
    JOURNAL OF LIGHTWAVE TECHNOLOGY, 2016, 34 (12) : 2905 - 2919
  • [27] AIX: A high performance and energy efficient inference accelerator on FPGA for a DNN-based commercial speech recognition
    Ahn, Minwook
    Hwang, Seok Joong
    Kim, Wonsub
    Jung, Seungrok
    Lee, Yeonbok
    Chung, Mookyoung
    Lim, Woohyung
    Kim, Youngjoon
    2019 DESIGN, AUTOMATION & TEST IN EUROPE CONFERENCE & EXHIBITION (DATE), 2019, : 1495 - 1500
  • [28] Design of Processing-in-Memory With Triple Computational Path and Sparsity Handling for Energy-Efficient DNN Training
    Han, Wontak
    Heo, Jaehoon
    Kim, Junsoo
    Lim, Sukbin
    Kim, Joo-Young
    IEEE JOURNAL ON EMERGING AND SELECTED TOPICS IN CIRCUITS AND SYSTEMS, 2022, 12 (02) : 354 - 366
  • [29] Energy-efficient Content-aware DNN Inference for Mobile Video via Deep Reinforcement Learning
    Guo, Guangfeng
    Zhang, Junxing
    IEEE INTERNATIONAL CONFERENCE ON COMMUNICATIONS (ICC 2022), 2022, : 763 - 768
  • [30] Energy-efficient computing-in-memory architecture for AI processor: device, circuit, architecture perspective
    Chang, Liang
    Li, Chenglong
    Zhang, Zhaomin
    Xiao, Jianbiao
    Liu, Qingsong
    Zhu, Zhen
    Li, Weihang
    Zhu, Zixuan
    Yang, Siqi
    Zhou, Jun
    SCIENCE CHINA-INFORMATION SCIENCES, 2021, 64 (06)