Sustainable AI Processing at the Edge

被引:3
作者
Ollivier, Sebastien [1 ]
Li, Sheng [2 ]
Tang, Yue [3 ]
Cahoon, Stephen [4 ]
Caginalp, Ryan [3 ]
Chaudhuri, Chayanika [3 ]
Zhou, Peipei [5 ]
Tang, Xulong [2 ]
Hu, Jingtong [5 ]
Jones, Alex K. [6 ]
机构
[1] Univ Pittsburgh, Pittsburgh, PA 15260 USA
[2] Univ Pittsburgh, Dept Comp Sci, Pittsburgh, PA 15260 USA
[3] Univ Pittsburgh, Elect & Comp Engn, Pittsburgh, PA 15260 USA
[4] Univ Pittsburgh, Comp Engn, Pittsburgh, PA 15260 USA
[5] Univ Pittsburgh, Dept Elect & Comp Engn, Pittsburgh, PA 15260 USA
[6] Univ Pittsburgh, Elect & Comp Engn & Comp Sci, Pittsburgh, PA 15260 USA
基金
美国国家科学基金会;
关键词
Artificial intelligence; Edge computing; Memory management; Sustainable development; Measurement; Fabrication; Costs; PERFORMANCE; ENERGY;
D O I
10.1109/MM.2022.3220399
中图分类号
TP3 [计算技术、计算机技术];
学科分类号
0812 ;
摘要
Edge computing is a popular paradigm for accelerating light- to medium-weight machine learning algorithms initiated from mobile devices without requiring the long communication latencies to send them to remote datacenters in the cloud. Edge servers primarily consider traditional concerns, such as size, weight, and power constraints for their installations. However, such metrics are not entirely sufficient to consider environmental impacts from computing given the significant contributions from embodied energy and carbon. In this article we explore the tradeoffs of hardware strategies for convolutional neural network acceleration engines considering inference and online training. In particular, we explore the use of mobile graphics processing unit (GPU) accelerators, recently released edge-class field-programmable gate arrays, and novel processing in memory (PIM) using dynamic random-access memory (DRAM) and emerging Racetrack memory. Given edge servers already employ DRAM and sometimes GPU accelerators, we consider the sustainability implications using breakeven analysis of replacing or augmenting DDR3 with Racetrack memory. We also consider the implications for provisioning edge servers with different accelerators using indifference analysis. While mobile GPUs are typically much more energy efficient, their significant embodied energy can make them less sustainable than PIM solutions in certain scenarios that consider activity time and compute effort.
引用
收藏
页码:19 / 28
页数:10
相关论文
共 23 条
  • [11] GreenChip: A tool for evaluating holistic sustainability of modern computing systems
    Kline, Donald, Jr.
    Parshook, Nikolas
    Ge, Xiaoyu
    Brunvand, Erik
    Melhem, Rami
    Chrysanthis, Panos K.
    Jones, Alex K.
    [J]. SUSTAINABLE COMPUTING-INFORMATICS & SYSTEMS, 2019, 22 : 322 - 332
  • [12] Mai T., NRELTP6A20524091
  • [13] Ollivier S., IEEE ACM T NETWORK, V42, P9
  • [14] CORUSCANT: Fast Efficient Processing-in-Racetrack Memories
    Ollivier, Sebastien
    Longofono, Stephen
    Dutta, Prayash
    Hu, Jingtong
    Bhanja, Sanjukta
    Jones, Alex K.
    [J]. 2022 55TH ANNUAL IEEE/ACM INTERNATIONAL SYMPOSIUM ON MICROARCHITECTURE (MICRO), 2022, : 784 - 798
  • [15] Popovich N., 2020, NY TIMES OCT
  • [16] Ambit: In-Memory Accelerator for Bulk Bitwise Operations Using Commodity DRAM Technology
    Seshadri, Vivek
    Lee, Donghyuk
    Mullins, Thomas
    Hassan, Hasan
    Boroumand, Amirali
    Kim, Jeremie
    Kozuch, Michael A.
    Mutlu, Onur
    Gibbons, Phillip B.
    Mowry, Todd C.
    [J]. 50TH ANNUAL IEEE/ACM INTERNATIONAL SYMPOSIUM ON MICROARCHITECTURE (MICRO), 2017, : 273 - 287
  • [17] EF-Train: Enable Efficient On-device CNN Training on FPGA through Data Reshaping for Online Adaptation or Personalization
    Tang, Yue
    Zhang, Xinyi
    Zhou, Peipei
    Hu, Jingtong
    [J]. ACM TRANSACTIONS ON DESIGN AUTOMATION OF ELECTRONIC SYSTEMS, 2022, 27 (05)
  • [18] Venkatesan R., 2012, ACMIEEE INT S LOW PO, P185, DOI DOI 10.1145/2333660.2333707
  • [19] Venkatesan R, 2013, DES AUT TEST EUROPE, P1825
  • [20] Opportunities for Nonvolatile Memory Systems in Extreme-Scale High-Performance Computing
    Vetter, Jeffrey S.
    Mittal, Sparsh
    [J]. COMPUTING IN SCIENCE & ENGINEERING, 2015, 17 (02) : 73 - 82