Multi-phase-quantization optimizer and its architecture for edge AI training

被引:0
|
作者
Akeno, Itsuki [1 ]
Yamazaki, Hiiro [1 ]
Asai, Tetsuya [2 ]
Ando, Kota [2 ]
机构
[1] Hokkaido Univ, Grad Sch Informat Sci & Technol, M BLDG 2F,Kita 14,Nishi 9,Kita Ku, Sapporo, Hokkaido 0600814, Japan
[2] Hokkaido Univ, Fac Informat Sci & Technol, M BLDG 2F,Kita 14,Nishi 9,Kita Ku, Sapporo, Hokkaido 0600814, Japan
来源
IEICE NONLINEAR THEORY AND ITS APPLICATIONS | 2025年 / 16卷 / 01期
关键词
artificial intelligence; edge AI; hardware architecture; neural network; optimizer; INTELLIGENCE;
D O I
10.1587/nolta.16.43
中图分类号
O1 [数学];
学科分类号
0701 ; 070101 ;
摘要
Developing hardware for Artificial intelligence (AI) training is vital. A hardwareoriented optimizer, named Holmes enables faster training with a smaller memory footprint. This study developed a hardware architecture that incorporates Holmes and benefits from parallelization and pipelining to achieve significant throughput improvement. We determined the required bit width for training and used it the architecture evaluation. We investigated scalability and the effectiveness of both Holmes and pipelining. The results proved the linear scalability of the memory footprint over the model size, reduction of the memory footprint by utilizing Holmes, drastic increase in throughput by pipelining and faster computing.
引用
收藏
页码:43 / 63
页数:21
相关论文
共 2 条
  • [1] Proposal and evaluation of recurrent neural network training by multi-phase quantization optimizer
    Yamazaki, Hiiro
    Akeno, Itsuki
    Nobori, Koki
    Asai, Tetsuya
    Ando, Kota
    IEICE NONLINEAR THEORY AND ITS APPLICATIONS, 2025, 16 (01): : 30 - 42
  • [2] MLFlash-CIM: Embedded Multi-Level NOR-Flash Cell based Computing in Memory Architecture for Edge AI Devices
    Zeng, Sitao
    Zhang, Yuxin
    Zhu, Zhiguo
    Qin, Zhaolong
    Dou, Chunmeng
    Si, Xin
    Li, Qiang
    2021 IEEE 3RD INTERNATIONAL CONFERENCE ON ARTIFICIAL INTELLIGENCE CIRCUITS AND SYSTEMS (AICAS), 2021,