共 23 条
[1]
CHEN Y H, KRISHNA T, EMER J S, Et al., Eyeriss: an energy-efficient reconfigurable accelerator for deep convolutional neural networks, IEEE Journal of Solid-State Circuits, 52, pp. 127-138, (2017)
[2]
JOUPPI N P, YOUNG C, PATIL N, Et al., In-datacenter performance analysis of a tensor processing unit, Proceedings of the 44th Annual International Symposium on Computer Architecture, pp. 1-12, (2017)
[3]
TALPES E, GORTI A, SACHDEV G S, Et al., Compute solution for Tesla’ s full self-driving computer, IEEE Micro, 40, pp. 25-35, (2020)
[4]
SHARMA H, PARK J, MAHAJAN D, Et al., From high-level deep neural models to FPGAs, The 49th Annual IEEE/ ACM International Symposium on Microarchitecture, 17, pp. 1-17:12, (2016)
[5]
ZHANG X, WANG J, ZHU C, Et al., DNNBuilder: an automated tool for building high-performance DNN hardware accelerators for FPGAs, Proceedings of the International Conference on Computer-Aided Design, (2018)
[6]
XU P, ZHANG X, HAO C, Et al., AutoDNNchip: an automated DNN chip predictor and builder for both FPGAs and ASICs, The 2020 ACM/ SIGDA International Symposium on Field-Programmable Gate Arrays, pp. 40-50, (2020)
[7]
KAO S, JEONG G, KRISHNA T., ConfuciuX: autonomous hardware resource assignment for DNN accelerators using reinforcement learning, The 53rd Annual IEEE/ ACM International Symposium on Microarchitecture, pp. 622-636, (2020)
[8]
YANG X, GAO M, LIU Q, Et al., Interstellar: using halide’ s scheduling language to analyze DNN accelerators [ C], Architectural Support for Programming Languages and Operating Systems, pp. 369-383, (2020)
[9]
XI S L, YAO Y, BHARDWAJ K, Et al., SMAUG: end-to-end full-stack simulation infrastructure for deep learning workloads, ACM Transactions on Architecture and Code Optimization, 17, pp. 1-26, (2020)
[10]
WU Y N, EMER J S, SZE V., Accelergy: an architecture-level energy estimation methodology for accelerator designs, Proceedings of the International Conference on Computer-Aided Design, pp. 1-8, (2019)