共 50 条
- [21] Optimization of General Matrix Multiply Library for Ternary Weight for Fast DNN Inference Journal of Signal Processing Systems, 2022, 94 : 929 - 943
- [22] The Perfect Match: Selecting Approximate Multipliers for Energy-Efficient Neural Network Inference 2023 IEEE 24TH INTERNATIONAL CONFERENCE ON HIGH PERFORMANCE SWITCHING AND ROUTING, HPSR, 2023,
- [23] Fast and fair split computing for accelerating deep neural network (DNN) inference ICT EXPRESS, 2025, 11 (01): : 47 - 52
- [24] TernGEMM: GEneral Matrix Multiply Library with Ternary Weights for Fast DNN Inference 2021 IEEE WORKSHOP ON SIGNAL PROCESSING SYSTEMS (SIPS 2021), 2021, : 111 - 116
- [25] Optimization of General Matrix Multiply Library for Ternary Weight for Fast DNN Inference JOURNAL OF SIGNAL PROCESSING SYSTEMS FOR SIGNAL IMAGE AND VIDEO TECHNOLOGY, 2022, 94 (10): : 929 - 943
- [26] AIN: Fast and Accurate Sequence Labeling with Approximate Inference Network PROCEEDINGS OF THE 2020 CONFERENCE ON EMPIRICAL METHODS IN NATURAL LANGUAGE PROCESSING (EMNLP), 2020, : 6019 - 6026
- [28] TLED: Training-Based Approximate Layer Exploration in DNNs with Efficient Multipliers 2024 INTERNATIONAL SYMPOSIUM OF ELECTRONICS DESIGN AUTOMATION, ISEDA 2024, 2024, : 247 - 252
- [29] STONNE: Enabling Cycle-Level Microarchitectural Simulation for DNN Inference Accelerators 2021 IEEE INTERNATIONAL SYMPOSIUM ON WORKLOAD CHARACTERIZATION (IISWC 2021), 2021, : 201 - 213
- [30] Towards Fast GPU-based Sparse DNN Inference: A Hybrid Compute Model 2022 IEEE HIGH PERFORMANCE EXTREME COMPUTING VIRTUAL CONFERENCE (HPEC), 2022,