Low-rank lottery tickets: finding efficient low-rank neural networks via matrix differential equations

被引:0
|
作者
Schotthoefer, Steffen [1 ]
Zangrando, Emanuele [2 ]
Kusch, Jonas [3 ]
Ceruti, Gianluca [4 ]
Tudisco, Francesco [2 ]
机构
[1] Karlsruhe Inst Technol, D-76131 Karlsruhe, Germany
[2] Gran Sasso Sci Inst, I-67100 Laquila, Italy
[3] Univ Innsbruck, A-6020 Innsbruck, Austria
[4] EPF Lausanne, CH-1015 Lausanne, Switzerland
来源
ADVANCES IN NEURAL INFORMATION PROCESSING SYSTEMS 35 (NEURIPS 2022) | 2022年
关键词
TIME INTEGRATION; APPROXIMATION; TUCKER;
D O I
暂无
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
Neural networks have achieved tremendous success in a large variety of applications. However, their memory footprint and computational demand can render them impractical in application settings with limited hardware or energy resources. In this work, we propose a novel algorithm to find efficient low-rank subnetworks. Remarkably, these subnetworks are determined and adapted already during the training phase and the overall time and memory resources required by both training and evaluating them are significantly reduced. The main idea is to restrict the weight matrices to a low-rank manifold and to update the low-rank factors rather than the full matrix during training. To derive training updates that are restricted to the prescribed manifold, we employ techniques from dynamic model order reduction for matrix differential equations. This allows us to provide approximation, stability, and descent guarantees. Moreover, our method automatically and dynamically adapts the ranks during training to achieve the desired approximation accuracy. The efficiency of the proposed method is demonstrated through a variety of numerical experiments on fully-connected and convolutional networks.
引用
收藏
页数:13
相关论文
共 50 条
  • [1] Numerical low-rank approximation of matrix differential equations
    Mena, Hermann
    Ostermann, Alexander
    Pfurtscheller, Lena-Maria
    Piazzola, Chiara
    JOURNAL OF COMPUTATIONAL AND APPLIED MATHEMATICS, 2018, 340 : 602 - 614
  • [2] Finding a low-rank basis in a matrix subspace
    Nakatsukasa, Yuji
    Soma, Tasuku
    Uschmajew, Andre
    MATHEMATICAL PROGRAMMING, 2017, 162 (1-2) : 325 - 361
  • [3] Finding a low-rank basis in a matrix subspace
    Yuji Nakatsukasa
    Tasuku Soma
    André Uschmajew
    Mathematical Programming, 2017, 162 : 325 - 361
  • [4] Low-rank passthrough neural networks
    Barone, Antonio Valerio Miceli
    DEEP LEARNING APPROACHES FOR LOW-RESOURCE NATURAL LANGUAGE PROCESSING (DEEPLO), 2018, : 77 - 86
  • [5] Low-Rank Matrix Recovery via Continuation-Based Approximate Low-Rank Minimization
    Zhang, Xiang
    Gao, Yongqiang
    Lan, Long
    Guo, Xiaowei
    Huang, Xuhui
    Luo, Zhigang
    PRICAI 2018: TRENDS IN ARTIFICIAL INTELLIGENCE, PT I, 2018, 11012 : 559 - 573
  • [6] Low-rank solvers for fractional differential equations
    Breiten, Tobias
    Simoncini, Valeria
    Stoll, Martin
    Electronic Transactions on Numerical Analysis, 2016, 45 : 107 - 132
  • [7] LOW-RANK SOLVERS FOR FRACTIONAL DIFFERENTIAL EQUATIONS
    Breiten, Tobias
    Simoncini, Valeria
    Stoll, Martin
    ELECTRONIC TRANSACTIONS ON NUMERICAL ANALYSIS, 2016, 45 : 107 - 132
  • [8] Efficient Map Prediction via Low-Rank Matrix Completion
    Chen, Zheng
    Bai, Shi
    Liu, Lantao
    2021 IEEE INTERNATIONAL CONFERENCE ON ROBOTICS AND AUTOMATION (ICRA 2021), 2021, : 13953 - 13959
  • [9] Localization of IoT Networks via Low-Rank Matrix Completion
    Luong Trung Nguyen
    Kim, Junhan
    Kim, Sangtae
    Shim, Byonghyo
    IEEE TRANSACTIONS ON COMMUNICATIONS, 2019, 67 (08) : 5833 - 5847
  • [10] Low-Rank Extragradient Method for Nonsmooth and Low-Rank Matrix Optimization Problems
    Garber, Dan
    Kaplan, Atara
    ADVANCES IN NEURAL INFORMATION PROCESSING SYSTEMS 34 (NEURIPS 2021), 2021, 34