Exploiting Low-Rank Tensor-Train Deep Neural Networks Based on Riemannian Gradient Descent With Illustrations of Speech Processing

被引:8
|
作者
Qi, Jun [1 ,2 ]
Yang, Chao-Han Huck [2 ]
Chen, Pin-Yu [3 ]
Tejedor, Javier [4 ]
机构
[1] Fudan Univ, Sch Informat Sci & Engn, Dept Elect Engn, Shanghai 200438, Peoples R China
[2] Georgia Inst Technol, Sch Elect & Comp Engn, Atlanta, GA 30332 USA
[3] IBM Res, Yorktown Height, NY 10598 USA
[4] CEU Univ, Univ San Pablo CEU, Inst Technol, Boadilla Del Monte 28668, Spain
关键词
Tensor-train network; speech enhancement; spoken command recognition; Riemannian gradient descent; low-rank tensor-train decomposition; tensor-train deep neural network; MEAN ABSOLUTE ERROR; ALGORITHMS; RMSE; MAE;
D O I
10.1109/TASLP.2022.3231714
中图分类号
O42 [声学];
学科分类号
070206 ; 082403 ;
摘要
This work focuses on designing low-complexity hybrid tensor networks by considering trade-offs between the model complexity and practical performance. Firstly, we exploit a low-rank tensor-train deep neural network (TT-DNN) to build an end-to-end deep learning pipeline, namely LR-TT-DNN. Secondly, a hybrid model combining LR-TT-DNN with a convolutional neural network (CNN), which is denoted as CNN+(LR-TT-DNN), is set up to boost the performance. Instead of randomly assigning large TT-ranks for TT-DNN, we leverage Riemannian gradient descent to determine a TT-DNN associated with small TT-ranks. Furthermore, CNN+(LR-TT-DNN) consists of convolutional layers at the bottom for feature extraction and several TT layers at the top to solve regression and classification problems. We separately assess the LR-TT-DNN and CNN+(LR-TT-DNN) models on speech enhancement and spoken command recognition tasks. Our empirical evidence demonstrates that the LR-TT-DNN and CNN+(LR-TT-DNN) models with fewer model parameters can outperform the TT-DNN and CNN+(TT-DNN) counterparts.
引用
收藏
页码:633 / 642
页数:10
相关论文
共 50 条
  • [41] Semi-Orthogonal Low-Rank Matrix Factorization for Deep Neural Networks
    Povey, Daniel
    Cheng, Gaofeng
    Wang, Yiming
    Li, Ke
    Xu, Hainan
    Yarmohamadi, Mahsa
    Khudanpur, Sanjeev
    19TH ANNUAL CONFERENCE OF THE INTERNATIONAL SPEECH COMMUNICATION ASSOCIATION (INTERSPEECH 2018), VOLS 1-6: SPEECH RESEARCH FOR EMERGING MARKETS IN MULTILINGUAL SOCIETIES, 2018, : 3743 - 3747
  • [42] Big Data Matrix Singular Value Decomposition Based on Low-Rank Tensor Train Decomposition
    Lee, Namgil
    Cichocki, Andrzej
    ADVANCES IN NEURAL NETWORKS - ISNN 2014, 2014, 8866 : 121 - 130
  • [43] Support Vector Machine based on Low-rank Tensor Train Decomposition for Big Data Applications
    Wang, Yongkang
    Zhang, Weicheng
    Yu, Zhuliang
    Gu, Zhenghui
    Liu, Hai
    Cai, Zhaoquan
    Wang, Congjun
    Gao, Shihan
    PROCEEDINGS OF THE 2017 12TH IEEE CONFERENCE ON INDUSTRIAL ELECTRONICS AND APPLICATIONS (ICIEA), 2017, : 850 - 853
  • [44] L1-norm low-rank linear approximation for accelerating deep neural networks: L1-norm low-rank linear approximation for accelerating deep neural networks
    Zhao Z.
    Wang H.
    Sun H.
    He Z.
    He, Zhihai (hezhi@missouri.edu), 2020, Elsevier B.V., Netherlands (400) : 216 - 226
  • [45] Toward Near-Real-Time Training With Semi-Random Deep Neural Networks and Tensor-Train Decomposition
    Syed, Humza
    Bryla, Ryan
    Majumder, Uttam
    Kudithipudi, Dhireesha
    IEEE JOURNAL OF SELECTED TOPICS IN APPLIED EARTH OBSERVATIONS AND REMOTE SENSING, 2021, 14 (14) : 8171 - 8179
  • [46] Exploiting deep neural networks for detection-based speech recognition
    Siniscalchi, Sabato Marco
    Yu, Dong
    Deng, Li
    Lee, Chin-Hui
    NEUROCOMPUTING, 2013, 106 : 148 - 157
  • [47] Iterative deep neural networks based on proximal gradient descent for image restoration
    Lv, Ting
    Pan, Zhenkuan
    Wei, Weibo
    Yang, Guangyu
    Song, Jintao
    Wang, Xuqing
    Sun, Lu
    Li, Qian
    Sun, Xiatao
    PLOS ONE, 2022, 17 (11):
  • [48] Low-Rank Tensor Train Dynamic Mode Decomposition: An Enhanced Multivariate Signal Processing Method for Mechanical Fault Diagnosis
    Yin, Yongqiang
    Yuan, Rui
    Lv, Yong
    Wu, Hongan
    Li, Hewenxuan
    Zhu, Weihang
    IEEE Transactions on Instrumentation and Measurement, 2025, 74
  • [49] Deep Neural Network Acceleration Based on Low-Rank Approximated Channel Pruning
    Chen, Zhen
    Chen, Zhibo
    Lin, Jianxin
    Liu, Sen
    Li, Weiping
    IEEE TRANSACTIONS ON CIRCUITS AND SYSTEMS I-REGULAR PAPERS, 2020, 67 (04) : 1232 - 1244
  • [50] Conjugate Gradient and Variance Reduction Based Online ADMM for Low-Rank Distributed Networks
    Chen, Yitong
    Jin, Danqi
    Chen, Jie
    Richard, Cedric
    Zhang, Wen
    IEEE SIGNAL PROCESSING LETTERS, 2025, 32 : 706 - 710