Exploiting Low-Rank Tensor-Train Deep Neural Networks Based on Riemannian Gradient Descent With Illustrations of Speech Processing

被引:8
|
作者
Qi, Jun [1 ,2 ]
Yang, Chao-Han Huck [2 ]
Chen, Pin-Yu [3 ]
Tejedor, Javier [4 ]
机构
[1] Fudan Univ, Sch Informat Sci & Engn, Dept Elect Engn, Shanghai 200438, Peoples R China
[2] Georgia Inst Technol, Sch Elect & Comp Engn, Atlanta, GA 30332 USA
[3] IBM Res, Yorktown Height, NY 10598 USA
[4] CEU Univ, Univ San Pablo CEU, Inst Technol, Boadilla Del Monte 28668, Spain
关键词
Tensor-train network; speech enhancement; spoken command recognition; Riemannian gradient descent; low-rank tensor-train decomposition; tensor-train deep neural network; MEAN ABSOLUTE ERROR; ALGORITHMS; RMSE; MAE;
D O I
10.1109/TASLP.2022.3231714
中图分类号
O42 [声学];
学科分类号
070206 ; 082403 ;
摘要
This work focuses on designing low-complexity hybrid tensor networks by considering trade-offs between the model complexity and practical performance. Firstly, we exploit a low-rank tensor-train deep neural network (TT-DNN) to build an end-to-end deep learning pipeline, namely LR-TT-DNN. Secondly, a hybrid model combining LR-TT-DNN with a convolutional neural network (CNN), which is denoted as CNN+(LR-TT-DNN), is set up to boost the performance. Instead of randomly assigning large TT-ranks for TT-DNN, we leverage Riemannian gradient descent to determine a TT-DNN associated with small TT-ranks. Furthermore, CNN+(LR-TT-DNN) consists of convolutional layers at the bottom for feature extraction and several TT layers at the top to solve regression and classification problems. We separately assess the LR-TT-DNN and CNN+(LR-TT-DNN) models on speech enhancement and spoken command recognition tasks. Our empirical evidence demonstrates that the LR-TT-DNN and CNN+(LR-TT-DNN) models with fewer model parameters can outperform the TT-DNN and CNN+(TT-DNN) counterparts.
引用
收藏
页码:633 / 642
页数:10
相关论文
共 50 条
  • [31] Low-rank tensor completion based on tensor train rank with partially overlapped sub-blocks and total variation
    He, Jingfei
    Yang, Zezhong
    Zheng, Xunan
    Zhang, Xiaoyue
    Li, Ao
    SIGNAL PROCESSING-IMAGE COMMUNICATION, 2024, 129
  • [32] Recurrent Neural Network Compression Based on Low-Rank Tensor Representation
    Tjandra, Andros
    Sakti, Sakriani
    Nakamura, Satoshi
    IEICE TRANSACTIONS ON INFORMATION AND SYSTEMS, 2020, E103D (02) : 435 - 449
  • [33] Low-Rank Gradient Descent for Memory-Efficient Training of Deep In-Memory Arrays
    Huang, Siyuan
    Hoskins, Brian D.
    Daniels, Matthew W.
    Stiles, Mark D.
    Adam, Gina C.
    ACM JOURNAL ON EMERGING TECHNOLOGIES IN COMPUTING SYSTEMS, 2023, 19 (02)
  • [34] An Analysis of Low-Rank Decomposition Selection for Deep Convolutional Neural Networks
    Liu, Baichen
    Jia, Huidi
    Han, Zhi
    Chen, Xi'ai
    Tang, Yandong
    INTELLIGENT ROBOTICS AND APPLICATIONS (ICIRA 2022), PT II, 2022, 13456 : 480 - 490
  • [35] Low-Rank Tensor Train and Self-Similarity Based Spectral CT Reconstruction
    Guo, Jie
    Yu, Xiaohuan
    Wang, Shaoyu
    Cai, Ailong
    Zheng, Zhizhong
    Liang, Ningning
    Li, Lei
    Yan, Bin
    IEEE ACCESS, 2024, 12 : 56368 - 56382
  • [36] Performance Analysis for Tensor-Train Decomposition to Deep Neural Network Based Vector-to-Vector Regression
    Qi, Jun
    Ma, Xiaoli
    Lee, Chin-Hui
    Du, Jun
    Siniscalchi, Sabato Marco
    2020 54TH ANNUAL CONFERENCE ON INFORMATION SCIENCES AND SYSTEMS (CISS), 2020, : 7 - 12
  • [37] Transformed Low-Rank Parameterization Can Help Robust Generalization for Tensor Neural Networks
    Wang, Andong
    Li, Chao
    Bai, Mingyuan
    Jin, Zhong
    Zhou, Guoxu
    Zhao, Qibin
    ADVANCES IN NEURAL INFORMATION PROCESSING SYSTEMS 36 (NEURIPS 2023), 2023,
  • [38] QUANTIZATION AND APPLICATION OF LOW-RANK TENSOR DECOMPOSITION BASED ON THE DEEP LEARNING MODEL
    Zhao, Jia
    3C TIC, 2023, 12 (01): : 330 - 350
  • [39] Fast nonnegative tensor factorization based on accelerated proximal gradient and low-rank approximation
    Zhang, Yu
    Zhou, Guoxu
    Zhao, Qibin
    Cichocki, Andrzej
    Wang, Xingyu
    NEUROCOMPUTING, 2016, 198 : 148 - 154
  • [40] EXTENDED LOW-RANK PLUS DIAGONAL ADAPTATION FOR DEEP AND RECURRENT NEURAL NETWORKS
    Zhao, Yong
    Li, Jinyu
    Kumar, Kshitiz
    Gong, Yifan
    2017 IEEE INTERNATIONAL CONFERENCE ON ACOUSTICS, SPEECH AND SIGNAL PROCESSING (ICASSP), 2017, : 5040 - 5044