A tensor compression algorithm using Tucker decomposition and dictionary dimensionality reduction

被引:2
作者
Gan, Chenquan [1 ,2 ,3 ]
Mao, Junwei [1 ,2 ,3 ]
Zhang, Zufan [1 ,2 ,3 ]
Zhu, Qingyi [4 ]
机构
[1] Chongqing Univ Posts & Telecommun, Sch Commun & Informat Engn, Chongqing 400065, Peoples R China
[2] Chongqing Key Lab Mobile Commun Technol, Chongqing, Peoples R China
[3] Minist Educ, Engn Res Ctr Mobile Commun, Chongqing, Peoples R China
[4] Chongqing Univ Posts & Telecommun, Sch Cyber Secur & Informat Law, Chongqing, Peoples R China
来源
INTERNATIONAL JOURNAL OF DISTRIBUTED SENSOR NETWORKS | 2020年 / 16卷 / 04期
关键词
Tensor signal compression; Tucker decomposition; sparse representation; dictionary learning; denoising ability;
D O I
10.1177/1550147720916408
中图分类号
TP [自动化技术、计算机技术];
学科分类号
0812 ;
摘要
Tensor compression algorithms play an important role in the processing of multidimensional signals. In previous work, tensor data structures are usually destroyed by vectorization operations, resulting in information loss and new noise. To this end, this article proposes a tensor compression algorithm using Tucker decomposition and dictionary dimensionality reduction, which mainly includes three parts: tensor dictionary representation, dictionary preprocessing, and dictionary update. Specifically, the tensor is respectively performed by the sparse representation and Tucker decomposition, from which one can obtain the dictionary, sparse coefficient, and core tensor. Furthermore, the sparse representation can be obtained through the relationship between sparse coefficient and core tensor. In addition, the dimensionality of the input tensor is reduced by using the concentrated dictionary learning. Finally, some experiments show that, compared with other algorithms, the proposed algorithm has obvious advantages in preserving the original data information and denoising ability.
引用
收藏
页数:10
相关论文
共 50 条
  • [21] GRAPH REGULARIZED NONNEGATIVE TUCKER DECOMPOSITION FOR TENSOR DATA REPRESENTATION
    Qiu, Yuning
    Zhou, Guoxu
    Zhang, Yu
    Xie, Shengli
    2019 IEEE INTERNATIONAL CONFERENCE ON ACOUSTICS, SPEECH AND SIGNAL PROCESSING (ICASSP), 2019, : 8613 - 8617
  • [22] Rank-Adaptive Tensor Completion Based on Tucker Decomposition
    Liu, Siqi
    Shi, Xiaoyu
    Liao, Qifeng
    ENTROPY, 2023, 25 (02)
  • [23] An incremental tensor tucker decomposition method based on model characteristics
    Qu C.-Y.
    Han J.-J.
    Kongzhi yu Juece/Control and Decision, 2024, 39 (07): : 2431 - 2437
  • [24] Personalized Tucker Decomposition: Modeling Commonality and Peculiarity on Tensor Data
    Hu, Jiuyun
    Shi, Naichen
    Al Kontar, Raed
    Yan, Hao
    TECHNOMETRICS, 2025,
  • [25] Nonlinear dimensionality reduction based on dictionary learning
    Zheng S.-L.
    Li Y.-X.
    Wei X.
    Peng X.-S.
    Zidonghua Xuebao/Acta Automatica Sinica, 2016, 42 (07): : 1065 - 1076
  • [26] Compressed Sensing based Speech Compression using Dictionary Learning and IRLS algorithm
    Derouaz, Wafa
    Meksen, Thouraya Merazi
    PROCEEDINGS 2018 3RD INTERNATIONAL CONFERENCE ON ELECTRICAL SCIENCES AND TECHNOLOGIES IN MAGHREB (CISTEM), 2018, : 614 - 618
  • [27] Perturbations of the Tcur Decomposition for Tensor Valued Data in the Tucker Format
    Maolin Che
    Juefei Chen
    Yimin Wei
    Journal of Optimization Theory and Applications, 2022, 194 : 852 - 877
  • [28] Simultaneous dimensionality reduction and dictionary learning for sparse representation based classification
    Yang, Bao-Qing
    Gu, Chao-Chen
    Wu, Kai-Jie
    Zhang, Tao
    Guan, Xin-Ping
    MULTIMEDIA TOOLS AND APPLICATIONS, 2017, 76 (06) : 8969 - 8990
  • [29] Simultaneous dimensionality reduction and dictionary learning for sparse representation based classification
    Bao-Qing Yang
    Chao-Chen Gu
    Kai-Jie Wu
    Tao Zhang
    Xin-Ping Guan
    Multimedia Tools and Applications, 2017, 76 : 8969 - 8990
  • [30] TUCKER DIMENSIONALITY REDUCTION OF THREE-DIMENSIONAL ARRAYS IN LINEAR TIME
    Oseledets, I. V.
    Savostianov, D. V.
    Tyrtyshnikov, E. E.
    SIAM JOURNAL ON MATRIX ANALYSIS AND APPLICATIONS, 2008, 30 (03) : 939 - 956