Machine Learning With Tree Tensor Networks, CP Rank Constraints, and Tensor Dropout

被引:1
|
作者
Chen, Hao [1 ]
Barthel, Thomas [2 ,3 ]
机构
[1] Swiss Fed Inst Technol, Dept Phys, CH-8093 Zurich, Switzerland
[2] Duke Univ, Dept Phys, Durham, NC 27708 USA
[3] Duke Univ, Duke Quantum Ctr, Durham, NC 27708 USA
关键词
Machine learning; image classification; tensor networks; tree tensor networks; CP rank; tensor dropout; MATRIX RENORMALIZATION-GROUP; STATES; APPROXIMATION; MODELS;
D O I
10.1109/TPAMI.2024.3396386
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
Tensor networks developed in the context of condensed matter physics try to approximate order-N tensors with a reduced number of degrees of freedom that is only polynomial in N and arranged as a network of partially contracted smaller tensors. As we have recently demonstrated in the context of quantum many-body physics, computation costs can be further substantially reduced by imposing constraints on the canonical polyadic (CP) rank of the tensors in such networks. Here, we demonstrate how tree tensor networks (TTN) with CP rank constraints and tensor dropout can be used in machine learning. The approach is found to outperform other tensor-network-based methods in Fashion-MNIST image classification. A low-rank TTN classifier with branching ratio b = 4 reaches a test set accuracy of 90.3% with low computation costs. Consisting of mostly linear elements, tensor network classifiers avoid the vanishing gradient problem of deep neural networks. The CP rank constraints have additional advantages: The number of parameters can be decreased and tuned more freely to control overfitting, improve generalization properties, and reduce computation costs. They allow us to employ trees with large branching ratios, substantially improving the representation power.
引用
收藏
页码:7825 / 7832
页数:8
相关论文
共 50 条
  • [1] Towards quantum machine learning with tensor networks
    Huggins, William
    Patil, Piyush
    Mitchell, Bradley
    Whaley, K. Birgitta
    Stoudenmire, E. Miles
    QUANTUM SCIENCE AND TECHNOLOGY, 2019, 4 (02)
  • [2] Tensor Dropout for Robust Learning
    Kolbeinsson, Arinbjorn
    Kossaifi, Jean
    Panagakis, Yannis
    Bulat, Adrian
    Anandkumar, Animashree
    Tzoulaki, Ioanna
    Matthews, Paul M.
    IEEE JOURNAL OF SELECTED TOPICS IN SIGNAL PROCESSING, 2021, 15 (03) : 630 - 640
  • [3] Learning with tree tensor networks: Complexity estimates and model selection
    Michel, Bertrand
    Nouy, Anthony
    BERNOULLI, 2022, 28 (02) : 910 - 936
  • [4] Tensor Convolutional Dictionary Learning With CP Low-Rank Activations
    Humbert, Pierre
    Oudre, Laurent
    Vayatis, Nicolas
    Audiffren, Julien
    IEEE TRANSACTIONS ON SIGNAL PROCESSING, 2022, 70 : 785 - 796
  • [5] Tensor Manifold with Tucker Rank Constraints
    Chang, Shih Yu
    Luo, Ziyan
    Qi, Liqun
    ASIA-PACIFIC JOURNAL OF OPERATIONAL RESEARCH, 2022, 39 (02)
  • [6] Tensor networks for explainable machine learning in cybersecurity
    Aizpurua, Borja
    Palmer, Samuel
    Orus, Roman
    NEUROCOMPUTING, 2025, 639
  • [7] Machine learning by unitary tensor network of hierarchical tree structure
    Liu, Ding
    Ran, Shi-Ju
    Wittek, Peter
    Peng, Cheng
    Garcia, Raul Blazquez
    Su, Gang
    Lewenstein, Maciej
    NEW JOURNAL OF PHYSICS, 2019, 21 (07)
  • [8] TensorKrowch: Smooth integration of tensor networks in machine learning
    Monturiol, Jose Ramon Pareja
    Perez-Garcia, David
    Pozas-Kerstjens, Alejandro
    QUANTUM, 2024, 8 : 1 - 20
  • [9] An Approximation of the CP-Rank of a Partially Sampled Tensor
    Ashraphijuo, Morteza
    Wang, Xiaodong
    Aggarwal, Vaneet
    2017 55TH ANNUAL ALLERTON CONFERENCE ON COMMUNICATION, CONTROL, AND COMPUTING (ALLERTON), 2017, : 604 - 611
  • [10] Learning relevant features of data with multi-scale tensor networks
    Stoudenmire, E. Miles
    QUANTUM SCIENCE AND TECHNOLOGY, 2018, 3 (03):