Multi-scale tensor network architecture for machine learning

被引:0
|
作者
Reyes, J.A. [1 ]
Stoudenmire, E.M. [2 ]
机构
[1] Department of Physics, University of Central Florida, University Blvd, Orlando,FL,32816, United States
[2] Center for Computational Quantum Physics, Flatiron Institute, 5th Avenue, New York,NY,10010, United States
来源
Machine Learning: Science and Technology | 2021年 / 2卷 / 03期
基金
美国国家科学基金会;
关键词
Adaptive algorithms - Data handling - Tensors - Maximum likelihood estimation - Network layers - Statistical mechanics - Learning algorithms - Network architecture - Matrix algebra;
D O I
暂无
中图分类号
学科分类号
摘要
We present an algorithm for supervised learning using tensor networks, employing a step of data pre-processing by coarse-graining through a sequence of wavelet transformations. These transformations are represented as a set of tensor network layers identical to those in a multi-scale entanglement renormalization ansatz tensor network. We perform supervised learning and regression tasks through a model based on a matrix product states (MPSs) acting on the coarse-grained data. Because the entire model consists of tensor contractions (apart from the initial non-linear feature map), we can adaptively fine-grain the optimized MPS model 'backwards' through the layers with essentially no loss in performance. The MPS itself is trained using an adaptive algorithm based on the density matrix renormalization group algorithm. We test our methods by performing a classification task on audio data and a regression task on temperature time-series data, studying the dependence of training accuracy on the number of coarse-graining layers and showing how fine-graining through the network may be used to initialize models which access finer-scale features. © 2021 The Author(s).
引用
收藏
相关论文
empty
未找到相关数据