Efficient training of energy-based models via spin-glass control

被引:7
作者
Pozas-Kerstjens, Alejandro [1 ]
Munoz-Gil, Gorka [2 ]
Pinol, Eloy [2 ,3 ]
Garcia-March, Miguel Angel [3 ]
Acin, Antonio [2 ,5 ]
Lewenstein, Maciej [2 ,5 ]
Grzybowski, Przemyslaw R. [4 ]
机构
[1] Univ Complutense Madrid, Dept Anal Matemat, Madrid 28040, Spain
[2] Barcelona Inst Sci & Technol, ICFO Inst Ciencies Foton, Barcelona 08860, Spain
[3] Univ Politecn Valencia, Inst Univ Matemat Pura & Aplicada, Valencia 46022, Spain
[4] Adam Mickiewicz Univ, Fac Phys, Umultowska 85, PL-61614 Poznan, Poland
[5] ICREA, Passeig Lluis Co 23, Barcelona 08010, Spain
来源
MACHINE LEARNING-SCIENCE AND TECHNOLOGY | 2021年 / 2卷 / 02期
关键词
unsupervised learning; Boltzmann machine; spin glass; statistical physics; physics-inspired machine learning; COHERENT ISING MACHINE; SYSTEMS; NETWORKS;
D O I
10.1088/2632-2153/abe807
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
We introduce a new family of energy-based probabilistic graphical models for efficient unsupervised learning. Its definition is motivated by the control of the spin-glass properties of the Ising model described by the weights of Boltzmann machines. We use it to learn the Bars and Stripes dataset of various sizes and the MNIST dataset, and show how they quickly achieve the performance offered by standard methods for unsupervised learning. Our results indicate that the standard initialization of Boltzmann machines with random weights equivalent to spin-glass models is an unnecessary bottleneck in the process of training. Furthermore, this new family allows for very easy access to low-energy configurations, which points to new, efficient training algorithms. The simplest variant of such algorithms approximates the negative phase of the log-likelihood gradient with no Markov chain Monte Carlo sampling costs at all, and with an accuracy sufficient to achieve good learning and generalization.
引用
收藏
页数:19
相关论文
共 44 条
[1]   Quantum Boltzmann Machine [J].
Amin, Mohammad H. ;
Andriyash, Evgeny ;
Rolfe, Jason ;
Kulchytskyy, Bohdan ;
Melko, Roger .
PHYSICAL REVIEW X, 2018, 8 (02)
[2]   SPIN-GLASS MODELS OF NEURAL NETWORKS [J].
AMIT, DJ ;
GUTFREUND, H .
PHYSICAL REVIEW A, 1985, 32 (02) :1007-1018
[3]   STORING INFINITE NUMBERS OF PATTERNS IN A SPIN-GLASS MODEL OF NEURAL NETWORKS [J].
AMIT, DJ ;
GUTFREUND, H ;
SOMPOLINSKY, H .
PHYSICAL REVIEW LETTERS, 1985, 55 (14) :1530-1533
[4]   Deep learning [J].
LeCun, Yann ;
Bengio, Yoshua ;
Hinton, Geoffrey .
NATURE, 2015, 521 (7553) :436-444
[5]  
[Anonymous], 1989, MODELING BRAIN FUNCT, DOI DOI 10.1017/CBO9780511623257
[6]  
[Anonymous], 2017, P NIPS 2017 AUT WORK
[7]  
[Anonymous], 2016, INT C LEARN REPR
[8]  
[Anonymous], 2016, J MACH LEARN RES
[9]  
[Anonymous], 1983, P IEEE C COMP VIS PA
[10]   ON THE COMPUTATIONAL-COMPLEXITY OF ISING SPIN-GLASS MODELS [J].
BARAHONA, F .
JOURNAL OF PHYSICS A-MATHEMATICAL AND GENERAL, 1982, 15 (10) :3241-3253