SurReal: Complex-Valued Learning as Principled Transformations on a Scaling and Rotation Manifold

被引:9
作者
Chakraborty, Rudrasis [1 ]
Xing, Yifei [1 ]
Yu, Stella X. [1 ]
机构
[1] Univ Calif Berkeley UC Berkeley, Int Comp Sci Inst ICSI, Berkeley, CA 94720 USA
关键词
Complex value; equivariance; Frechet mean; invariance; Riemannian manifold; NEURAL-NETWORK;
D O I
10.1109/TNNLS.2020.3030565
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
Complex-valued data are ubiquitous in signal and image processing applications, and complex-valued representations in deep learning have appealing theoretical properties. While these aspects have long been recognized, complex-valued deep learning continues to lag far behind its real-valued counterpart. We propose a principled geometric approach to complexvalued deep learning. Complex-valued data could often be subject to arbitrary complex-valued scaling; as a result, real and imaginary components could covary. Instead of treating complex values as two independent channels of real values, we recognize their underlying geometry: we model the space of complex numbers as a product manifold of nonzero scaling and planar rotations. Arbitrary complex-valued scaling naturally becomes a group of transitive actions on this manifold. We propose to extend the property instead of the form of real-valued functions to the complex domain. We define convolution as the weighted Frechet mean on the manifold that is equivariant to the group of scaling/rotation actions and define distance transform on the manifold that is invariant to the action group. The manifold perspective also allows us to define nonlinear activation functions, such as tangent ReLU and G-transport, as well as residual connections on the manifold-valued data. We dub our model SurReal, as our experiments on MSTAR and RadioML deliver high performance with only a fractional size of real- and complex-valued baseline models.
引用
收藏
页码:940 / 951
页数:12
相关论文
共 56 条
[41]   Understanding deep convolutional networks [J].
Mallat, Stephane .
PHILOSOPHICAL TRANSACTIONS OF THE ROYAL SOCIETY A-MATHEMATICAL PHYSICAL AND ENGINEERING SCIENCES, 2016, 374 (2065)
[42]   An extension of the back-propagation algorithm to complex numbers [J].
Nitta, T .
NEURAL NETWORKS, 1997, 10 (08) :1391-1415
[43]  
Nitta T, 2002, ICONIP'02: PROCEEDINGS OF THE 9TH INTERNATIONAL CONFERENCE ON NEURAL INFORMATION PROCESSING, P1099
[44]   Convolutional Radio Modulation Recognition Networks [J].
O'Shea, Timothy J. ;
Corgan, Johnathan ;
Clancy, T. Charles .
ENGINEERING APPLICATIONS OF NEURAL NETWORKS, EANN 2016, 2016, 629 :213-226
[45]   TENSOR-TRAIN DECOMPOSITION [J].
Oseledets, I. V. .
SIAM JOURNAL ON SCIENTIFIC COMPUTING, 2011, 33 (05) :2295-2317
[46]  
Parcollet Titouan, 2018, ARXIV180604418
[47]  
Reichert J., 1992, ICASSP-92: 1992 IEEE International Conference on Acoustics, Speech and Signal Processing (Cat. No.92CH3103-9), P221, DOI 10.1109/ICASSP.1992.226530
[48]  
Salehian H., 2015, MATH FDN COMPUT ANAT, V3
[49]   The Graph Neural Network Model [J].
Scarselli, Franco ;
Gori, Marco ;
Tsoi, Ah Chung ;
Hagenbuchner, Markus ;
Monfardini, Gabriele .
IEEE TRANSACTIONS ON NEURAL NETWORKS, 2009, 20 (01) :61-80
[50]   A Lightweight Convolutional Neural Network Based on Visual Attention for SAR Image Target Classification [J].
Shao, Jiaqi ;
Qu, Changwen ;
Li, Jianwei ;
Peng, Shujuan .
SENSORS, 2018, 18 (09)