Densely Connected G-invariant Deep Neural Networks with Signed Permutation Representations

被引:0
作者
Agrawal, Devanshu [1 ]
Ostrowski, James [1 ]
机构
[1] Univ Tennessee, Dept Ind & Syst Engn, Knoxville, TN 37996 USA
关键词
deep learning; group theory; neural network; skip connection; symmetry;
D O I
暂无
中图分类号
TP [自动化技术、计算机技术];
学科分类号
0812 ;
摘要
We introduce and investigate, for finite groups G, G-invariant deep neural network (G-DNN) architectures with ReLU activation that are densely connected- i.e., include all possible skip connections. In contrast to other G-invariant architectures in the literature, the preactivations of theG-DNNs presented here are able to transform by signed permu-tation representations (signed perm-reps) of G. Moreover, the individual layers of the G-DNNs are not required to be G-equivariant; instead, the preactivations are constrained to be G-equivariant functions of the network input in a way that couples weights across all layers. The result is a richer family of G-invariant architectures never seen previously. We derive an efficient implementation of G-DNNs after a reparameterization of weights, as well as necessary and sufficient conditions for an architecture to be "admissible"- i.e., non-degenerate and inequivalent to smaller architectures. We include code that allows a user to build a G-DNN interactively layer-by-layer, with the final architecture guaranteed to be admissible. We show that there are far more admissible G-DNN architectures than those accessible with the "concatenated ReLU" activation function from the literature. Finally, we apply G-DNNs to two example problems-(1) multiplication in {-1, 1} (with theoretical guarantees) and (2) 3D object classification-finding that the inclusion of signed perm-reps significantly boosts predictive performance compared to baselines with only ordinary (i.e., unsigned) perm-reps.
引用
收藏
页数:40
相关论文
共 34 条
  • [1] Group-equivariant autoencoder for identifying spontaneously broken symmetries
    Agrawal, Devanshu
    Del Maestro, Adrian
    Johnston, Steven
    Ostrowski, James
    [J]. PHYSICAL REVIEW E, 2023, 107 (05)
  • [2] Agrawal Devanshu, 2022, Advances in Neural Information Processing Systems., V35
  • [3] Geometric deep learning on molecular representations
    Atz, Kenneth
    Grisoni, Francesca
    Schneider, Gisbert
    [J]. NATURE MACHINE INTELLIGENCE, 2021, 3 (12) : 1023 - 1032
  • [4] STRUCTURE AND REPRESENTATIONS OF THE HYPEROCTAHEDRAL GROUP
    BAAKE, M
    [J]. JOURNAL OF MATHEMATICAL PHYSICS, 1984, 25 (11) : 3171 - 3182
  • [5] Balduzzi D, 2017, 34 INT C MACHINE LEA, V70
  • [6] Basu S, 2021, Arxiv, DOI arXiv:2104.04848
  • [7] Bouc Serge, 2000, HDB ALGEBRA, V2, P739
  • [8] Burnside W., 1911, THEORY GROUPS FINITE
  • [9] Cohen T. S., 2017, INT C LEARNING REPRE
  • [10] Cohen TS, 2019, ADV NEUR IN, V32