Efficient Hyperbolic Perceptron for Image Classification

被引:0
作者
Ahsan, Ahmad Omar [1 ]
Tang, Susanna [2 ]
Peng, Wei [2 ]
机构
[1] Univ Calgary, Dept Biomed Engn, Calgary, AB T2N 1N4, Canada
[2] Stanford Univ, Stanford Med, Stanford, CA 94305 USA
关键词
deep neural networks; hyperbolic geometry; image classification; Lorentz model; graphs;
D O I
10.3390/electronics12194027
中图分类号
TP [自动化技术、计算机技术];
学科分类号
0812 ;
摘要
Deep neural networks, often equipped with powerful auto-optimization tools, find widespread use in diverse domains like NLP and computer vision. However, traditional neural architectures come with specific inductive biases, designed to reduce parameter search space, cut computational costs, or introduce domain expertise into the network design. In contrast, multilayer perceptrons (MLPs) offer greater freedom and lower inductive bias than convolutional neural networks (CNNs), making them versatile for learning complex patterns. Despite their flexibility, most neural architectures operate in a flat Euclidean space, which may not be optimal for various data types, particularly those with hierarchical correlations. In this paper, we move one step further by introducing the hyperbolic Res-MLP (HR-MLP), an architecture extending the attention-free MLP to a non-Euclidean space. HR-MLP leverages fully hyperbolic layers for feature embeddings and end-to-end image classification. Our novel Lorentz cross-patch and cross-channel layers enable direct hyperbolic operations with fewer parameters, facilitating faster training and superior performance compared to Euclidean counterparts. Experimental results on CIFAR10, CIFAR100, and MiniImageNet confirm HR-MLP's competitive and improved performance.
引用
收藏
页数:16
相关论文
共 60 条
[1]  
Ahonen T, 2004, LECT NOTES COMPUT SC, V3021, P469
[2]  
[Anonymous], 2007, Proceedings of the international workshop on Workshop on multimedia information retrieval, DOI DOI 10.1145/1290082.1290111
[3]   Hyperbolic Image Segmentation [J].
Atigh, Mina Ghadimi ;
Schoep, Julian ;
Acar, Erman ;
Van Noord, Nanne ;
Mettes, Pascal .
2022 IEEE/CVF CONFERENCE ON COMPUTER VISION AND PATTERN RECOGNITION (CVPR 2022), 2022, :4443-4452
[4]  
Bachmann G, 2020, PR MACH LEARN RES, V119
[5]  
Becigneul G., 2019, P INT C LEARNING REP
[6]  
Beltrami E., 1868, Annali. di Mat., ser II, V2, P232, DOI 10.1007/BF02419615
[7]   Random forests [J].
Breiman, L .
MACHINE LEARNING, 2001, 45 (01) :5-32
[8]   Geometric Deep Learning Going beyond Euclidean data [J].
Bronstein, Michael M. ;
Bruna, Joan ;
LeCun, Yann ;
Szlam, Arthur ;
Vandergheynst, Pierre .
IEEE SIGNAL PROCESSING MAGAZINE, 2017, 34 (04) :18-42
[9]  
Cai H, 2024, Arxiv, DOI arXiv:2205.14756
[10]  
Cannon J. W., 1997, Flavors of geometry, V31, P2