Photonic neural networks and optics-informed deep learning fundamentals

被引:28
作者
Tsakyridis, Apostolos [1 ]
Moralis-Pegios, Miltiadis [1 ]
Giamougiannis, George [1 ]
Kirtas, Manos [1 ]
Passalis, Nikolaos [1 ]
Tefas, Anastasios [1 ]
Pleros, Nikos [1 ]
机构
[1] Aristotle Univ Thessaloniki, Dept Informat, Thessaloniki 54124, Greece
基金
欧盟地平线“2020”;
关键词
SILICON PHOTONICS; ARCHITECTURE; ELECTRONICS; INTEGRATION; MODULATORS; COMPACT; DESIGN; ANALOG; POWER;
D O I
10.1063/5.0169810
中图分类号
O43 [光学];
学科分类号
070207 ; 0803 ;
摘要
The recent explosive compute growth, mainly fueled by the boost of artificial intelligence (AI) and deep neural networks (DNNs), is currently instigating the demand for a novel computing paradigm that can overcome the insurmountable barriers imposed by conventional electronic computing architectures. Photonic neural networks (PNNs) implemented on silicon integration platforms stand out as a promising candidate to endow neural network (NN) hardware, offering the potential for energy efficient and ultra-fast computations through the utilization of the unique primitives of photonics, i.e., energy efficiency, THz bandwidth, and low-latency. Thus far, several demonstrations have revealed the huge potential of PNNs in performing both linear and non-linear NN operations at unparalleled speed and energy consumption metrics. Transforming this potential into a tangible reality for deep learning (DL) applications requires, however, a deep understanding of the basic PNN principles, requirements, and challenges across all constituent architectural, technological, and training aspects. In this Tutorial, we, initially, review the principles of DNNs along with their fundamental building blocks, analyzing also the key mathematical operations needed for their computation in photonic hardware. Then, we investigate, through an intuitive mathematical analysis, the interdependence of bit precision and energy efficiency in analog photonic circuitry, discussing the opportunities and challenges of PNNs. Followingly, a performance overview of PNN architectures, weight technologies, and activation functions is presented, summarizing their impact in speed, scalability, and power consumption. Finally, we provide a holistic overview of the optics-informed NN training framework that incorporates the physical properties of photonic building blocks into the training process in order to improve the NN classification accuracy and effectively elevate neuromorphic photonic hardware into high-performance DL computational settings.
引用
收藏
页数:25
相关论文
共 169 条
[1]   A Many-core Architecture for In-Memory Data Processing [J].
Agrawal, Sandeep R. ;
Idicula, Sam ;
Raghavan, Arun ;
Vlachos, Evangelos ;
Govindaraju, Venkatraman ;
Varadarajan, Venkatanathan ;
Balkesen, Cagri ;
Giannikis, Georgios ;
Roth, Charlie ;
Agarwal, Nipun ;
Sedlar, Eric .
50TH ANNUAL IEEE/ACM INTERNATIONAL SYMPOSIUM ON MICROARCHITECTURE (MICRO), 2017, :245-258
[2]   Scaling up silicon photonic-based accelerators: Challenges and opportunities [J].
Al-Qadasi, M. A. ;
Chrostowski, L. ;
Shastri, B. J. ;
Shekhar, S. .
APL PHOTONICS, 2022, 7 (02)
[3]   Optical RAM and integrated optical memories: a survey [J].
Alexoudi, Theoni ;
Kanellos, George Theodore ;
Pleros, Nikos .
LIGHT-SCIENCE & APPLICATIONS, 2020, 9 (01)
[4]   Sub-wavelength GHz-fast broadband ITO Mach-Zehnder modulator on silicon photonics [J].
Amin, Rubab ;
Maiti, Rishi ;
Gui, Yaliang ;
Suer, Can ;
Miscuglio, Mario ;
Heidari, Elham ;
Chen, Ray T. ;
Dalir, Hamed ;
Sorger, Volker J. .
OPTICA, 2020, 7 (04) :333-335
[5]   A Low Power, Fully Event-Based Gesture Recognition System [J].
Amir, Arnon ;
Taba, Brian ;
Berg, David ;
Melano, Timothy ;
McKinstry, Jeffrey ;
Di Nolfo, Carmelo ;
Nayak, Tapan ;
Andreopoulos, Alexander ;
Garreau, Guillaume ;
Mendoza, Marcela ;
Kusnitz, Jeff ;
Debole, Michael ;
Esser, Steve ;
Delbruck, Tobi ;
Flickner, Myron ;
Modha, Dharmendra .
30TH IEEE CONFERENCE ON COMPUTER VISION AND PATTERN RECOGNITION (CVPR 2017), 2017, :7388-7397
[6]  
[Anonymous], Intel Gaudi white paper
[7]  
[Anonymous], About us
[8]  
[Anonymous], Taking Powerful, Efficient Inference to the Edge Paradigms Seems, Capable of Stimulating Additional Advances Shaping Future Digital Computing Roadmaps
[9]  
[Anonymous], 2012, On the Difficulty of Training Recurrent Neural Networks, DOI DOI 10.48550/ARXIV.1211.5063
[10]  
[Anonymous], 2015, P IEEE INT C COMP VI