Residual D2NN: training diffractive deep neural networks via learnable light shortcuts

被引:80
作者
Dou, Hongkun [1 ]
Deng, Yue [1 ,7 ]
Yan, Tao [2 ,5 ]
Wu, Huaqiang [3 ,4 ]
Lin, Xing [2 ,3 ,5 ]
Dai, Qionghai [2 ,5 ,6 ]
机构
[1] Beihang Univ, Sch Astronaut, Beijing 100191, Peoples R China
[2] Tsinghua Univ, Dept Automat, Beijing 100084, Peoples R China
[3] Tsinghua Univ, Beijing Innovat Ctr Future Chip, Beijing 100084, Peoples R China
[4] Tsinghua Univ, Inst Microelect, Beijing 100084, Peoples R China
[5] Tsinghua Univ, Inst Brain & Cognit Sci, Beijing 100084, Peoples R China
[6] Tsinghua Univ, Beijing Natl Res Ctr Informat Sci & Technol, Beijing 100084, Peoples R China
[7] Beihang Univ, Beijing Adv Innovat Ctr Big Data & Brain Comp, Beijing 100191, Peoples R China
基金
中国国家自然科学基金;
关键词
Backpropagation;
D O I
10.1364/OL.389696
中图分类号
O43 [光学];
学科分类号
070207 ; 0803 ;
摘要
The diffractive deep neural network ((DNN)-N-2) has demonstrated its importance in performing various all-optical machine learning tasks, e.g., classification, segmentation, etc. However, deeper D(2)NNs that provide higher inference complexity are more difficult to train due to the problem of gradient vanishing. We introduce the residual D(2)NNs (Res-(DNN)-N-2), which enables us to train substantially deeper diffractive networks by constructing diffractive residual learning blocks to learn the residual mapping functions. Unlike the existing plain D(2)NNs, Res-D(2)NNs contribute to the design of a learnable light shortcut to directly connect the input and output between optical layers. Such a shortcut offers a direct path for gradient backpropagation in training, which is an effective way to alleviate the gradient vanishing issue on very deep diffractive neural networks. Experimental results on image classification and pixel super-resolution demonstrate the superiority of Res-D(2)NNs over the existing plain (DNN)-N-2 architectures. (C) 2020 Optical Society of America
引用
收藏
页码:2688 / 2691
页数:4
相关论文
共 19 条
[11]   Design of task-specific optical systems using broadband diffractive neural networks [J].
Luo, Yi ;
Mengu, Deniz ;
Yardimci, Nezih T. ;
Rivenson, Yair ;
Veli, Muhammed ;
Jarrahi, Mona ;
Ozcan, Aydogan .
LIGHT-SCIENCE & APPLICATIONS, 2019, 8 (1)
[12]   John von Neumann and the evolutionary growth of complexity: Looking backward, looking forward ... [J].
McMullin, B .
ARTIFICIAL LIFE, 2000, 6 (04) :347-361
[13]   FRACTIONAL FOURIER-TRANSFORMS AND THEIR OPTICAL IMPLEMENTATION .1. [J].
MENDLOVIC, D ;
OZAKTAS, HM .
JOURNAL OF THE OPTICAL SOCIETY OF AMERICA A-OPTICS IMAGE SCIENCE AND VISION, 1993, 10 (09) :1875-1881
[14]   Deep learning [J].
Rusk, Nicole .
NATURE METHODS, 2016, 13 (01) :35-35
[15]   Moore's Law: Past, present, and future [J].
Schaller, RR .
IEEE SPECTRUM, 1997, 34 (06) :52-+
[16]  
Shen YC, 2017, NAT PHOTONICS, V11, P441, DOI [10.1038/NPHOTON.2017.93, 10.1038/nphoton.2017.93]
[17]  
Szegedy C, 2017, AAAI CONF ARTIF INTE, P4278
[18]  
Vollgraf R, 2017, arXiv [Preprint]
[19]   Fourier-space Diffractive Deep Neural Network [J].
Yan, Tao ;
Wu, Jiamin ;
Zhou, Tiankuang ;
Xie, Hao ;
Xu, Feng ;
Fan, Jingtao ;
Fang, Lu ;
Lin, Xing ;
Dai, Qionghai .
PHYSICAL REVIEW LETTERS, 2019, 123 (02)