Deep Learning With TensorFlow: A Review

被引:297
作者
Pang, Bo [1 ]
Nijkamp, Erik [1 ]
Wu, Ying Nian [1 ]
机构
[1] UCLA, Dept Stat, 520 Portola Plaza, Los Angeles, CA 90095 USA
关键词
adaptive testing; computation; modeling; neural network; program evaluation; statistics; technology;
D O I
10.3102/1076998619872761
中图分类号
G40 [教育学];
学科分类号
040101 ; 120403 ;
摘要
This review covers the core concepts and design decisions of TensorFlow. TensorFlow, originally created by researchers at Google, is the most popular one among the plethora of deep learning libraries. In the field of deep learning, neural networks have achieved tremendous success and gained wide popularity in various areas. This family of models also has tremendous potential to promote data analysis and modeling for various problems in educational and behavioral sciences given its flexibility and scalability. We give the reader an overview of the basics of neural network models such as the multilayer perceptron, the convolutional neural network, and stochastic gradient descent, the most commonly used optimization method for neural network models. However, the implementation of these models and optimization algorithms is time-consuming and error-prone. Fortunately, TensorFlow greatly eases and accelerates the research and application of neural network models. We review several core concepts of TensorFlow such as graph construction functions, graph execution tools, and TensorFlow's visualization tool, TensorBoard. Then, we apply these concepts to build and train a convolutional neural network model to classify handwritten digits. This review is concluded by a comparison of low- and high-level application programming interfaces and a discussion of graphical processing unit support, distributed training, and probabilistic modeling with TensorFlow Probability library.
引用
收藏
页码:227 / 248
页数:22
相关论文
共 15 条
[1]  
Collobert R., 2008, P 25 INT C MACH LEAR, P160, DOI [https://doi.org/10.1145/1390156.1390177, DOI 10.1145/1390156.1390177, DOI 10.1145/1390156.1390177.ICML08]
[2]  
Cybenko G., 1989, Mathematics of Control, Signals, and Systems, V2, P303, DOI 10.1007/BF02551274
[3]  
Devlin J, 2019, 2019 CONFERENCE OF THE NORTH AMERICAN CHAPTER OF THE ASSOCIATION FOR COMPUTATIONAL LINGUISTICS: HUMAN LANGUAGE TECHNOLOGIES (NAACL HLT 2019), VOL. 1, P4171
[4]  
Geron A. l., 2017, Tech. Rep., V1st
[5]  
Goodfellow I, 2016, ADAPT COMPUT MACH LE, P1
[6]   APPROXIMATION CAPABILITIES OF MULTILAYER FEEDFORWARD NETWORKS [J].
HORNIK, K .
NEURAL NETWORKS, 1991, 4 (02) :251-257
[7]   ImageNet Classification with Deep Convolutional Neural Networks [J].
Krizhevsky, Alex ;
Sutskever, Ilya ;
Hinton, Geoffrey E. .
COMMUNICATIONS OF THE ACM, 2017, 60 (06) :84-90
[8]   Gradient-based learning applied to document recognition [J].
Lecun, Y ;
Bottou, L ;
Bengio, Y ;
Haffner, P .
PROCEEDINGS OF THE IEEE, 1998, 86 (11) :2278-2324
[9]  
LeNail A., 2019, J OPEN SOURCE SOFTWA, V4, P747
[10]   Deep learning in bioinformatics [J].
Min, Seonwoo ;
Lee, Byunghan ;
Yoon, Sungroh .
BRIEFINGS IN BIOINFORMATICS, 2017, 18 (05) :851-869