Neural network training fingerprint: visual analytics of the training process in classification neural networks

被引:3
作者
Ferreira, Martha Dais [3 ]
Cantareira, Gabriel D. [1 ,4 ]
de Mello, Rodrigo F. [2 ,5 ]
Paulovich, Fernando V. [3 ]
机构
[1] King's Coll London, Dept Informatics, London, England
[2] Univ Sao Paulo, Sao Carlos, Brazil
[3] Dalhousie Univ, Fac Comp Sci, Halifax, NS, Canada
[4] Kings Coll London, Dept Informat, London, England
[5] Univ Sao Paulo, Sao Carlos, Brazil
基金
加拿大自然科学与工程研究理事会;
关键词
Neural network visualization; Neural network training; Deep learning; Visual analytics; Visualization;
D O I
10.1007/s12650-021-00809-4
中图分类号
TP39 [计算机的应用];
学科分类号
081203 ; 0835 ;
摘要
The striking results of deep neural networks (DNN) have motivated its wide acceptance to tackle large datasets and complex tasks such as natural language processing, facial recognition, and artificial image generation. However, DNN parameters are often empirically selected on a trial-and-error approach without detailed information on convergence behavior. While some visualization techniques have been proposed to aid the comprehension of general-purpose neural networks, only a few explore the training process, lacking the ability to adequately display how abstract representations are formed and represent the influence of training parameters during this process. This paper describes neural network training fingerprint (NNTF), a visual analytics approach to investigate the training process of any neural network performing classification. NNTF allows understanding how classification decisions change along the training process, displaying information about convergence, oscillations, and training rates. We show its usefulness through case studies and demonstrate how it can support the analysis of training parameters.
引用
收藏
页码:593 / 612
页数:20
相关论文
共 40 条
[1]  
Abadi M, 2016, PROCEEDINGS OF OSDI'16: 12TH USENIX SYMPOSIUM ON OPERATING SYSTEMS DESIGN AND IMPLEMENTATION, P265
[2]  
Alencar A.B., 2012, SAC 12, P997, DOI DOI 10.1145/2245276.2245469
[3]  
Babiker HKB, 2017, ARXIV PREPRINT ARXIV
[4]   On Pixel-Wise Explanations for Non-Linear Classifier Decisions by Layer-Wise Relevance Propagation [J].
Bach, Sebastian ;
Binder, Alexander ;
Montavon, Gregoire ;
Klauschen, Frederick ;
Mueller, Klaus-Robert ;
Samek, Wojciech .
PLOS ONE, 2015, 10 (07)
[5]   Exploring Neural Network Hidden Layer Activity Using Vector Fields [J].
Cantareira, Gabriel D. ;
Etemad, Elham ;
Paulovich, Fernando V. .
INFORMATION, 2020, 11 (09) :1-15
[6]   OoDAnalyzer: Interactive Analysis of Out-of-Distribution Samples [J].
Chen, Changjian ;
Yuan, Jun ;
Lu, Yafeng ;
Liu, Yang ;
Su, Hang ;
Yuan, Songtao ;
Liu, Shixia .
IEEE TRANSACTIONS ON VISUALIZATION AND COMPUTER GRAPHICS, 2021, 27 (07) :3335-3349
[7]   Fast and reliable incremental dimensionality reduction for streaming data [J].
de Araujo Tiburtino Neves, Tacito Trindade ;
Martins, Rafael Messias ;
Coimbra, Danilo Barbosa ;
Kucher, Kostiantyn ;
Kerren, Andreas ;
Paulovich, Fernando, V .
COMPUTERS & GRAPHICS-UK, 2022, 102 :233-244
[8]  
Erhan D., 2009, Technical Report
[9]  
Goodfellow I, 2016, ADAPT COMPUT MACH LE, P1
[10]   Know Your Enemy: Identifying Quality Problems of Time Series Data [J].
Gschwandtner, Theresia ;
Erhart, Oliver .
2018 IEEE PACIFIC VISUALIZATION SYMPOSIUM (PACIFICVIS), 2018, :205-214