SEMI-SUPERVISED BOOTSTRAPPING APPROACH FOR NEURAL NETWORK FEATURE EXTRACTOR TRAINING

被引:0
作者
Grezl, Frantisek [1 ]
Karafiat, Martin [1 ]
机构
[1] Brno Univ Technol, Speech FIT & IT4I Ctr Excellence, CS-61090 Brno, Czech Republic
来源
2013 IEEE WORKSHOP ON AUTOMATIC SPEECH RECOGNITION AND UNDERSTANDING (ASRU) | 2013年
关键词
Semi-supervised training; bootstrapping; bottle-neck features;
D O I
暂无
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
This paper presents bootstrapping approach for neural network training. The neural networks serve as bottle-neck feature extractor for subsequent GMM-HMM recognizer. The recognizer is also used for transcription and confidence assignment of untranscribed data. Based on the confidence, segments are selected and mixed with supervised data and new NNs are trained. With this approach, it is possible to recover 40-55% of the difference between partially and fully transcribed data (3 to 5% absolute improvement over NN trained on supervised data only). Using 70-85% of automatically transcribed segments with the highest confidence was found optimal to achieve this result.
引用
收藏
页码:470 / 475
页数:6
相关论文
共 19 条
[1]  
[Anonymous], P INTERSPEECH
[2]  
[Anonymous], OPTIMIZATION
[3]  
Chen L., 2011, P ICASSP
[4]   Size matters: An empirical study of neural network training for large vocabulary continuous speech recognition [J].
Ellis, D ;
Morgan, N .
ICASSP '99: 1999 IEEE INTERNATIONAL CONFERENCE ON ACOUSTICS, SPEECH, AND SIGNAL PROCESSING, PROCEEDINGS VOLS I-VI, 1999, :1013-1016
[5]  
Grezl F., 2009, P INT 2009 SEP, P294
[6]  
HERMANSKY H, 2000, P ICASSP 2000 TURK
[7]  
Joachims T, 1999, MACHINE LEARNING, PROCEEDINGS, P200
[8]   Lightly supervised and unsupervised acoustic model training [J].
Lamel, L ;
Gauvain, JL ;
Adda, G .
COMPUTER SPEECH AND LANGUAGE, 2002, 16 (01) :115-129
[9]  
Malkinn J., 2009, P INTERSPEECH 2009 S
[10]  
Morgan N., 1995, IEEE SIGNAL PROC MAY, V12, P25