Information transmission analysis for continuous speech features

被引:2
作者
Oosthuizen, Dirk J. J. [1 ]
Hanekom, Johan J. [1 ]
机构
[1] Univ Pretoria, Dept Elect Elect & Comp Engn, Univ Rd, ZA-0002 Pretoria, South Africa
关键词
MUTUAL INFORMATION; RECOGNITION; CLASSIFICATION; VOWEL; ALGORITHMS;
D O I
10.1016/j.specom.2016.06.003
中图分类号
O42 [声学];
学科分类号
070206 ; 082403 ;
摘要
Acoustic features are important for the study of human speech perception and the development of computational speech classification algorithms. These areas can benefit from a technique that accurately quantifies the information content of an individual feature, as well as the degree to which this information is used by a listener or algorithm. Feature information transmission analysis (FITA) was developed to do this for categorical features, but complications arise when applying it to continuous features. Absolute information measures are bounded from above by values substantially lower than their theoretical maxima, the precision with which a feature is produced is ignored and a tedious manual step is introduced into the analysis process. This article presents an alternative approach (the continuous FITA) that addresses these complications effectively by representing continuous features in a more natural way. It is shown that this approach is well-suited to continuous feature information analysis and, furthermore, can be used to estimate redundancy in multiple features and information transmitted by combinations of features. The continuous FITA can quantify information in extracted features before an identification experiment has been conducted. It can aid in feature selection for computational speech classification systems, measure feature utilization by humans as well as computational algorithms and facilitate the investigation of the effect of noise or signal processing on feature information. (C) 2016 Elsevier B.V. All rights reserved.
引用
收藏
页码:53 / 66
页数:14
相关论文
共 33 条
[1]   Automated aural classification used for inter-species discrimination of cetaceans [J].
Binder, Carolyn M. ;
Hines, Paul C. .
JOURNAL OF THE ACOUSTICAL SOCIETY OF AMERICA, 2014, 135 (04) :2113-2125
[2]  
BLAMEY P J, 1989, Journal of Rehabilitation Research and Development, V26, P15
[3]  
Boersma P., 2018, Praat: doing phonetics by computer (Version 5.3) Computer software
[4]  
Botha Liesbeth., 1996, P I PHON SCI U AMST, V20, P65
[5]  
Carlson R., 1979, STL QPSR, V34, P19
[6]   Semi-automatic classification of bird vocalizations using spectral peak tracks [J].
Chen, Zhixin ;
Maher, Robert C. .
JOURNAL OF THE ACOUSTICAL SOCIETY OF AMERICA, 2006, 120 (05) :2974-2984
[7]   High-dimensional integration: The quasi-Monte Carlo way [J].
Dick, Josef ;
Kuo, Frances Y. ;
Sloan, Ian H. .
ACTA NUMERICA, 2013, 22 :133-288
[8]   Within-Subjects Comparison of the HiRes and Fidelity120 Speech Processing Strategies: Speech Perception and Its Relation to Place-Pitch Sensitivity [J].
Donaldson, Gail S. ;
Dawson, Patricia K. ;
Borden, Lamar Z. .
EAR AND HEARING, 2011, 32 (02) :238-250
[9]   ACOUSTIC CHARACTERISTICS OF AMERICAN ENGLISH VOWELS [J].
HILLENBRAND, J ;
GETTY, LA ;
CLARK, MJ ;
WHEELER, K .
JOURNAL OF THE ACOUSTICAL SOCIETY OF AMERICA, 1995, 97 (05) :3099-3111
[10]   Speech perception based on spectral peaks versus spectral shape [J].
Hillenbrand, James M. ;
Houde, Robert A. ;
Gayvert, Robert T. .
JOURNAL OF THE ACOUSTICAL SOCIETY OF AMERICA, 2006, 119 (06) :4041-4054