A psychoacoustic model for audio coding based on a cochlear filter bank

被引:0
作者
Baumgarte, F [1 ]
机构
[1] Agere Syst, Media Signal Proc Res, Murray Hill, NJ 07974 USA
来源
PROCEEDINGS OF THE 2001 IEEE WORKSHOP ON THE APPLICATIONS OF SIGNAL PROCESSING TO AUDIO AND ACOUSTICS | 2001年
关键词
D O I
10.1109/ASPAA.2001.969562
中图分类号
O42 [声学];
学科分类号
070206 ; 082403 ;
摘要
Perceptual audio coders use an estimated masked threshold for the determination of the maximum permissible just-inaudible noise level introduced by quantization. This estimate is derived from a psychoacoustic model mimicking the psychoacoustics of masking. Current applications use a uniform spectral decomposition as first stage of that model to approximate the frequency selectivity of the human auditory system. The availability of efficient implementations led to a virtually exclusive use of uniform decompositions in audio coding. However, the equal filter properties of the uniform sub-bands are not in line with the nonuniform auditory filters. This paper presents a psychoacoustic model based on an efficient nonuniform cochlear filter bank with a simplified less complex post-processing for estimating the masked threshold. Application results in audio coding show a significantly better performance in terms of bit rate and/or quality of the new model in comparison with other state-of-the-art models with a uniform spectral decomposition.
引用
收藏
页码:139 / 142
页数:4
相关论文
共 10 条
  • [1] BAUMGARTE F, 1995, 99 AES CONV NEW YORK
  • [2] BAUMGARTE F, 1997, 103 AES CONV NEW YOR
  • [3] BAUMGARTE F, 2001, P ICASSP 2001 SALT L
  • [4] BAUMGARTE F, 2000, PHYSL GEHORMODELL NA
  • [5] BRANDENBURG K, 1994, J AUDIO ENG SOC, V42, P780
  • [6] EDLER B, 2000, P ICASSP 2001 IST JU, P1881
  • [7] DERIVATION OF AUDITORY FILTER SHAPES FROM NOTCHED-NOISE DATA
    GLASBERG, BR
    MOORE, BCJ
    [J]. HEARING RESEARCH, 1990, 47 (1-2) : 103 - 138
  • [8] *ISO IEC, 1997, 138187 ISO IEC
  • [9] SINHA D, 1998, DIGIT SIGNAL PROCESS, V42, P42
  • [10] Zwicker E., 1999, PSYCHOACOUSTICS, DOI [10.1007/978-3-662-09562-1, DOI 10.1007/978-3-662-09562-1]