Neural Estimation of the Rate-Distortion Function With Applications to Operational Source Coding

被引:3
作者
Lei E. [1 ]
Hassani H. [1 ]
Bidokhti S.S. [1 ]
机构
[1] University of Pennsylvania, Department of Electrical and Systems Engineering, Philadelphia, 19104, PA
来源
IEEE Journal on Selected Areas in Information Theory | 2022年 / 3卷 / 04期
关键词
Generative models; lossy compression; neural compression; rate-distortion theory; reverse channel coding;
D O I
10.1109/JSAIT.2023.3273467
中图分类号
学科分类号
摘要
A fundamental question in designing lossy data compression schemes is how well one can do in comparison with the rate-distortion function, which describes the known theoretical limits of lossy compression. Motivated by the empirical success of deep neural network (DNN) compressors on large, real-world data, we investigate methods to estimate the rate-distortion function on such data, which would allow comparison of DNN compressors with optimality. While one could use the empirical distribution of the data and apply the Blahut-Arimoto algorithm, this approach presents several computational challenges and inaccuracies when the datasets are large and high-dimensional, such as the case of modern image datasets. Instead, we re-formulate the rate-distortion objective, and solve the resulting functional optimization problem using neural networks. We apply the resulting rate-distortion estimator, called NERD, on popular image datasets, and provide evidence that NERD can accurately estimate the rate-distortion function. Using our estimate, we show that the rate-distortion achievable by DNN compressors are within several bits of the rate-distortion function for real-world datasets. Additionally, NERD provides access to the rate-distortion achieving channel, as well as samples from its output marginal. Therefore, using recent results in reverse channel coding, we describe how NERD can be used to construct an operational one-shot lossy compression scheme with guarantees on the achievable rate and distortion. Experimental results demonstrate competitive performance with DNN compressors. © 2020 IEEE.
引用
收藏
页码:674 / 686
页数:12
相关论文
共 42 条
  • [1] Balle J., Laparra V., Simoncelli E.P., End-to-end optimized image compression, Proc. 5th Int. Conf. Learn. Represent., Toulon, France, (2017)
  • [2] Theis L., Shi W., Cunningham A., Huszar F., Lossy image compression with compressive autoencoders, Proc. Int. Conf. Learn. Represent, (2017)
  • [3] Agustsson E., Et al., Soft-to-hard vector quantization for end-toend learning compressible representations, Proc. NIPS, pp. 1141-1151, (2017)
  • [4] Balle J., Et al., Nonlinear transform coding, IEEE Trans. Special Topics Signal Process, 15, 2, pp. 339-353, (2021)
  • [5] Wagner A.B., Balle J., Neural networks optimally compress the sawbridge, Proc. Data Compress. Conf. (DCC), pp. 143-152, (2021)
  • [6] LeCun Y., Bengio Y., Hinton G., Deep learning, Nature, 521, 7553, pp. 436-444, (2015)
  • [7] Li C.T., Gamal A.E., Strong functional representation lemma and applications to coding theorems, Proc. IEEE Int. Symp. Inf. Theory (ISIT), pp. 589-593, (2017)
  • [8] Theis L., Yosri N., Algorithms for the communication of samples, (2021)
  • [9] Harsha P., Jain R., McAllester D., Radhakrishnan J., The communication complexity of correlation, IEEE Trans. Inf. Theory, 56, 1, pp. 438-449, (2010)
  • [10] Bennett C., Shor P., Smolin J., Thapliyal A., Entanglement-assisted capacity of a quantum channel and the reverse Shannon theorem, IEEE Trans. Inf. Theory, 48, 10, pp. 2637-2655, (2002)