[go: up one dir, main page]
More Web Proxy on the site http://driver.im/ skip to main content
research-article

Rate-Distortion-Perception Tradeoff Based on the Conditional-Distribution Perception Measure

Published: 25 September 2024 Publication History

Abstract

This paper studies the rate-distortion-perception (RDP) tradeoff for a memoryless source model in the asymptotic limit of large block-lengths. The perception measure is based on a divergence between the distributions of the source and reconstruction sequences conditioned on the encoder output, first proposed by Mentzer et al. We consider the case when there is no shared randomness between the encoder and the decoder and derive a single-letter characterization of the RDP function, for the case of discrete memoryless sources. This is in contrast to the marginal-distribution metric case (introduced by Blau and Michaeli), whose RDP characterization remains open when there is no shared randomness. The achievability scheme is based on lossy source coding with a posterior reference map. For the case of continuous valued sources under the squared error distortion measure and the squared quadratic Wasserstein perception measure, we also derive a single-letter characterization and show that the decoder can be restricted to a noise-adding mechanism. Interestingly, the RDP function characterized for the case of zero perception loss coincides with that of the marginal metric, and further zero perception loss can be achieved with a 3-dB penalty in minimum distortion. Finally we specialize to the case of Gaussian sources, and derive the RDP function for Gaussian vector case and propose a reverse water-filling type solution. We also partially characterize the RDP function for a mixture of Gaussian vector sources.

References

[1]
Y. Blau and T. Michaeli, “Rethinking lossy compression: The rate-distortion-perception tradeoff,” in Proc. Int. Conf. Mach. Learn., 2019, pp. 675–685.
[2]
T. M. Cover and J. A. Thomas, Elements of Information Theory. Hoboken, NJ, USA: Wiley, 1991.
[3]
Y. Blau and T. Michaeli, “The perception-distortion tradeoff,” in Proc. IEEE/CVF Conf. Comput. Vis. Pattern Recognit., Jun. 2018, pp. 6228–6237.
[4]
L. Theis and A. Wagner, “A coding theorem for the rate-distortion-perception function,” in Proc. Int. Conf. Learn. Represent. (ICLR), 2021, pp. 1–5.
[5]
G. Zhang, J. Qian, J. Chen, and A. Khisti, “Universal rate-distortion-perception representations for lossy compression,” in Proc. Adv. Neural Inf. Process. Syst., vol. 34, 2021, pp. 11517–11529.
[6]
R. Matsumoto, “Introducing the perception-distortion tradeoff into the rate-distortion theory of general information sources,” IEICE Commun. Exp., vol. 7, no. 11, pp. 427–431, 2018.
[7]
R. Matsumoto, “Rate-distortion-perception tradeoff of variable-length source coding for general information sources,” IEICE Commun. Exp., vol. 8, no. 2, pp. 38–42, 2019.
[8]
A. B. Wagner, “The rate-distortion-perception tradeoff: The role of common randomness,” 2022, arXiv:2202.04147.
[9]
J. Chen, L. Yu, J. Wang, W. Shi, Y. Ge, and W. Tong, “On the rate-distortion-perception function,” IEEE J. Sel. Areas Inf. Theory, vol. 3, no. 4, pp. 664–673, Dec. 2022.
[10]
N. Saldi, T. Linder, and S. Yüksel, “Output constrained lossy source coding with limited common randomness,” IEEE Trans. Inf. Theory, vol. 61, no. 9, pp. 4984–4998, Sep. 2015.
[11]
X. Niu, D. Gündüz, B. Bai, and W. Han, “Conditional rate-distortion-perception trade-off,” 2023, arXiv:2305.09318.
[12]
Y. Hamdi and D. Gündüz, “The rate-distortion-perception trade-off with side information,” 2023, arXiv:2305.13116.
[13]
F. Mentzer, E. Agustsson, J. Ballé, D. Minnen, N. Johnston, and G. Toderici, “Neural video compression using GANs for detail synthesis and propagation,” in Proc. Eur. Conf. Comput. Vis. (ECCV), 2022, pp. 562–578.
[14]
E. Agustsson, M. Tschannen, F. Mentzer, R. Timofte, and L. Van Gool, “Generative adversarial networks for extreme learned image compression,” in Proc. IEEE/CVF Int. Conf. Comput. Vis. (ICCV), Oct. 2019, pp. 221–231.
[15]
J. Ballé, V. Laparra, and E. P. Simoncelli, “End-to-end optimized image compression,” in Proc. Int. Conf. Learn. Represent. (ICLR), 2017, pp. 1–27.
[16]
L. Theis, W. Shi, A. Cunningham, and F. Huszár, “Lossy image compression with compressive autoencoders,” in Proc. Int. Conf. Learn. Represent. (ICLR), 2017, pp. 1–19.
[17]
F. Mentzer, E. Agustsson, M. Tschannen, R. Timofte, and L. V. Gool, “Conditional probability models for deep image compression,” in Proc. IEEE/CVF Conf. Comput. Vis. Pattern Recognit., Jun. 2018, pp. 4394–4402.
[18]
F. Mentzer, G. Toderici, M. Tschannen, and E. Agustsson, “High-fidelity generative image compression,” in Proc. Adv. Neural Inf. Process. Syst. (NeurIPS), 2020, pp. 11913–11924.
[19]
A. Golinski, R. Pourreza, Y. Yang, G. Sautiere, and T. S. Cohen, “Feedback recurrent autoencoder for video compression,” in Proc. Asian Conf. Comput. Vis., 2020, pp. 1–17.
[20]
H. Liu, G. Zhang, J. Chen, and A. Khisti, “Lossy compression with distribution shift as entropy constrained optimal transport,” in Proc. Int. Conf. Learn. Represent. (ICLR), 2022, pp. 1–34.
[21]
T. Xu et al., “Conditional perceptual quality preserving image compression,” 2023, arXiv:2308.08154.
[22]
T. A. Atif, M. A. Sohail, and S. S. Pradhan, “Lossy quantum source coding with a global error criterion based on a posterior reference map,” 2023, arXiv:2302.00625.
[23]
Z. Yan, F. Wen, R. Ying, C. Ma, and P. Liu, “On perceptual lossy compression: The cost of perceptual reconstruction and an optimal training framework,” in Proc. Int. Conf. Mach. Learn. (ICML), 2021, pp. 11682–11692.
[24]
L. Theis and E. Agustsson, “On the advantages of stochastic encoders,” in Proc. Int. Conf. Learn. Represent. (ICLR), 2021, pp. 1–8.
[25]
A. El Gamal and Y. H. Kim, Network Information Theory. Cambridge, U.K.: Cambridge Univ. Press, 2011.
[26]
D. Freirich, T. Michaeli, and R. Meir, “A theory of the distortion-perception tradeoff in Wasserstein space,” in Proc. Adv. Neural Inf. Process. Syst. (NeurIPS), vol. 34, 2021, pp. 25661–25672.
[27]
Z. Yan, F. Wen, and P. Liu, “Optimally controllable perceptual lossy compression,” in Proc. Int. Conf. Mach. Learn. (ICML), 2022, pp. 1–18.
[28]
J. Qian et al., “Rate-distortion-perception tradeoff for Gaussian vector sources,” 2024, arXiv:2406.18008.
[29]
D. C. Dowson and B. V. Landau, “The Fréchet distance between multivariate normal distributions,” J. Multivariate Anal., vol. 12, no. 3, pp. 450–455, Sep. 1982.
[30]
R. A. Horn and C. R. Johnson, Matrix Analysis. Cambridge, U.K.: Cambridge Univ. Press, 1985.
[31]
C. R. Givens and R. M. Shortt, “A class of Wasserstein metrics for probability distributions,” Michigan Math. J., vol. 31, no. 2, pp. 231–240, Jan. 1984.

Index Terms

  1. Rate-Distortion-Perception Tradeoff Based on the Conditional-Distribution Perception Measure
          Index terms have been assigned to the content through auto-classification.

          Recommendations

          Comments

          Please enable JavaScript to view thecomments powered by Disqus.

          Information & Contributors

          Information

          Published In

          cover image IEEE Transactions on Information Theory
          IEEE Transactions on Information Theory  Volume 70, Issue 12
          Dec. 2024
          936 pages

          Publisher

          IEEE Press

          Publication History

          Published: 25 September 2024

          Qualifiers

          • Research-article

          Contributors

          Other Metrics

          Bibliometrics & Citations

          Bibliometrics

          Article Metrics

          • 0
            Total Citations
          • 0
            Total Downloads
          • Downloads (Last 12 months)0
          • Downloads (Last 6 weeks)0
          Reflects downloads up to 11 Dec 2024

          Other Metrics

          Citations

          View Options

          View options

          Login options

          Media

          Figures

          Other

          Tables

          Share

          Share

          Share this Publication link

          Share on social media