[go: up one dir, main page]
More Web Proxy on the site http://driver.im/ skip to main content
research-article

Adaptive Information Bottleneck Guided Joint Source and Channel Coding for Image Transmission

Published: 01 August 2023 Publication History

Abstract

Joint source and channel coding (JSCC) for image transmission has attracted increasing attention due to its robustness and high efficiency. However, the existing deep JSCC research mainly focuses on minimizing the distortion between the transmitted and received information under a fixed number of available channels. Therefore, the transmitted rate may be far more than its required minimum value. In this paper, an adaptive information bottleneck (IB) guided joint source and channel coding (AIB-JSCC) method is proposed for image transmission. The goal of AIB-JSCC is to reduce the transmission rate while improving the image reconstruction quality. In particular, a new IB objective for image transmission is proposed so as to minimize the distortion and the transmission rate. A mathematically tractable lower bound on the proposed objective is derived, and then, adopted as the loss function of AIB-JSCC. To trade off compression and reconstruction quality, an adaptive algorithm is proposed to adjust the hyperparameter of the proposed loss function dynamically according to the distortion during the training. Experimental results show that AIB-JSCC can significantly reduce the required amount of transmitted data and improve the reconstruction quality and downstream task accuracy.

References

[1]
C. E. Shannon, “A mathematical theory of communication,” Bell Syst. Tech. J., vol. 27, no. 3, pp. 379–423, Jul. 1948.
[2]
E. Berlekamp, R. McEliece, and H. van Tilborg, “On the inherent intractability of certain coding problems (Corresp.),” IEEE Trans. Inf. Theory, vol. IT-24, no. 3, pp. 384–386, May 1978.
[3]
W. Saad, M. Bennis, and M. Chen, “A vision of 6G wireless systems: Applications, trends, technologies, and open research problems,” IEEE Netw., vol. 34, no. 3, pp. 134–142, May 2020.
[4]
Y. Yanget al., “Semantic communications with artificial intelligence tasks: Reducing bandwidth requirements and improving artificial intelligence task performance,” IEEE Ind. Electron. Mag., early access, May 26, 2022. 10.1109/MIE.2022.3174331.
[5]
M. Chafii, L. Bariah, S. Muhaidat, and M. Debbah, “Twelve scientific challenges for 6G: Rethinking the foundations of communications theory,” 2022, arXiv:2207.01843.
[6]
A. A. Boulogeorgos, J. M. Jornet, and A. Alexiou, “Directional terahertz communication systems for 6G: Fact check,” IEEE Veh. Technol. Mag., vol. 16, no. 4, pp. 68–77, Dec. 2021.
[7]
M. Sana and E. C. Strinati, “Learning semantics: An opportunity for effective 6G communications,” in Proc. IEEE 19th Annu. Consum. Commun. Netw. Conf. (CCNC), Jan. 2022, pp. 631–636.
[8]
V. Ziegler, H. Viswanathan, H. Flinck, M. Hoffmann, V. Räisänen, and K. Hätönen, “6G architecture to connect the worlds,” IEEE Access, vol. 8, pp. 173508–173520, 2020.
[9]
Z. Wan, Z. Gao, M. Di Renzo, and L. Hanzo, “The road to industry 4.0 and beyond: A communications-, information-, and operation technology collaboration perspective,” 2022, arXiv:2205.04741.
[10]
M. Chenet al., “Distributed learning in wireless networks: Recent progress and future challenges,” IEEE J. Sel. Areas Commun., vol. 39, no. 12, pp. 3579–3605, Dec. 2021.
[11]
R. G. Gallager, Information Theory and Reliable Communication, vol. 2. Cham, Switzerland: Springer, 1968.
[12]
M. Gastpar, B. Rimoldi, and M. Vetterli, “To code, or not to code: Lossy source-channel communication revisited,” IEEE Trans. Inf. Theory, vol. 49, no. 5, pp. 1147–1158, May 2003.
[13]
G. Cheung and A. Zakhor, “Bit allocation for joint source/channel coding of scalable video,” IEEE Trans. Image Process., vol. 9, no. 3, pp. 340–356, Mar. 2000.
[14]
S. Heinen and P. Vary, “Transactions papers source-optimized channel coding for digital transmission channels,” IEEE Trans. Commun., vol. 53, no. 4, pp. 592–600, Apr. 2005.
[15]
E. Bourtsoulatze, D. B. Kurka, and D. Gündüz, “Deep joint source-channel coding for wireless image transmission,” IEEE Trans. Cognit. Commun. Netw., vol. 5, no. 3, pp. 567–579, Sep. 2019.
[16]
D. B. Kurka and D. Gündüz, “DeepJSCC-f: Deep joint source-channel coding of images with feedback,” IEEE J. Sel. Areas Inf. Theory, vol. 1, no. 1, pp. 178–193, May 2020.
[17]
J. Xu, B. Ai, W. Chen, A. Yang, P. Sun, and M. Rodrigues, “Wireless image transmission using deep source channel coding with attention modules,” IEEE Trans. Circuits Syst. Video Technol., vol. 32, no. 4, pp. 2315–2328, Apr. 2022.
[18]
K. Choi, K. Tatwawadi, A. Grover, T. Weissman, and S. Ermon, “Neural joint source-channel coding,” in Proc. Int. Conf. Mach. Learn. Long Beach, CA, USA, Jun. 2019, pp. 1182–1192.
[19]
Y. Song, M. Xu, L. Yu, H. Zhou, S. Shao, and Y. Yu, “Infomax neural joint source-channel coding via adversarial bit flip,” in Proc. AAAI Conf. Artif. Intell. New York, NY, USA, Feb. 2020, pp. 5834–5841.
[20]
C. Szegedyet al., “Going deeper with convolutions,” in Proc. IEEE Conf. Comput. Vis. Pattern Recognit. (CVPR), Jun. 2015, pp. 1–9.
[21]
D. Minnen, J. Ballé, and G. D. Toderici, “Joint autoregressive and hierarchical priors for learned image compression,” in Proc. Neural Inform. Process. Syst., vol. 31. Montreal, QC, Canada, Dec. 2018, pp. 10794–10803.
[22]
J. Devlin, M.-W. Chang, K. Lee, and K. Toutanova, “BERT: Pre-training of deep bidirectional transformers for language understanding,” 2018, arXiv:1810.04805.
[23]
P. Jiang, C. Wen, S. Jin, and G. Y. Li, “Deep source-channel coding for sentence semantic transmission with HARQ,” IEEE Trans. Commun., vol. 70, no. 8, pp. 5225–5240, Aug. 2022.
[24]
Y. Wanget al., “Performance optimization for semantic communications: An attention-based reinforcement learning approach,” IEEE J. Sel. Areas Commun., vol. 40, no. 9, pp. 2598–2613, Sep. 2022.
[25]
N. Tishby, F. C. Pereira, and W. Bialek, “The information bottleneck method,” 2000, arXiv:physics/0004057.
[26]
A. A. Alemi, I. Fischer, J. V. Dillon, and K. Murphy, “Deep variational information bottleneck,” 2016, arXiv:1612.00410.
[27]
R. K. Mahabadi, Y. Belinkov, and J. Henderson, “Variational information bottleneck for effective low-resource fine-tuning,” 2021, arXiv:2106.05469.
[28]
Y. Duet al., “Learning to learn with variational information bottleneck for domain generalization,” in Proc. Eur. Conf. Comput. Vis., Aug. 2020, pp. 200–216.
[29]
N. Tishby and N. Zaslavsky, “Deep learning and the information bottleneck principle,” in Proc. IEEE Inf. Theory Workshop (ITW), Apr. 2015, pp. 1–5.
[30]
J. Lee, J. Choi, J. Mok, and S. Yoon, “Reducing information bottleneck for weakly supervised semantic segmentation,” in Proc. Neural Inform. Process. Syst., vol. 34, Dec. 2021, pp. 27408–27421.
[31]
R. G. Gallager, “Low-density parity-check codes,” IRE Trans. Inf. Theory, vol. 8, no. 1, pp. 21–28, Jan. 1962.
[32]
T. M. Cover, Elements Information Theory. Hoboken, NJ, USA: Wiley, 1999.
[33]
G. Romano and D. Ciuonzo, “Minimum-variance importance-sampling Bernoulli estimator for fast simulation of linear block codes over binary symmetric channels,” IEEE Trans. Wireless Commun., vol. 13, no. 1, pp. 486–496, Jan. 2014.
[34]
K. Podgorski, G. Simons, and Y.-W. Ma, “On estimation for a binary-symmetric channel,” IEEE Trans. Inf. Theory, vol. 44, no. 3, pp. 1260–1272, May 1998.
[35]
W. Huleihel and O. Ordentlich, “How to quantize n outputs of a binary symmetric channel to n–1 bits?” in Proc. IEEE Int. Symp. Inf. Theory (ISIT), Jun. 2017, pp. 91–95.
[36]
D. B. F. Agakov, “The IM algorithm: A variational approach to information maximization,” in Proc. Neural Inform. Process. Syst., Montreal, QC, Canada, Dec. 2004, p. 201.
[37]
A. Mnih and D. J. Rezende, “Variational inference for Monte Carlo objectives,” in Proc. 33rd Int. Conf. Mach. Learn. (ICML). New York, NY, USA, Jun. 2016, pp. 2188–2196.
[38]
P. Cheng, W. Hao, S. Dai, J. Liu, Z. Gan, and L. Carin, “CLUB: A contrastive log-ratio upper bound of mutual information,” in Proc. Int. Conf. Mach. Learn., Vienna, Austria, Jul. 2020, pp. 1779–1788.
[39]
Y. Li, P. Zhao, D. Wang, X. Xian, Y. Liu, and V. S. Sheng, “Learning disentangled user representation based on controllable VAE for recommendation,” in Proc. Int. Conf. Database Syst. Adv. Applicat., Taipei, Taiwan, Apr. 2021, pp. 179–194.
[40]
T. Wu, I. Fischer, I. L. Chuang, and M. Tegmark, “Learnability for the information bottleneck,” in Proc. Uncertainty Artif. Intell. Conf., Toronto, ON, Canada, Jul. 2020, pp. 1050–1060.
[41]
L. Bottou, “Large-scale machine learning with stochastic gradient descent,” in Proc. 19th Int. Conf. Comput. Statist., Paris, France, Aug. 2010, pp. 177–186.
[42]
Y. Lecun, L. Bottou, Y. Bengio, and P. Haffner, “Gradient-based learning applied to document recognition,” Proc. IEEE, vol. 86, no. 11, pp. 2278–2324, Nov. 1998.
[43]
B. M. Lake, R. Salakhutdinov, and J. B. Tenenbaum, “Human-level concept learning through probabilistic program induction,” Science, vol. 350, no. 6266, pp. 1332–1338, Dec. 2015.
[44]
A. Krizhevsky and G. Hinton, “Learning multiple layers of features from tiny images,” M.S. thesis, 2009. [Online]. Available: https://www.cs.toronto.edu/-kriz/learning-features-2009-TR.pdf
[45]
Y. Netzer, T. Wang, A. Coates, A. Bissacco, B. Wu, and A. Y. Ng, “Reading digits in natural images with unsupervised feature learning,” in Proc. NIPS Workshop, Granada, Spain, Dec. 2011, pp. 3730–3738.
[46]
G. K. Wallace, “The JPEG still picture compression standard,” IEEE Trans. Consum. Electron., vol. 38, no. 1, pp. 18–34, Feb. 1992.
[47]
M. Rabbani, Book Review: JPEG2000: Image Compression Fundamentals, Standards and Practice. Bellingham, WA, USA: SPIE, 2002.
[48]
Google. (2015). WebP Compression Study. [Online]. Available: https://developers.google.com/speed/webp/docs/webpstudy
[49]
F. Bellard. (2018). BPG Image Format. [Online]. Available: https://bellard.org/bpg/
[50]
Z. Cheng, H. Sun, M. Takeuchi, and J. Katto, “Learned image compression with discretized Gaussian mixture likelihoods and attention modules,” in Proc. IEEE/CVF Conf. Comput. Vis. Pattern Recognit. (CVPR), Jun. 2020, pp. 7936–7945.
[51]
F. Mentzer, G. D. Toderici, M. Tschannen, and E. Agustsson, “High-fidelity generative image compression,” in Proc. Neural Inf. Process. Syst., Dec. 2020, pp. 11913–11924.
[52]
M. Lentmaier, A. Sridharan, D. J. Costello, and K. S. Zigangirov, “Iterative decoding threshold analysis for LDPC convolutional codes,” IEEE Trans. Inf. Theory, vol. 56, no. 10, pp. 5274–5289, Oct. 2010.
[53]
J. Hagenauer, E. Offer, and L. Papke, “Iterative decoding of binary block and convolutional codes,” IEEE Trans. Inf. Theory, vol. 42, no. 2, pp. 429–445, Mar. 1996.
[54]
H. Vikalo, B. Hassibi, and T. Kailath, “Iterative decoding for MIMO channels via modified sphere decoding,” IEEE Trans. Wireless Commun., vol. 3, no. 6, pp. 2299–2311, Nov. 2004.
[55]
L. Van der Maaten and G. Hinton, “Visualizing data using t-SNE,” J. Mach. Learn. Res., vol. 9, no. 11, pp. 1–27, Nov. 2008.

Cited By

View all
  • (2024)Two-View Image Semantic Cooperative Nonorthogonal Transmission in Distributed Edge NetworksInternational Journal of Intelligent Systems10.1155/int/50810172024Online publication date: 1-Jan-2024

Recommendations

Comments

Please enable JavaScript to view thecomments powered by Disqus.

Information & Contributors

Information

Published In

cover image IEEE Journal on Selected Areas in Communications
IEEE Journal on Selected Areas in Communications  Volume 41, Issue 8
Aug. 2023
426 pages

Publisher

IEEE Press

Publication History

Published: 01 August 2023

Qualifiers

  • Research-article

Contributors

Other Metrics

Bibliometrics & Citations

Bibliometrics

Article Metrics

  • Downloads (Last 12 months)0
  • Downloads (Last 6 weeks)0
Reflects downloads up to 03 Feb 2025

Other Metrics

Citations

Cited By

View all
  • (2024)Two-View Image Semantic Cooperative Nonorthogonal Transmission in Distributed Edge NetworksInternational Journal of Intelligent Systems10.1155/int/50810172024Online publication date: 1-Jan-2024

View Options

View options

Figures

Tables

Media

Share

Share

Share this Publication link

Share on social media