Highly Efficient Lossless Coding for High Dynamic Range Red, Clear, Clear, Clear Image Sensors
<p>Piecewise linear representation in HDR image sensors.</p> "> Figure 2
<p>Color filter arrays: (<b>a</b>) monochrome (CCCC), (<b>b</b>) RGGB, (<b>c</b>) RCCC, (<b>d</b>) RCCB, (<b>e</b>) RGBC, (<b>f</b>) RYYC.</p> "> Figure 3
<p>Three sample parts of source RCCC image.</p> "> Figure 4
<p>Three sample parts of source RCCC image decomposed into two images with R and CCCC components (the second image is created with the CCC components while the missing values are interpolated).</p> "> Figure 5
<p>Three sample parts of source RCCC image decomposed into four images with all components R, C1, C2, C3 separated.</p> "> Figure 6
<p>Three sample parts of source RCCC image decomposed into three images: two small images with R and C1 components and a horizontally two times larger image comprising C2 and C3 components.</p> "> Figure 7
<p>RCCC lossless compression procedures with FFV1 codec for separated R, C1 and joined C2 and C3 components in: (<b>a</b>) intra-frame mode, (<b>b</b>) inter-frame mode.</p> ">
Abstract
:1. Introduction
2. Lossless Compression for ADAS and ADS
3. New Class of HDR Image Sensors
3.1. High-Dynamic Range Image Sensors
3.2. Novel Color Optical Filter Arrays
4. State-of-The-Art in the Lossless Compression of Images
5. Proposal of Four Procedures for Lossless Compression of Specific HDR RCCC Image Format
- First, RCCC image may be treated as a monochrome image and encoded directly (see Figure 3). Notice that the R component is spatially sub-sampled (consecutive R values do not occur in the direct neighborhood) and the difference between R and C components may be significant. Certainly, the same phenomenon is also valid for conventional video formats [20]. This reduces accuracy of prediction based on neighboring pixels. Horizontal neighborhood of C2 and C3 components (see Figure 2c) is closer than the vertical one, thus the prediction using the typical prediction masks is less accurate.
- Second, due to the informative differences between R and C components and, in consequence, expected errors in their joint prediction, R and CCC components can be decomposed into two separate images. Resolution of the image created with R components only is certainly ¼ of the resolution of the source RCCC image. We suggest that in the second image, created from the CCC components, the missing values (i.e., those previously occupied with R component values) should interpolated, resulting in the CCCC image preserving the original resolution (see Figure 4). In this method, the encoder has by 25% more data to encode than in the previous method, but inside the decomposed images the pixels are more similar in the neighborhood and the overall compression can be more effective. There are various ways possible to interpolate the missing C values. In order to maximize the compression ratio we propose to calculate them just with the predictor. In this way, the prediction errors are equal to zero for the interpolated C values.
- Third, all four components, i.e., R, C1, C2, C3 can be decomposed into four separate images with equal resolutions of ¼ of the source RCCC image resolution (see Figure 5). Compression of single components results in small prediction errors, but we lose information about the closest neighborhood between components C2 and C3. However, this method is the most regular and universal as it can be directly used to any color filter mosaic not only to RCCC (cf. Figure 2).
- Last but not least, components of the RCCC source image can be decomposed into three images: two small images R, C1 (both are just the same as in the previously described possibility) and to a horizontally two times larger image comprising C2 and C3 components (see Figure 6). Compression of C2 and C3 components together, due to their direct proximity, may be more efficient than the separate one.
6. Experiments on Lossless HDR RCCC Codec and Implementation of Coding Procedure
6.1. Methodology
- compression ratio:
- compression throughput of the input stream:
- decompression throughput of the output stream:
6.2. Lossless Video Codec
6.3. RCCC Video Database
6.4. Testbed and Experimental Setup
6.5. Selecting the Best of Lossless RCCC Codec Procedures
- The codec is designed and optimized for typical color images, i.e., RGGB, not RCCC. The misprediction rate during the compression is higher.
- In RGGB with 3 × 8 bit per pixel, there is only one source component, i.e., R, G, or B for one pixel (cf. Figure 2b). To obtain the full RGB pixel, two remaining components are to be interpolated. Interpolation means that there is no additional source information. Therefore the prediction is better, and in consequence the compression ratio, is higher. In the RCCC format each pixel carries independently measured information and the correlation between values is much smaller than in the case of the RGGB format.
- Good quality 8-bit per component images contain typically less noise than HDR ones. Due to the unpredictable noise, the prediction during the image compression is much less accurate. To prove it we additionally performed a special test for 8 bpp SDR RCCC images. The SDR RCCC images were obtained from the 12-bit PWL bit-length compressed HDR source sequences by removing the least four bits. In this case we achieved average value of CR equal to 3.4, which is similar to 3.3, reported in [15]. This shows that even for untypical RCCC format it is possible to achieve high CR values, comparable to the potential of FFV1 codec. On the other hand, the much lower values of CR for 12 bpp RCCC images confirm that the last four bits in the test recordings are really noised.
6.6. Efficient Implementation of the Best Lossless HDR RCCC Codec Procedure
6.7. Detailed Tests of the Final Implementation of Lossless HDR RCCC Codec Procedure
7. Conclusions
Author Contributions
Funding
Informed Consent Statement
Data Availability Statement
Conflicts of Interest
References
- Awan, F.M.; Saleem, Y.; Minerva, R.; Crespi, N. A Comparative Analysis of Machine/Deep Learning Models for Parking Space Availability Prediction. Sensors 2020, 20, 322. [Google Scholar] [CrossRef] [PubMed] [Green Version]
- Dabral, S.; Kamath, S.; Appia, V.; Mody, M.; Zhang, B.; Batur, U. Trends in camera based Automotive Driver Assistance Systems (ADAS). In Proceedings of the 2014 IEEE 57th International Midwest Symposium on Circuits and Systems (MWSCAS), College Station, TX, USA, 3–6 August 2014; pp. 1110–1115. [Google Scholar]
- Piniarski, K.; Pawłowski, P. Efficient pedestrian detection with enhanced object segmentation in far IR night vision. In Proceedings of the Signal Processing: Algorithms, Architectures, Arrangements, and Applications (SPA), Poznan, Poland, 20–22 September 2017; pp. 160–165. [Google Scholar]
- Komorkiewicz, M.; Turek, K.; Skruch, P.; Kryjak, T.; Gorgon, M. FPGA-based Hardware-in-the-Loop environment using video injection concept for camera-based systems in automotive applications. In Proceedings of the Confrence on Design and Architectures for Signal and Image Processing (DASIP), Rennes, France, 12–14 October 2016; pp. 183–190. [Google Scholar]
- Shen, X.; Chong, Z.J.; Pendleton, S.; Fu, G.M.J.; Qin, B.; Frazzoli, E.; Ang, M.H., Jr. Teleoperation of On-Road Vehicles via Immersive Telepresence Using Off-the-shelf Components. In Intelligent Autonomous Systems 13 Book; Springer International Publishing: Berlin/Heidelberg, Germany, 2016; pp. 1419–1433. [Google Scholar]
- Hosseini, A.; Lienkamp, M. Enhancing telepresence during the teleoperation of road vehicles using HMD-based mixed reality. In Proceedings of the IEEE Intelligent Vehicles Symposium, Gothenburg, Sweden, 19–22 June 2016; pp. 1366–1373. [Google Scholar]
- Asif, M.T.; Srinivasan, K.; Mitrovic, N.; Dauwels, J.; Jaillet, P. Near-Lossless Compression for Large Traffic Networks. IEEE Trans. Intell. Transp. Syst. 2015, 16, 1817–1826. [Google Scholar] [CrossRef] [Green Version]
- Agarwal, S.; Regentova, E.E.; Kachroo, P.; Verma, H. Multidimensional Compression of ITS Data Using Wavelet-Based Compression Techniques. IEEE Trans. Intell. Transp. Syst. 2017, 18, 1907–1917. [Google Scholar] [CrossRef]
- Hihara, H.; Moritani, K.; Inoue, M.; Hoshi, Y.; Iwasaki, A.; Takada, J.; Inada, H.; Suzuki, M.; Seki, T.; Ichikawa, S.; et al. Onboard Image Processing System for Hyperspectral Sensor. Sensors 2015, 15, 24926–24944. [Google Scholar] [CrossRef] [PubMed] [Green Version]
- Konstantinos, K.; Dianati, M. A Conceptual 5G Vehicular Networking Architecture. In 5G Mobile Communications Book; Springer International Publishing: Berlin/Heidelberg, Germany, 2017; pp. 595–623. [Google Scholar]
- Piniarski, K.; Pawłowski, P.; Dąbrowski, A. Efficient HDR tone-mapping for ADAS applications. In Proceedings of the 2019 Signal Processing: Algorithms, Architectures, Arrangements, and Applications (SPA), Poznan, Poland, 18–20 September 2019; pp. 325–330. [Google Scholar]
- Karanam, G. Interfacing Red/Clear Sensors to ADSP-BF609® Blackfin Processors, Analog Devices, Inc., Technical Notes, EE-358. May 2013, pp. 1–4. Available online: https://www.analog.com/media/en/technical-documentation/application-notes/EE358.pdf (accessed on 19 January 2021).
- Huang, H.; Lee, C.; Lin, H. Nighttime vehicle detection and tracking base on spatiotemporal analysis using RCCC sensor. In Proceedings of the IEEE 9th International Conference on Humanoid, Nanotechnology, Information Technology, Communication and Control, Environment and Management (HNICEM), Manila, Philippines, 1–3 December 2017; pp. 1–5. [Google Scholar]
- Lin, H.; Liao, P.; Chang, Y. Long-Distance Vehicle Detection Algorithm at Night for Driving Assistance. In Proceedings of the 2018 3rd IEEE International Conference on Intelligent Transportation Engineering (ICITE), Singapore, 3–5 September 2018; pp. 296–300. [Google Scholar]
- Pawłowski, P.; Piniarski, K.; Dąbrowski, A. Selection and tests of lossless and lossy video codecs for advanced driver-assistance systems. In Proceedings of the 2018 3rd IEEE International Conference on Intelligent Transportation Engineering (ICITE), Singapore, 3–5 September 2018; pp. 344–349. [Google Scholar]
- Ma, S.; Zhang, X.; Jia, C.; Zhao, Z.; Wang, S.; Wang, S. Image and Video Compression with Neural Networks: A Review. IEEE Trans. Circuits Syst. Video Technol. 2020, 30, 1683–1698. [Google Scholar] [CrossRef] [Green Version]
- Birman, R.; Segal, Y.; Hadar, O. Overview of Research in the field of Video Compression using Deep Neural Networks. Multimed. Tools Appl. 2020, 79, 11699–11722. [Google Scholar] [CrossRef]
- Babu, R.V.; Tom, M.; Wadekar, P. A survey on compressed domain video analysis techniques. Multimed. Tools Appl. 2016, 75, 1043–1078. [Google Scholar] [CrossRef]
- Choi, H.; Bajic, I.V. High efficiency compression for object detection. In Proceedings of the 2018 IEEE International Conference on Acoustics, Speech and Signal Processing (ICASSP), Calgary, AB, Canada, 15–20 April 2018; pp. 1729–1796. [Google Scholar]
- Kong, L.; Dai, R. Temporal-Fluctuation-Reduced Video Encoding for Object Detection in Wireless Surveillance Systems. In Proceedings of the IEEE International Symposium on Multimedia (ISM), San Jose, CA, USA, 11–13 December 2016; pp. 126–132. [Google Scholar]
- Molenaar, R.; van Bilsen, A.; van der Made, R.; de Vries, R. Full spectrum camera simulation for reliable virtual development and validation of ADAS and automated driving applications. In Proceedings of the IEEE Intelligent Vehicles Symposium (IV), Seoul, Korea, 28 June–1 July 2015; pp. 47–52. [Google Scholar]
- STMicroelectronics, VG6640, VD6640, Automotive 1.3 Megapixel High-Dynamic Range Image Sensor, Datasheet. 2018. Available online: https://www.st.com/resource/en/datasheet/vg6640.pdf (accessed on 8 December 2020).
- On Semiconductor, AR0132AT 1/3-Inch CMOS Digital Image Sensor, Datasheet, Rev. 9. 2016. Available online: https://www.onsemi.com/products/sensors/image-sensors-processors/image-sensors/ar0132at (accessed on 19 January 2021).
- Oh, M.; Velichko, S.; Johnson, S.; Guidash, M.; Chang, H.-C.; Tekleab, D.; Gravelle, B.; Nicholes, S.; Suryadevara, M.; Collins, D.; et al. Automotive 3.0 µm Pixel High Dynamic Range Sensor with LED Flicker Mitigation. Sensors 2020, 20, 1390. [Google Scholar] [CrossRef] [PubMed] [Green Version]
- Eilertsen, G.; Mantiuk, R.K.; Unger, J. A comparative review of tone-mapping algorithms for high dynamic range video. Comput. Graph. Forum 2017, 36, 565–592. [Google Scholar] [CrossRef] [Green Version]
- Thanigasalam, H. Imaging Interface Advancements and Development to Meet the Needs of Mobile and Mobile-Influenced Industries, MIPI Alliance. 2016. Available online: https://www.mipi.org/sites/default/files/Feb17-2016-MIPI-CSI-Webinar-Mobile-Beyond-Imaging-Interface-Development.pdf (accessed on 19 January 2021).
- Eytan, O.; Belman, E. High-Resolution Automotive Lens and Sensor. U.S. Patent Application Publication US 2019/0377110A1, 12 December 2019. [Google Scholar]
- Hussain, A.J.; Al-Fayadh, A.; Radi, N. Image compression techniques: A survey in lossless and lossy algorithms. Neurocomputing 2018, 300, 44–69. [Google Scholar] [CrossRef]
- Heindel, A.; Wige, E.; Kaup, A. Low-Complexity Enhancement Layer Compression for Scalable Lossless Video Coding Based on HEVC. IEEE Trans. Circuits Syst. Video Technol. 2017, 27, 1749–1760. [Google Scholar] [CrossRef]
- Choi, J.A.; Ho, Y.S. Efficient residual data coding in CABAC for HEVC lossless video compression. SIViP 2015, 9, 1055–1066. [Google Scholar] [CrossRef]
- Wang, Z.; Hou, Z.; Hu, R.; Xiao, J. A Lossless Recompression Approach for Video Streaming Transmission. IEEE Access 2019, 7, 35162–35172. [Google Scholar] [CrossRef]
- Bui, V.; Chang, L.; Li, D.; Hsu, L.; Chen, M.Y. Comparison of lossless video and image compression codecs for medical computed tomography datasets. In Proceedings of the 2016 IEEE International Conference on Big Data (Big Data), Washington, DC, USA, 5–8 December 2016; pp. 3960–3962. [Google Scholar] [CrossRef]
- Mentzer, F.; Agustsson, E.; Tschannen, M.; Timofte, R.; Van Gool, L. Practical Full Resolution Learned Lossless Image Compression. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR), Long Beach, CA, USA, 16–20 June 2019; pp. 10629–10638. [Google Scholar]
- Trifan, A.; Neves, A. A Survey on Lossless Compression of Bayer Color Filter Array Images. Int. J. Comput. Electr. Autom. Control Inf. Eng. 2016, 10, 585–590. [Google Scholar]
- Bazhyna, A.; Egiazarian, K. Lossless and near lossless compression of real color filter array data. IEEE Trans. Consum. Electron. 2008, 54, 1492–1500. [Google Scholar] [CrossRef]
- Richter, T.; Fößel, S. Bayer Pattern Compression with JPEG XS. In Proceedings of the 2019 IEEE International Conference on Image Processing (ICIP), Taipei, Taiwan, 22–25 September 2019; pp. 3177–3181. [Google Scholar] [CrossRef]
- Niedermayer, M. FFV1 Video Codec Specification. Available online: https://www.ffmpeg.org/~michael/ffv1.html (accessed on 8 December 2020).
- FFmpeg Multimedia Framework. Available online: https://www.ffmpeg.org/ (accessed on 19 January 2021).
- Daede, T.; Norkin, A.; Brailovskiy, I. Video Codec Testing and Quality Measurement. Available online: https://tools.ietf.org/id/draft-ietf-netvc-testing-06.html (accessed on 8 December 2020).
Sequence Number | Number of Frames | Description |
---|---|---|
1 | 1855 | urban, sunny |
2 | 1875 | urban, sunny |
3 | 515 | test ride in laboratory |
4 | 2011 | urban, sunny |
5 | 2051 | road crossing, winter |
6 | 2114 | Suburban, cloudy |
7 | 1761 | road crossing, sunny |
8 | 1707 | suburban, evening |
9 | 1909 | suburban, sunny |
10 | 1897 | departure from the property |
11 | 2149 | road crossing, cloudy |
Total frames | 19,844 |
Compression Mode | GOP | Coding Procedure | |||||||
---|---|---|---|---|---|---|---|---|---|
standard | 1 | Direct—RCCC | 1.47 | - | - | - | - | - | - |
standard | 1 | R images and interpolated CCCC | 1.97 | 1.98 | 1.96 | - | - | - | - |
standard | 1 | All four components (R, C, C, C) separated | 2.06 | 1.98 | - | 2.08 | 2.08 | 2.08 | - |
standard | 1 | Separated R, and joined components | 2.12 | 1.98 | - | 2.08 | - | - | 2.21 |
extended | 100 | Direct—RCCC | 1.50 | - | - | - | - | - | - |
extended | 100 | R images and interpolated CCCC | 2.04 | 2.12 | 2.01 | - | - | - | - |
extended | 100 | All four components (R, C, C, C) separated | 2.20 | 2.12 | - | 2.21 | 2.21 | 2.23 | - |
extended | 100 | Separated R, and joined components | 2.23 | 2.12 | - | 2.21 | - | - | 2.30 |
Compression Mode | ||||||||
---|---|---|---|---|---|---|---|---|
Coding Procedure | Standard (GOP = 1) | Extended (GOP = 100) | ||||||
Direct—RCCC | 73.6 | 41 | 77.3 | 44 | 72.8 | 41 | 16.6 | 9 |
R images and interpolated CCCC | 103.1 | 58 | 124.8 | 70 | 102.7 | 58 | 30.2 | 17 |
Separated all components | 115.9 | 65 | 130.9 | 74 | 116.3 | 66 | 81.8 | 46 |
Separated R, and component images | 116.6 | 66 | 132.7 | 75 | 116.2 | 65 | 50.6 | 29 |
Compression Mode | ||||||||||
---|---|---|---|---|---|---|---|---|---|---|
Sequence | Standard (GOP = 1) | Extended (GOP = 100) | ||||||||
Compression CR and | ||||||||||
1 | 2.02 | 110.9 | 63 | 127.3 | 72 | 2.14 | 110.7 | 62 | 51.1 | 29 |
2 | 2.15 | 114.2 | 64 | 135.2 | 76 | 2.27 | 112.8 | 64 | 55.2 | 31 |
3 | 1.69 | 96.9 | 55 | 110.9 | 62 | 1.81 | 98.3 | 55 | 42.4 | 24 |
4 | 2.27 | 117.9 | 66 | 139.7 | 79 | 2.38 | 117.3 | 66 | 52.6 | 30 |
5 | 2.31 | 125.1 | 70 | 141.7 | 80 | 2.42 | 124.1 | 70 | 50.9 | 29 |
6 | 2.33 | 127.4 | 72 | 143.1 | 81 | 2.43 | 126.1 | 71 | 49.8 | 28 |
7 | 2.05 | 116.1 | 65 | 130.7 | 74 | 2.18 | 115.6 | 65 | 50.5 | 28 |
8 | 1.91 | 108.5 | 61 | 122.5 | 69 | 2.03 | 108.9 | 61 | 48.6 | 27 |
9 | 2.13 | 119.4 | 67 | 132.1 | 74 | 2.24 | 119.5 | 67 | 53.5 | 30 |
10 | 2.16 | 120.3 | 68 | 134.5 | 76 | 2.27 | 120.1 | 68 | 52.7 | 30 |
11 | 2.31 | 125.6 | 71 | 142.1 | 80 | 2.42 | 124.3 | 70 | 49.9 | 28 |
Average | 2.12 | 116.6 | 66 | 132.7 | 75 | 2.23 | 116.2 | 65 | 50.6 | 29 |
Publisher’s Note: MDPI stays neutral with regard to jurisdictional claims in published maps and institutional affiliations. |
© 2021 by the authors. Licensee MDPI, Basel, Switzerland. This article is an open access article distributed under the terms and conditions of the Creative Commons Attribution (CC BY) license (http://creativecommons.org/licenses/by/4.0/).
Share and Cite
Pawłowski, P.; Piniarski, K.; Dąbrowski, A. Highly Efficient Lossless Coding for High Dynamic Range Red, Clear, Clear, Clear Image Sensors. Sensors 2021, 21, 653. https://doi.org/10.3390/s21020653
Pawłowski P, Piniarski K, Dąbrowski A. Highly Efficient Lossless Coding for High Dynamic Range Red, Clear, Clear, Clear Image Sensors. Sensors. 2021; 21(2):653. https://doi.org/10.3390/s21020653
Chicago/Turabian StylePawłowski, Paweł, Karol Piniarski, and Adam Dąbrowski. 2021. "Highly Efficient Lossless Coding for High Dynamic Range Red, Clear, Clear, Clear Image Sensors" Sensors 21, no. 2: 653. https://doi.org/10.3390/s21020653
APA StylePawłowski, P., Piniarski, K., & Dąbrowski, A. (2021). Highly Efficient Lossless Coding for High Dynamic Range Red, Clear, Clear, Clear Image Sensors. Sensors, 21(2), 653. https://doi.org/10.3390/s21020653