[go: up one dir, main page]
More Web Proxy on the site http://driver.im/
Next Article in Journal
Learning Entropy as a Learning-Based Information Concept
Next Article in Special Issue
A New Phylogenetic Inference Based on Genetic Attribute Reduction for Morphological Data
Previous Article in Journal
Effect of Ti/Ni Coating of Diamond Particles on Microstructure and Properties of High-Entropy Alloy/Diamond Composites
You seem to have javascript disabled. Please note that many of the page functionalities won't work as expected without javascript enabled.
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Quality-Oriented Perceptual HEVC Based on the Spatiotemporal Saliency Detection Model

1
Department of Information Engineering, Shanghai Maritime University, NO.1550, Haigang Ave. Shanghai 201306, China
2
Department of Electrical and Electronics Engineering, Tokushima University, 2-24, Shinkura-cho, Tokushima 770-8501, Japan
*
Author to whom correspondence should be addressed.
Entropy 2019, 21(2), 165; https://doi.org/10.3390/e21020165
Submission received: 20 January 2019 / Revised: 4 February 2019 / Accepted: 10 February 2019 / Published: 11 February 2019
(This article belongs to the Special Issue Application of Information Theory in Biomedical Data Mining)

Abstract

:
Perceptual video coding (PVC) can provide a lower bitrate with the same visual quality compared with traditional H.265/high efficiency video coding (HEVC). In this work, a novel H.265/HEVC-compliant PVC framework is proposed based on the video saliency model. Firstly, both an effective and efficient spatiotemporal saliency model is used to generate a video saliency map. Secondly, a perceptual coding scheme is developed based on the saliency map. A saliency-based quantization control algorithm is proposed to reduce the bitrate. Finally, the simulation results demonstrate that the proposed perceptual coding scheme shows its superiority in objective and subjective tests, achieving up to a 9.46% bitrate reduction with negligible subjective and objective quality loss. The advantage of the proposed method is the high quality adapted for a high-definition video application.

1. Introduction

High efficiency video coding (HEVC), also known as H.265, is the latest video coding standard. It was released in 2013 and achieves an average bitrate reduction of 50% for fixed video quality compared with the H.264/AVC which is a block-oriented motion-compensation-based video compression standard. As regards the limited channel bandwidth and storage capacity, its coding efficiency needs further improvement. A perceptual model based on the human visual system (HVS) can be integrated into the video coding framework to achieve a low bitrate with a high perceptual quality.
Saliency refers to the indication of the probability of human attention according to the visual stimulus, and the term saliency often appears indistinguishably in two types of studies with different approaches; the “different outputs” approach and the “different evaluations” approach [1]. The first is a scientific study that aims to implement the psychophysical findings of human visual attention systems. This approach is referred to as a saliency map, and computes how much attention a pixel in a given image or video attracts, and then compares the output of the method with actual human gaze data. These outputs are often in the form of a two-dimensional map. The second type is a form of engineering studies, simply designed to extract meaningful objects. This approach is referred to as salient object detection, which aims to estimate regions in a given image that contain meaningful objects, and utilizes ground truth region labels for evaluation. Salient object detection is commonly interpreted in computer vision as a process that includes two stages: (a) detecting the most salient object and (b) segmenting the accurate regions of that object. These outputs are often in the form of a binary map.
Recently, perceptual video coding (PVC) has attracted lots of attention [2,3,4]. The perceptual models in previous works have been divided into four categories: the visual attention (VA) model, the region-of-interest (ROI) model, the visual sensitivity (VS) model, and the cross-modal attention (CMA) model. A detailed description of these methods is as follows.
The main idea, according to the visual attention model, is to encode the importance of different regions with bottom-up or top-down attention. Itti presents a biologically-motivated algorithm to select visually-salient regions for video compression [5]. Gou et al. propose a novel multiresolution spatiotemporal saliency detection model to calculate the spatiotemporal saliency map [6]. This method can improve the coding efficiency significantly and works in real time. Wei et al. present an H.265/HEVC-compliant perceptual video coding scheme based on visual saliency [7]. This method can reduce the bitrate significantly with negligible perceptual quality loss. However, in previous work, the video-based visual saliency model (VSM) has proved to be unsophisticated for use in video coding. The main idea of the ROI model is to encode an area around the predicted attention-grabbing regions with higher quality compared to other less visually important regions. Xu et al. present a region-of-interest based H.265/HEVC coding approach to improve the perceived visual quality [8]. Grois et al. define a dynamic transition region between the region-of-interest and backgrounds [9], then a complexity-aware adaptive spatial preprocessing scheme is presented for the efficient scalable video coding. However, the above methods do not make a distinction between the different parts of the non-moving region. The main idea, as regards the visual sensitivity model, is to improve the coding performance with kinds of visual signal distortion sensitivity differences in HVS [10]. However, the computation complexity is higher when using these models. The main idea of the cross-modal attention model is to encode using multimodal interaction mechanisms [11]. However, it is not applicable for no voice mixing of the video sequence. Recently, some innovative ideas have been developed to improve the performance of video coding. Ferroukhi et al. propose a method combining bandelets and the set partitioning in hierarchical trees (SPIHT) algorithm for medical video coding [12], and this method has shown good high performances in terms of visual quality and peak signal-to-noise rate (PSNR) at low bitrates. Rouis et al. present a novel approach for perceptually guiding the rate-distortion optimization (RDO) process in H.265/HEVC [13], and this proposed method demonstrates a superior rate-distortion (R-D) performance over a compared approach adopting a similar scheme. Moreover, the learning-based perceptual video coding methods are proposed to improve video quality [14,15]. However, the computation complexity of these methods is high.
In summary, the research related to PVC has made great progress in recent years, but there are deficiencies. For example, there is lack of effective saliency detection computing models for video coding applications. Image-based saliency detection has been extensively studied over the past decades. However, the research related to video-based saliency detection is much less explored. The existing saliency detection models only consider some of the characteristics of the image, and some of the temporal and context information is ignored. Moreover, most of the computation models are based on the pixel level. The computation complexity is too high, making it is unfavorable for real-time video applications.
In this paper, we proposed a new saliency detection model for video compression. Firstly, a novel saliency object detection model is proposed to generate a spatiotemporal saliency map. Secondly, the H.265/HEVC-compliant perceptual coding framework is developed based on the spatiotemporal saliency map. This paper has the following contributions: (1) a new superpixel-based salient object model is proposed which incorporates spatial and temporal features. (2) the proposed perceptual coding method achieves higher compression rates with a good balance between subjective and objective quality compared to the H.265/HEVC reference model.

2. Proposed Perceptual Video Coding Algorithm

2.1. System Overview

Figure 1 shows the framework of the saliency-based H.265/HEVC. This perceptual video coding framework includes two parts: the perceptual module and the encoder. The perceptual module is used to generate the saliency map using the spatiotemporal information. The saliency map is inputted so the encoder can determine the quantization parameters.

2.2. Video-Based Spatiotemporal Saliency Detection Model

For monitoring video, the moving object is the most attractive point of attention for a human. Meanwhile, there are the static features in the video. Thus, the spatiotemporal saliency map generated by the visual attention model and moving object detection model can be explored to guide the bit allocation for video coding.
For the input video signal, the spatial saliency and temporal saliency are computed using the spatiotemporal information, respectively. Therefore, according to a certain weight coefficient, the spatial and temporal saliency are merged into the final saliency map.
A good deal of research effort has already been devoted to saliency models for images. The image saliency models can be used for spatial saliency detection in each video frame. In this paper, the spatial saliency computation is based on the Markov chain (MC) saliency detection model developed by Lu et al. [16], which is both effective and efficient. The main details of the MC saliency algorithm are as follows:
  • A single layer graph G ( V , E ) with superpixels is constructed, where V and E represent the nodes and edges of G, respectively. On the G, each node, including transient nodes and absorbing nodes, is connected to the transient nodes which neighbor it or share common boundaries with its neighboring nodes. Thus, the weight w i j of the edge e i j between adjacent nodes i and j is defined as
    w i j = e | | x i x j | | σ 2
    where x i and x j represent the mean of the two nodes in the CIELAB color space, and σ is a constant.
  • The nodes are renumbered so that the first t nodes are transient nodes and the last r nodes are absorbing nodes, and the affinity matrix A and degree matrix D can be expressed with w i j . Then, the transition matrix P on the sparsely connected graph can be calculated from matrix A and D. After that, using the Markov chain theory, the absorbed time for each transient state can be calculated based on the transition matrix P. Finally, the saliency map S can be generated by normalizing the absorbed time.
    The basic idea of the MC method is to detect the saliency using the time property in an absorbing Markov chain. The virtual boundary nodes are identified as absorbing nodes based on the prior boundary. The saliency is computed as the absorbed time to the absorbing nodes. On the basis of the MC saliency model, the spatial saliency value is represented as S S ( i ) for each pixel i.
Particularly for videos, the temporal information is a significant hint, and more context exists in the field of video saliency detection. In video saliency detection, the optical flow technique can detect motion accurately. M V x ( x , y ) and M V y ( x , y ) are represented as the horizontal and vertical motion vector by using the Lucas–Kanade (LK) algorithm [17]. Thus, the motion vector amplitude (MVA) M V ( x , y ) is calculated as
M V ( x , y ) = M V x ( x , y ) 2 + M V y ( x , y ) 2 .
Furthermore, the MVA is enhanced as
S m v ( x , y ) = α × M V ( x , y ) α × β if M V ( x , y ) > β 0 otherwise
where α and β are parameters, in this work, α = 10 and β = 2 . Finally, the S m v ( x , y ) value is clipped to the [0, 255] range:
S ^ m v ( x , y ) = 255 if S m v ( x , y ) > 255 S m v ( x , y ) others
Then, the S ^ m v ( x , y ) can be assigned the temporal saliency value S T ( i ) for pixel i ( x , y ) :
S T ( i ) = S ^ m v ( x , y ) .
The temporal saliency reflects the dynamic characteristics of the video, while the spatial saliency reflects the static characteristics of the video. When the spatial and temporal saliency maps are constructed, we use them to get the final video saliency map. Figure 2 shows the block diagram of the spatiotemporal saliency fusion framework. For the input video signal, the spatial saliency and temporal saliency are computed by using the spatiotemporal information, respectively. Thus, according to a certain weight coefficient, the spatial and temporal saliency are merged into the final saliency map.
Different fusion methods can be utilized, and the linear fusion is used in this paper. The final video saliency map S S T ( i ) is generated by
S S T ( i ) = ( 1 ω t ) × S S ( i ) + ω t × S T ( i )
where ω t is the weight of the temporal saliency, and ( 1 ω t ) is the weight of the spatial saliency. The weight ω t is determined by the contribution of temporal characteristics in the spatiotemporal saliency. The best experimental combination that we obtained for Equation (6) was achieved for ω t = 3/7 in this paper [18].
In the video, the moving objects are regarded as the salient region. Finally, the superposition of spatiotemporal saliency is taken as the final saliency with linear fusion. Figure 3d shows the spatiotemporal saliency for the third frame of the BasketballDrill sequence.
It is noted that this work is different from the existing method [6]. Firstly, the Markov chain-based saliency detection model in this work is a spatial domain approach, while the phase spectrum of quaternion Fourier transform (PQFT) model in the existing method [6] is a frequency domain one. However, the PQFT-based saliency model only highlights the boundaries of the salient object while the object interior still has a low saliency value. Secondly, the Markov chain-based saliency detection method in the proposed work can enhance the object interior significantly. Moreover, the proposed method is more effective in saliency detection. Figure 4 shows an example of the saliency maps calculated by PQFT and the proposed methods, and it was discovered that the proposed method can predict eye fixations better than the PQFT method.
Moreover, for good saliency detection, a model should meet at least the following two criteria: (a) Good detection: the probability of missing real salient regions and falsely marking the background as a salient region should be low. (b) High saliency value: the boundaries of salient regions and region interiors have a high saliency value. The advantages of the proposed work are as follows: (1) In this work, the optical flow technique is used to detect temporal saliency, which can detect moving objects precisely. (2) The Markov chain-based saliency detection method in the proposed work can enhance the region’s interior significantly. Thus, the generated saliency map with the proposed method can increase the coding performance.

2.3. Saliency-Based Quantization Control Algorithm

In the video, the moving objects are regarded as the salient region. Finally, the superposition of spatiotemporal saliency is taken as the final saliency with linear fusion.
In H.265/HEVC standards, the coding unit (CU) is a square region and is a node of the quad-tree partitioning of the coding tree units (CTU) [19]. The quad-tree partitioning structure allows recursive splitting into four equally sized nodes, and the minimum CU size is configured to be 32 × 32 , 16 × 16 , or 8 × 8 that of the luma samples. Coding unit tree is an algorithm for adaptive quantization parameters (QP). In order to control QP in the encoder, the saliency map can be partitioned into the block with a size of 64 × 64 . Then, the block-based saliency average value S ¯ S T ( x , y ) is calculated as
S ¯ S T ( x , y ) = x = 0 N y = 0 N S S T ( x , y ) N 2
where N is CU size, and S S T ( x , y ) is the saliency value of the pixel i ( x , y ) within a CU.
In H.265/HEVC standards, a QP is used to determine the quantization step, the setting range of QP is from 0 to 51. In order to associate the block-based saliency value with QP, the saliency value can be normalized to the range [0, 3]. In this paper, the min–max normalization method is used to generate the saliency level for the block-based saliency average value S ¯ S T ( x , y ) :
Δ S = 3 × S ¯ S T ( x , y ) S m i n S m a x S m i n
where Δ S is the calculated saliency level, and S m a x and S m i n are the maximums and minimums of S ¯ S T ( x , y ) , respectively.
When the Δ S value is high, the corresponding image can be encoded as a foreground. In contrast, when the Δ S value is low, the corresponding image can be encoded as a background. In the perceptual video coding process, the foreground needs to increase the allocated bitrate resources, while the background needs to reduce the allocated bitrate resources. In video coding, QP is used to adjust distortion. When the QP is larger, the image distortion is higher. Therefore, a high QP is used for the foreground coding, and a low QP is used for the background coding. On the basis of the saliency level Δ S , the setting quantized parameters ( Q P s e t ) can be dynamically adjusted in the encoding process. When Q P d e f is the default quantized parameter, and the Q P d e f has four values (22, 27, 32, and 37) in H.265/HEVC reference software. Thus, the saliency-based quantization control algorithm is shown in Algorithm 1. By using the proposed method, more image distortion is introduced to CUs with a low saliency while less to CUs with a high saliency.
Algorithm 1: Saliency-based quantization control algorithm
Entropy 21 00165 i001

3. Experiment Results

The proposed algorithm is implemented and verified based on the H.265/HEVC reference model HM 16.0. The quantization parameters are set to 22, 27, 32, and 37, respectively. The configuration profile is low-delay (LD).
The performance of the proposed algorithm is evaluated by the average bitrate saving (BS) and Bjontegarrd Delta PSNR (BD-PSNR) according to Reference [20]. The BS is calculated as
B S ( % ) = 1 4 i = 1 4 B R H M ( Q P i ) B R p r o ( Q P i ) B R H M ( Q P i ) × 100 %
where B R H M ( Q P i ) and B R p r o ( Q P i ) are the bitrates using the H.265/HEVC reference software and the proposed method with different Q P i .
The double stimulus continuous quality scale (DSCQS) method is used for subjective evaluation [21]. The subjects are presented with pairs of video sequences, where the first sequence is the H.265/HEVC reference video and the second sequence is the video with the proposed method. Secondly, a total of 24 naive viewers take part in the test campaign, and all viewers are screened for the correct level of visual acuity and color vision. Thirdly, viewers are expected to mark their visual quality score on an answer sheet with a quality rating scale over a defined scale that is defined in Table 1.
The subjective evaluation of video quality is evaluated by the difference mean opinion score (MOS) average values Delta MOS (D-MOS), which is calculated as
D-MOS = 1 4 i = 1 4 ( M O S p r o ( Q P i ) M O S H M ( Q P i ) )
where M O S p r o ( Q P i ) and M O S H M ( Q P i ) are the measured MOS values from the sequence encoded by the proposed method and the H.265/HEVC reference software HM 16.0 with different Q P i , respectively.
In the ablation study, the results of spatial-only, temporal-only, and the proposed spatiotemporal saliency-based quantization control scenarios are shown in Table 2. From the experimental results, it can be seen that the average bitrate saving and Bjontegarrd Delta PSNR loss are 6.51%, 0.059 dB, and 6.09%, 0.053 dB for the spatial-only, temporal-only scenarios, respectively. For the proposed spatiotemporal method, the bitrate can be reduced by 5.22% while the Bjontegarrd Delta PSNR loss is only 0.046 dB on average. Thus, the objective distortion of the proposed method is smaller compared with the spatial-only and temporal-only scenarios.
For our experiments, we selected four test sequences of different resolutions. The four test sequences were also selected by considering the diversity of spatial characteristics and temporal characteristics. The four test sequences were the ‘ P a r k S c e n e ’ sequence with a resolution of 1920 × 1080 , the ‘ V i d y o 1 ’ sequence with a resolution of 1280 × 720 , the ‘ B a s k e t b a l l D r i l l ’ sequence with a resolution of 832 × 480 , and the ‘ B a s k e t b a l l P a s s ’ sequence with a resolution of 416 × 240 . The test results of the proposed method are shown in Table 3, compared to H.265/HEVC reference software with four QP values of 22, 27, 32, and 37, It shows that the proposed PVC method can achieve 1.22%–8.84% bitrate reduction for the test sequences, where B i t r a t e R e d u c t i o n ( % ) = H M 16.0 B i t r a t e s P r o p o s e d B i t r a t e s H M 16.0 B i t r a t e s × 100 % . It is also observed in Table 3 that the bitrate reduction values are usually decreased as the QP value increases for the proposed method. This is because the distortions introduced by quantization errors are high enough at high QP values. It is noted in Table 3 that the average time saving is 0.16% for the test sequences ( T i m e S a v i n g ( % ) = H M 16.0 E n c o d i n g T i m e P r o p o s e d E n c o d i n g T i m e H M 16.0 E n c o d i n g T i m e × 100 % ), compared to the H.265/HEVC reference software. Thus, the proposed method is effective, and the computational cost of the proposed method is almost the same as the H.265/HEVC reference software.
When the Delta MOS values are smaller, the subjective qualities are closer to the original H.265/HEVC reference software. The fifth, eighth, and eleventh columns in Table 2 show the visual quality for the spatial-only, temporal-only, and spatiotemporal scenarios. It can be seen from the table that the average absolute Delta MOS value of the proposed method is even smaller than 0.1. Figure 5 shows the typical example of MOS curve for Vidyo1 and BasketballDrill sequences among the spatial-only, temporal-only, spatio-temporal methods, and the H.265/HEVC reference model. From these figures, it can be seen that the proposed method shows visual quality improvements over the H.265/HEVC reference model, spatiotemporal methods, and temporal-only scenarios. Furthermore, when QP is set to 37, Figure 6 shows the difference between the image of pure-HEVC and the image of the proposed reconstructed video for RaceHorsesC sequence. As can be seen from the graphic, at the low bit-rate, the visual quality of the proposed method is better than the visual quality of pure-HEVC.
Furthermore, the comparative results with previous work are shown in Table 4, with QP set to 32. For Bae’s work [10], the average bitrate saving and Delta MOS are 1.4%, 0.3, respectively. In contrast, the average bitrate can be reduced by 3.9% and the Delta MOS is 0.2 for the proposed scheme. Thus, the performance of the proposed method is superior to Bae’s work.
In summary, from the objective and subjective test results, the proposed perceptual video coding scheme can achieve higher compression rates with the balance between subjective and objective quality compared to the H.265/HEVC reference model.

4. Conclusions

In this paper, we proposed a novel perceptual video coding scheme based on the spatiotemporal saliency model. Simulation results demonstrate that the proposed perceptual coding scheme can save of 5.22% bitrate, while the Bjontegarrd Delta PSNR is only 0.046 dB on average. Moreover, the average absolute Delta MOS value of the proposed method is even smaller than 0.1.

Author Contributions

X.J. designed the algorithm, conducted all experiments, analyzed the results, and wrote the manuscript. T.S. conceived the algorithm and wrote the manuscript. D.Z. conducted the literature review and wrote the manuscript. T.K. conducted the literature review and wrote the manuscript. L.W. wrote the manuscript.

Funding

This research was funded by National Natural Science Foundation of China grant number 61701297, China Postdoctoral Science Foundation grant number 2018M641982, and JSPS KAKENHI grant number 17K00157.

Conflicts of Interest

The authors declare no conflict of interest.

References

  1. Kimura, A.; Yonetani, R.; Hirayama, T. Computational models of human visual attention and their implementations: A survey. IEICE Trans. Inf. Syst. 2013, E96-D, 562–578. [Google Scholar] [CrossRef]
  2. Lee, J.S.; Ebrahimi, T. Perceptual video compression: A survey. IEEE J. Sel. Top. Signal Proc. 2012, 6, 684–697. [Google Scholar] [CrossRef]
  3. Hadizadeh, H.; Bajic, I.V. Saliency-Aware video compression. IEEE Trans. Image Proc. 2014, 23, 19–33. [Google Scholar] [CrossRef] [PubMed]
  4. Valizadeh, S.; Nasiopoulos, P.; Ward, R.K. Perceptual rate distortion optimization of 3D-HEVC using PSNR-HVS. Multimed. Tools Appl. 2018, 77, 22985–23008. [Google Scholar] [CrossRef]
  5. Itti, L. Automatic foveation for video compression using a neurobiological model of visual attention. IEEE Trans. Image Proc. 2004, 13, 1304–1318. [Google Scholar] [CrossRef]
  6. Guo, C.; Zhang, L. A novel multiresolution spatiotemporal saliency detection model and its applications in image and video compression. IEEE Trans. Image Proc. 2010, 19, 185–198. [Google Scholar]
  7. Wei, H.; Zhou, X.; Zhou, W.; Yan, C.; Duan, Z.; Shan, N. Visual saliency based perceptual video coding in HEVC. In Proceedings of the IEEE International Symposium on Circuits and Systems (ISCAS), Montreal, QC, Canada, 22–25 May 2016; pp. 2547–2550. [Google Scholar]
  8. Xu, M.; Deng, X.; Li, S.; Wang, Z. Region-of-Interest based conversational HEVC coding with hierarchical perception model of face. IEEE J. Sel. Top. Signal Proc. 2014, 8, 475–489. [Google Scholar] [CrossRef]
  9. Grois, D.; Hadar, O. Complexity-aware adaptive preprocessing scheme for Region-of-Interest spatial scalable video coding. IEEE Trans. Circuits Syst. Video Technol. 2014, 24, 1025–1039. [Google Scholar] [CrossRef]
  10. Bae, S.H.; Kim, J.; Kim, M. HEVC-based perceptually adaptive video coding using a DCT-based local distortion detection probability model. IEEE Trans. Image Proc. 2016, 25, 3343–3357. [Google Scholar] [CrossRef] [PubMed]
  11. Lee, J.S.; Simone, F.D.; Ebrahimi, T. Efficient video coding based on audio-visual focus of attention. J. Vis. Comm. Image Represent. 2011, 22, 704–711. [Google Scholar] [CrossRef] [Green Version]
  12. Ferroukhi, M.; Ouahabi, A.; Attari, M.; Yassine, H.; Abdelmalik, T.A. Medical Video Coding Based on 2nd-Generation Wavelets: Performance Evaluation. Electronics 2019, 8, 88. [Google Scholar] [CrossRef]
  13. Rouis, K.; Larabi, M.C.; Tahar, J.B. Perceptually Adaptive Lagrangian Multiplier for HEVC Guided Rate-Distortion Optimization. IEEE Access. 2018, 6, 33589–33603. [Google Scholar] [CrossRef]
  14. Zhu, S.; Xu, Z. Spatiotemporal visual saliency guided perceptual high efficiency video coding with neural network. Neurocomputing 2018, 275, 511–522. [Google Scholar] [CrossRef]
  15. Ki, S.; Bae, S.H.; Kim, M.; Ko, H. Learning-based just-noticeable-quantization-distortion modeling for perceptual video coding. IEEE Trans. Image Proc. 2018, 27, 3178–3193. [Google Scholar] [CrossRef] [PubMed]
  16. Jiang, B.; Zhang, L.; Lu, H.; Yang, C.; Yang, M.H. Saliency detection via absorbing markov chain. In Proceedings of the IEEE International Conference on Computer Vision (CVPR), Sydney, NSW, Australia, 1–8 December 2013; pp. 1665–1672. [Google Scholar]
  17. Baker, S.; Scharstein, D.; Lewis, J.P.; Roth, S.; Black, M.J.; Szeliski, R. A database and evaluation methodology for optical flow. Int. J. Comput. Vis. 2011, 92, 1–31. [Google Scholar] [CrossRef]
  18. Tong, Y.; Cheikh, F.A.; Guraya, F.F.E.; Konik, H.; Tremeay, A. A spatiotemporal saliency model for video surveillance. Cogn. Comput. 2011, 3, 241–263. [Google Scholar]
  19. Rosewarne, C. High Efficiency Video Coding (HEVC) Test Model 16 (HM 16); Document JCTVC-V1002, JCT-VC. October 2015. Available online: http://phenix.int-evry.fr/jct/ (accessed on 11 February 2019).
  20. Bjontegaard, G. Calculation of Average PSNR Differences between RD-Curves, ITU-T SG16 Q.6, VCEG-M33, JVT. April 2001. Available online: https://www.itu.int/wftp3/av-arch/video-site/0104_Aus/ (accessed on 11 February 2019).
  21. Recommendation ITU-R BT.500-13, Methodology for the Subjective Assessment of the Quality of Television Pictures, ITU-R, Telecommunication Standardization Bureau. 2012. Available online: https://www.itu.int/dms_pubrec/itu-r/rec/bt/r-rec-bt.500-13-201201-i!!pdf-e.pdf (accessed on 11 February 2019).
Figure 1. Proposed system framework of the perceptual video coding.
Figure 1. Proposed system framework of the perceptual video coding.
Entropy 21 00165 g001
Figure 2. Spatiotemporal saliency fusion framework.
Figure 2. Spatiotemporal saliency fusion framework.
Entropy 21 00165 g002
Figure 3. The saliency fusion of BasketballDrill sequence.
Figure 3. The saliency fusion of BasketballDrill sequence.
Entropy 21 00165 g003
Figure 4. Saliency maps calculated by the phase spectrum of quaternion Fourier transform (PQFT) and proposed methods.
Figure 4. Saliency maps calculated by the phase spectrum of quaternion Fourier transform (PQFT) and proposed methods.
Entropy 21 00165 g004
Figure 5. Rate mean opinion score (MOS) curves for quantization control algorithm.
Figure 5. Rate mean opinion score (MOS) curves for quantization control algorithm.
Entropy 21 00165 g005
Figure 6. Visual quality comparison between the pure high efficiency video coding (HEVC) and the proposed method.
Figure 6. Visual quality comparison between the pure high efficiency video coding (HEVC) and the proposed method.
Entropy 21 00165 g006
Table 1. Scale.
Table 1. Scale.
ScaleMOS
Very annoying1
Annoying2
Slightly annoying3
Perception but not annoying4
Imperception5
Table 2. The results of quantization control algorithm.
Table 2. The results of quantization control algorithm.
Spatial-OnlyTemporal-OnlySpatio-Temporal (Proposed)
ClassSequenceBS(%)BD-PSNR(dB)D-MOSBS(%)BD-PSNR(dB)D-MOSBS(%)BD-PSNR(dB)D-MOS
1920 × 1080 ParkScene4.66−0.146−0.0594.13−0.1730.0294.75−0.1390.206
Cactus6.35−0.142−0.3235.80−0.129−0.2355.45−0.1400.030
BQTerrace11.01−0.155−0.2949.56−0.133−0.2069.46−0.1480.073
1280 × 720 Vidyo17.01−0.060−0.1476.20−0.010−0.0594.06−0.0480.162
Vidyo39.64−0.049−0.1627.59−0.001−0.0736.21−0.0570.118
Vidyo47.66−0.031−0.0886.47−0.015−0.0155.14−0.0250.162
High Res.Average7.72−0.097−0.1796.63−0.077−0.0935.85−0.0930.125
832 × 480 BasketballDrill5.78−0.0170.0294.94−0.0470.1173.06−0.0030.206
BQMall7.39−0.005−0.1776.590.012−0.1775.810.0160.132
PartyScene6.13−0.052−0.2366.08−0.068−0.1474.80−0.0440.103
RaceHorsesC0.99−0.115−0.1922.05−0.152−0.1033.31−0.0940.044
416 × 240 BasketballPass6.270.027−0.1335.980.020−0.0455.170.0580.014
BQSquare7.080.042−0.2388.340.043−0.1775.750.060−0.059
BlowingBubbles7.40−0.057−0.3237.49−0.033−0.2356.08−0.031−0.191
RaceHorses3.79−0.070−0.1764.03−0.059−0.0883.98−0.044−0.015
Low Res.Average5.60−0.031−0.1815.69−0.036−0.1074.74−0.0100.029
Average 6.51−0.059−0.1806.09−0.053−0.1015.22−0.0460.070
Table 3. The performance comparison with the H.265/HEVC reference software.
Table 3. The performance comparison with the H.265/HEVC reference software.
PSNR (dB)Bitrates (kbps)
Sequence
(Size)
QPHM16.0ProposedHM16.0Proposed Bitrate Reduction
(%)
Time Saving
(%)
ParkScene
( 1920 × 1080 )
2240.0239.5810,462.139805.686.272.73
2737.3337.004492.624270.774.941.04
3234.6534.412034.981955.443.91−1.37
3732.1631.98924.67888.673.89−1.50
Vidyo1
( 1280 × 720 )
2243.3443.053405.153104.268.840.18
2741.2541.071483.551422.434.12−1.71
3238.9238.82791.84778.561.68−2.81
3736.2136.12450.40443.171.61−3.96
BasketballDrill
( 832 × 480 )
2240.8040.564025.283828.774.884.67
2737.6637.491953.631871.794.190.80
3234.7234.63954.91936.291.95−1.15
3732.2332.17501.79495.651.22−1.30
BasketballPass
( 416 × 240 )
2241.4441.091086.851003.607.662.02
2737.8237.58560.75527.076.012.44
3234.4734.33282.05272.083.541.23
3731.4831.36145.73140.693.461.21
Average 4.260.16
Table 4. Comparisons of rate and Delta MOS values (QP = 32).
Table 4. Comparisons of rate and Delta MOS values (QP = 32).
Bae [10]Proposed
SequenceBS(%)D-MOSBS(%)D-MOS
BQTerrace3.60.95.90.1
Catus0.80.02.20.0
ParkScene0.8−0.33.30.1
BQMall0.90.24.20.4
Average1.40.33.90.2

Share and Cite

MDPI and ACS Style

Jiang, X.; Song, T.; Zhu, D.; Katayama, T.; Wang, L. Quality-Oriented Perceptual HEVC Based on the Spatiotemporal Saliency Detection Model. Entropy 2019, 21, 165. https://doi.org/10.3390/e21020165

AMA Style

Jiang X, Song T, Zhu D, Katayama T, Wang L. Quality-Oriented Perceptual HEVC Based on the Spatiotemporal Saliency Detection Model. Entropy. 2019; 21(2):165. https://doi.org/10.3390/e21020165

Chicago/Turabian Style

Jiang, Xiantao, Tian Song, Daqi Zhu, Takafumi Katayama, and Lu Wang. 2019. "Quality-Oriented Perceptual HEVC Based on the Spatiotemporal Saliency Detection Model" Entropy 21, no. 2: 165. https://doi.org/10.3390/e21020165

APA Style

Jiang, X., Song, T., Zhu, D., Katayama, T., & Wang, L. (2019). Quality-Oriented Perceptual HEVC Based on the Spatiotemporal Saliency Detection Model. Entropy, 21(2), 165. https://doi.org/10.3390/e21020165

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop