CN102999911B - Three-dimensional image quality objective evaluation method based on energy diagrams - Google Patents
Three-dimensional image quality objective evaluation method based on energy diagrams Download PDFInfo
- Publication number
- CN102999911B CN102999911B CN201210493072.5A CN201210493072A CN102999911B CN 102999911 B CN102999911 B CN 102999911B CN 201210493072 A CN201210493072 A CN 201210493072A CN 102999911 B CN102999911 B CN 102999911B
- Authority
- CN
- China
- Prior art keywords
- mrow
- dis
- org
- msubsup
- directions
- Prior art date
- Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
- Expired - Fee Related
Links
- 238000011156 evaluation Methods 0.000 title claims abstract description 92
- 238000010586 diagram Methods 0.000 title claims abstract description 40
- 230000000007 visual effect Effects 0.000 claims abstract description 41
- 230000008447 perception Effects 0.000 claims abstract description 9
- 230000004927 fusion Effects 0.000 claims abstract description 7
- 230000004044 response Effects 0.000 claims description 41
- 230000000873 masking effect Effects 0.000 claims description 39
- 238000000034 method Methods 0.000 claims description 39
- 230000008859 change Effects 0.000 claims description 18
- 238000001914 filtration Methods 0.000 claims description 18
- 230000008569 process Effects 0.000 claims description 16
- 230000006870 function Effects 0.000 claims description 10
- 238000012800 visualization Methods 0.000 claims description 10
- 238000005286 illumination Methods 0.000 claims description 4
- 238000012545 processing Methods 0.000 claims description 3
- 230000016776 visual perception Effects 0.000 abstract 1
- 238000013441 quality evaluation Methods 0.000 description 5
- 238000005516 engineering process Methods 0.000 description 4
- 230000006835 compression Effects 0.000 description 3
- 238000007906 compression Methods 0.000 description 3
- 241001182632 Akko Species 0.000 description 2
- 230000000694 effects Effects 0.000 description 2
- 238000013210 evaluation model Methods 0.000 description 2
- YYJNOYZRYGDPNH-MFKUBSTISA-N fenpyroximate Chemical compound C=1C=C(C(=O)OC(C)(C)C)C=CC=1CO/N=C/C=1C(C)=NN(C)C=1OC1=CC=CC=C1 YYJNOYZRYGDPNH-MFKUBSTISA-N 0.000 description 2
- 230000005540 biological transmission Effects 0.000 description 1
- 210000004556 brain Anatomy 0.000 description 1
- 238000004364 calculation method Methods 0.000 description 1
- 238000011161 development Methods 0.000 description 1
- 239000000284 extract Substances 0.000 description 1
- 238000012067 mathematical method Methods 0.000 description 1
- 238000011160 research Methods 0.000 description 1
Landscapes
- Image Analysis (AREA)
- Image Processing (AREA)
Abstract
The invention discloses a three-dimensional image quality objective evaluation method based on energy diagrams. The three-dimensional image quality objective evaluation method includes: firstly respectively calculating energy diagrams on different scales and directions of original undistorted three-dimensional images and energy diagrams on different scales and directions of fuzzy three-dimensional images to be evaluated to obtain the objective evaluation metrics of each pixel point of the fuzzy three-dimensional images to be evaluated, and then performing fusion on the objective evaluation metrics of each pixel point of the fuzzy three-dimensional images to be evaluated according to the energy diagrams and binocular minimum perceptible changing images to obtain image quality objective evaluation predicted values of the fuzzy three-dimensional images to be evaluated. The three-dimensional image quality objective evaluation method has the advantages of being capable of well reflecting visual perception characteristics of human visual systems by obtaining the energy diagrams of different scales and directions, performing fusion on the energy diagrams and the minimum perceptible changing images, and being capable of effectively improving objective evaluation results and correlation of subjective perception.
Description
Technical Field
The invention relates to an image quality evaluation method, in particular to a three-dimensional image quality objective evaluation method based on an energy map.
Background
With the rapid development of image coding technology and stereoscopic display technology, the stereoscopic image technology has received more and more extensive attention and application, and has become a current research hotspot. The stereo image technology utilizes the binocular parallax principle of human eyes, the left and right viewpoint images from the same scene are respectively and independently received by binoculars, and binocular parallax is formed through brain fusion, so that the stereo image with depth feeling and reality feeling is appreciated. Because of the influence of the acquisition system and the storage compression and transmission equipment, a series of distortions are inevitably introduced into the stereo image, and compared with a single-channel image, the stereo image needs to ensure the image quality of two channels simultaneously, so that the quality evaluation of the stereo image has very important significance. However, currently, an effective objective evaluation method for evaluating the quality of a stereoscopic image is lacking. Therefore, establishing an effective objective evaluation model of the quality of the stereo image has very important significance.
At present, a plane image quality evaluation method is generally directly applied to evaluating stereoscopic image quality, however, a process of fusing left and right viewpoint images of a stereoscopic image to generate a stereoscopic effect is not a simple process of superimposing the left and right viewpoint images and is difficult to express by a simple mathematical method, so that how to extract effective characteristic information from the stereoscopic image to simulate binocular stereoscopic fusion, how to modulate an objective evaluation result according to a visual masking characteristic of human eyes and a response characteristic of binocular energy intensity of the human eyes, so that the objective evaluation result is more perceptually accordant with a human visual system, and the problem needs to be researched and solved in the process of objectively evaluating the stereoscopic image quality.
Disclosure of Invention
The invention aims to provide an objective evaluation method for the quality of a three-dimensional image based on an energy map, which can effectively improve the correlation between objective evaluation results and subjective perception.
The technical scheme adopted by the invention for solving the technical problems is as follows: a three-dimensional image quality objective evaluation method based on an energy map is characterized in that the processing process is as follows: firstly, acquiring an energy diagram of the original undistorted stereo image in different scales and directions according to even symmetric frequency response and odd symmetric frequency response of each pixel point in the left viewpoint image and the right viewpoint image of the original undistorted stereo image in different scales and directions and a pixel value of each pixel point in a parallax image between the left viewpoint image and the right viewpoint image of the original undistorted stereo image, acquiring energy graphs of the distorted stereo image to be evaluated in different scales and directions according to even symmetric frequency response and odd symmetric frequency response of each pixel point in the left viewpoint image and the right viewpoint image of the distorted stereo image to be evaluated in different scales and directions and pixel values of each pixel point in a parallax image between the left viewpoint image and the right viewpoint image of the original undistorted stereo image; then obtaining an objective evaluation metric value of each pixel point in the distorted stereo image to be evaluated according to the two energy maps; obtaining an objective evaluation metric value of the distorted three-dimensional image to be evaluated for reflecting a binocular visual masking effect according to the objective evaluation metric value of each pixel point in the distorted three-dimensional image to be evaluated and the binocular minimum perceptible change image of the left viewpoint image of the distorted three-dimensional image to be evaluated, and obtaining an objective evaluation metric value of the distorted three-dimensional image to be evaluated for reflecting binocular energy intensity according to the objective evaluation metric value of each pixel point in the distorted three-dimensional image to be evaluated and energy maps of the distorted three-dimensional image to be evaluated in different scales and directions; and finally, fusing an objective evaluation metric value for reflecting binocular visual masking effect and an objective evaluation metric value for reflecting binocular energy intensity of the distorted three-dimensional image to be evaluated to obtain an objective evaluation prediction value of the image quality of the distorted three-dimensional image to be evaluated.
The method specifically comprises the following steps:
making SorgFor original undistorted stereo image, let SdisFor the distorted stereo image to be evaluated, SorgIs noted as { Lorg(x, y) }, adding SorgIs noted as { Rorg(x, y) }, adding SdisIs noted as { Ldis(x, y) }, adding SdisIs noted as { Rdis(x, y) }, wherein (x, y) denotes a coordinate position of a pixel point in the left viewpoint image and the right viewpoint image, x is 1. ltoreq. x.ltoreq.W, y is 1. ltoreq. y.ltoreq.H, W denotes a width of the left viewpoint image and the right viewpoint image, H denotes a height of the left viewpoint image and the right viewpoint image, L is Lorg(x, y) represents { L }orgThe coordinate position in (x, y) } is the pixel value of the pixel point with (x, y), Rorg(x, y) represents { RorgThe pixel value L of the pixel point with the coordinate position (x, y) in (x, y) } isdis(x, y) represents { L }dis(x, y) } the pixel value of the pixel point with the coordinate position of (x, y), Rdis(x, y) represents { RdisThe coordinate position in (x, y) is the pixel value of the pixel point of (x, y);
secondly, extracting { L by using visual masking effect of human stereoscopic vision perception on background illumination and contrastdis(x, y) } binocular minimum perceivable change image, notedWherein,represents { Ldis(x, y) } binocular minimum perceivable change imageThe middle coordinate position is the pixel value of the pixel point of (x, y);
(iii) calculating { Lorg(x,y)}、{Rorg(x,y)}、{Ldis(x,y)}、{RdisEven symmetric frequency response and odd symmetric frequency response of each pixel point in (x, y) } in different scales and directions; then obtain { Lorg(x,y)}、{Rorg(x,y)}、{Ldis(x,y)}、{RdisThe amplitude and the phase of each pixel point in (x, y) in different scales and directions; then according to { Lorg(x, y) } and { R }orgAmplitude and phase of each pixel point in (x, y) in different scales and directions, and { L }org(x, y) } and { R }org(x, y) } calculating S from the pixel value of each pixel in the parallax imageorgEnergy diagram at different scales and directions, and is recorded asAnd according to { Ldis(x, y) } and { R }disAmplitude and phase of each pixel point in (x, y) in different scales and directions, and { L }org(x, y) } and { R }org(x, y) } calculating S from the pixel value of each pixel in the parallax imagedisEnergy diagram at different scales and directions, and is recorded asWherein,to representThe pixel value of the pixel point with the middle coordinate position (x, y) in different scales and directions,to representThe pixel values of the pixel points with the middle coordinate positions (x, y) in different scales and directions, alpha represents the scale factor of the filter adopted by filtering, alpha is more than or equal to 1 and less than or equal to 4, theta represents the direction factor of the filter adopted by filtering, and theta is more than or equal to 1 and less than or equal to 4;
fourthly, according to SorgEnergy diagram at different scales and directionsAnd SdisEnergy diagram at different scales and directionsCalculating SdisThe objective evaluation metric value of each pixel point in SdisThe objective evaluation metric value of the pixel point with the middle coordinate position (x, y) is marked as Q (x, y), <math>
<mrow>
<mi>Q</mi>
<mrow>
<mo>(</mo>
<mi>x</mi>
<mo>,</mo>
<mi>y</mi>
<mo>)</mo>
</mrow>
<mo>=</mo>
<munderover>
<mi>Σ</mi>
<mrow>
<mi>θ</mi>
<mo>=</mo>
<mn>1</mn>
</mrow>
<mn>4</mn>
</munderover>
<munderover>
<mi>Σ</mi>
<mrow>
<mi>α</mi>
<mo>=</mo>
<mn>1</mn>
</mrow>
<mn>4</mn>
</munderover>
<mi>arccos</mi>
<mrow>
<mo>(</mo>
<mfrac>
<mrow>
<mn>2</mn>
<mo>×</mo>
<msubsup>
<mi>BE</mi>
<mrow>
<mi>α</mi>
<mo>,</mo>
<mi>θ</mi>
</mrow>
<mi>org</mi>
</msubsup>
<mrow>
<mo>(</mo>
<mi>x</mi>
<mo>,</mo>
<mi>y</mi>
<mo>)</mo>
</mrow>
<mo>×</mo>
<msubsup>
<mi>BE</mi>
<mrow>
<mi>α</mi>
<mo>,</mo>
<mi>θ</mi>
</mrow>
<mi>dis</mi>
</msubsup>
<mrow>
<mo>(</mo>
<mi>x</mi>
<mo>,</mo>
<mi>y</mi>
<mo>)</mo>
</mrow>
<mo>+</mo>
<msub>
<mi>T</mi>
<mn>1</mn>
</msub>
</mrow>
<mrow>
<msubsup>
<mi>BE</mi>
<mrow>
<mi>α</mi>
<mo>,</mo>
<mi>θ</mi>
</mrow>
<mi>org</mi>
</msubsup>
<msup>
<mrow>
<mo>(</mo>
<mi>x</mi>
<mo>,</mo>
<mi>y</mi>
<mo>)</mo>
</mrow>
<mn>2</mn>
</msup>
<mo>+</mo>
<msubsup>
<mi>BE</mi>
<mrow>
<mi>α</mi>
<mo>,</mo>
<mi>θ</mi>
</mrow>
<mi>dis</mi>
</msubsup>
<msup>
<mrow>
<mo>(</mo>
<mi>x</mi>
<mo>,</mo>
<mi>y</mi>
<mo>)</mo>
</mrow>
<mn>2</mn>
</msup>
<mo>+</mo>
<msub>
<mi>T</mi>
<mn>1</mn>
</msub>
</mrow>
</mfrac>
<mo>)</mo>
</mrow>
<mo>/</mo>
<mn>16</mn>
<mo>,</mo>
</mrow>
</math> wherein arccos () is an inverted cosine function, T1Is a control parameter;
according to SdisAnd the sum of objective evaluation metrics of each pixel point { Ldis(x, y) } binocular minimum perceivable change imageCalculating S using binocular visual masking effect of human visual systemdisThe objective evaluation metric value for reflecting the binocular visual masking effect is marked as Qbm, <math>
<mrow>
<msub>
<mi>Q</mi>
<mi>bm</mi>
</msub>
<mo>=</mo>
<mfrac>
<mrow>
<munder>
<mi>Σ</mi>
<mrow>
<mrow>
<mo>(</mo>
<mi>x</mi>
<mo>,</mo>
<mi>y</mi>
<mo>)</mo>
</mrow>
<mo>∈</mo>
<mi>Ω</mi>
</mrow>
</munder>
<mi>Q</mi>
<mrow>
<mo>(</mo>
<mi>x</mi>
<mo>,</mo>
<mi>y</mi>
<mo>)</mo>
</mrow>
<mo>×</mo>
<mrow>
<mo>(</mo>
<mn>1</mn>
<mo>/</mo>
<msubsup>
<mi>J</mi>
<mi>L</mi>
<mi>dis</mi>
</msubsup>
<mrow>
<mo>(</mo>
<mi>x</mi>
<mo>,</mo>
<mi>y</mi>
<mo>)</mo>
</mrow>
<mo>)</mo>
</mrow>
</mrow>
<mrow>
<munder>
<mi>Σ</mi>
<mrow>
<mrow>
<mo>(</mo>
<mi>x</mi>
<mo>,</mo>
<mi>y</mi>
<mo>)</mo>
</mrow>
<mo>∈</mo>
<mi>Ω</mi>
</mrow>
</munder>
<mrow>
<mo>(</mo>
<mn>1</mn>
<mo>/</mo>
<msubsup>
<mi>J</mi>
<mi>L</mi>
<mi>dis</mi>
</msubsup>
<mrow>
<mo>(</mo>
<mi>x</mi>
<mo>,</mo>
<mi>y</mi>
<mo>)</mo>
</mrow>
<mo>)</mo>
</mrow>
</mrow>
</mfrac>
<mo>,</mo>
</mrow>
</math> Wherein Ω represents a pixel domain range;
according to SdisAnd the objective evaluation metric value of each pixel point in the image data and SdisEnergy diagram at different scales and directionsCalculating S by using the response characteristics of the human visual system to the binocular energy intensitydisThe objective evaluation metric value for reflecting the binocular energy intensity is marked as Qbe, <math>
<mrow>
<msub>
<mi>Q</mi>
<mi>be</mi>
</msub>
<mo>=</mo>
<mfrac>
<mrow>
<munder>
<mi>Σ</mi>
<mrow>
<mrow>
<mo>(</mo>
<mi>x</mi>
<mo>,</mo>
<mi>y</mi>
<mo>)</mo>
</mrow>
<mo>∈</mo>
<mi>Ω</mi>
</mrow>
</munder>
<mi>Q</mi>
<mrow>
<mo>(</mo>
<mi>x</mi>
<mo>,</mo>
<mi>y</mi>
<mo>)</mo>
</mrow>
<mo>×</mo>
<mrow>
<mo>(</mo>
<mn>1</mn>
<mo>/</mo>
<msup>
<mi>BE</mi>
<mi>dis</mi>
</msup>
<mrow>
<mo>(</mo>
<mi>x</mi>
<mo>,</mo>
<mi>y</mi>
<mo>)</mo>
</mrow>
<mo>)</mo>
</mrow>
</mrow>
<mrow>
<munder>
<mi>Σ</mi>
<mrow>
<mrow>
<mo>(</mo>
<mi>x</mi>
<mo>,</mo>
<mi>y</mi>
<mo>)</mo>
</mrow>
<mo>∈</mo>
<mi>Ω</mi>
</mrow>
</munder>
<mrow>
<mo>(</mo>
<mn>1</mn>
<mo>/</mo>
<msup>
<mi>BE</mi>
<mi>dis</mi>
</msup>
<mrow>
<mo>(</mo>
<mi>x</mi>
<mo>,</mo>
<mi>y</mi>
<mo>)</mo>
</mrow>
<mo>)</mo>
</mrow>
</mrow>
</mfrac>
<mo>,</mo>
</mrow>
</math> Wherein, <math>
<mrow>
<msup>
<mi>BE</mi>
<mi>dis</mi>
</msup>
<mrow>
<mo>(</mo>
<mi>x</mi>
<mo>,</mo>
<mi>y</mi>
<mo>)</mo>
</mrow>
<mo>=</mo>
<munderover>
<mi>Σ</mi>
<mrow>
<mi>θ</mi>
<mo>=</mo>
<mn>1</mn>
</mrow>
<mn>4</mn>
</munderover>
<munderover>
<mi>Σ</mi>
<mrow>
<mi>α</mi>
<mo>=</mo>
<mn>1</mn>
</mrow>
<mn>4</mn>
</munderover>
<msubsup>
<mi>BE</mi>
<mrow>
<mi>α</mi>
<mo>,</mo>
<mi>θ</mi>
</mrow>
<mi>dis</mi>
</msubsup>
<mrow>
<mo>(</mo>
<mi>x</mi>
<mo>,</mo>
<mi>y</mi>
<mo>)</mo>
</mrow>
<mo>;</mo>
</mrow>
</math>
to SdisObjective evaluation metric Q for reflecting binocular visual masking effectbmAnd SdisObjective evaluation metric Q for reflecting binocular energy intensitybeCarrying out fusion to obtain SdisThe predicted value of the objective evaluation of image quality is marked as Qfin,Qfin=(Qbm)γ(Qbe)βWherein γ and β are weight parameters.
The concrete process of the second step is as follows:
② 1, calculating { LdisVisualization threshold set of luminance masking effect of (x, y) }, denoted as { T }l(x,y)},Wherein, Tl(x, y) represents { L }disVisualization threshold of brightness masking effect of pixel point with coordinate position (x, y) in (x, y) }Value, bgl(x, y) represents { L }disThe average value of the brightness of all pixel points in an NxN neighborhood window taking a pixel point with a coordinate position (x, y) as the center in (x, y) }, wherein N is more than or equal to 1;
2, calculating { LdisVisualization threshold set of contrast masking effect of (x, y) }, denoted as { T }c(x,y)},Tc(x,y)=K(bgl(x,y))+ehl(x, y) wherein Tc(x, y) represents { L }disThe coordinate position in (x, y) is the visual threshold value of the contrast masking effect of the pixel point of (x, y), ehl(x, y) represents the pair { L }disThe pixel points with the coordinate positions (x, y) in (x, y) are respectively subjected to edge filtering in the horizontal direction and the vertical direction to obtain an average gradient value K (bg)l(x,y))=-10-6×(0.7×bgl(x,y)2+32×bgl(x,y))+0.07;
2- (3) pairs of { LdisVisualization threshold set of luminance masking effects of (x, y) } { Tl(x, y) } and a visual threshold set of contrast masking effects { T }c(x, y) } to obtain { Ldis(x, y) } binocular minimum perceivable change image, notedWill be provided withThe pixel value of the pixel point with the middle coordinate position (x, y) is recorded as
And N =5 is taken in the step II-1.
The concrete process of the step III is as follows:
③ 1, adopting log-Garbor filter pair { Lorg(x, y) is filtered to obtain { L }orgEven symmetric frequency response and odd symmetric frequency response of each pixel point in (x, y) } in different scales and directions are converted into { L }orgEven symmetric frequency responses of pixel points with coordinate positions (x, y) in different scales and directions in (x, y) are recorded asWill { LorgThe odd symmetric frequency response of the pixel point with the coordinate position (x, y) in different scales and directions is recorded asWherein alpha represents the scale factor of the filter used for filtering, alpha is more than or equal to 1 and less than or equal to 4, theta represents the direction factor of the filter used for filtering, and theta is more than or equal to 1 and less than or equal to 4;
③ 2, calculate { LorgThe amplitude and phase of each pixel point in (x, y) in different scales and directions will be { L }orgThe amplitudes of the pixel points with the coordinate positions (x, y) in the (x, y) at different scales and directions are recorded as Will { LorgThe phase positions of the pixel points with the coordinate positions (x, y) in (x, y) at different scales and directions are recorded as <math>
<mrow>
<msubsup>
<mi>LP</mi>
<mrow>
<mi>α</mi>
<mo>,</mo>
<mi>θ</mi>
</mrow>
<mrow>
<mi>L</mi>
<mo>,</mo>
<mi>org</mi>
</mrow>
</msubsup>
<mrow>
<mo>(</mo>
<mi>x</mi>
<mo>,</mo>
<mi>y</mi>
<mo>)</mo>
</mrow>
<mo>=</mo>
<mi>arctan</mi>
<mrow>
<mo>(</mo>
<mfrac>
<mrow>
<msubsup>
<mi>e</mi>
<mrow>
<mi>α</mi>
<mo>,</mo>
<mi>θ</mi>
</mrow>
<mi>L</mi>
</msubsup>
<mrow>
<mo>(</mo>
<mi>x</mi>
<mo>,</mo>
<mi>y</mi>
<mo>)</mo>
</mrow>
</mrow>
<mrow>
<msubsup>
<mi>o</mi>
<mrow>
<mi>α</mi>
<mo>,</mo>
<mi>θ</mi>
</mrow>
<mi>L</mi>
</msubsup>
<mrow>
<mo>(</mo>
<mi>x</mi>
<mo>,</mo>
<mi>y</mi>
<mo>)</mo>
</mrow>
</mrow>
</mfrac>
<mo>)</mo>
</mrow>
<mo>,</mo>
</mrow>
</math> Wherein arctan () is an inverse tangent function;
③ 3, obtaining { L ] according to the steps from the third step to the third steporgThe operation process of the amplitude and the phase of each pixel point in (x, y) } in different scales and directions is carried out to obtain { R } in the same wayorg(x,y)}、{Ldis(x,y)}、{RdisThe amplitude and phase of each pixel point in (x, y) in different scales and directions will be { R }orgThe amplitudes of the pixel points with the coordinate positions (x, y) in the (x, y) at different scales and directions are recorded asWill { RorgThe phase positions of the pixel points with the coordinate positions (x, y) in (x, y) at different scales and directions are recorded asWill { LdisThe amplitudes of the pixel points with the coordinate positions (x, y) in the (x, y) at different scales and directions are recorded asWill { LdisThe phase positions of the pixel points with the coordinate positions (x, y) in (x, y) at different scales and directions are recorded asWill { RdisThe amplitudes of the pixel points with the coordinate positions (x, y) in the (x, y) at different scales and directions are recorded asWill { RdisThe phase positions of the pixel points with the coordinate positions (x, y) in (x, y) at different scales and directions are recorded as
Thirdly-4, calculating L by adopting a block matching methodorg(x, y) } and { R }org(x, y) } parallax images, noted asWherein,to representThe middle coordinate position is the pixel value of the pixel point of (x, y);
③ 5 according to { Lorg(x, y) } and { R }orgAmplitude and phase of each pixel point in (x, y) in different scales and directions, and { L }org(x, y) } and { R }orgParallax image between (x, y) } sCalculating the pixel value of each pixel point in SorgEnergy diagram at different scales and directions, and is recorded asWill be provided withThe pixel values of the pixel points with the middle coordinate position (x, y) in different scales and directions are recorded as <math>
<mrow>
<msubsup>
<mi>BE</mi>
<mrow>
<mi>α</mi>
<mo>,</mo>
<mi>θ</mi>
</mrow>
<mi>org</mi>
</msubsup>
<mrow>
<mo>(</mo>
<mi>x</mi>
<mo>,</mo>
<mi>y</mi>
<mo>)</mo>
</mrow>
<mo>=</mo>
<msup>
<mrow>
<mo>(</mo>
<msubsup>
<mi>LE</mi>
<mrow>
<mi>α</mi>
<mo>,</mo>
<mi>θ</mi>
</mrow>
<mrow>
<mi>L</mi>
<mo>,</mo>
<mi>org</mi>
</mrow>
</msubsup>
<mrow>
<mo>(</mo>
<mi>x</mi>
<mo>,</mo>
<mi>y</mi>
<mo>)</mo>
</mrow>
<mo>)</mo>
</mrow>
<mn>2</mn>
</msup>
<mo>+</mo>
<msup>
<mrow>
<mo>(</mo>
<msubsup>
<mi>LE</mi>
<mrow>
<mi>α</mi>
<mo>,</mo>
<mi>θ</mi>
</mrow>
<mrow>
<mi>R</mi>
<mo>,</mo>
<mi>org</mi>
</mrow>
</msubsup>
<mrow>
<mo>(</mo>
<mi>x</mi>
<mo>-</mo>
<msubsup>
<mi>d</mi>
<mi>org</mi>
<mi>L</mi>
</msubsup>
<mrow>
<mo>(</mo>
<mi>x</mi>
<mo>,</mo>
<mi>y</mi>
<mo>)</mo>
</mrow>
<mo>,</mo>
<mi>y</mi>
<mo>)</mo>
</mrow>
<mo>)</mo>
</mrow>
<mn>2</mn>
</msup>
</mrow>
</math> <math>
<mrow>
<mo>+</mo>
<msubsup>
<mi>LE</mi>
<mrow>
<mi>α</mi>
<mo>,</mo>
<mi>θ</mi>
</mrow>
<mrow>
<mi>L</mi>
<mo>,</mo>
<mi>org</mi>
</mrow>
</msubsup>
<mrow>
<mo>(</mo>
<mi>x</mi>
<mo>,</mo>
<mi>y</mi>
<mo>)</mo>
</mrow>
<mo>×</mo>
<msubsup>
<mi>LE</mi>
<mrow>
<mi>α</mi>
<mo>,</mo>
<mi>θ</mi>
</mrow>
<mrow>
<mi>R</mi>
<mo>,</mo>
<mi>org</mi>
</mrow>
</msubsup>
<mrow>
<mo>(</mo>
<mi>x</mi>
<mo>-</mo>
<msubsup>
<mi>d</mi>
<mi>org</mi>
<mi>L</mi>
</msubsup>
<mrow>
<mo>(</mo>
<mi>x</mi>
<mo>,</mo>
<mi>y</mi>
<mo>)</mo>
</mrow>
<mo>,</mo>
<mi>y</mi>
<mo>)</mo>
</mrow>
<mo>×</mo>
<mi>cos</mi>
<mrow>
<mo>(</mo>
<msubsup>
<mi>LP</mi>
<mrow>
<mi>α</mi>
<mo>,</mo>
<mi>θ</mi>
</mrow>
<mrow>
<mi>L</mi>
<mo>,</mo>
<mi>org</mi>
</mrow>
</msubsup>
<mrow>
<mo>(</mo>
<mi>x</mi>
<mo>,</mo>
<mi>y</mi>
<mo>)</mo>
</mrow>
<mo>-</mo>
<msubsup>
<mi>LP</mi>
<mrow>
<mi>α</mi>
<mo>,</mo>
<mi>θ</mi>
</mrow>
<mrow>
<mi>R</mi>
<mo>,</mo>
<mi>org</mi>
</mrow>
</msubsup>
<mrow>
<mo>(</mo>
<mi>x</mi>
<mo>-</mo>
<msubsup>
<mi>d</mi>
<mi>org</mi>
<mi>L</mi>
</msubsup>
<mrow>
<mo>(</mo>
<mi>x</mi>
<mo>,</mo>
<mi>y</mi>
<mo>)</mo>
</mrow>
<mo>,</mo>
<mi>y</mi>
<mo>)</mo>
</mrow>
</mrow>
<mo>)</mo>
</mrow>
</math> Wherein cos () is a cosine-taking function,represents { Rorg(x, y) } coordinate position ofThe amplitudes of the pixel points of (a) at different scales and directions,represents { Rorg(x, y) } coordinate position ofThe phases of the pixel points in different scales and directions;
③ 6 according to { Ldis(x, y) } and { R }disAmplitude and phase of each pixel point in (x, y) in different scales and directions, and { L }org(x, y) } and { R }orgParallax image between (x, y) } sCalculating the pixel value of each pixel point in SdisEnergy diagram at different scales and directions, and is recorded asWill be provided withThe pixel values of the pixel points with the middle coordinate position (x, y) in different scales and directions are recorded as <math>
<mrow>
<msubsup>
<mi>BE</mi>
<mrow>
<mi>α</mi>
<mo>,</mo>
<mi>θ</mi>
</mrow>
<mi>dis</mi>
</msubsup>
<mrow>
<mo>(</mo>
<mi>x</mi>
<mo>,</mo>
<mi>y</mi>
<mo>)</mo>
</mrow>
<mo>=</mo>
<msup>
<mrow>
<mo>(</mo>
<msubsup>
<mi>LE</mi>
<mrow>
<mi>α</mi>
<mo>,</mo>
<mi>θ</mi>
</mrow>
<mrow>
<mi>L</mi>
<mo>,</mo>
<mi>dis</mi>
</mrow>
</msubsup>
<mrow>
<mo>(</mo>
<mi>x</mi>
<mo>,</mo>
<mi>y</mi>
<mo>)</mo>
</mrow>
<mo>)</mo>
</mrow>
<mn>2</mn>
</msup>
<mo>+</mo>
<msup>
<mrow>
<mo>(</mo>
<msubsup>
<mi>LE</mi>
<mrow>
<mi>α</mi>
<mo>,</mo>
<mi>θ</mi>
</mrow>
<mrow>
<mi>R</mi>
<mo>,</mo>
<mi>dis</mi>
</mrow>
</msubsup>
<mrow>
<mo>(</mo>
<mi>x</mi>
<mo>-</mo>
<msubsup>
<mi>d</mi>
<mi>org</mi>
<mi>L</mi>
</msubsup>
<mrow>
<mo>(</mo>
<mi>x</mi>
<mo>,</mo>
<mi>y</mi>
<mo>)</mo>
</mrow>
<mo>,</mo>
<mi>y</mi>
<mo>)</mo>
</mrow>
<mo>)</mo>
</mrow>
<mn>2</mn>
</msup>
</mrow>
</math> <math>
<mrow>
<mo>+</mo>
<msubsup>
<mi>LE</mi>
<mrow>
<mi>α</mi>
<mo>,</mo>
<mi>θ</mi>
</mrow>
<mrow>
<mi>L</mi>
<mo>,</mo>
<mi>dis</mi>
</mrow>
</msubsup>
<mrow>
<mo>(</mo>
<mi>x</mi>
<mo>,</mo>
<mi>y</mi>
<mo>)</mo>
</mrow>
<mo>×</mo>
<msubsup>
<mi>LE</mi>
<mrow>
<mi>α</mi>
<mo>,</mo>
<mi>θ</mi>
</mrow>
<mrow>
<mi>R</mi>
<mo>,</mo>
<mi>dis</mi>
</mrow>
</msubsup>
<mrow>
<mo>(</mo>
<mi>x</mi>
<mo>-</mo>
<msubsup>
<mi>d</mi>
<mi>org</mi>
<mi>L</mi>
</msubsup>
<mrow>
<mo>(</mo>
<mi>x</mi>
<mo>,</mo>
<mi>y</mi>
<mo>)</mo>
</mrow>
<mo>,</mo>
<mi>y</mi>
<mo>)</mo>
</mrow>
<mo>×</mo>
<mi>cos</mi>
<msup>
<mrow>
<mo>(</mo>
<mi>L</mi>
<msubsup>
<mi>P</mi>
<mrow>
<mi>α</mi>
<mo>,</mo>
<mi>θ</mi>
</mrow>
<mrow>
<mi>L</mi>
<mo>,</mo>
<mi>dis</mi>
</mrow>
</msubsup>
<mrow>
<mo>(</mo>
<mi>x</mi>
<mo>,</mo>
<mi>y</mi>
<mo>)</mo>
</mrow>
<mo>-</mo>
<msubsup>
<mi>LP</mi>
<mrow>
<mi>α</mi>
<mo>,</mo>
<mi>θ</mi>
</mrow>
<mrow>
<mi>R</mi>
<mo>,</mo>
<mi>dis</mi>
</mrow>
</msubsup>
<mrow>
<mo>(</mo>
<mi>x</mi>
<mo>-</mo>
<msubsup>
<mi>d</mi>
<mi>org</mi>
<mi>L</mi>
</msubsup>
<mrow>
<mo>(</mo>
<mi>x</mi>
<mo>,</mo>
<mi>y</mi>
<mo>)</mo>
</mrow>
<mo>,</mo>
<mi>y</mi>
<mo>)</mo>
</mrow>
<mo>)</mo>
</mrow>
<mo>,</mo>
</msup>
</mrow>
</math> Wherein,represents { Rdis(x, y) } coordinate position ofThe amplitudes of the pixel points of (a) at different scales and directions,represents { Rdis(x, y) } coordinate position ofThe phase of the pixel point in different scales and directions.
Taking T1Is 16.
In the step (c), gamma =0.7755 and beta =0.0505 are taken.
Compared with the prior art, the invention has the advantages that:
1) according to the method, the energy graphs of the original undistorted stereo image in different scales and directions and the energy graphs of the distorted stereo image to be evaluated in different scales and directions are respectively calculated, so that the objective evaluation metric value of each pixel point in the distorted stereo image to be evaluated is obtained, and the evaluation result is more perceptually accordant with a human visual system.
2) The method obtains the binocular minimum perceptible change image according to the stereoscopic vision characteristics of human eyes, and fuses the objective evaluation metric values of each pixel point in the distorted stereoscopic image to be evaluated, so that the evaluation result can reflect the binocular vision masking effect, and the correlation between the objective evaluation result and subjective perception is effectively improved.
3) According to the method, objective evaluation metric values of all pixel points in the distorted three-dimensional image to be evaluated are fused according to the energy graphs in different scales and directions, so that the evaluation result can reflect the response characteristic of a human visual system to binocular energy intensity, and the correlation between the objective evaluation result and subjective perception is effectively improved.
Drawings
FIG. 1 is a block diagram of an overall implementation of the method of the present invention;
fig. 2a is a left viewpoint image of Akko (640 × 480 size) stereo image;
fig. 2b is a right viewpoint image of an Akko (size 640 × 480) stereoscopic image;
fig. 3a is a left viewpoint image of an altmobit (size 1024 × 768) stereoscopic image;
fig. 3b is a right view image of an altmobit (size 1024 × 768) stereoscopic image;
fig. 4a is a left viewpoint image of a balloon (size 1024 × 768) stereoscopic image;
fig. 4b is a right viewpoint image of a balloon (size 1024 × 768) stereoscopic image;
fig. 5a is a left viewpoint image of a Doorflower (size 1024 × 768) stereoscopic image;
fig. 5b is a right viewpoint image of a Doorflower (size 1024 × 768) stereoscopic image;
fig. 6a is a left view image of a Kendo (size 1024 × 768) stereoscopic image;
fig. 6b is a right view image of a Kendo (size 1024 × 768) stereoscopic image;
fig. 7a is a left view image of a LeaveLaptop (size 1024 × 768) stereoscopic image;
fig. 7b is a right view image of a LeaveLaptop (size 1024 × 768) stereoscopic image;
fig. 8a is a left viewpoint image of a lovedual 1 (size 1024 × 768) stereoscopic image;
fig. 8b is a right viewpoint image of a lovedual 1 (size 1024 × 768) stereoscopic image;
fig. 9a is a left view image of a newsapper (size 1024 × 768) stereoscopic image;
fig. 9b is a right view image of a newsapper (size 1024 × 768) stereoscopic image;
FIG. 10a is a left viewpoint image of Puppy (size 720 × 480) stereo image;
FIG. 10b is a right viewpoint image of Puppy (size 720 × 480) stereoscopic image;
fig. 11a is a left viewpoint image of a Soccer2 (size 720 × 480) stereoscopic image;
fig. 11b is a right viewpoint image of a Soccer2 (size 720 × 480) stereoscopic image;
fig. 12a is a left viewpoint image of a Horse (size 720 × 480) stereoscopic image;
fig. 12b is a right view image of a Horse (size 720 × 480) stereoscopic image;
fig. 13a is a left viewpoint image of an Xmas (size 640 × 480) stereoscopic image;
fig. 13b is a right view image of an Xmas (size 640 × 480) stereoscopic image;
fig. 14 is a scatter diagram of the difference between the objective evaluation prediction value of image quality and the average subjective score for each distorted stereoscopic image.
Detailed Description
The invention is described in further detail below with reference to the accompanying examples.
The invention provides a stereo image quality objective evaluation method based on an energy diagram, the overall implementation block diagram of which is shown in figure 1, and the processing process of the method is as follows: firstly, acquiring an energy diagram of the original undistorted stereo image in different scales and directions according to even symmetric frequency response and odd symmetric frequency response of each pixel point in the left viewpoint image and the right viewpoint image of the original undistorted stereo image in different scales and directions and a pixel value of each pixel point in a parallax image between the left viewpoint image and the right viewpoint image of the original undistorted stereo image, acquiring energy graphs of the distorted stereo image to be evaluated in different scales and directions according to even symmetric frequency response and odd symmetric frequency response of each pixel point in the left viewpoint image and the right viewpoint image of the distorted stereo image to be evaluated in different scales and directions and pixel values of each pixel point in a parallax image between the left viewpoint image and the right viewpoint image of the original undistorted stereo image; then obtaining an objective evaluation metric value of each pixel point in the distorted stereo image to be evaluated according to the two energy maps; obtaining an objective evaluation metric value of the distorted three-dimensional image to be evaluated for reflecting a binocular visual masking effect according to the objective evaluation metric value of each pixel point in the distorted three-dimensional image to be evaluated and the binocular minimum perceptible change image of the left viewpoint image of the distorted three-dimensional image to be evaluated, and obtaining an objective evaluation metric value of the distorted three-dimensional image to be evaluated for reflecting binocular energy intensity according to the objective evaluation metric value of each pixel point in the distorted three-dimensional image to be evaluated and energy maps of the distorted three-dimensional image to be evaluated in different scales and directions; and finally, fusing an objective evaluation metric value for reflecting binocular visual masking effect and an objective evaluation metric value for reflecting binocular energy intensity of the distorted three-dimensional image to be evaluated to obtain an objective evaluation prediction value of the image quality of the distorted three-dimensional image to be evaluated. The method specifically comprises the following steps:
making SorgFor original undistorted stereo image, let SdisFor the distorted stereo image to be evaluated, SorgIs noted as { Lorg(x, y) }, adding SorgRight viewpoint ofThe image is noted as { Rorg(x, y) }, adding SdisIs noted as { Ldis(x, y) }, adding SdisIs noted as { Rdis(x, y) }, wherein (x, y) denotes a coordinate position of a pixel point in the left viewpoint image and the right viewpoint image, x is 1. ltoreq. x.ltoreq.W, y is 1. ltoreq. y.ltoreq.H, W denotes a width of the left viewpoint image and the right viewpoint image, H denotes a height of the left viewpoint image and the right viewpoint image, L is Lorg(x, y) represents { L }orgThe coordinate position in (x, y) } is the pixel value of the pixel point with (x, y), Rorg(x, y) represents { RorgThe pixel value L of the pixel point with the coordinate position (x, y) in (x, y) } isdis(x, y) represents { L }disThe coordinate position in (x, y) } is the pixel value of the pixel point with (x, y), Rdis(x, y) represents { RdisAnd the coordinate position in the (x, y) is the pixel value of the pixel point of (x, y).
The human visual characteristics indicate that the human eye is imperceptible to a property or noise that changes less in the image unless the intensity of the change of the property or noise exceeds a threshold, which is the minimum perceptible distortion (JND). However, the visual masking effect of human eyes is a local effect, which is influenced by background illumination, texture complexity and other factors, and the brighter the background is, the more complex the texture is, and the higher the threshold value is. Therefore, the invention extracts L by using the visual masking effect of human stereoscopic vision perception on background illumination and contrastdis(x, y) } binocular minimum perceivable change image, notedWherein,represents { Ldis(x, y) } binocular minimum perceivable change imageThe middle coordinate position is the pixel value of the pixel point of (x, y).
In this embodiment, the specific process of step two is:
② 1, calculating { LdisVisualization threshold set of luminance masking effect of (x, y) }, denoted as { T }l(x,y)},Wherein, Tl(x, y) represents { L }disThe coordinate position in (x, y) is the visual threshold value of the brightness masking effect of the pixel point of (x, y), bgl(x, y) represents { L }disIn (x, y) }, the pixel with the coordinate position (x, y) as the center is the average value of the brightness of all the pixels in the N × N neighborhood window, where N is greater than or equal to 1, and in this embodiment, N =5 is taken.
2, calculating { LdisVisualization threshold set of contrast masking effect of (x, y) }, denoted as { T }c(x,y)},Tc(x,y)=K(bgl(x,y))+ehl(x, y) wherein Tc(x, y) represents { L }disThe coordinate position in (x, y) is the visual threshold value of the contrast masking effect of the pixel point of (x, y), ehl(x, y) represents the pair { L }disThe pixel points with the coordinate positions (x, y) in (x, y) are respectively subjected to edge filtering in the horizontal direction and the vertical direction to obtain an average gradient value K (bg)l(x,y))=-10-6×(0.7×bgl(x,y)2+32×bgl(x,y))+0.07。
2- (3) pairs of { LdisVisualization threshold set of luminance masking effects of (x, y) } { Tl(x, y) } and a visual threshold set of contrast masking effects { T }c(x, y) } to obtain { Ldis(x, y) } binocular minimum perceivable change image, notedWill be provided withThe pixel value of the pixel point with the middle coordinate position (x, y) is recorded as
(iii) calculating { Lorg(x,y)}、{Rorg(x,y)}、{Ldis(x,y)}、{RdisEven symmetric frequency response and odd symmetric frequency response of each pixel point in (x, y) } in different scales and directions; then obtain { Lorg(x,y)}、{Rorg(x,y)}、{Ldis(x,y)}、{RdisThe amplitude and the phase of each pixel point in (x, y) in different scales and directions; then according to { Lorg(x, y) } and { R }orgAmplitude and phase of each pixel point in (x, y) in different scales and directions, and { L }org(x, y) } and { R }org(x, y) } calculating S from the pixel value of each pixel in the parallax imageorgEnergy diagram at different scales and directions, and is recorded asAnd according to { Ldis(x, y) } and { R }disAmplitude and phase of each pixel point in (x, y) in different scales and directions, and { L }org(x, y) } and { R }org(x, y) } calculating S from the pixel value of each pixel in the parallax imagedisEnergy diagram at different scales and directions, and is recorded asWherein,to representThe pixel value of the pixel point with the middle coordinate position (x, y) in different scales and directions,to representAnd the pixel values of the pixel points with the middle coordinate positions (x, y) in different scales and directions, wherein alpha represents the scale factor of the filter adopted by filtering, alpha is more than or equal to 1 and less than or equal to 4, theta represents the direction factor of the filter adopted by filtering, and theta is more than or equal to 1 and less than or equal to 4.
In this embodiment, the specific process of step (c) is:
③ 1, adopting log-Garbor filter pair { Lorg(x, y) is filtered to obtain { L }orgEven symmetric frequency response and odd symmetric frequency response of each pixel point in (x, y) } in different scales and directions are converted into { L }orgEven symmetric frequency responses of pixel points with coordinate positions (x, y) in different scales and directions in (x, y) are recorded asWill { LorgThe odd symmetric frequency response of the pixel point with the coordinate position (x, y) in different scales and directions is recorded asWherein alpha represents the scale factor of the filter used for filtering, alpha is more than or equal to 1 and less than or equal to 4, theta represents the direction factor of the filter used for filtering, and theta is more than or equal to 1 and less than or equal to 4.
③ 2, calculate { LorgThe amplitude and phase of each pixel point in (x, y) in different scales and directions will be { L }orgThe amplitudes of the pixel points with the coordinate positions (x, y) in the (x, y) at different scales and directions are recorded as Will { LorgThe phase positions of the pixel points with the coordinate positions (x, y) in (x, y) at different scales and directions are recorded as <math>
<mrow>
<msubsup>
<mi>LP</mi>
<mrow>
<mi>α</mi>
<mo>,</mo>
<mi>θ</mi>
</mrow>
<mrow>
<mi>L</mi>
<mo>,</mo>
<mi>org</mi>
</mrow>
</msubsup>
<mrow>
<mo>(</mo>
<mi>x</mi>
<mo>,</mo>
<mi>y</mi>
<mo>)</mo>
</mrow>
<mo>=</mo>
<mi>arctan</mi>
<mrow>
<mo>(</mo>
<mfrac>
<mrow>
<msubsup>
<mi>e</mi>
<mrow>
<mi>α</mi>
<mo>,</mo>
<mi>θ</mi>
</mrow>
<mi>L</mi>
</msubsup>
<mrow>
<mo>(</mo>
<mi>x</mi>
<mo>,</mo>
<mi>y</mi>
<mo>)</mo>
</mrow>
</mrow>
<mrow>
<msubsup>
<mi>o</mi>
<mrow>
<mi>α</mi>
<mo>,</mo>
<mi>θ</mi>
</mrow>
<mi>L</mi>
</msubsup>
<mrow>
<mo>(</mo>
<mi>x</mi>
<mo>,</mo>
<mi>y</mi>
<mo>)</mo>
</mrow>
</mrow>
</mfrac>
<mo>)</mo>
</mrow>
<mo>,</mo>
</mrow>
</math> Wherein arctan () is an arctangent function.
③ 3, obtaining { L ] according to the steps from the third step to the third steporgThe operation process of the amplitude and the phase of each pixel point in (x, y) } in different scales and directions is carried out to obtain { R } in the same wayorg(x,y)}、{Ldis(x,y)}、{RdisThe amplitude and phase of each pixel point in (x, y) in different scales and directions will be { R }orgThe amplitudes of the pixel points with the coordinate positions (x, y) in the (x, y) at different scales and directions are recorded asWill { RorgThe phase positions of the pixel points with the coordinate positions (x, y) in (x, y) at different scales and directions are recorded asWill { LdisThe amplitudes of the pixel points with the coordinate positions (x, y) in the (x, y) at different scales and directions are recorded asWill { LdisThe phase positions of the pixel points with the coordinate positions (x, y) in (x, y) at different scales and directions are recorded asWill { RdisThe amplitudes of the pixel points with the coordinate positions (x, y) in the (x, y) at different scales and directions are recorded asWill { RdisThe phase positions of the pixel points with the coordinate positions (x, y) in (x, y) at different scales and directions are recorded asFor example: obtaining { LdisThe amplitudes of pixel points with (x, y) as coordinate positions in (x, y) in different scales and directionsAnd phaseThe process comprises the following steps: a. using log-Garbor filter pairs { Ldis(x, y) is filtered to obtain { L }disEven symmetric frequency response and odd symmetric frequency response of each pixel point in (x, y) } in different scales and directionsSymmetric frequency response, will { LdisEven symmetric frequency responses of pixel points with coordinate positions (x, y) in different scales and directions in (x, y) are recorded asWill { LdisThe odd symmetric frequency response of the pixel point with the coordinate position (x, y) in different scales and directions is recorded asWherein alpha represents the scale factor of the filter used for filtering, alpha is more than or equal to 1 and less than or equal to 4, theta represents the direction factor of the filter used for filtering, and theta is more than or equal to 1 and less than or equal to 4; calculation of { LdisThe amplitude and phase of each pixel point in (x, y) in different scales and directions will be { L }disThe amplitudes of the pixel points with the coordinate positions (x, y) in the (x, y) at different scales and directions are recorded as <math>
<mrow>
<msubsup>
<mi>LE</mi>
<mrow>
<mi>α</mi>
<mo>,</mo>
<mi>θ</mi>
</mrow>
<mrow>
<mi>L</mi>
<mo>,</mo>
<mi>dis</mi>
</mrow>
</msubsup>
<mrow>
<mo>(</mo>
<mi>x</mi>
<mo>,</mo>
<mi>y</mi>
<mo>)</mo>
</mrow>
<mo>=</mo>
<msqrt>
<msup>
<msubsup>
<mi>e</mi>
<mrow>
<mi>α</mi>
<mo>,</mo>
<mi>θ</mi>
</mrow>
<mi>L</mi>
</msubsup>
<mo>′</mo>
</msup>
<msup>
<mrow>
<mo>(</mo>
<mi>x</mi>
<mo>,</mo>
<mi>y</mi>
<mo>)</mo>
</mrow>
<mn>2</mn>
</msup>
<mo>+</mo>
<msup>
<msubsup>
<mi>o</mi>
<mrow>
<mi>α</mi>
<mo>,</mo>
<mi>θ</mi>
</mrow>
<mi>L</mi>
</msubsup>
<mo>′</mo>
</msup>
<msup>
<mrow>
<mo>(</mo>
<mi>x</mi>
<mo>,</mo>
<mi>y</mi>
<mo>)</mo>
</mrow>
<mn>2</mn>
</msup>
</msqrt>
<mo>,</mo>
</mrow>
</math> Will { LdisThe phase positions of the pixel points with the coordinate positions (x, y) in (x, y) at different scales and directions are recorded as <math>
<mrow>
<msubsup>
<mi>LP</mi>
<mrow>
<mi>α</mi>
<mo>,</mo>
<mi>θ</mi>
</mrow>
<mrow>
<mi>L</mi>
<mo>,</mo>
<mi>dis</mi>
</mrow>
</msubsup>
<mrow>
<mo>(</mo>
<mi>x</mi>
<mo>,</mo>
<mi>y</mi>
<mo>)</mo>
</mrow>
<mo>=</mo>
<mi>arctan</mi>
<mrow>
<mo>(</mo>
<mfrac>
<mrow>
<msup>
<msubsup>
<mi>e</mi>
<mrow>
<mi>α</mi>
<mo>,</mo>
<mi>θ</mi>
</mrow>
<mi>L</mi>
</msubsup>
<mo>′</mo>
</msup>
<mrow>
<mo>(</mo>
<mi>x</mi>
<mo>,</mo>
<mi>y</mi>
<mo>)</mo>
</mrow>
</mrow>
<mrow>
<msup>
<msubsup>
<mi>o</mi>
<mrow>
<mi>α</mi>
<mo>,</mo>
<mi>θ</mi>
</mrow>
<mi>L</mi>
</msubsup>
<mo>′</mo>
</msup>
<mrow>
<mo>(</mo>
<mi>x</mi>
<mo>,</mo>
<mi>y</mi>
<mo>)</mo>
</mrow>
</mrow>
</mfrac>
<mo>)</mo>
</mrow>
<mo>.</mo>
</mrow>
</math>
Thirdly-4, calculating L by adopting a block matching methodorg(x, y) } and { R }org(x, y) } parallax images, noted asWherein,to representThe middle coordinate position is the pixel value of the pixel point of (x, y).
③ 5 according to { Lorg(x, y) } and { R }orgAmplitude and phase of each pixel point in (x, y) in different scales and directions, and { L }org(x, y) } and { R }orgParallax image between (x, y) } sCalculating the pixel value of each pixel point in SorgEnergy diagram at different scales and directions, and is recorded asWill be provided withThe pixel values of the pixel points with the middle coordinate position (x, y) in different scales and directions are recorded as <math>
<mrow>
<msubsup>
<mi>BE</mi>
<mrow>
<mi>α</mi>
<mo>,</mo>
<mi>θ</mi>
</mrow>
<mi>org</mi>
</msubsup>
<mrow>
<mo>(</mo>
<mi>x</mi>
<mo>,</mo>
<mi>y</mi>
<mo>)</mo>
</mrow>
<mo>=</mo>
<msup>
<mrow>
<mo>(</mo>
<msubsup>
<mi>LE</mi>
<mrow>
<mi>α</mi>
<mo>,</mo>
<mi>θ</mi>
</mrow>
<mrow>
<mi>L</mi>
<mo>,</mo>
<mi>org</mi>
</mrow>
</msubsup>
<mrow>
<mo>(</mo>
<mi>x</mi>
<mo>,</mo>
<mi>y</mi>
<mo>)</mo>
</mrow>
<mo>)</mo>
</mrow>
<mn>2</mn>
</msup>
<mo>+</mo>
<msup>
<mrow>
<mo>(</mo>
<msubsup>
<mi>LE</mi>
<mrow>
<mi>α</mi>
<mo>,</mo>
<mi>θ</mi>
</mrow>
<mrow>
<mi>R</mi>
<mo>,</mo>
<mi>org</mi>
</mrow>
</msubsup>
<mrow>
<mo>(</mo>
<mi>x</mi>
<mo>-</mo>
<msubsup>
<mi>d</mi>
<mi>org</mi>
<mi>L</mi>
</msubsup>
<mrow>
<mo>(</mo>
<mi>x</mi>
<mo>,</mo>
<mi>y</mi>
<mo>)</mo>
</mrow>
<mo>,</mo>
<mi>y</mi>
<mo>)</mo>
</mrow>
<mo>)</mo>
</mrow>
<mn>2</mn>
</msup>
</mrow>
</math> <math>
<mrow>
<mo>+</mo>
<msubsup>
<mi>LE</mi>
<mrow>
<mi>α</mi>
<mo>,</mo>
<mi>θ</mi>
</mrow>
<mrow>
<mi>L</mi>
<mo>,</mo>
<mi>org</mi>
</mrow>
</msubsup>
<mrow>
<mo>(</mo>
<mi>x</mi>
<mo>,</mo>
<mi>y</mi>
<mo>)</mo>
</mrow>
<mo>×</mo>
<msubsup>
<mi>LE</mi>
<mrow>
<mi>α</mi>
<mo>,</mo>
<mi>θ</mi>
</mrow>
<mrow>
<mi>R</mi>
<mo>,</mo>
<mi>org</mi>
</mrow>
</msubsup>
<mrow>
<mo>(</mo>
<mi>x</mi>
<mo>-</mo>
<msubsup>
<mi>d</mi>
<mi>org</mi>
<mi>L</mi>
</msubsup>
<mrow>
<mo>(</mo>
<mi>x</mi>
<mo>,</mo>
<mi>y</mi>
<mo>)</mo>
</mrow>
<mo>,</mo>
<mi>y</mi>
<mo>)</mo>
</mrow>
<mo>×</mo>
<mi>cos</mi>
<mrow>
<mo>(</mo>
<msubsup>
<mi>LP</mi>
<mrow>
<mi>α</mi>
<mo>,</mo>
<mi>θ</mi>
</mrow>
<mrow>
<mi>L</mi>
<mo>,</mo>
<mi>org</mi>
</mrow>
</msubsup>
<mrow>
<mo>(</mo>
<mi>x</mi>
<mo>,</mo>
<mi>y</mi>
<mo>)</mo>
</mrow>
<mo>-</mo>
<msubsup>
<mi>LP</mi>
<mrow>
<mi>α</mi>
<mo>,</mo>
<mi>θ</mi>
</mrow>
<mrow>
<mi>R</mi>
<mo>,</mo>
<mi>org</mi>
</mrow>
</msubsup>
<mrow>
<mo>(</mo>
<mi>x</mi>
<mo>-</mo>
<msubsup>
<mi>d</mi>
<mi>org</mi>
<mi>L</mi>
</msubsup>
<mrow>
<mo>(</mo>
<mi>x</mi>
<mo>,</mo>
<mi>y</mi>
<mo>)</mo>
</mrow>
<mo>,</mo>
<mi>y</mi>
<mo>)</mo>
</mrow>
</mrow>
<mo>)</mo>
</mrow>
</math> Wherein cos () is a cosine-taking function,represents { Rorg(x, y) } coordinate position ofThe amplitudes of the pixel points of (a) at different scales and directions,represents { Rorg(x, y) } coordinate position ofThe phase of the pixel point in different scales and directions.
③ 6 according to { Ldis(x, y) } and { R }disAmplitude and phase of each pixel point in (x, y) in different scales and directions, and { L }org(x, y) } and { R }orgParallax image between (x, y) } sCalculating the pixel value of each pixel point in SdisEnergy diagram at different scales and directions, and is recorded asWill be provided withThe pixel values of the pixel points with the middle coordinate position (x, y) in different scales and directions are recorded as <math>
<mrow>
<msubsup>
<mi>BE</mi>
<mrow>
<mi>α</mi>
<mo>,</mo>
<mi>θ</mi>
</mrow>
<mi>dis</mi>
</msubsup>
<mrow>
<mo>(</mo>
<mi>x</mi>
<mo>,</mo>
<mi>y</mi>
<mo>)</mo>
</mrow>
<mo>=</mo>
<msup>
<mrow>
<mo>(</mo>
<msubsup>
<mi>LE</mi>
<mrow>
<mi>α</mi>
<mo>,</mo>
<mi>θ</mi>
</mrow>
<mrow>
<mi>L</mi>
<mo>,</mo>
<mi>dis</mi>
</mrow>
</msubsup>
<mrow>
<mo>(</mo>
<mi>x</mi>
<mo>,</mo>
<mi>y</mi>
<mo>)</mo>
</mrow>
<mo>)</mo>
</mrow>
<mn>2</mn>
</msup>
<mo>+</mo>
<msup>
<mrow>
<mo>(</mo>
<msubsup>
<mi>LE</mi>
<mrow>
<mi>α</mi>
<mo>,</mo>
<mi>θ</mi>
</mrow>
<mrow>
<mi>R</mi>
<mo>,</mo>
<mi>dis</mi>
</mrow>
</msubsup>
<mrow>
<mo>(</mo>
<mi>x</mi>
<mo>-</mo>
<msubsup>
<mi>d</mi>
<mi>org</mi>
<mi>L</mi>
</msubsup>
<mrow>
<mo>(</mo>
<mi>x</mi>
<mo>,</mo>
<mi>y</mi>
<mo>)</mo>
</mrow>
<mo>,</mo>
<mi>y</mi>
<mo>)</mo>
</mrow>
<mo>)</mo>
</mrow>
<mn>2</mn>
</msup>
</mrow>
</math> <math>
<mrow>
<mo>+</mo>
<msubsup>
<mi>LE</mi>
<mrow>
<mi>α</mi>
<mo>,</mo>
<mi>θ</mi>
</mrow>
<mrow>
<mi>L</mi>
<mo>,</mo>
<mi>dis</mi>
</mrow>
</msubsup>
<mrow>
<mo>(</mo>
<mi>x</mi>
<mo>,</mo>
<mi>y</mi>
<mo>)</mo>
</mrow>
<mo>×</mo>
<msubsup>
<mi>LE</mi>
<mrow>
<mi>α</mi>
<mo>,</mo>
<mi>θ</mi>
</mrow>
<mrow>
<mi>R</mi>
<mo>,</mo>
<mi>dis</mi>
</mrow>
</msubsup>
<mrow>
<mo>(</mo>
<mi>x</mi>
<mo>-</mo>
<msubsup>
<mi>d</mi>
<mi>org</mi>
<mi>L</mi>
</msubsup>
<mrow>
<mo>(</mo>
<mi>x</mi>
<mo>,</mo>
<mi>y</mi>
<mo>)</mo>
</mrow>
<mo>,</mo>
<mi>y</mi>
<mo>)</mo>
</mrow>
<mo>×</mo>
<mi>cos</mi>
<msup>
<mrow>
<mo>(</mo>
<mi>L</mi>
<msubsup>
<mi>P</mi>
<mrow>
<mi>α</mi>
<mo>,</mo>
<mi>θ</mi>
</mrow>
<mrow>
<mi>L</mi>
<mo>,</mo>
<mi>dis</mi>
</mrow>
</msubsup>
<mrow>
<mo>(</mo>
<mi>x</mi>
<mo>,</mo>
<mi>y</mi>
<mo>)</mo>
</mrow>
<mo>-</mo>
<msubsup>
<mi>LP</mi>
<mrow>
<mi>α</mi>
<mo>,</mo>
<mi>θ</mi>
</mrow>
<mrow>
<mi>R</mi>
<mo>,</mo>
<mi>dis</mi>
</mrow>
</msubsup>
<mrow>
<mo>(</mo>
<mi>x</mi>
<mo>-</mo>
<msubsup>
<mi>d</mi>
<mi>org</mi>
<mi>L</mi>
</msubsup>
<mrow>
<mo>(</mo>
<mi>x</mi>
<mo>,</mo>
<mi>y</mi>
<mo>)</mo>
</mrow>
<mo>,</mo>
<mi>y</mi>
<mo>)</mo>
</mrow>
<mo>)</mo>
</mrow>
<mo>,</mo>
</msup>
</mrow>
</math> Wherein,represents { Rdis(x, y) } coordinate position ofThe amplitudes of the pixel points of (a) at different scales and directions,represents { Rdis(x, y) } coordinate position ofThe phase of the pixel point in different scales and directions.
Fourthly, according to SorgEnergy diagram at different scales and directionsAnd SdisEnergy diagram at different scales and directionsCalculating SdisThe objective evaluation metric value of each pixel point in SdisThe objective evaluation metric values of all the pixel points in (1) are expressed as { Q (x, y) } by a set, and S is expressed asdisThe objective evaluation metric value of the pixel point with the middle coordinate position (x, y) is marked as Q (x, y), <math>
<mrow>
<mi>Q</mi>
<mrow>
<mo>(</mo>
<mi>x</mi>
<mo>,</mo>
<mi>y</mi>
<mo>)</mo>
</mrow>
<mo>=</mo>
<munderover>
<mi>Σ</mi>
<mrow>
<mi>θ</mi>
<mo>=</mo>
<mn>1</mn>
</mrow>
<mn>4</mn>
</munderover>
<munderover>
<mi>Σ</mi>
<mrow>
<mi>α</mi>
<mo>=</mo>
<mn>1</mn>
</mrow>
<mn>4</mn>
</munderover>
<mi>arccos</mi>
<mrow>
<mo>(</mo>
<mfrac>
<mrow>
<mn>2</mn>
<mo>×</mo>
<msubsup>
<mi>BE</mi>
<mrow>
<mi>α</mi>
<mo>,</mo>
<mi>θ</mi>
</mrow>
<mi>org</mi>
</msubsup>
<mrow>
<mo>(</mo>
<mi>x</mi>
<mo>,</mo>
<mi>y</mi>
<mo>)</mo>
</mrow>
<mo>×</mo>
<msubsup>
<mi>BE</mi>
<mrow>
<mi>α</mi>
<mo>,</mo>
<mi>θ</mi>
</mrow>
<mi>dis</mi>
</msubsup>
<mrow>
<mo>(</mo>
<mi>x</mi>
<mo>,</mo>
<mi>y</mi>
<mo>)</mo>
</mrow>
<mo>+</mo>
<msub>
<mi>T</mi>
<mn>1</mn>
</msub>
</mrow>
<mrow>
<msubsup>
<mi>BE</mi>
<mrow>
<mi>α</mi>
<mo>,</mo>
<mi>θ</mi>
</mrow>
<mi>org</mi>
</msubsup>
<msup>
<mrow>
<mo>(</mo>
<mi>x</mi>
<mo>,</mo>
<mi>y</mi>
<mo>)</mo>
</mrow>
<mn>2</mn>
</msup>
<mo>+</mo>
<msubsup>
<mi>BE</mi>
<mrow>
<mi>α</mi>
<mo>,</mo>
<mi>θ</mi>
</mrow>
<mi>dis</mi>
</msubsup>
<msup>
<mrow>
<mo>(</mo>
<mi>x</mi>
<mo>,</mo>
<mi>y</mi>
<mo>)</mo>
</mrow>
<mn>2</mn>
</msup>
<mo>+</mo>
<msub>
<mi>T</mi>
<mn>1</mn>
</msub>
</mrow>
</mfrac>
<mo>)</mo>
</mrow>
<mo>/</mo>
<mn>16</mn>
<mo>,</mo>
</mrow>
</math> wherein arccos () is an inverted cosine function, T1To control the parameters, in this embodiment, T is taken1Is 16.
The human visual system characteristic shows that human eyes are sensitive to distortion of the area with small binocular perceptibility change value. The invention is thus based on SdisAnd the sum of objective evaluation metrics of each pixel point { Ldis(x, y) } binocular minimum perceivable change imageCalculating S using binocular visual masking effect of human visual systemdisThe objective evaluation metric value for reflecting the binocular visual masking effect is marked as Qbm,Where Ω represents the pixel domain range.
The characteristics of the human visual system show that human eyes can generate stronger response to areas with larger binocular energy intensity. The invention is thus based on SdisAnd the objective evaluation metric value of each pixel point in the image data and SdisEnergy diagram at different scales and directionsCalculating S by using the response characteristics of the human visual system to the binocular energy intensitydisThe objective evaluation metric value for reflecting the binocular energy intensity is marked as Qbe, <math>
<mrow>
<msub>
<mi>Q</mi>
<mi>be</mi>
</msub>
<mo>=</mo>
<mfrac>
<mrow>
<munder>
<mi>Σ</mi>
<mrow>
<mrow>
<mo>(</mo>
<mi>x</mi>
<mo>,</mo>
<mi>y</mi>
<mo>)</mo>
</mrow>
<mo>∈</mo>
<mi>Ω</mi>
</mrow>
</munder>
<mi>Q</mi>
<mrow>
<mo>(</mo>
<mi>x</mi>
<mo>,</mo>
<mi>y</mi>
<mo>)</mo>
</mrow>
<mo>×</mo>
<mrow>
<mo>(</mo>
<mn>1</mn>
<mo>/</mo>
<msup>
<mi>BE</mi>
<mi>dis</mi>
</msup>
<mrow>
<mo>(</mo>
<mi>x</mi>
<mo>,</mo>
<mi>y</mi>
<mo>)</mo>
</mrow>
<mo>)</mo>
</mrow>
</mrow>
<mrow>
<munder>
<mi>Σ</mi>
<mrow>
<mrow>
<mo>(</mo>
<mi>x</mi>
<mo>,</mo>
<mi>y</mi>
<mo>)</mo>
</mrow>
<mo>∈</mo>
<mi>Ω</mi>
</mrow>
</munder>
<mrow>
<mo>(</mo>
<mn>1</mn>
<mo>/</mo>
<msup>
<mi>BE</mi>
<mi>dis</mi>
</msup>
<mrow>
<mo>(</mo>
<mi>x</mi>
<mo>,</mo>
<mi>y</mi>
<mo>)</mo>
</mrow>
<mo>)</mo>
</mrow>
</mrow>
</mfrac>
<mo>,</mo>
</mrow>
</math> Wherein, <math>
<mrow>
<msup>
<mi>BE</mi>
<mi>dis</mi>
</msup>
<mrow>
<mo>(</mo>
<mi>x</mi>
<mo>,</mo>
<mi>y</mi>
<mo>)</mo>
</mrow>
<mo>=</mo>
<munderover>
<mi>Σ</mi>
<mrow>
<mi>θ</mi>
<mo>=</mo>
<mn>1</mn>
</mrow>
<mn>4</mn>
</munderover>
<munderover>
<mi>Σ</mi>
<mrow>
<mi>α</mi>
<mo>=</mo>
<mn>1</mn>
</mrow>
<mn>4</mn>
</munderover>
<msubsup>
<mi>BE</mi>
<mrow>
<mi>α</mi>
<mo>,</mo>
<mi>θ</mi>
</mrow>
<mi>dis</mi>
</msubsup>
<mrow>
<mo>(</mo>
<mi>x</mi>
<mo>,</mo>
<mi>y</mi>
<mo>)</mo>
</mrow>
<mo>.</mo>
</mrow>
</math>
to SdisObjective evaluation metric Q for reflecting binocular visual masking effectbmAnd SdisObjective evaluation metric Q for reflecting binocular energy intensitybeCarrying out fusion to obtain SdisThe predicted value of the objective evaluation of image quality is marked as Qfin,Qfin=(Qbm)γ(Qbe)βWhere γ and β are weight parameters, in this embodiment, γ =0.7755 and β =0.0505 are taken.
In the present embodiment, 12 undistorted stereoscopic images shown in fig. 2a and 2b, fig. 3a and 3b, fig. 4a and 4b, fig. 5a and 5b, fig. 6a and 6b, fig. 7a and 7b, fig. 8a and 8b, fig. 9a and 9b, fig. 10a and 10b, fig. 11a and 11b, fig. 12a and 12b, and fig. 13a and 13b are used to create the stereoscopic image without distortionThe distorted stereo image S obtained by the method of the invention is analyzed by using 312 distorted stereo images under the conditions of different degrees of JPEG compression, JPEG2000 compression, Gaussian blur, white noise and H.264 coding distortiondisThe correlation between the predicted value and the average subjective score difference is objectively evaluated, wherein 60 distorted stereoscopic images are compressed by JPEG, 60 distorted stereoscopic images are compressed by JPEG2000, 60 distorted stereoscopic images are compressed by Gaussian Blur, 60 distorted stereoscopic images are blurred by White Noise (White Noise), and 72 distorted stereoscopic images are encoded by H.264. And respectively obtaining average subjective score difference values of 312 distorted stereo images by using the existing subjective quality evaluation method, and recording the average subjective score difference values as DMOS, DMOS =100-MOS, wherein MOS represents the subjective score average value, DMOS belongs to [0,100 ]]。
In this embodiment, 4 common objective parameters of the image quality evaluation method are used as evaluation indexes, that is, Pearson correlation coefficient (PLCC), Spearman correlation coefficient (SROCC), Kendall correlation coefficient (KROCC), mean square error (RMSE), accuracy of the stereo image evaluation objective model in which distortion is reflected by PLCC and rmocc, and monotonicity of SROCC and KROCC is reflected by KROCC under a nonlinear regression condition. The image quality objective evaluation predicted value of the distorted stereo image calculated according to the method is subjected to five-parameter Logistic function nonlinear fitting, and the higher the PLCC, SROCC and KROCC values are, the lower the RMSE value is, the better the correlation between the objective evaluation method and the average subjective score difference is. The PLCC, SROCC, KROCC and RMSE coefficients reflecting the performance of the three-dimensional image objective evaluation model are shown in the table 1, and the data listed in the table 1 shows that the correlation between the final image quality objective evaluation predicted value of the distorted three-dimensional image obtained by the method and the average subjective score difference value is very high, so that the objective evaluation result is fully consistent with the result of human eye subjective perception, and the effectiveness of the method is sufficiently proved.
Fig. 14 shows a scatter diagram of the difference between the objective evaluation prediction value of the image quality and the average subjective score of 312 distorted stereoscopic images, and the more concentrated the scatter is, the better the consistency between the objective evaluation result and the subjective perception is. As can be seen from fig. 14, the scatter diagram obtained by the method of the present invention is more concentrated, and the goodness of fit with the subjective evaluation data is higher.
TABLE 1 correlation between objective evaluation prediction value and subjective score of image quality of distorted stereoscopic image obtained by the method of the present invention
Claims (7)
1. A three-dimensional image quality objective evaluation method based on an energy map is characterized in that the processing process is as follows: firstly, acquiring an energy diagram of the original undistorted stereo image in different scales and directions according to even symmetric frequency response and odd symmetric frequency response of each pixel point in the left viewpoint image and the right viewpoint image of the original undistorted stereo image in different scales and directions and a pixel value of each pixel point in a parallax image between the left viewpoint image and the right viewpoint image of the original undistorted stereo image, acquiring energy graphs of the distorted stereo image to be evaluated in different scales and directions according to even symmetric frequency response and odd symmetric frequency response of each pixel point in the left viewpoint image and the right viewpoint image of the distorted stereo image to be evaluated in different scales and directions and pixel values of each pixel point in a parallax image between the left viewpoint image and the right viewpoint image of the original undistorted stereo image; then obtaining an objective evaluation metric value of each pixel point in the distorted stereo image to be evaluated according to the two energy maps; obtaining an objective evaluation metric value of the distorted three-dimensional image to be evaluated for reflecting a binocular visual masking effect according to the objective evaluation metric value of each pixel point in the distorted three-dimensional image to be evaluated and the binocular minimum perceptible change image of the left viewpoint image of the distorted three-dimensional image to be evaluated, and obtaining an objective evaluation metric value of the distorted three-dimensional image to be evaluated for reflecting binocular energy intensity according to the objective evaluation metric value of each pixel point in the distorted three-dimensional image to be evaluated and energy maps of the distorted three-dimensional image to be evaluated in different scales and directions; finally, objective evaluation metric values used for reflecting binocular visual masking effects and objective evaluation metric values used for reflecting binocular energy intensity of the distorted three-dimensional image to be evaluated are fused to obtain an objective evaluation prediction value of the image quality of the distorted three-dimensional image to be evaluated;
here, let SorgFor original undistorted stereo image, let SdisFor the distorted stereo image to be evaluated, SorgIs noted as { Lorg(x, y) }, adding SorgIs noted as { Rorg(x, y) }, adding SdisIs noted as { Ldis(x, y) }, adding SdisIs noted as { Rdis(x, y) }, wherein (x, y) denotes a coordinate position of a pixel point in the left viewpoint image and the right viewpoint image, x is 1. ltoreq. x.ltoreq.W, y is 1. ltoreq. y.ltoreq.H, W denotes a width of the left viewpoint image and the right viewpoint image, H denotes a height of the left viewpoint image and the right viewpoint image, L is Lorg(x, y) represents { L }orgThe coordinate position in (x, y) } is the pixel value of the pixel point with (x, y), Rorg(x, y) represents { Rorg(x, y) having (x, y) as the coordinate positionPixel value, L, of a pixel pointdis(x, y) represents { L }disThe coordinate position in (x, y) } is the pixel value of the pixel point with (x, y), Rdis(x, y) represents { RdisThe coordinate position in (x, y) is the pixel value of the pixel point of (x, y); then according to { Lorg(x, y) } and { R }orgAmplitude and phase of each pixel point in (x, y) in different scales and directions, and { L }org(x, y) } and { R }org(x, y) } calculating S from the pixel value of each pixel in the parallax imageorgEnergy diagram at different scales and directions, and is recorded asAnd according to { Ldis(x, y) } and { R }disAmplitude and phase of each pixel point in (x, y) in different scales and directions, and { L }org(x, y) } and { R }org(x, y) } calculating S from the pixel value of each pixel in the parallax imagedisEnergy diagram at different scales and directions, and is recorded asWherein,to representThe pixel value of the pixel point with the middle coordinate position (x, y) in different scales and directions,to representAnd the pixel values of the pixel points with the middle coordinate positions (x, y) in different scales and directions, wherein alpha represents the scale factor of the filter adopted by filtering, alpha is more than or equal to 1 and less than or equal to 4, theta represents the direction factor of the filter adopted by filtering, and theta is more than or equal to 1 and less than or equal to 4.
2. The method for objectively evaluating the quality of a stereoscopic image based on an energy map according to claim 1, characterized in that it comprises the following steps:
making SorgFor original undistorted stereo image, let SdisFor the distorted stereo image to be evaluated, SorgIs noted as { Lorg(x, y) }, adding SorgIs noted as { Rorg(x, y) }, adding SdisIs noted as { Ldis(x, y) }, adding SdisIs noted as { Rdis(x, y) }, wherein (x, y) denotes a coordinate position of a pixel point in the left viewpoint image and the right viewpoint image, x is 1. ltoreq. x.ltoreq.W, y is 1. ltoreq. y.ltoreq.H, W denotes a width of the left viewpoint image and the right viewpoint image, H denotes a height of the left viewpoint image and the right viewpoint image, L is Lorg(x, y) represents { L }orgThe coordinate position in (x, y) } is the pixel value of the pixel point with (x, y), Rorg(x, y) represents { RorgThe pixel value L of the pixel point with the coordinate position (x, y) in (x, y) } isdis(x, y) represents { L }disThe coordinate position in (x, y) } is the pixel value of the pixel point with (x, y), Rdis(x, y) represents { RdisThe coordinate position in (x, y) is the pixel value of the pixel point of (x, y);
secondly, extracting { L by using visual masking effect of human stereoscopic vision perception on background illumination and contrastdis(x, y) } binocular minimum perceivable change image, notedWherein,represents { Ldis(x, y) } binocular minimum perceivable change imageThe middle coordinate position is the pixel value of the pixel point of (x, y);
(iii) calculating { Lorg(x,y)}、{Rorg(x,y)}、{Ldis(x,y)}、{RdisEven symmetric frequency response and odd symmetric frequency response of each pixel point in (x, y) } in different scales and directions; then obtain { Lorg(x,y)}、{Rorg(x,y)}、{Ldis(x,y)}、{RdisThe amplitude and the phase of each pixel point in (x, y) in different scales and directions; then according to { Lorg(x, y) } and { R }orgAmplitude and phase of each pixel point in (x, y) in different scales and directions, and { L }org(x, y) } and { R }org(x, y) } calculating S from the pixel value of each pixel in the parallax imageorgEnergy diagram at different scales and directions, and is recorded asAnd according to { Ldis(x, y) } and { R }disAmplitude and phase of each pixel point in (x, y) in different scales and directions, and { L }org(x, y) } and { R }org(x, y) } calculating S from the pixel value of each pixel in the parallax imagedisEnergy diagram at different scales and directions, and is recorded asWherein,to representThe pixel value of the pixel point with the middle coordinate position (x, y) in different scales and directions,to representThe pixel values of the pixel points with the (x, y) coordinate positions in different scales and directions are determined, alpha represents the scale factor of the filter adopted by filtering, alpha is more than or equal to 1 and less than or equal to 4, theta represents the direction factor of the filter adopted by filtering, and theta is more than or equal to 1 and less than or equal to 4θ≤4;
Fourthly, according to SorgEnergy diagram at different scales and directionsAnd SdisEnergy diagram at different scales and directionsCalculating SdisThe objective evaluation metric value of each pixel point in SdisThe objective evaluation metric value of the pixel point with the middle coordinate position (x, y) is marked as Q (x, y), <math>
<mrow>
<mi>Q</mi>
<mrow>
<mo>(</mo>
<mi>x</mi>
<mo>,</mo>
<mi>y</mi>
<mo>)</mo>
</mrow>
<mo>=</mo>
<munderover>
<mi>Σ</mi>
<mrow>
<mi>θ</mi>
<mo>=</mo>
<mn>1</mn>
</mrow>
<mn>4</mn>
</munderover>
<munderover>
<mi>Σ</mi>
<mrow>
<mi>α</mi>
<mo>=</mo>
<mn>1</mn>
</mrow>
<mn>4</mn>
</munderover>
<mi>arccos</mi>
<mrow>
<mo>(</mo>
<mfrac>
<mrow>
<mn>2</mn>
<mo>×</mo>
<msubsup>
<mi>BE</mi>
<mrow>
<mi>α</mi>
<mo>,</mo>
<mi>θ</mi>
</mrow>
<mi>org</mi>
</msubsup>
<mrow>
<mo>(</mo>
<mi>x</mi>
<mo>,</mo>
<mi>y</mi>
<mo>)</mo>
</mrow>
<mo>×</mo>
<msubsup>
<mi>BE</mi>
<mrow>
<mi>α</mi>
<mo>,</mo>
<mi>θ</mi>
</mrow>
<mi>dis</mi>
</msubsup>
<mrow>
<mo>(</mo>
<mi>x</mi>
<mo>,</mo>
<mi>y</mi>
<mo>)</mo>
</mrow>
<mo>+</mo>
<msub>
<mi>T</mi>
<mn>1</mn>
</msub>
</mrow>
<mrow>
<msubsup>
<mi>BE</mi>
<mrow>
<mi>α</mi>
<mo>,</mo>
<mi>θ</mi>
</mrow>
<mi>org</mi>
</msubsup>
<msup>
<mrow>
<mo>(</mo>
<mi>x</mi>
<mo>,</mo>
<mi>y</mi>
<mo>)</mo>
</mrow>
<mn>2</mn>
</msup>
<mo>+</mo>
<msubsup>
<mi>BE</mi>
<mrow>
<mi>α</mi>
<mo>,</mo>
<mi>θ</mi>
</mrow>
<mi>dis</mi>
</msubsup>
<msup>
<mrow>
<mo>(</mo>
<mi>x</mi>
<mo>,</mo>
<mi>y</mi>
<mo>)</mo>
</mrow>
<mn>2</mn>
</msup>
<mo>+</mo>
<msub>
<mi>T</mi>
<mn>1</mn>
</msub>
</mrow>
</mfrac>
<mo>)</mo>
</mrow>
<mo>/</mo>
<mn>16</mn>
<mo>,</mo>
</mrow>
</math> wherein arccos () is an inverted cosine function, T1Is a control parameter;
according to SdisAnd the sum of objective evaluation metrics of each pixel point { Ldis(x, y) } binocular minimum perceivable change imageCalculating S using binocular visual masking effect of human visual systemdisThe objective evaluation metric value for reflecting the binocular visual masking effect is marked as Qbm,Wherein Ω represents a pixel domain range;
according to SdisAnd the objective evaluation metric value of each pixel point in the image data and SdisEnergy diagram at different scales and directionsCalculating S by using the response characteristics of the human visual system to the binocular energy intensitydisThe objective evaluation metric value for reflecting the binocular energy intensity is marked as Qbe,Wherein, <math>
<mrow>
<msup>
<mi>BE</mi>
<mi>dis</mi>
</msup>
<mrow>
<mo>(</mo>
<mi>x</mi>
<mo>,</mo>
<mi>y</mi>
<mo>)</mo>
</mrow>
<mo>=</mo>
<munderover>
<mi>Σ</mi>
<mrow>
<mi>θ</mi>
<mo>=</mo>
<mn>1</mn>
</mrow>
<mn>4</mn>
</munderover>
<munderover>
<mi>Σ</mi>
<mrow>
<mi>α</mi>
<mo>=</mo>
<mn>1</mn>
</mrow>
<mn>4</mn>
</munderover>
<msubsup>
<mi>BE</mi>
<mrow>
<mi>α</mi>
<mo>,</mo>
<mi>θ</mi>
</mrow>
<mi>dis</mi>
</msubsup>
<mrow>
<mo>(</mo>
<mi>x</mi>
<mo>,</mo>
<mi>y</mi>
<mo>)</mo>
</mrow>
<mo>;</mo>
</mrow>
</math>
to SdisObjective evaluation metric Q for reflecting binocular visual masking effectbmAnd SdisObjective evaluation metric Q for reflecting binocular energy intensitybeCarrying out fusion to obtain SdisThe predicted value of the objective evaluation of image quality is marked as Qfin,Qfin=(Qbm)γ(Qbe)βWherein γ and β are weight parameters.
3. The objective evaluation method for the quality of the stereo image based on the energy map as claimed in claim 2, characterized in that the concrete process of the step (II) is as follows:
② 1, calculating { LdisVisualization threshold set of luminance masking effect of (x, y) }, denoted as { T }l(x,y)},Wherein, Tl(x, y) represents { L }disThe coordinate position in (x, y) is the visual threshold value of the brightness masking effect of the pixel point of (x, y), bgl(x, y) represents { L }disThe average value of the brightness of all pixel points in an NxN neighborhood window taking a pixel point with a coordinate position (x, y) as the center in (x, y) }, wherein N is more than or equal to 1;
2, calculating { LdisVisualization threshold set of contrast masking effect of (x, y) }, denoted as { T }c(x,y)},Tc(x,y)=K(bgl(x,y))+ehl(x, y) wherein Tc(x, y) represents { L }disThe coordinate position in (x, y) is the visual threshold value of the contrast masking effect of the pixel point of (x, y), ehl(x, y) represents the pair { L }disThe pixel points with the coordinate positions (x, y) in (x, y) are respectively subjected to edge filtering in the horizontal direction and the vertical direction to obtain an average gradient value K (bg)l(x,y))=-10-6×(0.7×bgl(x,y)2+32×bgl(x,y))+0.07;
2- (3) pairs of { LdisVisualization threshold set of luminance masking effects of (x, y) } { Tl(x, y) } and a visual threshold set of contrast masking effects { T }c(x, y) } to obtain { Ldis(x, y) } binocular minimum perceivable change image, notedWill be provided withThe pixel value of the pixel point with the middle coordinate position (x, y) is recorded as
4. The method according to claim 3, wherein in step (2), N is 5.
5. The objective evaluation method for stereo image quality based on energy map according to any one of claims 2 to 4, characterized in that the concrete process of the third step is:
③ 1, adopting log-Garbor filter pair { Lorg(x, y) is filtered to obtain { L }orgEven symmetric frequency response and odd symmetric frequency response of each pixel point in (x, y) } in different scales and directions are converted into { L }orgEven symmetric frequency responses of pixel points with coordinate positions (x, y) in different scales and directions in (x, y) are recorded asWill { LorgThe odd symmetric frequency response of the pixel point with the coordinate position (x, y) in different scales and directions is recorded asWhere α represents the scale of the filter employed for filteringThe factor alpha is more than or equal to 1 and less than or equal to 4, theta represents the direction factor of the filter used for filtering, and theta is more than or equal to 1 and less than or equal to 4;
③ 2, calculate { LorgThe amplitude and phase of each pixel point in (x, y) in different scales and directions will be { L }orgThe amplitudes of the pixel points with the coordinate positions (x, y) in the (x, y) at different scales and directions are recorded as Will { LorgThe phase positions of the pixel points with the coordinate positions (x, y) in (x, y) at different scales and directions are recorded as Wherein arctan () is an inverse tangent function;
③ 3, obtaining { L ] according to the steps from the third step to the third steporgThe operation process of the amplitude and the phase of each pixel point in (x, y) } in different scales and directions is carried out to obtain { R } in the same wayorg(x,y)}、{Ldis(x,y)}、{RdisThe amplitude and phase of each pixel point in (x, y) in different scales and directions will be { R }orgThe amplitudes of the pixel points with the coordinate positions (x, y) in the (x, y) at different scales and directions are recorded asWill { RorgThe phase positions of the pixel points with the coordinate positions (x, y) in (x, y) at different scales and directions are recorded asWill { LdisThe amplitudes of the pixel points with the coordinate positions (x, y) in the (x, y) at different scales and directions are recorded asWill { LdisThe phase positions of the pixel points with the coordinate positions (x, y) in (x, y) at different scales and directions are recorded asWill { RdisThe amplitudes of the pixel points with the coordinate positions (x, y) in the (x, y) at different scales and directions are recorded asWill { RdisThe phase positions of the pixel points with the coordinate positions (x, y) in (x, y) at different scales and directions are recorded as
Thirdly-4, calculating L by adopting a block matching methodorg(x, y) } and { R }org(x, y) } parallax images, noted asWherein,to representThe middle coordinate position is the pixel value of the pixel point of (x, y);
③ 5 according to { Lorg(x, y) } and { R }orgAmplitude and phase of each pixel point in (x, y) in different scales and directions, and { L }org(x, y) } and { R }orgParallax image between (x, y) } sCalculating the pixel value of each pixel point in SorgEnergy diagram at different scales and directions, and is recorded asWill be provided withThe pixel values of the pixel points with the middle coordinate position (x, y) in different scales and directions are recorded as
③ 6 according to { Ldis(x, y) } and { R }disAmplitude and phase of each pixel point in (x, y) in different scales and directions, and { L }org(x, y) } and { R }orgParallax image between (x, y) } sCalculating the pixel value of each pixel point in SdisEnergy diagram at different scales and directions, and is recorded asWill be provided withThe pixel values of the pixel points with the middle coordinate position (x, y) in different scales and directions are recorded as <math>
<mrow>
<mfenced open='' close=''>
<mtable>
<mtr>
<mtd>
<msubsup>
<mi>BE</mi>
<mrow>
<mi>α</mi>
<mo>,</mo>
<mi>θ</mi>
</mrow>
<mi>dis</mi>
</msubsup>
<mrow>
<mo>(</mo>
<mi>x</mi>
<mo>,</mo>
<mi>y</mi>
<mo>)</mo>
</mrow>
<mo>=</mo>
<msup>
<mrow>
<mo>(</mo>
<msubsup>
<mi>LE</mi>
<mrow>
<mi>α</mi>
<mo>,</mo>
<mi>θ</mi>
</mrow>
<mrow>
<mi>L</mi>
<mo>,</mo>
<mi>dis</mi>
</mrow>
</msubsup>
<mrow>
<mo>(</mo>
<mi>x</mi>
<mo>,</mo>
<mi>y</mi>
<mo>)</mo>
</mrow>
<mo>)</mo>
</mrow>
<mn>2</mn>
</msup>
<mo>+</mo>
<msup>
<mrow>
<mo>(</mo>
<msubsup>
<mi>LE</mi>
<mrow>
<mi>α</mi>
<mo>,</mo>
<mi>θ</mi>
</mrow>
<mrow>
<mi>R</mi>
<mo>,</mo>
<mi>dis</mi>
</mrow>
</msubsup>
<mrow>
<mo>(</mo>
<mi>x</mi>
<mo>-</mo>
<msubsup>
<mi>d</mi>
<mi>org</mi>
<mi>L</mi>
</msubsup>
<mrow>
<mo>(</mo>
<mi>x</mi>
<mo>,</mo>
<mi>y</mi>
<mo>)</mo>
</mrow>
<mo>,</mo>
<mi>y</mi>
<mo>)</mo>
</mrow>
<mo>)</mo>
</mrow>
<mn>2</mn>
</msup>
</mtd>
</mtr>
<mtr>
<mtd>
<mo>+</mo>
<msubsup>
<mi>LE</mi>
<mrow>
<mi>α</mi>
<mo>,</mo>
<mi>θ</mi>
</mrow>
<mrow>
<mi>L</mi>
<mo>,</mo>
<mi>dis</mi>
</mrow>
</msubsup>
<mrow>
<mo>(</mo>
<mi>x</mi>
<mo>,</mo>
<mi>y</mi>
<mo>)</mo>
</mrow>
<mo>×</mo>
<msubsup>
<mi>LE</mi>
<mrow>
<mi>α</mi>
<mo>,</mo>
<mi>θ</mi>
</mrow>
<mrow>
<mi>R</mi>
<mo>,</mo>
<mi>dis</mi>
</mrow>
</msubsup>
<mrow>
<mo>(</mo>
<mi>x</mi>
<mo>-</mo>
<msubsup>
<mi>d</mi>
<mi>org</mi>
<mi>L</mi>
</msubsup>
<mrow>
<mo>(</mo>
<mi>x</mi>
<mo>,</mo>
<mi>y</mi>
<mo>)</mo>
</mrow>
<mo>,</mo>
<mi>Y</mi>
<mo>)</mo>
</mrow>
<mo>×</mo>
<mi>cos</mi>
<mrow>
<mo>(</mo>
<msubsup>
<mi>LP</mi>
<mrow>
<mi>α</mi>
<mo>,</mo>
<mi>θ</mi>
</mrow>
<mrow>
<mi>L</mi>
<mo>,</mo>
<mi>dis</mi>
</mrow>
</msubsup>
<mrow>
<mo>(</mo>
<mi>x</mi>
<mo>,</mo>
<mi>y</mi>
<mo>)</mo>
</mrow>
<mo>-</mo>
<msubsup>
<mi>LP</mi>
<mrow>
<mi>α</mi>
<mo>,</mo>
<mi>θ</mi>
</mrow>
<mrow>
<mi>R</mi>
<mo>,</mo>
<mi>dis</mi>
</mrow>
</msubsup>
<mrow>
<mo>(</mo>
<mi>x</mi>
<mo>-</mo>
<msubsup>
<mi>d</mi>
<mi>org</mi>
<mi>L</mi>
</msubsup>
<mrow>
<mo>(</mo>
<mi>x</mi>
<mo>,</mo>
<mi>y</mi>
<mo>)</mo>
</mrow>
<mo>,</mo>
<mi>y</mi>
<mo>)</mo>
</mrow>
<mo>)</mo>
</mrow>
</mtd>
</mtr>
</mtable>
</mfenced>
<mo>,</mo>
</mrow>
</math> Wherein,represents { Rdis(x, y) } coordinate position ofThe amplitudes of the pixel points of (a) at different scales and directions,represents { Rdis(x, y) } coordinate position ofThe phase of the pixel point in different scales and directions.
6. The objective evaluation method for stereo image quality based on energy map as claimed in claim 5, wherein T is taken in the step (iv)1Is 16.
7. The method as claimed in claim 6, wherein in step (c) γ 0.7755 and β 0.0505 are included.
Priority Applications (1)
Application Number | Priority Date | Filing Date | Title |
---|---|---|---|
CN201210493072.5A CN102999911B (en) | 2012-11-27 | 2012-11-27 | Three-dimensional image quality objective evaluation method based on energy diagrams |
Applications Claiming Priority (1)
Application Number | Priority Date | Filing Date | Title |
---|---|---|---|
CN201210493072.5A CN102999911B (en) | 2012-11-27 | 2012-11-27 | Three-dimensional image quality objective evaluation method based on energy diagrams |
Publications (2)
Publication Number | Publication Date |
---|---|
CN102999911A CN102999911A (en) | 2013-03-27 |
CN102999911B true CN102999911B (en) | 2015-06-03 |
Family
ID=47928445
Family Applications (1)
Application Number | Title | Priority Date | Filing Date |
---|---|---|---|
CN201210493072.5A Expired - Fee Related CN102999911B (en) | 2012-11-27 | 2012-11-27 | Three-dimensional image quality objective evaluation method based on energy diagrams |
Country Status (1)
Country | Link |
---|---|
CN (1) | CN102999911B (en) |
Families Citing this family (5)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
CN106683072B (en) * | 2015-11-09 | 2020-02-21 | 上海交通大学 | 3D image comfort level quality evaluation method and system based on PUP image |
CN105828061B (en) * | 2016-05-11 | 2017-09-29 | 宁波大学 | A kind of virtual view quality evaluating method of view-based access control model masking effect |
CN109285146B (en) * | 2018-08-30 | 2021-12-10 | 南京邮电大学 | Full-reference image quality evaluation method based on just noticeable distortion |
JP7168848B2 (en) * | 2018-11-21 | 2022-11-10 | 日本電信電話株式会社 | Evaluation device, evaluation method, and program. |
CN112353365B (en) * | 2021-01-14 | 2021-03-23 | 广州市诺以德医疗科技发展有限公司 | Binocular dynamic energy stereoscopic vision detection system and detection method |
Citations (3)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
CN102142145A (en) * | 2011-03-22 | 2011-08-03 | 宁波大学 | Image quality objective evaluation method based on human eye visual characteristics |
CN102547368A (en) * | 2011-12-16 | 2012-07-04 | 宁波大学 | Objective evaluation method for quality of stereo images |
CN102663747A (en) * | 2012-03-23 | 2012-09-12 | 宁波大学 | Stereo image objectivity quality evaluation method based on visual perception |
-
2012
- 2012-11-27 CN CN201210493072.5A patent/CN102999911B/en not_active Expired - Fee Related
Patent Citations (3)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
CN102142145A (en) * | 2011-03-22 | 2011-08-03 | 宁波大学 | Image quality objective evaluation method based on human eye visual characteristics |
CN102547368A (en) * | 2011-12-16 | 2012-07-04 | 宁波大学 | Objective evaluation method for quality of stereo images |
CN102663747A (en) * | 2012-03-23 | 2012-09-12 | 宁波大学 | Stereo image objectivity quality evaluation method based on visual perception |
Non-Patent Citations (2)
Title |
---|
《Research on Stereoscopic Images Quality Assessment Methods》;Wujie Zhou et al;《International Conference of China Communication and Information Technology》;20101001;273-279 * |
《一种基于最小可察觉失真的立体图像质量客观评价方法》;顾珊波等;《光电子·激光》;20120531;第23卷(第5期);999-1004 * |
Also Published As
Publication number | Publication date |
---|---|
CN102999911A (en) | 2013-03-27 |
Similar Documents
Publication | Publication Date | Title |
---|---|---|
CN102708567B (en) | Visual perception-based three-dimensional image quality objective evaluation method | |
CN103152600B (en) | Three-dimensional video quality evaluation method | |
CN103413298B (en) | A kind of objective evaluation method for quality of stereo images of view-based access control model characteristic | |
CN104036501B (en) | A kind of objective evaluation method for quality of stereo images based on rarefaction representation | |
CN102333233B (en) | Stereo image quality objective evaluation method based on visual perception | |
CN104243976B (en) | A kind of three-dimensional image objective quality evaluation method | |
CN104394403B (en) | A kind of stereoscopic video quality method for objectively evaluating towards compression artefacts | |
US20140064604A1 (en) | Method for objectively evaluating quality of stereo image | |
CN102843572B (en) | Phase-based stereo image quality objective evaluation method | |
CN102999911B (en) | Three-dimensional image quality objective evaluation method based on energy diagrams | |
CN104954778B (en) | Objective stereo image quality assessment method based on perception feature set | |
CN104658001A (en) | Non-reference asymmetric distorted stereo image objective quality assessment method | |
CN102903107B (en) | Three-dimensional picture quality objective evaluation method based on feature fusion | |
CN102521825B (en) | Three-dimensional image quality objective evaluation method based on zero watermark | |
CN104036502B (en) | A kind of without with reference to fuzzy distortion stereo image quality evaluation methodology | |
CN103136748B (en) | The objective evaluation method for quality of stereo images of a kind of feature based figure | |
CN105282543A (en) | Total blindness three-dimensional image quality objective evaluation method based on three-dimensional visual perception | |
CN103369348B (en) | Three-dimensional image quality objective evaluation method based on regional importance classification | |
CN103200420B (en) | Three-dimensional picture quality objective evaluation method based on three-dimensional visual attention | |
CN102999912B (en) | A kind of objective evaluation method for quality of stereo images based on distortion map | |
CN107360416A (en) | Stereo image quality evaluation method based on local multivariate Gaussian description | |
CN102737380B (en) | Stereo image quality objective evaluation method based on gradient structure tensor | |
CN103841411A (en) | Method for evaluating quality of stereo image based on binocular information processing | |
CN103745457A (en) | Stereo image objective quality evaluation method | |
CN103281556B (en) | Objective evaluation method for stereo image quality on the basis of image decomposition |
Legal Events
Date | Code | Title | Description |
---|---|---|---|
C06 | Publication | ||
PB01 | Publication | ||
C10 | Entry into substantive examination | ||
SE01 | Entry into force of request for substantive examination | ||
C14 | Grant of patent or utility model | ||
GR01 | Patent grant | ||
CF01 | Termination of patent right due to non-payment of annual fee | ||
CF01 | Termination of patent right due to non-payment of annual fee |
Granted publication date: 20150603 Termination date: 20171127 |