CN102903107B - Three-dimensional picture quality objective evaluation method based on feature fusion - Google Patents

Three-dimensional picture quality objective evaluation method based on feature fusion Download PDF

Info

Publication number
CN102903107B
CN102903107B CN201210357956.8A CN201210357956A CN102903107B CN 102903107 B CN102903107 B CN 102903107B CN 201210357956 A CN201210357956 A CN 201210357956A CN 102903107 B CN102903107 B CN 102903107B
Authority
CN
China
Prior art keywords
org
dis
pixel
mrow
image
Prior art date
Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
Expired - Fee Related
Application number
CN201210357956.8A
Other languages
Chinese (zh)
Other versions
CN102903107A (en
Inventor
邵枫
段芬芳
蒋刚毅
郁梅
李福翠
Current Assignee (The listed assignees may be inaccurate. Google has not performed a legal analysis and makes no representation or warranty as to the accuracy of the list.)
Jiangsu Qizhen Information Technology Service Co ltd
Original Assignee
Ningbo University
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Application filed by Ningbo University filed Critical Ningbo University
Priority to CN201210357956.8A priority Critical patent/CN102903107B/en
Publication of CN102903107A publication Critical patent/CN102903107A/en
Application granted granted Critical
Publication of CN102903107B publication Critical patent/CN102903107B/en
Expired - Fee Related legal-status Critical Current
Anticipated expiration legal-status Critical

Links

Landscapes

  • Image Processing (AREA)

Abstract

The invention discloses a three-dimensional picture quality objective evaluation method based on feature fusion. The method includes that at first a single-eye image of an original non-distortion three-dimensional image and a single-eye image of a distortion three-dimensional image to be evaluated are calculated respectively, by calculating mean value and a standard difference of pixel value in each pixel point in the two single-eye images, objective evaluation metric of each pixel point in the single-eye image of the distortion three-dimensional image to be evaluated is obtained, remarkable images of the two single-eye images and a distortion image between the two single-eye images are calculated respectively, the objective evaluation metric of each pixel point in the single-eye image of the distortion three-dimensional image to be evaluated is fused, and an image quality objective evaluation predicating value of the distortion three-dimensional image to be evaluated is obtained. The three-dimensional picture quality objective evaluation method has the advantages that the obtained single-eye images can simulate a double-eye three-dimensional fusion process well, and by fusing the remarkable images and the distortion images, relevance of objective evaluation results and subjective perception can be effectively improved.

Description

Stereo image quality objective evaluation method based on feature fusion
Technical Field
The invention relates to an image quality evaluation method, in particular to a stereo image quality objective evaluation method based on feature fusion.
Background
With the rapid development of image coding technology and stereoscopic display technology, the stereoscopic image technology has received more and more extensive attention and application, and has become a current research hotspot. The stereo image technology utilizes the binocular parallax principle of human eyes, the left and right viewpoint images from the same scene are respectively and independently received by binoculars, and binocular parallax is formed through brain fusion, so that the stereo image with depth feeling and reality feeling is appreciated. Due to the influence of an acquisition system and storage compression and transmission equipment, a series of distortions are inevitably introduced into the stereo image, and compared with a single-channel image, the stereo image needs to ensure the image quality of two channels simultaneously, so that the quality evaluation of the stereo image is of great significance. However, currently, there is no effective objective evaluation method for evaluating the quality of stereoscopic images. Therefore, establishing an effective objective evaluation model of the quality of the stereo image has very important significance.
The existing objective evaluation method for the quality of the stereo image is to directly apply the plane image quality evaluation method to the evaluation of the quality of the stereo image, however, the process of fusing the left and right viewpoint images of the stereo image to generate the stereo effect is not a simple process of superposing the left and right viewpoint images and is difficult to express by a simple mathematical method, so that how to effectively simulate the binocular stereo fusion in the stereo image quality evaluation process and how to extract effective characteristic information to fuse the evaluation result, the objective evaluation result is more in line with the human visual system, and the problem needs to be researched and solved in the process of evaluating the objective quality of the stereo image.
Disclosure of Invention
The invention aims to solve the technical problem of providing a three-dimensional image quality objective evaluation method based on feature fusion, which can effectively improve the correlation between objective evaluation results and subjective perception.
The technical scheme adopted by the invention for solving the technical problems is as follows: a stereo image quality objective evaluation method based on feature fusion is characterized in that the processing process is as follows: firstly, obtaining a single eye diagram of an original undistorted stereo image according to even symmetric frequency response and odd symmetric frequency response of each pixel point in a left viewpoint image and a right viewpoint image of the original undistorted stereo image at different scales and directions and a parallax image between the left viewpoint image and the right viewpoint image of the original undistorted stereo image; obtaining a single eye diagram of the distorted stereo image to be evaluated according to even symmetric frequency response and odd symmetric frequency response of each pixel point in the left viewpoint image and the right viewpoint image of the distorted stereo image to be evaluated in different scales and directions and a parallax image between the left viewpoint image and the right viewpoint image of the original undistorted stereo image; secondly, obtaining an objective evaluation metric value of each pixel point in the single eye diagram of the distorted stereo image to be evaluated according to the mean value and the standard deviation of the pixel values of each pixel point in the two single eye diagrams; thirdly, obtaining a corresponding saliency map according to the amplitude and the phase of the single eye map of the original undistorted stereo image; obtaining a corresponding saliency map according to the amplitude and the phase of the single eye map of the distorted stereo image to be evaluated; then, according to the two saliency maps and the distortion map between the two single eye maps, fusing objective evaluation metric values of each pixel point in the single eye map of the distorted three-dimensional image to be evaluated to obtain an objective evaluation prediction value of the image quality of the distorted three-dimensional image to be evaluated; and finally, obtaining the image quality objective evaluation predicted value of the distorted three-dimensional images with different distortion types and different distortion degrees according to the processing process.
The invention relates to a method for objectively evaluating the quality of a stereo image based on feature fusion, which comprises the following specific steps:
making SorgFor original undistorted stereo image, let SdisFor the distorted stereo image to be evaluated, SorgIs noted as { Lorg(x, y) }, adding SorgIs noted as { Rorg(x, y) }, adding SdisIs noted as { Ldis(x, y) }, adding SdisIs noted as { Rdis(x, y) }, where (x, y) here denotes the coordinate positions of pixel points in the left viewpoint image and the right viewpoint image, x is 1. ltoreq. W, y is 1. ltoreq. H, W denotes the widths of the left viewpoint image and the right viewpoint image, H denotes the heights of the left viewpoint image and the right viewpoint image, Lorg(x, y) represents { L }orgThe coordinate position in (x, y) } is the pixel value of the pixel point with (x, y), Rorg(x, y) represents { RorgThe pixel value L of the pixel point with the coordinate position (x, y) in (x, y) } isdis(x, y) represents { L }disThe coordinate position in (x, y) } is the pixel value of the pixel point with (x, y), Rdis(x, y) represents { RdisThe coordinate position in (x, y) is the pixel value of the pixel point of (x, y);
② according to { Lorg(x,y)}、{Rorg(x,y)}、{Ldis(x,y)}、{RdisEven symmetric frequency response and odd symmetric frequency response of each pixel point in (x, y) } in different scales and directions are correspondingly obtained to obtain { L }org(x,y)}、{Rorg(x,y)}、{Ldis(x,y)}、{RdisThe amplitude of each pixel in (x, y) } is then based on { L }org(x, y) } and { R }orgAmplitude of each pixel in (x, y) } and { Lorg(x, y) } and { R }org(x, y) } calculating S from the pixel value of each pixel in the parallax imageorgThe eye-independent diagram of (2), is marked as { CMorg(x, y) }, and according to { L }dis(x, y) } and { R }disAmplitude of each pixel in (x, y) } and { Lorg(x, y) } and { R }org(x, y) } calculating S from the pixel value of each pixel in the parallax imagedisThe eye-independent diagram of (2), is marked as { CMdis(x, y) }, wherein CMorg(x, y) denotes { CMorgThe pixel value of the pixel point with the coordinate position (x, y) in (x, y) }, CMdis(x, y) denotes { CMdisThe coordinate position in (x, y) is the pixel value of the pixel point of (x, y);
③ according to { CMorg(x, y) } and { CMdisCalculating the mean and standard deviation of the pixel values of each pixel in (x, y) } to calculate { CMdis(x, y) } the objective evaluation metric for each pixel point, will be { CMdisThe objective evaluation metric value of the pixel point with the coordinate position (x, y) in (x, y) is recorded as Qimage(x,y);
Fourthly, according to { CMorg(x, y) } amplitude and phase, calculating { CMorg(x, y) } significant graph, denoted as { SMorg(x, y) }, according to { CMdis(x, y) } amplitude and phase, calculating { CMdis(x, y) } significant graph, denoted as { SMdis(x, y) }, wherein, SMorg(x, y) denotes { SMorgThe pixel value of the pixel point whose coordinate position is (x, y), SMdis(x, y) denotes { SMdisThe coordinate position in (x, y) is the pixel value of the pixel point of (x, y);
fifthly, calculating { CMorg(x, y) } and { CMdisThe distortion map between (x, y) } is denoted as { DM (x, y) }, the pixel value of the pixel point with the coordinate position (x, y) in { DM (x, y) } is denoted as DM (x, y), and DM (x, y) ═ CMorg(x,y)-CMdis(x,y))2
According to { SMorg(x, y) } and { SMdis(x, y) } and { DM (x, y) }, for { CMdisThe objective evaluation metric values of all pixel points in (x, y) are fused to obtain SdisThe image quality objective evaluation predicted value of (1) is marked as Q, <math> <mrow> <mi>Q</mi> <mo>=</mo> <msup> <mrow> <mo>[</mo> <mfrac> <mrow> <munder> <mi>&Sigma;</mi> <mrow> <mrow> <mo>(</mo> <mi>x</mi> <mo>,</mo> <mi>y</mi> <mo>)</mo> </mrow> <mo>&Element;</mo> <mi>&Omega;</mi> </mrow> </munder> <msub> <mi>Q</mi> <mi>image</mi> </msub> <mrow> <mo>(</mo> <mi>x</mi> <mo>,</mo> <mi>y</mi> <mo>)</mo> </mrow> <mo>&times;</mo> <mi>SM</mi> <mrow> <mo>(</mo> <mi>x</mi> <mo>,</mo> <mi>y</mi> <mo>)</mo> </mrow> </mrow> <mrow> <munder> <mi>&Sigma;</mi> <mrow> <mrow> <mo>(</mo> <mi>x</mi> <mo>,</mo> <mi>y</mi> <mo>)</mo> </mrow> <mo>&Element;</mo> <mi>&Omega;</mi> </mrow> </munder> <mi>SM</mi> <mrow> <mo>(</mo> <mi>x</mi> <mo>,</mo> <mi>y</mi> <mo>)</mo> </mrow> </mrow> </mfrac> <mo>]</mo> </mrow> <mi>&gamma;</mi> </msup> <mo>&times;</mo> <msup> <mrow> <mo>[</mo> <mfrac> <mrow> <munder> <mi>&Sigma;</mi> <mrow> <mrow> <mo>(</mo> <mi>x</mi> <mo>,</mo> <mi>y</mi> <mo>)</mo> </mrow> <mo>&Element;</mo> <mi>&Omega;</mi> </mrow> </munder> <msub> <mi>Q</mi> <mi>image</mi> </msub> <mrow> <mo>(</mo> <mi>x</mi> <mo>,</mo> <mi>y</mi> <mo>)</mo> </mrow> <mo>&times;</mo> <mi>DM</mi> <mrow> <mo>(</mo> <mi>x</mi> <mo>,</mo> <mi>y</mi> <mo>)</mo> </mrow> </mrow> <mrow> <munder> <mi>&Sigma;</mi> <mrow> <mrow> <mo>(</mo> <mi>x</mi> <mo>,</mo> <mi>y</mi> <mo>)</mo> </mrow> <mo>&Element;</mo> <mi>&Omega;</mi> </mrow> </munder> <mi>DM</mi> <mrow> <mo>(</mo> <mi>x</mi> <mo>,</mo> <mi>y</mi> <mo>)</mo> </mrow> </mrow> </mfrac> <mo>]</mo> </mrow> <mi>&beta;</mi> </msup> <mo>,</mo> </mrow> </math> where Ω denotes a pixel domain range, and SM (x, y) is max (SM)org(x,y),SMdis(x, y)), max () is a function taking the maximum value, and γ and β are weight coefficients;
adopting n original undistorted stereo images, establishing a distorted stereo image set of the undistorted stereo images under different distortion types and different distortion degrees, wherein the distorted stereo image set comprises a plurality of distorted stereo images, respectively obtaining an average subjective score difference of each distorted stereo image in the distorted stereo image set by using a subjective quality evaluation method, and marking as DMOS (diffusion metal oxide semiconductor) which is 100-MOS (metal oxide semiconductor), wherein MOS represents a subjective score mean value, DMOS belongs to [0,100], and n is more than or equal to 1;
calculating S according to the steps from the first step to the sixth stepdisOperation of the image quality objective evaluation prediction valueAnd calculating the image quality objective evaluation prediction value of each distorted stereo image in the distorted stereo image set respectively.
The concrete process of the second step is as follows:
2-1, pair { Lorg(x, y) is filtered to obtain { L }orgEven symmetric frequency response and odd symmetric frequency response of each pixel point in (x, y) } in different scales and directions are converted into { L }orgEven symmetric frequency responses of pixel points with coordinate positions (x, y) in different scales and directions are recorded as eα,θ(x, y) will { LorgThe odd symmetric frequency response of the pixel point with the coordinate position (x, y) in different scales and directions is marked as oα,θ(x, y), wherein alpha represents the scale factor of the filter used for filtering, alpha is more than or equal to 1 and less than or equal to 4, theta represents the direction factor of the filter used for filtering, and theta is more than or equal to 1 and less than or equal to 4;
2 according to { LorgCalculating the even symmetric frequency response and the odd symmetric frequency response of each pixel point in (x, y) in different scales and directions, and calculating the { L }orgThe amplitude of each pixel in (x, y) } will be { LorgThe amplitude of the pixel point with the coordinate position (x, y) in (x, y) is recorded as <math> <mrow> <msubsup> <mi>GE</mi> <mi>org</mi> <mi>L</mi> </msubsup> <mrow> <mo>(</mo> <mi>x</mi> <mo>,</mo> <mi>y</mi> <mo>)</mo> </mrow> <mo>=</mo> <munderover> <mi>&Sigma;</mi> <mrow> <mi>&theta;</mi> <mo>=</mo> <mn>1</mn> </mrow> <mn>4</mn> </munderover> <munderover> <mi>&Sigma;</mi> <mrow> <mi>&alpha;</mi> <mo>=</mo> <mn>1</mn> </mrow> <mn>4</mn> </munderover> <msqrt> <msub> <mi>e</mi> <mrow> <mi>&alpha;</mi> <mo>,</mo> <mi>&theta;</mi> </mrow> </msub> <msup> <mrow> <mo>(</mo> <mi>x</mi> <mo>,</mo> <mi>y</mi> <mo>)</mo> </mrow> <mn>2</mn> </msup> <mo>+</mo> <msub> <mi>o</mi> <mrow> <mi>&alpha;</mi> <mo>,</mo> <mi>&theta;</mi> </mrow> </msub> <msup> <mrow> <mo>(</mo> <mi>x</mi> <mo>,</mo> <mi>y</mi> <mo>)</mo> </mrow> <mn>2</mn> </msup> </msqrt> <mo>;</mo> </mrow> </math>
② -3, obtaining { L ] according to the steps from ② -1 to ② -2orgOperation of amplitude of each pixel in (x, y) } acquires { R } in the same mannerorg(x,y)}、{Ldis(x, y) } and { R }dis(x, y) } the amplitude of each pixel point in the (x, y) } will be { RorgThe amplitude of the pixel point with the coordinate position (x, y) in (x, y) is recorded asWill { LdisThe amplitude of the pixel point with the coordinate position (x, y) in (x, y) is recorded asWill { RdisThe amplitude of the pixel point with the coordinate position (x, y) in (x, y) is recorded as
② 4, calculating { L by adopting a block matching methodorg(x, y) } and { R }org(x, y) } parallax images, noted asWherein,to representThe middle coordinate position is the pixel value of the pixel point of (x, y);
② 5 according to { Lorg(x, y) } and { R }orgAmplitude of each pixel in (x, y) } andcalculating the pixel value of each pixel point in SorgThe eye-independent diagram of (2), is marked as { CMorg(x, y) }, will { CMorgThe pixel value of the pixel point with the coordinate position (x, y) in (x, y) is marked as CMorg(x,y), <math> <mrow> <msub> <mi>CM</mi> <mi>org</mi> </msub> <mrow> <mo>(</mo> <mi>x</mi> <mo>,</mo> <mi>y</mi> <mo>)</mo> </mrow> <mo>=</mo> <mfrac> <mrow> <msubsup> <mi>GE</mi> <mi>org</mi> <mi>L</mi> </msubsup> <mrow> <mo>(</mo> <mi>x</mi> <mo>,</mo> <mi>y</mi> <mo>)</mo> </mrow> <mo>&times;</mo> <msub> <mi>L</mi> <mi>org</mi> </msub> <mrow> <mo>(</mo> <mi>x</mi> <mo>,</mo> <mi>y</mi> <mo>)</mo> </mrow> <mo>+</mo> <msubsup> <mi>GE</mi> <mi>org</mi> <mi>R</mi> </msubsup> <mrow> <mo>(</mo> <mi>x</mi> <mo>-</mo> <msubsup> <mi>d</mi> <mi>org</mi> <mi>L</mi> </msubsup> <mrow> <mo>(</mo> <mi>x</mi> <mo>,</mo> <mi>y</mi> <mo>)</mo> </mrow> <mo>,</mo> <mi>y</mi> <mo>)</mo> </mrow> <mo>&times;</mo> <msub> <mi>R</mi> <mi>org</mi> </msub> <mrow> <mo>(</mo> <mi>x</mi> <mo>-</mo> <msubsup> <mi>d</mi> <mi>org</mi> <mi>L</mi> </msubsup> <mrow> <mo>(</mo> <mi>x</mi> <mo>,</mo> <mi>y</mi> <mo>)</mo> </mrow> <mo>,</mo> <mi>y</mi> <mo>)</mo> </mrow> </mrow> <mrow> <msubsup> <mi>GE</mi> <mi>org</mi> <mi>L</mi> </msubsup> <mrow> <mo>(</mo> <mi>x</mi> <mo>,</mo> <mi>y</mi> <mo>)</mo> </mrow> <mo>+</mo> <msubsup> <mi>GE</mi> <mi>org</mi> <mi>R</mi> </msubsup> <mrow> <mo>(</mo> <mi>x</mi> <mo>-</mo> <msubsup> <mi>d</mi> <mi>org</mi> <mi>L</mi> </msubsup> <mrow> <mo>(</mo> <mi>x</mi> <mo>,</mo> <mi>y</mi> <mo>)</mo> </mrow> <mo>,</mo> <mi>y</mi> <mo>)</mo> </mrow> </mrow> </mfrac> <mo>,</mo> </mrow> </math> Wherein,represents { Rorg(x, y) } coordinate position ofThe amplitude of the pixel points of (a) is,represents { Rorg(x, y) } coordinate position ofThe pixel value of the pixel point of (1);
② 6 according to { Ldis(x, y) } and { R }disAmplitude of each pixel in (x, y) } andcalculating the pixel value of each pixel point in SdisThe eye-independent diagram of (2), is marked as { CMdis(x, y) }, will { CMdisThe pixel value of the pixel point with the coordinate position (x, y) in (x, y) is marked as CMdis(x,y), <math> <mrow> <msub> <mi>CM</mi> <mi>dis</mi> </msub> <mrow> <mo>(</mo> <mi>x</mi> <mo>,</mo> <mi>y</mi> <mo>)</mo> </mrow> <mo>=</mo> <mfrac> <mrow> <msubsup> <mi>GE</mi> <mi>dis</mi> <mi>L</mi> </msubsup> <mrow> <mo>(</mo> <mi>x</mi> <mo>,</mo> <mi>y</mi> <mo>)</mo> </mrow> <mo>&times;</mo> <msub> <mi>L</mi> <mi>dis</mi> </msub> <mrow> <mo>(</mo> <mi>x</mi> <mo>,</mo> <mi>y</mi> <mo>)</mo> </mrow> <mo>+</mo> <msubsup> <mi>GE</mi> <mi>dis</mi> <mi>R</mi> </msubsup> <mrow> <mo>(</mo> <mi>x</mi> <mo>-</mo> <msubsup> <mi>d</mi> <mi>org</mi> <mi>L</mi> </msubsup> <mrow> <mo>(</mo> <mi>x</mi> <mo>,</mo> <mi>y</mi> <mo>)</mo> </mrow> <mo>,</mo> <mi>y</mi> <mo>)</mo> </mrow> <mo>&times;</mo> <msub> <mi>R</mi> <mi>dis</mi> </msub> <mrow> <mo>(</mo> <mi>x</mi> <mo>-</mo> <msubsup> <mi>d</mi> <mi>org</mi> <mi>L</mi> </msubsup> <mrow> <mo>(</mo> <mi>x</mi> <mo>,</mo> <mi>y</mi> <mo>)</mo> </mrow> <mo>,</mo> <mi>y</mi> <mo>)</mo> </mrow> </mrow> <mrow> <msubsup> <mi>GE</mi> <mi>dis</mi> <mi>L</mi> </msubsup> <mrow> <mo>(</mo> <mi>x</mi> <mo>,</mo> <mi>y</mi> <mo>)</mo> </mrow> <mo>+</mo> <msubsup> <mi>GE</mi> <mi>dis</mi> <mi>R</mi> </msubsup> <mrow> <mo>(</mo> <mi>x</mi> <mo>-</mo> <msubsup> <mi>d</mi> <mi>org</mi> <mi>L</mi> </msubsup> <mrow> <mo>(</mo> <mi>x</mi> <mo>,</mo> <mi>y</mi> <mo>)</mo> </mrow> <mo>,</mo> <mi>y</mi> <mo>)</mo> </mrow> </mrow> </mfrac> <mo>,</mo> </mrow> </math> Wherein,represents { Rdis(x, y) } coordinate position ofThe amplitude of the pixel points of (a) is,represents { Rdis(x, y) } coordinate position ofThe pixel value of the pixel point of (1).
In the step II-1, the pairs { Lorg(x, y) } the filter used for the filtering process is a log-Garbor filter.
The concrete process of the step III is as follows:
③ 1, calculate { CMorg(x, y) } and { CMdisThe mean and standard deviation of the pixel values for each pixel in (x, y) } will be { CMorgThe coordinate position in (x, y) } is (x)1,y1) Respectively recording the mean value and the standard difference of the pixel values of the pixel points as muorg(x1,y1) And σorg(x1,y1) Will { CMdisThe coordinate position in (x, y) } is (x)1,y1) Respectively recording the mean value and the standard difference of the pixel values of the pixel points as mudis(x1,y1) And σdis(x1,y1),
<math> <mrow> <msub> <mi>&sigma;</mi> <mi>org</mi> </msub> <mrow> <mo>(</mo> <msub> <mi>x</mi> <mn>1</mn> </msub> <mo>,</mo> <msub> <mi>y</mi> <mn>1</mn> </msub> <mo>)</mo> </mrow> <mo>=</mo> <msqrt> <mfrac> <mrow> <munder> <mi>&Sigma;</mi> <mrow> <mrow> <mo>(</mo> <msub> <mi>x</mi> <mn>1</mn> </msub> <mo>,</mo> <msub> <mi>y</mi> <mn>1</mn> </msub> <mo>)</mo> </mrow> <mo>&Element;</mo> <mi>N</mi> <mrow> <mo>(</mo> <msub> <mi>x</mi> <mn>1</mn> </msub> <mo>,</mo> <msub> <mi>y</mi> <mn>1</mn> </msub> <mo>)</mo> </mrow> </mrow> </munder> <msup> <mrow> <mo>(</mo> <msub> <mi>CM</mi> <mi>org</mi> </msub> <mrow> <mo>(</mo> <msub> <mi>x</mi> <mn>1</mn> </msub> <mo>,</mo> <msub> <mi>y</mi> <mn>1</mn> </msub> <mo>)</mo> </mrow> <mo>-</mo> <msub> <mi>&mu;</mi> <mi>org</mi> </msub> <mrow> <mo>(</mo> <msub> <mi>x</mi> <mn>1</mn> </msub> <mo>,</mo> <msub> <mi>y</mi> <mn>1</mn> </msub> <mo>)</mo> </mrow> <mo>)</mo> </mrow> <mn>2</mn> </msup> </mrow> <mi>M</mi> </mfrac> </msqrt> <mo>,</mo> </mrow> </math>
<math> <mrow> <msub> <mi>&mu;</mi> <mi>dis</mi> </msub> <mrow> <mo>(</mo> <msub> <mi>x</mi> <mn>1</mn> </msub> <mo>,</mo> <msub> <mi>y</mi> <mn>1</mn> </msub> <mo>)</mo> </mrow> <mo>=</mo> <mfrac> <mrow> <munder> <mi>&Sigma;</mi> <mrow> <mrow> <mo>(</mo> <msub> <mi>x</mi> <mn>1</mn> </msub> <mo>,</mo> <msub> <mi>y</mi> <mn>1</mn> </msub> <mo>)</mo> </mrow> <mo>&Element;</mo> <mrow> <mo>(</mo> <msub> <mi>x</mi> <mn>1</mn> </msub> <mo>,</mo> <msub> <mi>y</mi> <mn>1</mn> </msub> <mo>)</mo> </mrow> </mrow> </munder> <msub> <mi>CM</mi> <mi>dis</mi> </msub> <mrow> <mo>(</mo> <msub> <mi>x</mi> <mn>1</mn> </msub> <mo>,</mo> <msub> <mi>y</mi> <mn>1</mn> </msub> <mo>)</mo> </mrow> </mrow> <mi>M</mi> </mfrac> <mo>,</mo> </mrow> </math>
<math> <mrow> <msub> <mi>&sigma;</mi> <mi>dis</mi> </msub> <mrow> <mo>(</mo> <msub> <mi>x</mi> <mn>1</mn> </msub> <mo>,</mo> <msub> <mi>y</mi> <mn>1</mn> </msub> <mo>)</mo> </mrow> <mo>=</mo> <msqrt> <mfrac> <mrow> <munder> <mi>&Sigma;</mi> <mrow> <mrow> <mo>(</mo> <msub> <mi>x</mi> <mn>1</mn> </msub> <mo>,</mo> <msub> <mi>y</mi> <mn>1</mn> </msub> <mo>)</mo> </mrow> <mo>&Element;</mo> <mi>N</mi> <mrow> <mo>(</mo> <msub> <mi>x</mi> <mn>1</mn> </msub> <mo>,</mo> <msub> <mi>y</mi> <mn>1</mn> </msub> <mo>)</mo> </mrow> </mrow> </munder> <msup> <mrow> <mo>(</mo> <msub> <mi>CM</mi> <mi>dis</mi> </msub> <mrow> <mo>(</mo> <msub> <mi>x</mi> <mn>1</mn> </msub> <mo>,</mo> <msub> <mi>y</mi> <mn>1</mn> </msub> <mo>)</mo> </mrow> <mo>-</mo> <msub> <mi>&mu;</mi> <mi>dis</mi> </msub> <mrow> <mo>(</mo> <msub> <mi>x</mi> <mn>1</mn> </msub> <mo>,</mo> <msub> <mi>y</mi> <mn>1</mn> </msub> <mo>)</mo> </mrow> <mo>)</mo> </mrow> <mn>2</mn> </msup> </mrow> <mi>M</mi> </mfrac> </msqrt> <mo>,</mo> </mrow> </math> Wherein x is more than or equal to 11≤W,1≤y1≤H,N(x1,y1) The coordinate position is shown as (x)1,y1) The pixel point of (a) is an 8 x 8 neighborhood window with the center, M represents N (x)1,y1) Number of inner pixels, CMorg(x1,y1) Representation { CMorgThe coordinate position in (x, y) } is (x)1,y1) Pixel value of the pixel point, CMdis(x1,y1) Representation { CMdisThe coordinate position in (x, y) } is (x)1,y1) The pixel value of the pixel point of (1);
③ 2 according to { CMorg(x, y) } and { CMdisCalculating the mean and standard deviation of the pixel values of each pixel in (x, y) } to calculate { CMdis(x, y) } the objective evaluation metric for each pixel point, will be { CMdisThe coordinate position in (x, y) } is (x)1,y1) The objective evaluation metric value of the pixel point is recorded as Qimage(x1,y1), <math> <mrow> <msub> <mi>Q</mi> <mi>image</mi> </msub> <mrow> <mo>(</mo> <msub> <mi>x</mi> <mn>1</mn> </msub> <mo>,</mo> <msub> <mi>y</mi> <mn>1</mn> </msub> <mo>)</mo> </mrow> <mo>=</mo> <mfrac> <mrow> <mn>4</mn> <mo>&times;</mo> <mrow> <mo>(</mo> <msub> <mi>&mu;</mi> <mi>org</mi> </msub> <mrow> <mo>(</mo> <msub> <mi>x</mi> <mn>1</mn> </msub> <mo>,</mo> <msub> <mi>y</mi> <mn>1</mn> </msub> <mo>)</mo> </mrow> <mo>&times;</mo> <msub> <mi>&mu;</mi> <mi>dis</mi> </msub> <mrow> <mo>(</mo> <msub> <mi>x</mi> <mn>1</mn> </msub> <mo>,</mo> <msub> <mi>y</mi> <mn>1</mn> </msub> <mo>)</mo> </mrow> <mo>)</mo> </mrow> <mo>&times;</mo> <mrow> <mo>(</mo> <msub> <mi>&sigma;</mi> <mi>org</mi> </msub> <mrow> <mo>(</mo> <msub> <mi>x</mi> <mn>1</mn> </msub> <mo>,</mo> <msub> <mi>y</mi> <mn>1</mn> </msub> <mo>)</mo> </mrow> <mo>&times;</mo> <msub> <mi>&sigma;</mi> <mi>dis</mi> </msub> <mrow> <mo>(</mo> <msub> <mi>x</mi> <mn>1</mn> </msub> <mo>,</mo> <msub> <mi>y</mi> <mn>1</mn> </msub> <mo>)</mo> </mrow> <mo>)</mo> </mrow> <mo>+</mo> <mi>C</mi> </mrow> <mrow> <mrow> <mo>(</mo> <msub> <mi>&mu;</mi> <mi>org</mi> </msub> <msup> <mrow> <mo>(</mo> <msub> <mi>x</mi> <mn>1</mn> </msub> <mo>,</mo> <msub> <mi>y</mi> <mn>1</mn> </msub> <mo>)</mo> </mrow> <mn>2</mn> </msup> <mo>+</mo> <msub> <mi>&mu;</mi> <mi>dis</mi> </msub> <msup> <mrow> <mo>(</mo> <msub> <mi>x</mi> <mn>1</mn> </msub> <mo>,</mo> <msub> <mi>y</mi> <mn>1</mn> </msub> <mo>)</mo> </mrow> <mn>2</mn> </msup> <mo>)</mo> </mrow> <mo>&times;</mo> <mrow> <mo>(</mo> <msub> <mi>&sigma;</mi> <mi>org</mi> </msub> <msup> <mrow> <mo>(</mo> <msub> <mi>x</mi> <mn>1</mn> </msub> <mo>,</mo> <msub> <mi>y</mi> <mn>1</mn> </msub> <mo>)</mo> </mrow> <mn>2</mn> </msup> <mo>+</mo> <msub> <mi>&sigma;</mi> <mi>dis</mi> </msub> <msup> <mrow> <mo>(</mo> <msub> <mi>x</mi> <mn>1</mn> </msub> <mo>,</mo> <msub> <mi>y</mi> <mn>1</mn> </msub> <mo>)</mo> </mrow> <mn>2</mn> </msup> <mo>)</mo> </mrow> <mo>+</mo> <mi>C</mi> </mrow> </mfrac> <mo>,</mo> </mrow> </math> Wherein C is a control parameter.
The specific process of the step IV is as follows:
tetra-1, pair { CMorg(x, y) performing discrete Fourier transform to obtain (CM)orgThe amplitude and phase of (x, y) } are denoted as { M, respectivelyorg(u, v) } and { Aorg(u, v) }, where u denotes the width of the amplitude or phase of the transform domain, v denotes the height of the amplitude or phase of the transform domain, 1. ltoreq. u.ltoreq.W, 1. ltoreq. v.ltoreq.H, Morg(u, v) represents { M }org(u, v) } the amplitude value of the pixel point with the coordinate position of (u, v), Aorg(u, v) represents { A }org(u, v) } the coordinate position is the phase value of the pixel point with (u, v);
fourthly-2, calculating { Morg(u, v) } amplitude of the high frequency component, denoted as { R }org(u, v) }, will { RorgThe amplitude value of the high-frequency component of the pixel point with the coordinate position (u, v) in (u, v) is recorded as Rorg(u,v),Rorg(u,v)=log(Morg(u,v))-hm(u,v)*log(Morg(u, v)), wherein log () is a logarithmic function based on e, e-2.718281828, "+" is the convolution operator, hm(u, v) represents an m × m mean filtering;
fourthly-3, according to { Rorg(u, v) } and { Aorg(u, v) } inverse discrete Fourier transform, and the obtained inverse transform image is taken as { CMorg(x, y) } significant graph, denoted as { SMorg(x, y) }, wherein, SMorg(x, y) denotes { SMorgThe coordinate position in (x, y) is the pixel value of the pixel point of (x, y);
fourthly-4, obtaining the { CM according to the steps from the fourth step-1 to the fourth step-3orgOperation of the saliency map of (x, y) } acquisition of { CM in the same mannerdis(x, y) } significant graph, denoted as { SMdis(x, y) }, wherein, SMdis(x, y) denotes { SMdis(x, y) } image with coordinate position (x, y)Pixel values of the pixel points.
Compared with the prior art, the invention has the advantages that:
1) the method of the invention respectively calculates the single eye diagram of the original undistorted stereo image and the single eye diagram of the distorted stereo image to be evaluated, and directly evaluates the single eye diagram of the distorted stereo image, thereby effectively simulating the binocular stereo fusion process and avoiding the process of linearly weighting the objective evaluation metric values of the left viewpoint image and the right viewpoint image.
2) The method can ensure that the evaluation result is more perceptually accordant with the human visual system by calculating the single eye diagram of the original undistorted stereo image, the salient diagram of the single eye diagram of the distorted stereo image to be evaluated and the distortion diagram between the two single eye diagrams and fusing the objective evaluation metric value of each pixel point in the single eye diagram of the distorted stereo image to be evaluated, thereby effectively improving the correlation between the objective evaluation result and the subjective perception.
Drawings
FIG. 1 is a block diagram of an overall implementation of the method of the present invention;
fig. 2a is a left viewpoint image of Akko (640 × 480 size) stereo image;
fig. 2b is a right viewpoint image of an Akko (size 640 × 480) stereoscopic image;
fig. 3a is a left viewpoint image of an altmobit (size 1024 × 768) stereoscopic image;
fig. 3b is a right view image of an altmobit (size 1024 × 768) stereoscopic image;
fig. 4a is a left viewpoint image of a balloon (size 1024 × 768) stereoscopic image;
fig. 4b is a right viewpoint image of a balloon (size 1024 × 768) stereoscopic image;
fig. 5a is a left viewpoint image of a Doorflower (size 1024 × 768) stereoscopic image;
fig. 5b is a right viewpoint image of a Doorflower (size 1024 × 768) stereoscopic image;
fig. 6a is a left view image of a Kendo (size 1024 × 768) stereoscopic image;
fig. 6b is a right view image of a Kendo (size 1024 × 768) stereoscopic image;
fig. 7a is a left view image of a LeaveLaptop (size 1024 × 768) stereoscopic image;
fig. 7b is a right view image of a LeaveLaptop (size 1024 × 768) stereoscopic image;
fig. 8a is a left viewpoint image of a lovedual 1 (size 1024 × 768) stereoscopic image;
fig. 8b is a right viewpoint image of a lovedual 1 (size 1024 × 768) stereoscopic image;
fig. 9a is a left view image of a newsapper (size 1024 × 768) stereoscopic image;
fig. 9b is a right view image of a newsapper (size 1024 × 768) stereoscopic image;
FIG. 10a is a left viewpoint image of Puppy (size 720 × 480) stereo image;
FIG. 10b is a right viewpoint image of Puppy (size 720 × 480) stereoscopic image;
fig. 11a is a left viewpoint image of a Soccer2 (size 720 × 480) stereoscopic image;
fig. 11b is a right viewpoint image of a Soccer2 (size 720 × 480) stereoscopic image;
fig. 12a is a left viewpoint image of a Horse (size 720 × 480) stereoscopic image;
fig. 12b is a right view image of a Horse (size 720 × 480) stereoscopic image;
fig. 13a is a left viewpoint image of an Xmas (size 640 × 480) stereoscopic image;
fig. 13b is a right view image of an Xmas (size 640 × 480) stereoscopic image;
fig. 14 is a scatter plot of the difference between the objective evaluation prediction value of image quality and the average subjective score for each distorted stereoscopic image in the set of distorted stereoscopic images.
Detailed Description
The invention is described in further detail below with reference to the accompanying examples.
The invention provides a method for objectively evaluating the quality of a stereo image based on feature fusion, the overall implementation block diagram of which is shown in figure 1, and the processing process is as follows: firstly, obtaining a single eye diagram of an original undistorted stereo image according to even symmetric frequency response and odd symmetric frequency response of each pixel point in a left viewpoint image and a right viewpoint image of the original undistorted stereo image at different scales and directions and a parallax image between the left viewpoint image and the right viewpoint image of the original undistorted stereo image; obtaining a single eye diagram of the distorted stereo image to be evaluated according to even symmetric frequency response and odd symmetric frequency response of each pixel point in the left viewpoint image and the right viewpoint image of the distorted stereo image to be evaluated in different scales and directions and a parallax image between the left viewpoint image and the right viewpoint image of the original undistorted stereo image; secondly, obtaining an objective evaluation metric value of each pixel point in the single eye diagram of the distorted stereo image to be evaluated according to the mean value and the standard deviation of the pixel values of each pixel point in the two single eye diagrams; thirdly, obtaining a corresponding saliency map according to the amplitude and the phase of the single eye map of the original undistorted stereo image; obtaining a corresponding saliency map according to the amplitude and the phase of the single eye map of the distorted stereo image to be evaluated; then, according to the two saliency maps and the distortion map between the two single eye maps, fusing objective evaluation metric values of each pixel point in the single eye map of the distorted three-dimensional image to be evaluated to obtain an objective evaluation prediction value of the image quality of the distorted three-dimensional image to be evaluated; and finally, obtaining the image quality objective evaluation predicted value of the distorted three-dimensional images with different distortion types and different distortion degrees according to the processing process.
The method specifically comprises the following steps:
making SorgFor original undistorted stereo image, let SdisFor the distorted stereo image to be evaluated, SorgIs noted as { Lorg(x, y) }, adding SorgIs noted as { Rorg(x, y) }, adding SdisIs noted as { Ldis(x, y) }, adding SdisIs noted as { Rdis(x, y) }, where (x, y) here denotes the coordinate positions of pixel points in the left viewpoint image and the right viewpoint image, x is 1. ltoreq. W, y is 1. ltoreq. H, W denotes the widths of the left viewpoint image and the right viewpoint image, H denotes the heights of the left viewpoint image and the right viewpoint image, Lorg(x, y) represents { L }orgThe coordinate position in (x, y) } is the pixel value of the pixel point with (x, y), Rorg(x, y) represents { RorgThe pixel value L of the pixel point with the coordinate position (x, y) in (x, y) } isdis(x, y) represents { L }disThe coordinate position in (x, y) } is the pixel value of the pixel point with (x, y), Rdis(x, y) represents { RdisAnd the coordinate position in the (x, y) is the pixel value of the pixel point of (x, y).
② according to { Lorg(x,y)}、{Rorg(x,y)}、{Ldis(x,y)}、{RdisEven symmetric frequency response and odd symmetric frequency response of each pixel point in (x, y) } in different scales and directions are correspondingly obtained to obtain { L }org(x,y)}、{Rorg(x,y)}、{Ldis(x,y)}、{RdisThe amplitude of each pixel in (x, y) } is then based on { L }org(x, y) } and { R }orgAmplitude of each pixel in (x, y) } and { Lorg(x, y) } and { R }org(x, y) } calculating S from the pixel value of each pixel in the parallax imageorgThe single eye diagram (cyclopean map) is recorded as{CMorg(x, y) }, and according to { L }dis(x, y) } and { R }disAmplitude of each pixel in (x, y) } and { Lorg(x, y) } and { R }org(x, y) } calculating S from the pixel value of each pixel in the parallax imagedisThe eye-independent diagram of (2), is marked as { CMdis(x, y) }, wherein CMorg(x, y) denotes { CMorgThe pixel value of the pixel point with the coordinate position (x, y) in (x, y) }, CMdis(x, y) denotes { CMdisAnd the coordinate position in the (x, y) is the pixel value of the pixel point of (x, y).
In this embodiment, the specific process of step two is:
2-1, pair { Lorg(x, y) is filtered to obtain { L }orgEven symmetric frequency response and odd symmetric frequency response of each pixel point in (x, y) } in different scales and directions are converted into { L }orgEven symmetric frequency responses of pixel points with coordinate positions (x, y) in different scales and directions are recorded as eα,θ(x, y) will { LorgThe odd symmetric frequency response of the pixel point with the coordinate position (x, y) in different scales and directions is marked as oα,θ(x, y), wherein alpha represents the scale factor of the filter used for filtering, alpha is more than or equal to 1 and less than or equal to 4, theta represents the direction factor of the filter used for filtering, and theta is more than or equal to 1 and less than or equal to 4.
Here, for { Lorg(x, y) } the filter used for the filtering process is a log-Garbor filter.
2 according to { LorgCalculating the even symmetric frequency response and the odd symmetric frequency response of each pixel point in (x, y) in different scales and directions, and calculating the { L }orgThe amplitude of each pixel in (x, y) } will be { LorgThe amplitude of the pixel point with the coordinate position (x, y) in (x, y) is recorded as <math> <mrow> <msubsup> <mi>GE</mi> <mi>org</mi> <mi>L</mi> </msubsup> <mrow> <mo>(</mo> <mi>x</mi> <mo>,</mo> <mi>y</mi> <mo>)</mo> </mrow> <mo>=</mo> <munderover> <mi>&Sigma;</mi> <mrow> <mi>&theta;</mi> <mo>=</mo> <mn>1</mn> </mrow> <mn>4</mn> </munderover> <munderover> <mi>&Sigma;</mi> <mrow> <mi>&alpha;</mi> <mo>=</mo> <mn>1</mn> </mrow> <mn>4</mn> </munderover> <msqrt> <msub> <mi>e</mi> <mrow> <mi>&alpha;</mi> <mo>,</mo> <mi>&theta;</mi> </mrow> </msub> <msup> <mrow> <mo>(</mo> <mi>x</mi> <mo>,</mo> <mi>y</mi> <mo>)</mo> </mrow> <mn>2</mn> </msup> <mo>+</mo> <msub> <mi>o</mi> <mrow> <mi>&alpha;</mi> <mo>,</mo> <mi>&theta;</mi> </mrow> </msub> <msup> <mrow> <mo>(</mo> <mi>x</mi> <mo>,</mo> <mi>y</mi> <mo>)</mo> </mrow> <mn>2</mn> </msup> </msqrt> <mo>.</mo> </mrow> </math>
② -3, obtaining { L ] according to the steps from ② -1 to ② -2orgOperation of amplitude of each pixel in (x, y) } acquires { R } in the same mannerorg(x,y)}、{Ldis(x, y) } and { R }dis(x, y) } the amplitude of each pixel point in the (x, y) } will be { RorgThe amplitude of the pixel point with the coordinate position (x, y) in (x, y) is recorded asWill { LdisThe amplitude of the pixel point with the coordinate position (x, y) in (x, y) is recorded asWill { RdisThe amplitude of the pixel point with the coordinate position (x, y) in (x, y) is recorded asIf obtain { LdisThe operation process of the amplitude of each pixel point in (x, y) } is as follows: 1) for { Ldis(x, y) is filtered to obtain { L }disEven symmetric frequency response and odd symmetric frequency response of each pixel point in (x, y) } in different scales and directions are converted into { L }disEven symmetric frequency responses of pixel points with coordinate positions (x, y) in different scales and directions are recorded as eα,θ' (x, y), will { LdisThe odd symmetric frequency response of the pixel point with the coordinate position (x, y) in different scales and directions is marked as oα,θ' (x, y), wherein alpha represents the scale factor of the filter used for filtering, 1 is less than or equal to alpha and less than or equal to 4, theta represents the direction factor of the filter used for filtering, and 1 is less than or equal to theta and less than or equal to 4; 2) according to { LdisCalculating the even symmetric frequency response and the odd symmetric frequency response of each pixel point in (x, y) in different scales and directions, and calculating the { L }disThe amplitude of each pixel in (x, y) } will be { LdisThe amplitude of the pixel point with the coordinate position (x, y) in (x, y) is recorded as <math> <mrow> <msubsup> <mi>GE</mi> <mi>dis</mi> <mi>L</mi> </msubsup> <mrow> <mo>(</mo> <mi>x</mi> <mo>,</mo> <mi>y</mi> <mo>)</mo> </mrow> <mo>=</mo> <munderover> <mi>&Sigma;</mi> <mrow> <mi>&theta;</mi> <mo>=</mo> <mn>1</mn> </mrow> <mn>4</mn> </munderover> <munderover> <mi>&Sigma;</mi> <mrow> <mi>&alpha;</mi> <mo>=</mo> <mn>1</mn> </mrow> <mn>4</mn> </munderover> <msqrt> <msup> <msub> <mi>e</mi> <mrow> <mi>&alpha;</mi> <mo>,</mo> <mi>&theta;</mi> </mrow> </msub> <mo>&prime;</mo> </msup> <msup> <mrow> <mo>(</mo> <mi>x</mi> <mo>,</mo> <mi>y</mi> <mo>)</mo> </mrow> <mn>2</mn> </msup> <mo>+</mo> <msup> <msub> <mi>o</mi> <mrow> <mi>&alpha;</mi> <mo>,</mo> <mi>&theta;</mi> </mrow> </msub> <mo>&prime;</mo> </msup> <msup> <mrow> <mo>(</mo> <mi>x</mi> <mo>,</mo> <mi>y</mi> <mo>)</mo> </mrow> <mn>2</mn> </msup> </msqrt> <mo>.</mo> </mrow> </math>
② 4, calculating { L by adopting a block matching methodorg(x, y) } and { R }org(x, y) } parallax images, noted asWherein,to representThe middle coordinate position is the pixel value of the pixel point of (x, y).
② 5 according to { Lorg(x, y) } and { R }orgAmplitude of each pixel in (x, y) } andcalculating the pixel value of each pixel point in SorgThe eye-independent diagram of (2), is marked as { CMorg(x, y) }, will { CMorgThe pixel value of the pixel point with the coordinate position (x, y) in (x, y) is marked as CMorg(x,y), <math> <mrow> <msub> <mi>CM</mi> <mi>org</mi> </msub> <mrow> <mo>(</mo> <mi>x</mi> <mo>,</mo> <mi>y</mi> <mo>)</mo> </mrow> <mo>=</mo> <mfrac> <mrow> <msubsup> <mi>GE</mi> <mi>org</mi> <mi>L</mi> </msubsup> <mrow> <mo>(</mo> <mi>x</mi> <mo>,</mo> <mi>y</mi> <mo>)</mo> </mrow> <mo>&times;</mo> <msub> <mi>L</mi> <mi>org</mi> </msub> <mrow> <mo>(</mo> <mi>x</mi> <mo>,</mo> <mi>y</mi> <mo>)</mo> </mrow> <mo>+</mo> <msubsup> <mi>GE</mi> <mi>org</mi> <mi>R</mi> </msubsup> <mrow> <mo>(</mo> <mi>x</mi> <mo>-</mo> <msubsup> <mi>d</mi> <mi>org</mi> <mi>L</mi> </msubsup> <mrow> <mo>(</mo> <mi>x</mi> <mo>,</mo> <mi>y</mi> <mo>)</mo> </mrow> <mo>,</mo> <mi>y</mi> <mo>)</mo> </mrow> <mo>&times;</mo> <msub> <mi>R</mi> <mi>org</mi> </msub> <mrow> <mo>(</mo> <mi>x</mi> <mo>-</mo> <msubsup> <mi>d</mi> <mi>org</mi> <mi>L</mi> </msubsup> <mrow> <mo>(</mo> <mi>x</mi> <mo>,</mo> <mi>y</mi> <mo>)</mo> </mrow> <mo>,</mo> <mi>y</mi> <mo>)</mo> </mrow> </mrow> <mrow> <msubsup> <mi>GE</mi> <mi>org</mi> <mi>L</mi> </msubsup> <mrow> <mo>(</mo> <mi>x</mi> <mo>,</mo> <mi>y</mi> <mo>)</mo> </mrow> <mo>+</mo> <msubsup> <mi>GE</mi> <mi>org</mi> <mi>R</mi> </msubsup> <mrow> <mo>(</mo> <mi>x</mi> <mo>-</mo> <msubsup> <mi>d</mi> <mi>org</mi> <mi>L</mi> </msubsup> <mrow> <mo>(</mo> <mi>x</mi> <mo>,</mo> <mi>y</mi> <mo>)</mo> </mrow> <mo>,</mo> <mi>y</mi> <mo>)</mo> </mrow> </mrow> </mfrac> <mo>,</mo> </mrow> </math> Wherein,represents { Rorg(x, y) } coordinate position ofThe amplitude of the pixel points of (a) is,represents { Rorg(x, y) } coordinate position ofThe pixel value of the pixel point of (1).
② 6 according to { Ldis(x, y) } and { R }disAmplitude of each pixel in (x, y) } andcalculating the pixel value of each pixel point in SdisThe eye-independent diagram of (2), is marked as { CMdis(x, y) }, will { CMdisThe pixel value of the pixel point with the coordinate position (x, y) in (x, y) is marked as CMdis(x,y), <math> <mrow> <msub> <mi>CM</mi> <mi>dis</mi> </msub> <mrow> <mo>(</mo> <mi>x</mi> <mo>,</mo> <mi>y</mi> <mo>)</mo> </mrow> <mo>=</mo> <mfrac> <mrow> <msubsup> <mi>GE</mi> <mi>dis</mi> <mi>L</mi> </msubsup> <mrow> <mo>(</mo> <mi>x</mi> <mo>,</mo> <mi>y</mi> <mo>)</mo> </mrow> <mo>&times;</mo> <msub> <mi>L</mi> <mi>dis</mi> </msub> <mrow> <mo>(</mo> <mi>x</mi> <mo>,</mo> <mi>y</mi> <mo>)</mo> </mrow> <mo>+</mo> <msubsup> <mi>GE</mi> <mi>dis</mi> <mi>R</mi> </msubsup> <mrow> <mo>(</mo> <mi>x</mi> <mo>-</mo> <msubsup> <mi>d</mi> <mi>org</mi> <mi>L</mi> </msubsup> <mrow> <mo>(</mo> <mi>x</mi> <mo>,</mo> <mi>y</mi> <mo>)</mo> </mrow> <mo>,</mo> <mi>y</mi> <mo>)</mo> </mrow> <mo>&times;</mo> <msub> <mi>R</mi> <mi>dis</mi> </msub> <mrow> <mo>(</mo> <mi>x</mi> <mo>-</mo> <msubsup> <mi>d</mi> <mi>org</mi> <mi>L</mi> </msubsup> <mrow> <mo>(</mo> <mi>x</mi> <mo>,</mo> <mi>y</mi> <mo>)</mo> </mrow> <mo>,</mo> <mi>y</mi> <mo>)</mo> </mrow> </mrow> <mrow> <msubsup> <mi>GE</mi> <mi>dis</mi> <mi>L</mi> </msubsup> <mrow> <mo>(</mo> <mi>x</mi> <mo>,</mo> <mi>y</mi> <mo>)</mo> </mrow> <mo>+</mo> <msubsup> <mi>GE</mi> <mi>dis</mi> <mi>R</mi> </msubsup> <mrow> <mo>(</mo> <mi>x</mi> <mo>-</mo> <msubsup> <mi>d</mi> <mi>org</mi> <mi>L</mi> </msubsup> <mrow> <mo>(</mo> <mi>x</mi> <mo>,</mo> <mi>y</mi> <mo>)</mo> </mrow> <mo>,</mo> <mi>y</mi> <mo>)</mo> </mrow> </mrow> </mfrac> <mo>,</mo> </mrow> </math> Wherein,represents { Rdis(x, y) } coordinate position ofThe amplitude of the pixel points of (a) is,represents { Rdis(x, y) } coordinate position ofThe pixel value of the pixel point of (1).
③ according to { CMorg(x, y) } and { CMdisCalculating the mean and standard deviation of the pixel values of each pixel in (x, y) } to calculate { CMdis(x, y) } the objective evaluation metric for each pixel point, will be { CMdisThe objective evaluation metric value of the pixel point with the coordinate position (x, y) in (x, y) is recorded as Qimage(x, y), will { CMdisThe objective evaluation metric values of all the pixel points in (x, y) are collectively expressed as { Q }image(x,y)}。
In this embodiment, the specific process of step three is:
③ 1, calculate { CMorg(x, y) } and { CMdisThe mean and standard deviation of the pixel values for each pixel in (x, y) } will be { CMorgThe coordinate position in (x, y) } is (x)1,y1) Respectively recording the mean value and the standard difference of the pixel values of the pixel points as muorg(x1,y1) And σorg(x1,y1) Will { CMdisThe coordinate position in (x, y) } is (x)1,y1) Respectively recording the mean value and the standard difference of the pixel values of the pixel points as mudis(x1,y1) And σdis(x1,y1),
<math> <mrow> <msub> <mi>&sigma;</mi> <mi>org</mi> </msub> <mrow> <mo>(</mo> <msub> <mi>x</mi> <mn>1</mn> </msub> <mo>,</mo> <msub> <mi>y</mi> <mn>1</mn> </msub> <mo>)</mo> </mrow> <mo>=</mo> <msqrt> <mfrac> <mrow> <munder> <mi>&Sigma;</mi> <mrow> <mrow> <mo>(</mo> <msub> <mi>x</mi> <mn>1</mn> </msub> <mo>,</mo> <msub> <mi>y</mi> <mn>1</mn> </msub> <mo>)</mo> </mrow> <mo>&Element;</mo> <mi>N</mi> <mrow> <mo>(</mo> <msub> <mi>x</mi> <mn>1</mn> </msub> <mo>,</mo> <msub> <mi>y</mi> <mn>1</mn> </msub> <mo>)</mo> </mrow> </mrow> </munder> <msup> <mrow> <mo>(</mo> <msub> <mi>CM</mi> <mi>org</mi> </msub> <mrow> <mo>(</mo> <msub> <mi>x</mi> <mn>1</mn> </msub> <mo>,</mo> <msub> <mi>y</mi> <mn>1</mn> </msub> <mo>)</mo> </mrow> <mo>-</mo> <msub> <mi>&mu;</mi> <mi>org</mi> </msub> <mrow> <mo>(</mo> <msub> <mi>x</mi> <mn>1</mn> </msub> <mo>,</mo> <msub> <mi>y</mi> <mn>1</mn> </msub> <mo>)</mo> </mrow> <mo>)</mo> </mrow> <mn>2</mn> </msup> </mrow> <mi>M</mi> </mfrac> </msqrt> <mo>,</mo> </mrow> </math>
<math> <mrow> <msub> <mi>&mu;</mi> <mi>dis</mi> </msub> <mrow> <mo>(</mo> <msub> <mi>x</mi> <mn>1</mn> </msub> <mo>,</mo> <msub> <mi>y</mi> <mn>1</mn> </msub> <mo>)</mo> </mrow> <mo>=</mo> <mfrac> <mrow> <munder> <mi>&Sigma;</mi> <mrow> <mrow> <mo>(</mo> <msub> <mi>x</mi> <mn>1</mn> </msub> <mo>,</mo> <msub> <mi>y</mi> <mn>1</mn> </msub> <mo>)</mo> </mrow> <mo>&Element;</mo> <mrow> <mo>(</mo> <msub> <mi>x</mi> <mn>1</mn> </msub> <mo>,</mo> <msub> <mi>y</mi> <mn>1</mn> </msub> <mo>)</mo> </mrow> </mrow> </munder> <msub> <mi>CM</mi> <mi>dis</mi> </msub> <mrow> <mo>(</mo> <msub> <mi>x</mi> <mn>1</mn> </msub> <mo>,</mo> <msub> <mi>y</mi> <mn>1</mn> </msub> <mo>)</mo> </mrow> </mrow> <mi>M</mi> </mfrac> <mo>,</mo> </mrow> </math>
<math> <mrow> <msub> <mi>&sigma;</mi> <mi>dis</mi> </msub> <mrow> <mo>(</mo> <msub> <mi>x</mi> <mn>1</mn> </msub> <mo>,</mo> <msub> <mi>y</mi> <mn>1</mn> </msub> <mo>)</mo> </mrow> <mo>=</mo> <msqrt> <mfrac> <mrow> <munder> <mi>&Sigma;</mi> <mrow> <mrow> <mo>(</mo> <msub> <mi>x</mi> <mn>1</mn> </msub> <mo>,</mo> <msub> <mi>y</mi> <mn>1</mn> </msub> <mo>)</mo> </mrow> <mo>&Element;</mo> <mi>N</mi> <mrow> <mo>(</mo> <msub> <mi>x</mi> <mn>1</mn> </msub> <mo>,</mo> <msub> <mi>y</mi> <mn>1</mn> </msub> <mo>)</mo> </mrow> </mrow> </munder> <msup> <mrow> <mo>(</mo> <msub> <mi>CM</mi> <mi>dis</mi> </msub> <mrow> <mo>(</mo> <msub> <mi>x</mi> <mn>1</mn> </msub> <mo>,</mo> <msub> <mi>y</mi> <mn>1</mn> </msub> <mo>)</mo> </mrow> <mo>-</mo> <msub> <mi>&mu;</mi> <mi>dis</mi> </msub> <mrow> <mo>(</mo> <msub> <mi>x</mi> <mn>1</mn> </msub> <mo>,</mo> <msub> <mi>y</mi> <mn>1</mn> </msub> <mo>)</mo> </mrow> <mo>)</mo> </mrow> <mn>2</mn> </msup> </mrow> <mi>M</mi> </mfrac> </msqrt> <mo>,</mo> </mrow> </math> Wherein x is more than or equal to 11≤W,1≤y1≤H,N(x1,y1) The coordinate position is shown as (x)1,y1) The pixel point of (a) is an 8 x 8 neighborhood window with the center, M represents N (x)1,y1) Number of inner pixels, CMorg(x1,y1) Representation { CMorgThe coordinate position in (x, y) } is (x)1,y1) Pixel value of the pixel point, CMdis(x1,y1) Representation { CMdisThe coordinate position in (x, y) } is (x)1,y1) The pixel value of the pixel point of (1).
③ 2 according to { CMorg(x, y) } and { CMdisCalculating the mean and standard deviation of the pixel values of each pixel in (x, y) } to calculate { CMdis(x, y) } the objective evaluation metric for each pixel point, will be { CMdisThe coordinate position in (x, y) } is (x)1,y1) The objective evaluation metric value of the pixel point is recorded as Qimage(x1,y1), <math> <mrow> <msub> <mi>Q</mi> <mi>image</mi> </msub> <mrow> <mo>(</mo> <msub> <mi>x</mi> <mn>1</mn> </msub> <mo>,</mo> <msub> <mi>y</mi> <mn>1</mn> </msub> <mo>)</mo> </mrow> <mo>=</mo> <mfrac> <mrow> <mn>4</mn> <mo>&times;</mo> <mrow> <mo>(</mo> <msub> <mi>&mu;</mi> <mi>org</mi> </msub> <mrow> <mo>(</mo> <msub> <mi>x</mi> <mn>1</mn> </msub> <mo>,</mo> <msub> <mi>y</mi> <mn>1</mn> </msub> <mo>)</mo> </mrow> <mo>&times;</mo> <msub> <mi>&mu;</mi> <mi>dis</mi> </msub> <mrow> <mo>(</mo> <msub> <mi>x</mi> <mn>1</mn> </msub> <mo>,</mo> <msub> <mi>y</mi> <mn>1</mn> </msub> <mo>)</mo> </mrow> <mo>)</mo> </mrow> <mo>&times;</mo> <mrow> <mo>(</mo> <msub> <mi>&sigma;</mi> <mi>org</mi> </msub> <mrow> <mo>(</mo> <msub> <mi>x</mi> <mn>1</mn> </msub> <mo>,</mo> <msub> <mi>y</mi> <mn>1</mn> </msub> <mo>)</mo> </mrow> <mo>&times;</mo> <msub> <mi>&sigma;</mi> <mi>dis</mi> </msub> <mrow> <mo>(</mo> <msub> <mi>x</mi> <mn>1</mn> </msub> <mo>,</mo> <msub> <mi>y</mi> <mn>1</mn> </msub> <mo>)</mo> </mrow> <mo>)</mo> </mrow> <mo>+</mo> <mi>C</mi> </mrow> <mrow> <mrow> <mo>(</mo> <msub> <mi>&mu;</mi> <mi>org</mi> </msub> <msup> <mrow> <mo>(</mo> <msub> <mi>x</mi> <mn>1</mn> </msub> <mo>,</mo> <msub> <mi>y</mi> <mn>1</mn> </msub> <mo>)</mo> </mrow> <mn>2</mn> </msup> <mo>+</mo> <msub> <mi>&mu;</mi> <mi>dis</mi> </msub> <msup> <mrow> <mo>(</mo> <msub> <mi>x</mi> <mn>1</mn> </msub> <mo>,</mo> <msub> <mi>y</mi> <mn>1</mn> </msub> <mo>)</mo> </mrow> <mn>2</mn> </msup> <mo>)</mo> </mrow> <mo>&times;</mo> <mrow> <mo>(</mo> <msub> <mi>&sigma;</mi> <mi>org</mi> </msub> <msup> <mrow> <mo>(</mo> <msub> <mi>x</mi> <mn>1</mn> </msub> <mo>,</mo> <msub> <mi>y</mi> <mn>1</mn> </msub> <mo>)</mo> </mrow> <mn>2</mn> </msup> <mo>+</mo> <msub> <mi>&sigma;</mi> <mi>dis</mi> </msub> <msup> <mrow> <mo>(</mo> <msub> <mi>x</mi> <mn>1</mn> </msub> <mo>,</mo> <msub> <mi>y</mi> <mn>1</mn> </msub> <mo>)</mo> </mrow> <mn>2</mn> </msup> <mo>)</mo> </mrow> <mo>+</mo> <mi>C</mi> </mrow> </mfrac> <mo>,</mo> </mrow> </math> Where C is a control parameter, in this embodiment, C is 0.01.
Fourthly, according to { CMorgSpectral redundancy characteristics of (x, y) }, i.e. according to { CMorg(x, y) } amplitude and phase, calculating { CMorg(x, y) } saliency map, denoted as { SM }org(x, y) }, according to { CMdisSpectral redundancy characteristics of (x, y) }, i.e. according to { CMdis(x, y) } amplitude and phase, calculating { CMdis(x,y)}Is marked as { SMdis(x, y) }, wherein, SMorg(x, y) denotes { SMorgThe pixel value of the pixel point whose coordinate position is (x, y), SMdis(x, y) denotes { SMdisAnd the coordinate position in the (x, y) is the pixel value of the pixel point of (x, y).
In this embodiment, the specific process of step iv is:
tetra-1, pair { CMorg(x, y) performing discrete Fourier transform to obtain (CM)orgThe amplitude and phase of (x, y) } are denoted as { M, respectivelyorg(u, v) } and { Aorg(u, v) }, where u denotes the width of the amplitude or phase of the transform domain, v denotes the height of the amplitude or phase of the transform domain, 1. ltoreq. u.ltoreq.W, 1. ltoreq. v.ltoreq.H, Morg(u, v) represents { M }org(u, v) } the amplitude value of the pixel point with the coordinate position of (u, v), Aorg(u, v) represents { A }orgAnd (u, v) the coordinate position in the (u, v) is the phase value of the pixel point of (u, v).
Fourthly-2, calculating { Morg(u, v) } amplitude of the high frequency component, denoted as { R }org(u, v) }, will { RorgThe amplitude value of the high-frequency component of the pixel point with the coordinate position (u, v) in (u, v) is recorded as Rorg(u,v),Rorg(u,v)=log(Morg(u,v))-hm(u,v)*log(Morg(u, v)), wherein log () is a logarithmic function based on e, e-2.718281828, "+" is the convolution operator, hm(u, v) represents an m × m average filter, and in the present embodiment, m is 3.
Fourthly-3, according to { Rorg(u, v) } and { Aorg(u, v) } inverse discrete Fourier transform, and the obtained inverse transform image is taken as { CMorg(x, y) } significant graph, denoted as { SMorg(x, y) }, wherein, SMorg(x, y) denotes { SMorgAnd the coordinate position in the (x, y) is the pixel value of the pixel point of (x, y).
Fourthly-4, obtaining the { CM according to the steps from the fourth step-1 to the fourth step-3orgOperation of the saliency map of (x, y) } acquisition of { CM in the same mannerdis(x, y) } significant graph, denoted as { SMdis(x, y) }, wherein, SMdis(x, y) denotes { SMdisAnd the coordinate position in the (x, y) is the pixel value of the pixel point of (x, y).
Fifthly, calculating { CMorg(x, y) } and { CMdisA distortion map (distortion map) between (x, y) } is recorded as { DM (x, y) }, a pixel value of a pixel point whose coordinate position is (x, y) in { DM (x, y) } is recorded as DM (x, y), and DM (x, y) } is recorded as (CM) in which the coordinate position is (x, y)org(x,y)-CMdis(x,y))2
According to { SMorg(x, y) } and { SMdis(x, y) } and { DM (x, y) }, for { CMdisThe objective evaluation metric values of all pixel points in (x, y) are fused to obtain SdisThe image quality objective evaluation predicted value of (1) is marked as Q, <math> <mrow> <mi>Q</mi> <mo>=</mo> <msup> <mrow> <mo>[</mo> <mfrac> <mrow> <munder> <mi>&Sigma;</mi> <mrow> <mrow> <mo>(</mo> <mi>x</mi> <mo>,</mo> <mi>y</mi> <mo>)</mo> </mrow> <mo>&Element;</mo> <mi>&Omega;</mi> </mrow> </munder> <msub> <mi>Q</mi> <mi>image</mi> </msub> <mrow> <mo>(</mo> <mi>x</mi> <mo>,</mo> <mi>y</mi> <mo>)</mo> </mrow> <mo>&times;</mo> <mi>SM</mi> <mrow> <mo>(</mo> <mi>x</mi> <mo>,</mo> <mi>y</mi> <mo>)</mo> </mrow> </mrow> <mrow> <munder> <mi>&Sigma;</mi> <mrow> <mrow> <mo>(</mo> <mi>x</mi> <mo>,</mo> <mi>y</mi> <mo>)</mo> </mrow> <mo>&Element;</mo> <mi>&Omega;</mi> </mrow> </munder> <mi>SM</mi> <mrow> <mo>(</mo> <mi>x</mi> <mo>,</mo> <mi>y</mi> <mo>)</mo> </mrow> </mrow> </mfrac> <mo>]</mo> </mrow> <mi>&gamma;</mi> </msup> <mo>&times;</mo> <msup> <mrow> <mo>[</mo> <mfrac> <mrow> <munder> <mi>&Sigma;</mi> <mrow> <mrow> <mo>(</mo> <mi>x</mi> <mo>,</mo> <mi>y</mi> <mo>)</mo> </mrow> <mo>&Element;</mo> <mi>&Omega;</mi> </mrow> </munder> <msub> <mi>Q</mi> <mi>image</mi> </msub> <mrow> <mo>(</mo> <mi>x</mi> <mo>,</mo> <mi>y</mi> <mo>)</mo> </mrow> <mo>&times;</mo> <mi>DM</mi> <mrow> <mo>(</mo> <mi>x</mi> <mo>,</mo> <mi>y</mi> <mo>)</mo> </mrow> </mrow> <mrow> <munder> <mi>&Sigma;</mi> <mrow> <mrow> <mo>(</mo> <mi>x</mi> <mo>,</mo> <mi>y</mi> <mo>)</mo> </mrow> <mo>&Element;</mo> <mi>&Omega;</mi> </mrow> </munder> <mi>DM</mi> <mrow> <mo>(</mo> <mi>x</mi> <mo>,</mo> <mi>y</mi> <mo>)</mo> </mrow> </mrow> </mfrac> <mo>]</mo> </mrow> <mi>&beta;</mi> </msup> <mo>,</mo> </mrow> </math> where Ω denotes a pixel domain range, and SM (x, y) is max (SM)org(x,y),SMdis(x, y)), max () is a function taking the maximum value, γ and β are weighting coefficients, and in the present embodiment, γ is 1.601, and β is 0.501.
And seventhly, establishing a distorted stereo image set under different distortion types and different distortion degrees by adopting n original undistorted stereo images, wherein the distorted stereo image set comprises a plurality of distorted stereo images, and respectively obtaining an average subjective score difference of each distorted stereo image in the distorted stereo image set by using a subjective quality evaluation method, and marking the average subjective score difference as DMOS (diffusion metal oxide semiconductor), wherein the DMOS is 100-MOS (metal oxide semiconductor), the MOS represents a subjective score mean value, and the DMOS belongs to [0,100], and n is more than or equal to 1.
In the present embodiment, a set of distorted stereoscopic images with different distortion degrees of different distortion types is established by using the stereoscopic images composed of fig. 2a and 2b, fig. 3a and 3b, fig. 4a and 4b, fig. 5a and 5b, fig. 6a and 6b, fig. 7a and 7b, fig. 8a and 8b, fig. 9a and 9b, fig. 10a and 10b, fig. 11a and 11b, fig. 12a and 12b, and fig. 13a and 13b, respectively, 12 (n is 12) undistorted stereoscopic images, which includes 252 distorted stereoscopic images of 4 distortion types in total, the total number of JPEG-compressed distorted stereoscopic images is 60, JPEG 2000-compressed distorted stereoscopic images is 60, Gaussian Blur (Gaussian Blur) distorted stereoscopic images is 60, and h.264-encoded distorted stereoscopic images is 72.
Calculating S according to the steps from the first step to the sixth stepdisThe image quality objective evaluation prediction value of each distorted stereo image in the distorted stereo image set is calculated respectively through the operation of the image quality objective evaluation prediction value.
The correlation between the objective evaluation prediction value of the image quality of the distorted stereo image obtained in the present embodiment and the average subjective score difference value is analyzed by using 252 distorted stereo images of 12 undistorted stereo images shown in fig. 2a to 13b under different degrees of JPEG compression, JPEG2000 compression, gaussian blur and h.264 coding distortion. Here, 4 common objective parameters of the evaluation method for evaluating image quality are used as evaluation indexes, that is, Pearson correlation coefficient (PLCC), Spearman correlation coefficient (SROCC), Kendall correlation coefficient (KROCC), mean square error (RMSE), stereo image evaluation objective model accuracy in which PLCC and RMSE reflect distortion, and SROCC and KROCC reflect monotonicity thereof under nonlinear regression conditions. The image quality objective evaluation predicted value of the distorted stereo image calculated according to the method is subjected to five-parameter Logistic function nonlinear fitting, and the higher the PLCC, SROCC and KROCC values are, the lower the RMSE value is, the better the correlation between the objective evaluation method and the average subjective score difference is. The Pearson correlation coefficient, the Spearman correlation coefficient, the Kendall correlation coefficient and the mean square error between the objective evaluation predicted value of the image quality and the subjective score of the distorted three-dimensional image obtained by respectively adopting the method and not adopting the method are compared, the comparison results are shown in tables 1, 2, 3 and 4, and as can be seen from tables 1, 2, 3 and 4, the correlation between the final objective evaluation predicted value of the image quality and the average subjective score difference value of the distorted three-dimensional image obtained by adopting the method is very high, which indicates that the objective evaluation result is more consistent with the result of human eye subjective perception, and the effectiveness of the method is enough to explain.
Fig. 14 shows a scatter diagram of the difference between the objective evaluation prediction value of the image quality of each distorted stereoscopic image in the distorted stereoscopic image set and the average subjective score, and the more concentrated the scatter is, the better the consistency between the objective evaluation result and the subjective perception is. As can be seen from fig. 14, the scatter diagram obtained by the method of the present invention is more concentrated, and the goodness of fit with the subjective evaluation data is higher.
TABLE 1 Pearson correlation coefficient comparison between objective evaluation prediction value and subjective score of image quality for distorted stereoscopic images obtained without and with the method of the present invention
TABLE 2 comparison of Spearman correlation coefficient between objective evaluation prediction value and subjective score of image quality for distorted stereo images obtained without and with the method of the present invention
TABLE 3 comparison of Kendall correlation coefficients between objective evaluation prediction values and subjective scores for image quality of distorted stereo images obtained without using the method of the present invention
TABLE 4 comparison of mean square error between objective evaluation prediction and subjective score of image quality for distorted stereoscopic images obtained with and without the method of the present invention

Claims (6)

1. A stereo image quality objective evaluation method based on feature fusion is characterized in that the processing process is as follows: firstly, obtaining a single eye diagram of an original undistorted stereo image according to even symmetric frequency response and odd symmetric frequency response of each pixel point in a left viewpoint image and a right viewpoint image of the original undistorted stereo image at different scales and directions and a parallax image between the left viewpoint image and the right viewpoint image of the original undistorted stereo image; obtaining a single eye diagram of the distorted stereo image to be evaluated according to even symmetric frequency response and odd symmetric frequency response of each pixel point in the left viewpoint image and the right viewpoint image of the distorted stereo image to be evaluated in different scales and directions and a parallax image between the left viewpoint image and the right viewpoint image of the original undistorted stereo image; secondly, obtaining an objective evaluation metric value of each pixel point in the single eye diagram of the distorted stereo image to be evaluated according to the mean value and the standard deviation of the pixel values of each pixel point in the two single eye diagrams; thirdly, obtaining a corresponding saliency map according to the amplitude and the phase of the single eye map of the original undistorted stereo image; obtaining a corresponding saliency map according to the amplitude and the phase of the single eye map of the distorted stereo image to be evaluated; then, according to the two saliency maps and the distortion map between the two single eye maps, fusing objective evaluation metric values of each pixel point in the single eye map of the distorted three-dimensional image to be evaluated to obtain an objective evaluation prediction value of the image quality of the distorted three-dimensional image to be evaluated; and finally, obtaining the image quality objective evaluation predicted value of the distorted three-dimensional images with different distortion types and different distortion degrees according to the processing process.
2. The method for objectively evaluating the quality of a stereoscopic image based on feature fusion according to claim 1, characterized in that it specifically includes the following steps:
making SorgFor original undistorted stereo image, let SdisFor the distorted stereo image to be evaluated, SorgIs noted as { Lorg(x, y) }, adding SorgIs noted as { Rorg(x, y) }, adding SdisIs noted as { Ldis(x, y) }, adding SdisIs noted as { Rdis(x, y) }, where (x, y) here denotes the coordinate positions of pixel points in the left viewpoint image and the right viewpoint image, x is 1. ltoreq. W, y is 1. ltoreq. H, W denotes the widths of the left viewpoint image and the right viewpoint image, H denotes the heights of the left viewpoint image and the right viewpoint image, Lorg(x, y) represents { L }orgThe coordinate position in (x, y) } is the pixel value of the pixel point with (x, y), Rorg(x, y) represents { Rorg(x, y) } middle coordinate positionSet as the pixel value, L, of the pixel point of (x, y)dis(x, y) represents { L }disThe coordinate position in (x, y) } is the pixel value of the pixel point with (x, y), Rdis(x, y) represents { RdisThe coordinate position in (x, y) is the pixel value of the pixel point of (x, y);
② according to { Lorg(x,y)}、{Rorg(x,y)}、{Ldis(x,y)}、{RdisEven symmetric frequency response and odd symmetric frequency response of each pixel point in (x, y) } in different scales and directions are correspondingly obtained to obtain { L }org(x,y)}、{Rorg(x,y)}、 {Ldis(x,y)}、{RdisThe amplitude of each pixel in (x, y) } is then based on { L }org(x, y) } and { R }orgAmplitude of each pixel in (x, y) } and { Lorg(x, y) } and { R }org(x, y) } calculating S from the pixel value of each pixel in the parallax imageorgThe eye-independent diagram of (2), is marked as { CMorg(x, y) }, and according to { L }dis(x, y) } and { R }disAmplitude of each pixel in (x, y) } and { Lorg(x, y) } and { R }org(x, y) } calculating S from the pixel value of each pixel in the parallax imagedisThe eye-independent diagram of (2), is marked as { CMdis(x, y) }, wherein CMorg(x, y) denotes { CMorgThe pixel value of the pixel point with the coordinate position (x, y) in (x, y) }, CMdis(x, y) denotes { CMdisThe coordinate position in (x, y) is the pixel value of the pixel point of (x, y);
③ according to { CMorg(x, y) } and { CMdisCalculating the mean and standard deviation of the pixel values of each pixel in (x, y) } to calculate { CMdis(x, y) } the objective evaluation metric for each pixel point, will be { CMdisThe objective evaluation metric value of the pixel point with the coordinate position (x, y) in (x, y) is recorded as Qimage(x,y);
Fourthly, according to { CMorg(x, y) } amplitude and phase, calculating { CMorg(x, y) } significant graph, denoted as { SMorg(x, y) }, according to { CMdis(x, y) } amplitude and phase, calculating { CMdis(x, y) } significant graph, denoted as { SMdis(x, y) }, wherein, SMorg(x, y) denotes { SMorg(x, y) } position of coordinate in (x, y) } spaceThe pixel value of the pixel point of (x, y), SMdis(x, y) denotes { SMdisThe coordinate position in (x, y) is the pixel value of the pixel point of (x, y);
fifthly, calculating { CMorg(x, y) } and { CMdisThe distortion map between (x, y) } is denoted as { DM (x, y) }, the pixel value of the pixel point with the coordinate position (x, y) in { DM (x, y) } is denoted as DM (x, y), and DM (x, y) ═ CMorg(x,y)-CMdis(x,y))2
According to { SMorg(x, y) } and { SMdis(x, y) } and { DM (x, y) }, for { CMdisThe objective evaluation metric values of all pixel points in (x, y) are fused to obtain SdisThe image quality objective evaluation predicted value of (1) is marked as Q,where Ω denotes a pixel domain range, and SM (x, y) is max (SM)org(x,y),SMdis(x, y)), max () is a function taking the maximum value, and γ and β are weight coefficients;
adopting n original undistorted stereo images, establishing a distorted stereo image set of the undistorted stereo images under different distortion types and different distortion degrees, wherein the distorted stereo image set comprises a plurality of distorted stereo images, respectively obtaining an average subjective score difference of each distorted stereo image in the distorted stereo image set by using a subjective quality evaluation method, and marking as DMOS (diffusion metal oxide semiconductor) which is 100-MOS (metal oxide semiconductor), wherein MOS represents a subjective score mean value, DMOS belongs to [0,100], and n is more than or equal to 1;
calculating S according to the steps from the first step to the sixth stepdisThe image quality objective evaluation prediction value of each distorted stereo image in the distorted stereo image set is calculated respectively through the operation of the image quality objective evaluation prediction value.
3. The objective evaluation method for stereo image quality based on feature fusion as claimed in claim 2, wherein the specific process of the step (II) is as follows:
2-1, pair { Lorg(x, y) is filtered to obtain { L }orgEach pixel point in (x, y) is at different scalesEven symmetric frequency response and odd symmetric frequency response of the sum direction will { L }orgEven symmetric frequency responses of pixel points with coordinate positions (x, y) in different scales and directions are recorded as eα,θ(x, y) will { LorgThe odd symmetric frequency response of the pixel point with the coordinate position (x, y) in different scales and directions is marked as oα,θ(x, y), wherein alpha represents the scale factor of the filter used for filtering, alpha is more than or equal to 1 and less than or equal to 4, theta represents the direction factor of the filter used for filtering, and theta is more than or equal to 1 and less than or equal to 4;
2 according to { LorgCalculating the even symmetric frequency response and the odd symmetric frequency response of each pixel point in (x, y) in different scales and directions, and calculating the { L }orgThe amplitude of each pixel in (x, y) } will be { LorgThe amplitude of the pixel point with the coordinate position (x, y) in (x, y) is recorded as
② -3, obtaining { L ] according to the steps from ② -1 to ② -2orgOperation of amplitude of each pixel in (x, y) } acquires { R } in the same mannerorg(x,y)}、{Ldis(x, y) } and { R }dis(x, y) } the amplitude of each pixel point in the (x, y) } will be { RorgThe amplitude of the pixel point with the coordinate position (x, y) in (x, y) is recorded asWill { LdisThe amplitude of the pixel point with the coordinate position (x, y) in (x, y) is recorded asWill { RdisThe amplitude of the pixel point with the coordinate position (x, y) in (x, y) is recorded as
② 4, adopting block matchingMethod calculation { Lorg(x, y) } and { R }org(x, y) } parallax images, noted asWherein,to representThe middle coordinate position is the pixel value of the pixel point of (x, y);
② 5 according to { Lorg(x, y) } and { R }orgAmplitude of each pixel in (x, y) } andcalculating the pixel value of each pixel point in SorgThe eye-independent diagram of (2), is marked as { CMorg(x, y) }, will { CMorgThe pixel value of the pixel point with the coordinate position (x, y) in (x, y) is marked as CMorg(x,y), Wherein,represents { Rorg(x, y) } coordinate position ofThe amplitude of the pixel points of (a) is,represents { Rorg(x, y) } coordinate position ofThe pixel value of the pixel point of (1);
② 6 according to { Ldis(x, y) } and { R }disEach pixel in (x, y) } and a pixel having a specific pixel areaAmplitude of the spot andcalculating the pixel value of each pixel point in SdisThe eye-independent diagram of (2), is marked as { CMdis(x, y) }, will { CMdisThe pixel value of the pixel point with the coordinate position (x, y) in (x, y) is marked as CMdis(x,y), Wherein,represents { Rdis(x, y) } coordinate position ofThe amplitude of the pixel points of (a) is,represents { Rdis(x, y) } coordinate position ofThe pixel value of the pixel point of (1).
4. The method according to claim 3, wherein the step (1) is performed by matching { L } -Lorg(x, y) } the filter used for the filtering process is a log-Garbor filter.
5. The objective evaluation method for stereo image quality based on feature fusion according to any one of claims 2 to 4, characterized in that the specific process of the third step is as follows:
③ 1, calculate { CMorg(x, y) } and { CMdisThe mean and standard deviation of the pixel values for each pixel in (x, y) } will be { CMorgThe coordinate position in (x, y) } is (x1,y1) Respectively recording the mean value and the standard difference of the pixel values of the pixel points as muorg(x1,y1) And σorg(x1,y1) Will { CMdisThe coordinate position in (x, y) } is (x)1,y1) Respectively recording the mean value and the standard difference of the pixel values of the pixel points as mudis(x1,y1) And σdis(x1,y1),
Wherein x is more than or equal to 11≤W,1≤y1≤H,N(x1,y1) The coordinate position is shown as (x)1,y1) The pixel point of (a) is an 8 x 8 neighborhood window with the center, M represents N (x)1,y1) Number of inner pixels, CMorg(x1,y1) Representation { CMorgThe coordinate position in (x, y) } is (x)1,y1) Pixel value of the pixel point, CMdis(x1,y1) Representation { CMdisThe coordinate position in (x, y) } is (x)1,y1) The pixel value of the pixel point of (1);
③ 2 according to { CMorg(x, y) } and { CMdisCalculating the mean and standard deviation of the pixel values of each pixel in (x, y) } to calculate { CMdis(x, y) } the objective evaluation metric for each pixel point, will be { CMdisThe coordinate position in (x, y) } is (x)1,y1) The objective evaluation metric value of the pixel point is recorded as Qimage(x1,y1), Wherein C is a control parameter.
6. The objective evaluation method for the quality of the stereo image based on the feature fusion as claimed in claim 5, wherein the specific process of the step (iv) is as follows:
tetra-1, pair { CMorg(x, y) performing discrete Fourier transform to obtain (CM)orgThe amplitude and phase of (x, y) } are denoted as { M, respectivelyorg(u, v) } and { Aorg(u, v) }, where u denotes the width of the amplitude or phase of the transform domain, v denotes the height of the amplitude or phase of the transform domain, 1. ltoreq. u.ltoreq.W, 1. ltoreq. v.ltoreq.H, Morg(u, v) represents { M }org(u, v) } the amplitude value of the pixel point with the coordinate position of (u, v), Aorg(u, v) represents { A }org(u, v) } the coordinate position is the phase value of the pixel point with (u, v);
fourthly-2, calculating { Morg(u, v) } amplitude of the high frequency component, denoted as { R }org(u, v) }, will { RorgThe amplitude value of the high-frequency component of the pixel point with the coordinate position (u, v) in (u, v) is recorded as Rorg(u,v), Rorg(u,v)=log(Morg(u,v))-hm(u,v)*log(Morg(u, v)), wherein log () is a logarithmic function based on e, e-2.718281828, "+" is the convolution operator, hm(u, v) represents an m × m mean filtering;
fourthly-3, according to { Rorg(u, v) } and { Aorg(u, v) } inverse discrete Fourier transform, and the obtained inverse transform image is taken as { CMorg(x, y) } significant graph, denoted as { SMorg(x, y) }, wherein, SMorg(x, y) denotes { SMorgThe coordinate position in (x, y) is the pixel value of the pixel point of (x, y);
fourthly-4, obtaining the { CM according to the steps from the fourth step-1 to the fourth step-3orgOperation of the saliency map of (x, y) } acquisition of { CM in the same mannerdis(x, y) } significant graph, denoted as { SMdis(x, y) }, wherein, SMdis(x, y) denotes { SMdis(x, y) of pixel points with (x, y) as coordinate positionsThe pixel value.
CN201210357956.8A 2012-09-24 2012-09-24 Three-dimensional picture quality objective evaluation method based on feature fusion Expired - Fee Related CN102903107B (en)

Priority Applications (1)

Application Number Priority Date Filing Date Title
CN201210357956.8A CN102903107B (en) 2012-09-24 2012-09-24 Three-dimensional picture quality objective evaluation method based on feature fusion

Applications Claiming Priority (1)

Application Number Priority Date Filing Date Title
CN201210357956.8A CN102903107B (en) 2012-09-24 2012-09-24 Three-dimensional picture quality objective evaluation method based on feature fusion

Publications (2)

Publication Number Publication Date
CN102903107A CN102903107A (en) 2013-01-30
CN102903107B true CN102903107B (en) 2015-07-08

Family

ID=47575320

Family Applications (1)

Application Number Title Priority Date Filing Date
CN201210357956.8A Expired - Fee Related CN102903107B (en) 2012-09-24 2012-09-24 Three-dimensional picture quality objective evaluation method based on feature fusion

Country Status (1)

Country Link
CN (1) CN102903107B (en)

Families Citing this family (7)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN103200420B (en) * 2013-03-19 2015-03-25 宁波大学 Three-dimensional picture quality objective evaluation method based on three-dimensional visual attention
CN103281556B (en) * 2013-05-13 2015-05-13 宁波大学 Objective evaluation method for stereo image quality on the basis of image decomposition
CN103369348B (en) * 2013-06-27 2015-03-25 宁波大学 Three-dimensional image quality objective evaluation method based on regional importance classification
CN106960432B (en) * 2017-02-08 2019-10-25 宁波大学 A kind of no reference stereo image quality evaluation method
CN107945151B (en) * 2017-10-26 2020-01-21 宁波大学 Repositioning image quality evaluation method based on similarity transformation
CN108694705B (en) * 2018-07-05 2020-12-11 浙江大学 Multi-frame image registration and fusion denoising method
CN109903273B (en) * 2019-01-30 2023-03-17 武汉科技大学 Stereo image quality objective evaluation method based on DCT domain characteristics

Citations (3)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN101378519A (en) * 2008-09-28 2009-03-04 宁波大学 Method for evaluating quality-lose referrence image quality base on Contourlet transformation
CN101610425A (en) * 2009-07-29 2009-12-23 清华大学 A kind of method and apparatus of evaluating stereo image quality
CN102170581A (en) * 2011-05-05 2011-08-31 天津大学 Human-visual-system (HVS)-based structural similarity (SSIM) and characteristic matching three-dimensional image quality evaluation method

Family Cites Families (1)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
JP4817246B2 (en) * 2006-07-31 2011-11-16 Kddi株式会社 Objective video quality evaluation system

Patent Citations (3)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN101378519A (en) * 2008-09-28 2009-03-04 宁波大学 Method for evaluating quality-lose referrence image quality base on Contourlet transformation
CN101610425A (en) * 2009-07-29 2009-12-23 清华大学 A kind of method and apparatus of evaluating stereo image quality
CN102170581A (en) * 2011-05-05 2011-08-31 天津大学 Human-visual-system (HVS)-based structural similarity (SSIM) and characteristic matching three-dimensional image quality evaluation method

Non-Patent Citations (2)

* Cited by examiner, † Cited by third party
Title
Quality Assessment of Stereoscopic Images;Alexandre Benoit,et al;《Quality Assessment of Stereoscopic Images》;20081014;1-13 *
基于小波图像融合的非对称失真;周武杰等;《光电工程》;20111130;第38卷(第11期);100-105 *

Also Published As

Publication number Publication date
CN102903107A (en) 2013-01-30

Similar Documents

Publication Publication Date Title
CN102903107B (en) Three-dimensional picture quality objective evaluation method based on feature fusion
CN103581661B (en) Method for evaluating visual comfort degree of three-dimensional image
CN104036501B (en) A kind of objective evaluation method for quality of stereo images based on rarefaction representation
CN103413298B (en) A kind of objective evaluation method for quality of stereo images of view-based access control model characteristic
CN102843572B (en) Phase-based stereo image quality objective evaluation method
CN102708567B (en) Visual perception-based three-dimensional image quality objective evaluation method
CN103517065B (en) Method for objectively evaluating quality of degraded reference three-dimensional picture
CN103136748B (en) The objective evaluation method for quality of stereo images of a kind of feature based figure
CN105282543B (en) Total blindness three-dimensional image quality objective evaluation method based on three-dimensional visual perception
CN104658001A (en) Non-reference asymmetric distorted stereo image objective quality assessment method
CN105407349A (en) No-reference objective three-dimensional image quality evaluation method based on binocular visual perception
CN105357519B (en) Quality objective evaluation method for three-dimensional image without reference based on self-similarity characteristic
CN104581143A (en) Reference-free three-dimensional picture quality objective evaluation method based on machine learning
CN103200420B (en) Three-dimensional picture quality objective evaluation method based on three-dimensional visual attention
CN103354617B (en) Boundary strength compressing image quality objective evaluation method based on DCT domain
Geng et al. A stereoscopic image quality assessment model based on independent component analysis and binocular fusion property
CN104954778A (en) Objective stereo image quality assessment method based on perception feature set
CN104036502A (en) No-reference fuzzy distorted stereo image quality evaluation method
CN102999912B (en) A kind of objective evaluation method for quality of stereo images based on distortion map
CN103369348B (en) Three-dimensional image quality objective evaluation method based on regional importance classification
CN103841411B (en) A kind of stereo image quality evaluation method based on binocular information processing
CN102999911B (en) Three-dimensional image quality objective evaluation method based on energy diagrams
CN102708568B (en) Stereoscopic image objective quality evaluation method on basis of structural distortion
CN104361583A (en) Objective quality evaluation method of asymmetrically distorted stereo images
CN103745457A (en) Stereo image objective quality evaluation method

Legal Events

Date Code Title Description
C06 Publication
PB01 Publication
C10 Entry into substantive examination
SE01 Entry into force of request for substantive examination
C14 Grant of patent or utility model
GR01 Patent grant
TR01 Transfer of patent right
TR01 Transfer of patent right

Effective date of registration: 20191218

Address after: Room 1,020, Nanxun Science and Technology Pioneering Park, No. 666 Chaoyang Road, Nanxun District, Huzhou City, Zhejiang Province, 313000

Patentee after: Huzhou You Yan Intellectual Property Service Co.,Ltd.

Address before: 315211 Zhejiang Province, Ningbo Jiangbei District Fenghua Road No. 818

Patentee before: Ningbo University

TR01 Transfer of patent right
TR01 Transfer of patent right

Effective date of registration: 20201229

Address after: 213001 3rd floor, Jinhu innovation center, No.8 Taihu Middle Road, Xinbei District, Changzhou City, Jiangsu Province

Patentee after: Jiangsu Qizhen Information Technology Service Co.,Ltd.

Address before: 313000 room 1020, science and Technology Pioneer Park, 666 Chaoyang Road, Nanxun Town, Nanxun District, Huzhou, Zhejiang.

Patentee before: Huzhou You Yan Intellectual Property Service Co.,Ltd.

CF01 Termination of patent right due to non-payment of annual fee
CF01 Termination of patent right due to non-payment of annual fee

Granted publication date: 20150708