CN104954778A - Objective stereo image quality assessment method based on perception feature set - Google Patents

Objective stereo image quality assessment method based on perception feature set Download PDF

Info

Publication number
CN104954778A
CN104954778A CN201510303868.3A CN201510303868A CN104954778A CN 104954778 A CN104954778 A CN 104954778A CN 201510303868 A CN201510303868 A CN 201510303868A CN 104954778 A CN104954778 A CN 104954778A
Authority
CN
China
Prior art keywords
dis
org
designated
pixel
coordinate position
Prior art date
Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
Granted
Application number
CN201510303868.3A
Other languages
Chinese (zh)
Other versions
CN104954778B (en
Inventor
郁梅
吕亚奇
彭宗举
陈芬
何美伶
刘姗姗
Current Assignee (The listed assignees may be inaccurate. Google has not performed a legal analysis and makes no representation or warranty as to the accuracy of the list.)
Ningbo University
Original Assignee
Ningbo University
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Application filed by Ningbo University filed Critical Ningbo University
Priority to CN201510303868.3A priority Critical patent/CN104954778B/en
Publication of CN104954778A publication Critical patent/CN104954778A/en
Application granted granted Critical
Publication of CN104954778B publication Critical patent/CN104954778B/en
Active legal-status Critical Current
Anticipated expiration legal-status Critical

Links

Landscapes

  • Image Analysis (AREA)

Abstract

The invention discloses an objective stereo image quality assessment method based on a perception feature set. According to the method, the distortion degree of a stereo image is measured according to the distortion degree of a visual perception feature map, a saliency map, a gradient map and an airspace just noticeable distortion map which are related to viewpoint perception quality are extracted, a disparity map related to third dimension and quality is extracted, distortion degrees of four perception feature maps are taken as features of the stereo image to create the perception feature set, a complicated human vision system is simulated according to a random forest machine learning algorithm, and feature parameters are fused. The objective stereo image quality assessment method has the advantages that change conditions of visual quality under the condition that the stereo image is influenced by various image processing and compressing methods can be reflected objectively, assessment performance of the method is not influenced by the content of the stereo image and the type of distortion, and conformity with subjective perception of human eyes is achieved.

Description

A kind of objective evaluation method for quality of stereo images based on perception feature set
Technical field
The present invention relates to a kind of image quality evaluating method, especially relate to a kind of objective evaluation method for quality of stereo images based on perception feature set.
Background technology
Stereo image quality evaluates the important component part that (Stereo Image Quality Assessment, SIQA) is three-dimensional video-frequency technology.Stereoscopic image/video is in collection, storage, process and inevitably can introduce noise or distortion in transmitting, and causes the decline of stereoscopic image/video quality, and therefore, the evaluation of stereoscopic image/video quality is the major issue of a needs research and solution.Stereo image quality evaluation method is generally divided into subjectivity and objectivity two class evaluation method.There is time-consuming, effort, shortcoming that cost is high because of it in subjective evaluation method, not easily realizes and apply; And method for objectively evaluating is integrated algorithm, model, can obtain the objective quality of stereo-picture quickly and easily, without the need to manual intervention, but its evaluation method is not yet ripe, and await further investigation, therefore method for objectively evaluating has become the emphasis of research.
In recent years, the research of stereo image quality objective evaluation achieves a series of achievement, two classes can be divided into generally: the first kind: the quality evaluation evaluation method of plane picture being directly applied to stereo-picture, if the people such as You are by the plane picture quality evaluating method of classics, as PSNR, MS-SSIM, VIF etc. are directly used in the evaluation of left visual point image and right visual point image, correspondence obtains the mass value of left visual point image and the mass value of right visual point image, get the quality of mean value as stereo-picture of two mass values again, but the quality evaluation of stereo-picture and plane picture makes a big difference, depth perception distortion level is also the key factor of the perceived quality affecting stereo-picture.Equations of The Second Kind: add the evaluation model that parallax information improves stereo-picture on the basis of plane picture quality evaluating method, as the people such as Yang add the antipode figure of stereo-picture to carry out stereo image quality evaluation; And for example the people such as Benoit is by depth information and plane picture quality evaluating method combining assessment stereo image quality; The people such as Hachicha propose the stereo image quality evaluation method based on the proper discernable distortion of binocular, and weigh relief change with binocular fusion competition, but, the third dimension of stereo-picture is the result of binocular fusion, binocular competition, binocular suppression, its mechanism is very complicated, therefore relief measurement is very difficult, and this also causes the consistency of these methods and subjective perception lower.These stereo image quality evaluation methods are all obtain global quality by local quality above, or the objective quality of stereo-picture is obtained with linear or nonlinear Feature fusion, but human visual system is extremely complicated system, these evaluation methods all cannot the human visual system of simulate complexity, causes evaluating accuracy lower.
Summary of the invention
Technical problem to be solved by this invention is to provide a kind of objective evaluation method for quality of stereo images based on perception feature set, and it can improve the correlation between objective evaluation result and subjective perception effectively.
The present invention solves the problems of the technologies described above adopted technical scheme: a kind of objective evaluation method for quality of stereo images based on perception feature set, it is characterized in that comprising the following steps:
1. I is made orgrepresent original undistorted stereo-picture, make I disrepresent the stereo-picture of distortion to be evaluated, by I orgleft visual point image be designated as L org, by I orgright visual point image be designated as R org, by I disleft visual point image be designated as L dis, by I disright visual point image be designated as R dis;
2. adopt frequency modulation conspicuousness detection algorithm, obtain L org, R org, L disand R disrespective remarkable figure, correspondence is designated as with then calculate with mean square error, be designated as MSE sal L = 1 M × N Σ i = 1 M Σ j = 1 N ( S org L ( i , j ) - S dis L ( i , j ) ) 2 ; Equally, calculate with mean square error, be designated as MSE sal R = 1 M × N Σ i = 1 M Σ j = 1 N ( S org R ( i , j ) - S dis R ( i , j ) ) 2 ;
Wherein, M represents I organd I diswidth, N represents I organd I disheight, 1≤i≤M, 1≤j≤N, represent middle coordinate position is the pixel value of the pixel of (i, j), represent middle coordinate position is the pixel value of the pixel of (i, j), represent middle coordinate position is the pixel value of the pixel of (i, j), represent middle coordinate position is the pixel value of the pixel of (i, j);
3. adopt horizontal Sobel operator, obtain L org, R org, L disand R disrespective horizontal gradient figure, correspondence is designated as with and adopt vertical Sobel operator, obtain L org, R org, L disand R disrespective vertical gradient map, correspondence is designated as with then basis with obtain L orggradient map, be designated as will middle coordinate position is that the pixel value of the pixel of (i, j) is designated as G org L ( i , j ) = G h L , org ( i , j ) 2 + G v L , org ( i , j ) 2 ; And according to with obtain R orggradient map, be designated as will middle coordinate position is that the pixel value of the pixel of (i, j) is designated as G org R ( i , j ) = G h R , org ( i , j ) 2 + G v R , org ( i , j ) 2 ; According to with obtain L disgradient map, be designated as will middle coordinate position is that the pixel value of the pixel of (i, j) is designated as G dis L ( i , j ) = G h L , dis ( i , j ) 2 + G v L , dis ( i , j ) 2 ; According to with obtain R disgradient map, be designated as will middle coordinate position is that the pixel value of the pixel of (i, j) is designated as G dis R ( i , j ) = G h R , dis ( i , j ) 2 + G v R , dis ( i , j ) 2 ; Calculate again with mean square error, be designated as MSE gra L = 1 M × N Σ i = 1 M Σ j = 1 N ( S org L ( i , j ) - S dis L ( i , j ) ) 2 ; Equally, calculate with mean square error, be designated as MSE gra R = 1 M × N Σ i = 1 M Σ j = 1 N ( G org R ( i , j ) - G dis R ( i , j ) ) 2 ;
Wherein, represent middle coordinate position is the pixel value of the pixel of (i, j), represent middle coordinate position is the pixel value of the pixel of (i, j), represent middle coordinate position is the pixel value of the pixel of (i, j), represent middle coordinate position is the pixel value of the pixel of (i, j), represent middle coordinate position is the pixel value of the pixel of (i, j), represent middle coordinate position is the pixel value of the pixel of (i, j), represent middle coordinate position is the pixel value of the pixel of (i, j), represent middle coordinate position is the pixel value of the pixel of (i, j);
4. adopt the proper discernable distortion model in spatial domain, obtain L org, R org, L disand R disthe proper discernable distortion map in respective spatial domain, correspondence is designated as with then calculate with mean square error, be designated as MSE JND L = 1 M × N Σ i = 1 M Σ j = 1 N ( J org L ( i , j ) - J dis L ( i , j ) ) 2 ; Equally, calculate with mean square error, be designated as MSE JND R = 1 M × N Σ i = 1 M Σ j = 1 N ( J org R ( i , j ) - J dis R ( i , j ) ) 2 ;
Wherein, represent middle coordinate position is the pixel value of the pixel of (i, j), represent middle coordinate position is the pixel value of the pixel of (i, j), represent middle coordinate position is the pixel value of the pixel of (i, j), represent middle coordinate position is the pixel value of the pixel of (i, j);
5. adopt light stream matching method, obtain I orghorizontal parallax amplitude figure and I orgvertical parallax amplitude figure, correspondence is designated as with then basis with obtain I orgdisparity map, be designated as D org, by D orgmiddle coordinate position is that the pixel value of the pixel of (i, j) is designated as D org(i, j), equally, adopt light stream matching method, obtain I dishorizontal parallax amplitude figure and I disvertical parallax amplitude figure, correspondence is designated as with then basis with obtain I disdisparity map, be designated as D dis, by D dismiddle coordinate position is that the pixel value of the pixel of (i, j) is designated as D dis(i, j), calculate D afterwards organd D dismean square error, be designated as MSE dsp, MSE dsp = 1 M × N Σ i = 1 M Σ j = 1 N ( D org ( i , j ) - D dis ( i , j ) ) 2 ;
Wherein, represent middle coordinate position is the pixel value of the pixel of (i, j), represent middle coordinate position is the pixel value of the pixel of (i, j), represent middle coordinate position is the pixel value of the pixel of (i, j), represent middle coordinate position is the pixel value of the pixel of (i, j);
6. will with the sets definition that arranged in sequence is formed is I disperception feature set, be designated as P, P = { MSE sal L , MSE sal R , MSE gra L , MSE gra R , MSE JND L , MSE JND R , MSE dsp } ;
7. adopt n original undistorted stereo-picture, set up its distortion stereo-picture set under the different distortion level of different type of distortion, using this distortion stereo-picture set as training set, training set comprises several distortion stereo-pictures; The mean subjective suggestion of the every width distortion stereo-picture then utilizing subjective quality assessment method evaluation to go out in training set is divided, and the mean subjective suggestion of the jth width distortion stereo-picture in training set is divided and is designated as MOS j; Again according to step 1. to step process 6., obtain the perception feature set of the every width distortion stereo-picture in training set in an identical manner, the perception feature set of the jth width distortion stereo-picture in training set be designated as P j;
Wherein, n>=1,1≤j≤S, S represents total width number of the distortion stereo-picture comprised in training set, MOS j∈ [0,5];
8. random forest machine learning algorithm is adopted, the perception feature set of all distortion stereo-pictures in training set is trained, make through training the regression function value that obtains to divide with corresponding mean subjective suggestion between error minimum, structure obtains random forest training pattern;
9. according to constructing the random forest training pattern obtained, to I disperception feature set P test, prediction obtain I disevaluating objective quality predicted value, be designated as Q dis, Q dis=MOD (P), wherein, the function representation form that MOD () is random forest training pattern.
Compared with prior art, the invention has the advantages that:
1) the inventive method weighs the distortion level of stereo-picture with the distortion level of visually-perceptible characteristic pattern, extract the remarkable figure relevant to viewpoint perceived quality, the proper discernable distortion map of gradient map and spatial domain, extract the disparity map relevant to three-dimensional perceived quality, using the structural feature perception feature set of the distortion level of four kinds of Perception Features figure as stereo-picture, use the human visual system of random forest machine learning algorithm Simulation of Complex again, carry out the fusion of characteristic parameter, the inventive method can reflect that stereo-picture is subject to the situation of change of various image procossing and the lower visual quality of compression method impact objectively, and the assess performance of the inventive method is not subject to the impact of stereoscopic image content and type of distortion, consistent with the subjective perception of human eye.
2) the inventive method is trained by the perception feature set of random forest machine learning algorithm to distortion stereo-picture, structure obtains random forest training pattern, again according to constructing the random forest training pattern obtained, the perception feature set of distortion stereo-picture to be evaluated is tested, prediction obtains the evaluating objective quality predicted value of distortion stereo-picture to be evaluated, this characteristic parameter that makes is with the evaluating objective quality predicted value of the amalgamation mode predicted distortion stereo-picture of the best, avoid the complicated simulation process of correlation properties to human visual system and mechanism, and because the perception feature set of training and the perception feature set of test are separate, therefore test result depending on unduly training data can be avoided, thus the correlation that can effectively improve between objective evaluation result and subjective perception.
Accompanying drawing explanation
Fig. 1 be the inventive method totally realize block diagram;
Fig. 2 a is the left visual point image of original undistorted horse stereo-picture;
The left visual point image of distortion that Fig. 2 b obtains after JPEG compression for the image shown in Fig. 2 a;
Fig. 2 c is the remarkable figure of the image shown in Fig. 2 a;
Fig. 2 d is the remarkable figure of the image shown in Fig. 2 b;
Fig. 2 e is the gradient map of the image shown in Fig. 2 a;
Fig. 2 f is the gradient map of the image shown in Fig. 2 b;
The proper discernable distortion map in spatial domain that Fig. 2 g is the image shown in Fig. 2 a;
The proper discernable distortion map in spatial domain that Fig. 2 h is the image shown in Fig. 2 b;
Fig. 2 i is the disparity map of stereo-picture corresponding to the image shown in Fig. 2 a;
Fig. 2 j is the disparity map of stereo-picture corresponding to the image shown in Fig. 2 b;
Fig. 3 a is the left visual point image of Akko (being of a size of 640 × 480) stereo-picture;
Fig. 3 b is the left visual point image of Altmoabit (being of a size of 1024 × 768) stereo-picture;
Fig. 3 c is the left visual point image of Balloons (being of a size of 1024 × 768) stereo-picture;
Fig. 3 d is the left visual point image of Doorflower (being of a size of 1024 × 768) stereo-picture;
Fig. 3 e is the left visual point image of Kendo (being of a size of 1024 × 768) stereo-picture;
Fig. 3 f is the left visual point image of LeaveLaptop (being of a size of 1024 × 768) stereo-picture;
Fig. 3 g is the left visual point image of Lovebierd1 (being of a size of 1024 × 768) stereo-picture;
Fig. 3 h is the left visual point image of Newspaper (being of a size of 1024 × 768) stereo-picture;
Fig. 3 i is the left visual point image of Puppy (being of a size of 720 × 480) stereo-picture;
Fig. 3 j is the left visual point image of Soccer2 (being of a size of 720 × 480) stereo-picture;
Fig. 3 k is the left visual point image of Horse (being of a size of 480 × 270) stereo-picture;
Fig. 3 l is the left visual point image of Xmas (being of a size of 640 × 480) stereo-picture.
Embodiment
Below in conjunction with accompanying drawing embodiment, the present invention is described in further detail.
The sensitivity of distortion in human eye stereoscopic image is different, therefore the consistency of the evaluation method of global quality and subjective perception is obtained by local quality lower, and Perception Features figure is the shallow-layer reflection of stereo-picture in human nervous system, that stereo-picture is experienced the most intuitively in human eye, when the distortion of stereo-picture causes Perception Features figure distortion, human eye also can discover these distortions sensitively, therefore, the objective evaluation method for quality of stereo images that the present invention proposes extracts four kinds of Perception Features figure, comprise remarkable figure, gradient map, the proper discernable distortion map in spatial domain and disparity map, and formed perception feature set using its distortion level as characteristic parameter, weigh the distortion level of stereo-picture.The method of the objective evaluation method for quality of stereo images machine learning that the present invention proposes carries out Fusion Features, obtains the highest feature high dimensional nonlinear Fusion Model of accuracy, thus reach higher forecasting accuracy by training random forest machine learning algorithm.
A kind of objective evaluation method for quality of stereo images based on perception feature set that the present invention proposes, it totally realizes block diagram as shown in Figure 1, and it comprises the following steps:
1. I is made orgrepresent original undistorted stereo-picture, make I disrepresent the stereo-picture of distortion to be evaluated, by I orgleft visual point image be designated as L org, by I orgright visual point image be designated as R org, by I disleft visual point image be designated as L dis, by I disright visual point image be designated as R dis.
Fig. 2 a gives the left visual point image of original undistorted horse stereo-picture, and Fig. 2 b gives the left visual point image of distortion that the image shown in Fig. 2 a obtains after JPEG compression.
2. adopt existing frequency modulation conspicuousness detection algorithm, obtain L org, R org, L disand R disrespective remarkable figure, correspondence is designated as with then calculate with mean square error, be designated as MSE sal L = 1 M × N Σ i = 1 M Σ j = 1 N ( S org L ( i , j ) - S dis L ( i , j ) ) 2 ; Equally, calculate with mean square error, be designated as MSE sal R = 1 M × N Σ i = 1 M Σ j = 1 N ( S org R ( i , j ) - S dis R ( i , j ) ) 2 .
Wherein, M represents I organd I diswidth, N represents I organd I disheight, 1≤i≤M, 1≤j≤N, represent middle coordinate position is the pixel value of the pixel of (i, j), represent middle coordinate position is the pixel value of the pixel of (i, j), represent middle coordinate position is the pixel value of the pixel of (i, j), represent middle coordinate position is the pixel value of the pixel of (i, j).
Remarkable figure, Fig. 2 d that Fig. 2 c gives the image shown in Fig. 2 a gives the remarkable figure of the image shown in Fig. 2 b.
3. adopt existing horizontal Sobel operator, obtain L org, R org, L disand R disrespective horizontal gradient figure, correspondence is designated as with and adopt existing vertical Sobel operator, obtain L org, R org, L disand R disrespective vertical gradient map, correspondence is designated as with then basis with obtain L orggradient map, be designated as will middle coordinate position is that the pixel value of the pixel of (i, j) is designated as G org L ( i , j ) = G h L , org ( i , j ) 2 + G v L , org ( i , j ) 2 ; And according to with obtain R orggradient map, be designated as will middle coordinate position is that the pixel value of the pixel of (i, j) is designated as G org R ( i , j ) = G h R , org ( i , j ) 2 + G v R , org ( i , j ) 2 ; According to with obtain L disgradient map, be designated as will middle coordinate position is that the pixel value of the pixel of (i, j) is designated as G dis L ( i , j ) = G h L , dis ( i , j ) 2 + G v L , dis ( i , j ) 2 ; According to with obtain R disgradient map, be designated as will middle coordinate position is that the pixel value of the pixel of (i, j) is designated as G dis R ( i , j ) = G h R , dis ( i , j ) 2 + G v R , dis ( i , j ) 2 ; Calculate again with mean square error, be designated as MSE gra L = 1 M × N Σ i = 1 M Σ j = 1 N ( S org L ( i , j ) - S dis L ( i , j ) ) 2 ; Equally, calculate with mean square error, be designated as MSE gra R = 1 M × N Σ i = 1 M Σ j = 1 N ( G org R ( i , j ) - G dis R ( i , j ) ) 2 .
Wherein, represent middle coordinate position is the pixel value of the pixel of (i, j), represent middle coordinate position is the pixel value of the pixel of (i, j), represent middle coordinate position is the pixel value of the pixel of (i, j), represent middle coordinate position is the pixel value of the pixel of (i, j), represent middle coordinate position is the pixel value of the pixel of (i, j), represent middle coordinate position is the pixel value of the pixel of (i, j), represent middle coordinate position is the pixel value of the pixel of (i, j), represent middle coordinate position is the pixel value of the pixel of (i, j).
Fig. 2 e gives the gradient map of the image shown in Fig. 2 a, and Fig. 2 f gives the gradient map of the image shown in Fig. 2 b.
4. adopt proper discernable distortion (Just-Noticeable-Distortion, the JND) model in existing spatial domain, obtain L org, R org, L disand R disthe proper discernable distortion map in respective spatial domain, correspondence is designated as with then calculate with mean square error, be designated as MSE JND L = 1 M × N Σ i = 1 M Σ j = 1 N ( J org L ( i , j ) - J dis L ( i , j ) ) 2 ; Equally, calculate with mean square error, be designated as MSE JND R = 1 M × N Σ i = 1 M Σ j = 1 N ( J org R ( i , j ) - J dis R ( i , j ) ) 2 .
Wherein, represent middle coordinate position is the pixel value of the pixel of (i, j), represent middle coordinate position is the pixel value of the pixel of (i, j), represent middle coordinate position is the pixel value of the pixel of (i, j), represent middle coordinate position is the pixel value of the pixel of (i, j).
At this, by L org, R org, L disand R disrespectively as pending image, then the detailed process obtaining the proper discernable distortion map in spatial domain of pending image is:
4.-1, pending image is designated as I;
-2 4., obtain the proper discernable distortion map of brightness of I, be designated as JND l, by JND lmiddle coordinate position is that the pixel value of the pixel of (i, j) is designated as JND l(i, j), wherein, represent that in I, coordinate position is the background value of the pixel of (i, j);
-3 4., obtain the proper discernable distortion map of texture of I, be designated as JND t, by JND tmiddle coordinate position is that the pixel value of the pixel of (i, j) is designated as JND t(i, j), JND t(i, j)=η G (i, j) W e(i, j), wherein, η is regulatory factor, gets η=0.01 at this, and G (i, j) represents that in I, coordinate position is the maximum gradient mean value of pixel under the high-pass filtering operator of different directions of (i, j), grad k(i, j) represents that in I, coordinate position is the gradient mean value of pixel under the high-pass filtering operator in a kth direction of (i, j), and four direction is horizontal direction, vertical direction and two diagonals respectively, W e(i, j) represents that in I, coordinate position is the Weighted Edges factor of the pixel of (i, j), W e ( i , j ) = 0.0001 I ‾ ( i , j ) + 0.115 ;
4.-4, according to JND land JND t, obtain the proper discernable distortion map in spatial domain of I, be designated as JND s, by JND smiddle coordinate position is that the pixel value of the pixel of (i, j) is designated as JND s(i, j), JND s(i, j)=JND l(i, j)+JND t(i, j)-0.3 × min{JND l(i, j), JND t(i, j) }, wherein, min () is for getting minimum value function.
Fig. 2 g gives the proper discernable distortion map in spatial domain of the image shown in Fig. 2 a, and Fig. 2 h gives the proper discernable distortion map in spatial domain of the image shown in Fig. 2 b.
5. adopt existing light stream matching method, obtain I orghorizontal parallax amplitude figure and I orgvertical parallax amplitude figure, correspondence is designated as with then basis with obtain I orgdisparity map, be designated as D org, by D orgmiddle coordinate position is that the pixel value of the pixel of (i, j) is designated as D org(i, j), equally, adopt existing light stream matching method, obtain I dishorizontal parallax amplitude figure and I disvertical parallax amplitude figure, correspondence is designated as with then basis with obtain I disdisparity map, be designated as D dis, by D dismiddle coordinate position is that the pixel value of the pixel of (i, j) is designated as D dis(i, j), calculate D afterwards organd D dismean square error, be designated as MSE dsp, MSE dsp = 1 M × N Σ i = 1 M Σ j = 1 N ( D org ( i , j ) - D dis ( i , j ) ) 2 .
Wherein, represent middle coordinate position is the pixel value of the pixel of (i, j), represent middle coordinate position is the pixel value of the pixel of (i, j), represent middle coordinate position is the pixel value of the pixel of (i, j), represent middle coordinate position is the pixel value of the pixel of (i, j).
Fig. 2 i gives the disparity map of stereo-picture corresponding to image shown in Fig. 2 a, and Fig. 2 j gives the disparity map of stereo-picture corresponding to image shown in Fig. 2 b.
6. will with the sets definition that arranged in sequence is formed is I disperception feature set, be designated as P, P = { MSE sal L , MSE sal R , MSE gra L , MSE gra R , MSE JND L , MSE JND R , MSE dsp } .
7. adopt n original undistorted stereo-picture, set up its distortion stereo-picture set under the different distortion level of different type of distortion, using this distortion stereo-picture set as training set, training set comprises several distortion stereo-pictures; The mean subjective suggestion of the every width distortion stereo-picture then utilizing existing subjective quality assessment method evaluation to go out in training set is divided, and the mean subjective suggestion of the jth width distortion stereo-picture in training set is divided and is designated as MOS j; Again according to step 1. to step process 6., obtain the perception feature set of the every width distortion stereo-picture in training set in an identical manner, the perception feature set of the jth width distortion stereo-picture in training set be designated as P j.
Wherein, n>=1, as got n=1000,1≤j≤S, S represents total width number of the distortion stereo-picture comprised in training set, MOS j∈ [0,5].
8. existing random forest machine learning algorithm is adopted, the perception feature set of all distortion stereo-pictures in training set is trained, make through training the regression function value that obtains to divide with corresponding mean subjective suggestion between error minimum, structure obtains random forest training pattern.
9. according to constructing the random forest training pattern obtained, to I disperception feature set P test, prediction obtain I disevaluating objective quality predicted value, be designated as Q dis, Q dis=MOD (P), wherein, the function representation form that MOD () is random forest training pattern.
For further illustrating feasibility and the validity of the inventive method, the inventive method is tested.
Random Forest model is made up of 20000 decision trees, and the characteristic of structure decision tree is 2.
Adopt 12 undistorted stereo-pictures shown in Fig. 3 a to Fig. 3 l, set up its distortion stereo-picture set under the different distortion level of 5 kinds of different type of distortion, comprise the stereo-picture of the distortion stereo-picture of 60 width JPEG compressed encodings distortion (JPEG), the distortion stereo-picture of 60 width JP2000 compressed encodings distortion (JP2K), the distortion stereo-picture of 60 width white Gaussian noise distortions (WN), the distortion stereo-picture of 60 width Gaussian Blur distortions (GB) and 72 width 312 width distortions H.264 in coding distortion (H.264) situation.The correlation that the respective evaluating objective quality predicted value of stereo-picture of the distortion that analysis and utilization the inventive method obtains and mean subjective are marked between difference.By randomly draw in the stereo-picture of above-mentioned 312 width distortions 80% the stereo-picture composing training collection of distortion, the stereo-picture of the distortion of residue 20% forms test set; Then the mean subjective suggestion utilizing subjective quality assessment method evaluation to go out the stereo-picture of the every width distortion in training set is divided; Again according to step 1. to step process 6., obtain the perception feature set of the stereo-picture of the every width distortion in training set in an identical manner; Then random forest machine learning algorithm is adopted, the perception feature set of all distortion stereo-pictures in training set is trained, make through training the regression function value that obtains to divide with corresponding mean subjective suggestion between error minimum, structure obtains random forest training pattern; Afterwards according to constructing the random forest training pattern obtained, test the perception feature set of the stereo-picture of the every width distortion in test set, prediction obtains the evaluating objective quality predicted value of the stereo-picture of the every width distortion in test set.
Here, utilize 3 of evaluate image quality evaluating method conventional objective parameters as evaluation index, i.e. Pearson linear correlation property coefficient (Pearson Linear Correlation Coefficients, PLCC), Spearman rank correlation coefficient (Spearman Rank Order Correlation coefficient, and root-mean-square error (Rooted Mean Squared Error, RMSE) SROCC).The span of PLCC and SROCC is [0,1], and its value, more close to 1, shows that evaluation method is better, otherwise, poorer; RMSE value is less, and represent that the prediction of evaluation method is more accurate, performance is better, otherwise, then poorer.Represent PLCC, SROCC and RMSE coefficient of assess performance as listed in table 1.From the data listed by table 1, PLCC and SROCC value is all more than 0.94, RMSE is lower than 5.8, that is, evaluating objective quality predicted value and the mean subjective correlation of marking between difference of the stereo-picture of the distortion utilizing the inventive method to obtain are very high, show that the result of objective evaluation result and human eye subjective perception is more consistent, be enough to the validity that the inventive method is described.
The correlation that the evaluating objective quality predicted value of the stereo-picture of the distortion that table 1 calculates by the inventive method and mean subjective are marked between difference

Claims (1)

1., based on an objective evaluation method for quality of stereo images for perception feature set, it is characterized in that comprising the following steps:
1. I is made orgrepresent original undistorted stereo-picture, make I disrepresent the stereo-picture of distortion to be evaluated, by I orgleft visual point image be designated as L org, by I orgright visual point image be designated as R org, by I disleft visual point image be designated as L dis, by I disright visual point image be designated as R dis;
2. adopt frequency modulation conspicuousness detection algorithm, obtain L org, R org, L disand R disrespective remarkable figure, correspondence is designated as with then calculate with mean square error, be designated as MSE sal L = 1 M × N Σ i = 1 M Σ j = 1 N ( S org L ( i , j ) - S dis L ( i , j ) ) 2 ; Equally, calculate with mean square error, be designated as MSE sal R = 1 M × N Σ i = 1 M Σ j = 1 N ( S org R ( i , j ) - S dis R ( i , j ) ) 2 ;
Wherein, M represents I organd I diswidth, N represents I organd I disheight, 1≤i≤M, 1≤j≤N, represent middle coordinate position is the pixel value of the pixel of (i, j), represent middle coordinate position is the pixel value of the pixel of (i, j), represent middle coordinate position is the pixel value of the pixel of (i, j), represent middle coordinate position is the pixel value of the pixel of (i, j);
3. adopt horizontal Sobel operator, obtain L org, R org, L disand R disrespective horizontal gradient figure, correspondence is designated as with and adopt vertical Sobel operator, obtain L org, R org, L disand R disrespective vertical gradient map, correspondence is designated as with then basis with obtain L orggradient map, be designated as will middle coordinate position is that the pixel value of the pixel of (i, j) is designated as G org L ( i , j ) = G h L , org ( i , j ) 2 + G v L , org ( i , j ) 2 ; And according to with obtain R orggradient map, be designated as will middle coordinate position is that the pixel value of the pixel of (i, j) is designated as G org R ( i , j ) = G h R , org ( i , j ) 2 + G v R , org ( i , j ) 2 ; According to with obtain L disgradient map, be designated as will middle coordinate position is that the pixel value of the pixel of (i, j) is designated as G dis L ( i , j ) = G h L , dis ( i , j ) 2 + G v L , dis ( i , j ) 2 ; According to with obtain R disgradient map, be designated as will middle coordinate position is that the pixel value of the pixel of (i, j) is designated as G dis R ( i , j ) = G h R , dis ( i , j ) 2 + G v R , dis ( i , j ) 2 ; Calculate again with mean square error, be designated as MSE gra L = 1 M × N Σ i = 1 M Σ j = 1 N ( G org L ( i , j ) - G dis L ( i , j ) ) 2 ; Equally, calculate with mean square error, be designated as MSE gra R = 1 M × N Σ i = 1 M Σ j = 1 N ( G org R ( i , j ) - G dis R ( i , j ) ) 2 ;
Wherein, represent middle coordinate position is the pixel value of the pixel of (i, j), represent middle coordinate position is the pixel value of the pixel of (i, j), represent middle coordinate position is the pixel value of the pixel of (i, j), represent middle coordinate position is the pixel value of the pixel of (i, j), represent middle coordinate position is the pixel value of the pixel of (i, j), represent middle coordinate position is the pixel value of the pixel of (i, j), represent middle coordinate position is the pixel value of the pixel of (i, j), represent middle coordinate position is the pixel value of the pixel of (i, j);
4. adopt the proper discernable distortion model in spatial domain, obtain L org, R org, L disand R disthe proper discernable distortion map in respective spatial domain, correspondence is designated as with then calculate with mean square error, be designated as MSE JND L = 1 M × N Σ i = 1 M Σ j = 1 N ( J org L ( i , j ) - J dis L ( i , j ) ) 2 ; Equally, calculate with mean square error, be designated as MSE JND R = 1 M × N Σ i = 1 M Σ j = 1 N ( J org R ( i , j ) - J dis R ( i , j ) ) 2 ;
Wherein, represent middle coordinate position is the pixel value of the pixel of (i, j), represent middle coordinate position is the pixel value of the pixel of (i, j), represent middle coordinate position is the pixel value of the pixel of (i, j), represent middle coordinate position is the pixel value of the pixel of (i, j);
5. adopt light stream matching method, obtain I orghorizontal parallax amplitude figure and I orgvertical parallax amplitude figure, correspondence is designated as with then basis with obtain I orgdisparity map, be designated as D org, by D orgmiddle coordinate position is that the pixel value of the pixel of (i, j) is designated as D org(i, j), equally, adopt light stream matching method, obtain I dishorizontal parallax amplitude figure and I disvertical parallax amplitude figure, correspondence is designated as with then basis with obtain I disdisparity map, be designated as D dis, by D dismiddle coordinate position is that the pixel value of the pixel of (i, j) is designated as D dis(i, j), D dis ( i , j ) = ( D h dis ( i , j ) ) 2 + ( D v dis ( i , j ) ) 2 ; Calculate D afterwards organd D dismean square error, be designated as MSE dsp, MSE dsp = 1 M × N Σ i = 1 M Σ j = 1 N ( D org ( i , j ) - D dis ( i , j ) ) 2 ;
Wherein, represent middle coordinate position is the pixel value of the pixel of (i, j), represent middle coordinate position is the pixel value of the pixel of (i, j), represent middle coordinate position is the pixel value of the pixel of (i, j), represent middle coordinate position is the pixel value of the pixel of (i, j);
6. will and MSE dspthe sets definition that arranged in sequence is formed is I disperception feature set, be designated as P, P = { MSE sal L , MSE sal R , MSE gra L , MSE gra R , MSE JND L , MSE JND R , MSE dsp } ;
7. adopt n original undistorted stereo-picture, set up its distortion stereo-picture set under the different distortion level of different type of distortion, using this distortion stereo-picture set as training set, training set comprises several distortion stereo-pictures; The mean subjective suggestion of the every width distortion stereo-picture then utilizing subjective quality assessment method evaluation to go out in training set is divided, and the mean subjective suggestion of the jth width distortion stereo-picture in training set is divided and is designated as MOS j; Again according to step 1. to step process 6., obtain the perception feature set of the every width distortion stereo-picture in training set in an identical manner, the perception feature set of the jth width distortion stereo-picture in training set be designated as P j;
Wherein, n>=1,1≤j≤S, S represents total width number of the distortion stereo-picture comprised in training set, MOS j∈ [0,5];
8. random forest machine learning algorithm is adopted, the perception feature set of all distortion stereo-pictures in training set is trained, make through training the regression function value that obtains to divide with corresponding mean subjective suggestion between error minimum, structure obtains random forest training pattern;
9. according to constructing the random forest training pattern obtained, to I disperception feature set P test, prediction obtain I disevaluating objective quality predicted value, be designated as Q dis, Q dis=MOD (P), wherein, the function representation form that MOD () is random forest training pattern.
CN201510303868.3A 2015-06-04 2015-06-04 Objective stereo image quality assessment method based on perception feature set Active CN104954778B (en)

Priority Applications (1)

Application Number Priority Date Filing Date Title
CN201510303868.3A CN104954778B (en) 2015-06-04 2015-06-04 Objective stereo image quality assessment method based on perception feature set

Applications Claiming Priority (1)

Application Number Priority Date Filing Date Title
CN201510303868.3A CN104954778B (en) 2015-06-04 2015-06-04 Objective stereo image quality assessment method based on perception feature set

Publications (2)

Publication Number Publication Date
CN104954778A true CN104954778A (en) 2015-09-30
CN104954778B CN104954778B (en) 2017-05-24

Family

ID=54169076

Family Applications (1)

Application Number Title Priority Date Filing Date
CN201510303868.3A Active CN104954778B (en) 2015-06-04 2015-06-04 Objective stereo image quality assessment method based on perception feature set

Country Status (1)

Country Link
CN (1) CN104954778B (en)

Cited By (9)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN105205504A (en) * 2015-10-04 2015-12-30 北京航空航天大学 Image interest region quality evaluation index learning method based on data driving
CN105376563A (en) * 2015-11-17 2016-03-02 浙江科技学院 No-reference three-dimensional image quality evaluation method based on binocular fusion feature similarity
CN105611272A (en) * 2015-12-28 2016-05-25 宁波大学 Eye exactly perceptible stereo image distortion analyzing method based on texture complexity
CN106412569A (en) * 2016-09-28 2017-02-15 宁波大学 No-reference multi-distortion three-dimensional image quality evaluation method based on feature selection
CN106780476A (en) * 2016-12-29 2017-05-31 杭州电子科技大学 A kind of stereo-picture conspicuousness detection method based on human-eye stereoscopic vision characteristic
CN106803952A (en) * 2017-01-20 2017-06-06 宁波大学 With reference to the cross validation depth map quality evaluating method of JND model
CN108401150A (en) * 2018-03-22 2018-08-14 浙江科技学院 A kind of compressed sensing reconstruction algorithm statistic of attribute evaluation method of analog vision subjective perception
CN109872305A (en) * 2019-01-22 2019-06-11 浙江科技学院 It is a kind of based on Quality Map generate network without reference stereo image quality evaluation method
CN114567776A (en) * 2022-02-21 2022-05-31 宁波职业技术学院 Video low-complexity coding method based on panoramic visual perception characteristic

Citations (4)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN103200420A (en) * 2013-03-19 2013-07-10 宁波大学 Three-dimensional picture quality objective evaluation method based on three-dimensional visual attention
CN103581661A (en) * 2013-10-28 2014-02-12 宁波大学 Method for evaluating visual comfort degree of three-dimensional image
KR20140148080A (en) * 2013-06-21 2014-12-31 한국과학기술원 Stereoscopic imaging method and system for visually comfortable 3D images
CN104394403A (en) * 2014-11-04 2015-03-04 宁波大学 A compression-distortion-oriented stereoscopic video quality objective evaluating method

Patent Citations (4)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN103200420A (en) * 2013-03-19 2013-07-10 宁波大学 Three-dimensional picture quality objective evaluation method based on three-dimensional visual attention
KR20140148080A (en) * 2013-06-21 2014-12-31 한국과학기술원 Stereoscopic imaging method and system for visually comfortable 3D images
CN103581661A (en) * 2013-10-28 2014-02-12 宁波大学 Method for evaluating visual comfort degree of three-dimensional image
CN104394403A (en) * 2014-11-04 2015-03-04 宁波大学 A compression-distortion-oriented stereoscopic video quality objective evaluating method

Non-Patent Citations (6)

* Cited by examiner, † Cited by third party
Title
PMFS: A Perceptual Modulated Feature Similarity Metric for Stereoscopic Image Quality Assessment;Wujie Zhou ET AL;《IEEE SIGNAL PROCESSING LETTERS》;20140831;第21卷(第8期);第1003-1006页 *
WUJIE ZHOU ET AL: "PMFS: A Perceptual Modulated Feature Similarity Metric for Stereoscopic Image Quality Assessment", 《IEEE SIGNAL PROCESSING LETTERS》 *
一种基于人眼视觉特性的视频质量评价算法;朱宏等;《计算机辅助设计与图形学学报》;20140530;第26卷(第5期);第776-781页 *
基于显著性分析的立体图像视觉舒适度预测;邵枫等;《光学精密工程》;20100630;第22卷(第6期);第1631-1638页 *
朱宏等: "一种基于人眼视觉特性的视频质量评价算法", 《计算机辅助设计与图形学学报》 *
邵枫等: "基于显著性分析的立体图像视觉舒适度预测", 《光学精密工程》 *

Cited By (13)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN105205504A (en) * 2015-10-04 2015-12-30 北京航空航天大学 Image interest region quality evaluation index learning method based on data driving
CN105205504B (en) * 2015-10-04 2018-09-18 北京航空航天大学 A kind of image attention regional quality evaluation index learning method based on data-driven
CN105376563A (en) * 2015-11-17 2016-03-02 浙江科技学院 No-reference three-dimensional image quality evaluation method based on binocular fusion feature similarity
CN105611272A (en) * 2015-12-28 2016-05-25 宁波大学 Eye exactly perceptible stereo image distortion analyzing method based on texture complexity
CN106412569B (en) * 2016-09-28 2017-12-15 宁波大学 A kind of selection of feature based without referring to more distortion stereo image quality evaluation methods
CN106412569A (en) * 2016-09-28 2017-02-15 宁波大学 No-reference multi-distortion three-dimensional image quality evaluation method based on feature selection
CN106780476A (en) * 2016-12-29 2017-05-31 杭州电子科技大学 A kind of stereo-picture conspicuousness detection method based on human-eye stereoscopic vision characteristic
CN106803952A (en) * 2017-01-20 2017-06-06 宁波大学 With reference to the cross validation depth map quality evaluating method of JND model
CN108401150A (en) * 2018-03-22 2018-08-14 浙江科技学院 A kind of compressed sensing reconstruction algorithm statistic of attribute evaluation method of analog vision subjective perception
CN109872305A (en) * 2019-01-22 2019-06-11 浙江科技学院 It is a kind of based on Quality Map generate network without reference stereo image quality evaluation method
CN109872305B (en) * 2019-01-22 2020-08-18 浙江科技学院 No-reference stereo image quality evaluation method based on quality map generation network
CN114567776A (en) * 2022-02-21 2022-05-31 宁波职业技术学院 Video low-complexity coding method based on panoramic visual perception characteristic
CN114567776B (en) * 2022-02-21 2023-05-05 宁波职业技术学院 Video low-complexity coding method based on panoramic visual perception characteristics

Also Published As

Publication number Publication date
CN104954778B (en) 2017-05-24

Similar Documents

Publication Publication Date Title
CN104954778A (en) Objective stereo image quality assessment method based on perception feature set
CN102333233B (en) Stereo image quality objective evaluation method based on visual perception
CN105208374B (en) A kind of non-reference picture assessment method for encoding quality based on deep learning
CN103581661B (en) Method for evaluating visual comfort degree of three-dimensional image
CN104902267B (en) No-reference image quality evaluation method based on gradient information
CN102209257A (en) Stereo image quality objective evaluation method
CN108428227A (en) Non-reference picture quality appraisement method based on full convolutional neural networks
CN101562675B (en) No-reference image quality evaluation method based on Contourlet transform
CN101610425B (en) Method for evaluating stereo image quality and device
CN104811691B (en) A kind of stereoscopic video quality method for objectively evaluating based on wavelet transformation
CN105407349A (en) No-reference objective three-dimensional image quality evaluation method based on binocular visual perception
CN107635136B (en) View-based access control model perception and binocular competition are without reference stereo image quality evaluation method
CN104394403B (en) A kind of stereoscopic video quality method for objectively evaluating towards compression artefacts
CN102547368A (en) Objective evaluation method for quality of stereo images
CN102708567B (en) Visual perception-based three-dimensional image quality objective evaluation method
CN104902268B (en) Based on local tertiary mode without with reference to three-dimensional image objective quality evaluation method
CN104658001A (en) Non-reference asymmetric distorted stereo image objective quality assessment method
Geng et al. A stereoscopic image quality assessment model based on independent component analysis and binocular fusion property
CN107330873B (en) Stereo image quality objective evaluation method based on multi-scale binocular fusion and local feature extraction
CN103873854A (en) Method for determining number of stereoscopic image subjective assessment testees and experiment data
CN105357519A (en) Quality objective evaluation method for three-dimensional image without reference based on self-similarity characteristic
CN104361583A (en) Objective quality evaluation method of asymmetrically distorted stereo images
CN105894507A (en) Image quality evaluation method based on image information content natural scenario statistical characteristics
CN106023152B (en) It is a kind of without with reference to objective evaluation method for quality of stereo images
CN102999911B (en) Three-dimensional image quality objective evaluation method based on energy diagrams

Legal Events

Date Code Title Description
C06 Publication
PB01 Publication
C10 Entry into substantive examination
SE01 Entry into force of request for substantive examination
GR01 Patent grant
GR01 Patent grant