CN105163111A - Method for evaluating visual comfort of three-dimensional image based on classification of scene modes - Google Patents

Method for evaluating visual comfort of three-dimensional image based on classification of scene modes Download PDF

Info

Publication number
CN105163111A
CN105163111A CN201510571897.8A CN201510571897A CN105163111A CN 105163111 A CN105163111 A CN 105163111A CN 201510571897 A CN201510571897 A CN 201510571897A CN 105163111 A CN105163111 A CN 105163111A
Authority
CN
China
Prior art keywords
stereo
picture
scene mode
pixel
screen
Prior art date
Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
Granted
Application number
CN201510571897.8A
Other languages
Chinese (zh)
Other versions
CN105163111B (en
Inventor
郁梅
应宏微
王颖
蒋刚毅
Current Assignee (The listed assignees may be inaccurate. Google has not performed a legal analysis and makes no representation or warranty as to the accuracy of the list.)
Xuancheng Youdu Technology Service Co., Ltd.
Original Assignee
Ningbo University
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Application filed by Ningbo University filed Critical Ningbo University
Priority to CN201510571897.8A priority Critical patent/CN105163111B/en
Publication of CN105163111A publication Critical patent/CN105163111A/en
Application granted granted Critical
Publication of CN105163111B publication Critical patent/CN105163111B/en
Active legal-status Critical Current
Anticipated expiration legal-status Critical

Links

Abstract

The invention discloses a method for evaluating the visual comfort of a three-dimensional image based on classification of scene modes. The method comprises the following steps: determining ten kinds of scene modes of the three-dimensional image according to depth positions, where a foreground object and a background area in a natural scene are; then, determining the scene mode, which each three-dimensional image belongs to, in a three-dimensional image database, and establishing a visual comfort evaluation mode in each scene mode; and firstly determining the scene mode, which the three-dimensional image to be evaluated belongs to, then, selecting the visual comfort evaluation mode in the scene mode, calculating a visual comfort evaluation predication value of the three-dimensional image to be evaluated, and finally, correcting the visual comfort evaluation predication value of the three-dimensional image to be evaluated. The method disclosed by the invention has the advantages that: perception of a human visual system to the visual comfort of the three-dimensional image can be precisely reflected through the visual comfort evaluation modes respectively established based on ten kinds of scenes; and thus, the consistency between an objective evaluation result and human eye subjective perception can be effectively improved.

Description

A kind of stereo image vision comfort level evaluation method based on scene mode classification
Technical field
The present invention relates to a kind of stereo-picture Quality of experience evaluation method, especially relate to a kind of stereo image vision comfort level evaluation method based on scene mode classification.
Background technology
Along with the fast development of Stereoscopic Video Presentation technology and high-quality stereoscopic video content acquiring technology, visual experience quality (the QoE of three-dimensional video-frequency, qualityofexperience) be a major issue in three-dimensional video-frequency system, and visual comfort (VC, visualcomfort) is one of key factor of the visual experience quality affecting three-dimensional video-frequency.At present, the visual experience matter quantifier elimination of stereoscopic video/stereo-picture mainly concentrates on the impact of research contents distortion for visual experience quality, and this respect has created many achievements in research.The research of stereo image vision comfort level is the physiological discomfort caused under stereoscopic image content non-distortion prerequisite, the visual comfort objective evaluation model obtained for improve beholder visual experience quality, instruct the making of 3D content and post-processed to have a very important role.
Existing stereo image vision comfort level evaluation method is mainly by extracting the left and right visual point image of stereo-picture and the feature of anaglyph, such as parallax, gradient etc., the method then in Using statistics or the algorithm of machine learning set up objective models.But cause the various factors of discomfort on physiology not study yet thoroughly in the content of current stereoscopic image, cause the consistency between objective evaluation result and human eye subjective perception poor.For obtaining the good visual comfort evaluation method of evaluation effect, need to segment all kinds of natural scene, and on this basis the various characteristics of image of Accurate Analysis to the influence degree of visual comfort.
Summary of the invention
Technical problem to be solved by this invention is to provide a kind of stereo image vision comfort level evaluation method based on scene mode classification, and it can improve the consistency between objective evaluation result and human eye subjective perception effectively.
The present invention solves the problems of the technologies described above adopted technical scheme: a kind of stereo image vision comfort level evaluation method based on scene mode classification, is characterized in that comprising the following steps:
1. determine 10 kinds of scene modes of the natural scene shown by three-dimensional display, the 1st kind of scene mode in screen, to be that the foreground target in the right viewpoint parallax gray level image of stereo-picture is convex be in that non-comfort zone, background region are recessed is in non-comfort zone in screen; 2nd kind of scene mode in screen, to be that the foreground target in the right viewpoint parallax gray level image of stereo-picture is convex be in that non-comfort zone, background region are convex is in non-comfort zone in screen; 3rd kind of scene mode in screen, to be that the foreground target in the right viewpoint parallax gray level image of stereo-picture is recessed be in that non-comfort zone, background region are recessed is in non-comfort zone in screen; 4th kind of scene mode in screen, to be that the foreground target in the right viewpoint parallax gray level image of stereo-picture is convex be in that non-comfort zone, background region are recessed is in comfort zone in screen; 5th kind of scene mode in screen, to be that the foreground target in the right viewpoint parallax gray level image of stereo-picture is convex be in that non-comfort zone, background region are convex is in comfort zone in screen; 6th kind of scene mode in screen, to be that the foreground target in the right viewpoint parallax gray level image of stereo-picture is convex be in that comfort zone, background region are recessed is in non-comfort zone in screen; 7th kind of scene mode in screen, to be that the foreground target in the right viewpoint parallax gray level image of stereo-picture is recessed be in that comfort zone, background region are recessed is in non-comfort zone in screen; 8th kind of scene mode in screen, to be that the foreground target in the right viewpoint parallax gray level image of stereo-picture is convex be in that comfort zone, background region are recessed is in comfort zone in screen; 9th kind of scene mode in screen, to be that the foreground target in the right viewpoint parallax gray level image of stereo-picture is convex be in that comfort zone, background region are convex is in comfort zone in screen; 10th kind of scene mode in screen, to be that the foreground target in the right viewpoint parallax gray level image of stereo-picture is recessed be in that comfort zone, background region are recessed is in comfort zone in screen;
2. every width stereo-picture is selected to have a stereoscopic image data storehouse of the mean subjective scoring average of visual comfort; Then according to the parallactic angle of foreground target in the right viewpoint parallax gray level image of the every width stereo-picture in stereoscopic image data storehouse and the parallactic angle of background region, the scene mode belonging to every width stereo-picture in stereoscopic image data storehouse is determined; Again according to the parallactic angle of foreground target in the respective right viewpoint parallax gray level image of all stereo-pictures belonging to often kind of scene mode in stereoscopic image data storehouse and the parallactic angle of background region, set up the visual comfort evaluation model under often kind of scene mode, the visual comfort evaluation model under n-th kind of scene mode is described below: SMMO n=(4.2028-u (Q n))-v (Q n) × D a(Q n)+0.1912 × ln (W a)-0.0208 × D a(Q n) × ln (W a), wherein, 1≤n≤10, SMMO nrepresent the output of the visual comfort evaluation model under n-th kind of scene mode, u (Q n) and v (Q n) be constant, D a(Q n) represent that a width to be entered belongs to the global disparity angle of the right viewpoint parallax gray level image of the stereo-picture to be evaluated of n-th kind of scene mode, D a(Q n)=Q n× | f a|+(1-Q n) × | b a|, Q nrepresent the weight under n-th kind of scene mode, f arepresent that a width to be entered belongs to the parallactic angle of the foreground target in the right viewpoint parallax gray level image of the stereo-picture to be evaluated of n-th kind of scene mode, b arepresent that a width to be entered belongs to the parallactic angle of the background region in the right viewpoint parallax gray level image of the stereo-picture to be evaluated of n-th kind of scene mode, W arepresent that a width to be entered belongs to the width angle of the foreground target in the right viewpoint parallax gray level image of the stereo-picture to be evaluated of n-th kind of scene mode, symbol " || " is the symbol that takes absolute value;
3. the right viewpoint parallax gray level image of stereo-picture to be evaluated is designated as { d r(x, y) }, wherein, 1≤x≤W, 1≤y≤H, W represents the width of stereo-picture to be evaluated, it is consistent with the width of the every width stereo-picture in stereoscopic image data storehouse, H represents the height of stereo-picture to be evaluated, and it is consistent with the height of the every width stereo-picture in stereoscopic image data storehouse, d r(x, y) represents { d r(x, y) } in coordinate position be the pixel value of the pixel of (x, y); Then adopt and step 2. in identical mode, determine the scene mode belonging to stereo-picture to be evaluated; Scene mode again belonging to stereo-picture to be evaluated, chooses the visual comfort evaluation model under this scene mode; Then according to the visual comfort evaluation model under this scene mode and { d r(x, y) } in the parallactic angle, { d of foreground target r(x, y) } in the parallactic angle of background region and { d r(x, y) } in the width angle of foreground target, calculate the visual comfort evaluation and foreca value of stereo-picture to be evaluated, be designated as smmo, suppose that stereo-picture to be evaluated belongs to n-th kind of scene mode, then smmo=(4.2028-u (Q n))-v (Q n) × D a' (Q n)+0.1912 × ln (W a')-0.0208 × D a' (Q n) × ln (W a'), wherein, D a' (Q n) represent { d r(x, y) } global disparity angle, D a' (Q n)=Q n× | f a' |+(1-Q n) × | b a' |, f a' represent { d r(x, y) } in the parallactic angle of foreground target, b a' represent { d r(x, y) } in the parallactic angle of background region, W a' represent { d r(x, y) } in the width angle of foreground target;
4. the visual comfort evaluation and foreca value smmo of stereo-picture to be evaluated is revised, the revised visual comfort evaluation and foreca value of stereo-picture to be evaluated is designated as smm, wherein, max () is for getting max function, and P represents the tortuosity attenuation coefficient of the visual comfort evaluation model under the scene mode belonging to stereo-picture to be evaluated, S r' represent { d r(x, y) } in the average prospect line hop count of foreground target, S c' represent { d r(x, y) } in the average prospect alignment hop count of foreground target, T frepresent the parallactic angle threshold value of setting, T rrepresent the prospect line hop count threshold value of setting, T crepresent the prospect alignment hop count threshold value of setting.
Described step is middle Q 2. nacquisition process be:
-1 2., total width number of supposing the stereo-picture belonging to n-th kind of scene mode in stereoscopic image data storehouse is M', wherein, and M' >=1;
2.-2, q is made ninitial value be 0.1, make Q ninitial value be 0;
2. the weight value-3, calculated under n-th kind of scene mode is q ntime stereoscopic image data storehouse in belong to the global disparity angle of the right viewpoint parallax gray level image of every width stereo-picture of n-th kind of scene mode, be q by the weight value under n-th kind of scene mode ntime stereoscopic image data storehouse in belong to the right viewpoint parallax gray level image of the m' width stereo-picture of n-th kind of scene mode global disparity angle be designated as D m', a(q n), D m', a(q n)=q n× | f m', a|+(1-q n) × | b m', a|, wherein, 1≤m'≤M', f m', aand b m', abelong to the parallactic angle of foreground target in the right viewpoint parallax gray level image of the m' width stereo-picture of n-th kind of scene mode and the parallactic angle of background region in corresponding expression stereoscopic image data storehouse, symbol " || " is the symbol that takes absolute value;
2. the weight value-4, calculated under n-th kind of scene mode is q ntime stereoscopic image data storehouse in belong to n-th kind of scene mode every width stereo-picture eliminate the comfort level objective evaluation value of parallax linear effect, be q by the weight value under n-th kind of scene mode ntime stereoscopic image data storehouse in belong to n-th kind of scene mode the m' width stereo-picture comfort level objective evaluation value that eliminates parallax linear effect be designated as VC h(D m', a(q n), W a), VC h(D m', a(q n), W a)=4.2028+0.1912 × ln (W a)-0.0208 × D m', a(q n) × ln (W a); Then the weight value calculated under n-th kind of scene mode is q ntime stereoscopic image data storehouse in belong to the comfort level objective evaluation value of the parallax linear effect of every width stereo-picture of n-th kind of scene mode, be q by the weight value under n-th kind of scene mode ntime stereoscopic image data storehouse in belong to the parallax linear effect of the m' width stereo-picture of n-th kind of scene mode comfort level objective evaluation value be designated as ERR m', ERR m'=VC h(D m', a(q n), W a)-MOS m', wherein, MOS m'represent that the weight value under n-th kind of scene mode is q ntime stereoscopic image data storehouse in belong to the comfort level subjective assessment value of the m' width stereo-picture of n-th kind of scene mode; Adopting least square method again, is q to the weight value under n-th kind of scene mode ntime stereoscopic image data storehouse in belong to n-th kind of scene mode the respective global disparity angle of right viewpoint parallax gray level image of all stereo-pictures and the comfort level objective evaluation value of respective parallax linear effect carry out linear fit, the weight value obtained under n-th kind of scene mode is q ntime stereoscopic image data storehouse in belong to a fitting a straight line equation corresponding to all stereo-pictures of n-th kind of scene mode: ERR=u (q n)+v (q n) × D a(q n), wherein, ERR is for representing the dependent variable of parallax to the linear effect of objective comfort level, u (q n) and v (q n) be ERR=u (q n)+v (q n) × D a(q n) in constant, D a(q n) be for representing that the weight value under n-th kind of scene mode is q ntime stereoscopic image data storehouse in belong to the independent variable at the global disparity angle of the right viewpoint parallax gray level image of arbitrary width stereo-picture of n-th kind of scene mode;
2. be-5, q according to the weight value under n-th kind of scene mode ntime stereoscopic image data storehouse in belong to a fitting a straight line equation corresponding to all stereo-pictures of n-th kind of scene mode, the weight value obtained under n-th kind of scene mode is q ntime stereoscopic image data storehouse in belong to the comfort level objective evaluation equation of all stereo-pictures of n-th kind of scene mode: SMMO n(q n)=VC h(D a(q n), W a)-ERR=(4.2028-u (q n))-v (q n) × D a(q n)+0.1912 × ln (W a)-0.0208 × D a(q n) × ln (W a), wherein, SMMO n(q n) for having superposed the comfort level objective evaluation value of parallax linear effect and parallax non-linear effects, VC h(D a(q n), W a) represent that the weight value under n-th kind of scene mode is q ntime stereoscopic image data storehouse in belong to n-th kind of scene mode arbitrary width stereo-picture eliminate the comfort level objective evaluation value of parallax linear effect; The weight value of then measuring under n-th kind of scene mode is q ntime stereoscopic image data storehouse in belong to the comfort level objective evaluation value of all stereo-pictures of n-th kind of scene mode and the fitting degree of corresponding comfort level subjective assessment value;
2.-6, q is made n=q n+ 0.1, then return step 2.-3 continuation execution, until q nterminate when equaling 1.1, corresponding fitting degree when the weight value obtained under n-th kind of scene mode is respectively 0.1,0.2,0.3,0.4,0.5,0.6,0.7,0.8,0.9,1.0, then by weight value assignment corresponding for best fit degree to Q n, wherein, q n=q n"=" in+0.1 is assignment;
Described step is middle u (Q 2. n) value be Monomial coefficient in fitting a straight line equation corresponding to best fit degree; Described step is middle v (Q 2. n) value be constant in fitting a straight line equation corresponding to best fit degree.
The obtain manner of the foreground target in the right viewpoint parallax gray level image of every width stereo-picture of described step 2. in neutral body image data base and background region and described step 3. in stereo-picture to be evaluated right viewpoint parallax gray level image in foreground target identical with the obtain manner of background region, using the every width stereo-picture in stereoscopic image data storehouse and stereo-picture to be evaluated all as pending stereo-picture, foreground target in the right viewpoint parallax gray level image of then pending stereo-picture and the acquisition process of background region are: the grey level histogram obtaining the right viewpoint parallax gray level image of pending stereo-picture, then maximum variance between clusters is adopted, and the grey level histogram of right viewpoint parallax gray level image according to pending stereo-picture, obtain the intensity slicing threshold value of the right viewpoint parallax gray level image of pending stereo-picture, be designated as P seg, again pixel value in the right viewpoint parallax gray level image of pending stereo-picture is more than or equal to P segpixel be defined as foreground pixel point, and pixel value in the right viewpoint parallax gray level image of pending stereo-picture is less than P segpixel be defined as background pixel, the last foreground target be made up of all foreground pixel points in the right viewpoint parallax gray level image of pending stereo-picture, forms the background region in the right viewpoint parallax gray level image of pending stereo-picture by all background pixels.
The intensity slicing threshold value P of the right viewpoint parallax gray level image of described pending stereo-picture segacquisition process be:
X1, set up intensity slicing threshold value find target function, be designated as T, T=w f× (μ-μ f) 2+ w b× (μ-μ b) 2, wherein, w frepresent that total number of the foreground pixel point in the right viewpoint parallax gray level image of pending stereo-picture accounts for the ratio of total number of the pixel in the right viewpoint parallax gray level image of pending stereo-picture, μ frepresent the average of the pixel value of all foreground pixel points in the right viewpoint parallax gray level image of pending stereo-picture, w brepresent that total number of the background pixel in the right viewpoint parallax gray level image of pending stereo-picture accounts for the ratio of total number of the pixel in the right viewpoint parallax gray level image of pending stereo-picture, μ brepresent the average of the pixel value of all background pixels in the right viewpoint parallax gray level image of pending stereo-picture, μ represents the average of the pixel value of all pixels in the right viewpoint parallax gray level image of pending stereo-picture, μ=w f× μ f+ w b× μ b, Hist dg () represents the grey level histogram { Hist of the right viewpoint parallax gray level image of pending stereo-picture d(g) } in pixel value be the number of the pixel of g, 0≤g min≤ g≤g max≤ 255, g minrepresent the grey level histogram { Hist of the right viewpoint parallax gray level image of pending stereo-picture d(g) } in the number of the pixel minimum pixel value that is greater than 0, g maxrepresent the grey level histogram { Hist of the right viewpoint parallax gray level image of pending stereo-picture d(g) } in the number of the pixel max pixel value that is greater than 0;
X2, at interval [g min, g max] interior traversal t, the value of t when making T maximum is defined as the intensity slicing threshold value P of the right viewpoint parallax gray level image of pending stereo-picture seg.
Described step 2. in determine to determine that the mode of the scene mode belonging to stereo-picture to be evaluated is identical during the mode of the scene mode belonging to every width stereo-picture in stereoscopic image data storehouse and described step are 3., using the every width stereo-picture in stereoscopic image data storehouse and stereo-picture to be evaluated all as pending stereo-picture, then determine that the detailed process of the scene mode belonging to pending stereo-picture is:
Z1, for the foreground target in the right viewpoint parallax gray level image of pending stereo-picture, if f a *>1 °, then determine that foreground target is convex and be in non-comfort zone in screen; If 0 °≤f a *<1 °, then to determine that foreground target is convex and be in comfort zone in screen; If-1 ° of <f a *<0 °, then to determine that foreground target is recessed and be in comfort zone in screen; If f a *<-1 °, then determine that foreground target is recessed and be in non-comfort zone in screen;
For the background region in the right viewpoint parallax gray level image of pending stereo-picture, if b a *>1 °, then determine that background region is convex and be in non-comfort zone in screen; If 0 °≤b a *<1 °, then to determine that background region is convex and be in comfort zone in screen; If-1 ° of < b a *<0 °, then to determine that background region is recessed and be in comfort zone in screen; If b a *<-1 °, then determine that background region is recessed and be in non-comfort zone in screen;
Wherein, f a *represent the parallactic angle of the foreground target in the right viewpoint parallax gray level image of pending stereo-picture, f a *=f-k, b a *represent the parallactic angle of the background region in the right viewpoint parallax gray level image of pending stereo-picture, b a *=b-k, f represent the convergent angle of human eye binocular viewing foreground target, b represents the convergent angle of human eye binocular viewing background region, k represents the adjustment angle of human eye binocular, arctan () is tan of negating, p represents the interpupillary distance of human eye binocular, L represents the width of display, N represents the horizontal resolution of display, h represents the distance of human eye to display, F represents the mean parallax amplitude of the foreground target in the right viewpoint parallax gray level image of pending stereo-picture, and F is in units of pixel F = &Sigma; x = 1 W &Sigma; y = 1 H &epsiv; ( d R * ( x , y ) - P s e g ) &times; d R * ( x , y ) &Sigma; x = 1 W &Sigma; y = 1 H &epsiv; ( d R * ( x , y ) - P s e g ) , &epsiv; ( d R * ( x , y ) - P s e g ) = 1 , d R * ( x , y ) - P s e g &GreaterEqual; 0 0 , d R * ( x , y ) - P s e g < 0 , B represents the mean parallax amplitude of the background region in the right viewpoint parallax gray level image of pending stereo-picture, B in units of pixel, B = &Sigma; x = 1 W &Sigma; y = 1 H &epsiv; ( P s e g - d R * ( x , y ) ) &times; d R * ( x , y ) &Sigma; x = 1 W &Sigma; y = 1 H &epsiv; ( P s e g - d R * ( x , y ) ) , &epsiv; ( P s e g - d R * ( x , y ) ) = 1 , P s e g - d R * ( x , y ) &GreaterEqual; 0 0 , P s e g - d R * ( x , y ) < 0 , D r *(x, y) represents that in the right viewpoint parallax gray level image of pending stereo-picture, coordinate position is the pixel value of the pixel of (x, y);
Z2, according to the foreground target in the right viewpoint parallax gray level image of pending stereo-picture relative to the concavity and convexity of screen and whether be in comfort zone, background region relative to screen concavity and convexity and whether be in comfort zone, determine the scene mode belonging to pending stereo-picture, be specially: if foreground target is convex be in that non-comfort zone, background region are recessed is in non-comfort zone in screen in screen, then determine that pending stereo-picture belongs to the 1st kind of scene mode; If foreground target is convex be in that non-comfort zone, background region are convex is in non-comfort zone in screen in screen, then determine that pending stereo-picture belongs to the 2nd kind of scene mode; If foreground target is recessed be in that non-comfort zone, background region are recessed is in non-comfort zone in screen in screen, then determine that pending stereo-picture belongs to the 3rd kind of scene mode; If foreground target is convex be in that non-comfort zone, background region are recessed is in comfort zone in screen in screen, then determine that pending stereo-picture belongs to the 4th kind of scene mode; If foreground target is convex be in that non-comfort zone, background region are convex is in comfort zone in screen in screen, then determine that pending stereo-picture belongs to the 5th kind of scene mode; If foreground target is convex be in that comfort zone, background region are recessed is in non-comfort zone in screen in screen, then determine that pending stereo-picture belongs to the 6th kind of scene mode; If foreground target is recessed be in that comfort zone, background region are recessed is in non-comfort zone in screen in screen, then determine that pending stereo-picture belongs to the 7th kind of scene mode; If foreground target is convex be in that comfort zone, background region are recessed is in comfort zone in screen in screen, then determine that pending stereo-picture belongs to the 8th kind of scene mode; If foreground target is convex be in that comfort zone, background region are convex is in comfort zone in screen in screen, then determine that pending stereo-picture belongs to the 9th kind of scene mode; If foreground target is recessed be in that comfort zone, background region are recessed is in comfort zone in screen in screen, then determine that pending stereo-picture belongs to the 10th kind of scene mode.
Described step 3. in { d r(x, y) } in the width angle of foreground target wherein, arctan () is tan of negating, and h represents the distance of human eye to display, W frontrepresent { d r(x, y) } in the mean breadth of foreground target, W frontacquisition process be:
A1, to { d r(x, y) } carry out binary conversion treatment, obtain { d r(x, y) } binary image, be designated as { BI (x, y) }, the pixel value being the pixel of (x, y) by coordinate position in { BI (x, y) } is designated as BI (x, y), as { d r(x, y) } in coordinate position be the pixel of (x, y) when belonging to foreground target, make BI (x, y)=1, as { d r(x, y) } in coordinate position be the pixel of (x, y) when belonging to background region, make BI (x, y)=0;
A2, to { BI (x, y) } line scanning is carried out, for { BI (x, y) row } is capable, scan to the right from the 1st pixel that row is capable, when to scan the 1st pixel value be the pixel of 1, using the original position of this pixel as the 1st section of prospect width line segment, and the row coordinate of this pixel is designated as x 1, continuing scanning to the right and obtain the 1st section of prospect width line segment until scanning till pixel value is the pixel of 0, is that the row coordinate of the pixel of 0 is designated as x by this pixel value 2, the length of the 1st section of prospect width line segment is designated as WL 1, WL 1=x 2-x 1; Continue scan to the right, in the mode identical with acquisition the 1st section of prospect width line segment, obtain row capable in all prospect width line segments; Wherein, the initial value of row is 1,1≤row≤H, 1≤x 1<x 2≤ W;
All prospect width line segments that a3, all prospect width line segments length in { BI (x, y) } being less than 0.002W and length are greater than 0.995W are removed; Then sort to remaining all prospect width line segments by length order from small to large, the prospect width line segment getting middle 80% forms prospect width line segment aggregate, is designated as { WL' n, wherein, WL' nrepresent { WL' nin the length of n-th section of prospect width line segment, 1≤n≤N', N' represents { WL' nin total hop count of prospect width line segment of comprising;
A4, calculating { WL' nin the mean value of length of all prospect width line segments, be designated as W front', then by W front' as { d r(x, y) } in the mean breadth W of foreground target fronteven, W front=W front'.
Described step 4. in { d r(x, y) } in the average prospect line hop count of foreground target { d r(x, y) } in the average prospect alignment hop count of foreground target wherein, 1≤row≤H, 1≤col≤W, RLQ rowrepresent { d r(x, y) } in row capable in the quantity of prospect line section, &epsiv; ( RLQ r o w ) = { 1 , RLQ r o w > 0 0 , RLQ r o w &le; 0 , CLQ colrepresent { d r(x, y) } in col row in the quantity of prospect alignment section, &epsiv; ( CLQ c o l ) = 1 , C L Q c o l > 0 0 , C L Q c o l &le; 0 ;
Wherein, RLQ rowand CLQ colacquisition process be:
B1, to { d r(x, y) } carry out binary conversion treatment, obtain { d r(x, y) } binary image, be designated as { BI (x, y) }, the pixel value being the pixel of (x, y) by coordinate position in { BI (x, y) } is designated as BI (x, y), as { d r(x, y) } in coordinate position be the pixel of (x, y) when belonging to foreground target, make BI (x, y)=1, as { d r(x, y) } in coordinate position be the pixel of (x, y) when belonging to background region, make BI (x, y)=0;
B2, to { BI (x, y) 2 dilation operations of mathematical morphology, 4 erosion operations, 2 dilation operations } are carried out successively, obtain { BI'(x, y) }, wherein, BI'(x, y) represent { BI'(x, y) }, coordinate position is the pixel value of the pixel of (x, y);
The quantity of effective prospect line section in often row in b3, statistics { BI'(x, y) }, by capable for the row in { BI'(x, y) } effectively the quantity of prospect line section be designated as RLQ row', RLQ row' acquisition process be: b3-1, make RLQ row' initial value be 0; B3-2, from { BI'(x, y) the 1st pixel that the row } is capable starts to scan to the right, when to scan the 1st pixel value be the pixel of 1, using the original position of this pixel as the 1st section of prospect line section, and the row coordinate of this pixel is designated as x 1', continuing scanning to the right and obtain the 1st section of prospect line section until scanning till pixel value is the pixel of 0, is that the row coordinate of the pixel of 0 is designated as x by this pixel value 2', if the length x of the 1st section of prospect line section 2'-x 1' be greater than 0.005W, then determine that the 1st section of prospect line section is effective prospect line section, and make RLQ row'=RLQ row'+1, if the length x of the 1st section of prospect line section 2'-x 1' be less than or equal to 0.005W, then determine that the 1st section of prospect line section is invalid prospect line section; Continue to scan to the right, to determine whether the 1st section of prospect line section is the mode that effective prospect line section is identical with acquisition the 1st section of prospect line section, obtain row capable in all effective prospect line section, and statistics obtain row capable in the amount R LQ of effective prospect line section row'; Wherein, the initial value of row is 1,1≤row≤H, 1≤x 1' <x 2'≤W, RLQ row'=RLQ row"=" in '+1 is assignment;
Equally, the quantity of effective prospect alignment section in the often row in statistics { BI'(x, y) }, by capable for the col in { BI'(x, y) } effectively the quantity of prospect alignment section be designated as CLQ col', CLQ col' acquisition process be: b3-1), make CLQ col' initial value be 0; B3-2), from { BI'(x, 1st pixel of the col row y) } starts downward scanning, when to scan the 1st pixel value be the pixel of 1, using the original position of this pixel as the 1st section of prospect alignment section, and the row-coordinate of this pixel is designated as y 1', continuing scanning downwards and obtain the 1st section of prospect alignment section until scanning till pixel value is the pixel of 0, is that the row-coordinate of the pixel of 0 is designated as y by this pixel value 2', if the length y of the 1st section of prospect alignment section 2'-y 1' be greater than 0.005H, then determine that the 1st section of prospect alignment section is effective prospect alignment section, and make CLQ col'=CLQ col'+1, if the length y of the 1st section of prostatitis alignment section 2'-y 1' be less than or equal to 0.005H, then determine that the 1st section of prospect alignment section is invalid prospect alignment section; Continue scanning downwards, to determine whether the 1st section of prospect alignment section is the mode that effective prospect alignment section is identical with acquisition the 1st section of prospect alignment section, obtain col capable in all effective prospect alignment section, and statistics obtains the quantity CLQ of effective prospect alignment section in col row col'; Wherein, the initial value of col is 1,1≤col≤W, 1≤y 1' <y 2'≤H, CLQ col'=CLQ col"=" in '+1 is assignment;
B4, make RLQ row=RLQ row', make CLQ col=CLQ col'.
Described step 4. in get P=1.6, T f=2.0, T r=2, T c=1.5.
Compared with prior art, the invention has the advantages that: foreground target and the depth location residing for background region in the natural scene of the inventive method according to three-dimensional display, determine 10 kinds of scene modes of stereo-picture, this can fully demonstrate the degree of depth factor of object in stereo-picture and the non-linear relation of visual comfort, thus the visual comfort evaluation model set up respectively based on 10 kinds of scene modes can reflect the perception of the visual comfort of human visual system's stereoscopic image more accurately, thus the consistency that can effectively improve between objective evaluation result and human eye subjective perception.
Accompanying drawing explanation
Fig. 1 be the inventive method totally realize block diagram;
Fig. 2 is the geometric representation of f, b, k, p, h, F, B;
Fig. 3 is W ' ageometric representation;
Fig. 4 a is the weight value under the 4th kind of scene mode belongs to Linear Fit Chart corresponding to all stereo-pictures of the 4th kind of scene mode when being 0.1 in stereoscopic image data storehouse, abscissa is global disparity angle, and ordinate is the objective comfortable angle value of parallax linear effect;
Fig. 4 b is the weight value under the 4th kind of scene mode belongs to Linear Fit Chart corresponding to all stereo-pictures of the 4th kind of scene mode when being 0.2 in stereoscopic image data storehouse, abscissa is global disparity angle, and ordinate is the objective comfortable angle value of parallax linear effect;
Fig. 4 c is the weight value under the 4th kind of scene mode belongs to Linear Fit Chart corresponding to all stereo-pictures of the 4th kind of scene mode when being 0.3 in stereoscopic image data storehouse, abscissa is global disparity angle, and ordinate is the objective comfortable angle value of parallax linear effect;
Fig. 4 d is the weight value under the 4th kind of scene mode belongs to Linear Fit Chart corresponding to all stereo-pictures of the 4th kind of scene mode when being 0.4 in stereoscopic image data storehouse, abscissa is global disparity angle, and ordinate is the objective comfortable angle value of parallax linear effect;
Fig. 4 e is the weight value under the 4th kind of scene mode belongs to Linear Fit Chart corresponding to all stereo-pictures of the 4th kind of scene mode when being 0.5 in stereoscopic image data storehouse, abscissa is global disparity angle, and ordinate is the objective comfortable angle value of parallax linear effect;
Fig. 4 f is the weight value under the 4th kind of scene mode belongs to Linear Fit Chart corresponding to all stereo-pictures of the 4th kind of scene mode when being 0.6 in stereoscopic image data storehouse, abscissa is global disparity angle, and ordinate is the objective comfortable angle value of parallax linear effect;
Fig. 4 g is the weight value under the 4th kind of scene mode belongs to Linear Fit Chart corresponding to all stereo-pictures of the 4th kind of scene mode when being 0.7 in stereoscopic image data storehouse, abscissa is global disparity angle, and ordinate is the objective comfortable angle value of parallax linear effect;
Fig. 4 h is the weight value under the 4th kind of scene mode belongs to Linear Fit Chart corresponding to all stereo-pictures of the 4th kind of scene mode when being 0.8 in stereoscopic image data storehouse, abscissa is global disparity angle, and ordinate is the objective comfortable angle value of parallax linear effect;
Fig. 4 i is the weight value under the 4th kind of scene mode belongs to Linear Fit Chart corresponding to all stereo-pictures of the 4th kind of scene mode when being 0.9 in stereoscopic image data storehouse, abscissa is global disparity angle, and ordinate is the objective comfortable angle value of parallax linear effect;
Fig. 4 j is the weight value under the 4th kind of scene mode belongs to Linear Fit Chart corresponding to all stereo-pictures of the 4th kind of scene mode when being 1.0 in stereoscopic image data storehouse, abscissa is global disparity angle, and ordinate is the objective comfortable angle value of parallax linear effect;
Fig. 5 is that the weight value under the 4th kind of scene mode belongs to fitting degree corresponding to all stereo-pictures of the 4th kind of scene mode when being respectively 0.1 ~ 1.0 (fitting degree is weighed with MAE in stereoscopic image data storehouse, when weight value is 0.7, MAE is minimum, and the weight value that namely best fit degree is corresponding is 0.7);
Fig. 6 a is the weight value under the 5th kind of scene mode belongs to Linear Fit Chart corresponding to all stereo-pictures of the 5th kind of scene mode when being 0.1 in stereoscopic image data storehouse, abscissa is global disparity angle, and ordinate is the objective comfortable angle value of parallax linear effect;
Fig. 6 b is the weight value under the 5th kind of scene mode belongs to Linear Fit Chart corresponding to all stereo-pictures of the 5th kind of scene mode when being 0.2 in stereoscopic image data storehouse, abscissa is global disparity angle, and ordinate is the objective comfortable angle value of parallax linear effect;
Fig. 6 c is the weight value under the 5th kind of scene mode belongs to Linear Fit Chart corresponding to all stereo-pictures of the 5th kind of scene mode when being 0.3 in stereoscopic image data storehouse, abscissa is global disparity angle, and ordinate is the objective comfortable angle value of parallax linear effect;
Fig. 6 d is the weight value under the 5th kind of scene mode belongs to Linear Fit Chart corresponding to all stereo-pictures of the 5th kind of scene mode when being 0.4 in stereoscopic image data storehouse, abscissa is global disparity angle, and ordinate is the objective comfortable angle value of parallax linear effect;
Fig. 6 e is the weight value under the 5th kind of scene mode belongs to Linear Fit Chart corresponding to all stereo-pictures of the 5th kind of scene mode when being 0.5 in stereoscopic image data storehouse, abscissa is global disparity angle, and ordinate is the objective comfortable angle value of parallax linear effect;
Fig. 6 f is the weight value under the 5th kind of scene mode belongs to Linear Fit Chart corresponding to all stereo-pictures of the 5th kind of scene mode when being 0.6 in stereoscopic image data storehouse, abscissa is global disparity angle, and ordinate is the objective comfortable angle value of parallax linear effect;
Fig. 6 g is the weight value under the 5th kind of scene mode belongs to Linear Fit Chart corresponding to all stereo-pictures of the 5th kind of scene mode when being 0.7 in stereoscopic image data storehouse, abscissa is global disparity angle, and ordinate is the objective comfortable angle value of parallax linear effect;
Fig. 6 h is the weight value under the 5th kind of scene mode belongs to Linear Fit Chart corresponding to all stereo-pictures of the 5th kind of scene mode when being 0.8 in stereoscopic image data storehouse, abscissa is global disparity angle, and ordinate is the objective comfortable angle value of parallax linear effect;
Fig. 6 i is the weight value under the 5th kind of scene mode belongs to Linear Fit Chart corresponding to all stereo-pictures of the 5th kind of scene mode when being 0.9 in stereoscopic image data storehouse, abscissa is global disparity angle, and ordinate is the objective comfortable angle value of parallax linear effect;
Fig. 6 j is the weight value under the 5th kind of scene mode belongs to Linear Fit Chart corresponding to all stereo-pictures of the 5th kind of scene mode when being 1.0 in stereoscopic image data storehouse, abscissa is global disparity angle, and ordinate is the objective comfortable angle value of parallax linear effect;
Fig. 7 is that the weight value under the 5th kind of scene mode belongs to fitting degree corresponding to all stereo-pictures of the 5th kind of scene mode when being respectively 0.1 ~ 1.0 (fitting degree is weighed with MAE in stereoscopic image data storehouse, when weight value is 0.6, MAE is minimum, and the weight value that namely best fit degree is corresponding is 0.6);
Fig. 8 a is the weight value under the 8th kind of scene mode belongs to Linear Fit Chart corresponding to all stereo-pictures of the 8th kind of scene mode when being 0.1 in stereoscopic image data storehouse, abscissa is global disparity angle, and ordinate is the objective comfortable angle value of parallax linear effect;
Fig. 8 b is the weight value under the 8th kind of scene mode belongs to Linear Fit Chart corresponding to all stereo-pictures of the 8th kind of scene mode when being 0.2 in stereoscopic image data storehouse, abscissa is global disparity angle, and ordinate is the objective comfortable angle value of parallax linear effect;
Fig. 8 c is the weight value under the 8th kind of scene mode belongs to Linear Fit Chart corresponding to all stereo-pictures of the 8th kind of scene mode when being 0.3 in stereoscopic image data storehouse, abscissa is global disparity angle, and ordinate is the objective comfortable angle value of parallax linear effect;
Fig. 8 d is the weight value under the 8th kind of scene mode belongs to Linear Fit Chart corresponding to all stereo-pictures of the 8th kind of scene mode when being 0.4 in stereoscopic image data storehouse, abscissa is global disparity angle, and ordinate is the objective comfortable angle value of parallax linear effect;
Fig. 8 e is the weight value under the 8th kind of scene mode belongs to Linear Fit Chart corresponding to all stereo-pictures of the 8th kind of scene mode when being 0.5 in stereoscopic image data storehouse, abscissa is global disparity angle, and ordinate is the objective comfortable angle value of parallax linear effect;
Fig. 8 f is the weight value under the 8th kind of scene mode belongs to Linear Fit Chart corresponding to all stereo-pictures of the 8th kind of scene mode when being 0.6 in stereoscopic image data storehouse, abscissa is global disparity angle, and ordinate is the objective comfortable angle value of parallax linear effect;
Fig. 8 g is the weight value under the 8th kind of scene mode belongs to Linear Fit Chart corresponding to all stereo-pictures of the 8th kind of scene mode when being 0.7 in stereoscopic image data storehouse, abscissa is global disparity angle, and ordinate is the objective comfortable angle value of parallax linear effect;
Fig. 8 h is the weight value under the 8th kind of scene mode belongs to Linear Fit Chart corresponding to all stereo-pictures of the 8th kind of scene mode when being 0.8 in stereoscopic image data storehouse, abscissa is global disparity angle, and ordinate is the objective comfortable angle value of parallax linear effect;
Fig. 8 i is the weight value under the 8th kind of scene mode belongs to Linear Fit Chart corresponding to all stereo-pictures of the 8th kind of scene mode when being 0.9 in stereoscopic image data storehouse, abscissa is global disparity angle, and ordinate is the objective comfortable angle value of parallax linear effect;
Fig. 8 j is the weight value under the 8th kind of scene mode belongs to Linear Fit Chart corresponding to all stereo-pictures of the 8th kind of scene mode when being 1.0 in stereoscopic image data storehouse, abscissa is global disparity angle, and ordinate is the objective comfortable angle value of parallax linear effect;
Fig. 9 is that the weight value under the 8th kind of scene mode belongs to fitting degree corresponding to all stereo-pictures of the 8th kind of scene mode when being respectively 0.1 ~ 1.0 (fitting degree is weighed with MAE in stereoscopic image data storehouse, when weight value is 0.6, MAE is minimum, and the weight value that namely best fit degree is corresponding is 0.6);
Figure 10 a is the weight value under the 9th kind of scene mode belongs to Linear Fit Chart corresponding to all stereo-pictures of the 9th kind of scene mode when being 0.1 in stereoscopic image data storehouse, abscissa is global disparity angle, and ordinate is the objective comfortable angle value of parallax linear effect;
Figure 10 b is the weight value under the 9th kind of scene mode belongs to Linear Fit Chart corresponding to all stereo-pictures of the 9th kind of scene mode when being 0.2 in stereoscopic image data storehouse, abscissa is global disparity angle, and ordinate is the objective comfortable angle value of parallax linear effect;
Figure 10 c is the weight value under the 9th kind of scene mode belongs to Linear Fit Chart corresponding to all stereo-pictures of the 9th kind of scene mode when being 0.3 in stereoscopic image data storehouse, abscissa is global disparity angle, and ordinate is the objective comfortable angle value of parallax linear effect;
Figure 10 d is the weight value under the 9th kind of scene mode belongs to Linear Fit Chart corresponding to all stereo-pictures of the 9th kind of scene mode when being 0.4 in stereoscopic image data storehouse, abscissa is global disparity angle, and ordinate is the objective comfortable angle value of parallax linear effect;
Figure 10 e is the weight value under the 9th kind of scene mode belongs to Linear Fit Chart corresponding to all stereo-pictures of the 9th kind of scene mode when being 0.5 in stereoscopic image data storehouse, abscissa is global disparity angle, and ordinate is the objective comfortable angle value of parallax linear effect;
Figure 10 f is the weight value under the 9th kind of scene mode belongs to Linear Fit Chart corresponding to all stereo-pictures of the 9th kind of scene mode when being 0.6 in stereoscopic image data storehouse, abscissa is global disparity angle, and ordinate is the objective comfortable angle value of parallax linear effect;
Figure 10 g is the weight value under the 9th kind of scene mode belongs to Linear Fit Chart corresponding to all stereo-pictures of the 9th kind of scene mode when being 0.7 in stereoscopic image data storehouse, abscissa is global disparity angle, and ordinate is the objective comfortable angle value of parallax linear effect;
Figure 10 h is the weight value under the 9th kind of scene mode belongs to Linear Fit Chart corresponding to all stereo-pictures of the 9th kind of scene mode when being 0.8 in stereoscopic image data storehouse, abscissa is global disparity angle, and ordinate is the objective comfortable angle value of parallax linear effect;
Figure 10 i is the weight value under the 9th kind of scene mode belongs to Linear Fit Chart corresponding to all stereo-pictures of the 9th kind of scene mode when being 0.9 in stereoscopic image data storehouse, abscissa is global disparity angle, and ordinate is the objective comfortable angle value of parallax linear effect;
Figure 10 j is the weight value under the 9th kind of scene mode belongs to Linear Fit Chart corresponding to all stereo-pictures of the 9th kind of scene mode when being 1.0 in stereoscopic image data storehouse, abscissa is global disparity angle, and ordinate is the objective comfortable angle value of parallax linear effect;
Figure 11 is that the weight value under the 9th kind of scene mode belongs to fitting degree corresponding to all stereo-pictures of the 9th kind of scene mode when being respectively 0.1 ~ 1.0 (fitting degree is weighed with MAE in stereoscopic image data storehouse, when weight value is 0.5, MAE is minimum, and the weight value that namely best fit degree is corresponding is 0.5).
Embodiment
Below in conjunction with accompanying drawing embodiment, the present invention is described in further detail.
A kind of stereo image vision comfort level evaluation method based on scene mode classification that the present invention proposes, it totally realizes block diagram as shown in Figure 1, and it comprises the following steps:
1. determine 10 kinds of scene modes of the natural scene shown by three-dimensional display, the 1st kind of scene mode in screen, to be that the foreground target in the right viewpoint parallax gray level image of stereo-picture is convex be in that non-comfort zone, background region are recessed is in non-comfort zone in screen; 2nd kind of scene mode in screen, to be that the foreground target in the right viewpoint parallax gray level image of stereo-picture is convex be in that non-comfort zone, background region are convex is in non-comfort zone in screen; 3rd kind of scene mode in screen, to be that the foreground target in the right viewpoint parallax gray level image of stereo-picture is recessed be in that non-comfort zone, background region are recessed is in non-comfort zone in screen; 4th kind of scene mode in screen, to be that the foreground target in the right viewpoint parallax gray level image of stereo-picture is convex be in that non-comfort zone, background region are recessed is in comfort zone in screen; 5th kind of scene mode in screen, to be that the foreground target in the right viewpoint parallax gray level image of stereo-picture is convex be in that non-comfort zone, background region are convex is in comfort zone in screen; 6th kind of scene mode in screen, to be that the foreground target in the right viewpoint parallax gray level image of stereo-picture is convex be in that comfort zone, background region are recessed is in non-comfort zone in screen; 7th kind of scene mode in screen, to be that the foreground target in the right viewpoint parallax gray level image of stereo-picture is recessed be in that comfort zone, background region are recessed is in non-comfort zone in screen; 8th kind of scene mode in screen, to be that the foreground target in the right viewpoint parallax gray level image of stereo-picture is convex be in that comfort zone, background region are recessed is in comfort zone in screen; 9th kind of scene mode in screen, to be that the foreground target in the right viewpoint parallax gray level image of stereo-picture is convex be in that comfort zone, background region are convex is in comfort zone in screen; 10th kind of scene mode in screen, to be that the foreground target in the right viewpoint parallax gray level image of stereo-picture is recessed be in that comfort zone, background region are recessed is in comfort zone in screen.
2. every width stereo-picture is selected to have a stereoscopic image data storehouse of the mean subjective scoring average of visual comfort; Then according to the parallactic angle of foreground target in the right viewpoint parallax gray level image of the every width stereo-picture in stereoscopic image data storehouse and the parallactic angle of background region, the scene mode belonging to every width stereo-picture in stereoscopic image data storehouse is determined; Again according to the parallactic angle of foreground target in the respective right viewpoint parallax gray level image of all stereo-pictures belonging to often kind of scene mode in stereoscopic image data storehouse and the parallactic angle of background region, set up the visual comfort evaluation model under often kind of scene mode, the visual comfort evaluation model under n-th kind of scene mode is described below: SMMO n=(4.2028-u (Q n))-v (Q n) × D a(Q n)+0.1912 × ln (W a)-0.0208 × D a(Q n) × ln (W a), wherein, 1≤n≤10, SMMO nrepresent the output of the visual comfort evaluation model under n-th kind of scene mode, u (Q n) and v (Q n) be constant, u (Q n) and v (Q n) value by the u (q in step 2.-4 n) and v (q n) determine, D a(Q n) represent that a width to be entered belongs to the global disparity angle of the right viewpoint parallax gray level image of the stereo-picture to be evaluated of n-th kind of scene mode, D a(Q n)=Q n× | f a|+(1-Q n) × | b a|, Q nrepresent the weight under n-th kind of scene mode, f arepresent that a width to be entered belongs to the parallactic angle of the foreground target in the right viewpoint parallax gray level image of the stereo-picture to be evaluated of n-th kind of scene mode, b arepresent that a width to be entered belongs to the parallactic angle of the background region in the right viewpoint parallax gray level image of the stereo-picture to be evaluated of n-th kind of scene mode, W arepresent that a width to be entered belongs to the width angle of the foreground target in the right viewpoint parallax gray level image of the stereo-picture to be evaluated of n-th kind of scene mode, symbol " || " is the symbol that takes absolute value.
In this particular embodiment, step 2. middle Q nacquisition process be:
-1 2., total width number of supposing the stereo-picture belonging to n-th kind of scene mode in stereoscopic image data storehouse is M', wherein, and M' >=1.
2.-2, q is made ninitial value be 0.1, make Q ninitial value be 0.
2. the weight value-3, calculated under n-th kind of scene mode is q ntime stereoscopic image data storehouse in belong to the global disparity angle of the right viewpoint parallax gray level image of every width stereo-picture of n-th kind of scene mode, be q by the weight value under n-th kind of scene mode ntime stereoscopic image data storehouse in belong to the right viewpoint parallax gray level image of the m' width stereo-picture of n-th kind of scene mode global disparity angle be designated as D m', a(q n), D m', a(q n)=q n× | f m', a|+(1-q n) × | b m', a|, wherein, 1≤m'≤M', f m', aand b m', abelong to the parallactic angle of foreground target in the right viewpoint parallax gray level image of the m' width stereo-picture of n-th kind of scene mode and the parallactic angle of background region in corresponding expression stereoscopic image data storehouse, symbol " || " is the symbol that takes absolute value.
2. the weight value-4, calculated under n-th kind of scene mode is q ntime stereoscopic image data storehouse in belong to n-th kind of scene mode every width stereo-picture eliminate the comfort level objective evaluation value of parallax linear effect, be q by the weight value under n-th kind of scene mode ntime stereoscopic image data storehouse in belong to n-th kind of scene mode the m' width stereo-picture comfort level objective evaluation value that eliminates parallax linear effect be designated as VC h(D m', a(q n), W a), VC h(D m', a(q n), W a)=4.2028+0.1912 × ln (W a)-0.0208 × D m', a(q n) × ln (W a); Then the weight value calculated under n-th kind of scene mode is q ntime stereoscopic image data storehouse in belong to the comfort level objective evaluation value of the parallax linear effect of every width stereo-picture of n-th kind of scene mode, be q by the weight value under n-th kind of scene mode ntime stereoscopic image data storehouse in belong to the parallax linear effect of the m' width stereo-picture of n-th kind of scene mode comfort level objective evaluation value be designated as ERR m', ERR m'=VC h(D m', a(q n), W a)-MOS m', wherein, MOS m'represent that the weight value under n-th kind of scene mode is q ntime stereoscopic image data storehouse in belong to the comfort level subjective assessment value of the m' width stereo-picture of n-th kind of scene mode, MOS m'value known; Adopting existing least square method again, is q to the weight value under n-th kind of scene mode ntime stereoscopic image data storehouse in belong to n-th kind of scene mode the respective global disparity angle of right viewpoint parallax gray level image of all stereo-pictures and the comfort level objective evaluation value of respective parallax linear effect carry out linear fit, the weight value obtained under n-th kind of scene mode is q ntime stereoscopic image data storehouse in belong to a fitting a straight line equation corresponding to all stereo-pictures of n-th kind of scene mode: ERR=u (q n)+v (q n) × D a(q n), wherein, ERR is for representing the dependent variable of parallax to the linear effect of objective comfort level, u (q n) and v (q n) be ERR=u (q n)+v (q n) × D a(q n) middle constant, u (q n) and v (q n) value be matching result out, D a(q n) be for representing that the weight value under n-th kind of scene mode is q ntime stereoscopic image data storehouse in belong to the independent variable at the global disparity angle of the right viewpoint parallax gray level image of arbitrary width stereo-picture of n-th kind of scene mode.
2. be-5, q according to the weight value under n-th kind of scene mode ntime stereoscopic image data storehouse in belong to a fitting a straight line equation corresponding to all stereo-pictures of n-th kind of scene mode, the weight value obtained under n-th kind of scene mode is q ntime stereoscopic image data storehouse in belong to the comfort level objective evaluation equation of all stereo-pictures of n-th kind of scene mode: SMMO n(q n)=VC h(D a(q n), W a)-ERR=(4.2028-u (q n))-v (q n) × D a(q n)+0.1912 × ln (W a)-0.0208 × D a(q n) × ln (W a), wherein, SMMO n(q n) for having superposed the comfort level objective evaluation value of parallax linear effect and parallax non-linear effects, VC h(D a(q n), W a) represent that the weight value under n-th kind of scene mode is q ntime stereoscopic image data storehouse in belong to n-th kind of scene mode arbitrary width stereo-picture eliminate the comfort level objective evaluation value of parallax linear effect; Then the weight value adopting existing average absolute value error (MAE) to measure under n-th kind of scene mode is q ntime stereoscopic image data storehouse in belong to the comfort level objective evaluation value of all stereo-pictures of n-th kind of scene mode and the fitting degree of corresponding comfort level subjective assessment value.
2.-6, q is made n=q n+ 0.1, then return step 2.-3 continuation execution, until q nterminate when equaling 1.1, corresponding fitting degree when the weight value obtained under n-th kind of scene mode is respectively 0.1,0.2,0.3,0.4,0.5,0.6,0.7,0.8,0.9,1.0, then by weight value assignment corresponding for best fit degree to Q n, wherein, q n=q n"=" in+0.1 is assignment.
In this particular embodiment, step 2. middle u (Q n) value be Monomial coefficient in fitting a straight line equation corresponding to best fit degree, v (Q n) value be constant in fitting a straight line equation corresponding to best fit degree.
At this, for the stereoscopic image data storehouse that Korea Advanced Institute of Science and Technology image and video system laboratory (IVYLAB) provide, this stereoscopic image data storehouse comprises the right viewpoint parallax gray level image of 120 width stereo-pictures and correspondence, this stereoscopic image data storehouse contains the indoor and outdoors stereo-picture of various scene depth, and gives the mean subjective scoring average of the visual comfort of every width stereo-picture, calculate the parallactic angle of foreground target in the right viewpoint parallax gray level image of the every width stereo-picture in this stereoscopic image data storehouse and the parallactic angle of background region, obtain foreground target and background region separately relative to the concavity and convexity of screen, and whether be positioned at the characteristic in euphorosia district, can find that the stereo-picture in this stereoscopic image data storehouse is only limitted to the 4th kind, 5th kind, 8th kind and the 9th kind of scene mode, therefore with the 4th kind, 5th kind, 8th kind and the 9th kind of scene mode are that example sets up the 4th kind, 5th kind, visual comfort evaluation model under 8th kind and the 9th kind of scene mode.
The Linear Fit Chart that all stereo-pictures of the 4th kind of scene mode are corresponding is belonged in stereoscopic image data storehouse when Fig. 4 a to Fig. 4 j correspondence weight value given under the 4th kind of scene mode is respectively 0.1 ~ 1.0, (fitting degree is weighed with MAE to belong to fitting degree corresponding to all stereo-pictures of the 4th kind of scene mode when Fig. 5 weight value given under the 4th kind of scene mode is respectively 0.1 ~ 1.0 in stereoscopic image data storehouse, when weight value is 0.7, MAE is minimum, and the weight value that namely best fit degree is corresponding is 0.7); The Linear Fit Chart that all stereo-pictures of the 5th kind of scene mode are corresponding is belonged in stereoscopic image data storehouse when Fig. 6 a to Fig. 6 j correspondence weight value given under the 5th kind of scene mode is respectively 0.1 ~ 1.0, (fitting degree is weighed with MAE to belong to fitting degree corresponding to all stereo-pictures of the 5th kind of scene mode when Fig. 7 weight value given under the 5th kind of scene mode is respectively 0.1 ~ 1.0 in stereoscopic image data storehouse, when weight value is 0.6, MAE is minimum, and the weight value that namely best fit degree is corresponding is 0.6); The Linear Fit Chart that all stereo-pictures of the 8th kind of scene mode are corresponding is belonged in stereoscopic image data storehouse when Fig. 8 a to Fig. 8 j correspondence weight value given under the 8th kind of scene mode is respectively 0.1 ~ 1.0, (fitting degree is weighed with MAE to belong to fitting degree corresponding to all stereo-pictures of the 8th kind of scene mode when Fig. 9 weight value given under the 8th kind of scene mode is respectively 0.1 ~ 1.0 in stereoscopic image data storehouse, when weight value is 0.6, MAE is minimum, and the weight value that namely best fit degree is corresponding is 0.6); The Linear Fit Chart that all stereo-pictures of the 9th kind of scene mode are corresponding is belonged in stereoscopic image data storehouse when Figure 10 a to Figure 10 j correspondence weight value given under the 9th kind of scene mode is respectively 0.1 ~ 1.0, (fitting degree is weighed with MAE to belong to fitting degree corresponding to all stereo-pictures of the 9th kind of scene mode when Figure 11 weight value given under the 9th kind of scene mode is respectively 0.1 ~ 1.0 in stereoscopic image data storehouse, when weight value is 0.5, MAE is minimum, and the weight value that namely best fit degree is corresponding is 0.5).
Be respectively according to the above weight calculated under the 4th kind, the 5th kind, the 8th kind and the 9th kind of scene mode: Q 4=0.7, Q 5=0.6, Q 8=0.6, Q 9visual comfort evaluation model under=0.5,4th kind, the 5th kind, the 8th kind and the 9th kind of scene mode is respectively:
SMMO 4=4.3938-0.6652×D a(Q 4)+0.1912×ln(W a)-0.0208×D a(Q 4)×ln(W a)
SMMO 5=4.2326-0.7120×D a(Q 5)+0.1912×ln(W a)-0.0208×D a(Q 5)×ln(W a)
SMMO 8=4.5232-0.8918×D a(Q 8)+0.1912×ln(W a)-0.0208×D a(Q 8)×ln(W a)
SMMO 9=5.4614-2.6905×D a(Q 9)+0.1912×ln(W a)-0.0208×D a(Q 9)×ln(W a)
3. the left viewpoint coloured image of stereo-picture to be evaluated is designated as { I l(x, y) }, the right viewpoint coloured image of stereo-picture to be evaluated is designated as { I r(x, y) }, the right viewpoint parallax gray level image of stereo-picture to be evaluated is designated as { d r(x, y) }, wherein, 1≤x≤W, 1≤y≤H, W represents the width of stereo-picture to be evaluated, it is consistent with the width of the every width stereo-picture in stereoscopic image data storehouse, H represents the height of stereo-picture to be evaluated, and it is consistent with the height of the every width stereo-picture in stereoscopic image data storehouse, I l(x, y) represents { I l(x, y) } in coordinate position be the pixel value of the pixel of (x, y), I r(x, y) represents { I r(x, y) } in coordinate position be the pixel value of the pixel of (x, y), d r(x, y) represents { d r(x, y) } in coordinate position be the pixel value of the pixel of (x, y); Then adopt and step 2. in identical mode, determine the scene mode belonging to stereo-picture to be evaluated; Scene mode again belonging to stereo-picture to be evaluated, chooses the visual comfort evaluation model under this scene mode; Then according to the visual comfort evaluation model under this scene mode and { d r(x, y) } in the parallactic angle, { d of foreground target r(x, y) } in the parallactic angle of background region and { d r(x, y) } in the width angle of foreground target, calculate the visual comfort evaluation and foreca value of stereo-picture to be evaluated, be designated as smmo, suppose that stereo-picture to be evaluated belongs to n-th kind of scene mode, then smmo=(4.2028-u (Q n))-v (Q n) × D a' (Q n)+0.1912 × ln (W a')-0.0208 × D a' (Q n) × ln (W a'), wherein, D a' (Q n) represent { d r(x, y) } global disparity angle, D a' (Q n)=Q n× | f a' |+(1-Q n) × | b a' |, f a' represent { d r(x, y) } in the parallactic angle of foreground target, b a' represent { d r(x, y) } in the parallactic angle of background region, W a' represent { d r(x, y) } in the width angle of foreground target.
In this particular embodiment, step 3. in { d r(x, y) } in the width angle of foreground target wherein, arctan () is tan of negating, and h represents the distance of human eye to display, W frontrepresent { d r(x, y) } in the mean breadth of foreground target, W frontacquisition process be:
A1, to { d r(x, y) } carry out binary conversion treatment, obtain { d r(x, y) } binary image, be designated as { BI (x, y) }, the pixel value being the pixel of (x, y) by coordinate position in { BI (x, y) } is designated as BI (x, y), as { d r(x, y) } in coordinate position be the pixel of (x, y) when belonging to foreground target, make BI (x, y)=1, as { d r(x, y) } in coordinate position be the pixel of (x, y) when belonging to background region, make BI (x, y)=0.
A2, to { BI (x, y) } line scanning is carried out, for { BI (x, y) row } is capable, scan to the right from the 1st pixel that row is capable, when to scan the 1st pixel value be the pixel of 1, using the original position of this pixel as the 1st section of prospect width line segment, and the row coordinate of this pixel is designated as x 1, continuing scanning to the right and obtain the 1st section of prospect width line segment until scanning till pixel value is the pixel of 0, is that the row coordinate of the pixel of 0 is designated as x by this pixel value 2, the length of the 1st section of prospect width line segment is designated as WL 1, WL 1=x 2-x 1; Continue scan to the right, in the mode identical with acquisition the 1st section of prospect width line segment, obtain row capable in all prospect width line segments; Wherein, the initial value of row is 1,1≤row≤H, 1≤x 1<x 2≤ W.
A3, remove { BI (x, y) the prospect width line segment of length exception }, namely prospect width line segment short especially and prospect width line segment long is especially removed, be about to length in { BI (x, y) } be less than all prospect width line segments that all prospect width line segments of 0.002W and length is greater than 0.995W and remove; Then sort to remaining all prospect width line segments by length order from small to large, the prospect width line segment getting middle 80% forms prospect width line segment aggregate, is designated as { WL' n, wherein, WL' nrepresent { WL' nin the length of n-th section of prospect width line segment, 1≤n≤N', N' represents { WL' nin total hop count of prospect width line segment of comprising.
A4, calculating { WL' nin the mean value of length of all prospect width line segments, be designated as W front', then by W front' as { d r(x, y) } in the mean breadth W of foreground target fronteven, W front=W front'.
Fig. 3 gives W a' geometric representation.
4. the visual comfort evaluation and foreca value smmo of stereo-picture to be evaluated is revised, the revised visual comfort evaluation and foreca value of stereo-picture to be evaluated is designated as smm, wherein, max () is for getting max function, and P represents the tortuosity attenuation coefficient of the visual comfort evaluation model under the scene mode belonging to stereo-picture to be evaluated, S r' represent { d r(x, y) } in the average prospect line hop count of foreground target, S c' represent { d r(x, y) } in the average prospect alignment hop count of foreground target, T frepresent the parallactic angle threshold value of setting, T rrepresent the prospect line hop count threshold value of setting, T crepresent the prospect alignment hop count threshold value of setting, get P=1.6, T in the present embodiment f=2.0, T r=2, T c=1.5.
In this particular embodiment, step 4. in { d r(x, y) } in the average prospect line hop count of foreground target S r &prime; = &Sigma; r o w = 1 H RLQ r o w &Sigma; r o w = 1 H &epsiv; ( RLQ r o w ) , { d r(x, y) } in the average prospect alignment hop count of foreground target S c &prime; = &Sigma; c o l = 1 W CLQ c o l &Sigma; c o l = 1 W &epsiv; ( CLQ c o l ) , Wherein, S r' and S c' be { d r(x, y) } in the tortuosity parameter of foreground target, 1≤row≤H, 1≤col≤W, RLQ rowrepresent { d r(x, y) } in row capable in the quantity of prospect line section, &epsiv; ( RLQ r o w ) = { 1 , RLQ r o w > 0 0 , RLQ r o w &le; 0 , CLQ colrepresent { d r(x, y) } in col row in the quantity of prospect alignment section, &epsiv; ( CLQ c o l ) = 1 , C L Q c o l > 0 0 , C L Q c o l &le; 0 .
Wherein, RLQ rowand CLQ colacquisition process be:
B1, to { d r(x, y) } carry out binary conversion treatment, obtain { d r(x, y) } binary image, be designated as { BI (x, y) }, the pixel value being the pixel of (x, y) by coordinate position in { BI (x, y) } is designated as BI (x, y), as { d r(x, y) } in coordinate position be the pixel of (x, y) when belonging to foreground target, make BI (x, y)=1, as { d r(x, y) } in coordinate position be the pixel of (x, y) when belonging to background region, make BI (x, y)=0.
B2, to { BI (x, y) 2 dilation operations of mathematical morphology, 4 erosion operations, 2 dilation operations } are carried out successively, obtain { BI'(x, y) }, wherein, BI'(x, y) represent { BI'(x, y) }, coordinate position is the pixel value of the pixel of (x, y).
The quantity of effective prospect line section in often row in b3, statistics { BI'(x, y) }, by capable for the row in { BI'(x, y) } effectively the quantity of prospect line section be designated as RLQ row', RLQ row' acquisition process be: b3-1, make RLQ row' initial value be 0; B3-2, from { BI'(x, y) the 1st pixel that the row } is capable starts to scan to the right, when to scan the 1st pixel value be the pixel of 1, using the original position of this pixel as the 1st section of prospect line section, and the row coordinate of this pixel is designated as x 1', continuing scanning to the right and obtain the 1st section of prospect line section until scanning till pixel value is the pixel of 0, is that the row coordinate of the pixel of 0 is designated as x by this pixel value 2', if the length x of the 1st section of prospect line section 2'-x 1' be greater than 0.005W, then determine that the 1st section of prospect line section is effective prospect line section, and make RLQ row'=RLQ row'+1, if the length x of the 1st section of prospect line section 2'-x 1' be less than or equal to 0.005W, then determine that the 1st section of prospect line section is invalid prospect line section; Continue to scan to the right, to determine whether the 1st section of prospect line section is the mode that effective prospect line section is identical with acquisition the 1st section of prospect line section, obtain row capable in all effective prospect line section, and statistics obtain row capable in the amount R LQ of effective prospect line section row'; Wherein, the initial value of row is 1,1≤row≤H, 1≤x 1' <x 2'≤W, RLQ row'=RLQ row"=" in '+1 is assignment.
Equally, the quantity of effective prospect alignment section in the often row in statistics { BI'(x, y) }, by capable for the col in { BI'(x, y) } effectively the quantity of prospect alignment section be designated as CLQ col', CLQ col' acquisition process be: b3-1), make CLQ col' initial value be 0; B3-2), from { BI'(x, 1st pixel of the col row y) } starts downward scanning, when to scan the 1st pixel value be the pixel of 1, using the original position of this pixel as the 1st section of prospect alignment section, and the row-coordinate of this pixel is designated as y 1', continuing scanning downwards and obtain the 1st section of prospect alignment section until scanning till pixel value is the pixel of 0, is that the row-coordinate of the pixel of 0 is designated as y by this pixel value 2', if the length y of the 1st section of prospect alignment section 2'-y 1' be greater than 0.005H, then determine that the 1st section of prospect alignment section is effective prospect alignment section, and make CLQ col'=CLQ col'+1, if the length y of the 1st section of prostatitis alignment section 2'-y 1' be less than or equal to 0.005H, then determine that the 1st section of prospect alignment section is invalid prospect alignment section; Continue scanning downwards, to determine whether the 1st section of prospect alignment section is the mode that effective prospect alignment section is identical with acquisition the 1st section of prospect alignment section, obtain col capable in all effective prospect alignment section, and statistics obtains the quantity CLQ of effective prospect alignment section in col row col'; Wherein, the initial value of col is 1,1≤col≤W, 1≤y 1' <y 2'≤H, CLQ col'=CLQ col"=" in '+1 is assignment.
B4, make RLQ row=RLQ row', make CLQ col=CLQ col'.
In above-mentioned, the obtain manner of the foreground target in the right viewpoint parallax gray level image of every width stereo-picture of step 2. in neutral body image data base and background region and step 3. in stereo-picture to be evaluated right viewpoint parallax gray level image in foreground target identical with the obtain manner of background region, using the every width stereo-picture in stereoscopic image data storehouse and stereo-picture to be evaluated all as pending stereo-picture, foreground target in the right viewpoint parallax gray level image of then pending stereo-picture and the acquisition process of background region are: the grey level histogram obtaining the right viewpoint parallax gray level image of pending stereo-picture, then existing maximum variance between clusters (OTSU) is adopted, and the grey level histogram of right viewpoint parallax gray level image according to pending stereo-picture, obtain the intensity slicing threshold value of the right viewpoint parallax gray level image of pending stereo-picture, be designated as P seg, again pixel value in the right viewpoint parallax gray level image of pending stereo-picture is more than or equal to P segpixel be defined as foreground pixel point, and pixel value in the right viewpoint parallax gray level image of pending stereo-picture is less than P segpixel be defined as background pixel, the last foreground target be made up of all foreground pixel points in the right viewpoint parallax gray level image of pending stereo-picture, forms the background region in the right viewpoint parallax gray level image of pending stereo-picture by all background pixels.
At this, the intensity slicing threshold value P of the right viewpoint parallax gray level image of pending stereo-picture segacquisition process be:
X1, set up intensity slicing threshold value find target function, be designated as T, T=w f× (μ-μ f) 2+ w b× (μ-μ b) 2, wherein, w frepresent that total number of the foreground pixel point in the right viewpoint parallax gray level image of pending stereo-picture accounts for the ratio of total number of the pixel in the right viewpoint parallax gray level image of pending stereo-picture, μ frepresent the average of the pixel value of all foreground pixel points in the right viewpoint parallax gray level image of pending stereo-picture, w brepresent that total number of the background pixel in the right viewpoint parallax gray level image of pending stereo-picture accounts for the ratio of total number of the pixel in the right viewpoint parallax gray level image of pending stereo-picture, μ brepresent the average of the pixel value of all background pixels in the right viewpoint parallax gray level image of pending stereo-picture, μ represents the average of the pixel value of all pixels in the right viewpoint parallax gray level image of pending stereo-picture, μ=w f× μ f+ w b× μ b, Hist dg () represents the grey level histogram { Hist of the right viewpoint parallax gray level image of pending stereo-picture d(g) } in pixel value be the number of the pixel of g, 0≤g min≤ g≤g max≤ 255, g minrepresent the grey level histogram { Hist of the right viewpoint parallax gray level image of pending stereo-picture d(g) } in the number of the pixel minimum pixel value that is greater than 0, g maxrepresent the grey level histogram { Hist of the right viewpoint parallax gray level image of pending stereo-picture d(g) } in the number of the pixel max pixel value that is greater than 0; Grey level histogram { the Hist of the right viewpoint parallax gray level image of pending stereo-picture d(g) } in abscissa represent the number that gray scale, the ordinate of 0 ~ 255 represent the number of pixel, each pillar represents the pixel of different grey-scale.
X2, at interval [g min, g max] interior traversal t, the value of t when making T maximum is defined as the intensity slicing threshold value P of the right viewpoint parallax gray level image of pending stereo-picture seg.
In above-mentioned, step 2. in determine to determine that the mode of the scene mode belonging to stereo-picture to be evaluated is identical during the mode of the scene mode belonging to every width stereo-picture in stereoscopic image data storehouse and step are 3., using the every width stereo-picture in stereoscopic image data storehouse and stereo-picture to be evaluated all as pending stereo-picture, then determine that the detailed process of the scene mode belonging to pending stereo-picture is:
Z1, for the foreground target in the right viewpoint parallax gray level image of pending stereo-picture, if f a *>1 °, then determine that foreground target is convex and be in non-comfort zone in screen; If 0 °≤f a *<1 °, then to determine that foreground target is convex and be in comfort zone in screen; If-1 ° of <f a *<0 °, then to determine that foreground target is recessed and be in comfort zone in screen; If f a *<-1 °, then determine that foreground target is recessed and be in non-comfort zone in screen.
For the background region in the right viewpoint parallax gray level image of pending stereo-picture, if b a *>1 °, then determine that background region is convex and be in non-comfort zone in screen; If 0 °≤b a *<1 °, then to determine that background region is convex and be in comfort zone in screen; If-1 ° of <b a *<0 °, then to determine that background region is recessed and be in comfort zone in screen; If b a *<-1 °, then determine that background region is recessed and be in non-comfort zone in screen.
Wherein, f a *represent the parallactic angle of the foreground target in the right viewpoint parallax gray level image of pending stereo-picture, f a *=f-k, b a *represent the parallactic angle of the background region in the right viewpoint parallax gray level image of pending stereo-picture, b a *=b-k, f represent the convergent angle of human eye binocular viewing foreground target, b represents the convergent angle of human eye binocular viewing background region, k represents the adjustment angle of human eye binocular, arctan () is tan of negating, p represents the interpupillary distance of human eye binocular, L represents the width of display, N represents the horizontal resolution of display, h represents the distance of human eye to display, F represents the mean parallax amplitude of the foreground target in the right viewpoint parallax gray level image of pending stereo-picture, and F is in units of pixel F = &Sigma; x = 1 W &Sigma; y = 1 H &epsiv; ( d R * ( x , y ) - P s e g ) &times; d R * ( x , y ) &Sigma; x = 1 W &Sigma; y = 1 H &epsiv; ( d R * ( x , y ) - P s e g ) , &epsiv; ( d R * ( x , y ) - P s e g ) = 1 , d R * ( x , y ) - P s e g &GreaterEqual; 0 0 , d R * ( x , y ) - P s e g < 0 , B represents the mean parallax amplitude of the background region in the right viewpoint parallax gray level image of pending stereo-picture, B in units of pixel, B = &Sigma; x = 1 W &Sigma; y = 1 H &epsiv; ( P s e g - d R * ( x , y ) ) &times; d R * ( x , y ) &Sigma; x = 1 W &Sigma; y = 1 H &epsiv; ( P s e g - d R * ( x , y ) ) , &epsiv; ( P s e g - d R * ( x , y ) ) = 1 , P s e g - d R * ( x , y ) &GreaterEqual; 0 0 , P s e g - d R * ( x , y ) < 0 , D r *(x, y) represents that in the right viewpoint parallax gray level image of pending stereo-picture, coordinate position is the pixel value of the pixel of (x, y), and Fig. 2 gives the geometric representation of f, b, k, p, h, F, B.
Z2, according to the foreground target in the right viewpoint parallax gray level image of pending stereo-picture relative to the concavity and convexity of screen and whether be in comfort zone, background region relative to screen concavity and convexity and whether be in comfort zone, determine the scene mode belonging to pending stereo-picture, be specially: if foreground target is convex be in that non-comfort zone, background region are recessed is in non-comfort zone, i.e. f in screen in screen a *>1 ° and b a *<-1 °, then determine that pending stereo-picture belongs to the 1st kind of scene mode; If foreground target is convex be in that non-comfort zone, background region are convex is in non-comfort zone, i.e. f in screen in screen a *>1 ° and b a *>1 °, then determine that pending stereo-picture belongs to the 2nd kind of scene mode; If foreground target is recessed be in that non-comfort zone, background region are recessed is in non-comfort zone, i.e. f in screen in screen a *<-1 ° and b a *<-1 °, then determine that pending stereo-picture belongs to the 3rd kind of scene mode; If foreground target is convex be in that non-comfort zone, background region are recessed is in comfort zone, i.e. f in screen in screen a *>1 ° and-1 °≤b a *<0 °, then determine that pending stereo-picture belongs to the 4th kind of scene mode; If foreground target is convex be in that non-comfort zone, background region are convex is in comfort zone, i.e. f in screen in screen a *>1 ° and 0 °≤b a *≤ 1 °, then determine that pending stereo-picture belongs to the 5th kind of scene mode; If foreground target is convex be in that comfort zone, background region are recessed is in non-comfort zone, i.e. 0 °≤f in screen in screen a *≤ 1 ° and b a *<-1 °, then determine that pending stereo-picture belongs to the 6th kind of scene mode; If foreground target is recessed be in that comfort zone, background region are recessed is in non-comfort zone, namely-1 °≤f in screen in screen a *<0 ° and b a *<-1 °, then determine that pending stereo-picture belongs to the 7th kind of scene mode; If foreground target is convex be in that comfort zone, background region are recessed is in comfort zone, i.e. 0 °≤f in screen in screen a *≤ 1 ° and-1 °≤b a *<0 °, then determine that pending stereo-picture belongs to the 8th kind of scene mode; If foreground target is convex be in that comfort zone, background region are convex is in comfort zone, i.e. 0 °≤f in screen in screen a *≤ 1 ° and 0 °≤b a *≤ 1 °, then determine that pending stereo-picture belongs to the 9th kind of scene mode; If foreground target is recessed be in that comfort zone, background region are recessed is in comfort zone, namely-1 °≤f in screen in screen a *<0 ° and-1 °≤b a *<0 °, then determine that pending stereo-picture belongs to the 10th kind of scene mode.
In the present embodiment, utilize 5 of evaluate image quality evaluating method conventional objective parameters as evaluation index, namely Pearson correlation coefficient (the Pearsonlinearcorrelationcoefficient under nonlinear regression condition, PLCC), Spearman coefficient correlation (Spearmanrankordercorrelationcoefficient, SROCC), Kendall coefficient correlation (Kendallrank-ordercorrelationcoefficient, KROCC), average absolute value error (MeanAbsoluteError, MAE), root-mean-square error (rootmeansquarederror, RMSE).PLCC reflects the correlation of objective evaluation predicted value and subjective assessment value; SROCC and KROCC reflects monotonicity, the consistency of objective evaluation predicted value and subjective assessment value; MAE and RMSE reflects the accuracy of objective evaluation predicted value.The inventive method is utilized to obtain 120 width stereo-pictures final visual comfort evaluation and foreca value separately, five parameter Logistic function nonlinear fittings are done to the final visual comfort evaluation and foreca value of this 120 width stereo-picture, PLCC, SROCC and KROCC value is higher, and the correlation that the evaluation result of the less explanation the inventive method of MAE and RMSE value and mean subjective are marked between average is better.Table 1 gives the evaluation index value of the inventive method, illustrates that the result of objective evaluation result and the human eye subjective perception utilizing the inventive method to obtain is more consistent, describes the validity of the inventive method well.
Table 1 evaluation index value
Evaluation index PLCC SROCC KROCC MAE RMSE
0.919 0.909 0.748 0.237 0.318

Claims (8)

1., based on a stereo image vision comfort level evaluation method for scene mode classification, it is characterized in that comprising the following steps:
1. determine 10 kinds of scene modes of the natural scene shown by three-dimensional display, the 1st kind of scene mode in screen, to be that the foreground target in the right viewpoint parallax gray level image of stereo-picture is convex be in that non-comfort zone, background region are recessed is in non-comfort zone in screen; 2nd kind of scene mode in screen, to be that the foreground target in the right viewpoint parallax gray level image of stereo-picture is convex be in that non-comfort zone, background region are convex is in non-comfort zone in screen; 3rd kind of scene mode in screen, to be that the foreground target in the right viewpoint parallax gray level image of stereo-picture is recessed be in that non-comfort zone, background region are recessed is in non-comfort zone in screen; 4th kind of scene mode in screen, to be that the foreground target in the right viewpoint parallax gray level image of stereo-picture is convex be in that non-comfort zone, background region are recessed is in comfort zone in screen; 5th kind of scene mode in screen, to be that the foreground target in the right viewpoint parallax gray level image of stereo-picture is convex be in that non-comfort zone, background region are convex is in comfort zone in screen; 6th kind of scene mode in screen, to be that the foreground target in the right viewpoint parallax gray level image of stereo-picture is convex be in that comfort zone, background region are recessed is in non-comfort zone in screen; 7th kind of scene mode in screen, to be that the foreground target in the right viewpoint parallax gray level image of stereo-picture is recessed be in that comfort zone, background region are recessed is in non-comfort zone in screen; 8th kind of scene mode in screen, to be that the foreground target in the right viewpoint parallax gray level image of stereo-picture is convex be in that comfort zone, background region are recessed is in comfort zone in screen; 9th kind of scene mode in screen, to be that the foreground target in the right viewpoint parallax gray level image of stereo-picture is convex be in that comfort zone, background region are convex is in comfort zone in screen; 10th kind of scene mode in screen, to be that the foreground target in the right viewpoint parallax gray level image of stereo-picture is recessed be in that comfort zone, background region are recessed is in comfort zone in screen;
2. every width stereo-picture is selected to have a stereoscopic image data storehouse of the mean subjective scoring average of visual comfort; Then according to the parallactic angle of foreground target in the right viewpoint parallax gray level image of the every width stereo-picture in stereoscopic image data storehouse and the parallactic angle of background region, the scene mode belonging to every width stereo-picture in stereoscopic image data storehouse is determined; Again according to the parallactic angle of foreground target in the respective right viewpoint parallax gray level image of all stereo-pictures belonging to often kind of scene mode in stereoscopic image data storehouse and the parallactic angle of background region, set up the visual comfort evaluation model under often kind of scene mode, the visual comfort evaluation model under n-th kind of scene mode is described below: SMMO n=(4.2028-u (Q n))-v (Q n) × D a(Q n)+0.1912 × ln (W a)-0.0208 × D a(Q n) × ln (W a), wherein, 1≤n≤10, SMMO nrepresent the output of the visual comfort evaluation model under n-th kind of scene mode, u (Q n) and v (Q n) be constant, D a(Q n) represent that a width to be entered belongs to the global disparity angle of the right viewpoint parallax gray level image of the stereo-picture to be evaluated of n-th kind of scene mode, D a(Q n)=Q n× | f a|+(1-Q n) × | b a|, Q nrepresent the weight under n-th kind of scene mode, f arepresent that a width to be entered belongs to the parallactic angle of the foreground target in the right viewpoint parallax gray level image of the stereo-picture to be evaluated of n-th kind of scene mode, b arepresent that a width to be entered belongs to the parallactic angle of the background region in the right viewpoint parallax gray level image of the stereo-picture to be evaluated of n-th kind of scene mode, W arepresent that a width to be entered belongs to the width angle of the foreground target in the right viewpoint parallax gray level image of the stereo-picture to be evaluated of n-th kind of scene mode, symbol " || " is the symbol that takes absolute value;
3. the right viewpoint parallax gray level image of stereo-picture to be evaluated is designated as { d r(x, y) }, wherein, 1≤x≤W, 1≤y≤H, W represents the width of stereo-picture to be evaluated, it is consistent with the width of the every width stereo-picture in stereoscopic image data storehouse, H represents the height of stereo-picture to be evaluated, and it is consistent with the height of the every width stereo-picture in stereoscopic image data storehouse, d r(x, y) represents { d r(x, y) } in coordinate position be the pixel value of the pixel of (x, y); Then adopt and step 2. in identical mode, determine the scene mode belonging to stereo-picture to be evaluated; Scene mode again belonging to stereo-picture to be evaluated, chooses the visual comfort evaluation model under this scene mode; Then according to the visual comfort evaluation model under this scene mode and { d r(x, y) } in the parallactic angle, { d of foreground target r(x, y) } in the parallactic angle of background region and { d r(x, y) } in the width angle of foreground target, calculate the visual comfort evaluation and foreca value of stereo-picture to be evaluated, be designated as smmo, suppose that stereo-picture to be evaluated belongs to n-th kind of scene mode, then smmo=(4.2028-u (Q n))-v (Q n) × D a' (Q n)+0.1912 × ln (W a')-0.0208 × D a' (Q n) × ln (W a'), wherein, D a' (Q n) represent { d r(x, y) } global disparity angle, D a' (Q n)=Q n× | f a' |+(1-Q n) × | b a' |, f a' represent { d r(x, y) } in the parallactic angle of foreground target, b a' represent { d r(x, y) } in the parallactic angle of background region, W a' represent { d r(x, y) } in the width angle of foreground target;
4. the visual comfort evaluation and foreca value smmo of stereo-picture to be evaluated is revised, the revised visual comfort evaluation and foreca value of stereo-picture to be evaluated is designated as smm, wherein, max () is for getting max function, and P represents the tortuosity attenuation coefficient of the visual comfort evaluation model under the scene mode belonging to stereo-picture to be evaluated, S r' represent { d r(x, y) } in the average prospect line hop count of foreground target, S c' represent { d r(x, y) } in the average prospect alignment hop count of foreground target, T frepresent the parallactic angle threshold value of setting, T rrepresent the prospect line hop count threshold value of setting, T crepresent the prospect alignment hop count threshold value of setting.
2. a kind of stereo image vision comfort level evaluation method based on scene mode classification according to claim 1, is characterized in that described step 2. middle Q nacquisition process be:
-1 2., total width number of supposing the stereo-picture belonging to n-th kind of scene mode in stereoscopic image data storehouse is M', wherein, and M' >=1;
2.-2, q is made ninitial value be 0.1, make Q ninitial value be 0;
2. the weight value-3, calculated under n-th kind of scene mode is q ntime stereoscopic image data storehouse in belong to the global disparity angle of the right viewpoint parallax gray level image of every width stereo-picture of n-th kind of scene mode, be q by the weight value under n-th kind of scene mode ntime stereoscopic image data storehouse in belong to the right viewpoint parallax gray level image of the m' width stereo-picture of n-th kind of scene mode global disparity angle be designated as D m', a(q n), D m', a(q n)=q n× | f m', a|+(1-q n) × | b m', a|, wherein, 1≤m'≤M', f m', aand b m', abelong to the parallactic angle of foreground target in the right viewpoint parallax gray level image of the m' width stereo-picture of n-th kind of scene mode and the parallactic angle of background region in corresponding expression stereoscopic image data storehouse, symbol " || " is the symbol that takes absolute value;
2. the weight value-4, calculated under n-th kind of scene mode is q ntime stereoscopic image data storehouse in belong to n-th kind of scene mode every width stereo-picture eliminate the comfort level objective evaluation value of parallax linear effect, be q by the weight value under n-th kind of scene mode ntime stereoscopic image data storehouse in belong to n-th kind of scene mode the m' width stereo-picture comfort level objective evaluation value that eliminates parallax linear effect be designated as VC h(D m', a(q n), W a), VC h(D m', a(q n), W a)=4.2028+0.1912 × ln (W a)-0.0208 × D m', a(q n) × ln (W a); Then the weight value calculated under n-th kind of scene mode is q ntime stereoscopic image data storehouse in belong to the comfort level objective evaluation value of the parallax linear effect of every width stereo-picture of n-th kind of scene mode, be q by the weight value under n-th kind of scene mode ntime stereoscopic image data storehouse in belong to the parallax linear effect of the m' width stereo-picture of n-th kind of scene mode comfort level objective evaluation value be designated as ERR m', ERR m'=VC h(D m', a(q n), W a)-MOS m', wherein, MOS m'represent that the weight value under n-th kind of scene mode is q ntime stereoscopic image data storehouse in belong to the comfort level subjective assessment value of the m' width stereo-picture of n-th kind of scene mode; Adopting least square method again, is q to the weight value under n-th kind of scene mode ntime stereoscopic image data storehouse in belong to n-th kind of scene mode the respective global disparity angle of right viewpoint parallax gray level image of all stereo-pictures and the comfort level objective evaluation value of respective parallax linear effect carry out linear fit, the weight value obtained under n-th kind of scene mode is q ntime stereoscopic image data storehouse in belong to a fitting a straight line equation corresponding to all stereo-pictures of n-th kind of scene mode: ERR=u (q n)+v (q n) × D a(q n), wherein, ERR is for representing the dependent variable of parallax to the linear effect of objective comfort level, u (q n) and v (q n) be ERR=u (q n)+v (q n) × D a(q n) in constant, D a(q n) be for representing that the weight value under n-th kind of scene mode is q ntime stereoscopic image data storehouse in belong to the independent variable at the global disparity angle of the right viewpoint parallax gray level image of arbitrary width stereo-picture of n-th kind of scene mode;
2. be-5, q according to the weight value under n-th kind of scene mode ntime stereoscopic image data storehouse in belong to a fitting a straight line equation corresponding to all stereo-pictures of n-th kind of scene mode, the weight value obtained under n-th kind of scene mode is q ntime stereoscopic image data storehouse in belong to the comfort level objective evaluation equation of all stereo-pictures of n-th kind of scene mode: SMMO n(q n)=VC h(D a(q n), W a)-ERR=(4.2028-u (q n))-v (q n) × D a(q n)+0.1912 × ln (W a)-0.0208 × D a(q n) × ln (W a), wherein, SMMO n(q n) for having superposed the comfort level objective evaluation value of parallax linear effect and parallax non-linear effects, VC h(D a(q n), W a) represent that the weight value under n-th kind of scene mode is q ntime stereoscopic image data storehouse in belong to n-th kind of scene mode arbitrary width stereo-picture eliminate the comfort level objective evaluation value of parallax linear effect; The weight value of then measuring under n-th kind of scene mode is q ntime stereoscopic image data storehouse in belong to the comfort level objective evaluation value of all stereo-pictures of n-th kind of scene mode and the fitting degree of corresponding comfort level subjective assessment value;
2.-6, q is made n=q n+ 0.1, then return step 2.-3 continuation execution, until q nterminate when equaling 1.1, corresponding fitting degree when the weight value obtained under n-th kind of scene mode is respectively 0.1,0.2,0.3,0.4,0.5,0.6,0.7,0.8,0.9,1.0, then by weight value assignment corresponding for best fit degree to Q n, wherein, q n=q n"=" in+0.1 is assignment;
Described step is middle u (Q 2. n) value be Monomial coefficient in fitting a straight line equation corresponding to best fit degree; Described step is middle v (Q 2. n) value be constant in fitting a straight line equation corresponding to best fit degree.
3. a kind of stereo image vision comfort level evaluation method based on scene mode classification according to claim 1 and 2, it is characterized in that the foreground target in the right viewpoint parallax gray level image of stereo-picture to be evaluated during the obtain manner of the foreground target in the right viewpoint parallax gray level image of every width stereo-picture of described step 2. in neutral body image data base and background region and described step are 3. is identical with the obtain manner of background region, using the every width stereo-picture in stereoscopic image data storehouse and stereo-picture to be evaluated all as pending stereo-picture, foreground target in the right viewpoint parallax gray level image of then pending stereo-picture and the acquisition process of background region are: the grey level histogram obtaining the right viewpoint parallax gray level image of pending stereo-picture, then maximum variance between clusters is adopted, and the grey level histogram of right viewpoint parallax gray level image according to pending stereo-picture, obtain the intensity slicing threshold value of the right viewpoint parallax gray level image of pending stereo-picture, be designated as P seg, again pixel value in the right viewpoint parallax gray level image of pending stereo-picture is more than or equal to P segpixel be defined as foreground pixel point, and pixel value in the right viewpoint parallax gray level image of pending stereo-picture is less than P segpixel be defined as background pixel, the last foreground target be made up of all foreground pixel points in the right viewpoint parallax gray level image of pending stereo-picture, forms the background region in the right viewpoint parallax gray level image of pending stereo-picture by all background pixels.
4. a kind of stereo image vision comfort level evaluation method based on scene mode classification according to claim 3, is characterized in that the intensity slicing threshold value P of the right viewpoint parallax gray level image of described pending stereo-picture segacquisition process be:
X1, set up intensity slicing threshold value find target function, be designated as T, T=w f× (μ-μ f) 2+ w b× (μ-μ b) 2, wherein, w frepresent that total number of the foreground pixel point in the right viewpoint parallax gray level image of pending stereo-picture accounts for the ratio of total number of the pixel in the right viewpoint parallax gray level image of pending stereo-picture, μ frepresent the average of the pixel value of all foreground pixel points in the right viewpoint parallax gray level image of pending stereo-picture, w brepresent that total number of the background pixel in the right viewpoint parallax gray level image of pending stereo-picture accounts for the ratio of total number of the pixel in the right viewpoint parallax gray level image of pending stereo-picture, μ brepresent the average of the pixel value of all background pixels in the right viewpoint parallax gray level image of pending stereo-picture, μ represents the average of the pixel value of all pixels in the right viewpoint parallax gray level image of pending stereo-picture, μ=w f× μ f+ w b× μ b, Hist dg () represents the grey level histogram { Hist of the right viewpoint parallax gray level image of pending stereo-picture d(g) } in pixel value be the number of the pixel of g, 0≤g min≤ g≤g max≤ 255, g minrepresent the grey level histogram { Hist of the right viewpoint parallax gray level image of pending stereo-picture d(g) } in the number of the pixel minimum pixel value that is greater than 0, g maxrepresent the grey level histogram { Hist of the right viewpoint parallax gray level image of pending stereo-picture d(g) } in the number of the pixel max pixel value that is greater than 0;
X2, at interval [g min, g max] interior traversal t, the value of t when making T maximum is defined as the intensity slicing threshold value P of the right viewpoint parallax gray level image of pending stereo-picture seg.
5. a kind of stereo image vision comfort level evaluation method based on scene mode classification according to claim 3, it is characterized in that determining during described step 2. to determine that the mode of the scene mode belonging to stereo-picture to be evaluated is identical during the mode of the scene mode belonging to every width stereo-picture in stereoscopic image data storehouse and described step are 3., using the every width stereo-picture in stereoscopic image data storehouse and stereo-picture to be evaluated all as pending stereo-picture, then determine that the detailed process of the scene mode belonging to pending stereo-picture is:
Z1, for the foreground target in the right viewpoint parallax gray level image of pending stereo-picture, if f a *>1 °, then determine that foreground target is convex and be in non-comfort zone in screen; If 0 °≤f a *<1 °, then to determine that foreground target is convex and be in comfort zone in screen; If-1 ° of <f a *<0 °, then to determine that foreground target is recessed and be in comfort zone in screen; If f a *<-1 °, then determine that foreground target is recessed and be in non-comfort zone in screen;
For the background region in the right viewpoint parallax gray level image of pending stereo-picture, if b a *>1 °, then determine that background region is convex and be in non-comfort zone in screen; If 0 °≤b a *<1 °, then to determine that background region is convex and be in comfort zone in screen; If-1 ° of <b a *<0 °, then to determine that background region is recessed and be in comfort zone in screen; If b a *<-1 °, then determine that background region is recessed and be in non-comfort zone in screen;
Wherein, f a *represent the parallactic angle of the foreground target in the right viewpoint parallax gray level image of pending stereo-picture, f a *=f-k, b a *represent the parallactic angle of the background region in the right viewpoint parallax gray level image of pending stereo-picture, b a *=b-k, f represent the convergent angle of human eye binocular viewing foreground target, b represents the convergent angle of human eye binocular viewing background region, k represents the adjustment angle of human eye binocular, arctan () is tan of negating, p represents the interpupillary distance of human eye binocular, L represents the width of display, N represents the horizontal resolution of display, h represents the distance of human eye to display, F represents the mean parallax amplitude of the foreground target in the right viewpoint parallax gray level image of pending stereo-picture, and F is in units of pixel F = &Sigma; x = 1 W &Sigma; y = 1 H &epsiv; ( d R * ( x , y ) - P s e g ) &times; d R * ( x , y ) &Sigma; x = 1 W &Sigma; y = 1 H &epsiv; ( d R * ( x , y ) - P s e g ) , &epsiv; ( d R * ( x , y ) - P s e g ) = 1 , d R * ( x , y ) - P s e g &GreaterEqual; 0 0 , d R * ( x , y ) - P s e g < 0 , B represents the mean parallax amplitude of the background region in the right viewpoint parallax gray level image of pending stereo-picture, B in units of pixel, B = &Sigma; x = 1 W &Sigma; y = 1 H &epsiv; ( P s e g - d R * ( x , y ) ) &times; d R * ( x , y ) &Sigma; x = 1 W &Sigma; y = 1 H &epsiv; ( P s e g - d R * ( x , y ) ) , &epsiv; ( P s e g - d R * ( x , y ) ) = 1 , P s e g - d R * ( x , y ) &GreaterEqual; 0 0 , P s e g - d R * ( x , y ) < 0 , D r *(x, y) represents that in the right viewpoint parallax gray level image of pending stereo-picture, coordinate position is the pixel value of the pixel of (x, y);
Z2, according to the foreground target in the right viewpoint parallax gray level image of pending stereo-picture relative to the concavity and convexity of screen and whether be in comfort zone, background region relative to screen concavity and convexity and whether be in comfort zone, determine the scene mode belonging to pending stereo-picture, be specially: if foreground target is convex be in that non-comfort zone, background region are recessed is in non-comfort zone in screen in screen, then determine that pending stereo-picture belongs to the 1st kind of scene mode; If foreground target is convex be in that non-comfort zone, background region are convex is in non-comfort zone in screen in screen, then determine that pending stereo-picture belongs to the 2nd kind of scene mode; If foreground target is recessed be in that non-comfort zone, background region are recessed is in non-comfort zone in screen in screen, then determine that pending stereo-picture belongs to the 3rd kind of scene mode; If foreground target is convex be in that non-comfort zone, background region are recessed is in comfort zone in screen in screen, then determine that pending stereo-picture belongs to the 4th kind of scene mode; If foreground target is convex be in that non-comfort zone, background region are convex is in comfort zone in screen in screen, then determine that pending stereo-picture belongs to the 5th kind of scene mode; If foreground target is convex be in that comfort zone, background region are recessed is in non-comfort zone in screen in screen, then determine that pending stereo-picture belongs to the 6th kind of scene mode; If foreground target is recessed be in that comfort zone, background region are recessed is in non-comfort zone in screen in screen, then determine that pending stereo-picture belongs to the 7th kind of scene mode; If foreground target is convex be in that comfort zone, background region are recessed is in comfort zone in screen in screen, then determine that pending stereo-picture belongs to the 8th kind of scene mode; If foreground target is convex be in that comfort zone, background region are convex is in comfort zone in screen in screen, then determine that pending stereo-picture belongs to the 9th kind of scene mode; If foreground target is recessed be in that comfort zone, background region are recessed is in comfort zone in screen in screen, then determine that pending stereo-picture belongs to the 10th kind of scene mode.
6. a kind of stereo image vision comfort level evaluation method based on scene mode classification according to claim 5, is characterized in that { d during described step 3. r(x, y) } in the width angle of foreground target wherein, arctan () is tan of negating, and h represents the distance of human eye to display, W frontrepresent { d r(x, y) } in the mean breadth of foreground target, W frontacquisition process be:
A1, to { d r(x, y) } carry out binary conversion treatment, obtain { d r(x, y) } binary image, be designated as { BI (x, y) }, the pixel value being the pixel of (x, y) by coordinate position in { BI (x, y) } is designated as BI (x, y), as { d r(x, y) } in coordinate position be the pixel of (x, y) when belonging to foreground target, make BI (x, y)=1, as { d r(x, y) } in coordinate position be the pixel of (x, y) when belonging to background region, make BI (x, y)=0;
A2, to { BI (x, y) } line scanning is carried out, for { BI (x, y) row } is capable, scan to the right from the 1st pixel that row is capable, when to scan the 1st pixel value be the pixel of 1, using the original position of this pixel as the 1st section of prospect width line segment, and the row coordinate of this pixel is designated as x 1, continuing scanning to the right and obtain the 1st section of prospect width line segment until scanning till pixel value is the pixel of 0, is that the row coordinate of the pixel of 0 is designated as x by this pixel value 2, the length of the 1st section of prospect width line segment is designated as WL 1, WL 1=x 2-x 1; Continue scan to the right, in the mode identical with acquisition the 1st section of prospect width line segment, obtain row capable in all prospect width line segments; Wherein, the initial value of row is 1,1≤row≤H, 1≤x 1<x 2≤ W;
All prospect width line segments that a3, all prospect width line segments length in { BI (x, y) } being less than 0.002W and length are greater than 0.995W are removed; Then sort to remaining all prospect width line segments by length order from small to large, the prospect width line segment getting middle 80% forms prospect width line segment aggregate, is designated as { WL' n, wherein, WL' nrepresent { WL' nin the length of n-th section of prospect width line segment, 1≤n≤N', N' represents { WL' nin total hop count of prospect width line segment of comprising;
A4, calculating { WL' nin the mean value of length of all prospect width line segments, be designated as W front', then by W front' as { d r(x, y) } in the mean breadth W of foreground target fronteven, W front=W front'.
7. a kind of stereo image vision comfort level evaluation method based on scene mode classification according to claim 6, is characterized in that { d during described step 4. r(x, y) } in the average prospect line hop count of foreground target { d r(x, y) } in the average prospect alignment hop count of foreground target wherein, 1≤row≤H, 1≤col≤W, RLQ rowrepresent { d r(x, y) } in row capable in the quantity of prospect line section, &epsiv; ( RLQ r o w ) = { 1 , RLQ r o w > 0 0 , RLQ r o w &le; 0 , CLQ colrepresent { d r(x, y) } in col row in the quantity of prospect alignment section, &epsiv; ( CLQ c o l ) = 1 , C L Q c o l > 0 0 , C L Q c o l &le; 0 ;
Wherein, RLQ rowand CLQ colacquisition process be:
B1, to { d r(x, y) } carry out binary conversion treatment, obtain { d r(x, y) } binary image, be designated as { BI (x, y) }, the pixel value being the pixel of (x, y) by coordinate position in { BI (x, y) } is designated as BI (x, y), as { d r(x, y) } in coordinate position be the pixel of (x, y) when belonging to foreground target, make BI (x, y)=1, as { d r(x, y) } in coordinate position be the pixel of (x, y) when belonging to background region, make BI (x, y)=0;
B2, to { BI (x, y) 2 dilation operations of mathematical morphology, 4 erosion operations, 2 dilation operations } are carried out successively, obtain { BI'(x, y) }, wherein, BI'(x, y) represent { BI'(x, y) }, coordinate position is the pixel value of the pixel of (x, y);
The quantity of effective prospect line section in often row in b3, statistics { BI'(x, y) }, by capable for the row in { BI'(x, y) } effectively the quantity of prospect line section be designated as RLQ row', RLQ row' acquisition process be: b3-1, make RLQ row' initial value be 0; B3-2, from { BI'(x, y) the 1st pixel that the row } is capable starts to scan to the right, when to scan the 1st pixel value be the pixel of 1, using the original position of this pixel as the 1st section of prospect line section, and the row coordinate of this pixel is designated as x 1', continuing scanning to the right and obtain the 1st section of prospect line section until scanning till pixel value is the pixel of 0, is that the row coordinate of the pixel of 0 is designated as x by this pixel value 2', if the length x of the 1st section of prospect line section 2'-x 1' be greater than 0.005W, then determine that the 1st section of prospect line section is effective prospect line section, and make RLQ row'=RLQ row'+1, if the length x of the 1st section of prospect line section 2'-x 1' be less than or equal to 0.005W, then determine that the 1st section of prospect line section is invalid prospect line section; Continue to scan to the right, to determine whether the 1st section of prospect line section is the mode that effective prospect line section is identical with acquisition the 1st section of prospect line section, obtain row capable in all effective prospect line section, and statistics obtain row capable in the amount R LQ of effective prospect line section row'; Wherein, the initial value of row is 1,1≤row≤H, 1≤x 1' <x 2'≤W, RLQ row'=RLQ row"=" in '+1 is assignment;
Equally, the quantity of effective prospect alignment section in the often row in statistics { BI'(x, y) }, by capable for the col in { BI'(x, y) } effectively the quantity of prospect alignment section be designated as CLQ col', CLQ col' acquisition process be: b3-1), make CLQ col' initial value be 0; B3-2), from { BI'(x, 1st pixel of the col row y) } starts downward scanning, when to scan the 1st pixel value be the pixel of 1, using the original position of this pixel as the 1st section of prospect alignment section, and the row-coordinate of this pixel is designated as y 1', continuing scanning downwards and obtain the 1st section of prospect alignment section until scanning till pixel value is the pixel of 0, is that the row-coordinate of the pixel of 0 is designated as y by this pixel value 2', if the length y of the 1st section of prospect alignment section 2'-y 1' be greater than 0.005H, then determine that the 1st section of prospect alignment section is effective prospect alignment section, and make CLQ col'=CLQ col'+1, if the length y of the 1st section of prostatitis alignment section 2'-y 1' be less than or equal to 0.005H, then determine that the 1st section of prospect alignment section is invalid prospect alignment section; Continue scanning downwards, to determine whether the 1st section of prospect alignment section is the mode that effective prospect alignment section is identical with acquisition the 1st section of prospect alignment section, obtain col capable in all effective prospect alignment section, and statistics obtains the quantity CLQ of effective prospect alignment section in col row col'; Wherein, the initial value of col is 1,1≤col≤W, 1≤y 1' <y 2'≤H, CLQ col'=CLQ col"=" in '+1 is assignment;
B4, make RLQ row=RLQ row', make CLQ col=CLQ col'.
8. a kind of stereo image vision comfort level evaluation method based on scene mode classification according to claim 7, is characterized in that getting P=1.6, T during described step 4. f=2.0, T r=2, T c=1.5.
CN201510571897.8A 2015-09-10 2015-09-10 Method for evaluating visual comfort of three-dimensional image based on classification of scene modes Active CN105163111B (en)

Priority Applications (1)

Application Number Priority Date Filing Date Title
CN201510571897.8A CN105163111B (en) 2015-09-10 2015-09-10 Method for evaluating visual comfort of three-dimensional image based on classification of scene modes

Applications Claiming Priority (1)

Application Number Priority Date Filing Date Title
CN201510571897.8A CN105163111B (en) 2015-09-10 2015-09-10 Method for evaluating visual comfort of three-dimensional image based on classification of scene modes

Publications (2)

Publication Number Publication Date
CN105163111A true CN105163111A (en) 2015-12-16
CN105163111B CN105163111B (en) 2017-01-25

Family

ID=54803852

Family Applications (1)

Application Number Title Priority Date Filing Date
CN201510571897.8A Active CN105163111B (en) 2015-09-10 2015-09-10 Method for evaluating visual comfort of three-dimensional image based on classification of scene modes

Country Status (1)

Country Link
CN (1) CN105163111B (en)

Cited By (3)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN108182444A (en) * 2017-12-08 2018-06-19 深圳英飞拓科技股份有限公司 The method and device of video quality diagnosis based on scene classification
CN109167988A (en) * 2018-08-29 2019-01-08 长春理工大学 A kind of stereo image vision comfort level evaluation method based on D+W model and contrast
CN111860691A (en) * 2020-07-31 2020-10-30 福州大学 Professional stereoscopic video visual comfort degree classification method based on attention and recurrent neural network

Citations (4)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN102724544A (en) * 2012-06-18 2012-10-10 清华大学 System and method for subjective assessment of stereoscopic videos
CN103096125A (en) * 2013-02-22 2013-05-08 吉林大学 Stereoscopic video visual comfort evaluation method based on region segmentation
KR20140148080A (en) * 2013-06-21 2014-12-31 한국과학기술원 Stereoscopic imaging method and system for visually comfortable 3D images
CN104853185A (en) * 2015-06-06 2015-08-19 吉林大学 Stereo video comfort evaluation method combining multiple parallaxes with motion

Patent Citations (4)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN102724544A (en) * 2012-06-18 2012-10-10 清华大学 System and method for subjective assessment of stereoscopic videos
CN103096125A (en) * 2013-02-22 2013-05-08 吉林大学 Stereoscopic video visual comfort evaluation method based on region segmentation
KR20140148080A (en) * 2013-06-21 2014-12-31 한국과학기술원 Stereoscopic imaging method and system for visually comfortable 3D images
CN104853185A (en) * 2015-06-06 2015-08-19 吉林大学 Stereo video comfort evaluation method combining multiple parallaxes with motion

Non-Patent Citations (3)

* Cited by examiner, † Cited by third party
Title
HONGWEI YING ET AL: "New Stereo Visual Comfort Assessment Method Based on Scene Mode Classification", 《IEEE XPLORE DIGITAL LIBRARY》 *
姜求平等: "基于视觉重要区域的立体图像视觉舒适度客观评价方法", 《电子与信息学报》 *
邵枫等: "基于显著性分析的立体图像视觉舒适度预测", 《光学精密工程》 *

Cited By (5)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN108182444A (en) * 2017-12-08 2018-06-19 深圳英飞拓科技股份有限公司 The method and device of video quality diagnosis based on scene classification
CN109167988A (en) * 2018-08-29 2019-01-08 长春理工大学 A kind of stereo image vision comfort level evaluation method based on D+W model and contrast
CN109167988B (en) * 2018-08-29 2020-06-26 长春理工大学 Stereo image visual comfort evaluation method based on D + W model and contrast
CN111860691A (en) * 2020-07-31 2020-10-30 福州大学 Professional stereoscopic video visual comfort degree classification method based on attention and recurrent neural network
CN111860691B (en) * 2020-07-31 2022-06-14 福州大学 Stereo video visual comfort degree classification method based on attention and recurrent neural network

Also Published As

Publication number Publication date
CN105163111B (en) 2017-01-25

Similar Documents

Publication Publication Date Title
CN103581661B (en) Method for evaluating visual comfort degree of three-dimensional image
CN103096125B (en) Stereoscopic video visual comfort evaluation method based on region segmentation
CN103347196B (en) Method for evaluating stereo image vision comfort level based on machine learning
CN104658002B (en) Non-reference image objective quality evaluation method
CN104394403B (en) A kind of stereoscopic video quality method for objectively evaluating towards compression artefacts
CN102750695A (en) Machine learning-based stereoscopic image quality objective assessment method
CN104811691B (en) A kind of stereoscopic video quality method for objectively evaluating based on wavelet transformation
CN105407349A (en) No-reference objective three-dimensional image quality evaluation method based on binocular visual perception
CN104036501A (en) Three-dimensional image quality objective evaluation method based on sparse representation
CN103780895B (en) A kind of three-dimensional video quality evaluation method
CN103281554B (en) Video objective quality evaluation method based on human eye visual characteristics
CN103136748B (en) The objective evaluation method for quality of stereo images of a kind of feature based figure
CN105282543B (en) Total blindness three-dimensional image quality objective evaluation method based on three-dimensional visual perception
CN104202594B (en) A kind of method for evaluating video quality based on 3 D wavelet transformation
KR20130081835A (en) Method and system for analyzing a quality of three-dimensional image
CN105163111A (en) Method for evaluating visual comfort of three-dimensional image based on classification of scene modes
CN109788275A (en) Naturality, structure and binocular asymmetry are without reference stereo image quality evaluation method
CN104036493A (en) No-reference image quality evaluation method based on multifractal spectrum
CN107330873A (en) Objective evaluation method for quality of stereo images based on multiple dimensioned binocular fusion and local shape factor
CN102708568B (en) Stereoscopic image objective quality evaluation method on basis of structural distortion
CN104361583A (en) Objective quality evaluation method of asymmetrically distorted stereo images
CN104144339B (en) A kind of matter based on Human Perception is fallen with reference to objective evaluation method for quality of stereo images
CN103745457B (en) A kind of three-dimensional image objective quality evaluation method
CN108848365B (en) A kind of reorientation stereo image quality evaluation method
CN105898279B (en) A kind of objective evaluation method for quality of stereo images

Legal Events

Date Code Title Description
C06 Publication
PB01 Publication
C10 Entry into substantive examination
SE01 Entry into force of request for substantive examination
C14 Grant of patent or utility model
GR01 Patent grant
TR01 Transfer of patent right
TR01 Transfer of patent right

Effective date of registration: 20190924

Address after: 242000 Meixi Road and Wolong Lane Intersection of Ningbo Feichuan Office, Xuancheng City, Anhui Province

Patentee after: Xuancheng Youdu Technology Service Co., Ltd.

Address before: 315211 Zhejiang Province, Ningbo Jiangbei District Fenghua Road No. 818

Patentee before: Ningbo University