CN110944165A - Stereoscopic image visual comfort level improving method combining perceived depth quality - Google Patents

Stereoscopic image visual comfort level improving method combining perceived depth quality Download PDF

Info

Publication number
CN110944165A
CN110944165A CN201911105732.6A CN201911105732A CN110944165A CN 110944165 A CN110944165 A CN 110944165A CN 201911105732 A CN201911105732 A CN 201911105732A CN 110944165 A CN110944165 A CN 110944165A
Authority
CN
China
Prior art keywords
image
equal
value
pixel point
visual
Prior art date
Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
Granted
Application number
CN201911105732.6A
Other languages
Chinese (zh)
Other versions
CN110944165B (en
Inventor
应宏微
郁梅
池碧蔚
张君君
蒋刚毅
Current Assignee (The listed assignees may be inaccurate. Google has not performed a legal analysis and makes no representation or warranty as to the accuracy of the list.)
Ningbo University
Original Assignee
Ningbo University
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Application filed by Ningbo University filed Critical Ningbo University
Priority to CN201911105732.6A priority Critical patent/CN110944165B/en
Publication of CN110944165A publication Critical patent/CN110944165A/en
Application granted granted Critical
Publication of CN110944165B publication Critical patent/CN110944165B/en
Active legal-status Critical Current
Anticipated expiration legal-status Critical

Links

Images

Classifications

    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04NPICTORIAL COMMUNICATION, e.g. TELEVISION
    • H04N17/00Diagnosis, testing or measuring for television systems or their details

Landscapes

  • Engineering & Computer Science (AREA)
  • Health & Medical Sciences (AREA)
  • Biomedical Technology (AREA)
  • General Health & Medical Sciences (AREA)
  • Multimedia (AREA)
  • Signal Processing (AREA)
  • Processing Or Creating Images (AREA)

Abstract

The invention discloses a method for improving the visual comfort of a three-dimensional image by combining perceived depth quality, which comprises the steps of firstly establishing a visual satisfaction objective evaluation model for guiding the improvement process, wherein the model integrates the visual comfort and the perceived depth quality objective evaluation consisting of a perceived absolute distance and a perceived relative distance; redrawing an original stereo image needing to improve visual comfort by adopting a viewing distance nonlinear translation method, wherein the redrawn stereo image has the characteristic of unchanged virtual scene geometric proportion; and then respectively calculating the visual satisfaction and the objective evaluation value of the visual comfort of the redrawn three-dimensional image, wherein when the visual satisfaction reaches the maximum value in the redrawing for many times and the visual comfort value is greater than that of the original three-dimensional image, the redrawn three-dimensional image is the optimal three-dimensional image, and the improved three-dimensional image improves the visual comfort and simultaneously keeps better perception depth quality and is more consistent with the preference of a human visual system.

Description

Stereoscopic image visual comfort level improving method combining perceived depth quality
Technical Field
The invention relates to a method for improving the visual comfort of a stereoscopic image, in particular to a method for improving the visual comfort of a stereoscopic image by combining perceived depth quality.
Background
With the rapid development of stereoscopic video display technology and high-quality stereoscopic video content acquisition technology, the visual satisfaction of stereoscopic images/videos has become an important issue in the design of stereoscopic systems. Visual Comfort (VC) of a stereoscopic image is one of important factors affecting the visual satisfaction of the stereoscopic image. The research on the visual comfort of the stereo image is physiological fatigue and discomfort caused on the premise that the content of the stereo image is not distorted; the research on the Perceived Depth Quality (PDQ) of a stereo image is that the stereo image brings the stereoscopic impression, immersion feeling and presence feeling to a viewer, and the visual comfort and the perceived depth quality are important factors influencing the visual satisfaction of the stereo image. The method for improving the visual comfort of the research stereo image plays an important role in improving the visual satisfaction of a viewer and guiding the production and post-processing of 3D content.
The visual comfort of the stereoscopic image is closely related to the binocular parallax, the binocular parallax is a cause for generating stereoscopic sensation, and the overlarge binocular parallax is an important factor for generating visual discomfort. The existing objective evaluation method for the visual comfort of the stereo image mainly extracts the parallax features of the stereo image, such as parallax amplitude, parallax gradient, relative parallax, parallax range and the like, then establishes an objective evaluation model for the visual comfort by using a statistical method or a machine learning algorithm, and when the model is used for guiding the manufacturing and post-processing of 3D content, binocular parallax can be excessively reduced so as to enable the visual comfort to be optimal, but the stereoscopic impression, immersion impression, presence impression and reality impression are weaker, namely, the perceived depth quality is weakened, so that the visual satisfaction is reduced. Therefore, a method for improving the comfort level of the stereoscopic image with the perceived depth quality fused is required to be researched, so that the visual comfort level of the stereoscopic image is improved, and the overall visual satisfaction degree of the stereoscopic image is also improved.
Disclosure of Invention
The technical problem to be solved by the invention is to provide a method for improving the visual comfort of a stereoscopic image by combining with the perceived depth quality, which effectively improves the visual comfort of the stereoscopic image and greatly improves the overall visual satisfaction of the stereoscopic image.
The technical scheme adopted by the invention for solving the technical problems is as follows: a stereoscopic image visual comfort level improving method combining perceived depth quality is characterized by comprising four processes of stereoscopic image redrawing based on viewing distance nonlinear translation, modeling of a stereoscopic image visual satisfaction degree objective evaluation model, objective evaluation of the visual satisfaction degree and the visual comfort level of a stereoscopic image and optimal stereoscopic image discrimination;
the method comprises the following specific steps of:
step 1_ 1: will have a width of WimageAnd has a height HimageThe stereo image to be improved in visual comfort is recorded as S0Will S0The left view image and the right view image are correspondingly denoted as SL0And SR0(ii) a Then obtain the SR0Is recorded as DR0(ii) a Then calculating DR0The screen parallax value of each pixel point in (1) is calculated as DR0The screen parallax value of the pixel point with the middle coordinate position (x, y) is recorded as DRS0(x, y); then according to DR0Calculating S as the screen parallax value of each pixel point0The distance between each point in the formed virtual scene and the human eye, i.e. calculating S0Viewing distance, S, of each point in the formed virtual scene0The viewing distance of a point with a coordinate position (x, y) in the formed virtual scene is recorded as VD0(x,y),
Figure BDA0002271231850000021
Finally, the S is0The matrix formed by the viewing distances of all the points in the virtual scene is denoted as VD0(ii) a Wherein x represents the column coordinate of the pixel point, y represents the row coordinate of the pixel point, x is more than or equal to 0 and is more than Wimage,0≤y<HimageP represents a binocular pupillary distance of human eyes, and T represents a distance from the human eyes to a 3D display for displaying a stereoscopic image;
step 1_ 2: let j denote the number of translation steps, and the value range of j is more than or equal to 0 and less than or equal to N; s when j is 00All points in the formed virtual scene do not translate in the depth direction; when j is more than or equal to 1 and less than or equal to N, enabling S to be0The point with the minimum viewing distance in the formed virtual scene is translated by a distance of j multiplied by delta Z in the depth direction; then let S0All the points except the point with the minimum viewing distance in the formed virtual scene are subjected to nonlinear translation in the depth direction to obtain a new virtual scene, and then a matrix formed by the viewing distances of all the points in the new virtual scene is obtained and recorded as VDjWill VDjThe value of the element with the middle subscript position (x, y), that is, the viewing distance of the point with the coordinate position (x, y) in the new virtual scene is recorded as VDj(x,y),
Figure BDA0002271231850000031
Wherein j is a positive integer, N is a positive integer, and N belongs to [10,20 ]]And deltaz represents the step size of the translation,
Figure BDA0002271231850000032
Figure BDA0002271231850000033
represents VD0Minimum value of (1);
step 1_ 3: j is more than or equal to 1 and less than or equal to N according to VDjCalculating the screen parallax value of each pixel point in the parallax image of the right viewpoint image of the new three-dimensional image corresponding to the new virtual scene, and recording the screen parallax value of the pixel point with the coordinate position (x, y) in the parallax image of the right viewpoint image of the new three-dimensional image as DRSj(x,y),
Figure BDA0002271231850000034
Then calculating the parallax image of the right viewpoint image of the new stereo image and recording the parallax image as DRjWill DRjThe pixel value of the pixel point with the middle coordinate position (x, y) is recorded as DRj(x,y),
Figure BDA0002271231850000035
DRj(x, y) is also a new perspective viewThe parallax value of a pixel point with a coordinate position of (x, y) in the right viewpoint image of the image, and the screen width of a 3D display for displaying a stereoscopic image is W3DAnd a screen height of the 3D display for displaying the stereoscopic image is H3D
Step 1_ 4: calculating each pixel point and SL in the left viewpoint image of the new stereoscopic image when j is more than or equal to 1 and less than or equal to N0The relation between the corresponding pixel points on the column coordinates, the pixel point with the coordinate position of (x, y) in the left viewpoint image of the new stereo image and the SL0The relation between the pixel points with the (x, y) middle coordinate position on the column coordinate is recorded as Δ L (x, y), and Δ L (x, y) is (DR)0(x,y)-DRj(x, y))/2; also, each pixel point and SR in the right view image of the new stereoscopic image are calculated0The relation between the corresponding pixel points on the column coordinates, the pixel point with the coordinate position of (x, y) in the right viewpoint image of the new three-dimensional image and the SR0The relation between the pixel points with the (x, y) middle coordinate position on the column coordinate is recorded as Δ R (x, y), and Δ R (x, y) — (DR)0(x,y)-DRj(x, y))/2; wherein, DR0(x, y) denotes DR0The pixel value DR of the pixel point with the middle coordinate position (x, y)0(x, y) is also SR0The middle coordinate position is the parallax value of the pixel point of (x, y);
step 1_ 5: obtaining SL when j is more than or equal to 1 and less than or equal to N0The left viewpoint image obtained after redrawing is marked as SLj(ii) a And obtaining the SR0The right viewpoint image obtained after redrawing is marked as SRj(ii) a Then from SLjAnd SRjForm a redrawn three-dimensional image, denoted as Sj(ii) a Wherein SLjAnd SRjThe acquisition process comprises the following steps:
step 1_5 a: SR'jRepresents SR0If x + delta R (x, y) < W is more than or equal to 0, the right viewpoint image obtained after the first redrawingimageThen SR 'is calculated'jThe pixel value of the pixel point with the middle coordinate position of (x, y) is recorded as SR'j(x,y),SR'j(x,y)=SR0(x + Δ R (x, y), y); if x + Δ R (x, y) < 0 or x + Δ R (x, y) ≧ WimageThen SR'jDetermining the pixel point with the middle coordinate position (x, y) as a cavity pixel point; wherein, SR0(x + Δ R (x, y), y) represents SR0The middle coordinate position is the pixel value of the pixel point of (x + delta R (x, y), y);
step 1_5 b: SR 'judgment'jIf the hollow pixel point exists in the buffer, SR 'is carried out'jAnew noted SR'jThen, step 1_5c is performed; if the empty pixel exists, each empty pixel is processed, and SR 'is set'jIf the pixel point with the middle coordinate position (x, y) is a cavity pixel point, then x + DR is more than or equal to 00(x,y)+ΔR(x,y)<WimageCalculating the redrawn pixel value of the hole pixel point to be equal to SL0(x+DR0(x, y) + Δ R (x, y), y), when x + DR0(x, y) + Δ R (x, y) < 0 or x + DR0(x,y)+ΔR(x,y)≥WimageThe hollow pixel points are not redrawn, and an SR is obtained after all the hollow pixel points are processed0The right viewpoint image obtained after the second redrawing is marked as SR "jThen, step 1_5c is performed; wherein SL0(x+DR0(x, y) + Δ R (x, y), y) represents SL0The middle coordinate position is (x + DR)0(x, y) + Δ R (x, y), y);
step 1_5 c: line of SL'jRepresents SL0If x-DR is not less than 0, the left viewpoint image obtained after the first redrawingj(x,y)<WimageThen calculate SL'jThe pixel value of the pixel point with the middle coordinate position of (x, y) is recorded as SL'j(x,y),SL'j(x,y)=SR”j(x-DRj(x, y), y); if x-DRj(x, y) < 0 or x-DRj(x,y)≥WimageThen SL 'is prepared'jDetermining the pixel point with the middle coordinate position (x, y) as a cavity pixel point; wherein, SR "j(x-DRj(x, y), y) represents SR "jThe middle coordinate position is (x-DR)j(x, y), y) pixel values of the pixels;
step 1_5 d: judging SL'jIf there is a hole pixel point, if there is no hole pixel point, SL'jAnew as SL "jThen, step 1_5e is performed; if the signal exists, processing each hollow pixel point, and setting SL'jIf the pixel point with the middle coordinate position (x, y) is a cavity pixel point, then when x + delta L (x, y) < W is more than or equal to 0imageCalculating the redrawn pixel value of the hole pixel point to be equal to SL0(x + Δ L (x, y), y) when x + Δ L (x, y) < 0 or x + Δ L (x, y) ≧ WimageThe hollow pixel points are not redrawn, and SL is obtained after all the hollow pixel points are processed0The left viewpoint image obtained after the second redrawing is marked as SL "jThen, step 1_5e is performed; wherein SL0(x + Δ L (x, y), y) represents SL0The middle coordinate position is the pixel value of the pixel point of (x + delta L (x, y), y);
step 1_5 e: if SL "jIf no hollow pixel point exists in the interface, SL is used "jAs SLj(ii) a If SR'jIf no hollow pixel exists in the image, then SR is carried out "jAs SRj(ii) a If SL "jAnd SR'jIf there are still void pixel points, then use image repairing technique to repair each void pixel point, and use SL "jThe left viewpoint image obtained after the patching is used as SLjIs prepared from SR "jThe right viewpoint image obtained after patching is used as SRj
The modeling part of the three-dimensional image visual satisfaction objective evaluation model comprises the following specific steps:
step 2_ 1: selecting M pieces of width WimageAnd has a height HimageThe stereo images of (a) form a stereo image set; wherein M is a positive integer, and M is more than or equal to 20;
step 2_ 2: defining the ith stereo image to be processed currently in the stereo image set as a current stereo image; wherein i is a positive integer, i is more than or equal to 1 and less than or equal to M;
step 2_ 3: recording the current stereo image as phii,0(ii) a Then changing the value of j from 1 to N, repeatedly executing the j values for 1 time in the same way according to the processes from step 1_1 to step 1_5, and repeatedly executing the j values from 1 to N for N times to obtain phii,0Corresponding N redrawing three-dimensional images, and obtaining phi when the number of translation steps is ji,0The corresponding redrawn stereo image is recorded as phii,jWhile obtaining phii,0A matrix VD of viewing distances of all points in the formed virtual scenei,0And each redrawn stereoscopic image pairThe viewing distance of all points in the new virtual scene is used to form a matrix, and phii,jThe matrix formed by the viewing distances of all the points in the corresponding new virtual scene is recorded as VDi,j
Step 2_ 4: taking the next to-be-processed stereo image in the stereo image set as a current stereo image by setting i to i + 1; then, returning to the step 2_3 to continue executing until N redrawn three-dimensional images corresponding to each three-dimensional image in the three-dimensional image set are obtained, and forming a training set by all three-dimensional images in the three-dimensional image set and the obtained M multiplied by N redrawn three-dimensional images; wherein, the' in the i-i +1 is an assignment symbol;
step 2_ 5: obtaining an average subjective evaluation mean value of 3D visual satisfaction of each three-dimensional image in a training set; then, the set formed by the mean subjective evaluation average of the 3D visual satisfaction of all the stereo images in the training set is recorded as { VSMOSi,jI is more than or equal to 1 and less than or equal to M, and j is more than or equal to 0 and less than or equal to N }; VSMOS when j is 0i,0Represents phii,0The average subjective evaluation mean value of 3D visual satisfaction is that when j is more than or equal to 1 and less than or equal to N, VSMOSi,jPhi obtained when the number of steps representing translation is ji,0Corresponding redrawn stereo image phii,jThe average subjective evaluation mean value of the 3D visual satisfaction degree comprises the comprehensive effect of the visual comfort degree and the perception depth;
step 2_ 6: calculating the objective evaluation value of the visual comfort degree of each three-dimensional image in the training set according to the objective evaluation model of the visual comfort degree; then, the set of objective evaluation values of visual comfort of all the stereo images in the training set is recorded as { VCA }i,jI is more than or equal to 1 and less than or equal to M, and j is more than or equal to 0 and less than or equal to N }; where, when j is 0, VCAi,0Represents phii,0The objective evaluation value of visual comfort is VCA when j is more than or equal to 1 and less than or equal to Ni,jPhi obtained when the number of steps representing translation is ji,0Corresponding redrawn stereo image phii,jThe objective evaluation value of the visual comfort of the user;
step 2_ 7: according to the set { VDi,jI is more than or equal to 1 and less than or equal to M, j is more than or equal to 0 and less than or equal to N, and calculating the objective evaluation value of the perceived absolute distance of each three-dimensional image in the training set; then all stereo images in the training set are combinedIs given as { APDA (adaptive distance estimation and estimation) in the set of objective evaluation values of perceived absolute distancei,jI is more than or equal to 1 and less than or equal to M, and j is more than or equal to 0 and less than or equal to N }; wherein, VD is when j is more than or equal to 1 and less than or equal to Ni,jPhi obtained when the number of steps representing translation is ji,0Corresponding redrawn stereo image phii,jAPDA when j is 0i,0Represents phii,0The perceived absolute distance objective evaluation value of (1) is more than or equal to j and less than or equal to N, and APDAi,jPhi obtained when the number of steps representing translation is ji,0Corresponding redrawn stereo image phii,jThe perceived absolute distance objective evaluation value of (1);
step 2_ 8: according to the set { VDi,jI is more than or equal to 1 and less than or equal to M, j is more than or equal to 0 and less than or equal to N, and calculating the objective evaluation value of the perception relative distance of each three-dimensional image in the training set; then, a set of objective evaluation values of perceptual relative distances of all stereo images in the training set is recorded as { RPDAi,jI is more than or equal to 1 and less than or equal to M, and j is more than or equal to 0 and less than or equal to N }; wherein, when j is 0, RPDAi,0Represents phii,0The perceived relative distance objective evaluation value of (1) is more than or equal to j and less than or equal to N, and the RPDA is used for evaluating the distancei,jPhi obtained when the number of steps representing translation is ji,0Corresponding redrawn stereo image phii,jThe perceived relative distance objective evaluation value of (1);
step 2_ 9: using a training model of ε -SVR in VSMOSi,jI is more than or equal to 1 and less than or equal to M, j is more than or equal to 0 and less than or equal to N and { VCAi,j|1≤i≤M,0≤j≤N}、{APDAi,j|1≤i≤M,0≤j≤N}、{RPDAi,jFitting is carried out between i is more than or equal to 1 and less than or equal to M and j is more than or equal to 0 and less than or equal to N to obtain the objective evaluation model of the visual satisfaction degree of the stereo image, and the description is as follows: STFA (VCA, APDA, RPDA); the kernel function of the training model epsilon-SVR uses a histogram interaction type, a width coefficient gamma is 1/256, a penalty coefficient C is 4, an insensitive loss coefficient epsilon is 0.1, STFA () is a function representation form of a stereoscopic image visual satisfaction objective evaluation model, VCA, APDA and RPDA are all input of the stereoscopic image visual satisfaction objective evaluation model, VCA is used for representing a visual comfort objective evaluation value of a stereoscopic image, APDA is used for representing a perception absolute distance objective evaluation value of the stereoscopic image, and RPDA is used for representing a perception relative distance objective evaluation value of the stereoscopic image;
the objective evaluation part of the visual satisfaction and the visual comfort of the stereo image comprises the following specific steps:
step 3_ 1: s is obtained in the same manner as the procedure of step 2_6jObjective evaluation value of visual comfort of (1), noted as VCAj(ii) a And obtains S in the same manner as the procedure of step 2_7jIs recorded as APDAj(ii) a S is obtained in the same manner as the procedure of step 2_8jThe objective evaluation value of the perceived relative distance is marked as RPDAj(ii) a Then VCAj、APDAj、RPDAjInputting the three-dimensional image visual satisfaction objective evaluation model obtained in the step 2-9 to obtain SjIs recorded as STFAj,STFAj=STFA(VCAj,APDAj,RPDAj) (ii) a Wherein j is more than or equal to 0 and less than or equal to N;
the specific steps of the optimal stereo image discrimination part are as follows:
step 4_ 1: directly performing step 4_2 when j is 0 or j is 1; when j is more than or equal to 2 and less than or equal to N, the pointer points to Sj-1Judging STFAj-1>STFAj-2、STFAj-1>STFAj、VCAj-1>VCA0Whether or not to satisfy simultaneously, if so, then Sj-1Is determined as S0The optimal stereo image with improved visual comfort is obtained; otherwise, directly executing the step 4_ 2; wherein, STFAj-1S obtained when the number of steps representing translation is j-10Corresponding redrawn three-dimensional image Sj-1Is the visual satisfaction value of (1), STFA when j is 2j-2Is STFA0And when j is more than 2 and less than or equal to N, STFAj-2S obtained when the number of steps representing translation is j-20Corresponding redrawn three-dimensional image Sj-2A visual satisfaction value of;
step 4_ 2: let j equal j +1, then repeat step 1_2 to step 1_5 and step 3_2 and step 4_ 1; wherein, j is the assigned symbol in j + 1.
In the step 1_1, the first step,
Figure BDA0002271231850000081
wherein,DR0(x, y) denotes DR0The pixel value DR of the pixel point with the middle coordinate position (x, y)0(x, y) is also SR0The parallax value of the pixel point with the middle coordinate position (x, y) and the screen width W of the 3D display for displaying the three-dimensional image3DAnd a screen height of the 3D display for displaying the stereoscopic image is H3D
In step 2_5, the specific process of obtaining the average subjective evaluation mean value of the 3D visual satisfaction of each stereo image in the training set comprises:
step 2_5 a: selecting more than 10 testees containing gender of male and female to participate in subjective experiments, wherein the visual system of each tester meets the condition that the stereoscopic vision is less than 60 arcsec and passes the color test;
step 2_5 b: making a distance between each subject and a 3D display for displaying a stereoscopic image 3 times a screen height of the 3D display, and enabling each subject to view the stereoscopic image displayed on the 3D display in head-up; informing each subject of a scoring standard, wherein the scoring standard is two subjective visual perceptions of visual comfort and perception depth of the comprehensive stereo image, and scoring on the whole;
step 2_5 c: enabling the 3D display to sequentially display each three-dimensional image in the training set, and displaying for 2-3 times in a co-circulation mode, enabling the display time of each three-dimensional image on the 3D display to be 5-10 seconds, enabling the 3D display to display the next three-dimensional image after the testee has a rest for 3-10 seconds, and enabling the 3D display to display a middle gray image during the rest of the testee; each subject looks up to view the stereoscopic images displayed on the 3D display, and when each stereoscopic image in the training set is displayed in turn in the last pass, each subject scores the 3D visual satisfaction level of the stereoscopic image just displayed on the 3D display during the rest period according to the absolute category rating in the international standards ITU-T p.910 and ITU-T p.911, 5 points represent "very satisfactory", 4 points represent "satisfactory", 3 points represent "normal", 2 points represent "unsatisfactory", 1 point represents "very unsatisfactory";
step 2_5 d: after each subject scores the 3D visual satisfaction level of each stereo image in the training set, removing abnormal scores and keeping normal scores according to a screening method disclosed in the international standard ITU-R BT.500-11;
step 2_5 e: and taking the average value of all normal scores of the stereo images as the average subjective score average value of the 3D visual satisfaction of the stereo images aiming at each stereo image in the training set.
In the step 2_7, APDAi,jThe acquisition process comprises the following steps:
step 2_7 a: for VDi,jThe values of all elements in the sequence are sorted from big to small or from small to big; then, the smallest top 1% value is removed from the values of all the sorted elements; finding out the minimum value from the values of all the elements left, and recording the minimum value as
Figure BDA0002271231850000091
Step 2_7 b: computing
Figure BDA0002271231850000092
In the step 2_8, RPDAi,jThe acquisition process comprises the following steps:
step 2_8 a: for VDi,jThe values of all elements in the sequence are sorted from big to small or from small to big; then, the smallest top 1% value and the largest top 1% value are removed from the values of all the sorted elements; finding out the minimum and maximum values from the values of all the elements left, and recording the minimum and maximum values as corresponding values
Figure BDA0002271231850000101
And
Figure BDA0002271231850000102
step 2_8 b: computing
Figure BDA0002271231850000103
Wherein e represents a natural base number, Gi,jRepresenting the perceptual relative enhancement coefficient(s),
Figure BDA0002271231850000104
Figure BDA0002271231850000105
to represent
Figure BDA0002271231850000106
The perceived distance obtained by the conversion is converted,
Figure BDA0002271231850000107
Figure BDA0002271231850000108
to represent
Figure BDA0002271231850000109
The perceived distance obtained by the conversion is converted,
Figure BDA00022712318500001010
Figure BDA00022712318500001011
when j is 0
Figure BDA00022712318500001012
The perceived distance obtained by the conversion is converted,
Figure BDA00022712318500001013
Figure BDA00022712318500001014
when j is 0
Figure BDA00022712318500001015
The perceived distance obtained by the conversion is converted,
Figure BDA00022712318500001016
compared with the prior art, the invention has the advantages that:
the method adopts the visual satisfaction objective evaluation model to guide redrawing of the three-dimensional image, so that the redrawn three-dimensional image is optimal, wherein the created visual satisfaction objective evaluation model integrates the visual comfort and perception depth quality evaluation of the three-dimensional image, and the satisfaction of the overall visual experience of the finally obtained redrawn three-dimensional image is optimal while the visual comfort is improved; in addition, the established viewing distance nonlinear translation method has the advantage of keeping the consistency of the geometric proportion of the virtual scene before and after the redrawing of the stereo image, namely, the sense of reality on the geometric proportion is kept, and the perceived depth quality is improved.
Drawings
FIG. 1 is a block diagram of an overall implementation of the method of the present invention;
FIG. 2 is a three-dimensional image S to be viewed with enhanced comfort0Viewing distance VD of point with coordinate position (x, y) in formed virtual scene0(x, y) and S0Right view image SR0Is a parallax image DR0Screen parallax value DRS of pixel point with (x, y) as middle coordinate position0(x, y), a geometric schematic diagram of the relationship between the distance T from the human eye to the 3D display and the binocular pupillary distance p of the human eye;
fig. 3 is a block diagram of an implementation of an objective evaluation model of the visual satisfaction degree of a stereoscopic image in the method of the present invention.
Detailed Description
The invention is described in further detail below with reference to the accompanying examples.
The invention provides a method for improving the visual comfort of a three-dimensional image by combining perceived depth quality, which has the overall realization block diagram as shown in figure 1 and is characterized by comprising four processes of three-dimensional image redrawing based on viewing distance nonlinear translation, modeling of a three-dimensional image visual satisfaction objective evaluation model, objective evaluation of the visual satisfaction and the visual comfort of the three-dimensional image and optimal three-dimensional image discrimination.
The method comprises the following specific steps of:
step 1_ 1: will have a width of WimageAnd has a height HimageThe stereo image to be improved in visual comfort is recorded as S0Will S0Left view image and right view image ofIs correspondingly denoted as SL0And SR0(ii) a Then obtain the SR0Is recorded as DR0(ii) a Then calculating DR0The screen parallax value of each pixel point in (1) is calculated as DR0The screen parallax value of the pixel point with the middle coordinate position (x, y) is recorded as DRS0(x, y); then according to DR0Calculating S as the screen parallax value of each pixel point0The distance between each point in the formed virtual scene and the human eye, i.e. calculating S0Viewing distance, S, of each point in the formed virtual scene0The viewing distance of a point with a coordinate position (x, y) in the formed virtual scene is recorded as VD0(x,y),
Figure BDA0002271231850000111
Finally, the S is0The matrix formed by the viewing distances of all the points in the virtual scene is denoted as VD0(ii) a Wherein x represents the column coordinate of the pixel point, y represents the row coordinate of the pixel point, x is more than or equal to 0 and is more than Wimage,0≤y<HimageP denotes a binocular pupillary distance of human eyes, and T denotes a distance from the human eyes looking down to a 3D display for displaying a stereoscopic image.
In this embodiment, in step 1_1,
Figure BDA0002271231850000121
wherein, DR0(x, y) denotes DR0The pixel value DR of the pixel point with the middle coordinate position (x, y)0(x, y) is also SR0The parallax value of the pixel point with the middle coordinate position (x, y) and the screen width W of the 3D display for displaying the three-dimensional image3DAnd a screen height of the 3D display for displaying the stereoscopic image is H3D
Here, DR0Obtaining by the prior art; fig. 2 shows a stereo image S to be visually enhanced in comfort0Viewing distance VD of point with coordinate position (x, y) in formed virtual scene0(x, y) and S0Right view image SR0Is a parallax image DR0Screen parallax value DRS of pixel point with (x, y) as middle coordinate position0(x, y) head-up to the human eyeA geometrical schematic diagram of the relationship between the distance T down to the 3D display, the binocular pupillary distance p of the human eye.
Step 1_ 2: let j denote the number of translation steps, and the value range of j is more than or equal to 0 and less than or equal to N; s when j is 00All points in the formed virtual scene do not translate in the depth direction; when j is more than or equal to 1 and less than or equal to N, enabling S to be0The point with the minimum viewing distance in the formed virtual scene is translated by a distance of j multiplied by delta Z in the depth direction; then let S0All the points except the point with the minimum viewing distance in the formed virtual scene are subjected to nonlinear translation in the depth direction to obtain a new virtual scene, and then a matrix formed by the viewing distances of all the points in the new virtual scene is obtained and recorded as VDjWill VDjThe value of the element with the middle subscript position (x, y), that is, the viewing distance of the point with the coordinate position (x, y) in the new virtual scene is recorded as VDj(x,y),
Figure BDA0002271231850000122
Wherein j is a positive integer, N is a positive integer, and N belongs to [10,20 ]]And deltaz represents the step size of the translation,
Figure BDA0002271231850000123
Figure BDA0002271231850000124
represents VD0Minimum value of (1).
Step 1_ 3: j is more than or equal to 1 and less than or equal to N according to VDjCalculating the screen parallax value of each pixel point in the parallax image of the right viewpoint image of the new three-dimensional image corresponding to the new virtual scene, and recording the screen parallax value of the pixel point with the coordinate position (x, y) in the parallax image of the right viewpoint image of the new three-dimensional image as DRSj(x,y),
Figure BDA0002271231850000131
Then calculating the parallax image of the right viewpoint image of the new stereo image and recording the parallax image as DRjWill DRjThe pixel value of the pixel point with the middle coordinate position (x, y) is recorded as DRj(x,y),
Figure BDA0002271231850000132
DRj(x, y) is also the parallax value of the pixel point with the coordinate position (x, y) in the right viewpoint image of the new stereo image, and the screen width of the 3D display for displaying the stereo image is W3DAnd a screen height of the 3D display for displaying the stereoscopic image is H3D
Step 1_ 4: calculating each pixel point and SL in the left viewpoint image of the new stereoscopic image when j is more than or equal to 1 and less than or equal to N0The relation between the corresponding pixel points on the column coordinates, the pixel point with the coordinate position of (x, y) in the left viewpoint image of the new stereo image and the SL0The relation between the pixel points with the (x, y) middle coordinate position on the column coordinate is recorded as Δ L (x, y), and Δ L (x, y) is (DR)0(x,y)-DRj(x, y))/2; also, each pixel point and SR in the right view image of the new stereoscopic image are calculated0The relation between the corresponding pixel points on the column coordinates, the pixel point with the coordinate position of (x, y) in the right viewpoint image of the new three-dimensional image and the SR0The relation between the pixel points with the (x, y) middle coordinate position on the column coordinate is recorded as Δ R (x, y), and Δ R (x, y) — (DR)0(x,y)-DRj(x, y))/2; wherein, DR0(x, y) denotes DR0The pixel value DR of the pixel point with the middle coordinate position (x, y)0(x, y) is also SR0And the middle coordinate position is the parallax value of the pixel point of (x, y).
Step 1_ 5: obtaining SL when j is more than or equal to 1 and less than or equal to N0The left viewpoint image obtained after redrawing is marked as SLj(ii) a And obtaining the SR0The right viewpoint image obtained after redrawing is marked as SRj(ii) a Then from SLjAnd SRjForm a redrawn three-dimensional image, denoted as Sj(ii) a Wherein SLjAnd SRjThe acquisition process comprises the following steps:
step 1_5 a: SR'jRepresents SR0If x + delta R (x, y) < W is more than or equal to 0, the right viewpoint image obtained after the first redrawingimageThen SR 'is calculated'jThe pixel value of the pixel point with the middle coordinate position of (x, y) is recorded as SR'j(x,y),SR'j(x,y)=SR0(x + Δ R (x, y), y); if x + Δ R (x, y) < 0 or x + Δ R (x, y) ≧ WimageThen SR'jDetermining the pixel point with the middle coordinate position (x, y) as a cavity pixel point; wherein, SR0(x + Δ R (x, y), y) represents SR0The middle coordinate position is the pixel value of the pixel point of (x + Δ R (x, y), y).
Step 1_5 b: SR 'judgment'jIf the hollow pixel point exists in the buffer, SR 'is carried out'jAnew noted SR'jThen, step 1_5c is performed; if the empty pixel exists, each empty pixel is processed, and SR 'is set'jIf the pixel point with the middle coordinate position (x, y) is a cavity pixel point, then x + DR is more than or equal to 00(x,y)+ΔR(x,y)<WimageCalculating the redrawn pixel value of the hole pixel point to be equal to SL0(x+DR0(x, y) + Δ R (x, y), y), when x + DR0(x, y) + Δ R (x, y) < 0 or x + DR0(x,y)+ΔR(x,y)≥WimageThe hollow pixel points are not redrawn, and an SR is obtained after all the hollow pixel points are processed0The right viewpoint image obtained after the second redrawing is marked as SR "jThen, step 1_5c is performed; wherein SL0(x+DR0(x, y) + Δ R (x, y), y) represents SL0The middle coordinate position is (x + DR)0(x, y) + Δ R (x, y), y).
Step 1_5 c: line of SL'jRepresents SL0If x-DR is not less than 0, the left viewpoint image obtained after the first redrawingj(x,y)<WimageThen calculate SL'jThe pixel value of the pixel point with the middle coordinate position of (x, y) is recorded as SL'j(x,y),SL'j(x,y)=SR”j(x-DRj(x, y), y); if x-DRj(x, y) < 0 or x-DRj(x,y)≥WimageThen SL 'is prepared'jDetermining the pixel point with the middle coordinate position (x, y) as a cavity pixel point; wherein, SR "j(x-DRj(x, y), y) represents SR "jThe middle coordinate position is (x-DR)j(x, y), y) pixel values of the pixels.
Step 1_5 d: judging SL'jIf there is a hole pixel point, if there is no hole pixel point, SL'jAnew as SL "jThen, step 1_5e is performed; if the signal exists, processing each hollow pixel point, and setting SL'jIf the pixel point with the middle coordinate position (x, y) is a cavity pixel point, then when x + delta L (x, y) < W is more than or equal to 0imageCalculating the redrawn pixel value of the hole pixel point to be equal to SL0(x + Δ L (x, y), y) when x + Δ L (x, y) < 0 or x + Δ L (x, y) ≧ WimageThe hollow pixel points are not redrawn, and SL is obtained after all the hollow pixel points are processed0The left viewpoint image obtained after the second redrawing is marked as SL "jThen, step 1_5e is performed; wherein SL0(x + Δ L (x, y), y) represents SL0The middle coordinate position is the pixel value of the pixel point of (x + Δ L (x, y), y).
Step 1_5 e: if SL "jIf no hollow pixel point exists in the interface, SL is used "jAs SLj(ii) a If SR'jIf no hollow pixel exists in the image, then SR is carried out "jAs SRj(ii) a If SL "jAnd SR'jIf there are still hollow pixel points, then use OpenCV library or other existing image repairing technique to repair each hollow pixel point, and use SL "jThe left viewpoint image obtained after the patching is used as SLjIs prepared from SR "jThe right viewpoint image obtained after patching is used as SRj
As shown in fig. 3, the specific steps of the modeling part of the objective evaluation model for the visual satisfaction of the stereoscopic image are as follows:
step 2_ 1: selecting M pieces of width WimageAnd has a height HimageThe stereo images of (a) form a stereo image set; wherein M is a positive integer, and M is more than or equal to 20.
Step 2_ 2: defining the ith stereo image to be processed currently in the stereo image set as a current stereo image; wherein i is a positive integer, and i is more than or equal to 1 and less than or equal to M.
Step 2_ 3: recording the current stereo image as phii,0(ii) a Then changing the value of j from 1 to N, repeatedly executing the j values for 1 time in the same way according to the processes from step 1_1 to step 1_5, and repeatedly executing the j values from 1 to N for N times to obtain phii,0Corresponding N redrawing three-dimensional images, and obtaining phi when the number of translation steps is ji,0The corresponding redrawn stereo image is recorded as phii,jAt the same time obtainTo phii,0A matrix VD of viewing distances of all points in the formed virtual scenei,0Viewing distances of all points in the new virtual scene corresponding to each redrawn stereo image are formed into a matrix, and phi is calculatedi,jThe matrix formed by the viewing distances of all the points in the corresponding new virtual scene is recorded as VDi,j
Step 2_ 4: taking the next to-be-processed stereo image in the stereo image set as a current stereo image by setting i to i + 1; then, returning to the step 2_3 to continue executing until N redrawn three-dimensional images corresponding to each three-dimensional image in the three-dimensional image set are obtained, and forming a training set by all three-dimensional images in the three-dimensional image set and the obtained M multiplied by N redrawn three-dimensional images; wherein, in i +1, "═ is an assigned symbol.
Step 2_ 5: obtaining an average subjective evaluation mean value of 3D visual satisfaction of each three-dimensional image in a training set; then, the set formed by the mean subjective evaluation average of the 3D visual satisfaction of all the stereo images in the training set is recorded as { VSMOSi,jI is more than or equal to 1 and less than or equal to M, and j is more than or equal to 0 and less than or equal to N }; VSMOS when j is 0i,0Represents phii,0The average subjective evaluation mean value of 3D visual satisfaction is that when j is more than or equal to 1 and less than or equal to N, VSMOSi,jPhi obtained when the number of steps representing translation is ji,0Corresponding redrawn stereo image phii,jThe average subjective evaluation mean value of the 3D visual satisfaction degree comprises the comprehensive effect of the visual comfort degree and the perception depth.
In this embodiment, in step 2_5, the specific process of obtaining the average subjective evaluation mean of the 3D visual satisfaction of each stereo image in the training set includes:
step 2_5 a: more than 10 subjects including gender were selected to participate in the subjective experiment, each subject's visual system satisfied stereoscopic vision less than 60 arcsec and passed the color test.
Step 2_5 b: making a distance between each subject and a 3D display for displaying a stereoscopic image 3 times a screen height of the 3D display, and enabling each subject to view the stereoscopic image displayed on the 3D display in head-up; and informing each subject of a scoring standard, wherein the scoring standard is two subjective visual perceptions of visual comfort and perception depth of the comprehensive stereo image, and scoring on the whole.
Step 2_5 c: enabling the 3D display to sequentially display each three-dimensional image in the training set, and displaying for 2-3 times in a co-circulation mode, enabling the display time of each three-dimensional image on the 3D display to be 5-10 seconds, enabling the 3D display to display the next three-dimensional image after the testee has a rest for 3-10 seconds, and enabling the 3D display to display a middle gray image during the rest of the testee; each subject looks up to view the stereoscopic images displayed on the 3D display, and when each stereoscopic image in the training set is displayed in turn on the last pass, each subject scores the 3D visual satisfaction level of the stereoscopic image just displayed on the 3D display during the rest period according to the absolute category rating in the international standards ITU-T p.910 and ITU-T p.911, 5 scores represent "very satisfactory", 4 scores represent "satisfactory", 3 scores represent "normal", 2 scores represent "unsatisfactory", and 1 score represents "very unsatisfactory".
Here, it is sufficient that (M +1) × N stereoscopic images are circularly displayed for 2-3 times, and if the number of times of circular display is too many, visual fatigue of a subject is easily caused, so that the problem of inaccuracy exists in the last time of scoring; the display time of each stereoscopic image on the 3D display may be determined according to the size of the stereoscopic image, and the display time of the large-sized stereoscopic image may be generally set to be longer, for example, to be 10 seconds.
Step 2_5 d: after each subject scored the 3D visual satisfaction rating of each stereoscopic image in the training set, the abnormal scores were removed and the normal scores were retained according to the screening method disclosed in international standard ITU-R bt.500-11.
Step 2_5 e: and taking the average value of all normal scores of the stereo images as the average subjective score average value of the 3D visual satisfaction of the stereo images aiming at each stereo image in the training set.
Step 2_ 6: calculating the objective evaluation value of the visual comfort degree of each three-dimensional image in the training set according to the objective evaluation model of the visual comfort degree; then look at all stereo images in the training setThe set of objective evaluation values of perceptual comfort is denoted as { VCA }i,jI is more than or equal to 1 and less than or equal to M, and j is more than or equal to 0 and less than or equal to N }; where, when j is 0, VCAi,0Represents phii,0The objective evaluation value of visual comfort is VCA when j is more than or equal to 1 and less than or equal to Ni,jPhi obtained when the number of steps representing translation is ji,0Corresponding redrawn stereo image phii,jThe visual comfort of the subject is evaluated objectively.
Here, the objective evaluation model of visual comfort was derived from the thesis: jung, h.sohn, s.lee, h.w.park, and y.m.ro. "compressing Visual display comfort of Stereoscopic Images using human attention Model", IEEE Transactions on Circuits & Systems for Video Technology, vol.23, No.12, pp.2077-2082,2013 (published in journal of circuit and system of Video Technology, volume 23, phase 12, page 2077 and 2082,2013, by the institute of electrical and electronics engineers, using the attention Model to predict the Visual comfort of Stereoscopic Images).
Step 2_ 7: according to the set { VDi,jI is more than or equal to 1 and less than or equal to M, j is more than or equal to 0 and less than or equal to N, and calculating the objective evaluation value of the perceived absolute distance of each three-dimensional image in the training set; then, the set of objective evaluation values of perceived absolute distance of all stereo images in the training set is recorded as { APDAi,jI is more than or equal to 1 and less than or equal to M, and j is more than or equal to 0 and less than or equal to N }; wherein, VD is when j is more than or equal to 1 and less than or equal to Ni,jPhi obtained when the number of steps representing translation is ji,0Corresponding redrawn stereo image phii,jAPDA when j is 0i,0Represents phii,0The perceived absolute distance objective evaluation value of (1) is more than or equal to j and less than or equal to N, and APDAi,jPhi obtained when the number of steps representing translation is ji,0Corresponding redrawn stereo image phii,jThe perceived absolute distance objective evaluation value of (1).
In this embodiment, in step 2_7, APDAi,jThe acquisition process comprises the following steps:
step 2_7 a: for VDi,jThe values of all elements in the sequence are sorted from big to small or from small to big; then, the smallest top 1% value is removed from the values of all the sorted elements; finding out the minimum value from the values of all the elements left, and recording the minimum value as
Figure BDA0002271231850000181
Step 2_7 b: computing
Figure BDA0002271231850000182
Step 2_ 8: according to the set { VDi,jI is more than or equal to 1 and less than or equal to M, j is more than or equal to 0 and less than or equal to N, and calculating the objective evaluation value of the perception relative distance of each three-dimensional image in the training set; then, a set of objective evaluation values of perceptual relative distances of all stereo images in the training set is recorded as { RPDAi,jI is more than or equal to 1 and less than or equal to M, and j is more than or equal to 0 and less than or equal to N }; wherein, when j is 0, RPDAi,0Represents phii,0The perceived relative distance objective evaluation value of (1) is more than or equal to j and less than or equal to N, and the RPDA is used for evaluating the distancei,jPhi obtained when the number of steps representing translation is ji,0Corresponding redrawn stereo image phii,jThe perceived relative distance of (1) is objectively evaluated.
In this embodiment, in step 2_8, RPDAi,jThe acquisition process comprises the following steps:
step 2_8 a: for VDi,jThe values of all elements in the sequence are sorted from big to small or from small to big; then, the smallest top 1% value and the largest top 1% value are removed from the values of all the sorted elements; finding out the minimum and maximum values from the values of all the elements left, and recording the minimum and maximum values as corresponding values
Figure BDA0002271231850000183
And
Figure BDA0002271231850000184
step 2_8 b: computing
Figure BDA0002271231850000185
Wherein e represents a natural base number, Gi,jRepresenting the perceptual relative enhancement coefficient(s),
Figure BDA0002271231850000186
Figure BDA0002271231850000187
to represent
Figure BDA0002271231850000188
The perceived distance obtained by the conversion is converted,
Figure BDA0002271231850000189
Figure BDA00022712318500001810
to represent
Figure BDA00022712318500001811
The perceived distance obtained by the conversion is converted,
Figure BDA00022712318500001812
Figure BDA0002271231850000191
when j is 0
Figure BDA0002271231850000192
The perceived distance obtained by the conversion is converted,
Figure BDA0002271231850000193
Figure BDA0002271231850000194
when j is 0
Figure BDA0002271231850000195
The perceived distance obtained by the conversion is converted,
Figure BDA0002271231850000196
step 2_ 9: adopts a classical training model of epsilon-SVR in VSMOSi,jI is more than or equal to 1 and less than or equal to M, j is more than or equal to 0 and less than or equal to N and { VCAi,j|1≤i≤M,0≤j≤N}、{APDAi,j|1≤i≤M,0≤j≤N}、{RPDAi,jFitting is carried out between i is more than or equal to 1 and less than or equal to M and j is more than or equal to 0 and less than or equal to N to obtain the objective evaluation model of the visual satisfaction degree of the stereo image, and the description is as follows: STFA (VCA, APDA, RPDA); wherein, trainingThe kernel function of the model epsilon-SVR is a histogram interaction type, the width coefficient gamma is 1/256, the penalty coefficient C is 4, the insensitive loss coefficient epsilon is 0.1, STFA () is a function representation form of a stereoscopic image visual satisfaction objective evaluation model, VCA, APDA and RPDA are all input of the stereoscopic image visual satisfaction objective evaluation model, VCA is used for representing a visual comfort objective evaluation value of a stereoscopic image, APDA is used for representing a perception absolute distance objective evaluation value of the stereoscopic image, and RPDA is used for representing a perception relative distance objective evaluation value of the stereoscopic image.
The objective evaluation part of the visual satisfaction and the visual comfort of the stereo image comprises the following specific steps:
step 3_ 1: s is obtained in the same manner as the procedure of step 2_6jObjective evaluation value of visual comfort of (1), noted as VCAj(ii) a And obtains S in the same manner as the procedure of step 2_7jIs recorded as APDAj(ii) a S is obtained in the same manner as the procedure of step 2_8jThe objective evaluation value of the perceived relative distance is marked as RPDAj(ii) a Then VCAj、APDAj、RPDAjInputting the three-dimensional image visual satisfaction objective evaluation model obtained in the step 2-9 to obtain SjIs recorded as STFAj,STFAj=STFA(VCAj,APDAj,RPDAj) (ii) a Wherein j is more than or equal to 0 and less than or equal to N.
The specific steps of the optimal stereo image discrimination part are as follows:
step 4_ 1: directly performing step 4_2 when j is 0 or j is 1; when j is more than or equal to 2 and less than or equal to N, the pointer points to Sj-1Judging STFAj-1>STFAj-2、STFAj-1>STFAj、VCAj-1>VCA0Whether or not to satisfy simultaneously, if so, then Sj-1Is determined as S0The optimal stereo image with improved visual comfort is obtained; otherwise, directly executing the step 4_ 2; wherein, STFAj-1S obtained when the number of steps representing translation is j-10Corresponding redrawn three-dimensional image Sj-1Is the visual satisfaction value of (1), STFA when j is 2j-2Is thatSTFA0And when j is more than 2 and less than or equal to N, STFAj-2S obtained when the number of steps representing translation is j-20Corresponding redrawn three-dimensional image Sj-2A visual satisfaction value of.
Step 4_ 2: let j equal j +1, then repeat step 1_2 to step 1_5 and step 3_2 and step 4_ 1; wherein, j is the assigned symbol in j + 1.
For S0Setting the appropriate step length for redrawing, automatically calculating and storing the three-dimensional image S after each redrawingjObjective visual comfort evaluation value VCAjAnd visual satisfaction value STFAj(ii) a If STFAjIs a maximum point and VCAjGreater than VCA0Then the step number is not increased, and S is the samejIs the optimal stereo image, i.e. S0The optimal result after the visual comfort is improved; if STFAjIf the maximum value point is not reached, the step number continues to increase.
To further illustrate the feasibility and effectiveness of the method of the present invention, experiments were conducted on the method of the present invention.
Here, taking a stereo image database provided by korea scientific technical institute image and video systems laboratory (IVY LAB) as an example, the stereo image database includes 120 stereo images and parallax images of right viewpoint images of each stereo image, includes indoor and outdoor stereo images of various scene depths, and gives an average subjective evaluation mean of 3D visual satisfaction of each stereo image.
In this experiment, 22 stereo images in which the mean subjective evaluation of 3D visual satisfaction in the stereo image database was lower than 3 points were selected, and the numbers were: 2. 28, 29, 30, 32, 33, 35, 39, 46, 47, 49, 50, 51, 52, 53, 55, 70, 73, 74, 101, 102, 103. And respectively redrawing the 22 original stereo images based on the stereo images subjected to nonlinear translation based on the viewing distance, wherein the step length is set to be 100mm, the step number is 12 steps, obtaining 22 multiplied by 12 to 264 redrawn stereo images in total, and forming a stereo image set by 286 stereo images of the 22 original stereo images and the 264 redrawn stereo images in total.
In order to verify the performance of the stereoscopic image visual satisfaction objective evaluation model in the method, 18 original stereoscopic images and 234 redrawn stereoscopic images corresponding to the original stereoscopic images are randomly selected from a stereoscopic image set to form a training set, the remaining 4 original images in the stereoscopic image set and 52 redrawn stereoscopic images corresponding to the original stereoscopic images form a test set, and 7315 times of cross verification are needed. 4 common objective parameters of the evaluation method for evaluating the image quality are used as evaluation indexes, namely Pearson correlation coefficient (PLCC), Spearman correlation coefficient (SROCC), Mean Absolute Error (MAE) and Root Mean Square Error (RMSE) under the condition of nonlinear regression. PLCC reflects the correlation between objective evaluation results and subjective evaluation values; SROCC reflects monotonicity and consistency of objective evaluation results and subjective evaluation values; MAE and RMSE reflect the accuracy of the objective evaluation results. The method of the invention is utilized to obtain the visual satisfaction value of each three-dimensional image in the three-dimensional image set, and the five-parameter Logistic function nonlinear fitting is carried out on the visual satisfaction values of the 286 three-dimensional images, and the higher the PLCC and SROCC values are, and the smaller the MAE and RMSE values are, the better the correlation between the objective evaluation result of the three-dimensional image visual satisfaction objective evaluation model and the average subjective evaluation mean value is. Table 1 shows performance evaluation index values of the stereoscopic image visual satisfaction objective evaluation model.
TABLE 1 evaluation of performance index values
Performance evaluation index PLCC SROCC MAE RMSE
Index value 0.9439 0.9349 0.2643 0.3308
As can be seen from table 1, the objective evaluation result obtained by using the stereoscopic image visual satisfaction objective evaluation model is consistent with the result of subjective perception of human eyes, and the effectiveness of the stereoscopic image visual satisfaction objective evaluation model is well illustrated.

Claims (5)

1. A stereoscopic image visual comfort level improving method combining perceived depth quality is characterized by comprising four processes of stereoscopic image redrawing based on viewing distance nonlinear translation, modeling of a stereoscopic image visual satisfaction degree objective evaluation model, objective evaluation of the visual satisfaction degree and the visual comfort level of a stereoscopic image and optimal stereoscopic image discrimination;
the method comprises the following specific steps of:
step 1_ 1: will have a width of WimageAnd has a height HimageThe stereo image to be improved in visual comfort is recorded as S0Will S0The left view image and the right view image are correspondingly denoted as SL0And SR0(ii) a Then obtain the SR0Is recorded as DR0(ii) a Then calculating DR0The screen parallax value of each pixel point in (1) is calculated as DR0The screen parallax value of the pixel point with the middle coordinate position (x, y) is recorded as DRS0(x, y); then according to DR0Calculating S as the screen parallax value of each pixel point0The distance between each point in the formed virtual scene and the human eye, i.e. calculating S0Viewing distance, S, of each point in the formed virtual scene0Viewing of a point in the virtual scene formed with a coordinate position of (x, y)The distance is recorded as VD0(x,y),
Figure FDA0002271231840000011
Finally, the S is0The matrix formed by the viewing distances of all the points in the virtual scene is denoted as VD0(ii) a Wherein x represents the column coordinate of the pixel point, y represents the row coordinate of the pixel point, x is more than or equal to 0 and is more than Wimage,0≤y<HimageP represents a binocular pupillary distance of human eyes, and T represents a distance from the human eyes to a 3D display for displaying a stereoscopic image;
step 1_ 2: let j denote the number of translation steps, and the value range of j is more than or equal to 0 and less than or equal to N; s when j is 00All points in the formed virtual scene do not translate in the depth direction; when j is more than or equal to 1 and less than or equal to N, enabling S to be0The point with the minimum viewing distance in the formed virtual scene is translated by a distance of j multiplied by delta Z in the depth direction; then let S0All the points except the point with the minimum viewing distance in the formed virtual scene are subjected to nonlinear translation in the depth direction to obtain a new virtual scene, and then a matrix formed by the viewing distances of all the points in the new virtual scene is obtained and recorded as VDjWill VDjThe value of the element with the middle subscript position (x, y), that is, the viewing distance of the point with the coordinate position (x, y) in the new virtual scene is recorded as VDj(x,y),
Figure FDA0002271231840000021
Wherein j is a positive integer, N is a positive integer, and N belongs to [10,20 ]]And deltaz represents the step size of the translation,
Figure FDA0002271231840000022
Figure FDA0002271231840000023
represents VD0Minimum value of (1);
step 1_ 3: j is more than or equal to 1 and less than or equal to N according to VDjCalculating the screen parallax value of each pixel point in the parallax image of the right viewpoint image of the new three-dimensional image corresponding to the new virtual scene, and converting the right viewpoint image of the new three-dimensional image into the screen parallax valueThe screen parallax value of the pixel point with the coordinate position (x, y) in the parallax image is recorded as DRSj(x,y),
Figure FDA0002271231840000024
Then calculating the parallax image of the right viewpoint image of the new stereo image and recording the parallax image as DRjWill DRjThe pixel value of the pixel point with the middle coordinate position (x, y) is recorded as DRj(x,y),
Figure FDA0002271231840000025
DRj(x, y) is also the parallax value of the pixel point with the coordinate position (x, y) in the right viewpoint image of the new stereo image, and the screen width of the 3D display for displaying the stereo image is W3DAnd a screen height of the 3D display for displaying the stereoscopic image is H3D
Step 1_ 4: calculating each pixel point and SL in the left viewpoint image of the new stereoscopic image when j is more than or equal to 1 and less than or equal to N0The relation between the corresponding pixel points on the column coordinates, the pixel point with the coordinate position of (x, y) in the left viewpoint image of the new stereo image and the SL0The relation between the pixel points with the (x, y) middle coordinate position on the column coordinate is recorded as Δ L (x, y), and Δ L (x, y) is (DR)0(x,y)-DRj(x, y))/2; also, each pixel point and SR in the right view image of the new stereoscopic image are calculated0The relation between the corresponding pixel points on the column coordinates, the pixel point with the coordinate position of (x, y) in the right viewpoint image of the new three-dimensional image and the SR0The relation between the pixel points with the (x, y) middle coordinate position on the column coordinate is recorded as Δ R (x, y), and Δ R (x, y) — (DR)0(x,y)-DRj(x, y))/2; wherein, DR0(x, y) denotes DR0The pixel value DR of the pixel point with the middle coordinate position (x, y)0(x, y) is also SR0The middle coordinate position is the parallax value of the pixel point of (x, y);
step 1_ 5: obtaining SL when j is more than or equal to 1 and less than or equal to N0The left viewpoint image obtained after redrawing is marked as SLj(ii) a And obtaining the SR0The right viewpoint image obtained after redrawing is marked as SRj(ii) a Then from SLjAnd SRjForm a redrawn three-dimensional image, denoted as Sj(ii) a Wherein SLjAnd SRjThe acquisition process comprises the following steps:
step 1_5 a: SR'jRepresents SR0If x + delta R (x, y) < W is more than or equal to 0, the right viewpoint image obtained after the first redrawingimageThen SR 'is calculated'jThe pixel value of the pixel point with the middle coordinate position of (x, y) is recorded as SR'j(x,y),SR'j(x,y)=SR0(x + Δ R (x, y), y); if x + Δ R (x, y) < 0 or x + Δ R (x, y) ≧ WimageThen SR'jDetermining the pixel point with the middle coordinate position (x, y) as a cavity pixel point; wherein, SR0(x + Δ R (x, y), y) represents SR0The middle coordinate position is the pixel value of the pixel point of (x + delta R (x, y), y);
step 1_5 b: SR 'judgment'jIf the hollow pixel point exists in the buffer, SR 'is carried out'jAnew noted SR'jThen, step 1_5c is performed; if the empty pixel exists, each empty pixel is processed, and SR 'is set'jIf the pixel point with the middle coordinate position (x, y) is a cavity pixel point, then x + DR is more than or equal to 00(x,y)+ΔR(x,y)<WimageCalculating the redrawn pixel value of the hole pixel point to be equal to SL0(x+DR0(x, y) + Δ R (x, y), y), when x + DR0(x, y) + Δ R (x, y) < 0 or x + DR0(x,y)+ΔR(x,y)≥WimageThe hollow pixel points are not redrawn, and an SR is obtained after all the hollow pixel points are processed0The right viewpoint image obtained after the second redrawing is marked as SR "jThen, step 1_5c is performed; wherein SL0(x+DR0(x, y) + Δ R (x, y), y) represents SL0The middle coordinate position is (x + DR)0(x, y) + Δ R (x, y), y);
step 1_5 c: line of SL'jRepresents SL0If x-DR is not less than 0, the left viewpoint image obtained after the first redrawingj(x,y)<WimageThen calculate SL'jThe pixel value of the pixel point with the middle coordinate position of (x, y) is recorded as SL'j(x,y),SL'j(x,y)=SR”j(x-DRj(x, y), y); if x-DRj(x, y) < 0 or x-DRj(x,y)≥WimageThen SL 'is prepared'jDetermining the pixel point with the middle coordinate position (x, y) as a cavity pixel point; wherein, SR "j(x-DRj(x, y), y) represents SR "jThe middle coordinate position is (x-DR)j(x, y), y) pixel values of the pixels;
step 1_5 d: judging SL'jIf there is a hole pixel point, if there is no hole pixel point, SL'jAnew as SL "jThen, step 1_5e is performed; if the signal exists, processing each hollow pixel point, and setting SL'jIf the pixel point with the middle coordinate position (x, y) is a cavity pixel point, then when x + delta L (x, y) < W is more than or equal to 0imageCalculating the redrawn pixel value of the hole pixel point to be equal to SL0(x + Δ L (x, y), y) when x + Δ L (x, y) < 0 or x + Δ L (x, y) ≧ WimageThe hollow pixel points are not redrawn, and SL is obtained after all the hollow pixel points are processed0The left viewpoint image obtained after the second redrawing is marked as SL "jThen, step 1_5e is performed; wherein SL0(x + Δ L (x, y), y) represents SL0The middle coordinate position is the pixel value of the pixel point of (x + delta L (x, y), y);
step 1_5 e: if SL "jIf no hollow pixel point exists in the interface, SL is used "jAs SLj(ii) a If SR'jIf no hollow pixel exists in the image, then SR is carried out "jAs SRj(ii) a If SL "jAnd SR'jIf there are still void pixel points, then use image repairing technique to repair each void pixel point, and use SL "jThe left viewpoint image obtained after the patching is used as SLjIs prepared from SR "jThe right viewpoint image obtained after patching is used as SRj
The modeling part of the three-dimensional image visual satisfaction objective evaluation model comprises the following specific steps:
step 2_ 1: selecting M pieces of width WimageAnd has a height HimageThe stereo images of (a) form a stereo image set; wherein M is a positive integer, and M is more than or equal to 20;
step 2_ 2: defining the ith stereo image to be processed currently in the stereo image set as a current stereo image; wherein i is a positive integer, i is more than or equal to 1 and less than or equal to M;
step 2_ 3: recording the current stereo image as phii,0(ii) a Then changing the value of j from 1 to N, repeatedly executing the j values for 1 time in the same way according to the processes from step 1_1 to step 1_5, and repeatedly executing the j values from 1 to N for N times to obtain phii,0Corresponding N redrawing three-dimensional images, and obtaining phi when the number of translation steps is ji,0The corresponding redrawn stereo image is recorded as phii,jWhile obtaining phii,0A matrix VD of viewing distances of all points in the formed virtual scenei,0Viewing distances of all points in the new virtual scene corresponding to each redrawn stereo image are formed into a matrix, and phi is calculatedi,jThe matrix formed by the viewing distances of all the points in the corresponding new virtual scene is recorded as VDi,j
Step 2_ 4: taking the next to-be-processed stereo image in the stereo image set as a current stereo image by setting i to i + 1; then, returning to the step 2_3 to continue executing until N redrawn three-dimensional images corresponding to each three-dimensional image in the three-dimensional image set are obtained, and forming a training set by all three-dimensional images in the three-dimensional image set and the obtained M multiplied by N redrawn three-dimensional images; wherein, the' in the i-i +1 is an assignment symbol;
step 2_ 5: obtaining an average subjective evaluation mean value of 3D visual satisfaction of each three-dimensional image in a training set; then, the set formed by the mean subjective evaluation average of the 3D visual satisfaction of all the stereo images in the training set is recorded as { VSMOSi,jI is more than or equal to 1 and less than or equal to M, and j is more than or equal to 0 and less than or equal to N }; VSMOS when j is 0i,0Represents phii,0The average subjective evaluation mean value of 3D visual satisfaction is that when j is more than or equal to 1 and less than or equal to N, VSMOSi,jPhi obtained when the number of steps representing translation is ji,0Corresponding redrawn stereo image phii,jThe average subjective evaluation mean value of the 3D visual satisfaction degree comprises the comprehensive effect of the visual comfort degree and the perception depth;
step 2_ 6: calculating the objective evaluation value of the visual comfort degree of each three-dimensional image in the training set according to the objective evaluation model of the visual comfort degree; then look at all stereo images in the training setThe set of objective evaluation values of perceptual comfort is denoted as { VCA }i,jI is more than or equal to 1 and less than or equal to M, and j is more than or equal to 0 and less than or equal to N }; where, when j is 0, VCAi,0Represents phii,0The objective evaluation value of visual comfort is VCA when j is more than or equal to 1 and less than or equal to Ni,jPhi obtained when the number of steps representing translation is ji,0Corresponding redrawn stereo image phii,jThe objective evaluation value of the visual comfort of the user;
step 2_ 7: according to the set { VDi,jI is more than or equal to 1 and less than or equal to M, j is more than or equal to 0 and less than or equal to N, and calculating the objective evaluation value of the perceived absolute distance of each three-dimensional image in the training set; then, the set of objective evaluation values of perceived absolute distance of all stereo images in the training set is recorded as { APDAi,jI is more than or equal to 1 and less than or equal to M, and j is more than or equal to 0 and less than or equal to N }; wherein, VD is when j is more than or equal to 1 and less than or equal to Ni,jPhi obtained when the number of steps representing translation is ji,0Corresponding redrawn stereo image phii,jAPDA when j is 0i,0Represents phii,0The perceived absolute distance objective evaluation value of (1) is more than or equal to j and less than or equal to N, and APDAi,jPhi obtained when the number of steps representing translation is ji,0Corresponding redrawn stereo image phii,jThe perceived absolute distance objective evaluation value of (1);
step 2_ 8: according to the set { VDi,jI is more than or equal to 1 and less than or equal to M, j is more than or equal to 0 and less than or equal to N, and calculating the objective evaluation value of the perception relative distance of each three-dimensional image in the training set; then, a set of objective evaluation values of perceptual relative distances of all stereo images in the training set is recorded as { RPDAi,jI is more than or equal to 1 and less than or equal to M, and j is more than or equal to 0 and less than or equal to N }; wherein, when j is 0, RPDAi,0Represents phii,0The perceived relative distance objective evaluation value of (1) is more than or equal to j and less than or equal to N, and the RPDA is used for evaluating the distancei,jPhi obtained when the number of steps representing translation is ji,0Corresponding redrawn stereo image phii,jThe perceived relative distance objective evaluation value of (1);
step 2_ 9: using a training model of ε -SVR in VSMOSi,jI is more than or equal to 1 and less than or equal to M, j is more than or equal to 0 and less than or equal to N and { VCAi,j|1≤i≤M,0≤j≤N}、{APDAi,j|1≤i≤M,0≤j≤N}、{RPDAi,jFitting is carried out between i is more than or equal to 1 and less than or equal to M and j is more than or equal to 0 and less than or equal to N to obtain the objective evaluation model of the visual satisfaction degree of the stereo image, and the description is as follows: STFA (VCA, APDA, RPDA);the kernel function of the training model epsilon-SVR uses a histogram interaction type, a width coefficient gamma is 1/256, a penalty coefficient C is 4, an insensitive loss coefficient epsilon is 0.1, STFA () is a function representation form of a stereoscopic image visual satisfaction objective evaluation model, VCA, APDA and RPDA are all input of the stereoscopic image visual satisfaction objective evaluation model, VCA is used for representing a visual comfort objective evaluation value of a stereoscopic image, APDA is used for representing a perception absolute distance objective evaluation value of the stereoscopic image, and RPDA is used for representing a perception relative distance objective evaluation value of the stereoscopic image;
the objective evaluation part of the visual satisfaction and the visual comfort of the stereo image comprises the following specific steps:
step 3_ 1: s is obtained in the same manner as the procedure of step 2_6jObjective evaluation value of visual comfort of (1), noted as VCAj(ii) a And obtains S in the same manner as the procedure of step 2_7jIs recorded as APDAj(ii) a S is obtained in the same manner as the procedure of step 2_8jThe objective evaluation value of the perceived relative distance is marked as RPDAj(ii) a Then VCAj、APDAj、RPDAjInputting the three-dimensional image visual satisfaction objective evaluation model obtained in the step 2-9 to obtain SjIs recorded as STFAj,STFAj=STFA(VCAj,APDAj,RPDAj) (ii) a Wherein j is more than or equal to 0 and less than or equal to N;
the specific steps of the optimal stereo image discrimination part are as follows:
step 4_ 1: directly performing step 4_2 when j is 0 or j is 1; when j is more than or equal to 2 and less than or equal to N, the pointer points to Sj-1Judging STFAj-1>STFAj-2、STFAj-1>STFAj、VCAj-1>VCA0Whether or not to satisfy simultaneously, if so, then Sj-1Is determined as S0The optimal stereo image with improved visual comfort is obtained; otherwise, directly executing the step 4_ 2; wherein, STFAj-1S obtained when the number of steps representing translation is j-10Corresponding redrawn three-dimensional image Sj-1Is the visual satisfaction value of (1), STFA when j is 2j-2Is STFA0And when j is more than 2 and less than or equal to N, STFAj-2S obtained when the number of steps representing translation is j-20Corresponding redrawn three-dimensional image Sj-2A visual satisfaction value of;
step 4_ 2: let j equal j +1, then repeat step 1_2 to step 1_5 and step 3_2 and step 4_ 1; wherein, j is the assigned symbol in j + 1.
2. The method for improving the visual comfort of a stereoscopic image with combination of perceived depth quality as claimed in claim 1, wherein in step 1_1,
Figure FDA0002271231840000071
wherein, DR0(x, y) denotes DR0The pixel value DR of the pixel point with the middle coordinate position (x, y)0(x, y) is also SR0The parallax value of the pixel point with the middle coordinate position (x, y) and the screen width W of the 3D display for displaying the three-dimensional image3DAnd a screen height of the 3D display for displaying the stereoscopic image is H3D
3. The method according to claim 1 or 2, wherein in step 2_5, the specific process of obtaining the mean subjective evaluation of the 3D visual satisfaction of each stereoscopic image in the training set comprises:
step 2_5 a: selecting more than 10 testees containing gender of male and female to participate in subjective experiments, wherein the visual system of each tester meets the condition that the stereoscopic vision is less than 60 arcsec and passes the color test;
step 2_5 b: making a distance between each subject and a 3D display for displaying a stereoscopic image 3 times a screen height of the 3D display, and enabling each subject to view the stereoscopic image displayed on the 3D display in head-up; informing each subject of a scoring standard, wherein the scoring standard is two subjective visual perceptions of visual comfort and perception depth of the comprehensive stereo image, and scoring on the whole;
step 2_5 c: enabling the 3D display to sequentially display each three-dimensional image in the training set, and displaying for 2-3 times in a co-circulation mode, enabling the display time of each three-dimensional image on the 3D display to be 5-10 seconds, enabling the 3D display to display the next three-dimensional image after the testee has a rest for 3-10 seconds, and enabling the 3D display to display a middle gray image during the rest of the testee; each subject looks up to view the stereoscopic images displayed on the 3D display, and when each stereoscopic image in the training set is displayed in turn in the last pass, each subject scores the 3D visual satisfaction level of the stereoscopic image just displayed on the 3D display during the rest period according to the absolute category rating in the international standards ITU-T p.910 and ITU-T p.911, 5 points represent "very satisfactory", 4 points represent "satisfactory", 3 points represent "normal", 2 points represent "unsatisfactory", 1 point represents "very unsatisfactory";
step 2_5 d: after each subject scores the 3D visual satisfaction level of each stereo image in the training set, removing abnormal scores and keeping normal scores according to a screening method disclosed in the international standard ITU-R BT.500-11;
step 2_5 e: and taking the average value of all normal scores of the stereo images as the average subjective score average value of the 3D visual satisfaction of the stereo images aiming at each stereo image in the training set.
4. The method according to claim 3, wherein in step 2_7, APDA is used for improving the visual comfort of the stereoscopic image with the combination of perceived depth qualityi,jThe acquisition process comprises the following steps:
step 2_7 a: for VDi,jThe values of all elements in the sequence are sorted from big to small or from small to big; then, the smallest top 1% value is removed from the values of all the sorted elements; finding out the minimum value from the values of all the elements left, and recording the minimum value as
Figure FDA0002271231840000081
Step 2_7 b: computing
Figure FDA0002271231840000082
5. The method according to claim 3, wherein in step 2_8, RPDA is used for improving the visual comfort of the stereoscopic image with the combination of perceived depth qualityi,jThe acquisition process comprises the following steps:
step 2_8 a: for VDi,jThe values of all elements in the sequence are sorted from big to small or from small to big; then, the smallest top 1% value and the largest top 1% value are removed from the values of all the sorted elements; finding out the minimum and maximum values from the values of all the elements left, and recording the minimum and maximum values as corresponding values
Figure FDA0002271231840000091
And
Figure FDA0002271231840000092
step 2_8 b: computing
Figure FDA0002271231840000093
Wherein e represents a natural base number, Gi,jRepresenting the perceptual relative enhancement coefficient(s),
Figure FDA0002271231840000094
Figure FDA0002271231840000095
to represent
Figure FDA0002271231840000096
The perceived distance obtained by the conversion is converted,
Figure FDA0002271231840000097
Figure FDA0002271231840000098
to represent
Figure FDA0002271231840000099
The perceived distance obtained by the conversion is converted,
Figure FDA00022712318400000910
Figure FDA00022712318400000911
when j is 0
Figure FDA00022712318400000912
The perceived distance obtained by the conversion is converted,
Figure FDA00022712318400000913
Figure FDA00022712318400000914
when j is 0
Figure FDA00022712318400000915
The perceived distance obtained by the conversion is converted,
Figure FDA00022712318400000916
CN201911105732.6A 2019-11-13 2019-11-13 Stereoscopic image visual comfort level improving method combining perceived depth quality Active CN110944165B (en)

Priority Applications (1)

Application Number Priority Date Filing Date Title
CN201911105732.6A CN110944165B (en) 2019-11-13 2019-11-13 Stereoscopic image visual comfort level improving method combining perceived depth quality

Applications Claiming Priority (1)

Application Number Priority Date Filing Date Title
CN201911105732.6A CN110944165B (en) 2019-11-13 2019-11-13 Stereoscopic image visual comfort level improving method combining perceived depth quality

Publications (2)

Publication Number Publication Date
CN110944165A true CN110944165A (en) 2020-03-31
CN110944165B CN110944165B (en) 2021-02-19

Family

ID=69907465

Family Applications (1)

Application Number Title Priority Date Filing Date
CN201911105732.6A Active CN110944165B (en) 2019-11-13 2019-11-13 Stereoscopic image visual comfort level improving method combining perceived depth quality

Country Status (1)

Country Link
CN (1) CN110944165B (en)

Cited By (1)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN111669563A (en) * 2020-06-19 2020-09-15 福州大学 Stereo image visual comfort enhancement method based on reinforcement learning

Citations (5)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
WO2013042392A1 (en) * 2011-09-21 2013-03-28 シャープ株式会社 Three-dimensional image evaluation device
KR101350641B1 (en) * 2012-07-13 2014-01-14 서울과학기술대학교 산학협력단 Autostereoscopic 3-d image display evaluation method and autostereoscopic 3-d image display evaluation apparatus
CN108848365A (en) * 2018-05-31 2018-11-20 宁波大学 A kind of reorientation stereo image quality evaluation method
CN110246111A (en) * 2018-12-07 2019-09-17 天津大学青岛海洋技术研究院 Based on blending image with reinforcing image without reference stereo image quality evaluation method
CN110351548A (en) * 2019-06-27 2019-10-18 天津大学 Stereo image quality evaluation method based on deep learning and disparity map weighting guidance

Patent Citations (5)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
WO2013042392A1 (en) * 2011-09-21 2013-03-28 シャープ株式会社 Three-dimensional image evaluation device
KR101350641B1 (en) * 2012-07-13 2014-01-14 서울과학기술대학교 산학협력단 Autostereoscopic 3-d image display evaluation method and autostereoscopic 3-d image display evaluation apparatus
CN108848365A (en) * 2018-05-31 2018-11-20 宁波大学 A kind of reorientation stereo image quality evaluation method
CN110246111A (en) * 2018-12-07 2019-09-17 天津大学青岛海洋技术研究院 Based on blending image with reinforcing image without reference stereo image quality evaluation method
CN110351548A (en) * 2019-06-27 2019-10-18 天津大学 Stereo image quality evaluation method based on deep learning and disparity map weighting guidance

Cited By (2)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN111669563A (en) * 2020-06-19 2020-09-15 福州大学 Stereo image visual comfort enhancement method based on reinforcement learning
CN111669563B (en) * 2020-06-19 2021-06-25 福州大学 Stereo image visual comfort enhancement method based on reinforcement learning

Also Published As

Publication number Publication date
CN110944165B (en) 2021-02-19

Similar Documents

Publication Publication Date Title
Park et al. 3D visual discomfort prediction: vergence, foveation, and the physiological optics of accommodation
CN106097327B (en) In conjunction with the objective evaluation method for quality of stereo images of manifold feature and binocular characteristic
EP2345993A2 (en) Image processing apparatus, image processing method and program
CN104811693B (en) A kind of stereo image vision comfort level method for objectively evaluating
CN103986925B (en) based on the stereoscopic video visual comfort evaluation method of luminance compensation
CN105407349A (en) No-reference objective three-dimensional image quality evaluation method based on binocular visual perception
CN104509087B (en) Binocular vision experiences strengthening system
CN108520510B (en) No-reference stereo image quality evaluation method based on overall and local analysis
CN108449596A (en) A kind of 3D stereo image quality appraisal procedures of fusion aesthetics and comfort level
CN114648482A (en) Quality evaluation method and system for three-dimensional panoramic image
CN111882516B (en) Image quality evaluation method based on visual saliency and deep neural network
CN111641822B (en) Method for evaluating quality of repositioning stereo image
CN104361583A (en) Objective quality evaluation method of asymmetrically distorted stereo images
CN105488792B (en) Based on dictionary learning and machine learning without referring to stereo image quality evaluation method
CN110944165B (en) Stereoscopic image visual comfort level improving method combining perceived depth quality
CN103745457B (en) A kind of three-dimensional image objective quality evaluation method
CN108848365B (en) A kind of reorientation stereo image quality evaluation method
Kim et al. Visual comfort aware-reinforcement learning for depth adjustment of stereoscopic 3d images
Li et al. Stereoscopic 3D Image Retargeting Quality Assessment.
CN106682599A (en) Stereo image visual saliency extraction method based on sparse representation
CN110728630A (en) Internet image processing method based on augmented reality and augmented reality glasses
CN110944166B (en) Objective evaluation method for stereoscopic image visual satisfaction
CN105430370B (en) Stereoscopic image visual comfort evaluation method based on ranking learning
Liu et al. Learning-based quality assessment of retargeted stereoscopic images
Su et al. A visual comfort assessment approach of stereoscopic images based on random forest regressor

Legal Events

Date Code Title Description
PB01 Publication
PB01 Publication
SE01 Entry into force of request for substantive examination
SE01 Entry into force of request for substantive examination
GR01 Patent grant
GR01 Patent grant