CN104243970A - 3D drawn image objective quality evaluation method based on stereoscopic vision attention mechanism and structural similarity - Google Patents

3D drawn image objective quality evaluation method based on stereoscopic vision attention mechanism and structural similarity Download PDF

Info

Publication number
CN104243970A
CN104243970A CN201310565886.XA CN201310565886A CN104243970A CN 104243970 A CN104243970 A CN 104243970A CN 201310565886 A CN201310565886 A CN 201310565886A CN 104243970 A CN104243970 A CN 104243970A
Authority
CN
China
Prior art keywords
image
quality evaluation
window
saliency
matching
Prior art date
Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
Pending
Application number
CN201310565886.XA
Other languages
Chinese (zh)
Inventor
张冬冬
黄佳禾
臧笛
刘典
陈艳毓
Current Assignee (The listed assignees may be inaccurate. Google has not performed a legal analysis and makes no representation or warranty as to the accuracy of the list.)
Tongji University
Original Assignee
Tongji University
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Application filed by Tongji University filed Critical Tongji University
Priority to CN201310565886.XA priority Critical patent/CN104243970A/en
Publication of CN104243970A publication Critical patent/CN104243970A/en
Pending legal-status Critical Current

Links

Landscapes

  • Testing, Inspecting, Measuring Of Stereoscopic Televisions And Televisions (AREA)
  • Image Analysis (AREA)

Abstract

The invention discloses a 3D drawn image objective quality evaluation method based on a stereoscopic vision attention mechanism and structural similarity and relates to the technical field of image quality evaluation. The position relation between an original reference image and a 3D drawn image is obtained in a matching mode through an SIFT (scale-invariant feature transform) model, and multiple concrete distortions such as the compression distortion, the transmission distortion and the drawing distortion are measured through an SSIM model. The stereoscopic vision attention mechanism of a human eye vision system is reflected by drawing a 3D saliency map of the image through calculation, the saliency average value of all SIFT matching windows is worked out to serve as the saliency factor, the saliency factor serves as the weight of each window to perform weighted summation on all SSIM values, and then the final quality evaluation result is obtained. The result shows that the objective quality evaluation method is highly relevant to the subjective feeling.

Description

Objective quality evaluation method of 3D (three-dimensional) drawing image based on stereoscopic vision attention mechanism and structural similarity
Technical Field
The invention relates to the technical field of image quality evaluation.
Technical Field
Image quality evaluation methods can be roughly classified into two types: the first category is subjective quality assessment methods, and many standards have been established so far, such as in document 1 (see International Telecommunication Union (ITU) Radio communication Sector: 'method for the objective assessment of the quality of images', ITU-R BT.500-11, January 2002). These standards require a common observer without prior experience, in the context of which the image to be evaluated (and in some cases the original image as a reference) is observed and then scored for quality. However, the subjective quality evaluation method requires a strict test environment, a considerable number of observers, and time and labor consumption, so that the method cannot be applied in real time and is often used for research in laboratories.
The second category is objective quality assessment methods, which use algorithms instead of humans as the quality criteria for assessment. The Video Quality Experts Group (VQEG) to which the international telecommunications union belongs proposes 3 types of Video Quality testing methods, i.e., Video Quality metrics of full reference, partial reference, and no reference data. The full reference method provides all original video for reference, and the quality of the distorted video is obtained by comparing the distorted video with the original video. The partial reference evaluation method only needs to use partial original sequence information of the video and judge the quality of the video by comparing the characteristics of the video. In the non-reference evaluation method, no reference video can be provided, and the quality of data is judged by analyzing the data of the video to be tested, so that the method is difficult to realize, and the research result is less.
Currently, most image quality assessment studies focus on full-reference and partial-reference assessment methods. For example, in document 2 (see z. Wang, a.c. Bovik, h.r. Sheikh and e.simocell, "IQA: From error visibility to Structural Similarity", IEEE trans. Image Process, vol. 13, No. 4, pp. 600 + 612, apr. 2004.), a Structural Similarity (SSIM) model is proposed, which simulates the theory of Structural Similarity in the human visual system as a whole and can effectively evaluate the perceived quality of a 2D Image by using the measure of Structural information as an approximation of the perceived quality of the Image. Document 3 (c.t.e.r. headwage, s.t. world, s. dog, et al, "Quality evaluation of color plus depth map-based stereoscopic video", IEEE Journal of Selected videos in Signal Processing, 2009, 3(2): 304-318.) evaluates the Quality of left and right viewpoint videos of a stereoscopic video represented by a color video plus depth video based on the SSIM model. The method is based on the premise that the drawn virtual viewpoint video has an original reference video, and the quality evaluation model is designed. In practical application, an original reference video of a virtual viewpoint often does not exist, and the model cannot be applied.
With the development of the visual attention mechanism research in recent years, scholars propose various visual attention detection models, such as: document 4 (L. Itti, C. Koch, E. Niebur, et al, "" A model of saliency-based visual engagement for rapid scene analysis "".IEEE Trans. on Pattern Analysis and Machine IntelligenceITTI model of vol.20, No.11, pp. 1254-1259, 1998), reference 5 (X. Hou, L. Zhang, "clinical detection: A spectral residual approach"IEEE Conference on Computer Vision and Pattern Recognition,2007. CVPR'07, 2007: 1-8), document 6 (j. Harel, c. Koch, p. Perona, "Graph-based visual representation",Advances in neural information processing systems. 2006: 545-552) GBVS (Graph-based visual representation). Since visual attention is an important influence factor of image quality evaluation, there have been started trainees who add a visual attention mechanism to the process of quality evaluation, for example, document 7 (gazing, anping, gazing, zhuyu, zhangyang, stereoscopic image quality evaluation of visual attention [ J]The China graphic image newspaper 2012, 17(006): 722) 725) detects texture maps and depth maps of two paths of images of a stereo image respectively based on an ITTI model to extract interested areas, performs weighted quality evaluation on the left path of image and the right path of image according to the interested areas, and takes the average value of the quality evaluation results of the two paths of images as the final evaluation result of the stereo image. The method also assumes that two paths of images of the stereo image have original reference images.
Since many existing 3D cameras usually only acquire a color image of one view angle and a depth image corresponding to the color image, another view angle image required for synthesizing a stereoscopic image is often generated by drawing using a dibr (depth image based rendering) technology. In this case, the quality of the 3D rendering image is evaluated without an original rendering view angle image being available for reference, and the original color image and the 3D rendering image collected by the 3D camera available for reference are not a single view angle, and have a certain parallax. This situation renders the conventional 3D rendering image quality evaluation method useless. It is therefore necessary to design a new objective quality assessment scheme for 3D rendered images.
Disclosure of Invention
The invention provides a novel objective quality evaluation method of a 3D drawing image by combining a stereoscopic vision attention mechanism, the position relation of an original reference image and the 3D drawing image is matched through an SIFT (Scale-invariant feature transform) model, and various specific distortions such as compression distortion, transmission distortion, drawing distortion and the like are measured by using an SSIM (small Scale-invariant feature transform) model. The three-dimensional visual attention mechanism of a human visual system is reflected by calculating a 3D saliency map of a drawn image, the mean value of the saliency of each SIFT matching window is calculated as a saliency factor, and the mean value is used as the weight of each window to perform weighted summation on all SSIM values to obtain a final evaluation quality result.
Therefore, the implementation steps of the technical scheme provided by the invention are as follows:
A3D rendering image objective quality evaluation method based on a stereoscopic vision attention mechanism adopts the following technical scheme, and comprises the following steps:
step S1: and drawing the original color image and the original depth image into a reference virtual viewpoint image for reference.
Step S2: SIFT points are respectively extracted from the original color image and the reference virtual viewpoint image.
Step S3: and (5) establishing a matching relation between the SIFT points by using the characteristic values of the SIFT points obtained in the step (S2) to form a matching point pair. In order to reduce the error, an abnormal point removing operation is performed.
Step S4: and carrying out saliency detection on the original color image and the depth image by using the stereoscopic vision attention model to obtain a saliency map of the stereoscopic image.
Step S5: taking windows from the original color image and the 3D rendered image to be measured, respectively, with the pixel position of the SIFT matching point pair obtained in step S3 as the center. Only one pixel window is reserved with overlapping pixels, and repeated calculation is avoided.
Step S6: SSIM calculation is performed in the matching window obtained in step S5 to measure various distortions in the window.
Step S7: the weight value of the matching window pair obtained in step S5, that is, the weight of the SSIM value in this window pair, is defined as the average value of the saliency values of the window in the original color image obtained in step S4, divided by the average value of the ownership weight values.
Step S8: weighted averaging is performed on all SSIM values obtained in step S6.
The key technical points reflected by the technical scheme are as follows:
1. in many practical 3D rendered images, there is generally a small displacement between the 3D rendered image to be measured and the original color image, which makes the conventional objective quality evaluation method ineffective. The invention proposes to use a SIFT model to establish a matching relationship between an original color image and a 3D rendered image. Each pair of SIFT points marks that the two pixels are at the same position in the actual scene.
2. The invention considers the stereoscopic vision attention mechanism of the human eye vision system in the process of objective quality evaluation. And calculating a 3D saliency map of the drawn image, using the average value of the saliency of each window as a saliency factor, and using the final saliency factor as the weight of each SSIM value. The method can effectively reflect the three-dimensional attention mechanism of the human eye vision system, so that the final objective quality evaluation result is more matched with the perception result of the human eye vision system.
The method has the beneficial effects that: the SIFT model is used for establishing the matching relation between the original color image and the 3D drawn image, so that the condition that the traditional objective quality evaluation method is ineffective is avoided. Meanwhile, the invention also considers the stereoscopic vision attention mechanism, so that the objective evaluation quality is more matched with the human visual system. The result shows that the objective quality evaluation method provided by the invention has high correlation with subjective feeling.
Drawings
FIG. 1 is a block diagram of the objective quality assessment method of 3D rendered images based on the stereoscopic vision attention mechanism and structural similarity of the present invention.
Fig. 2 is an original color image of middle of an example of the present invention.
FIG. 3 is the original depth image of the middle example of the present invention.
Fig. 4 is a 3D rendering image to be evaluated of the invention example middle.
Fig. 5 is a perspective saliency map result of the example middle of the present invention.
Detailed Description
The invention is further illustrated by the following specific examples in conjunction with the accompanying drawings:
the example provided by the invention adopts MATLAB R2010a as a simulation experiment platform, takes a bmp grayscale image middle of 1396 × 1110 as a selected test image, and is described in detail below by combining each step:
step (1), the original color and original depth images of 1396 × 1110, bmp format are selected as input images, which are rendered as reference virtual viewpoint images for reference.
Step (2), SIFT points are respectively extracted from the original color image and the reference virtual viewpoint image:
and (3) establishing a matching relation between the SIFT points by using the characteristic values of the SIFT points obtained in the step (2) to form a matching point pair. In order to reduce the error, an abnormal point removing operation is performed. When the following formula is satisfied, the pair of points is considered as an abnormal point:
wherein, N is the number of SIFT matching point pairs before removing the abnormality.Is the firstThe physical distance of the matching point pairs, i.e. the physical distance of the positions of the two points in the original color image and the reference virtual image.Andthe mean and standard deviation, respectively, of all distances.
And (4) carrying out significance detection on the original color image and the depth image by using the stereoscopic vision attention model to obtain a significance map of the stereoscopic image. Through a traditional 2D saliency map method, an original color map and a corresponding depth map are utilized to jointly find a saliency map of a stereoscopic image.
In the example, the image is subjected to saliency detection by using a spectral residual method and a GBVS method, and the stereo saliency mapIs the combination of the two:
wherein,is a saliency map obtained from a depth map using the GBVS method,is a saliency map obtained from a color map using a spectral residual method.Is thatThe weight of (a) is determined,is thatIn this example, we set both to 0.5.
And (5) taking windows from the original color image and the 3D drawn image to be detected respectively by taking the pixel position of the SIFT matching point pair obtained in the step (3) as the center. Only one of the overlapping windows remains. In this example, we set the window size to 11 × 11.
And (6) carrying out SSIM calculation in the matching window obtained in the step (5) and measuring distortion in the window. We consider a matching pair of windows as signal X, signal Y, respectively.
Are the average luminance of the signals X and Y respectively,are the standard deviations of the signals X and Y respectively,is the signal difference between signals X and Y.,Set small enough to avoid a denominator of 0.
Step (7), the weight value of the matching window pair, that is, the weight of the SSIM value in this window pair, is defined as the correlation value of the average significant value of the window taken in the original color image on the stereo significant map obtained in step (4):
is the firstThe average of the pairs of matching windows on the stereoscopic saliency map.Is the average of the average saliency values for all matching window pairs. And N is the number of the matching window pairs obtained in the step (5).
Step (8) of averaging all SSIM values obtained in step (6) with weights to obtain final objective quality
Wherein N is the number of matching windows after the outliers are removed,andare respectively the firstSSIM values and weights for the individual matching window pairs.
The objective quality of the 3D drawn image is obtained by integrating all the steps, the reference environment, the stereoscopic vision attention mechanism and the like in practical application are considered comprehensively by the threshold, and the result matched with the human eye vision system is obtained.

Claims (1)

1. A3D rendering image objective quality evaluation method based on a stereoscopic vision attention mechanism is characterized by comprising the following steps:
step S1: drawing the original color image and the original depth image into a reference virtual viewpoint image for reference;
step S2: SIFT points are respectively extracted from the original color image and the reference virtual viewpoint image;
step S3: establishing a matching relationship between the SIFT points by using the characteristic values of the SIFT points obtained in the step S2 to form matching point pairs, and performing abnormal point removing work for reducing errors;
step S4: carrying out saliency detection on the original color image and the depth image by using a stereoscopic vision attention model to obtain a saliency map of the stereoscopic image;
step S5: taking the pixel position of the SIFT matching point pair obtained in the step S3 as a center, respectively taking windows from the original color image and the 3D drawn image to be detected, and only one window with overlapped pixels is reserved to avoid repeated calculation;
step S6: performing SSIM calculation in the matching window obtained in the step S5 to measure various distortions in the window;
step S7: the weight value of the matching window pair obtained in step S5, that is, the weight of the SSIM value in this window pair, is defined as the average value of the significant values of the window taken in the original color image on the stereoscopic significant image obtained in step S4 divided by the average value of the ownership weight values;
step S8: weighted averaging is performed on all SSIM values obtained in step S6.
CN201310565886.XA 2013-11-14 2013-11-14 3D drawn image objective quality evaluation method based on stereoscopic vision attention mechanism and structural similarity Pending CN104243970A (en)

Priority Applications (1)

Application Number Priority Date Filing Date Title
CN201310565886.XA CN104243970A (en) 2013-11-14 2013-11-14 3D drawn image objective quality evaluation method based on stereoscopic vision attention mechanism and structural similarity

Applications Claiming Priority (1)

Application Number Priority Date Filing Date Title
CN201310565886.XA CN104243970A (en) 2013-11-14 2013-11-14 3D drawn image objective quality evaluation method based on stereoscopic vision attention mechanism and structural similarity

Publications (1)

Publication Number Publication Date
CN104243970A true CN104243970A (en) 2014-12-24

Family

ID=52231194

Family Applications (1)

Application Number Title Priority Date Filing Date
CN201310565886.XA Pending CN104243970A (en) 2013-11-14 2013-11-14 3D drawn image objective quality evaluation method based on stereoscopic vision attention mechanism and structural similarity

Country Status (1)

Country Link
CN (1) CN104243970A (en)

Cited By (6)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN106097256A (en) * 2016-05-31 2016-11-09 南京邮电大学 A kind of video image fuzziness detection method based on Image Blind deblurring
CN106341677A (en) * 2015-07-07 2017-01-18 中国科学院深圳先进技术研究院 Virtual viewpoint video quality evaluation method
CN106469448A (en) * 2015-06-26 2017-03-01 康耐视公司 Carry out automatic industrial inspection using 3D vision
CN107122787A (en) * 2017-02-14 2017-09-01 北京理工大学 A kind of image scaling quality evaluating method of feature based fusion
CN110276082A (en) * 2019-06-06 2019-09-24 百度在线网络技术(北京)有限公司 Translation processing method and device based on dynamic window
CN113724182A (en) * 2020-05-21 2021-11-30 无锡科美达医疗科技有限公司 No-reference video quality evaluation method based on expansion convolution and attention mechanism

Citations (1)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN102073985B (en) * 2010-12-23 2012-05-09 清华大学 Method and device for objectively evaluating scaled image quality by matching pixel points

Patent Citations (1)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN102073985B (en) * 2010-12-23 2012-05-09 清华大学 Method and device for objectively evaluating scaled image quality by matching pixel points

Non-Patent Citations (2)

* Cited by examiner, † Cited by third party
Title
MARC DECOMBAS ET AL.: "A New Object Based Quality Metric Based On SIFT And SSIM", 《2012 IEEE INTERNATIONAL CONFERENCE ON IMAGE PROCESSING (ICIP 2012)》 *
杨亚洲: "基于局部不变特征的图像质量评价", 《计算机应用》 *

Cited By (9)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN106469448A (en) * 2015-06-26 2017-03-01 康耐视公司 Carry out automatic industrial inspection using 3D vision
CN106469448B (en) * 2015-06-26 2021-07-30 康耐视公司 Automated industrial inspection with 3D vision
CN106341677A (en) * 2015-07-07 2017-01-18 中国科学院深圳先进技术研究院 Virtual viewpoint video quality evaluation method
CN106341677B (en) * 2015-07-07 2018-04-20 中国科学院深圳先进技术研究院 Virtual view method for evaluating video quality
CN106097256A (en) * 2016-05-31 2016-11-09 南京邮电大学 A kind of video image fuzziness detection method based on Image Blind deblurring
CN106097256B (en) * 2016-05-31 2019-01-01 南京邮电大学 A kind of video image fuzziness detection method based on Image Blind deblurring
CN107122787A (en) * 2017-02-14 2017-09-01 北京理工大学 A kind of image scaling quality evaluating method of feature based fusion
CN110276082A (en) * 2019-06-06 2019-09-24 百度在线网络技术(北京)有限公司 Translation processing method and device based on dynamic window
CN113724182A (en) * 2020-05-21 2021-11-30 无锡科美达医疗科技有限公司 No-reference video quality evaluation method based on expansion convolution and attention mechanism

Similar Documents

Publication Publication Date Title
CN105389554B (en) Living body determination method and equipment based on recognition of face
CN109118470B (en) Image quality evaluation method and device, terminal and server
CN105279372B (en) A kind of method and apparatus of determining depth of building
CN101996407B (en) Colour calibration method for multiple cameras
CN102665086B (en) Method for obtaining parallax by using region-based local stereo matching
CN109523506B (en) Full-reference stereo image quality objective evaluation method based on visual salient image feature enhancement
CN107578403A (en) The stereo image quality evaluation method of binocular view fusion is instructed based on gradient information
CN104243970A (en) 3D drawn image objective quality evaluation method based on stereoscopic vision attention mechanism and structural similarity
CN101610425B (en) Method for evaluating stereo image quality and device
CN109345502B (en) Stereo image quality evaluation method based on disparity map stereo structure information extraction
CN109255358B (en) 3D image quality evaluation method based on visual saliency and depth map
CN107274483A (en) A kind of object dimensional model building method
CN105654142B (en) Based on natural scene statistics without reference stereo image quality evaluation method
CN110189294B (en) RGB-D image significance detection method based on depth reliability analysis
CN109242834A (en) It is a kind of based on convolutional neural networks without reference stereo image quality evaluation method
CN105404888A (en) Saliency object detection method integrated with color and depth information
CN111597933B (en) Face recognition method and device
WO2022126674A1 (en) Method and system for evaluating quality of stereoscopic panoramic image
CN107590444A (en) Detection method, device and the storage medium of static-obstacle thing
CN107016698A (en) Based on tapered plane smooth binocular solid matching process and device
CN112802081A (en) Depth detection method and device, electronic equipment and storage medium
Jeong et al. Visual comfort assessment of stereoscopic images using deep visual and disparity features based on human attention
CN104038752B (en) Multi-view point video rectangular histogram color correction based on three-dimensional Gaussian mixed model
CN109978928B (en) Binocular vision stereo matching method and system based on weighted voting
CN108062765A (en) Binocular image processing method, imaging device and electronic equipment

Legal Events

Date Code Title Description
C06 Publication
PB01 Publication
C10 Entry into substantive examination
SE01 Entry into force of request for substantive examination
C02 Deemed withdrawal of patent application after publication (patent law 2001)
WD01 Invention patent application deemed withdrawn after publication

Application publication date: 20141224