CN108391121B - No-reference stereo image quality evaluation method based on deep neural network - Google Patents

No-reference stereo image quality evaluation method based on deep neural network Download PDF

Info

Publication number
CN108391121B
CN108391121B CN201810375052.5A CN201810375052A CN108391121B CN 108391121 B CN108391121 B CN 108391121B CN 201810375052 A CN201810375052 A CN 201810375052A CN 108391121 B CN108391121 B CN 108391121B
Authority
CN
China
Prior art keywords
distorted
image
neural network
deep neural
quality
Prior art date
Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
Active
Application number
CN201810375052.5A
Other languages
Chinese (zh)
Other versions
CN108391121A (en
Inventor
陈志波
周玮
李卫平
Current Assignee (The listed assignees may be inaccurate. Google has not performed a legal analysis and makes no representation or warranty as to the accuracy of the list.)
University of Science and Technology of China USTC
Original Assignee
University of Science and Technology of China USTC
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Application filed by University of Science and Technology of China USTC filed Critical University of Science and Technology of China USTC
Priority to CN201810375052.5A priority Critical patent/CN108391121B/en
Publication of CN108391121A publication Critical patent/CN108391121A/en
Application granted granted Critical
Publication of CN108391121B publication Critical patent/CN108391121B/en
Active legal-status Critical Current
Anticipated expiration legal-status Critical

Links

Images

Classifications

    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04NPICTORIAL COMMUNICATION, e.g. TELEVISION
    • H04N17/00Diagnosis, testing or measuring for television systems or their details
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06TIMAGE DATA PROCESSING OR GENERATION, IN GENERAL
    • G06T7/00Image analysis
    • G06T7/0002Inspection of images, e.g. flaw detection
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06TIMAGE DATA PROCESSING OR GENERATION, IN GENERAL
    • G06T2207/00Indexing scheme for image analysis or image enhancement
    • G06T2207/10Image acquisition modality
    • G06T2207/10004Still image; Photographic image
    • G06T2207/10012Stereo images
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06TIMAGE DATA PROCESSING OR GENERATION, IN GENERAL
    • G06T2207/00Indexing scheme for image analysis or image enhancement
    • G06T2207/20Special algorithmic details
    • G06T2207/20021Dividing image into blocks, subimages or windows
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06TIMAGE DATA PROCESSING OR GENERATION, IN GENERAL
    • G06T2207/00Indexing scheme for image analysis or image enhancement
    • G06T2207/20Special algorithmic details
    • G06T2207/20081Training; Learning
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06TIMAGE DATA PROCESSING OR GENERATION, IN GENERAL
    • G06T2207/00Indexing scheme for image analysis or image enhancement
    • G06T2207/20Special algorithmic details
    • G06T2207/20084Artificial neural networks [ANN]
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06TIMAGE DATA PROCESSING OR GENERATION, IN GENERAL
    • G06T2207/00Indexing scheme for image analysis or image enhancement
    • G06T2207/30Subject of image; Context of image processing
    • G06T2207/30168Image quality inspection
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04NPICTORIAL COMMUNICATION, e.g. TELEVISION
    • H04N13/00Stereoscopic video systems; Multi-view video systems; Details thereof
    • H04N2013/0074Stereoscopic image analysis

Landscapes

  • Engineering & Computer Science (AREA)
  • Quality & Reliability (AREA)
  • Biomedical Technology (AREA)
  • General Health & Medical Sciences (AREA)
  • Multimedia (AREA)
  • Signal Processing (AREA)
  • Health & Medical Sciences (AREA)
  • Computer Vision & Pattern Recognition (AREA)
  • Physics & Mathematics (AREA)
  • General Physics & Mathematics (AREA)
  • Theoretical Computer Science (AREA)
  • Image Analysis (AREA)
  • Testing, Inspecting, Measuring Of Stereoscopic Televisions And Televisions (AREA)

Abstract

The invention discloses a no-reference stereo image quality evaluation method based on a deep neural network, which comprises the steps of simultaneously inputting left and right visual distortion image blocks into a double-flow deep neural network structure by considering the interaction principle of fusion between left and right vision and parallax information in a multilayer structure of a human visual system and using a no-reference 3D stereo image quality evaluation algorithm of the deep neural network based on end-to-end double-flow interaction, and combining discriminant feature extraction and regression learning into an end-to-end optimization process, thereby achieving the purpose of effectively predicting the perception quality of distorted stereo images.

Description

No-reference stereo image quality evaluation method based on deep neural network
Technical Field
The invention relates to the technical field of deep learning, in particular to a no-reference stereo image quality evaluation method based on a deep neural network.
Background
With the rapid popularization and development of 3D multimedia technologies including 3D movies, etc., 3D stereoscopic images have gradually entered people's daily lives, and viewing 3D stereoscopic images can create an immersive visual experience that 2D images lack. Meanwhile, due to the additional depth perception and asymmetric distortion existing between the left view image and the right view image, the 3D stereoscopic image quality evaluation is more challenging, namely a more complex binocular vision mechanism needs to be considered when a 3D stereoscopic image quality evaluation algorithm is designed.
Early full-reference 3D stereoscopic image quality evaluation algorithms originated from 2D image quality evaluation methods, such as article [1] (a. benoit, p.le call, p.campisi, and r.cousseau.quality assessment of stereoscopic images, 2008(1):659024,2009), and article [2] (j.you, l.xing, a.perkis, and x.wang.stereoscopic image for stereoscopic images based on 2D image quality measurements and analysis, of International workstation Video Processing and analysis, electronic devices, 2010, respectively, were used to evaluate the final image quality using the overall depth evaluation algorithm, 2010, USA, respectively. However, these depth learning methods for 2D images are not suitable for stereoscopic image quality evaluation because a binocular vision mechanism needs to be considered when evaluating stereoscopic image quality.
Later more sophisticated algorithms proposed the need to consider binocular visual characteristics of the human visual system, such as contrast masking effect [3] (p. gorley and n. hollliman. stereo Image quality metrics and compression. in proc. spie, volume 6803, page 680305,2008) and binocular binding behavior [4] (y. h. lin and j. l. wu. quality assessment of stereo 3D Image compression by binocular integration devices. ieee transformations on Image Processing,23(4): 1547-.
Since the original undistorted stereo image in the actual scene cannot be obtained usually, it is necessary to provide a non-reference 3D stereo image quality evaluation algorithm, which includes quality evaluation algorithms for specific distortion types and in a general sense, and mainly extracts discriminant features from the distorted stereo image manually based on human visual features, natural scene statistics, and the like, and then evaluates the perceptual quality of the stereo image through a Regression learning model such as Support Vector Regression (Support Vector Regression). Among them, the general non-reference stereo Image quality evaluation algorithm is more general as in article [5] (M.J.Chen, L.K.Cormac, and A.C.Bovik.No. reference quality assessment of national stereo images, IEEE Transactions on Image Processing,22(9): 3379-.
For the application of deep learning in quality evaluation, currently, much work is to apply deep learning to 2D image quality evaluation, and these technologies are mainly divided into training based on image blocks and training based on whole images according to a training strategy. The image block-based training method is to divide the original image into smaller image blocks and then to regress and learn the perceptual quality of each input image block through a deep neural network, such as the non-reference convolutional neural network (NR-CNN) [6] (L.Kang, P.Ye, Y.Li, and D.Doermann.Consortional neural network for no-reference image quality assessment. in Proceedings of the IEEE conference on Computer Vision and Pattern Recognition (CVPR), pages 1733-. Another training approach based on the whole image is to train to evaluate the quality of the whole image by aggregating the features of the merged image blocks or the prediction scores of each image block.
However, the performance of the above-mentioned article [5] and the no-reference convolutional neural network (NR-CNN) approach [6] is low.
Disclosure of Invention
The invention aims to provide a no-reference stereo image quality evaluation method based on a deep neural network.
The purpose of the invention is realized by the following technical scheme:
a no-reference stereo image quality evaluation method based on a deep neural network comprises the following steps:
dividing left and right view images forming all distorted stereo images into non-overlapping distorted image blocks respectively to obtain a plurality of left and right view distorted image block pairs;
correspondingly inputting the left and right visual distortion image block pairs obtained by division into a pre-constructed double-current input interactive deep neural network, training a network model, and obtaining the one-dimensional perception quality of the corresponding left and right visual distortion image block pairs through regression learning;
when the quality of the distorted stereo image is predicted, the quality of each left-view and right-view distorted image block pair is predicted by using a trained network model, and then the quality of all the left-view and right-view distorted image block pairs in each distorted stereo image is averaged to obtain the quality of each distorted stereo image.
According to the technical scheme provided by the invention, by considering the interaction principle of fusion between left and right views and parallax information in a multilayer structure of a human eye vision system and using a non-reference 3D (three-dimensional) image quality evaluation algorithm of a deep neural network based on end-to-end double-flow interaction, left and right view distorted image blocks are simultaneously input into the double-flow deep neural network structure, and the distinguishing characteristic extraction and regression learning are combined into an end-to-end optimization process, so that the aim of effectively predicting the perception quality of the distorted three-dimensional image is fulfilled.
Drawings
In order to more clearly illustrate the technical solutions of the embodiments of the present invention, the drawings needed to be used in the description of the embodiments are briefly introduced below, and it is obvious that the drawings in the following description are only some embodiments of the present invention, and it is obvious for those skilled in the art to obtain other drawings based on the drawings without creative efforts.
Fig. 1 is a flowchart of a method for evaluating quality of a non-reference stereo image based on a deep neural network according to an embodiment of the present invention;
fig. 2 is a flowchart of an embodiment of a method for evaluating quality of a reference-free stereo image based on a deep neural network according to the present invention;
FIG. 3 is a schematic diagram of interaction of a multi-layer network in a dual-stream input interactive deep neural network according to an embodiment of the present invention;
FIG. 4 is a distorted stereo image and its corresponding fusion image and difference image according to an embodiment of the present invention;
fig. 5 is a schematic diagram illustrating that the distortion quality of a local image block can be effectively evaluated according to the present invention;
fig. 6 is a process of optimizing the loss function for all distortions on the LIVE database according to the present invention.
Detailed Description
The technical solutions in the embodiments of the present invention are clearly and completely described below with reference to the drawings in the embodiments of the present invention, and it is obvious that the described embodiments are only a part of the embodiments of the present invention, and not all embodiments. All other embodiments, which can be derived by a person skilled in the art from the embodiments of the present invention without making any creative effort, shall fall within the protection scope of the present invention.
The embodiment of the invention provides a no-reference stereo image quality evaluation method based on a deep neural network, which mainly comprises the following steps as shown in figure 1:
step 1, dividing left and right view images forming all distorted stereo images into non-overlapping distorted image blocks respectively to obtain a plurality of left and right view distorted image block pairs.
In the embodiment of the invention, the non-overlapping distorted image blocks can be divided into the size of 32x32, and the amount of training data can be effectively increased by obtaining a plurality of left-view and right-view distorted image block pairs.
In addition, in the embodiment of the present invention, the quality of the distorted stereoscopic image is also used as the label value of the training sample of all left-right view distorted image block pairs obtained after the distorted stereoscopic image is divided.
And 2, correspondingly inputting the left and right visual distortion image block pairs obtained by division into a pre-constructed double-current input interactive deep neural network, training a network model, and obtaining the one-dimensional perception quality of the corresponding left and right visual distortion image block pairs through regression learning.
Because the visual cortex of human eyes is a layered structure and the interaction of left and right vision exists in a multilayer visual area, the left and right sub-networks in the network frame designed by the method realize the interaction in a plurality of convolution layers, namely, the fusion images and the differential images corresponding to the feature maps are connected, the fusion capability and the parallax information of the left and right vision are respectively represented, and finally, the full connection layers of the sub-networks are also interactively connected, so that the one-dimensional perception quality corresponding to the input distorted image block pair is obtained through regression learning.
And 3, when the quality of the distorted stereo image is predicted, predicting the quality of each left-view and right-view distorted image block pair by using the trained network model, and averaging the quality of all the left-view and right-view distorted image block pairs in each distorted stereo image to obtain the quality of each distorted stereo image.
Because uniform distortion exists, in the embodiment of the present invention, the distorted stereoscopic image to be finally predicted may be calculated by averaging the qualities of all left and right distorted image block pairs in the distorted stereoscopic image.
The skilled person can understand that after the left-view and right-view distorted image blocks corresponding to the 'composition of all distorted stereo images' are used for training the network model, the trained network model can be used for performing quality evaluation on the 'distorted stereo images to be predicted'; the "distorted stereo image to be predicted" may be the aforementioned "all distorted stereo images are composed", or may be other distorted stereo images; meanwhile, the "each left-right view distorted image block" mentioned in step 3 is the left-right view distorted image block obtained by dividing the "distorted stereoscopic image to be predicted" in the same manner as in step 1.
For the sake of understanding, the following description is further made with reference to the accompanying drawings.
Fig. 2 shows a flow chart of the present invention. The interactive deep neural network framework based on double-flow input can perform end-to-end training learning feature expression by respectively inputting left and right distorted stereo image block pairs and regress to obtain a perception quality score.
In an embodiment of the present invention, the dual-stream input interactive deep neural network includes two identical sub-networks, which respectively correspond to input left-view and right-view distorted image blocks, that is, visual information streams representing left and right views, where each sub-network is obtained by adjusting a first convolution layer, a second convolution layer, a third convolution layer (that is, a third convolution layer, a fourth convolution layer and a fifth convolution layer), a third convolution layer, and a first and a second fully connected layers, respectively, by using AlexNet (document krishevsky, i.e., sutsche, and g.e. hinton, "imaging with horizontal constraint network works," in advance in new information processing systems,2012, pp.1097-1105). And the second convolutional layer and the fifth convolutional layer (namely the last convolutional layer in the three convolutional layers) and the second fully-connected layer adopt the fusion interaction of two sub-networks, and finally a quality score is obtained.
During training, the divided left and right distorted stereo images are respectively input to the two sub-networks, the parameters are propagated in the forward direction and then updated in the backward direction, and the mean square difference between the predicted value and the true value is minimized, so that the learning network parameters are continuously updated in an iterative manner.
In the dual-stream input interactive deep neural network, two convolutional layers (the second convolutional layer and the last convolutional layer in the third convolutional layer) of two sub-networks and one fully-connected layer (the second fully-connected layer) correspond to each other to perform the interactive connection of the sub-networks, as shown in fig. 3.
The interaction of the convolution layer is to firstly perform the following operation on the characteristic diagram of the corresponding left-right view network to obtain a fused image S+Sum-difference image S-Re-connecting the fused image S+Sum-difference image S-
S+=Fl+Fr
S-=Fl-Fr
Wherein, Fl、FrThe left and right distortion image blocks are corresponding left and right view characteristic diagrams;
the interaction of the full-connection layer is to connect the last full-connection layer of each sub-network (namely, the second full-connection layer of each sub-network);
correspondingly inputting the left and right visual distortion image block pairs obtained by division into a pre-constructed double-flow input interactive deep neural network for forward propagation training, and finally calculating the Euclidean loss value according to a training sample label as a minimum objective function in the training process:
Figure BDA0001639400500000051
Figure BDA0001639400500000052
wherein, f (P)li,Pri(ii) a w) represents a left-right view distorted image block pair (P) under the parameter wli,Pri) W is a parameter of the double-current input interactive deep neural network, and needs to be updated iteratively, w' is an updated network parameter, F represents a 2 norm, yiIs a left-right view distorted image block pair (P)li,Pri) Corresponding training sample label values;
it is seen from fig. 4 that the fusion and difference images for different distortion types have discriminability, so that the effective quality features can be learned through the dual-stream input interactive deep neural network proposed by the present invention.
Through the double-current input interactive deep neural network learning iteration, a trained network model can be obtained, and due to uniform distortion, the quality of all left-view and right-view distorted image block pairs in each distorted stereo image can be averaged, so that the quality of each distorted stereo image is obtained, and the calculation formula is as follows:
Figure BDA0001639400500000061
wherein, f (P)li,Pri(ii) a w') denotes a left-right view distorted image block pair (P) under the parameter wli,Pri) P denotes the number of left and right view distorted image block pairs in each distorted stereoscopic image.
The scheme of the invention can effectively evaluate the distortion quality of the local image block, thereby effectively predicting the quality of the distorted three-dimensional image. In fig. 5, (a), (b), (c), and (d) are JPEG-compressed symmetrically-distorted stereoscopic images, blurred symmetrically-distorted stereoscopic images, fast-fading symmetrically-distorted stereoscopic images, and white-noise symmetrically-distorted stereoscopic images, respectively, and the prediction qualities of the corresponding distorted image blocks are 46.128, 53.495, 66.804, and 74.013, with higher prediction values indicating lower visual perception qualities.
On the other hand, various tests are also performed based on the scheme of the embodiment of the invention.
As shown in fig. 6, for the above-mentioned scheme of the embodiment of the present invention, the process of optimizing the loss function for all distortions on the LIVE database can be seen to be well converged. In fig. 5, curve 1 (thick line) corresponds to LIVE database 1, and curve 2 (thin line) corresponds to LIVE database 2.
The performance comparison results of the above-mentioned solutions of the embodiments of the present invention and the solutions mentioned in the background art on the stereo image LIVE database for all distortions are shown in table 1.
Figure BDA0001639400500000062
Figure BDA0001639400500000071
Table 1 comparison of the performance of the present invention with other solutions on the stereo image LIVE database for all distortions
The results of comparing the performance of the related schemes of the present invention and the background art on symmetrically distorted and asymmetrically distorted stereo images are shown in table 2.
Name of algorithm Symmetrical distortion Asymmetric distortion
Article [5]] 0.918 0.834
Non-reference convolutional neural network [6] 0.590 0.633
The invention 0.979 0.927
TABLE 2 comparison of the Performance of the related schemes of the present invention and the background art on symmetrically distorted and asymmetrically distorted stereo images
The results of the performance comparison between the related schemes of the present invention and the background art on the cross database test are shown in table 3.
Figure BDA0001639400500000072
TABLE 3 comparison of Performance on Cross database testing of related protocols as mentioned in the present invention and background
The results of comparing the performance of the present invention with that of the prior art solutions for different distortion types are shown in table 4.
Figure BDA0001639400500000073
Figure BDA0001639400500000081
Table 4 comparison of the performance of the various schemes of the invention and the background art with respect to different distortion types
According to the comparison result, the performance of the scheme of the embodiment of the invention is far better than that of each scheme mentioned in the background technology.
According to the scheme of the embodiment of the invention, by considering the interaction principle of fusion between left and right views and parallax information in a multilayer structure of a human eye vision system, a non-reference 3D (three-dimensional) image quality evaluation algorithm of a deep neural network based on end-to-end double-flow interaction is used, distorted image blocks of the left and right views are simultaneously input into the double-flow deep neural network structure, and the judgment feature extraction and regression learning are combined into an end-to-end optimization process, so that the purpose of effectively predicting the perception quality of the distorted three-dimensional image is achieved.
Through the above description of the embodiments, it is clear to those skilled in the art that the above embodiments can be implemented by software, and can also be implemented by software plus a necessary general hardware platform. With this understanding, the technical solutions of the embodiments can be embodied in the form of a software product, which can be stored in a non-volatile storage medium (which can be a CD-ROM, a usb disk, a removable hard disk, etc.), and includes several instructions for enabling a computer device (which can be a personal computer, a server, or a network device, etc.) to execute the methods according to the embodiments of the present invention.
The above description is only for the preferred embodiment of the present invention, but the scope of the present invention is not limited thereto, and any changes or substitutions that can be easily conceived by those skilled in the art within the technical scope of the present invention are included in the scope of the present invention. Therefore, the protection scope of the present invention shall be subject to the protection scope of the claims.

Claims (4)

1. A no-reference stereo image quality evaluation method based on a deep neural network is characterized in that the interaction principle of fusion between left and right views and parallax information in a multilayer structure of a human visual system is considered, and the method comprises the following steps:
dividing left and right view images forming all distorted stereo images into non-overlapping distorted image blocks respectively to obtain a plurality of left and right view distorted image block pairs;
correspondingly inputting the left and right visual distortion image block pairs obtained by division into a pre-constructed double-current input interactive deep neural network, training a network model, and obtaining the one-dimensional perception quality of the corresponding left and right visual distortion image block pairs through regression learning;
when the quality of the distorted stereo image is predicted, predicting the one-dimensional perception quality of each left-view and right-view distorted image block pair by using a trained network model, and averaging the one-dimensional perception qualities of all the left-view and right-view distorted image block pairs in each distorted stereo image to obtain the one-dimensional perception quality of each distorted stereo image;
the double-current input interactive deep neural network comprises two identical sub-networks which are used for inputting left and right visual distortion image blocks respectively and correspondingly; each sub-network comprises a first convolution layer, a first pooling layer, a second convolution layer, a second pooling layer, a third convolution layer, a third pooling layer, a first full-connection layer and a second full-connection layer which are arranged in sequence;
in the double-current input interactive deep neural network, the second convolution layer of the two sub-networks, the last convolution layer of the third convolution layer and the second full-connection layer are correspondingly connected with each other in an interactive mode;
the interaction of the convolution layer is to firstly perform the following operation on the characteristic diagram of the corresponding left-right view network to obtain a fused image S+Sum-difference image S-Re-connecting the fused image S+Sum-difference image S-
S+=Fl+Fr
S-=Fl-Fr
Wherein, Fl、FrThe left and right distortion image blocks are corresponding left and right view characteristic diagrams;
the interaction of the fully-connected layer is to connect the second fully-connected layers of the sub-networks.
2. The method according to claim 1, wherein the quality of the distorted stereo image is used as the label value of the training sample of all left-right view distorted image block pairs obtained after the distorted stereo image is divided.
3. The method according to claim 1, wherein the deep neural network-based no-reference stereo image quality evaluation method comprises,
correspondingly inputting the left and right visual distortion image block pairs obtained by division into a pre-constructed double-flow input interactive deep neural network for forward propagation training, and finally calculating the Euclidean loss value according to a training sample label as a minimum objective function in the training process:
Figure FDA0002522090450000021
Figure FDA0002522090450000022
wherein, f (P)li,Pri(ii) a w) represents a left-right view distorted image block pair (P) under the parameter wli,Pri) W is a parameter of the double-current input interactive deep neural network, and needs to be updated iteratively, w' is an updated network parameter, F represents a 2 norm, yiIs a left-right view distorted image block pair (P)li,Pri) The corresponding training sample label value.
4. The method according to claim 1, wherein the quality of all left-right view distorted image block pairs in each distorted stereo image is averaged, and the calculation formula for obtaining the quality of each distorted stereo image is as follows:
Figure FDA0002522090450000023
wherein, f (P)li,Pri(ii) a w') denotes a left-right view distorted image block pair (P) under the parameter wli,Pri) P denotes the number of left and right view distorted image block pairs in each distorted stereoscopic image.
CN201810375052.5A 2018-04-24 2018-04-24 No-reference stereo image quality evaluation method based on deep neural network Active CN108391121B (en)

Priority Applications (1)

Application Number Priority Date Filing Date Title
CN201810375052.5A CN108391121B (en) 2018-04-24 2018-04-24 No-reference stereo image quality evaluation method based on deep neural network

Applications Claiming Priority (1)

Application Number Priority Date Filing Date Title
CN201810375052.5A CN108391121B (en) 2018-04-24 2018-04-24 No-reference stereo image quality evaluation method based on deep neural network

Publications (2)

Publication Number Publication Date
CN108391121A CN108391121A (en) 2018-08-10
CN108391121B true CN108391121B (en) 2020-10-27

Family

ID=63065726

Family Applications (1)

Application Number Title Priority Date Filing Date
CN201810375052.5A Active CN108391121B (en) 2018-04-24 2018-04-24 No-reference stereo image quality evaluation method based on deep neural network

Country Status (1)

Country Link
CN (1) CN108391121B (en)

Families Citing this family (9)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN109784358B (en) * 2018-11-23 2023-07-11 南京航空航天大学 No-reference image quality evaluation method integrating artificial features and depth features
CN110246111B (en) * 2018-12-07 2023-05-26 天津大学青岛海洋技术研究院 No-reference stereoscopic image quality evaluation method based on fusion image and enhanced image
CN109492759B (en) * 2018-12-17 2022-05-20 北京百度网讯科技有限公司 Neural network model prediction method, device and terminal
CN109714592A (en) * 2019-01-31 2019-05-03 天津大学 Stereo image quality evaluation method based on binocular fusion network
CN110738645B (en) * 2019-10-11 2022-06-10 浙江科技学院 3D image quality detection method based on convolutional neural network
CN111127435B (en) * 2019-12-25 2022-11-15 福州大学 No-reference image quality evaluation method based on double-current convolution neural network
US20210233259A1 (en) * 2020-01-28 2021-07-29 Ssimwave Inc. No-reference visual media assessment combining deep neural networks and models of human visual system and video content/distortion analysis
CN113128517B (en) * 2021-03-22 2023-06-13 西北大学 Tone mapping image mixed visual feature extraction model establishment and quality evaluation method
CN113393461B (en) * 2021-08-16 2021-12-07 北京大学第三医院(北京大学第三临床医学院) Method and system for screening metaphase chromosome image quality based on deep learning

Citations (6)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN105160678A (en) * 2015-09-02 2015-12-16 山东大学 Convolutional-neural-network-based reference-free three-dimensional image quality evaluation method
CN106650699A (en) * 2016-12-30 2017-05-10 中国科学院深圳先进技术研究院 CNN-based face detection method and device
CN106815579A (en) * 2017-01-22 2017-06-09 深圳市唯特视科技有限公司 A kind of motion detection method based on multizone double fluid convolutional neural networks model
CN106845471A (en) * 2017-02-20 2017-06-13 深圳市唯特视科技有限公司 A kind of vision significance Forecasting Methodology based on generation confrontation network
CN107067465A (en) * 2017-04-14 2017-08-18 深圳市唯特视科技有限公司 A kind of 3-D view synthetic method that network is generated based on checking transition diagram picture
CN107481188A (en) * 2017-06-23 2017-12-15 珠海经济特区远宏科技有限公司 A kind of image super-resolution reconstructing method

Patent Citations (6)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN105160678A (en) * 2015-09-02 2015-12-16 山东大学 Convolutional-neural-network-based reference-free three-dimensional image quality evaluation method
CN106650699A (en) * 2016-12-30 2017-05-10 中国科学院深圳先进技术研究院 CNN-based face detection method and device
CN106815579A (en) * 2017-01-22 2017-06-09 深圳市唯特视科技有限公司 A kind of motion detection method based on multizone double fluid convolutional neural networks model
CN106845471A (en) * 2017-02-20 2017-06-13 深圳市唯特视科技有限公司 A kind of vision significance Forecasting Methodology based on generation confrontation network
CN107067465A (en) * 2017-04-14 2017-08-18 深圳市唯特视科技有限公司 A kind of 3-D view synthetic method that network is generated based on checking transition diagram picture
CN107481188A (en) * 2017-06-23 2017-12-15 珠海经济特区远宏科技有限公司 A kind of image super-resolution reconstructing method

Non-Patent Citations (1)

* Cited by examiner, † Cited by third party
Title
一种基于卷积神经网络的无参考立体图像质量评估算法;瞿晨非等;《中国科技论文在线》;20150827;第1-3节 *

Also Published As

Publication number Publication date
CN108391121A (en) 2018-08-10

Similar Documents

Publication Publication Date Title
CN108391121B (en) No-reference stereo image quality evaluation method based on deep neural network
Chen et al. No-reference quality assessment of natural stereopairs
CN109360178B (en) Fusion image-based non-reference stereo image quality evaluation method
CN110060236B (en) Stereoscopic image quality evaluation method based on depth convolution neural network
Shao et al. Blind image quality assessment for stereoscopic images using binocular guided quality lookup and visual codebook
CN108769671B (en) Stereo image quality evaluation method based on self-adaptive fusion image
CN109831664B (en) Rapid compressed stereo video quality evaluation method based on deep learning
CN110246111B (en) No-reference stereoscopic image quality evaluation method based on fusion image and enhanced image
Su et al. Visual quality assessment of stereoscopic image and video: challenges, advances, and future trends
CN108520510B (en) No-reference stereo image quality evaluation method based on overall and local analysis
CN109523513A (en) Based on the sparse stereo image quality evaluation method for rebuilding color fusion image
Geng et al. A stereoscopic image quality assessment model based on independent component analysis and binocular fusion property
Yan et al. Blind stereoscopic image quality assessment by deep neural network of multi-level feature fusion
Shao et al. Toward simultaneous visual comfort and depth sensation optimization for stereoscopic 3-D experience
CN111915589A (en) Stereo image quality evaluation method based on hole convolution
CN105376563A (en) No-reference three-dimensional image quality evaluation method based on binocular fusion feature similarity
Jiang et al. 3D Visual Attention for Stereoscopic Image Quality Assessment.
CN108259893B (en) Virtual reality video quality evaluation method based on double-current convolutional neural network
CN114648482A (en) Quality evaluation method and system for three-dimensional panoramic image
CN116485741A (en) No-reference image quality evaluation method, system, electronic equipment and storage medium
Liu et al. Blind stereoscopic image quality assessment accounting for human monocular visual properties and binocular interactions
CN108492275B (en) No-reference stereo image quality evaluation method based on deep neural network
CN105488792B (en) Based on dictionary learning and machine learning without referring to stereo image quality evaluation method
CN105898279A (en) Stereoscopic image quality objective evaluation method
Kim et al. Visual comfort aware-reinforcement learning for depth adjustment of stereoscopic 3d images

Legal Events

Date Code Title Description
PB01 Publication
PB01 Publication
SE01 Entry into force of request for substantive examination
SE01 Entry into force of request for substantive examination
GR01 Patent grant
GR01 Patent grant