CN108470336B - Stereo image quality evaluation method based on stack type automatic encoder - Google Patents

Stereo image quality evaluation method based on stack type automatic encoder Download PDF

Info

Publication number
CN108470336B
CN108470336B CN201810272419.0A CN201810272419A CN108470336B CN 108470336 B CN108470336 B CN 108470336B CN 201810272419 A CN201810272419 A CN 201810272419A CN 108470336 B CN108470336 B CN 108470336B
Authority
CN
China
Prior art keywords
monocular
diagram
binocular
extracting
image
Prior art date
Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
Active
Application number
CN201810272419.0A
Other languages
Chinese (zh)
Other versions
CN108470336A (en
Inventor
杨嘉琛
赵洋
Current Assignee (The listed assignees may be inaccurate. Google has not performed a legal analysis and makes no representation or warranty as to the accuracy of the list.)
Tianjin University
Original Assignee
Tianjin University
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Application filed by Tianjin University filed Critical Tianjin University
Priority to CN201810272419.0A priority Critical patent/CN108470336B/en
Publication of CN108470336A publication Critical patent/CN108470336A/en
Application granted granted Critical
Publication of CN108470336B publication Critical patent/CN108470336B/en
Active legal-status Critical Current
Anticipated expiration legal-status Critical

Links

Images

Classifications

    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06TIMAGE DATA PROCESSING OR GENERATION, IN GENERAL
    • G06T7/00Image analysis
    • G06T7/0002Inspection of images, e.g. flaw detection
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06TIMAGE DATA PROCESSING OR GENERATION, IN GENERAL
    • G06T2207/00Indexing scheme for image analysis or image enhancement
    • G06T2207/10Image acquisition modality
    • G06T2207/10004Still image; Photographic image
    • G06T2207/10012Stereo images
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06TIMAGE DATA PROCESSING OR GENERATION, IN GENERAL
    • G06T2207/00Indexing scheme for image analysis or image enhancement
    • G06T2207/20Special algorithmic details
    • G06T2207/20016Hierarchical, coarse-to-fine, multiscale or multiresolution image processing; Pyramid transform
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06TIMAGE DATA PROCESSING OR GENERATION, IN GENERAL
    • G06T2207/00Indexing scheme for image analysis or image enhancement
    • G06T2207/20Special algorithmic details
    • G06T2207/20048Transform domain processing
    • G06T2207/20064Wavelet transform [DWT]
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06TIMAGE DATA PROCESSING OR GENERATION, IN GENERAL
    • G06T2207/00Indexing scheme for image analysis or image enhancement
    • G06T2207/20Special algorithmic details
    • G06T2207/20081Training; Learning
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06TIMAGE DATA PROCESSING OR GENERATION, IN GENERAL
    • G06T2207/00Indexing scheme for image analysis or image enhancement
    • G06T2207/20Special algorithmic details
    • G06T2207/20212Image combination
    • G06T2207/20221Image fusion; Image merging
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06TIMAGE DATA PROCESSING OR GENERATION, IN GENERAL
    • G06T2207/00Indexing scheme for image analysis or image enhancement
    • G06T2207/30Subject of image; Context of image processing
    • G06T2207/30168Image quality inspection

Abstract

The invention relates to a three-dimensional image quality evaluation method based on a stack type automatic encoder, which comprises the following steps: synthesizing a monocular diagram by using the left view and the right view; will be leftThe right view is converted into a binocular fusion graph and a binocular difference graph which are mutually irrelevant; calculating a monocular diagram ICMSCN coefficient of
Figure DDA0001612881420000011
And the monocular diagram is at the horizontal H, vertical V and main diagonal D1And the secondary diagonal D2MSCN neighborhood coefficients in four directions; extracting a primary feature vector from the monocular image; extracting a primary feature vector of the binocular fusion map; extracting features of the binocular disparity map according to the same method; training a stack type automatic encoder to obtain deep features; training a corresponding support vector regression machine; and obtaining the quality score of the stereo image. The invention improves the accuracy of the objective evaluation method.

Description

Stereo image quality evaluation method based on stack type automatic encoder
Technical Field
The invention belongs to the field of image processing, and relates to an objective evaluation method for quality of a non-reference stereo image.
Background
With the rapid development of multimedia technology in recent years, stereoscopic display technology is widely used. Meanwhile, compared with a plane image, the stereo image has stronger visual perception to people and brings more real audio-visual perception and telepresence experience to people, so that stereo image/video processing research is widely concerned by people. However, due to the influence of factors such as equipment and processing means, the image/video distortion is inevitably caused during the acquisition, compression, transmission and storage processes of the stereo image/video, and the quality of the stereo image/video is further influenced. Therefore, it is important to research an evaluation method capable of effectively evaluating the quality of a stereoscopic image. Although subjective quality evaluation is a reliable evaluation method, the subjective evaluation method is easily interfered by human and external environmental factors, the evaluation result is not stable enough, and a large amount of manpower and material resources are consumed. Compared with subjective evaluation, objective evaluation evaluates the quality of images in a software mode, does not need participants and a large number of subjective tests, is simple to operate, and is highly related to the subjective evaluation, so that the objective evaluation is more and more concerned by related researchers.
The current stereo image quality evaluation is mainly divided into three types of methods according to whether to refer to an original image. The first type is full-reference stereo image quality evaluation, the second type is partial-reference quality evaluation, and the two methods need to refer to an original image or make objective evaluation on the stereo image by partial information of the original image, so that the method has great limitation. The third type is no reference quality evaluation, and the method does not need to refer to an original image and is the method which is most suitable for practical situations. Due to inaccuracy of acquisition of depth information of the stereo image and insufficient consideration of binocular characteristics, evaluation of the quality of the stereo image is still a hotspot and a difficulty of current research.
Disclosure of Invention
The invention aims to overcome the defects of the prior art and establish a non-reference stereo image quality evaluation method which is based on a stack type automatic encoder and fully considers the visual characteristics of human eyes. The invention extracts the basic characteristics which accord with the human eye three-dimensional perception from the monocular diagram, the binocular fusion diagram and the binocular difference diagram, and converts the characteristics into the deep level characteristics which accord with the human eye characteristics through the stacked self-encoder, thereby being capable of making more comprehensive and accurate objective evaluation on the quality of the three-dimensional image. The technical scheme is as follows:
a three-dimensional image quality evaluation method based on a stack type automatic encoder comprises the following steps:
the first step is as follows: monocular picture I synthesized by using left and right viewsC
The second step is that: converting the left view and the right view into a binocular fusion graph and a binocular difference graph which are mutually irrelevant;
the third step: extracting primary feature vectors of a monocular diagram
(1) Calculating a monocular diagram ICMSCN coefficient of
Figure BDA0001612881400000011
And the monocular diagram is at the horizontal H, vertical V and main diagonal D1And the secondary diagonal D2MSCN neighborhood coefficients in four directions;
(2) extracting MSCN neighborhood coefficients of the monocular image in 4 directions of horizontal, vertical, main diagonal and secondary diagonal, and defining as follows:
Figure BDA0001612881400000021
Figure BDA0001612881400000022
Figure BDA0001612881400000023
Figure BDA0001612881400000024
(3) fitting the MSCN coefficient histogram of the monocular diagram and MSCN neighborhood coefficient histograms of the monocular diagram in four directions of horizontal, vertical, main diagonal and secondary diagonal by using an asymmetric generalized Gaussian distribution model (AGGD); taking the mean value, the variance, the shape parameter and the size parameter of the 5 AGGD models as the characteristics of the monocular diagram, and extracting 20 characteristics; fitting the gradient amplitude of the monocular diagram by using the Weber distribution, and extracting the shape parameters and the size parameters of the Weber distribution as 2 characteristics of the monocular diagram; furthermore, a monocular image is subjected to DCT transformation, 5 features of block, sharpness, smoothness, kurtosis and Jansen Shannon divergence JSD in a DCT domain are extracted, and a 27-dimensional primary feature vector P is extracted from the monocular imageC (i)
The fourth step: extracting primary feature vectors of binocular fusion map
(1) Performing joint adaptive normalization JAN on a gradient magnitude graph GM and a Laplace transform graph LOG of a binocular fusion graph, quantizing the normalized gradient magnitude graph and the Laplace transform graph into an M level and an N level respectively, and calculating the marginal probability density P of the two quantized graphsGM、PLOGAnd independently distributed QGM、QLOGHere, M ═ N ═ 5, so 20 features are extracted;
(2) performing wavelet decomposition on the binocular fusion map in 6 directions of 2-level size, theta epsilon {0 degrees, 30 degrees, 60 degrees, 90 degrees, 120 degrees and 150 degrees }; respectively fitting 6 directions by using 6 generalized Gaussian distribution GGDs (generalized Gaussian distribution) with wavelet coefficients of 2-level sizes in each direction, and taking shape parameters of the GGDs as the characteristics of a binocular fusion graph to obtain 6 characteristics; fitting wavelet coefficients of all sizes and all directions by using a GGD, and extracting shape parameters of the GGD to serve as 1 characteristic of a binocular fusion graph; extracting 27-dimensional feature vector from common binocular fusion image
Figure BDA0001612881400000025
The fifth step: extracting features from the binocular difference image according to the same method, and extracting 27-dimensional primary feature vector from the binocular difference image
Figure BDA0001612881400000026
And a sixth step: training stack type automatic encoder
Selecting training data, left-right view
Figure BDA0001612881400000027
And
Figure BDA0001612881400000028
respectively training a stacked automatic encoder SAE-C of a single-eye diagram, a stacked automatic encoder SAE-S of a binocular fusion diagram and a stacked automatic encoder SAE-D of a binocular difference diagram as samples;
the seventh step: three stacked self-encoders will
Figure BDA0001612881400000031
And
Figure BDA0001612881400000032
encoding deep features as abstractions, respectively
Figure BDA0001612881400000033
And
Figure BDA0001612881400000034
wherein the number of units of each layer of SAE-C, SAE-S and SAE-D is 27-25-20-15;
eighth step: building training set and test set, using left and right views in training set
Figure BDA0001612881400000035
With corresponding MOS training
Figure BDA0001612881400000036
A corresponding support vector regression (SVR-C); in the same way, utilize
Figure BDA0001612881400000037
With corresponding MOS training
Figure BDA0001612881400000038
A corresponding support vector regression (SVR-SD);
the ninth step: for the test set, the quality scores of the monocular map, the binocular fusion map/binocular difference map are predicted using SVR-C, SVR-SD, and then the stereoscopic image quality scores are formed by weighting.
The method for evaluating the objective quality of the stereo image provided by the invention is based on a binocular inhibition and depth perception mechanism, and establishes a non-reference stereo video quality objective evaluation model by using a stack type automatic encoder. The obtained objective evaluation result and the subjective evaluation result of the quality of the three-dimensional image have high consistency, and the quality of the three-dimensional image can be reflected more accurately.
Drawings
Fig. 1 is a flowchart of a method for evaluating quality of a stereoscopic image based on a stacked automatic encoder.
Detailed Description
The invention provides a method for evaluating objective quality of a non-reference stereo image, which is provided on the basis of a stereo image perception theory and a sum-difference channel theory. In order to extract effective stereo perception features, the invention respectively extracts a monocular diagram (I)C) Binocular fusion picture (I)S) And binocular disparity map (I)D) Primary eigenvectors of the three stereo images
Figure BDA0001612881400000039
Then, under an unsupervised condition, three Stacked Auto-encoders (SAE) are trained by using the primary features of the three stereo images; the three trained SAEs are then used to encode the primary feature vector of the stereo image, which is converted into a deep feature vector more suitable for stereo image quality assessment
Figure BDA00016128814000000310
The quality score of the deep characteristic vector of the stereo image to be measured is predicted by fitting the deep characteristic vector of the stereo image and a subjective quality score (MOS) corresponding to the stereo image. The deep feature vectors can reflect the distortion degree of the stereo image, so that the quality evaluation is carried out on the distorted stereo image, and the method comprises the following steps:
the first step is as follows: monocular rendering using left and right views
Combining binocular features, fusing left and right views according to binocular competition characteristics to obtain a monocular diagram, wherein the monocular diagram is defined as:
IC(i,j)=WL(i,j)IL(i,j)+WR((i+d),j)IR((i+d),j) (1)
wherein ILAnd IRRespectively representing a left view and a right view, d is parallax, WLAnd WRThe weights of the left view and the right view are obtained by normalizing Gabor filter energy response assignment and are defined as:
Figure BDA0001612881400000041
Figure BDA0001612881400000042
wherein is GELAnd GERRepresenting the energy response values of the left and right views in all dimensions and directions, respectively.
The second step is that: converting left and right views into unrelated binocular fusion image ISAnd binocular disparity map IDDefined as:
Figure BDA0001612881400000043
IS=|IL-IR| (5)
the third step: extracting primary feature vectors of a monocular diagram
Calculating Mean Subtracted Contrast Normalized (MSCN) coefficients for the monocular image, and the monocular image is in horizontal (H), vertical (V), main diagonal (D)1) And secondary diagonal (D)2) The MSCN neighborhood coefficients in 4 directions are calculated by the following method:
(1) let ICHas a size of BxD, MSCN coefficient
Figure BDA0001612881400000044
Comprises the following steps:
Figure BDA0001612881400000045
in the formula (I), the compound is shown in the specification,
Figure BDA0001612881400000046
i is 1,2, …, B, j is 1,2, …, D, γ is constant, the denominator part is added with a constant in order to avoid instability when the denominator approaches zero in image flat areas, ω is { ω ═ ω { (ω) } ωk,l-K, …, K; l ═ L, …, L } is a two-dimensional circularly symmetric gaussian weighting function;
(2) MSCN neighborhood coefficients of the monocular diagram in 4 directions of horizontal, vertical, main diagonal and secondary diagonal are defined as:
Figure BDA0001612881400000047
Figure BDA0001612881400000048
Figure BDA0001612881400000049
Figure BDA0001612881400000051
then, fitting the MSCN coefficient histogram of the monocular diagram and the MSCN neighborhood coefficient histograms of 4 directions of the horizontal, vertical, main diagonal and secondary diagonal of the monocular diagram by using an Asymmetric Generalized Gaussian Distribution (AGGD) model; taking the mean value, the variance, the shape parameter and the size parameter of the 5 AGGD models as the characteristics of the monocular diagram, and extracting 20 characteristics; fitting the gradient amplitude of the monocular diagram by using the Weber distribution, and extracting the shape parameters and the size parameters of the Weber distribution as 2 characteristics of the monocular diagram; furthermore, the monocular image was DCT-transformed, and 5 features of block, sharpness, smoothness, kurtosis, and Jensen Shannon Divergence (JSD) in the DCT domain were extracted.
According to the steps, 27-dimensional primary feature vectors are extracted from the monocular image
Figure BDA0001612881400000052
The fourth step: extracting primary feature vectors of binocular fusion map
Performing combined adaptive normalization (JAN) on a gradient magnitude Graph (GM) and a Laplace transform graph (LOG) of the binocular fusion graph, respectively quantizing the normalized gradient magnitude graph and the Laplace transform graph into an M level and an N level, and calculating edge probability densities (P) of the two types of graphsGM、PLOG) And independent distribution (Q)GM、QLOG). Here, M is 5, so 20 features are extracted. The specific calculation method is as follows:
(1) GM graph of binocular fusion graph (G)I) And LOG transformation graph (L)I) The joint normalization calculation method comprises the following steps:
Figure BDA0001612881400000053
(2) calculating the local adaptive normalization factor of each (i, j) pixel by the following method:
Figure BDA0001612881400000054
wherein omegai,jIs the local center window of the (i, j) pixel, and ω (l, k) is the weight, satisfying
Figure BDA0001612881400000055
For GI、LINormalization is carried out
Figure BDA0001612881400000056
C ═ 0.2 is a constant for ensuring stability.
(3) Will be provided with
Figure BDA0001612881400000057
Quantization is M level and N level. Calculating the edge probability density P of GM and LOGGMAnd PLOG
Figure BDA0001612881400000058
Figure BDA0001612881400000059
Wherein, Km,n=P(G=gm,L=ln) Where M is 1,2, …, M, N is 1,2, …, N, where M is 5, 10 features are extracted.
(4) Calculating independent distribution function Q of GM and LOGGMAnd QLOG
Figure BDA0001612881400000061
Figure BDA0001612881400000062
Where M is 1,2, …, M, N is 1,2, …, N, M is 5, so 10 features are extracted.
In addition, wavelet decomposition is carried out on the binocular fusion map in 2-level sizes (alpha belongs to {1, 2}), 6 directions (theta belongs to {0 degrees, 30 degrees, 60 degrees, 90 degrees, 120 degrees, 150 degrees) } by utilizing the controllable pyramid; respectively fitting wavelet coefficients in 6 directions (each direction has 2-level size) by using 6 Generalized Gaussian Distributions (GGDs), and taking shape parameters of the GGDs as the characteristics of a binocular fusion diagram to obtain 6 characteristics; and fitting the wavelet coefficients of all sizes and all directions by using a GGD, and extracting shape parameters of the GGD to serve as 1 characteristic of the binocular fusion image.
According to the steps, 27-dimensional feature vectors are extracted from the binocular fusion map
Figure BDA0001612881400000063
The fifth step: the method for extracting the features of the binocular difference image is completely the same as that of the binocular fusion image, and 27-dimensional primary feature vectors are extracted from the binocular difference image by referring to the fourth step
Figure BDA0001612881400000064
And a sixth step: training stack type automatic encoder
Randomly select left and right views of 50% in the image library to train 3 SAEs. Will be viewed from left to right
Figure BDA0001612881400000065
And
Figure BDA0001612881400000066
as samples, a stacked automatic encoder (SAE-C) of a single-eye diagram, a stacked automatic encoder (SAE-S) of a double-eye fusion diagram and a stacked automatic encoder (SAE-D) of a double-eye difference diagram are trained respectively, the three SAE training methods are the same, and the procedures are as follows:
(1) the feature vector x is input into a stacked self-encoder (SAE) as input layer data, and is encoded by a first layer self-encoder in SAE (the encoding function is
Figure BDA0001612881400000067
) Mapping x to the first hidden layer h1Then using a decoder (decoding function is
Figure BDA0001612881400000068
) And h1Reconstructing the input layer to obtain a reconstructed vector
Figure BDA0001612881400000069
Training the first hidden layer h of an auto-encoder by minimizing the reconstruction error1When the reconstruction error tends to be stable, the first layer coding function is obtained
Figure BDA00016128814000000610
And h1Removing the reconstruction layer;
(2) h is to be1Repeating the step (1) as the input of the second layer automatic encoder to obtain the second layer encoding function
Figure BDA00016128814000000611
And a second hidden layer h2(ii) a After the training layer by layer, the trained stack type self-encoder with 3 hidden layers is finally obtained.
The seventh step: the trained SAE-C, SAE-S and SAE-D will
Figure BDA00016128814000000612
And
Figure BDA00016128814000000613
encoding deep features as abstractions, respectively
Figure BDA0001612881400000071
And
Figure BDA0001612881400000072
wherein the number of units of each layer of SAE-C, SAE-S and SAE-D is 27-25-20-15.
Eighth step: randomly selecting 80% of left and right views in the image library as a training set, and utilizing the left and right views in the training set
Figure BDA0001612881400000073
Training with corresponding subjective evaluation values (MOS)
Figure BDA0001612881400000074
A corresponding support vector regression machine (SVR-C); use in the same way
Figure BDA0001612881400000075
With corresponding MOS training
Figure BDA0001612881400000076
Corresponding support vector regression machine (SVR-SD).
The ninth step: and (3) taking left and right views of the left and right 20 percent of the images in the image library as a test set, and predicting local mass fractions of the monocular image, the binocular fusion image and the binocular difference image by utilizing SVR-C, SVR-SD (singular value decomposition-singular
Figure BDA0001612881400000077
The stereo image quality score Q is then formed by weighting(i)The weight is W, and the calculation formula is as follows:
Figure BDA0001612881400000078

Claims (1)

1. a three-dimensional image quality evaluation method based on a stack type automatic encoder comprises the following steps:
the first step is as follows: monocular picture I synthesized by using left and right viewsC
The second step is that: converting the left view and the right view into a binocular fusion graph and a binocular difference graph which are mutually irrelevant;
the third step: extracting primary feature vectors of a monocular diagram
(1) Calculating a monocular diagram ICMSCN coefficient of
Figure FDA0001612881390000011
And the monocular diagram is at the horizontal H, vertical V and main diagonal D1And the secondary diagonal D2MSCN neighborhood coefficients in four directions;
(2) extracting MSCN neighborhood coefficients of the monocular image in 4 directions of horizontal, vertical, main diagonal and secondary diagonal, and defining as follows:
Figure FDA0001612881390000012
Figure FDA0001612881390000013
Figure FDA0001612881390000014
Figure FDA0001612881390000015
(3) fitting the MSCN coefficient histogram of the monocular diagram and MSCN neighborhood coefficient histograms of the monocular diagram in four directions of horizontal, vertical, main diagonal and secondary diagonal by using an asymmetric generalized Gaussian distribution model (AGGD); taking the mean value, the variance, the shape parameter and the size parameter of the 5 AGGD models as the characteristics of the monocular diagram, and extracting 20 characteristics; fitting the gradient amplitude of the monocular diagram by using the Weber distribution, and extracting the shape parameters and the size parameters of the Weber distribution as 2 characteristics of the monocular diagram; furthermore, a monocular image is subjected to DCT transformation, 5 features of block, sharpness, smoothness, kurtosis and Jansen Shannon divergence JSD in a DCT domain are extracted, and a 27-dimensional primary feature vector is extracted from the monocular image
Figure FDA0001612881390000016
The fourth step: extracting primary feature vectors of binocular fusion map
(1) Performing joint adaptive normalization JAN on a gradient magnitude graph GM and a Laplace transform graph LOG of a binocular fusion graph, quantizing the normalized gradient magnitude graph and the Laplace transform graph into an M level and an N level respectively, and calculating the marginal probability density P of the two quantized graphsGM、PLOGAnd independently distributed QGM、QLOGWhere M is equal to N is equal to 5, soExtracting 20 features;
(2) performing wavelet decomposition on the binocular fusion map in 6 directions of 2-level size, theta, epsilon, 0 degrees, 30 degrees, 60 degrees, 90 degrees, 120 degrees and 150 degrees; respectively fitting 6 directions by using 6 generalized Gaussian distribution GGDs (generalized Gaussian distribution) with wavelet coefficients of 2-level sizes in each direction, and taking shape parameters of the GGDs as the characteristics of a binocular fusion graph to obtain 6 characteristics; fitting wavelet coefficients of all sizes and all directions by using a GGD, and extracting shape parameters of the GGD to serve as 1 characteristic of a binocular fusion graph; extracting 27-dimensional feature vector from common binocular fusion image
Figure FDA0001612881390000017
The fifth step: extracting features from the binocular difference image according to the same method, and extracting 27-dimensional primary feature vector from the binocular difference image
Figure FDA0001612881390000021
And a sixth step: training stack type automatic encoder
Selecting training data, left-right view
Figure FDA0001612881390000022
And
Figure FDA0001612881390000023
respectively training a stacked automatic encoder SAE-C of a single-eye diagram, a stacked automatic encoder SAE-S of a binocular fusion diagram and a stacked automatic encoder SAE-D of a binocular difference diagram as samples;
the seventh step: three stacked self-encoders will
Figure FDA0001612881390000024
And
Figure FDA0001612881390000025
encoding deep features as abstractions, respectively
Figure FDA0001612881390000026
And
Figure FDA0001612881390000027
wherein the number of units of each layer of SAE-C, SAE-S and SAE-D is 27-25-20-15;
eighth step: building training set and test set, using left and right views in training set
Figure FDA0001612881390000028
With corresponding MOS training
Figure FDA0001612881390000029
A corresponding support vector regression (SVR-C); in the same way, utilize
Figure FDA00016128813900000210
With corresponding MOS training
Figure FDA00016128813900000211
A corresponding support vector regression (SVR-SD);
the ninth step: for the test set, the quality scores of the monocular map, the binocular fusion map/binocular difference map are predicted using SVR-C, SVR-SD, and then the stereoscopic image quality scores are formed by weighting.
CN201810272419.0A 2018-03-29 2018-03-29 Stereo image quality evaluation method based on stack type automatic encoder Active CN108470336B (en)

Priority Applications (1)

Application Number Priority Date Filing Date Title
CN201810272419.0A CN108470336B (en) 2018-03-29 2018-03-29 Stereo image quality evaluation method based on stack type automatic encoder

Applications Claiming Priority (1)

Application Number Priority Date Filing Date Title
CN201810272419.0A CN108470336B (en) 2018-03-29 2018-03-29 Stereo image quality evaluation method based on stack type automatic encoder

Publications (2)

Publication Number Publication Date
CN108470336A CN108470336A (en) 2018-08-31
CN108470336B true CN108470336B (en) 2021-06-29

Family

ID=63262344

Family Applications (1)

Application Number Title Priority Date Filing Date
CN201810272419.0A Active CN108470336B (en) 2018-03-29 2018-03-29 Stereo image quality evaluation method based on stack type automatic encoder

Country Status (1)

Country Link
CN (1) CN108470336B (en)

Families Citing this family (3)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN109584203A (en) * 2018-09-29 2019-04-05 天津大学 Reorientation image quality evaluating method based on deep learning and semantic information
CN111818329B (en) * 2020-06-24 2021-08-13 天津大学 Video quality evaluation method based on stack type adaptive encoder
CN112233089B (en) * 2020-10-14 2022-10-25 西安交通大学 No-reference stereo mixed distortion image quality evaluation method

Citations (4)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN103152600A (en) * 2013-03-08 2013-06-12 天津大学 Three-dimensional video quality evaluation method
CN104994375A (en) * 2015-07-08 2015-10-21 天津大学 Three-dimensional image quality objective evaluation method based on three-dimensional visual saliency
CN107578403A (en) * 2017-08-22 2018-01-12 浙江大学 The stereo image quality evaluation method of binocular view fusion is instructed based on gradient information
CN107679543A (en) * 2017-02-22 2018-02-09 天津大学 Sparse autocoder and extreme learning machine stereo image quality evaluation method

Family Cites Families (1)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
JP6547275B2 (en) * 2014-10-29 2019-07-24 株式会社リコー INFORMATION PROCESSING SYSTEM, INFORMATION PROCESSING DEVICE, INFORMATION PROCESSING METHOD, AND PROGRAM

Patent Citations (4)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN103152600A (en) * 2013-03-08 2013-06-12 天津大学 Three-dimensional video quality evaluation method
CN104994375A (en) * 2015-07-08 2015-10-21 天津大学 Three-dimensional image quality objective evaluation method based on three-dimensional visual saliency
CN107679543A (en) * 2017-02-22 2018-02-09 天津大学 Sparse autocoder and extreme learning machine stereo image quality evaluation method
CN107578403A (en) * 2017-08-22 2018-01-12 浙江大学 The stereo image quality evaluation method of binocular view fusion is instructed based on gradient information

Also Published As

Publication number Publication date
CN108470336A (en) 2018-08-31

Similar Documents

Publication Publication Date Title
CN108961198B (en) Underwater image synthesis method of multi-grid generation countermeasure network and application thereof
Cao et al. Visual quality of compressed mesh and point cloud sequences
Maalouf et al. CYCLOP: A stereo color image quality assessment metric
CN108470336B (en) Stereo image quality evaluation method based on stack type automatic encoder
CN109523513B (en) Stereoscopic image quality evaluation method based on sparse reconstruction color fusion image
CN108765414B (en) No-reference stereo image quality evaluation method based on wavelet decomposition and natural scene statistics
CN108391121B (en) No-reference stereo image quality evaluation method based on deep neural network
CN107635136B (en) View-based access control model perception and binocular competition are without reference stereo image quality evaluation method
CN109345502B (en) Stereo image quality evaluation method based on disparity map stereo structure information extraction
CN109831664B (en) Rapid compressed stereo video quality evaluation method based on deep learning
Ma et al. Reduced-reference stereoscopic image quality assessment using natural scene statistics and structural degradation
CN108520510B (en) No-reference stereo image quality evaluation method based on overall and local analysis
Shao et al. No-reference view synthesis quality prediction for 3-D videos based on color–depth interactions
Shao et al. Toward simultaneous visual comfort and depth sensation optimization for stereoscopic 3-D experience
CN114648482A (en) Quality evaluation method and system for three-dimensional panoramic image
CN109257592B (en) Stereoscopic video quality objective evaluation method based on deep learning
Jeong et al. Visual comfort assessment of stereoscopic images using deep visual and disparity features based on human attention
JP2012208759A (en) Method and program for improving accuracy of three-dimensional shape model
CN106682599B (en) Sparse representation-based stereo image visual saliency extraction method
CN103905812A (en) Texture/depth combination up-sampling method
CN108648186B (en) No-reference stereo image quality evaluation method based on primary visual perception mechanism
Xiao et al. No-reference quality assessment of stereoscopic video based on deep frequency perception
Kim et al. Cnn-based blind quality prediction on stereoscopic images via patch to image feature pooling
CN111178163B (en) Stereoscopic panoramic image salient region prediction method based on cube projection format
Fang et al. A spatial-temporal weighted method for asymmetrically distorted stereo video quality assessment

Legal Events

Date Code Title Description
PB01 Publication
PB01 Publication
SE01 Entry into force of request for substantive examination
SE01 Entry into force of request for substantive examination
GR01 Patent grant
GR01 Patent grant