CN107809631A - The wavelet field method for evaluating video quality eliminated based on background - Google Patents

The wavelet field method for evaluating video quality eliminated based on background Download PDF

Info

Publication number
CN107809631A
CN107809631A CN201710926882.8A CN201710926882A CN107809631A CN 107809631 A CN107809631 A CN 107809631A CN 201710926882 A CN201710926882 A CN 201710926882A CN 107809631 A CN107809631 A CN 107809631A
Authority
CN
China
Prior art keywords
mrow
msub
video
quality
frame
Prior art date
Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
Granted
Application number
CN201710926882.8A
Other languages
Chinese (zh)
Other versions
CN107809631B (en
Inventor
张淑芳
黄小琴
Current Assignee (The listed assignees may be inaccurate. Google has not performed a legal analysis and makes no representation or warranty as to the accuracy of the list.)
Tianjin University
Original Assignee
Tianjin University
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Application filed by Tianjin University filed Critical Tianjin University
Priority to CN201710926882.8A priority Critical patent/CN107809631B/en
Publication of CN107809631A publication Critical patent/CN107809631A/en
Application granted granted Critical
Publication of CN107809631B publication Critical patent/CN107809631B/en
Active legal-status Critical Current
Anticipated expiration legal-status Critical

Links

Classifications

    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04NPICTORIAL COMMUNICATION, e.g. TELEVISION
    • H04N17/00Diagnosis, testing or measuring for television systems or their details

Landscapes

  • Engineering & Computer Science (AREA)
  • Health & Medical Sciences (AREA)
  • Biomedical Technology (AREA)
  • General Health & Medical Sciences (AREA)
  • Multimedia (AREA)
  • Signal Processing (AREA)
  • Compression Or Coding Systems Of Tv Signals (AREA)

Abstract

The invention discloses a kind of wavelet field method for evaluating video quality eliminated based on background, the first step:Video sequence global qualitySecond step:The local quality of video is calculated using background null method3rd step:The oeverall quality of video is calculated, i.e., according to the local quality and global quality of resulting video sequence, video quality evaluation modelIt is contemplated that improve the uniformity of the subjective quality assessment of video objective quality evaluation and human eye;Can there is preferable video evaluation performance for different type of distortion, different scenes, and the algorithm has relatively low complexity, can realize real-time quality assessment.

Description

The wavelet field method for evaluating video quality eliminated based on background
Technical field
The present invention relates to video quality evaluation field, more particularly to a kind of wavelet field video quality evaluation model.
Background technology
Video quality evaluation algorithm can weigh different degrees of distortion, turn into the popular research side of video field in recent years To.Method for evaluating video quality can be divided into subjective assessment and objective evaluation, and subjective quality assessment is mainly imitated by human eye vision Fruit evaluates video quality, it is considered to be most reliable quality evaluating method.Therefore generally use subjective evaluation result with it is objective Performance Evaluating Indexes of the uniformity of evaluation result as objective evaluation algorithm.
The perceived quality of distortion video is global quality and the coefficient result of local quality.Global quality is observer For the rough impression of video quality, obtained after being averaged by the quality of video sequence whole frame;Local quality is mainly by human eye The characteristic such as mass change determines between visual attention location, sequence.By calculating the overall situation, the local quality of video sequence respectively, depending on The oeverall quality of frequency.
A variety of objective video quality evaluation methods, wherein average poor (MSE), Y-PSNR in the world be present at present (PSNR), structural similarity (SSIM) and Multi-scale model similitude (MS-SSIM) due to model it is simple, in image and video matter Extensive utilization in amount evaluation, but the uniformity of these methods and human eye visual perception is poor.Kalpana et al. is improved to this, Multichannel Decomposition is carried out to image using Gabor filter, estimation is carried out to each passage, it is proposed that based drive space-time Domain video quality evaluation algorithm (MOVIE).Phong et al. combines image quality evaluation model M AD, time-space domain correlation and base In the model of human visual system, the spatial distortion to video sequence, time-space domain distortion are weighed respectively, so as to propose base In spatial domain and the video quality evaluation algorithm (VIS3) of time-space domain slice gradient similarity.Regarded according to human visual system in understanding Relied primarily on during frequency in image various marginal texture information the characteristics of, PengYan et al. proposes to be based on Spatial-temporal slice gradient The method for evaluating video quality of similarity.These methods can obtain higher accuracy, but complexity is high, limits model Practicality.
Different from image quality evaluation, video quality evaluation is not only related to spatial domain distortion, also related to the change of time domain. Distortion information of the change comprising movable information and time domain of time domain, and observer is when video observe, notice understands quilt Emergent object, the object of motion intense and time domain distortion information such as ghost image etc. are influenceed, that is to say, that foreground information is One of focal point of human eye.Frame difference method is a kind of simple Moving object extraction method, has that context update is fast, it is adaptive should be able to The features such as power is strong.Frame difference method is widely used in extraction video sequence due to its simple and quick characteristic in video quality evaluation Movable information.Loh et al. subtracts each other forward direction two frame of the present frame respectively with reference video of reference video, distortion video, proposes A kind of time-domain video quality evaluating method based on SSIM, this method speed is fast, but with human eye visual perception uniformity, general Property it is all poor because frame difference method is for larger, solid colour moving target, it is possible to produced in target internal empty Hole, so as to can not completely extract moving target.But direct application can influence the accuracy of model in video quality evaluation, Because in the case where continuous multiple frames have identical distortion information, frame difference method can filter out these distortion informations, and identical distortion The cumulative influence to video quality of type is bigger.
Due to the complexity of human visual system, the video quality evaluation algorithm that there is currently is in ageing and accuracy Good equilibrium is not reached.
The content of the invention
Consider high attention rate of the human visual system for marginal texture information, moving object and distortion information, the present invention A kind of wavelet field method for evaluating video quality eliminated based on background is proposed, considers that human visual system believes for marginal texture The high attention rate of breath, moving object and distortion information, the movable information of video sequence is extracted with improved frame difference method, and structure includes The time-space domain video block of movable information and distortion information, it is real with reference to advantage of the Haar wavelet transformations in extraction marginal information Wavelet field method for evaluating video quality (BSWQ) is showed.
A kind of wavelet field method for evaluating video quality eliminated based on background of the present invention, this method are comprised the following steps:
The first step:Global quality is calculated, using 4 grades of haar wavelet transforms (DWT) respectively to reference video RNAnd mistake True video DNEach frame decomposed, its decomposition coefficient represents as follows respectively:
CR(λ, θ, t, i)=DWT (Rt) (1)
CD(λ, θ, t, i)=DWT (Dt) (2)
Wherein, CR(λ,θ,t,i),CD(λ, θ, t, i) represents the wavelet decomposition of t frames in reference video and distortion video respectively Coefficient, t ∈ [1, N], Rt,DtT frames in reference, distortion video sequence are represented respectively;{ λ, θ } represents to be used for thumbnail respectively The coefficient subband of different scale and different directions, θ=2,3,4 represent horizontal, clinodiagonal and vertical subband respectively, the table of θ=1 Show approximation subband;I represents t frames, λ yardsticks, the position for the wavelet coefficient that direction is θ;
By the wavelet field different scale of each frame of video and the detail subbands coefficient of different directions, reference is obtained respectively and is regarded Fringing coefficient E under frequency and distortion video different scaleR(λ,t,i),ED(λ,t,i):
Calculate the fringing coefficient similitude that t frames yardstick is λ respectively using above formula:
In formula, T is normal number, and ESIM (λ, t, i) represents t frames in reference, distortion video, λ yardsticks, the office at the i of position Portion's similitude, works as ER,EDWhen just the same, value 1;
By calculating the standard deviation of local similarity, the quality ESIMD of the different scale of video sequence single-frame images is obtained (λ,t):
Wherein, NcFor the coefficient sum in the coefficient matrix of t frames, λ yardsticks in video.
The single frames quality of video sequence is:
Wherein, l represents haar wavelet transformation series, and value herein is 4.
Then, video sequence global quality QglobalIt is expressed as:
Second step:The local quality of video is calculated using background null method, i.e., first by reference video and distortion video with Continuous 3 frame of video sequence is one group, is configured to every group of nonoverlapping video block mutually;Secondly, with reference to average background method, will join Background of the frame of video class mean as the frame group is examined, is substituted in frame difference method using former frame as background, with the frame of centre one of frame group As the embodiment of spatial domain distortion position, remaining two frame subtracts background and obtains prospect frame to embody the feature of time-space domain, so as to by Between frame and two prospect frames collectively form time-space domain video block;
Reference video RNWith distortion video DNT frames in corresponding time-space domain video blockCalculation formula is as follows:
Wherein, BgThe background image of g group frame of video groups is represented, g ∈ { 1,2 ..., floor (N/3) }, N represent video The totalframes of sequence, m represent the position of the relatively whole video sequence of intermediate frame of frame group, and m=3g-1, t represent that present frame is relative The position of whole video sequence;
The method that video sequence single frames quality is calculated according to formula (8), to each frame quality in the video block of video time-space domain Weighed, so as to obtain the quality Q of frame of video groupg
The acquired multigroup time-space domain video block quality of video sequence is ranked up, and extracts second-rate part Final local quality Qs of the H% as video sequencelocal
Wherein, H represents the quality set { Q of video sequence frame group1,Q2,...,Qg,...,Qfloor(N/3)Middle by sequence The H% of worst part set afterwards, NHRepresent the element sum in set;
3rd step:Calculate the oeverall quality of video, i.e., according to the local quality and global quality of resulting video sequence, Video quality evaluation Model B SWQ is as follows:
The video quality evaluation algorithm of the wavelet field proposed by the present invention eliminated based on background is intended to improve video objective matter Amount evaluation and the uniformity of the subjective quality assessment of human eye;The algorithm can for different type of distortion, different scenes There is preferable video evaluation performance, and the algorithm has relatively low complexity, can realize real-time quality assessment.
Brief description of the drawings
Fig. 1 is wavelet sub-band coefficient index figure;
Fig. 2 is the wavelet field method for evaluating video quality overall flow figure eliminated based on background of the present invention;
Fig. 3 is BSWQ objective scorings and DMOS matched curve figure.
Embodiment
Embodiments of the present invention are described in further detail below in conjunction with accompanying drawing.
Concrete implementation step is as follows:
The first step:Calculate global quality:From human visual system curve, human eye it is most sensitive be midband, It is exactly the principal outline of image.Suitable image outline extracting method has important for video, image quality evaluation model Effect.Wavelet transformation is a kind of method extracted to frequency information, can be the subband under different scale by picture breakdown Image, so as to which the marginal information of image to be expressed as to the wavelet coefficient under different scale.Haar wavelet transformations are low with its complexity With effect it is good the advantages of image, video quality evaluation and compression in be widely used, based on this, the present invention use 4 grades Haar wavelet transforms (DWT) are respectively to reference video RNWith distortion video DNEach frame decomposed, its decomposition coefficient point Biao Shi not be as follows:
CR(λ, θ, t, i)=DWT (Rt) (1)
CD(λ, θ, t, i)=DWT (Dt) (2)
Wherein, CR(λ,θ,t,i),CD(λ, θ, t, i) represents the wavelet decomposition of t frames in reference video and distortion video respectively Coefficient, t ∈ [1, N], Rt,DtT frames in reference, distortion video sequence are represented respectively;{ λ, θ } represents to be used for thumbnail respectively The coefficient subband of different scale and different directions, θ=2,3,4 represent horizontal, clinodiagonal and vertical subband respectively, the table of θ=1 Show approximation subband;By taking 2 grades of discrete wavelet transformations as an example, it indexes as shown in Figure 1.I represents that t frames, λ yardsticks, direction are the small of θ The position of wave system number.
By the wavelet field different scale of each frame of video and the detail subbands coefficient of different directions, reference is obtained respectively and is regarded Fringing coefficient E under frequency and distortion video different scaleR(λ,t,i),ED(λ,t,i):
Calculate the fringing coefficient similitude that t frames yardstick is λ respectively using above formula:
In formula, T is normal number, mainly serves for ensuring ESIM stability, and ESIM (λ, t, i) represents reference, distortion video In t frames, λ yardsticks, the local similarity at the i of position, work as ER,EDWhen just the same, value 1.
By calculating the standard deviation of local similarity, the quality ESIMD of the different scale of video sequence single-frame images is obtained (λ,t):
Wherein, NcFor the coefficient sum in the coefficient matrix of t frames, λ yardsticks in video.
The single frames quality of video sequence is:
Wherein, l represents haar wavelet transformation series, and value herein is 4.
Then, video sequence global quality QglobalIt is expressed as:
Second step:Calculate the local quality of video:Used is background null method, and the specific practice is:It will refer to first Video and distortion video using continuous 3 frame of video sequence as one group, be configured to every group mutually nonoverlapping video block (due to video The image more than frame of encoding and decoding standard (such as H.264, HEVC) generally use 2 carries out estimation as reference frame and motion is mended Repay, i.e. present frame and the correlation of preceding a later frame is most strong);Secondly, with reference to average background method, using reference video frame class mean as The background of the frame group, substitute in frame difference method using former frame as the method for background, so as to avoid the distortion information of successive frame While filtering out, the foreground information of video sequence is effectively extracted.From masking effect, the signal under background, in different skies Between position, the visibility of signal has obvious otherness.When distortion information is located in the higher background of complexity, to video The influence of quality degradation is simultaneously little, thus the locus of distortion information has material impact for video quality evaluation.According to Video quality evaluation is the result of space-time domain information interaction, to consider background complexity, movable information and distortion letter simultaneously The influence for video quality is ceased, background is subtracted using the frame of centre one of frame group as the embodiment of spatial domain distortion position, remaining two frame Prospect frame is obtained to embody the feature of time-space domain, so as to which intermediate frame and two prospect frames are collectively formed into time-space domain video block.Ginseng Examine video RNWith distortion video DNT frames in corresponding time-space domain video blockCalculation is as follows:
Wherein, BgThe background image of g group frame of video groups is represented, g ∈ { 1,2 ..., floor (N/3) }, N represent video The totalframes of sequence, m represent the position of the relatively whole video sequence of intermediate frame of frame group, and m=3g-1, t represent that present frame is relative The position of whole video sequence.
The method that video sequence single frames quality is calculated according to formula (8), to each frame quality in the video block of video time-space domain Weighed, so as to obtain the quality Q of frame of video groupg
The value of quality is bigger, and the quality of the frame of video group is poorer.
By the time-domain perceptual quality of distortion video is generally determined by partial frame second-rate in video sequence, therefore Herein using worst quality convergence method, the acquired multigroup time-space domain video block quality of video sequence is ranked up first, and Extract final local quality Qs of the H% as video sequence of second-rate partlocal
Wherein, H represents the quality set { Q of video sequence frame group1,Q2,...,Qg,...,Qfloor(N/3)Middle by sequence The H% of worst part set afterwards, NHRepresent the element sum in set;
3rd step:Calculate the oeverall quality of video:According to the local quality and global quality of resulting video sequence, depending on Frequency Environmental Evaluation Model BSWQ is as follows:
.The predicted value of model is bigger, and video quality is higher.
Specific embodiment is illustrated below:
1) intend choosing T=1700, H=15;
2) global quality of video sequence and then according to formula (1)-(9) is calculated;
3) video sequence is divided into one group of 3 frame, using formula (10) (11) (12) to handling respectively frame group, built Time-space domain video block, so calculated with formula (13) calculate respectively each frame group quality;
4) after each frame group quality is obtained, using formula (14) calculate video sequence local quality;
5) global quality and local quality of combining video sequences, using formula (15) calculate video sequence total constitution Amount.
6) performance test:
The quality evaluating method proposed is tested from LIVE video databases, wherein including 10 kinds of different scenes Reference video and 150 distortion video sequences.Every kind of video source includes the type of distortion (Wireless of 4 kinds of varying levels Distortion, H.264 IP distortions, compression and MPEG-2 compressions), IP distortions have 3 kinds of varying levels, and its excess-three kind type of distortion is each Different degrees of distortion in own 4, i.e., the reference video under every kind of scene all contain 15 distortion videos.This algorithm uses video Two kinds commonly used in 4 kinds of evaluation indexes that Quality Expert's group (VQEG) proposes are used as evaluation index:Spearman rank order phase Relation number (SROCC), Pearson's linearly dependent coefficient (PLCC).Larger SROCC values and PLCC values represent video quality evaluation Algorithm has preferable accuracy and uniformity.
Table 1 is evaluation performance of the inventive method to different type of distortion videos, it can be seen that using background elimination algorithm To all there is good performance to various type of distortion videos, there is preferable robustness.
Table 2 is evaluation performance of the inventive method for 150 distortion videos, it can be seen that using background elimination algorithm With good versatility.
Table 3 be the inventive method for 250 frame video pa2_25fps.yuv run time, indicate the algorithm Real-Time Evaluation available for video.
Table 1
Table 2
Evaluation index SROCC PLCC
BSWQ 0.8265 0.8437
Table 3
Quality evaluating method Time (s)
The algorithm proposed 28.58
Fig. 3 by proposition Environmental Evaluation Model BSWQ predicted value and LIVE video libraries in difference mean subjective fraction (DMOS) scatter diagram.In figure solid line be Logistic function pair video sequences objective evaluation result and subjective data it is non-thread Property matched curve, includes Wireless distortions, H.264 IP distortions, distortion, MPEG distortions.If discrete point is uniformly distributed in plan Close around curve, then it represents that the predicted value of model and the correlation of subjective data are stronger.

Claims (1)

1. a kind of wavelet field method for evaluating video quality eliminated based on background, it is characterised in that this method comprises the following steps:
The first step:Global quality is calculated, using 4 grades of haar wavelet transforms respectively to reference video RNWith distortion video DN's Each frame is decomposed, and its decomposition coefficient represents as follows respectively:
CR(λ, θ, t, i)=DWT (Rt) (1)
CD(λ, θ, t, i)=DWT (Dt) (2)
Wherein, CR(λ,θ,t,i),CD(λ, θ, t, i) represents the coefficient of wavelet decomposition of t frames in reference video and distortion video respectively, t∈[1,N]、Rt,DtT frames in reference, distortion video sequence are represented respectively;{ λ, θ } represents to be used for thumbnail difference chi respectively The coefficient subband of degree and different directions, θ=2,3,4 represent horizontal, clinodiagonal and vertical subband respectively, and θ=1 represents approximate Subband;I represents t frames, λ yardsticks, the position for the wavelet coefficient that direction is θ;
By the wavelet field different scale of each frame of video and the detail subbands coefficient of different directions, obtain respectively reference video and Fringing coefficient E under distortion video different scaleR(λ,t,i),ED(λ,t,i):
<mrow> <msub> <mi>E</mi> <mi>R</mi> </msub> <mrow> <mo>(</mo> <mi>&amp;lambda;</mi> <mo>,</mo> <mi>t</mi> <mo>,</mo> <mi>i</mi> <mo>)</mo> </mrow> <mo>=</mo> <msqrt> <mrow> <munderover> <mo>&amp;Sigma;</mo> <mrow> <mi>&amp;theta;</mi> <mo>=</mo> <mn>2</mn> </mrow> <mn>4</mn> </munderover> <msubsup> <mi>C</mi> <mi>R</mi> <mn>2</mn> </msubsup> <mrow> <mo>(</mo> <mi>&amp;lambda;</mi> <mo>,</mo> <mi>&amp;theta;</mi> <mo>,</mo> <mi>t</mi> <mo>,</mo> <mi>i</mi> <mo>)</mo> </mrow> </mrow> </msqrt> <mo>-</mo> <mo>-</mo> <mo>-</mo> <mrow> <mo>(</mo> <mn>3</mn> <mo>)</mo> </mrow> </mrow>
<mrow> <msub> <mi>E</mi> <mi>D</mi> </msub> <mrow> <mo>(</mo> <mi>&amp;lambda;</mi> <mo>,</mo> <mi>t</mi> <mo>,</mo> <mi>i</mi> <mo>)</mo> </mrow> <mo>=</mo> <msqrt> <mrow> <munderover> <mo>&amp;Sigma;</mo> <mrow> <mi>&amp;theta;</mi> <mo>=</mo> <mn>2</mn> </mrow> <mn>4</mn> </munderover> <msubsup> <mi>C</mi> <mi>D</mi> <mn>2</mn> </msubsup> <mrow> <mo>(</mo> <mi>&amp;lambda;</mi> <mo>,</mo> <mi>&amp;theta;</mi> <mo>,</mo> <mi>t</mi> <mo>,</mo> <mi>i</mi> <mo>)</mo> </mrow> </mrow> </msqrt> <mo>-</mo> <mo>-</mo> <mo>-</mo> <mrow> <mo>(</mo> <mn>4</mn> <mo>)</mo> </mrow> </mrow>
Calculate the fringing coefficient similitude that t frames yardstick is λ respectively using above formula:
<mrow> <mi>E</mi> <mi>S</mi> <mi>I</mi> <mi>M</mi> <mrow> <mo>(</mo> <mi>&amp;lambda;</mi> <mo>,</mo> <mi>t</mi> <mo>,</mo> <mi>i</mi> <mo>)</mo> </mrow> <mo>=</mo> <mfrac> <mrow> <mn>2</mn> <mo>&amp;CenterDot;</mo> <msub> <mi>E</mi> <mi>R</mi> </msub> <mrow> <mo>(</mo> <mi>&amp;lambda;</mi> <mo>,</mo> <mi>t</mi> <mo>,</mo> <mi>i</mi> <mo>)</mo> </mrow> <mo>&amp;CenterDot;</mo> <msub> <mi>E</mi> <mi>D</mi> </msub> <mrow> <mo>(</mo> <mi>&amp;lambda;</mi> <mo>,</mo> <mi>t</mi> <mo>,</mo> <mi>i</mi> <mo>)</mo> </mrow> <mo>+</mo> <mi>T</mi> </mrow> <mrow> <msubsup> <mi>E</mi> <mi>R</mi> <mn>2</mn> </msubsup> <mrow> <mo>(</mo> <mi>&amp;lambda;</mi> <mo>,</mo> <mi>t</mi> <mo>,</mo> <mi>i</mi> <mo>)</mo> </mrow> <mo>+</mo> <msubsup> <mi>E</mi> <mi>D</mi> <mn>2</mn> </msubsup> <mrow> <mo>(</mo> <mi>&amp;lambda;</mi> <mo>,</mo> <mi>t</mi> <mo>,</mo> <mi>i</mi> <mo>)</mo> </mrow> <mo>+</mo> <mi>T</mi> </mrow> </mfrac> <mo>-</mo> <mo>-</mo> <mo>-</mo> <mrow> <mo>(</mo> <mn>5</mn> <mo>)</mo> </mrow> </mrow>
In formula, T is normal number, and ESIM (λ, t, i) represents t frames in reference, distortion video, λ yardsticks, the Local Phase at the i of position Like property, work as ER,EDWhen just the same, value 1;
By calculating the standard deviation of local similarity, obtain the different scale of video sequence single-frame images quality ESIMD (λ, t):
<mrow> <mi>E</mi> <mi>S</mi> <mi>I</mi> <mi>M</mi> <mi>D</mi> <mrow> <mo>(</mo> <mi>&amp;lambda;</mi> <mo>,</mo> <mi>t</mi> <mo>)</mo> </mrow> <mo>=</mo> <msqrt> <mrow> <mfrac> <mn>1</mn> <msub> <mi>N</mi> <mi>c</mi> </msub> </mfrac> <munderover> <mo>&amp;Sigma;</mo> <mrow> <mi>i</mi> <mo>=</mo> <mn>1</mn> </mrow> <msub> <mi>N</mi> <mi>c</mi> </msub> </munderover> <msup> <mrow> <mo>(</mo> <mi>E</mi> <mi>S</mi> <mi>I</mi> <mi>M</mi> <mo>(</mo> <mrow> <mi>&amp;lambda;</mi> <mo>,</mo> <mi>t</mi> <mo>,</mo> <mi>i</mi> </mrow> <mo>)</mo> <mo>-</mo> <mi>E</mi> <mi>S</mi> <mi>I</mi> <mi>M</mi> <mi>M</mi> <mo>(</mo> <mrow> <mi>&amp;lambda;</mi> <mo>,</mo> <mi>t</mi> </mrow> <mo>)</mo> <mo>)</mo> </mrow> <mn>2</mn> </msup> </mrow> </msqrt> <mo>-</mo> <mo>-</mo> <mo>-</mo> <mrow> <mo>(</mo> <mn>6</mn> <mo>)</mo> </mrow> </mrow>
<mrow> <mi>E</mi> <mi>S</mi> <mi>I</mi> <mi>M</mi> <mi>M</mi> <mrow> <mo>(</mo> <mi>&amp;lambda;</mi> <mo>,</mo> <mi>t</mi> <mo>)</mo> </mrow> <mo>=</mo> <mfrac> <mn>1</mn> <msub> <mi>N</mi> <mi>c</mi> </msub> </mfrac> <munderover> <mo>&amp;Sigma;</mo> <mrow> <mi>i</mi> <mo>=</mo> <mn>1</mn> </mrow> <msub> <mi>N</mi> <mi>c</mi> </msub> </munderover> <mi>E</mi> <mi>S</mi> <mi>I</mi> <mi>M</mi> <mrow> <mo>(</mo> <mi>&amp;lambda;</mi> <mo>,</mo> <mi>t</mi> <mo>,</mo> <mi>i</mi> <mo>)</mo> </mrow> <mo>-</mo> <mo>-</mo> <mo>-</mo> <mrow> <mo>(</mo> <mn>7</mn> <mo>)</mo> </mrow> </mrow>
Wherein, NcFor the coefficient sum in the coefficient matrix of t frames, λ yardsticks in video.
The single frames quality of video sequence is:
<mrow> <msub> <mi>Q</mi> <mi>t</mi> </msub> <mrow> <mo>(</mo> <msub> <mi>R</mi> <mi>t</mi> </msub> <mo>,</mo> <msub> <mi>D</mi> <mi>t</mi> </msub> <mo>)</mo> </mrow> <mo>=</mo> <mfrac> <mn>1</mn> <mi>l</mi> </mfrac> <munderover> <mo>&amp;Sigma;</mo> <mrow> <mi>&amp;lambda;</mi> <mo>=</mo> <mn>1</mn> </mrow> <mi>l</mi> </munderover> <mi>E</mi> <mi>S</mi> <mi>I</mi> <mi>M</mi> <mi>D</mi> <mrow> <mo>(</mo> <mi>&amp;lambda;</mi> <mo>,</mo> <mi>t</mi> <mo>)</mo> </mrow> <mo>-</mo> <mo>-</mo> <mo>-</mo> <mrow> <mo>(</mo> <mn>8</mn> <mo>)</mo> </mrow> </mrow>
Wherein, l represents haar wavelet transformation series, and value herein is 4.
Then, video sequence global quality QglobalIt is expressed as:
<mrow> <msub> <mi>Q</mi> <mrow> <mi>g</mi> <mi>l</mi> <mi>o</mi> <mi>b</mi> <mi>a</mi> <mi>l</mi> </mrow> </msub> <mo>=</mo> <mfrac> <mn>1</mn> <mi>N</mi> </mfrac> <munderover> <mo>&amp;Sigma;</mo> <mrow> <mi>t</mi> <mo>=</mo> <mn>1</mn> </mrow> <mi>N</mi> </munderover> <msub> <mi>Q</mi> <mi>t</mi> </msub> <mrow> <mo>(</mo> <msub> <mi>R</mi> <mi>t</mi> </msub> <mo>,</mo> <msub> <mi>D</mi> <mi>t</mi> </msub> <mo>)</mo> </mrow> <mo>-</mo> <mo>-</mo> <mo>-</mo> <mrow> <mo>(</mo> <mn>9</mn> <mo>)</mo> </mrow> </mrow>
Second step:The local quality of video is calculated using background null method, i.e., first by reference video and distortion video with video 3 frames of Sequentially continuous are one group, are configured to every group of nonoverlapping video block mutually;Secondly, with reference to average background method, it will refer to and regard Background of the frequency frame class mean as the frame group, substitute frame difference method in using former frame as background, using the frame of centre one of frame group as The embodiment of spatial domain distortion position, remaining two frame subtract background and obtain prospect frame to embody the feature of time-space domain, so as to by intermediate frame Time-space domain video block is collectively formed with two prospect frames;
Reference video RNWith distortion video DNT frames F in corresponding time-space domain video blockt R,Ft DCalculation formula is as follows:
<mrow> <msubsup> <mi>F</mi> <mi>t</mi> <mi>R</mi> </msubsup> <mo>=</mo> <mfenced open = "{" close = ""> <mtable> <mtr> <mtd> <mrow> <msub> <mi>R</mi> <mi>t</mi> </msub> <mo>-</mo> <msub> <mi>B</mi> <mi>g</mi> </msub> <mo>,</mo> <mi>t</mi> <mo>&amp;Element;</mo> <mo>&amp;lsqb;</mo> <mi>m</mi> <mo>-</mo> <mn>1</mn> <mo>,</mo> <mi>m</mi> <mo>)</mo> <mo>&amp;cup;</mo> <mo>(</mo> <mi>m</mi> <mo>,</mo> <mi>m</mi> <mo>+</mo> <mn>1</mn> <mo>&amp;rsqb;</mo> </mrow> </mtd> </mtr> <mtr> <mtd> <mrow> <msub> <mi>R</mi> <mi>t</mi> </msub> <mo>,</mo> <mi>t</mi> <mo>=</mo> <mi>m</mi> </mrow> </mtd> </mtr> </mtable> </mfenced> <mo>-</mo> <mo>-</mo> <mo>-</mo> <mrow> <mo>(</mo> <mn>10</mn> <mo>)</mo> </mrow> </mrow>
<mrow> <msubsup> <mi>F</mi> <mi>t</mi> <mi>D</mi> </msubsup> <mo>=</mo> <mfenced open = "{" close = ""> <mtable> <mtr> <mtd> <mrow> <msub> <mi>D</mi> <mi>t</mi> </msub> <mo>-</mo> <msub> <mi>B</mi> <mi>g</mi> </msub> <mo>,</mo> <mi>t</mi> <mo>&amp;Element;</mo> <mo>&amp;lsqb;</mo> <mi>m</mi> <mo>-</mo> <mn>1</mn> <mo>,</mo> <mi>m</mi> <mo>)</mo> <mo>&amp;cup;</mo> <mo>(</mo> <mi>m</mi> <mo>,</mo> <mi>m</mi> <mo>+</mo> <mn>1</mn> <mo>&amp;rsqb;</mo> </mrow> </mtd> </mtr> <mtr> <mtd> <mrow> <msub> <mi>D</mi> <mi>t</mi> </msub> <mo>,</mo> <mi>t</mi> <mo>=</mo> <mi>m</mi> </mrow> </mtd> </mtr> </mtable> </mfenced> <mo>-</mo> <mo>-</mo> <mo>-</mo> <mrow> <mo>(</mo> <mn>11</mn> <mo>)</mo> </mrow> </mrow>
<mrow> <msub> <mi>B</mi> <mi>g</mi> </msub> <mo>=</mo> <mfrac> <mn>1</mn> <mn>3</mn> </mfrac> <mo>&amp;CenterDot;</mo> <mrow> <mo>(</mo> <munderover> <mo>&amp;Sigma;</mo> <mrow> <mi>t</mi> <mo>=</mo> <mi>m</mi> <mo>-</mo> <mn>1</mn> </mrow> <mrow> <mi>m</mi> <mo>+</mo> <mn>1</mn> </mrow> </munderover> <msub> <mi>R</mi> <mi>t</mi> </msub> <mo>)</mo> </mrow> <mo>-</mo> <mo>-</mo> <mo>-</mo> <mrow> <mo>(</mo> <mn>12</mn> <mo>)</mo> </mrow> </mrow>
Wherein, BgThe background image of g group frame of video groups is represented, g ∈ { 1,2 ..., floor (N/3) }, N represent video sequence Totalframes, m represent the position of the relatively whole video sequence of intermediate frame of frame group, and m=3g-1, t represent that present frame entirely regards relatively The position of frequency sequence;
The method that video sequence single frames quality is calculated according to formula (8), each frame quality in the video block of video time-space domain is carried out Weigh, so as to obtain the quality Q of frame of video groupg
<mrow> <msub> <mi>Q</mi> <mi>g</mi> </msub> <mo>=</mo> <mfrac> <mn>1</mn> <mn>3</mn> </mfrac> <munderover> <mo>&amp;Sigma;</mo> <mrow> <mi>t</mi> <mo>=</mo> <mi>m</mi> <mo>-</mo> <mn>1</mn> </mrow> <mrow> <mi>m</mi> <mo>+</mo> <mn>1</mn> </mrow> </munderover> <msub> <mi>Q</mi> <mi>t</mi> </msub> <mrow> <mo>(</mo> <msubsup> <mi>F</mi> <mi>t</mi> <mi>R</mi> </msubsup> <mo>,</mo> <msubsup> <mi>F</mi> <mi>t</mi> <mi>D</mi> </msubsup> <mo>)</mo> </mrow> <mo>-</mo> <mo>-</mo> <mo>-</mo> <mrow> <mo>(</mo> <mn>13</mn> <mo>)</mo> </mrow> </mrow>
The acquired multigroup time-space domain video block quality of video sequence is ranked up, and the H% for extracting second-rate part makees For the final local quality Q of video sequencelocal
<mrow> <msub> <mi>Q</mi> <mrow> <mi>l</mi> <mi>o</mi> <mi>c</mi> <mi>a</mi> <mi>l</mi> </mrow> </msub> <mo>=</mo> <mfrac> <mn>1</mn> <msub> <mi>N</mi> <mi>H</mi> </msub> </mfrac> <munder> <mo>&amp;Sigma;</mo> <mrow> <mi>k</mi> <mo>&amp;Element;</mo> <mi>H</mi> </mrow> </munder> <msub> <mi>Q</mi> <mi>k</mi> </msub> <mo>-</mo> <mo>-</mo> <mo>-</mo> <mrow> <mo>(</mo> <mn>14</mn> <mo>)</mo> </mrow> </mrow>
Wherein, H represents the quality set { Q of video sequence frame group1,Q2,...,Qg,...,Qfloor(N/3)In it is worst after sequence Partial H% set, NHRepresent the element sum in set;
3rd step:The oeverall quality of video is calculated, i.e., according to the local quality and global quality of resulting video sequence, video Environmental Evaluation Model BSWQ is as follows:
<mrow> <mi>B</mi> <mi>S</mi> <mi>W</mi> <mi>Q</mi> <mo>=</mo> <mfrac> <mn>1</mn> <msqrt> <mrow> <mfrac> <mn>1</mn> <mn>2</mn> </mfrac> <mrow> <mo>(</mo> <msub> <mi>Q</mi> <mrow> <mi>g</mi> <mi>l</mi> <mi>o</mi> <mi>b</mi> <mi>a</mi> <mi>l</mi> </mrow> </msub> <mo>+</mo> <msub> <mi>Q</mi> <mrow> <mi>l</mi> <mi>o</mi> <mi>c</mi> <mi>a</mi> <mi>l</mi> </mrow> </msub> <mo>)</mo> </mrow> </mrow> </msqrt> </mfrac> <mo>-</mo> <mo>-</mo> <mo>-</mo> <mrow> <mo>(</mo> <mn>15</mn> <mo>)</mo> </mrow> <mo>.</mo> </mrow>
CN201710926882.8A 2017-10-08 2017-10-08 The wavelet field method for evaluating video quality eliminated based on background Active CN107809631B (en)

Priority Applications (1)

Application Number Priority Date Filing Date Title
CN201710926882.8A CN107809631B (en) 2017-10-08 2017-10-08 The wavelet field method for evaluating video quality eliminated based on background

Applications Claiming Priority (1)

Application Number Priority Date Filing Date Title
CN201710926882.8A CN107809631B (en) 2017-10-08 2017-10-08 The wavelet field method for evaluating video quality eliminated based on background

Publications (2)

Publication Number Publication Date
CN107809631A true CN107809631A (en) 2018-03-16
CN107809631B CN107809631B (en) 2019-05-14

Family

ID=61584092

Family Applications (1)

Application Number Title Priority Date Filing Date
CN201710926882.8A Active CN107809631B (en) 2017-10-08 2017-10-08 The wavelet field method for evaluating video quality eliminated based on background

Country Status (1)

Country Link
CN (1) CN107809631B (en)

Cited By (1)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN113709453A (en) * 2021-09-13 2021-11-26 北京车和家信息技术有限公司 Video quality evaluation method, device, equipment and medium

Citations (2)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN104811691A (en) * 2015-04-08 2015-07-29 宁波大学 Stereoscopic video quality objective evaluation method based on wavelet transformation
CN104918039A (en) * 2015-05-05 2015-09-16 四川九洲电器集团有限责任公司 Image quality evaluation method and image quality evaluation system

Patent Citations (2)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN104811691A (en) * 2015-04-08 2015-07-29 宁波大学 Stereoscopic video quality objective evaluation method based on wavelet transformation
CN104918039A (en) * 2015-05-05 2015-09-16 四川九洲电器集团有限责任公司 Image quality evaluation method and image quality evaluation system

Non-Patent Citations (2)

* Cited by examiner, † Cited by third party
Title
SHADIYA P ETA.: "A perceptual distortion measure for video quality assessment using wavelet and statistical measures", 《IEEE INTERNATIONAL CONFERENCE ON SIGNAL PROCESSING,INFROMATICS,COMMUNICATION AND ENERGY SYSTEMS》 *
YOU J ETA.: "Attention modeling for video quality assessment:Balancing global quality and local quality", 《IEEE INTERNATIONAL CONFERENCE ON MULTIMEDIA AND EXPO.TRONDHEIM》 *

Cited By (2)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN113709453A (en) * 2021-09-13 2021-11-26 北京车和家信息技术有限公司 Video quality evaluation method, device, equipment and medium
CN113709453B (en) * 2021-09-13 2023-09-08 北京车和家信息技术有限公司 Video quality assessment method, device, equipment and medium

Also Published As

Publication number Publication date
CN107809631B (en) 2019-05-14

Similar Documents

Publication Publication Date Title
Li et al. Spatiotemporal statistics for video quality assessment
Moorthy et al. Efficient motion weighted spatio-temporal video SSIM index
Xu et al. No-reference video quality assessment via feature learning
Maalouf et al. CYCLOP: A stereo color image quality assessment metric
Seshadrinathan et al. A structural similarity metric for video based on motion models
US8027542B2 (en) High speed video action recognition and localization
US7733372B2 (en) Method and system for video quality measurements
CN104219525B (en) Perception method for video coding based on conspicuousness and minimum discernable distortion
US8345971B2 (en) Method and system for spatial-temporal denoising and demosaicking for noisy color filter array videos
CN104023227B (en) A kind of objective evaluation method of video quality based on spatial domain and spatial structure similitude
CN104243973B (en) Video perceived quality non-reference objective evaluation method based on areas of interest
CN107241607B (en) Visual perception coding method based on multi-domain JND model
CN104378636B (en) A kind of video encoding method and device
CN103037217B (en) The image detected in interpolated image damages
CN109831664B (en) Rapid compressed stereo video quality evaluation method based on deep learning
CN106875389B (en) Stereo video quality evaluation method based on motion significance
CN105635743A (en) Minimum noticeable distortion method and system based on saliency detection and total variation
CN105049851A (en) Channel no-reference image quality evaluation method based on color perception
Moorthy et al. Visual perception and quality assessment
CN107071423A (en) Application process of the vision multi-channel model in stereoscopic video quality objective evaluation
CN105828064A (en) No-reference video quality evaluation method integrating local and global temporal and spatial characteristics
CN108846800A (en) A kind of non-reference picture quality appraisement method of image super-resolution rebuilding
CN106412571A (en) Video quality evaluation method based on gradient similarity standard deviation
CN107492085B (en) Stereo image quality evaluation method based on dual-tree complex wavelet transform
CN107809631A (en) The wavelet field method for evaluating video quality eliminated based on background

Legal Events

Date Code Title Description
PB01 Publication
PB01 Publication
SE01 Entry into force of request for substantive examination
SE01 Entry into force of request for substantive examination
GR01 Patent grant
GR01 Patent grant