WO2013086654A1 - Method and apparatus for video quality measurement - Google Patents

Method and apparatus for video quality measurement Download PDF

Info

Publication number
WO2013086654A1
WO2013086654A1 PCT/CN2011/002096 CN2011002096W WO2013086654A1 WO 2013086654 A1 WO2013086654 A1 WO 2013086654A1 CN 2011002096 W CN2011002096 W CN 2011002096W WO 2013086654 A1 WO2013086654 A1 WO 2013086654A1
Authority
WO
WIPO (PCT)
Prior art keywords
video
value
average
macroblocks
quality score
Prior art date
Application number
PCT/CN2011/002096
Other languages
French (fr)
Inventor
Fan Zhang
Ning Liao
Kai Xie
Zhibo Chen
Original Assignee
Thomson Licensing
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Priority to JP2014546260A priority Critical patent/JP2015505196A/en
Priority to MX2014007041A priority patent/MX2014007041A/en
Priority to CN201180075448.4A priority patent/CN103999461A/en
Priority to BR112014014349A priority patent/BR112014014349A2/en
Priority to AU2011383036A priority patent/AU2011383036B2/en
Priority to US14/364,144 priority patent/US9961340B2/en
Priority to RU2014128824/08A priority patent/RU2595917C2/en
Priority to KR1020147015877A priority patent/KR20140102215A/en
Application filed by Thomson Licensing filed Critical Thomson Licensing
Priority to EP11877571.7A priority patent/EP2792144B1/en
Priority to PCT/CN2011/002096 priority patent/WO2013086654A1/en
Priority to CA2857367A priority patent/CA2857367A1/en
Publication of WO2013086654A1 publication Critical patent/WO2013086654A1/en
Priority to ZA2014/03615A priority patent/ZA201403615B/en
Priority to HK15103844.5A priority patent/HK1203722A1/en

Links

Classifications

    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04NPICTORIAL COMMUNICATION, e.g. TELEVISION
    • H04N17/00Diagnosis, testing or measuring for television systems or their details
    • H04N17/004Diagnosis, testing or measuring for television systems or their details for digital television systems
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04NPICTORIAL COMMUNICATION, e.g. TELEVISION
    • H04N19/00Methods or arrangements for coding, decoding, compressing or decompressing digital video signals
    • H04N19/10Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using adaptive coding
    • H04N19/102Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using adaptive coding characterised by the element, parameter or selection affected or controlled by the adaptive coding
    • H04N19/124Quantisation
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04NPICTORIAL COMMUNICATION, e.g. TELEVISION
    • H04N19/00Methods or arrangements for coding, decoding, compressing or decompressing digital video signals
    • H04N19/10Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using adaptive coding
    • H04N19/134Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using adaptive coding characterised by the element, parameter or criterion affecting or controlling the adaptive coding
    • H04N19/154Measured or subjectively estimated visual quality after decoding, e.g. measurement of distortion
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04NPICTORIAL COMMUNICATION, e.g. TELEVISION
    • H04N19/00Methods or arrangements for coding, decoding, compressing or decompressing digital video signals
    • H04N19/10Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using adaptive coding
    • H04N19/169Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using adaptive coding characterised by the coding unit, i.e. the structural portion or semantic portion of the video signal being the object or the subject of the adaptive coding
    • H04N19/17Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using adaptive coding characterised by the coding unit, i.e. the structural portion or semantic portion of the video signal being the object or the subject of the adaptive coding the unit being an image region, e.g. an object
    • H04N19/176Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using adaptive coding characterised by the coding unit, i.e. the structural portion or semantic portion of the video signal being the object or the subject of the adaptive coding the unit being an image region, e.g. an object the region being a block, e.g. a macroblock

Definitions

  • This invention relates to video quality measurement, in particular with the assessment of a compressed video without reference to a copy of the original uncompressed video.
  • IPTV Internet protocol television
  • video programs have different format stages during their life circle.
  • a video encoder compresses the video program to a bit stream, also referred to as an elementary stream (ES).
  • ES is further packetized into a transport stream (TS) and finally transmitted in an IP channel.
  • TS transport stream
  • Video quality can be measured using data that are obtained by accessing the transport stream, elementary stream or decoded video.
  • using the transport stream is generally the fastest but the least accurate, since it has the smallest amount of video data available; using the decoded video is often accurate but the slowest, since decoding the video is computationally expensive; using the elementary stream can achieve a tradeoff between the accuracy and the computational complexity.
  • Currently, particularly video quality measurement based on the elementary stream is being investigated.
  • Video compression generally employs quantization techniques. Quantization is a lossy compression technique by means of limiting the precision of signal values. It is well known that quantization is a significant factor to artifact visibility, and the quantization parameter (QP) is a powerful predictor to the video quality. Various functions of video quality with respect to QP have been provided in the literature, such as linear function [1 , 2] and exponential function [3]. However, they are insufficiently accurate for the relatively large and/or the relatively small QP level, and thus their results are not satisfactory for low-bandwidth or high-fidelity applications. The content complexity is another critical factor for video quality measurement (VQM). Visual artifacts in complex videos are more likely to be tolerated by the human eye, and thus show better quality. Therefore, content complexity in combination with QP can improve the accuracy of quality measurement, compared to using the QP alone.
  • VQM video quality measurement
  • content complexity may be quantified as the variance, the gradient, or the edge filter response of pixel values, or their combinations.
  • the traditional methods have at least the following disadvantages.
  • Such features are not tightly correlated with human visual perception.
  • a video with large content complexity may have not only rich texture and irregular motion, but also many edges and/or regular motion.
  • visual artifacts are more likely to be tolerated in texture and irregularly (i.e., stochastical- ly) moving regions, but ordinarily more attractive and visible in edges or regularly (i.e., constantly) moving regions.
  • Second, such features can hardly be computed until the pixels are recovered after full decoding. Thus, the traditional complexity measurement is computational expensive since it requires full decoding of the video.
  • the present invention solves at least the problem of improving accuracy and/or efficiency of video quality measurement.
  • the method for accurately predicting video quality uses a rational function of the QP, which is corrected by a correction function that depends on content unpredictability (CU).
  • the correction function is a power function of the CU.
  • a first advantage of the invention is high accuracy of video quality prediction, which is confirmed by subjective experiments conforming to ITU-T SG 12 [5] as well as statistical verifications.
  • the accuracy is ensured by at least two features.
  • One is calculating an average QP and using a rational function of the average QP for predicting a basic impairment due to quantization. This function can capture the effect of quality saturation at both the relatively large and the relatively small QP levels, and thus provides a sufficiently accurate result.
  • the other feature is using a power function of CU to correct the QP-based prediction, which improves the prediction accuracy further.
  • the CU as a video feature, can discriminate irregular changes from regular changes and from "no change" in a video signal. Consequently, the CU is more powerful to capture the influence of content features on perceived quality.
  • the invention also provides a fast algorithm to estimate the CU from the video elementary stream, which results in a second advantage.
  • the second advantage is that the method requires only the elementary stream of a video instead of fully decoded video, and thus is computationally less expensive than known methods.
  • the VQM can be deployed e.g. in user terminals, set-top boxes, home gateways, routers, or video streaming servers, so as to monitor the video quality and provide feedback for service planning.
  • the present invention in one aspect, concerns a method for estimating video quality of an encoded video stream that comprises macroblock prediction residuals, wherein the method comprises a step of calculating, for a video sequence comprising a plurality of frames, an overall quality score Q from a basic quality score Qb multiplied by a correction factor Qc, wherein the basic quality score Qb is obtained from an average quantization parameter QP over the frames of the video sequence, and wherein the correction factor Qc is obtained from an average pixel variance within the prediction residuals of macroblocks in the video sequence.
  • the invention concerns an apparatus for estimating video quality of an encoded video stream that comprises macroblock prediction residuals, the apparatus comprising a processing element for calculating, for a video sequence comprising a plurality of frames, an overall quality score Q from a basic quality 5 score Qb multiplied by a correction factor Qc, wherein the basic quality score Qb is obtained from an average quantization parameter value over the frames of the video sequence, and wherein the correction factor Qc is obtained from an average pixel variance within the prediction residuals of macroblocks in the video sequence.
  • the apparatus comprises one or more of the following: a quantization parameter parser for detecting and averaging the quantization parameters of the macroblocks (MBs), wherein the quantization parameter parser provides the quantization parameter of the MBs and said average quantization i s parameter value;
  • a quantization parameter parser for detecting and averaging the quantization parameters of the macroblocks (MBs), wherein the quantization parameter parser provides the quantization parameter of the MBs and said average quantization i s parameter value;
  • a DCT coefficients parser for detecting and extracting the DCT coefficients of the MBs
  • a Content Unpredictability calculator for calculating, according to the DCT coefficients of the MBs and the quantization parameter of the MBs, a first value0 CU
  • a de-packetizer for de-packetizing (or de-packing) the Transport Stream, wherein an Elementary Stream (ES) is obtained, and
  • the invention relates to a computer readable medium having executable instructions stored thereon to cause a computer to perform a method comprising a step of calculating, for a video sequence comprising a plurality of0 frames, an overall quality score Q from a basic quality score Qb multiplied by a correction factor Qc, wherein the basic quality score Qb is obtained from an average quantization parameter ⁇ QP over the frames of the video sequence, and wherein the correction factor Qc is obtained from an average pixel variance within the prediction residuals of macroblocks in the video sequence.
  • Fig.1 the structure of a video quality measurement tool
  • Fig.6 improved quality prediction according to the invention especially at a large
  • Fig.7 a flow chart of video quality measurement. Detailed description of the invention
  • Fig.1 shows the structure of a video quality measurement (VQM) tool 106 within a video transmission system 100.
  • Typical format stages 01 ,102,103 of a video program are as follows: a video encoder 104 compresses the video programs 101 to a bit stream, also referred to as an elementary stream (ES) (in Fig.1 exemplarily only within video encoder 104, therefore not shown). The ES is further packetized into a transport stream (TS) 103 and then transmitted in a transmission channel, e.g. an IP channel.
  • a transmission channel e.g. an IP channel.
  • the VQM tool 106 de-packetizes the video TS 103, thus obtaining the video ES, then parses and averages the QP and obtains a content unpredictability CU value from the obtained video ES, and finally predicts the video quality Q from the QP and the CU value. More details are given below.
  • this process is fast and does not require fully decoding the video.
  • the video quality is predicted by a function that is a multiplication of two sub- functions.
  • the value of a first sub-function determines the basic impairment due to quantization, i.e. a basic quality score.
  • the first sub-function is a rational function of the QP.
  • the second sub-function is a correction factor, and its value quantifies the influence of content unpredictability (CU) on the perceived quality.
  • the second sub-function is a power function about the CU of the video, as further specified below.
  • the CU is a value associated with a video, and advantageously can be computed from the video ES as described below, specifically by using the quantized DCT coefficients of the video.
  • the CU of a video reflects the intrinsic features of the content, i.e. provides a value that characterizes the content. Thus, it can be used for determining the content's impact on the perceived video quality.
  • QPs and quantized DCT coefficients are recovered after selective entropy decoding in a selective entropy decoding module 108.
  • Full decoding of the video which would include complete run-length decoding, de-quantization, inverse discrete cosine transform (IDCT) and residual
  • the video quality prediction of the present invention is of the "no-reference" NR (or non-intrusive) type. That is, it does not need to access a copy of the original uncompressed video. Further, there is no need for the quality prediction of the present invention to fully decode the video.
  • the prediction uses data that can directly be extracted from a video elementary stream.
  • VQM video quality measurement
  • the input of the VQM tool 106 may be a transport stream 103, which may be generated by a video encoder and packetizer 104 from video programs 101.
  • the video encoder and packetizer 104 may but needs not be part of the invention.
  • a video ES 102 comprising encoded video programs 101 is directly input to the VQM tool 106.
  • the output 115 of the VQM tool 106 is a predicted quality score Q associated with the input video.
  • a depacketizer 107 parses the received transport stream 103 to obtain the video elementary stream 102.
  • video features including the average QP 112 and the global CU 113 are obtained by selective entropy decoding in a selective entropy decoder 108, parsing the selectively entropy decoded data in a QP parser 109 and a DCT coefficients parser 110, wherein the QP parser 109 provides the average QP 112, and calculating the global CU in a CU calculator 111.
  • a quality predictor module 114 determines a quality score Q according to the video features via a predetermined measurement function.
  • the video features are obtained from the output of the selective entropy decoder 108 by two simultaneously executing function blocks, or threads.
  • the QP parser 109 picks up (i.e. extracts) the QPs of each MB and provides the QPs to the CU calculator 111. Further, the QP parser 109 averages the QPs over a plurality of MBs and outputs the resulting average value 112.
  • the DCT coefficients parser 110 picks up (i.e.
  • the CU calculator 111 calculates the local CU according to the DCT coefficients from the DCT coefficients parser 110 and the corresponding QP from the QP parser 109. Finally, the CU calculator 111 averages the local CUs and outputs the global CU 113 obtained by averaging all local CUs.
  • the measurement function works in the quality estimator module 114 according to a mathematical model that depicts the video quality against at least two video features, including the average QP 112 and the global CU 113.
  • the measurement function is defined as:
  • Q h is the basic quality level due to quantization
  • Q c is an item to correct h according to the Content Unpredictability of the video
  • Q is the final predicted quality
  • the basic quality level due to quantization Q h and the correction factor Q c for correcting the basic quality level Q b according to the CU are calculated according to
  • the predetermined parameters are positive, rational and selected from the following ranges:
  • b 3 1 ,65.
  • Figs. 2-4 shows for exemplary video sequences different views of fitting a subjective quality with the measurement function.
  • Fig.2 shows a holistic view.
  • Figs. 2-4 depict fitting the subjective quality scores against the average QP and the global CU.
  • Each point corresponds to one of the 64 videos from a subjective database.
  • the compressed videos from different original videos are marked by different markers, as illustrated by the legend in Fig.3.
  • Subjective quality is rated by 24 reliable subjects during a subjective experiment conforming to ITU-T SG 12 [5].
  • the average QP and the global CU are of the corresponding video are rated.
  • Fig. 2 shows how the predicted quality scores match with the subjective quality scores via the measurement function according to Eq. (1-3), which in this view forms a surface.
  • Fig.3 shows a left side view, and Fig.4 a right side view of the surface (sectional drawings).
  • the experimental result of the correlation between J O the predicted quality scores and the subjective scores confirms that the
  • each macroblock has a 15 QP.
  • the QP indexes a predefined table of the quantization step QSTEP, which is the linear quantizer used for quantizing the DCT coefficients inside a current MB.
  • the average QP 112 is the mean or average QPs among all the MBs.
  • Fig. 5 shows fitting (i.e. mapping) the basic impairment (i.e. left side view as in Fig.3) to different functions of the average QP.
  • Fig. 5 shows the subjective quality against the average QP alone.
  • Fig.5 a fitting the basic impairment with a linear function is depicted.
  • Fig.5 b fitting the basic impairment with an exponential function is depicted.
  • Fig.5 c fitting the basic impairment with a rational function is depicted.
  • Each compressed video is depicted as a point.
  • Fig. 5 shows fitting (i.e. mapping) the basic impairment (i.e. left side view as in Fig.3) to different functions of the average QP.
  • Fig. 5 shows the subjective quality against the average QP alone.
  • Fig.5 a fitting the basic impairment with a linear function is depicted.
  • Fig.5 b fitting the basic impairment with an exponential function is depicted.
  • Fig.5 c fitting the basic impairment with a rational function is
  • FIG. 5 shows fitting the points (quality against QP) in Fig.5 a) with a linear function, as in [1 ,2], in Fig.5 b) with an exponential function, as in [3], and in Fig.5 c) with a rational function as in Eq. (2), according to the0 present invention.
  • Regression analysis shows that the rational function is closest to the position of the points.
  • the mapping the quality to a rational function according to the invention is a better solution than to an exponential or linear function, since it can, firstly, match the quality saturation trends at small QP levels and at large QP levels, and secondly, capture the slow saturation (large turn) at the small QP levels but also the sharp saturation (steep turn) at the large QP levels.
  • the present invention uses content unpredictability (CU) to discriminate irregular changes from regular changes or no changes in a video signal.
  • Traditional content complexity is computed with the pixel information of a video, while CU is computed with the
  • the local CU is defined as the pixel variance of (intra- or inter-) prediction residuals of the macroblocks
  • the global CU is defined as the average of the local CUs of several or all macroblocks in the video.
  • Inter prediction in H.264, H.263, H.261 , MPEG-4, MPEG-2, etc.
  • intra prediction in H.264, MPEG-4, etc.
  • compression i s techniques that exploit in a video predictability that results from redundancy.
  • Prediction residuals usually preserve the irregularity information, which can hardly be predicted from the temporal-spatial neighborhood. Therefore, in the invention, the variance of residuals is a suitable indicator of content unpredictability.
  • bit rate is affected by many factors (e.g. DCT coefficient distribution, motion vector, etc.).
  • estimating CU by bit rate suffers from the interference of many factors that are unrelated to CU. It is therefore an advantage of the present invention that CU is5 predicted as defined above, and not just according to the bit rate. Further,
  • CU can be fast computed from data inside the ES, without full video decoding.
  • the variance of a residual MB is approximated by the
  • Block variance is
  • Local CU is proportional to the Euclidean norm of quantized AC coefficients plus a weighted variance of DC coefficients, where the weight is to balance the different scale factor in 4x4 and 8x8 IDCT (integer discrete cosine transform).
  • the quantization is compensated according to the QP in the present invention.
  • the local CU For a residual macroblock with 4x4 transform blocks, the local CU is defined as
  • the local CU For a residual macroblock with 8x8 transform blocks, the local CU is defined as
  • QP is the QP value of the current macroblock
  • AC ⁇ is the i AC coefficient value of the ; th block inside the current macroblock
  • DCj is the DC coefficient value of the / lh block inside the current macroblock
  • constant ci is associated with QSTEP scaling and 2 Qi>f > > is to compensate quantization.
  • the constants 2 and c 3 are for balancing the scale factor in 4x4 and 8x8 transform respectively.
  • the global CU is the average local CUs among all MBs. Exemplary parameters for calculating CU
  • Fig.5c Fitting every eight compressed videos corresponding to the same original video with different rational function curves in Fig. 3 is more accurate than fitting the total of 64 compressed videos with a unique rational function curve in Fig. 5c). Moreover, the eight rational function curves are similar to each other and only differ at their amplitudes. The amplitude can be predicted by a power function of the global CU, as shown in Fig. 3 (i.e. the right side view of Fig. 2). The power function is defined by Eq. (3). Although the VQM tool 106 and the exact measurement function of Eq. (1-3) can ensure maximum accuracy, the invention may in one embodiment contain the QP-based prediction alone and provide fairly good accuracy and lower computational cost than known solutions.
  • the embodiment may comprise at least the QP parser 109 and quality prediction module 114, i.e. without the DCT coefficient parser 110 and CU calculator 111.
  • the VQM tool may in one embodiment also comprise the depacketizer 107 and the selective entropy decoder 108.
  • the selective entropy decoder 108 may handle the QPs only in l-frames of video, but ignore the QPs in P and B frames.
  • the quality prediction module 1 14 executes only Eq. (2) to determine the final video quality.
  • Fig.6 shows a diagram that illustrates the improved quality prediction, according to the invention, particularly at large QP levels.
  • the method according to the invention can predict video quality accurately for normal videos, it may output unique results for certain special input. E.g., given a compressed video with a constant QP of 45, if the QP is set to be 51 or larger, a better quality score is achieved. This effect is advantageous since higher QP means higher
  • QP-based methods often use a monotonic function with respect to QP, and a larger QP always leads to a worse quality score, i.e. QP of 51 should usually be worse than QP of 45.
  • the rational function Eq. (2) monotonically decreases with the QP when the QP is smaller than 46, but increases with the QP when the QP is larger than 46. This is shown in Fig. 6. That is, a QP of e.g. 51 is better than a QP of e.g. 45.
  • Fig.7 shows a flow chart of a video quality measurement method 706. It is a
  • the encoded video stream comprises MB prediction residuals, and comprises at least a step of calculating 714, for a video sequence VS comprising one or more frames, an overall quality score Q from a basic quality score Qb multiplied by a l o correction factor Qc,
  • the basic quality score Qb is obtained from an average quantization parameter QP over the frames of the video sequence
  • the correction factor Qc is obtained from an average pixel variance CU
  • the method further comprises in the QP parsing step 709 also a step of averaging the QPs within a MB for determining an average QP, denoted as QP.
  • the method further comprises steps of calculating, according5 to the DCT coefficients and the quantization parameter of each MB, a first value CU
  • the first value is the pixel variance of the prediction residuals.
  • the invention can be used for video quality assessment, perceptual video coding, planning on video streaming, etc.
  • the video quality prediction is applicable to videos that have been compressed by DCT plus quantization.
  • One embodiment is designed for the video compressed according to the H.264 standard.
  • the invention relates to other codecs, such as H.261 , H.263, MPEG-2, MPEG-4 etc.
  • the invention relates to a no-reference, ES-based video quality measurement tool.
  • the tool or a corresponding method, can operate in user terminals, set-top boxes, home gateways, routers, or video streaming servers, using the QP and the quantized DCT (discrete cosine transform) coefficients.
  • CU is computed according to the DCT coefficients and the QP.
  • the video quality is l o then computed from a function about the QP and CU.
  • the parameters of the function are predetermined by multiple regression analysis on the subjective database which is specially built conforming to ITU-T SG 12.

Landscapes

  • Engineering & Computer Science (AREA)
  • Health & Medical Sciences (AREA)
  • Biomedical Technology (AREA)
  • General Health & Medical Sciences (AREA)
  • Multimedia (AREA)
  • Signal Processing (AREA)
  • Compression Or Coding Systems Of Tv Signals (AREA)
  • Testing, Inspecting, Measuring Of Stereoscopic Televisions And Televisions (AREA)
  • Compression, Expansion, Code Conversion, And Decoders (AREA)

Abstract

Accuracy and efficiency of video quality measurement are major problems to be solved. According to the invention, a method (506) for accurately predicting video quality uses a rational function of the quantization parameter QP, which is corrected by a correction function that depends on content unpredictability CU. Exemplarily, the correction function is a power function of the CU. Both QP and CU can be computed (511) from the video elementary stream, without full decoding the video. This ensures high efficiency.

Description

Method and apparatus for video quality measurement
Field of the invention This invention relates to video quality measurement, in particular with the assessment of a compressed video without reference to a copy of the original uncompressed video.
Background
In IPTV (Internet protocol television), video programs have different format stages during their life circle. A video encoder compresses the video program to a bit stream, also referred to as an elementary stream (ES). The ES is further packetized into a transport stream (TS) and finally transmitted in an IP channel. Video quality can be measured using data that are obtained by accessing the transport stream, elementary stream or decoded video. Among the three types of measurement, using the transport stream is generally the fastest but the least accurate, since it has the smallest amount of video data available; using the decoded video is often accurate but the slowest, since decoding the video is computationally expensive; using the elementary stream can achieve a tradeoff between the accuracy and the computational complexity. Currently, particularly video quality measurement based on the elementary stream is being investigated.
Video compression generally employs quantization techniques. Quantization is a lossy compression technique by means of limiting the precision of signal values. It is well known that quantization is a significant factor to artifact visibility, and the quantization parameter (QP) is a powerful predictor to the video quality. Various functions of video quality with respect to QP have been provided in the literature, such as linear function [1 , 2] and exponential function [3]. However, they are insufficiently accurate for the relatively large and/or the relatively small QP level, and thus their results are not satisfactory for low-bandwidth or high-fidelity applications. The content complexity is another critical factor for video quality measurement (VQM). Visual artifacts in complex videos are more likely to be tolerated by the human eye, and thus show better quality. Therefore, content complexity in combination with QP can improve the accuracy of quality measurement, compared to using the QP alone.
Traditionally, as in [4], content complexity may be quantified as the variance, the gradient, or the edge filter response of pixel values, or their combinations. The traditional methods have at least the following disadvantages.
First, such features are not tightly correlated with human visual perception. A video with large content complexity may have not only rich texture and irregular motion, but also many edges and/or regular motion. For human eyes, visual artifacts are more likely to be tolerated in texture and irregularly (i.e., stochastical- ly) moving regions, but ordinarily more attractive and visible in edges or regularly (i.e., constantly) moving regions. Second, such features can hardly be computed until the pixels are recovered after full decoding. Thus, the traditional complexity measurement is computational expensive since it requires full decoding of the video.
Summary of the Invention
The present invention solves at least the problem of improving accuracy and/or efficiency of video quality measurement. According to the invention, the method for accurately predicting video quality uses a rational function of the QP, which is corrected by a correction function that depends on content unpredictability (CU). In various embodiments, the correction function is a power function of the CU. Both QP and CU can be computed from the video elementary stream, without fully decoding the video. Advantageously, this ensures high efficiency.
A first advantage of the invention is high accuracy of video quality prediction, which is confirmed by subjective experiments conforming to ITU-T SG 12 [5] as well as statistical verifications. The accuracy is ensured by at least two features. One is calculating an average QP and using a rational function of the average QP for predicting a basic impairment due to quantization. This function can capture the effect of quality saturation at both the relatively large and the relatively small QP levels, and thus provides a sufficiently accurate result. The other feature is using a power function of CU to correct the QP-based prediction, which improves the prediction accuracy further.
Specifically, the CU, as a video feature, can discriminate irregular changes from regular changes and from "no change" in a video signal. Consequently, the CU is more powerful to capture the influence of content features on perceived quality. The invention also provides a fast algorithm to estimate the CU from the video elementary stream, which results in a second advantage.
The second advantage is that the method requires only the elementary stream of a video instead of fully decoded video, and thus is computationally less expensive than known methods.
With the accuracy and low computational cost, the VQM can be deployed e.g. in user terminals, set-top boxes, home gateways, routers, or video streaming servers, so as to monitor the video quality and provide feedback for service planning.
The present invention, in one aspect, concerns a method for estimating video quality of an encoded video stream that comprises macroblock prediction residuals, wherein the method comprises a step of calculating, for a video sequence comprising a plurality of frames, an overall quality score Q from a basic quality score Qb multiplied by a correction factor Qc, wherein the basic quality score Qb is obtained from an average quantization parameter QP over the frames of the video sequence, and wherein the correction factor Qc is obtained from an average pixel variance within the prediction residuals of macroblocks in the video sequence. In one aspect, the invention concerns an apparatus for estimating video quality of an encoded video stream that comprises macroblock prediction residuals, the apparatus comprising a processing element for calculating, for a video sequence comprising a plurality of frames, an overall quality score Q from a basic quality 5 score Qb multiplied by a correction factor Qc, wherein the basic quality score Qb is obtained from an average quantization parameter value over the frames of the video sequence, and wherein the correction factor Qc is obtained from an average pixel variance within the prediction residuals of macroblocks in the video sequence.
10
In various embodiments, the apparatus comprises one or more of the following: a quantization parameter parser for detecting and averaging the quantization parameters of the macroblocks (MBs), wherein the quantization parameter parser provides the quantization parameter of the MBs and said average quantization i s parameter value;
a DCT coefficients parser for detecting and extracting the DCT coefficients of the MBs;
a Content Unpredictability calculator for calculating, according to the DCT coefficients of the MBs and the quantization parameter of the MBs, a first value0 CU| that represents a local Content Unpredictability and a second value CUg that represents a global Content Unpredictability, wherein the second value is calculated by averaging the first values;
a de-packetizer for de-packetizing (or de-packing) the Transport Stream, wherein an Elementary Stream (ES) is obtained, and
5 a selective entropy decoder for decoding at least portions of the ES.
In one aspect, the invention relates to a computer readable medium having executable instructions stored thereon to cause a computer to perform a method comprising a step of calculating, for a video sequence comprising a plurality of0 frames, an overall quality score Q from a basic quality score Qb multiplied by a correction factor Qc, wherein the basic quality score Qb is obtained from an average quantization parameter ~QP over the frames of the video sequence, and wherein the correction factor Qc is obtained from an average pixel variance within the prediction residuals of macroblocks in the video sequence.
Advantageous embodiments of the invention are disclosed in the dependent claims, the following description and the figures.
Brief description of the drawings
Exemplary embodiments of the invention are described with reference to the accompanying drawings, which show in
Fig.1 the structure of a video quality measurement tool;
Figs.2-4 different views of fitting a subjective quality with the measurement
function;
Fig.5 fitting basic impairment with different functions of the QP;
Fig.6 improved quality prediction according to the invention especially at a large
QP level; and
Fig.7 a flow chart of video quality measurement. Detailed description of the invention
In one embodiment of the invention, Fig.1 shows the structure of a video quality measurement (VQM) tool 106 within a video transmission system 100. Typical format stages 01 ,102,103 of a video program are as follows: a video encoder 104 compresses the video programs 101 to a bit stream, also referred to as an elementary stream (ES) (in Fig.1 exemplarily only within video encoder 104, therefore not shown). The ES is further packetized into a transport stream (TS) 103 and then transmitted in a transmission channel, e.g. an IP channel. The VQM tool 106, in principle, de-packetizes the video TS 103, thus obtaining the video ES, then parses and averages the QP and obtains a content unpredictability CU value from the obtained video ES, and finally predicts the video quality Q from the QP and the CU value. More details are given below. Advantageously, this process is fast and does not require fully decoding the video. The video quality is predicted by a function that is a multiplication of two sub- functions. The value of a first sub-function determines the basic impairment due to quantization, i.e. a basic quality score. In one embodiment, the first sub-function is a rational function of the QP. Specifically, its denominator is polynomial about the average QP of the video, while its numerator is the denominator lessened by a constant (e.g. 1). The second sub-function is a correction factor, and its value quantifies the influence of content unpredictability (CU) on the perceived quality. Preferably, the second sub-function is a power function about the CU of the video, as further specified below.
CU is a value associated with a video, and advantageously can be computed from the video ES as described below, specifically by using the quantized DCT coefficients of the video. The CU of a video reflects the intrinsic features of the content, i.e. provides a value that characterizes the content. Thus, it can be used for determining the content's impact on the perceived video quality.
In one embodiment, QPs and quantized DCT coefficients are recovered after selective entropy decoding in a selective entropy decoding module 108. Full decoding of the video, which would include complete run-length decoding, de-quantization, inverse discrete cosine transform (IDCT) and residual
compensation, is generally not required.
The video quality prediction of the present invention is of the "no-reference" NR (or non-intrusive) type. That is, it does not need to access a copy of the original uncompressed video. Further, there is no need for the quality prediction of the present invention to fully decode the video. In one embodiment, the prediction uses data that can directly be extracted from a video elementary stream.
In the following, the workflow of the video quality measurement (VQM) tool 106 shown in Fig.1 is described in detail.
The input of the VQM tool 106 may be a transport stream 103, which may be generated by a video encoder and packetizer 104 from video programs 101. The video encoder and packetizer 104 may but needs not be part of the invention. In another embodiment, a video ES 102 comprising encoded video programs 101 is directly input to the VQM tool 106. The output 115 of the VQM tool 106 is a predicted quality score Q associated with the input video.
Within the tool 106, in an embodiment capable of processing transport streams, first a depacketizer 107 parses the received transport stream 103 to obtain the video elementary stream 102. Second, video features including the average QP 112 and the global CU 113 are obtained by selective entropy decoding in a selective entropy decoder 108, parsing the selectively entropy decoded data in a QP parser 109 and a DCT coefficients parser 110, wherein the QP parser 109 provides the average QP 112, and calculating the global CU in a CU calculator 111. Finally, a quality predictor module 114 determines a quality score Q according to the video features via a predetermined measurement function.
Specifically, the video features are obtained from the output of the selective entropy decoder 108 by two simultaneously executing function blocks, or threads. In one thread, the QP parser 109 picks up (i.e. extracts) the QPs of each MB and provides the QPs to the CU calculator 111. Further, the QP parser 109 averages the QPs over a plurality of MBs and outputs the resulting average value 112. In the other thread, first the DCT coefficients parser 110 picks up (i.e. extracts) the DCT coefficients of each MB, and then the CU calculator 111 calculates the local CU according to the DCT coefficients from the DCT coefficients parser 110 and the corresponding QP from the QP parser 109. Finally, the CU calculator 111 averages the local CUs and outputs the global CU 113 obtained by averaging all local CUs.
In the following, a holistic view of the measurement function is described. The measurement function works in the quality estimator module 114 according to a mathematical model that depicts the video quality against at least two video features, including the average QP 112 and the global CU 113. In one
embodiment, the measurement function is defined as:
Figure imgf000009_0001
where Qh is the basic quality level due to quantization, Qc is an item to correct h according to the Content Unpredictability of the video, and Q is the final predicted quality.
In embodiments, the basic quality level due to quantization Qh and the correction factor Qc for correcting the basic quality level Qb according to the CU are calculated according to
Figure imgf000009_0002
^b i + ¾ X ¾P - a,)«i (2) Qc = b2 X CU^ + b3 (3) where QP is the average QP value, CUg is the global CU value, and ai,a2,a3,bi,b2 and b3 are predetermined parameters.
Preferably, the predetermined parameters are positive, rational and selected from the following ranges:
2 < ai < 6 ; a particularly advantageous value is ai=4.
10"5 < a2£ 10"4; a particularly advantageous value is a2=0,00005 (5-10"5).
30 < 33≤ 75; a particularly advantageous value is a3=49.
0,1 < b-i < 0,3; a particularly advantageous value is b-i=0,2
0,1≤ b2 < 0,3; a particularly advantageous value is b2=0,18 and
1 < b3 1 2; a particularly advantageous value is b3=1 ,65.
High accuracy of the measurement function is confirmed by experimental results as shown in Figs. 2-4, which shows for exemplary video sequences different views of fitting a subjective quality with the measurement function. Fig.2 shows a holistic view. In particular, Figs. 2-4 depict fitting the subjective quality scores against the average QP and the global CU. Each point corresponds to one of the 64 videos from a subjective database. The 64 (=8x8) videos are generated by comoressina eiaht different oriainal videos at eiaht different compression ratios. The compressed videos from different original videos are marked by different markers, as illustrated by the legend in Fig.3. Subjective quality is rated by 24 reliable subjects during a subjective experiment conforming to ITU-T SG 12 [5]. The average QP and the global CU are of the corresponding video are rated.
5
Fig. 2 shows how the predicted quality scores match with the subjective quality scores via the measurement function according to Eq. (1-3), which in this view forms a surface. Fig.3 shows a left side view, and Fig.4 a right side view of the surface (sectional drawings). The experimental result of the correlation between J O the predicted quality scores and the subjective scores confirms that the
measurement function can accurately predict the video quality.
Average QP
For H.264, as also for other compression standards, each macroblock (MB) has a 15 QP. The QP indexes a predefined table of the quantization step QSTEP, which is the linear quantizer used for quantizing the DCT coefficients inside a current MB. The average QP 112 is the mean or average QPs among all the MBs.
Basic impairment by quantization
0 A difference between the QP-based function of the invention and existing
solutions is illustrated in Fig. 5, which shows fitting (i.e. mapping) the basic impairment (i.e. left side view as in Fig.3) to different functions of the average QP. Note that Fig. 5 shows the subjective quality against the average QP alone. In Fig.5 a), fitting the basic impairment with a linear function is depicted. In Fig.5 b),5 fitting the basic impairment with an exponential function is depicted. In Fig.5 c), fitting the basic impairment with a rational function is depicted. Each compressed video is depicted as a point. Fig. 5 shows fitting the points (quality against QP) in Fig.5 a) with a linear function, as in [1 ,2], in Fig.5 b) with an exponential function, as in [3], and in Fig.5 c) with a rational function as in Eq. (2), according to the0 present invention. Regression analysis shows that the rational function is closest to the position of the points. In other words, the mapping the quality to a rational function according to the invention is a better solution than to an exponential or linear function, since it can, firstly, match the quality saturation trends at small QP levels and at large QP levels, and secondly, capture the slow saturation (large turn) at the small QP levels but also the sharp saturation (steep turn) at the large QP levels.
5 Content unpredictability
Different from the existing solutions based on content complexity, the present invention uses content unpredictability (CU) to discriminate irregular changes from regular changes or no changes in a video signal. Traditional content complexity is computed with the pixel information of a video, while CU is computed with the
10 residual information of a video. For the present invention, the local CU is defined as the pixel variance of (intra- or inter-) prediction residuals of the macroblocks, and the global CU is defined as the average of the local CUs of several or all macroblocks in the video. Inter prediction (in H.264, H.263, H.261 , MPEG-4, MPEG-2, etc.) and intra prediction (in H.264, MPEG-4, etc.) are compression i s techniques that exploit in a video predictability that results from redundancy.
Prediction residuals usually preserve the irregularity information, which can hardly be predicted from the temporal-spatial neighborhood. Therefore, in the invention, the variance of residuals is a suitable indicator of content unpredictability. 0 Known solutions, even if they are aware of the importance of CU, like [2], yet estimate CU according to the bit rate of video. However, bit rate is affected by many factors (e.g. DCT coefficient distribution, motion vector, etc.). Thus, estimating CU by bit rate suffers from the interference of many factors that are unrelated to CU. It is therefore an advantage of the present invention that CU is5 predicted as defined above, and not just according to the bit rate. Further,
different from the content complexity being computed after full decoding, CU can be fast computed from data inside the ES, without full video decoding.
In one embodiment, the variance of a residual MB is approximated by the
0 variance and the mean (average) of the blocks of the MB. Block variance is
theoretically equal to the Euclidean norm of all de-quantized AC coefficients. Local CU is proportional to the Euclidean norm of quantized AC coefficients plus a weighted variance of DC coefficients, where the weight is to balance the different scale factor in 4x4 and 8x8 IDCT (integer discrete cosine transform). To further approximate local CU, the quantization is compensated according to the QP in the present invention.
For a residual macroblock with 4x4 transform blocks, the local CU is defined as
Figure imgf000012_0001
For a residual macroblock with 8x8 transform blocks, the local CU is defined as
Figure imgf000012_0002
where QP is the QP value of the current macroblock, AC^ is the i AC coefficient value of the ; th block inside the current macroblock, DCj is the DC coefficient value of the / lh block inside the current macroblock, constant ci is associated with QSTEP scaling and 2Qi>f >> is to compensate quantization. The constants 2 and c3 are for balancing the scale factor in 4x4 and 8x8 transform respectively. Finally, the global CU is the average local CUs among all MBs. Exemplary parameters for calculating CU| are ci=3, c2=15, C3=3.
Perceived quality corrected by CU
An advantage of introducing CU in the VQM is that the quality as predicted by the QP alone is corrected, and therefore improved, which leads to higher accuracy of the VQM. This is confirmed by the experimental results shown in Fig.3 and
Fig.5c). Fitting every eight compressed videos corresponding to the same original video with different rational function curves in Fig. 3 is more accurate than fitting the total of 64 compressed videos with a unique rational function curve in Fig. 5c). Moreover, the eight rational function curves are similar to each other and only differ at their amplitudes. The amplitude can be predicted by a power function of the global CU, as shown in Fig. 3 (i.e. the right side view of Fig. 2). The power function is defined by Eq. (3). Although the VQM tool 106 and the exact measurement function of Eq. (1-3) can ensure maximum accuracy, the invention may in one embodiment contain the QP-based prediction alone and provide fairly good accuracy and lower computational cost than known solutions.
The embodiment may comprise at least the QP parser 109 and quality prediction module 114, i.e. without the DCT coefficient parser 110 and CU calculator 111. As mentioned above, the VQM tool may in one embodiment also comprise the depacketizer 107 and the selective entropy decoder 108. The selective entropy decoder 108 may handle the QPs only in l-frames of video, but ignore the QPs in P and B frames. The quality prediction module 1 14 executes only Eq. (2) to determine the final video quality.
Fig.6 shows a diagram that illustrates the improved quality prediction, according to the invention, particularly at large QP levels. Although the method according to the invention can predict video quality accurately for normal videos, it may output unique results for certain special input. E.g., given a compressed video with a constant QP of 45, if the QP is set to be 51 or larger, a better quality score is achieved. This effect is advantageous since higher QP means higher
compression, and it is achieved because of the following two reasons.
First, traditional QP-based methods often use a monotonic function with respect to QP, and a larger QP always leads to a worse quality score, i.e. QP of 51 should usually be worse than QP of 45.
Second, the rational function Eq. (2) monotonically decreases with the QP when the QP is smaller than 46, but increases with the QP when the QP is larger than 46. This is shown in Fig. 6. That is, a QP of e.g. 51 is better than a QP of e.g. 45.
Further, the following effect is achieved. Given a compressed video, when the largest AC coefficient in each transform block is increased by 1 , then a better quality is achieved. This is because, first, such modification acts like a video watermark, i.e. it almost does not change both the frequently-used features (e.g. QP, bit rate) in the elementary stream and the decoded video, except for increasing the Euclidean norm of AC coefficients and hence CU&. Second, the method of the invention can detect the quality change, since it uses the high-order moments (including variance) of AC coefficients. Third, the method of the invention will predict a better quality since cu£ (and hence Qc) increases.
5 Fig.7 shows a flow chart of a video quality measurement method 706. It is a
method for estimating video quality of an encoded video stream, wherein the encoded video stream comprises MB prediction residuals, and comprises at least a step of calculating 714, for a video sequence VS comprising one or more frames, an overall quality score Q from a basic quality score Qb multiplied by a l o correction factor Qc,
wherein the basic quality score Qb is obtained from an average quantization parameter QP over the frames of the video sequence, and wherein the correction factor Qc is obtained from an average pixel variance CU| within the prediction residuals of macroblocks in the video sequence. In one embodiment, the method
15 comprises one or more of a de-packetizing step 707, a selective entropy decoding step 708, a QP parsing step 709, a DCT coefficients parsing step 710, a step for calculating 71 1 the local CU and the global CU, and a video quality prediction step 714. 0 In one embodiment, the method further comprises in the QP parsing step 709 also a step of averaging the QPs within a MB for determining an average QP, denoted as QP.
In one embodiment, the method further comprises steps of calculating, according5 to the DCT coefficients and the quantization parameter of each MB, a first value CU| that represents a local Content Unpredictability and a second value CUg that represents a global Content Unpredictability, wherein the second value is calculated by averaging Avg the first values. The first value is the pixel variance of the prediction residuals.
0
The invention can be used for video quality assessment, perceptual video coding, planning on video streaming, etc. In principle, the video quality prediction is applicable to videos that have been compressed by DCT plus quantization. One embodiment is designed for the video compressed according to the H.264 standard. In other embodiments, the invention relates to other codecs, such as H.261 , H.263, MPEG-2, MPEG-4 etc.
5 In an embodiment, the invention relates to a no-reference, ES-based video quality measurement tool. The tool, or a corresponding method, can operate in user terminals, set-top boxes, home gateways, routers, or video streaming servers, using the QP and the quantized DCT (discrete cosine transform) coefficients. CU is computed according to the DCT coefficients and the QP. The video quality is l o then computed from a function about the QP and CU. In one embodiment, the parameters of the function are predetermined by multiple regression analysis on the subjective database which is specially built conforming to ITU-T SG 12.
While there has been shown, described, and pointed out fundamental novel 15 features of the present invention as applied to preferred embodiments thereof, it will be understood that various omissions and substitutions and changes in the apparatus and method described, in the form and details of the devices disclosed, and in their operation, may be made by those skilled in the art without departing from the spirit of the present invention. It is expressly intended that all
0 combinations of those elements that perform substantially the same function in substantially the same way to achieve the same results are within the scope of the invention. Substitutions of elements from one described embodiment to another are also fully intended and contemplated. It will be understood that the present invention has been described purely by way of example, and modifications of5 detail can be made without departing from the scope of the invention.
Each feature disclosed in the description and (where appropriate) the claims and drawings may be provided independently or in any appropriate combination.
Features may, where appropriate be implemented in hardware, software, or a0 combination of the two. Reference numerals appearing in the claims are by way of illustration only and shall have no limiting effect on the scope of the claims. Cited References
[1] A. G. Davis, "Video quality measurement." U.S. Patent application, publication No. US 2008/03171 1 A1 published Dec. 25, 2008.
5 [2] F. Yang, S. Wan, Q, Xie et. al., "No-reference quality assessment for
networked video via primary analysis of bit stream." IEEE Trans. Circuits Syst. Video Technol. vol. 20, no. 11 , pp. 1544 - 1554, Nov. 2010.
[3] M. N. Garcia, R. Schleicher, A. Raake, "Towards a content-based parametric i o video quality model for IPTV", in VPQM, 2010.
[4] K. Yamagishi, T. Kawano, and T. Hayashi, "Hybrid video-quality-estimation model for IPTV services." in GLOBECOM, 2009.
15 [5] ITU TD 469-GEN, "P.NAMS Test Plan", http://www.itu.int/md/T09-SG12- 110118-TD-GEN-0469/en, Jan. 2011

Claims

. A method for estimating video quality of an encoded video stream, the encoded video stream comprising macroblock prediction residuals, the method comprising a step of
calculating (714), for a video sequence (VS) comprising a plurality of frames, an overall quality score Q from a basic quality score Qb multiplied by a correction factor Qc,
wherein the basic quality score Qb is obtained from an average quantization parameter ~Q P of the frames of the video sequence, and wherein the correction factor Qc is obtained from an average pixel variance within the prediction residuals of macroblocks in the video sequence.
2. Method according to claim 1 , further comprising steps of performing QP
parsing (509) and averaging the QPs within a MB for determining an average QP.
3. Method according to any of claims 1-2, further comprising steps of calculating, according to the DCT coefficients and the quantization parameter of the macroblocks, a first value CU| that represents a local Content Unpredictability, the first value being a pixel variance of the prediction residuals, and a second value CUg that represents a global Content Unpredictability, wherein the second value is calculated by averaging the first values.
4. Method according to any of the claims 1-3, further comprising steps of
- determining said basic quality score Qb according to a linear function of the form Xt/( +Xb), with Xb depending on said average quantization parameter QP; and
- determining said correction factor Qc according to a rational function of the form b2 xc + b3, with b2,b3 being constants and xc being a power function of said average pixel variance within the prediction residuals.
5. Method according to claim 4, with xb = a2 (QP - a3)a1 for said basic quality score Qb, wherein ai,a2,a3 are positive rational parameters.
Method according to claim 5, wherein the positive parameters are within the following ranges:
2 < a! < 6
10-6 < a2 < 10-*
30 < a3 < 75.
Method according to claim 4, 5 or 6, with xc = (CUg)b1 for said correction factor Qc, wherein b1.b2.b3 are positive parameters.
Method according to claim 4 or 7, wherein the parameters are within the following ranges:
Figure imgf000018_0001
0,1 < b2 < 0,3 and
1 < b3 < 2.
Method according to any of claims 1-8, wherein the method operates on a Transport Stream level, further comprising steps of
- parsing and depacketizing (107) the Transport Stream (103), wherein an Elementary Stream (102) is obtained, and
- decoding at least portions of the Elementary Stream (102) by a selective entropy decoder (108) for obtaining said quantization parameters and pixel values.
10. An apparatus (106) for estimating video quality of an encoded video stream, the encoded video stream comprising macroblock prediction residuals, the apparatus comprising
a processing element (114) for calculating, for a video sequence comprising a plurality of frames, an overall quality score Q from a basic quality score Qb multiplied by a correction factor Qc,
wherein the basic quality score Qb is obtained from an average quantization parameter value (112) over the frames of the video sequence, and wherein the correction factor Qc is obtained from an average pixel variance (1 3) within the prediction residuals of macroblocks in the video sequence.
11. Apparatus according to claim 10, further comprising a quantization parameter parser (109) for detecting and averaging the quantization parameters of the macroblocks, the quantization parameter parser providing the quantization parameters of the macroblocks and said average quantization parameter value.
12. Apparatus according to claim 10 or 11 , further comprising a DCT coefficients parser (1 10) for detecting and extracting the DCT coefficients of the macroblocks.
13. Apparatus according to one of the claims 10-12, further comprising a Content Unpredictability calculator (11 1) for calculating, according to the DCT coefficients of the macroblocks and the quantization parameters of the macroblocks, a first value CUi that represents a local Content Unpredictability and a second value CUg that represents a global Content Unpredictability, wherein the second value is calculated by averaging the first values.
14. Apparatus according to one of the claims 10-13, wherein the apparatus
operates on Transport Stream level, further comprising a depacketizer (107) for de-packetizing the Transport Stream (103), wherein an Elementary Stream (102) is obtained, and a selective entropy decoder (108) for decoding at least portions of the Elementary Stream.
15. Apparatus according to one of the claims 10-14,wherein
- said basic quality score Qb is determined according to a linear function of the form xb/(1 +Xb), with xb depending on said average quantization parameter QP; and
- said correction factor Qc is determined according to a rational function of the form b2 xc + b3, with b2,b3 being constants and xc being a power function of said average pixel variance within the prediction residuals.
PCT/CN2011/002096 2011-12-15 2011-12-15 Method and apparatus for video quality measurement WO2013086654A1 (en)

Priority Applications (13)

Application Number Priority Date Filing Date Title
RU2014128824/08A RU2595917C2 (en) 2011-12-15 2011-12-15 Method and apparatus for video quality measurement
CN201180075448.4A CN103999461A (en) 2011-12-15 2011-12-15 Method and apparatus for video quality measurement
BR112014014349A BR112014014349A2 (en) 2011-12-15 2011-12-15 method and apparatus for measuring video quality
AU2011383036A AU2011383036B2 (en) 2011-12-15 2011-12-15 Method and apparatus for video quality measurement
US14/364,144 US9961340B2 (en) 2011-12-15 2011-12-15 Method and apparatus for video quality measurement
JP2014546260A JP2015505196A (en) 2011-12-15 2011-12-15 Method and apparatus for video quality measurement
KR1020147015877A KR20140102215A (en) 2011-12-15 2011-12-15 Method and apparatus for video quality measurement
MX2014007041A MX2014007041A (en) 2011-12-15 2011-12-15 Method and apparatus for video quality measurement.
EP11877571.7A EP2792144B1 (en) 2011-12-15 2011-12-15 Method and apparatus for video quality measurement
PCT/CN2011/002096 WO2013086654A1 (en) 2011-12-15 2011-12-15 Method and apparatus for video quality measurement
CA2857367A CA2857367A1 (en) 2011-12-15 2011-12-15 Method and apparatus for video quality measurement
ZA2014/03615A ZA201403615B (en) 2011-12-15 2014-05-19 Method and apparatus for video quality measurement
HK15103844.5A HK1203722A1 (en) 2011-12-15 2015-04-21 Method and apparatus for video quality measurement

Applications Claiming Priority (1)

Application Number Priority Date Filing Date Title
PCT/CN2011/002096 WO2013086654A1 (en) 2011-12-15 2011-12-15 Method and apparatus for video quality measurement

Publications (1)

Publication Number Publication Date
WO2013086654A1 true WO2013086654A1 (en) 2013-06-20

Family

ID=48611776

Family Applications (1)

Application Number Title Priority Date Filing Date
PCT/CN2011/002096 WO2013086654A1 (en) 2011-12-15 2011-12-15 Method and apparatus for video quality measurement

Country Status (13)

Country Link
US (1) US9961340B2 (en)
EP (1) EP2792144B1 (en)
JP (1) JP2015505196A (en)
KR (1) KR20140102215A (en)
CN (1) CN103999461A (en)
AU (1) AU2011383036B2 (en)
BR (1) BR112014014349A2 (en)
CA (1) CA2857367A1 (en)
HK (1) HK1203722A1 (en)
MX (1) MX2014007041A (en)
RU (1) RU2595917C2 (en)
WO (1) WO2013086654A1 (en)
ZA (1) ZA201403615B (en)

Families Citing this family (5)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
KR101825761B1 (en) 2010-03-31 2018-03-22 톰슨 라이센싱 3d disparity maps
US9600923B2 (en) 2011-05-26 2017-03-21 Thomson Licensing Scale-independent maps
CN106713901B (en) * 2015-11-18 2018-10-19 华为技术有限公司 A kind of method for evaluating video quality and device
CN109089115B (en) * 2018-05-28 2022-07-01 北方工业大学 Method for encoding 360-degree video in HEVC
US10924741B2 (en) * 2019-04-15 2021-02-16 Novatek Microelectronics Corp. Method of determining quantization parameters

Citations (3)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN1656823A (en) * 2002-05-24 2005-08-17 皇家飞利浦电子股份有限公司 A method and system for estimating objective quality of compressed video data
CN101895752A (en) * 2010-07-07 2010-11-24 清华大学 Video transmission method, system and device based on visual quality of images
WO2011043793A1 (en) * 2009-10-05 2011-04-14 Thomson Licensing Methods and apparatus for embedded quantization parameter adjustment in video encoding and decoding

Family Cites Families (17)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
GB0228556D0 (en) 2002-12-06 2003-01-15 British Telecomm Video quality measurement
US7170933B2 (en) 2002-12-13 2007-01-30 International Business Machines Corporation Method and system for objective quality assessment of image and video streams
US8311113B2 (en) * 2004-11-12 2012-11-13 Broadcom Corporation Method and system for using motion prediction to equalize video quality across intra-coded frames
WO2007066066A2 (en) 2005-12-05 2007-06-14 British Telecommunications Public Limited Company Non-intrusive video quality measurement
US20080019669A1 (en) * 2006-07-18 2008-01-24 Sahra Reza Girshick Automatically editing video data
EP2074827A4 (en) 2006-10-19 2011-06-22 Ericsson Telefon Ab L M A method of determining video quality
RU2420022C2 (en) * 2006-10-19 2011-05-27 Телефонактиеболагет Лм Эрикссон (Пабл) Method to detect video quality
EP1937002B1 (en) 2006-12-21 2017-11-01 Rohde & Schwarz GmbH & Co. KG Method and device for estimating the image quality of compressed images and/or video sequences
JP4635016B2 (en) 2007-02-16 2011-02-16 株式会社東芝 Information processing apparatus and inter prediction mode determination method
CN101682794B (en) 2007-05-11 2012-05-02 皇家飞利浦电子股份有限公司 Method, apparatus and system for processing depth-related information
JP2009260940A (en) * 2008-03-21 2009-11-05 Nippon Telegr & Teleph Corp <Ntt> Method, device, and program for objectively evaluating video quality
JP4861371B2 (en) * 2008-06-27 2012-01-25 日本電信電話株式会社 Video quality estimation apparatus, method, and program
EP2144449A1 (en) 2008-07-07 2010-01-13 BRITISH TELECOMMUNICATIONS public limited company Video quality measurement
US20100110199A1 (en) 2008-11-03 2010-05-06 Stefan Winkler Measuring Video Quality Using Partial Decoding
US20100316131A1 (en) 2009-06-12 2010-12-16 Motorola, Inc. Macroblock level no-reference objective quality estimation of video
JP5484140B2 (en) * 2010-03-17 2014-05-07 Kddi株式会社 Objective image quality evaluation device for video quality
EP2373049A1 (en) * 2010-03-31 2011-10-05 British Telecommunications Public Limited Company Video quality measurement

Patent Citations (3)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN1656823A (en) * 2002-05-24 2005-08-17 皇家飞利浦电子股份有限公司 A method and system for estimating objective quality of compressed video data
WO2011043793A1 (en) * 2009-10-05 2011-04-14 Thomson Licensing Methods and apparatus for embedded quantization parameter adjustment in video encoding and decoding
CN101895752A (en) * 2010-07-07 2010-11-24 清华大学 Video transmission method, system and device based on visual quality of images

Also Published As

Publication number Publication date
EP2792144A4 (en) 2015-08-05
AU2011383036B2 (en) 2017-03-16
JP2015505196A (en) 2015-02-16
RU2595917C2 (en) 2016-08-27
EP2792144B1 (en) 2017-02-01
AU2011383036A1 (en) 2014-06-26
EP2792144A1 (en) 2014-10-22
HK1203722A1 (en) 2015-10-30
CN103999461A (en) 2014-08-20
US9961340B2 (en) 2018-05-01
CA2857367A1 (en) 2013-06-20
US20140334555A1 (en) 2014-11-13
RU2014128824A (en) 2016-02-10
KR20140102215A (en) 2014-08-21
MX2014007041A (en) 2014-09-16
ZA201403615B (en) 2016-02-24
BR112014014349A2 (en) 2017-06-13

Similar Documents

Publication Publication Date Title
US9426475B2 (en) Scene change detection using sum of variance and estimated picture encoding cost
US10009611B2 (en) Visual quality measure for real-time video processing
AU2011381970B2 (en) Video quality measurement
US20110255589A1 (en) Methods of compressing data and methods of assessing the same
WO2010004238A1 (en) Video quality measurement
JP5133290B2 (en) Video encoding apparatus and decoding apparatus
EP2792144B1 (en) Method and apparatus for video quality measurement
JP2005527159A (en) Method and system for estimating objective quality of compressed video data
WO2012013777A2 (en) Method and apparatus for assessing the quality of a video signal during encoding or compressing of the video signal
KR20040060980A (en) Method and system for detecting intra-coded pictures and for extracting intra DCT precision and macroblock-level coding parameters from uncompressed digital video
US9591304B2 (en) Evaluation of perceptual visual quality
KR20060127159A (en) System and method for global indication of mpeg impairments in compressed digital video
WO2014032463A1 (en) Method and apparatus for estimating content complexity for video quality assessment
JP4309703B2 (en) Coding error estimation device
JP4133788B2 (en) Coding error estimation method and coding error estimation apparatus
EP4005217B1 (en) System and method to estimate blockiness in transform-based video encoding
Chen et al. Macroblock layer rate control based on structural similarity and mean absolute difference for H. 264
Wang et al. Content-related features for video quality assessment based on bit streams
Arrivukannamma et al. A study on CODEC quality metric in video compression techniques
CN116132697A (en) Image blocking effect detection method, system, equipment and storage medium
JP2007129628A (en) Encoding error measurement device and encoding error measurement program
Menkovski et al. Objective QoE Models
KR20170067301A (en) Adaptive watermarking system and method using quantization parameter

Legal Events

Date Code Title Description
121 Ep: the epo has been informed by wipo that ep was designated in this application

Ref document number: 11877571

Country of ref document: EP

Kind code of ref document: A1

ENP Entry into the national phase

Ref document number: 2857367

Country of ref document: CA

ENP Entry into the national phase

Ref document number: 2014546260

Country of ref document: JP

Kind code of ref document: A

WWE Wipo information: entry into national phase

Ref document number: 14364144

Country of ref document: US

ENP Entry into the national phase

Ref document number: 20147015877

Country of ref document: KR

Kind code of ref document: A

REEP Request for entry into the european phase

Ref document number: 2011877571

Country of ref document: EP

WWE Wipo information: entry into national phase

Ref document number: MX/A/2014/007041

Country of ref document: MX

Ref document number: 2011877571

Country of ref document: EP

WWE Wipo information: entry into national phase

Ref document number: IDP00201403453

Country of ref document: ID

NENP Non-entry into the national phase

Ref country code: DE

ENP Entry into the national phase

Ref document number: 2011383036

Country of ref document: AU

Date of ref document: 20111215

Kind code of ref document: A

ENP Entry into the national phase

Ref document number: 2014128824

Country of ref document: RU

Kind code of ref document: A

REG Reference to national code

Ref country code: BR

Ref legal event code: B01A

Ref document number: 112014014349

Country of ref document: BR

ENP Entry into the national phase

Ref document number: 112014014349

Country of ref document: BR

Kind code of ref document: A2

Effective date: 20140612