GB2440004A - Fine granularity scalability encoding using a prediction signal formed using a weighted combination of the base layer and difference data - Google Patents

Fine granularity scalability encoding using a prediction signal formed using a weighted combination of the base layer and difference data Download PDF

Info

Publication number
GB2440004A
GB2440004A GB0613675A GB0613675A GB2440004A GB 2440004 A GB2440004 A GB 2440004A GB 0613675 A GB0613675 A GB 0613675A GB 0613675 A GB0613675 A GB 0613675A GB 2440004 A GB2440004 A GB 2440004A
Authority
GB
United Kingdom
Prior art keywords
data
reference data
base layer
enhancement layer
generate
Prior art date
Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
Withdrawn
Application number
GB0613675A
Other versions
GB0613675D0 (en
Inventor
Leszek Cieplinski
Current Assignee (The listed assignees may be inaccurate. Google has not performed a legal analysis and makes no representation or warranty as to the accuracy of the list.)
Mitsubishi Electric Information Technology Corp
Mitsubishi Electric R&D Centre Europe BV Netherlands
Original Assignee
Mitsubishi Electric Information Technology Corp
Mitsubishi Electric Information Technology Center Europe BV Nederlands
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Application filed by Mitsubishi Electric Information Technology Corp, Mitsubishi Electric Information Technology Center Europe BV Nederlands filed Critical Mitsubishi Electric Information Technology Corp
Priority to GB0613675A priority Critical patent/GB2440004A/en
Publication of GB0613675D0 publication Critical patent/GB0613675D0/en
Priority to JP2009518877A priority patent/JP2009543490A/en
Priority to EP07787315A priority patent/EP2047685A2/en
Priority to US12/373,270 priority patent/US20090252229A1/en
Priority to PCT/EP2007/057040 priority patent/WO2008006829A2/en
Priority to CNA2007800260100A priority patent/CN101548549A/en
Publication of GB2440004A publication Critical patent/GB2440004A/en
Withdrawn legal-status Critical Current

Links

Classifications

    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04NPICTORIAL COMMUNICATION, e.g. TELEVISION
    • H04N19/00Methods or arrangements for coding, decoding, compressing or decompressing digital video signals
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04NPICTORIAL COMMUNICATION, e.g. TELEVISION
    • H04N19/00Methods or arrangements for coding, decoding, compressing or decompressing digital video signals
    • H04N19/30Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using hierarchical techniques, e.g. scalability
    • H04N19/34Scalability techniques involving progressive bit-plane based encoding of the enhancement layer, e.g. fine granular scalability [FGS]
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04NPICTORIAL COMMUNICATION, e.g. TELEVISION
    • H04N19/00Methods or arrangements for coding, decoding, compressing or decompressing digital video signals
    • H04N19/10Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using adaptive coding
    • H04N19/102Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using adaptive coding characterised by the element, parameter or selection affected or controlled by the adaptive coding
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04NPICTORIAL COMMUNICATION, e.g. TELEVISION
    • H04N19/00Methods or arrangements for coding, decoding, compressing or decompressing digital video signals
    • H04N19/10Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using adaptive coding
    • H04N19/102Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using adaptive coding characterised by the element, parameter or selection affected or controlled by the adaptive coding
    • H04N19/103Selection of coding mode or of prediction mode
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04NPICTORIAL COMMUNICATION, e.g. TELEVISION
    • H04N19/00Methods or arrangements for coding, decoding, compressing or decompressing digital video signals
    • H04N19/10Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using adaptive coding
    • H04N19/102Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using adaptive coding characterised by the element, parameter or selection affected or controlled by the adaptive coding
    • H04N19/117Filters, e.g. for pre-processing or post-processing
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04NPICTORIAL COMMUNICATION, e.g. TELEVISION
    • H04N19/00Methods or arrangements for coding, decoding, compressing or decompressing digital video signals
    • H04N19/10Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using adaptive coding
    • H04N19/134Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using adaptive coding characterised by the element, parameter or criterion affecting or controlling the adaptive coding
    • H04N19/136Incoming video signal characteristics or properties
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04NPICTORIAL COMMUNICATION, e.g. TELEVISION
    • H04N19/00Methods or arrangements for coding, decoding, compressing or decompressing digital video signals
    • H04N19/10Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using adaptive coding
    • H04N19/134Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using adaptive coding characterised by the element, parameter or criterion affecting or controlling the adaptive coding
    • H04N19/136Incoming video signal characteristics or properties
    • H04N19/137Motion inside a coding unit, e.g. average field, frame or block difference
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04NPICTORIAL COMMUNICATION, e.g. TELEVISION
    • H04N19/00Methods or arrangements for coding, decoding, compressing or decompressing digital video signals
    • H04N19/10Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using adaptive coding
    • H04N19/134Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using adaptive coding characterised by the element, parameter or criterion affecting or controlling the adaptive coding
    • H04N19/136Incoming video signal characteristics or properties
    • H04N19/137Motion inside a coding unit, e.g. average field, frame or block difference
    • H04N19/139Analysis of motion vectors, e.g. their magnitude, direction, variance or reliability
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04NPICTORIAL COMMUNICATION, e.g. TELEVISION
    • H04N19/00Methods or arrangements for coding, decoding, compressing or decompressing digital video signals
    • H04N19/10Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using adaptive coding
    • H04N19/134Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using adaptive coding characterised by the element, parameter or criterion affecting or controlling the adaptive coding
    • H04N19/136Incoming video signal characteristics or properties
    • H04N19/14Coding unit complexity, e.g. amount of activity or edge presence estimation
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04NPICTORIAL COMMUNICATION, e.g. TELEVISION
    • H04N19/00Methods or arrangements for coding, decoding, compressing or decompressing digital video signals
    • H04N19/10Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using adaptive coding
    • H04N19/169Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using adaptive coding characterised by the coding unit, i.e. the structural portion or semantic portion of the video signal being the object or the subject of the adaptive coding
    • H04N19/17Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using adaptive coding characterised by the coding unit, i.e. the structural portion or semantic portion of the video signal being the object or the subject of the adaptive coding the unit being an image region, e.g. an object
    • H04N19/176Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using adaptive coding characterised by the coding unit, i.e. the structural portion or semantic portion of the video signal being the object or the subject of the adaptive coding the unit being an image region, e.g. an object the region being a block, e.g. a macroblock
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04NPICTORIAL COMMUNICATION, e.g. TELEVISION
    • H04N19/00Methods or arrangements for coding, decoding, compressing or decompressing digital video signals
    • H04N19/10Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using adaptive coding
    • H04N19/169Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using adaptive coding characterised by the coding unit, i.e. the structural portion or semantic portion of the video signal being the object or the subject of the adaptive coding
    • H04N19/182Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using adaptive coding characterised by the coding unit, i.e. the structural portion or semantic portion of the video signal being the object or the subject of the adaptive coding the unit being a pixel
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04NPICTORIAL COMMUNICATION, e.g. TELEVISION
    • H04N19/00Methods or arrangements for coding, decoding, compressing or decompressing digital video signals
    • H04N19/20Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using video object coding
    • H04N19/29Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using video object coding involving scalability at the object level, e.g. video object layer [VOL]
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04NPICTORIAL COMMUNICATION, e.g. TELEVISION
    • H04N19/00Methods or arrangements for coding, decoding, compressing or decompressing digital video signals
    • H04N19/30Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using hierarchical techniques, e.g. scalability
    • H04N19/33Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using hierarchical techniques, e.g. scalability in the spatial domain
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04NPICTORIAL COMMUNICATION, e.g. TELEVISION
    • H04N7/00Television systems
    • H04N7/24Systems for the transmission of television signals using pulse code modulation
    • H04N7/26
    • H04N7/2601
    • H04N7/26021
    • H04N7/26122
    • H04N7/26127
    • H04N7/26132
    • H04N7/26138
    • H04N7/26686
    • H04N7/26691

Landscapes

  • Engineering & Computer Science (AREA)
  • Multimedia (AREA)
  • Signal Processing (AREA)
  • Compression Or Coding Systems Of Tv Signals (AREA)
  • Compression, Expansion, Code Conversion, And Decoders (AREA)

Abstract

An improved adaptive reference fine granularity scalability (AR-FGS) system for MPEG video is described where the prediction signal is formed using a weighted combination of the base layer data and the difference data, the weighting being dependent on the characteristics of the data being encoded or decoded. This is then used to provide scalable video to different applications. The characteristics can depend on the number of coefficients changing in the enhancement layer, the magnitudes of the values in the difference data between the reference block data in the enhancement layer and base layer, transformation differences between the reference block data in the enhancement and base layer or magnitudes of differences in the motion field or with the motion vectors.

Description

<p>Image Encoding and Decoding The present invention relates to the field
of image encoding and decoding, and more particularly to the</p>
<p>field of video compression encoding and decoding.</p>
<p>Scalable video coding aims to address the diversity of video communications networks and end-user interests, by compressing the original video content in such a way that efficient reconstruction at different bit-rates, frame-rates and display resolutions from the same bitstream is supported. Bit-rate scalability refers to the ability to reconstruct a compressed video over a fine gradation of bit-rates, without loss of compression efficiency. This allows a single compressed bitstream to be accessed by multiple users, each user utilizing all of his/her available bandwidth. Without rate-scalability, several versions of the same video data would have to be made available on the network, significantly increasing the storage and transmission burden. Other important forms of scalability include spatial resolution and frame-rate (temporal resolution) scalability. These allow the compressed video to be efficiently reconstructed at various display resolutions, thereby catering for the different capabilities of all sorts of end-user devices.</p>
<p>The current draft of the emerging scalable video coding standard (which will become ISO/IEC 14496-lO/AND2 and ITU-T Recommendation H.264 annex F; the current draft, Joint Draft 6 can be found in the Joint Video Team document JVT-S20l) supports a specific form of bitrate scalability called fine granularity scalability (FGS), which allows the bitstream to be cut at essentially any bitrate. This is achieved by performing the coding of transform coefficients using a form of progressive refinement. This technique orders the coefficient bits in the blocks in a nearly rate-distortion optimal fashion and introduces efficient signalling of the order that the refinements are transmitted in. This means that when some bits are dropped, the remaining bits allow for as good reconstruction of the original block as possible given the nunther of bits left. A more detailed description of the idea of fine granularity siability, as implemented in the previous MPEG standard can be found in Overview of Fine Granularity Scalability in MPEG-4 Video Standard by Weipeng Li, published in IEEE Transactions on Circuits and Systems for Video Technology, vol. 11, no. 3, March 2001.</p>
<p>The processing flow for the fine granularity scalability scheme is illustrated in Figure 1, for the case of a single enhancement layer. The coding process can be considered in two parts. The coding of the base layer follows the familiar pattern for a non-scalable coding as used in e.g. MPEG-4 AVC, where ME stands for motion estimation, MC for motion compensation, T for spatial transform, and Q for quantisation.</p>
<p>For the enhancement (FGS) layer, the difference between the original difference frame and the reconstructed base layer difference frame is transformed using the spatial transform and quantised with the quantisation step equal to half the quantisation step used in the encoding of the base layer. The quantised transform coefficients are then coded using a modified entropy coding technique called progressive refinement, which allows for the enhancement layer bitstream to be cut at an arbitrary point. As defined in the current draft of the MPEG-4 SVC standard, this truncation can be performed in a number of ways: 3 - 1. Dropping of whole progressive refinement network adaptation layer (NAL) units corresponding to complete FGS layers. This only applies when multiple FGS layers are used.</p>
<p>2. Simple truncation, where the last progressive refinement NAL unit for the highest spatio-temporal level in the bitstream is truncated by the percentage necessary to satisfy the bitrate constraint.</p>
<p>3. Quality layers, where progressive refinement NAL units are assigned quality layer identifiers which are transmitted either in the NAL units themselves or in a separate message. In this case, instead of truncating only the highest NAL unit, all the NAL units with the maximum possible quality layer identifier are truncated by the same percentage.</p>
<p>An important point in the scheme described above is that the reference for the motion compensation of both the base layer and enhancement layer is the reconstructed base layer frame. This is illustrated in Figure 2, where F(t, n) is the transform coefficient value at time t in layer n, and has important consequences, which will be discussed below.</p>
<p>The goal of adaptive-reference fine granularity scalability (AR-FGS) is to improve the performance of FGS for low-delay applications, where only P pictures are used. The problem with FGS in this application scenario is that when the FGS layers are removed from the bitstream in order to adjust the bitrate, an error is introduced due to the use of the resulting frames with degraded reconstruction quality as reference frames for motion compensation. As the motion compensation is repeated, the error accumulates in a process commonly referred to as prediction drift. As noted above, this problem is solved in the "regular" version of FGS coding in the current draft of the MPEG- 4 SVC standard by using only base layer reference frame as motion compensation reference. This solution avoids the drift problem but results in reduced compression efficiency.</p>
<p>In "Robust and Efficient Scalable Video Coding with Leaky Prediction", presented at the IEEE International Conference on Image Processing 2002, Han and Girod proposed to overcome this problem by introducing so-called leaky prediction. This is a modification of the usual motion compensation scheme, in which the prediction signal is formed as a weighted average of the base-layer reference picture and the enhancement layer reference picture. A similar technique has been adopted for the current draft of the MPEG-4 SVC standard (see the Joint Draft version 6 referred to above and also the Joint Scalable Video Model version 6 in the Joint Video Team document JVT-S202). The details of the scheme are different depending on the characteristics of the currently processed base layer block coefficients as described below.</p>
<p>When all coefficients for the current block in the base layer are 0, the processing is performed in the spatial domain as illustrated in Figure 3. First, the difference between the reference block (i.e. the blocks from the reference frame used for motion compensation of the current block) in the enhancement layer and the base layer is calculated: D(t-l, n) = R(t-l, n) -R(t-1, 0), where D(t, n) denotes the difference for frame at time t and R(t, n) is the reconstructed pixel value at time t in layer n (spatial index is omitted for clarity).</p>
<p>The resulting differential reference block is then scaled and added to the base layer reconstruction to create reference block P(t, n) P(t, n) = R (t, 0) + a*D(t_1 n) which is then used as the reference for the current block FGS layer. The weight a is a parameter controlling the amount of information from the enhancement layer reference picture that is used for prediction. In general. the reference frame does not have to correspond to a time t-l if multiple reference frames are used. It should be noted that, since this is a P-type macroblock and all the coefficients in the base layer are 0, the reconstructed base layer for the current block is exactly the same as the reconstructed reference block (at time t-l).</p>
<p>When there are non-zero coefficients in the base layer block, the enhancement layer coefficients are processed in the transform domain as illustrated in Figure 4. For the coefficients which were non-zero in the base layer, rio enhancement layer contribution from the reference frame is added. For the coefficients that have 0 values in the base layer, a similar weighted average as in the case of zero block is calculated, but this time in the transform domain. Thus, an additional step is introduced, in which the transform is performed on the reference difference block D(t-l, n), resulting in a block of transform coefficients FD(t-l, n). These coefficients are then further adjusted depending on the value of the base layer coefficients at corresponding locations in the current block FR(t, 0) . The coefficients for which the corresponding base layer current block coefficients are 0 are also set to 0, whereas the coefficients corresponding to non-zero base layer current block coefficients are scaled by a weight /3. 0 6</p>
<p>FD'(t-l, n) = 13*FD(t_1 n).</p>
<p>The resulting block of coefficients is then inverse-transformed to obtain the differential reference block D'(t-l, n), which is finally added to the base layer reconstruction to create reference block P(t, n) P(t, n) = R(t, 0) + D'(t-l, n), which is then used as the reference for the current block FGS layer.</p>
<p>The design described is based on a trade-off between compression efficiency, which is improved by utilising more information from the enhancement layer and the control of the prediction drift, which is aggravated by this process. It is argued that the impact of drift is smaller for the pixels/coefficients, for which the base layer does not change between the reference and current frame and therefore they can use enhanced reference.</p>
<p>The parameter a is quantised to 5 bits and sent in the slice header as max_dif f_ref_scale_for_zero_base_block.</p>
<p>The parameter is also quantised to 5 bits and sent in the same structure as max_diff_ref_scale_for_zero_base_coeff. The presence of both of them is controlled by the adaptive_ref_f gs_f lag.</p>
<p>Further refinements in the use of the weighting factors are defined. The context-adaptive binary arithmetic coder (CABAC) coding context is used for further classification of the all-zero blocks. If the context is non-zero, it means that some neighbouring blocks have non-zero coefficients in the enhancement layer and thus the probability of coefficients becoming non-zero in the current block is higher. Therefore, the value of a is decreased so that less enhancement layer signal 0 7 should be used to from prediction. For the case of blocks with non-zero coefficients in the base layer, the enhancement layer signal is only added when there are no more than 4 such coefficients and the value of 13 is adjusted depending on their number.</p>
<p>The present invention aims to improve the known adaptive-reference fine granularity scalability encoders and decoders, and does this in one aspect by taking advantage of further information available for adjusting the weighting of the components in the prediction.</p>
<p>According to the present invention there is provided an image sequence encoding/decoding apparatus/method in which the classification of the coefficients is improved by taking into consideration the probability of their corresponding reference block coefficients changing in the enhancement layer. This is based on the observation that the impact of dropping of the bits from the reference slice on the prediction mismatch in the areas where more coefficients change in the enhancement layer is stronger than in the areas where few or no coefficients change. While the reference block enhancement layer coefficients are not available to the decoder when the corresponding progressive refinement NAL unit has been dropped or truncated, this does not pose a problem, as the reference block adjustment is only performed when the block is available. Thus, no additional prediction mismatch is introduced by the proposed weight factor adjustment.</p>
<p>Embodiments of the invention will now be described, by way of example only, with reference to the accompanying drawings, in which: Figure 1 shows a block diagram of a quality-scalable video codec; Figure 2 shows reference block formation in non-adaptive FGS; Figure 3 shows reference block formation for an all-zero block; Figure 4 shows reference block formation for zero coefficients in a non-zero block; Figure 5 shows a block diagram of the decision process in an embodiment; and Figure 6 illustrates a macroblock with 16 motion vectors in an embodiment.</p>
<p>The embodiments set out below may comprise hardware, software or a combination of hardware and software to perform the described processes. Accordingly, an embodiment may be attained by supplying programming instructions to a programmable processing appatarus, for example, as data stored on a data storage medium (such as an optical CD RaM, semiconductor ROM, magnetic recording medium, etc) , and/or as a signal (for example an electrical or optical signal) input to the programmable processing apparatus, for example from a remote database, by transmission over a communication network such as the Internet or by transmission through the atmosphere.</p>
<p>Although the generation of a reference block in an encoder will be described below, the reference frame is generated in the same way in the decoder in embodiments of the invention.</p>
<p>First Embodiment The first embodiment of the present invention does not change the current technique for determining the values of the parameters a and f3. Instead, they are treated as the initial values for the slice, and the embodiment performs further processing to adjust the values on a block basis depending on the characteristics of the reference block. More precisely, the weighting for the enhancement reference frame for blocks for which the reference block has few (no) coefficients changing in the enhancement layer, is increased. Conversely, this weighting is decreased from its initial value for blocks with many coefficients changing in the enhancement layer.</p>
<p>A specific implementation is as follows.</p>
<p>For the case when all coefficients are 0 in the base layer, the value of a is adjusted as follows: 1. If all the coefficients in the enhancement layer of the reference block are 0, the enhancement-layer reference block is identical to the base-layer current block and the value of a is immaterial. This is useful as it allows the omission of the computation of the weighted average for these blocks, thus reducing the complexity.</p>
<p>2. The same applies to pixels for which the impact of the coefficient change on the reconstructed pixel value is 0.</p>
<p>3. The value of the reconstructed sample of the reference blocks changes in the enhancement layer.</p>
<p>In this case the same formula is used as in the current draft of MPEG-4 SVC, but the value of a is 0 10 changed proportionally to the change in the magnitude of the reconstructed sample.</p>
<p>Depending on the complexity (memory) requirements, it may not be practical to separately consider the impact of coefficient enhancements on pixels in the reference block. In that case, case 2 is not treated separately and the calculation is performed in the same way as for case 3. Similarly, it may not be practical to adjust the weight on a per-pixel basis in case 3. In that case, the weight is adjusted based on the average magnitude of the change in the reconstructed sample value.</p>
<p>For the case when not all the coefficients are 0 in the base layer block, the weight 3 is adjusted as follows: 1. Similarly as in the case above, if all the coefficients in the enhancement layer of the reference block are 0, the difference block D(t-l, n) is zero and no adjustment of the base layer reference is needed.</p>
<p>2. If the coefficient changes in the reference block enhancement layer, the weighting of the reference block is decreased proportionally to the change in the value of the corresponding coefficient of difference block FD(t-l, n).</p>
<p>3. If the coefficient does not change, the weighting of the reference block is left unchanged.</p>
<p>Similarly as in the previous case, if computational complexity needs to be constrained, the adjustment of the weight can be performed on a block basis, based on the average magnitude of the coefficient change.</p>
<p>This is expected to be less of a problem than for the previous case because the processing is adjusted on a coefficient by coefficient basis anyway.</p>
<p>In both cases, an appropriate clipping is applied to the weighting factors to ensure that they remain in the allowed range. In a particular implementation, the adjustments for both cases are made in steps of 1/16 per unit difference in the pixel or coefficient value of the appropriate enhancement layer coefficient.</p>
<p>The decision process for the implementation described above is illustrated in Figure 5.</p>
<p>In the described implementation, the adjustment is proportional to the magnitude of the change in the reference block enhancement layer. In an alternative implementation, the relationship may be nonlinear, including setting the value of the appropriate weighting factor to 0 or 1 if a predetermined threshold is reached.</p>
<p>Second Embodiment Another aspect of adaptive-reference FGS is the complexity increase caused by the introduction of the second motion compensation/prediction loop. In an alternative implementation, the design is changed so that instead of weighted prediction, only a selection between base layer and reference picture enhancement layer is used, which means that only one of the two motion compensation processes needs to be invoked for a given block. Thus, there is no need to calculate the differential reference blocks D(t-1, n) and D'(t-l, n) as described above. Instead, the reference block P(t, n) is simply a copy of either the base layer block R(t, 0) or the enhancement layer reference block R (t-1, n) depending on whether there are any non-zero coefficients in the base layer reference block. More particularly, the reference block P(t, n) is a copy of the enhancement layer reference block R (t-l, n) if all the coefficients in the base layer reference block are zero, while it is a copy of the base layer block R(t, 0) if not all of the coefficients in the base layer reference block are zero.</p>
<p>To offset the loss of precision caused by coarser quantisation of the weights, the alternative implementation uses a finer granularity of their adjustment. where the weights are changed per block rather than per slice.</p>
<p>In one implementation, the weights are adjusted based on the characteristics of signal that are known to both the encoder and decoder, which means that no explicit signalling is required on a macroblock level.</p>
<p>In an alternative implementation, the flags are sent in the bitstream. While this has a cost in the bandwidth required, it is not very significant if efficient entropy coding is employed, particularly at the higher bitrates. In addition to helping improve the coding efficiency, this variant also allows the implementation of "partial decoder refresh". That is, the effect of the prediction drift can be controlled on a macroblock by macroblock basis, thus limiting its impact on coding efficiency and particularly reducing the visual impact perceived when a whole frame is encoded with lower compression efficiency.</p>
<p>Third Embodiment In a third embodiment, the weights cx and 13 are adapted</p>
<p>based on the properties of the motion field in the</p>
<p>vicinity of the currently processed block.</p>
<p>Specifically, the motion vectors of the surrounding blocks are compared to the motion vector of the current block and the weights are adjusted based on a measure of difference between them. In one implementation, the measure used is based on the magnitude of the difference of the current block motion vector and the surrounding motion vectors. This magnitude can be *calculated as the average squared difference of the current motion vectors and the surrounding motion vectors, i.e.: M = -+ ( - } where N is the number of surrounding blocks taken into consideration, vC' is the motion vector of the current block, v1 is the motion vector of the i-th surrounding block, and x and y denote the components of the motion vectors. Other measures of difference can also be used.</p>
<p>One example is a similar formula as above, with the exception that square root of the magnitude is used in the summation, i.e. M = -v + -v)2} 0 The amount of adjustment can be specified to be proportional to the value of the difference measure used. Alternatively non-linear dependency can be specified, including specification of a look-up table for the values of adjustment depending on the value of the difference measure. A specific example of a look-up</p>
<p>table is as follows:</p>
<p>* Decrease weight by 6/32 if M>64 * Otherwise, decrease weight by 4/32 if M>32 * Otherwise, decrease weight by 3/32 if M>16 * Otherwise, decrease weight by 2/32 if M>8 * Otherwise, decrease weight by 1/32 if M>O. * 14</p>
<p>When no motion vectors are transmitted (so-called SKIP macroblock, where no motion vectors or transform coefficients are sent and the macroblock from the previous frame is simply copied over), it is impossible to calculate the measure of difference as defined above. Since this indicates that there is little change between the previous and current frame at the current macroblock position, the weight can be either left unchanged or increased in this case.</p>
<p>The selection of the set of the surrounding blocks to be used in the calculation of the measure of difference depends on complexity considerations, the position of the current macroblock in the frame and the position of the current block in the current macroblock.</p>
<p>In order to limit the complexity, the first implementation only uses the blocks within the current macroblock for the calculation of the measure of difference. The maximum number of motion vectors in a macroblock is 16 (one for each of the 4x4 blocks) Depending on the position of the block within the macroblock, the number of surrounding blocks then varies between 3 and 8 (top, left, right, bottom, top-left, top-right, bottom-left and bottom-right), depending on the position of the block in the macroblock. This is illustrated in Figure 6, where the blocks labelled A, D, M and P have 3 available surrounding blocks, B, C, E, H, I, L, N and 0 have 5 available surrounding blocks, and F, G, J and K have 8 available surrounding blocks each. The cases where fewer than 16 motion vectors are used can be treated similarly, by copying treating the motion vectors corresponding to larger (e.g. 8x8) blocks as if they were a corresponding number (e.g. 4) motion vectors for the corresponding 4x4 blocks.</p>
<p>In a second implementation, the information from all or some of the previously processed macroblocks (top-left, top, top-right and left to the current macroblock) is also used in the calculation of the measure of difference. This means that additional blocks become available for the calculation of the measure of difference for the blocks A, B, C, D, E, I and M in Figure 6. When all the previously processed macroblocks are used, blocks A, B, C, E and I have 8 available surrounding blocks, block D has 6 available surrounding blocks nd block M has 5 available surrounding blocks.</p>
<p>In an alternative implementation, the measure of difference is adjusted to take into consideration the distance between the current block and each of the surrounding blocks. For example, using the set of blocks shown in Figure 6, and taking block F as the current block, the blocks B, E, G and J are used with more weight than blocks A, C, I and K. A simplified implementation calculates the measure of distance only once for the whole macroblock and then uses this value for all the blocks within the macroblock.</p>
<p>Modifications Many modifications can be made to the embodiments described above within the scope of the accompanying claims.</p>
<p>For example, in an alternative implementation, the value of the adjustment of the weighting parameters is based on a combination of factors, one of which is the change of the values of the reference block enhancement layer coefficients described above. These other factors can be, for example, the macroblock type and prediction modes or the arithmetic coder coding context as</p>
<p>described in the prior art. More specifically, the</p>
<p>adjustments can simply be added together to form the total adjustment, where the adjustments from different factors may have different granularity.</p>
<p>In another alternative implementation, the strength of the modification of the weighting factor depends on its initial value. More precisely, if the initial value of the weighting factor is small, the changes are also small and as the initial value increases the changes are allowed to be made larger. This is done in order to allow for better control of the drift. When the drift is a problem (e.g. long distance between intra pictures), it is more important to be able to control its strength with higher precision. More precisely, such a scheme is implemented by scaling the change by the initial value of the weight. That is, instead of changing the weights a and /3 by fixed amount (e.g. 1/16), they are changed by an amount proportional to the initial value obtained from the bitstream elements max_di f f_ref_scale_far_zero_base_block and max_diff_ref_scale_for_zero_base_coeff, e.g. a = (1 + where a0 is the initial value of the weight and y is the strength of adjustment based on the change of magnitude as described above. If adjustments from multiple factors are combined, the formula can be expanded to e.g. a = (1 + y) (1 + ô)ao, where y and ö are the contributions corresponding to the different factors, e.g. enhancement layer coefficient magnitudes, arithmetic coder coding context, or macroblock type and prediction mode. I 17</p>
<p>This is just one example of non-uniform quantisation of the weight values, more general schemes could also be considered. For example, the quantisation step used in the quantisation of the block transform coefficients has a significant impact on the drift properties. In an alternative implementation, the calculation of adjustment of the weights takes this into account.</p>
<p>When only a part of the FGS layer is received by the decoder, the quality of the decoding can be improved if the amount of the data available as proportion of the total amount of data in the FGS layer is known or estimated. In that case, the amount of adjustment of the weight values can be additionally adjusted to correspond more closely to the optimal weighting.</p>

Claims (1)

  1. <p>CLIAIMS</p>
    <p>1. An improved adaptive reference fine granularity scalability encoding or decoding method in which a prediction signal is formed using a weighted combination of base layer data and difference data, with the amount of weighting being dependent upon characteristics of the data.</p>
    <p>I</p>
    <p>2. A method according to Claim 1, wherein the weighting is dependent upon the number of coefficients changing in the enhancement layer.</p>
    <p>3. A method according to Claim 1 or Claim 2, wherein the weighting is dependent upon the magnitudes of the values in the difference data.</p>
    <p>4. A method according to any preceding claim, wherein the difference data comprises a difference between reference block data in the enhancement layer and the base layer.</p>
    <p>5. A method according to any of Claims 1 to 3, wherein the difference data comprises transformation coefficients defining transformation differences between reference block data in the enhancement layer and the base layer.</p>
    <p>6. A method of generating enhancement layer prediction data in an adaptive reference fine granularity scalability encoder or decoder, the method comprising: generating reference data by combining data in accordance with scaling parameters such that the scaling parameters are set in dependence upon the probability of reference data values changing in the enhancement layer; and O 19 generating the prediction data in dependence upon the generated reference data.</p>
    <p>7. An adaptive reference fine granularity scalability encoding or decoding method, comprising: calculating a weighting factor for use in combining difference data with base layer data; adjusting the weighting factor in dependence upon the magnitude of the difference data; combining the difference data and the base layer data in accordance with the adjusted weighting factor to generate reference data; and generating enhancement layer prediction data using the generated reference data.</p>
    <p>8. A method according to Claim 7, wherein the weighting factor is adjusted in linear dependence upon the magnitude of the difference data.</p>
    <p>9. A method according to Claim 7 or Claim 8, wherein the weighting factor is adjusted by an amount dependent upon its non-adjusted value and the magnitude of the difference data.</p>
    <p>10. A method according to any of Claims 7 to 9, wherein the weighting factor is reduced by an amount which increases as the magnitude of the difference data increases.</p>
    <p>11. A method of encoding or decoding a sequence of images in accordance with base layer data and enhancement layer data, wherein: the enhancement layer data is generated by comparing reference data with base layer data to generate prediction error data; and the reference data is generated by: calculating a difference between reference data in the enhancement layer and reference data in the base layer to generate differential reference data; determining the magnitude of the values in the differential reference data scaling the differential reference data in dependence upon the determined magnitude values to generate scaled differential reference data; and combining the scaled differential reference data with base layer reconstructed data.</p>
    <p>12. A method according to Claim 11, wherein the process of scaling the differential reference data comprises scaling each value by a respective scaling factor set in dependence upon the magnitude of the value in the differential reference data.</p>
    <p>13. A method according to Claim 11, wherein the process of scaling the differential reference data comprises scaling the values by a scaling factor set in dependence upon an average of the magnitudes of the values in the differential reference data.</p>
    <p>14. A method of encoding or decoding a sequence of images in accordance with base layer data and enhancement layer data, wherein: the enhancement layer data is generated by comparing reference data with base layer data to generate prediction error data; and the reference data is generated by: determining whether all of the values of reference data in the enhancement layer are 0; in a case where the values are not all 0: (i) calculating a difference between the reference data in the enhancement layer and reference data in the base layer to generate differential reference data; (ii) scaling the differential reference data; and (iii) combining the scaled differential reference data with base layer reconstructed data; in a case where all of the values are 0: omitting processes (i), (ii) and (iii) and using reference data from the base layer.</p>
    <p>15. A method of encoding or decoding a sequence of images in accordance with base layer data and enhancement layer data, wherein: the enhancement layer data is generated by comparing reference data with base layer data to generate prediction error data; and the reference data is generated by: processing values of reference data in the enhancement layer to identify non-zero values and zero values; for each non-zero value: (1) calculating a difference between the reference data value in the enhancement layer and the corresponding value in reference data in the base layer to generate a difference value; (ii) scaling the difference value; and (iii) combining the scaled difference value with a corresponding value in base layer reconstructed data; for each zero value: omitting process (i), (ii) and (iii) and using reference data from the base layer.</p>
    <p>16. A method of encoding or decoding a sequence of images in accordance with base layer data and enhancement layer data, wherein: the enhancement layer data is generated by comparing reference data with base layer data to generate prediction error data; and the reference data is generated by: calculating a difference between reference data in the enhancement layer and reference data in the base layer to generate differential reference data; transforming the differential reference data to generate transform coefficients; determining the magnitude of the transform coefficients; scaling the transform coefficients in dependence upon the determined magnitude values to generate scaled transform coefficients; inverse-transforming the scaled transform coefficients to obtain modified differential reference data; and combining the modified differential reference data with base layer reconstructed data.</p>
    <p>17. A method according to Claim 16, wherein the process of scaling the transform coefficients comprises scaling each transform coefficient by a respective scaling factor set in dependence upon the magnitude of the transform coefficient.</p>
    <p>18. A method according to Claim 16, wherein the process of scaling the transform coefficients comprises scaling the transform coefficients by a scaling factor set in dependence upon an average of the magnitudes of the transform coefficients.</p>
    <p>19. A method of encoding or decoding a sequence of images in accordance with base layer data and enhancement layer data, wherein: the enhancement layer data is generated by comparing reference data with base layer data to generate prediction error data; and the reference data is generated by: determining whether all of the values of reference data in the enhancement layer are 0; in a case where the values are not all zero: Ci) calculating a difference between the reference data in the enhancement layer and reference data in the base layer to generate differential reference data; (ii) transforming the differential reference data to generate transform coefficients; (iii) scaling the transform coefficients; (iv) inverse-transforming the scaled transform coefficients to obtain modified differential reference data; and (v) combining the modified differential reference data with base layer reconstructed data; in a case where all of the values are 0: omitting processes (i) , (ii) , (iii) , (iv) and (v) and using reference data from the base layer.</p>
    <p>20. A method of generating enhancement layer prediction data in an adaptive reference fine granularity scalability encoder or decoder, the method comprising: determining whether there are any non-zero coefficients in the base layer reference block; in a case where there is at least one non-zero coefficient, using a base layer block as a reference block to generate the enhancement layer prediction data; in a case where there are no non-zero coefficients, using an enhancement layer reference block to generate the enhancement layer prediction data.</p>
    <p>21. An improved adaptive reference fine granularity scalability encoding or decoding method in which a prediction signal is formed using a weighted combination of base layer data and difference data, with the amount of weighting being dependent upon properties of the motion field in the vicinity of the currently processed block.</p>
    <p>22. A method according to Claim 21, wherein the weighting is dependent upon the magnitudes of</p>
    <p>differences in the motion field.</p>
    <p>23. A method according to Claim 21 or Claim 22, wherein the difference data comprises a difference between reference block data in the enhancement layer and the base layer.</p>
    <p>24. A method according to Claim 21 or Claim 22, wherein the difference data comprises transformation coefficients defining transformation differences between reference block data in the enhancement layer and the base layer.</p>
    <p>25. A method of generating enhancement layer prediction data in an adaptive reference fine granularity scalability encoder or decoder, the method comprising: generating reference data by combining data in accordance with scaling parameters such that the scaling parameters are set in dependence upon differences between motion vectors of a currently processed block and motion vectors of at least one surrounding block; and generating the prediction data in dependence upon the generated reference data.</p>
    <p>26. An adaptive reference fine granularity scalability encoding or decoding method, comprising processing a block of data by: calculating a weighting factor for use in combining difference data with base layer data; comparing motion vectors of the block with motion vectors of a plurality of surrounding blocks to determine differences therebetween; adjusting the weighting factor in dependence upon the differences between the motion vectors; combining the difference data and the base layer data in accordance with the adjusted weighting factor to generate reference data; and generating enhancement layer prediction data using the generated reference data.</p>
    <p>27. A method according to Claim 26, wherein the weighting factor is adjusted in linear dependence upon the differences between the motion vectors.</p>
    <p>28. A method of encoding or decoding a sequence of images in accordance with base layer data and enhancement layer data, wherein: the enhancement layer data is generated by comparing reference data with base layer data to generate prediction error data; and the reference data is generated by: calculating a difference between reference data in the enhancement layer and reference data in the base layer to generate differential reference data; comparing motion vectors of a current block with motion vectors of a plurality of surrounding blocks to calculate a measure of the motion vector differences; scaling the differential reference data in dependence upon the calculated motion vector difference measure to generate scaled differential reference data; and combining the scaled differential reference data with base layer reconstructed data.</p>
    <p>29. A method of encoding or decoding a sequence of images in accordance with base layer data and enhancement layer data, wherein: the enhancement layer data is generated by comparing reference data with base layer data to generate prediction error data; and the reference data is generated by: calculating a difference between reference data in the enhancement layer and reference data in the base layer to generate differential reference data; transforming the differential reference data to generate transform coefficients; S 26 comparing motion vectors of a current block with motion vectors of a plurality of surrounding blocks to calculate a measure of the motion vector differences; scaling the transform coefficients in dependence upon the calculated motion vector difference measure to generate scaled transform coefficients; inverse-transforming the scaled transform coefficients to obtain modified differential reference data; and combining the modified differential reference data with base layer reconstructed data.</p>
    <p>30. A method according to Claim 28 or Claim 29, wherein the process of comparing motion vectors comprises comparing motion vectors of the currently processed block with motion vectors of surrounding blocks in the same macroblock.</p>
    <p>31. A method according to Claim 30, wherein the process of comparing motion vectors comprises comparing motion vectors of the currently processed block with motion vectors of surrounding blocks in the same macroblock and also motion vectors of at least one previously processed macroblock.</p>
    <p>32. A method according to any of Claims 28 to 31, further comprising weighting the differences between the motion vectors of the currently processed block and each surrounding block in dependence upon a measure of the distance between the currently processed block and the surrounding block.</p>
    <p>33. A method according to Claim 32, wherein the differences between the motion vectors are weighted such that motion vector differences between blocks with a smaller distance therebetween contribute more to the motion vector difference measure than differences between the motion vectors of blocks with a larger distance therebetween.</p>
    <p>34. A storage medium storing computer program instructions to program a programmable processing apparatus to become operable to perform a method as set out in at least one of Claims 1 to 33.</p>
    <p>35. A signal carrying computer program instructions to program a programmable processing apparatus to become operable to perform a method as set out in at least one of Claims 1 to 33.</p>
    <p>36. An improved adaptive reference fine granularity scalability encoder or decoder, comprising means for generating a prediction signal using a weighted combination of base layer data and difference data, with the amount of weighting being dependent upon characteristics of the data.</p>
    <p>37. An adaptive reference fine granularity scalability encoder or decoder, having means for generating enhancement layer prediction data comprising: means for generating reference data by combining data in accordance with scaling parameters such that the scaling parameters are set in dependence upon the probability of reference data values changing in the enhancement layer; and means for generating the prediction data in dependence upon the generated reference data.</p>
    <p>38. An adaptive reference fine granularity scalability encoder or decoder, comprising: means for calculating a weighting factor for use in combining difference data with base layer data; means for adjusting the weighting factor in dependence upon the magnitude of the difference data; means for combining the difference data and the base layer data in accordance with the adjusted weighting factor to generate reference data; and means for generating enhancement layer prediction data using the generated reference data.</p>
    <p>39. An encoder or decoder for encoding or decoding a sequence of images in accordance with base layer data and enhancement layer data, comprising: reference data generating means, comprising: means for calculating a difference between reference data in the enhancement layer and reference data in the base layer to generate differential reference data; means for determining the magnitude of the values in the differential reference data; means for scaling the differential reference data in dependence upon the determined magnitude values to generate scaled differential reference data; and means for combining the scaled differential reference data with base layer reconstructed data; and means for generating enhancement layer data by comparing reference data generated by the reference data generating means with base layer data to generate prediction error data.</p>
    <p>40. An encoder or decoder for encoding or decoding a sequence of images in accordance with base layer data and enhancement layer data, comprising: reference data generating means operable to generate reference data by: determining whether all of the values of reference data in the enhancement layer are 0; in a case where the values are not all 0: (i) calculating a difference between the reference data in the enhancement layer and reference data in the base layer to generate differential reference data; (ii) scaling the differential reference data; and (iii) combining the scaled differential reference data with base layer reconstructed data; in a case where all of the values are 0: omitting processes (i), (ii) and (iii) and using reference data from the base layer; and means for generating enhancement layer data by comparing reference data generated by the reference data generating means with base layer data to generate prediction error data.</p>
    <p>41. An encoder or decoder for encoding or decoding a sequence of images in accordance with base layer data and enhancement layer data, comprising: reference data generating means operable to generate reference data by: processing values of reference data in the enhancement layer to identify non-zero values and zero values; for each non-zero value: (i) calculating a difference between the reference data value in the enhancement layer and the corresponding value in reference data in the base layer to generate a difference value; (ii) scaling the difference value; and (iii) combining the scaled difference value with a corresponding value in base layer reconstructed data; for each zero value: omitting process (1), (ii) and (iii) and using reference data from the base layer; and means for generating enhancement layer data by comparing reference data generated by the reference data generating means with base layer data to generate prediction error data. S 30</p>
    <p>42. An encoder or decoder for encoding or decoding a sequence of images in accordance with base layer data and enhancement layer data, comprising: reference data generating means comprising: means for calculating a difference between reference data in the enhancement layer and reference data in the base layer to generate differential reference data; means for transforming the differential reference data to generate transform coefficients; means for determining the magnitude of the transform coefficients; means for scaling the transform coefficients in dependence upon the determined magnitude values to generate scaled transform coefficients; means for inverse-transforming the scaled transform coefficients to obtain modified differential reference data; and means for combining the modified differential reference data with base layer reconstructed data; and means for generating enhancement layer data by comparing reference data generated by the reference data generating means with base layer data to generate prediction error data.</p>
    <p>43. An encoder or decoder for encoding or decoding a sequence of images in accordance with base layer data and enhancement layer data, comprising: reference data generating means operable to generate reference data by: determining whether all of the values of reference data in the enhancement layer are 0; in a case where the values are not all zero: (1) calculating a difference between the reference data in the enhancement layer and reference data in the base layer to generate differential reference data; S 31 (ii) transforming the differential reference data to generate transform coefficients; (iii) scaling the transform coefficients; (iv) inverse-transforming the scaled transform coefficients to obtain modified differential reference data; and (v) combining the modified differential reference data with base layer reconstructed data; in a case where all of the values are 0: omitting processes (i), (ii), (iii), (iv) and (v) and using reference data from the base layer; and means for generating enhancement layer data by comparing reference data generated by the reference data generating means with base layer data to generate prediction error data.</p>
    <p>44. An adaptive reference fine granularity scalability encoder or decoder, having enhancement layer prediction data generating means operable to generate enhancement layer prediction data by: determining whether there are any non-zero coefficients in the base layer reference block; in a case where there is at least one non-zero coefficient, using a base layer block as a reference block to generate the enhancement layer prediction data; in a case where there are no non-zero coefficients, using an enhancement layer reference block to generate the enhancement layer prediction data.</p>
    <p>45. An improved adaptive reference fine granularity scalability encoder or decoder, comprising means for generating a prediction signal using a weighted combination of base layer data and difference data, with the amount of weighting being dependent upon properties of the motion field in the vicinity of the currently processed block. 32 -</p>
    <p>46. An adaptive reference fine granularity scalability encoder or decoder, having means for generating enhancement layer prediction data comprising: means for generating reference data by combining data in accordance with scaling parameters such that the scaling parameters are set in dependence upon differences between motion vectors of a currently processed block and motion vectors of at least one surrounding block; and means for generating the prediction data in dependence upon the generated reference data.</p>
    <p>47. An adaptive reference fine granularity scalability encoder or decoder, comprising: means for calculating a weighting factor for use in combining difference data with base layer data; means for comparing motion vectors of a block with motion vectors of a plurality of surrounding blocks to determine differences therebetween; means for adjusting the weighting factor in dependence upon the differences between the motion vectors; means for combining the difference data and the base layer data in accordance with the adjusted weighting factor to generate reference data; and means for generating enhancement layer prediction data using the generated reference data.</p>
    <p>48. An encoder or decoder for encoding or decoding a sequence of images in accordance with base layer data and enhancement layer data, comprising: reference data generating means, comprising: means for calculating a difference between reference data in the enhancement layer and reference data in the base layer to generate differential reference data; means for comparing motion vectors of a current block with motion vectors of a plurality of surrounding blocks to calculate a measure of the motion vector differences; means for scaling the differential reference data in dependence upon the calculated motion vector difference measure to generate scaled differential reference data; and means for combining the scaled differential reference data with base layer reconstructed data; and means for generating enhancement layer data by comparing reference data generated by the reference data generating means with base layer data to generate prediction error data.</p>
    <p>49. An encoder or decoder for encoding or decoding a sequence of images in accordance with base layer data and enhancement layer data, comprising: reterence data generating means, comprising: means for calculating a difference between reference data in the enhancement layer and reference data in the base layer to generate differential reference data; means for transforming the differential reference data to generate transform coefficients; means for comparing motion vectors of a current block with motion vectors of a plurality of surrounding blocks to calculate a measure of the motion vector differences; means for scaling the transform coefficients in dependence upon the calculated motion vector difference measure to generate scaled transform coefficients; means for inverse-transforming the scaled transform coefficients to obtain modified differential reference data; and means for combining the modified differential reference data with base layer reconstructed data; and means for generating enhancement layer data by comparing reference data generated by the reference data generating means with base layer data to generate prediction error data.</p>
GB0613675A 2006-07-10 2006-07-10 Fine granularity scalability encoding using a prediction signal formed using a weighted combination of the base layer and difference data Withdrawn GB2440004A (en)

Priority Applications (6)

Application Number Priority Date Filing Date Title
GB0613675A GB2440004A (en) 2006-07-10 2006-07-10 Fine granularity scalability encoding using a prediction signal formed using a weighted combination of the base layer and difference data
JP2009518877A JP2009543490A (en) 2006-07-10 2007-07-10 Image encoding and decoding
EP07787315A EP2047685A2 (en) 2006-07-10 2007-07-10 Fine granular scalable image encoding and decoding
US12/373,270 US20090252229A1 (en) 2006-07-10 2007-07-10 Image encoding and decoding
PCT/EP2007/057040 WO2008006829A2 (en) 2006-07-10 2007-07-10 Fine granular scalable image encoding and decoding
CNA2007800260100A CN101548549A (en) 2006-07-10 2007-07-10 Image encoding and decoding

Applications Claiming Priority (1)

Application Number Priority Date Filing Date Title
GB0613675A GB2440004A (en) 2006-07-10 2006-07-10 Fine granularity scalability encoding using a prediction signal formed using a weighted combination of the base layer and difference data

Publications (2)

Publication Number Publication Date
GB0613675D0 GB0613675D0 (en) 2006-08-16
GB2440004A true GB2440004A (en) 2008-01-16

Family

ID=36926762

Family Applications (1)

Application Number Title Priority Date Filing Date
GB0613675A Withdrawn GB2440004A (en) 2006-07-10 2006-07-10 Fine granularity scalability encoding using a prediction signal formed using a weighted combination of the base layer and difference data

Country Status (6)

Country Link
US (1) US20090252229A1 (en)
EP (1) EP2047685A2 (en)
JP (1) JP2009543490A (en)
CN (1) CN101548549A (en)
GB (1) GB2440004A (en)
WO (1) WO2008006829A2 (en)

Cited By (1)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
GB2492396A (en) * 2011-06-30 2013-01-02 Canon Kk Decoding a Scalable Video Bit-Stream

Families Citing this family (16)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US20080013624A1 (en) * 2006-07-14 2008-01-17 Samsung Electronics Co., Ltd. Method and apparatus for encoding and decoding video signal of fgs layer by reordering transform coefficients
US20080013623A1 (en) * 2006-07-17 2008-01-17 Nokia Corporation Scalable video coding and decoding
EP2154893A1 (en) * 2008-08-13 2010-02-17 Thomson Licensing Method for modifying a reference block of a reference image, method for encoding or decoding a block of an image by help of a reference block and device therefor and storage medium or signal carrying a block encoded by help of a modified reference block
CN102150432A (en) * 2008-09-17 2011-08-10 夏普株式会社 Scalable video stream decoding apparatus and scalable video stream generating apparatus
JP5446198B2 (en) * 2008-10-03 2014-03-19 富士通株式会社 Image prediction apparatus and method, image encoding apparatus, and image decoding apparatus
GB2486692B (en) * 2010-12-22 2014-04-16 Canon Kk Method for encoding a video sequence and associated encoding device
US9392274B2 (en) * 2012-03-22 2016-07-12 Qualcomm Incorporated Inter layer texture prediction for video coding
WO2013147495A1 (en) * 2012-03-26 2013-10-03 엘지전자 주식회사 Scalable video encoding/decoding method and apparatus
US9854259B2 (en) 2012-07-09 2017-12-26 Qualcomm Incorporated Smoothing of difference reference picture
US20150237372A1 (en) * 2012-10-08 2015-08-20 Samsung Electronics Co., Ltd. Method and apparatus for coding multi-layer video and method and apparatus for decoding multi-layer video
US10097825B2 (en) * 2012-11-21 2018-10-09 Qualcomm Incorporated Restricting inter-layer prediction based on a maximum number of motion-compensated layers in high efficiency video coding (HEVC) extensions
CN104104957B (en) * 2013-04-08 2018-03-16 华为技术有限公司 Coding/decoding method, coding method, decoding apparatus and code device
FR3029055B1 (en) * 2014-11-24 2017-01-13 Ateme IMAGE ENCODING METHOD AND EQUIPMENT FOR IMPLEMENTING THE METHOD
KR102379182B1 (en) * 2015-11-20 2022-03-24 삼성전자주식회사 Apparatus and method for continuous data compression
KR102331933B1 (en) 2016-05-12 2021-11-29 엘지전자 주식회사 Method and apparatus for processing a video signal using coefficient derived reconstruction
US20220279185A1 (en) * 2021-02-26 2022-09-01 Lemon Inc. Methods of coding images/videos with alpha channels

Citations (8)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
WO2001030085A1 (en) * 1999-10-15 2001-04-26 Koninklijke Philips Electronics N.V. System and method for encoding and decoding the residual signal for fine granular scalable video
US20030156638A1 (en) * 2002-02-21 2003-08-21 Koninklijke Philips Electronics N.V. Minimizing drift in motion-compensation fine granular scalable structures
US20050201462A1 (en) * 2004-03-09 2005-09-15 Nokia Corporation Method and device for motion estimation in scalable video editing
WO2007018669A1 (en) * 2005-07-21 2007-02-15 Thomson Licensing Method and apparatus for weighted prediction for scalable video coding
WO2007047271A2 (en) * 2005-10-12 2007-04-26 Thomson Licensing Methods and apparatus for weighted prediction in scalable video encoding and decoding
US20070147494A1 (en) * 2005-12-22 2007-06-28 Victor Company Of Japan, Ltd. Video-signal layered coding and decoding methods, apparatuses, and programs
US20070177664A1 (en) * 2006-01-19 2007-08-02 Samsung Electronics Co., Ltd. Entropy encoding/decoding method and apparatus
US20070201551A1 (en) * 2006-01-09 2007-08-30 Nokia Corporation System and apparatus for low-complexity fine granularity scalable video coding with motion compensation

Patent Citations (8)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
WO2001030085A1 (en) * 1999-10-15 2001-04-26 Koninklijke Philips Electronics N.V. System and method for encoding and decoding the residual signal for fine granular scalable video
US20030156638A1 (en) * 2002-02-21 2003-08-21 Koninklijke Philips Electronics N.V. Minimizing drift in motion-compensation fine granular scalable structures
US20050201462A1 (en) * 2004-03-09 2005-09-15 Nokia Corporation Method and device for motion estimation in scalable video editing
WO2007018669A1 (en) * 2005-07-21 2007-02-15 Thomson Licensing Method and apparatus for weighted prediction for scalable video coding
WO2007047271A2 (en) * 2005-10-12 2007-04-26 Thomson Licensing Methods and apparatus for weighted prediction in scalable video encoding and decoding
US20070147494A1 (en) * 2005-12-22 2007-06-28 Victor Company Of Japan, Ltd. Video-signal layered coding and decoding methods, apparatuses, and programs
US20070201551A1 (en) * 2006-01-09 2007-08-30 Nokia Corporation System and apparatus for low-complexity fine granularity scalable video coding with motion compensation
US20070177664A1 (en) * 2006-01-19 2007-08-02 Samsung Electronics Co., Ltd. Entropy encoding/decoding method and apparatus

Cited By (1)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
GB2492396A (en) * 2011-06-30 2013-01-02 Canon Kk Decoding a Scalable Video Bit-Stream

Also Published As

Publication number Publication date
WO2008006829A3 (en) 2009-03-05
GB0613675D0 (en) 2006-08-16
JP2009543490A (en) 2009-12-03
EP2047685A2 (en) 2009-04-15
WO2008006829A2 (en) 2008-01-17
US20090252229A1 (en) 2009-10-08
CN101548549A (en) 2009-09-30

Similar Documents

Publication Publication Date Title
GB2440004A (en) Fine granularity scalability encoding using a prediction signal formed using a weighted combination of the base layer and difference data
KR102438062B1 (en) Multidimensional Quantization Techniques for Video Coding/Decoding Systems
US20060120450A1 (en) Method and apparatus for multi-layered video encoding and decoding
US20150326896A1 (en) Techniques for hdr/wcr video coding
EP1680925A1 (en) Foveated video coding and transcoding system and method for mono or stereoscopic images
EP1921863A2 (en) An encoder
MXPA06002367A (en) Process and arrangement for encoding video pictures.
US20090074082A1 (en) System And Method Of Error Control For Video Coding
Wiegand et al. Video coding: Part II of fundamentals of source and video coding
EP2051525A1 (en) Bandwidth and content dependent transmission of scalable video layers
Dominguez et al. The H. 264 video coding standard
EP1867172A1 (en) Method for encoding at least one digital picture, encoder, computer program product
Sanz-Rodriguez et al. RBF-based QP estimation model for VBR control in H. 264/SVC
Shojaei et al. FJND-based fuzzy rate control of scalable video for streaming applications
WO1998053613A1 (en) Apparatus, method and computer readable medium for scalable coding of video information
Sanz-Rodriguez et al. In-layer multibuffer framework for rate-controlled scalable video coding
KR101375302B1 (en) Apparatus and method of processing multimedia data
Macnicol et al. Scalable video coding by stream morphing
Seo et al. Rate control scheme for consistent video quality in scalable video codec
Tsai et al. Joint-layer encoder optimization for HEVC scalable extensions
Shen et al. Transcoding to FGS streams from H. 264/AVC hierarchical B-pictures
Atta Optimal bit allocation for subband video coding
Wu et al. An enhanced initialization method for rate control of H. 264 based on image quality balance
Atta OPTIMAL BIT ALLOCATION FOR SUBBAND IMAGE/VIDEO CODING
Cheong et al. A new scanning method for h. 264 based fine granular scalable video coding

Legal Events

Date Code Title Description
WAP Application withdrawn, taken to be withdrawn or refused ** after publication under section 16(1)