KR20140106450A - Method and apparatus for scalable video encoding considering memory bandwidth and calculation complexity, method and apparatus for scalable video decoding considering memory bandwidth and calculation complexity - Google Patents
Method and apparatus for scalable video encoding considering memory bandwidth and calculation complexity, method and apparatus for scalable video decoding considering memory bandwidth and calculation complexity Download PDFInfo
- Publication number
- KR20140106450A KR20140106450A KR1020140022179A KR20140022179A KR20140106450A KR 20140106450 A KR20140106450 A KR 20140106450A KR 1020140022179 A KR1020140022179 A KR 1020140022179A KR 20140022179 A KR20140022179 A KR 20140022179A KR 20140106450 A KR20140106450 A KR 20140106450A
- Authority
- KR
- South Korea
- Prior art keywords
- interpolation filtering
- prediction
- unit
- encoding
- image
- Prior art date
Links
Images
Classifications
-
- H—ELECTRICITY
- H04—ELECTRIC COMMUNICATION TECHNIQUE
- H04N—PICTORIAL COMMUNICATION, e.g. TELEVISION
- H04N19/00—Methods or arrangements for coding, decoding, compressing or decompressing digital video signals
- H04N19/10—Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using adaptive coding
- H04N19/102—Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using adaptive coding characterised by the element, parameter or selection affected or controlled by the adaptive coding
- H04N19/103—Selection of coding mode or of prediction mode
- H04N19/105—Selection of the reference unit for prediction within a chosen coding or prediction mode, e.g. adaptive choice of position and number of pixels used for prediction
-
- H—ELECTRICITY
- H04—ELECTRIC COMMUNICATION TECHNIQUE
- H04N—PICTORIAL COMMUNICATION, e.g. TELEVISION
- H04N19/00—Methods or arrangements for coding, decoding, compressing or decompressing digital video signals
- H04N19/10—Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using adaptive coding
- H04N19/102—Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using adaptive coding characterised by the element, parameter or selection affected or controlled by the adaptive coding
- H04N19/117—Filters, e.g. for pre-processing or post-processing
-
- H—ELECTRICITY
- H04—ELECTRIC COMMUNICATION TECHNIQUE
- H04N—PICTORIAL COMMUNICATION, e.g. TELEVISION
- H04N19/00—Methods or arrangements for coding, decoding, compressing or decompressing digital video signals
- H04N19/10—Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using adaptive coding
- H04N19/169—Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using adaptive coding characterised by the coding unit, i.e. the structural portion or semantic portion of the video signal being the object or the subject of the adaptive coding
- H04N19/17—Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using adaptive coding characterised by the coding unit, i.e. the structural portion or semantic portion of the video signal being the object or the subject of the adaptive coding the unit being an image region, e.g. an object
- H04N19/176—Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using adaptive coding characterised by the coding unit, i.e. the structural portion or semantic portion of the video signal being the object or the subject of the adaptive coding the unit being an image region, e.g. an object the region being a block, e.g. a macroblock
-
- H—ELECTRICITY
- H04—ELECTRIC COMMUNICATION TECHNIQUE
- H04N—PICTORIAL COMMUNICATION, e.g. TELEVISION
- H04N19/00—Methods or arrangements for coding, decoding, compressing or decompressing digital video signals
- H04N19/10—Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using adaptive coding
- H04N19/169—Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using adaptive coding characterised by the coding unit, i.e. the structural portion or semantic portion of the video signal being the object or the subject of the adaptive coding
- H04N19/187—Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using adaptive coding characterised by the coding unit, i.e. the structural portion or semantic portion of the video signal being the object or the subject of the adaptive coding the unit being a scalable video layer
-
- H—ELECTRICITY
- H04—ELECTRIC COMMUNICATION TECHNIQUE
- H04N—PICTORIAL COMMUNICATION, e.g. TELEVISION
- H04N19/00—Methods or arrangements for coding, decoding, compressing or decompressing digital video signals
- H04N19/30—Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using hierarchical techniques, e.g. scalability
- H04N19/34—Scalability techniques involving progressive bit-plane based encoding of the enhancement layer, e.g. fine granular scalability [FGS]
-
- H—ELECTRICITY
- H04—ELECTRIC COMMUNICATION TECHNIQUE
- H04N—PICTORIAL COMMUNICATION, e.g. TELEVISION
- H04N19/00—Methods or arrangements for coding, decoding, compressing or decompressing digital video signals
- H04N19/50—Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using predictive coding
- H04N19/503—Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using predictive coding involving temporal prediction
- H04N19/51—Motion estimation or motion compensation
- H04N19/513—Processing of motion vectors
- H04N19/517—Processing of motion vectors by encoding
- H04N19/52—Processing of motion vectors by encoding by predictive encoding
Abstract
Description
The present invention relates to a scalable video coding method and a decoding method. And more particularly, to a scalable video encoding method and a decoding method for optimizing a memory bandwidth and an arithmetic amount when interlaced prediction is performed.
Background of the Invention [0002] As the development and dissemination of hardware capable of playing back and storing high-resolution or high-definition video content increases the need for video codecs to effectively encode or decode high-definition or high-definition video content. According to the conventional video codec, video is encoded according to a limited encoding method based on a macroblock of a predetermined size.
The image data in the spatial domain is transformed into coefficients in the frequency domain using frequency conversion. The video codec divides an image into blocks of a predetermined size for fast calculation of frequency conversion, performs DCT conversion on each block, and encodes frequency coefficients on a block-by-block basis. Compared to image data in the spatial domain, coefficients in the frequency domain have a form that is easy to compress. In particular, since the image pixel values of the spatial domain are expressed by prediction errors through inter prediction or intra prediction of the video codec, many data can be converted to 0 when the frequency transformation is performed on the prediction error. Video codecs reduce the amount of data by replacing consecutively repeated data with small-sized data.
The multi-layer video codec decodes the base layer video and one or more enhancement layer video. The amount of data in this layer video and enhancement layer video can be reduced by eliminating temporal / spatial redundancy in the base layer video and enhancement layer video, respectively. It is possible to decode only the base layer video or decode both the base layer video and the enhancement layer video according to the playback capability of the receiving end.
Various embodiments of a scalable video encoding method and apparatus for limiting the number of interpolation filtering to minimize the increase of memory bandwidth and computation amount generated in interpolation filtering performed for inter prediction and inter-layer prediction are provided. There are provided various embodiments of a scalable video decoding method and apparatus in which inter prediction and inter-layer prediction are limited in an enhancement layer under predetermined conditions.
A scalable video encoding method according to various embodiments includes determining a reference layer image from among base layer images to inter-layer predict an enhancement layer image; Performing inter-layer (IL) interpolation filtering on the determined reference layer image to generate an upsampled reference layer image; And if the upsampled reference layer image is determined through the IL interpolation filtering, decide not to perform inter prediction between enhancement layer images for the enhancement layer image, And encoding the residue component.
According to various embodiments, the sum of the first calculation amount of MC interpolation filtering for inter prediction between the base layer images and the second calculation amount of MC interpolation filtering for inter prediction between the enhancement layer images, The computation amount of the interpolation filtering for interlayer prediction can be limited so as not to be large.
The step of encoding the residue component between the enhancement layer images according to various embodiments may include encoding a reference index indicating that the reference image of the enhancement layer image is the upsampled reference image, And coding the motion vector to represent zero.
According to various embodiments, at least one of the number of taps of the MC interpolation filter for MC interpolation filtering for inter prediction, the number of taps of the IL interpolation filter for the IL interpolation filtering, and the size of the prediction unit of the enhancement layer image, , The number of MC interpolation filtering and the number of IL interpolation filtering may be limited.
The scalable video decoding method according to various embodiments includes the steps of: obtaining a reference index indicating a residue component and a reference layer image for inter-layer prediction of an enhancement layer video; Determining, based on the reference index, not to perform inter prediction between enhancement layer images and determining the reference layer image from among base layer images; Generating an upsampled reference layer image by performing IL interpolation filtering on the determined reference layer image; And reconstructing the enhancement layer image using the residue component of the interlaced prediction and the upsampled reference layer image.
According to various embodiments, when the reference index of the enhancement layer image indicates the upsampled reference image, the step of determining the reference layer image may include determining a motion vector for inter-prediction between the enhancement layer images to be 0 .
According to various embodiments, a scalable video encoding apparatus includes: a base layer encoding unit for performing inter-prediction on base layer images; And determining that the inter-prediction between the enhancement layer images is not performed for the enhancement layer image if the up-sampled reference layer image is determined by performing IL interpolation filtering on the determined reference layer image, And an enhancement layer encoder for encoding a residue component between the enhanced layer video and the enhancement layer video.
A scalable video encoding apparatus according to various embodiments includes a base layer decoding unit for performing motion compensation to reconstruct base layer images; And determining a reference index indicating a reference layer index and a residue component for inter-layer prediction of an enhancement layer image, determines not to perform inter-prediction between enhancement layer images based on the reference index, And performing an IL interpolation filtering on the determined reference layer image to generate an upsampled reference layer image and restoring the enhancement layer image using the residue component of the interlayer prediction and the upsampled reference layer image And a layer decoding unit.
The present invention provides a computer-readable recording medium on which a computer program for implementing a scalable video encoding method according to various embodiments is recorded.
The present invention provides a computer-readable recording medium on which a computer program for implementing a scalable video decoding method according to various embodiments is recorded.
1 shows a block diagram of a scalable video coding apparatus according to various embodiments.
2 shows a block diagram of a scalable video decoding apparatus according to various embodiments.
3 shows a detailed structure of a scalable video encoding apparatus according to various embodiments.
4 shows an individual prediction structure of base layer images and enhancement layer images.
FIG. 5 shows an interlayer prediction structure of base layer images and enhancement layer images.
Figure 6 shows the bandwidth of memory required for interpolation filtering for a block.
7 shows a memory access pattern.
FIG. 8 illustrates a memory bandwidth for MC interpolation filtering that varies according to the inter prediction mode and the block size according to an embodiment.
FIG. 9 illustrates a memory bandwidth for IL interpolation filtering that varies with block size according to one embodiment.
FIG. 10 shows the number of interpolation filtering performed in base layer coding and enhancement layer coding, without restriction.
FIG. 11 illustrates the number of interpolation filtering performed in base layer coding and enhancement layer coding under a predetermined condition according to an embodiment.
Figure 12 lists combinations of MC interpolation filtering and IL interpolation filtering that can be performed under certain conditions according to various embodiments.
13 shows a flow chart of a scalable video coding method according to various embodiments.
14 shows a flowchart of a scalable video decoding method according to various embodiments.
15 shows a block diagram of a video coding apparatus based on a coding unit according to a tree structure according to various embodiments.
16 shows a block diagram of a video decoding apparatus based on a coding unit according to a tree structure according to various embodiments.
Figure 17 shows the concept of a coding unit according to various embodiments.
FIG. 18 shows a block diagram of an image encoding unit based on an encoding unit according to various embodiments.
FIG. 19 shows a block diagram of an image decoding unit based on an encoding unit according to various embodiments.
20 illustrates depth-specific encoding units and partitions according to various embodiments.
Figure 21 illustrates the relationship between an encoding unit and a conversion unit, according to various embodiments.
Figure 22 illustrates depth-specific encoding information, in accordance with various embodiments.
FIG. 23 shows a depth encoding unit according to various embodiments.
Figures 24, 25 and 26 show the relationship of the coding unit, the prediction unit and the conversion unit according to various embodiments.
FIG. 27 shows the relationship between the encoding unit, the prediction unit, and the conversion unit according to the encoding information information in Table 1. FIG.
28 illustrates a physical structure of a disk on which a program according to various embodiments is stored.
29 shows a disk drive for recording and reading a program using a disk.
30 shows the overall structure of a content supply system for providing a content distribution service.
31 and 32 illustrate an external structure and an internal structure of a mobile phone to which the video encoding method and the video decoding method of the present invention are applied according to various embodiments.
33 shows a digital broadcasting system to which a communication system is applied.
34 shows a network structure of a cloud computing system using a video encoding apparatus and a video decoding apparatus according to various embodiments.
A scalable video encoding apparatus, a scalable video decoding apparatus, a scalable video encoding method, and a scalable video decoding method according to various embodiments are described below with reference to Figs. 1 to 14. 15 to 27, a video coding apparatus, a video coding apparatus, a video coding method, and a video decoding method based on a coding unit of a tree structure according to various embodiments are disclosed. Various embodiments in which the scalable video encoding method, the scalable video decoding method, the video encoding method, and the video decoding method according to the embodiments of Figs. 1 to 27 are applicable will be described with reference to Figs. 28 to 34. Fig.
Hereinafter, 'video' may be a still image of a video or a video, that is, a video itself.
Hereinafter, 'sample' means data to be processed as data assigned to a sampling position of an image. For example, pixel values assigned to pixels in an image in the spatial domain may be samples.
Hereinafter, the symbol refers to a value of each syntax determined by performing encoding on an image. The bitstreams generated by performing entropy encoding on the symbols may be successively output to generate a bitstream. The entropy decoding is performed on the bit strings parsed from the bit stream to recover the symbols, and if decoding is performed using the symbols, the images can be reconstructed.
First, referring to FIG. 1 to FIG. 14, a scalable video encoding apparatus, a scalable video encoding method, and a scalable video decoding apparatus and a scalable video decoding method according to an embodiment are disclosed.
1 shows a block diagram of a scalable
The scalable
The scalable
The base
The enhancement
For example, according to a scalable video coding scheme based on spatial scalability, low resolution images can be encoded as base layer images, and high resolution images can be encoded as enhancement layer images. The encoding result of the base layer images may be output to the base layer stream and the encoding result of the enhancement layer images may be output to the enhancement layer stream.
As another example, according to the scalable video coding scheme based on SNR (Scalable Video Coding), the resolution and size of the base layer images and the enhancement layer images are the same, but the difference in the quantization parameter QP have. The larger the QP, the larger the quantization interval, and thus the quality of the reconstructed image is lowered. The low-quality images with relatively large QPs are encoded as the base layer images, and the high-quality images with relatively small QPs can be encoded as enhancement layer images.
As another example, multi-view video can be encoded according to a scalable video coding scheme. Left view images may be coded as base layer images, and right view images may be coded as enhancement layer images. Alternatively, the center view images, the left view images, and the right view images are coded. Among them, the central view images are coded as the base layer images, the left view images are the first enhancement layer images, Enhancement layer images.
As another example, a scalable video coding scheme may be performed according to Temporal Hierarchical Prediction based on temporal scalability. A base layer stream including encoded information generated by encoding images of a basic frame rate may be output. It is possible to further encode the images of the high frame rate by referring to the images of the basic frame rate and output the enhancement layer stream including the encoding information of the high frame rate.
The scalable
In addition, the scalable
The interlayer prediction structure will be described later with reference to FIG.
The scalable
Inter prediction and inter-layer prediction may be performed based on a data unit of an encoding unit, a prediction unit, or a conversion unit.
The base
The enhancement
The enhancement
The enhancement
For example, the restored image of the base layer image allocated with the same POC (Picture Order Count) as the enhancement layer image may be determined as the reference image. Also, of the blocks of the base layer reconstructed image, it may be determined that a block located corresponding to the position of the current block in the enhancement layer image is a reference block. The enhancement
The enhancement
Hereinafter, the case where the enhancement layer images are coded with reference to the base layer images according to the interlayer prediction will be described in detail.
The base
The enhancement
If the resolution is different between the base layer image and the enhancement layer image, such as spatial scalability, the image size is also different. Therefore, in order to generate the reference layer image for the enhancement layer image, the enhancement
In general, interpolation filtering can be performed to determine a reference block on a sub-pixel basis even when inter prediction on a sub-pixel basis is performed.
Hereinafter, the interpolation filtering for inter prediction of an encoding unit will be referred to as 'MC (Motion Interpolation) interpolation filtering', and the interpolation filtering for inter-layer prediction will be referred to as 'IL (Inter-Layer) interpolation filtering'. MC interpolation filtering and IL interpolation filtering are collectively referred to as 'interpolation filtering'.
In general, since interpolation filtering is performed using surrounding samples of the current sample, not only block samples but also some samples of adjacent blocks are required for interpolation filtering for the current block. Therefore, as the number of interpolation filtering for inter prediction and inter-layer prediction increases, the memory bandwidth and computation burden are greatly increased.
Therefore, in order to reduce the memory bandwidth and the computation burden, the
According to an exemplary embodiment, an up-sampled base layer reconstructed image, which is a reference layer image, is required for the inter-layer prediction structure. Therefore, the base
The enhancement
For example, when the scalable
The MC interpolation filtering and the number of times of performing the IL interpolation filtering can be used to evaluate the memory bandwidth or computation amount required for inter prediction or inter-layer prediction. Hereinafter, the number of times of filtering referred to in the present specification indicates the number of times that interpolation filtering is performed for one sample.
As described above, in order to perform the interlayer prediction for the enhancement layer image once, MC interpolation filtering for inter prediction, which is performed in the base layer, is performed once, and a basic layer restored image is generated for inter- The IL interpolation filtering for up-sampling may occur once.
In a typical individual layer prediction structure, if inter prediction is performed once in the base layer and inter prediction is performed once in the enhancement layer, interpolation filtering is performed at least twice in total by performing MC interpolation filtering and IL interpolation filtering once, respectively .
Therefore, when the interlayer prediction for the enhancement layer image is performed, the MC interpolation filtering for the base layer inter prediction and the IL interpolation filtering for the inter layer prediction are performed. Should not be added.
Therefore, when the up-sampled reference layer image is determined through the IL interpolation filtering, the enhancement
Accordingly, the enhancement
The scalable
For example, interpolation filtering for blocks of size 8x8 or larger can be limited to i) a combination of two 8-tap MC interpolation filtering or ii) a combination of 8 tap MC interpolation filtering and one 8 tap IL interpolation filtering .
As another example, interpolation filtering for blocks of size 4x8 or greater may be combined with iii) one 8-tap IL interpolation filtering, or iv) two 6 tap IL interpolation filtering or v) two 4 tap IL interpolation filtering, or vi) 3
Another example is interpolation filtering for blocks of size 8x16 or more, vii) a combination of two 8-tap MC interpolation filtering and a 4 tap IL interpolation filtering, or viii) four 2-tap MC interpolation filtering and 4 2-tap IL interpolation filtering , Or ix) a combination of two 8-tap MC interpolation filtering and two 2-tap IL interpolation filtering, or x) a combination of 8 8-tap MC interpolation filtering and 8 tap IL interpolation filtering.
The scalable
The scalable
The scalable
Therefore, the base
The scalable
2 shows a block diagram of a scalable
The scalable
The scalable
For example, the scalable
As another example, the scalable
As another example, multi-view video may be decoded according to a scalable video coding scheme. When the stereoscopic video stream is received in multiple layers, the left-view images can be reconstructed by decoding the base layer stream. The right view image can be restored by further decoding the enhancement layer stream to the base layer stream.
Alternatively, when the multi-view video stream is received in multiple layers, the central view images can be reconstructed by decoding the base layer stream. The left view image can be restored by further decoding the first enhancement layer stream in the base layer stream. The second enhancement layer stream may be further decoded in the base layer stream to restore the right view images.
As another example, a scalable video coding scheme based on temporal scalability can be performed. Images of the basic frame rate can be reconstructed by decoding the base layer stream. A higher frame rate image can be restored by further decoding the enhancement layer stream to the base layer stream.
The scalable
For example, the scalable
For each layer stream, the reconstructed images can be reconstructed by referring to reconstructed images predicted through inter prediction of the same layer, and performing motion compensation for the current image. Motion compensation refers to an operation of reconstructing a reconstructed image of a current image by synthesizing a residual image of a current image with a reference image determined using a motion vector of the current image.
In addition, the scalable
The interlayer prediction structure will be described later with reference to FIG.
The scalable
The base
The base
The base
The enhancement
Specifically, the base
According to the interlayer prediction structure, an enhancement layer prediction image can be generated using samples of a base layer reconstruction image. The enhancement
As described above, the enhancement
Hereinafter, the case where the enhancement layer images are decoded using the base layer reconstructed images according to the interlayer prediction will be described in detail.
The base
The enhancement
If the resolution is different between the base layer image and the enhancement layer image, such as spatial scalability, the image size is also different. Therefore, in order to generate the reference layer image for the enhancement layer image, the enhancement
Also, when inter prediction on the sub-pixel basis is performed, interpolation filtering can be performed to determine a reference block on a sub-pixel basis.
Therefore, in order to reduce the memory bandwidth and the calculation burden in the enhancement
As described above, for the inter-layer prediction structure, an up-sampled base layer reconstructed image which is a reference layer image is required. Therefore, the base
The enhancement
In the scalable
As described above with reference to FIG. 1, when the inter-layer prediction structure is compared with the individual layer prediction structure, when inter-layer prediction is performed for the enhancement layer image, MC interpolation filtering and inter- No additional interpolation filtering should be added in addition to the IL interpolation filtering.
Therefore, when the up-sampled reference layer image is determined through the IL interpolation filtering, the enhancement
For example, the enhancement
Therefore, the enhancement
The scalable
For example, interpolation filtering for blocks of size 8x8 or larger can be limited to i) a combination of two 8-tap MC interpolation filtering or ii) a combination of 8 tap MC interpolation filtering and one 8 tap IL interpolation filtering .
As another example, interpolation filtering for blocks of size 4x8 or greater may be combined with iii) one 8-tap IL interpolation filtering, or iv) two 6 tap IL interpolation filtering or v) two 4 tap IL interpolation filtering, or vi) 3
Another example is interpolation filtering for blocks of size 8x16 or more, vii) a combination of two 8-tap MC interpolation filtering and a 4 tap IL interpolation filtering, or viii) four 2-tap MC interpolation filtering and 4 2-tap IL interpolation filtering , Or ix) a combination of two 8-tap MC interpolation filtering and two 2-tap IL interpolation filtering, or x) a combination of 8 8-tap MC interpolation filtering and 8 tap IL interpolation filtering.
Therefore, the base
The scalable
The scalable
The scalable
According to the scalable
Therefore, in the scalable
Similarly, in the scalable
3 shows the detailed structure of the scalable
The
The base
The input image (low-resolution image, high-resolution image) is divided into a maximum encoding unit, an encoding unit, a prediction unit, a conversion unit, and the like through the
The residue components between the prediction unit and the surrounding image are input to the conversion /
The scaling /
In particular, in the inter mode, the in-
De-blocking filtering is filtering to mitigate blocking of data units, and SAO filtering is filtering to compensate pixel values that are modified by data encoding and decoding. The data filtered by the in-
In this way, the above-described encoding operation can be repeated for each encoding unit of the input image.
For the inter-layer prediction, the enhancement
The
When the
In order to encode the image, various encoding modes for the encoding unit, the prediction unit, and the conversion unit can be set. For example, as an encoding mode for an encoding unit, depth or split flag and the like can be set. The prediction mode, the partition type, the intra direction information, the reference list information, and the like can be set as the encoding mode for the prediction unit. As an encoding mode for the conversion unit, conversion depth or division information and the like can be set.
The base
The
For example, the enhancement
Similar to the
If the enhancement
The memory bandwidths of the individual layer prediction structure and the interlayer prediction structure are compared with reference to FIGS.
4 shows an individual prediction structure of base layer images and enhancement layer images.
According to the individual prediction structure, inter prediction can be performed in the base layer and the enhancement layer, respectively.
That is, in the base layer, the inter-prediction for the base layer image BL (45) is performed using at least one of the reference images (48, 49) belonging to the L0 reference list and the reference images Can be performed. Also, in the enhancement layer, the inter-prediction for the enhancement layer video EL (40) is performed using at least one of the reference images (43, 44) belonging to the L0 reference list and the reference images Can be performed.
In order to perform motion prediction or motion compensation on a sub-pixel basis, interpolation filtering is required on the reference image. Therefore, in order to perform inter prediction on the current image in each layer, MC interpolation filtering is performed once in the base layer, and MC interpolation filtering is performed once in the enhancement layer.
FIG. 5 shows an interlayer prediction structure of base layer images and enhancement layer images.
According to the interlayer prediction structure, inter prediction is performed in the base layer, and inter prediction and inter layer prediction can be performed in the enhancement layer.
That is, inter prediction for the base layer image BL (45) in the base layer can be performed. MC interpolation filtering can be performed once for sub-pixel motion prediction or motion compensation in the base layer.
In the enhancement layer, at least one of the reference images (53, 54) belonging to the L0 reference list and the reference images (51, 52) belonging to the L1 reference list is used, Can be performed. In the enhancement layer, MC interpolation filtering may be performed once for sub-pixel-based motion prediction or motion compensation.
Also, the up-sampled
In order to perform motion prediction or motion compensation on a sub-pixel basis, interpolation filtering is required on the reference image. Therefore, in order to encode the current image in each layer, interpolation filtering may be performed once for inter prediction of the base layer, and interpolation filtering may be performed for inter prediction of the enhancement layer.
Hereinafter, the computational complexity of the individual prediction structure and the interlayer prediction structure will be compared. The computational complexity can be evaluated in terms of the memory bandwidth required for the computation, the number of operations of multiplication and addition, the dynamic range of the sample to be computed, the memory size in which the filter coefficients are stored, and the computational latency. In the present specification, the computational complexity is evaluated using the memory bandwidth and the amount of computation (the number of computations) required for inter prediction and inter-layer prediction.
Hereinafter, the memory efficiency of the interpolation filtering occurring in the individual prediction structure and the interlayer prediction structure will be described in detail with reference to FIGS. 6 and 7. FIG.
Figure 6 shows the bandwidth of memory required for interpolation filtering for a block.
First, we want to determine how many adjacent pixels should be stored in memory for inter prediction of one sample in a block.
In the case of unidirectional prediction, interpolation filtering in one direction L0 or L1 is required. In the case of bidirectional prediction, a memory bandwidth is needed in which adjacent pixels for interpolation filtering in both directions L0 and L1 can be stored.
In the case of inter-layer prediction, a memory bandwidth is required in which adjacent pixels for interpolation filtering for up-sampling from the resolution of the base layer image can be stored.
When inter-layer prediction and inter prediction (unidirectional or bidirectional) are combined, both memory bandwidth for interpolation filtering for inter-layer prediction and memory bandwidth for interpolation filtering for inter prediction are both required.
When intercolor interpolation is used, the required memory bandwidth increases in proportion to the number of color components stored at different locations in the memory.
6, the width and height of the interpolation block including the samples to be interpolated are denoted by W and H, respectively. The width and height of the memory pattern representing the sample area that the memory can read at one time are denoted by w and h, respectively.
The filter length (number of filter taps) of the interpolation filter for the luma block is denoted by TL, and the filter length of the interpolation filter for the chroma block is denoted by TC.
The resolution of the enhancement layer to be predicted is S_EL, and the resolution of the base layer to be the reference object is represented by S_BL.
And the ratio S (= S_BL / S_EL) of the enhancement layer and the base layer. x 2 Spatial scalability can be expressed as
The
The memory bandwidth means the maximum amount of data that can be accessed at one time by accessing the memory. The larger the memory bandwidth required to perform inter prediction or inter-layer prediction, the lower the memory efficiency. For example, the closer S is to 1, the lower the memory efficiency.
The most memory bandwidth is required in the two-dimensional interpolation filtering in which the vertical direction performs interpolation filtering and the vertical direction interpolation filtering are sequentially performed. For example, the pixel blocks necessary for the two-dimensional interpolation filtering are determined by extending from the
That is, for the interpolation filtering of the
In order to perform the interpolation filtering of the
If the enhancement layer and the reference layer block have different spatial resolutions, scaling according to the resolution ratio is required. That is, the size of the memory area for interpolation filtering of the
7 shows a memory access pattern.
Also, instead of reading the samples one at a time from the memory, it is possible to read as many memory patterns as the width w and the height h. However, if the upper left corner of the
When the memory bandwidth is the largest, the memory efficiency is reduced to the lowest when the size of the memory pattern is added to the memory area for interpolation filtering.
That is, the maximum memory bandwidth required for interpolation filtering for the
Finally, in the enhancement layer block, the maximum memory bandwidth required for interpolation filtering for one sample containing all of the YCbCr components can be determined by the following equation.
((W + TL-1) / S + w-1) * ((H + TL-1) / S + h-1) + / 2 + TC-1) / S + h-1) / W / H
The above-mentioned memory bandwidth is the memory size to be accessed in interpolation filtering for unidirectional inter prediction or inter-layer prediction. For bidirectional inter prediction, twice the memory bandwidth will be required compared to the memory bandwidth described above. When inter-layer prediction and inter prediction are combined, a memory bandwidth equal to the sum of the memory bandwidths in all interpolation filtering is required.
Further, according to the above-described equation with reference to Figs. 6 and 7, it can be envisaged that the memory bandwidth varies depending on the block size and the inter prediction mode.
Accordingly, the scalable
FIG. 8 illustrates a memory bandwidth for MC interpolation filtering that varies according to the inter prediction mode and the block size according to an embodiment.
Referring to FIG. 8, the memory bandwidth required for interpolation filtering for prediction of the enhancement layer decreases from left to right.
The partition type of the block indicates the classification according to the size and type of the block.
According to one embodiment, bidirectional inter prediction is limited to 4x8 blocks and 8x4 in the enhancement layer.
On the other hand, bidirectional inter prediction for 8x8 blocks, 4x16 blocks, 16x4 blocks, unidirectional inter prediction for 4x8 blocks, bidirectional inter prediction for 8x16 blocks, and unidirectional inter prediction for 8x4 blocks are allowed in the enhancement layer. Also, according to the listed order, the memory bandwidth required in the interpolation filtering for each inter prediction is gradually reduced.
FIG. 9 illustrates a memory bandwidth for IL interpolation filtering that varies with block size according to one embodiment.
If two-dimensional IL interpolation filtering is performed for inter-layer prediction, the memory bandwidth required for IL interpolation filtering is equal to the memory bandwidth for MC interpolation filtering for unidirectional prediction. However, the smaller the size of the predicted block, the larger the memory bandwidth for IL interpolation filtering.
9, according to the order of 4x8 block, 8x4 block, 8x8 block, 4x16 block, 16x4 block, 8x16 block, 8x32 block, the memory bandwidth for IL interpolation filtering can be reduced.
The scalable
Hereinafter, with reference to FIGS. 10 and 11, the number of interpolation filtering required for base layer coding and enhancement layer coding is compared in the case of no constraint.
FIG. 10 shows the number of interpolation filtering performed in base layer coding and enhancement layer coding, without restriction.
Unidirectional prediction or bi-directional prediction can be performed first for motion compensation, but bi-directional prediction requires twice as much memory bandwidth. It is assumed that motion compensation of bidirectional prediction is performed in order to assume a case where memory bandwidth is most required in motion compensation.
First, in the two-layer individual coding structure, base layer inter prediction and enhancement layer inter prediction are performed. Motion compensation can be performed in the base layer inter prediction and the enhancement layer inter prediction, respectively.
For the motion compensation, horizontal reference MC interpolation filtering and vertical direction MC interpolation filtering are necessary because the reference block is determined for each subpixel. In addition, in order to compensate for motion in the L0 direction and the L1 direction, MC interpolation filtering in the horizontal direction and MC interpolation filtering in the vertical direction in each prediction direction can be performed.
Therefore, for the base layer inter prediction in the two-layer independent coding structure, horizontal MC interpolation filtering and vertical MC interpolation filtering can be performed in the L0 direction. In addition, horizontal MC interpolation filtering and vertical MC interpolation filtering can be performed in the L0 direction for enhanced layer inter prediction in the two-layer independent coding structure. Therefore, in the two-layer individual coding structure, the memory bandwidth and calculation burden required for four interpolation filtering can occur.
In the next two-layer reference coding structure, a prediction scheme may be different depending on whether interlayer prediction is performed.
First, if inter-layer prediction is not performed even in the case of a two-layer reference coding structure, horizontal MC interpolation filtering and vertical MC interpolation filtering are performed in the L0 direction for the base layer inter- Horizontal MC interpolation filtering and vertical MC interpolation filtering can be performed. For enhancement layer inter prediction, horizontal MC interpolation filtering and vertical MC interpolation filtering are performed in the L0 direction, and horizontal MC interpolation filtering and vertical MC interpolation filtering are performed in the L1 direction. That is, memory bandwidth and computation burden for up to 8 interpolation filters may be required.
However, if inter-layer prediction is performed in the two-layer reference coding structure, horizontal IL interpolation filtering and vertical direction IL interpolation filtering may be further performed to generate an upsampled reference image for inter-layer prediction . Therefore, when interlayer prediction is performed, a memory bandwidth and computation burden for a maximum of 10 interpolation filters at maximum may be required.
FIG. 11 illustrates the number of interpolation filtering performed in base layer coding and enhancement layer coding under a predetermined condition according to an embodiment.
The scalable
The scalable
In the scalable
Therefore, according to FIG. 11, the scalable
In addition, when the reference index for the current layer image indicates the up-sampled reference layer image, that is, when the inter-layer prediction is performed, the scalable
11, the sum of the first calculation amount of MC interpolation filtering for inter prediction between base layer images and the second calculation amount of MC interpolation filtering for inter prediction between enhancement layer images, The computation amount of interpolation filtering for prediction may not be large.
According to another embodiment, if both the inter-layer prediction and the inter-prediction can be performed in the enhancement layer, the scalable
Figure 12 lists combinations of MC interpolation filtering and IL interpolation filtering that can be performed under certain conditions according to various embodiments.
According to the H.265 / HEVC standard, an 8-tap interpolation filter is used for MC interpolation filtering of a luma block when inter prediction is performed on a block of 8x8 or larger size, and a 4-tap interpolation filter is used for MC interpolation filtering of a chroma block. Is used. According to the H.264 standard, 6-tap interpolation filter is used for MC interpolation filtering of luma block when inter prediction is performed for blocks 4x4 or larger in size, and 2-tap interpolation filter is used for MC interpolation filtering of chroma block do.
The scalable
The scalable
The scalable
In the case of performing inter-layer prediction using a 6-tap or 4-tap IL interpolation filter for a 4x8 or larger luma block, two IL interpolation filtering is allowed, but inter prediction can be limited. For the chroma block of the block, two 2-tap IL interpolation filtering may be allowed.
Three IL interpolation filtering is allowed when performing inter-layer prediction using a 2-tap IL interpolation filter for a 4x8 or larger size luma block, but inter prediction can be limited. Three times of 2-tap IL interpolation filtering may be allowed for the chroma block of the corresponding block.
The scalable
In addition, when the scalable
In addition, when the scalable
In addition, when the scalable
As described above with respect to various combinations of inter prediction and interlayer prediction that can be performed in the enhancement layer, it is confirmed that the minimum block size in which the combination of interlayer prediction and inter prediction can be allowed is 8x8.
13 shows a flow chart of a scalable video coding method according to various embodiments.
In
First, the scalable
In
The scalable
If the scalable
In
The scalable
14 shows a flowchart of a scalable video decoding method according to various embodiments.
In
In
When the reference index of the enhancement layer image indicates the upsampled reference image, the scalable
First, the scalable
In
The scalable
If performing an inter-layer prediction on one of the enhancement layer images, then the total interpolation filtering may be limited to four MC interpolation filtering at
In
The scalable
Accordingly, the sum of the first calculation amount of MC interpolation filtering for inter-prediction between base layer images and the second calculation amount of MC interpolation filtering for inter prediction between the enhancement layer images, The computation amount of the interpolation filtering can be adjusted so as not to be large.
In the scalable
In principle, encoding / decoding processes for base layer images and encoding / decoding processes for enhancement layer images are separately performed in the encoding / decoding process for multi-layer video. That is, when interlaced prediction occurs in the multi-layer video, the coding / decoding result of the single layer video can be cross-referenced, but a separate coding / decoding process occurs for each layer video.
Therefore, for convenience of explanation, the video encoding process and the video decoding process based on the encoding units of the tree structure described below with reference to FIGS. 15 to 27 are a video encoding process and a video decoding process for single layer video, . However, as described above with reference to FIGS. 1 to 14, interlayer prediction and compensation between the basic view images and the enhancement layer images are performed for video stream encoding / decoding.
Therefore, in order for the encoding unit 12 of the scalable
Similarly, in order for the decoding unit 26 of the scalable
FIG. 15 shows a block diagram of a
The
The encoding
An encoding unit according to an embodiment may be characterized by a maximum size and a depth. The depth indicates the number of times the coding unit is spatially divided from the maximum coding unit. As the depth increases, the depth coding unit can be divided from the maximum coding unit to the minimum coding unit. The depth of the maximum encoding unit is the highest depth and the minimum encoding unit can be defined as the least significant encoding unit. As the depth of the maximum encoding unit increases, the size of the depth-dependent encoding unit decreases, so that the encoding unit of the higher depth may include a plurality of lower-depth encoding units.
As described above, according to the maximum size of an encoding unit, the image data of the current picture is divided into a maximum encoding unit, and each maximum encoding unit may include encoding units divided by depth. Since the maximum encoding unit according to an embodiment is divided by depth, image data of a spatial domain included in the maximum encoding unit can be hierarchically classified according to depth.
The maximum depth for limiting the total number of times the height and width of the maximum encoding unit can be hierarchically divided and the maximum size of the encoding unit may be preset.
The encoding
The image data in the maximum encoding unit is encoded based on the depth encoding unit according to at least one depth below the maximum depth, and the encoding results based on the respective depth encoding units are compared. As a result of the comparison of the encoding error of the depth-dependent encoding unit, the depth with the smallest encoding error can be selected. At least one coding depth may be determined for each maximum coding unit.
As the depth of the maximum encoding unit increases, the encoding unit is hierarchically divided and divided, and the number of encoding units increases. In addition, even if encoding units of the same depth included in one maximum encoding unit, the encoding error of each data is measured and it is determined whether or not the encoding unit is divided into lower depths. Therefore, even if the data included in one maximum coding unit has a different coding error according to the position, the coding depth can be determined depending on the position. Accordingly, one or more coding depths may be set for one maximum coding unit, and data of the maximum coding unit may be divided according to one or more coding depth encoding units.
Therefore, the
The maximum depth according to one embodiment is an index related to the number of divisions from the maximum encoding unit to the minimum encoding unit. The first maximum depth according to an exemplary embodiment may indicate the total number of division from the maximum encoding unit to the minimum encoding unit. The second maximum depth according to an exemplary embodiment may represent the total number of depth levels from the maximum encoding unit to the minimum encoding unit. For example, when the depth of the maximum encoding unit is 0, the depth of the encoding unit in which the maximum encoding unit is divided once may be set to 1, and the depth of the encoding unit that is divided twice may be set to 2. In this case, if the coding unit divided four times from the maximum coding unit is the minimum coding unit, since the depth levels of
Prediction encoding and conversion of the maximum encoding unit can be performed. Likewise, predictive coding and conversion are performed on the basis of the depth coding unit for each maximum coding unit and for each depth below the maximum depth.
Since the number of coding units per depth is increased every time the maximum coding unit is divided by depth, the coding including prediction coding and conversion should be performed for every depth coding unit as the depth increases. For convenience of explanation, predictive encoding and conversion will be described based on a current encoding unit of at least one of the maximum encoding units.
The
For example, the
For predictive coding of the maximum coding unit, predictive coding may be performed based on a coding unit of coding depth according to an embodiment, i.e., a coding unit which is not further divided. Hereinafter, the more unfragmented encoding units that are the basis of predictive encoding will be referred to as 'prediction units'. The partition in which the prediction unit is divided may include a data unit in which at least one of the height and the width of the prediction unit and the prediction unit is divided. A partition is a data unit in which a prediction unit of a coding unit is divided, and a prediction unit may be a partition having the same size as a coding unit.
For example, if the encoding unit of size 2Nx2N (where N is a positive integer) is not further divided, it is a prediction unit of size 2Nx2N, and the size of the partition may be 2Nx2N, 2NxN, Nx2N, NxN, and the like. The partition type according to an embodiment is not limited to symmetric partitions in which the height or width of a prediction unit is divided by a symmetric ratio, but also partitions partitioned asymmetrically, such as 1: n or n: 1, Partitioned partitions, arbitrary type partitions, and the like.
The prediction mode of the prediction unit may be at least one of an intra mode, an inter mode, and a skip mode. For example, intra mode and inter mode can be performed for partitions of 2Nx2N, 2NxN, Nx2N, NxN sizes. In addition, the skip mode can be performed only for a partition of 2Nx2N size. Encoding is performed independently for each prediction unit within an encoding unit, and a prediction mode having the smallest encoding error can be selected.
In addition, the
The conversion unit in the encoding unit is also recursively divided into smaller conversion units in a similar manner to the encoding unit according to the tree structure according to the embodiment, And can be partitioned according to the conversion unit.
For a conversion unit according to one embodiment, a conversion depth indicating the number of times of division until the conversion unit is divided by the height and width of the encoding unit can be set. For example, if the size of the conversion unit of the current encoding unit of size 2Nx2N is 2Nx2N, the conversion depth is set to 0 if the conversion depth is 0, if the conversion unit size is NxN, and if the conversion unit size is N / 2xN / 2, . That is, a conversion unit according to the tree structure can be set for the conversion unit according to the conversion depth.
The coding information according to the coding depth needs not only the coding depth but also prediction related information and conversion related information. Therefore, the coding
The encoding unit, the prediction unit / partition, and the conversion unit determination method according to the tree structure of the maximum encoding unit according to the embodiment will be described later in detail with reference to FIGS. 17 to 27.
The encoding
The
The encoded image data may be a result of encoding residual data of the image.
The information on the depth-dependent coding mode may include coding depth information, partition type information of a prediction unit, prediction mode information, size information of a conversion unit, and the like.
The coding depth information can be defined using depth division information indicating whether or not coding is performed at the lower depth coding unit without coding at the current depth. If the current depth of the current encoding unit is the encoding depth, the current encoding unit is encoded in the current depth encoding unit, so that the division information of the current depth can be defined so as not to be further divided into lower depths. On the other hand, if the current depth of the current encoding unit is not the encoding depth, the encoding using the lower depth encoding unit should be tried. Therefore, the division information of the current depth may be defined to be divided into the lower depth encoding units.
If the current depth is not the encoding depth, encoding is performed on the encoding unit divided into lower-depth encoding units. Since there are one or more lower-level coding units in the current-depth coding unit, the coding is repeatedly performed for each lower-level coding unit so that recursive coding can be performed for each coding unit of the same depth.
Since the coding units of the tree structure are determined in one maximum coding unit and information on at least one coding mode is determined for each coding unit of coding depth, information on at least one coding mode is determined for one maximum coding unit . Since the data of the maximum encoding unit is hierarchically divided according to the depth and the depth of encoding may be different for each position, information on the encoding depth and the encoding mode may be set for the data.
Accordingly, the
The minimum unit according to an exemplary embodiment is a square data unit having a minimum coding unit having the lowest coding depth divided into quadrants. The minimum unit according to an exemplary embodiment may be a maximum size square data unit that can be included in all coding units, prediction units, partition units, and conversion units included in the maximum coding unit.
For example, the encoding information output through the
Information on the maximum size of a coding unit defined for each picture, slice or GOP, and information on the maximum depth can be inserted into a header, a sequence parameter set, or a picture parameter set of a bitstream.
Information on the maximum size of the conversion unit allowed for the current video and information on the minimum size of the conversion unit can also be output through a header, a sequence parameter set, or a picture parameter set or the like of the bit stream. The
According to the simplest embodiment of the
Therefore, the
Therefore, if an image having a very high image resolution or a very large data amount is encoded in units of existing macroblocks, the number of macroblocks per picture becomes excessively large. This increases the amount of compression information generated for each macroblock, so that the burden of transmission of compressed information increases and the data compression efficiency tends to decrease. Therefore, the video encoding apparatus according to an embodiment can increase the maximum size of the encoding unit in consideration of the image size, and adjust the encoding unit in consideration of the image characteristic, so that the image compression efficiency can be increased.
The scalable
When the
Even when the
FIG. 16 shows a block diagram of a
A
Definitions of various terms such as coding unit, depth, prediction unit, conversion unit, and information on various coding modes for the decoding operation of the
The receiving
Also, the image data and encoding
Information on the coding depth and coding mode per coding unit can be set for one or more coding depth information, and the information on the coding mode for each coding depth is divided into partition type information of the coding unit, prediction mode information, The size information of the image data, and the like. In addition, as the encoding depth information, depth-based segmentation information may be extracted.
The encoding depth and encoding mode information extracted by the image data and encoding
The encoding information for the encoding depth and the encoding mode according to the embodiment may be allocated for a predetermined data unit among the encoding unit, the prediction unit and the minimum unit. Therefore, the image data and the encoding
The image
The image
In addition, the image
The image
In other words, the encoding information set for the predetermined unit of data among the encoding unit, the prediction unit and the minimum unit is observed, and the data units holding the encoding information including the same division information are collected, and the image
The scalable
When the base layer video stream is received, the video
When the enhancement layer video stream is received, the video
As a result, the
Accordingly, even if an image with a high resolution or an excessively large amount of data is used, the information on the optimal encoding mode transmitted from the encoding end is used, and the image data is efficiently encoded according to the encoding unit size and encoding mode, Can be decoded and restored.
Figure 17 shows the concept of a coding unit according to various embodiments.
An example of an encoding unit is that the size of an encoding unit is represented by a width x height, and may include 32x32, 16x16, and 8x8 from an encoding unit having a size of 64x64. The encoding unit of size 64x64 can be divided into the partitions of size 64x64, 64x32, 32x64, 32x32, and the encoding unit of size 32x32 is the partitions of size 32x32, 32x16, 16x32, 16x16 and the encoding unit of size 16x16 is the size of 16x16 , 16x8, 8x16, and 8x8, and a size 8x8 encoding unit can be divided into partitions of size 8x8, 8x4, 4x8, and 4x4.
With respect to the
It is preferable that the maximum size of the coding size is relatively large in order to improve the coding efficiency as well as to accurately characterize the image characteristics when the resolution or the data amount is large. Therefore, the maximum size of the
Since the maximum depth of the
Since the maximum depth of the
FIG. 18 shows a block diagram of an
The
The data output from the
The
In particular, the
FIG. 19 shows a block diagram of an
The
The
The data in the spatial domain that has passed through the
In order to decode the image data in the image
A
In particular, the
The coding operation of Fig. 18 and the decoding operation of Fig. 19 are described above for the video stream coding operation and the decoding operation in a single layer, respectively. Therefore, if the scalable
20 illustrates depth-specific encoding units and partitions according to various embodiments.
The
The
That is, the
Prediction units and partitions of coding units are arranged along the horizontal axis for each depth. That is, if the
Likewise, the prediction unit of the
Likewise, the prediction unit of the
Likewise, the prediction unit of the
The encoding
The number of coding units per depth to include data of the same range and size increases as the depth of the coding unit increases. For example, for data containing one coding unit at
For each depth-of-field coding, encoding is performed for each prediction unit of the depth-dependent coding unit along the horizontal axis of the
Figure 21 illustrates the relationship between an encoding unit and a conversion unit, according to various embodiments.
The
For example, in the
In addition, the data of the 64x64 encoding unit 710 is converted into 32x32, 16x16, 8x8, and 4x4 conversion units each having a size of 64x64 or smaller, and then a conversion unit having the smallest error with the original is selected .
Figure 12 shows depth-specific encoding information, in accordance with various embodiments.
The
The partition type information 800 represents information on the type of partition in which the prediction unit of the current encoding unit is divided, as a data unit for predictive encoding of the current encoding unit. For example, the current encoding unit CU_0 of size 2Nx2N may be any one of a
The prediction mode information 810 indicates a prediction mode of each partition. For example, it is determined whether the partition indicated by the information 800 relating to the partition type is predictive-encoded in one of the
In addition, the information 820 on the conversion unit size indicates whether to perform conversion based on which conversion unit the current encoding unit is to be converted. For example, the conversion unit may be one of a first
The video data and encoding
FIG. 23 shows a depth encoding unit according to various embodiments.
Partition information may be used to indicate changes in depth. The division information indicates whether the current-depth encoding unit is divided into lower-depth encoding units.
The
For each partition type, predictive encoding should be repeatedly performed for each partition of size 2N_0x2N_0, two 2N_0xN_0 partitions, two N_0x2N_0 partitions, and four N_0xN_0 partitions. For a partition of size 2N_0x2N_0, size N_0x2N_0, size 2N_0xN_0 and size N_0xN_0, predictive coding can be performed in intra mode and inter mode. The skip mode can be performed only on the partition of size 2N_0x2N_0 with predictive coding.
If the encoding error caused by one of the
If the coding error by the
A
If the encoding error by the
If the maximum depth is d, the depth-based coding unit is set up to the depth d-1, and the division information can be set up to the depth d-2. That is, when the encoding is performed from the depth d-2 to the depth d-1, the prediction encoding of the
(D-1) x2N_ (d-1), two size 2N_ (d-1) xN_ (d-1) partitions, and two sizes N_ (d-1) and the partition of four sizes N_ (d-1) xN_ (d-1), the partition type in which the minimum coding error occurs can be retrieved .
Even if the coding error by the
The
In this way, the minimum coding error of each of the
The video data and encoding
Figures 24, 25 and 26 show the relationship of the coding unit, the prediction unit and the conversion unit according to various embodiments.
The coding unit 1010 is coding units for coding depth determined by the
When the depth of the maximum encoding unit is 0, the depth of the
Some
The image data of a
Thus, for each maximum encoding unit, the encoding units are recursively performed for each encoding unit hierarchically structured in each region, and the optimal encoding unit is determined, so that encoding units according to the recursive tree structure can be constructed. The encoding information may include division information for the encoding unit, partition type information, prediction mode information, and conversion unit size information. Table 1 below shows an example that can be set in the
Inter
Skip (2Nx2N only)
2NxN
Nx2N
NxN
2NxnD
nLx2N
nRx2N
(Symmetrical partition type)
N / 2xN / 2
(Asymmetric partition type)
The
The division information indicates whether the current encoding unit is divided into low-depth encoding units. If the division information of the current depth d is 0, since the depth at which the current encoding unit is not further divided into the current encoding unit is the encoding depth, the partition type information, prediction mode, and conversion unit size information are defined . When it is necessary to further divide by one division according to the division information, encoding should be performed independently for each of four divided sub-depth coding units.
The prediction mode may be represented by one of an intra mode, an inter mode, and a skip mode. Intra mode and inter mode can be defined in all partition types, and skip mode can be defined only in partition type 2Nx2N.
The partition type information indicates symmetrical partition types 2Nx2N, 2NxN, Nx2N and NxN in which the height or width of the predicted unit is divided into symmetric proportions and asymmetric partition types 2NxnU, 2NxnD, nLx2N, and nRx2N divided by the asymmetric ratio . Asymmetric partition types 2NxnU and 2NxnD are respectively divided into heights 1: 3 and 3: 1, and asymmetric partition types nLx2N and nRx2N are respectively divided into widths of 1: 3 and 3: 1.
The conversion unit size can be set to two kinds of sizes in the intra mode and two kinds of sizes in the inter mode. That is, if the conversion unit division information is 0, the size of the conversion unit is set to the size 2Nx2N of the current encoding unit. If the conversion unit division information is 1, a conversion unit of the size where the current encoding unit is divided can be set. Also, if the partition type for the current encoding unit of size 2Nx2N is a symmetric partition type, the size of the conversion unit may be set to NxN, or N / 2xN / 2 if it is an asymmetric partition type.
The encoding information of the encoding units according to the tree structure according to an exemplary embodiment may be allocated to at least one of encoding units, prediction units, and minimum unit units of the encoding depth. The coding unit of the coding depth may include one or more prediction units and minimum units having the same coding information.
Therefore, if encoding information held in adjacent data units is checked, it can be confirmed whether or not the encoded information is included in the encoding unit of the same encoding depth. In addition, since the encoding unit of the encoding depth can be identified by using the encoding information held by the data unit, the distribution of encoding depths within the maximum encoding unit can be inferred.
Therefore, in this case, when the current encoding unit is predicted with reference to the neighboring data unit, the encoding information of the data unit in the depth encoding unit adjacent to the current encoding unit can be directly referenced and used.
In another embodiment, when predictive encoding is performed with reference to a current encoding unit with reference to a surrounding encoding unit, data adjacent to the current encoding unit in the depth encoding unit is encoded using the encoding information of adjacent encoding units The surrounding encoding unit may be referred to by being searched.
FIG. 27 shows the relationship between the encoding unit, the prediction unit, and the conversion unit according to the encoding mode information in Table 1.
The
The TU size flag is a kind of conversion index, and the size of the conversion unit corresponding to the conversion index can be changed according to the prediction unit type or partition type of the coding unit.
For example, when the partition type information is set to one of the symmetric
When the partition type information is set to one of the asymmetric
The TU size flag described above with reference to FIG. 27 is a flag having a value of 0 or 1, but the conversion unit division information according to the embodiment is not limited to a 1-bit flag, , 1, 2, 3, etc., and the conversion unit may be divided hierarchically. The conversion unit partition information can be used as an embodiment of the conversion index.
In this case, if the conversion unit division information according to the embodiment is used together with the maximum size of the conversion unit and the minimum size of the conversion unit, the size of the conversion unit actually used can be expressed. The
For example, if (a) the current encoding unit is 64x64 and the maximum conversion unit size is 32x32, (a-1) when the conversion unit division information is 0, the size of the conversion unit is 32x32, When the division information is 1, the size of the conversion unit is 16x16, (a-3) When the conversion unit division information is 2, the size of the conversion unit can be set to 8x8.
As another example, (b) if the current encoding unit is 32x32 and the minimum conversion unit size is 32x32, the size of the conversion unit may be set to 32x32 when the conversion unit division information is 0, Since the size can not be smaller than 32x32, further conversion unit division information can not be set.
As another example, (c) if the current encoding unit is 64x64 and the maximum conversion unit division information is 1, the conversion unit division information may be 0 or 1, and other conversion unit division information can not be set.
Therefore, when the maximum conversion unit division information is defined as 'MaxTransformSizeIndex', the minimum conversion unit size is defined as 'MinTransformSize', and the conversion unit size when the conversion unit division information is 0 is defined as 'RootTuSize', the minimum conversion unit The size 'CurrMinTuSize' can be defined as the following relation (1).
CurrMinTuSize
= max (MinTransformSize, RootTuSize / (2 ^ MaxTransformSizeIndex)) (1)
'RootTuSize', which is the conversion unit size when the conversion unit division information is 0 as compared with the minimum conversion unit size 'CurrMinTuSize' possible in the current encoding unit, can represent the maximum conversion unit size that can be adopted by the system. That is, according to the relational expression (1), 'RootTuSize / (2 ^ MaxTransformSizeIndex)' is obtained by dividing 'RootTuSize', which is the conversion unit size in the case where the conversion unit division information is 0, by the number corresponding to the maximum conversion unit division information Unit size, and 'MinTransformSize' is the minimum conversion unit size, so a smaller value of these may be the minimum conversion unit size 'CurrMinTuSize' that is currently available in the current encoding unit.
The maximum conversion unit size RootTuSize according to an exemplary embodiment may vary depending on the prediction mode.
For example, if the current prediction mode is the inter mode, RootTuSize can be determined according to the following relation (2). In the relation (2), 'MaxTransformSize' indicates the maximum conversion unit size and 'PUSize' indicates the current prediction unit size.
RootTuSize = min (MaxTransformSize, PUSize) (2)
That is, if the current prediction mode is the inter mode, 'RootTuSize' which is the conversion unit size when the conversion unit division information is 0 can be set to a smaller value of the maximum conversion unit size and the current prediction unit size.
If the prediction mode of the current partition unit is the intra mode, if the prediction mode is the mode, 'RootTuSize' can be determined according to the following relation (3). 'PartitionSize' represents the size of the current partition unit.
RootTuSize = min (MaxTransformSize, PartitionSize) (3)
That is, if the current prediction mode is the intra mode, 'RootTuSize' which is the conversion unit size when the conversion unit division information is 0 can be set to a smaller value among the maximum conversion unit size and the size of the current partition unit.
However, it should be noted that the present maximum conversion unit size 'RootTuSize' according to one embodiment that varies according to the prediction mode of the partition unit is only one embodiment, and the factor for determining the current maximum conversion unit size is not limited thereto.
According to a video coding technique based on the coding units of the tree structure described above with reference to FIGS. 15 to 27, video data of a spatial region is encoded for each coding unit of a tree structure, and a video decoding technique based on coding units of a tree structure Decoding is performed for each maximum encoding unit according to the motion vector, and the video data in the spatial domain is reconstructed, and the video and the video, which is a picture sequence, can be reconstructed. The restored video can be played back by the playback apparatus, stored in a storage medium, or transmitted over a network.
The above-described embodiments of the present invention can be embodied in a general-purpose digital computer that can be embodied as a program that can be executed by a computer and operates the program using a computer-readable recording medium. The computer-readable recording medium includes a storage medium such as a magnetic storage medium (e.g., ROM, floppy disk, hard disk, etc.), optical reading medium (e.g., CD ROM,
For convenience of explanation, the scalable video encoding method and / or the video encoding method described above with reference to FIGS. 1 to 27 are collectively referred to as 'video encoding method of the present invention'. In addition, the scalable video decoding method and / or video decoding method described above with reference to FIGS. 1 to 27 is referred to as 'video decoding method of the present invention'
The video encoding apparatus composed of the scalable
An embodiment in which the computer-readable storage medium on which the program according to one embodiment is stored is
Figure 28 illustrates the physical structure of a
A computer system achieved using the above-described video encoding method and a storage medium storing a program for implementing the video decoding method will be described below with reference to FIG.
Fig. 29 shows a
A program for implementing at least one of the video coding method and the video decoding method of the present invention may be stored in a memory card, a ROM cassette, or a solid state drive (SSD) as well as the
A system to which the video coding method and the video decoding method according to the above-described embodiments are applied will be described later.
30 shows the overall structure of a
The
However, the
The video camera 12300 is an imaging device that can capture a video image such as a digital video camera. The
The video camera 12300 may be connected to the streaming server 11300 via the
The video data photographed by the camera 12600 may also be transmitted to the streaming server 11300 via the computer 12100. [ The camera 12600 is an imaging device that can capture both still images and video images like a digital camera. The video data received from the camera 12600 may be encoded by the camera 12600 or the computer 12100. [ The software for video encoding and decoding may be stored in a computer readable recording medium such as a CD-ROM disk, a floppy disk, a hard disk drive, an SSD, or a memory card, to which the computer 12100 can access.
Also, when video is taken by a camera mounted on the
The video data can be encoded by a large scale integrated circuit (LSI) system mounted on the video camera 12300, the
In a
Clients are devices capable of decoding encoded content data and may be, for example, a computer 12100, a
The video encoding apparatus and the video decoding apparatus of the present invention can be applied to the encoding operation and the decode operation of the independent devices included in the
One embodiment of the
31 shows an external structure of a
The
32 shows the internal structure of the
The
The
A digital signal is generated in the
For example, the sound signal obtained by the
When a text message such as e-mail is transmitted in the data communication mode, the text data of the message is input using the
In order to transmit the image data in the data communication mode, the image data photographed by the
The structure of the
The multiplexing /
In the process of receiving communication data from the outside of the
When the
In a data communication mode, when data of an accessed video file is received from a web site of the Internet, a signal received from the
In order to decode the multiplexed data received via the
The structure of the
Accordingly, the video data of the video file accessed from the web site of the Internet can be displayed on the display screen 1252. [ At the same time, the sound processing unit 1265 can also convert the audio data to an analog sound signal and provide an analog sound signal to the speaker 1258. [ Accordingly, the audio data included in the video file accessed from the web site of the Internet can also be played back on the speaker 1258. [
The cellular phone 1250 or another type of communication terminal may be a transmitting terminal including both the video coding apparatus and the video decoding apparatus of the present invention or a transmitting terminal including only the video coding apparatus of the present invention described above, Only the receiving terminal may be included.
The communication system of the present invention is not limited to the above-described structure with reference to Fig. For example, FIG. 33 shows a digital broadcasting system to which a communication system according to various embodiments is applied. The digital broadcasting system according to the embodiment of FIG. 33 can receive digital broadcasting transmitted through a satellite or a terrestrial network by using the video encoding apparatus and the video decoding apparatus of the present invention.
Specifically, the broadcasting station 12890 transmits the video data stream to the communication satellite or broadcast satellite 12900 through radio waves. The broadcast satellite 12900 transmits the broadcast signal, and the broadcast signal is received by the satellite broadcast receiver by the antenna 12860 in the home. In each assumption, the encoded video stream may be decoded and played back by a
By implementing the video decoding apparatus of the present invention in the reproducing apparatus 12830, the reproducing apparatus 12830 can read and decode the encoded video stream recorded in the
The video decoding apparatus of the present invention may be installed in the set-top box 12870 connected to the antenna 12860 for satellite / terrestrial broadcast or the cable antenna 12850 for cable TV reception. The output data of the set-top box 12870 can also be played back on the
As another example, the video decoding apparatus of the present invention may be mounted on the
An automobile 12920 having an appropriate antenna 12910 may receive a signal transmitted from the satellite 12800 or the
The video signal can be encoded by the video encoding apparatus of the present invention and recorded and stored in the storage medium. Specifically, the video signal may be stored in the
The car navigation system 12930 may not include the
34 shows a network structure of a cloud computing system using a video encoding apparatus and a video decoding apparatus according to various embodiments.
The cloud computing system of the present invention may include a
The cloud computing system provides an on demand outsourcing service of computing resources through an information communication network such as the Internet according to a request of a user terminal. In a cloud computing environment, service providers integrate computing resources in data centers that are in different physical locations into virtualization technologies to provide services to users. Service users do not install and use computing resources such as application, storage, OS, security, etc. in the terminals owned by each user, but instead use services in the virtual space created through virtualization technology Can be selected and used as desired.
A user terminal of a specific service user accesses the
The
The
Information on the moving image stored in the
The
At this time, the user terminal may include the video decoding apparatus of the present invention described above with reference to Figs. 1 to 27. As another example, the user terminal may include the video encoding apparatus of the present invention described above with reference to Figs. Also, the user terminal may include both the video encoding apparatus and the video decoding apparatus of the present invention described above with reference to Figs. 1 to 27.
Various embodiments in which the video coding method and the video decoding method, video coding apparatus and video decoding apparatus described above with reference to Figs. 1 to 27 are utilized have been described in Figs. 28 to 34. Fig. However, various embodiments in which the video encoding method and the video decoding method described above with reference to Figs. 1 to 27 are stored in a storage medium or in which a video encoding apparatus and a video decoding apparatus are implemented in a device, .
It will be understood by those skilled in the art that various embodiments disclosed herein may be embodied in various forms without departing from the essential characteristics of the embodiments disclosed herein. Therefore, the disclosed embodiments should be considered in an illustrative rather than a restrictive sense. The scope of the present disclosure is set forth in the following claims, rather than the foregoing description, and all differences within the scope of equivalents thereof should be construed as being included within the scope of the present disclosure.
Claims (18)
Determining a reference layer image among base layer images to inter-layer predict an enhancement layer image;
Performing inter-layer (IL) interpolation filtering on the determined reference layer image to generate an upsampled reference layer image; And
Sampled reference layer image is determined through the IL interpolation filtering, the inter-prediction between the enhancement layer images is determined not to be performed for the enhancement layer image, and the inter-prediction between the upsampled reference layer image and the enhancement layer image Wherein the quantization step includes the steps of:
Performing four MC interpolation filtering including interpolation filtering in the horizontal direction and interpolation filtering in the vertical direction in the L0 prediction direction and the L1 prediction direction with respect to the base layer image,
Wherein the step of generating the upsampled reference layer video comprises:
Performing two IL interpolation filtering including horizontal directional IL interpolation filtering and vertical directional IL interpolation filtering on the determined reference layer image,
Wherein the interpolation filtering for inter-layer prediction of the enhancement layer image is limited to the MC interpolation filtering for the four times and the IL interpolation filtering for the two times.
Layer interpolation for an interlayer prediction of the enhancement layer image, the sum of a first calculation amount of MC interpolation filtering for inter prediction between the base layer images and a second calculation amount of MC interpolation filtering for inter prediction between the enhancement layer images, And the amount of filtering is limited so as not to be large.
Encoding a reference index indicating that the reference image of the enhancement layer image is the upsampled reference image and encoding a motion vector for inter-prediction between the enhancement layer images to indicate 0, A method for encoding a video signal.
In the enhancement layer video, whether to perform inter prediction is determined based on the size, shape, and prediction direction of the block,
Wherein the number of taps of the IL interpolation filter for the IL interpolation filtering is limited not to be greater than the number of taps of the MC interpolation filter for the MC interpolation filtering.
Based on at least one of a number of taps of a MC interpolation filter for MC interpolation filtering for inter-prediction for the enhancement layer image, a number of taps of the IL interpolation filter for IL interpolation filtering, and a size of a prediction unit of the enhancement layer image, Wherein the number of MC interpolation filtering and the number of IL interpolation filtering are limited.
Interpolation filtering for blocks of size 8x8 or greater is limited to a combination of i) two 8-tap MC interpolation filtering or ii) one 8-tap MC interpolation filtering and one 8-tap IL interpolation filtering,
I) a combination of one 8-tap IL interpolation filtering, or iv) a combination of two 6-tap IL interpolation filtering, or v) a combination of two 4-tap IL interpolation filtering, or vi) 2-tap IL interpolation filtering,
(Vii) a combination of two 8-tap MC interpolation filtering and one 4-tap IL interpolation filtering, or viii) a combination of 4 2-tap MC interpolation filtering and 4 2-tap IL interpolation filtering , Or ix) a combination of two 8-tap MC interpolation filtering and 2 2-tap IL interpolation filtering, or x) a combination of 8 8-tap MC interpolation filtering and 1 8-tap IL interpolation filtering. A method for encoding a video signal.
Obtaining a reference index indicating a residue component and a reference layer image for inter-layer prediction of an enhancement layer image;
Determining, based on the reference index, not to perform inter prediction between enhancement layer images and determining the reference layer image from among base layer images;
Generating an upsampled reference layer image by performing IL interpolation filtering on the determined reference layer image; And
And reconstructing the enhancement layer image using the residue component of the interlaced prediction and the upsampled reference layer image.
Performing four MC interpolation filtering including horizontal interpolation filtering and vertical interpolation filtering in the L0 prediction direction and the L1 prediction direction with respect to the base layer video,
Wherein the step of generating the upsampled reference layer video comprises:
Performing two IL interpolation filtering including horizontal directional IL interpolation filtering and vertical directional IL interpolation filtering on the determined reference layer image,
Wherein the interpolation filtering for inter-layer prediction of the enhancement layer image is limited to the MC interpolation filtering and the two IL interpolation filtering.
Layer interpolation for an interlayer prediction of the enhancement layer image, the sum of a first calculation amount of MC interpolation filtering for inter prediction between the base layer images and a second calculation amount of MC interpolation filtering for inter prediction between the enhancement layer images, And the amount of computation of the filtering is not large.
And determining a motion vector for inter-prediction between the enhancement layer images to be 0 if the reference index of the enhancement layer image indicates the upsampled reference image.
In the enhancement layer video, whether to perform inter prediction is determined based on the size, shape, and prediction direction of the block,
Wherein the number of taps of the IL interpolation filter for the IL interpolation filtering is limited not to be greater than the number of taps of the MC interpolation filter for the MC interpolation filtering.
Based on at least one of a number of taps of a MC interpolation filter for MC interpolation filtering for inter-prediction for the enhancement layer image, a number of taps of the IL interpolation filter for IL interpolation filtering, and a size of a prediction unit of the enhancement layer image, Wherein the number of MC interpolation filtering and the number of IL interpolation filtering are limited.
Interpolation filtering for blocks of size 8x8 or greater is limited to a combination of i) two 8-tap MC interpolation filtering or ii) one 8-tap MC interpolation filtering and one 8-tap IL interpolation filtering,
I) a combination of one 8-tap IL interpolation filtering, or iv) a combination of two 6-tap IL interpolation filtering, or v) a combination of two 4-tap IL interpolation filtering, or vi) 2-tap IL interpolation filtering,
(Vii) a combination of two 8-tap MC interpolation filtering and one 4-tap IL interpolation filtering, or viii) a combination of 4 2-tap MC interpolation filtering and 4 2-tap IL interpolation filtering , Or ix) a combination of two 8-tap MC interpolation filtering and 2 2-tap IL interpolation filtering, or x) a combination of 8 8-tap MC interpolation filtering and 1 8-tap IL interpolation filtering. A method for decoding a good video.
A base layer encoding unit for performing inter-prediction on base layer images; And
Determines an up-sampled reference layer video image by performing IL interpolation filtering on the determined reference layer video image, determines that inter-prediction between enhancement layer video images is not performed for the enhancement layer video image, And an enhancement layer encoder for encoding the residue component between the enhancement layer images.
A base layer decoding unit for performing motion compensation to restore base layer images; And
Determining whether to perform inter-prediction between enhancement layer images based on the reference index if a reference index indicating a residue component and a reference layer image for inter-layer prediction of the enhancement layer image is obtained; An enhancement layer for restoring the enhancement layer image using the residue component of the interlaced prediction and the upsampled reference layer image by performing IL interpolation filtering on the reference layer image to generate an upsampled reference layer image, And a decoding unit for decoding the encoded video data.
Applications Claiming Priority (2)
Application Number | Priority Date | Filing Date | Title |
---|---|---|---|
US201313768838A | 2013-02-25 | 2013-02-25 | |
US61/768,838 | 2013-02-25 |
Publications (1)
Publication Number | Publication Date |
---|---|
KR20140106450A true KR20140106450A (en) | 2014-09-03 |
Family
ID=51754877
Family Applications (1)
Application Number | Title | Priority Date | Filing Date |
---|---|---|---|
KR1020140022179A KR20140106450A (en) | 2013-02-25 | 2014-02-25 | Method and apparatus for scalable video encoding considering memory bandwidth and calculation complexity, method and apparatus for scalable video decoding considering memory bandwidth and calculation complexity |
Country Status (1)
Country | Link |
---|---|
KR (1) | KR20140106450A (en) |
-
2014
- 2014-02-25 KR KR1020140022179A patent/KR20140106450A/en not_active Application Discontinuation
Similar Documents
Publication | Publication Date | Title |
---|---|---|
KR101838112B1 (en) | Method and apparatus for inter-prediction and motion-compensation | |
KR101712109B1 (en) | Method and apparatus for inter layer video decoding using depth-based disparity vector, method and apparatus for inter layer video encoding using depth-based disparity vector. | |
KR102273025B1 (en) | Method and apparatus for scalable video encoding using switchable de-noising filtering, Method and apparatus for scalable video decoding using switchable de-noising filtering | |
KR20140089487A (en) | Method and apparatus for scalable video encoding using image upsampling based on phase-shift, method and apparatus for scalable video decoding using image upsampling based on phase-shift | |
KR20160132859A (en) | Method and device for configuring merge candidate list for decoding and encoding of interlayer video | |
KR20140043037A (en) | Method and apparatus for compensating sample adaptive offset for encoding inter layer prediction error | |
KR20140122196A (en) | Method and apparatus for video encoding, method and apparatus for video decoding | |
KR20140034053A (en) | Method and appratus for inter-layer encoding of prediction information in scalable video encoding based on coding units of tree structure, method and appratus for inter-layer decoding of prediction information in scalable video decoding based on coding units of tree structure | |
KR20140091493A (en) | Method and apparatus for video encoding for illumination compensation, method and apparatus for video decoding for illu mination compensation | |
KR20170023000A (en) | Method and device for transmitting prediction mode of depth image for interlayer video encoding and decoding | |
KR20140007293A (en) | Method and apparatus for multi-layer video encoding for random access, method and apparatus for multi-layer video decoding for random access | |
KR20140122195A (en) | Method and apparatus for video encoding for Determining Prediction Candidate, method and apparatus for Determining Inter Prediction Candidate | |
KR20140122202A (en) | Method and apparatus for video stream encoding according to layer ID extention, method and apparatus for video stream decoding according to layer ID extention | |
KR20130116832A (en) | Method and appratus for multiview video encoding based on coding units of tree structure, method and appratus for multiview video decoding based on coding units of tree structure | |
KR20150105264A (en) | Method and apparatus for inter layer video decoding for performing a prediction based on sub-block and method and apparatus for inter layer video encoding for performing a prediction based on sub-block | |
KR20150043217A (en) | Method and apparatus for multi-layer video encoding, method and apparatus for multi-layer video decoding | |
EP2961167A1 (en) | Device and method for scalable video encoding considering memory bandwidth and computational quantity, and device and method for scalable video decoding | |
KR20150012223A (en) | Method and apparatus for determining motion vector | |
KR20170019363A (en) | Inter-layer video encoding method for compensating for luminance difference and device therefor, and video decoding method and device therefor | |
KR20130119379A (en) | Method and apparatus for multiview video encoding using reference picture list for multiview video prediction, method and apparatus for multiview video decoding using reference picture list for multiview video prediction | |
KR101770300B1 (en) | Method and apparatus for video encoding, method and apparatus for video decoding | |
KR20150043218A (en) | Method and apparatus for multi-layer video encoding for encoding auxiliary pictures, method and apparatus for multi-layer video decoding for decoding auxiliary pictures | |
KR20150043226A (en) | Method and apparatus for depth intra encoding and method and apparatus for depth intra decoding | |
KR20150083444A (en) | Apparatus and Method for Scalable Vedio Encoding/Decoding | |
KR20150073132A (en) | Method and apparatus for video encoding for illumination compensation, method and apparatus for video decoding for residual prediction |
Legal Events
Date | Code | Title | Description |
---|---|---|---|
WITN | Withdrawal due to no request for examination |