WO2015005621A1 - 비디오 신호 처리 방법 및 장치 - Google Patents
비디오 신호 처리 방법 및 장치 Download PDFInfo
- Publication number
- WO2015005621A1 WO2015005621A1 PCT/KR2014/006043 KR2014006043W WO2015005621A1 WO 2015005621 A1 WO2015005621 A1 WO 2015005621A1 KR 2014006043 W KR2014006043 W KR 2014006043W WO 2015005621 A1 WO2015005621 A1 WO 2015005621A1
- Authority
- WO
- WIPO (PCT)
- Prior art keywords
- tile
- picture
- indicating whether
- layer
- flag indicating
- Prior art date
Links
Images
Classifications
-
- H—ELECTRICITY
- H04—ELECTRIC COMMUNICATION TECHNIQUE
- H04N—PICTORIAL COMMUNICATION, e.g. TELEVISION
- H04N19/00—Methods or arrangements for coding, decoding, compressing or decompressing digital video signals
- H04N19/10—Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using adaptive coding
- H04N19/169—Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using adaptive coding characterised by the coding unit, i.e. the structural portion or semantic portion of the video signal being the object or the subject of the adaptive coding
- H04N19/187—Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using adaptive coding characterised by the coding unit, i.e. the structural portion or semantic portion of the video signal being the object or the subject of the adaptive coding the unit being a scalable video layer
-
- H—ELECTRICITY
- H04—ELECTRIC COMMUNICATION TECHNIQUE
- H04N—PICTORIAL COMMUNICATION, e.g. TELEVISION
- H04N19/00—Methods or arrangements for coding, decoding, compressing or decompressing digital video signals
- H04N19/10—Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using adaptive coding
- H04N19/102—Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using adaptive coding characterised by the element, parameter or selection affected or controlled by the adaptive coding
- H04N19/117—Filters, e.g. for pre-processing or post-processing
-
- H—ELECTRICITY
- H04—ELECTRIC COMMUNICATION TECHNIQUE
- H04N—PICTORIAL COMMUNICATION, e.g. TELEVISION
- H04N19/00—Methods or arrangements for coding, decoding, compressing or decompressing digital video signals
- H04N19/10—Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using adaptive coding
- H04N19/134—Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using adaptive coding characterised by the element, parameter or criterion affecting or controlling the adaptive coding
- H04N19/157—Assigned coding mode, i.e. the coding mode being predefined or preselected to be further used for selection of another element or parameter
-
- H—ELECTRICITY
- H04—ELECTRIC COMMUNICATION TECHNIQUE
- H04N—PICTORIAL COMMUNICATION, e.g. TELEVISION
- H04N19/00—Methods or arrangements for coding, decoding, compressing or decompressing digital video signals
- H04N19/10—Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using adaptive coding
- H04N19/169—Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using adaptive coding characterised by the coding unit, i.e. the structural portion or semantic portion of the video signal being the object or the subject of the adaptive coding
- H04N19/17—Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using adaptive coding characterised by the coding unit, i.e. the structural portion or semantic portion of the video signal being the object or the subject of the adaptive coding the unit being an image region, e.g. an object
-
- H—ELECTRICITY
- H04—ELECTRIC COMMUNICATION TECHNIQUE
- H04N—PICTORIAL COMMUNICATION, e.g. TELEVISION
- H04N19/00—Methods or arrangements for coding, decoding, compressing or decompressing digital video signals
- H04N19/30—Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using hierarchical techniques, e.g. scalability
- H04N19/33—Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using hierarchical techniques, e.g. scalability in the spatial domain
-
- H—ELECTRICITY
- H04—ELECTRIC COMMUNICATION TECHNIQUE
- H04N—PICTORIAL COMMUNICATION, e.g. TELEVISION
- H04N19/00—Methods or arrangements for coding, decoding, compressing or decompressing digital video signals
- H04N19/44—Decoders specially adapted therefor, e.g. video decoders which are asymmetric with respect to the encoder
-
- H—ELECTRICITY
- H04—ELECTRIC COMMUNICATION TECHNIQUE
- H04N—PICTORIAL COMMUNICATION, e.g. TELEVISION
- H04N19/00—Methods or arrangements for coding, decoding, compressing or decompressing digital video signals
- H04N19/70—Methods or arrangements for coding, decoding, compressing or decompressing digital video signals characterised by syntax aspects related to video coding, e.g. related to compression standards
-
- H—ELECTRICITY
- H04—ELECTRIC COMMUNICATION TECHNIQUE
- H04N—PICTORIAL COMMUNICATION, e.g. TELEVISION
- H04N19/00—Methods or arrangements for coding, decoding, compressing or decompressing digital video signals
- H04N19/80—Details of filtering operations specially adapted for video compression, e.g. for pixel interpolation
- H04N19/82—Details of filtering operations specially adapted for video compression, e.g. for pixel interpolation involving filtering within a prediction loop
-
- H—ELECTRICITY
- H04—ELECTRIC COMMUNICATION TECHNIQUE
- H04N—PICTORIAL COMMUNICATION, e.g. TELEVISION
- H04N19/00—Methods or arrangements for coding, decoding, compressing or decompressing digital video signals
- H04N19/90—Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using coding techniques not provided for in groups H04N19/10-H04N19/85, e.g. fractals
- H04N19/96—Tree coding, e.g. quad-tree coding
Definitions
- the present invention relates to a method and apparatus for processing a video signal, and more particularly, to a video signal processing method and apparatus for encoding or decoding a video signal.
- Compression coding refers to a series of signal processing techniques for transmitting digitized information through a communication line or for storing in a form suitable for a storage medium.
- the object of compression encoding includes objects such as voice, video, text, and the like.
- a technique of performing compression encoding on an image is called video image compression.
- Compression coding on a video signal is performed by removing redundant information in consideration of spatial correlation, temporal correlation, and stochastic correlation.
- An object of the present invention is to improve the coding efficiency of a video signal.
- the present invention seeks to provide an efficient coding method for a scalable video signal.
- the video signal processing method receiving a scalable video signal comprising a base layer and an enhancement layer; Receiving a flag indicating whether a tile of the enhancement layer is used; When the flag indicating whether the tile is used indicates that a picture referring to the corresponding video parameter set includes a plurality of tiles, a flag indicating whether to apply an inter-tile loop filter and a flag indicating whether the inter-layer tile boundary is aligned are received. Doing; And decoding a picture of the enhancement layer based on a received flag indicating whether the inter-tile loop filter is applied and a flag value indicating whether the inter-layer tile boundary is aligned. Characterized in that it comprises a.
- a demultiplexer for receiving a scalable video signal including a base layer and an enhancement layer; A base layer decoder for decoding a picture of the base layer; And an enhancement layer decoder configured to decode a picture of the enhancement layer using the decoded base layer picture, wherein the enhancement layer decoder receives a flag indicating whether a tile of the enhancement layer is used, When the flag indicating whether the tile is used indicates that a picture referring to the corresponding video parameter set includes a plurality of tiles, a flag indicating whether to apply an inter-tile loop filter and a flag indicating whether the inter-layer tile boundary is aligned are received. The picture of the enhancement layer is decoded based on the received flag indicating whether the inter-tile loop filter is applied and a flag value indicating whether the inter-layer tile boundary is aligned.
- FIG. 1 is a schematic block diagram of a video signal encoder apparatus according to an embodiment of the present invention.
- FIG. 2 is a schematic block diagram of a video signal decoder device according to an embodiment of the present invention.
- FIG. 3 illustrates an example of splitting a coding unit according to an embodiment of the present invention.
- FIG. 4 illustrates an embodiment of a method for hierarchically representing the division structure of FIG. 3.
- FIG. 5 is a diagram illustrating prediction units of various sizes and shapes according to an embodiment of the present invention.
- FIG. 6 illustrates an embodiment in which one picture is divided into a plurality of slices.
- FIG. 7 illustrates an embodiment in which one picture is divided into a plurality of tiles.
- FIG. 8 is a schematic block diagram of a scalable video coding system according to an embodiment of the present invention.
- FIG 9 illustrates an example of bitstreams of two layers in which tile structures are aligned with each other.
- FIG. 10 illustrates some syntax of a bitstream in accordance with an embodiment of the present invention.
- FIG. 11 illustrates some syntax of a bitstream according to another embodiment of the present invention.
- FIG. 12 illustrates some syntax of a bitstream according to another embodiment of the present invention.
- Coding can be interpreted as encoding or decoding in some cases, and information is a term that includes values, parameters, coefficients, elements, and the like. May be interpreted otherwise, the present invention is not limited thereto.
- 'Unit' is used to refer to a basic unit of image (picture) processing or a specific position of a picture, and in some cases, may be used interchangeably with terms such as 'block', 'partition' or 'region'.
- a unit may be used as a concept including a coding unit, a prediction unit, and a transform unit.
- the encoding apparatus 100 of the present invention is largely composed of a transformer 110, a quantizer 115, an inverse quantizer 120, an inverse transformer 125, a filter 130, and a predictor ( 150 and the entropy coding unit 160.
- the transform unit 110 obtains a transform coefficient value by converting the pixel value of the received video signal.
- a discrete cosine transform DCT
- a wavelet transform may be used.
- the discrete cosine transform divides the input picture signal into blocks having a predetermined size to perform the transform.
- the coding efficiency may vary depending on the distribution and the characteristics of the values in the transform domain.
- the quantization unit 115 quantizes the transform coefficient value output from the transform unit 110.
- the inverse quantization unit 120 inverse quantizes the transform coefficient value, and the inverse transform unit 125 restores the original pixel value by using the inverse quantized transform coefficient value.
- the filtering unit 130 performs a filtering operation for improving the quality of the reconstructed picture.
- a deblocking filter and an adaptive loop filter may be included.
- the filtered picture is output or stored in a decoded picture buffer 156 for use as a reference picture.
- the prediction unit 150 predicts the picture using a region already coded, and adds a residual value between the original picture and the predicted picture to the predicted picture to improve coding efficiency.
- the method of obtaining is used.
- the intra predictor 152 performs intra prediction within the current picture, and the inter predictor 154 predicts the current picture using a reference picture stored in the decoded picture buffer 156.
- the intra prediction unit 152 performs intra prediction from the reconstructed regions in the current picture, and transmits intra coding information to the entropy coding unit 160.
- the inter predictor 154 may further include a motion estimator 154a and a motion compensator 154b.
- the motion estimation unit 154a obtains a motion vector value of the current region by referring to the restored specific region.
- the motion estimator 154a transmits the position information (reference frame, motion vector, etc.) of the reference region to the entropy coding unit 160 to be included in the bitstream.
- the motion compensation unit 154b performs inter-screen motion compensation by using the motion vector value transmitted from the motion estimation unit 154a.
- the entropy coding unit 160 entropy codes the quantized transform coefficients, inter picture encoding information, intra picture encoding information, and reference region information input from the inter prediction unit 154 to generate a video signal bitstream.
- a variable length coding (VLC) scheme may be used.
- the variable length coding (VLC) scheme converts input symbols into consecutive codewords, which may have a variable length. For example, frequently occurring symbols are represented by short codewords and infrequently occurring symbols by long codewords.
- a context-based adaptive variable length coding (CAVLC) method may be used as a variable length coding method.
- Arithmetic coding converts consecutive data symbols into a single prime number, which can obtain the optimal fractional bits needed to represent each symbol.
- Context-based Adaptive Binary Arithmetic Code (CABAC) may be used as the arithmetic coding.
- CABAC Context-based Adaptive Binary Arithmetic Code
- the generated bitstream is encapsulated in a NAL unit.
- the NAL unit includes a coded slice segment, which consists of an integer number of coding tree units.
- the bitstream In order to decode the bitstream in the video decoder, the bitstream must first be divided into NAL unit units, and then each separated NAL unit must be decoded.
- the decoding apparatus 200 of the present invention largely includes an entropy decoding unit 210, an inverse quantization unit 220, an inverse transform unit 225, a filtering unit 230, and a prediction unit 250.
- the entropy decoding unit 210 entropy decodes the video signal bitstream and extracts transform coefficients, motion vectors, and the like for each region.
- the inverse quantization unit 220 inverse quantizes the entropy decoded transform coefficient, and the inverse transform unit 225 restores the original pixel value by using the inverse quantized transform coefficient.
- the filtering unit 230 performs filtering on the picture to improve the picture quality. This may include a deblocking filter to reduce block distortion and / or an adaptive loop filter to remove distortion of the entire picture.
- the filtered picture is output or stored in a decoded picture buffer (256) for use as a reference picture for the next frame.
- the predictor 250 of the present invention includes an intra predictor 252 and an inter predictor 254, and includes a coding type decoded by the entropy decoder 210 described above, a transform coefficient for each region, The prediction picture is reconstructed using information such as a motion vector.
- the intra prediction unit 252 performs the intra prediction from the decoded samples in the current picture.
- the inter prediction unit 254 generates the predictive picture using the reference picture and the motion vector stored in the decoded picture buffer 256.
- the inter predictor 254 may again include a motion estimator 254a and a motion compensator 254b.
- the motion estimator 254a obtains a motion vector indicating the positional relationship between the current block and the reference block of the reference picture used for coding and transfers the motion vector to the motion compensator 254b.
- the predicted value output from the intra predictor 252 or the inter predictor 254 and the pixel value output from the inverse transform unit 225 are added to generate a reconstructed video frame.
- a coding unit is a process such as intra / inter prediction, transform, quantization and / or entropy coding in the processing of the video signal described above.
- the size of the coding unit used in coding one picture may not be constant.
- the coding unit may have a rectangular shape, and one coding unit may be further divided into several coding units.
- FIG. 3 illustrates an example of splitting a coding unit according to an embodiment of the present invention.
- one coding unit having a size of 2N ⁇ 2N may be divided into four coding units having a size of N ⁇ N.
- the splitting of such coding units can be done recursively, and not all coding units need to be split in the same form.
- FIG. 4 illustrates an embodiment of a method for hierarchically representing a division structure of a coding unit illustrated in FIG. 3 using a flag value.
- Information indicating whether a coding unit is divided may be allocated to a value of '1' when the corresponding unit is divided and '0' when it is not divided.
- a flag value indicating whether to split is 1, a coding unit corresponding to a corresponding node is divided into 4 coding units, and when 0, a processing process for the coding unit is performed without being divided further. Can be.
- the structure of the coding unit described above may be represented using a recursive tree structure. That is, a coding unit split into another coding unit with one picture or maximum size coding unit as a root has as many child nodes as the number of split coding units. Thus, coding units that are no longer split become leaf nodes. Assuming that only square division is possible for one coding unit, one coding unit may be divided into up to four other coding units, so the tree representing the coding unit may be in the form of a quad tree.
- an optimal coding unit size is selected according to characteristics (eg, resolution) of a video picture or in consideration of coding efficiency, and information about the same or information capable of deriving the coding unit may be included in the bitstream.
- characteristics eg, resolution
- information about the same or information capable of deriving the coding unit may be included in the bitstream.
- the size of the largest coding unit and the maximum depth of the tree can be defined.
- the minimum coding unit size can be obtained using the above information.
- the minimum coding unit size and the maximum depth of the tree may be defined in advance, and the maximum coding unit size may be derived and used. Since the size of the unit changes in a multiple of 2 in square division, the size of the actual coding unit is represented by a logarithm of the base of 2, thereby improving transmission efficiency.
- the decoder may obtain information indicating whether the current coding unit is split. This information can be obtained only if the information is obtained (transmitted) under certain conditions, thereby increasing efficiency. For example, the condition that the current coding unit can be divided is that the current coding unit is divided by the current coding unit size at the current position is smaller than the picture size and the current unit size is larger than the preset minimum coding unit size. Information indicating whether or not it can be obtained.
- the size of the coding unit to be split is half of the current coding unit, and is split into four square coding units based on the current processing position. The above process can be repeated for each divided coding units.
- Picture prediction (motion compensation) for coding is directed to coding units (i.e. leaf nodes of the coding unit tree) that are no longer divided.
- the basic unit for performing such prediction is hereinafter referred to as a prediction unit or a prediction block.
- the prediction unit may have a form of square, rectangle, or the like within the coding unit.
- one prediction unit is not split (2N X 2N), or NXN, 2N XN, NX 2N, 2N XN / 2, 2N X 3N / 2, N / 2 X 2N, 3N as shown in FIG. / 2 X 2N and the like can be divided into various sizes and shapes.
- the possible division forms of the prediction unit may be defined differently in the intra coding unit and the inter coding unit.
- the bitstream may include information about whether the prediction unit is divided or in what form. Or this information may be derived from other information.
- the term unit used in the present specification may be used as a term to replace the prediction unit, which is a basic unit for performing prediction.
- the present invention is not limited thereto and may be broadly understood as a concept including the coding unit.
- the decoded portion of the current picture or other pictures in which the current unit is included may be used to reconstruct the current unit in which decoding is performed.
- a picture (slice) that uses up to one motion vector and a reference index to predict each unit of an inter picture (slice) is called a predictive picture or P picture (slice), and up to two motion vectors and a reference index
- a picture using a slice is called a bi-predictive picture or a B picture.
- the intra prediction unit performs intra prediction to predict the pixel value of the target unit from the reconstructed regions in the current picture.
- the pixel value of the current unit can be predicted from the encoded pixels of units located at the top, left, top left and / or top right with respect to the current unit.
- the inter prediction unit performs inter prediction for predicting a pixel value of a target unit by using information of other reconstructed pictures other than the current picture.
- a picture used for prediction is referred to as a reference picture.
- Which reference region is used to predict the current unit in the inter prediction process may be indicated by using an index indicating a reference picture including the reference region, motion vector information, and the like.
- the inter prediction may include forward direction prediction, backward direction prediction, and bi-prediction.
- Forward prediction is prediction using one reference picture displayed (or output) before the current picture in time
- backward prediction means prediction using one reference picture displayed (or output) after the current picture in time.
- one set of motion information eg, a motion vector and a reference picture index
- up to two reference regions may be used.
- the two reference regions may exist in the same reference picture or may exist in different pictures, respectively. That is, up to two sets of motion information (eg, a motion vector and a reference picture index) may be used in the bi-prediction method, wherein two motion vectors may have the same reference picture index or may have different reference picture indexes. It may be.
- the reference pictures may be displayed (or output) before or after the current picture in time.
- the reference unit of the current unit may be obtained using the motion vector and the reference picture index.
- the reference unit exists in a reference picture having the reference picture index.
- a pixel value or an interpolated value of a unit specified by the motion vector may be used as a predictor of the current unit.
- an 8-tap interpolation filter may be used for the luminance signal and a 4-tap interpolation filter may be used for the chrominance signal.
- motion compensation is performed to predict the texture of the current unit from a previously decoded picture.
- a reference picture list may be configured of pictures used for inter prediction for the current picture.
- two reference picture lists are required, and in the following, each of them is referred to as reference picture list 0 (or L0) and reference picture list 1 (or L1).
- One picture may be divided into slices, slice segments, tiles, and the like.
- a slice indicates a horizontal division of a picture
- a tile indicates a horizontal division and / or a vertical division of the picture.
- 6 and 7 illustrate various embodiments in which a picture is divided.
- FIG. 6 illustrates an embodiment in which one picture is divided into a plurality of slices (Slice 0 and Slice 1).
- the thick line represents the slice boundary and the dotted line represents the slice segment boundary.
- a slice may consist of one independent slice segment or a set of at least one dependent slice segment contiguous with one independent slice segment.
- a slice segment is a sequence of coding tree units (CTUs) 30. That is, an independent or dependent slice segment consists of at least one CTU 30.
- CTUs coding tree units
- slice 0 is composed of a total of three slice segments, which are composed of an independent slice segment including four CTUs, a dependent slice segment including 35 CTUs, and another dependent slice segment including 15 CTUs.
- slice 1 is composed of one independent slice segment including 42 CTUs.
- FIG. 7 illustrates an embodiment in which one picture is divided into a plurality of tiles (Tile 0 and Tile 1).
- the thick line represents the tile boundary and the dotted line represents the slice segment boundary.
- a tile is a sequence of CTUs 30, like slices, and has a rectangular shape.
- one picture is divided into two tiles, that is, tile 0 and tile 1.
- the picture consists of one slice, and includes one independent slice segment and four consecutive slice segments in succession.
- one tile may be divided into a plurality of slices. That is, one tile may be composed of CTUs included in one or more slices.
- one slice may be composed of CTUs included in one or more tiles.
- each slice and tile must satisfy at least one of the following conditions. i) All CTUs included in one slice belong to the same tile. ii) All CTUs included in one tile belong to the same slice.
- one picture may be divided into slices and / or tiles, and each partition (slice, tile) may be encoded or decoded in parallel.
- FIG. 8 shows a schematic block diagram of a scalable video coding (or scalable high efficiency video coding) system according to an embodiment of the invention.
- the scalable video coding scheme is a compression method for providing video content hierarchically in terms of spatial, temporal and / or image quality according to various user environments such as network conditions or terminal resolutions in various multimedia environments. Spatial scalability may be supported by encoding the same picture with different resolutions for each layer, and temporal hierarchicality may be implemented by adjusting the screen refresh rate per second of the picture.
- the quality hierarchical structure may provide pictures of various image quality by encoding different quantization parameters for each layer. In this case, a picture sequence having a low resolution, frames per second and / or a low quality is called a base layer, and a picture sequence having a relatively high resolution, frames per second and / or a high quality is called an enhancement layer.
- the scalable video coding system includes an encoding device 300 and a decoding device 400.
- the encoding apparatus 300 includes a base layer encoding unit 100a, an enhancement layer encoding unit 100b, and a multiplexer 180
- the decoding apparatus 400 includes a demultiplexer 280 and a base layer decoding unit 200a.
- an enhancement layer decoder 200b may generate a base bitstream by compressing the input signal X (n).
- the enhancement layer encoding unit 100b may generate an enhancement layer bitstream using the information generated by the input signal X (n) and the base layer encoding unit 100a.
- the multiplexer 180 generates a scalable bitstream using the base layer bitstream and the enhancement layer bitstream.
- the basic configuration of the base layer encoding unit 100a and the enhancement layer encoding unit 100b may be the same as or similar to that of the encoding apparatus 100 illustrated in FIG. 1.
- the inter prediction unit of the enhancement layer encoding unit 100b may perform inter prediction using the motion information generated by the base layer encoding unit 100a.
- the decoded picture buffer DPB of the enhancement layer encoder 100b may sample and store a picture stored in the decoded picture buffer DPB of the base layer encoder 100a. The sampling may include resampling, upsampling, and the like as described below.
- the scalable bitstream generated as described above is transmitted to the decoding device 400 through a predetermined channel, and the transmitted scalable bitstream is enhanced by the demultiplexer 280 of the decoding device 400 and the base layer bitstream. It can be divided into.
- the base layer decoder 200a receives the base layer bitstream, restores the base layer bitstream, and generates an output signal Xb (n).
- the enhancement layer decoding unit 200b receives the enhancement layer bitstream and generates an output signal Xe (n) with reference to the signal reconstructed by the base layer decoding unit 200a.
- Basic configurations of the base layer decoder 200a and the enhancement layer decoder 200b may be the same as or similar to those of the decoder 200 shown in FIG. 2.
- the inter predictor of the enhancement layer decoder 200b may perform inter prediction using the motion information generated by the base layer decoder 200a.
- the decoded picture buffer DPB of the enhancement layer decoder 200b may sample and store a picture stored in the decoded picture buffer DPB of the base layer decoder 200a. The sampling may include resampling, upsampling, and the like.
- Inter-layer prediction means predicting a picture signal of an upper layer using motion information, syntax information, and / or texture information of a lower layer.
- the lower layer referred to for encoding the upper layer may be referred to as a reference layer.
- the enhancement layer may be coded using the base layer as a reference layer.
- the reference unit of the base layer may be enlarged or reduced through sampling.
- Sampling may mean changing image resolution or quality.
- the sampling may include re-sampling, down-sampling, up-sampling, and the like.
- intra samples may be resampled to perform inter-layer prediction.
- image resolution may be reduced by regenerating pixel data using a down sampling filter, which is called down sampling.
- image resolution can be increased by creating additional pixel data using an upsampling filter, called upsampling.
- upsampling an upsampling filter
- the single loop method decodes only the picture of the layer to be actually reproduced, and the lower layer does not decode the other pictures except the intra unit. Therefore, in the enhancement layer, motion vectors, syntax information, and the like of the lower layer can be referred to, but texture information of units other than the intra unit cannot be referred to.
- the multi-loop method is a method of restoring not only the current layer but also all lower layers thereof. Therefore, when the multi-loop method is used, all texture information may be referred to as well as syntax information of a lower layer.
- a reference layer includes a direct reference layer and an indirect reference layer.
- the direct reference layer refers to a layer used directly for inter-layer prediction of another upper layer.
- the indirect reference layer refers to a layer that is not directly used for inter-layer prediction of another upper layer but indirectly. That is, the indirect reference layer includes a direct or indirect reference layer with respect to the direct reference layer of the upper layer.
- one layer and a reference layer (ie, an enhancement layer and a base layer) of the corresponding layer may be divided into a plurality of slices and / or a plurality of tiles.
- each slice and tile consists of a set of CTUs having the same size.
- the term “partition” may be used as a concept including both a slice and a tile for dividing a picture. In this case, a situation in which partition structures (or boundaries) of pictures in a plurality of layers or all layers are aligned with each other may often occur.
- FIG. 9 shows an example of a bitstream of two layers in which tile structures are aligned with each other.
- Each picture shown in FIG. 9 shows a base layer picture 40a, an up-sampled base layer picture 40b, and an enhancement layer picture 40c from below.
- Each picture is divided into four tiles.
- the tile structure between layers may be said to be aligned. Therefore, in order to align the tile structure of two or more layers, the following conditions must be satisfied.
- Tiles of the same position in each layer may be allocated to the same processing core of the decoder and processed. For example, assuming that the decoder has four processing cores, tiles having the same identifier in the base layer picture 40a and the enhancement layer picture 40c of FIG. 9 may be processed by the same core. Therefore, processing may be sequentially performed from tiles located in lower layers (ie, base layers).
- the tile structure of the inter-layer pictures is the same, it is not necessary to repeatedly transmit information on the tile structure of the upper layer to the decoder. This is because the tile structure of the upper layer is simply predictable from the tile structure of the nearest dependent layer or reference layer of the layer.
- various information related to the tile structure between the layers may be included in the bitstream.
- the information may be included in a video parameter set (VPS), a sequence parameter set (SPS), a picture parameter set (PPS), or extension data thereof, or may be included in Supplemental Enhancement Information (SEI).
- SEI Supplemental Enhancement Information
- the information may be included in the utilization information (VPS_VUI) data of the video parameter set.
- FIG. 10 illustrates some syntax of a bitstream according to an embodiment of the present invention. The syntax of FIG. 10 will now be described in detail.
- the decoding apparatus may receive a 'tiles_enabled_flag' indicating whether a tile is used.
- 'Tiles indicating whether tiles are used' indicates whether the associated picture includes a plurality of tiles, i) a PPS, its extended data (ie PPS level), and / or ii) a VPS, its extended Data, VPS_VUI (ie, VPS level) and the like.
- the flag may indicate whether each picture referring to the corresponding PPS includes a plurality of tiles.
- each picture referring to the corresponding PPS may include a plurality of tiles. However, when the flag value is 0, each picture referring to the corresponding PPS may include only one tile.
- the flag may indicate whether each picture referring to the corresponding VPS includes a plurality of tiles. According to an embodiment, the flag may indicate hierarchically whether the related picture includes a plurality of tiles. For example, the flag indicating whether the tile is used may be included in the PPS level and the VPS level, respectively. The flag indicating whether the tile is used for the VPS level may indicate a value of the flag indicating whether the tile is used for each PPS referred to by at least one picture that refers to the VPS.
- the decoding apparatus may include a 'flag indicating whether the tile boundaries between layers are aligned' ('tiles_structure_aligned_flag') and 'flag indicating whether the tile loop filter is applied'. ('loop_filter_across_tiles_enabled_flag') may be received.
- a flag indicating whether or not the inter-layer tile boundary is aligned indicates whether the tile boundary (or structure) of the current layer pictures matches the tile boundary (or structure) of the nearest direct reference layer pictures. Indicates. Specifically, when the 'tiles_structure_aligned_flag' value is 1, if any two samples of the current layer picture belong to the same tile, the samples of juxtaposition of the corresponding reference layer picture should also belong to the same tile. Also, if any two samples of the current layer picture belong to different tiles, the samples of each juxtaposition of the corresponding reference layer picture should belong to different tiles. If the 'tiles_structure_aligned_flag' value is 0, this restriction may not apply. If the 'tiles_structure_aligned_flag' information does not exist in the bitstream, the corresponding flag may be inferred to zero.
- a non-base layer having a layer identifier greater than zero may have one or more direct reference layers.
- 'tiles_structre_aligned_flag' may indicate whether the tile boundary between the specific upper layer and each direct reference layer corresponding thereto is aligned. Specific embodiments thereof will be described later.
- the decoding apparatus may determine 'tile column count information' ('num_tile_columns_minus1'), 'tile row count information' ('num_tile_rows_minus1'), and 'flag indicating whether the tile boundary is uniformly spaced' (' uniform_spacing_flag ') may be received.
- 'Tile column number information' ('num_tile_columns_minus1') and 'tile row number information' ('num_tile_rows_minus1') indicate the number of columns and rows of tiles for dividing pictures, respectively.
- the 'num_tile_columns_minus1' and 'num_tile_rows_minus1' may be inferred to have the same values as the corresponding values by referring to 'num_tile_columns_minus1' and 'num_tile_rows_minus1' of the reference layer, respectively.
- a 'flag indicating whether the tile boundary is uniform' indicates whether the column and column spacings of the tile boundary are uniform throughout the picture.
- the flag value indicates that the column spacing and the row spacing of the tile boundary are uniformly distributed.
- the flag value is 0, it indicates that the column and / or row spacing of the tile boundary is not uniformly distributed, and the width of each column and the height of each row are 'width information of the tile column' (column_width_minus1 [i]) and It may be directly signaled through 'height information of the tile row' (row_heignt_minus1 [i]).
- 'Tile column width information' ('column_width_minus1 [i]') represents the width of the i-th tile column
- 'tile row height information' ('row_heignt_minus1 [i]') represents the height of the i-th tile row.
- the values of 'column_width_minus1 [i]' and 'row_heignt_minus1 [i]' may be expressed in CTU units.
- 'column_width_minus1 [i]' and 'row_heignt_minus1 [i]' may be inferred from information of the direct reference layer, respectively. That is, the current 'column_width_minus1 [i]' may be obtained by scaling the picture width ratio between the current layer and the direct reference layer to the 'column_width_minus1 [i]' value of the direct reference layer. Similarly, the current 'row_heignt_minus1 [i]' may be obtained by scaling a picture height ratio between the current layer and the direct reference layer to the 'row_heignt_minus1 [i]' value of the direct reference layer.
- 'flag indicating whether the loop filter is applied to the tile' indicates whether in-loop filtering across a tile boundary may be performed in a related picture.
- An in-loop filter is a filter that is applied to a reconstructed picture to produce a picture output to a playback device and inserted into a decoded picture buffer.
- the in-loop filter includes a deblocking filter and a Sample Adaptive Offset (SAO) filter.
- 'loop_filter_across_tiles_enabled_flag' may be included in i) PPS, its extended data (ie, PPS level), and / or ii) VPS, its extended data, VPS_VUI (ie, VPS level), etc., as described above.
- the flag may indicate whether in-loop filtering across a tile boundary may be performed in pictures referring to the PPS. If the flag value is 1, in-loop filtering may be performed across the tile boundary in pictures referring to the corresponding PPS.
- the flag when the flag value is 0, in-loop filtering that crosses a tile boundary in pictures referring to the PPS may not be performed.
- the flag when 'loop_filter_across_tiles_enabled_flag' is included in the VPS level, the flag may indicate whether in-loop filtering across a tile boundary may be performed in pictures referring to the VPS.
- the flag may indicate hierarchically whether in-loop filtering across a tile boundary can be performed, for example, a 'flag indicating whether a tile loop filter is applied to the PPS level and the VPS level', respectively. May be included.
- the 'flag indicating whether the loop loop filter is applied to the VPS level' may indicate a value of the 'flag indicating whether the loop loop filter is applied to each PPS' referenced by at least one picture referring to the corresponding VPS.
- the decoding apparatus of the present invention may perform in-loop filtering based on the flag value obtained by the above-described method. Meanwhile, in the embodiment of FIG. 10 and the embodiment described below, the information indicated by the flag values 0 and 1 may be defined in reverse. That is, the value represented by each flag in the present invention is only an embodiment for describing the present invention, and the present invention is not limited thereto.
- FIG. 11 illustrates some syntax of a bitstream according to another embodiment of the present invention.
- the same or corresponding parts as those of the embodiment of FIG. 10 will be omitted.
- 'a flag indicating whether a tile filter is applied to a tile' may be received when the 'flag indicating whether the tile boundaries are aligned between layers' ('tiles_structure_aligned_flag') is 0. have. That is, the flag may be received only when the tile boundary of the current layer picture and the direct reference layer picture do not match.
- the value of 'loop_filter_across_tiles_enabled_flag' may be set to be the same as the 'loop_filter_across_tiles_enabled_flag' value of the nearest direct reference layer.
- FIG. 12 is a diagram illustrating some syntax of a bitstream according to another embodiment of the present invention.
- the same or corresponding parts as those of the embodiment of FIG. 10 will be omitted.
- the decoding apparatus of the present invention receives a 'flag indicating whether tiles are used' ('tiles_enabled_flag') and 'flags indicating whether to apply the tile loop filter' based on the flag value ('loop_filter_across_tiles_enabled_flag'). And 'flags indicating whether tile boundaries are aligned' ('tiles_structure_aligned_flag [i] [j]').
- the decoding apparatus may receive 'loop_filter_across_tiles_enabled_flag' and 'tiles_structure_aligned_flag [i] [j]'.
- 'flag_filter_across_tiles_enabled_flag' indicating whether to apply a loop filter on a tile day indicates whether in-loop filtering across a tile boundary may be performed on a related picture.
- the flag may be included in the PPS level and / or the VPS level, and when included in both levels, the flags may indicate hierarchically whether in-loop filtering is performed across a tile boundary. Specific embodiments thereof are as described with reference to FIG. 10.
- the decoding apparatus receives a 'flag indicating whether the tile boundary is aligned' ('tiles_structure_aligned_flag [i] [j]').
- a non-base layer having a layer identifier greater than zero may have one or more direct reference layers.
- 'tiles_structre_aligned_flag [i] [j]' may indicate whether the tile boundary between the specific upper layer (or an enhancement layer) and each of the direct reference layers corresponding thereto is aligned.
- the number of direct reference layers of that layer ('num_direct_ref_layers [layer_id [i]]' ) Directly obtains a layer identifier ('layerIdx') of the reference layer.
- 'flags indicating whether tiles are used' in each layer (ie, 'tiles_enabled_flag [i]', Check 'tiles_enabled_flag [layerIdx]').
- the decoding apparatus may indicate a flag indicating whether the tile boundary is aligned between the upper layer and the corresponding direct reference layer ('tiles_structure_aligned_flag [i'). ] [j] ').
- 'tiles_structure_aligned_flag [i] [j]' indicates whether the tile boundary is aligned between the enhancement layer having the layer identifier 'i' and the 'j' th direct reference layer of the corresponding layer.
- the present invention can be applied to process and output video signals.
Landscapes
- Engineering & Computer Science (AREA)
- Multimedia (AREA)
- Signal Processing (AREA)
- Compression Or Coding Systems Of Tv Signals (AREA)
Abstract
Description
Claims (5)
- 비디오 신호 처리 방법에 있어서,베이스 레이어와 인핸스먼트 레이어를 포함하는 스케일러블 비디오 신호를 수신하는 단계;상기 인핸스먼트 레이어의 타일 사용 여부를 나타내는 플래그를 수신하는 단계;상기 타일 사용 여부를 나타내는 플래그가 해당 비디오 파라미터 셋을 참조하는 픽쳐가 복수의 타일을 포함함을 나타낼 경우, 타일간 루프 필터 적용 여부를 나타내는 플래그 및 레이어간 타일 경계의 얼라인 여부를 나타내는 플래그를 수신하는 단계; 및수신된 상기 타일간 루프 필터 적용 여부를 나타내는 플래그 및 레이어간 타일 경계의 얼라인 여부를 나타내는 플래그 값에 기초하여 상기 인핸스먼트 레이어의 픽쳐를 디코딩하는 단계;를 포함하는 것을 특징으로 하는 비디오 신호 처리 방법.
- 제1 항에 있어서,상기 레이어간 타일 경계의 얼라인 여부를 나타내는 플래그는,상기 인핸스먼트 레이어 및 해당 인핸스먼트 레이어의 적어도 하나의 직접 참조 레이어 간의 타일 경계가 얼라인 되어 있는지 여부를 나타내는 것을 특징으로 하는 비디오 신호 처리 방법.
- 제2 항에 있어서,상기 레이어간 타일 경계의 얼라인 여부를 나타내는 플래그는,상기 인핸스먼트 레이어의 타일 사용 여부를 나타내는 플래그 및 상기 직접 참조 레이어의 타일 사용 여부를 나타내는 플래그가 모두 복수의 타일을 사용함을 나타낼 경우 수신되는 것을 특징으로 하는 비디오 신호 처리 방법.
- 제1 항에 있어서,상기 타일간 루프 필터 적용 여부를 나타내는 플래그는 해당 비디오 파라미터 셋을 참조하는 픽쳐들에서 타일 경계를 가로지르는 인-루프 필터링이 수행될 수 있는지 여부를 나타내는 것을 특징으로 하는 비디오 신호 처리 방법.
- 베이스 레이어와 인핸스먼트 레이어를 포함하는 스케일러블 비디오 신호를 수신하는 디멀티 플렉서;상기 베이스 레이어의 픽쳐를 디코딩 하는 베이스 레이어 디코더; 및상기 디코딩 된 베이스 레이어의 픽쳐를 이용하여 상기 인핸스먼트 레이어의 픽쳐를 디코딩 하는 인핸스먼트 레이어 디코더를 포함하되,상기 인핸스먼트 레이어 디코더는, 상기 인핸스먼트 레이어의 타일 사용 여부를 나타내는 플래그를 수신하고, 상기 타일 사용 여부를 나타내는 플래그가 해당 비디오 파라미터 셋을 참조하는 픽쳐가 복수의 타일을 포함함을 나타낼 경우, 타일간 루프 필터 적용 여부를 나타내는 플래그 및 레이어간 타일 경계의 얼라인 여부를 나타내는 플래그를 수신하고, 수신된 상기 타일간 루프 필터 적용 여부를 나타내는 플래그 및 레이어간 타일 경계의 얼라인 여부를 나타내는 플래그 값에 기초하여 상기 인핸스먼트 레이어의 픽쳐를 디코딩 하는 것을 특징으로 하는 비디오 신호 처리 장치.
Priority Applications (5)
Application Number | Priority Date | Filing Date | Title |
---|---|---|---|
CN201480030183.XA CN105284114A (zh) | 2013-07-07 | 2014-07-07 | 视频信号处理方法及装置 |
CA2909601A CA2909601A1 (en) | 2013-07-07 | 2014-07-07 | Method and apparatus for processing video signal |
US14/784,954 US20160080753A1 (en) | 2013-07-07 | 2014-07-07 | Method and apparatus for processing video signal |
KR1020157029984A KR20160002786A (ko) | 2013-07-07 | 2014-07-07 | 비디오 신호 처리 방법 및 장치 |
EP14823575.7A EP3021586A4 (en) | 2013-07-07 | 2014-07-07 | METHOD AND APPARATUS FOR PROCESSING VIDEO SIGNAL |
Applications Claiming Priority (2)
Application Number | Priority Date | Filing Date | Title |
---|---|---|---|
US201361843424P | 2013-07-07 | 2013-07-07 | |
US61/843,424 | 2013-07-07 |
Publications (1)
Publication Number | Publication Date |
---|---|
WO2015005621A1 true WO2015005621A1 (ko) | 2015-01-15 |
Family
ID=52280241
Family Applications (2)
Application Number | Title | Priority Date | Filing Date |
---|---|---|---|
PCT/KR2014/006044 WO2015005622A1 (ko) | 2013-07-07 | 2014-07-07 | 비디오 신호 처리 방법 및 장치 |
PCT/KR2014/006043 WO2015005621A1 (ko) | 2013-07-07 | 2014-07-07 | 비디오 신호 처리 방법 및 장치 |
Family Applications Before (1)
Application Number | Title | Priority Date | Filing Date |
---|---|---|---|
PCT/KR2014/006044 WO2015005622A1 (ko) | 2013-07-07 | 2014-07-07 | 비디오 신호 처리 방법 및 장치 |
Country Status (6)
Country | Link |
---|---|
US (1) | US20160080753A1 (ko) |
EP (1) | EP3021586A4 (ko) |
KR (1) | KR20160002786A (ko) |
CN (1) | CN105284114A (ko) |
CA (1) | CA2909601A1 (ko) |
WO (2) | WO2015005622A1 (ko) |
Cited By (1)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
US11762136B2 (en) | 2011-03-03 | 2023-09-19 | Enchroma, Inc. | Multi-band color vision filters and method by LP-optimization |
Families Citing this family (17)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
JP6261215B2 (ja) * | 2013-07-12 | 2018-01-17 | キヤノン株式会社 | 画像符号化装置、画像符号化方法及びプログラム、画像復号装置、画像復号方法及びプログラム |
KR20200045012A (ko) * | 2013-07-15 | 2020-04-29 | 소니 주식회사 | 상호작용성을 위한 모션-구속된 타일 세트들 sei 메시지의 확장들 |
WO2015037922A1 (ko) * | 2013-09-10 | 2015-03-19 | 주식회사 케이티 | 스케일러블 비디오 신호 인코딩/디코딩 방법 및 장치 |
US10091519B2 (en) | 2013-10-14 | 2018-10-02 | Electronics And Telecommunications Research Institute | Multilayer-based image encoding/decoding method and apparatus |
US10045036B2 (en) * | 2013-10-22 | 2018-08-07 | Kt Corporation | Method and device for encoding/decoding multi-layer video signal |
KR20150046742A (ko) * | 2013-10-22 | 2015-04-30 | 주식회사 케이티 | 멀티 레이어 비디오 신호 인코딩/디코딩 방법 및 장치 |
US9967576B2 (en) * | 2013-10-29 | 2018-05-08 | Kt Corporation | Multilayer video signal encoding/decoding method and device |
WO2016188447A1 (en) * | 2015-05-28 | 2016-12-01 | Hfi Innovation Inc. | Method and apparatus for using a current picture as a reference picture |
US10999602B2 (en) | 2016-12-23 | 2021-05-04 | Apple Inc. | Sphere projected motion estimation/compensation and mode decision |
US11259046B2 (en) | 2017-02-15 | 2022-02-22 | Apple Inc. | Processing of equirectangular object data to compensate for distortion by spherical projections |
US10924747B2 (en) | 2017-02-27 | 2021-02-16 | Apple Inc. | Video coding techniques for multi-view video |
US11093752B2 (en) | 2017-06-02 | 2021-08-17 | Apple Inc. | Object tracking in multi-view video |
US10754242B2 (en) | 2017-06-30 | 2020-08-25 | Apple Inc. | Adaptive resolution and projection format in multi-direction video |
US20190005709A1 (en) * | 2017-06-30 | 2019-01-03 | Apple Inc. | Techniques for Correction of Visual Artifacts in Multi-View Images |
PT3931748T (pt) * | 2019-03-11 | 2024-06-27 | Huawei Tech Co Ltd | Endereços de parcela baseados em subgravura na codificação vídeo |
GB201913403D0 (en) * | 2019-09-17 | 2019-10-30 | Canon Kk | Method and apparatus for encoding and decoding a video stream with subpictures |
WO2021185311A1 (en) * | 2020-03-20 | 2021-09-23 | Mediatek Inc. | Method and apparatus for signaling tile and slice partition information in image and video coding |
Citations (2)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
US20130114735A1 (en) * | 2011-11-04 | 2013-05-09 | Qualcomm Incorporated | Video coding with network abstraction layer units that include multiple encoded picture partitions |
US20130163660A1 (en) * | 2011-07-01 | 2013-06-27 | Vidyo Inc. | Loop Filter Techniques for Cross-Layer prediction |
Family Cites Families (6)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
KR100878812B1 (ko) * | 2005-05-26 | 2009-01-14 | 엘지전자 주식회사 | 영상신호의 레이어간 예측에 대한 정보를 제공하고 그정보를 이용하는 방법 |
KR101029854B1 (ko) * | 2006-01-11 | 2011-04-15 | 노키아 코포레이션 | 스케일러블 비디오 코딩에서 픽쳐들의 역방향-호환 집합 |
US8699583B2 (en) * | 2006-07-11 | 2014-04-15 | Nokia Corporation | Scalable video coding and decoding |
US9819949B2 (en) * | 2011-12-16 | 2017-11-14 | Microsoft Technology Licensing, Llc | Hardware-accelerated decoding of scalable video bitstreams |
US9294776B2 (en) * | 2013-03-05 | 2016-03-22 | Qualcomm Incorporated | Parallel processing for video coding |
US20140254681A1 (en) * | 2013-03-08 | 2014-09-11 | Nokia Corporation | Apparatus, a method and a computer program for video coding and decoding |
-
2014
- 2014-07-07 CN CN201480030183.XA patent/CN105284114A/zh active Pending
- 2014-07-07 KR KR1020157029984A patent/KR20160002786A/ko not_active Application Discontinuation
- 2014-07-07 CA CA2909601A patent/CA2909601A1/en not_active Abandoned
- 2014-07-07 US US14/784,954 patent/US20160080753A1/en not_active Abandoned
- 2014-07-07 WO PCT/KR2014/006044 patent/WO2015005622A1/ko active Application Filing
- 2014-07-07 EP EP14823575.7A patent/EP3021586A4/en not_active Withdrawn
- 2014-07-07 WO PCT/KR2014/006043 patent/WO2015005621A1/ko active Application Filing
Patent Citations (2)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
US20130163660A1 (en) * | 2011-07-01 | 2013-06-27 | Vidyo Inc. | Loop Filter Techniques for Cross-Layer prediction |
US20130114735A1 (en) * | 2011-11-04 | 2013-05-09 | Qualcomm Incorporated | Video coding with network abstraction layer units that include multiple encoded picture partitions |
Non-Patent Citations (3)
Title |
---|
GARY J. SULLIVAN ET AL.: "Overview of the High Efficiency Video Coding (HEVC) Standard", IEEE TRANSACTIONS ON CIRCUITS AND SYSTEMS FOR VIDEO TECHNOLOGY, vol. 22, no. 12, December 2012 (2012-12-01), pages 1649 - 1668, XP011487803 * |
KARSTEN SUHRING ET AL.: "Indication of tile boundary alignment", JOINT COLLABORATIVE TEAM ON VIDEO CODING (JCT-VC) OF ITU-T SG 16 WP 3 AND ISO/IEC JTC 1/SC 29/WG 11, 12TH MEETING, 14 January 2013 (2013-01-14) - 23 January 2013 (2013-01-23), GENEVA, CH, XP030113685 * |
KARSTEN SUHRING ET AL.: "Tile boundary alignment and inter-layer prediction constraints for SHVC and MV-HEVC", JOINT COLLABORATIVE TEAM ON VIDEO CODING (JCT- VC) OF ITU-T SG 16 WP 3 AND ISO/IEC JTC 1/SC 29/WG 11, 13TH MEETING, 18 April 2013 (2013-04-18) - 26 April 2013 (2013-04-26), INCHEON, KR, XP055119146 * |
Cited By (1)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
US11762136B2 (en) | 2011-03-03 | 2023-09-19 | Enchroma, Inc. | Multi-band color vision filters and method by LP-optimization |
Also Published As
Publication number | Publication date |
---|---|
EP3021586A4 (en) | 2016-12-07 |
WO2015005622A9 (ko) | 2015-03-12 |
WO2015005622A1 (ko) | 2015-01-15 |
KR20160002786A (ko) | 2016-01-08 |
EP3021586A1 (en) | 2016-05-18 |
CA2909601A1 (en) | 2015-01-15 |
US20160080753A1 (en) | 2016-03-17 |
CN105284114A (zh) | 2016-01-27 |
Similar Documents
Publication | Publication Date | Title |
---|---|---|
WO2015005621A1 (ko) | 비디오 신호 처리 방법 및 장치 | |
WO2014171768A1 (ko) | 비디오 신호 처리 방법 및 장치 | |
CN111988619B (zh) | 视频编解码方法、装置、计算机设备和存储介质 | |
CN113892270B (zh) | 视频编解码方法、装置和存储介质 | |
WO2014171770A1 (ko) | 비디오 신호 처리 방법 및 장치 | |
KR102079803B1 (ko) | 영상 디코딩 방법 및 이를 이용하는 장치 | |
KR20230154285A (ko) | 참조 픽처 리샘플링을 위한 시그널링 | |
KR102136358B1 (ko) | 영상 정보 시그널링 방법 및 장치와 이를 이용한 디코딩 방법 및 장치 | |
EP3985970A1 (en) | Video signal processing method and device using block dpcm prediction method | |
CN113424532B (zh) | 视频编解码方法、装置和存储介质 | |
CN113994676B (zh) | 视频解码方法、装置以及电子设备 | |
WO2021137445A1 (ko) | 비디오 신호 처리를 위한 변환 커널 결정 방법 및 이를 위한 장치 | |
CN114830673A (zh) | 用于多个层的共享解码器图片缓冲器 | |
KR20220127308A (ko) | 비디오 코딩을 위한 방법 및 장치 | |
CN113574895A (zh) | 帧间位置相关的预测组合模式的改进 | |
US11445206B2 (en) | Method and apparatus for video coding | |
CN115152208A (zh) | 视频解码的方法和设备 | |
WO2014171771A1 (ko) | 비디오 신호 처리 방법 및 장치 | |
CN117015972B (zh) | 视频编码和解码的方法、装置及存储介质 | |
KR20210033858A (ko) | 비디오 코덱에서 인트라 예측에 따른 이차 변환 커널의 유도 및 매핑 방법 | |
KR20200096385A (ko) | 코딩블록과 변환 블록 간 화면내 예측 모드를 통합 사용하는 비디오 신호 처리 방법 및 장치 | |
KR20210035413A (ko) | 비디오 코덱에서 행렬 기반 인트라 예측에 따른 명시적 주변환 커널 적용 방법 | |
AU2024219800A1 (en) | Method and device for processing video signal by using picture header | |
CN116368801A (zh) | 交叉分量变换系数水平重构 | |
KR20200054052A (ko) | 화면내 블록 복사를 사용하는 비디오 신호 처리 방법 및 장치 |
Legal Events
Date | Code | Title | Description |
---|---|---|---|
WWE | Wipo information: entry into national phase |
Ref document number: 201480030183.X Country of ref document: CN |
|
121 | Ep: the epo has been informed by wipo that ep was designated in this application |
Ref document number: 14823575 Country of ref document: EP Kind code of ref document: A1 |
|
ENP | Entry into the national phase |
Ref document number: 2909601 Country of ref document: CA |
|
WWE | Wipo information: entry into national phase |
Ref document number: 14784954 Country of ref document: US |
|
ENP | Entry into the national phase |
Ref document number: 20157029984 Country of ref document: KR Kind code of ref document: A |
|
WWE | Wipo information: entry into national phase |
Ref document number: 2014823575 Country of ref document: EP |
|
NENP | Non-entry into the national phase |
Ref country code: DE |