WO2015056941A1 - Procédé et appareil de codage/décodage d'images sur base multicouche - Google Patents

Procédé et appareil de codage/décodage d'images sur base multicouche Download PDF

Info

Publication number
WO2015056941A1
WO2015056941A1 PCT/KR2014/009580 KR2014009580W WO2015056941A1 WO 2015056941 A1 WO2015056941 A1 WO 2015056941A1 KR 2014009580 W KR2014009580 W KR 2014009580W WO 2015056941 A1 WO2015056941 A1 WO 2015056941A1
Authority
WO
WIPO (PCT)
Prior art keywords
tile
related information
layer
flag
picture
Prior art date
Application number
PCT/KR2014/009580
Other languages
English (en)
Korean (ko)
Inventor
강정원
이진호
이하현
최진수
김진웅
Original Assignee
한국전자통신연구원
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Application filed by 한국전자통신연구원 filed Critical 한국전자통신연구원
Priority to US15/028,570 priority Critical patent/US10091519B2/en
Priority to CN201910476959.5A priority patent/CN110225349B/zh
Priority to CN201480062557.6A priority patent/CN105723712B/zh
Priority claimed from KR1020140137482A external-priority patent/KR102290091B1/ko
Publication of WO2015056941A1 publication Critical patent/WO2015056941A1/fr
Priority to US16/039,766 priority patent/US10701379B2/en
Priority to US16/039,996 priority patent/US10701380B2/en
Priority to US16/040,119 priority patent/US10701381B2/en
Priority to US16/040,222 priority patent/US10701382B2/en

Links

Images

Classifications

    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04NPICTORIAL COMMUNICATION, e.g. TELEVISION
    • H04N19/00Methods or arrangements for coding, decoding, compressing or decompressing digital video signals
    • H04N19/10Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using adaptive coding
    • H04N19/169Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using adaptive coding characterised by the coding unit, i.e. the structural portion or semantic portion of the video signal being the object or the subject of the adaptive coding
    • H04N19/17Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using adaptive coding characterised by the coding unit, i.e. the structural portion or semantic portion of the video signal being the object or the subject of the adaptive coding the unit being an image region, e.g. an object
    • H04N19/174Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using adaptive coding characterised by the coding unit, i.e. the structural portion or semantic portion of the video signal being the object or the subject of the adaptive coding the unit being an image region, e.g. an object the region being a slice, e.g. a line of blocks or a group of blocks
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04NPICTORIAL COMMUNICATION, e.g. TELEVISION
    • H04N19/00Methods or arrangements for coding, decoding, compressing or decompressing digital video signals
    • H04N19/30Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using hierarchical techniques, e.g. scalability

Definitions

  • the present invention relates to image encoding and decoding, and more particularly, to a method of representing tiles and pictures when one or more layers exist in a video bitstream.
  • the supporting performance is also diversified by device.
  • the network in which information is transmitted is also diversified not only in the external structure such as wired and wireless networks, but also in functions such as the type of information to be transmitted, the amount and speed of information.
  • the user selects a terminal and a network to be used according to a desired function, and the spectrum of terminals and networks provided to the user by the enterprise is also diversified.
  • Ultra High Definition which has more than four times the resolution of HDTV
  • the demand for technology for compressing and processing higher resolution and higher quality images is increasing.
  • an inter prediction technique for predicting a pixel value included in a current picture from a previous and / or subsequent picture in time, another pixel included in the current picture using pixel information in the current picture
  • An intra prediction technique for predicting a value an entropy encoding technique for allocating a short sign to a high frequency symbol and a long code to a low frequency symbol may be used.
  • the quality, size, frame, etc. of the supported images need to be diversified accordingly.
  • the present invention provides an image encoding / decoding method and apparatus capable of improving image encoding / decoding efficiency.
  • the present invention provides a method and apparatus for signaling tile related information in scalable video coding that can improve encoding / decoding efficiency.
  • the present invention provides a method and apparatus for signaling information related to representation format in scalable video coding that can improve encoding / decoding efficiency.
  • an image decoding method supporting a plurality of layers includes acquiring information indicating whether tile related information exists and acquiring the tile related information based on information indicating whether the tile related information exists.
  • the tile related information may be information indicating whether a position of a tile existing in a picture of a current layer corresponds to a position of a tile existing in a picture of a reference layer referenced by the current layer.
  • an image decoding method supporting a plurality of layers may further include obtaining information indicating whether chroma format related information and bit-depth related information exist, and whether the chrominance format related information and bit depth related information exist. And obtaining the color difference format related information and the bit depth related information based on the information indicating the.
  • FIG. 1 is a block diagram illustrating a configuration of an image encoding apparatus according to an embodiment of the present invention.
  • FIG. 2 is a block diagram illustrating a configuration of an image decoding apparatus according to an embodiment of the present invention.
  • FIG. 3 is a conceptual diagram schematically illustrating an embodiment of a scalable video coding structure using multiple layers to which the present invention can be applied.
  • FIG. 4 is a diagram illustrating a tile area that is removed and emptied in one picture.
  • FIG. 5 is a flowchart schematically illustrating a method of acquiring tile related information in a scalable video coding structure supporting a plurality of layers according to an embodiment of the present invention.
  • FIG. 6 is a flowchart schematically illustrating a method of acquiring representation format related information in a scalable video coding structure supporting a plurality of layers according to an embodiment of the present invention.
  • first and second may be used to describe various configurations, but the configurations are not limited by the terms. The terms are used to distinguish one configuration from another.
  • first configuration may be referred to as the second configuration, and similarly, the second configuration may also be referred to as the first configuration.
  • each component shown in the embodiments of the present invention are independently shown to represent different characteristic functions, and do not mean that each component is made of separate hardware or one software component unit.
  • each component is listed as a component for convenience of description, and at least two of the components may form one component, or one component may be divided into a plurality of components to perform a function.
  • the integrated and separated embodiments of each component are also included in the scope of the present invention without departing from the spirit of the present invention.
  • the components may not be essential components for performing essential functions in the present invention, but may be optional components for improving performance.
  • the present invention can be implemented including only the components essential for implementing the essentials of the present invention except for the components used for improving performance, and the structure including only the essential components except for the optional components used for improving performance. Also included in the scope of the present invention.
  • FIG. 1 is a block diagram illustrating a configuration of an image encoding apparatus according to an embodiment of the present invention.
  • a scalable video encoding apparatus supporting a multi-layer structure may be implemented by extending a general image encoding apparatus having a single layer structure.
  • the block diagram of FIG. 1 shows an embodiment of an image encoding apparatus that can be the basis of a scalable video encoding apparatus applicable to a multilayer structure.
  • the image encoding apparatus 100 may include an inter predictor 110, an intra predictor 120, a switch 115, a subtractor 125, a transformer 130, a quantizer 140, and entropy.
  • An encoder 150, an inverse quantizer 160, an inverse transformer 170, an adder 175, a filter 180, and a reference picture buffer 190 are included.
  • the image encoding apparatus 100 may encode an input image in an intra mode or an inter mode and output a bitstream.
  • the switch 115 may be switched to intra, and in the inter mode, the switch 115 may be switched to inter.
  • Intra prediction means intra prediction and inter prediction means inter prediction.
  • the image encoding apparatus 100 may generate a prediction block for an input block of an input image and then encode a residual between the input block and the prediction block.
  • the input image may mean an original picture.
  • the intra predictor 120 may use a sample value of an already encoded / decoded block around the current block as a reference sample.
  • the intra prediction unit 120 may perform spatial prediction using the reference sample and generate prediction samples for the current block.
  • the inter prediction unit 110 may identify a motion vector specifying a reference block having a smallest difference from an input block (current block) in a reference picture stored in the reference picture buffer 190 during a motion prediction process. You can get it.
  • the inter prediction unit 110 may generate a prediction block for the current block by performing motion compensation using the motion vector and the reference picture stored in the reference picture buffer 190.
  • inter prediction applied in inter mode may include inter layer prediction.
  • the inter prediction unit 110 may construct an inter-layer reference picture by sampling a picture of the reference layer, and perform inter-layer prediction by including an inter-layer reference picture in the reference picture list.
  • the reference relationship between layers may be signaled through information specifying dependencies between layers.
  • sampling applied to the reference layer picture may mean generation of a reference sample by copying a sample from the reference layer picture.
  • sampling applied to the reference layer picture may mean upsampling.
  • an interlayer reference picture may be configured by upsampling a reconstructed picture of a reference layer between layers that support scalability with respect to resolution.
  • the encoding apparatus may transmit information specifying the layer to which the picture to be used as the inter-layer reference picture belongs, to the decoding apparatus.
  • a layer referred to in inter-layer prediction that is, a picture used for prediction of the current block in the reference layer, may be a picture of the same access unit (AU) as the current picture (picture to be predicted in the current layer).
  • AU access unit
  • the subtractor 125 may generate a residual block by the difference between the input block and the generated prediction block.
  • the transform unit 130 may output a transform coefficient by performing a transform on the residual block.
  • the transform coefficient may mean a coefficient value generated by performing transform on the residual block and / or the residual signal.
  • a quantized transform coefficient level generated by applying quantization to a transform coefficient may also be referred to as a transform coefficient.
  • the transform unit 130 may omit the transform on the residual block.
  • the quantization unit 140 may quantize the input transform coefficient according to a quantization parameter or a quantization parameter, and output a quantized coefficient. Quantized coefficients may be referred to as quantized transform coefficient levels. In this case, the quantization unit 140 may quantize the input transform coefficients using the quantization matrix.
  • the entropy encoder 150 may output a bitstream by entropy encoding the values calculated by the quantizer 140 or the encoding parameter values calculated in the encoding process according to a probability distribution.
  • the entropy encoder 150 may entropy encode information (eg, syntax elements) for video decoding in addition to the pixel information of the video.
  • the encoding parameter is information necessary for encoding and decoding, and may include information that may be inferred in the encoding or decoding process as well as information encoded in the encoding apparatus and delivered to the decoding apparatus, such as a syntax element.
  • encoding parameters may include intra / inter prediction modes, moving / motion vectors, reference image indexes, coding block patterns, presence or absence of residual signals, transform coefficients, quantized transform coefficients, quantization parameters, block sizes, block division information, and the like. Or statistics.
  • the residual signal may mean a difference between the original signal and the prediction signal, and a signal in which the difference between the original signal and the prediction signal is transformed or a signal in which the difference between the original signal and the prediction signal is transformed and quantized It may mean.
  • the residual signal may be referred to as a residual block in block units.
  • the entropy encoder 150 may use an encoding method such as exponential golomb, context-adaptive variable length coding (CAVLC), or context-adaptive binary arithmetic coding (CABAC). For example, the entropy encoder 150 may perform entropy encoding by using a variable length coding (VLC) table. In addition, the entropy encoder 150 derives a binarization method of a target symbol and a probability model of a target symbol / bin, and then performs entropy encoding using the derived binarization method or a probability model. You may.
  • VLC variable length coding
  • the quantized coefficients may be inversely quantized by the inverse quantizer 160 and inversely transformed by the inverse transformer 170.
  • the inverse quantized and inverse transformed coefficients are added to the prediction block through the adder 175 and a reconstructed block is generated.
  • the reconstruction block passes through the filter unit 180, and the filter unit 180 applies at least one or more of a deblocking filter, a sample adaptive offset (SAO), and an adaptive loop filter (ALF) to the reconstructed block or reconstructed picture. can do.
  • the filter unit 180 may be referred to as an adaptive in-loop filter.
  • the deblocking filter can remove block distortion generated at the boundary between blocks.
  • SAO can add an appropriate offset to the pixel value to compensate for coding errors.
  • the ALF may perform filtering based on a value obtained by comparing the reconstructed image with the original image.
  • the reconstructed block that has passed through the filter unit 180 may be stored in the reference picture buffer 190.
  • FIG. 2 is a block diagram illustrating a configuration of an image decoding apparatus according to an embodiment of the present invention.
  • a scalable video decoding apparatus supporting a multi-layer structure may be implemented by extending a general image decoding apparatus having a single layer structure.
  • the block diagram of FIG. 2 illustrates an embodiment of an image decoding apparatus that may be the basis of a scalable video decoding apparatus applicable to a multilayer structure.
  • the image decoding apparatus 200 may include an entropy decoder 210, an inverse quantizer 220, an inverse transformer 230, an intra predictor 240, an inter predictor 250, and an adder 255. ), A filter unit 260 and a reference picture buffer 270.
  • the image decoding apparatus 200 may receive a bitstream output from the encoder and perform decoding in an intra mode or an inter mode, and output a reconstructed image, that is, a reconstructed image.
  • the switch In the intra mode, the switch may be switched to intra, and in the inter mode, the switch may be switched to inter.
  • the image decoding apparatus 200 may obtain a reconstructed residual block from the received bitstream, generate a prediction block, and then add the reconstructed residual block and the prediction block to generate a reconstructed block, that is, a reconstruction block. .
  • the entropy decoder 210 may entropy decode the input bitstream according to a probability distribution, and output information such as quantized coefficients and syntax elements.
  • the quantized coefficients are inversely quantized by the inverse quantizer 220 and inversely transformed by the inverse transformer 230.
  • a reconstructed residual block may be generated.
  • the inverse quantization unit 220 may apply a quantization matrix to the quantized coefficients.
  • the intra predictor 240 may perform spatial prediction using sample values of blocks already decoded around the current block, and generate prediction samples for the current block.
  • the inter prediction unit 250 may generate a prediction block for the current block by performing motion compensation using the reference picture stored in the motion vector and the reference picture buffer 270.
  • inter prediction applied in inter mode may include inter layer prediction.
  • the inter prediction unit 250 may construct an interlayer reference picture by sampling a picture of the reference layer, and perform interlayer prediction by including an interlayer reference picture in a reference picture list.
  • the reference relationship between layers may be signaled through information specifying dependencies between layers.
  • sampling applied to the reference layer picture may mean generation of a reference sample by copying a sample from the reference layer picture.
  • sampling applied to the reference layer picture may mean upsampling.
  • the inter-layer reference picture may be configured by upsampling the reconstructed picture of the reference layer.
  • information specifying a layer to which a picture to be used as an inter-layer reference picture belongs may be transmitted from the encoding apparatus to the decoding apparatus.
  • a layer referred to in inter-layer prediction that is, a picture used for prediction of the current block in the reference layer, may be a picture of the same access unit (AU) as the current picture (picture to be predicted in the current layer).
  • AU access unit
  • the reconstructed residual block and the predictive block are added at the adder 255 to generate a reconstructed block.
  • the residual sample and the predictive sample are added to generate a reconstructed sample or a reconstructed picture.
  • the reconstructed picture is filtered by the filter unit 260.
  • the filter unit 260 may apply at least one or more of the deblocking filter, SAO, and ALF to the reconstructed block or the reconstructed picture.
  • the filter unit 260 outputs a modified or filtered reconstructed picture.
  • the reconstructed picture may be stored in the reference picture buffer 270 and used for inter prediction.
  • the image decoding apparatus 200 may further include a parser (not shown) that parses information related to an encoded image included in the bitstream.
  • the parser may include the entropy decoder 210 or may be included in the entropy decoder 210. Such a parser may also be implemented as one component of the decoder.
  • one encoding device / decoding device processes both encoding / decoding for multiple layers. However, this is for convenience of description.
  • the encoding device / decoding device may be configured for each layer.
  • the encoding device / decoding device of the upper layer may perform encoding / decoding of the corresponding upper layer by using the information of the upper layer and the information of the lower layer.
  • the prediction unit (inter prediction unit) of the upper layer may perform intra prediction or inter prediction on the current block by using pixel information or picture information of the upper layer, and receives picture information reconstructed from the lower layer and Inter prediction (inter layer prediction) may be performed on the current block of the upper layer by using the interlayer prediction.
  • Inter prediction inter layer prediction
  • the encoding device / decoding device performs encoding / decoding on the current layer by using information of another layer regardless of whether the device is configured for each layer or one device processes multiple layers. can do.
  • the layer may include a view.
  • information of another layer between layers specified as having dependence by information specifying dependency between layers is not merely performed by using information of a lower layer.
  • Inter layer prediction may be performed using.
  • FIG. 3 is a conceptual diagram schematically illustrating an embodiment of a scalable video coding structure using multiple layers to which the present invention can be applied.
  • a GOP Group of Picture
  • FIG. 3 a GOP (Group of Picture) represents a picture group, that is, a group of pictures.
  • a transmission medium In order to transmit image data, a transmission medium is required, and its performance varies depending on the transmission medium according to various network environments.
  • a scalable video coding method may be provided for application to such various transmission media or network environments.
  • a video coding method that supports scalability removes redundancy between layers by using texture information, motion information, and residual signals between layers.
  • a coding method that improves encoding and decoding performance has various scales in terms of spatial, temporal, image quality (or quality), and view according to ambient conditions such as transmission bit rate, transmission error rate, and system resources. It can provide the ability.
  • Scalable video coding may be performed using multiple layers structure to provide a bitstream applicable to various network situations.
  • the scalable video coding structure may include a base layer that compresses and processes image data using a general image decoding method, and compresses the image data using both the decoding information of the base layer and a general image decoding method. May include an enhancement layer for processing.
  • the base layer may be referred to as a base layer or may be referred to as a lower layer.
  • the enhancement layer may be referred to as an enhancement layer or a higher layer.
  • the lower layer may mean a layer that supports a lower level of scalability (spatial, temporal, image quality, or viewpoint scalability) than a specific layer, and the upper layer may have a higher level of scalability than a specific layer. It may mean a layer that supports (spatial, temporal, image quality, or view scalability).
  • a layer referred to for encoding / decoding of another layer may be referred to as a reference layer (reference layer), and a layer encoded / decoded using another layer may be referred to as a current layer (current layer).
  • the reference layer may be lower than the current layer, and the current layer may be higher than the reference layer.
  • the layer may be based on spatial (eg, image size), temporal (eg, decoding order, image output order, frame rate), image quality, complexity, view, etc. Means a set of image and bitstream to be distinguished.
  • the base layer may be defined as a standard definition (SD), a frame rate of 15 Hz, and a 1 Mbps bit rate
  • the first enhancement layer may be a high definition (HD), a frame rate of 30 Hz, and a 3.9 Mbps bit rate
  • the second enhancement layer may be defined as 4K-UHD (ultra high definition), a frame rate of 60 Hz, and a bit rate of 27.2 Mbps.
  • the format, frame rate, bit rate, etc. are exemplary and may be determined differently as necessary.
  • the number of hierarchies used is not limited to this embodiment and may be determined differently according to a situation. For example, if the transmission bandwidth is 4 Mbps, the frame rate of the first enhancement layer HD may be reduced and transmitted at 15 Hz or less.
  • the scalable video coding method may provide temporal, spatial, image quality, and view scalability by the method described above in the embodiment of FIG. 3.
  • scalable video coding has the same meaning as scalable video encoding in terms of encoding and scalable video decoding in terms of decoding.
  • HEVC High Efficiency Video Coding
  • HEVC High Efficiency Video Coding
  • parallel processing is possible by dividing one picture into a plurality of tiles and independently encoding / decoding each divided tile in one picture. Therefore, information indicating how a picture in a bitstream is configured in a tile form is required.
  • tile related information is described in a picture parameter set (PPS).
  • PPS picture parameter set
  • SPS Sequence Parameter Set
  • the present invention provides a method for efficiently encoding / decoding tile related information through signaling of correlation between layers when a picture is composed of one or more tiles for each layer in the hierarchical bitstream.
  • the present invention also provides a method for efficiently signaling picture representation related information in scalable video coding.
  • a tile refers to a rectangular area composed of coding tree blocks or coding tree units in a picture and may be a sequence of coding tree blocks or coding tree units.
  • a tile always consists of integer coding tree units.
  • the coding tree block or the coding tree unit may be a processing unit of samples in the picture.
  • one picture may be divided into two tiles by a vertical tile boundary in the picture.
  • Each divided tile includes integer coding tree units and may be a rectangular area.
  • scalable HEVC (SHVC) standard for supporting scalable video coding
  • MV-HEVC multiview HEVC
  • information related to tiles can be signaled in a video parameter set (VPS) and a PPS.
  • VPS video parameter set
  • PPS video parameter set
  • the spatial resolution of the reference layer when the spatial resolution of the reference layer does not match the spatial resolution of the corresponding layer, the spatial resolution of the reference layer is scaled to be the same as the spatial resolution of the corresponding layer, and the positions of the tiles included in the scaled reference layer It may be determined whether to correspond / match with the position of the tiles of the layer based on the.
  • tile_boundaries_aligned_flag [i] [j] is used as a flag indicating whether the positions of the tiles included in the layer are encoded / decoded to correspond / match the positions of the tiles included in the reference layers of the layer.
  • tile_boundaries_aligned_flag [i] [j] indicates whether the i-th layer is aligned with the j-th reference layer and the position of the tile (the boundary of the tiles). However, when only one tile is included for each layer in the entire bitstream, it is not necessary to signal tile_boundaries_aligned_flag [i] [j].
  • tile_boundaries_aligned_flag [i] [j] only when there is a picture including at least two tiles.
  • Table 1 is an example of a syntax illustrating a method of signaling tile_boundaries_aligned_flag [i] [j] according to an embodiment of the present invention.
  • vps_tiles_enabled_flag when the vps_tiles_enabled_flag has a value of 1, it means that a picture including two or more tiles among pictures referring to the corresponding VPS may exist.
  • vps_tiles_enabled_flag has a value of 0, it means that all pictures of the CVS (Coded Video Stream) contain only one tile.
  • tile_boundaries_aligned_flag [i] [j] may be signaled.
  • Table 2 is another example of a syntax illustrating a method of signaling tile_boundaries_aligned_flag [i] [j] according to an embodiment of the present invention.
  • tiles_not_in_use_flag when tiles_not_in_use_flag has a value of 1, it means that all pictures of the CVS include only one tile. When tiles_not_in_use_flag has a value of 0, it means that a picture including two or more tiles among pictures referring to the corresponding VPS may exist.
  • tile_boundaries_aligned_flag [i] [j] may be signaled.
  • Table 3 is another example of a syntax illustrating a method of signaling tile_boundaries_aligned_flag [i] [j] according to an embodiment of the present invention.
  • vps_tiles_enabled_flag [i] when vps_tiles_enabled_flag [i] has a value of 1, it means that each picture of the i-th layer includes at least two tiles. When vps_tiles_enabled_flag [i] has a value of 0, it means that all pictures of the i-th layer include only one tile.
  • tile_boundaries_aligned_flag [i] [j] may be signaled.
  • Table 4 is an example of a syntax that describes a method for signaling in a PPS using vps_tiles_enabled_flag signaled in the VPS of Table 1 according to an embodiment of the present invention.
  • tiles_enabled_flag when tiles_enabled_flag has a value of 1, it means that all pictures referring to the corresponding PPS include two or more tiles. When tiles_enabled_flag has a value of 0, it means that all pictures referring to the PPS include only one tile. If the tiles_enabled_flag value is not signaled, the tiles_enabled_flag value is assumed to be zero.
  • vps_tiles_enabled_flag As shown in Table 4, whether tiles_enabled_flag is signaled or not may be determined by vps_tiles_enabled_flag. However, regardless of vps_tiles_enabled_flag, when tiles_enabled_flag is signaled as shown in Table 5, the value of tiles_enabled_flag should be 1 only when the value of vps_tiles_enabled_flag is 1. In other words, when the value of vps_tiles_enabled_flag is 0, the value of tiles_enabled_flag should always be 0.
  • Table 6 is an example of a syntax that describes a method for signaling sps_tiles_enabled_flag without signaling vps_tiles_enabled_flag in an SPS according to an embodiment of the present invention.
  • sps_tiles_enabled_flag when sps_tiles_enabled_flag has a value of 1, it means that all pictures referring to the corresponding SPS include two or more tiles. When sps_tiles_enabled_flag has a value of 0, it means that all pictures referring to the corresponding SPS include one tile.
  • PPS may signal tiles_enabled_flag as shown in Table 7 below.
  • tiles_enabled_flag has the following restrictions.
  • the present invention proposes a method of signaling tiles_enabled_flag in the SPS without signaling in the PPS when the layer identifier (nuh_layer_id) value is greater than 0 in order to make the PPSs included in the CVS consistent and simple.
  • Table 8 is an example of a syntax that describes a method for signaling tiles_enabled_flag in an SPS based on layer identifier (nuh_layer_id) information according to an embodiment of the present invention.
  • Table 9 is another example of a syntax illustrating a method of signaling tiles_enabled_flag based on layer identifier (nuh_layer_id) information according to an embodiment of the present invention.
  • tiles_enabled_flag when tiles_enabled_flag has a value of 1, it means that all pictures referring to the corresponding PPS include two or more tiles. When tiles_enabled_flag has a value of 0, it means that all pictures referring to the corresponding PPS include only one tile. If the tiles_enabled_flag value is not signaled, the tiles_enabled_flag value is inferred as the sps_tiles_enabled_flag value.
  • the tiles_enabled_flag value should have the same value for all PPSs activated in a specific layer of the CVS.
  • Some CVSs may be complicated to decode in real time in some decoders. If the tile related information is delivered to the decoder, the decoder can know in advance whether it can decode the CVS. That is, tile related information signaled in the VPS may be useful for session negotiation. In addition, if pictures in a particular layer have the same tile setting, that is, in the form of the same default tile, it may not be necessary to signal tile related information in every PPS.
  • Table 10 is an example of a syntax that describes a method for signaling tile related information in a VPS according to an embodiment of the present invention.
  • tiles_not_in_use_flag has a value of 1
  • all the pictures of the CVS include only one tile.
  • tiles_not_in_use_flag has a value of 0
  • default_tile_info_flag [i] When default_tile_info_flag [i] has a value of 1, it may mean that the picture of the i-th layer has a default tile setting. When default_tile_info_flag [i] has a value of 0, it indicates that a picture of the i-th layer may or may not have a default tile setting.
  • a value of num_tile_columns_minus1 [i] plus 1 represents the number of tile columns of a partitioning picture of the i-th layer. If num_tile_columns_minus1 [i] is not signaled, num_tile_columns_minus1 [i] is assumed to be 0.
  • a value of num_tile_rows_minus1 [i] plus 1 represents the number of tile rows dividing a picture of the i-th layer. If num_tile_rows_minus1 [i] is not signaled, num_tile_rows_minus1 [i] is assumed to be zero.
  • uniform_spacing_flag [i] When uniform_spacing_flag [i] has a value of 1, it means that a tile column and a tile row are positioned at regular intervals in a picture of the i-th layer. If Uniform_spacing_flag [i] has a value of 0, it means that the tile column and the tile row are not located at regular intervals in the picture of the i-th layer, where column_width_minus1 of each tile [i] [j] and row_height_minus1 [i] [j] may be used to signal the width of the column and the height of the row. If the uniform_spacing_flag value does not exist, the uniform_spacing_flag value may be assumed to be 1.
  • row_height_minus1 [i] [j] plus 1 indicates the height of the row of the j-th tile of the i-th layer in units of coding tree blocks.
  • loop_filter_across_tiles_enabled_flag [i] When loop_filter_across_tiles_enabled_flag [i] has a value of 1, it indicates that in-loop filtering is performed on the tile boundary portion in the picture of the i-th layer. When loop_filter_across_tiles_enabled_flag [i] has a value of 0, it indicates that in-loop filtering is not performed on the tile boundary portion of the picture of the i-th layer.
  • In-loop filtering operations include a deblocking filter and an adaptive offset filter operation. If the loop_filter_across_tiles_enabled_flag value is not signaled, the loop_filter_across_tiles_enabled_flag value is assumed to be 1.
  • supplemental enhancement information SEI
  • the PPS may be modified to update only tile-related information signaled by the PPS when necessary.
  • Table 11 is an example of a syntax that describes a method of updating tile related information signaled by a PPS according to an embodiment of the present invention only when necessary.
  • tiles_enabled_flag when tiles_enabled_flag has a value of 1, it means that all pictures referring to the corresponding PPS include two or more tiles. When tiles_enabled_flag has a value of 0, it means that all pictures referring to the PPS include only one tile. If the tiles_enabled_flag value is not signaled, the tiles_enabled_flag value is inferred to a value of zero.
  • update_tiles_info_flag When update_tiles_info_flag has a value of 1, it indicates that tile related information of pictures referring to the corresponding PPS is updated. When update_tiles_info_flag has a value of 0, it indicates that tile related information of pictures referring to the corresponding PPS is not updated.
  • Table 12 is another example of a syntax that describes a method of signaling tile related information in a VPS according to an embodiment of the present invention.
  • the update mechanism of the PPS can be applied only to the number of tiles and the loop filter signaling, as shown in Table 13 below.
  • Table 14 is another example of a syntax illustrating a method of signaling tile related information in a VPS according to an embodiment of the present invention.
  • Table 15 is another example of a syntax illustrating a method of signaling tile related information in a VPS according to an embodiment of the present invention.
  • min_max_tile_info_flag [i] when min_max_tile_info_flag [i] has a value of 1, it means that the maximum tile number and the minimum tile number of the picture of the i-th layer included in the bitstream are signaled. If min_max_tile_info_flag [i] has a value of 0, it means that the maximum tile number and the minimum tile number of the picture of the i-th layer included in the bitstream are not signaled.
  • max_num_tile_columns_minus1 [i] plus 1 represents the maximum number of tile columns of a partitioning picture of the i-th layer. If max_num_tile_columns_minus1 [i] is not signaled, max_num_tile_columns_minus1 [i] is assumed to be zero.
  • min_num_tile_columns_minus1 [i] plus 1 represents the minimum number of tile columns of a partitioning picture of the i-th layer. If min_num_tile_columns_minus1 [i] is not signaled, min_num_tile_columns_minus1 [i] is assumed to be zero.
  • max_num_tile_rows_minus1 [i] plus 1 represents the maximum number of tile rows dividing a picture of the i-th layer. If max_num_tile_rows_minus1 [i] is not signaled, max_num_tile_rows_minus1 [i] is assumed to be zero.
  • min_num_tile_rows_minus1 [i] plus 1 represents the minimum number of tile rows of partitioning pictures of the i-th layer. If min_num_tile_rows_minus1 [i] is not signaled, min_num_tile_rows_minus1 [i] is assumed to be zero.
  • Table 16 is another example of a syntax illustrating a method of signaling tile related information in a VPS according to an embodiment of the present invention.
  • max_tile_info_flag [i] when max_tile_info_flag [i] has a value of 1, it means that the maximum number of tiles of the picture of the i-th layer included in the bitstream is signaled.
  • max_tile_info_flag [i] has a value of 0, it means that the maximum number of tiles of the picture of the i-th layer included in the bitstream is not signaled.
  • min_tile_info_flag [i] When min_tile_info_flag [i] has a value of 1, it means that the minimum number of tiles of the picture of the i-th layer included in the bitstream is signaled. If min_tile_info_flag [i] has a value of 0, it means that the minimum number of tiles of the picture of the i-th layer included in the bitstream is not signaled.
  • the fourth embodiment and the fifth embodiment may be signaled in vps_vui ().
  • the tile settings of pictures in a particular layer will be the same, so other ways of describing tile related information can be described in the SPS. If the tile related information is not modified in the VPS, the SPS may be modified as shown in Table 17 below.
  • Table 17 is an example of a syntax that describes a method for signaling tile related information in an SPS according to an embodiment of the present invention.
  • tile set later intentionally removed and empty.
  • This tile area may later be filled by a device present in the content delivery path (used for advertising boxes, etc.). To this end, it would be helpful if the decoder could know which tile regions (ie, independently decodable regions) were removed.
  • the current "inter-layer constrained tile sets SEI messages" describe only the independently decodable region, and does not describe which region is removed from the transmission path.
  • the present invention provides a method for describing information about a tile region so that a decodable region and a removed region within a picture can be identified.
  • a flag may be added to each defined tile area to indicate whether a tile area in a picture exists.
  • Table 18 is an example of syntax illustrating a method of using a flag indicating whether a tile area exists according to an embodiment of the present invention.
  • present_flag [i] when present_flag [i] is a value of 1, it means that a recognized tile set exists in the CVS. If present_flag [i] is a value of 0, it means that there is no tile region recognized in the CVS.
  • the present invention proposes a method for improving the expression of the "representation format" described in the VPS and SPS.
  • Table 19 is an example of a syntax that describes a method for signaling a representation format according to an embodiment of the present invention.
  • vps_rep_format_idx [i] represents an index of a list of rep_format () syntax applied to the i-th layer.
  • Table 19 may be modified as shown in Table 20 below.
  • Table 20 is another example of a syntax illustrating a method of signaling a representation format according to an embodiment of the present invention.
  • information about a representation format may be signaled as shown in Table 21 below to prevent signaling of redundant information.
  • Table 21 shows an example of a representation format syntax for improving signaling of duplicate information according to an embodiment of the present invention.
  • chroma_size_vps_present_flag when chroma_size_vps_present_flag has a value of 1, it indicates that information related to chroma format and picture size is included in the syntax.
  • chroma_size_vps_present_flag has a value of 0, it means that the chroma format and the picture size related information refer to the previously signaled rep_format ().
  • the value of chroma_size_vps_present_flag of the first rep_format () must be 1.
  • bit_depth_vps_present_flag When bit_depth_vps_present_flag has a value of 1, it indicates that bit-depth related information is included in the syntax. When bit_depth_vps_present_flag has a value of 0, it means that bit-depth related information refers to previously signaled rep_format (). The value of bit_depth_vps_present_flag of the first rep_format () must be 1.
  • information about a representation format may be signaled as shown in Table 22 below to prevent signaling of redundant information.
  • Table 22 is another example showing a representation format syntax for improving signaling of redundant information according to an embodiment of the present invention.
  • depth_chroma_vps_present_flag when depth_chroma_vps_present_flag has a value of 1, it indicates that information related to chroma format and bit-depth is included in the syntax.
  • the depth_chroma_vps_present_flag has a value of 0, it means that the chroma format and the bit-depth related information refer to the previously signaled rep_format ().
  • previously signaled rep_format () may mean rep_format () applied to a layer immediately below the layer to which the current rep_format () is applied.
  • the value of depth_chroma_vps_present_flag of the first rep_format () must be 1.
  • update_re_format_flag is replaced with update_bit_depth_flag indicating that update_chroma_size_flag and bit-depth related update information exist, indicating that update information related to the chroma format and picture size exists, as shown in Table 23 below. It can also be.
  • Table 23 is an example of a syntax that describes a method for signaling chrominance format, picture size, and bit depth related update information in an SPS according to an embodiment of the present invention.
  • update_chroma_size_flag has a value of 1
  • chroma_format_idc, separate_colour_plane_flag, pic_width_in_luma_samples, and pic_height_in_luma_samples are signaled in the SPS, and all layers in which the nuh_layer_id value referring to the SPS is greater than 0 use the signaled values. .
  • a layer whose nuh_layer_id value is greater than 0 does not use chroma_format_idc, separate_colour_plane_flag, pic_width_in_luma_samples, and pic_height_in_luma_samples related values signaled in the VPS.
  • update_chroma_size_flag has a value of 0, it indicates that chroma_format_idc, separate_colour_plane_flag, pic_width_in_luma_samples, and pic_height_in_luma_samples values are not signaled in the SPS, and that all layers referring to that SPS are signaled in the VPS: chroma_format_idc, separate_colour_ight_flag, pic_colour_plane_flag_information, do. If update_chroma_size_flag is not signaled, the update_chroma_size_flag value is assumed to be 1.
  • nuh_layer_id layerIdCurr> 0
  • chroma_format_idc separate_colour_plane_flag
  • chroma_format_idc, separate_colour_plane_flag, pic_width_in_luma_samples, pic_height_in_luma_samples values are vps_ref_format_idx [j] -th rep_format () of chroma_format_id_fla_pic_height_flag_picma_flag_pic_plane_flag
  • j is LayerIdxInVps [layerIdCurr]
  • chroma_format_idc, separate_colour_plane_flag, pic_width_in_luma_samples, and pic_height_in_luma_samples values signaled by the activated SPS are ignored.
  • chroma_format_idc When an active non-base layer refers to an SPS that is also used in the base layer, chroma_format_idc, separate_colour_plane_flag, pic_width_in_luma_samples, pic_height_in_luma_samples values are inferred from the VPS, where the SPS has a nuh_layer_id value of zero.
  • chroma_format_idc, separate_colour_plane_flag, pic_width_in_luma_samples, and pic_height_in_luma_samples values signaled in the activated SPS are applied.
  • nuh_layer_id value of the activated SPS is greater than 0, the following applies.
  • chroma_format_idc If the value of update_chroma_size_flag is 0, chroma_format_idc, separate_colour_plane_flag, pic_width_in_luma_samples, pic_height_in_luma_samples values are vsp_ref_format_idx [j] -th rep_format () of chroma_format_idc, separate_colour_fla_in_flag_information_flag. In this case, j is LayerIdxInVps [layerIdCurr].
  • update_chroma_size_flag chroma_format_idc, separate_colour_plane_flag, pic_width_in_luma_samples, pic_height_in_luma_samples values should be chroma_format_idc, separate_color_flag, separate_colour_plane_flag, separate_color_height_in_lu__s_in_lus
  • j is LayerIdxInVps [layerIdCurr].
  • update_bit_depth_flag has a value of 1
  • bit_depth_luma_minus8 and bit_depth_chroma_minus8 are signaled in the SPS, and all layers whose nuh_layer_id value referring to the SPS is greater than 0 use the signaled values.
  • all layers whose nuh_layer_id value referring to the SPS is greater than 0 use the signaled values.
  • values related to bit_depth_luma_minus8 and bit_depth_chroma_minus8 signaled by the VPS are not used.
  • update_bit_depth_flag When update_bit_depth_flag has a value of 0, it indicates that bit_depth_luma_minus8 and bit_depth_chroma_minus8 values are not signaled to the SPS, and all layers referring to the SPS use bit_depth_luma_minus8 and bit_depth_chroma_minus8 values that are signaled by the VPS. If update_bit_depth_flag is not signaled, the update_bit_depth_flag value is assumed to be 1.
  • bit_depth_luma_minus8 and bit_depth_chroma_minus8 may be regarded or restricted as the following values.
  • bit_depth_luma_minus8 and bit_depth_chroma_minus8 values are regarded as bit_depth_luma_minus8 and bit_depth_chroma_minus8 of vps_ref_format_idx [j] -th rep_format () of the activated VPS.
  • j is LayerIdxInVps [layerIdCurr]
  • bit_depth_luma_minus8 and bit_depth_chroma_minus8 values signaled by the activated SPS are ignored.
  • bit_depth_luma_minus8 and bit_depth_chroma_minus8 values are inferred from the VPS, where the SPS has a nuh_layer_id value of zero.
  • bit_depth_luma_minus8 and bit_depth_chroma_minus8 values signaled in the activated SPS are applied.
  • nuh_layer_id value of the activated SPS is greater than 0, the following applies.
  • bit_depth_luma_minus8 and bit_depth_chroma_minus8 values are considered bit_depth_luma_minus8 and bit_depth_chroma_minus8 of vps_ref_format_idx [j] -th rep_format () of the activated VPS.
  • j is LayerIdxInVps [layerIdCurr].
  • bit_depth_luma_minus8 and bit_depth_chroma_minus8 should be less than or equal to bit_depth_luma_minus8 of vps_ref_format_idx [j] -th rep_format () of the activated VPS.
  • j is LayerIdxInVps [layerIdCurr].
  • the number of temporal sub-layers per layer included in the bitstream may be signaled in the VPS extension.
  • Table 24 is an example of a syntax that describes a method for signaling the number of maximum temporal sub-layers according to an embodiment of the present invention.
  • sub_layers_vps_max_minus1 [i] indicates the maximum number of temporal sub-layers present in the layer whose nuh_layer_id value is layer_id_in_nuh [i].
  • the value of sub_layers_vps_max_minus1 [i] has a range from 0 to vps_max_sub_layers_minus1, and if the value of sub_layer_vps_max_minus1 [i] is not signaled, the value of sub_layers_vps_max_minus1 [i] is inferred as the value of vps_max_sub_layers_minus1.
  • sps_max_sub_layers_minus1 for indicating the maximum number of temporal sub-layers that the layer can have is signaled.
  • the SPS does not separately signal sps_max_sub_layers_minus1 to indicate the maximum number of temporal sub-layers that the layer may have when the layer referring to the SPS has a nuh_layer_id value greater than zero.
  • the value of vps_max_sub_layers_minus1 + 1 which is the number of temporal sub-layers present in the entire bitstream, is determined, not the number of temporal sub-layers of the corresponding layer.
  • the sps_max_sub_layers_minus1 value when the nuh_layer_id value is greater than 0 is determined as the sub_layers_vps_max_minus1 [i] value as follows. can do.
  • sps_max_sub_layers_minus1 + 1 indicates the maximum number of temporal sub-layers that the layer can have when the layer referring to the SPS has nuh_layer_id greater than zero.
  • sps_max_sub_layers_minus1 has a value between 0 and 6. If sps_max_sub_layers_minus1 is not signaled, the sps_max_sub_layers_minus1 value is inferred to be equal to the sub_layers_vps_max_minus1 [i] value of the i-th layer whose nuh_layer_id value is layer_id_in_nuh_ [i].
  • the vps_temporal_id_nesting_flag signaled in the VPS is used to indicate that switching-up of the temporal sublayer is possible for all layers of the bitstream. That is, when the value of vps_max_sub_layers_minus1 is greater than 0, the value of vps_temporal_id_nesting_flag indicates whether additional inter prediction is restricted for the CVS (coded video bitstream) referring to the corresponding VPS. If the value of vps_max_sub_layers_minus1 is 0, vps_temporal_id_nesting_flag must have a value of 1.
  • Table 25 is an example of a syntax illustrating a method of signaling a flag indicating whether switching up is possible for each layer according to an embodiment of the present invention.
  • vps_layer_temporal_id_nesting_flag [i] indicates whether inter-screen prediction is additionally restricted to the picture of the i-th layer when sub_layers_vps_max_minus1 [i] is greater than zero.
  • sub_layers_vps_max_minus [i] is 0, vps_layer_temporal_id_nesting_flag [i] must have a value of 1. If vps_layer_temporal_id_nesting_flag [i] is not signaled, it is inferred that vps_layer_temporal_id_nesting_flag [i] is equal to the vps_temporal_id_nesting_flag value.
  • semantics of sps_temporal_id_nesting_flag signaled in the SPS may be modified as follows.
  • sps_temporal_id_nesting_flag indicates whether additional inter prediction is restricted to CVS (coded video stream) referring to the SPS when sps_max_sub_layers_minus1 is greater than zero.
  • the sps_temporal_id_nesting_flag value When the vps_temporal_id_nesting_flag value is 1, the sps_temporal_id_nesting_flag value must be 1. If sps_max_sub_layers_minus1 is 0, the sps_temporal_id_nesting_flag value must be 1. If the sps_temporal_id_nesting_flag value is not signaled, it is inferred that the sps_temporal_id_nesting_flag value is the same as the vps_layer_temporal_id_nesting_flag [i] value of the i-th layer whose nuh_layer_id value is layer_id_in_nuh [i].
  • FIG. 5 is a flowchart schematically illustrating a method of acquiring tile related information in a scalable video coding structure supporting a plurality of layers according to an embodiment of the present invention.
  • the method of FIG. 5 may be performed by the image decoding apparatus (more specifically, the entropy decoding unit) of FIG. 2 described above. In the present embodiment, it is described as being performed by the decoding apparatus for convenience of explanation, but the reverse process may be performed by the encoding apparatus.
  • the decoding apparatus obtains information indicating whether tile related information exists (S500).
  • the decoding apparatus may receive a scalable bitstream and entropy decode it to obtain information indicating whether tile related information is present.
  • the information indicating whether the tile related information exists may be information indicating whether a picture including two or more tiles among the pictures in the layer exists as described above.
  • information indicating whether tile related information exists may be set as tile related information.
  • the information indicating whether the tile related information exists is related to the tile The information may be set to no present.
  • the decoding apparatus obtains tile related information based on information indicating whether tile related information exists (S510).
  • the decoding apparatus may obtain the entropy decoding of the tile related information.
  • the tile related information may be information indicating whether the position of the tile existing in the picture of the current layer corresponds to the position of the tile existing in the picture of the reference layer referenced by the current layer.
  • the tile related information may be tile_boundaries_aligned_flag.
  • Information indicating whether the tile related information exists and the tile related information may be transmitted from the encoder through the VPS, PPS, SPS and the like.
  • the process of acquiring the tile related information based on the information indicating whether the tile related information exists may be performed as in the syntaxes described in Tables 1 to 7, or It is also applicable to several combinations of these.
  • FIG. 6 is a flowchart schematically illustrating a method of acquiring representation format related information in a scalable video coding structure supporting a plurality of layers according to an embodiment of the present invention.
  • the method of FIG. 6 may be performed by the image decoding apparatus (more specifically, the entropy decoding unit) of FIG. 2 described above. In the present embodiment, it is described as being performed by the decoding apparatus for convenience of explanation, but the reverse process may be performed by the encoding apparatus.
  • the decoding apparatus obtains information indicating whether color information related to chroma format and bit depth related information exist (S600).
  • the decoding apparatus may receive a scalable bitstream and entropy decode it to obtain information indicating whether color difference format related information and bit depth related information exist.
  • the decoding apparatus obtains the color difference format related information and the bit depth related information based on the information indicating whether the color difference format related information and the bit depth related information exist.
  • the decoding apparatus may obtain the entropy decoding by using the color difference format related information and the bit depth related information.
  • the color difference format related information is information about color difference and luma sampling, and there are monochrome sampling, 4: 2: 0 sampling, 4: 2: 2 sampling, and 4: 4: 4 sampling color difference formats.
  • Motochrome sampling refers to a sample array consisting only of luma arrays.
  • 4: 2: 0 sampling means that each of the two color difference arrays is half the height and half the width of the luma array.
  • 4: 2: 2 sampling means that each of the two chroma arrays has a height and a half width of the luma array.
  • 4: 4: 4 Sampling means that each of the two color difference arrays is equal to the height and width of the luma array, or treated as monochrome sampling.
  • the color difference format related information refers to previously signaled color difference format related information
  • the bit depth related information is related to the previously signaled bit depth related information. You can refer to the information.
  • the information indicating whether the color difference format related information and the bit depth related information exist, the color difference format related information, and the bit depth related information include a rep_format () which is a syntax for describing information related to VPS, PPS, SPS, and expression format. Can be transmitted from the encoder.
  • the method according to the present invention described above may be stored in a computer-readable recording medium that is produced as a program for execution on a computer, and examples of the computer-readable recording medium include ROM, RAM, CD-ROM, magnetic tape , Floppy disks, optical data storage devices, and the like, and also include those implemented in the form of carrier waves (eg, transmission over the Internet).
  • the computer readable recording medium can be distributed over network coupled computer systems so that the computer readable code is stored and executed in a distributed fashion.
  • functional programs, codes, and code segments for implementing the method can be easily inferred by programmers in the art to which the present invention belongs.
  • the methods are described based on a flowchart as a series of steps or blocks, but the present invention is not limited to the order of steps, and certain steps may occur in a different order or at the same time than other steps described above. Can be. Also, one of ordinary skill in the art appreciates that the steps shown in the flowcharts are not exclusive, that other steps may be included, or that one or more steps in the flowcharts may be deleted without affecting the scope of the present invention. I can understand.

Landscapes

  • Engineering & Computer Science (AREA)
  • Multimedia (AREA)
  • Signal Processing (AREA)
  • Compression Or Coding Systems Of Tv Signals (AREA)

Abstract

L'invention concerne un procédé de décodage d'images et un appareil de décodage d'images qui prennent en charge une pluralité de couches. Le procédé de décodage d'images comporte les étapes consistant à: obtenir des informations indiquant s'il existe ou non des informations liées à des pavés; et obtenir les informations liées aux pavés en se basant sur les informations indiquant si les informations liées à des pavés existent ou non.
PCT/KR2014/009580 2013-10-14 2014-10-13 Procédé et appareil de codage/décodage d'images sur base multicouche WO2015056941A1 (fr)

Priority Applications (7)

Application Number Priority Date Filing Date Title
US15/028,570 US10091519B2 (en) 2013-10-14 2014-10-13 Multilayer-based image encoding/decoding method and apparatus
CN201910476959.5A CN110225349B (zh) 2013-10-14 2014-10-13 用于支持多层的视频编码/解码方法和计算机可读记录介质
CN201480062557.6A CN105723712B (zh) 2013-10-14 2014-10-13 基于多层的图像编码/解码方法和设备
US16/039,766 US10701379B2 (en) 2013-10-14 2018-07-19 Multilayer-based image encoding/decoding method and apparatus
US16/039,996 US10701380B2 (en) 2013-10-14 2018-07-19 Multilayer-based image encoding/decoding method and apparatus
US16/040,119 US10701381B2 (en) 2013-10-14 2018-07-19 Multilayer-based image encoding/decoding method and apparatus
US16/040,222 US10701382B2 (en) 2013-10-14 2018-07-19 Multilayer-based image encoding/decoding method and apparatus

Applications Claiming Priority (14)

Application Number Priority Date Filing Date Title
KR20130122252 2013-10-14
KR10-2013-0122252 2013-10-14
KR10-2013-0124205 2013-10-17
KR20130124205 2013-10-17
KR20130127469 2013-10-25
KR10-2013-0127469 2013-10-25
KR20130128202 2013-10-28
KR10-2013-0128202 2013-10-28
KR10-2014-0000426 2014-01-02
KR20140000426 2014-01-02
KR10-2014-0004131 2014-01-13
KR20140004131 2014-01-13
KR1020140137482A KR102290091B1 (ko) 2013-10-14 2014-10-13 다계층 기반의 영상 부호화/복호화 방법 및 장치
KR10-2014-0137482 2014-10-13

Related Child Applications (5)

Application Number Title Priority Date Filing Date
US15/028,570 A-371-Of-International US10091519B2 (en) 2013-10-14 2014-10-13 Multilayer-based image encoding/decoding method and apparatus
US16/040,222 Continuation US10701382B2 (en) 2013-10-14 2018-07-19 Multilayer-based image encoding/decoding method and apparatus
US16/039,996 Continuation US10701380B2 (en) 2013-10-14 2018-07-19 Multilayer-based image encoding/decoding method and apparatus
US16/039,766 Continuation US10701379B2 (en) 2013-10-14 2018-07-19 Multilayer-based image encoding/decoding method and apparatus
US16/040,119 Continuation US10701381B2 (en) 2013-10-14 2018-07-19 Multilayer-based image encoding/decoding method and apparatus

Publications (1)

Publication Number Publication Date
WO2015056941A1 true WO2015056941A1 (fr) 2015-04-23

Family

ID=52828327

Family Applications (1)

Application Number Title Priority Date Filing Date
PCT/KR2014/009580 WO2015056941A1 (fr) 2013-10-14 2014-10-13 Procédé et appareil de codage/décodage d'images sur base multicouche

Country Status (1)

Country Link
WO (1) WO2015056941A1 (fr)

Cited By (8)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
WO2017007208A1 (fr) * 2015-07-03 2017-01-12 에스케이텔레콤 주식회사 Dispositif et procédé d'extraction d'image à partir d'une image à haute résolution
JP2017513352A (ja) * 2014-03-18 2017-05-25 クゥアルコム・インコーポレイテッドQualcomm Incorporated マルチレイヤビットストリームのためのsps時間idネスティング情報の導出
WO2018004317A1 (fr) * 2016-07-01 2018-01-04 에스케이텔레콤 주식회사 Procédé et dispositif de génération de flux binaire d'image pour la diffusion en continu d'images à haute résolution
US10623816B2 (en) 2015-07-03 2020-04-14 Sk Telecom Co., Ltd. Method and apparatus for extracting video from high resolution video
US10743000B2 (en) 2016-07-01 2020-08-11 Sk Telecom Co., Ltd. Video bitstream generation method and device for high-resolution video streaming
CN114208165A (zh) * 2019-08-10 2022-03-18 北京字节跳动网络技术有限公司 子图片解码中的缓冲器管理
CN115088256A (zh) * 2020-02-14 2022-09-20 Lg电子株式会社 选择性地用信号通知滤波可用信息的图像编码/解码方法和装置以及发送比特流的方法
US11956432B2 (en) 2019-10-18 2024-04-09 Beijing Bytedance Network Technology Co., Ltd Interplay between subpictures and in-loop filtering

Citations (5)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
WO2012088211A1 (fr) * 2010-12-21 2012-06-28 Docomo Communications Laboratories Usa Inc. Codage prédictif intra amélioré utilisant des représentations planes
WO2013042884A1 (fr) * 2011-09-19 2013-03-28 엘지전자 주식회사 Procédé de codage/décodage d'image et dispositif associé
US20130083850A1 (en) * 2008-07-02 2013-04-04 Samsung Electronics Co., Ltd. Image encoding method and device, and decoding method and device therefor
KR20130047650A (ko) * 2011-10-28 2013-05-08 삼성전자주식회사 비디오의 인트라 예측 방법 및 장치
KR20130079261A (ko) * 2011-12-30 2013-07-10 (주)휴맥스 3차원 영상 부호화 방법 및 장치, 및 복호화 방법 및 장치

Patent Citations (5)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US20130083850A1 (en) * 2008-07-02 2013-04-04 Samsung Electronics Co., Ltd. Image encoding method and device, and decoding method and device therefor
WO2012088211A1 (fr) * 2010-12-21 2012-06-28 Docomo Communications Laboratories Usa Inc. Codage prédictif intra amélioré utilisant des représentations planes
WO2013042884A1 (fr) * 2011-09-19 2013-03-28 엘지전자 주식회사 Procédé de codage/décodage d'image et dispositif associé
KR20130047650A (ko) * 2011-10-28 2013-05-08 삼성전자주식회사 비디오의 인트라 예측 방법 및 장치
KR20130079261A (ko) * 2011-12-30 2013-07-10 (주)휴맥스 3차원 영상 부호화 방법 및 장치, 및 복호화 방법 및 장치

Cited By (14)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US11388441B2 (en) 2014-03-18 2022-07-12 Qualcomm Incorporated Derivation of SPS temporal ID nesting information for multi-layer bitstreams
JP2017513352A (ja) * 2014-03-18 2017-05-25 クゥアルコム・インコーポレイテッドQualcomm Incorporated マルチレイヤビットストリームのためのsps時間idネスティング情報の導出
US10623816B2 (en) 2015-07-03 2020-04-14 Sk Telecom Co., Ltd. Method and apparatus for extracting video from high resolution video
WO2017007208A1 (fr) * 2015-07-03 2017-01-12 에스케이텔레콤 주식회사 Dispositif et procédé d'extraction d'image à partir d'une image à haute résolution
WO2018004317A1 (fr) * 2016-07-01 2018-01-04 에스케이텔레콤 주식회사 Procédé et dispositif de génération de flux binaire d'image pour la diffusion en continu d'images à haute résolution
US10743000B2 (en) 2016-07-01 2020-08-11 Sk Telecom Co., Ltd. Video bitstream generation method and device for high-resolution video streaming
US10893278B2 (en) 2016-07-01 2021-01-12 Sk Telecom Co., Ltd. Video bitstream generation method and device for high-resolution video streaming
CN114208165A (zh) * 2019-08-10 2022-03-18 北京字节跳动网络技术有限公司 子图片解码中的缓冲器管理
CN114208165B (zh) * 2019-08-10 2023-12-15 北京字节跳动网络技术有限公司 子图片解码中的缓冲器管理
US12047558B2 (en) 2019-08-10 2024-07-23 Beijing Bytedance Network Technology Co., Ltd. Subpicture dependent signaling in video bitstreams
US12075030B2 (en) 2019-08-10 2024-08-27 Beijing Bytedance Network Technology Co., Ltd. Subpicture dependent signaling in video bitstreams
US11956432B2 (en) 2019-10-18 2024-04-09 Beijing Bytedance Network Technology Co., Ltd Interplay between subpictures and in-loop filtering
US11962771B2 (en) 2019-10-18 2024-04-16 Beijing Bytedance Network Technology Co., Ltd Syntax constraints in parameter set signaling of subpictures
CN115088256A (zh) * 2020-02-14 2022-09-20 Lg电子株式会社 选择性地用信号通知滤波可用信息的图像编码/解码方法和装置以及发送比特流的方法

Similar Documents

Publication Publication Date Title
WO2013157828A1 (fr) Procédé et dispositif de décodage destinés à un train de bits comportant une pluralité de couches
WO2014003379A1 (fr) Procédé de décodage d'image et appareil l'utilisant
WO2017086746A1 (fr) Procédé et appareil d'encodage/décodage d'un mode de prédiction intra-écran.
WO2014092407A1 (fr) Procédé de décodage d'image et appareil l'utilisant
WO2015056941A1 (fr) Procédé et appareil de codage/décodage d'images sur base multicouche
WO2014010943A1 (fr) Procédé et dispositif de codage/décodage d'image
WO2014038906A1 (fr) Procédé de décodage d'image et appareil utilisant celui-ci
WO2014104725A1 (fr) Procédé de codage/décodage d'images et appareil l'utilisant
WO2021040492A1 (fr) Procédé et dispositif permettant de coder un coefficient de transformée dans un système de codage vidéo/image
WO2015009036A1 (fr) Procédé et appareil de prédiction inter-couche en fonction d'informations de sous-couche temporelle
WO2021040319A1 (fr) Procédé et appareil pour dériver un paramètre rice dans un système de codage vidéo/image
WO2014107069A1 (fr) Procédé et dispositif d'encodage/de décodage d'image
WO2020231139A1 (fr) Codage vidéo ou d'image basé sur une cartographie de luminance et une mise à l'échelle chromatique
WO2020180122A1 (fr) Codage de vidéo ou d'images sur la base d'un modèle à alf analysé conditionnellement et d'un modèle de remodelage
WO2020017785A1 (fr) Procédé de décodage d'images à l'aide d'informations liées à l'intra-prédiction dans un système de codage d'images et appareil associé
WO2021066618A1 (fr) Codage d'image ou de vidéo basé sur la signalisation d'informations liées au saut de transformée et au codage de palette
WO2021066609A1 (fr) Codage d'image ou de vidéo basé sur un élément de syntaxe avancée lié au codage de saut de transformée et de palette
WO2021241963A1 (fr) Procédé de codage d'image sur la base d'informations de poc et d'un indicateur d'image non de référence dans un système de codage de vidéo ou d'image
WO2021235759A1 (fr) Procédé de codage d'image basé sur une architecture multicouche dans un système de codage d'image ou de vidéo
WO2021182802A1 (fr) Procédé de codage d'image basé sur des informations relatives à un pavé et des informations relatives à une tranche dans un système de codage de vidéo ou d'image
WO2021201513A1 (fr) Procédé et dispositif de codage/décodage d'image destinés à signaler des informations relatives à des ptl, dpb et hrd dans un sps, et support d'enregistrement lisible par ordinateur stockant un flux binaire
WO2013169049A1 (fr) Procédé de prédiction inter-couches et appareil utilisant ledit procédé
WO2020185036A1 (fr) Procédé et appareil de traitement de signal vidéo
WO2021066610A1 (fr) Codage d'image ou de vidéo sur la base de données relatives à un codage de saut de transformée et de palette
WO2021141462A1 (fr) Procédé et appareil de codage/décodage d'images utilisant la bdpcm, et support d'enregistrement stockant un flux binaire

Legal Events

Date Code Title Description
121 Ep: the epo has been informed by wipo that ep was designated in this application

Ref document number: 14853492

Country of ref document: EP

Kind code of ref document: A1

WWE Wipo information: entry into national phase

Ref document number: 15028570

Country of ref document: US

NENP Non-entry into the national phase

Ref country code: DE

122 Ep: pct application non-entry in european phase

Ref document number: 14853492

Country of ref document: EP

Kind code of ref document: A1