WO2021047590A1 - Signaling of subpicture structures - Google Patents
Signaling of subpicture structures Download PDFInfo
- Publication number
- WO2021047590A1 WO2021047590A1 PCT/CN2020/114508 CN2020114508W WO2021047590A1 WO 2021047590 A1 WO2021047590 A1 WO 2021047590A1 CN 2020114508 W CN2020114508 W CN 2020114508W WO 2021047590 A1 WO2021047590 A1 WO 2021047590A1
- Authority
- WO
- WIPO (PCT)
- Prior art keywords
- subpicture
- sequence
- video
- subpictures
- video pictures
- Prior art date
Links
- 230000011664 signaling Effects 0.000 title claims abstract description 15
- 238000000034 method Methods 0.000 claims abstract description 38
- 230000015654 memory Effects 0.000 description 22
- 238000012545 processing Methods 0.000 description 21
- 230000008569 process Effects 0.000 description 19
- 238000001914 filtration Methods 0.000 description 7
- 101150114515 CTBS gene Proteins 0.000 description 6
- 238000013139 quantization Methods 0.000 description 6
- 241000023320 Luma <angiosperm> Species 0.000 description 4
- 230000003044 adaptive effect Effects 0.000 description 4
- 239000011449 brick Substances 0.000 description 4
- OSWPMRLSEDHDFF-UHFFFAOYSA-N methyl salicylate Chemical compound COC(=O)C1=CC=CC=C1O OSWPMRLSEDHDFF-UHFFFAOYSA-N 0.000 description 4
- 230000008901 benefit Effects 0.000 description 3
- 238000004590 computer program Methods 0.000 description 3
- 230000003287 optical effect Effects 0.000 description 3
- 238000010276 construction Methods 0.000 description 2
- 238000000605 extraction Methods 0.000 description 2
- 238000009499 grossing Methods 0.000 description 2
- 238000013507 mapping Methods 0.000 description 2
- 239000013598 vector Substances 0.000 description 2
- 238000013459 approach Methods 0.000 description 1
- 238000003491 array Methods 0.000 description 1
- 230000008859 change Effects 0.000 description 1
- 230000000295 complement effect Effects 0.000 description 1
- 230000006835 compression Effects 0.000 description 1
- 238000007906 compression Methods 0.000 description 1
- 239000002355 dual-layer Substances 0.000 description 1
- 230000006870 function Effects 0.000 description 1
- 239000004973 liquid crystal related substance Substances 0.000 description 1
- 238000012986 modification Methods 0.000 description 1
- 230000004048 modification Effects 0.000 description 1
- 239000003607 modifier Substances 0.000 description 1
- 238000005192 partition Methods 0.000 description 1
- 230000002093 peripheral effect Effects 0.000 description 1
- 230000001681 protective effect Effects 0.000 description 1
- 239000007787 solid Substances 0.000 description 1
- 238000000638 solvent extraction Methods 0.000 description 1
- 230000003068 static effect Effects 0.000 description 1
- 230000002123 temporal effect Effects 0.000 description 1
Images
Classifications
-
- H—ELECTRICITY
- H04—ELECTRIC COMMUNICATION TECHNIQUE
- H04N—PICTORIAL COMMUNICATION, e.g. TELEVISION
- H04N19/00—Methods or arrangements for coding, decoding, compressing or decompressing digital video signals
- H04N19/70—Methods or arrangements for coding, decoding, compressing or decompressing digital video signals characterised by syntax aspects related to video coding, e.g. related to compression standards
-
- H—ELECTRICITY
- H04—ELECTRIC COMMUNICATION TECHNIQUE
- H04N—PICTORIAL COMMUNICATION, e.g. TELEVISION
- H04N19/00—Methods or arrangements for coding, decoding, compressing or decompressing digital video signals
- H04N19/10—Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using adaptive coding
- H04N19/102—Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using adaptive coding characterised by the element, parameter or selection affected or controlled by the adaptive coding
- H04N19/119—Adaptive subdivision aspects, e.g. subdivision of a picture into rectangular or non-rectangular coding blocks
-
- H—ELECTRICITY
- H04—ELECTRIC COMMUNICATION TECHNIQUE
- H04N—PICTORIAL COMMUNICATION, e.g. TELEVISION
- H04N19/00—Methods or arrangements for coding, decoding, compressing or decompressing digital video signals
- H04N19/10—Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using adaptive coding
- H04N19/134—Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using adaptive coding characterised by the element, parameter or criterion affecting or controlling the adaptive coding
- H04N19/167—Position within a video image, e.g. region of interest [ROI]
-
- H—ELECTRICITY
- H04—ELECTRIC COMMUNICATION TECHNIQUE
- H04N—PICTORIAL COMMUNICATION, e.g. TELEVISION
- H04N19/00—Methods or arrangements for coding, decoding, compressing or decompressing digital video signals
- H04N19/10—Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using adaptive coding
- H04N19/169—Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using adaptive coding characterised by the coding unit, i.e. the structural portion or semantic portion of the video signal being the object or the subject of the adaptive coding
-
- H—ELECTRICITY
- H04—ELECTRIC COMMUNICATION TECHNIQUE
- H04N—PICTORIAL COMMUNICATION, e.g. TELEVISION
- H04N19/00—Methods or arrangements for coding, decoding, compressing or decompressing digital video signals
- H04N19/10—Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using adaptive coding
- H04N19/169—Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using adaptive coding characterised by the coding unit, i.e. the structural portion or semantic portion of the video signal being the object or the subject of the adaptive coding
- H04N19/17—Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using adaptive coding characterised by the coding unit, i.e. the structural portion or semantic portion of the video signal being the object or the subject of the adaptive coding the unit being an image region, e.g. an object
-
- H—ELECTRICITY
- H04—ELECTRIC COMMUNICATION TECHNIQUE
- H04N—PICTORIAL COMMUNICATION, e.g. TELEVISION
- H04N19/00—Methods or arrangements for coding, decoding, compressing or decompressing digital video signals
- H04N19/10—Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using adaptive coding
- H04N19/169—Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using adaptive coding characterised by the coding unit, i.e. the structural portion or semantic portion of the video signal being the object or the subject of the adaptive coding
- H04N19/17—Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using adaptive coding characterised by the coding unit, i.e. the structural portion or semantic portion of the video signal being the object or the subject of the adaptive coding the unit being an image region, e.g. an object
- H04N19/174—Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using adaptive coding characterised by the coding unit, i.e. the structural portion or semantic portion of the video signal being the object or the subject of the adaptive coding the unit being an image region, e.g. an object the region being a slice, e.g. a line of blocks or a group of blocks
-
- H—ELECTRICITY
- H04—ELECTRIC COMMUNICATION TECHNIQUE
- H04N—PICTORIAL COMMUNICATION, e.g. TELEVISION
- H04N19/00—Methods or arrangements for coding, decoding, compressing or decompressing digital video signals
- H04N19/10—Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using adaptive coding
- H04N19/169—Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using adaptive coding characterised by the coding unit, i.e. the structural portion or semantic portion of the video signal being the object or the subject of the adaptive coding
- H04N19/17—Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using adaptive coding characterised by the coding unit, i.e. the structural portion or semantic portion of the video signal being the object or the subject of the adaptive coding the unit being an image region, e.g. an object
- H04N19/176—Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using adaptive coding characterised by the coding unit, i.e. the structural portion or semantic portion of the video signal being the object or the subject of the adaptive coding the unit being an image region, e.g. an object the region being a block, e.g. a macroblock
-
- H—ELECTRICITY
- H04—ELECTRIC COMMUNICATION TECHNIQUE
- H04N—PICTORIAL COMMUNICATION, e.g. TELEVISION
- H04N19/00—Methods or arrangements for coding, decoding, compressing or decompressing digital video signals
- H04N19/46—Embedding additional information in the video signal during the compression process
-
- H—ELECTRICITY
- H04—ELECTRIC COMMUNICATION TECHNIQUE
- H04N—PICTORIAL COMMUNICATION, e.g. TELEVISION
- H04N19/00—Methods or arrangements for coding, decoding, compressing or decompressing digital video signals
- H04N19/60—Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using transform coding
- H04N19/61—Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using transform coding in combination with predictive coding
-
- H—ELECTRICITY
- H04—ELECTRIC COMMUNICATION TECHNIQUE
- H04N—PICTORIAL COMMUNICATION, e.g. TELEVISION
- H04N19/00—Methods or arrangements for coding, decoding, compressing or decompressing digital video signals
- H04N19/85—Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using pre-processing or post-processing specially adapted for video compression
- H04N19/86—Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using pre-processing or post-processing specially adapted for video compression involving reduction of coding artifacts, e.g. of blockiness
- H04N19/865—Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using pre-processing or post-processing specially adapted for video compression involving reduction of coding artifacts, e.g. of blockiness with detection of the former encoding block subdivision in decompressed video
Definitions
- the present disclosure relates generally to video coding.
- the present disclosure relates to methods of signaling subpicture structures.
- High-efficiency video coding is the latest international video coding standard developed by the Joint Collaborative Team on Video Coding (JCT-VC) .
- the input video signal is predicted from the reconstructed signal, which is derived from the coded picture regions.
- the prediction residual signal is processed by a linear transform.
- the transform coefficients are quantized and entropy coded together with other side information in the bitstream.
- the reconstructed signal is generated from the prediction signal and the reconstructed residual signal after inverse transform on the de-quantized transform coefficients.
- the reconstructed signal is further processed by in-loop filtering for removing coding artifacts.
- the decoded pictures are stored in the frame buffer for predicting the future pictures in the input video signal.
- a coded picture is partitioned into non-overlapped square block regions represented by the associated coding tree units (CTUs) .
- a coded picture can be represented by a collection of slices, each comprising an integer number of CTUs. The individual CTUs in a slice are processed in a raster scanning order.
- a bi-predictive (B) slice may be decoded using intra prediction or inter prediction using at most two motion vectors and reference indices to predict the sample values of each block.
- a predictive (P) slice is decoded using intra prediction or inter prediction using at most one motion vector and reference index to predict the sample values of each block.
- An intra (I) slice is decoded using intra prediction only.
- a CTU can be partitioned into multiple non-overlapped coding units (CUs) using the recursive quadtree (QT) structure to adapt to various local motion and texture characteristics.
- a CTU may also be partitioned into one or multiple smaller size CUs by a quadtree with nested multi-type tree using binary and ternary split.
- the resulting CU partitions can be in square or rectangular shapes.
- One or more prediction units are specified for each CU.
- the prediction unit together with the associated CU syntax, works as a basic unit for signaling the predictor information.
- the specified prediction process is employed to predict the values of the associated pixel samples inside the PU.
- a CU can be further partitioned using the residual quadtree (RQT) structure for representing the associated prediction residual signal.
- the leaf nodes of the RQT correspond to the transform units (TUs) .
- a transform unit is comprised of a transform block (TB) of luma samples of size 8x8, 16x16, or 32x32 or four transform blocks of luma samples of size 4x4, and two corresponding transform blocks of chroma samples of a picture in 4: 2: 0 color format.
- An integer transform is applied to a transform block and the level values of quantized coefficients together with other side information are entropy coded in the bitstream.
- coding tree block CB
- CB coding block
- PB prediction block
- TB transform block
- Some embodiments provide a method for signaling subpicture structures for coded video.
- a video decoder receives data from a bitstream to be decoded as a sequence of video pictures.
- the video decoder receives from the bitstream subpicture specification for one or more subpictures in the sequence of video pictures.
- the subpicture specification identifies a position and a size for each subpicture by providing an index that identifies a coding tree unit (CTU) in the order of raster scan within a picture for the subpicture.
- CTU coding tree unit
- a syntax element in a sequence parameter set (SPS) of the sequence of video pictures indicates that one or more subpicture is present in the sequence of video pictures.
- the SPS may also include a syntax element that specifies the number of subpictures of the sequence of video pictures, as well as an identifier for each subpicture. Identifiers of subpictures may also be signaled in slice headers and/or picture parameter sets (PPSs) of video pictures in the sequence of video pictures.
- PPSs picture parameter sets
- a syntax element in the PPS of a video picture in the sequence of video picture indicates that all slices of the video picture are rectangular.
- the identified CTU in raster scan within a picture is at a corner (e.g., top-left or bottom-right) of the subpicture.
- the index is assigned to a subpicture grid in raster scan within a picture, and different subpicture grids are assigned different indices.
- the index identifies a subpicture grid that is defined to be one CTU or correspond to one CTU, such that the boundaries of the subpicture grids are defined along boundaries of CTUs.
- the indices of subpicture positions and sizes are signaled in a SPS of the sequence of video pictures.
- FIGS. 1a-e conceptually illustrate CTB or CTU based grid units being used to specify subpictures of a video sequence.
- FIG. 2 illustrates CTU or CTB-based subpicture grids that are indexed in the order of raster scan within a picture for specifying subpictures.
- FIG. 3 illustrates an example video encoder that support subpictures.
- FIG. 4 conceptually illustrates portions of the video encoder that implements signaling for subpictures.
- FIG. 5 conceptually illustrates a process for providing subpicture specification at a video encoder.
- FIG. 6 illustrates an example video decoder that support subpictures.
- FIG. 7 conceptually illustrates portions of the video decoder that implements signaling for subpictures.
- FIG. 8 conceptually illustrates a process for processing subpicture specification at a video decoder.
- FIG. 9 conceptually illustrates an electronic system with which some embodiments of the present disclosure are implemented.
- a subpicture is a rectangular region of one or more slices within a picture and a slice consists of one or more tiles/bricks. Each tile/brick is CTU aligned. When subpictures are present in a picture, the number of the subpictures may be greater than or equal to 2. Slices forming a subpicture may be rectangular. In some embodiments, grid in units of CTBs is used to specify subpicture structures within a picture by indicating the bottom-right subpicture grid index in the order of raster scan within a picture for each subpicture.
- a video encoder may signal the specification of a subpicture (and a video decoder may receive the specification of the subpicture) .
- Table 1A is an example syntax table for sequence parameter set (SPS) raw byte sequence payload (RBSP) that signal subpicture information:
- Table 1A SPS signaling subpicture information
- Table 1B is another example syntax table for sequence parameter set (SPS) raw byte sequence payload (RBSP) that provide specification of a subpicture:
- Table 1C is another example syntax table for sequence parameter set (SPS) raw byte sequence payload (RBSP) that provide specification of a subpicture:
- Table 1D is another example syntax table for sequence parameter set (SPS) raw byte sequence payload (RBSP) that provide specification of a subpicture:
- the syntax element subpics_present_flag being 1 indicates that subpicture parameters are present in the present in the SPS RBSP syntax. In some embodiments, when subpics_present_flag is equal to 1, the value of rect_slice_flag is set to 1. The syntax element subpics_present_flag being 0 indicates that subpicture parameters are not present in the present in the SPS RBSP syntax. In some embodiments, when a bitstream is the result of a sub-bitstream extraction process and contains only a subset of the subpictures of the input bitstream to the sub-bitstream extraction process, the value of subpics_present_flag may be set to 1 in the RBSP of the SPS.
- max_subpics_minus2 plus 2 specifies the maximum number of subpictures that may be present in a coded video sequence (CVS) .
- max_subpics_minus2 is constrained to be in the range of 0 to 254. The value of 255 is reserved for future use.
- the syntax element num_subpics_minus1 plus 1 specifies the number of subpictures that may be present in a CVS.
- the value of num_subpics_minus1 is constrained to be in the range of 0 to 254. The value of 255 is reserved for future use.
- the number of subpictures present in a CVS is directly signaled in the SPS through the syntax element num_subpics_minus1.
- the syntax element num_subpics_minus2 plus 2 specifies the number of subpictures that may be present in a CVS.
- the value of num_subpics_minus2 is constrained to be in the range of 0 to 254. The value of 255 is reserved for future use.
- the syntax element subpic_grid_col_width_minus1 plus 1 specifies the width of each element of the subpicture identifier grid, excluding the right-most grid column of the picture, in units of CtbSizeY.
- the length of the syntax element is Ceil (Log2 (pic_width_max_in_luma_samples /CtbSizeY) ) bits.
- subpic_grid_row_width_minus1 is inferred to be 0.
- the syntax element subpic_grid_row_height_minus1 plus 1 specifies the height of each element of the subpicture identifier grid, excluding the bottom grid row of the picture, in units of CtbSizeY.
- the length of the syntax element is Ceil (Log2 (pic_height_max_in_luma_samples /CtbSizeY) ) bits.
- subpic_grid_row_height_minus1 is inferred to be 0.
- the syntax element bottom_right_subpic_grid_idx_length_minus1 plus 1 specifies the number of bits used to represent the syntax element bottom_right_subpic_grid_idx_delta [i] .
- the value of bottom_right_subpic_grid_idx_length_minus1 shall be in the range of 0 to Ceil (Log2 (NumSubPicGridRows *NumSubPicGridCols) ) -1, inclusive.
- variable bottom_right_subpic_grid_idx_delta [i] when i is greater than 0 specifies the difference between the subpicture grid index of the bottom-right corner of the i-th subpicture and and the subpicture grid index of the bottom-right corner of the (i –1) -th subpicture.
- the variable bottom_right_subpic_grid_idx_delta [0] specifies the subpic grid index of the bottom right corner of the 0-th subpicture.
- syntax element subpic_grid_idx_delta_sign_flag [i] indicates a positive sign for bottom_right_subpic_grid_idx_delta [i] .
- sign_bottom_right_subpic_grid_idx_delta [i] indicates a negative sign for bottom_right_subpic_grid_idx_delta [i] .
- FIGS. 1a-e conceptually illustrate CTB or CTU based grid units being used to specify subpictures of a video sequence.
- FIG. 1a illustrates a video sequence (CVS) 100 that includes several video pictures. The video pictures in the sequence 100, such as a picture 110, are divided into CTUs.
- FIG. 1b illustrates the pictures of the sequence being divided into subpicture grids for specifying subpictures. Each subpicture grid 120 correspond to an integer number of CTUs such that each subpicture grid is defined based on boundaries of CTUs or CTBs.
- FIG. 1c illustrates an example in which each subpicture grid 120 correspond to exactly one CTU or CTB.
- FIG. 1a illustrates a video sequence (CVS) 100 that includes several video pictures. The video pictures in the sequence 100, such as a picture 110, are divided into CTUs.
- FIG. 1b illustrates the pictures of the sequence being divided into subpicture grids for specifying subpictures.
- Each subpicture grid 120 correspond to an
- each subpicture grid 120 corresponds to 2x1 CTUs or CTBs.
- FIG. 1e illustrates an example in which each subpicture grid 120 correspond to 2x3 CTUs or CTBs.
- SPS syntax elements such as subpic_grid_col_width_minus1 and subpic_grid_row_height_minus1 defines the subpicture grids in terms of CTUs or CTBs.
- the size of each CTU or CTB is signaled in the SPS in a syntax element log2_ctu_size_minus5.
- FIG. 2 illustrates CTU or CTB-based subpicture grids that are indexed for specifying subpictures.
- Each subpicture grid in the picture 110 corresponds to an index that can be used when a video codec is signaling the size (e.g., width, height) and/or position of a subpicture.
- the specified elements of the subpicture identifier grid are indexed in the order of ranster scan. In some embodiments, the subpictures are also indexed in the order of raster scan.
- the pictures of the sequence 100 are defined to have four subpictures 210, 220, 230, and 240 that are defined by using the CTU or CTB-based subpicture grids.
- the indices associated with the subpicture grids are used to specify the sizes and positions of the subpictures 210-240.
- the position of each subpicture is specified based on the index associated with the subpicture grid of a corner of the subpicture (e.g., bottom-right corner or top-left corner) .
- the position of the subpicture 210 is index 27
- the position of the subpicture 220 is index 36
- the position of the subpicture 230 is index 79
- the position of the subpicture 240 is the index 84.
- SPS syntax elements such as bottom_right_subpic_grid_idx_delta [i] and subpic_grid_idx_delta_sign_flag [i] are used to specify the bottom-right corner position of one subpicture by referencing the bottom right corner position of another subpicture.
- each CTU/CTB-based subpicture grid may be associated with a X-index and a Y-index, and the position of each subpicture may be specified by the X and Y indices of the subpicture grid at the top-left corner of the subpicture.
- the maximum number of subpictures (in a CVS) may be specified in a Video Parameter Set (VPS) .
- VPS Video Parameter Set Table 3 below is an example syntax table of a VPS that specifies the maximum number of subpictures in a CVS.
- the syntax element vps_max_subpics_minus2 plus 2 (or vps_max_subpics_minus1 plus 1) specifies the maximum allowed number of subpictures in each CVS referring to the VPS.
- the syntax element vps_max_subpics_minus2 is constrained to be in the range of 0 to 254. The value of 255 is reserved for future use.
- subpicture related parameters are signaled in a Picture Parameter Set (PPS) .
- PPS Picture Parameter Set
- Table 4 below illustrates an example syntax table of a PPS that includes subpicture information.
- a syntax element rect_slice_flag in the PPS being 0 specifies that bricks within each slice are in raster scan order and the slice information is not signalled in the PPS.
- the syntax element rect_slice_flag being 1 specifies that bricks within each slice cover a rectangular region of the picture and the slice information is signalled in the PPS.
- subpics_present_flag is equal to 1
- the value of rect_slice_flag shall be equal to 1.
- the value of rect_slice_flag is set to 1.
- the syntax element rect_slice_flag is inferred to be equal to 1.
- syntax element signalled_subpic_id_flag 1 specifies that the subpicture ID for each subpicture is signalled.
- syntax element signalled_subpic_id_flag 0 specifies that subpicture IDs are not signalled.
- the value of signalled_subpic_id_flag is inferred to be equal to 0.
- the syntax element signalled_subpic_id_length_minus1 plus 1 specifies the number of bits used to represent the syntax element subpic_id [i] when present, and the syntax element subpicture_id in slice headers.
- the value of signalled_subpic_id_length_minus1 is constrained to be in the range of 0 to 7, inclusive.
- the value of signalled_subpic_id_length_minus1 is inferred to be equal to Ceil (Log2 (Max (2, num_subpics_minus1 + 1) ) ) -1.
- the syntax element subpic_id [i] specifies the subpicture ID of the i-th subpicture.
- the length of the subpuic_id [i] syntax element is signalled_subpic_id_length_minus1 + 1 bits.
- the value of subpic_id [i] is inferred to be equal to i, for each i in the range of 0 to num_subpics_minus1, inclusive.
- the slices with the same subpicture ID collectively form the subpicture region.
- the syntax element subpicture_id specifies the subpicture ID to which the the current slice belongs.
- the length of the subpic_id syntax element is Ceil (Log2 (num_subpics_minus1 + 1) ) bits.
- the value of subpicture_id is a mapping to subpic_id [i] specified in the PPS.
- the signalled subpicture ID, subpic_id [i] for the i-th subpicture may be added in the PPS and the signalled subpicture ID, subpicure_id, mapping to subpic_id [i]in the PPS, may be added in the slice headers.
- Table 5 below illustrates an example syntax table of a slice header that includes subpicture information.
- subpicture IDs may be alternatively signaled in slice headers, PPS, and/or SPS of the CVS.
- the signalled subpicture ID, subpic_id [i] , for the i-th subpicture may be signaled in the SPS (rather than the PPS) according to Table 6 below:
- FIG. 3 illustrates an example video encoder 300 that support subpictures.
- the video encoder 300 receives input video signal from a video source 305 and encodes the signal into bitstream 395.
- the video encoder 300 has several components or modules for encoding the signal from the video source 305, at least including some components selected from a transform module 310, a quantization module 311, an inverse quantization module 314, an inverse transform module 315, an intra-picture estimation module 320, an intra-prediction module 325, a motion compensation module 330, a motion estimation module 335, an in-loop filter 345, a reconstructed picture buffer 350, a MV buffer 365, and a MV prediction module 375, and an entropy encoder 390.
- the motion compensation module 330 and the motion estimation module 335 are part of an inter-prediction module 340.
- the modules 310 –390 are modules of software instructions being executed by one or more processing units (e.g., a processor) of a computing device or electronic apparatus. In some embodiments, the modules 310 –390 are modules of hardware circuits implemented by one or more integrated circuits (ICs) of an electronic apparatus. Though the modules 310 –390 are illustrated as being separate modules, some of the modules can be combined into a single module.
- the video source 305 provides a raw video signal that presents pixel data of each video frame without compression.
- a subtractor 308 computes the difference between the raw video pixel data of the video source 305 and the predicted pixel data 313 from the motion compensation module 330 or intra-prediction module 325.
- the transform module 310 converts the difference (or the residual pixel data or residual signal 309) into transform coefficients (e.g., by performing Discrete Cosine Transform, or DCT) .
- the quantization module 311 quantizes the transform coefficients into quantized data (or quantized coefficients) 312, which is encoded into the bitstream 395 by the entropy encoder 390.
- the inverse quantization module 314 de-quantizes the quantized data (or quantized coefficients) 312 to obtain transform coefficients, and the inverse transform module 315 performs inverse transform on the transform coefficients to produce reconstructed residual 319.
- the reconstructed residual 319 is added with the predicted pixel data 313 to produce reconstructed pixel data 317.
- the reconstructed pixel data 317 is temporarily stored in a line buffer (not illustrated) for intra-picture prediction and spatial MV prediction.
- the reconstructed pixels are filtered by the in-loop filter 345 and stored in the reconstructed picture buffer 350.
- the reconstructed picture buffer 350 is a storage external to the video encoder 300.
- the reconstructed picture buffer 350 is a storage internal to the video encoder 300.
- the intra-picture estimation module 320 performs intra-prediction based on the reconstructed pixel data 317 to produce intra prediction data.
- the intra-prediction data is provided to the entropy encoder 390 to be encoded into bitstream 395.
- the intra-prediction data is also used by the intra-prediction module 325 to produce the predicted pixel data 313.
- the motion estimation module 335 performs inter-prediction by producing MVs to reference pixel data of previously decoded frames stored in the reconstructed picture buffer 350. These MVs are provided to the motion compensation module 330 to produce predicted pixel data.
- the video encoder 300 uses MV prediction to generate predicted MVs, and the difference between the MVs used for motion compensation and the predicted MVs is encoded as residual motion data and stored in the bitstream 395.
- the MV prediction module 375 generates the predicted MVs based on reference MVs that were generated for encoding previously video frames, i.e., the motion compensation MVs that were used to perform motion compensation.
- the MV prediction module 375 retrieves reference MVs from previous video frames from the MV buffer 365.
- the video encoder 300 stores the MVs generated for the current video frame in the MV buffer 365 as reference MVs for generating predicted MVs.
- the MV prediction module 375 uses the reference MVs to create the predicted MVs.
- the predicted MVs can be computed by spatial MV prediction or temporal MV prediction.
- the difference between the predicted MVs and the motion compensation MVs (MC MVs) of the current frame (residual motion data) are encoded into the bitstream 395 by the entropy encoder 390.
- the entropy encoder 390 encodes various parameters and data into the bitstream 395 by using entropy-coding techniques such as context-adaptive binary arithmetic coding (CABAC) or Huffman encoding.
- CABAC context-adaptive binary arithmetic coding
- the entropy encoder 390 encodes various header elements, flags, along with the quantized transform coefficients 312, and the residual motion data as syntax elements into the bitstream 395.
- the bitstream 395 is in turn stored in a storage device or transmitted to a decoder over a communications medium such as a network.
- the in-loop filter 345 performs filtering or smoothing operations on the reconstructed pixel data 317 to reduce the artifacts of coding, particularly at boundaries of pixel blocks.
- the filtering operation performed includes sample adaptive offset (SAO) .
- the filtering operations include adaptive loop filter (ALF) .
- FIG. 4 conceptually illustrates portions of the video encoder 300 that implements signaling for subpictures.
- the entropy encoder 390 receives signaling from the video source 305, which specifies parameters 410 for subpictures that are to be present in a currently coded sequence 420 of video pictures. These parameters 410 may indicate the number of subpictures that are present in a currently coded sequence. The parameters 410 may also indicate the positions and the geometries (height, width, and size) of each subpicture. The quantized coefficients 312 of the different subpictures are provided to the data path of the encoder 300.
- the entropy encoder 390 Based on the parameters for the subpictures, the entropy encoder 390 generates syntax elements that serve as specification for the subpictures in the bitstream 395. These syntax elements may include identifiers of subpictures, number of subpictures, as well as subpicture positions and geometries that are specified in terms of CTB/CTU-based subpicture grids. These syntax elements may be stored in the SPS of the currently coded video sequence 420, the PPSs of individual pictures in the video sequence, the slice headers of individual slices in the pictures of the sequence, and/or VPS of the entire video. Examples of these syntax elements are described by reference to Tables 1A-1D and 3-6 above.
- FIG. 5 conceptually illustrates a process 500 for providing subpicture specification at a video encoder.
- one or more processing units e.g., a processor
- the encoder 300 performs the process 500 by executing instructions stored in a computer readable medium.
- an electronic apparatus implementing the decoder 300 performs the process 500.
- the encoder receives (at block 510) data to be encoded as a sequence of video pictures in a bitstream.
- the encoder signals (at block 520) in the bitstream subpicture specification for one or more subpictures in the sequence of video pictures.
- a syntax element in a SPS of the sequence of video pictures indicates that one or more subpicture is present in the sequence of video pictures.
- the SPS may also include a syntax element that specifies the number of subpictures of the sequence of video pictures, as well as an identifier for each subpicture. Identifiers of subpictures may also be signaled in slice headers and/or PPS of video pictures in the sequence of video pictures.
- a syntax element in the PPS of a video picture in the sequence of video picture indicates that all slices of the video picture are rectangular.
- the encoder identifies (at block 530) a position and a size for each subpicture by providing an index that identifies a CTU for the subpicture830.
- the identified CTU is at a corner (e.g., top-left or bottom-right) of the subpicture.
- the index is assigned to a subpicture grid, and different subpicture grids are assigned different indices.
- the index identifies a subpicture grid that is defined to be one CTU or correspond to one CTU, such that the boundaries of the subpicture grids are defined along boundaries of CTUs.
- the indices of subpicture positions are signaled in a SPS of the sequence of video pictures.
- the encoder encodes (at block 540) each subpicture for the sequence of video pictures according to the subpicture specification.
- FIG. 6 illustrates an example video decoder 600 that support subpictures.
- the video decoder 600 is an image-decoding or video-decoding circuit that receives a bitstream 695 and decodes the content of the bitstream into pixel data of video frames for display.
- the video decoder 600 has several components or modules for decoding the bitstream 695, including some components selected from an inverse quantization module 611, an inverse transform module 610, an intra-prediction module 625, a motion compensation module 630, an in-loop filter 645, a decoded picture buffer 650, a MV buffer 665, a MV prediction module 675, and a parser 690.
- the motion compensation module 630 is part of an inter-prediction module 640.
- the modules 610 –690 are modules of software instructions being executed by one or more processing units (e.g., a processor) of a computing device. In some embodiments, the modules 610 –690 are modules of hardware circuits implemented by one or more ICs of an electronic apparatus. Though the modules 610 –690 are illustrated as being separate modules, some of the modules can be combined into a single module.
- the parser 690 receives the bitstream 695 and performs initial parsing according to the syntax defined by a video-coding or image-coding standard.
- the parsed syntax element includes various header elements, flags, as well as quantized data (or quantized coefficients) 612.
- the parser 690 parses out the various syntax elements by using entropy-coding techniques such as context-adaptive binary arithmetic coding (CABAC) or Huffman encoding.
- CABAC context-adaptive binary arithmetic coding
- Huffman encoding Huffman encoding
- the inverse quantization module 611 de-quantizes the quantized data (or quantized coefficients) 612 to obtain transform coefficients, and the inverse transform module 610 performs inverse transform on the transform coefficients 616 to produce reconstructed residual signal 619.
- the reconstructed residual signal 619 is added with predicted pixel data 613 from the intra-prediction module 625 or the motion compensation module 630 to produce decoded pixel data 617.
- the decoded pixels data are filtered by the in-loop filter 645 and stored in the decoded picture buffer 650.
- the decoded picture buffer 650 is a storage external to the video decoder 600.
- the decoded picture buffer 650 is a storage internal to the video decoder 600.
- the intra-prediction module 625 receives intra-prediction data from bitstream 695 and according to which, produces the predicted pixel data 613 from the decoded pixel data 617 stored in the decoded picture buffer 650.
- the decoded pixel data 617 is also stored in a line buffer (not illustrated) for intra-picture prediction and spatial MV prediction.
- the content of the decoded picture buffer 650 is used for display.
- a display device 655 either retrieves the content of the decoded picture buffer 650 for display directly, or retrieves the content of the decoded picture buffer to a display buffer.
- the display device receives pixel values from the decoded picture buffer 650 through a pixel transport.
- the motion compensation module 630 produces predicted pixel data 613 from the decoded pixel data 617 stored in the decoded picture buffer 650 according to motion compensation MVs (MC MVs) . These motion compensation MVs are decoded by adding the residual motion data received from the bitstream 695 with predicted MVs received from the MV prediction module 675.
- MC MVs motion compensation MVs
- the MV prediction module 675 generates the predicted MVs based on reference MVs that were generated for decoding previous video frames, e.g., the motion compensation MVs that were used to perform motion compensation.
- the MV prediction module 675 retrieves the reference MVs of previous video frames from the MV buffer 665.
- the video decoder 600 stores the motion compensation MVs generated for decoding the current video frame in the MV buffer 665 as reference MVs for producing predicted MVs.
- the in-loop filter 645 performs filtering or smoothing operations on the decoded pixel data 617 to reduce the artifacts of coding, particularly at boundaries of pixel blocks.
- the filtering operation performed includes sample adaptive offset (SAO) .
- the filtering operations include adaptive loop filter (ALF) .
- FIG. 7 conceptually illustrates portions of the video decoder 600 that implements signaling for subpictures.
- the entropy decoder 690 provides quantized coefficients 612 to the data path of the video decoder 600, which in turn generates pixel data to be displayed at the display device 655 for different subpictures.
- the display device may display the received pixel data according to subpicture parameters 710 for subpictures that are to be present in a currently coded sequence of pictures 720.
- the parameters 710 may also indicate the number of subpictures that are to be present in the currently coded sequence. These parameters may also indicate the positions and the geometries (size, height, width) of each subpicture.
- the entropy decoder 690 provide the subpicture parameters based on syntax elements decoded from the bitstream 695.
- the entropy decoder (parser) 690 receives syntax elements that serve as specification for the subpictures from the bitstream 695. These syntax elements may include identifiers of subpictures, number of subpictures, as well as subpicture positions and geometries that are specified in terms of CTB/CTU-based subpicture grids. These syntax elements may be stored in the SPS of the currently coded video sequence 720, the PPSs of individual pictures in the video sequence, the slice headers of individual slices in the pictures of the sequence, and/or VPS of the entire video. Examples of these syntax elements are described by reference to Tables 1A-1D and 3-6 above.
- FIG. 8 conceptually illustrates a process 800 for processing subpicture specification at a video decoder.
- one or more processing units e.g., a processor
- the decoder 600 performs the process 800 by executing instructions stored in a computer readable medium.
- an electronic apparatus implementing the decoder 600 performs the process 800.
- the decoder receives (at block 810) from a bitstream to be decoded as a sequence of video pictures.
- a syntax element in a SPS of the sequence of video pictures indicates that one or more subpicture is present in the sequence of video pictures.
- the SPS may also include a syntax element that specifies the number of subpictures of the sequence of video pictures, as well as an identifier for each subpicture. Identifiers of subpictures may also be signaled in slice headers and/or PPS of video pictures in the sequence of video pictures.
- a syntax element in the PPS of a video picture in the sequence of video picture indicates that all slices of the video picture are rectangular.
- the decoder identifies (at block 830) a position and a size for each subpicture by providing an index that identifies a CTU for the subpicture.
- the identified CTU is at a corner (e.g., top-left or bottom-right) of the subpicture.
- the index is assigned to a subpicture grid, and different subpicture grids are assigned different indices.
- the index identifies a subpicture grid that is defined to be one CTU or correspond to one CTU, such that the boundaries of the subpicture grids are defined along boundaries of CTUs.
- the indices of subpicture positions are signaled in a SPS of the sequence of video pictures.
- the decoder reconstructs (at block 840) each subpicture for the sequence of video pictures according to the subpicture specification.
- Computer readable storage medium also referred to as computer readable medium
- these instructions are executed by one or more computational or processing unit (s) (e.g., one or more processors, cores of processors, or other processing units) , they cause the processing unit (s) to perform the actions indicated in the instructions.
- computational or processing unit e.g., one or more processors, cores of processors, or other processing units
- Examples of computer readable media include, but are not limited to, CD-ROMs, flash drives, random-access memory (RAM) chips, hard drives, erasable programmable read only memories (EPROMs) , electrically erasable programmable read-only memories (EEPROMs) , etc.
- the computer readable media does not include carrier waves and electronic signals passing wirelessly or over wired connections.
- the term “software” is meant to include firmware residing in read-only memory or applications stored in magnetic storage which can be read into memory for processing by a processor.
- multiple software inventions can be implemented as sub-parts of a larger program while remaining distinct software inventions.
- multiple software inventions can also be implemented as separate programs.
- any combination of separate programs that together implement a software invention described here is within the scope of the present disclosure.
- the software programs when installed to operate on one or more electronic systems, define one or more specific machine implementations that execute and perform the operations of the software programs.
- FIG. 9 conceptually illustrates an electronic system 900 with which some embodiments of the present disclosure are implemented.
- the electronic system 900 may be a computer (e.g., a desktop computer, personal computer, tablet computer, etc. ) , phone, PDA, or any other sort of electronic device.
- Such an electronic system includes various types of computer readable media and interfaces for various other types of computer readable media.
- Electronic system 900 includes a bus 905, processing unit (s) 910, a graphics-processing unit (GPU) 915, a system memory 920, a network 925, a read-only memory 930, a permanent storage device 935, input devices 940, and output devices 945.
- the bus 905 collectively represents all system, peripheral, and chipset buses that communicatively connect the numerous internal devices of the electronic system 900.
- the bus 905 communicatively connects the processing unit (s) 910 with the GPU 915, the read-only memory 930, the system memory 920, and the permanent storage device 935.
- the processing unit (s) 910 retrieves instructions to execute and data to process in order to execute the processes of the present disclosure.
- the processing unit (s) may be a single processor or a multi-core processor in different embodiments. Some instructions are passed to and executed by the GPU 915.
- the GPU 915 can offload various computations or complement the image processing provided by the processing unit (s) 910.
- the read-only-memory (ROM) 930 stores static data and instructions that are used by the processing unit (s) 910 and other modules of the electronic system.
- the permanent storage device 935 is a read-and-write memory device. This device is a non-volatile memory unit that stores instructions and data even when the electronic system 900 is off. Some embodiments of the present disclosure use a mass-storage device (such as a magnetic or optical disk and its corresponding disk drive) as the permanent storage device 935.
- the system memory 920 is a read-and-write memory device. However, unlike storage device 935, the system memory 920 is a volatile read-and-write memory, such a random access memory.
- the system memory 920 stores some of the instructions and data that the processor uses at runtime.
- processes in accordance with the present disclosure are stored in the system memory 920, the permanent storage device 935, and/or the read-only memory 930.
- the various memory units include instructions for processing multimedia clips in accordance with some embodiments. From these various memory units, the processing unit (s) 910 retrieves instructions to execute and data to process in order to execute the processes of some embodiments.
- the bus 905 also connects to the input and output devices 940 and 945.
- the input devices 940 enable the user to communicate information and select commands to the electronic system.
- the input devices 940 include alphanumeric keyboards and pointing devices (also called “cursor control devices” ) , cameras (e.g., webcams) , microphones or similar devices for receiving voice commands, etc.
- the output devices 945 display images generated by the electronic system or otherwise output data.
- the output devices 945 include printers and display devices, such as cathode ray tubes (CRT) or liquid crystal displays (LCD) , as well as speakers or similar audio output devices. Some embodiments include devices such as a touchscreen that function as both input and output devices.
- CTR cathode ray tubes
- LCD liquid crystal displays
- bus 905 also couples electronic system 900 to a network 925 through a network adapter (not shown) .
- the computer can be a part of a network of computers (such as a local area network ( “LAN” ) , a wide area network ( “WAN” ) , or an Intranet, or a network of networks, such as the Internet.
- LAN local area network
- WAN wide area network
- Intranet a network of networks
- Some embodiments include electronic components, such as microprocessors, storage and memory that store computer program instructions in a machine-readable or computer-readable medium (alternatively referred to as computer-readable storage media, machine-readable media, or machine-readable storage media) .
- computer-readable media include RAM, ROM, read-only compact discs (CD-ROM) , recordable compact discs (CD-R) , rewritable compact discs (CD-RW) , read-only digital versatile discs (e.g., DVD-ROM, dual-layer DVD-ROM) , a variety of recordable/rewritable DVDs (e.g., DVD-RAM, DVD-RW, DVD+RW, etc.
- the computer-readable media may store a computer program that is executable by at least one processing unit and includes sets of instructions for performing various operations. Examples of computer programs or computer code include machine code, such as is produced by a compiler, and files including higher-level code that are executed by a computer, an electronic component, or a microprocessor using an interpreter.
- ASICs application specific integrated circuits
- FPGAs field programmable gate arrays
- integrated circuits execute instructions that are stored on the circuit itself.
- PLDs programmable logic devices
- ROM read only memory
- RAM random access memory
- the terms “computer” , “server” , “processor” , and “memory” all refer to electronic or other technological devices. These terms exclude people or groups of people.
- display or displaying means displaying on an electronic device.
- the terms “computer readable medium, ” “computer readable media, ” and “machine readable medium” are entirely restricted to tangible, physical objects that store information in a form that is readable by a computer. These terms exclude any wireless signals, wired download signals, and any other ephemeral signals.
- any two components so associated can also be viewed as being “operably connected” , or “operably coupled” , to each other to achieve the desired functionality, and any two components capable of being so associated can also be viewed as being “operably couplable” , to each other to achieve the desired functionality.
- operably couplable include but are not limited to physically mateable and/or physically interacting components and/or wirelessly interactable and/or wirelessly interacting components and/or logically interacting and/or logically interactable components.
Landscapes
- Engineering & Computer Science (AREA)
- Multimedia (AREA)
- Signal Processing (AREA)
- Compression Or Coding Systems Of Tv Signals (AREA)
- Television Signal Processing For Recording (AREA)
Priority Applications (4)
Application Number | Priority Date | Filing Date | Title |
---|---|---|---|
MX2022002854A MX2022002854A (es) | 2019-09-10 | 2020-09-10 | Señalización de estructuras subimagen. |
CN202080063649.1A CN114375579A (zh) | 2019-09-10 | 2020-09-10 | 子图像结构的发送技术 |
KR1020227009223A KR20220045231A (ko) | 2019-09-10 | 2020-09-10 | 서브화상 구조의 시그널링 |
EP20863420.4A EP4005205A4 (en) | 2019-09-10 | 2020-09-10 | SIGNALING OF SUB-PICTURE STRUCTURES |
Applications Claiming Priority (6)
Application Number | Priority Date | Filing Date | Title |
---|---|---|---|
US201962898127P | 2019-09-10 | 2019-09-10 | |
US62/898,127 | 2019-09-10 | ||
US201962898620P | 2019-09-11 | 2019-09-11 | |
US62/898,620 | 2019-09-11 | ||
US17/016,257 US11284077B2 (en) | 2019-09-10 | 2020-09-09 | Signaling of subpicture structures |
US17/016,257 | 2020-09-09 |
Publications (1)
Publication Number | Publication Date |
---|---|
WO2021047590A1 true WO2021047590A1 (en) | 2021-03-18 |
Family
ID=74866107
Family Applications (1)
Application Number | Title | Priority Date | Filing Date |
---|---|---|---|
PCT/CN2020/114508 WO2021047590A1 (en) | 2019-09-10 | 2020-09-10 | Signaling of subpicture structures |
Country Status (6)
Country | Link |
---|---|
EP (1) | EP4005205A4 (es) |
KR (1) | KR20220045231A (es) |
CN (1) | CN114375579A (es) |
MX (1) | MX2022002854A (es) |
TW (1) | TWI750802B (es) |
WO (1) | WO2021047590A1 (es) |
Citations (4)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
WO2018221368A1 (ja) * | 2017-05-31 | 2018-12-06 | シャープ株式会社 | 動画像復号装置、及び動画像符号化装置 |
US20180376126A1 (en) * | 2017-06-26 | 2018-12-27 | Nokia Technologies Oy | Apparatus, a method and a computer program for omnidirectional video |
CN109587478A (zh) * | 2017-09-29 | 2019-04-05 | 华为技术有限公司 | 一种媒体信息的处理方法及装置 |
US20190174141A1 (en) * | 2012-02-02 | 2019-06-06 | Texas Instruments Incorporated | Sub-pictures for pixel rate balancing on multi-core platforms |
Family Cites Families (2)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
WO2016203881A1 (ja) * | 2015-06-18 | 2016-12-22 | シャープ株式会社 | 算術復号装置及び算術符号化装置 |
US11363307B2 (en) * | 2019-08-08 | 2022-06-14 | Hfi Innovation Inc. | Video coding with subpictures |
-
2020
- 2020-09-10 KR KR1020227009223A patent/KR20220045231A/ko not_active Application Discontinuation
- 2020-09-10 MX MX2022002854A patent/MX2022002854A/es unknown
- 2020-09-10 WO PCT/CN2020/114508 patent/WO2021047590A1/en unknown
- 2020-09-10 EP EP20863420.4A patent/EP4005205A4/en active Pending
- 2020-09-10 CN CN202080063649.1A patent/CN114375579A/zh active Pending
- 2020-09-10 TW TW109131117A patent/TWI750802B/zh active
Patent Citations (4)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
US20190174141A1 (en) * | 2012-02-02 | 2019-06-06 | Texas Instruments Incorporated | Sub-pictures for pixel rate balancing on multi-core platforms |
WO2018221368A1 (ja) * | 2017-05-31 | 2018-12-06 | シャープ株式会社 | 動画像復号装置、及び動画像符号化装置 |
US20180376126A1 (en) * | 2017-06-26 | 2018-12-27 | Nokia Technologies Oy | Apparatus, a method and a computer program for omnidirectional video |
CN109587478A (zh) * | 2017-09-29 | 2019-04-05 | 华为技术有限公司 | 一种媒体信息的处理方法及装置 |
Non-Patent Citations (7)
Title |
---|
BENJAMIN BROSS , JIANLE CHEN , SHAN LIU: "Versatile Video Coding (Draft 6)", 127. MPEG MEETING; 20190708 - 20190712; GOTHENBURG; (MOTION PICTURE EXPERT GROUP OR ISO/IEC JTC1/SC29/WG11), GOTHENBURG SE, no. JVET-O2001-VE, 31 July 2019 (2019-07-31), Gothenburg SE, pages 1 - 455, XP030208568 * |
BROSS B ET AL., HIGH EFFICIENCY VIDEO CODING (HEVC) TEXT SPECIFICATION DRAFT 6 |
CHRISTIAN FELDMAN ET AL., EFFICIENT STREAMING-REASSEMBLING FOR VIDEO CONFERENCING APPLICATIONS USING TILES IN HEVC |
J. BOYCE (INTEL), LIDONG XU (INTEL): "AHG12: Sub-pictures and sub-picture sets with level derivation", 15. JVET MEETING; 20190703 - 20190712; GOTHENBURG; (THE JOINT VIDEO EXPLORATION TEAM OF ISO/IEC JTC1/SC29/WG11 AND ITU-T SG.16 ), 25 June 2019 (2019-06-25), XP030219894 * |
R. SKUPIN, Y. SANCHEZ, K. SUEHRING, T. SCHIERL (HHI): "AHG12: On filtering of independently coded region", 15. JVET MEETING; 20190703 - 20190712; GOTHENBURG; (THE JOINT VIDEO EXPLORATION TEAM OF ISO/IEC JTC1/SC29/WG11 AND ITU-T SG.16 ), no. m48619, 9 July 2019 (2019-07-09), XP030219731 * |
See also references of EP4005205A4 |
Y-K WANG ET AL., AHG12: SUB-PICTURE BASED VIDEO CODING |
Also Published As
Publication number | Publication date |
---|---|
EP4005205A4 (en) | 2023-08-30 |
TW202116068A (zh) | 2021-04-16 |
KR20220045231A (ko) | 2022-04-12 |
TWI750802B (zh) | 2021-12-21 |
EP4005205A1 (en) | 2022-06-01 |
CN114375579A (zh) | 2022-04-19 |
MX2022002854A (es) | 2023-04-14 |
Similar Documents
Publication | Publication Date | Title |
---|---|---|
US11546587B2 (en) | Adaptive loop filter with adaptive parameter set | |
WO2021088855A1 (en) | Signaling high-level information in video and image coding | |
US10887594B2 (en) | Entropy coding of coding units in image and video data | |
US11284077B2 (en) | Signaling of subpicture structures | |
US11924426B2 (en) | Signaling block partitioning of image and video | |
US11936890B2 (en) | Video coding using intra sub-partition coding mode | |
US11405649B2 (en) | Specifying slice chunks of a slice within a tile | |
US11778235B2 (en) | Signaling coding of transform-skipped blocks | |
US20210176475A1 (en) | Specifying Layout In Video Pictures | |
US11785214B2 (en) | Specifying video picture information | |
WO2021164736A1 (en) | Constraints for inter-layer referencing | |
US10999604B2 (en) | Adaptive implicit transform setting | |
WO2021047590A1 (en) | Signaling of subpicture structures | |
WO2023197998A1 (en) | Extended block partition types for video coding | |
WO2023198110A1 (en) | Block partitioning image and video data | |
WO2023236775A1 (en) | Adaptive coding image and video data | |
WO2024222411A1 (en) | Entropy coding transform blocks | |
WO2024169978A1 (en) | Ctu slice-based parallel processing in video coding | |
WO2024016955A1 (en) | Out-of-boundary check in video coding | |
WO2023241347A1 (en) | Adaptive regions for decoder-side intra mode derivation and prediction | |
WO2023198105A1 (en) | Region-based implicit intra mode derivation and prediction | |
WO2023198187A1 (en) | Template-based intra mode derivation and prediction |
Legal Events
Date | Code | Title | Description |
---|---|---|---|
121 | Ep: the epo has been informed by wipo that ep was designated in this application |
Ref document number: 20863420 Country of ref document: EP Kind code of ref document: A1 |
|
ENP | Entry into the national phase |
Ref document number: 2020863420 Country of ref document: EP Effective date: 20220224 |
|
NENP | Non-entry into the national phase |
Ref country code: DE |
|
ENP | Entry into the national phase |
Ref document number: 20227009223 Country of ref document: KR Kind code of ref document: A |