WO2019146718A1 - Dispositif et procédé de codage, et dispositif et procédé de décodage - Google Patents

Dispositif et procédé de codage, et dispositif et procédé de décodage Download PDF

Info

Publication number
WO2019146718A1
WO2019146718A1 PCT/JP2019/002341 JP2019002341W WO2019146718A1 WO 2019146718 A1 WO2019146718 A1 WO 2019146718A1 JP 2019002341 W JP2019002341 W JP 2019002341W WO 2019146718 A1 WO2019146718 A1 WO 2019146718A1
Authority
WO
WIPO (PCT)
Prior art keywords
block
sub
predetermined value
unit
encoding
Prior art date
Application number
PCT/JP2019/002341
Other languages
English (en)
Japanese (ja)
Inventor
安倍 清史
西 孝啓
遠間 正真
龍一 加納
Original Assignee
パナソニック インテレクチュアル プロパティ コーポレーション オブ アメリカ
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Application filed by パナソニック インテレクチュアル プロパティ コーポレーション オブ アメリカ filed Critical パナソニック インテレクチュアル プロパティ コーポレーション オブ アメリカ
Publication of WO2019146718A1 publication Critical patent/WO2019146718A1/fr

Links

Images

Classifications

    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04NPICTORIAL COMMUNICATION, e.g. TELEVISION
    • H04N19/00Methods or arrangements for coding, decoding, compressing or decompressing digital video signals
    • H04N19/10Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using adaptive coding
    • H04N19/102Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using adaptive coding characterised by the element, parameter or selection affected or controlled by the adaptive coding
    • H04N19/119Adaptive subdivision aspects, e.g. subdivision of a picture into rectangular or non-rectangular coding blocks
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04NPICTORIAL COMMUNICATION, e.g. TELEVISION
    • H04N19/00Methods or arrangements for coding, decoding, compressing or decompressing digital video signals
    • H04N19/10Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using adaptive coding
    • H04N19/134Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using adaptive coding characterised by the element, parameter or criterion affecting or controlling the adaptive coding
    • H04N19/157Assigned coding mode, i.e. the coding mode being predefined or preselected to be further used for selection of another element or parameter
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04NPICTORIAL COMMUNICATION, e.g. TELEVISION
    • H04N19/00Methods or arrangements for coding, decoding, compressing or decompressing digital video signals
    • H04N19/10Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using adaptive coding
    • H04N19/169Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using adaptive coding characterised by the coding unit, i.e. the structural portion or semantic portion of the video signal being the object or the subject of the adaptive coding
    • H04N19/17Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using adaptive coding characterised by the coding unit, i.e. the structural portion or semantic portion of the video signal being the object or the subject of the adaptive coding the unit being an image region, e.g. an object
    • H04N19/176Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using adaptive coding characterised by the coding unit, i.e. the structural portion or semantic portion of the video signal being the object or the subject of the adaptive coding the unit being an image region, e.g. an object the region being a block, e.g. a macroblock

Definitions

  • the present disclosure relates to an encoding device, a decoding device, an encoding method, and a decoding method.
  • HEVC High-Efficiency Video Coding
  • JCT-VC Joint Collaborative Team on Video Coding
  • the present disclosure aims to provide an encoding device, a decoding device, an encoding method, or a decoding method that can realize further improvement.
  • An encoding apparatus includes a circuit and a memory, and the circuit divides the processing target block into sub blocks using the memory and performs motion compensation in units of sub blocks.
  • encoding processing is performed by switching the horizontal and vertical lengths of the sub block based on the size of the processing target block.
  • a decoding device includes a circuit and a memory, and the circuit divides the processing target block into sub blocks using the memory and performs motion compensation in units of sub blocks.
  • the decoding process is performed by switching the horizontal and vertical lengths of the sub block based on the size of the processing target block.
  • the present disclosure can provide a coding device, a decoding device, a coding method or a decoding method that can realize further improvement.
  • FIG. 1 is a block diagram showing a functional configuration of the coding apparatus according to the first embodiment.
  • FIG. 2 is a diagram showing an example of block division in the first embodiment.
  • FIG. 3 is a table showing transform basis functions corresponding to each transform type.
  • FIG. 4A is a view showing an example of the shape of a filter used in ALF.
  • FIG. 4B is a view showing another example of the shape of a filter used in ALF.
  • FIG. 4C is a view showing another example of the shape of a filter used in ALF.
  • FIG. 5A is a diagram illustrating 67 intra prediction modes in intra prediction.
  • FIG. 5B is a flowchart for describing an outline of predicted image correction processing by OBMC processing.
  • FIG. 5A is a diagram illustrating 67 intra prediction modes in intra prediction.
  • FIG. 5B is a flowchart for describing an outline of predicted image correction processing by OBMC processing.
  • FIG. 5C is a conceptual diagram for describing an outline of predicted image correction processing by OBMC processing.
  • FIG. 5D is a diagram illustrating an example of FRUC.
  • FIG. 6 is a diagram for describing pattern matching (bilateral matching) between two blocks along a motion trajectory.
  • FIG. 7 is a diagram for describing pattern matching (template matching) between a template in a current picture and a block in a reference picture.
  • FIG. 8 is a diagram for explaining a model assuming uniform linear motion.
  • FIG. 9A is a diagram for describing derivation of a motion vector in units of sub blocks based on motion vectors of a plurality of adjacent blocks.
  • FIG. 9B is a diagram for describing an overview of motion vector derivation processing in the merge mode.
  • FIG. 9A is a diagram for describing derivation of a motion vector in units of sub blocks based on motion vectors of a plurality of adjacent blocks.
  • FIG. 9B is a diagram for describing an
  • FIG. 9C is a conceptual diagram for describing an overview of DMVR processing.
  • FIG. 9D is a diagram for describing an outline of a predicted image generation method using luminance correction processing by LIC processing.
  • FIG. 10 is a block diagram showing a functional configuration of the decoding apparatus according to the first embodiment.
  • FIG. 11 is a flowchart showing a first example of the inter prediction process in the first embodiment.
  • FIG. 12 is a diagram for describing a specific example of the size of the sub block obtained by dividing the processing target block.
  • FIG. 13 is a diagram for explaining an ATM VP mode which is an example of the sub book unit merge mode.
  • FIG. 14 is a diagram for describing an STMVP mode which is another example of the sub book unit merge mode.
  • FIG. 15 is an overall configuration diagram of a content supply system for realizing content distribution service.
  • FIG. 16 is a diagram illustrating an example of a coding structure at the time of scalable coding.
  • FIG. 17 is a diagram illustrating an example of a coding structure at the time of scalable coding.
  • FIG. 18 is a diagram showing an example of a display screen of a web page.
  • FIG. 19 is a view showing an example of a display screen of a web page.
  • FIG. 20 is a diagram illustrating an example of a smartphone.
  • FIG. 21 is a block diagram showing a configuration example of a smartphone.
  • Embodiment 1 First, an outline of the first embodiment will be described as an example of an encoding apparatus and a decoding apparatus to which the process and / or the configuration described in each aspect of the present disclosure described later can be applied.
  • Embodiment 1 is merely an example of an encoding apparatus and a decoding apparatus to which the processing and / or configuration described in each aspect of the present disclosure can be applied, and the processing and / or configuration described in each aspect of the present disclosure Can also be implemented in an encoding apparatus and a decoding apparatus different from those in the first embodiment.
  • the manner of implementation of the processing and / or configuration described in each aspect of the present disclosure is not limited to the above example.
  • the process may be implemented in an apparatus used for a purpose different from that of the moving picture / image coding apparatus or the moving picture / image decoding apparatus disclosed in the first embodiment, or the processing and / or configuration described in each aspect. May be implemented alone. Also, the processes and / or configurations described in the different embodiments may be implemented in combination.
  • FIG. 1 is a block diagram showing a functional configuration of coding apparatus 100 according to the first embodiment.
  • the encoding device 100 is a moving image / image coding device that encodes a moving image / image in units of blocks.
  • the encoding apparatus 100 is an apparatus for encoding an image in units of blocks, and includes a dividing unit 102, a subtracting unit 104, a converting unit 106, a quantizing unit 108, and entropy coding.
  • Unit 110 inverse quantization unit 112, inverse transformation unit 114, addition unit 116, block memory 118, loop filter unit 120, frame memory 122, intra prediction unit 124, inter prediction unit 126, And a prediction control unit 128.
  • the encoding device 100 is realized by, for example, a general-purpose processor and a memory.
  • the processor controls the division unit 102, the subtraction unit 104, the conversion unit 106, the quantization unit 108, the entropy coding unit 110, and the dequantization unit 112.
  • the inverse transform unit 114, the addition unit 116, the loop filter unit 120, the intra prediction unit 124, the inter prediction unit 126, and the prediction control unit 128 function.
  • coding apparatus 100 includes division section 102, subtraction section 104, conversion section 106, quantization section 108, entropy coding section 110, inverse quantization section 112, inverse conversion section 114, addition section 116, and loop filter section 120. , And may be realized as one or more dedicated electronic circuits corresponding to the intra prediction unit 124, the inter prediction unit 126, and the prediction control unit 128.
  • the intra prediction unit 124 may correct the pixel value after intra prediction based on the gradient of the reference pixel in the horizontal / vertical directions. Intra prediction with such correction is sometimes called position dependent intra prediction combination (PDPC). Information indicating the presence or absence of application of PDPC (for example, called a PDPC flag) is signaled, for example, at CU level. Note that the signaling of this information need not be limited to the CU level, but may be at other levels (eg, sequence level, picture level, slice level, tile level or CTU level).
  • the inter prediction signal may be generated using not only the motion information of the current block obtained by the motion search but also the motion information of the adjacent block. Specifically, the inter prediction signal is generated in units of sub blocks in the current block by weighting and adding a prediction signal based on motion information obtained by motion search and a prediction signal based on motion information of an adjacent block. It may be done.
  • Such inter prediction (motion compensation) may be called OBMC (overlapped block motion compensation).
  • the motion vector (MV_L) of the encoded left adjacent block is applied to the current block to obtain a predicted image (Pred_L), and the predicted image and Pred_L are weighted and superimposed. Perform the first correction of the image.
  • the processing target block may be a prediction block unit or a sub block unit obtained by further dividing the prediction block.
  • obmc_flag is a signal indicating whether to apply the OBMC process.
  • the encoding apparatus it is determined whether the encoding target block belongs to a complex area of motion, and if it belongs to a complex area of motion, the value 1 is set as obmc_flag. The encoding is performed by applying the OBMC processing, and when not belonging to the complex region of motion, the value 0 is set as the obmc_flag and the encoding is performed without applying the OBMC processing.
  • the decoding apparatus decodes the obmc_flag described in the stream to switch whether to apply the OBMC process according to the value and performs decoding.
  • the mode in which motion estimation is performed on the side of the decoding apparatus may be referred to as a pattern matched motion vector derivation (PMMVD) mode or a frame rate up-conversion (FRUC) mode.
  • PMMVD pattern matched motion vector derivation
  • FRUC frame rate up-conversion
  • FIG. 5D An example of the FRUC process is shown in FIG. 5D.
  • a plurality of candidate lists (which may be common to the merge list) each having a predicted motion vector are generated Be done.
  • the best candidate MV is selected from among the plurality of candidate MVs registered in the candidate list. For example, an evaluation value of each candidate included in the candidate list is calculated, and one candidate is selected based on the evaluation value.
  • first pattern matching or second pattern matching is used as pattern matching.
  • the first pattern matching and the second pattern matching may be referred to as bilateral matching and template matching, respectively.
  • FIG. 6 is a diagram for explaining an example of pattern matching (bilateral matching) between two blocks along a motion trajectory.
  • First pattern matching among pairs of two blocks in two reference pictures (Ref0, Ref1) which are two blocks along the motion trajectory of the current block (Cur block), Two motion vectors (MV0, MV1) are derived by searching for the most matching pair. Specifically, for the current block, a reconstructed image at a designated position in the first encoded reference picture (Ref 0) designated by the candidate MV, and a symmetric MV obtained by scaling the candidate MV at a display time interval.
  • FIG. 7 is a diagram for explaining an example of pattern matching (template matching) between a template in a current picture and a block in a reference picture.
  • the current block (Cur Pic) is searched for in the reference picture (Ref 0) for a block that most closely matches a block adjacent to the current block (Cur block).
  • Motion vectors are derived.
  • a FRUC flag Information indicating whether to apply such a FRUC mode (for example, called a FRUC flag) is signaled at the CU level.
  • a signal for example, called a FRUC mode flag
  • a method of pattern matching for example, first pattern matching or second pattern matching
  • the signaling of these pieces of information need not be limited to the CU level, but may be at other levels (eg, sequence level, picture level, slice level, tile level, CTU level or subblock level) .
  • This mode is sometimes referred to as a bi-directional optical flow (BIO) mode.
  • BIO bi-directional optical flow
  • FIG. 8 is a diagram for explaining a model assuming uniform linear motion.
  • (v x , v y ) indicate velocity vectors
  • ⁇ 0 and ⁇ 1 indicate the time between the current picture (Cur Pic) and two reference pictures (Ref 0 and Ref 1 ), respectively.
  • (MVx 0 , MVy 0 ) indicates a motion vector corresponding to the reference picture Ref 0
  • (MVx 1 , MVy 1 ) indicates a motion vector corresponding to the reference picture Ref 1 .
  • the optical flow equation is: (i) the time derivative of the luminance value, (ii) the product of the horizontal velocity and the horizontal component of the spatial gradient of the reference image, and (iii) the vertical velocity and the spatial gradient of the reference image Indicates that the product of the vertical components of and the sum of is equal to zero.
  • a motion vector in units of blocks obtained from a merge list or the like is corrected in units of pixels.
  • the motion vector may be derived on the decoding device side by a method different from the derivation of the motion vector based on a model assuming uniform linear motion.
  • motion vectors may be derived on a subblock basis based on motion vectors of a plurality of adjacent blocks.
  • This mode is sometimes referred to as affine motion compensation prediction mode.
  • FIG. 9A is a diagram for describing derivation of a motion vector in units of sub blocks based on motion vectors of a plurality of adjacent blocks.
  • the current block includes sixteen 4 ⁇ 4 subblocks.
  • the motion vector v 0 of the upper left corner control point of the current block is derived based on the motion vector of the adjacent block
  • the motion vector v 1 of the upper right corner control point of the current block is derived based on the motion vector of the adjacent subblock Be done.
  • the motion vector (v x , v y ) of each sub block in the current block is derived according to the following equation (2).
  • the derivation method of the motion vector of the upper left and upper right control points may include several different modes.
  • Information indicating such an affine motion compensation prediction mode (for example, called an affine flag) is signaled at the CU level. Note that the signaling of the information indicating this affine motion compensation prediction mode need not be limited to the CU level, and other levels (eg, sequence level, picture level, slice level, tile level, CTU level or subblock level) ) May be.
  • the prediction control unit 128 selects one of the intra prediction signal and the inter prediction signal, and outputs the selected signal as a prediction signal to the subtraction unit 104 and the addition unit 116.
  • one prediction MV is selected from among the plurality of prediction MVs registered in the prediction MV list, and it is determined as the MV of the current block.
  • merge_idx which is a signal indicating which prediction MV has been selected, is described in the stream and encoded.
  • FIG. 9D is a diagram for describing an outline of a predicted image generation method using luminance correction processing by LIC processing.
  • an MV for obtaining a reference image corresponding to a current block to be coded is derived from a reference picture which is a coded picture.
  • a predicted image for a block to be encoded is generated.
  • a predicted image is generated from a plurality of reference pictures, and is similar to the reference image acquired from each reference picture. After performing luminance correction processing by a method, a predicted image is generated.
  • FIG. 10 is a block diagram showing a functional configuration of decoding apparatus 200 according to Embodiment 1.
  • the decoding device 200 is a moving image / image decoding device that decodes a moving image / image in units of blocks.
  • the decoding apparatus 200 includes an entropy decoding unit 202, an inverse quantization unit 204, an inverse conversion unit 206, an addition unit 208, a block memory 210, a loop filter unit 212, and a frame memory 214. , An intra prediction unit 216, an inter prediction unit 218, and a prediction control unit 220.
  • the inverse quantization unit 204 inversely quantizes the quantization coefficient of the block to be decoded (hereinafter referred to as a current block), which is an input from the entropy decoding unit 202. Specifically, the dequantization part 204 dequantizes the said quantization coefficient about each of the quantization coefficient of a current block based on the quantization parameter corresponding to the said quantization coefficient. Then, the dequantization unit 204 outputs the dequantized quantization coefficient (that is, transform coefficient) of the current block to the inverse transformation unit 206.
  • a current block which is an input from the entropy decoding unit 202.
  • the dequantization part 204 dequantizes the said quantization coefficient about each of the quantization coefficient of a current block based on the quantization parameter corresponding to the said quantization coefficient. Then, the dequantization unit 204 outputs the dequantized quantization coefficient (that is, transform coefficient) of the current block to the inverse transformation unit 206.
  • the inverse transform unit 206 restores the prediction error by inversely transforming the transform coefficient that is the input from the inverse quantization unit 204.
  • the inverse transform unit 206 determines the current block based on the deciphered transformation type information. Inverse transform coefficients of
  • the inverse transform unit 206 applies inverse retransformation to the transform coefficients.
  • the addition unit 208 adds the prediction error, which is the input from the inverse conversion unit 206, and the prediction sample, which is the input from the prediction control unit 220, to reconstruct the current block. Then, the adding unit 208 outputs the reconstructed block to the block memory 210 and the loop filter unit 212.
  • the block memory 210 is a storage unit for storing a block within a picture to be decoded (hereinafter referred to as a current picture) which is a block referred to in intra prediction. Specifically, the block memory 210 stores the reconstructed block output from the adding unit 208.
  • one filter is selected from the plurality of filters based on the local gradient direction and activity, The selected filter is applied to the reconstruction block.
  • the frame memory 214 is a storage unit for storing a reference picture used for inter prediction, and may be referred to as a frame buffer. Specifically, the frame memory 214 stores the reconstructed block filtered by the loop filter unit 212.
  • the intra prediction unit 216 refers to a block in the current picture stored in the block memory 210 to perform intra prediction based on the intra prediction mode read from the coded bit stream, thereby generating a prediction signal (intra prediction Signal). Specifically, the intra prediction unit 216 generates an intra prediction signal by performing intra prediction with reference to samples (for example, luminance value, color difference value) of a block adjacent to the current block, and performs prediction control on the intra prediction signal. Output to unit 220.
  • the intra prediction unit 216 may predict the chrominance component of the current block based on the luminance component of the current block. .
  • the inter prediction unit 218 predicts the current block with reference to the reference picture stored in the frame memory 214.
  • the prediction is performed in units of the current block or subblocks (for example, 4 ⁇ 4 blocks) in the current block.
  • the inter prediction unit 218 generates an inter prediction signal of the current block or sub block by performing motion compensation using motion information (for example, a motion vector) read from the coded bit stream, and generates an inter prediction signal. It is output to the prediction control unit 220.
  • the inter prediction unit 218 determines not only the motion information of the current block obtained by the motion search but also the motion information of the adjacent block. Use to generate an inter prediction signal.
  • the inter prediction unit 218 derives a motion vector based on a model assuming uniform linear motion. Also, in the case where the information deciphered from the coded bit stream indicates that the affine motion compensation prediction mode is applied, the inter prediction unit 218 performs motion vectors in units of sub blocks based on motion vectors of a plurality of adjacent blocks. Derive
  • the prediction control unit 220 selects one of the intra prediction signal and the inter prediction signal, and outputs the selected signal to the addition unit 208 as a prediction signal.
  • the inter prediction unit 126 can select a prediction mode from a plurality of modes (normal mode, merge mode, affine mode, etc.), and derives a motion vector (MV) using the selected prediction mode.
  • the plurality of prediction modes are a first type of prediction mode (normal mode, block unit merge mode, etc.) for deriving a motion vector (MV) in prediction block units, and MV in sub block units obtained by further dividing the prediction block.
  • MV Motion vector
  • sub-block unit merge mode there are an ATMVP (Advanced Temporal Motion Vector Prediction) mode, an STMVP (Spatial-Temporal Motion Vector Prediction) mode, and the like.
  • the inter prediction unit 126 determines the size of the prediction block to be processed. Then, the sub-block size is determined according to (S106), and the prediction block is divided into sub-blocks at that size.
  • the block size is defined by the horizontal and vertical lengths of the block. Details of the determination of the sub block size in step S106 will be described later.
  • the inter prediction unit 126 derives an MV for merge mode in a sub-block unit or an MV for affine mode in a sub-block unit (S107, S108). Then, in loop processing in units of subblocks (S109 to S111), the inter prediction unit 126 performs MC processing in units of subblocks using the MV derived in step S107 or step S108 (S110). Generate a predicted image of
  • the inter prediction unit 126 determines, for example, the horizontal and vertical lengths of the sub block based on the following formula.
  • Sub block size (horizontal) MAX (M1, predicted block size (horizontal) / 2 N1 )
  • Sub block size (vertical) MAX (M2, predicted block size (vertical) / 2 N2 )
  • the inter prediction unit 126 first calculates the quotient of the first division that divides the predicted block size in the horizontal direction by the first power operation (2 N1 ).
  • the base of the first power operation is 2, and the exponent of the first power operation is N1.
  • the inter prediction unit 126 sets the calculated quotient of the first division as the sub-block size in the horizontal direction.
  • the inter prediction unit 126 sets M1 as the horizontal sub-block size.
  • the inter prediction unit 126 calculates the quotient of the second division that divides the predicted block size in the vertical direction by the second power operation (2 N2 ).
  • the base of the second power operation is 2, and the exponent of the second power operation is N2.
  • the inter prediction unit 126 sets the calculated quotient of the second division as the vertical sub-block size.
  • the inter prediction unit 126 sets M2 as the vertical sub-block size.
  • N1 and N2 may be the same value or different values.
  • M1 and M2 may have the same value or different values.
  • M1 and M2 may be determined based on the minimum block size (for example, 4 ⁇ 4 pixels) that can perform motion compensation processing.
  • M1 and M2 may be determined as the horizontal size (for example, 4 pixels) and the vertical size (for example, 4 pixels) of the minimum block size allowed for motion compensation processing, respectively.
  • the sub-block size determination (S106) is performed for both sub-block unit merge mode and affine mode, but the sub-block size determination may be performed only in one of the modes. Good. For example, determination of subblock size is performed in merge mode in units of subblocks, and in affine mode, a prediction block may be always divided into subblocks of fixed size without determination of subblock size, or vice versa It may be Also, the determination of the sub-block size may be similarly performed for the merge mode of the sub-block unit and the second type prediction mode for deriving the MV in sub-block units other than the affine mode.
  • process flow of FIG. 11 is an example, and part of the processes described may be removed or processes not described may be added.
  • FIG. 12 is a diagram for describing a specific example of the size of the sub block obtained by dividing the processing target block.
  • the prediction block size is 32 ⁇ 32 pixels
  • the prediction block is divided into 4 ⁇ 4 pixel sub blocks.
  • the prediction block and the sub-block both have a square of 1: 1 in width: length.
  • the prediction block size is 64 ⁇ 32 pixels
  • the prediction block is divided into subblocks of 8 ⁇ 4 pixels.
  • the prediction block and the sub-block both have a rectangle of 2: 1 horizontal: vertical.
  • the prediction block size is 16 ⁇ 32 pixels
  • the prediction block is divided into subblocks of 4 ⁇ 4 pixels.
  • the prediction block is a rectangle of 1: 2 in horizontal: vertical, but the sub-block is a square of 1: 1 in horizontal: vertical.
  • the prediction block size is 32 ⁇ 32 pixels
  • the prediction block is divided into sub blocks of 8 ⁇ 8 pixels.
  • the prediction block and the sub-block both have a square of 1: 1 in width: length.
  • the prediction block size is 64 ⁇ 32 pixels
  • the prediction block is divided into subblocks of 16 ⁇ 8 pixels.
  • the prediction block and the sub-block both have a rectangle of 2: 1 horizontal: vertical.
  • the prediction block size is 16 ⁇ 32 pixels
  • the prediction block is divided into 4 ⁇ 8 pixel sub blocks.
  • both of the prediction block and the sub-block become a rectangle having a width: height of 1: 2.
  • N1 and N2 increase, the sub-block size is reduced. Also, if the values of M1 and M2 decrease, smaller subblock sizes are allowed. That is, increasing N1 and N2 and / or decreasing M1 and M2 divides the prediction block into smaller sub-blocks. On the other hand, if N1 and N2 are decreased and / or M1 and M2 are increased, the prediction block is divided into larger sub-blocks.
  • the present invention is not limited to this example.
  • the vertical and horizontal lengths of the sub blocks may be determined based on only one of the two.
  • the vertical and horizontal lengths of the sub-blocks may be determined based on the ratio of vertical and horizontal lengths rather than the absolute values of vertical and horizontal lengths of the prediction block to be processed.
  • the vertical and horizontal lengths of the sub-blocks may be determined such that the ratio of vertical and horizontal lengths of the prediction block and the vertical and horizontal lengths of the sub-blocks are the same.
  • the entropy coding unit 110 of the coding apparatus 100 of the information related to M1, M2, N1 and N2 in FIGS. 11 and 12 is a sequence header area, a picture header area, a slice header area or auxiliary information of a coded bit stream. It may be encoded in the area.
  • the entropy decoding unit 202 of the decoding device 200 may decode the information on M1, M2, N1, and N2 from those regions.
  • the coding apparatus 100 switches the values of M1, M2, N1, and N2 according to the processing capacity of the coding apparatus 100 and the processing capacity of the decoding apparatus 200 to which the coding bit stream is transmitted from the coding apparatus 100.
  • the encoding apparatus 100 may determine the values of M1, M2, N1, and N2 based on the processing capabilities of the encoding apparatus 100 and the decoding apparatus 200. For example, if the processing capability is high, the encoding apparatus 100 sets the values of M1, M2, N1, and N2 so that they can be divided into small subblocks because many processes can be performed in a fixed processing time. May be On the other hand, if the processing capability is low, encoding apparatus 100 can set many values of M1, M2, N1, and N2 so that they can be divided into large subblocks because they can not perform many processings in a fixed processing time. Good.
  • the coding apparatus 100 may switch the values of M1, M2, N1, and N2 according to the profile or level information assigned to the coded bit stream. That is, the encoding apparatus 100 may determine the values of M1, M2, N1 and N2 based on the profile or level information. For example, in a profile and level assuming that decoding apparatus 200 has sufficient processing capability, encoding apparatus 100 sets the values of M1, M2, N1, and N2 so as to be divided into small subblocks. It is also good. On the other hand, in the profile and level assuming that the decoding apparatus 200 has insufficient processing power, the encoding apparatus 100 sets the values of M1, M2, N1 and N2 so as to be divided into large subblocks. May be
  • the coding apparatus 100 may switch the values of M1, M2, N1, and N2 according to the size of the current picture to be coded. That is, the encoding apparatus 100 may determine the values of M1, M2, N1, and N2 based on the size of the current picture to be encoded. For example, when the picture size is small, the encoding apparatus 100 may set the values of M1, M2, N1, and N2 so as to be divided into small subblocks so as to be able to follow finer motion. On the other hand, when the picture size is large, the encoding apparatus 100 divides the values of M1, M2, N1, and N2 so as to be divided into large subblocks so as not to be unnecessarily dragged by local fine motions. It may be set.
  • the information on M1, M2, N1 and N2 to be encoded into the bit stream may not be a signal of information strictly indicating them, and may be encoded as a signal having another meaning. For example, by directly associating information on M1, M2, N1, and N2 with profiles and levels in advance and defining them in advance, M1, M2, N1, and N2 can be identified with only signals indicating the profiles and levels.
  • information indicating the values of M1, M2, N1 and N2 themselves may be encoded, or information indicating a difference value from the values of M1, M2, N1 and N2 defined in advance in a standard etc. may be encoded. It is also good.
  • encoding apparatus 100 and decoding apparatus 200 divide the processing target block (prediction block) into sub blocks and perform motion compensation in units of sub blocks.
  • encoding / decoding processing can be performed by switching the horizontal and vertical lengths of the sub block based on the size of the processing target block. That is, the encoding apparatus 100 and the decoding apparatus 200 can perform motion compensation on subblocks of variable sizes depending on the size of the processing target block.
  • each embodiment may be realized by centralized processing using a single device (system), or may be realized by distributed processing using a plurality of devices. Good.
  • the processor that executes the program may be singular or plural. That is, centralized processing may be performed, or distributed processing may be performed.
  • the server not only encodes a two-dimensional moving image, but also automatically encodes a still image based on scene analysis of the moving image or at a time designated by the user and transmits it to the receiving terminal. It is also good. Furthermore, if the server can acquire relative positional relationship between the imaging terminals, the three-dimensional shape of the scene is not only determined based on the two-dimensional moving image but also the video of the same scene captured from different angles. Can be generated. Note that the server may separately encode three-dimensional data generated by a point cloud or the like, or an image to be transmitted to the receiving terminal based on a result of recognizing or tracking a person or an object using the three-dimensional data. Alternatively, it may be generated by selecting or reconfiguring from videos taken by a plurality of terminals.
  • the user can enjoy the scene by arbitrarily selecting each video corresponding to each photographing terminal, or from the three-dimensional data reconstructed using a plurality of images or videos, the video of the arbitrary viewpoint You can also enjoy the extracted content.
  • the sound may be picked up from a plurality of different angles as well as the video, and the server may multiplex the sound from a specific angle or space with the video and transmit it according to the video.
  • the server superimposes virtual object information in the virtual space on camera information in the real space based on the three-dimensional position or the movement of the user's viewpoint.
  • the decoding apparatus may acquire or hold virtual object information and three-dimensional data, generate a two-dimensional image according to the movement of the user's viewpoint, and create superimposed data by smoothly connecting.
  • the decoding device transmits the motion of the user's viewpoint to the server in addition to the request for virtual object information, and the server creates superimposed data in accordance with the motion of the viewpoint received from the three-dimensional data held in the server.
  • the superimposed data may be encoded and distributed to the decoding device.
  • the picture may be divided into tiles or the like according to the meaning of an object or the like in the image, and the decoding side may be configured to decode only a part of the area by selecting the tile to be decoded.
  • the decoding side can position the desired object based on the meta information. And determine the tile that contains the object. For example, as shown in FIG. 17, meta-information is stored using a data storage structure different from pixel data, such as an SEI message in HEVC. This meta information indicates, for example, the position, size, or color of the main object.
  • the receiving terminal when transmitting or receiving still image or video data such as two-dimensional or three-dimensional map information for automatic traveling or driving assistance of a car, the receiving terminal is added as image information belonging to one or more layers as meta information Information on weather or construction may also be received, and these may be correlated and decoded.
  • the meta information may belong to the layer or may be simply multiplexed with the image data.
  • the receiving terminal since a car including a receiving terminal, a drone or an airplane moves, the receiving terminal transmits the position information of the receiving terminal at the time of reception request to seamlessly receive and decode while switching the base stations ex106 to ex110. Can be realized.
  • the receiving terminal can dynamically switch how much meta information is received or how much map information is updated according to the user's selection, the user's situation or the state of the communication band. become.
  • the client can receive, decode, and reproduce the encoded information transmitted by the user in real time.
  • the server may change and encode the face of a person at the periphery of the screen, or the inside of a house, etc. into an image out of focus.
  • the server recognizes whether or not the face of a person different from the person registered in advance appears in the image to be encoded, and if so, performs processing such as mosaicing the face portion. May be Alternatively, the user designates a person or background area desired to process an image from the viewpoint of copyright etc.
  • preprocessing or post-processing of encoding replaces the designated area with another video or blurs the focus. It is also possible to perform such processing. If it is a person, it is possible to replace the image of the face part while tracking the person in the moving image.
  • the present invention is not limited to the content supply system ex100 via the Internet ex101, but is also applicable to at least a moving picture coding apparatus (image coding apparatus) or a moving picture decoding apparatus (image decoding apparatus) of the above embodiments Can be incorporated. There is a difference in that it is multicast-oriented with respect to the configuration in which the content supply system ex100 can be easily unicasted, since multiplexed data in which video and sound are multiplexed is transmitted on broadcast radio waves using satellites etc. Similar applications are possible for the encoding process and the decoding process.
  • the smartphone ex115 performs processing such as call and data communication based on control of the main control unit ex460 having a CPU, a ROM, a RAM, and the like.
  • the audio signal collected by the audio input unit ex456 is converted to a digital audio signal by the audio signal processing unit ex454, spread spectrum processing is performed by the modulation / demodulation unit ex452, and digital analog conversion is performed by the transmission / reception unit ex451.
  • transmission is performed via the antenna ex450.
  • the received data is amplified and subjected to frequency conversion processing and analog-to-digital conversion processing, subjected to spectrum despreading processing by modulation / demodulation unit ex452, and converted to an analog sound signal by sound signal processing unit ex454.
  • Output from In the data communication mode text, still images, or video data are sent to the main control unit ex460 via the operation input control unit ex462 by the operation of the operation unit ex466 or the like of the main unit, and transmission and reception processing is similarly performed.
  • the video signal processing unit ex 455 executes the video signal stored in the memory unit ex 467 or the video signal input from the camera unit ex 465 as described above.
  • the multiplexing / demultiplexing unit ex453 multiplexes in order to decode multiplexed data received via the antenna ex450.
  • the multiplexed data is divided into a bit stream of video data and a bit stream of audio data, and the encoded video data is supplied to the video signal processing unit ex455 via the synchronization bus ex470, and The converted audio data is supplied to the audio signal processing unit ex 454.
  • the smartphone ex115 is described as an example, in addition to a transmission / reception terminal having both an encoder and a decoder as a terminal, it is referred to as a transmitting terminal having only the encoder and a receiving terminal having only the decoder.
  • a transmitting terminal having only the encoder
  • a receiving terminal having only the decoder.
  • multiplexed data in which audio data is multiplexed with video data is received or transmitted, but in multiplexed data, character data related to video other than audio data is also described. It may be multiplexed, or video data itself may be received or transmitted, not multiplexed data.
  • the present disclosure is applicable to, for example, television receivers, digital video recorders, car navigation systems, mobile phones, digital cameras, digital video cameras, and the like.

Landscapes

  • Engineering & Computer Science (AREA)
  • Multimedia (AREA)
  • Signal Processing (AREA)
  • Compression Or Coding Systems Of Tv Signals (AREA)

Abstract

L'invention concerne un dispositif de codage (100) pourvu d'un circuit et d'une mémoire. Dans un mode de prédiction inter dans lequel un bloc à traiter est divisé en sous-blocs et une compensation de mouvement (MC) est effectuée sous-bloc par sous-bloc, le circuit utilise la mémoire pour commuter les longueurs horizontale et verticale des sous-blocs sur la base de la taille du bloc à traiter, et pour effectuer un traitement de codage.
PCT/JP2019/002341 2018-01-29 2019-01-24 Dispositif et procédé de codage, et dispositif et procédé de décodage WO2019146718A1 (fr)

Applications Claiming Priority (2)

Application Number Priority Date Filing Date Title
US201862623133P 2018-01-29 2018-01-29
US62/623,133 2018-01-29

Publications (1)

Publication Number Publication Date
WO2019146718A1 true WO2019146718A1 (fr) 2019-08-01

Family

ID=67395480

Family Applications (1)

Application Number Title Priority Date Filing Date
PCT/JP2019/002341 WO2019146718A1 (fr) 2018-01-29 2019-01-24 Dispositif et procédé de codage, et dispositif et procédé de décodage

Country Status (1)

Country Link
WO (1) WO2019146718A1 (fr)

Citations (3)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
WO2015048459A1 (fr) * 2013-09-26 2015-04-02 Qualcomm Incorporated Prédiction temporelle de vecteur de mouvement basée sur une unité de sous-prédiction (pu) en hevc et conception de sous-pu en 3d hevc
WO2016160605A1 (fr) * 2015-03-27 2016-10-06 Qualcomm Incorporated Dérivation d'informations de mouvement pour sous-blocs en codage vidéo
WO2016201094A1 (fr) * 2015-06-11 2016-12-15 Qualcomm Incorporated Prédiction de vecteur de mouvement par sous-unité de prédiction au moyen d'informations de mouvement spatiales et/ou temporelles

Patent Citations (3)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
WO2015048459A1 (fr) * 2013-09-26 2015-04-02 Qualcomm Incorporated Prédiction temporelle de vecteur de mouvement basée sur une unité de sous-prédiction (pu) en hevc et conception de sous-pu en 3d hevc
WO2016160605A1 (fr) * 2015-03-27 2016-10-06 Qualcomm Incorporated Dérivation d'informations de mouvement pour sous-blocs en codage vidéo
WO2016201094A1 (fr) * 2015-06-11 2016-12-15 Qualcomm Incorporated Prédiction de vecteur de mouvement par sous-unité de prédiction au moyen d'informations de mouvement spatiales et/ou temporelles

Non-Patent Citations (2)

* Cited by examiner, † Cited by third party
Title
JIANLE CHEN ET AL.: "Algorithm Description of Joint Exploration Test Model 1", JOINT VIDEO EXPLORATION TEAM (JVET) OF ITU-T SG 16 WP 3, February 2016 (2016-02-01), Geneva, CH, pages i - ii, 1-25, XP030150000 *
WEI-JUNG CHIEN ET AL.: "Extension of Advanced Temporal Motion Vector Predictor (ATMVP),", ITU - TELECOMMUNICATIONS STANDARDIZATION SECTOR STUDY GROUP 16 QUESTION 6 VIDEO CODING EXPERTS GROUP (VCEG, June 2015 (2015-06-01), Warsaw, Poland, pages 1 - 4, XP030003891 *

Similar Documents

Publication Publication Date Title
JP6946419B2 (ja) 復号装置、復号方法及びプログラム
JP2020198649A (ja) 符号化装置、復号装置、符号化方法及び復号方法
JP7339890B2 (ja) 符号化装置及び復号装置
WO2019022099A1 (fr) Dispositif de codage, dispositif de décodage, procédé de codage et procédé de décodage
JP7087030B2 (ja) 符号化装置、復号装置、符号化方法及び復号方法
JP2022066195A (ja) 符号化装置及び符号化方法
JP2022008413A (ja) 復号装置、符号化装置及び記録媒体
JP7026747B2 (ja) 復号装置及び復号方法
WO2019138998A1 (fr) Dispositif de codage, dispositif de décodage, procédé de codage et procédé de décodage
JP7017580B2 (ja) 復号装置、画像復号装置及び復号方法
WO2019013235A1 (fr) Dispositif de codage, procédé de codage, dispositif de décodage et procédé de décodage
WO2019013236A1 (fr) Dispositif de codage, procédé de codage, dispositif de décodage et procédé de décodage
JP2022093625A (ja) 符号化装置、復号装置、符号化方法、及び復号方法
JP2021101576A (ja) 復号装置及び記憶媒体
JP2023126387A (ja) 符号化装置及び符号化方法
WO2019069782A1 (fr) Dispositif de codage, dispositif de décodage, procédé de codage et procédé de décodage
JP7371055B2 (ja) 符号化装置、及び復号装置
JP2021180494A (ja) 復号装置、復号方法、及び、非一時的記憶媒体
WO2019009314A1 (fr) Dispositif de codage, dispositif de décodage, procédé de codage, et procédé de décodage
WO2019069902A1 (fr) Dispositif de codage, dispositif de décodage, procédé de codage et procédé de décodage
WO2019065329A1 (fr) Dispositif et procédé de codage, et dispositif et procédé de décodage
WO2019059107A1 (fr) Dispositif d'encodage, dispositif de décodage, procédé d'encodage et procédé de décodage
WO2019021803A1 (fr) Dispositif de codage, dispositif de décodage, procédé de codage, et procédé de décodage
WO2019031369A1 (fr) Dispositif de codage, dispositif de décodage, procédé de codage et procédé de décodage
WO2019244809A1 (fr) Dispositif de codage, dispositif de décodage, procédé de codage, et procédé de décodage

Legal Events

Date Code Title Description
121 Ep: the epo has been informed by wipo that ep was designated in this application

Ref document number: 19743486

Country of ref document: EP

Kind code of ref document: A1

NENP Non-entry into the national phase

Ref country code: DE

122 Ep: pct application non-entry in european phase

Ref document number: 19743486

Country of ref document: EP

Kind code of ref document: A1