WO2011013253A1 - Dispositif produisant un signal de prévision à l’aide d’une prédiction de compensation de mouvement de transformation géométrique, dispositif de codage d’image variant dans le temps et dispositif de décodage d’image variant dans le temps - Google Patents

Dispositif produisant un signal de prévision à l’aide d’une prédiction de compensation de mouvement de transformation géométrique, dispositif de codage d’image variant dans le temps et dispositif de décodage d’image variant dans le temps Download PDF

Info

Publication number
WO2011013253A1
WO2011013253A1 PCT/JP2009/063692 JP2009063692W WO2011013253A1 WO 2011013253 A1 WO2011013253 A1 WO 2011013253A1 JP 2009063692 W JP2009063692 W JP 2009063692W WO 2011013253 A1 WO2011013253 A1 WO 2011013253A1
Authority
WO
WIPO (PCT)
Prior art keywords
geometric transformation
prediction
transformation parameter
unit
block
Prior art date
Application number
PCT/JP2009/063692
Other languages
English (en)
Japanese (ja)
Inventor
昭行 谷沢
太一郎 塩寺
中條 健
Original Assignee
株式会社 東芝
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Application filed by 株式会社 東芝 filed Critical 株式会社 東芝
Priority to PCT/JP2009/063692 priority Critical patent/WO2011013253A1/fr
Publication of WO2011013253A1 publication Critical patent/WO2011013253A1/fr

Links

Images

Classifications

    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06TIMAGE DATA PROCESSING OR GENERATION, IN GENERAL
    • G06T3/00Geometric image transformations in the plane of the image
    • G06T3/02Affine transformations
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06TIMAGE DATA PROCESSING OR GENERATION, IN GENERAL
    • G06T7/00Image analysis
    • G06T7/20Analysis of motion
    • G06T7/223Analysis of motion using block-matching
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04NPICTORIAL COMMUNICATION, e.g. TELEVISION
    • H04N19/00Methods or arrangements for coding, decoding, compressing or decompressing digital video signals
    • H04N19/10Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using adaptive coding
    • H04N19/102Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using adaptive coding characterised by the element, parameter or selection affected or controlled by the adaptive coding
    • H04N19/103Selection of coding mode or of prediction mode
    • H04N19/105Selection of the reference unit for prediction within a chosen coding or prediction mode, e.g. adaptive choice of position and number of pixels used for prediction
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04NPICTORIAL COMMUNICATION, e.g. TELEVISION
    • H04N19/00Methods or arrangements for coding, decoding, compressing or decompressing digital video signals
    • H04N19/10Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using adaptive coding
    • H04N19/169Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using adaptive coding characterised by the coding unit, i.e. the structural portion or semantic portion of the video signal being the object or the subject of the adaptive coding
    • H04N19/17Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using adaptive coding characterised by the coding unit, i.e. the structural portion or semantic portion of the video signal being the object or the subject of the adaptive coding the unit being an image region, e.g. an object
    • H04N19/174Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using adaptive coding characterised by the coding unit, i.e. the structural portion or semantic portion of the video signal being the object or the subject of the adaptive coding the unit being an image region, e.g. an object the region being a slice, e.g. a line of blocks or a group of blocks
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04NPICTORIAL COMMUNICATION, e.g. TELEVISION
    • H04N19/00Methods or arrangements for coding, decoding, compressing or decompressing digital video signals
    • H04N19/10Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using adaptive coding
    • H04N19/169Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using adaptive coding characterised by the coding unit, i.e. the structural portion or semantic portion of the video signal being the object or the subject of the adaptive coding
    • H04N19/17Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using adaptive coding characterised by the coding unit, i.e. the structural portion or semantic portion of the video signal being the object or the subject of the adaptive coding the unit being an image region, e.g. an object
    • H04N19/176Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using adaptive coding characterised by the coding unit, i.e. the structural portion or semantic portion of the video signal being the object or the subject of the adaptive coding the unit being an image region, e.g. an object the region being a block, e.g. a macroblock
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04NPICTORIAL COMMUNICATION, e.g. TELEVISION
    • H04N19/00Methods or arrangements for coding, decoding, compressing or decompressing digital video signals
    • H04N19/10Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using adaptive coding
    • H04N19/189Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using adaptive coding characterised by the adaptation method, adaptation tool or adaptation type used for the adaptive coding
    • H04N19/196Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using adaptive coding characterised by the adaptation method, adaptation tool or adaptation type used for the adaptive coding being specially adapted for the computation of encoding parameters, e.g. by averaging previously computed encoding parameters
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04NPICTORIAL COMMUNICATION, e.g. TELEVISION
    • H04N19/00Methods or arrangements for coding, decoding, compressing or decompressing digital video signals
    • H04N19/46Embedding additional information in the video signal during the compression process
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04NPICTORIAL COMMUNICATION, e.g. TELEVISION
    • H04N19/00Methods or arrangements for coding, decoding, compressing or decompressing digital video signals
    • H04N19/50Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using predictive coding
    • H04N19/503Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using predictive coding involving temporal prediction
    • H04N19/51Motion estimation or motion compensation
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04NPICTORIAL COMMUNICATION, e.g. TELEVISION
    • H04N19/00Methods or arrangements for coding, decoding, compressing or decompressing digital video signals
    • H04N19/50Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using predictive coding
    • H04N19/503Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using predictive coding involving temporal prediction
    • H04N19/51Motion estimation or motion compensation
    • H04N19/513Processing of motion vectors
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04NPICTORIAL COMMUNICATION, e.g. TELEVISION
    • H04N19/00Methods or arrangements for coding, decoding, compressing or decompressing digital video signals
    • H04N19/60Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using transform coding
    • H04N19/61Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using transform coding in combination with predictive coding
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04NPICTORIAL COMMUNICATION, e.g. TELEVISION
    • H04N19/00Methods or arrangements for coding, decoding, compressing or decompressing digital video signals
    • H04N19/70Methods or arrangements for coding, decoding, compressing or decompressing digital video signals characterised by syntax aspects related to video coding, e.g. related to compression standards

Definitions

  • the present invention relates to a prediction signal generation apparatus and a moving picture coding method for deriving a geometric transformation parameter using motion information of neighboring blocks and a prediction target block, and performing a geometric transformation prediction process for the prediction target block based on the derived geometric transformation parameter.
  • the present invention relates to a device and a video decoding device.
  • H.264 a moving picture coding method that has greatly improved coding efficiency has been jointly developed by ITU-T and ISO / IEC. H. H.264 and ISO / IEC 14496-10 (hereinafter referred to as “H.264”).
  • H. In H.264 prediction processing, conversion processing, and entropy coding processing are performed in units of rectangular blocks (for example, 16 ⁇ 16 pixels, 8 ⁇ 8 pixels). For this reason, H.C. In H.264, when an object that cannot be expressed by a rectangular block is predicted, the prediction efficiency is increased by selecting a smaller prediction block (for example, 4 ⁇ 4 pixels).
  • Methods for effectively predicting such objects include a method of preparing a plurality of prediction patterns in a rectangular block, and a method of applying motion compensation using affine transformation to a deformed object.
  • Japanese Patent Application Laid-Open No. 2007-312397 uses an object motion model as an affine transformation model, and calculates the optimum affine transformation parameters for each block to be predicted, thereby taking into consideration the enlargement / reduction / rotation of the object.
  • a video frame transfer method using prediction is disclosed.
  • the method described in Kordasiewicz et al. Uses affine transformation using motion vectors of eight types of adjacent blocks adjacent to a pixel block to be predicted, such as up, down, left, and right, and a motion vector calculated from the pixel block to be predicted. In order to estimate the parameters, it is necessary to re-encode the frame a plurality of times in order to obtain an optimal motion vector. On the other hand, when only the motion vector is calculated in units of frames from the original image, this conventional method is not optimal from the viewpoint of code amount and encoding distortion, and there is a problem that encoding efficiency decreases.
  • the present invention reduces a motion detection process necessary for estimating a geometric transformation parameter used for geometric transformation motion compensated prediction and improves prediction efficiency without increasing a code amount, a moving picture coding apparatus, a moving picture, and the like. It is an object of the present invention to provide a decoding device and a moving image decoding device.
  • prediction selection information indicating whether to use a first geometric transformation parameter or a second geometric transformation parameter indicating information related to an image shape by geometric transformation of a pixel block is set.
  • One setting unit and one or more second blocks of which the prediction signal generation process has already been completed among the plurality of first adjacent blocks adjacent to one pixel block of the plurality of pixel blocks into which the image signal is divided An acquisition unit that acquires motion information of neighboring blocks or the geometric transformation parameters; a derivation unit that derives predicted geometric transformation parameters of the one pixel block from the geometric transformation parameters of the one or more second neighboring blocks; Deriving the first geometric transformation parameter by a predetermined method from the derived geometric transformation parameter derivation value and the predicted geometric transformation parameter, A second setting unit for setting, a third setting unit for setting the second geometric transformation parameter based on motion information of the one pixel block and the one or more second adjacent blocks, A prediction signal is generated by performing a geometric transformation process using the first geometric transformation parameter or the second geometric transformation parameter indicated by the selection information on
  • the prediction selection information indicating whether to use the first geometric transformation parameter or the second geometric transformation parameter is encoded using the prediction signal generation device described above.
  • a derived value of the geometric transformation parameter is derived from the first geometric transformation parameter and the predicted geometric transformation parameter by a predetermined method.
  • a moving image decoding apparatus that decodes moving image encoded data obtained by encoding an input image signal in units of a plurality of pixel blocks, and performs decoding processing by a prescribed method.
  • Motion information or pixel block of one or more second adjacent blocks that have already been decoded among a plurality of first adjacent blocks adjacent to one pixel block of the plurality of pixel blocks into which the image signal is divided
  • a motion information acquisition unit for acquiring a geometric conversion parameter indicating information related to the shape of an image by geometric conversion of the image, and decoding selection information indicating whether to use the first geometric conversion parameter or the second geometric conversion parameter
  • a derivation unit for deriving a predicted geometric transformation parameter of the one pixel block from the geometric transformation parameters of two adjacent blocks, a derivation value of the decoded geometric transformation parameter, and the predicted geometric transformation
  • the first and second embodiments relate to a video encoding device
  • the third and fourth embodiments relate to a video decoding device.
  • the video encoding apparatus described in the following embodiment divides each frame constituting an input image signal into a plurality of pixel blocks, performs encoding processing on the divided pixel blocks, and performs compression encoding. It is a device that outputs a code string.
  • the moving picture coding apparatus 100 is connected to the coding control unit 114.
  • the subtractor 101 calculates a difference between the input image signal 115 and the prediction signal 206, and outputs a prediction error signal 116.
  • the output terminal of the subtractor 101 is connected to the transform / quantization unit 102.
  • the transform / quantization unit 102 includes, for example, an orthogonal transformer (discrete cosine transformer) and a quantizer, performs orthogonal transform (discrete cosine transform) on the prediction error signal 116, quantizes it, and transforms it into a transform coefficient 117.
  • the output terminal of the transform / quantization unit 102 is connected to the inverse quantization / inverse transform unit 103 and the entropy coding unit 112.
  • the inverse quantization / inverse transform unit 103 includes an inverse quantizer and an inverse orthogonal transformer (inverse discrete cosine transformer), inversely quantizes the transform coefficient 117, and performs inverse orthogonal transform to restore the decoded prediction error signal 118. .
  • the output terminal of the inverse quantization / inverse transform unit 103 is connected to the adder 104.
  • the adder 104 adds the decoded prediction error signal 118 and the prediction signal 206 to generate a decoded image signal 119.
  • the output terminal of the adder 104 is connected to the reference image memory 105.
  • the reference image memory 105 stores the decoded image signal 119 as a reference image signal.
  • the output terminal of the reference image memory 105 is connected to the motion information search unit 106, the intra prediction signal generation device 107, and the inter prediction signal generation device 109.
  • the motion information search unit 106 uses the input image signal 115 and the reference image signal 207 to calculate motion information (motion vector) 210 suitable for the prediction target block.
  • the output terminal of the motion information search unit 106 is connected to the derived parameter derivation unit 108 and the inter prediction signal generation device 109.
  • the derivation parameter derivation unit 108 includes a predictive geometric transformation parameter derivation unit 601 and a converter 602.
  • the derivation parameter derivation unit 108 has a function of deriving a predictive geometric transformation parameter to be described later, It has a function of calculating derived parameters using information 210.
  • the output terminal of the derived parameter deriving unit 108 is connected to the inter prediction signal generation device 109 and the entropy coding unit 112.
  • the inter prediction signal generation device 109 has a function of generating a prediction signal 206 using the input motion information 210, derived parameter information 108, reference image signal 207, and prediction selection information 123.
  • the intra prediction signal generation device 107 has a function of performing intra prediction using the input reference image signal 207.
  • Output terminals of the intra prediction signal generation device 107 and the inter prediction signal generation device 109 are connected to terminals of the prediction separation switch 110, respectively.
  • the prediction selection unit 111 sets the prediction selection information 123 according to the prediction mode controlled by the encoding control unit 114.
  • the output terminal of the prediction selection unit 111 is connected to the inter prediction signal generation device 109, the prediction separation switch 110, and the entropy encoding unit 112.
  • the prediction separation switch 110 switches between the intra prediction signal generation device 107 and the inter prediction signal generation device 109 according to the prediction selection information 123 of the prediction selection unit 111.
  • the switching terminal of the prediction separation switch 110 is connected to the subtractor 101 and the adder 104, and introduces the prediction signal of the intra prediction signal generation device 107 or the inter prediction signal generation device 109 to the subtractor 101 and the adder 104.
  • the entropy encoding unit 112 includes an encoder and a multiplexer, and entropy-encodes and multiplexes the transform coefficient 117, the derived parameter information 211, and the prediction selection information 123.
  • the output terminal of the entropy encoding unit 112 is connected to the output buffer 113.
  • the output buffer 113 temporarily stores the multiplexed data and outputs it as encoded data 129 according to the output timing managed by the encoding control unit 114.
  • the moving image encoding apparatus 100 having the above configuration performs an intra prediction (intraframe prediction) or inter prediction (interframe prediction) encoding process on the input image signal 115 based on the encoding parameter input from the encoding control unit 114.
  • the prediction signal 206 is generated, and the encoded data 129 is output. That is, an input image signal 115 of a moving image or a still image is input to the moving image encoding apparatus 100 after being divided into pixel blocks, for example, macroblocks.
  • the input image signal is one encoding processing unit including both a frame and a field. In the present embodiment, an example in which a frame is used as one encoding processing unit will be described.
  • the moving picture encoding apparatus 100 performs encoding in a plurality of prediction modes with different block sizes and generation methods of the prediction signal 206.
  • the generation method of the prediction signal 206 is roughly divided into intra prediction (intraframe prediction) in which a prediction signal is generated only within a frame to be encoded, and prediction using a plurality of temporally different reference frames.
  • intra prediction intraframe prediction
  • inter prediction an example in which a prediction signal is generated using inter prediction will be described in detail.
  • the macro block is set to the basic processing block size of the encoding process.
  • the macroblock is typically a 16 ⁇ 16 pixel block shown in FIG. 3B, for example, but may be a 32 ⁇ 32 pixel block unit or an 8 ⁇ 8 pixel block unit.
  • the shape of the macroblock does not necessarily need to be a square lattice.
  • the encoding target block or macroblock of the input image signal 115 is simply referred to as a “prediction target block”.
  • the input image signal 115 is input to the subtractor 101.
  • the subtracter 101 further receives a prediction signal 206 corresponding to each prediction mode output from the prediction separation switch 110.
  • the subtractor 101 calculates a prediction error signal 116 obtained by subtracting the prediction signal 206 from the input image signal 115.
  • the prediction error signal 116 is input to the transform / quantization unit 102.
  • the transform / quantization unit 102 performs orthogonal transform such as discrete cosine transform (DCT) on the prediction error signal 116 to generate transform coefficients.
  • DCT discrete cosine transform
  • the transformation in the transformation / quantization unit 102 is H.264.
  • discrete sine transform, wavelet transform, or component analysis is included.
  • the transform / quantization unit 102 quantizes the transform coefficient in accordance with quantization information represented by a quantization parameter, a quantization matrix, and the like given by the encoding control unit 114.
  • the transform / quantization unit 102 outputs the quantized transform coefficient 117 to the entropy coding unit 112 and also outputs it to the inverse quantization / inverse transform unit 103.
  • the entropy encoding unit 112 performs entropy encoding, for example, Huffman encoding or arithmetic encoding, on the quantized transform coefficient 117.
  • the entropy encoding unit 112 further performs entropy encoding on various encoding parameters used when the encoding target block including the prediction information output from the encoding control unit 114 is encoded. . As a result, encoded data 129 is generated.
  • the encoding parameter is a parameter required for decoding prediction information, information on transform coefficients, information on quantization, and the like.
  • the encoding parameter of the prediction target block is held in an internal memory of the encoding control unit 114, and is used when the prediction target block is used as an adjacent block of another pixel block.
  • the encoded data 129 generated and multiplexed by the entropy encoding unit 112 is output from the moving image encoding apparatus 100, temporarily stored in the output buffer 113, and then managed by the encoding control unit 114.
  • the encoded data 129 is output according to the output timing.
  • the encoded data 129 is sent to, for example, a storage system (storage medium) or a transmission system (communication line) (not shown).
  • the quantization information corresponding to the quantization information used in the transform / quantization unit 102 is loaded from the internal memory of the encoding control unit 114 and subjected to inverse quantization processing.
  • the quantization information is, for example, a parameter represented by a quantization parameter, a quantization matrix, or the like.
  • the inverse quantization / inverse transform unit 103 further reproduces the decoded prediction error signal 118 by performing inverse orthogonal transform such as inverse discrete cosine transform (IDCT) on the transform coefficient after inverse quantization.
  • inverse orthogonal transform such as inverse discrete cosine transform (IDCT)
  • the decoded prediction error signal 118 is input to the adder 104.
  • the adder 104 adds the decoded prediction error signal 118 and the prediction signal 206 output from the prediction separation switch 110 to generate a decoded image signal 119.
  • the decoded image signal 119 is a local decoded image signal.
  • the decoded image signal 119 is stored as the reference image signal 207 in the reference image memory 105.
  • the reference image signal 207 stored in the reference image memory 105 is output to the motion information search unit 106, the intra prediction signal generation device 107, the inter prediction signal generation device 109, etc., and is referred to when performing prediction.
  • the motion information search unit 106 uses the input image signal 115 and the reference image signal 207 to calculate motion information 210 suitable for the prediction target block.
  • the motion information may be represented by an affine transformation parameter, for example.
  • the motion information can be represented by a motion vector.
  • the motion information 210 may be, for example, a predicted value for predicting an affine transformation parameter with another affine transformation parameter or the like, or may be a predicted value for predicting a motion vector with another motion vector or the like.
  • motion information including geometric deformation between images is used as motion information.
  • the motion information search unit 106 searches for motion information 210 (affine transformation parameters and motion vectors) by performing a search such as block matching between the prediction target block of the input image signal 115 and the interpolated image of the reference image signal 207. calculate.
  • a search such as block matching between the prediction target block of the input image signal 115 and the interpolated image of the reference image signal 207.
  • As an evaluation criterion for matching for example, a value obtained by accumulating a difference between the input image signal 115 and the interpolated image after matching for each pixel, a value obtained by adding a difference value between the calculated affine transformation parameter and the center of search, or the like. Use.
  • the motion information 210 may be determined by using a value obtained by converting the difference between the predicted image and the original image, taking into account the size of the motion vector or the affine transformation parameter, Alternatively, the determination may be made by taking into account the code amount of the affine transformation parameter or the like. Moreover, you may utilize costs, such as Formula (1) and Formula (2) mentioned later. Further, the matching may be performed through a search within the matching range based on search range information provided from the outside of the moving image encoding apparatus 100, or may be performed hierarchically for each pixel accuracy.
  • the motion information 210 calculated for a plurality of reference image signals in this way is input to the inter prediction signal generation device 109 and used to generate the prediction signal 206.
  • the plurality of reference image signals are locally decoded images having different display times.
  • the calculated motion information 210 is output to the derived parameter derivation unit 108.
  • the derived parameter derivation unit 108 includes a predicted geometric transformation parameter derivation unit 601 and a converter 602, derives a predicted geometric transformation parameter to be described later, and inputs the predicted geometric transformation parameter and the input motion.
  • the derived parameter is calculated using the information 210.
  • the converter 602 may be, for example, a subtracter.
  • the converter 602 may be an adder, a multiplier, a divider, a converter that performs conversion using a predetermined matrix, and a combination of these.
  • achieves may be sufficient.
  • the converter 602 will be described as a subtractor.
  • the derived parameter information 211 derived by the derived parameter deriving unit 108 is output to the entropy encoding unit 112, and after being subjected to entropy encoding, is multiplexed into encoded data. Furthermore, the motion information 210 obtained by encoding the target pixel block is stored in the internal memory of the encoding control unit 114, and is appropriately loaded from the inter prediction signal generation device 109 and used.
  • the reference image signal 207 stored in the reference image memory 105 is output to the intra prediction signal generation device 107.
  • intra prediction is performed using the input reference image signal 207.
  • a prediction signal is generated by performing pixel interpolation in the prediction direction such as the vertical direction and the horizontal direction using an encoded reference pixel value adjacent to the prediction target block.
  • the interpolated pixel value may be copied in a predetermined prediction direction.
  • the generated prediction signal 206 is output to the prediction separation switch 110.
  • the inter prediction signal generation apparatus 109 generates a prediction signal 206 using the input motion information 210, derived parameter information 108, reference image signal 207, and prediction selection information 123.
  • the generated prediction signal 206 is output to the prediction separation switch 110.
  • the prediction separation switch 110 selects the output terminal of the intra prediction signal generation device 107 and the output terminal of the inter prediction signal generation device 109 according to the prediction selection information 123.
  • the switch is connected to the intra prediction signal generation device 107.
  • the prediction selection information 123 is inter prediction
  • the switch is connected to the inter prediction signal generation device 109.
  • An example of the prediction selection information 123 is shown in FIG.
  • the prediction selection unit 111 sets the prediction selection information 123 according to the prediction mode controlled by the encoding control unit 114.
  • the prediction mode intra prediction or inter prediction can be selected, and a plurality of modes may exist for each.
  • the encoding control unit 114 controls which mode is selected.
  • the prediction signal 206 may be generated for all prediction modes, and one prediction mode may be selected from these, or the prediction mode may be limited according to the characteristics of the input image.
  • the prediction selection information 123 is determined using a cost such as the following equation.
  • the code amount (for example, the code amount of the derivation parameter 211 and the code amount of the prediction block size) required for the prediction information when the prediction mode is selected is OH, and the absolute difference (prediction error) between the input image signal 115 and the prediction signal 206 If the SAD is defined as the absolute cumulative sum of the signals 116, the following determination formula is used.
  • K is the cost and ⁇ is a constant.
  • is a Lagrangian undetermined multiplier determined based on the quantization scale and the value of the quantization parameter. In this determination formula, the mode giving the value with the smallest cost K is selected as the optimum prediction mode.
  • the prediction selection information 123 may be determined using (a) only prediction information or (b) only SAD instead of the formula (1), and Hadamard transformation is applied to these (a) and (b). A value or a value close to it may be used.
  • a provisional encoding unit is prepared, and the amount of code when the prediction error signal 116 generated in the prediction mode by the provisional encoding unit is actually encoded, the input image signal 115 and the decoded image signal
  • the prediction selection information 123 may be determined using a square error with 119.
  • the judgment formula in this case is as follows.
  • J is an encoding cost
  • D is an encoding distortion representing a square error between the input image signal 114 and the decoded image signal 118.
  • R represents a code amount estimated by provisional encoding.
  • the cost J of Equation (2) When the encoding cost J of Equation (2) is used, provisional encoding and local decoding processing are required for each prediction mode, so that the circuit scale or calculation amount increases. However, since a more accurate code amount and encoding distortion are used, high encoding efficiency can be maintained.
  • the cost may be calculated using only R or only D instead of the expression (2), or the cost function may be created using a value approximating R or D.
  • inter prediction signal generation device 109 Next, the inter prediction signal generation device 109 will be described with reference to FIG.
  • the inter prediction signal generation device 109 includes a prediction separation switch 201, a geometric transformation prediction unit 202, a second geometric transformation parameter derivation unit 203, a first geometric transformation parameter derivation unit 204, and a predicted geometric transformation parameter derivation unit 205.
  • the second geometric transformation parameter derivation unit 203 uses the motion information 210 of the prediction target block output from the motion information search unit 106 and the motion information of the encoded block stored in the encoding control unit 114 to perform prediction.
  • the second geometric transformation parameter 208 of the target block is derived.
  • the motion information 210 stored in the encoding control unit 114 is a motion vector included in the motion information 210 of the adjacent block, and is hereinafter referred to as “adjacent motion vector”.
  • the second geometric transformation parameter derivation unit 203 includes a motion information acquisition unit 501 and a second parameter derivation unit 502.
  • the motion information acquisition unit 501 determines an adjacent block from which motion information is acquired from among a plurality of adjacent blocks, and acquires motion information of the adjacent block, for example, a motion vector.
  • the second parameter derivation unit 502 derives a second geometric transformation parameter from the motion vector of the adjacent block.
  • FIG. 6A shows an example in which the sizes of prediction target blocks and adjacent blocks (for example, 16 ⁇ 16 pixel blocks) match.
  • a hatched pixel block p is a pixel block that has already been encoded or predicted (hereinafter referred to as “predicted pixel block”).
  • a block c with dot hatching indicates a prediction target block, and a pixel block n displayed in white is an uncoded pixel (unpredicted) block.
  • X represents an encoding (prediction) target pixel block.
  • the adjacent block A is the adjacent block on the left of the prediction target block X
  • the adjacent block B is the adjacent block on the prediction target block X
  • the adjacent block C is the adjacent block on the upper right of the prediction target block X
  • the adjacent block D is This is an adjacent block at the upper left of the prediction target block X.
  • the adjacent motion vector held in the internal memory of the encoding control unit 114 is only the motion vector of the predicted pixel block. As shown in FIG. 3A, since the pixel block is encoded and predicted from the upper left to the lower right, when the pixel block X is predicted, the right and lower pixel blocks are still encoded. Has not been made. Therefore, an adjacent motion vector cannot be derived from these adjacent blocks.
  • 6B to 6E are diagrams illustrating examples of adjacent blocks when the prediction target block is an 8 ⁇ 8 pixel block.
  • bold lines represent macroblock boundaries.
  • 6B is a pixel block located at the upper left in the macro block
  • FIG. 6C is a pixel block located at the upper right in the macro block
  • FIG. 6D is a pixel block located at the lower left in the macro block
  • FIG. An example is shown in which each pixel block located at the lower right in the block is a prediction target block.
  • the position of the adjacent block changes according to the encoding order of the 8 ⁇ 8 pixel block.
  • the pixel block becomes an encoded pixel block and is used as an adjacent block of the pixel block to be processed later.
  • the pixel block located at the upper right of the encoded pixel block is set as an adjacent block.
  • adjacent blocks having the closest Euclidean distance to the prediction target block X are referred to as adjacent blocks A, B, C, and D, respectively.
  • adjacent blocks A, B, C, and D adjacent blocks having the closest Euclidean distance to the prediction target block X
  • adjacent blocks A, B, C, and D are referred to as adjacent blocks A, B, C, and D, respectively.
  • the nearest block among the upper right adjacent blocks is the adjacent block C.
  • a block having the shortest Euclidean distance with respect to the prediction target block X is set as an adjacent block.
  • the block is 16 ⁇ 16 pixels and 8 ⁇ 8 pixels has been described as an example.
  • a square pixel block such as 32 ⁇ 32 pixels, 4 ⁇ 4 pixels, or the like is used using a similar framework.
  • Adjacent blocks may be determined for rectangular pixel blocks such as 8 pixels and 8 ⁇ 16 pixels.
  • adjacent blocks may be defined more widely.
  • a pixel block on the left of the adjacent block A may be used, or a pixel block further on the adjacent block B may be used.
  • the second geometric transformation parameter 208 is derived by the second parameter deriving unit 502.
  • the adjacent motion vectors held by the adjacent blocks are defined by equations (3) to (6), respectively.
  • the motion information 210 provided from the motion information search unit 106 is defined by equation (7). Note that the motion information 210 indicates a motion vector of the prediction target block X.
  • the second geometric transformation parameter 208 is derived using the motion vector and the adjacent motion vector represented by the equations (3) to (7).
  • the transformation formula is expressed by the following formula (8).
  • the parameters (c, f) correspond to motion vectors
  • the parameters (a, b, d, e) indicate parameters associated with geometric deformation.
  • u and v indicate the coordinates of the encoding target block
  • x and y indicate the coordinates of the reference image. If the parameter (a, b, d, e) is (1, 0, 0, 1), it means that it is the same as the motion compensation of the parallel movement model (formula (18) described later).
  • affine transformation was shown here as geometric transformation
  • geometric such as a bilinear transformation, Helmert transformation, secondary conformal transformation, projective transformation, and three-dimensional projective transformation corresponding to geometric transformation
  • a transformation may be used.
  • the required number of parameters varies depending on the geometric transformation to be used, but a suitable geometric transformation can be selected according to the nature of the image to be applied, depending on the amount of code when the parameters are encoded and the corresponding geometric transformation pattern. Should be selected.
  • affine transformation will be described.
  • equation (8) coordinates (x, y) are converted to coordinates (u, v) by affine transformation.
  • Six parameters a, b, c, d, e, and f included in Expression (8) represent geometric transformation parameters.
  • affine transformation since these six types of parameters are derived from adjacent vectors, six or more input values are required.
  • a geometric transformation parameter is derived by the following equation (9).
  • the motion vector is 1 ⁇ 4 precision
  • the precision of the parameters (a, b, d, e) is 1/64.
  • ax and ay are variables based on the size of the prediction target block, and are calculated by the following equation (10).
  • mb_size_x and mb_size_y indicate the horizontal and vertical sizes of the macroblock.
  • Equation (8) shows an example in which a, b, d, and e are obtained as real numbers, respectively, but by determining the calculation accuracy of these parameters in advance, an integer can be easily obtained as in Equation (9). Is possible.
  • the predicted geometric transformation parameter derivation unit 205 includes a motion information acquisition unit 701 and a predicted geometric transformation parameter calculation unit 702 as shown in FIG.
  • the motion information acquisition unit 501 determines an adjacent block in the same procedure as the motion information acquisition unit 501 held by the second geometric transformation derivation unit 203. However, the motion information calculated from the adjacent block is a geometric transformation parameter.
  • the motion information acquisition unit 701 derives a prediction geometric transformation parameter 212 of the prediction target block using the motion information 210 of the encoded block stored in the encoding control unit 114.
  • the motion information 210 stored in the encoding control unit 114 is motion information 210 of adjacent encoded blocks, and is hereinafter referred to as “adjacent geometric transformation parameter”.
  • a method for deriving the predicted geometric transformation parameter 212 will be described with reference to FIG. 6A.
  • the adjacent geometric transformation parameters held by the adjacent blocks are defined by equations (11) to (14), respectively.
  • ap indicates an affine transformation parameter, which is a six-dimensional parameter as shown in Equation (8).
  • the predicted geometric transformation parameter calculation unit 702 calculates a predicted geometric transformation parameter by median processing using the spatial correlation between the prediction target block and the adjacent block.
  • pred_ap represents a predicted geometric transformation parameter.
  • the function affine_median () is a function that takes the median value of the 6-dimensional affine transformation parameters. Moreover, you may determine a prediction geometric transformation parameter using following Formula.
  • Equation (16) indicates a scalar median.
  • T in equation (16) means transposition.
  • equation (16) a predicted geometric transformation parameter is derived by taking a median value for each geometric transformation parameter component.
  • four blocks that have already been encoded (further located on the left, upper located, upper left, and upper right) are adjacent.
  • a geometric transformation parameter is derived using Equation (9) as a block. It is also possible for ap A to use this geometric transformation parameter as an adjacent geometric transformation parameter. It is also possible to derive predicted geometric transformation parameters by re-deriving geometric transformation parameters for adjacent blocks B, C, and D in the same manner.
  • the predicted geometric transformation parameter 212 derived by the predicted geometric transformation parameter derivation unit 205 is output to the first geometric transformation parameter derivation unit 204.
  • the first geometric transformation parameter derivation unit 204 adds the inputted geometric transformation parameter derivation parameter information 211 of the prediction target block and the predicted geometric transformation parameter 212.
  • the added parameter becomes the first geometric transformation parameter 209 and is output to the geometric transformation prediction unit 202.
  • the first geometric transformation parameter deriving unit 204 derives the first geometric transformation parameter 209 by adding the predicted geometric transformation parameter 212 and the derivation parameter information 211 of the geometric transformation parameter of the prediction target block.
  • the predicted geometric transformation parameter is ⁇ pred_ap
  • the first geometric transformation parameter derivation unit 204 adds the predicted geometric transformation parameter 212 ( ⁇ pred_ap) and the derived parameter information 211 to obtain the first geometric transformation parameter 209. To derive.
  • the prediction geometric transformation parameter 212 and the first geometric transformation parameter derivation unit 204 define the derivation formula so as to reduce the information amount of the first geometric transformation parameter 209 calculated in the prediction target block as much as possible. .
  • the geometric transformation prediction unit 202 has a function of generating a prediction signal using the input geometric transformation parameters and the reference image signal 207.
  • the geometric transformation prediction unit 202 includes a geometric transformation unit 401 and an interpolation unit 402 as shown in FIG.
  • the geometric transformation unit 202 performs geometric transformation on the reference image signal 207 and calculates the position of the predicted pixel.
  • the interpolation unit 402 calculates the predicted pixel value corresponding to the fractional position of the predicted pixel obtained by the geometric transformation by interpolation or the like.
  • the prediction target block is a square pixel block CR composed of pixels indicated by ⁇ .
  • the corresponding pixel of motion compensation prediction is indicated by ⁇ .
  • a pixel block MER composed of pixels indicated by ⁇ is square.
  • the pixel corresponding to the geometric transformation prediction is indicated by x, and the pixel block GTR composed of these pixels is, for example, a parallelogram.
  • the region after motion compensation and the region after geometric transformation describe the corresponding region of the reference image signal according to the coordinates of the frame to be encoded.
  • the geometric transformation prediction it is possible to generate a prediction signal in accordance with deformation such as rotation, enlargement / reduction, shearing, and mirror transformation of the rectangular pixel block.
  • the geometric transformation unit 401 calculates coordinates (u, v) after the geometric transformation using the input geometric transformation parameters using the equation (8).
  • the calculated coordinates (u, v) after geometric transformation are real values. Therefore, the predicted value is generated by interpolating the luminance value corresponding to the coordinates (u, v) from the reference image signal.
  • an interpolation method is used.
  • the interpolation method is expressed by the following equation (17).
  • R (x, y) the integer pixel value of the used reference image signal
  • a new prediction signal is generated by applying interpolation for each coordinate in the prediction target block subjected to geometric transformation.
  • the geometric transformation prediction unit 202 can also generate a prediction signal using a conventional translation model.
  • the coordinate conversion formula of the translation model is expressed by formula (19).
  • Equation (19) is the same as the parameter (a, b, d, e) in Equation (8) being (1, 0, 0, 1).
  • the geometric transformation prediction unit 202 can calculate the equation (2) regardless of the values of the first geometric transformation parameter 208 and the second geometric transformation parameter 209.
  • the conventional motion compensated prediction can be realized by changing 8) to Equation (19) and deriving coordinates.
  • bilinear interpolation is used as an interpolation method.
  • nearest neighbor interpolation cubic convolution interpolation
  • linear filter interpolation linear filter interpolation
  • Lagrange interpolation Lagrange interpolation
  • spline spline
  • Any interpolation method such as an interpolation method or a Lanczos interpolation method may be applied.
  • the prediction separation switch 201 switches the output terminals of the two prediction signals 206 output from the geometric transformation prediction unit 202. That is, the prediction separation switch 201 uses the prediction selection information 213 (FIG. 1) as the output terminal of the prediction signal 215 generated by the first geometric transformation parameter 209 and the output terminal of the prediction signal 214 generated by the second geometric transformation parameter 208. In accordance with 123).
  • Examples of the prediction selection information 213 and 123 are shown in FIG.
  • the skip mode is selected. In the skip mode, transform coefficients, motion vectors, etc. are not encoded. For this reason, it means that the first geometric transformation prediction that needs to encode the additional motion information 210 is not selected.
  • the index of the prediction selection information 213 is 9, intra prediction is selected. In this case, since the output terminal of the prediction separation switch 110 is connected to the intra prediction signal generation device 107, this means that the inter prediction signal generation device 109 does not need to perform the prediction signal generation processing.
  • index tables may be divided into a plurality of tables, or a plurality of index tables may be integrated. Moreover, it is not always necessary to use the same term, and it may be arbitrarily changed depending on the form to be used. Furthermore, each element described in the index table may be changed so as to be described by an independent flag.
  • the prediction separation switch 201 sets the prediction selection information 213 to the first geometric transformation parameter. It is determined whether it is 209 (S502). When this determination is YES, the prediction separation switch 201 switches the output terminal of the geometric transformation prediction unit 202 to the first prediction signal 215. On the other hand, when the determination is NO, the prediction separation switch 201 switches the output terminal of the geometric transformation prediction unit 202 to the second prediction signal 215.
  • the motion information acquisition unit 701 in the predicted geometric transformation parameter derivation unit 205 determines an adjacent block based on the motion information 210 input from the outside (S508). Using the determined adjacent block motion information 210, an adjacent geometric transformation parameter is derived (S509). Receiving the derived adjacent geometric transformation parameter, the predicted geometric transformation parameter derivation unit 702 derives the predicted geometric transformation parameter 212 using equation (15) or equation (16) (S510). The predicted geometric transformation parameter 212 is output to the first geometric transformation parameter derivation unit 204.
  • the first geometric transformation parameter derivation unit 204 derives the first geometric transformation parameter 209 using the derived parameter information 211 and the predicted geometric transformation parameter 212 input from the outside (S511).
  • the first geometric transformation parameter 209 is input to the geometric transformation prediction unit 202, and the geometric transformation unit 401 derives the coordinates after geometric transformation using the equation (8) (S512).
  • the interpolation unit 402 Based on the calculated coordinates, the interpolation unit 402 performs an interpolation process on the reference image signal 207 input from the outside using Expression (18) to generate a first prediction signal 215 ( S513).
  • the first prediction signal 215 is output to the outside via the prediction separation switch 201 to which the output terminal is connected (S515), and the first geometric transformation parameter 209 used in the prediction target block is stored in the memory (S514). ).
  • the first geometric transformation parameter 209 held in the memory is used as an adjacent geometric transformation parameter or an adjacent motion vector of the next block (S517).
  • step S502 the motion information acquisition unit 501 in the second geometric transformation parameter derivation unit 203 determines an adjacent block based on the motion information 210 input from the outside (S503). ). An adjacent motion vector is derived using the determined adjacent block motion information 210 (S504). Receiving the derived adjacent motion vector, the second parameter deriving unit 502 derives the second geometric transformation parameter 208 using the equations (9) to (10) (S505).
  • the second geometric transformation parameter 208 is input to the geometric transformation prediction unit 202, and the geometric transformation unit 401 derives coordinates after the geometric transformation using the equation (8) (S506). Based on the derived coordinates, the interpolation unit 402 performs an interpolation process on the reference image signal 207 input from the outside using Expression (18) to generate the second prediction signal 214 ( S507).
  • the second prediction signal 214 is output to the outside via the prediction separation switch 201 (S515), and the second geometric transformation parameter 208 used in the prediction target block is stored in the memory (S514).
  • the second geometric transformation parameter 208 held in the memory is used as an adjacent geometric transformation parameter or an adjacent motion vector of the next block (S517).
  • the syntax 1600 mainly has three parts.
  • the high-level syntax 1601 has higher layer syntax information that is equal to or higher than a slice.
  • the slice level syntax 1602 has information necessary for decoding for each slice, and the macroblock level syntax 1603 has information necessary for decoding for each macroblock.
  • High level syntax 1601 includes sequence and picture level syntax, such as sequence parameter set syntax 1604 and picture parameter set syntax 1605.
  • the slice level syntax 1602 includes a slice header syntax 1606, a slice data syntax 1607, and the like.
  • the macroblock level syntax 1603 includes a macroblock layer syntax 1608, a macroblock prediction syntax 1609, and the like.
  • slice_affine_motion_prediction_flag is a syntax element indicating whether to apply geometric transformation prediction to a slice.
  • the geometric transformation prediction unit 202 does not use the parameters (a, b, d, e) in Equation (8) but uses Equation (19) for this slice.
  • H.264 represents motion compensation prediction using a translation model used in H.264 and the like, and parameters (c, f) correspond to motion vectors.
  • this flag is 0, it is the same as the motion compensation prediction of the conventional translation model.
  • slice_affine_motion_prediction_flag 1
  • the prediction separation switch 201 dynamically switches the prediction signal as indicated by the prediction selection information 213 in the slice.
  • mb_skip_flag is a flag indicating whether or not the macroblock is encoded in the skip mode. In the skip mode, transform coefficients, motion vectors, etc. are not encoded. For this reason, the first geometric transformation prediction is not applied to the skip mode.
  • AvailAffineMode is an internal parameter indicating whether or not the second geometric transformation prediction can be used in the macroblock. When AvailAffineMode is 0, it means that the prediction selection information 213 is set not to use the second geometric transformation prediction. When the adjacent motion vector of the adjacent block and the motion vector of the prediction target block have the same value, AvailAffineMode is 0. Otherwise, AvailAffineMode is 1.
  • the setting of AvailAffineMode can also be set using an adjacent geometric transformation parameter or an adjacent motion vector. For example, when the adjacent motion vector points in a completely different direction, there is a possibility that an object boundary exists in the adjacent block of the current prediction target block. Therefore, it is possible to set AvailAffineMode to 0.
  • mb_affine_motion_skip_flag indicating whether to use the second geometric transformation prediction or the motion compensation prediction is encoded.
  • mb_affine_motion_skip_flag 1, it means that the second geometric transformation prediction is applied to the skip mode.
  • mb_affine_motion_skip_flag 0, it means that motion compensation prediction is applied using equation (19).
  • mb_type indicates macroblock type information. That is, whether the current macroblock is intra-coded, inter-coded, what block shape is being predicted, whether the prediction direction is unidirectional prediction or bidirectional prediction, etc. Contains information.
  • the mb_type is passed to the macroblock prediction syntax and the submacroblock prediction syntax indicating the syntax of the subblock in the macroblock.
  • mb_affine_pred_flag indicates whether the first geometric transformation prediction or the second geometric transformation prediction is used in the block.
  • the prediction selection information 213 is set to use the second geometric transformation parameter.
  • the prediction selection information 213 is set to use the first geometric transformation parameter.
  • NumMbPart () is an internal function that returns the number of block divisions specified in mb_type. It is 1 for a 16 ⁇ 16 pixel block, 2 for an 8 ⁇ 16 pixel block, and 2 for an 8 ⁇ 8 pixel block. In the case, 4 is output.
  • mv_l0 and mv_l1 indicate motion vector difference information in the macroblock.
  • the motion vector information is a value set by the motion information search unit 106 and obtained by taking a difference from a predicted motion vector not disclosed in the present embodiment.
  • mvd_l0_affine and mvd_l1_affine indicate derived parameters in the macroblock, and indicate difference information of components (a, b, d, e) excluding motion vectors of affine transformation parameters. This syntax element is encoded only when the first geometric transformation parameter is selected.
  • mb_affine_pred_flag indicates whether to use the first geometric transformation prediction or the second geometric transformation prediction in the block.
  • the prediction selection information 213 is set to use the second geometric transformation parameter.
  • the prediction selection information 213 is set to use the first geometric transformation parameter.
  • mv_l0 and mv_l1 indicate motion vector difference information in the sub macroblock.
  • the motion vector information is a value set by the motion information search unit 106 and obtained by taking a difference from a predicted motion vector not disclosed in the present embodiment.
  • Mvd_l0_affine and mvd_l1_affine in the figure indicate derived parameters in the sub-macroblock, and indicate difference information of components (a, b, d, e) excluding the motion vectors of the affine transformation parameters.
  • This syntax element is encoded only when the first geometric transformation parameter is selected.
  • the conventional motion compensated prediction or the second geometric transformation prediction using the translation model can be selected, and other inter predictions can be selected.
  • the first geometric transformation prediction or the second geometric transformation prediction can be selected.
  • syntax elements not defined in the present embodiment may be inserted between lines in the syntax tables shown in FIGS. 12 to 17, and descriptions regarding other conditional branches may be included.
  • the syntax table may be divided into a plurality of tables, or a plurality of syntax tables may be integrated. Moreover, it is not always necessary to use the same term, and it may be arbitrarily changed depending on the form to be used.
  • two geometric transformation parameters indicating information related to the shape of the image by the geometric transformation of the pixel block that is, the first geometric transformation parameter and the second geometric transformation parameter are derived, and any of these geometric transformation parameters is derived.
  • a prediction signal is generated by performing motion compensation prediction using a geometric transformation parameter selected according to prediction selection information indicating whether or not to select.
  • FIG. 18 is a diagram illustrating an example of the macroblock layer syntax 1607.
  • Mb_type shown in the figure indicates macroblock type information. That is, whether the current macroblock is intra-coded, inter-coded, what block shape is being predicted, whether the prediction direction is unidirectional prediction or bidirectional prediction, etc. Contains information.
  • the mb_type is passed to the macroblock prediction syntax and the submacroblock prediction syntax indicating the syntax of the subblock in the macroblock.
  • mb_additional_affine_motion_flag indicates flag information for selecting whether to use the first geometric transformation parameter or the second geometric transformation parameter for the prediction target block. When this flag is 0, the second geometric transformation parameter is used, and when this flag is 1, the first geometric transformation parameter is used.
  • mb_affine_pred_flag uses a geometric transformation prediction including a first geometric transformation prediction or a second geometric transformation prediction, or uses a motion compensated prediction of a translation model.
  • the prediction selection information 213 is set to use the motion compensated prediction of the translation model regardless of the mb_additional_affine_motion_flag.
  • the prediction selection information 213 is set to use the first geometric transformation parameter or the second geometric transformation parameter according to the flag information of mb_additional_affine_motion_flag.
  • mb_affine_pred_flag is a block that uses geometric transformation prediction including first geometric transformation prediction or second geometric transformation prediction, or uses motion compensation prediction of a translation model. Indicates whether or not When the flag is 0, the prediction selection information 213 is set to use the motion compensated prediction of the translation model regardless of the mb_additional_affine_motion_flag. On the other hand, when the flag is 1, the prediction selection information 213 is set to use the first geometric transformation parameter or the second geometric transformation parameter according to the flag information of mb_additional_affine_motion_flag.
  • the conventional motion compensation or the second geometric transformation prediction using the translation model can be selected. Then, at the macroblock level, it is determined whether the first geometric transformation prediction or the second geometric transformation prediction is used, and the geometric transformation prediction including the first geometric transformation prediction or the second geometric transformation prediction at the sub macroblock level. Or using motion compensated prediction of a translation model.
  • syntax elements not defined in the present embodiment may be inserted between lines in the syntax tables shown in FIGS. 18 to 20, and descriptions regarding other conditional branches may be included.
  • the syntax table may be divided into a plurality of tables, or a plurality of syntax tables may be integrated. Moreover, it is not always necessary to use the same term, and it may be arbitrarily changed depending on the form to be used.
  • the video decoding device 300 decodes, for example, encoded data generated by the video encoding device according to the first embodiment.
  • the video decoding device 300 decodes the encoded data 311 stored in the input buffer 301 and outputs a decoded image signal 317 to the output buffer 309.
  • the encoded data 311 is encoded data that is transmitted from, for example, the moving image encoding apparatus 100, transmitted through a storage system or a transmission system, once stored in the input buffer 301, and multiplexed.
  • the video decoding device 300 includes an entropy decoding unit 302, an inverse quantization / inverse conversion unit 303, an adder 304, a reference image memory 305, an intra prediction signal generation device 306, an inter prediction signal generation device 307, and a prediction separation switch 308.
  • the moving picture decoding apparatus 300 is also connected to the input buffer 301, the output buffer 309, and the decoding control unit 310.
  • the entropy decoding unit 302 decodes the encoded data 311 by syntax analysis based on the syntax for each frame or field.
  • the entropy decoding unit 302 sequentially entropy-decodes the code string of each syntax, and reproduces the motion information 315, the derived parameter information 316, the encoding parameters of the decoding target block, and the like.
  • the encoding parameter includes all parameters necessary for decoding such as prediction information, information on transform coefficients, information on quantization, and the like.
  • the transform coefficient decoded by the entropy decoding unit 302 is input to an inverse quantization / inverse transform unit 303 including an inverse quantizer and an inverse transformer.
  • Various information relating to the quantization decoded by the entropy decoding unit 302, that is, the quantization parameter and the quantization matrix are set in the internal memory of the decoding control unit 310 and loaded when used as an inverse quantization process.
  • the inverse quantization process is first performed by the inverse quantizer using the information on the loaded quantization.
  • the inverse quantized transform coefficient is then subjected to inverse transform processing, for example, inverse discrete cosine transform, by an inverse transformer.
  • inverse transform processing for example, inverse discrete cosine transform
  • the inverse orthogonal transform has been described.
  • the inverse quantization / inverse transform unit 303 performs the corresponding inverse quantization and inverse wavelet transform. Good.
  • the restored prediction error signal 312 is input to the adder 304.
  • the adder 304 adds the prediction error signal 312 and the prediction signal 416 output from the prediction separation switch 308 described later to generate a decoded image signal 317.
  • the generated decoded image signal 317 is output from the moving image decoding apparatus 300, temporarily stored in the output buffer 317, and then output according to the output timing managed by the decoding control unit 310.
  • the decoded image signal 317 is stored in the reference image memory 305 and becomes a reference image signal 313.
  • the reference image signal 313 is sequentially read from the reference image memory 305 for each frame or each field, and is input to the intra prediction signal generation device 306 or the inter prediction signal generation device 307.
  • the motion information 315 used in the decoding target pixel block is stored in the decoding control unit 310, and is appropriately loaded from the decoding control unit 310 and used in the inter prediction signal generation processing of the next block.
  • the intra prediction signal generation device 306 has the same function and configuration as the intra prediction signal generation device 107 in the video encoding device 100 shown in FIG. That is, the intra prediction signal generation device 306 performs intra prediction using the input reference image signal 313.
  • a prediction signal is generated by performing pixel interpolation in the prediction direction such as the vertical direction and the horizontal direction using an encoded reference pixel value adjacent to the prediction target block.
  • the interpolated pixel value may be copied in a predetermined prediction direction.
  • the generated prediction signal 416 is output to the prediction separation switch 308.
  • the inter prediction signal generation device 307 has the same function and configuration as the inter prediction signal generation device 109 shown in FIGS. 1 and 2 and FIGS. That is, the inter prediction signal generation device 307 generates the prediction signal 416 using the input motion information 315, derivation parameter information 316, reference image signal 313, and prediction selection information 314.
  • the motion information 315, the derived parameter information 316, the reference image signal 313, and the prediction selection information 314 are the motion information 210, the derived parameter information 211, the reference image signal 207, and the input to the inter prediction signal generation device 109 of the video encoding device 100.
  • the prediction signal 416 is generated in the inter prediction signal generation device 109 shown in FIG.
  • the generated prediction signal 416 is output to the prediction separation switch 308.
  • the prediction separation switch 308 selects the output terminal of the intra prediction signal generation device 308 and the output terminal of the inter prediction signal generation device 307 according to the prediction selection information 314.
  • the switch is connected to the intra prediction signal generation device 308.
  • the prediction selection information 314 is inter prediction, the switch is connected to the inter prediction signal generation device 307.
  • the prediction selection information 314 is the same as the prediction selection information 123 set by the prediction selection unit 111 of the video encoding device 100, and is shown in FIG.
  • the encoded data 311 decoded by the video decoding device 300 may have the same syntax structure as that of the video encoding device 100.
  • the same syntax as in FIGS. 12 to 17 is used.
  • the syntax 1600 has mainly three parts as shown in FIG.
  • the high-level syntax 1601 has higher layer syntax information that is equal to or higher than a slice.
  • the slice level syntax 1602 has information necessary for decoding for each slice, and the macroblock level syntax 1603 has information necessary for decoding for each macroblock.
  • High level syntax 1601 includes sequence and picture level syntax, such as sequence parameter set syntax 1604 and picture parameter set syntax 1605.
  • the slice level syntax 1602 includes a slice header syntax 1606, a slice data syntax 1607, and the like.
  • the macroblock level syntax 1603 includes a macroblock layer syntax 1608, a macroblock prediction syntax 1609, and the like.
  • slice_affine_motion_prediction_flag is a syntax element indicating whether to apply geometric transformation prediction to a slice.
  • the geometric transformation prediction unit 202 does not use the parameters (a, b, d, e) in Equation (8) but uses Equation (19) for this slice.
  • H.264 represents motion compensation prediction using a translation model used in H.264 and the like, and parameters (c, f) correspond to motion vectors.
  • this flag is 0, it is the same as the motion compensation prediction of the conventional translation model.
  • slice_affine_motion_prediction_flag 1
  • the prediction separation switch 201 dynamically switches the prediction signal as indicated by the prediction selection information 314 in the slice.
  • mb_skip_flag is a flag indicating whether or not the macroblock is encoded in the skip mode. In the skip mode, transform coefficients, motion vectors, etc. are not encoded. For this reason, the first geometric transformation prediction is not applied to the skip mode.
  • AvailAffineMode is an internal parameter indicating whether or not the second geometric transformation prediction can be used in the macroblock. When AvailAffineMode is 0, it means that the prediction selection information 314 is set not to use the second geometric transformation prediction. When the adjacent motion vector of the adjacent block and the motion vector of the prediction target block have the same value, AvailAffineMode is 0. Otherwise, AvailAffineMode is 1.
  • the setting of AvailAffineMode can also be set using an adjacent geometric transformation parameter or an adjacent motion vector. For example, when the adjacent motion vector points in a completely different direction, there is a possibility that an object boundary exists in the adjacent block of the current prediction target block. Therefore, it is possible to set AvailAffineMode to 0.
  • mb_affine_motion_skip_flag indicating whether to use the second geometric transformation prediction or the motion compensation prediction is encoded.
  • mb_affine_motion_skip_flag 1, it means that the second geometric transformation prediction is applied to the skip mode.
  • mb_affine_motion_skip_flag 0, it means that motion compensation prediction is applied using equation (19).
  • mb_type indicates macroblock type information. That is, whether the current macroblock is intra-coded, inter-coded, what block shape is being predicted, whether the prediction direction is unidirectional prediction or bidirectional prediction, etc. Contains information.
  • the mb_type is passed to the macroblock prediction syntax and the submacroblock prediction syntax indicating the syntax of the subblock in the macroblock.
  • mb_affine_pred_flag indicates whether the first geometric transformation prediction or the second geometric transformation prediction is used in the block.
  • the prediction selection information 314 is set to use the second geometric transformation parameter.
  • the prediction selection information 314 is set to use the first geometric transformation parameter.
  • NumMbPart () is an internal function that returns the number of block divisions specified in mb_type. It is 1 for a 16 ⁇ 16 pixel block, 2 for an 8 ⁇ 16 pixel block, and 2 for an 8 ⁇ 8 pixel block. In the case, 4 is output.
  • mv_l0 and mv_l1 indicate motion vector difference information in the macroblock.
  • the motion vector information is a value that is set by the motion information search unit 106 of the video encoding device 100 and is obtained by taking a difference from a predicted motion vector that is not disclosed in the present embodiment.
  • mvd_l0_affine and mvd_l1_affine indicate derived parameters in the macroblock, and indicate difference information of components (a, b, d, e) excluding motion vectors of affine transformation parameters. This syntax element is encoded only when the first geometric transformation parameter is selected.
  • mb_affine_pred_flag indicates whether to use the first geometric transformation prediction or the second geometric transformation prediction in the block.
  • the prediction selection information 314 is set to use the second geometric transformation parameter.
  • the prediction selection information 314 is set to use the first geometric transformation parameter.
  • mv_l0 and mv_l1 indicate motion vector difference information in the sub macroblock.
  • the motion vector information is a value that is set by the motion information search unit 106 of the video encoding device 100 and is obtained by taking a difference from a predicted motion vector that is not disclosed in the present embodiment.
  • Mvd_l0_affine and mvd_l1_affine in the figure indicate derived parameters in the sub-macroblock, and indicate difference information of components (a, b, d, e) excluding the motion vectors of the affine transformation parameters.
  • This syntax element is encoded only when the first geometric transformation parameter is selected.
  • the syntax structure shown in FIGS. 12 to 17 can select the conventional motion compensation prediction or the second geometric transformation prediction using the translation model when the prediction target pixel block is in the skip mode, and other inter predictions.
  • the first geometric transformation prediction or the second geometric transformation prediction can be selected.
  • syntax elements not defined in the present embodiment may be inserted between lines in the syntax tables shown in FIGS. 12 to 17, and descriptions regarding other conditional branches may be included.
  • the syntax table may be divided into a plurality of tables, or a plurality of syntax tables may be integrated. Moreover, it is not always necessary to use the same term, and it may be arbitrarily changed depending on the form to be used.
  • mb_type indicates macroblock type information. That is, whether the current macroblock is intra-coded, inter-coded, what block shape is being predicted, whether the prediction direction is unidirectional prediction or bidirectional prediction, etc. Contains information.
  • the mb_type is passed to the macroblock prediction syntax and the submacroblock prediction syntax indicating the syntax of the subblock in the macroblock.
  • mb_additional_affine_motion_flag indicates flag information for selecting whether to use the first geometric transformation parameter or the second geometric transformation parameter for the prediction target block. When this flag is 0, the second geometric transformation parameter is used, and when this flag is 1, the first geometric transformation parameter is used.
  • mb_affine_pred_flag uses a geometric transformation prediction including a first geometric transformation prediction or a second geometric transformation prediction or uses a motion compensation prediction of a translation model in a block.
  • the prediction selection information 314 is set to use the motion compensation prediction of the translation model regardless of the mb_additional_affine_motion_flag.
  • the prediction selection information 314 is set to use the first geometric transformation parameter or the second geometric transformation parameter according to the flag information of mb_additional_affine_motion_flag.
  • mb_affine_pred_flag is a block that uses geometric transformation prediction including first geometric transformation prediction or second geometric transformation prediction, or uses motion compensation prediction of a translation model. Indicates whether or not When the flag is 0, the prediction selection information 314 is set to use the motion compensation prediction of the translation model regardless of the mb_additional_affine_motion_flag. On the other hand, when the flag is 1, the prediction selection information 314 is set to use the first geometric transformation parameter or the second geometric transformation parameter according to the flag information of mb_additional_affine_motion_flag.
  • the conventional motion compensation or the second geometric transformation prediction using the translation model can be selected. Then, at the macroblock level, it is determined whether the first geometric transformation prediction or the second geometric transformation prediction is used, and the geometric transformation prediction including the first geometric transformation prediction or the second geometric transformation prediction at the sub macroblock level. Or using motion compensated prediction of a translation model.
  • syntax elements not defined in the present embodiment may be inserted between lines in the syntax tables shown in FIGS. 18 to 20, and descriptions regarding other conditional branches may be included.
  • the syntax table may be divided into a plurality of tables, or a plurality of syntax tables may be integrated. Moreover, it is not always necessary to use the same term, and it may be arbitrarily changed depending on the form to be used.
  • the processing target frame is divided into short blocks of 16 ⁇ 16 pixel size or the like, and as shown in FIG.
  • the case of encoding / decoding has been described, but the encoding order and decoding order are not limited to this.
  • encoding and decoding may be performed sequentially from the lower right to the upper left, or encoding and decoding may be performed sequentially from the center of the screen toward the spiral.
  • encoding and decoding may be performed in order from the upper right to the lower left, or encoding and decoding may be performed in order from the peripheral part to the center part of the screen.
  • the block size has been described as a 4 ⁇ 4 pixel block and an 8 ⁇ 8 pixel block.
  • the prediction target block does not need to have a uniform block shape, and 16 ⁇ Any block size such as an 8 pixel block, an 8 ⁇ 16 pixel block, an 8 ⁇ 4 pixel block, or a 4 ⁇ 8 pixel block may be used.
  • the luminance signal and the color difference signal are not divided and described as an example limited to one color signal component.
  • different prediction methods may be used, or the same prediction method may be used.
  • the prediction method selected for the color difference signal is encoded or decoded by the same method as the luminance signal.
  • the present invention is not limited to the above-described embodiment as it is, and can be embodied by modifying constituent elements without departing from the scope of the invention in the implementation stage.
  • various inventions can be formed by appropriately combining a plurality of components disclosed in the embodiment. For example, some components may be deleted from all the components shown in the embodiment.
  • constituent elements over different embodiments may be appropriately combined.
  • the method of the present invention described in the above embodiment can be executed by a computer, and as a program that can be executed by the computer, a magnetic disk (flexible disk, hard disk, etc.), an optical disk (CD-ROM, DVD) It is also possible to store and distribute in a recording medium such as a semiconductor memory.
  • the prediction signal generation device the moving image encoding device, and the moving image decoding device according to the embodiment of the present invention, the error of the parameter generated by the derivation of the geometric transformation parameter used for the geometric transformation motion compensation prediction is reduced.
  • a prediction signal generation device a moving image encoding device, and a moving image decoding device that suppress error propagation and improve prediction efficiency without increasing the code amount.
  • the following prediction signal generation method, video encoding method, and video decoding method can be provided.
  • a prediction signal generation method for generating a prediction signal sets prediction selection information indicating whether to use a first geometric conversion parameter or a second geometric conversion parameter indicating information related to the shape of an image obtained by pixel block geometric conversion. And one or more second adjacent blocks for which prediction signal generation processing has already been completed among a plurality of first adjacent blocks adjacent to one pixel block of the plurality of pixel blocks into which the image signal is divided And obtaining the first geometric transformation parameter and the second geometric transformation parameter, and predictive geometric transformation of the one pixel block from the geometric transformation parameters of the one or more second neighboring blocks
  • the moving image encoding method includes a step of setting prediction selection information indicating whether to use a first geometric transformation parameter or a second geometric transformation parameter indicating information related to an image shape by geometric transformation of a pixel block; Motion information of one or more second adjacent blocks for which prediction signal generation processing has already been completed among a plurality of first adjacent blocks adjacent to one pixel block of the plurality of pixel blocks into which the signal is divided, or Obtaining a first geometric transformation parameter and the second geometric transformation parameter, and deriving a predicted geometric transformation parameter of the one pixel block from the geometric transformation parameters of the one or more second neighboring blocks. And a derived method of the geometric transformation parameter input from the outside and the predicted geometric transformation parameter by a predetermined method.
  • Prediction is performed by performing a geometric transformation process using the first geometric transformation parameter or the second geometric transformation parameter indicated in the selection information on a reference image signal when performing motion compensation on the one pixel block.
  • a step of generating a signal, a step of encoding the prediction selection information indicating whether to use the first geometric transformation parameter or the second geometric transformation parameter, and the first geometric transformation parameter are selected.
  • Derived to includes a step of encoding the derived value, a step of encoding the differential signal of the input image signal and the prediction signal.
  • a moving image decoding method for decoding moving image encoded data obtained by encoding an input image signal in units of a plurality of pixel blocks and performing decoding processing by a prescribed method includes a plurality of pixels obtained by dividing an input image signal.
  • the motion information of one or more second neighboring blocks that have already been decoded, or the shape of the image by the geometric transformation of the pixel block Obtaining a first geometric transformation parameter and a second geometric transformation parameter indicating information; and decoding selection information indicating whether to use the first geometric transformation parameter or the second geometric transformation parameter Decoding a derived value of the geometric transformation parameter if the first geometric transformation parameter is selected; and the one or more second neighbors Deriving a predicted geometric transformation parameter of the one pixel block from the geometric transformation parameter of the block, and deriving the predicted geometric transformation parameter from the decoded derived value of the geometric transformation parameter and the predicted geometric transformation parameter in a predetermined manner.
  • Deriving and setting one geometric transformation parameter Setting the second geometric transformation parameter based on motion information of the one pixel block and the one or more second neighboring blocks; A prediction signal obtained by performing a geometric transformation process using the first geometric transformation parameter or the second geometric transformation parameter indicated in the selection information on a reference image signal when performing motion compensation for the one pixel block. Generating.
  • the moving image encoding device, the moving image decoding device, the moving image encoding method, and the moving image decoding method according to the present invention are useful for highly efficient moving image encoding, and in particular, geometric transformation motion compensation. It is suitable for moving picture coding that reduces the motion detection process necessary for estimating the geometric transformation parameter used for prediction.

Landscapes

  • Engineering & Computer Science (AREA)
  • Multimedia (AREA)
  • Signal Processing (AREA)
  • Theoretical Computer Science (AREA)
  • Physics & Mathematics (AREA)
  • General Physics & Mathematics (AREA)
  • Computer Vision & Pattern Recognition (AREA)
  • Computing Systems (AREA)
  • Compression Or Coding Systems Of Tv Signals (AREA)

Abstract

La présente invention concerne un dispositif produisant un signal de prévision comprenant : une unité de définition servant à définir des informations de sélection de prévision indiquant les paramètres à utiliser parmi des premiers et seconds paramètres de transformation géométrique indiquant les informations relatives à la forme d’une image selon la transformation géométrique d’un bloc de pixels ; une unité d’acquisition servant à acquérir les informations de mouvement ou le paramètre de transformation géométrique d’un ou de plusieurs seconds blocs adjacents, qui ont déjà été soumis au traitement produisant un signal de prévision, du premier bloc adjacent qui est adjacent à un des blocs de pixels ayant des signaux d’image divisés ; une unité de déduction servant à déduire les paramètres déduits de la transformation géométrique d’un bloc de pixels à partir des paramètres de transformation géométrique du second bloc adjacent ; une unité de définition servant à déduire le premier paramètre de transformation géométrique par le biais d’un procédé prédéterminé à partir des valeurs déduites des paramètres de transformation géométrique et des paramètres de transformation géométrique prévus et servant à définir le paramètre déduit ; une unité de définition servant à définir les seconds paramètres de transformation géométrique sur la base des informations de mouvement d’un bloc de pixels et du second bloc adjacent ; et une unité de production servant à soumettre les signaux d’image de référence au moment de l’exécution des compensations de mouvement pour un bloc de pixels, à un traitement de transformation géométrique à l’aide des premiers et seconds paramètres de transformation géométrique indiqués dans les informations de sélection, afin de produire de cette façon un signal de prévision.
PCT/JP2009/063692 2009-07-31 2009-07-31 Dispositif produisant un signal de prévision à l’aide d’une prédiction de compensation de mouvement de transformation géométrique, dispositif de codage d’image variant dans le temps et dispositif de décodage d’image variant dans le temps WO2011013253A1 (fr)

Priority Applications (1)

Application Number Priority Date Filing Date Title
PCT/JP2009/063692 WO2011013253A1 (fr) 2009-07-31 2009-07-31 Dispositif produisant un signal de prévision à l’aide d’une prédiction de compensation de mouvement de transformation géométrique, dispositif de codage d’image variant dans le temps et dispositif de décodage d’image variant dans le temps

Applications Claiming Priority (1)

Application Number Priority Date Filing Date Title
PCT/JP2009/063692 WO2011013253A1 (fr) 2009-07-31 2009-07-31 Dispositif produisant un signal de prévision à l’aide d’une prédiction de compensation de mouvement de transformation géométrique, dispositif de codage d’image variant dans le temps et dispositif de décodage d’image variant dans le temps

Publications (1)

Publication Number Publication Date
WO2011013253A1 true WO2011013253A1 (fr) 2011-02-03

Family

ID=43528928

Family Applications (1)

Application Number Title Priority Date Filing Date
PCT/JP2009/063692 WO2011013253A1 (fr) 2009-07-31 2009-07-31 Dispositif produisant un signal de prévision à l’aide d’une prédiction de compensation de mouvement de transformation géométrique, dispositif de codage d’image variant dans le temps et dispositif de décodage d’image variant dans le temps

Country Status (1)

Country Link
WO (1) WO2011013253A1 (fr)

Cited By (14)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
GB2504069A (en) * 2012-07-12 2014-01-22 Canon Kk Intra-prediction using a parametric displacement transformation
JP2018509087A (ja) * 2015-02-16 2018-03-29 華為技術有限公司Huawei Technologies Co.,Ltd. ビデオ画像符号化方法、ビデオ画像復号方法、符号化デバイス、および復号デバイス
WO2018067823A1 (fr) * 2016-10-05 2018-04-12 Qualcomm Incorporated Prédiction de vecteur de mouvement destinée à des modèles de mouvement affine dans un codage vidéo
KR20180043830A (ko) * 2015-09-29 2018-04-30 후아웨이 테크놀러지 컴퍼니 리미티드 이미지 예측 방법 및 장치
JP2018529255A (ja) * 2015-08-29 2018-10-04 ホアウェイ・テクノロジーズ・カンパニー・リミテッド イメージ予測方法および装置
WO2019245228A1 (fr) * 2018-06-18 2019-12-26 엘지전자 주식회사 Procédé et dispositif de traitement de signal vidéo utilisant la prédiction affine de mouvement
US10560712B2 (en) 2016-05-16 2020-02-11 Qualcomm Incorporated Affine motion prediction for video coding
GB2577318A (en) * 2018-09-21 2020-03-25 Canon Kk Video coding and decoding
WO2020058954A1 (fr) * 2018-09-23 2020-03-26 Beijing Bytedance Network Technology Co., Ltd. Représentation d'un modèle affine
CN111066324A (zh) * 2017-08-03 2020-04-24 Lg 电子株式会社 利用仿射预测处理视频信号的方法和装置
WO2020108560A1 (fr) * 2018-11-30 2020-06-04 Mediatek Inc. Procédés et appareils de traitement vidéo pour déterminer des vecteurs de mouvement pour la mémorisation dans des systèmes de codage vidéo
US11778226B2 (en) 2018-10-22 2023-10-03 Beijing Bytedance Network Technology Co., Ltd Storage of motion information for affine mode
US11877001B2 (en) 2017-10-10 2024-01-16 Qualcomm Incorporated Affine prediction in video coding
WO2024066332A1 (fr) * 2022-09-27 2024-04-04 Beijing Xiaomi Mobile Software Co., Ltd. Codage/décodage de données d'image vidéo

Citations (4)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
JPH09266574A (ja) * 1996-01-22 1997-10-07 Matsushita Electric Ind Co Ltd 画像符号化装置および画像復号化装置
JP2000138935A (ja) * 1998-10-29 2000-05-16 Fujitsu Ltd 動きベクトル符号化装置および復号装置
JP2005244503A (ja) * 2004-02-25 2005-09-08 Sony Corp 画像情報符号化装置および画像情報符号化方法
JP2007312397A (ja) * 2007-05-25 2007-11-29 Nokia Corp 通信システムにおけるビデオフレーム転送方法と装置

Patent Citations (4)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
JPH09266574A (ja) * 1996-01-22 1997-10-07 Matsushita Electric Ind Co Ltd 画像符号化装置および画像復号化装置
JP2000138935A (ja) * 1998-10-29 2000-05-16 Fujitsu Ltd 動きベクトル符号化装置および復号装置
JP2005244503A (ja) * 2004-02-25 2005-09-08 Sony Corp 画像情報符号化装置および画像情報符号化方法
JP2007312397A (ja) * 2007-05-25 2007-11-29 Nokia Corp 通信システムにおけるビデオフレーム転送方法と装置

Cited By (43)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
GB2504069B (en) * 2012-07-12 2015-09-16 Canon Kk Method and device for predicting an image portion for encoding or decoding of an image
US9779516B2 (en) 2012-07-12 2017-10-03 Canon Kabushiki Kaisha Method and device for predicting an image portion for encoding or decoding of an image
GB2504069A (en) * 2012-07-12 2014-01-22 Canon Kk Intra-prediction using a parametric displacement transformation
JP2018509087A (ja) * 2015-02-16 2018-03-29 華為技術有限公司Huawei Technologies Co.,Ltd. ビデオ画像符号化方法、ビデオ画像復号方法、符号化デバイス、および復号デバイス
US10349079B2 (en) 2015-02-16 2019-07-09 Huawei Technologies Co., Ltd. Video image encoding method, video image decoding method, encoding device, and decoding device
US11979559B2 (en) 2015-08-29 2024-05-07 Huawei Technologies Co., Ltd. Image prediction method and device
US10880543B2 (en) 2015-08-29 2020-12-29 Huawei Technologies Co., Ltd. Image prediction method and device
US11368678B2 (en) 2015-08-29 2022-06-21 Huawei Technologies Co., Ltd. Image prediction method and device
JP2018529255A (ja) * 2015-08-29 2018-10-04 ホアウェイ・テクノロジーズ・カンパニー・リミテッド イメージ予測方法および装置
KR20200057120A (ko) * 2015-09-29 2020-05-25 후아웨이 테크놀러지 컴퍼니 리미티드 이미지 예측 방법 및 장치
US12058365B2 (en) 2015-09-29 2024-08-06 Huawei Technologies Co., Ltd. Image prediction method and apparatus
KR20180043830A (ko) * 2015-09-29 2018-04-30 후아웨이 테크놀러지 컴퍼니 리미티드 이미지 예측 방법 및 장치
US11323736B2 (en) 2015-09-29 2022-05-03 Huawei Technologies Co., Ltd. Image prediction method and apparatus
US10560712B2 (en) 2016-05-16 2020-02-11 Qualcomm Incorporated Affine motion prediction for video coding
US11503324B2 (en) 2016-05-16 2022-11-15 Qualcomm Incorporated Affine motion prediction for video coding
US10448010B2 (en) 2016-10-05 2019-10-15 Qualcomm Incorporated Motion vector prediction for affine motion models in video coding
RU2718225C1 (ru) * 2016-10-05 2020-03-31 Квэлкомм Инкорпорейтед Предсказание вектора движения для аффинных моделей движения в кодировании видео
US11082687B2 (en) 2016-10-05 2021-08-03 Qualcomm Incorporated Motion vector prediction for affine motion models in video coding
WO2018067823A1 (fr) * 2016-10-05 2018-04-12 Qualcomm Incorporated Prédiction de vecteur de mouvement destinée à des modèles de mouvement affine dans un codage vidéo
EP3758378A1 (fr) * 2016-10-05 2020-12-30 QUALCOMM Incorporated Prédiction de vecteur de mouvement destinée à des modèles de mouvement affine dans un codage vidéo
US11736707B2 (en) 2017-08-03 2023-08-22 Lg Electronics Inc. Method and apparatus for processing video signal using affine prediction
CN111066324B (zh) * 2017-08-03 2024-05-28 Oppo广东移动通信有限公司 利用仿射预测处理视频信号的方法和装置
CN111066324A (zh) * 2017-08-03 2020-04-24 Lg 电子株式会社 利用仿射预测处理视频信号的方法和装置
US11877001B2 (en) 2017-10-10 2024-01-16 Qualcomm Incorporated Affine prediction in video coding
CN112567749A (zh) * 2018-06-18 2021-03-26 Lg电子株式会社 使用仿射运动预测来处理视频信号的方法和装置
US12069295B2 (en) 2018-06-18 2024-08-20 Lg Electronics Inc. Method and device for processing video signal using affine motion prediction
US11140410B2 (en) 2018-06-18 2021-10-05 Lg Electronics Inc. Method and device for processing video signal using affine motion prediction
WO2019245228A1 (fr) * 2018-06-18 2019-12-26 엘지전자 주식회사 Procédé et dispositif de traitement de signal vidéo utilisant la prédiction affine de mouvement
CN112567749B (zh) * 2018-06-18 2024-03-26 Lg电子株式会社 使用仿射运动预测来处理视频信号的方法和装置
US11632567B2 (en) 2018-06-18 2023-04-18 Lg Electronics Inc. Method and device for processing video signal using affine motion prediction
GB2577318B (en) * 2018-09-21 2021-03-10 Canon Kk Video coding and decoding
GB2577318A (en) * 2018-09-21 2020-03-25 Canon Kk Video coding and decoding
US11870974B2 (en) 2018-09-23 2024-01-09 Beijing Bytedance Network Technology Co., Ltd Multiple-hypothesis affine mode
WO2020058954A1 (fr) * 2018-09-23 2020-03-26 Beijing Bytedance Network Technology Co., Ltd. Représentation d'un modèle affine
US11909953B2 (en) 2018-09-23 2024-02-20 Beijing Bytedance Network Technology Co., Ltd Representation of affine model
US11778226B2 (en) 2018-10-22 2023-10-03 Beijing Bytedance Network Technology Co., Ltd Storage of motion information for affine mode
US11785242B2 (en) 2018-11-30 2023-10-10 Hfi Innovation Inc. Video processing methods and apparatuses of determining motion vectors for storage in video coding systems
WO2020108560A1 (fr) * 2018-11-30 2020-06-04 Mediatek Inc. Procédés et appareils de traitement vidéo pour déterminer des vecteurs de mouvement pour la mémorisation dans des systèmes de codage vidéo
CN113170174B (zh) * 2018-11-30 2024-04-12 寰发股份有限公司 视频编码系统中用于决定储存用运动向量的视频处理方法和装置
CN113170174A (zh) * 2018-11-30 2021-07-23 联发科技股份有限公司 视频编码系统中用于决定储存用运动向量的视频处理方法和装置
US11290739B2 (en) 2018-11-30 2022-03-29 Mediatek Inc. Video processing methods and apparatuses of determining motion vectors for storage in video coding systems
TWI737055B (zh) * 2018-11-30 2021-08-21 聯發科技股份有限公司 視訊編碼系統中用於決定儲存用運動向量的視訊處理方法和裝置
WO2024066332A1 (fr) * 2022-09-27 2024-04-04 Beijing Xiaomi Mobile Software Co., Ltd. Codage/décodage de données d'image vidéo

Similar Documents

Publication Publication Date Title
WO2011013253A1 (fr) Dispositif produisant un signal de prévision à l’aide d’une prédiction de compensation de mouvement de transformation géométrique, dispositif de codage d’image variant dans le temps et dispositif de décodage d’image variant dans le temps
KR101984764B1 (ko) 영상 부호화/복호화 장치 및 방법
JP6615287B2 (ja) 画像復号装置
JP6667609B2 (ja) 画像符号化装置、画像符号化方法、画像復号装置および画像復号方法
KR101362757B1 (ko) 인터 컬러 보상을 이용한 영상의 부호화 방법 및 장치,복호화 방법 및 장치
JP7012809B2 (ja) 画像符号化装置、動画像復号装置、動画像符号化データ及び記録媒体
JP5061179B2 (ja) 照明変化補償動き予測符号化および復号化方法とその装置
KR101670532B1 (ko) 픽처 시퀀스를 나타내는 스트림을 디코딩하는 방법, 픽처 시퀀스를 코딩하는 방법, 및 코딩 데이터 스트럭처
JP4844449B2 (ja) 動画像符号化装置、方法、プログラム、動画像復号化装置、方法、およびプログラム
WO2012176381A1 (fr) Appareil de codage d'images en mouvement, appareil de décodage d'images en mouvement, procédé de codage d'images en mouvement et procédé de décodage d'images en mouvement
JP5310614B2 (ja) 動画像符号化装置、動画像符号化方法及び動画像復号装置ならびに動画像復号方法
WO2014163200A1 (fr) Appareil de codage d'image en couleur, appareil de décodage d'image en couleur, procédé de codage d'image en couleur, et procédé de décodage d'image en couleur
JP2010011075A (ja) 動画像符号化及び動画像復号化の方法及び装置
WO2010090335A1 (fr) Dispositif de codage de film cinématographique et dispositif de décodage de film cinématographique utilisant la prédiction de compensation de mouvements par transformation géométrique
JP2005086834A (ja) フレームシーケンスを符号化する方法、フレームシーケンスを復号する方法、その方法を実施する装置、その方法を実行するコンピュータプログラム、およびそのコンピュータプログラムを格納する記憶媒体
KR20150135457A (ko) 복수의 입력 화상을 인코딩하는 방법, 프로그램을 격납하는 기억 매체 및 장치
JP2010183162A (ja) 動画像符号化装置
WO2012176387A1 (fr) Dispositif de codage vidéo, dispositif de décodage vidéo, procédé de codage vidéo et procédé de décodage vidéo
JP2008193501A (ja) 画像符号化装置及び画像符号化方法
JP5533885B2 (ja) 動画像符号化装置および動画像復号装置
JP2023086397A (ja) イントラ予測装置、復号装置、及びプログラム
WO2013077305A1 (fr) Dispositif de décodage d'image, procédé de décodage d'image, dispositif de codage d'image
JP2009111733A (ja) 画像符号化方法、画像符号化装置、及び画像符号化プログラム
KR20150022953A (ko) 참조 프레임 생성 방법 및 장치와 그를 이용한 영상 부호화/복호화 방법 및 장치
KR20150022952A (ko) 참조 프레임 생성 방법 및 장치와 그를 이용한 영상 부호화/복호화 방법 및 장치

Legal Events

Date Code Title Description
121 Ep: the epo has been informed by wipo that ep was designated in this application

Ref document number: 09847844

Country of ref document: EP

Kind code of ref document: A1

DPE2 Request for preliminary examination filed before expiration of 19th month from priority date (pct application filed from 20040101)
NENP Non-entry into the national phase

Ref country code: DE

122 Ep: pct application non-entry in european phase

Ref document number: 09847844

Country of ref document: EP

Kind code of ref document: A1

NENP Non-entry into the national phase

Ref country code: JP