WO2011155171A1 - Dispositif de codage d'image animée, procédé de codage d'image animée et programme de codage d'image animée, et dispositif de décodage d'image animée, procédé de décodage d'image animée et programme de décodage d'image animée - Google Patents

Dispositif de codage d'image animée, procédé de codage d'image animée et programme de codage d'image animée, et dispositif de décodage d'image animée, procédé de décodage d'image animée et programme de décodage d'image animée Download PDF

Info

Publication number
WO2011155171A1
WO2011155171A1 PCT/JP2011/003170 JP2011003170W WO2011155171A1 WO 2011155171 A1 WO2011155171 A1 WO 2011155171A1 JP 2011003170 W JP2011003170 W JP 2011003170W WO 2011155171 A1 WO2011155171 A1 WO 2011155171A1
Authority
WO
WIPO (PCT)
Prior art keywords
block
reference image
motion vector
prediction
motion
Prior art date
Application number
PCT/JP2011/003170
Other languages
English (en)
Japanese (ja)
Inventor
上田 基晴
Original Assignee
株式会社Jvcケンウッド
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Priority claimed from JP2010132184A external-priority patent/JP2011259230A/ja
Priority claimed from JP2010132183A external-priority patent/JP2011259229A/ja
Application filed by 株式会社Jvcケンウッド filed Critical 株式会社Jvcケンウッド
Publication of WO2011155171A1 publication Critical patent/WO2011155171A1/fr

Links

Images

Classifications

    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04NPICTORIAL COMMUNICATION, e.g. TELEVISION
    • H04N19/00Methods or arrangements for coding, decoding, compressing or decompressing digital video signals
    • H04N19/60Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using transform coding
    • H04N19/61Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using transform coding in combination with predictive coding
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04NPICTORIAL COMMUNICATION, e.g. TELEVISION
    • H04N19/00Methods or arrangements for coding, decoding, compressing or decompressing digital video signals
    • H04N19/10Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using adaptive coding
    • H04N19/102Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using adaptive coding characterised by the element, parameter or selection affected or controlled by the adaptive coding
    • H04N19/103Selection of coding mode or of prediction mode
    • H04N19/105Selection of the reference unit for prediction within a chosen coding or prediction mode, e.g. adaptive choice of position and number of pixels used for prediction
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04NPICTORIAL COMMUNICATION, e.g. TELEVISION
    • H04N19/00Methods or arrangements for coding, decoding, compressing or decompressing digital video signals
    • H04N19/10Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using adaptive coding
    • H04N19/169Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using adaptive coding characterised by the coding unit, i.e. the structural portion or semantic portion of the video signal being the object or the subject of the adaptive coding
    • H04N19/17Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using adaptive coding characterised by the coding unit, i.e. the structural portion or semantic portion of the video signal being the object or the subject of the adaptive coding the unit being an image region, e.g. an object
    • H04N19/176Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using adaptive coding characterised by the coding unit, i.e. the structural portion or semantic portion of the video signal being the object or the subject of the adaptive coding the unit being an image region, e.g. an object the region being a block, e.g. a macroblock
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04NPICTORIAL COMMUNICATION, e.g. TELEVISION
    • H04N19/00Methods or arrangements for coding, decoding, compressing or decompressing digital video signals
    • H04N19/50Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using predictive coding
    • H04N19/503Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using predictive coding involving temporal prediction
    • H04N19/51Motion estimation or motion compensation
    • H04N19/513Processing of motion vectors

Definitions

  • the present invention relates to a moving picture signal encoding technique and decoding technique.
  • High-efficiency coding of moving images includes the correlation between pixels that are spatially adjacent in the same frame of a moving image signal, and the correlation between temporally adjacent frames and fields, as represented by MPEG4-AVC. A method of compressing information using it is used.
  • a local decoded image of an already encoded frame is used as a reference image for a target image that is a target frame to be encoded, and a two-dimensional block having a predetermined size is used.
  • a motion amount (hereinafter referred to as “motion vector”) between the target image and the reference image is detected in units (hereinafter referred to as “target block”), and a predicted image based on the target block and the motion vector is detected.
  • the generated motion compensated prediction is used.
  • the size of a target block in a 16 ⁇ 16 pixel two-dimensional block (hereinafter referred to as “macroblock”), which is a unit of encoding processing, is made variable, and a motion vector for each target block is obtained.
  • motion compensated prediction it is necessary to encode and transmit the generated motion vector, and in order to prevent an increase in the amount of information due to the motion vector, a predicted motion vector predicted from the motion vector for a decoded block around the target block By encoding using values, it is possible to use motion compensated prediction called direct mode in which no motion vector is transmitted.
  • both the encoding side and the decoding side detect a motion vector between reference images, and the motion vector is A method of generating a predicted motion vector of a target block and assuming a direct mode on the assumption that it is continuous in time is also presented.
  • the first problem is the degradation of the quality of the motion compensated predicted image due to the degradation of the quality of the decoded image used as the reference image, especially the degradation mixed in the motion compensated predicted image when high-compression encoding is performed. While the component deteriorates the prediction accuracy, it is necessary to encode information for restoring the deteriorated component as a prediction difference, and the amount of information is increasing.
  • the second problem is that the motion vector prediction is not accurate enough for image signals with little temporal and spatial motion continuity, and the predicted image quality when using the direct mode is effective. It is a point not to do. This degradation occurs when adjacent blocks have different motions across the target object, and the motion vector used for prediction when the motion is large in time has moved corresponding to the motion of the original target block This degradation occurs because a block of positions is assumed. Similarly, when the motion changes with time, the prediction is not successful and deterioration occurs.
  • a reference image used for motion compensation prediction can be switched in units of a target block from a plurality of reference images, and intra prediction that performs prediction within a screen without using motion compensation prediction is selected. Is also possible. For this reason, there are many cases where there is no adjacent motion vector for the corresponding reference image used for motion vector prediction in the adjacent block, and the motion vector prediction accuracy in the direct mode further decreases.
  • the third problem is an increase in the amount of code required for motion vector transmission when using prediction using two reference images or motion compensated prediction in fine block units.
  • prediction deterioration is smoothed by adding the reference images, and the influence of the deterioration component can be reduced.
  • To increase Also, in motion compensation in fine block units, it is possible to obtain appropriate motion according to the boundary of the object, and the accuracy of the predicted image is improved, but it is necessary to transmit motion vectors in fine units.
  • the amount of code increases.
  • Patent Document 1 is a technique presented to solve the second problem described above.
  • a motion vector obtained between reference images is a target block.
  • the motion vector prediction accuracy is improved because the motion passes through the position, but if the motion is not spatially uniform, it is the predicted motion vector obtained without using the target block information Therefore, the motion is different from that of the target block, and the prediction is not sufficient.
  • a motion vector detection process over a wide range between reference images is required for both the encoding device and the decoding device, which causes a problem that the amount of calculation increases.
  • the present invention has been made in view of such a situation, and an object of the present invention is to compensate for motion compensation by improving the quality of a predicted image while suppressing an increase in the amount of calculation in an encoding device and a decoding device in a mode in which no motion vector is transmitted. It is to provide a technique for improving the efficiency of prediction.
  • a moving picture coding apparatus uses a motion compensation prediction of a first reference picture used for a motion compensation process of a coding target block using a peripheral block of the coding target block. Based on the information used for the first motion vector, a motion vector value predicted from the selected first reference image is selected according to the frequency of the reference image used for motion compensation prediction in the neighboring blocks.
  • a reference reference image determination unit (107) to be acquired as a value, a first reference block extracted from the first reference image using the first motion vector value, and a predetermined region of at least one other reference image
  • a reference image synthesis unit (108 and 109) that generates a synthesized reference block, and the synthesized reference block is a predicted block, and the predicted block is the coding target block.
  • the reference reference image determination unit (107) determines a motion vector value for a reference image that is not used for motion compensated prediction in a peripheral block of an encoding target block, a motion vector value for a reference image used for motion compensated prediction, and a local part.
  • a motion vector value for a reference image that is not used for motion compensation prediction is determined by calculating using the decoded image signal and determining whether to add a reference image that is not used for motion compensation prediction to the first reference image selection candidate according to the calculated motion vector value.
  • the frequency of the reference image not used for the motion compensation prediction is increased by adding the calculated motion vector value to the motion vector candidates for the reference image not used for the motion compensation prediction.
  • the first reference image is selected above, and the motion vector value added to the candidate in the selected first reference image is used. It may generate the first motion vector value each (FIG. 12, S1202 ⁇ S1225).
  • a reference image to be used for the second reference block is selected, and a synthesized reference block obtained by synthesizing the first reference block and the second reference block is generated, and the synthesized reference block is set as a prediction block, or the first It may be determined whether to use the reference block as a prediction block (FIG. 13, S1303 to S1308).
  • the first reference image used for the motion compensation process of the encoding target block is used for the motion compensation prediction in the peripheral block based on the information used for the motion compensation prediction in the peripheral block of the encoding target block.
  • Selecting according to the frequency of the selected reference image obtaining a motion vector value predicted from the selected first reference image as a first motion vector value, and using the first motion vector value
  • Generating a synthesized reference block obtained by synthesizing the first reference block extracted from the first reference image and a predetermined region of at least one other reference image and using the synthesized reference block as a prediction block, the prediction Encoding a prediction difference block obtained by subtracting the block from the encoding target block.
  • Still another aspect of the present invention is a video decoding device.
  • This apparatus includes a mode detection unit (203) for detecting mode information indicating that the reference image designation information at the time of encoding and the motion vector designation information at the time of encoding are not transmitted, and the mode detection unit.
  • the mode detection unit When the mode information is detected, the first reference image and the first motion vector value used for the motion compensation process of the decoding target block are used for the motion compensation prediction in the peripheral blocks of the decoding target block.
  • a reference reference image determination unit (214) to select based on the first reference block extracted from the first reference image using the first motion vector value, and a predetermined region of at least one other reference image
  • a reference image synthesis unit (215 and 216) that generates a synthesized reference block, and the synthesized reference block as a predicted block.
  • the reference reference image determination unit (214) decodes a motion vector value for a reference image that is not used for motion compensated prediction in a peripheral block of the encoding target block, and a motion vector value for a reference image used for motion compensated prediction. It is calculated using the image signal, and it is determined whether or not to add a reference image that is not used for the motion compensation prediction to the first reference image selection candidate according to the calculated motion vector value. In this case, by adding the calculated motion vector value to a motion vector candidate for a reference image that is not used for the motion compensation prediction, the frequency of the reference image that is not used for the motion compensation prediction is increased.
  • the first reference image is selected by using the motion vector value added to the candidate in the selected first reference image. Serial may generate a first motion vector values (Fig. 12, S1202 ⁇ S1225).
  • a reference image to be used for the second reference block is selected, and a synthesized reference block obtained by synthesizing the first reference block and the second reference block is generated, and the synthesized reference block is set as a prediction block, or the first It may be determined whether to use the reference block as a prediction block (FIG. 13, S1303 to S1308).
  • Still another aspect of the present invention is a moving picture decoding method.
  • a decoded image is generated by adding the prediction block and a prediction difference block decoded from the decoding target block That and a step.
  • the present invention it is possible to reduce the motion vectors to be encoded and suppress the increase in the amount of calculation in the encoding device and the decoding device, while improving the quality of the predicted image and improving the efficiency of motion compensation prediction.
  • FIG. 5 is a diagram showing an example of a motion vector detection range between reference images in Embodiment 1 of the present invention. It is a block diagram which shows the structure of the moving image encoder of Embodiment 2 of this invention. It is a block diagram which shows the structure of the moving image decoding apparatus of Embodiment 2 of this invention.
  • Embodiment 2 of this invention It is a flowchart explaining operation
  • Embodiment 2 of this invention it is a figure which shows an example of the selection result of the bidirectional
  • FIG. 1 is a block diagram showing a configuration of the moving picture encoding apparatus according to the first embodiment.
  • the moving picture coding apparatus includes an input terminal 100, an input picture buffer 101, a block division unit 102, an intra-frame prediction unit 103, a motion vector detection unit 104, and a motion compensation prediction unit 105.
  • Motion vector prediction unit 106 Motion vector prediction unit 106, standard reference image determination unit 107, inter-reference image motion vector detection unit 108, direct motion compensation prediction unit 109, prediction mode determination unit 110, subtractor 111, orthogonal transform unit 112, quantization unit 113, An inverse quantization unit 114, an inverse orthogonal transform unit 115, an adder 116, an intra-frame decoded image memory 117, a decoded reference image memory 118, an entropy encoding unit 119, a stream buffer 120, an output terminal 121, and a code amount control unit 122 are included.
  • the inter-reference image motion vector detection unit 108 and the direct motion compensation prediction unit 109 are examples of the “reference image synthesis unit” of the video encoding device of the present invention.
  • the same processing as that constituting a coding process in a moving picture coding apparatus such as MPEG4-AVC can be applied.
  • the digital image signal input from the input terminal 100 is stored in the input image buffer 101.
  • the digital image signal stored in the input image buffer 101 is supplied to the block dividing unit 102, and is cut out as an encoding target block in units of macroblocks composed of 16 ⁇ 16 pixels.
  • the block dividing unit 102 supplies the extracted encoding target block to the intra-frame prediction unit 103, the motion vector detection unit 104, the motion compensation prediction unit 105, the direct motion compensation prediction unit 109, and the subtractor 111.
  • the block to be encoded input from the block division unit 102 and the decoded image of the frame that has been encoded on the entire screen and stored in the decoded reference image memory 118 are input as the reference image. Then, motion estimation is performed between the encoding target block and the reference image. As a general motion estimation process, a reference image at a position moved by a predetermined movement amount from the same position on the screen is cut out, and a movement amount that minimizes a prediction error when the image is used as a prediction block is determined as a motion vector. As a value, a block matching process that is obtained while changing the movement amount is used. The detected motion vector value is output to the motion compensation prediction unit 105.
  • the motion compensation prediction unit 105 receives the motion vector value obtained by the motion vector detection unit 104, and receives motion compensated prediction images for a plurality of block sizes of 16 ⁇ 16 or less and a plurality of reference images from the decoded reference image memory 118.
  • the prediction signal with the least difference information to be encoded is selected and the prediction signal with the least difference information to be encoded is selected for the encoding target block acquired and input from the block dividing unit 102.
  • the motion compensation prediction unit 105 outputs the selected motion compensation prediction mode and the prediction signal to the prediction mode determination unit 110.
  • the motion vector prediction unit 106 calculates a predicted motion vector value using the motion vectors of the surrounding encoded blocks, and supplies the motion vector detection unit 104 and the motion compensated prediction unit 105.
  • the motion vector detection unit 104 takes into account the amount of code necessary for encoding the difference between the motion vector predicted value and the motion vector value, and determines the optimal motion vector value. To detect.
  • the motion compensation prediction unit 105 takes into account the amount of code required when encoding the difference between the motion vector prediction value and the motion vector value, and the reference image used as the block unit for the optimal motion compensation prediction and Select a motion vector value.
  • the motion vector prediction unit 106 further stores the stored motion vectors of the surrounding encoded blocks, the reference image that is the target of the motion vector, and the information indicating the block size of the motion compensation, the standard reference image determination unit 107. Output to.
  • the reference reference image determination unit 107 is a motion vector that is motion compensated prediction information of a surrounding encoded block input from the motion vector prediction unit 106, and a reference image or motion compensation block that is a target of the motion vector. Based on the size, it has a function of determining a reference image that can predict a motion vector for an encoding target block and calculating a predicted motion vector value for the determined reference image. Information indicating the reference image corresponding to the calculated predicted motion vector value is output to the inter-reference image motion vector detection unit 108. The detailed operation of the reference reference image determination unit 107 will be described later.
  • the inter-reference image motion vector detection unit 108 uses the decoded reference image memory 118 for the motion compensated prediction based on the information indicating the reference image corresponding to the predicted motion vector value input from the standard reference image determination unit 107. Extract reference blocks.
  • the inter-reference image motion vector detection unit 108 calculates an error value by block matching or the like for the motion vector between the extracted first reference block and another reference image, and refers to a motion vector having a small value. Calculated as an inter-image motion vector.
  • the inter-reference image motion vector detection unit 108 outputs the second reference block extracted from the other reference image based on the calculated inter-reference image motion vector to the direct motion compensation prediction unit 109 together with the first reference block.
  • the inter-reference image motion vector detection unit 108 calculates a motion vector between another reference image and the encoding target block, which is calculated from the predicted motion vector value for the reference reference image, the predicted motion vector value, and the inter-reference image motion vector. The value is also output to the direct motion compensation prediction unit 109.
  • the detailed operation of the inter-reference image motion vector detection unit 108 will be described later.
  • the direct motion compensation prediction unit 109 performs motion averaging for the direct mode in which no motion vector is transmitted by averaging the first reference block and the second reference block input from the inter-reference image motion vector detection unit 108 for each pixel.
  • a compensated prediction block is generated and output to the prediction mode determination unit 110.
  • a prediction motion vector value for the reference reference image and a motion vector value between another reference image and the encoding target block are also predicted by the prediction mode determination unit. To 110.
  • the prediction mode determination unit 110 uses the prediction mode and the prediction image for each prediction method input from the intra-frame prediction unit 103, the motion compensation prediction unit 105, and the direct motion compensation prediction unit 109, and the encoding input from the block division unit 102. For the target block, a prediction signal with the least amount of difference information to be encoded is selected, a prediction image block for the selected prediction method is output to the subtractor 111 and the adder 116, and the entropy encoding unit 119 is Thus, prediction mode information as additional information and information that requires encoding according to the prediction mode are output.
  • the prediction mode determination unit 110 outputs, to the motion vector prediction unit 106, the motion vector used in the selected prediction method, the reference image that is the target of the motion vector, and the motion compensation block size and the like.
  • the output information is used to generate a predicted motion vector value in the subsequent encoding target block.
  • the subtractor 111 calculates the difference between the encoding target block supplied from the block dividing unit 102 and the predicted image block supplied from the prediction mode determining unit 110, and supplies the result to the orthogonal transform unit 112 as a difference block. .
  • the orthogonal transform unit 112 generates DCT coefficients corresponding to the orthogonally transformed frequency component signal by performing DCT transform on the difference block in units of 4 ⁇ 4 pixels or 8 ⁇ 8 pixels. Further, the orthogonal transform unit 112 collects the generated DCT coefficients in units of macroblocks and outputs them to the quantization unit 113.
  • the quantization unit 113 performs quantization by dividing the DCT coefficient by a different value for each frequency component.
  • the quantization unit 113 supplies the quantized DCT coefficient to the inverse quantization unit 114 and the entropy encoding unit 119.
  • the inverse quantization unit 114 performs inverse quantization by multiplying the quantized DCT coefficient input from the quantization unit 113 by a value divided at the time of quantization, and the result of the inverse quantization is obtained.
  • the decoded DCT coefficient is output to the inverse orthogonal transform unit 115.
  • the inverse orthogonal transform unit 115 performs inverse DCT processing to generate a decoded difference block.
  • the inverse orthogonal transform unit 115 supplies the decoded difference block to the adder 116.
  • the adder 116 adds the prediction image block supplied from the prediction mode determination unit 110 and the decoded difference block supplied from the inverse orthogonal transform unit 115 to generate a local decoding block.
  • the local decoded block generated by the adder 116 is stored in the intra-frame decoded image memory 117 and the decoded reference image memory 118 in a form subjected to inverse block conversion.
  • adaptive filtering is applied to block boundaries where coding distortion of each block tends to appear as a boundary. In some cases, processing to be performed is performed.
  • the entropy encoding unit 119 converts the quantized DCT coefficient supplied from the quantization unit 113, the prediction mode information supplied from the prediction mode determination unit 110, and information that requires encoding according to the prediction mode.
  • variable length encoding of each information is performed. Specifically, intra prediction mode and prediction block size information in the case of intra-frame prediction, and prediction block size, reference image designation information, and motion vector in the case of motion compensation prediction and synthesized image motion compensation prediction, The difference value from the predicted motion vector value is information that requires encoding.
  • Information subjected to variable length coding is output to the stream buffer 120 from the entropy coding unit 119 as a coded bit stream.
  • the encoded bit stream stored in the stream buffer 120 is output to a recording medium or a transmission path via the output terminal 121.
  • the code amount control unit 122 is supplied with the code amount of the encoded bit stream stored in the stream buffer 120 and compared with the target code amount. In order to approach the target code amount, the quantization level (quantization scale) of the quantization unit 113 is controlled.
  • FIG. 2 is a configuration diagram of the moving picture decoding apparatus according to the first embodiment.
  • the moving picture decoding apparatus includes an input terminal 200, a stream buffer 201, an entropy decoding unit 202, a prediction mode decoding unit 203, a prediction image selection unit 204, an inverse quantization unit 205, and an inverse.
  • the inter-reference image motion vector detection unit 215 and the direct motion compensation prediction unit 216 are examples of the “reference image synthesis unit” of the video decoding device of the present invention.
  • the operation is the same as that of the same functional block of the moving picture coding apparatus shown in FIG. 1, thereby generating a motion compensated prediction block for the direct mode without transmission of additional information.
  • the same processing as the processing blocks constituting the decoding processing in the moving image decoding apparatus such as MPEG4-AVC can be applied.
  • the encoded bit stream input from the input terminal 200 is supplied to the stream buffer 201, and the stream buffer 201 absorbs the code amount variation of the encoded bit stream and is supplied to the entropy decoding unit 202 in a predetermined unit such as a frame.
  • the entropy decoding unit 202 performs variable-length decoding on the encoded prediction mode information, the additional information corresponding to the prediction mode, and the quantized DCT coefficient from the encoded bitstream input via the stream buffer 201. Then, the quantized DCT coefficient is output to the inverse quantization unit 205, and the prediction mode information and additional information corresponding to the prediction mode are output to the prediction mode decoding unit 203.
  • the local decoding process of the moving image coding apparatus According to the first embodiment of the present invention. Processing similar to that of a certain inverse quantization unit 114, inverse orthogonal transform unit 115, adder 116, intra-frame decoded image memory 117, and decoded reference image memory 118 is performed.
  • the decoded image stored in the decoded reference image memory 209 is displayed as a decoded image signal on the display device via the output terminal 210.
  • the motion vector prediction decoding part 212 when motion compensation prediction is selected as a prediction mode from the prediction mode information input from the entropy decoding part 202 and the additional information corresponding to the prediction mode, the motion vector prediction decoding part 212
  • the prediction mode information indicating the predicted block unit or the prediction mode information indicating the direct mode and the decoded difference vector value in the case of the motion compensation prediction mode are output and the prediction image selection unit 204 is also predicted.
  • the prediction mode decoding unit 203 provides information indicating that the prediction mode decoding unit 203 has selected to the intra-frame prediction unit 211, the motion compensation prediction unit 213, and the direct motion compensation prediction unit 216 according to the decoded prediction mode information. Outputs additional information according to the mode.
  • the prediction image selection unit 204 is output from any of the intra-frame prediction unit 211, the motion compensation prediction unit 213, and the direct motion compensation prediction unit 216 according to the prediction mode information input from the prediction mode decoding unit 203.
  • a predicted image for the decoding target block is selected and output to the adder 207.
  • the intra-frame prediction unit 211 receives the intra prediction mode as additional information according to the prediction mode from the prediction mode decoding unit 203, and according to the intra prediction mode.
  • the decoded image of the region where decoding is completed is input to the periphery of the decoding target block stored in the intra-frame decoded image memory 208, and prediction using the intra-frame correlation is performed in the same intra prediction mode as the encoding device. Done.
  • the intra-frame prediction unit 211 outputs the intra-frame prediction image generated by the prediction to the prediction image selection unit 204.
  • the motion vector predictive decoding unit 212 uses the motion vector of the neighboring decoded block for the decoded difference vector value input from the prediction mode decoding unit 203, and performs the motion prediction using the same method as that performed by the encoding device. A vector value is calculated, and a value obtained by adding the difference vector value and the predicted motion vector value is output to the motion compensated prediction unit 213 as a motion vector value of the decoding target block.
  • the motion vector is decoded by the number encoded according to the block unit of the prediction process indicated in the motion compensation prediction mode or the composite image motion compensation prediction mode.
  • the motion vector predictive decoding unit 212 further stores the stored motion vectors of the surrounding encoded blocks, the reference image that is the target of the motion vector, and information indicating the block size of motion compensation, as a reference reference image determination unit. To 214.
  • the motion compensation prediction unit 213 generates a motion compensated prediction image from the decoded reference image memory 209 from the motion vector value input from the motion vector prediction decoding unit 212, and sends the generated motion compensation prediction image to the prediction image selection unit 204. Output.
  • the reference reference image determination unit 214 is the motion compensation prediction information of the surrounding encoded blocks input from the motion vector prediction decoding unit 212, the motion vector, the reference image that is the target of the motion vector, and motion compensation. Based on the block size, it has a function of determining a reference image that can predict a motion vector for a decoding target block and calculating a predicted motion vector value for the determined reference image. Information indicating the reference image corresponding to the calculated predicted motion vector value is output to the inter-reference image motion vector detection unit 215.
  • the inter-reference image motion vector detection unit 215 uses the decoded reference image memory 209 for the first motion compensation prediction based on information indicating the reference image corresponding to the predicted motion vector value input from the standard reference image determination unit 214. Extract reference blocks.
  • the inter-reference image motion vector detection unit 215 calculates an error value by block matching or the like for the motion vector between the extracted first reference block and another reference image, and refers to a motion vector having a small value. Calculated as an inter-image motion vector.
  • the inter-reference image motion vector detection unit 215 outputs the second reference block extracted from the other reference images based on the calculated inter-reference image motion vector together with the first reference block to the direct motion compensation prediction unit 216.
  • the inter-reference image motion vector detecting unit 215 calculates a motion vector between another reference image and the encoding target block, which is calculated from the predicted motion vector value for the reference reference image, the predicted motion vector value, and the inter-reference image motion vector.
  • the value is also output to the direct motion compensation prediction unit 216.
  • Detailed operations of the reference reference image determination unit 214 and the inter-reference image motion vector detection unit 215 are paired with the reference reference image determination unit 107 and the inter-reference image motion vector detection unit 108 in the moving image encoding apparatus of the first embodiment. The detailed operation will be described later.
  • the direct motion compensation prediction unit 216 averages the first reference block and the second reference block input from the inter-reference image motion vector detection unit 215 for each pixel, thereby performing motion for the direct mode in which no motion vector is transmitted.
  • a compensated prediction block is generated and output to the prediction image selection unit 204, and a motion vector value between a prediction motion vector value for the reference reference image and another reference image and the current block is encoded to the motion vector prediction decoding unit 212.
  • Output The output information is used to generate a predicted motion vector value in the subsequent decoding target block.
  • FIG. 3 (c) is a conceptual diagram showing direct motion compensation prediction.
  • 3 (a) and 3 (b) are conceptual diagrams of motion compensation prediction using a plurality of reference images used in MPEG4-AVC.
  • FIG. 3 (a) detects a motion vector between two reference pictures called bi-directional prediction and a coding target block, transmits a motion vector for each reference picture,
  • the average value of reference blocks indicated by motion vectors is used as a predicted image.
  • a prediction image having a function of removing an encoding degradation component as a motion adaptive filter in the time direction by synthesizing two reference images and a function of following a minute luminance change component of an encoding object by averaging. Can be generated.
  • a plurality of decoded reference images are stored, and a reference image number and a motion vector used for prediction are transmitted in a predetermined block unit to adaptively select the reference image.
  • a reference image number and a motion vector used for prediction are transmitted in a predetermined block unit to adaptively select the reference image.
  • four decoded images are secured as reference images, and two prediction images are acquired using the reference image 1 and the reference image 3, and bidirectional prediction is performed. ing.
  • mvL0 and mvL1 take the difference value from the predicted motion vector and are transmitted.
  • FIG. 3 (b) is a technique for performing prediction using two reference images without transmitting a motion vector, which is called a temporal direct mode.
  • an anchor block which is a temporally closest reference image (referred to as reference image 3 here) reproduced after the encoding target image, is located at the same position as the encoding target block.
  • reference image 3 a temporally closest reference image reproduced after the encoding target image.
  • a block is generated by motion compensated prediction from the reference image 2
  • it is encoded from a motion vector mvCol between the reference image 2 and the reference image 3 on the assumption that the motion is temporally continuous.
  • a motion vector value mvL0 between the target block and the reference image 2 and a motion vector value mvL1 between the encoding target block and the reference image 3 are generated, and bidirectional prediction is performed using the motion vector.
  • a prediction image obtained by synthesizing two reference images can be generated without transmitting the motion vectors mvL0 and mvL1, but when mvCol is large as shown in FIG. 3b), it is expressed by mvCol. Therefore, when there is no spatial motion continuity, appropriate mvL0 and mvL1 cannot be calculated.
  • the motion vectors mvL0 and mvL1 are implicitly generated only when the motion is temporally continuous, and an appropriate motion vector can be selected even when the motion vector value is less continuous in time. It cannot be generated, and the temporal direct mode does not function effectively unless it is limited in terms of spatial and temporal movement.
  • both the encoding side and the decoding side detect motion in a block existing at a symmetrical position around the encoding target block between reference images.
  • This is a method of generating a motion vector having temporal continuity passing through the position of the encoding target block by recreating mvCol, but the temporal continuity is limited in that the motion is uniform as in the temporal direct mode.
  • mvL0 and mvL1 are generated and the temporal continuity is small, it does not function effectively like the temporal direct mode.
  • the prediction configuration of the direct motion compensated prediction determines a reference reference image (reference image 3 in FIG. 3C) from a plurality of existing reference images, and sets the reference reference image.
  • the predicted motion vector value between the current block and the current block is calculated as mvL0.
  • motion vector detection between reference images for another reference image is performed on the encoding side and decoding.
  • the motion vector value mvL1 between the encoding target block and the reference image 1 is calculated from the inter-reference image motion vector mvInterRef, and bi-directional prediction using two reference images using the motion vectors mvL0 and mvL1 Take the configuration to do.
  • a reference image that can predict a motion vector most appropriately is selected as a reference image using peripheral decoded motion compensation prediction information.
  • the point of tracking and implicitly calculating the other motion vector value is the point of the direct motion compensated prediction method in the moving picture coding apparatus and the moving picture decoding apparatus according to the embodiment of the present invention. Even for video signals with little temporal continuity, more appropriate bi-directional prediction can be realized without the transmission of motion vectors, and coding efficiency can be greatly improved. .
  • FIG. 4A and FIG. 4B are flowcharts showing the direct motion compensated prediction processing procedure in the encoding device and decoding device of Embodiment 1, and the flow of prediction processing will be described.
  • 4A is the encoding side
  • FIG. 4B is the decoding side.
  • a standard reference image for an encoding target block is determined (S400). Subsequently, a predicted motion vector value for the standard reference image is calculated (S401). Subsequently, the motion vector value mvL0 is used as the predicted motion vector value, a reference reference image at a position moved by mvL0 from the position in the screen of the encoding target block is acquired, and a first predicted block is generated (S402).
  • a motion vector is detected between the first prediction block and another reference image, and an inter-reference image motion vector value mvInterRef is calculated (S403).
  • the method of explicitly specifying in a predetermined unit or the reference image closest to the encoding target block that is in the temporal position across the encoding target block is acquired. It is possible to use the method to do.
  • a reference ID information indicating the other reference image (hereinafter referred to as a reference ID) with respect to the standard reference image is transmitted as screen unit information in the same unit as the slice header in MPEG4-AVC.
  • mvInterRef From the inter-reference image motion vector value mvInterRef, another reference image moved by mvInterRef from the in-screen position of the first prediction block is acquired, and a second prediction block is generated (S404).
  • the motion vector value has two components, horizontal and vertical, and each value is calculated by the above formula.
  • first prediction block and the second prediction block are averaged for each pixel, and the prediction block subjected to bi-directional prediction is output as a direct motion compensation prediction block together with the motion vector values mvL0 and mvL1 (S406). ).
  • a predicted motion vector value for the reference reference image is calculated (S412), and a first predicted block is generated using the predicted motion vector value as the motion vector value mvL0 (S413).
  • a motion vector is detected with another reference image, a motion vector value mvInterRef between reference images is calculated (S414), and the motion vector value mvInterRef between reference images is 2 prediction blocks are generated (S415), and mvL1 is calculated by adding mvL0 and mvInterRef (S416).
  • first prediction block and the second prediction block are added and averaged for each pixel, and the prediction block subjected to bidirectional prediction is output as a direct motion compensated prediction block together with the motion vector values mvL0 and mvL1 (S417). ).
  • the decoded image signal is used as the reference image for these operations, the same operation can be performed on the encoding side and the decoding side, and the decoding is performed without transmitting the motion vector values mvL0, mvInterRef, and mvL1. Can be reproduced on the side.
  • the motion compensated prediction information adjacent to the encoding target block is stored in a form in which the motion compensated prediction block size is combined (S501). Specific behavior will be described with reference to FIG. FIG. 6A shows the relationship between the encoding target block and the adjacent decoded block. If the prediction block size of direct motion compensated prediction is 16 ⁇ 16, the encoding process is performed from left to right and from top to bottom in units of 16 ⁇ 16 macroblocks, so left (A), top ( The motion compensated prediction information for the four macroblocks B), upper right (C), and upper left (D) is adjacent block information.
  • Each macroblock is motion-compensated with various block sizes. As shown in FIG. 6B, a plurality of motion compensations are performed in a macroblock that is motion-compensated and predicted in units smaller than the macroblock. Predictive information exists.
  • a configuration is adopted in which the motion compensated prediction information is weighted and evaluated by the block size, and each block position adjacent to the encoding target for the minimum motion compensated prediction block is taken.
  • the motion compensation prediction information used for motion compensation of the portion is stored.
  • motion compensation prediction of MPEG4-AVC a minimum 4 ⁇ 4 pixel unit is used when the adaptive motion compensation block size in the sub macroblock is used, and a minimum 8 ⁇ 8 pixel unit is used when the adaptive motion compensation block size is not used.
  • 8 ⁇ 8 pixels are the minimum motion compensation prediction block size, but the same processing can be realized in the case of 4 ⁇ 4 pixels.
  • FIG. 6C shows the motion compensation prediction information to be stored when the motion compensation prediction of the neighboring blocks as shown in FIG.
  • Motion compensated prediction includes unidirectional prediction performed using only one reference image and bidirectional prediction performed using two reference images.
  • unidirectional prediction one motion is used as motion compensated prediction information at the same position.
  • a vector value and one reference ID are stored.
  • bidirectional prediction two motion vector values and two reference IDs are stored.
  • the encoding target block has a configuration of 16 ⁇ 8 blocks on the left, 8 ⁇ 16 blocks on the top, 8 ⁇ 8 blocks on the top right, and 16 ⁇ 16 blocks on the top left.
  • the upper macroblock is unidirectionally predicted only by the first reference, and the upper left macroblock is unidirectionally predicted only by the second reference.
  • mvA1, mvA2 and the respective reference IDs for the first reference are stored as the motion compensation prediction information of the left macroblock
  • the mvA1, mvA2 and the respective reference IDs for the second reference are stored.
  • MvB1, mvB2 and their respective reference IDs for the first reference, mvC3 and reference ID for the first reference as motion compensation prediction information for the upper right macroblock, mvC3 and reference ID for the second reference, and motion compensation prediction information for the upper left macroblock A total of nine pieces of motion compensation prediction information of mvD and reference ID for the second reference are stored.
  • the reference IDs of the stored motion compensation prediction information are tabulated for each reference image to calculate the most reference images (S502). For example, when motion compensation prediction of 16 ⁇ 16 pixels is selected in the left or upper macroblock, the same motion compensation prediction is stored when the motion compensation prediction block is divided into the smallest motion compensation prediction blocks. The information is weighted according to the size. Further, the first reference and the second reference are aggregated without distinction, and the number of times of application for each reference image finally indicated by the motion vector is counted.
  • the reference image when there is one reference image (S503: YES), the reference image is selected as a standard reference image (S504). If there are a plurality of reference images (S503: NO), the reference image is preferentially selected as a reference image closer to the encoding target block, and if the distance is equal, the code is preceded in time. It is determined that the converted reference image has a smaller number of times of predictive encoding (S505).
  • the quantization value when the reference image is encoded is acquired on the encoding side / decoding side, and the reference image is subjected to finer quantization.
  • the reference image that is most frequently referred to as motion compensated prediction in the peripheral region is judged to have high reliability as the reference image quality and the temporal correlation with the image in which the block to be encoded exists are high.
  • it is a point of the standard reference image determination process a large amount of information that can be used as a motion vector prediction candidate can be acquired, and the motion vector prediction accuracy is increased.
  • a process of calculating a predicted motion vector value for the selected standard reference image is performed.
  • the horizontal and vertical median values of the motion vector values in the motion compensated prediction are calculated ( S509).
  • the method for calculating the median is the same as the method for generating a predicted vector value in MPEG4-AVC, but is suitable for prediction because the motion vector value is weighted to improve the priority of a motion vector having a large area.
  • prediction with higher accuracy than that of a conventional motion vector can be performed.
  • the median value is acquired (S509) after excluding the motion vector value farthest from the average of the two central motion vector values (S508).
  • the motion compensation prediction information whose reference ID is the reference reference image is two or less (S506: NO)
  • the average value of the motion vector values is calculated (S511) in two cases (S510: YES).
  • the motion vector value is set as a predicted motion vector value (S512).
  • the calculated reference ID and predicted motion vector value are output to the inter-reference image motion vector detection units 108 and 215 (S513), and the standard reference image determination process ends.
  • inter-reference image motion vector detecting unit shown in FIG. 7 will be described with reference to the inter-reference image motion vector detecting process, which operates according to the inter-reference image motion vector detecting units 108 and 215 and S403 and S414 in the processing flowchart of FIG. This is performed using the internal configuration diagram and the processing flowchart shown in FIG.
  • the inter-reference image motion vector detection unit 108 includes a standard reference image acquisition unit 700, a motion vector detection range setting unit 701, a standard reference image memory 702, a reference image acquisition unit 703, and a block matching evaluation unit 704.
  • the standard reference image acquisition unit 700 encodes the standard reference image encoding target from the decoded reference image memory 118.
  • the image block that has moved mvL0 from the block is cut out, and the first reference block is acquired (S800).
  • the acquired reference block is stored in the standard reference image memory 702.
  • the motion vector detection range setting unit 701 determines the reference ID of the second reference image based on the reference ID of the standard reference image and the predicted motion vector value mvL0 input from the standard reference image determination unit 107 (S801). ).
  • the other reference ID for the standard reference image is transmitted in units similar to the slice header in MPEG4-AVC as information in units of screens, and the corresponding information is referred to.
  • the reference ID of the second reference image can be determined.
  • a detection range of a motion vector between reference images to be detected between the standard reference image and the second reference image is set (S802).
  • the detection range it is also possible to take the entire area of the second reference image as the motion vector detection range for the first reference block, and by performing detection processing with the same definition as the encoding device and the decoding device.
  • a detection range as shown in FIG. 9 is set.
  • FIG. 9 is an example of a motion vector detection range between reference images in the first embodiment.
  • the input time of the encoding target image is Poc_Cur
  • the input time of the standard reference image is Poc_Ref1
  • the input time of the second reference image is Poc_Ref2
  • the motion vector prediction value between the encoding target block and the second reference image when it is assumed that the motion is temporally continuous is set.
  • an appropriate reference block of the second reference image can be obtained by searching for a motion vector for a specific region with the search position as the center. To be able to get In the example shown in FIG. 9, an area of ⁇ 8 pixels is designated as the specific area.
  • the reference image acquisition unit 703 acquires the reference block of the second reference image in the motion vector detection range specified by the motion vector detection range setting unit 701 from the decoded reference image memory 118 (S803), and the block matching evaluation unit 704. Output to.
  • the block matching evaluation unit 704 calculates an error sum for each pixel between the first reference block stored in the standard reference image memory 702 and the reference block of the second reference image input from the reference image acquisition unit 703. Then, a reference block having a smaller total sum and a motion vector value when the reference block is acquired are stored (S804).
  • the stored reference block is set as the second reference block, and the motion vector value is set as the inter-reference image motion vector.
  • the value mvInterRef is set (S806).
  • the detection accuracy of motion vectors between reference images it is possible to apply a method of implicitly detecting motion vectors with the same detection accuracy in the encoding device and the decoding device, but detection of motion vectors for each frame or reference image used. It is also possible to use a technique of transmitting accuracy as encoded information.
  • the detection accuracy is 1/4 pixel accuracy.
  • the calculated inter-reference image motion vector value mvInterRef is output to the direct motion compensation prediction unit 109 together with mvL0, the first reference block, and the second reference block (S807).
  • a function of selecting a reference image suitable for prediction based on the motion compensated prediction information that has been encoded in the peripheral blocks of the encoding target block By having both the encoding side and the decoding side, the accuracy of the motion vector that is implicitly generated and the reference image to be used can be made more appropriate, and the prediction block in which the obtained reference image is extracted as the standard reference image And a function to perform bi-directional prediction using the detected motion vector on both the encoding side and the decoding side, so that additional information such as a motion vector can be obtained. Without transmission, it is possible to improve the prediction efficiency for a moving image signal accompanied by a temporal change in motion, and to improve the coding efficiency.
  • Embodiment 2 Next, a video encoding device and a video decoding device according to Embodiment 2 will be described. Also in the second embodiment, the configurations of the moving image encoding device and the moving image decoding device are the same as those of the first embodiment, but as shown in FIGS. 10 and 11, the reference reference image determination units 107 and 214 are used. Is different in that the reference image signal is input from the decoded reference image memories 118 and 209 and the adjacent intra-frame decoded signal from the intra-frame decoded image memories 117 and 208 is used for the reference image determination process.
  • FIG. 12 shows a process flowchart of the reference image determination process in the second embodiment
  • FIG. 13 shows a reference image in the second embodiment.
  • a process flowchart of the inter-motion vector detection process is shown, and the operation of the second embodiment will be described. Note that, as in the first embodiment, these functions function by performing similar processing on the encoding side and decoding side, and therefore will be described as a common configuration.
  • the motion vector for the reference image that is not used with respect to the motion compensated prediction information used in the peripheral blocks of the encoding target block.
  • the second embodiment is based on the error evaluation value between reference images when detecting a motion vector between the first reference block acquired from the standard reference image and a plurality of other reference images. It has a function of determining which reference image is used for bidirectional prediction and determining whether bidirectional prediction is performed in units smaller than the size of the reference block.
  • the motion compensated prediction motion first used in the peripheral block of the encoding target block (decoding target block in the case of a decoding device) is the same as in the first embodiment.
  • Vector values, reference IDs, and motion compensation block size information are acquired (S1200), and motion compensation prediction information adjacent to the block to be encoded is stored in a form in which the motion compensation prediction block size is combined (S1201).
  • the motion vector value of the adjacent block is set as a candidate vector value (S1205).
  • the target block uses a reference image of another reference ID (S1206: YES)
  • the motion vector value in the reference ID closest to the reference image that is the target in the reference ID is defined as the encoding target block.
  • the value reduced or enlarged according to the distance ratio is set as a candidate vector value (S1207).
  • the reference image and the intra-frame decoded image are obtained from the reference image memory and the intra-frame decoded image memory, and the motion compensation is performed on the peripheral blocks using the candidate vector.
  • An error value at the time of prediction is calculated (S1208). Although a plurality of candidate vector values can be taken, error values are evaluated for all, and the candidate motion vector value having the smallest error value is set as a temporary motion vector value (S1209).
  • the error value is compared with a threshold value ⁇ . If the error value is smaller than ⁇ (S1210: YES), The motion vector value is determined as motion compensation prediction information and stored as a motion vector of the target reference image (S1211). On the other hand, when the error value is greater than or equal to the threshold value ⁇ (S1210: NO), the temporary motion vector value is not adopted as the motion compensation prediction information, and it is determined that there is no information on the target reference image (S1212).
  • the threshold value ⁇ it is possible to transmit a fixed value as an implicit value (eg, 2 ⁇ block size) or information in units of slices, but an error when the target block is originally motion compensated predicted. If the value is calculated using the decoded image and compared with the value, for example, 4/3 or less, it can be determined that the value is reliable.
  • the stored motion compensated prediction information which is the same process as in the first embodiment. Are counted by reference image, and the most reference images are calculated (S1214).
  • the first reference and the second reference are also aggregated without distinction, and finally the number of times of application for each reference image indicated by the motion vector is counted.
  • the reference image when there is one reference image that is the most similar to the first embodiment (S1215: YES), the reference image is selected as a standard reference image (S1216). If there are a plurality of reference images (S1216: NO), the reference image is preferentially selected as a reference image closer to the encoding target block, and if the distance is equal, the code is preceded in time. It is determined that the converted reference image has a smaller number of times of predictive encoding (S1217).
  • the process of calculating a predicted motion vector value for the selected reference reference image is the same as that in the first embodiment, but since the adopted candidate vector values are stored, a predicted vector suitable for more motion vector values. Can be calculated.
  • the horizontal and vertical median values of the motion vector values in the motion compensated prediction are calculated ( S1221). If an even-numbered motion vector is acquired (S1219: YES), the motion vector value farthest from the average of the two central motion vector values is excluded (S1220), and then the median value is acquired (S1221).
  • the motion vector value is set as a predicted motion vector value (S1224).
  • the calculated reference ID and predicted motion vector value are output to the inter-reference image motion vector detection units 108 and 215 (S1225), and the standard reference image determination process is terminated.
  • the inter-reference image motion vector detection process will be described.
  • the reference ID and predicted motion vector of the reference reference image input from the reference reference image determination unit 107 by the reference reference image acquisition unit 700.
  • the image block moved mvL0 from the coding target block of the standard reference image is extracted from the decoded reference image memory 118, and the first reference block is acquired (S1300).
  • the acquired reference block is stored in the standard reference image memory 702.
  • the motion vector detection range setting unit 701 based on the reference ID and the predicted motion vector value mvL0 of the standard reference image input from the standard reference image determination unit 107,
  • the detection range of the motion vector between the reference images is set (S1301).
  • the detection method can be performed in the same manner as in the first embodiment, but is different in that it is performed on a plurality of reference images.
  • the reference image acquisition unit 703 acquires, from the decoded reference image memory 118, the reference blocks in the motion vector detection range in the corresponding reference image specified by the motion vector detection range setting unit 701 for all reference images ( S1302) and output to the block matching evaluation unit 704.
  • the block matching evaluation unit 704 blocks the error sum for each pixel between the first reference block stored in the standard reference image memory 702 and the reference block of the corresponding reference image input from the reference image acquisition unit 703. The calculation is performed for each unit smaller than the size, and the reference block having a small sum by the small unit and the motion vector value when the reference block is acquired are stored (S1303).
  • the block size of the first reference block is 16 ⁇ 16 pixels
  • the unit smaller than the block size is 4 ⁇ 4 pixels
  • reference to each reference image is performed for 16 small block sizes.
  • An inter-image motion vector value mvInterRef and an error evaluation value in block matching are stored.
  • the error evaluation values of all reference images are compared for each 4 ⁇ 4 block, and the reference image having the smallest value is determined.
  • the second reference image is determined (S1305).
  • the error evaluation value it is possible to use the error evaluation value at the time of block matching as it is, but it is also possible to make a comparison with the error value from which the DC component of the error is removed. It is possible to generate a bidirectional prediction signal that makes use of a change in DC.
  • the reference image and motion vector conforming to the block shape as shown in the example of FIG. A value can be generated. It is possible to realize a configuration in which an appropriate predicted image is synthesized for each part with respect to a reference block of a reference image serving as a reference.
  • the image of the second reference block for the block for which unidirectional prediction is performed is not used in the direct motion compensation prediction unit. For example, if the same pixel data of the first reference block is stored, the average value is calculated for all the blocks in the block. Even when applied to the value of, a unidirectional prediction block can be generated.
  • the reference ID and reference inter-picture motion vector value mvInterRef in units of 4 ⁇ 4 blocks, and information indicating unidirectional prediction or bidirectional prediction, together with mvL0, the first reference block, and the second reference block, It outputs to the direct motion compensation prediction part 109 (S1310).
  • the motion vector result for the reference image that is not used is obtained with respect to the motion compensated prediction information used in the peripheral blocks of the coding target block.
  • a reference image that more appropriately reflects the importance of motion compensated prediction information by having a function to evaluate using the decoded image with spatially and temporally adjacent motion vector information on both the encoding side and the decoding side The accuracy of the standard reference image determination process in Embodiment 1 can be improved, and the direct mode motion compensation prediction efficiency generated using the result can be further improved.
  • Embodiment 2 motion vector detection between reference images is performed between a plurality of reference images for a motion compensated prediction block of a reference image serving as a reference, and bi-directional prediction is performed based on error evaluation values.
  • a video that accompanies a screen change such as a scene change or when an object is hidden by having both the encoding side and the decoding side have a function of determining whether the other reference image is to be determined and performing bidirectional prediction. It is possible to generate a motion compensated prediction signal using only reference image information suitable for the signal without additional information, and the motion compensation prediction efficiency in the direct mode can be further improved.
  • a function of selecting a reference image suitable for prediction (reference numerals 107, 215, 5)) on both the encoding side and the decoding side, a reference image suitable for the direct mode in which motion compensation prediction is performed without transmitting motion vector information can be recognized without additional information, so that prediction efficiency is improved. Therefore, it is possible to realize moving picture coding with improved coding efficiency.
  • a motion vector between another reference image is detected for a motion compensated prediction block of a reference image serving as a reference (see reference numerals 108, 215, FIG. 7, and FIG. 8). )
  • a motion compensated prediction block of another reference image based on the detected motion vector and performing bi-directional prediction (see reference numerals 109 and 216) on both the encoding side and the decoding side, motion vector information Therefore, it is possible to improve the prediction efficiency for a moving image signal accompanied by a temporal change in motion without transmitting the image, and to realize moving image coding with improved coding efficiency.
  • an error evaluation value between reference images when a motion vector between a plurality of other reference images is detected with respect to a motion compensated prediction block of a reference image serving as a reference is obtained.
  • both the encoding side and the decoding side have a function (see S1303 to S1308 in FIG. 13) for determining whether to perform bi-prediction and determination of the other reference image for bi-directional prediction.
  • a motion compensated prediction signal that uses only reference image information suitable for moving image signals accompanying screen changes such as scene changes or when objects are hidden can be generated without additional information, improving coding efficiency. It is possible to realize the encoded video encoding.
  • the motion vector results for the reference images that are not used are spatially and temporally adjacent to the motion compensated prediction information used in the peripheral blocks of the encoding target block.
  • the moving picture encoding apparatus and moving picture decoding apparatus presented as the first, second, third, fourth, and fifth embodiments are physically a CPU (central processing unit), a memory, and the like.
  • a CPU central processing unit
  • the program can be provided by being recorded on a computer-readable recording medium, provided from a server through a wired or wireless network, or provided as data broadcasting of terrestrial or satellite digital broadcasting. is there.
  • the present invention can be used for a video signal encoding technique and decoding technique.

Landscapes

  • Engineering & Computer Science (AREA)
  • Multimedia (AREA)
  • Signal Processing (AREA)
  • Compression Or Coding Systems Of Tv Signals (AREA)

Abstract

Selon l'invention, une unité de détermination d'image de référence standard (108) sélectionne une première image de référence utilisée dans un traitement de compensation de mouvement d'un bloc à coder conformément à la fréquence d'une image de référence utilisée dans une prédiction par compensation de mouvement pour des blocs entourant le bloc à coder, et obtient une valeur de vecteur de mouvement prédite à partir de la première image de référence sélectionnée à titre de première valeur de vecteur de mouvement. Une unité de synthèse d'image de référence génère un bloc de référence de synthèse qui a été synthétisé à partir d'un bloc de référence qui a été extrait de la première image de référence à l'aide de la première valeur de vecteur de mouvement et d'une zone prédéterminée d'au moins une autre image de référence. Une unité de codage entropique (118), en utilisant le bloc de référence de synthèse à titre de bloc prédit, code un bloc différence prédit dans lequel le bloc prédit a été soustrait du bloc à coder.
PCT/JP2011/003170 2010-06-09 2011-06-06 Dispositif de codage d'image animée, procédé de codage d'image animée et programme de codage d'image animée, et dispositif de décodage d'image animée, procédé de décodage d'image animée et programme de décodage d'image animée WO2011155171A1 (fr)

Applications Claiming Priority (4)

Application Number Priority Date Filing Date Title
JP2010132184A JP2011259230A (ja) 2010-06-09 2010-06-09 動画像復号装置、動画像復号方法及び動画像復号プログラム
JP2010-132183 2010-06-09
JP2010-132184 2010-06-09
JP2010132183A JP2011259229A (ja) 2010-06-09 2010-06-09 動画像符号化装置、動画像符号化方法及び動画像符号化プログラム

Publications (1)

Publication Number Publication Date
WO2011155171A1 true WO2011155171A1 (fr) 2011-12-15

Family

ID=45097788

Family Applications (1)

Application Number Title Priority Date Filing Date
PCT/JP2011/003170 WO2011155171A1 (fr) 2010-06-09 2011-06-06 Dispositif de codage d'image animée, procédé de codage d'image animée et programme de codage d'image animée, et dispositif de décodage d'image animée, procédé de décodage d'image animée et programme de décodage d'image animée

Country Status (1)

Country Link
WO (1) WO2011155171A1 (fr)

Citations (5)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
JP2004048711A (ja) * 2002-05-22 2004-02-12 Matsushita Electric Ind Co Ltd 動画像の符号化方法および復号化方法およびデータ記録媒体
JP2004129191A (ja) * 2002-10-04 2004-04-22 Lg Electronics Inc Bピクチャのダイレクトモード動きベクトル演算方法
JP2008092352A (ja) * 2006-10-03 2008-04-17 Ntt Docomo Inc 画像予測符号化装置、画像予測符号化方法、画像予測符号化プログラム、画像予測復号装置、画像予測復号方法及び画像予測復号プログラム
JP2008154015A (ja) * 2006-12-19 2008-07-03 Hitachi Ltd 復号化方法および符号化方法
WO2009126260A1 (fr) * 2008-04-11 2009-10-15 Thomson Licensing Procédés et appareil permettant de prévoir une concordance de gabarit (tmp) dans le codage et décodage de données vidéo

Patent Citations (5)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
JP2004048711A (ja) * 2002-05-22 2004-02-12 Matsushita Electric Ind Co Ltd 動画像の符号化方法および復号化方法およびデータ記録媒体
JP2004129191A (ja) * 2002-10-04 2004-04-22 Lg Electronics Inc Bピクチャのダイレクトモード動きベクトル演算方法
JP2008092352A (ja) * 2006-10-03 2008-04-17 Ntt Docomo Inc 画像予測符号化装置、画像予測符号化方法、画像予測符号化プログラム、画像予測復号装置、画像予測復号方法及び画像予測復号プログラム
JP2008154015A (ja) * 2006-12-19 2008-07-03 Hitachi Ltd 復号化方法および符号化方法
WO2009126260A1 (fr) * 2008-04-11 2009-10-15 Thomson Licensing Procédés et appareil permettant de prévoir une concordance de gabarit (tmp) dans le codage et décodage de données vidéo

Similar Documents

Publication Publication Date Title
US10110902B2 (en) Method and apparatus for encoding/decoding motion vector
US8428136B2 (en) Dynamic image encoding method and device and program using the same
US8204127B2 (en) Method and apparatus for encoding and decoding image by using multiple reference-based motion prediction
KR101227601B1 (ko) 시차 벡터 예측 방법, 그 방법을 이용하여 다시점 동영상을부호화 및 복호화하는 방법 및 장치
JP4527677B2 (ja) 動画像符号化方法、動画像符号化装置、動画像符号化プログラム
US11917157B2 (en) Image encoding/decoding method and device for performing PROF, and method for transmitting bitstream
KR101388902B1 (ko) 모션 추정을 위한 기법들
WO2010082463A1 (fr) Appareil et procédé de codage d'image et appareil et procédé de décodage d'image
JP2011147049A (ja) 画像処理装置および方法、並びにプログラム
KR20120025111A (ko) 참조 화소들의 특성을 이용하여 예측모드 정보의 전송을 생략할 수 있는 화면 내 예측 부호화/복호화 장치 및 그 방법
WO2011070730A1 (fr) Dispositif de codage vidéo et dispositif de décodage vidéo
JP2007531444A (ja) ビデオデータのための動き予測及びセグメンテーション
Kamp et al. Fast decoder side motion vector derivation for inter frame video coding
WO2012056924A1 (fr) Dispositif de traitement d'image et procédé de traitement d'image
JP6036972B2 (ja) 動画像復号装置、動画像復号方法、及び動画像復号プログラム、並びに、受信装置、受信方法、及び受信プログラム
JP5737384B2 (ja) 動画像復号装置、動画像復号方法、及び動画像復号プログラム、並びに、受信装置、受信方法、及び受信プログラム
JP2011259230A (ja) 動画像復号装置、動画像復号方法及び動画像復号プログラム
WO2011155171A1 (fr) Dispositif de codage d'image animée, procédé de codage d'image animée et programme de codage d'image animée, et dispositif de décodage d'image animée, procédé de décodage d'image animée et programme de décodage d'image animée
WO2012090425A1 (fr) Dispositif de codage d'image animée, procédé de codage d'image animée et programme de codage d'image animée, ainsi que dispositif de décodage d'image animée, procédé de décodage d'image animée et programme de décodage d'image animée
JP2011259229A (ja) 動画像符号化装置、動画像符号化方法及び動画像符号化プログラム
WO2011099242A1 (fr) Dispositif de codage d'images, dispositif de décodage d'images, procédé de codage d'images et procédé de décodage d'images
JP2011217104A (ja) 動画像符号化装置、動画像符号化方法及び動画像符号化プログラム
JP2012138838A (ja) 動画像復号装置、動画像復号方法、及び動画像復号プログラム
JP2011217105A (ja) 動画像復号装置、動画像復号方法及び動画像復号プログラム
JP2012138837A (ja) 動画像符号化装置、動画像符号化方法、及び動画像符号化プログラム

Legal Events

Date Code Title Description
121 Ep: the epo has been informed by wipo that ep was designated in this application

Ref document number: 11792130

Country of ref document: EP

Kind code of ref document: A1

NENP Non-entry into the national phase

Ref country code: DE

122 Ep: pct application non-entry in european phase

Ref document number: 11792130

Country of ref document: EP

Kind code of ref document: A1