WO2012147306A1 - Dispositif de codage d'image animée, procédé de codage d'image animée, programme de codage d'image animée, dispositif de décodage d'image animée, procédé de décodage d'image animée et programme de décodage d'image animée - Google Patents

Dispositif de codage d'image animée, procédé de codage d'image animée, programme de codage d'image animée, dispositif de décodage d'image animée, procédé de décodage d'image animée et programme de décodage d'image animée Download PDF

Info

Publication number
WO2012147306A1
WO2012147306A1 PCT/JP2012/002649 JP2012002649W WO2012147306A1 WO 2012147306 A1 WO2012147306 A1 WO 2012147306A1 JP 2012002649 W JP2012002649 W JP 2012002649W WO 2012147306 A1 WO2012147306 A1 WO 2012147306A1
Authority
WO
WIPO (PCT)
Prior art keywords
block
adjacent
prediction
reference destination
partition
Prior art date
Application number
PCT/JP2012/002649
Other languages
English (en)
Japanese (ja)
Inventor
福島 茂
西谷 勝義
上田 基晴
徹 熊倉
Original Assignee
株式会社Jvcケンウッド
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Priority claimed from JP2011101951A external-priority patent/JP5703942B2/ja
Priority claimed from JP2011101952A external-priority patent/JP5703943B2/ja
Application filed by 株式会社Jvcケンウッド filed Critical 株式会社Jvcケンウッド
Publication of WO2012147306A1 publication Critical patent/WO2012147306A1/fr

Links

Images

Classifications

    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04NPICTORIAL COMMUNICATION, e.g. TELEVISION
    • H04N19/00Methods or arrangements for coding, decoding, compressing or decompressing digital video signals
    • H04N19/46Embedding additional information in the video signal during the compression process
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04NPICTORIAL COMMUNICATION, e.g. TELEVISION
    • H04N19/00Methods or arrangements for coding, decoding, compressing or decompressing digital video signals
    • H04N19/50Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using predictive coding
    • H04N19/503Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using predictive coding involving temporal prediction
    • H04N19/51Motion estimation or motion compensation
    • H04N19/513Processing of motion vectors
    • H04N19/517Processing of motion vectors by encoding
    • H04N19/52Processing of motion vectors by encoding by predictive encoding

Definitions

  • the present invention relates to a moving picture encoding and decoding technique, and more particularly to a moving picture encoding and decoding technique using motion compensation prediction.
  • the moving picture coding method that divides a picture into rectangular blocks, represented by MPEG (Moving Picture Experts ⁇ Group), and performs motion estimation and compensation between blocks, the coding amount of the motion vector generated in each block
  • MPEG Motion Picture Experts ⁇ Group
  • a prediction process is performed on the motion vector.
  • the motion vector detected in units of macroblocks is taken as a difference from the motion vector of the macroblock encoded immediately before, and the amount of code is reduced by encoding the difference vector. Yes.
  • MPEG-4 AVC / H. H.264 uses the fact that motion vectors have a strong correlation with the motion vectors of neighboring blocks, and performs prediction from neighboring blocks and encodes the difference vector to reduce the amount of code. .
  • the median is calculated from the motion vectors of adjacent blocks on the left, top and top right of the block to be processed, and motion vector prediction is realized by taking the difference from the median.
  • the motion information of the block to be processed and the neighboring already-encoded neighboring blocks are the same in the standard work of moving picture coding in Patent Document 1 and recent ISO / IEC and ITU-T. If so, the block to be processed does not encode its own motion information, uses the motion information of the adjacent block for encoding, and encodes additional information specifying the adjacent block having the motion information to be referenced. Thus, attempts have been made to reduce the amount of code of motion information. Such a method is called “merge” and has attracted attention as a method for reducing the code amount of motion information.
  • a 1-bit flag (whether to use the motion information of the block to be processed itself detected by motion vector detection or to use the motion information of an adjacent block by the merge method as additional information for each block)
  • merge_lu_flag a 1-bit flag that defines whether to refer to the left or right of neighboring blocks when the merge method is applied. Entropy coding using arithmetic codes is performed on these flags.
  • a short code is assigned to data having a high appearance rate
  • a long code is assigned to data having a low appearance rate
  • data compression is performed to reduce the code amount.
  • the flag merge_lu_flag depends on the texture in the screen, since a code is assigned in advance to select the left or top as the reference destination adjacent block, even if the occurrence frequency of the reference destination selection is biased Therefore, there is a problem that the code amount cannot be sufficiently reduced because the encoding is performed with equal probability.
  • the present invention has been made in view of such circumstances, and an object of the present invention is to provide a technique for improving the coding efficiency by reducing the code amount of a flag representing a reference destination of motion information.
  • a moving image encoding device is a moving image encoding device that encodes the moving image using a motion vector in units of blocks obtained by dividing each picture of the moving image. Any one of the adjacent blocks determined based on an adjacency condition between the encoding target block and the adjacent block is moved from among a plurality of encoded adjacent blocks adjacent to the encoding target block.
  • a reference destination designating unit (113) that designates a reference block for referring to information, intra-picture prediction, inter-picture prediction using a motion vector of an encoding target block, and inter-picture using a motion vector of an adjacent block
  • the inter-picture prediction using the motion vector of the adjacent block is selected as the prediction method from the prediction, it is referred to by the selected prediction method.
  • a prediction selection unit that generates, instead of motion information, reference destination valid information indicating whether or not a contact block matches the reference destination block specified by the reference destination specifying unit, and information indicating the selected prediction method (112).
  • This method is a moving image encoding method for encoding the moving image using a motion vector in units of blocks obtained by dividing each picture of the moving image, and a plurality of encoded adjacent regions adjacent to the encoding target block.
  • a moving picture decoding apparatus is a moving picture decoding apparatus that decodes a bitstream in which the moving picture is encoded using a motion vector in units of blocks obtained by dividing each picture of the moving picture.
  • a variable for decoding the information indicating the prediction method and the reference destination valid information from the bitstream including information indicating the method and reference destination valid information indicating the validity of the adjacent block referred to by the selected prediction method Using the long decoding unit (201), intra-picture prediction, inter-picture prediction using a motion vector of a decoding target block, and one motion vector of a plurality of decoded adjacent blocks adjacent to the decoding target block
  • a prediction selection unit (209) that selects one of prediction methods based on information indicating the prediction method from among the inter-image predictions; and the prediction selection Therefore, when inter-picture prediction using a motion vector of an adjacent block is selected as a prediction method, any one adjacent block determined based on an adjacent condition between the decoding target block and the adjacent block is moved.
  • a motion information selection unit (210) that determines
  • Another aspect of the present invention is a video decoding method.
  • This method is a moving image decoding method for decoding a bitstream in which the moving image is encoded using a motion vector in units of blocks obtained by dividing each picture of the moving image, and is selected as information indicating a prediction method
  • a variable length decoding step for decoding the information indicating the prediction method and the reference destination valid information from the bitstream including the reference destination valid information indicating the validity of the adjacent block referred to by the prediction method, and intra prediction
  • the prediction method includes: inter-picture prediction using a motion vector of a decoding target block; and inter-picture prediction using any one of a plurality of decoded adjacent blocks adjacent to the decoding target block.
  • a prediction selection step of selecting one of the prediction methods based on the information indicated, and the prediction selection step When inter-picture prediction using a motion vector of a block is selected, a reference for referring to motion information for any one adjacent block determined based on an adjacent condition between the decoding target block and the adjacent block
  • the present invention it is possible to improve the coding efficiency by reducing the code amount of the flag representing the reference destination of the motion information.
  • FIG. 10A is a flowchart showing details of the calculation processing of the number of candidates NumMergeCandidates in FIG. 9, and FIG. 10B is a diagram showing an example of a reference destination partition adjacent to the processing target partition. It is a flowchart explaining operation
  • FIG. 13A and FIG. 13B are flowcharts for explaining processing executed in accordance with a flag value indicating a predicted reference destination. It is a flowchart explaining the operation
  • Embodiments of the present invention relate to moving picture coding technology, and in particular, improve the efficiency of motion information coding in moving picture coding in which a picture is divided into rectangular blocks and motion estimation and compensation are performed in units of blocks between pictures. Therefore, the motion information of the block to be processed is obtained by using the motion information such as the motion vector of the encoded adjacent block around the block to be processed and the reference picture number as the motion information of the block to be processed.
  • the present invention relates to a technique for reducing the amount of codes by encoding additional information indicating adjacent blocks to be referred to without encoding.
  • the reference destination of motion information is predicted using information such as the size of neighboring blocks and the prediction mode, and is determined by the prediction.
  • a flag merge_proble_flag that represents a match between the adjacent block of the reference destination and the adjacent block actually selected by encoding is defined. If the determination rate by prediction is high, the probability of occurrence of the bit represented by the flag merge_proble_flag increases, so the number of bits representing the flag decreases, and the amount of generated code can be reduced.
  • the flag merge_proble_flag itself representing the match with the above-mentioned reference destination prediction is defined on the assumption that the coincidence accuracy of the determination of the adjacent block that becomes the reference destination by the reference destination prediction is high.
  • the flag merge_direc_flag having the same function as the merge_lu_flag only when the prediction of the reference destination is indefinite. It is also possible to offset the increase in the code amount due to the flag merge_direc_flag by reducing the code amount required for the flag merge_proble_flag, and to relatively improve the encoding efficiency.
  • FIG. 1 is a block diagram showing a configuration of a moving picture encoding apparatus according to an embodiment.
  • the moving image encoding apparatus includes a motion vector detection unit 101, a subtraction unit 102, an orthogonal transform / quantization unit 103, a variable length encoding unit 104, an inverse quantization / inverse orthogonal transform unit 105, and a motion compensation unit 106.
  • the intra prediction unit 111 performs intra prediction from the input encoding target block and the encoded decoded block image adjacent to the encoding target block stored in the memory 110, and determines the intra prediction block. Output.
  • the reference destination specifying unit 113 specifies any one of a plurality of adjacent blocks adjacent to the encoding target block stored in the memory 110 as a reference destination adjacent block when performing the merge processing, and the reference destination adjacent The motion vector of the block is given to the motion compensation unit 106.
  • the motion vector detection unit 101 refers to the input image signal and the reference image signal stored in the memory 110 to perform matching between blocks in units of blocks to detect a motion vector, and to detect the detected motion vector. This is given to the motion compensation unit 106.
  • the motion compensation unit 106 generates a prediction image using the motion vector of the block to be encoded detected by the motion vector detection unit 101 or the motion vector of the adjacent block specified by the reference destination specifying unit 113.
  • the weighted prediction unit 107 adaptively multiplies the prediction image generated by the motion compensation unit 106 by a weight coefficient to generate a final prediction image, and provides the prediction selection unit 112 with it.
  • the prediction selection unit 112 performs intra-screen prediction by the intra-screen prediction unit 111, inter-image prediction using the motion vector of the block to be encoded by the motion compensation unit 106, and the reference destination adjacent block specified by the reference destination specifying unit 113.
  • One prediction method with the smallest code amount is selected from the inter-image prediction using the motion vector, and a prediction image by the selected prediction method is given to the subtraction unit 102 and the addition unit 108.
  • the reference destination specifying unit 113 selects the adjacent block that is the reference destination of the merge process.
  • the prediction process is executed, and the prediction result of the merge reference destination is given to the prediction selection unit 112.
  • the prediction selecting unit 112 determines whether or not the reference destination adjacent block predicted by the reference destination specifying unit 113 matches the reference destination adjacent block used in the actually selected prediction method, and the reference destination adjacent block
  • the reference destination valid information indicating the validity of the prediction result is supplied to the variable length coding unit 104 together with the information indicating the selected prediction method.
  • the subtraction unit 102 generates a residual signal by subtracting the image to be encoded and the predicted image, and supplies the residual signal to the orthogonal transform / quantization unit 103.
  • the orthogonal transform / quantization unit 103 performs orthogonal transform and quantization on the residual signal, generates a transform signal, and supplies the transform signal to the variable length coding unit 104 and the inverse quantization / inverse orthogonal transform unit 105.
  • the variable length coding unit 104 entropy codes the residual signal that has been orthogonally transformed and quantized.
  • the variable length encoding unit 104 also encodes reference destination valid information generated by the prediction selection unit 112 and information indicating the selected prediction method, and outputs a bit stream including an encoded image.
  • the inverse quantization / inverse orthogonal transform unit 105 performs inverse quantization and inverse orthogonal transform on the transform signal received from the orthogonal transform / quantization unit 103 to return to the original residual signal.
  • the adding unit 108 adds the predicted image and the residual signal, generates a decoded image, and supplies the decoded image to the deblocking filter unit 109.
  • the deblocking filter unit 109 performs a process of reducing block distortion due to encoding on the decoded image and stores it in the memory 110.
  • the memory 110 stores information on decoded images and already encoded images.
  • FIG. 2 is a block diagram showing a configuration of a moving picture decoding apparatus according to an embodiment corresponding to the moving picture encoding apparatus of FIG.
  • the moving picture decoding apparatus includes a variable length decoding unit 201, an inverse quantization / inverse orthogonal transform unit 202, a motion compensation unit 203, a weighted prediction unit 204, an addition unit 205, a deblocking filter unit 206, and a memory 207.
  • the decoding process of the moving picture decoding apparatus in FIG. 2 corresponds to the decoding process provided in the moving picture encoding apparatus in FIG. 1, so that the inverse quantization / inverse orthogonal transform unit 202 in FIG.
  • Each configuration of the motion compensation unit 203, the weighted prediction unit 204, the addition unit 205, the deblocking filter unit 206, the memory 207, and the in-screen prediction unit 208 is the inverse quantization / inverse of the moving picture coding apparatus in FIG.
  • the orthogonal transform unit 105, the motion compensation unit 106, the weighted prediction unit 107, the addition unit 108, the deblocking filter unit 109, the memory 110, and the in-screen prediction unit 111 have functions corresponding to the respective configurations.
  • the variable length decoding unit 201 decodes the bitstream and outputs information on the prediction residual signal and the motion vector, provides the prediction residual signal to the inverse quantization / inverse orthogonal transform unit 202, and provides information on the motion vector as the motion information.
  • Information about the coding mode is given to the selection unit 210 to the intra-screen / inter-screen prediction selection unit 209.
  • the inverse quantization / inverse orthogonal transform unit 202 performs inverse orthogonal transform and inverse quantization on the prediction residual signal decoded by the variable length decoding unit 201.
  • the addition unit 205 decodes the image signal by adding the prediction residual component inversely transformed by the inverse quantization / inverse orthogonal transformation unit 202 and the prediction image calculated by the weighted prediction unit 204, and decodes the image signal. This is applied to the blocking filter unit 206.
  • the deblocking filter unit 206 performs a process of reducing block distortion due to encoding on the decoded image and stores it in the memory 207.
  • the intra-screen / inter-screen prediction selection unit 209 determines whether the encoding mode is intra-screen prediction or inter-screen prediction, and if it is intra-screen prediction, instructs the intra-screen prediction unit 208 to execute the intra-screen prediction process. In the case of inter-screen prediction, the motion information selection unit 210 is instructed to execute inter-screen prediction processing.
  • the motion information selection unit 210 selects whether or not motion information such as a motion vector is due to merge processing. If it is due to merge processing, the motion information selection unit 210 is based on predetermined adjacent conditions based on reference destination valid information. It is determined whether the adjacent block to be determined can be designated as a reference block.
  • the intra prediction unit 208 performs intra prediction from the input decoding target block and the decoded block image adjacent to the decoding target block stored in the memory 207, and passes the intra prediction block to the adding unit 205.
  • the motion compensation unit 203 generates a prediction image using the decoded motion vector of the decoding target block when the inter-screen prediction is not merge processing, that is, is based on normal motion vector detection processing.
  • the motion compensation unit 203 performs prediction using the motion vector of the reference destination adjacent block predicted by the reference destination prediction of the motion information. Generate an image.
  • the weighted prediction unit 204 adaptively multiplies the prediction image generated by the motion compensation unit 203 by a weighting factor to generate a final prediction image, which is given to the addition unit 205.
  • the prediction method of an adjacent block of a reference destination is performed in the prediction selection unit 112 of the video encoding device in FIG. 1 and the motion information selection unit 210 of the video decoding device in FIG.
  • the prediction selection unit 112 of the video encoding device from among the motion information detected by the normal motion vector detection process or the inter-screen prediction based on the motion information of the adjacent block selected by the merge process, and the intra-screen prediction, One encoding mode is selected that has the least amount of generated code, the best image quality, or both. Further, if the selected encoding mode is inter-screen prediction, it is determined whether the motion information is detected by a normal motion vector detection process or is selected by a merge process, and the merge process is performed. In this case, the reference destination prediction of the motion information is performed, the identity between the predicted reference destination adjacent block and the adjacent block selected by the merging process is determined, and the result is encoded as reference destination valid information. Is transmitted. When the merge process is not performed, normal motion vector prediction is performed, and a difference motion vector between the motion vector and the predicted motion vector and other motion information are encoded and transmitted.
  • the intra-screen / inter-screen prediction selection unit 209 determines the inter-screen based on the encoding mode and motion information decoded from the bit stream by the variable length decoding unit 201. If it is determined to be prediction, it is further determined whether or not the inter-screen prediction is due to merge processing. If it is determined to be due to merge processing, motion information reference destination prediction is performed. The motion information is acquired by determining the identity between the predicted reference adjacent block and the adjacent block specified by the decoded additional information. In the following embodiments, details of motion information reference destination prediction will be described.
  • Example 1 Prior to describing an example of a motion information reference destination prediction method, terms used in this example will be described.
  • the screen is equally divided into square rectangular blocks of the same size.
  • This block is called a coding block and is the basis of processing when performing coding and decoding.
  • the coding block can be divided into four blocks with a small block size in order to optimize the coding process.
  • a coded block that is divided into equal screen sizes in the screen shown in FIG. 3 is called a maximum coded block, and the inside of which is divided into four according to the coding conditions is collectively called a coded block.
  • An encoded block having a minimum size that cannot be further divided into four is referred to as a minimum encoded block.
  • a block that performs this motion compensation is called a prediction block.
  • the prediction block is represented by the same size as the coding block, and according to the motion compensation, the prediction block is regarded as one block without being divided, and is divided into two in the horizontal or vertical direction. Divided into four parts by equal division.
  • the mode corresponding to the division type is defined according to the size after the division, and is shown in FIG.
  • Each area obtained by dividing the prediction block is called a partition.
  • numbers starting from 0 are assigned to the partitions existing in the prediction block in the zigzag scan order. This number is called a partition number and is represented by puPartIdx.
  • the number described in the partition of the prediction block in FIG. 4 represents the partition number of the partition.
  • a motion information reference destination prediction method will be described with reference to the drawings.
  • the motion information reference destination prediction method is performed in any of encoding and decoding processes for each partition constituting a prediction block.
  • the motion information is referred from the surrounding partition adjacent to the partition of the prediction block (partition to be processed in FIG. 5) defined for motion compensation within the coding block in the same picture. Select the destination partition.
  • a partition that has already been encoded before encoding the partition to be processed is a candidate. Since encoding is performed in block units in the raster scan order from the upper left to the lower right of the screen, here, in principle, the left or upper partition adjacent to the periphery of the partition to be processed is considered as a reference destination candidate. The following explanation will be based on this assumption unless otherwise noted.
  • FIG. 6 is an enlarged view of the area represented by the thick dotted circle in FIG.
  • FIG. 6 shows a partition to be processed and a partition adjacent to it.
  • nA represents the total number of partition groups adjacent to the left
  • nB represents the total number of partition groups adjacent immediately above.
  • the reference destination on the left or directly above is determined based on the following rule. Determine candidate partitions.
  • the top partition A0 among them is set as a reference destination candidate.
  • the leftmost partition B0 among them is set as a reference destination candidate.
  • FIG. 7 shows a first syntax pattern described in a sequence parameter set (SPS) of a moving image bit stream.
  • the sequence parameter set is a header in which information related to coding of the entire sequence is described.
  • the motion information reference destination prediction method When performing inter-screen prediction over the entire sequence, the motion information reference destination prediction method according to the present embodiment is applied.
  • a first flag inter_merge_flag indicating whether or not to perform is set.
  • FIG. 8 shows a second syntax pattern described for each prediction block in the slice.
  • the prediction mode of the prediction block is inter-screen prediction (MODE_INTER)
  • the number of effective neighboring partitions NumMergeCandidates is obtained for each partition in the prediction block, and when NumMergeCandidates exceeds 0, merging is performed in this partition.
  • a second flag merge_flag [i] indicating whether or not to apply is set.
  • i represents the partition number in the prediction block.
  • the second flag merge_flag [i] is not coded and is not described in the bitstream when the motion information reference prediction method according to the present embodiment is not applied, that is, when the inter_merge_flag is false (0).
  • the third flag merge_proble_flag [i] does not apply merging, that is, when merge_flag [i] is false (0), the motion information detected by the normal motion vector detection is used, so the third flag merge_proble_flag [i] is used. i] does not need to be determined. Also, even when NumMergeCandidates is 1, it is not encoded. This is because, if the number of effective partitions adjacent to each other is one, one of the partitions becomes a reference destination partition, so that the motion information of the reference destination partition is determined without transmitting merge_proble_flag [i].
  • the reference destination specifying unit 113 determines whether or not to use a partition adjacent to the left and directly above the processing target partition as a reference destination of the merge processing. The operation of the reference destination specifying unit 113 will be described with reference to the flowchart of FIG. First, the number NumMergeCandidates of peripheral partitions adjacent to the partition to be processed is calculated (S101).
  • FIG. 10A is a flowchart showing details of the calculation processing of the number of candidates NumMergeCandidates.
  • NumMergeCandidates is set to 0 (S201).
  • a partition located to the left of the partition to be processed is set as a reference destination partition (S202). It is determined from the position information in the screen of the partition to be processed whether or not the reference destination partition exists in the screen (S203).
  • FIG. 10 (b) shows an example of a reference destination partition (rectangle represented by diagonal lines in the figure) adjacent to the partition to be processed (gray rectangle in the figure).
  • a reference destination partition rectangle represented by diagonal lines in the figure
  • FIG. 10 (b) shows an example of a reference destination partition (rectangle represented by diagonal lines in the figure) adjacent to the partition to be processed (gray rectangle in the figure).
  • the partition to be processed is located at the upper left, there is no reference destination partition, only when it is located at the top of the screen, only at the left, when located at the left edge of the screen, only at the top, other positions Then it is on the left and top.
  • the reference destination partition does not exist in the screen (No in S203)
  • the subsequent processing is skipped and the process proceeds to Step S207.
  • the reference destination specifying unit 113 acquires the motion information of the reference destination partition (S204).
  • the reference destination designating unit 113 reads the encoded motion information corresponding to the reference destination partition from the memory 110. Based on the read motion information of the reference destination partition, the reference destination specifying unit 113 determines whether or not the encoding mode of the reference destination partition is intra (S205). If it is intra, the process proceeds to S207. If it is not intra, NumMergeCandidates is incremented by 1 (S206).
  • Step S207 It is determined whether the setting of the reference destination partition is on the partition to be processed. If the reference destination partition is set on the processing target partition, the process ends and outputs NumMergeCandidates. When the reference destination partition is not set on the processing target partition, the reference destination partition is set again on the processing target partition, and the determination processing after step S203 is continued (S208).
  • the number of candidates NumMergeCandidates is calculated as described above.
  • the reference destination specifying unit 113 acquires the motion information of the reference destination partition from the memory 110 (S105). ).
  • the reference destination specifying unit 113 supplies the acquired motion information of the reference destination partition to the motion compensation unit 106, and the motion compensation unit 106 performs motion compensation of the processing target partition based on the motion information of the specified reference destination partition. .
  • the reference destination designating unit 113 reads out and acquires the motion information of the partitions adjacent to the left and above the processing target partition from the memory 110 (S104).
  • the reference destination specifying unit 113 compares the acquired motion information of the left and upper adjacent partitions (S106). Specifically, the prediction direction, the reference picture number, and the motion vector constituting the motion information are compared. When the motion information of the left and upper adjacent partitions is exactly the same (Yes in S106), the number of candidates NumMergeCandidates is changed to 1 (S107), and the reference destination designating unit 113 sets either the left or the upper adjacent The motion information of the partition to be supplied is supplied to the motion compensation unit 106 as the motion information of the reference destination partition. Here, the motion information of the partition adjacent to the left is selected. When the motion information of the left and upper adjacent partitions is different (Yes in S106), the reference destination specifying unit 113 supplies the two pieces of motion information to the motion compensation unit 106, respectively.
  • the reference destination specifying unit 113 supplies the motion information of the partition of the reference block selected by the merging process to the motion compensation unit 106, and the motion compensation unit 106 specifies the partition of the specified reference block. Motion compensation is performed based on the motion information. Also, the motion compensation unit 106 performs normal motion compensation based on the motion vector for each partition of the prediction block detected by the motion vector detection unit 101.
  • the in-screen prediction unit 111 performs in-plane prediction using pixel correlation in the screen. A prediction image generated by these three prediction methods is input to the prediction selection unit 112.
  • the prediction selection unit 112 calculates a generated code amount of additional information such as a residual image and motion information generated by the difference between the image to be encoded and the three input prediction images, and makes a prediction that minimizes the code amount Select a method.
  • the prediction selection unit 112 sets a flag merge_dirc_flag that indicates which one of the left or the upper is used as the motion information of the reference destination partition selected by the merge process.
  • merge_dirc_flag indicates which one of the left or the upper is used as the motion information of the reference destination partition selected by the merge process.
  • the prediction selection unit 112 sets a flag for merge processing defined by the above-described syntax when the prediction method by merge processing is selected.
  • FIG. 11 is a flowchart illustrating a procedure in which the prediction selection unit 112 sets a flag for merge processing.
  • merge_flag is set to 1 (S302).
  • the number of motion information selected by the merge process is checked (S304). Specifically, it is determined whether or not the number of adjacent partition candidates NumMergeCandidates to be specified by the reference destination specifying unit 113 is greater than one.
  • the process proceeds to prediction of a reference destination partition (S305).
  • a reference destination prediction method for motion information described later is used. Based on information such as the size of the partition adjacent to the left and top of the target partition and the length of the side where the target partition and the adjacent partition are in contact with each other, the reference partition is predicted and predicted A flag pred_direc_flag indicating the reference destination partition is output.
  • the flag pred_dirc_flag indicating the predicted reference destination partition indicates the selection destination on the left (1) or above (0) as the adjacent partition of the reference destination, as in the case of merge_direct_flag, and the reference destination partition cannot be selected by prediction.
  • Indefinite (2) is set.
  • the flag pred_direc_flag indicating the predicted reference destination partition set in this way is determined (S306).
  • the flag pred_dirc_flag indicating the predicted reference destination partition is 2 when prediction is impossible in the prediction of the reference destination partition (Yes in S306), and at this time, it is selected as the prediction method that actually minimizes the generated code amount.
  • the merge_direc_flag indicating the reference destination partition of the motion information used in the prediction performed by the merge processing is set in the flag merge_proble_flag indicating the reference destination valid information, and the process ends (S307).
  • the flag merge_proble_flag indicating the reference destination valid information directly indicates the reference destination partition in the merge process as in the conventional case.
  • merge_direc_flag represents a reference destination partition of motion information used in prediction by merge processing selected as a prediction method that actually minimizes the amount of generated code, and pred_direc_flag has the same value if the prediction of the reference destination partition is correct.
  • the flag merge_proble_flag is set to 1 (S309).
  • pred_dirc_flag has a different value. In this case, the flag merge_proble_flag is set to 0 (S310).
  • merge_proble_flag When merge_proble_flag indicating success / failure of the merge destination partition is used, merge_proble_flag becomes 1 when the merge destination partition is predicted, and merge_proble_flag becomes 0 only when the prediction is off. In other words, merge_proble_flag can be transmitted with a smaller amount of information by using arithmetic coding using the fact that merge_proble_flag has a higher probability of 1 and merge_proble_flag has a higher probability of 1 as the merge destination partition is predicted. it can.
  • the first flag inter_merge_flag described in the header of the SPS is described in the prediction block.
  • the second flag merge_flag and the third flag merge_proble_flag are set and encoded.
  • the motion information reference destination prediction method according to the present embodiment is used for the entire bitstream sequence. It is determined whether inter-screen prediction by merge processing is applied. When inter_merge_flag is true (1), merge processing is applied. When inter_merge_flag is false (0), merge processing is ignored and inter-screen prediction based on motion information decoded from a conventional bitstream is performed. . In the following description, it is assumed that inter_merge_flag is true (1), that is, the merge processing using the motion information reference destination prediction method according to the present embodiment is applied.
  • the intra-screen / inter-screen prediction selection unit 209 determines whether to select intra-screen or inter-screen prediction with reference to the prediction mode for each prediction block in the slice.
  • the intra prediction MODE_INTRA
  • the process proceeds to the intra prediction unit 208
  • the inter prediction MODE_INTER
  • the process proceeds to the motion information selection unit 210.
  • the motion information selection unit 210 uses, as information decoded from the bit stream for each partition of the prediction block, the motion information detected by the motion vector detection of encoding or the reference destination partition selected by the merge process. Either one of the motion information is selected, and the selected motion information is output to the motion compensation unit 203.
  • FIG. 12 is a flowchart showing a selection procedure of the motion information selection unit 210, which will be described with reference to this figure.
  • the number NumMergeCandidates of peripheral partitions adjacent to the partition to be processed is calculated (S401). Since the calculation procedure is the same as S101 in encoding, it is omitted.
  • the motion information selection unit 210 outputs the motion information decoded from the bitstream as in the conventional case.
  • merge_flag decoded by the variable length decoding unit 201 is read (S403), and selection of motion information is determined based on merge_flag (S404).
  • merge_flag false (0)
  • the motion information selection unit 210 outputs motion information decoded from the bitstream as in the conventional case.
  • the motion information of the reference destination partition is acquired from the memory 207 in which the decoded information is recorded (S407).
  • the acquired motion information is supplied to the motion compensation unit 203, and motion compensation is performed based on the motion information.
  • the motion information of the partitions adjacent to the left and above the partition to be processed is read from the memory 207 and acquired (S406).
  • the obtained motion information of the left and upper adjacent partitions is compared (S408). Specifically, the prediction direction, the reference picture number, and the motion vector constituting the motion information are compared. If the motion information is exactly the same, the motion information of either the left or upper adjacent partition is acquired and supplied to the motion compensation unit 203 (S410).
  • the motion information of the partition adjacent to the left is selected as in the encoding.
  • the process proceeds to prediction of the reference destination partition (S409).
  • a motion information reference destination prediction method which will be described later, is used as in the case of encoding.
  • the target partition is predicted based on information such as the size of the partition adjacent to the processing target partition and the left and top of the periphery, and the length of the side where the processing target partition and the adjacent partition are in contact with each other.
  • the flag pred_dirc_flag indicating the referred reference partition is output.
  • the flag pred_dirc_flag indicating the predicted reference destination partition is represented by left (1) or upper (0) when the adjacent reference destination partition is selected as in the encoding, and the reference destination partition can be selected by prediction. If not, undefined (2) is set.
  • pred_direc_flag output is determined (S411).
  • the case where pred_dirc_flag is 2 is a case where prediction is impossible due to prediction of the reference destination partition, and at this time, the process proceeds to case 1 (S412).
  • pred_dirc_flag is not 2
  • the process proceeds to case 2 (S413).
  • FIG. 13A shows a flowchart of the process of case 1 in step S412.
  • the prediction result is undefined in the prediction of the reference destination partition, and the prediction result of the reference destination is not shown. Therefore, the partition adjacent to the left or above indicated by merge_proble_flag is selected as the reference destination partition as in the conventional case. This is because, in the encoding side, merge_direct_flag indicating the actual reference destination partition is set in merge_proble_flag.
  • the merge_proble_flag decoded by the variable length decoding unit 201 is read (S421), and a reference destination partition is selected based on the merge_proble_flag (S422).
  • pred_dirc_flag 1
  • the left adjacent partition S423
  • merge_proble_flag 0
  • the upper adjacent partition is selected as the reference destination partition (S424).
  • FIG. 13B shows a flowchart of the process of case 2 in step S413.
  • the probability of the flag pred_direc_flag indicating the predicted reference destination partition is represented by the flag merge_proble_flag indicating the reference destination validity. That is, merge_proble_flag is represented by 1 (true) if the reference destination partition indicated by pred_dirc_flag is correct, and merge_proble_flag is represented by 0 (false) if not correct.
  • the merge_proble_flag decoded by the variable length decoding unit 201 is read (S431), and the selection of the reference destination partition is determined based on the merge_proble_flag (S432).
  • merge_proble_flag 1
  • a partition having a value indicated by pred_dirc_flag determined by prediction of the reference destination partition is selected as the reference destination partition (S433).
  • merge_proble_flag 0
  • a partition opposite to the value indicated by pred_direc_flag is selected as a reference destination partition. For example, if pred_dirc_flag is 0 (left), the upper adjacent partition is set as the reference destination partition (S434).
  • the motion information of the reference destination partition selected in the processing of case 1 and case 2 is acquired (S415), supplied to the motion compensation unit 203, and the process ends.
  • the reference is made by the motion information reference destination prediction method described later. Select the destination partition.
  • a decoded image is generated by adding the predicted image generated by motion compensation from the motion information of the selected reference destination partition and the residual signal decoded from the bitstream.
  • FIG. 14 is a flowchart for explaining a motion information reference destination prediction method according to the present embodiment.
  • the process shown in FIG. 14 is performed in units of partitions.
  • a partition that refers to motion information is determined for the partition to be processed, and if a partition is determined as a reference destination, the process does not proceed to the next process and ends.
  • each process will be described step by step.
  • FIG. 15 is a flowchart showing a boundary determination process with adjacent surrounding partitions in step S500. Details of the processing will be described with reference to FIG.
  • motion boundary detection is performed between a partition to be processed and neighboring partitions around it (S501).
  • FIG. 16 shows an arrangement example in which the processing target partition is X and the left and upper adjacent partitions are A and B, respectively.
  • the rectangular size of the partition to be processed is represented by width wx and height hx
  • the height of the adjacent left partition is represented by pa
  • the width of the adjacent upper partition is represented by pb.
  • the boundary between the processing target partition and the left and upper adjacent partitions is determined by comparing the height hx of the processing target partition with the height pa of the adjacent left partition and the width wx of the processing target partition. This is done by comparing the width pb of the upper partition.
  • an edge where an adjacent partition is adjacent to the target partition is called a “motion boundary”.
  • FIG. 18 is a flowchart showing the adjacent side length comparison process with the adjacent partition in step S510. Details of the processing will be described with reference to FIG.
  • the length of the adjacent side in contact with the left or upper partition adjacent to the processing target partition is calculated (S511).
  • the adjacent side length to the left partition A is represented by L (A)
  • the adjacent side length to the upper partition B is represented by L (B) (the hatched portion in FIG. 16 is adjacent).
  • the adjacent side length L (A) with the left partition A is the smaller of the height hx of the partition X to be processed and the height pa of the left partition, and is expressed by the following equation.
  • L (A) min (pa, hx)
  • the function min (a, b) is a function for selecting the smaller one of a and b.
  • the adjacent side length L (B) with the upper partition B is the smaller one of the width wx of the partition X to be processed and the height pb of the upper partition, and is expressed by the following equation.
  • L (B) min (pb, wx)
  • pa and pb are the lengths of the motion boundaries obtained in step S500.
  • the obtained adjacent side lengths L (A) and L (B) are compared, and a reference destination partition is selected (S512). If L (A) and L (B) are equal, the reference destination partition cannot be determined, so the reference destination partition is not selected and is undefined (S513), and the process ends. If L (A) and L (B) are not equal, L (A) and L (B) are compared in size (S514). If L (A) is greater than L (B), A is selected (S515) and the process ends. If L (A) is smaller than L (B), B is selected (S516) and the process ends.
  • the determination results classified according to the motion boundary and side length conditions of the left and upper partitions adjacent to the partition to be processed are as follows.
  • the summary is shown in the table of FIG.
  • the selection column in the table of FIG. 19 indicates the partition to be referred to, and the symbol “-” represents the case where the reference partition is indefinite.
  • the fifth boundary comparison condition “pb ⁇ wx and pa ⁇ hx” in FIG. 19 is arranged as shown in FIG. 20, and the lengths of the motion boundaries of adjacent partitions A and B are both subject to processing. This is the case when the width and height of the partition do not match.
  • the reference destination is determined by comparing the lengths of adjacent side lengths in contact with the partition to be processed.
  • the adjacent side lengths L (A) and L (B) of the adjacent partitions A and B are the sides pa and pb that are in contact with the partition X to be processed, so that the reference partition is determined by comparing pa and pb. Determined.
  • FIG. 21 shows an example of the processing target and the arrangement of adjacent partitions under the boundary comparison condition “pb> wx and pa> hx”. As shown in FIG. 4, since the size and shape of the partition inside the prediction block are defined, the size and shape of the partition meeting the boundary comparison condition “pb> wx and pa> hx” shown in FIG. Limited.
  • the prediction result of the determined reference destination partition is output as a flag pred_dirc_flag indicating the reference destination partition predicted in the encoding and decoding processes.
  • pred_direc_flag is the same as merge_direc_flag, when it is selected as a reference destination adjacent partition, “1 (left)” or “0 (upper)” indicates the reference destination, and the reference destination partition cannot be selected by prediction. Is set to “2 (undefined)”.
  • the occurrence of the bit represented by the flag merge_proble_flag used for designating the reference destination is biased and generated by predicting the reference block in the merge process.
  • the amount of codes can be reduced, and the encoding efficiency can be improved.
  • the reference destination flag has a frequency of 1 ⁇ 2 and the arithmetic coding cannot be used.
  • the predicted reference is By using the flag merge_proble_flag indicating the probability of the previous, if the prediction is correct, the bit frequency of the flag can be biased, and the amount of generated codes can be reduced in arithmetic coding.
  • the adjacent block that matches the size of the encoding target block, or the adjacent block having a longer side length is highly correlated with the encoding target block, so that the reference block It is possible to improve the accuracy of prediction by predicting.
  • Example 2 A second example of the motion information reference destination prediction method according to the embodiment of the present invention will be described. The difference from the first embodiment is that a determination process based on the residual signal of the left or upper partition adjacent to the processing target partition is added as a determination condition for the reference destination prediction.
  • FIG. 22 is a flowchart for explaining the operation of the second embodiment of the motion information reference destination prediction method.
  • the processes of steps S600 and S610 shown in FIG. 22 are the same as the processes of steps S500 and S510 of the first embodiment shown in FIG.
  • a boundary determination is made with the surrounding partition adjacent to the left or upper side of the partition to be processed (S600), and the length of the common part of adjacent sides of the left or upper partition adjacent to the processing target partition ("adjacent" The side length ") is determined (S610).
  • FIG. 23 is a diagram illustrating a pixel region for calculating a boundary residual between adjacent partitions.
  • A be the left partition adjacent to partition X to be processed
  • B be the upper partition.
  • Each partition A, B is assumed to be composed of a pixel residual signal after motion compensation prediction in the encoding process or a pixel residual signal after inverse quantization and inverse orthogonal transformation in the decoding process.
  • the pixel area of the partition A that is in contact with the partition X is represented by an area having a height of min (ha, hx) and a width of 1 pixel located at the left boundary of the partition X, and a residual signal of the pixel.
  • ⁇ (A) is expressed by the following equation, where ⁇ (A) is the sum of absolute values of.
  • ai represents the residual signal of the pixel in partition A adjacent to partition X.
  • the number of pixels in the partition A adjacent to the partition X is the smaller of the number of pixels ha representing the height of the partition A and the height hx of the partition X, and is represented by a function min (ha, hx).
  • the pixel area of the partition B that is in contact with the partition X is represented by an area having a width of min (wb, wx) and a height of 1 pixel located at the upper boundary of the partition X. If the sum of absolute values of residual signals is ⁇ (B), ⁇ (B) is expressed by the following equation.
  • bi represents the residual signal of the pixel in partition B adjacent to partition X.
  • the number of pixels in the partition B adjacent to the partition X is the smaller of the number of pixels wb representing the width of the partition B and the width wx of the partition X, and is represented by a function min (wb, wx).
  • FIG. 24 is a flowchart for explaining the operation of determining the reference destination partition based on the boundary residual between adjacent partitions.
  • normalization of the calculated boundary residual is performed (S622). As shown in FIG. 23, when the adjacent side lengths of the partitions A and B adjacent to the partition X are different, it is not appropriate to directly compare the boundary residual of the partition A and the boundary residual of the partition B. As shown in the following equation, normalization is performed by dividing the boundary residuals of the partitions A and B by the adjacent side lengths of the partitions A and B, respectively.
  • the process proceeds to calculation of the residual ratio (S627).
  • the residual ratios ⁇ (A) and ⁇ (B) of the partitions A and B are the boundary residuals ⁇ (A) and ⁇ (B) before normalization of the partitions A and B calculated in step S621. ) To the total residual of partitions A and B is expressed by the following equation.
  • Ai and Bi are residual signals of pixels in partitions A and B adjacent to partition X
  • nA and nB are the total number of pixels in partitions A and B.
  • an area that is not adjacent to the partition X to be processed is not a target.
  • the number of pixels wa representing the width of the partition A is compared with the number of pixels hb representing the height of the partition B, and the smaller one is defined as the width of the target region.
  • the number of pixels nB in the target area of partition B is given by the following equation.
  • nA is a rectangular area indicated by light gray in the thick dotted line in the partition A of FIG. 23, and the denominator of the calculation formula of ⁇ (A) is the absolute value of the residual signal of the pixel in this rectangular area. Expressed as the sum.
  • nB covers the entire area of partition B in FIG. 23, and the denominator part of the calculation formula of ⁇ (B) is expressed as the sum of absolute values of residual signals of pixels in partition B.
  • the boundary residuals ⁇ (A) and ⁇ (B) before normalization are sums of absolute values of residual signals of pixels adjacent to the partition X, respectively, and the number of pixels adjacent to the partition X is min (ha, hx).
  • the residual ratios ⁇ (A) and ⁇ (B) are calculated with the same number of pixel ratios, and ⁇ (A) and ⁇ (B) are the partition X Represents the ratio of the residual signals of adjacent pixels to the residual signals of the pixels in partitions A and B. If this ratio is high, the motion compensation prediction near the pixel adjacent to the partition X is not applied, that is, the motion information used for the motion compensation prediction is not appropriate for the prediction near the pixel adjacent to the partition X. Means that. That is, it is determined that the motion information of the adjacent partition having the higher residual ratio is not referred to.
  • the residual ratio calculated in this way is compared. It is determined whether or not ⁇ (A) and ⁇ (B) are equal (S628). If ⁇ (A) and ⁇ (B) are not equal, ⁇ (A) and ⁇ (B) have a magnitude relationship, so a magnitude comparison is performed. Here, it is determined whether or not ⁇ (A) is smaller than ⁇ (B) (S629). If ⁇ (A) is smaller than ⁇ (B), A is selected (S630). Otherwise, B is selected (S631). If ⁇ (A) and ⁇ (B) are equal, there is no difference in the residual that occurs regardless of which one is selected, so A is selected (S632), and the process ends.
  • pred_direc_flag indicates a reference destination with “1 (left)” or “0 (upper)” when it is selected as an adjacent partition to be referred to.
  • the reason for selecting A in step S632 is that A (left) is assigned to bit "1" in this embodiment. This is to reduce the amount of generated code in encoding by biasing the occurrence frequency of pred_direc_flag to “1” as prediction of a reference destination partition of motion information.
  • step S623 shown in FIG. 24 strict determination as to whether or not the boundary residuals ⁇ (A) and ⁇ (B) are equal is performed using ⁇ (A) and ⁇ (B The determination may be changed based on whether or not the absolute value difference
  • may be a fixed value or may be a variable value that is changed according to the value of the quantization parameter Qp, for example.
  • the difference determination of the boundary residual and the residual ratio is performed in order, but the boundary residual and the residual ratio are determined separately, and the final reference destination partition is selected from the results. You may do it.
  • the boundary residual and the residual ratio may be collected and a new evaluation criterion may be generated and determined. For example, the determination may be made based on the evaluation value generated by the combination of the boundary residual and the residual ratio as in the following equation.
  • variable x is A or B
  • is a weighting coefficient.
  • may be a fixed value, or may be set as a variable value according to a partition, a slice unit, a slice type, or the like.
  • prediction of a reference destination is performed by predicting a reference block based on the magnitude of a residual signal of a block adjacent to the block to be encoded. Accuracy can be further increased.
  • Example 3 A third example of the motion information reference destination prediction method according to the embodiment of the present invention will be described.
  • the difference from the first embodiment is that the partition represented by the flag pred_direc_flag determined and output by prediction of the reference destination partition is selected as the reference destination partition.
  • the flag pred_direc_flag based on prediction of the reference destination partition is 0 or 1
  • the left (pred_direc_flag is “1”) or the top (pred_dirc_flag is “0”) is selected as the reference destination adjacent partition.
  • FIG. 26 is a flowchart for explaining a procedure for setting a flag for merge processing.
  • the prediction method selected by the prediction selection unit 112 is prediction based on motion information selected in the merge process (S701).
  • merge_flag is set to 0 and the process ends (S703). If the prediction is based on the motion information selected in the merge process, merge_flag is set to 1 (S702).
  • the number of motion information selected by the merge process is checked (S704). It is determined whether or not the number of adjacent partition candidates NumMergeCandidates to be specified by the reference destination specifying unit 113 is greater than one. When the number of candidates NumMergeCandidates is 1 or less, it is assumed that merge processing is selected, and there is always one adjacent partition of a valid reference destination, and either the left or the upper can be specified. Therefore, only merge_flag is set and the process ends.
  • the process proceeds to prediction of a reference destination partition (S705).
  • the motion information reference destination prediction method described in the first embodiment is used. Based on information such as the size of the partition adjacent to the left and top of the target partition and the length of the side where the target partition and the adjacent partition are in contact with each other, the reference destination partition is predicted, and the flag pred_dirc_flag is set. Is output.
  • the flag pred_dirc_flag indicating the predicted reference destination partition represents the selection destination on the left (1) or above (0) as the adjacent partition of the reference destination, and the reference destination partition cannot be selected by prediction. Indeterminate (2) is set.
  • the flag pred_direc_flag indicating the predicted reference destination partition set in this way is determined (S706).
  • the flag pred_dirc_flag indicating the predicted reference destination partition is 2 when prediction is impossible in the prediction of the reference destination partition, and at this time, it depends on the merge process selected as the prediction method that actually minimizes the generated code amount.
  • the process sets the merge_dir_flag indicating the reference partition of the motion information used in the motion compensation prediction, and the process ends (S707).
  • pred_direc_flag is not 2
  • pred_direc_flag directly indicates a reference destination partition, and when “0” is selected, a partition adjacent to the left is selected.
  • the first flag inter_merge_flag described in the header of the SPS is described in the prediction block.
  • the second flag merge_flag and the third flag merge_direc_flag are set and encoded.
  • the motion information reference destination prediction method according to the present embodiment is used for the entire bitstream sequence. It is determined whether inter-screen prediction by merge processing is applied. When inter_merge_flag is true (1), merge processing is applied. When inter_merge_flag is false (0), merge processing is ignored and inter-screen prediction based on motion information decoded from a conventional bitstream is performed. . In the following description, it is assumed that inter_merge_flag is true (1), that is, the merge processing using the motion information reference destination prediction method according to the present embodiment is applied.
  • the intra-screen / inter-screen prediction selection unit 209 determines whether to select intra-screen or inter-screen prediction with reference to the prediction mode for each prediction block in the slice.
  • the intra prediction MODE_INTRA
  • the process proceeds to the intra prediction unit 208
  • the inter prediction MODE_INTER
  • the process proceeds to the motion information selection unit 210.
  • the motion information selection unit 210 uses, as information decoded from the bit stream for each partition of the prediction block, the motion information detected by the motion vector detection of encoding or the reference destination partition selected by the merge process. Either one of the motion information is selected, and the selected motion information is output to the motion compensation unit 203.
  • FIG. 27 is a flowchart showing the selection procedure of the motion information selection unit 210, which will be described with reference to this figure.
  • the number NumMergeCandidates of neighboring partitions adjacent to the partition to be processed is calculated (S801). Since the calculation procedure is the same as S701 in encoding, it is omitted.
  • the motion information selection unit 210 outputs the motion information decoded from the bitstream as in the conventional case.
  • merge_flag decoded by the variable length decoding unit 201 is read (S803), and the selection of motion information is determined based on the merge_flag (S804).
  • merge_flag is false (0), the motion information selection unit 210 outputs motion information decoded from the bitstream as in the conventional case.
  • merge_flag is true (1), the process proceeds to processing for selecting motion information of the reference partition in the merge processing.
  • the motion information of the reference destination partition is acquired from the memory 207 in which the decoded information is recorded (S807).
  • the acquired motion information is supplied to the motion compensation unit 203, and motion compensation is performed based on the motion information.
  • the motion information of the partitions adjacent to the left and above the processing target partition is read from the memory 207 and acquired (S806).
  • the obtained motion information of the left and upper adjacent partitions is compared (S808). Specifically, the prediction direction, the reference picture number, and the motion vector constituting the motion information are compared. If the motion information is exactly the same, the motion information of either the left or upper adjacent partition is acquired and output to the motion compensation unit 203 (S810).
  • the motion information of the partition adjacent to the left is selected as in the encoding.
  • the process proceeds to prediction of the reference destination partition (S809).
  • the motion information reference destination prediction method described in the first embodiment is used, similarly to the encoding.
  • the target partition is predicted based on information such as the size of the partition adjacent to the processing target partition and the left and top of the periphery, and the length of the side where the processing target partition and the adjacent partition are in contact with each other.
  • the flag pred_dirc_flag indicating the referred reference partition is output.
  • the flag pred_dirc_flag indicating the predicted reference destination partition is represented by left (1) or upper (0) when the adjacent reference destination partition is selected as in the encoding, and the reference destination partition can be selected by prediction. If not, undefined (2) is set.
  • the pred_direc_flag output is determined (S811).
  • the pred_dirc_flag is 2 when prediction is impossible due to prediction of the reference destination partition, and the reference destination partition cannot be determined. Therefore, merge_direc_flag decoded by the variable length decoding unit 201 is read (S812). Based on merge_direc_flag, the selection of the reference destination partition is determined (S813). If merge_dirc_flag is 1, the left adjacent partition (S814) is selected. If merge_dirc_flag is 0, the upper adjacent partition is selected as the reference destination partition (S815).
  • pred_direc_flag when pred_direc_flag is not 2, the value indicated by pred_direc_flag directly indicates the adjacent partition of the reference destination. That is, if pred_dirc_flag is 1, the left adjacent partition is selected, and if merge_proble_flag is 0, the upper adjacent partition is selected as a reference destination partition. The motion information of the reference destination partition thus selected is acquired, supplied to the motion compensation unit 203, and the process ends.
  • the reference is made by the above-described motion information reference destination prediction method. Select the destination partition.
  • a decoded image is generated by adding the predicted image generated by motion compensation from the motion information of the selected reference destination partition and the residual signal decoded from the bitstream.
  • the determination processing based on the residual signal of the left or upper partition adjacent to the partition to be processed described in the second embodiment can be applied in the same manner.
  • the third embodiment it is obtained by prediction of reference destination without defining merge_proble_flag itself, assuming that the matching accuracy of the determination of the adjacent block that is the reference destination by the prediction of the reference destination is high.
  • the reference destination neighboring block may be indicated by the determination result, and the reference destination neighboring block may be indicated by the flag merge_dirc_flag only when the prediction of the reference destination is indefinite.
  • the increase in the code amount due to merge_direct_flag can be offset by the reduction in the code amount required for merge_proble_flag, and the encoding efficiency can be relatively improved.
  • the left or upper adjacent partition is always designated as the reference destination partition in the prediction of the reference destination partition, so that it is not determined that the reference destination partition is indeterminate by the prediction of the reference destination partition. May be.
  • it is determined as indefinite in the prediction of the reference destination partition it is predetermined in advance that the adjacent partition on the left or the upper side is uniquely specified. Since it is not necessary to encode and transmit the flag merge_dirc_flag indicating the reference destination partition when it is determined indefinitely by partition prediction, it is possible to improve the encoding efficiency.
  • the reference destination prediction method of motion information in decoding corresponding to encoding the same reference destination partition as that on the encoding side is predicted, and the left or upper adjacent partition determined by prediction of the reference destination partition is used as the reference destination partition. select.
  • the moving image encoded stream output from the moving image encoding apparatus of the embodiment described above has a specific data format so that it can be decoded according to the encoding method used in the embodiment. Therefore, the moving picture decoding apparatus corresponding to the moving picture encoding apparatus can decode the encoded stream of this specific data format.
  • the encoded stream When a wired or wireless network is used to exchange an encoded stream between a moving image encoding device and a moving image decoding device, the encoded stream is converted into a data format suitable for the transmission form of the communication path. It may be transmitted.
  • a video transmission apparatus that converts the encoded stream output from the video encoding apparatus into encoded data in a data format suitable for the transmission form of the communication channel and transmits the encoded data to the network, and receives the encoded data from the network Then, a moving image receiving apparatus that restores the encoded stream and supplies the encoded stream to the moving image decoding apparatus is provided.
  • the moving image transmitting apparatus is a memory that buffers the encoded stream output from the moving image encoding apparatus, a packet processing unit that packetizes the encoded stream, and transmission that transmits the packetized encoded data via the network.
  • the moving image receiving apparatus generates a coded stream by packetizing the received data, a receiving unit that receives the packetized coded data via a network, a memory that buffers the received coded data, and packet processing. And a packet processing unit provided to the video decoding device.
  • the above processing relating to encoding and decoding can be realized as a transmission, storage, and reception device using hardware, and is also stored in a ROM (Read Only Memory), a flash memory, or the like. It can also be realized by firmware or software such as a computer.
  • the firmware program and software program can be provided by recording them on a computer-readable recording medium, provided from a server through a wired or wireless network, or provided as a data broadcast of terrestrial or satellite digital broadcasting. Is also possible.
  • the present invention can be used for moving picture encoding and decoding techniques.

Landscapes

  • Engineering & Computer Science (AREA)
  • Multimedia (AREA)
  • Signal Processing (AREA)
  • Compression Or Coding Systems Of Tv Signals (AREA)

Abstract

Selon l'invention, une unité de spécification de destination de référence (113) spécifie, parmi une pluralité de blocs adjacents déjà codés adjacents à un bloc à coder, n'importe lequel des blocs adjacents, déterminés sur la base de la condition d'adjacence entre le bloc à coder et les blocs adjacents, à titre de bloc destination de référence utilisé pour se rapporter à des informations de mouvement. Lorsque, parmi une prédiction intra-image, une prédiction inter-image à l'aide du vecteur de mouvement du bloc à coder, et une autre prédiction inter-image à l'aide du vecteur de mouvement d'un bloc adjacent, l'autre prédiction inter-image est sélectionnée à titre de procédé de prédiction, une unité de sélection de prédiction (112) génère, à la place des informations de mouvement, des informations de validité de destination de référence indiquant si le bloc adjacent auquel se rapporte le procédé de prédiction sélectionné et le bloc destination de référence spécifié par l'unité de spécification de destination de référence (113) concordent ou non, et des informations indiquant le procédé de prédiction sélectionné.
PCT/JP2012/002649 2011-04-28 2012-04-17 Dispositif de codage d'image animée, procédé de codage d'image animée, programme de codage d'image animée, dispositif de décodage d'image animée, procédé de décodage d'image animée et programme de décodage d'image animée WO2012147306A1 (fr)

Applications Claiming Priority (4)

Application Number Priority Date Filing Date Title
JP2011101951A JP5703942B2 (ja) 2011-04-28 2011-04-28 動画像符号化装置、動画像符号化方法及び動画像符号化プログラム
JP2011-101951 2011-04-28
JP2011-101952 2011-04-28
JP2011101952A JP5703943B2 (ja) 2011-04-28 2011-04-28 動画像復号装置、動画像復号方法及び動画像復号プログラム

Publications (1)

Publication Number Publication Date
WO2012147306A1 true WO2012147306A1 (fr) 2012-11-01

Family

ID=47071839

Family Applications (1)

Application Number Title Priority Date Filing Date
PCT/JP2012/002649 WO2012147306A1 (fr) 2011-04-28 2012-04-17 Dispositif de codage d'image animée, procédé de codage d'image animée, programme de codage d'image animée, dispositif de décodage d'image animée, procédé de décodage d'image animée et programme de décodage d'image animée

Country Status (1)

Country Link
WO (1) WO2012147306A1 (fr)

Cited By (1)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN107431806A (zh) * 2015-03-19 2017-12-01 Lg 电子株式会社 用于处理视频信号的方法及其设备

Citations (2)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
JP2010524397A (ja) * 2007-04-12 2010-07-15 トムソン ライセンシング ビデオ・エンコーディング、およびデコーディングのためのスキップ/ダイレクトモードのコンテクストに依存するマージのための方法と装置
JP2010529811A (ja) * 2007-06-12 2010-08-26 トムソン ライセンシング スライスデータ用のマルチパスビデオのシンタックス構造をサポートする方法及び装置

Patent Citations (2)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
JP2010524397A (ja) * 2007-04-12 2010-07-15 トムソン ライセンシング ビデオ・エンコーディング、およびデコーディングのためのスキップ/ダイレクトモードのコンテクストに依存するマージのための方法と装置
JP2010529811A (ja) * 2007-06-12 2010-08-26 トムソン ライセンシング スライスデータ用のマルチパスビデオのシンタックス構造をサポートする方法及び装置

Non-Patent Citations (2)

* Cited by examiner, † Cited by third party
Title
MARTIN WINKEN ET AL.: "Description of video coding technology proposal by Fraunhofer HHI", JOINT COLLABORATIVE TEAM ON VIDEO CODING (JCT-VC) OF ITU-T SG16 WP3 AND ISO/IEC JTC1/ SC29/WG11 1ST MEETING, 15 April 2010 (2010-04-15), DRESDEN, DE, pages 9 - 10 *
SHAO LI ET AL.: "Simplification and improvement of merge mode coding", JOINT COLLABORATIVE TEAM ON VIDEO CODING (JCT-VC) OF ITU-T SG16 WP3 AND ISO/IEC JTC1/SC29/WG11 5TH MEETING, 16 March 2011 (2011-03-16), GENEVA, CH *

Cited By (2)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN107431806A (zh) * 2015-03-19 2017-12-01 Lg 电子株式会社 用于处理视频信号的方法及其设备
CN107431806B (zh) * 2015-03-19 2020-07-31 Lg 电子株式会社 用于处理视频信号的方法及其设备

Similar Documents

Publication Publication Date Title
KR101949818B1 (ko) 영상 복호화 방법 및 컴퓨터로 읽을 수 있는 기록 매체
US9053544B2 (en) Methods and apparatuses for encoding/decoding high resolution images
US20240146959A1 (en) Image encoding method using a skip mode, and a device using the method
KR101336475B1 (ko) 동화상 부호화 장치 및 동화상 복호 장치
KR101641400B1 (ko) 복호 장치 및 방법
EP2863637B1 (fr) Dispositif de décodage vidéo, procédé de décodage vidéo, programme de décodage vidéo
WO2010001917A1 (fr) Dispositif et procédé de traitement d'image
JP2010135864A (ja) 画像符号化方法及び装置並びに画像復号化方法及び装置
KR20110073263A (ko) 인트라 예측 부호화 방법 및 부호화 방법, 그리고 상기 방법을 수행하는 인트라 예측 부호화 장치 및 인트라 예측 복호화 장치
JP7311635B2 (ja) コーデック方法、デバイス及びその機器
JP2012142702A (ja) 画像処理装置および方法、並びにプログラム
WO2012042654A1 (fr) Procédé de décodage d'image, procédé de codage d'image, dispositif de décodage d'image, dispositif de codage d'image, programme de décodage d'image et programme de codage d'image
JP6825506B2 (ja) 動画像符号化装置、動画像符号化方法、動画像符号化用コンピュータプログラム、動画像復号装置及び動画像復号方法ならびに動画像復号用コンピュータプログラム
JP2023138826A (ja) イントラ予測装置、画像符号化装置、画像復号装置、及びプログラム
KR101603887B1 (ko) Hevc에서의 고속 cu 결정을 통한 부호화방법
JP5703943B2 (ja) 動画像復号装置、動画像復号方法及び動画像復号プログラム
WO2012147306A1 (fr) Dispositif de codage d'image animée, procédé de codage d'image animée, programme de codage d'image animée, dispositif de décodage d'image animée, procédé de décodage d'image animée et programme de décodage d'image animée
JP5703942B2 (ja) 動画像符号化装置、動画像符号化方法及び動画像符号化プログラム
JP2009049969A (ja) 動画像符号化装置及び方法並びに動画像復号化装置及び方法
KR20140124920A (ko) 움직임 벡터 예측을 이용한 비디오 부호화/복호화 방법 및 장치
JP2013168932A (ja) 画像復号装置、画像復号方法及び画像復号プログラム
JP2010034946A (ja) 動画像符号化装置および復号装置
WO2013108634A1 (fr) Dispositif de codage d'image, procédé de codage d'image, programme de codage d'image, dispositif de décodage d'image, procédé de décodage d'image et programme de décodage d'image
KR20140124437A (ko) 움직임 정보 부호화/복호화 방법 및 장치
JP2014239504A (ja) 復号化装置

Legal Events

Date Code Title Description
121 Ep: the epo has been informed by wipo that ep was designated in this application

Ref document number: 12776108

Country of ref document: EP

Kind code of ref document: A1

NENP Non-entry into the national phase

Ref country code: DE

122 Ep: pct application non-entry in european phase

Ref document number: 12776108

Country of ref document: EP

Kind code of ref document: A1