WO2012147306A1 - Moving picture encoding device, moving picture encoding method, moving picture encoding program, moving picture decoding device, moving picture decoding method, and moving picture decoding program - Google Patents

Moving picture encoding device, moving picture encoding method, moving picture encoding program, moving picture decoding device, moving picture decoding method, and moving picture decoding program Download PDF

Info

Publication number
WO2012147306A1
WO2012147306A1 PCT/JP2012/002649 JP2012002649W WO2012147306A1 WO 2012147306 A1 WO2012147306 A1 WO 2012147306A1 JP 2012002649 W JP2012002649 W JP 2012002649W WO 2012147306 A1 WO2012147306 A1 WO 2012147306A1
Authority
WO
WIPO (PCT)
Prior art keywords
block
adjacent
prediction
reference destination
partition
Prior art date
Application number
PCT/JP2012/002649
Other languages
French (fr)
Japanese (ja)
Inventor
福島 茂
西谷 勝義
上田 基晴
徹 熊倉
Original Assignee
株式会社Jvcケンウッド
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Priority claimed from JP2011101951A external-priority patent/JP5703942B2/en
Priority claimed from JP2011101952A external-priority patent/JP5703943B2/en
Application filed by 株式会社Jvcケンウッド filed Critical 株式会社Jvcケンウッド
Publication of WO2012147306A1 publication Critical patent/WO2012147306A1/en

Links

Images

Classifications

    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04NPICTORIAL COMMUNICATION, e.g. TELEVISION
    • H04N19/00Methods or arrangements for coding, decoding, compressing or decompressing digital video signals
    • H04N19/46Embedding additional information in the video signal during the compression process
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04NPICTORIAL COMMUNICATION, e.g. TELEVISION
    • H04N19/00Methods or arrangements for coding, decoding, compressing or decompressing digital video signals
    • H04N19/50Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using predictive coding
    • H04N19/503Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using predictive coding involving temporal prediction
    • H04N19/51Motion estimation or motion compensation
    • H04N19/513Processing of motion vectors
    • H04N19/517Processing of motion vectors by encoding
    • H04N19/52Processing of motion vectors by encoding by predictive encoding

Definitions

  • the present invention relates to a moving picture encoding and decoding technique, and more particularly to a moving picture encoding and decoding technique using motion compensation prediction.
  • the moving picture coding method that divides a picture into rectangular blocks, represented by MPEG (Moving Picture Experts ⁇ Group), and performs motion estimation and compensation between blocks, the coding amount of the motion vector generated in each block
  • MPEG Motion Picture Experts ⁇ Group
  • a prediction process is performed on the motion vector.
  • the motion vector detected in units of macroblocks is taken as a difference from the motion vector of the macroblock encoded immediately before, and the amount of code is reduced by encoding the difference vector. Yes.
  • MPEG-4 AVC / H. H.264 uses the fact that motion vectors have a strong correlation with the motion vectors of neighboring blocks, and performs prediction from neighboring blocks and encodes the difference vector to reduce the amount of code. .
  • the median is calculated from the motion vectors of adjacent blocks on the left, top and top right of the block to be processed, and motion vector prediction is realized by taking the difference from the median.
  • the motion information of the block to be processed and the neighboring already-encoded neighboring blocks are the same in the standard work of moving picture coding in Patent Document 1 and recent ISO / IEC and ITU-T. If so, the block to be processed does not encode its own motion information, uses the motion information of the adjacent block for encoding, and encodes additional information specifying the adjacent block having the motion information to be referenced. Thus, attempts have been made to reduce the amount of code of motion information. Such a method is called “merge” and has attracted attention as a method for reducing the code amount of motion information.
  • a 1-bit flag (whether to use the motion information of the block to be processed itself detected by motion vector detection or to use the motion information of an adjacent block by the merge method as additional information for each block)
  • merge_lu_flag a 1-bit flag that defines whether to refer to the left or right of neighboring blocks when the merge method is applied. Entropy coding using arithmetic codes is performed on these flags.
  • a short code is assigned to data having a high appearance rate
  • a long code is assigned to data having a low appearance rate
  • data compression is performed to reduce the code amount.
  • the flag merge_lu_flag depends on the texture in the screen, since a code is assigned in advance to select the left or top as the reference destination adjacent block, even if the occurrence frequency of the reference destination selection is biased Therefore, there is a problem that the code amount cannot be sufficiently reduced because the encoding is performed with equal probability.
  • the present invention has been made in view of such circumstances, and an object of the present invention is to provide a technique for improving the coding efficiency by reducing the code amount of a flag representing a reference destination of motion information.
  • a moving image encoding device is a moving image encoding device that encodes the moving image using a motion vector in units of blocks obtained by dividing each picture of the moving image. Any one of the adjacent blocks determined based on an adjacency condition between the encoding target block and the adjacent block is moved from among a plurality of encoded adjacent blocks adjacent to the encoding target block.
  • a reference destination designating unit (113) that designates a reference block for referring to information, intra-picture prediction, inter-picture prediction using a motion vector of an encoding target block, and inter-picture using a motion vector of an adjacent block
  • the inter-picture prediction using the motion vector of the adjacent block is selected as the prediction method from the prediction, it is referred to by the selected prediction method.
  • a prediction selection unit that generates, instead of motion information, reference destination valid information indicating whether or not a contact block matches the reference destination block specified by the reference destination specifying unit, and information indicating the selected prediction method (112).
  • This method is a moving image encoding method for encoding the moving image using a motion vector in units of blocks obtained by dividing each picture of the moving image, and a plurality of encoded adjacent regions adjacent to the encoding target block.
  • a moving picture decoding apparatus is a moving picture decoding apparatus that decodes a bitstream in which the moving picture is encoded using a motion vector in units of blocks obtained by dividing each picture of the moving picture.
  • a variable for decoding the information indicating the prediction method and the reference destination valid information from the bitstream including information indicating the method and reference destination valid information indicating the validity of the adjacent block referred to by the selected prediction method Using the long decoding unit (201), intra-picture prediction, inter-picture prediction using a motion vector of a decoding target block, and one motion vector of a plurality of decoded adjacent blocks adjacent to the decoding target block
  • a prediction selection unit (209) that selects one of prediction methods based on information indicating the prediction method from among the inter-image predictions; and the prediction selection Therefore, when inter-picture prediction using a motion vector of an adjacent block is selected as a prediction method, any one adjacent block determined based on an adjacent condition between the decoding target block and the adjacent block is moved.
  • a motion information selection unit (210) that determines
  • Another aspect of the present invention is a video decoding method.
  • This method is a moving image decoding method for decoding a bitstream in which the moving image is encoded using a motion vector in units of blocks obtained by dividing each picture of the moving image, and is selected as information indicating a prediction method
  • a variable length decoding step for decoding the information indicating the prediction method and the reference destination valid information from the bitstream including the reference destination valid information indicating the validity of the adjacent block referred to by the prediction method, and intra prediction
  • the prediction method includes: inter-picture prediction using a motion vector of a decoding target block; and inter-picture prediction using any one of a plurality of decoded adjacent blocks adjacent to the decoding target block.
  • a prediction selection step of selecting one of the prediction methods based on the information indicated, and the prediction selection step When inter-picture prediction using a motion vector of a block is selected, a reference for referring to motion information for any one adjacent block determined based on an adjacent condition between the decoding target block and the adjacent block
  • the present invention it is possible to improve the coding efficiency by reducing the code amount of the flag representing the reference destination of the motion information.
  • FIG. 10A is a flowchart showing details of the calculation processing of the number of candidates NumMergeCandidates in FIG. 9, and FIG. 10B is a diagram showing an example of a reference destination partition adjacent to the processing target partition. It is a flowchart explaining operation
  • FIG. 13A and FIG. 13B are flowcharts for explaining processing executed in accordance with a flag value indicating a predicted reference destination. It is a flowchart explaining the operation
  • Embodiments of the present invention relate to moving picture coding technology, and in particular, improve the efficiency of motion information coding in moving picture coding in which a picture is divided into rectangular blocks and motion estimation and compensation are performed in units of blocks between pictures. Therefore, the motion information of the block to be processed is obtained by using the motion information such as the motion vector of the encoded adjacent block around the block to be processed and the reference picture number as the motion information of the block to be processed.
  • the present invention relates to a technique for reducing the amount of codes by encoding additional information indicating adjacent blocks to be referred to without encoding.
  • the reference destination of motion information is predicted using information such as the size of neighboring blocks and the prediction mode, and is determined by the prediction.
  • a flag merge_proble_flag that represents a match between the adjacent block of the reference destination and the adjacent block actually selected by encoding is defined. If the determination rate by prediction is high, the probability of occurrence of the bit represented by the flag merge_proble_flag increases, so the number of bits representing the flag decreases, and the amount of generated code can be reduced.
  • the flag merge_proble_flag itself representing the match with the above-mentioned reference destination prediction is defined on the assumption that the coincidence accuracy of the determination of the adjacent block that becomes the reference destination by the reference destination prediction is high.
  • the flag merge_direc_flag having the same function as the merge_lu_flag only when the prediction of the reference destination is indefinite. It is also possible to offset the increase in the code amount due to the flag merge_direc_flag by reducing the code amount required for the flag merge_proble_flag, and to relatively improve the encoding efficiency.
  • FIG. 1 is a block diagram showing a configuration of a moving picture encoding apparatus according to an embodiment.
  • the moving image encoding apparatus includes a motion vector detection unit 101, a subtraction unit 102, an orthogonal transform / quantization unit 103, a variable length encoding unit 104, an inverse quantization / inverse orthogonal transform unit 105, and a motion compensation unit 106.
  • the intra prediction unit 111 performs intra prediction from the input encoding target block and the encoded decoded block image adjacent to the encoding target block stored in the memory 110, and determines the intra prediction block. Output.
  • the reference destination specifying unit 113 specifies any one of a plurality of adjacent blocks adjacent to the encoding target block stored in the memory 110 as a reference destination adjacent block when performing the merge processing, and the reference destination adjacent The motion vector of the block is given to the motion compensation unit 106.
  • the motion vector detection unit 101 refers to the input image signal and the reference image signal stored in the memory 110 to perform matching between blocks in units of blocks to detect a motion vector, and to detect the detected motion vector. This is given to the motion compensation unit 106.
  • the motion compensation unit 106 generates a prediction image using the motion vector of the block to be encoded detected by the motion vector detection unit 101 or the motion vector of the adjacent block specified by the reference destination specifying unit 113.
  • the weighted prediction unit 107 adaptively multiplies the prediction image generated by the motion compensation unit 106 by a weight coefficient to generate a final prediction image, and provides the prediction selection unit 112 with it.
  • the prediction selection unit 112 performs intra-screen prediction by the intra-screen prediction unit 111, inter-image prediction using the motion vector of the block to be encoded by the motion compensation unit 106, and the reference destination adjacent block specified by the reference destination specifying unit 113.
  • One prediction method with the smallest code amount is selected from the inter-image prediction using the motion vector, and a prediction image by the selected prediction method is given to the subtraction unit 102 and the addition unit 108.
  • the reference destination specifying unit 113 selects the adjacent block that is the reference destination of the merge process.
  • the prediction process is executed, and the prediction result of the merge reference destination is given to the prediction selection unit 112.
  • the prediction selecting unit 112 determines whether or not the reference destination adjacent block predicted by the reference destination specifying unit 113 matches the reference destination adjacent block used in the actually selected prediction method, and the reference destination adjacent block
  • the reference destination valid information indicating the validity of the prediction result is supplied to the variable length coding unit 104 together with the information indicating the selected prediction method.
  • the subtraction unit 102 generates a residual signal by subtracting the image to be encoded and the predicted image, and supplies the residual signal to the orthogonal transform / quantization unit 103.
  • the orthogonal transform / quantization unit 103 performs orthogonal transform and quantization on the residual signal, generates a transform signal, and supplies the transform signal to the variable length coding unit 104 and the inverse quantization / inverse orthogonal transform unit 105.
  • the variable length coding unit 104 entropy codes the residual signal that has been orthogonally transformed and quantized.
  • the variable length encoding unit 104 also encodes reference destination valid information generated by the prediction selection unit 112 and information indicating the selected prediction method, and outputs a bit stream including an encoded image.
  • the inverse quantization / inverse orthogonal transform unit 105 performs inverse quantization and inverse orthogonal transform on the transform signal received from the orthogonal transform / quantization unit 103 to return to the original residual signal.
  • the adding unit 108 adds the predicted image and the residual signal, generates a decoded image, and supplies the decoded image to the deblocking filter unit 109.
  • the deblocking filter unit 109 performs a process of reducing block distortion due to encoding on the decoded image and stores it in the memory 110.
  • the memory 110 stores information on decoded images and already encoded images.
  • FIG. 2 is a block diagram showing a configuration of a moving picture decoding apparatus according to an embodiment corresponding to the moving picture encoding apparatus of FIG.
  • the moving picture decoding apparatus includes a variable length decoding unit 201, an inverse quantization / inverse orthogonal transform unit 202, a motion compensation unit 203, a weighted prediction unit 204, an addition unit 205, a deblocking filter unit 206, and a memory 207.
  • the decoding process of the moving picture decoding apparatus in FIG. 2 corresponds to the decoding process provided in the moving picture encoding apparatus in FIG. 1, so that the inverse quantization / inverse orthogonal transform unit 202 in FIG.
  • Each configuration of the motion compensation unit 203, the weighted prediction unit 204, the addition unit 205, the deblocking filter unit 206, the memory 207, and the in-screen prediction unit 208 is the inverse quantization / inverse of the moving picture coding apparatus in FIG.
  • the orthogonal transform unit 105, the motion compensation unit 106, the weighted prediction unit 107, the addition unit 108, the deblocking filter unit 109, the memory 110, and the in-screen prediction unit 111 have functions corresponding to the respective configurations.
  • the variable length decoding unit 201 decodes the bitstream and outputs information on the prediction residual signal and the motion vector, provides the prediction residual signal to the inverse quantization / inverse orthogonal transform unit 202, and provides information on the motion vector as the motion information.
  • Information about the coding mode is given to the selection unit 210 to the intra-screen / inter-screen prediction selection unit 209.
  • the inverse quantization / inverse orthogonal transform unit 202 performs inverse orthogonal transform and inverse quantization on the prediction residual signal decoded by the variable length decoding unit 201.
  • the addition unit 205 decodes the image signal by adding the prediction residual component inversely transformed by the inverse quantization / inverse orthogonal transformation unit 202 and the prediction image calculated by the weighted prediction unit 204, and decodes the image signal. This is applied to the blocking filter unit 206.
  • the deblocking filter unit 206 performs a process of reducing block distortion due to encoding on the decoded image and stores it in the memory 207.
  • the intra-screen / inter-screen prediction selection unit 209 determines whether the encoding mode is intra-screen prediction or inter-screen prediction, and if it is intra-screen prediction, instructs the intra-screen prediction unit 208 to execute the intra-screen prediction process. In the case of inter-screen prediction, the motion information selection unit 210 is instructed to execute inter-screen prediction processing.
  • the motion information selection unit 210 selects whether or not motion information such as a motion vector is due to merge processing. If it is due to merge processing, the motion information selection unit 210 is based on predetermined adjacent conditions based on reference destination valid information. It is determined whether the adjacent block to be determined can be designated as a reference block.
  • the intra prediction unit 208 performs intra prediction from the input decoding target block and the decoded block image adjacent to the decoding target block stored in the memory 207, and passes the intra prediction block to the adding unit 205.
  • the motion compensation unit 203 generates a prediction image using the decoded motion vector of the decoding target block when the inter-screen prediction is not merge processing, that is, is based on normal motion vector detection processing.
  • the motion compensation unit 203 performs prediction using the motion vector of the reference destination adjacent block predicted by the reference destination prediction of the motion information. Generate an image.
  • the weighted prediction unit 204 adaptively multiplies the prediction image generated by the motion compensation unit 203 by a weighting factor to generate a final prediction image, which is given to the addition unit 205.
  • the prediction method of an adjacent block of a reference destination is performed in the prediction selection unit 112 of the video encoding device in FIG. 1 and the motion information selection unit 210 of the video decoding device in FIG.
  • the prediction selection unit 112 of the video encoding device from among the motion information detected by the normal motion vector detection process or the inter-screen prediction based on the motion information of the adjacent block selected by the merge process, and the intra-screen prediction, One encoding mode is selected that has the least amount of generated code, the best image quality, or both. Further, if the selected encoding mode is inter-screen prediction, it is determined whether the motion information is detected by a normal motion vector detection process or is selected by a merge process, and the merge process is performed. In this case, the reference destination prediction of the motion information is performed, the identity between the predicted reference destination adjacent block and the adjacent block selected by the merging process is determined, and the result is encoded as reference destination valid information. Is transmitted. When the merge process is not performed, normal motion vector prediction is performed, and a difference motion vector between the motion vector and the predicted motion vector and other motion information are encoded and transmitted.
  • the intra-screen / inter-screen prediction selection unit 209 determines the inter-screen based on the encoding mode and motion information decoded from the bit stream by the variable length decoding unit 201. If it is determined to be prediction, it is further determined whether or not the inter-screen prediction is due to merge processing. If it is determined to be due to merge processing, motion information reference destination prediction is performed. The motion information is acquired by determining the identity between the predicted reference adjacent block and the adjacent block specified by the decoded additional information. In the following embodiments, details of motion information reference destination prediction will be described.
  • Example 1 Prior to describing an example of a motion information reference destination prediction method, terms used in this example will be described.
  • the screen is equally divided into square rectangular blocks of the same size.
  • This block is called a coding block and is the basis of processing when performing coding and decoding.
  • the coding block can be divided into four blocks with a small block size in order to optimize the coding process.
  • a coded block that is divided into equal screen sizes in the screen shown in FIG. 3 is called a maximum coded block, and the inside of which is divided into four according to the coding conditions is collectively called a coded block.
  • An encoded block having a minimum size that cannot be further divided into four is referred to as a minimum encoded block.
  • a block that performs this motion compensation is called a prediction block.
  • the prediction block is represented by the same size as the coding block, and according to the motion compensation, the prediction block is regarded as one block without being divided, and is divided into two in the horizontal or vertical direction. Divided into four parts by equal division.
  • the mode corresponding to the division type is defined according to the size after the division, and is shown in FIG.
  • Each area obtained by dividing the prediction block is called a partition.
  • numbers starting from 0 are assigned to the partitions existing in the prediction block in the zigzag scan order. This number is called a partition number and is represented by puPartIdx.
  • the number described in the partition of the prediction block in FIG. 4 represents the partition number of the partition.
  • a motion information reference destination prediction method will be described with reference to the drawings.
  • the motion information reference destination prediction method is performed in any of encoding and decoding processes for each partition constituting a prediction block.
  • the motion information is referred from the surrounding partition adjacent to the partition of the prediction block (partition to be processed in FIG. 5) defined for motion compensation within the coding block in the same picture. Select the destination partition.
  • a partition that has already been encoded before encoding the partition to be processed is a candidate. Since encoding is performed in block units in the raster scan order from the upper left to the lower right of the screen, here, in principle, the left or upper partition adjacent to the periphery of the partition to be processed is considered as a reference destination candidate. The following explanation will be based on this assumption unless otherwise noted.
  • FIG. 6 is an enlarged view of the area represented by the thick dotted circle in FIG.
  • FIG. 6 shows a partition to be processed and a partition adjacent to it.
  • nA represents the total number of partition groups adjacent to the left
  • nB represents the total number of partition groups adjacent immediately above.
  • the reference destination on the left or directly above is determined based on the following rule. Determine candidate partitions.
  • the top partition A0 among them is set as a reference destination candidate.
  • the leftmost partition B0 among them is set as a reference destination candidate.
  • FIG. 7 shows a first syntax pattern described in a sequence parameter set (SPS) of a moving image bit stream.
  • the sequence parameter set is a header in which information related to coding of the entire sequence is described.
  • the motion information reference destination prediction method When performing inter-screen prediction over the entire sequence, the motion information reference destination prediction method according to the present embodiment is applied.
  • a first flag inter_merge_flag indicating whether or not to perform is set.
  • FIG. 8 shows a second syntax pattern described for each prediction block in the slice.
  • the prediction mode of the prediction block is inter-screen prediction (MODE_INTER)
  • the number of effective neighboring partitions NumMergeCandidates is obtained for each partition in the prediction block, and when NumMergeCandidates exceeds 0, merging is performed in this partition.
  • a second flag merge_flag [i] indicating whether or not to apply is set.
  • i represents the partition number in the prediction block.
  • the second flag merge_flag [i] is not coded and is not described in the bitstream when the motion information reference prediction method according to the present embodiment is not applied, that is, when the inter_merge_flag is false (0).
  • the third flag merge_proble_flag [i] does not apply merging, that is, when merge_flag [i] is false (0), the motion information detected by the normal motion vector detection is used, so the third flag merge_proble_flag [i] is used. i] does not need to be determined. Also, even when NumMergeCandidates is 1, it is not encoded. This is because, if the number of effective partitions adjacent to each other is one, one of the partitions becomes a reference destination partition, so that the motion information of the reference destination partition is determined without transmitting merge_proble_flag [i].
  • the reference destination specifying unit 113 determines whether or not to use a partition adjacent to the left and directly above the processing target partition as a reference destination of the merge processing. The operation of the reference destination specifying unit 113 will be described with reference to the flowchart of FIG. First, the number NumMergeCandidates of peripheral partitions adjacent to the partition to be processed is calculated (S101).
  • FIG. 10A is a flowchart showing details of the calculation processing of the number of candidates NumMergeCandidates.
  • NumMergeCandidates is set to 0 (S201).
  • a partition located to the left of the partition to be processed is set as a reference destination partition (S202). It is determined from the position information in the screen of the partition to be processed whether or not the reference destination partition exists in the screen (S203).
  • FIG. 10 (b) shows an example of a reference destination partition (rectangle represented by diagonal lines in the figure) adjacent to the partition to be processed (gray rectangle in the figure).
  • a reference destination partition rectangle represented by diagonal lines in the figure
  • FIG. 10 (b) shows an example of a reference destination partition (rectangle represented by diagonal lines in the figure) adjacent to the partition to be processed (gray rectangle in the figure).
  • the partition to be processed is located at the upper left, there is no reference destination partition, only when it is located at the top of the screen, only at the left, when located at the left edge of the screen, only at the top, other positions Then it is on the left and top.
  • the reference destination partition does not exist in the screen (No in S203)
  • the subsequent processing is skipped and the process proceeds to Step S207.
  • the reference destination specifying unit 113 acquires the motion information of the reference destination partition (S204).
  • the reference destination designating unit 113 reads the encoded motion information corresponding to the reference destination partition from the memory 110. Based on the read motion information of the reference destination partition, the reference destination specifying unit 113 determines whether or not the encoding mode of the reference destination partition is intra (S205). If it is intra, the process proceeds to S207. If it is not intra, NumMergeCandidates is incremented by 1 (S206).
  • Step S207 It is determined whether the setting of the reference destination partition is on the partition to be processed. If the reference destination partition is set on the processing target partition, the process ends and outputs NumMergeCandidates. When the reference destination partition is not set on the processing target partition, the reference destination partition is set again on the processing target partition, and the determination processing after step S203 is continued (S208).
  • the number of candidates NumMergeCandidates is calculated as described above.
  • the reference destination specifying unit 113 acquires the motion information of the reference destination partition from the memory 110 (S105). ).
  • the reference destination specifying unit 113 supplies the acquired motion information of the reference destination partition to the motion compensation unit 106, and the motion compensation unit 106 performs motion compensation of the processing target partition based on the motion information of the specified reference destination partition. .
  • the reference destination designating unit 113 reads out and acquires the motion information of the partitions adjacent to the left and above the processing target partition from the memory 110 (S104).
  • the reference destination specifying unit 113 compares the acquired motion information of the left and upper adjacent partitions (S106). Specifically, the prediction direction, the reference picture number, and the motion vector constituting the motion information are compared. When the motion information of the left and upper adjacent partitions is exactly the same (Yes in S106), the number of candidates NumMergeCandidates is changed to 1 (S107), and the reference destination designating unit 113 sets either the left or the upper adjacent The motion information of the partition to be supplied is supplied to the motion compensation unit 106 as the motion information of the reference destination partition. Here, the motion information of the partition adjacent to the left is selected. When the motion information of the left and upper adjacent partitions is different (Yes in S106), the reference destination specifying unit 113 supplies the two pieces of motion information to the motion compensation unit 106, respectively.
  • the reference destination specifying unit 113 supplies the motion information of the partition of the reference block selected by the merging process to the motion compensation unit 106, and the motion compensation unit 106 specifies the partition of the specified reference block. Motion compensation is performed based on the motion information. Also, the motion compensation unit 106 performs normal motion compensation based on the motion vector for each partition of the prediction block detected by the motion vector detection unit 101.
  • the in-screen prediction unit 111 performs in-plane prediction using pixel correlation in the screen. A prediction image generated by these three prediction methods is input to the prediction selection unit 112.
  • the prediction selection unit 112 calculates a generated code amount of additional information such as a residual image and motion information generated by the difference between the image to be encoded and the three input prediction images, and makes a prediction that minimizes the code amount Select a method.
  • the prediction selection unit 112 sets a flag merge_dirc_flag that indicates which one of the left or the upper is used as the motion information of the reference destination partition selected by the merge process.
  • merge_dirc_flag indicates which one of the left or the upper is used as the motion information of the reference destination partition selected by the merge process.
  • the prediction selection unit 112 sets a flag for merge processing defined by the above-described syntax when the prediction method by merge processing is selected.
  • FIG. 11 is a flowchart illustrating a procedure in which the prediction selection unit 112 sets a flag for merge processing.
  • merge_flag is set to 1 (S302).
  • the number of motion information selected by the merge process is checked (S304). Specifically, it is determined whether or not the number of adjacent partition candidates NumMergeCandidates to be specified by the reference destination specifying unit 113 is greater than one.
  • the process proceeds to prediction of a reference destination partition (S305).
  • a reference destination prediction method for motion information described later is used. Based on information such as the size of the partition adjacent to the left and top of the target partition and the length of the side where the target partition and the adjacent partition are in contact with each other, the reference partition is predicted and predicted A flag pred_direc_flag indicating the reference destination partition is output.
  • the flag pred_dirc_flag indicating the predicted reference destination partition indicates the selection destination on the left (1) or above (0) as the adjacent partition of the reference destination, as in the case of merge_direct_flag, and the reference destination partition cannot be selected by prediction.
  • Indefinite (2) is set.
  • the flag pred_direc_flag indicating the predicted reference destination partition set in this way is determined (S306).
  • the flag pred_dirc_flag indicating the predicted reference destination partition is 2 when prediction is impossible in the prediction of the reference destination partition (Yes in S306), and at this time, it is selected as the prediction method that actually minimizes the generated code amount.
  • the merge_direc_flag indicating the reference destination partition of the motion information used in the prediction performed by the merge processing is set in the flag merge_proble_flag indicating the reference destination valid information, and the process ends (S307).
  • the flag merge_proble_flag indicating the reference destination valid information directly indicates the reference destination partition in the merge process as in the conventional case.
  • merge_direc_flag represents a reference destination partition of motion information used in prediction by merge processing selected as a prediction method that actually minimizes the amount of generated code, and pred_direc_flag has the same value if the prediction of the reference destination partition is correct.
  • the flag merge_proble_flag is set to 1 (S309).
  • pred_dirc_flag has a different value. In this case, the flag merge_proble_flag is set to 0 (S310).
  • merge_proble_flag When merge_proble_flag indicating success / failure of the merge destination partition is used, merge_proble_flag becomes 1 when the merge destination partition is predicted, and merge_proble_flag becomes 0 only when the prediction is off. In other words, merge_proble_flag can be transmitted with a smaller amount of information by using arithmetic coding using the fact that merge_proble_flag has a higher probability of 1 and merge_proble_flag has a higher probability of 1 as the merge destination partition is predicted. it can.
  • the first flag inter_merge_flag described in the header of the SPS is described in the prediction block.
  • the second flag merge_flag and the third flag merge_proble_flag are set and encoded.
  • the motion information reference destination prediction method according to the present embodiment is used for the entire bitstream sequence. It is determined whether inter-screen prediction by merge processing is applied. When inter_merge_flag is true (1), merge processing is applied. When inter_merge_flag is false (0), merge processing is ignored and inter-screen prediction based on motion information decoded from a conventional bitstream is performed. . In the following description, it is assumed that inter_merge_flag is true (1), that is, the merge processing using the motion information reference destination prediction method according to the present embodiment is applied.
  • the intra-screen / inter-screen prediction selection unit 209 determines whether to select intra-screen or inter-screen prediction with reference to the prediction mode for each prediction block in the slice.
  • the intra prediction MODE_INTRA
  • the process proceeds to the intra prediction unit 208
  • the inter prediction MODE_INTER
  • the process proceeds to the motion information selection unit 210.
  • the motion information selection unit 210 uses, as information decoded from the bit stream for each partition of the prediction block, the motion information detected by the motion vector detection of encoding or the reference destination partition selected by the merge process. Either one of the motion information is selected, and the selected motion information is output to the motion compensation unit 203.
  • FIG. 12 is a flowchart showing a selection procedure of the motion information selection unit 210, which will be described with reference to this figure.
  • the number NumMergeCandidates of peripheral partitions adjacent to the partition to be processed is calculated (S401). Since the calculation procedure is the same as S101 in encoding, it is omitted.
  • the motion information selection unit 210 outputs the motion information decoded from the bitstream as in the conventional case.
  • merge_flag decoded by the variable length decoding unit 201 is read (S403), and selection of motion information is determined based on merge_flag (S404).
  • merge_flag false (0)
  • the motion information selection unit 210 outputs motion information decoded from the bitstream as in the conventional case.
  • the motion information of the reference destination partition is acquired from the memory 207 in which the decoded information is recorded (S407).
  • the acquired motion information is supplied to the motion compensation unit 203, and motion compensation is performed based on the motion information.
  • the motion information of the partitions adjacent to the left and above the partition to be processed is read from the memory 207 and acquired (S406).
  • the obtained motion information of the left and upper adjacent partitions is compared (S408). Specifically, the prediction direction, the reference picture number, and the motion vector constituting the motion information are compared. If the motion information is exactly the same, the motion information of either the left or upper adjacent partition is acquired and supplied to the motion compensation unit 203 (S410).
  • the motion information of the partition adjacent to the left is selected as in the encoding.
  • the process proceeds to prediction of the reference destination partition (S409).
  • a motion information reference destination prediction method which will be described later, is used as in the case of encoding.
  • the target partition is predicted based on information such as the size of the partition adjacent to the processing target partition and the left and top of the periphery, and the length of the side where the processing target partition and the adjacent partition are in contact with each other.
  • the flag pred_dirc_flag indicating the referred reference partition is output.
  • the flag pred_dirc_flag indicating the predicted reference destination partition is represented by left (1) or upper (0) when the adjacent reference destination partition is selected as in the encoding, and the reference destination partition can be selected by prediction. If not, undefined (2) is set.
  • pred_direc_flag output is determined (S411).
  • the case where pred_dirc_flag is 2 is a case where prediction is impossible due to prediction of the reference destination partition, and at this time, the process proceeds to case 1 (S412).
  • pred_dirc_flag is not 2
  • the process proceeds to case 2 (S413).
  • FIG. 13A shows a flowchart of the process of case 1 in step S412.
  • the prediction result is undefined in the prediction of the reference destination partition, and the prediction result of the reference destination is not shown. Therefore, the partition adjacent to the left or above indicated by merge_proble_flag is selected as the reference destination partition as in the conventional case. This is because, in the encoding side, merge_direct_flag indicating the actual reference destination partition is set in merge_proble_flag.
  • the merge_proble_flag decoded by the variable length decoding unit 201 is read (S421), and a reference destination partition is selected based on the merge_proble_flag (S422).
  • pred_dirc_flag 1
  • the left adjacent partition S423
  • merge_proble_flag 0
  • the upper adjacent partition is selected as the reference destination partition (S424).
  • FIG. 13B shows a flowchart of the process of case 2 in step S413.
  • the probability of the flag pred_direc_flag indicating the predicted reference destination partition is represented by the flag merge_proble_flag indicating the reference destination validity. That is, merge_proble_flag is represented by 1 (true) if the reference destination partition indicated by pred_dirc_flag is correct, and merge_proble_flag is represented by 0 (false) if not correct.
  • the merge_proble_flag decoded by the variable length decoding unit 201 is read (S431), and the selection of the reference destination partition is determined based on the merge_proble_flag (S432).
  • merge_proble_flag 1
  • a partition having a value indicated by pred_dirc_flag determined by prediction of the reference destination partition is selected as the reference destination partition (S433).
  • merge_proble_flag 0
  • a partition opposite to the value indicated by pred_direc_flag is selected as a reference destination partition. For example, if pred_dirc_flag is 0 (left), the upper adjacent partition is set as the reference destination partition (S434).
  • the motion information of the reference destination partition selected in the processing of case 1 and case 2 is acquired (S415), supplied to the motion compensation unit 203, and the process ends.
  • the reference is made by the motion information reference destination prediction method described later. Select the destination partition.
  • a decoded image is generated by adding the predicted image generated by motion compensation from the motion information of the selected reference destination partition and the residual signal decoded from the bitstream.
  • FIG. 14 is a flowchart for explaining a motion information reference destination prediction method according to the present embodiment.
  • the process shown in FIG. 14 is performed in units of partitions.
  • a partition that refers to motion information is determined for the partition to be processed, and if a partition is determined as a reference destination, the process does not proceed to the next process and ends.
  • each process will be described step by step.
  • FIG. 15 is a flowchart showing a boundary determination process with adjacent surrounding partitions in step S500. Details of the processing will be described with reference to FIG.
  • motion boundary detection is performed between a partition to be processed and neighboring partitions around it (S501).
  • FIG. 16 shows an arrangement example in which the processing target partition is X and the left and upper adjacent partitions are A and B, respectively.
  • the rectangular size of the partition to be processed is represented by width wx and height hx
  • the height of the adjacent left partition is represented by pa
  • the width of the adjacent upper partition is represented by pb.
  • the boundary between the processing target partition and the left and upper adjacent partitions is determined by comparing the height hx of the processing target partition with the height pa of the adjacent left partition and the width wx of the processing target partition. This is done by comparing the width pb of the upper partition.
  • an edge where an adjacent partition is adjacent to the target partition is called a “motion boundary”.
  • FIG. 18 is a flowchart showing the adjacent side length comparison process with the adjacent partition in step S510. Details of the processing will be described with reference to FIG.
  • the length of the adjacent side in contact with the left or upper partition adjacent to the processing target partition is calculated (S511).
  • the adjacent side length to the left partition A is represented by L (A)
  • the adjacent side length to the upper partition B is represented by L (B) (the hatched portion in FIG. 16 is adjacent).
  • the adjacent side length L (A) with the left partition A is the smaller of the height hx of the partition X to be processed and the height pa of the left partition, and is expressed by the following equation.
  • L (A) min (pa, hx)
  • the function min (a, b) is a function for selecting the smaller one of a and b.
  • the adjacent side length L (B) with the upper partition B is the smaller one of the width wx of the partition X to be processed and the height pb of the upper partition, and is expressed by the following equation.
  • L (B) min (pb, wx)
  • pa and pb are the lengths of the motion boundaries obtained in step S500.
  • the obtained adjacent side lengths L (A) and L (B) are compared, and a reference destination partition is selected (S512). If L (A) and L (B) are equal, the reference destination partition cannot be determined, so the reference destination partition is not selected and is undefined (S513), and the process ends. If L (A) and L (B) are not equal, L (A) and L (B) are compared in size (S514). If L (A) is greater than L (B), A is selected (S515) and the process ends. If L (A) is smaller than L (B), B is selected (S516) and the process ends.
  • the determination results classified according to the motion boundary and side length conditions of the left and upper partitions adjacent to the partition to be processed are as follows.
  • the summary is shown in the table of FIG.
  • the selection column in the table of FIG. 19 indicates the partition to be referred to, and the symbol “-” represents the case where the reference partition is indefinite.
  • the fifth boundary comparison condition “pb ⁇ wx and pa ⁇ hx” in FIG. 19 is arranged as shown in FIG. 20, and the lengths of the motion boundaries of adjacent partitions A and B are both subject to processing. This is the case when the width and height of the partition do not match.
  • the reference destination is determined by comparing the lengths of adjacent side lengths in contact with the partition to be processed.
  • the adjacent side lengths L (A) and L (B) of the adjacent partitions A and B are the sides pa and pb that are in contact with the partition X to be processed, so that the reference partition is determined by comparing pa and pb. Determined.
  • FIG. 21 shows an example of the processing target and the arrangement of adjacent partitions under the boundary comparison condition “pb> wx and pa> hx”. As shown in FIG. 4, since the size and shape of the partition inside the prediction block are defined, the size and shape of the partition meeting the boundary comparison condition “pb> wx and pa> hx” shown in FIG. Limited.
  • the prediction result of the determined reference destination partition is output as a flag pred_dirc_flag indicating the reference destination partition predicted in the encoding and decoding processes.
  • pred_direc_flag is the same as merge_direc_flag, when it is selected as a reference destination adjacent partition, “1 (left)” or “0 (upper)” indicates the reference destination, and the reference destination partition cannot be selected by prediction. Is set to “2 (undefined)”.
  • the occurrence of the bit represented by the flag merge_proble_flag used for designating the reference destination is biased and generated by predicting the reference block in the merge process.
  • the amount of codes can be reduced, and the encoding efficiency can be improved.
  • the reference destination flag has a frequency of 1 ⁇ 2 and the arithmetic coding cannot be used.
  • the predicted reference is By using the flag merge_proble_flag indicating the probability of the previous, if the prediction is correct, the bit frequency of the flag can be biased, and the amount of generated codes can be reduced in arithmetic coding.
  • the adjacent block that matches the size of the encoding target block, or the adjacent block having a longer side length is highly correlated with the encoding target block, so that the reference block It is possible to improve the accuracy of prediction by predicting.
  • Example 2 A second example of the motion information reference destination prediction method according to the embodiment of the present invention will be described. The difference from the first embodiment is that a determination process based on the residual signal of the left or upper partition adjacent to the processing target partition is added as a determination condition for the reference destination prediction.
  • FIG. 22 is a flowchart for explaining the operation of the second embodiment of the motion information reference destination prediction method.
  • the processes of steps S600 and S610 shown in FIG. 22 are the same as the processes of steps S500 and S510 of the first embodiment shown in FIG.
  • a boundary determination is made with the surrounding partition adjacent to the left or upper side of the partition to be processed (S600), and the length of the common part of adjacent sides of the left or upper partition adjacent to the processing target partition ("adjacent" The side length ") is determined (S610).
  • FIG. 23 is a diagram illustrating a pixel region for calculating a boundary residual between adjacent partitions.
  • A be the left partition adjacent to partition X to be processed
  • B be the upper partition.
  • Each partition A, B is assumed to be composed of a pixel residual signal after motion compensation prediction in the encoding process or a pixel residual signal after inverse quantization and inverse orthogonal transformation in the decoding process.
  • the pixel area of the partition A that is in contact with the partition X is represented by an area having a height of min (ha, hx) and a width of 1 pixel located at the left boundary of the partition X, and a residual signal of the pixel.
  • ⁇ (A) is expressed by the following equation, where ⁇ (A) is the sum of absolute values of.
  • ai represents the residual signal of the pixel in partition A adjacent to partition X.
  • the number of pixels in the partition A adjacent to the partition X is the smaller of the number of pixels ha representing the height of the partition A and the height hx of the partition X, and is represented by a function min (ha, hx).
  • the pixel area of the partition B that is in contact with the partition X is represented by an area having a width of min (wb, wx) and a height of 1 pixel located at the upper boundary of the partition X. If the sum of absolute values of residual signals is ⁇ (B), ⁇ (B) is expressed by the following equation.
  • bi represents the residual signal of the pixel in partition B adjacent to partition X.
  • the number of pixels in the partition B adjacent to the partition X is the smaller of the number of pixels wb representing the width of the partition B and the width wx of the partition X, and is represented by a function min (wb, wx).
  • FIG. 24 is a flowchart for explaining the operation of determining the reference destination partition based on the boundary residual between adjacent partitions.
  • normalization of the calculated boundary residual is performed (S622). As shown in FIG. 23, when the adjacent side lengths of the partitions A and B adjacent to the partition X are different, it is not appropriate to directly compare the boundary residual of the partition A and the boundary residual of the partition B. As shown in the following equation, normalization is performed by dividing the boundary residuals of the partitions A and B by the adjacent side lengths of the partitions A and B, respectively.
  • the process proceeds to calculation of the residual ratio (S627).
  • the residual ratios ⁇ (A) and ⁇ (B) of the partitions A and B are the boundary residuals ⁇ (A) and ⁇ (B) before normalization of the partitions A and B calculated in step S621. ) To the total residual of partitions A and B is expressed by the following equation.
  • Ai and Bi are residual signals of pixels in partitions A and B adjacent to partition X
  • nA and nB are the total number of pixels in partitions A and B.
  • an area that is not adjacent to the partition X to be processed is not a target.
  • the number of pixels wa representing the width of the partition A is compared with the number of pixels hb representing the height of the partition B, and the smaller one is defined as the width of the target region.
  • the number of pixels nB in the target area of partition B is given by the following equation.
  • nA is a rectangular area indicated by light gray in the thick dotted line in the partition A of FIG. 23, and the denominator of the calculation formula of ⁇ (A) is the absolute value of the residual signal of the pixel in this rectangular area. Expressed as the sum.
  • nB covers the entire area of partition B in FIG. 23, and the denominator part of the calculation formula of ⁇ (B) is expressed as the sum of absolute values of residual signals of pixels in partition B.
  • the boundary residuals ⁇ (A) and ⁇ (B) before normalization are sums of absolute values of residual signals of pixels adjacent to the partition X, respectively, and the number of pixels adjacent to the partition X is min (ha, hx).
  • the residual ratios ⁇ (A) and ⁇ (B) are calculated with the same number of pixel ratios, and ⁇ (A) and ⁇ (B) are the partition X Represents the ratio of the residual signals of adjacent pixels to the residual signals of the pixels in partitions A and B. If this ratio is high, the motion compensation prediction near the pixel adjacent to the partition X is not applied, that is, the motion information used for the motion compensation prediction is not appropriate for the prediction near the pixel adjacent to the partition X. Means that. That is, it is determined that the motion information of the adjacent partition having the higher residual ratio is not referred to.
  • the residual ratio calculated in this way is compared. It is determined whether or not ⁇ (A) and ⁇ (B) are equal (S628). If ⁇ (A) and ⁇ (B) are not equal, ⁇ (A) and ⁇ (B) have a magnitude relationship, so a magnitude comparison is performed. Here, it is determined whether or not ⁇ (A) is smaller than ⁇ (B) (S629). If ⁇ (A) is smaller than ⁇ (B), A is selected (S630). Otherwise, B is selected (S631). If ⁇ (A) and ⁇ (B) are equal, there is no difference in the residual that occurs regardless of which one is selected, so A is selected (S632), and the process ends.
  • pred_direc_flag indicates a reference destination with “1 (left)” or “0 (upper)” when it is selected as an adjacent partition to be referred to.
  • the reason for selecting A in step S632 is that A (left) is assigned to bit "1" in this embodiment. This is to reduce the amount of generated code in encoding by biasing the occurrence frequency of pred_direc_flag to “1” as prediction of a reference destination partition of motion information.
  • step S623 shown in FIG. 24 strict determination as to whether or not the boundary residuals ⁇ (A) and ⁇ (B) are equal is performed using ⁇ (A) and ⁇ (B The determination may be changed based on whether or not the absolute value difference
  • may be a fixed value or may be a variable value that is changed according to the value of the quantization parameter Qp, for example.
  • the difference determination of the boundary residual and the residual ratio is performed in order, but the boundary residual and the residual ratio are determined separately, and the final reference destination partition is selected from the results. You may do it.
  • the boundary residual and the residual ratio may be collected and a new evaluation criterion may be generated and determined. For example, the determination may be made based on the evaluation value generated by the combination of the boundary residual and the residual ratio as in the following equation.
  • variable x is A or B
  • is a weighting coefficient.
  • may be a fixed value, or may be set as a variable value according to a partition, a slice unit, a slice type, or the like.
  • prediction of a reference destination is performed by predicting a reference block based on the magnitude of a residual signal of a block adjacent to the block to be encoded. Accuracy can be further increased.
  • Example 3 A third example of the motion information reference destination prediction method according to the embodiment of the present invention will be described.
  • the difference from the first embodiment is that the partition represented by the flag pred_direc_flag determined and output by prediction of the reference destination partition is selected as the reference destination partition.
  • the flag pred_direc_flag based on prediction of the reference destination partition is 0 or 1
  • the left (pred_direc_flag is “1”) or the top (pred_dirc_flag is “0”) is selected as the reference destination adjacent partition.
  • FIG. 26 is a flowchart for explaining a procedure for setting a flag for merge processing.
  • the prediction method selected by the prediction selection unit 112 is prediction based on motion information selected in the merge process (S701).
  • merge_flag is set to 0 and the process ends (S703). If the prediction is based on the motion information selected in the merge process, merge_flag is set to 1 (S702).
  • the number of motion information selected by the merge process is checked (S704). It is determined whether or not the number of adjacent partition candidates NumMergeCandidates to be specified by the reference destination specifying unit 113 is greater than one. When the number of candidates NumMergeCandidates is 1 or less, it is assumed that merge processing is selected, and there is always one adjacent partition of a valid reference destination, and either the left or the upper can be specified. Therefore, only merge_flag is set and the process ends.
  • the process proceeds to prediction of a reference destination partition (S705).
  • the motion information reference destination prediction method described in the first embodiment is used. Based on information such as the size of the partition adjacent to the left and top of the target partition and the length of the side where the target partition and the adjacent partition are in contact with each other, the reference destination partition is predicted, and the flag pred_dirc_flag is set. Is output.
  • the flag pred_dirc_flag indicating the predicted reference destination partition represents the selection destination on the left (1) or above (0) as the adjacent partition of the reference destination, and the reference destination partition cannot be selected by prediction. Indeterminate (2) is set.
  • the flag pred_direc_flag indicating the predicted reference destination partition set in this way is determined (S706).
  • the flag pred_dirc_flag indicating the predicted reference destination partition is 2 when prediction is impossible in the prediction of the reference destination partition, and at this time, it depends on the merge process selected as the prediction method that actually minimizes the generated code amount.
  • the process sets the merge_dir_flag indicating the reference partition of the motion information used in the motion compensation prediction, and the process ends (S707).
  • pred_direc_flag is not 2
  • pred_direc_flag directly indicates a reference destination partition, and when “0” is selected, a partition adjacent to the left is selected.
  • the first flag inter_merge_flag described in the header of the SPS is described in the prediction block.
  • the second flag merge_flag and the third flag merge_direc_flag are set and encoded.
  • the motion information reference destination prediction method according to the present embodiment is used for the entire bitstream sequence. It is determined whether inter-screen prediction by merge processing is applied. When inter_merge_flag is true (1), merge processing is applied. When inter_merge_flag is false (0), merge processing is ignored and inter-screen prediction based on motion information decoded from a conventional bitstream is performed. . In the following description, it is assumed that inter_merge_flag is true (1), that is, the merge processing using the motion information reference destination prediction method according to the present embodiment is applied.
  • the intra-screen / inter-screen prediction selection unit 209 determines whether to select intra-screen or inter-screen prediction with reference to the prediction mode for each prediction block in the slice.
  • the intra prediction MODE_INTRA
  • the process proceeds to the intra prediction unit 208
  • the inter prediction MODE_INTER
  • the process proceeds to the motion information selection unit 210.
  • the motion information selection unit 210 uses, as information decoded from the bit stream for each partition of the prediction block, the motion information detected by the motion vector detection of encoding or the reference destination partition selected by the merge process. Either one of the motion information is selected, and the selected motion information is output to the motion compensation unit 203.
  • FIG. 27 is a flowchart showing the selection procedure of the motion information selection unit 210, which will be described with reference to this figure.
  • the number NumMergeCandidates of neighboring partitions adjacent to the partition to be processed is calculated (S801). Since the calculation procedure is the same as S701 in encoding, it is omitted.
  • the motion information selection unit 210 outputs the motion information decoded from the bitstream as in the conventional case.
  • merge_flag decoded by the variable length decoding unit 201 is read (S803), and the selection of motion information is determined based on the merge_flag (S804).
  • merge_flag is false (0), the motion information selection unit 210 outputs motion information decoded from the bitstream as in the conventional case.
  • merge_flag is true (1), the process proceeds to processing for selecting motion information of the reference partition in the merge processing.
  • the motion information of the reference destination partition is acquired from the memory 207 in which the decoded information is recorded (S807).
  • the acquired motion information is supplied to the motion compensation unit 203, and motion compensation is performed based on the motion information.
  • the motion information of the partitions adjacent to the left and above the processing target partition is read from the memory 207 and acquired (S806).
  • the obtained motion information of the left and upper adjacent partitions is compared (S808). Specifically, the prediction direction, the reference picture number, and the motion vector constituting the motion information are compared. If the motion information is exactly the same, the motion information of either the left or upper adjacent partition is acquired and output to the motion compensation unit 203 (S810).
  • the motion information of the partition adjacent to the left is selected as in the encoding.
  • the process proceeds to prediction of the reference destination partition (S809).
  • the motion information reference destination prediction method described in the first embodiment is used, similarly to the encoding.
  • the target partition is predicted based on information such as the size of the partition adjacent to the processing target partition and the left and top of the periphery, and the length of the side where the processing target partition and the adjacent partition are in contact with each other.
  • the flag pred_dirc_flag indicating the referred reference partition is output.
  • the flag pred_dirc_flag indicating the predicted reference destination partition is represented by left (1) or upper (0) when the adjacent reference destination partition is selected as in the encoding, and the reference destination partition can be selected by prediction. If not, undefined (2) is set.
  • the pred_direc_flag output is determined (S811).
  • the pred_dirc_flag is 2 when prediction is impossible due to prediction of the reference destination partition, and the reference destination partition cannot be determined. Therefore, merge_direc_flag decoded by the variable length decoding unit 201 is read (S812). Based on merge_direc_flag, the selection of the reference destination partition is determined (S813). If merge_dirc_flag is 1, the left adjacent partition (S814) is selected. If merge_dirc_flag is 0, the upper adjacent partition is selected as the reference destination partition (S815).
  • pred_direc_flag when pred_direc_flag is not 2, the value indicated by pred_direc_flag directly indicates the adjacent partition of the reference destination. That is, if pred_dirc_flag is 1, the left adjacent partition is selected, and if merge_proble_flag is 0, the upper adjacent partition is selected as a reference destination partition. The motion information of the reference destination partition thus selected is acquired, supplied to the motion compensation unit 203, and the process ends.
  • the reference is made by the above-described motion information reference destination prediction method. Select the destination partition.
  • a decoded image is generated by adding the predicted image generated by motion compensation from the motion information of the selected reference destination partition and the residual signal decoded from the bitstream.
  • the determination processing based on the residual signal of the left or upper partition adjacent to the partition to be processed described in the second embodiment can be applied in the same manner.
  • the third embodiment it is obtained by prediction of reference destination without defining merge_proble_flag itself, assuming that the matching accuracy of the determination of the adjacent block that is the reference destination by the prediction of the reference destination is high.
  • the reference destination neighboring block may be indicated by the determination result, and the reference destination neighboring block may be indicated by the flag merge_dirc_flag only when the prediction of the reference destination is indefinite.
  • the increase in the code amount due to merge_direct_flag can be offset by the reduction in the code amount required for merge_proble_flag, and the encoding efficiency can be relatively improved.
  • the left or upper adjacent partition is always designated as the reference destination partition in the prediction of the reference destination partition, so that it is not determined that the reference destination partition is indeterminate by the prediction of the reference destination partition. May be.
  • it is determined as indefinite in the prediction of the reference destination partition it is predetermined in advance that the adjacent partition on the left or the upper side is uniquely specified. Since it is not necessary to encode and transmit the flag merge_dirc_flag indicating the reference destination partition when it is determined indefinitely by partition prediction, it is possible to improve the encoding efficiency.
  • the reference destination prediction method of motion information in decoding corresponding to encoding the same reference destination partition as that on the encoding side is predicted, and the left or upper adjacent partition determined by prediction of the reference destination partition is used as the reference destination partition. select.
  • the moving image encoded stream output from the moving image encoding apparatus of the embodiment described above has a specific data format so that it can be decoded according to the encoding method used in the embodiment. Therefore, the moving picture decoding apparatus corresponding to the moving picture encoding apparatus can decode the encoded stream of this specific data format.
  • the encoded stream When a wired or wireless network is used to exchange an encoded stream between a moving image encoding device and a moving image decoding device, the encoded stream is converted into a data format suitable for the transmission form of the communication path. It may be transmitted.
  • a video transmission apparatus that converts the encoded stream output from the video encoding apparatus into encoded data in a data format suitable for the transmission form of the communication channel and transmits the encoded data to the network, and receives the encoded data from the network Then, a moving image receiving apparatus that restores the encoded stream and supplies the encoded stream to the moving image decoding apparatus is provided.
  • the moving image transmitting apparatus is a memory that buffers the encoded stream output from the moving image encoding apparatus, a packet processing unit that packetizes the encoded stream, and transmission that transmits the packetized encoded data via the network.
  • the moving image receiving apparatus generates a coded stream by packetizing the received data, a receiving unit that receives the packetized coded data via a network, a memory that buffers the received coded data, and packet processing. And a packet processing unit provided to the video decoding device.
  • the above processing relating to encoding and decoding can be realized as a transmission, storage, and reception device using hardware, and is also stored in a ROM (Read Only Memory), a flash memory, or the like. It can also be realized by firmware or software such as a computer.
  • the firmware program and software program can be provided by recording them on a computer-readable recording medium, provided from a server through a wired or wireless network, or provided as a data broadcast of terrestrial or satellite digital broadcasting. Is also possible.
  • the present invention can be used for moving picture encoding and decoding techniques.

Landscapes

  • Engineering & Computer Science (AREA)
  • Multimedia (AREA)
  • Signal Processing (AREA)
  • Compression Or Coding Systems Of Tv Signals (AREA)

Abstract

A reference destination specifying unit (113) specifies, from among a plurality of already encoded adjacent blocks adjacent to a block to be encoded, any one of the adjacent blocks determined based on the adjacent condition between the block to be encoded and the adjacent blocks as a reference destination block used for referring to motion information. When, from among an intra-image prediction, an inter-image prediction using the motion vector of the block to be encoded, and another inter-image prediction using the motion vector of an adjacent block, the another inter-image prediction is selected as a prediction method, a prediction selecting unit (112) generates, in place of the motion information, reference destination valid information indicating whether the adjacent block referred to by the selected prediction method and the reference destination block specified by the reference destination specifying unit (113) match each other or not and information indicating the selected prediction method.

Description

動画像符号化装置、動画像符号化方法及び動画像符号化プログラム、並びに動画像復号装置、動画像復号方法及び動画像復号プログラムMoving picture encoding apparatus, moving picture encoding method, moving picture encoding program, moving picture decoding apparatus, moving picture decoding method, and moving picture decoding program
 本発明は、動画像の符号化及び復号技術に関し、特に動き補償予測を利用した動画像の符号化及び復号技術に関する。 The present invention relates to a moving picture encoding and decoding technique, and more particularly to a moving picture encoding and decoding technique using motion compensation prediction.
 MPEG(Moving Picture Experts Group)に代表される、ピクチャを矩形ブロックに分割し、ピクチャ間でブロック単位に動き推定、補償を行う動画像符号化方式では、各ブロックで生成される動きベクトルの符号量を削減する為に、動きベクトルに対して予測処理が行われる。MPEG-2では、マクロブロック単位に検出される動きベクトルは、直前に符号化されたマクロブロックの動きベクトルとの差分がとられ、その差分ベクトルを符号化することで、符号量を削減している。 In the moving picture coding method that divides a picture into rectangular blocks, represented by MPEG (Moving Picture Experts 、 Group), and performs motion estimation and compensation between blocks, the coding amount of the motion vector generated in each block In order to reduce the motion vector, a prediction process is performed on the motion vector. In MPEG-2, the motion vector detected in units of macroblocks is taken as a difference from the motion vector of the macroblock encoded immediately before, and the amount of code is reduced by encoding the difference vector. Yes.
 MPEG-4 AVC/H.264では、動きベクトルが周囲の隣接ブロックの動きベクトルと強い相関があることを利用して、周囲の隣接ブロックからの予測を行い、その差分ベクトルを符号化することによって符号量を削減している。具体的には、処理対象のブロックの左、上及び右上の隣接ブロックの動きベクトルから中央値を算出し、その中央値との差分をとることで動きベクトルの予測を実現している。 MPEG-4 AVC / H. H.264 uses the fact that motion vectors have a strong correlation with the motion vectors of neighboring blocks, and performs prediction from neighboring blocks and encodes the difference vector to reduce the amount of code. . Specifically, the median is calculated from the motion vectors of adjacent blocks on the left, top and top right of the block to be processed, and motion vector prediction is realized by taking the difference from the median.
 この予測により、動きベクトルの符号量は削減されるものの、他の動き情報は処理対象のブロック毎に符号化されるので、周囲の隣接ブロックと同一の動き情報を有していても、重複して符号化されることになり、効率的な符号化に至っていないという課題があった。この課題に対して、特許文献1や最近のISO/IEC及びITU-Tにおける動画像符号化の規格作業の中で、処理対象のブロックと周囲の既符号化済みの隣接ブロックの動き情報が同一であれば、処理対象のブロックは自身の動き情報を符号化せずに、隣接ブロックの動き情報を符号化に使用することにし、参照する動き情報を備える隣接ブロックを指定する付加情報を符号化することで、動き情報の符号量を削減することが試みられている。このような手法は、マージ(merge)と呼ばれ、動き情報の符号量削減の一手法として注目されている。 Although the amount of code of the motion vector is reduced by this prediction, other motion information is encoded for each block to be processed, so even if it has the same motion information as the adjacent neighboring blocks, it overlaps. Therefore, there is a problem that efficient encoding has not been achieved. In response to this problem, the motion information of the block to be processed and the neighboring already-encoded neighboring blocks are the same in the standard work of moving picture coding in Patent Document 1 and recent ISO / IEC and ITU-T. If so, the block to be processed does not encode its own motion information, uses the motion information of the adjacent block for encoding, and encodes additional information specifying the adjacent block having the motion information to be referenced. Thus, attempts have been made to reduce the amount of code of motion information. Such a method is called “merge” and has attracted attention as a method for reducing the code amount of motion information.
特開平10-276439号公報JP-A-10-276439
 上述したマージ手法の一例として、処理対象のブロックに対して、既に符号化済みの左に隣接するブロック或いは真上に隣接するブロックの動き情報を参照するモデルが挙げられる。この場合、ブロック毎に付加情報として、動きベクトル検出により検出された処理対象のブロック自身の動き情報を使用するか、或いはマージ手法により隣接ブロックの動き情報を使用するかを指定する1bitのフラグ(以下、merge_flagと呼ぶ)と、マージ手法を適用する場合に、隣接ブロックのうち左或いは真上のどちらを参照するかを指定する1bitのフラグ(以下、merge_lu_flagと呼ぶ)を定義する。これらのフラグに対して算術符号によるエントロピー符号化が行われる。一般に、エントロピー符号化では出現率の高いデータに短い符号、逆に出現率が低いデータに長い符号を割り当ててデータ圧縮を行い符号量の削減を行う。しかしながら、フラグmerge_lu_flagは画面内のテクスチャにも因るが、参照先の隣接ブロックとして左或いは上を選択に予め符号を割り当てているので、参照先選択の発生頻度に偏りがある場合であっても、等確率で符号化してしまうために十分な符号量削減がなされないという課題があった。 As an example of the merging method described above, there is a model that refers to motion information of a block adjacent to the left that has already been encoded or a block immediately adjacent to the block to be processed. In this case, a 1-bit flag (whether to use the motion information of the block to be processed itself detected by motion vector detection or to use the motion information of an adjacent block by the merge method as additional information for each block) In the following, a 1-bit flag (hereinafter referred to as merge_lu_flag) that defines whether to refer to the left or right of neighboring blocks when the merge method is applied is defined. Entropy coding using arithmetic codes is performed on these flags. In general, in entropy coding, a short code is assigned to data having a high appearance rate, and a long code is assigned to data having a low appearance rate, and data compression is performed to reduce the code amount. However, although the flag merge_lu_flag depends on the texture in the screen, since a code is assigned in advance to select the left or top as the reference destination adjacent block, even if the occurrence frequency of the reference destination selection is biased Therefore, there is a problem that the code amount cannot be sufficiently reduced because the encoding is performed with equal probability.
 本発明はこうした状況に鑑みてなされたものであり、その目的は、動き情報の参照先を表すフラグの符号量の削減を図って符号化効率を向上させる技術を提供することにある。 The present invention has been made in view of such circumstances, and an object of the present invention is to provide a technique for improving the coding efficiency by reducing the code amount of a flag representing a reference destination of motion information.
 上記課題を解決するために、本発明のある態様の動画像符号化装置は、動画像の各ピクチャを分割したブロック単位で動きベクトルを用いて前記動画像を符号化する動画像符号化装置であって、符号化対象ブロックに隣接する複数の符号化済みの隣接ブロックの中から、前記符号化対象ブロックと前記隣接ブロックとの隣接条件に基づいて判定されるいずれか一つの隣接ブロックを、動き情報を参照するための参照先ブロックとして指定する参照先指定部(113)と、画像内予測、符号化対象ブロックの動きベクトルを利用した画像間予測、および隣接ブロックの動きベクトルを利用した画像間予測の中から、予測方法として、隣接ブロックの動きベクトルを利用した画像間予測が選択された場合、選択された予測方法により参照される隣接ブロックと前記参照先指定部により指定された前記参照先ブロックとが一致するか否かを表す参照先有効情報と選択された予測方法を示す情報とを動き情報に代わって生成する予測選択部(112)とを含む。 In order to solve the above-described problem, a moving image encoding device according to an aspect of the present invention is a moving image encoding device that encodes the moving image using a motion vector in units of blocks obtained by dividing each picture of the moving image. Any one of the adjacent blocks determined based on an adjacency condition between the encoding target block and the adjacent block is moved from among a plurality of encoded adjacent blocks adjacent to the encoding target block. A reference destination designating unit (113) that designates a reference block for referring to information, intra-picture prediction, inter-picture prediction using a motion vector of an encoding target block, and inter-picture using a motion vector of an adjacent block When the inter-picture prediction using the motion vector of the adjacent block is selected as the prediction method from the prediction, it is referred to by the selected prediction method. A prediction selection unit that generates, instead of motion information, reference destination valid information indicating whether or not a contact block matches the reference destination block specified by the reference destination specifying unit, and information indicating the selected prediction method (112).
 本発明の別の態様は、動画像符号化方法である。この方法は、動画像の各ピクチャを分割したブロック単位で動きベクトルを用いて前記動画像を符号化する動画像符号化方法であって、符号化対象ブロックに隣接する複数の符号化済みの隣接ブロックの中から、前記符号化対象ブロックと前記隣接ブロックとの隣接条件に基づいて判定されるいずれか一つの隣接ブロックを、動き情報を参照するための参照先ブロックとして指定する参照先指定ステップと、画像内予測、符号化対象ブロックの動きベクトルを利用した画像間予測、および隣接ブロックの動きベクトルを利用した画像間予測の中から、予測方法として、隣接ブロックの動きベクトルを利用した画像間予測が選択された場合、選択された予測方法により参照される隣接ブロックと前記参照先指定ステップにより指定された前記参照先ブロックとが一致するか否かを表す参照先有効情報と選択された予測方法を示す情報とを動き情報に代わって生成する予測選択ステップとを含む。 Another aspect of the present invention is a video encoding method. This method is a moving image encoding method for encoding the moving image using a motion vector in units of blocks obtained by dividing each picture of the moving image, and a plurality of encoded adjacent regions adjacent to the encoding target block. A reference destination designating step of designating any one adjacent block determined based on an adjacency condition between the encoding target block and the adjacent block as a reference destination block for referring to motion information from among the blocks; Intra-picture prediction, inter-picture prediction using the motion vector of the encoding target block, and inter-picture prediction using the motion vector of the adjacent block, inter-picture prediction using the motion vector of the adjacent block as a prediction method Is selected, the neighboring block referred to by the selected prediction method and the reference designated by the reference destination designation step. And a prediction selection step of the previous block is produced on behalf of the referenced enable information and motion information indicating the selected prediction method information indicating whether match.
 本発明のある態様の動画像復号装置は、動画像の各ピクチャを分割したブロック単位で動きベクトルを用いて前記動画像が符号化されたビットストリームを復号する動画像復号装置であって、予測方法を示す情報と選択された予測方法により参照される隣接ブロックの有効性を示す参照先有効情報とが含まれる前記ビットストリームから前記予測方法を示す情報と前記参照先有効情報とを復号する可変長復号部(201)と、画像内予測、復号対象ブロックの動きベクトルを利用した画像間予測、および前記復号対象ブロックに隣接する複数の復号済みの隣接ブロックのいずれか一つの動きベクトルを利用した画像間予測の中から、前記予測方法を示す情報に基づいていずれかの予測方法を選択する予測選択部(209)と、前記予測選択部により、予測方法として、隣接ブロックの動きベクトルを利用した画像間予測が選択された場合、前記復号対象ブロックと前記隣接ブロックとの隣接条件に基づいて判定されるいずれか一つの隣接ブロックを、動き情報を参照するための参照先ブロックとして指定できるか否かを前記参照先有効情報に基づいて判定する動き情報選択部(210)とを含む。 A moving picture decoding apparatus according to an aspect of the present invention is a moving picture decoding apparatus that decodes a bitstream in which the moving picture is encoded using a motion vector in units of blocks obtained by dividing each picture of the moving picture. A variable for decoding the information indicating the prediction method and the reference destination valid information from the bitstream including information indicating the method and reference destination valid information indicating the validity of the adjacent block referred to by the selected prediction method Using the long decoding unit (201), intra-picture prediction, inter-picture prediction using a motion vector of a decoding target block, and one motion vector of a plurality of decoded adjacent blocks adjacent to the decoding target block A prediction selection unit (209) that selects one of prediction methods based on information indicating the prediction method from among the inter-image predictions; and the prediction selection Therefore, when inter-picture prediction using a motion vector of an adjacent block is selected as a prediction method, any one adjacent block determined based on an adjacent condition between the decoding target block and the adjacent block is moved. A motion information selection unit (210) that determines whether it can be designated as a reference block for referring to information based on the reference destination valid information.
 本発明の別の態様は、動画像復号方法である。この方法は、動画像の各ピクチャを分割したブロック単位で動きベクトルを用いて前記動画像が符号化されたビットストリームを復号する動画像復号方法であって、予測方法を示す情報と選択された予測方法により参照される隣接ブロックの有効性を示す参照先有効情報とが含まれる前記ビットストリームから前記予測方法を示す情報と前記参照先有効情報とを復号する可変長復号ステップと、画像内予測、復号対象ブロックの動きベクトルを利用した画像間予測、および前記復号対象ブロックに隣接する複数の復号済みの隣接ブロックのいずれか一つの動きベクトルを利用した画像間予測の中から、前記予測方法を示す情報に基づいていずれかの予測方法を選択する予測選択ステップと、前記予測選択ステップにより、予測方法として、隣接ブロックの動きベクトルを利用した画像間予測が選択された場合、前記復号対象ブロックと前記隣接ブロックとの隣接条件に基づいて判定されるいずれか一つの隣接ブロックを、動き情報を参照するための参照先ブロックとして指定できるか否かを前記参照先有効情報に基づいて判定する動き情報選択ステップとを含む。 Another aspect of the present invention is a video decoding method. This method is a moving image decoding method for decoding a bitstream in which the moving image is encoded using a motion vector in units of blocks obtained by dividing each picture of the moving image, and is selected as information indicating a prediction method A variable length decoding step for decoding the information indicating the prediction method and the reference destination valid information from the bitstream including the reference destination valid information indicating the validity of the adjacent block referred to by the prediction method, and intra prediction The prediction method includes: inter-picture prediction using a motion vector of a decoding target block; and inter-picture prediction using any one of a plurality of decoded adjacent blocks adjacent to the decoding target block. A prediction selection step of selecting one of the prediction methods based on the information indicated, and the prediction selection step, When inter-picture prediction using a motion vector of a block is selected, a reference for referring to motion information for any one adjacent block determined based on an adjacent condition between the decoding target block and the adjacent block A motion information selection step of determining whether or not it can be designated as a destination block based on the reference destination valid information.
 なお、以上の構成要素の任意の組合せ、本発明の表現を方法、装置、システム、記録媒体、コンピュータプログラムなどの間で変換したものもまた、本発明の態様として有効である。 It should be noted that an arbitrary combination of the above-described components and a conversion of the expression of the present invention between a method, an apparatus, a system, a recording medium, a computer program, and the like are also effective as an aspect of the present invention.
 本発明によれば、動き情報の参照先を表すフラグの符号量の削減を図って符号化効率を向上させることができる。 According to the present invention, it is possible to improve the coding efficiency by reducing the code amount of the flag representing the reference destination of the motion information.
実施の形態に係る動き情報の参照先予測方法を具備した動画像符号化装置の構成を示すブロック図である。It is a block diagram which shows the structure of the moving image encoder which comprised the reference destination prediction method of the motion information which concerns on embodiment. 実施の形態に係る動き情報の参照先予測方法を具備した動画像復号装置の構成を示すブロック図である。It is a block diagram which shows the structure of the moving image decoding apparatus which comprised the reference destination prediction method of the motion information which concerns on embodiment. 実施の形態における符号化ブロックを説明する為の図である。It is a figure for demonstrating the encoding block in embodiment. 実施の形態における予測ブロックの形状の種類を説明する為の図である。It is a figure for demonstrating the kind of shape of the prediction block in embodiment. 実施の形態における動き情報の参照先予測方法の処理対象となるパーティションの位置を説明する為の図である。It is a figure for demonstrating the position of the partition used as the process target of the reference destination prediction method of the motion information in embodiment. 実施の形態における動き情報の参照先予測方法の処理対象となるパーティションの周辺を説明する為の図である。It is a figure for demonstrating the periphery of the partition used as the process target of the reference destination prediction method of the motion information in embodiment. 実施の形態における動き情報の参照先予測方法をシーケンスレベルで実行するか否かを決定するビットストリームのシンタックスパターンを示す図である。It is a figure which shows the syntax pattern of the bit stream which determines whether the reference destination prediction method of the motion information in embodiment is performed at a sequence level. 実施の形態における動き情報の参照先予測方法を予測ブロックレベルで実行するか否かを決定するビットストリームのシンタックスパターンを示す図である。It is a figure which shows the syntax pattern of the bit stream which determines whether the reference destination prediction method of the motion information in embodiment is performed at a prediction block level. 図1の参照先指定部の動作を説明するフローチャートである。It is a flowchart explaining operation | movement of the reference designation | designated part of FIG. 図10(a)は、図9の候補数NumMergeCandidatesの算出処理の詳細を示すフローチャートであり、図10(b)は、処理対象のパーティションに隣接する参照先パーティションの例を示す図である。FIG. 10A is a flowchart showing details of the calculation processing of the number of candidates NumMergeCandidates in FIG. 9, and FIG. 10B is a diagram showing an example of a reference destination partition adjacent to the processing target partition. 図1の予測選択部の動作を説明するフローチャートである。It is a flowchart explaining operation | movement of the prediction selection part of FIG. 図2の動き情報選択部の動作を説明するフローチャートである。It is a flowchart explaining operation | movement of the motion information selection part of FIG. 図13(a)、図13(b)は、予測された参照先を示すフラグの値に応じて実行される処理を説明するフローチャートである。FIG. 13A and FIG. 13B are flowcharts for explaining processing executed in accordance with a flag value indicating a predicted reference destination. 本実施の形態における動き情報の参照先予測方法の第1の実施例の動作を説明するフローチャートである。It is a flowchart explaining the operation | movement of the 1st Example of the reference destination prediction method of the motion information in this Embodiment. 図14の隣接するパーティションとの境界判定の詳細動作を説明するフローチャートである。It is a flowchart explaining the detailed operation | movement of a boundary determination with the adjacent partition of FIG. 処理対象のパーティションと隣接するパーティションの配置と定義を説明する図である。It is a figure explaining arrangement | positioning and a definition of the partition adjacent to the partition of a process target. 隣接するパーティションが複数配置された場合の動き境界の定義を説明する図である。It is a figure explaining the definition of a motion boundary when two or more adjacent partitions are arrange | positioned. 図14の隣接するパーティションとの隣接辺長比較の詳細動作を説明するフローチャートである。It is a flowchart explaining the detailed operation | movement of adjacent edge length comparison with the adjacent partition of FIG. 処理対象パーティションと隣接するパーティションの隣接条件に基づいて判定される参照先パーティションを説明する図である。It is a figure explaining the reference destination partition determined based on the adjacency condition of the partition adjacent to a process target partition. 処理対象パーティションと隣接するパーティションの隣接条件の一例を示す図である。It is a figure which shows an example of the adjacency conditions of the partition adjacent to a process target partition. 処理対象パーティションと隣接するパーティションの隣接条件の別の例を示す図である。It is a figure which shows another example of the adjacency conditions of the partition adjacent to a process target partition. 動き情報の参照先予測方法の第2の実施例の動作を説明するフローチャートである。It is a flowchart explaining operation | movement of the 2nd Example of the reference destination prediction method of motion information. 隣接するパーティションの境界残差を算出する画素領域を示す図である。It is a figure which shows the pixel area | region which calculates the boundary residual of an adjacent partition. 隣接するパーティションの境界残差に基づく参照先パーティションの判定の動作を説明するフローチャートである。It is a flowchart explaining the operation | movement of the determination of the reference destination partition based on the boundary residual of an adjacent partition. 動き情報の参照先予測方法の第3の実施例において、予測ブロックレベルで実行するか否かを決定するビットストリームのシンタックスパターンを示す図である。It is a figure which shows the syntax pattern of the bit stream which determines whether it performs in the prediction block level in the 3rd Example of the reference destination prediction method of motion information. 動き情報の参照先予測方法の第3の実施例の動作を説明するフローチャートである。It is a flowchart explaining operation | movement of the 3rd Example of the reference destination prediction method of motion information. 第3の実施例の動き情報選択部の動作を説明するフローチャートである。It is a flowchart explaining operation | movement of the motion information selection part of a 3rd Example.
 本発明の実施の形態は、動画像の符号化技術に関し、特にピクチャを矩形ブロックに分割し、ピクチャ間でブロック単位に動き推定、補償を行う動画像符号化における動き情報の符号化効率を向上させる為に、処理対象のブロックの周囲の符号化済みの隣接ブロックの動きベクトルや参照先ピクチャ番号等の動き情報を処理対象のブロックの動き情報として使用することで、処理対象のブロックの動き情報を符号化せず、参照する隣接ブロックを示す付加情報を符号化することによって符号量を削減する技術に関する。 Embodiments of the present invention relate to moving picture coding technology, and in particular, improve the efficiency of motion information coding in moving picture coding in which a picture is divided into rectangular blocks and motion estimation and compensation are performed in units of blocks between pictures. Therefore, the motion information of the block to be processed is obtained by using the motion information such as the motion vector of the encoded adjacent block around the block to be processed and the reference picture number as the motion information of the block to be processed. The present invention relates to a technique for reducing the amount of codes by encoding additional information indicating adjacent blocks to be referred to without encoding.
 本発明の実施の形態では、等しい頻度で発生する上述のフラグmerge_lu_flagに代わり、動き情報の参照先を周辺の隣接ブロックのサイズや予測モード等の情報を用いて予測を行い、その予測により判定された参照先の隣接ブロックと、実際に符号化によって選択された隣接ブロックとの一致を表すフラグmerge_probable_flagを定義する。予測による判定率が高ければ、フラグmerge_probable_flagで表されるビットの発生確率が高くなるので、フラグを表すビット数が小さくなり、発生符号量を削減することが可能となる。 In the embodiment of the present invention, instead of the above-described flag merge_lu_flag that occurs at the same frequency, the reference destination of motion information is predicted using information such as the size of neighboring blocks and the prediction mode, and is determined by the prediction. In addition, a flag merge_proble_flag that represents a match between the adjacent block of the reference destination and the adjacent block actually selected by encoding is defined. If the determination rate by prediction is high, the probability of occurrence of the bit represented by the flag merge_proble_flag increases, so the number of bits representing the flag decreases, and the amount of generated code can be reduced.
 更に、本発明の別の実施の形態では、参照先の予測による参照先となる隣接ブロックの判定の一致精度が高いものとして、前述の参照先の予測との一致を表すフラグmerge_probable_flagそのものを定義せず、参照先の予測により得られた判定結果で参照先の隣接ブロックを示し、参照先の予測が不定の場合にのみmerge_lu_flagと同様の機能を有するフラグmerge_direc_flagで参照先の隣接ブロックを示すことで、フラグmerge_direc_flagによる符号量の増加をフラグmerge_probable_flagに要する符号量の削減で相殺し、相対的に符号化効率を向上させることも可能である。 Furthermore, in another embodiment of the present invention, the flag merge_proble_flag itself representing the match with the above-mentioned reference destination prediction is defined on the assumption that the coincidence accuracy of the determination of the adjacent block that becomes the reference destination by the reference destination prediction is high. In addition, by indicating the adjacent block of the reference destination in the determination result obtained by the prediction of the reference destination, and indicating the adjacent block of the reference destination with the flag merge_direc_flag having the same function as the merge_lu_flag only when the prediction of the reference destination is indefinite. It is also possible to offset the increase in the code amount due to the flag merge_direc_flag by reducing the code amount required for the flag merge_proble_flag, and to relatively improve the encoding efficiency.
 以下、本発明の実施の形態に係る動画像符号化装置及び動画像復号装置について図面を参照して説明する。 Hereinafter, a moving picture coding apparatus and a moving picture decoding apparatus according to embodiments of the present invention will be described with reference to the drawings.
 図1は実施の形態に係る動画像符号化装置の構成を示すブロックである。実施の形態の動画像符号化装置は、動きベクトル検出部101、減算部102、直交変換・量子化部103、可変長符号化部104、逆量子化・逆直交変換部105、動き補償部106、重み付き予測部107、加算部108、デブロッキング・フィルタ部109、メモリ110、画面内予測部111、予測選択部112、および参照先指定部113を備える。 FIG. 1 is a block diagram showing a configuration of a moving picture encoding apparatus according to an embodiment. The moving image encoding apparatus according to the embodiment includes a motion vector detection unit 101, a subtraction unit 102, an orthogonal transform / quantization unit 103, a variable length encoding unit 104, an inverse quantization / inverse orthogonal transform unit 105, and a motion compensation unit 106. , A weighted prediction unit 107, an addition unit 108, a deblocking filter unit 109, a memory 110, an in-screen prediction unit 111, a prediction selection unit 112, and a reference destination designation unit 113.
 画面内予測部111は、入力される符号化対象ブロックと、メモリ110に格納された符号化対象ブロックに隣接する符号化済みの復号ブロック画像とから、面内予測を行い、面内予測ブロックを出力する。 The intra prediction unit 111 performs intra prediction from the input encoding target block and the encoded decoded block image adjacent to the encoding target block stored in the memory 110, and determines the intra prediction block. Output.
 参照先指定部113は、メモリ110に格納された符号化対象ブロックに隣接する複数の隣接ブロックのいずれか一つを、マージ処理を行う場合の参照先の隣接ブロックとして指定し、その参照先隣接ブロックの動きベクトルを動き補償部106に与える。 The reference destination specifying unit 113 specifies any one of a plurality of adjacent blocks adjacent to the encoding target block stored in the memory 110 as a reference destination adjacent block when performing the merge processing, and the reference destination adjacent The motion vector of the block is given to the motion compensation unit 106.
 動きベクトル検出部101は、入力された画像信号と、メモリ110に格納された参照画像信号とを参照して、ピクチャ間でブロック単位にマッチングを行い動きベクトルを検出し、検出された動きベクトルを動き補償部106に与える。 The motion vector detection unit 101 refers to the input image signal and the reference image signal stored in the memory 110 to perform matching between blocks in units of blocks to detect a motion vector, and to detect the detected motion vector. This is given to the motion compensation unit 106.
 動き補償部106は、動きベクトル検出部101により検出された符号化対象ブロックの動きベクトルまたは参照先指定部113により指定された隣接ブロックの動きベクトルを用いて予測画像を生成する。重み付き予測部107は、動き補償部106により生成された予測画像に適応的に重み係数を掛け算して最終的な予測画像を生成し、予測選択部112に与える。 The motion compensation unit 106 generates a prediction image using the motion vector of the block to be encoded detected by the motion vector detection unit 101 or the motion vector of the adjacent block specified by the reference destination specifying unit 113. The weighted prediction unit 107 adaptively multiplies the prediction image generated by the motion compensation unit 106 by a weight coefficient to generate a final prediction image, and provides the prediction selection unit 112 with it.
 予測選択部112は、画面内予測部111による画面内予測、動き補償部106による符号化対象ブロックの動きベクトルを利用した画像間予測、および参照先指定部113により指定された参照先隣接ブロックの動きベクトルを利用した画像間予測の中から、符号量が最小となる予測方法を一つ選択し、選択された予測方法による予測画像を減算部102と加算部108に与える。 The prediction selection unit 112 performs intra-screen prediction by the intra-screen prediction unit 111, inter-image prediction using the motion vector of the block to be encoded by the motion compensation unit 106, and the reference destination adjacent block specified by the reference destination specifying unit 113. One prediction method with the smallest code amount is selected from the inter-image prediction using the motion vector, and a prediction image by the selected prediction method is given to the subtraction unit 102 and the addition unit 108.
 符号量が最小となる予測方法として、参照先隣接ブロックの動きベクトルを利用した画像間予測が予測選択部112によって選択された場合、参照先指定部113は、マージ処理の参照先の隣接ブロックを予測する処理を実行し、マージ参照先の予測結果を予測選択部112に与える。予測選択部112は、参照先指定部113により予測された参照先隣接ブロックと、実際に選択された予測方法で利用された参照先隣接ブロックとが一致するかどうかを判定し、参照先隣接ブロックの予測結果の有効性を示す参照先有効情報を、選択された予測方法を示す情報とともに可変長符号化部104に供給する。 As a prediction method that minimizes the code amount, when the inter-picture prediction using the motion vector of the reference destination adjacent block is selected by the prediction selection unit 112, the reference destination specifying unit 113 selects the adjacent block that is the reference destination of the merge process. The prediction process is executed, and the prediction result of the merge reference destination is given to the prediction selection unit 112. The prediction selecting unit 112 determines whether or not the reference destination adjacent block predicted by the reference destination specifying unit 113 matches the reference destination adjacent block used in the actually selected prediction method, and the reference destination adjacent block The reference destination valid information indicating the validity of the prediction result is supplied to the variable length coding unit 104 together with the information indicating the selected prediction method.
 減算部102は、符号化する画像と予測画像との引き算を行い残差信号を生成し、直交変換・量子化部103に与える。直交変換・量子化部103は、残差信号に対して直交変換及び量子化を行い変換信号を生成し、可変長符号化部104と逆量子化・逆直交変換部105に与える。可変長符号化部104は、直交変換及び量子化された残差信号をエントロピー符号化する。可変長符号化部104は、予測選択部112によって生成される参照先有効情報と選択された予測方法を示す情報も符号化し、符号化画像を含むビットストリームを出力する。 The subtraction unit 102 generates a residual signal by subtracting the image to be encoded and the predicted image, and supplies the residual signal to the orthogonal transform / quantization unit 103. The orthogonal transform / quantization unit 103 performs orthogonal transform and quantization on the residual signal, generates a transform signal, and supplies the transform signal to the variable length coding unit 104 and the inverse quantization / inverse orthogonal transform unit 105. The variable length coding unit 104 entropy codes the residual signal that has been orthogonally transformed and quantized. The variable length encoding unit 104 also encodes reference destination valid information generated by the prediction selection unit 112 and information indicating the selected prediction method, and outputs a bit stream including an encoded image.
 逆量子化・逆直交変換部105は、直交変換・量子化部103から受け取った変換信号を逆量子化及び逆直交変換して元の残差信号に戻す。加算部108は、予測画像と残差信号を加算して復号画像を生成し、デブロッキング・フィルタ部109に与える。デブロッキング・フィルタ部109は、復号画像に対して符号化によるブロック歪を減少させる処理を施し、メモリ110に格納する。メモリ110は、復号画像や既に符号化された画像の情報を記憶する。 The inverse quantization / inverse orthogonal transform unit 105 performs inverse quantization and inverse orthogonal transform on the transform signal received from the orthogonal transform / quantization unit 103 to return to the original residual signal. The adding unit 108 adds the predicted image and the residual signal, generates a decoded image, and supplies the decoded image to the deblocking filter unit 109. The deblocking filter unit 109 performs a process of reducing block distortion due to encoding on the decoded image and stores it in the memory 110. The memory 110 stores information on decoded images and already encoded images.
 図2は図1の動画像符号化装置に対応した実施の形態に係る動画像復号装置の構成を示すブロックである。実施の形態の動画像復号装置は、可変長復号部201、逆量子化・逆直交変換部202、動き補償部203、重み付き予測部204、加算部205、デブロッキング・フィルタ部206、メモリ207、画面内予測部208、画面内/画面間予測選択部209、および動き情報選択部210を備える。 FIG. 2 is a block diagram showing a configuration of a moving picture decoding apparatus according to an embodiment corresponding to the moving picture encoding apparatus of FIG. The moving picture decoding apparatus according to the embodiment includes a variable length decoding unit 201, an inverse quantization / inverse orthogonal transform unit 202, a motion compensation unit 203, a weighted prediction unit 204, an addition unit 205, a deblocking filter unit 206, and a memory 207. , An intra-screen prediction unit 208, an intra-screen / inter-screen prediction selection unit 209, and a motion information selection unit 210.
 図2の動画像復号装置の復号処理は、図1の動画像符号化装置の内部に設けられている復号処理に対応するものであるから、図2の逆量子化・逆直交変換部202、動き補償部203、重み付き予測部204、加算部205、デブロッキング・フィルタ部206、メモリ207、および画面内予測部208の各構成は、図1の動画像符号化装置の逆量子化・逆直交変換部105、動き補償部106、重み付き予測部107、加算部108、デブロッキング・フィルタ部109、メモリ110、および画面内予測部111の各構成とそれぞれ対応する機能を有する。 The decoding process of the moving picture decoding apparatus in FIG. 2 corresponds to the decoding process provided in the moving picture encoding apparatus in FIG. 1, so that the inverse quantization / inverse orthogonal transform unit 202 in FIG. Each configuration of the motion compensation unit 203, the weighted prediction unit 204, the addition unit 205, the deblocking filter unit 206, the memory 207, and the in-screen prediction unit 208 is the inverse quantization / inverse of the moving picture coding apparatus in FIG. The orthogonal transform unit 105, the motion compensation unit 106, the weighted prediction unit 107, the addition unit 108, the deblocking filter unit 109, the memory 110, and the in-screen prediction unit 111 have functions corresponding to the respective configurations.
 可変長復号部201は、ビットストリームを復号して予測残差信号、動きベクトルに関する情報を出力し、予測残差信号を逆量子化・逆直交変換部202に与え、動きベクトルに関する情報を動き情報選択部210に、符号化モードに関する情報を画面内/画面間予測選択部209に与える。 The variable length decoding unit 201 decodes the bitstream and outputs information on the prediction residual signal and the motion vector, provides the prediction residual signal to the inverse quantization / inverse orthogonal transform unit 202, and provides information on the motion vector as the motion information. Information about the coding mode is given to the selection unit 210 to the intra-screen / inter-screen prediction selection unit 209.
 逆量子化・逆直交変換部202は、可変長復号部201で復号された予測残差信号に対して逆直交変換及び逆量子化を行う。加算部205は、逆量子化・逆直交変換部202により逆変換された予測残差成分と、重み付き予測部204により算出された予測画像とを加算することにより、画像信号を復号し、デブロッキング・フィルタ部206に与える。デブロッキング・フィルタ部206は、復号画像に対して符号化によるブロック歪を減少させる処理を施し、メモリ207に格納する。 The inverse quantization / inverse orthogonal transform unit 202 performs inverse orthogonal transform and inverse quantization on the prediction residual signal decoded by the variable length decoding unit 201. The addition unit 205 decodes the image signal by adding the prediction residual component inversely transformed by the inverse quantization / inverse orthogonal transformation unit 202 and the prediction image calculated by the weighted prediction unit 204, and decodes the image signal. This is applied to the blocking filter unit 206. The deblocking filter unit 206 performs a process of reducing block distortion due to encoding on the decoded image and stores it in the memory 207.
 画面内/画面間予測選択部209は、符号化モードが画面内予測か画面間予測かを判定し、画面内予測であれば、画面内予測部208に画面内予測処理を実行するよう指示し、画面間予測であれば、動き情報選択部210に画面間予測処理を実行するよう指示する。 The intra-screen / inter-screen prediction selection unit 209 determines whether the encoding mode is intra-screen prediction or inter-screen prediction, and if it is intra-screen prediction, instructs the intra-screen prediction unit 208 to execute the intra-screen prediction process. In the case of inter-screen prediction, the motion information selection unit 210 is instructed to execute inter-screen prediction processing.
 動き情報選択部210は、動きベクトルなどの動き情報がマージ処理によるものであるか否かを選択し、マージ処理によるものである場合、参照先有効情報にもとづいて、所定の隣接条件にもとづいて判定される隣接ブロックを参照先ブロックとして指定できるかどうかを判定する。 The motion information selection unit 210 selects whether or not motion information such as a motion vector is due to merge processing. If it is due to merge processing, the motion information selection unit 210 is based on predetermined adjacent conditions based on reference destination valid information. It is determined whether the adjacent block to be determined can be designated as a reference block.
 画面内予測部208は、入力される復号対象ブロックと、メモリ207に格納された復号対象ブロックに隣接する復号ブロック画像とから、面内予測を行い、面内予測ブロックを加算部205に渡す。 The intra prediction unit 208 performs intra prediction from the input decoding target block and the decoded block image adjacent to the decoding target block stored in the memory 207, and passes the intra prediction block to the adding unit 205.
 動き補償部203は、画面間予測がマージ処理でない、すなわち通常の動きベクトル検出処理によるものである場合、復号対象ブロックの復号された動きベクトルを用いて予測画像を生成する。動き補償部203は、画面間予測がマージ処理によるものであり、参照先有効情報が肯定的である場合、動き情報の参照先予測により予測される参照先の隣接ブロックの動きベクトルを用いて予測画像を生成する。 The motion compensation unit 203 generates a prediction image using the decoded motion vector of the decoding target block when the inter-screen prediction is not merge processing, that is, is based on normal motion vector detection processing. When the inter-frame prediction is based on the merge process and the reference destination valid information is affirmative, the motion compensation unit 203 performs prediction using the motion vector of the reference destination adjacent block predicted by the reference destination prediction of the motion information. Generate an image.
 重み付き予測部204は、動き補償部203により生成された予測画像に適応的に重み係数を掛け算して最終的な予測画像を生成し、加算部205に与える。 The weighted prediction unit 204 adaptively multiplies the prediction image generated by the motion compensation unit 203 by a weighting factor to generate a final prediction image, which is given to the addition unit 205.
 実施の形態に係る参照先の隣接ブロックの予測方法は、図1の動画像符号化装置の予測選択部112及び図2の動画像復号装置の動き情報選択部210において実施される。 The prediction method of an adjacent block of a reference destination according to the embodiment is performed in the prediction selection unit 112 of the video encoding device in FIG. 1 and the motion information selection unit 210 of the video decoding device in FIG.
 動画像符号化装置の予測選択部112では、通常の動きベクトル検出処理により検出された動き情報またはマージ処理により選択された隣接ブロックの動き情報に基づく画面間予測、および画面内予測の中から、発生符号量が最も少ないか、画質が最も良い、あるいはその両方を満たす符号化モードが一つ選択される。さらに、選択された符号化モードが画面間予測であれば、動き情報が通常の動きベクトル検出処理により検出されたものであるか、マージ処理により選択されたものであるかを判定し、マージ処理の場合は、動き情報の参照先予測が行われ、予測された参照先の隣接ブロックとマージ処理により選択された隣接ブロックとの同一性を判定し、その結果が参照先有効情報として符号化・伝送される。マージ処理でない場合は、通常の動きベクトルの予測が行われ、動きベクトルと予測動きベクトルとの差分動きベクトルとその他の動き情報が符号化・伝送される。 In the prediction selection unit 112 of the video encoding device, from among the motion information detected by the normal motion vector detection process or the inter-screen prediction based on the motion information of the adjacent block selected by the merge process, and the intra-screen prediction, One encoding mode is selected that has the least amount of generated code, the best image quality, or both. Further, if the selected encoding mode is inter-screen prediction, it is determined whether the motion information is detected by a normal motion vector detection process or is selected by a merge process, and the merge process is performed. In this case, the reference destination prediction of the motion information is performed, the identity between the predicted reference destination adjacent block and the adjacent block selected by the merging process is determined, and the result is encoded as reference destination valid information. Is transmitted. When the merge process is not performed, normal motion vector prediction is performed, and a difference motion vector between the motion vector and the predicted motion vector and other motion information are encoded and transmitted.
 一方、動画像復号装置の動き情報選択部210では、可変長復号部201にてビットストリームから復号された符号化モードや動き情報に基づいて、画面内/画面間予測選択部209にて画面間予測と判定された場合、更にその画面間予測がマージ処理によるものか否かが判定され、マージ処理によるものであると判定された場合、動き情報の参照先予測が行われ、この予測方法により予測された参照先の隣接ブロックと復号された付加情報で指定される隣接ブロックとの同一性を判定することで動き情報を取得する。以下の実施例では、動き情報の参照先予測の詳細について説明する。 On the other hand, in the motion information selection unit 210 of the video decoding device, the intra-screen / inter-screen prediction selection unit 209 determines the inter-screen based on the encoding mode and motion information decoded from the bit stream by the variable length decoding unit 201. If it is determined to be prediction, it is further determined whether or not the inter-screen prediction is due to merge processing. If it is determined to be due to merge processing, motion information reference destination prediction is performed. The motion information is acquired by determining the identity between the predicted reference adjacent block and the adjacent block specified by the decoded additional information. In the following embodiments, details of motion information reference destination prediction will be described.
[実施例1]
 動き情報の参照先予測方法の実施例を説明する前に、本実施例において使用する用語を説明しておく。
[Example 1]
Prior to describing an example of a motion information reference destination prediction method, terms used in this example will be described.
 (符号化ブロックについて)
 実施の形態では、図3に示されるように、画面内を同一サイズの正方の矩形ブロックにて均等分割する。このブロックを符号化ブロックと呼び、符号化及び復号を行う際の処理の基本となる。符号化ブロックは画面内のテクスチャに応じて、符号化処理を最適にすべく、符号化ブロック内を4分割して、ブロックサイズの小さい符号化ブロックにすることが出来る。図3で示される画面内を均等サイズで分割される符号化ブロックを最大符号化ブロックと呼び、その内部を符号化条件に応じて4分割したものを総じて符号化ブロックとする。符号化ブロックをこれ以上4分割出来ない最小のサイズとなる符号化ブロックを最小符号化ブロックと呼ぶこととする。
(About coding block)
In the embodiment, as shown in FIG. 3, the screen is equally divided into square rectangular blocks of the same size. This block is called a coding block and is the basis of processing when performing coding and decoding. According to the texture in the screen, the coding block can be divided into four blocks with a small block size in order to optimize the coding process. A coded block that is divided into equal screen sizes in the screen shown in FIG. 3 is called a maximum coded block, and the inside of which is divided into four according to the coding conditions is collectively called a coded block. An encoded block having a minimum size that cannot be further divided into four is referred to as a minimum encoded block.
 (予測ブロックについて)
 画面内をブロックに分割して動き補償を行う場合、動き補償のブロックサイズをより小さくした方が、よりきめ細かい予測を行うことが出来ることから、いくつかのブロックサイズの中から最適なものを選択して、ブロック内部を分割して動き補償を行う仕組みを取り入れている。この動き補償を行うブロックを予測ブロックと呼ぶ。予測ブロックは符号化ブロックと同一のサイズで表され、動き補償に応じて、予測ブロック内部を分割せず1ブロックとみなす場合を最大とし、水平或いは垂直方向に2分割したもの、水平と垂直の均等分割により4分割したものに分けられる。分割後のサイズに応じて、分割タイプに対応したモードが定義されており、図4に示される。
(About prediction block)
When motion compensation is performed by dividing the screen into blocks, the smaller the motion compensation block size, the more detailed prediction can be made, so select the optimal block size from several block sizes. Thus, a mechanism for performing motion compensation by dividing the inside of the block is adopted. A block that performs this motion compensation is called a prediction block. The prediction block is represented by the same size as the coding block, and according to the motion compensation, the prediction block is regarded as one block without being divided, and is divided into two in the horizontal or vertical direction. Divided into four parts by equal division. The mode corresponding to the division type is defined according to the size after the division, and is shown in FIG.
 (パーティションについて)
 予測ブロックを分割した個々の領域はパーティションと呼ばれる。予測ブロック内部において、パーティションを管理する為に、0から開始する番号を、予測ブロック内部に存在するパーティションに対してジグザグスキャン順に割り当てる。この番号はパーティション番号と呼ばれ、puPartIdxで表される。図4の予測ブロックのパーティションの中に記述された数字は、そのパーティションのパーティション番号を表す。
(About partitions)
Each area obtained by dividing the prediction block is called a partition. In order to manage the partitions in the prediction block, numbers starting from 0 are assigned to the partitions existing in the prediction block in the zigzag scan order. This number is called a partition number and is represented by puPartIdx. The number described in the partition of the prediction block in FIG. 4 represents the partition number of the partition.
 実施の形態による動き情報の参照先予測方法を図面を用いて説明する。動き情報の参照先予測方法は、予測ブロックを構成するパーティション単位に、符号化及び復号の処理の何れでも実施される。 A motion information reference destination prediction method according to an embodiment will be described with reference to the drawings. The motion information reference destination prediction method is performed in any of encoding and decoding processes for each partition constituting a prediction block.
 図5に示されるように、同一ピクチャ内の符号化ブロック内部を動き補償する為に定義される予測ブロックのパーティション(図5中の処理対象のパーティション)に隣接する周囲のパーティションから動き情報の参照先となるパーティションを選択する。 As shown in FIG. 5, the motion information is referred from the surrounding partition adjacent to the partition of the prediction block (partition to be processed in FIG. 5) defined for motion compensation within the coding block in the same picture. Select the destination partition.
 参照先となるパーティションは、復号においても使用することから、処理対象のパーティションの符号化の前に既に符号化されたパーティションが候補となる。符号化は画面の左上から右下にラスタスキャン順にブロック単位で行うことから、ここでは原則的に処理対象のパーティションの周囲に隣接する左或いは真上のパーティションを参照先の候補として考えることとし、以下断りが無い限りこの仮定を前提として説明することとする。 Since the partition to be referred to is also used in decoding, a partition that has already been encoded before encoding the partition to be processed is a candidate. Since encoding is performed in block units in the raster scan order from the upper left to the lower right of the screen, here, in principle, the left or upper partition adjacent to the periphery of the partition to be processed is considered as a reference destination candidate. The following explanation will be based on this assumption unless otherwise noted.
 図5中の太点線の円内で表される領域を拡大したものが図6である。 FIG. 6 is an enlarged view of the area represented by the thick dotted circle in FIG.
 図6は、処理対象のパーティションとそれに隣接するパーティションを示す。予測動きベクトルの候補は、処理対象のパーティションの左側に隣接するパーティションAk(k=0,…,nA-1)から構成されるパーティション群A、および真上に隣接するパーティションBk(k=0,…,nB-1)から構成されるパーティション群Bの2つのパーティション群からそれぞれ動き情報の参照先となるパーティションを選択する。ここで、nAは左に隣接するパーティション群の総数、nBは真上に隣接するパーティション群の総数を表す。 FIG. 6 shows a partition to be processed and a partition adjacent to it. Predicted motion vector candidates include a partition group A composed of a partition Ak (k = 0,..., NA−1) adjacent to the left side of the partition to be processed, and a partition Bk (k = 0, .., NB-1) are selected from the two partition groups of partition group B, each of which is a reference destination of motion information. Here, nA represents the total number of partition groups adjacent to the left, and nB represents the total number of partition groups adjacent immediately above.
 パーティションのサイズは動き補償によって変わるため、図6に示されるように、処理対象のパーティションとその周囲に隣接するパーティションのサイズが異なる場合は、次の規則に基づいて左或いは真上の参照先の候補となるパーティションを決定する。 Since the size of the partition changes depending on motion compensation, as shown in FIG. 6, when the size of the partition to be processed and the neighboring partitions are different, the reference destination on the left or directly above is determined based on the following rule. Determine candidate partitions.
 ・処理対象のパーティションに対して左に複数の隣接するパーティションが存在する場合は、その中の一番上のパーティションA0を参照先の候補とする。
 ・処理対象のパーティションに対して真上に複数の隣接するパーティションが存在する場合は、その中の一番左のパーティションB0を参照先の候補とする。
When there are a plurality of adjacent partitions on the left with respect to the partition to be processed, the top partition A0 among them is set as a reference destination candidate.
When there are a plurality of adjacent partitions directly above the processing target partition, the leftmost partition B0 among them is set as a reference destination candidate.
 (シンタックスについて)
 まず、本実施例に係る動き情報の参照先予測方法を備える動画像符号化装置により符号化される動画像のビットストリームのシンタックスについて説明する。
(About syntax)
First, the syntax of a bit stream of a moving image that is encoded by a moving image encoding device that includes the motion information reference destination prediction method according to the present embodiment will be described.
 図7は動画像のビットストリームのシーケンス・パラメータ・セット(SPS)に記述される第1のシンタックスパターンを示す。シーケンス・パラメータ・セットはシーケンス全体の符号化に関わる情報が記述されるヘッダーであり、シーケンス全体に渡って、画面間予測を行う場合、本実施の形態に係る動き情報の参照先予測方法を適用するかどうかを示す第1フラグinter_merge_flagが設置される。 FIG. 7 shows a first syntax pattern described in a sequence parameter set (SPS) of a moving image bit stream. The sequence parameter set is a header in which information related to coding of the entire sequence is described. When performing inter-screen prediction over the entire sequence, the motion information reference destination prediction method according to the present embodiment is applied. A first flag inter_merge_flag indicating whether or not to perform is set.
 図8はスライスの中の予測ブロック単位に記述される第2のシンタックスパターンを示す。予測ブロックの予測モードが画面間予測(MODE_INTER)の場合、予測ブロック内のパーティション毎に、周囲に隣接する有効なパーティションの数NumMergeCandidatesが求められ、NumMergeCandidatesが0を超える場合に、このパーティションにてマージを適用するか否かを表す第2フラグmerge_flag[i]が設置される。ここで、iは予測ブロックの中のパーティション番号を表す。第2フラグmerge_flag[i]は本実施の形態に係る動き情報の参照先予測方法が適用されない場合、即ち、inter_merge_flagが偽(0)の場合には符号化されず、ビットストリーム中に記述されない。 FIG. 8 shows a second syntax pattern described for each prediction block in the slice. When the prediction mode of the prediction block is inter-screen prediction (MODE_INTER), the number of effective neighboring partitions NumMergeCandidates is obtained for each partition in the prediction block, and when NumMergeCandidates exceeds 0, merging is performed in this partition. A second flag merge_flag [i] indicating whether or not to apply is set. Here, i represents the partition number in the prediction block. The second flag merge_flag [i] is not coded and is not described in the bitstream when the motion information reference prediction method according to the present embodiment is not applied, that is, when the inter_merge_flag is false (0).
 次に、第2フラグmerge_flag[i]が真(1)で且つNumMergeCandidatesが1を超える場合に、本実施の形態に係る動き情報の参照先予測方法にて選択される参照先の隣接パーティションが正しいか否かを表す第3フラグmerge_probable_flag[i]が設置される。第3フラグmerge_probable_flag[i]はマージを適用しない場合、即ちmerge_flag[i]が偽(0)の場合は、通常の動きベクトル検出にて検出された動き情報を使用するので、第3フラグmerge_probable_flag[i]にて判定する必要が無い。また、NumMergeCandidatesが1の場合にも符号化されない。周囲に隣接する有効なパーティションの数が1つであれば、その1つが参照先のパーティションとなるので、merge_probable_flag[i]を伝送せずとも参照先するパーティションの動き情報が確定する為である。 Next, when the second flag merge_flag [i] is true (1) and NumMergeCandidates exceeds 1, the reference destination adjacent partition selected by the motion information reference destination prediction method according to the present embodiment is correct. A third flag merge_proble_flag [i] indicating whether or not is set. When the third flag merge_proble_flag [i] does not apply merging, that is, when merge_flag [i] is false (0), the motion information detected by the normal motion vector detection is used, so the third flag merge_proble_flag [i] is used. i] does not need to be determined. Also, even when NumMergeCandidates is 1, it is not encoded. This is because, if the number of effective partitions adjacent to each other is one, one of the partitions becomes a reference destination partition, so that the motion information of the reference destination partition is determined without transmitting merge_proble_flag [i].
 尚、NumMergeCandidatesの算出については、後述する。 In addition, calculation of NumMergeCandidates will be described later.
 以下では説明を簡単にする為に、第2フラグ及び第3フラグの配列部分を取り除いた変数を使用することとする。 In the following, in order to simplify the explanation, a variable obtained by removing the array part of the second flag and the third flag will be used.
 (符号化における動き情報の参照先予測)
 上述のシンタックスに基づき、動画像のビットストリームを符号化する動画像符号化装置における、実施の形態に係る動き情報の参照先予測方法の動作を説明する。動き情報の参照先予測方法をシーケンス全体に渡って、画面間予測に適用する場合には、シーケンス・パラメータ・セット(SPS)に記載される第1フラグinter_merge_flagが真(1)に設定される。次に、スライスの中の予測モードが画面間予測(MODE_INTER)である予測ブロックのパーティション毎に、本実施の形態に係る動き情報の参照先予測方法の適用/不適用が切り替えられる。
(Reference prediction of motion information in encoding)
Based on the above-described syntax, the operation of the motion information reference destination prediction method according to the embodiment in the video encoding device that encodes a video bitstream will be described. When the motion information reference prediction method is applied to inter-screen prediction over the entire sequence, the first flag inter_merge_flag described in the sequence parameter set (SPS) is set to true (1). Next, application / non-application of the motion information reference destination prediction method according to the present embodiment is switched for each partition of a prediction block whose prediction mode in the slice is inter-frame prediction (MODE_INTER).
 参照先指定部113は、処理対象のパーティションの左及び真上に隣接するパーティションをマージ処理の参照先として使用するか否かの判定を行う。図9のフローチャートを参照して参照先指定部113の動作を説明する。まず、処理対象のパーティションに隣接する周辺のパーティションの候補数NumMergeCandidatesを算出する(S101)。 The reference destination specifying unit 113 determines whether or not to use a partition adjacent to the left and directly above the processing target partition as a reference destination of the merge processing. The operation of the reference destination specifying unit 113 will be described with reference to the flowchart of FIG. First, the number NumMergeCandidates of peripheral partitions adjacent to the partition to be processed is calculated (S101).
 図10(a)は、候補数NumMergeCandidatesの算出処理の詳細を示すフローチャートである。最初にNumMergeCandidatesを0に設定する(S201)。次に参照先パーティションとして処理対象のパーティションの左に位置するパーティションを設定する(S202)。処理対象のパーティションの画面内の位置情報から参照先パーティションが画面内に存在するかどうかを判定する(S203)。 FIG. 10A is a flowchart showing details of the calculation processing of the number of candidates NumMergeCandidates. First, NumMergeCandidates is set to 0 (S201). Next, a partition located to the left of the partition to be processed is set as a reference destination partition (S202). It is determined from the position information in the screen of the partition to be processed whether or not the reference destination partition exists in the screen (S203).
 図10(b)は、処理対象のパーティション(図中の灰色矩形)に隣接する参照先パーティション(図中の斜線で表される矩形)の例を示す。同図に示すように、処理対象のパーティションが左上に位置する場合には参照先パーティションは存在せず、画面上端に位置する場合は左のみ、画面左端に位置する場合は上のみ、その他の位置では左と上に存在する。参照先パーティションが画面内に存在しない場合(S203のNo)、以降の処理を飛ばしてステップS207に進む。 FIG. 10 (b) shows an example of a reference destination partition (rectangle represented by diagonal lines in the figure) adjacent to the partition to be processed (gray rectangle in the figure). As shown in the figure, when the partition to be processed is located at the upper left, there is no reference destination partition, only when it is located at the top of the screen, only at the left, when located at the left edge of the screen, only at the top, other positions Then it is on the left and top. When the reference destination partition does not exist in the screen (No in S203), the subsequent processing is skipped and the process proceeds to Step S207.
 参照先パーティションが画面内に存在する場合(S203のYes)、参照先指定部113は、参照先パーティションの動き情報を取得する(S204)。参照先指定部113は、メモリ110から参照先パーティションに対応する符号化済みの動き情報を読み出す。読み出された参照先パーティションの動き情報に基づき、参照先指定部113は、参照先パーティションの符号化モードがイントラか否かを判定する(S205)。イントラの場合はS207に進み、イントラでない場合はNumMergeCandidatesを1加算する(S206)。 When the reference destination partition exists in the screen (Yes in S203), the reference destination specifying unit 113 acquires the motion information of the reference destination partition (S204). The reference destination designating unit 113 reads the encoded motion information corresponding to the reference destination partition from the memory 110. Based on the read motion information of the reference destination partition, the reference destination specifying unit 113 determines whether or not the encoding mode of the reference destination partition is intra (S205). If it is intra, the process proceeds to S207. If it is not intra, NumMergeCandidates is incremented by 1 (S206).
 参照先パーティションの設定が処理対象のパーティションの上であるかどうかを判定する(S207)。参照先パーティションが処理対象のパーティションの上に設定されている場合は終了し、NumMergeCandidatesを出力する。参照先パーティションが処理対象のパーティションの上に設定されていない場合は、参照先パーティションを処理対象のパーティションの上に設定し直して、ステップS203以降の判定処理を継続する(S208)。 It is determined whether the setting of the reference destination partition is on the partition to be processed (S207). If the reference destination partition is set on the processing target partition, the process ends and outputs NumMergeCandidates. When the reference destination partition is not set on the processing target partition, the reference destination partition is set again on the processing target partition, and the determination processing after step S203 is continued (S208).
 以上のようにして、候補数NumMergeCandidatesが算出される。 The number of candidates NumMergeCandidates is calculated as described above.
 次に、図9に戻り、算出された候補数NumMergeCandidatesが0個より大きいか否かの判定が行われる(S102)。候補数NumMergeCandidatesが0の場合、マージ処理の参照先となる隣接するパーティションが存在しないので、マージ処理により選択された動き情報に基づく動き補償は行わないこととする。候補数NumMergeCandidatesが0より大きい場合、次のステップS103の判定に進む。 Next, returning to FIG. 9, it is determined whether or not the calculated number of candidates NumMergeCandidates is greater than 0 (S102). When the number of candidates NumMergeCandidates is 0, there is no adjacent partition as a reference destination of the merge process, and therefore motion compensation based on the motion information selected by the merge process is not performed. If the number of candidates NumMergeCandidates is greater than 0, the process proceeds to the next step S103.
 候補数NumMergeCandidatesが1より大きいか否かの判定が行われる(S103)。候補数NumMergeCandidatesが1個の場合(S103のNo)、その1個がマージ処理の参照先パーティションとなるので、参照先指定部113は、その参照先パーティションの動き情報をメモリ110から取得する(S105)。参照先指定部113は、取得した参照先パーティションの動き情報を動き補償部106に供給し、動き補償部106は、指定された参照先パーティションの動き情報にもとづいて処理対象パーティションの動き補償を行う。 It is determined whether the candidate number NumMergeCandidates is greater than 1 (S103). If the number of candidates NumMergeCandidates is 1 (No in S103), one of them becomes a reference partition for the merge process, and the reference destination specifying unit 113 acquires the motion information of the reference destination partition from the memory 110 (S105). ). The reference destination specifying unit 113 supplies the acquired motion information of the reference destination partition to the motion compensation unit 106, and the motion compensation unit 106 performs motion compensation of the processing target partition based on the motion information of the specified reference destination partition. .
 NumMergeCandidatesが1より大きい、即ち2の場合(S103のYes)、参照先指定部113は、処理対象のパーティションの左と上に隣接するパーティションの動き情報をメモリ110から読み出し、取得する(S104)。 If NumMergeCandidates is greater than 1, that is, 2 (Yes in S103), the reference destination designating unit 113 reads out and acquires the motion information of the partitions adjacent to the left and above the processing target partition from the memory 110 (S104).
 参照先指定部113は、取得された左と上に隣接するパーティションの動き情報を比較する(S106)。具体的には、動き情報を構成する予測方向、参照ピクチャ番号及び動きベクトルを比較する。左と上に隣接するパーティションの動き情報が全く同一の場合(S106のYes)、候補数NumMergeCandidatesを1に変更して(S107)、参照先指定部113は、左または上のどちらか一方の隣接するパーティションの動き情報を参照先パーティションの動き情報として動き補償部106に供給する。ここでは、左に隣接するパーティションの動き情報を選択することにする。左と上に隣接するパーティションの動き情報が異なる場合(S106のYes)、参照先指定部113は、2個の動き情報をそれぞれ動き補償部106に供給する。 The reference destination specifying unit 113 compares the acquired motion information of the left and upper adjacent partitions (S106). Specifically, the prediction direction, the reference picture number, and the motion vector constituting the motion information are compared. When the motion information of the left and upper adjacent partitions is exactly the same (Yes in S106), the number of candidates NumMergeCandidates is changed to 1 (S107), and the reference destination designating unit 113 sets either the left or the upper adjacent The motion information of the partition to be supplied is supplied to the motion compensation unit 106 as the motion information of the reference destination partition. Here, the motion information of the partition adjacent to the left is selected. When the motion information of the left and upper adjacent partitions is different (Yes in S106), the reference destination specifying unit 113 supplies the two pieces of motion information to the motion compensation unit 106, respectively.
 以上のように、参照先指定部113は、マージ処理により選択された参照先ブロックのパーティションの動き情報を動き補償部106に供給し、動き補償部106は、指定された参照先ブロックのパーティションの動き情報に基づいて動き補償を行う。また、動き補償部106は、動きベクトル検出部101により検出された予測ブロックのパーティション毎の動きベクトルにもとづいて通常の動き補償を行う。また、画面内予測部111は、画面内の画素相関を利用して面内予測を行う。これら3つの予測方法により生成される予測画像が予測選択部112に入力される。 As described above, the reference destination specifying unit 113 supplies the motion information of the partition of the reference block selected by the merging process to the motion compensation unit 106, and the motion compensation unit 106 specifies the partition of the specified reference block. Motion compensation is performed based on the motion information. Also, the motion compensation unit 106 performs normal motion compensation based on the motion vector for each partition of the prediction block detected by the motion vector detection unit 101. The in-screen prediction unit 111 performs in-plane prediction using pixel correlation in the screen. A prediction image generated by these three prediction methods is input to the prediction selection unit 112.
 予測選択部112は、符号化する画像と入力された3つの予測画像との差分により生成される残差画像と動き情報等の付加情報の発生符号量を計算し、符号量が最小となる予測方法を選択する。 The prediction selection unit 112 calculates a generated code amount of additional information such as a residual image and motion information generated by the difference between the image to be encoded and the three input prediction images, and makes a prediction that minimizes the code amount Select a method.
 ここで、マージ処理により選択された参照先パーティションの動き情報に基づく動き補償による予測方法が選択された場合について説明する。この時、予測選択部112は、マージ処理により選択された参照先パーティションの動き情報として左または上のどちらを使用したかを示すフラグmerge_direc_flagを設定する。merge_direc_flagが1の場合は左に隣接するパーティションの動き情報を、0の場合は上に隣接するパーティションの動き情報を選択することとして、以下説明する。 Here, a case where a prediction method based on motion compensation based on the motion information of the reference partition selected by the merge process is selected will be described. At this time, the prediction selection unit 112 sets a flag merge_dirc_flag that indicates which one of the left or the upper is used as the motion information of the reference destination partition selected by the merge process. In the following description, the motion information of the partition adjacent to the left is selected when merge_dirc_flag is 1, and the motion information of the upper adjacent partition is selected when 0.
 予測選択部112は、マージ処理による予測方法を選択した場合、上述したシンタックスで定義されるマージ処理用のフラグを設定する。図11は、予測選択部112がマージ処理用のフラグを設定する手順を説明するフローチャートである。 The prediction selection unit 112 sets a flag for merge processing defined by the above-described syntax when the prediction method by merge processing is selected. FIG. 11 is a flowchart illustrating a procedure in which the prediction selection unit 112 sets a flag for merge processing.
 最初に予測選択部112により選択された予測方法がマージ処理で選択された動き情報に基づく予測であるか否かの判定が行われる(S301)。画像内予測や通常の動きベクトル検出を用いた画面間予測の場合(S301のNo)、merge_flagを0に設定して、終了する(S303)。 First, it is determined whether or not the prediction method selected by the prediction selection unit 112 is prediction based on motion information selected in the merge process (S301). In the case of inter-screen prediction using intra-picture prediction or normal motion vector detection (No in S301), merge_flag is set to 0 and the process ends (S303).
 マージ処理で選択された動き情報に基づく予測であれば(S301のYes)、merge_flagを1に設定する(S302)。次に、マージ処理で選択された動き情報の数を調べる(S304)。具体的には、上述した参照先指定部113が指定する参照先となる隣接するパーティションの候補数NumMergeCandidatesが1より大きいか否かの判定を行う。 If the prediction is based on the motion information selected in the merge process (Yes in S301), merge_flag is set to 1 (S302). Next, the number of motion information selected by the merge process is checked (S304). Specifically, it is determined whether or not the number of adjacent partition candidates NumMergeCandidates to be specified by the reference destination specifying unit 113 is greater than one.
 候補数NumMergeCandidatesが1以下の場合(S304のNo)、マージ処理を選択していることが前提であり、必ず有効な参照先の隣接するパーティションが1つ存在し、左または上のどちらか一方に特定することができるため、merge_flagのみを設定し、終了する。 If the number of candidates NumMergeCandidates is 1 or less (No in S304), it is assumed that merge processing is selected, and there is always one adjacent partition of a valid reference destination, either on the left or above Since it can be specified, only merge_flag is set and the process ends.
 候補数NumMergeCandidatesが1より大きい場合、即ち2個の隣接するパーティションの動き情報の候補があった場合(S304のYes)、参照先パーティションの予測に進む(S305)。ここで、参照先パーティションの予測では、後述する動き情報の参照先予測方法が用いられる。処理対象のパーティションの左及び上に隣接するパーティションのサイズや、処理対象パーティションと隣接パーティションが互いに接している辺の長さ等の情報に基づいて、参照先パーティションの予測が行われ、予測された参照先パーティションを示すフラグpred_direc_flagが出力される。 If the number of candidates NumMergeCandidates is larger than 1, that is, if there is a candidate for motion information of two adjacent partitions (Yes in S304), the process proceeds to prediction of a reference destination partition (S305). Here, in prediction of the reference destination partition, a reference destination prediction method for motion information described later is used. Based on information such as the size of the partition adjacent to the left and top of the target partition and the length of the side where the target partition and the adjacent partition are in contact with each other, the reference partition is predicted and predicted A flag pred_direc_flag indicating the reference destination partition is output.
 予測された参照先パーティションを示すフラグpred_direc_flagはmerge_direc_flagと同様に、参照先の隣接するパーティションとして左(1)或いは上(0)で選択先を表すとともに、予測によって参照先パーティションが選択できなかった場合は、不定(2)が設定される。 The flag pred_dirc_flag indicating the predicted reference destination partition indicates the selection destination on the left (1) or above (0) as the adjacent partition of the reference destination, as in the case of merge_direct_flag, and the reference destination partition cannot be selected by prediction. Indefinite (2) is set.
 次に、このようにして設定された、予測された参照先パーティションを示すフラグpred_direc_flagの判定を行う(S306)。予測された参照先パーティションを示すフラグpred_direc_flagが2となるのは参照先パーティションの予測において予測不可の場合であり(S306のYes)、この時、実際に発生符号量が最小となる予測方法として選択されたマージ処理による予測で用いた動き情報の参照先パーティションを表すmerge_direc_flagを、参照先有効情報を示すフラグmerge_probable_flagに設定して、終了する(S307)。この場合、参照先有効情報を示すフラグmerge_probable_flagは、マージ処理において従来と同様に参照先のパーティションを直接示すことになる。 Next, the flag pred_direc_flag indicating the predicted reference destination partition set in this way is determined (S306). The flag pred_dirc_flag indicating the predicted reference destination partition is 2 when prediction is impossible in the prediction of the reference destination partition (Yes in S306), and at this time, it is selected as the prediction method that actually minimizes the generated code amount. The merge_direc_flag indicating the reference destination partition of the motion information used in the prediction performed by the merge processing is set in the flag merge_proble_flag indicating the reference destination valid information, and the process ends (S307). In this case, the flag merge_proble_flag indicating the reference destination valid information directly indicates the reference destination partition in the merge process as in the conventional case.
 予測された参照先パーティションを示すフラグpred_direc_flagが2でない場合(S306のNo)、フラグmerge_direc_flagとpred_direc_flagの比較を行う(S308)。merge_direc_flagは実際に発生符号量が最小となる予測方法として選択されたマージ処理による予測で用いた動き情報の参照先パーティションを表し、参照先パーティションの予測が当たっていればpred_direc_flagは同一の値になるので、その場合フラグmerge_probable_flagを1に設定する(S309)。一方、参照先パーティションの予測が外れていればpred_direc_flagは異なる値になるので、その場合フラグmerge_probable_flagを0に設定する(S310)。 When the flag pred_dirc_flag indicating the predicted reference destination partition is not 2 (No in S306), the flags merge_dirc_flag and pred_dirc_flag are compared (S308). merge_direc_flag represents a reference destination partition of motion information used in prediction by merge processing selected as a prediction method that actually minimizes the amount of generated code, and pred_direc_flag has the same value if the prediction of the reference destination partition is correct. In this case, the flag merge_proble_flag is set to 1 (S309). On the other hand, if the prediction of the reference destination partition is not correct, pred_dirc_flag has a different value. In this case, the flag merge_proble_flag is set to 0 (S310).
 マージ先パーティションの予測の成否を示すmerge_probable_flagを用いると、マージ先パーティションの予測が当たった場合、merge_probable_flagは1となり、予測がはずれた場合のみmerge_probable_flagは0となる。つまり、マージ先パーティションの予測が当たるほどmerge_probable_flagは1である確率が高く、merge_probable_flagが1である確率が高いことを利用した算術符号化を用いることにより、より少ない情報量でmerge_probable_flagを伝送することができる。 When merge_proble_flag indicating success / failure of the merge destination partition is used, merge_proble_flag becomes 1 when the merge destination partition is predicted, and merge_proble_flag becomes 0 only when the prediction is off. In other words, merge_proble_flag can be transmitted with a smaller amount of information by using arithmetic coding using the fact that merge_proble_flag has a higher probability of 1 and merge_proble_flag has a higher probability of 1 as the merge destination partition is predicted. it can.
 以上のようにして、パーティション毎にマージ処理により選択された隣接するパーティションの動き情報に基づく動き補償予測が選択された場合、SPSのヘッダーに記載される第1フラグinter_merge_flag、予測ブロックに記載される第2フラグmerge_flag及び第3フラグmerge_probable_flagが設定され、符号化される。 As described above, when motion-compensated prediction based on the motion information of adjacent partitions selected by the merge process for each partition is selected, the first flag inter_merge_flag described in the header of the SPS is described in the prediction block. The second flag merge_flag and the third flag merge_proble_flag are set and encoded.
 (復号における動き情報の参照先予測)
 上述のシンタックスに基づき、上記の動画像符号化装置により符号化された動画像のビットストリームを復号する動画像復号装置における、実施の形態に係る動き情報の参照先予測方法の動作を説明する。
(Prediction of motion information reference in decoding)
Based on the syntax described above, an operation of the motion information reference destination prediction method according to the embodiment in the video decoding device that decodes the bit stream of the video encoded by the video encoding device will be described. .
 最初に、可変長復号部201により復号されるビットストリームのSPSに記述されているフラグinter_merge_flagから、ビットストリームのシーケンス全体に対して、本実施の形態に係る動き情報の参照先予測方法を使用したマージ処理による画面間予測が適用されるか否かを判定する。inter_merge_flagが真(1)の場合、マージ処理が適用され、inter_merge_flagが偽(0)の場合、マージ処理が無視され、従来通りのビットストリームから復号された動き情報に基づく画面間予測が実施される。以下、inter_merge_flagが真(1)、即ち、本実施の形態に係る動き情報の参照先予測方法を使用したマージ処理が適用されているものとして説明する。 First, from the flag inter_merge_flag described in the SPS of the bitstream decoded by the variable length decoding unit 201, the motion information reference destination prediction method according to the present embodiment is used for the entire bitstream sequence. It is determined whether inter-screen prediction by merge processing is applied. When inter_merge_flag is true (1), merge processing is applied. When inter_merge_flag is false (0), merge processing is ignored and inter-screen prediction based on motion information decoded from a conventional bitstream is performed. . In the following description, it is assumed that inter_merge_flag is true (1), that is, the merge processing using the motion information reference destination prediction method according to the present embodiment is applied.
 次に、画面内/画面間予測選択部209により、スライスの中の予測ブロック毎に予測モードを参照して画面内または画面間予測の何れを選択するかが判定される。画面内予測(MODE_INTRA)が選択された場合は画面内予測部208の処理に進み、画面間予測(MODE_INTER)が選択された場合は動き情報選択部210の処理に進む。 Next, the intra-screen / inter-screen prediction selection unit 209 determines whether to select intra-screen or inter-screen prediction with reference to the prediction mode for each prediction block in the slice. When the intra prediction (MODE_INTRA) is selected, the process proceeds to the intra prediction unit 208, and when the inter prediction (MODE_INTER) is selected, the process proceeds to the motion information selection unit 210.
 ここで、動き情報選択部210は、予測ブロックのパーティション毎にビットストリームから復号された情報として、符号化の動きベクトル検出にて検出された動き情報、或いはマージ処理により選択された参照先パーティションの動き情報のどちらか一方を選択し、選択された動き情報を動き補償部203に出力する。 Here, the motion information selection unit 210 uses, as information decoded from the bit stream for each partition of the prediction block, the motion information detected by the motion vector detection of encoding or the reference destination partition selected by the merge process. Either one of the motion information is selected, and the selected motion information is output to the motion compensation unit 203.
 図12は、動き情報選択部210の選択手順を示すフローチャートであり、この図を参照して説明する。まず、処理対象のパーティションに隣接する周辺のパーティションの候補数NumMergeCandidatesが算出される(S401)。算出手順は符号化におけるS101と同様であるので割愛する。 FIG. 12 is a flowchart showing a selection procedure of the motion information selection unit 210, which will be described with reference to this figure. First, the number NumMergeCandidates of peripheral partitions adjacent to the partition to be processed is calculated (S401). Since the calculation procedure is the same as S101 in encoding, it is omitted.
 算出された候補数NumMergeCandidatesが0個より大きいか否かが判定される(S402)。0個の場合、マージ処理の参照先となる隣接するパーティションが存在しないので、動き情報選択部210は、従来同様ビットストリームから復号された動き情報を出力する。0個より大きい場合は、可変長復号部201により復号されたmerge_flagが読み込まれ(S403)、merge_flagに基づき動き情報の選択を判定する(S404)。merge_flagが偽(0)の場合、動き情報選択部210は、従来同様ビットストリームから復号された動き情報を出力する。merge_flagが真(1)の場合、マージ処理での参照先パーティションの動き情報を選択する為の処理に移る。 It is determined whether or not the calculated number of candidates NumMergeCandidates is greater than 0 (S402). In the case of 0, since there is no adjacent partition that becomes the reference destination of the merge process, the motion information selection unit 210 outputs the motion information decoded from the bitstream as in the conventional case. When the number is larger than 0, merge_flag decoded by the variable length decoding unit 201 is read (S403), and selection of motion information is determined based on merge_flag (S404). When merge_flag is false (0), the motion information selection unit 210 outputs motion information decoded from the bitstream as in the conventional case. When merge_flag is true (1), the process proceeds to processing for selecting motion information of the reference partition in the merge processing.
 再度、候補数NumMergeCandidatesが1個より大きいか否かの判定が行われる(S405)。候補数NumMergeCandidatesが1個の場合、その1個がマージ処理の参照先パーティションとなるので、その参照先パーティションの動き情報を、復号済み情報が記録されているメモリ207から取得する(S407)。取得された動き情報は動き補償部203に供給され、その動き情報にもとづいて動き補償が行われる。 Again, it is determined whether or not the number of candidates NumMergeCandidates is larger than one (S405). When the number of candidates NumMergeCandidates is one, one of them becomes a reference destination partition for the merge process, so the motion information of the reference destination partition is acquired from the memory 207 in which the decoded information is recorded (S407). The acquired motion information is supplied to the motion compensation unit 203, and motion compensation is performed based on the motion information.
 候補数NumMergeCandidatesが1個より大きい場合、即ち2個の場合、処理対象のパーティションの左と上に隣接するパーティションの動き情報をメモリ207から読み出し、取得する(S406)。取得された左と上に隣接するパーティションの動き情報を比較する(S408)。具体的には、動き情報を構成する予測方向、参照ピクチャ番号及び動きベクトルを比較する。動き情報が全く同一の場合、左或いは上のどちらか一方の隣接するパーティションの動き情報を取得して、動き補償部203に供給する(S410)。ここでは、符号化と同様に左に隣接するパーティションの動き情報を選択することにする。 When the number of candidates NumMergeCandidates is larger than 1, that is, two, the motion information of the partitions adjacent to the left and above the partition to be processed is read from the memory 207 and acquired (S406). The obtained motion information of the left and upper adjacent partitions is compared (S408). Specifically, the prediction direction, the reference picture number, and the motion vector constituting the motion information are compared. If the motion information is exactly the same, the motion information of either the left or upper adjacent partition is acquired and supplied to the motion compensation unit 203 (S410). Here, the motion information of the partition adjacent to the left is selected as in the encoding.
 動き情報が異なる場合は参照先パーティションの予測に進む(S409)。参照先パーティションの予測は、符号化と同様に、後述する動き情報の参照先予測方法が使用される。処理対象のパーティションと周辺の左及び上に隣接するパーティションのサイズや、処理対象パーティションと隣接パーティションが互いに接している辺の長さ等の情報に基づいて、参照先パーティションの予測が行われ、予測された参照先パーティションを示すフラグpred_direc_flagが出力される。 If the motion information is different, the process proceeds to prediction of the reference destination partition (S409). For prediction of the reference destination partition, a motion information reference destination prediction method, which will be described later, is used as in the case of encoding. The target partition is predicted based on information such as the size of the partition adjacent to the processing target partition and the left and top of the periphery, and the length of the side where the processing target partition and the adjacent partition are in contact with each other. The flag pred_dirc_flag indicating the referred reference partition is output.
 予測された参照先パーティションを示すフラグpred_direc_flagは、符号化と同様に参照先の隣接するパーティションが選択される場合は左(1)或いは上(0)で表され、予測によって参照先パーティションが選択出来なかった場合は不定(2)が設定される。 The flag pred_dirc_flag indicating the predicted reference destination partition is represented by left (1) or upper (0) when the adjacent reference destination partition is selected as in the encoding, and the reference destination partition can be selected by prediction. If not, undefined (2) is set.
 次に出力されたpred_direc_flagの判定を行う(S411)。pred_direc_flagが2となるのは参照先パーティションの予測で予測不可の場合であり、この時、case1の処理に進む(S412)。一方、pred_direc_flagが2でない場合は、case2の処理に進む(S413)。 Next, the pred_direc_flag output is determined (S411). The case where pred_dirc_flag is 2 is a case where prediction is impossible due to prediction of the reference destination partition, and at this time, the process proceeds to case 1 (S412). On the other hand, if pred_dirc_flag is not 2, the process proceeds to case 2 (S413).
 図13(a)は、ステップS412のcase1の処理のフローチャートを示す。case1では参照先パーティションの予測では予測結果が不定であり、参照先の予測結果が示されない。そこで、従来と同様にmerge_probable_flagで示される左或いは上に隣接するパーティションを参照先パーティションとして選択する。この場合、符号化側でmerge_probable_flagには、実際の参照先パーティションを示すmerge_direc_flagが設定されているからである。 FIG. 13A shows a flowchart of the process of case 1 in step S412. In case 1, the prediction result is undefined in the prediction of the reference destination partition, and the prediction result of the reference destination is not shown. Therefore, the partition adjacent to the left or above indicated by merge_proble_flag is selected as the reference destination partition as in the conventional case. This is because, in the encoding side, merge_direct_flag indicating the actual reference destination partition is set in merge_proble_flag.
 可変長復号部201により復号されたmerge_probable_flagが読み込まれ(S421)、merge_probable_flagに基づき参照先パーティションを選択する(S422)。pred_direc_flagが1の場合、左に隣接するパーティション(S423)、merge_probable_flagが0の場合、上に隣接するパーティションを参照先パーティションとして選択する(S424)。 The merge_proble_flag decoded by the variable length decoding unit 201 is read (S421), and a reference destination partition is selected based on the merge_proble_flag (S422). When pred_dirc_flag is 1, the left adjacent partition (S423) is selected, and when merge_proble_flag is 0, the upper adjacent partition is selected as the reference destination partition (S424).
 図13(b)は、ステップS413のcase2の処理のフローチャートを示す。case2の処理の場合、予測された参照先パーティションを示すフラグpred_direc_flagの確からしさは、参照先有効性を示すフラグmerge_probable_flagによって表される。即ち、pred_direc_flagで示される参照先パーティションの選択が正しければmerge_probable_flagは1(真)で表され、正しくなければmerge_probable_flagは0(偽)で表されることになる。 FIG. 13B shows a flowchart of the process of case 2 in step S413. In the case 2 process, the probability of the flag pred_direc_flag indicating the predicted reference destination partition is represented by the flag merge_proble_flag indicating the reference destination validity. That is, merge_proble_flag is represented by 1 (true) if the reference destination partition indicated by pred_dirc_flag is correct, and merge_proble_flag is represented by 0 (false) if not correct.
 可変長復号部201により復号されたmerge_probable_flagが読み込まれ(S431)、merge_probable_flagに基づき参照先パーティションの選択を判定する(S432)。merge_probable_flagが1の場合、参照先パーティションの予測で判定されたpred_direc_flagの示す値のパーティションを参照先パーティションとして選択される(S433)。merge_probable_flagが0の場合、pred_direc_flagの示す値と逆のパーティションを参照先パーティションとして選択する。例えば、pred_direc_flagが0(左)であれば、上に隣接するパーティションを参照先パーティションとする(S434)。 The merge_proble_flag decoded by the variable length decoding unit 201 is read (S431), and the selection of the reference destination partition is determined based on the merge_proble_flag (S432). When merge_proble_flag is 1, a partition having a value indicated by pred_dirc_flag determined by prediction of the reference destination partition is selected as the reference destination partition (S433). When merge_proble_flag is 0, a partition opposite to the value indicated by pred_direc_flag is selected as a reference destination partition. For example, if pred_dirc_flag is 0 (left), the upper adjacent partition is set as the reference destination partition (S434).
 図12に戻り、case1及びcase2の処理で選択された参照先パーティションの動き情報が取得され(S415)、動き補償部203に供給され、終了する。 Returning to FIG. 12, the motion information of the reference destination partition selected in the processing of case 1 and case 2 is acquired (S415), supplied to the motion compensation unit 203, and the process ends.
 以上のように、可変長復号部201により復号された各フラグと復号過程での演算結果に応じてマージ処理による画面間予測が選択された場合、後述する動き情報の参照先予測方法によって、参照先パーティションを選択する。選択された参照先パーティションの動き情報から動き補償により生成される予測画像とビットストリームから復号された残差信号とを加算することで復号画像が生成される。 As described above, when inter-screen prediction by merge processing is selected according to each flag decoded by the variable-length decoding unit 201 and the calculation result in the decoding process, the reference is made by the motion information reference destination prediction method described later. Select the destination partition. A decoded image is generated by adding the predicted image generated by motion compensation from the motion information of the selected reference destination partition and the residual signal decoded from the bitstream.
 (動き情報の参照先予測方法)
 図14は、本実施の形態に係る動き情報の参照先予測方法を説明するフローチャートである。図14に示される処理過程は、パーティション単位で実施される。各過程で処理対象のパーティションに対して動き情報を参照するパーティションの判定が行われ、参照先としてパーティションが決定された場合には次の過程の判定に進まず、終了する。以下、各過程を順を追って説明する。
(Motion information reference destination prediction method)
FIG. 14 is a flowchart for explaining a motion information reference destination prediction method according to the present embodiment. The process shown in FIG. 14 is performed in units of partitions. In each process, a partition that refers to motion information is determined for the partition to be processed, and if a partition is determined as a reference destination, the process does not proceed to the next process and ends. Hereinafter, each process will be described step by step.
 まず、処理対象のパーティションの左或いは上に隣接する周囲のパーティションとの境界判定を行う(S500)。図15は、ステップS500の隣接する周囲のパーティションとの境界判定処理を示すフローチャートである。この図を用いて処理の詳細を説明する。 First, a boundary determination with a neighboring partition adjacent to the left or top of the partition to be processed is performed (S500). FIG. 15 is a flowchart showing a boundary determination process with adjacent surrounding partitions in step S500. Details of the processing will be described with reference to FIG.
 最初に処理対象のパーティションとその周囲に隣接するパーティションの動き境界検出を行う(S501)。 First, motion boundary detection is performed between a partition to be processed and neighboring partitions around it (S501).
 図16は、処理対象のパーティションをXとし、左及び上に隣接するパーティションをそれぞれA及びBとした場合の配置例を示す。処理対象のパーティションの矩形サイズを幅wx、高さhxで表し、隣接する左のパーティションの高さをpa、隣接する上のパーティションの幅をpbで表す。 FIG. 16 shows an arrangement example in which the processing target partition is X and the left and upper adjacent partitions are A and B, respectively. The rectangular size of the partition to be processed is represented by width wx and height hx, the height of the adjacent left partition is represented by pa, and the width of the adjacent upper partition is represented by pb.
 処理対象のパーティションと左および上に隣接するパーティションの境界の判定は、処理対象のパーティションの高さhxと隣接する左のパーティションの高さpaの比較、及び処理対象のパーティションの幅wxと隣接する上のパーティションの幅pbの比較により実施される。 The boundary between the processing target partition and the left and upper adjacent partitions is determined by comparing the height hx of the processing target partition with the height pa of the adjacent left partition and the width wx of the processing target partition. This is done by comparing the width pb of the upper partition.
 隣接するパーティションが処理対象のパーティションと隣接する辺を「動き境界」と呼ぶ。ここで、動き境界とは、単に隣接する左或いは上の一つのパーティションの高さ或いは幅だけを評価するだけでなく、図17に示されるように、例えば処理対象のパーティションの上に隣接するパーティションが複数存在する場合、参照先の候補となる一番左にあるパーティションB0とその右隣のパーティションB1の動き情報が全く同一であると判定されたなら、動き境界をパーティションB1の位置まで延長し、更に右隣のパーティションBk(k=2,3,...nB-1)について同様の判定を行い、動き情報が同一でないと判定されたパーティションの一つ前のパーティションの位置までの幅を加算した値を動き境界の長さとして定義する。 An edge where an adjacent partition is adjacent to the target partition is called a “motion boundary”. Here, the motion boundary is not only simply evaluating the height or width of one adjacent left or upper partition, but as shown in FIG. 17, for example, a partition adjacent to the partition to be processed. If the motion information of the leftmost partition B0, which is a candidate for the reference destination, and the partition information B1 adjacent to the right are determined to be exactly the same, the motion boundary is extended to the position of the partition B1. Further, the same determination is performed for the partition Bk (k = 2, 3,..., NB-1) on the right side, and the width to the position of the previous partition of the partition determined to have the same motion information is calculated. The added value is defined as the length of the motion boundary.
 図17に示される一例では、B0とB1が同一の動き情報を備えるものと判定された場合であり、結果として上に隣接するパーティションの動き境界の長さは、B0とB1の各パーティションの幅を加算したもの、即ちpb=pb0+pb1となる。こうして求められた動き境界の長さが処理対象のパーティションの高さまたは幅と一致するか否かを判定する。 In the example shown in FIG. 17, it is determined that B0 and B1 have the same motion information. As a result, the length of the motion boundary between the upper adjacent partitions is the width of each partition of B0 and B1. , I.e., pb = pb0 + pb1. It is determined whether the length of the motion boundary thus obtained matches the height or width of the partition to be processed.
 まず、隣接する左のパーティションAと上のパーティションBと両方の動き境界の長さが処理対象のパーティションXの高さまたは幅と一致するか否かの判定を行う(S502)。両方とも一致する場合は参照先パーティションを判定することができないので、次の判定ステップS510に進む。両方とも一致しない場合は、AまたはBのどちらか一方が一致するか否かの判定を行う(S503)。A,Bともに一致しない場合は判定が出来ないので、次の判定ステップS510に進む。AまたはBのどちらか一方の動き境界の長さが処理対象のパーティションXの高さまたは幅と一致する場合、一致するパーティションを参照先として選択し(S504)、次の判定ステップS510には進まずに終了する。これは、動き境界の長さが処理対象のパーティションXの高さまたは幅に一致する場合、その隣接するパーティションと処理対象のパーティションとは動き情報が同一である確率が高いことを利用した参照先パーティションの予測である。 First, it is determined whether or not the length of the motion boundary between the adjacent left partition A and the upper partition B matches the height or width of the processing target partition X (S502). If both match, the reference destination partition cannot be determined, and the process proceeds to the next determination step S510. If both do not match, it is determined whether either A or B matches (S503). If A and B do not match, the determination cannot be made, and the process proceeds to the next determination step S510. If the length of the motion boundary of either A or B matches the height or width of the processing target partition X, the matching partition is selected as a reference destination (S504), and the process proceeds to the next determination step S510. Finish first. This is because, when the length of the motion boundary matches the height or width of the processing target partition X, the reference destination using the fact that the adjacent partition and the processing target partition are likely to have the same motion information. Partition prediction.
 次に、図14に戻り、処理対象のパーティションと隣接する左或いは上のパーティションの互いに隣接する辺の共通部分の長さ(以下、「隣接辺長」と呼ぶ)による判定を行う(S510)。処理対象のパーティションは、隣接辺長が長いパーティションの動きとの相関性が高いとの推定に基づき判定する。図18は、ステップS510の隣接するパーティションとの隣接辺長比較処理を示すフローチャートである。この図を用いて処理の詳細を説明する。 Next, returning to FIG. 14, a determination is made based on the length of the common part of adjacent sides of the left or upper partition adjacent to the processing target partition (hereinafter referred to as “adjacent side length”) (S510). The partition to be processed is determined based on the estimation that the correlation with the motion of the partition having a long adjacent side length is high. FIG. 18 is a flowchart showing the adjacent side length comparison process with the adjacent partition in step S510. Details of the processing will be described with reference to FIG.
 最初に、処理対象のパーティションが隣接する左或いは上のパーティションと接している隣接辺長を算出する(S511)。図16で示されるように、左のパーティションAとの隣接辺長はL(A)、上のパーティションBとの隣接辺長はL(B)で表される(図16中の斜線部が隣接部を表す)。左のパーティションAとの隣接辺長L(A)は処理対象のパーティションXの高さhxと左のパーティションの高さpaとのうち小さい方の長さとなり、次式で表される。
 L(A)=min(pa,hx)
ここで、関数min(a,b)はa或いはbのうち小さい方を選択する関数である。一方、上のパーティションBとの隣接辺長L(B)は処理対象のパーティションXの幅wxと上のパーティションの高さpbとのうち小さい方の長さとなり、次式で表される。
 L(B)=min(pb,wx)
但し、pa及びpbはステップS500にて求められた動き境界の長さである。
First, the length of the adjacent side in contact with the left or upper partition adjacent to the processing target partition is calculated (S511). As shown in FIG. 16, the adjacent side length to the left partition A is represented by L (A), and the adjacent side length to the upper partition B is represented by L (B) (the hatched portion in FIG. 16 is adjacent). Part). The adjacent side length L (A) with the left partition A is the smaller of the height hx of the partition X to be processed and the height pa of the left partition, and is expressed by the following equation.
L (A) = min (pa, hx)
Here, the function min (a, b) is a function for selecting the smaller one of a and b. On the other hand, the adjacent side length L (B) with the upper partition B is the smaller one of the width wx of the partition X to be processed and the height pb of the upper partition, and is expressed by the following equation.
L (B) = min (pb, wx)
Here, pa and pb are the lengths of the motion boundaries obtained in step S500.
 求められた隣接辺長L(A)とL(B)の比較を行い、参照先のパーティションを選択する(S512)。L(A)とL(B)とが等しい場合、参照先のパーティションを判定することができないので、参照先のパーティションを選択せず不定とし(S513)、終了する。L(A)とL(B)とが等しくない場合、L(A)とL(B)の大小比較を行う(S514)。L(A)がL(B)より大きい場合、Aを選択し(S515)、終了する。L(A)がL(B)より小さい場合、Bを選択し(S516)、終了する。 The obtained adjacent side lengths L (A) and L (B) are compared, and a reference destination partition is selected (S512). If L (A) and L (B) are equal, the reference destination partition cannot be determined, so the reference destination partition is not selected and is undefined (S513), and the process ends. If L (A) and L (B) are not equal, L (A) and L (B) are compared in size (S514). If L (A) is greater than L (B), A is selected (S515) and the process ends. If L (A) is smaller than L (B), B is selected (S516) and the process ends.
 以上述べた隣接パーティションとの境界判定/(S500)及び隣接辺長比較(S510)により、処理対象のパーティションと隣接する左及び上のパーティションの動き境界や辺長の条件で場合分けした判定結果を纏めたものを図19の表に示す。図19の表中の選択欄が参照先となるパーティションを示しており、記号”-”は参照先パーティションが不定の場合を表す。 Based on the boundary determination with the adjacent partition described above / (S500) and the adjacent side length comparison (S510), the determination results classified according to the motion boundary and side length conditions of the left and upper partitions adjacent to the partition to be processed are as follows. The summary is shown in the table of FIG. The selection column in the table of FIG. 19 indicates the partition to be referred to, and the symbol “-” represents the case where the reference partition is indefinite.
 ここでは、図19の表の中の条件に基づく判定の例を幾つか説明する。例えば、図19の5つ目の境界比較条件”pb<wx且つpa<hx”は、図20に示されるような配置で、隣接するパーティションA,Bの動き境界の長さがともに処理対象のパーティションの幅および高さと一致していない場合である。この場合、動き境界一致による判定は出来ないので、処理対象のパーティションに接している隣接辺長の長さの比較により参照先を判定する。隣接するパーティションA及びBの隣接辺長L(A)及びL(B)はそれぞれ処理対象のパーティションXに接している辺pa及びpbとなるので、paとpbの比較により、参照先のパーティションが判定される。 Here, some examples of determination based on the conditions in the table of FIG. 19 will be described. For example, the fifth boundary comparison condition “pb <wx and pa <hx” in FIG. 19 is arranged as shown in FIG. 20, and the lengths of the motion boundaries of adjacent partitions A and B are both subject to processing. This is the case when the width and height of the partition do not match. In this case, since the determination based on the motion boundary matching cannot be performed, the reference destination is determined by comparing the lengths of adjacent side lengths in contact with the partition to be processed. The adjacent side lengths L (A) and L (B) of the adjacent partitions A and B are the sides pa and pb that are in contact with the partition X to be processed, so that the reference partition is determined by comparing pa and pb. Determined.
 また、図19の9つ目の境界比較条件”pb>wx且つpa>hx”も、隣接するパーティションA,Bの動き境界の長さがともに処理対象のパーティションの幅および高さと一致していない場合である。この場合も、動き境界一致による判定は出来ないので、処理対象のパーティションに接している隣接辺長の長さの比較により参照先を判定する。図21は境界比較条件”pb>wx且つpa>hx”での処理対象及び隣接するパーティションの配置の一例を示したものである。図4に示されるように、予測ブロック内部のパーティションのサイズや形状は定義されている為、図21で示される境界比較条件”pb>wx且つpa>hx”に合致するパーティションのサイズや形状は限定される。この場合、隣接するパーティションに形状に関わらず、処理対象のパーティションはN×Nに限定される。その為、隣接するパーティションA及びBの隣接辺長L(A)及びL(B)はともにNとなる。L(A)=L(B)=Nになり、隣接辺長による比較からは判定が出来ないので、不定と判定する。 In addition, in the ninth boundary comparison condition “pb> wx and pa> hx” in FIG. 19, the length of the motion boundary between the adjacent partitions A and B does not match the width and height of the partition to be processed. Is the case. Also in this case, since the determination based on the motion boundary matching cannot be performed, the reference destination is determined by comparing the lengths of the adjacent sides in contact with the processing target partition. FIG. 21 shows an example of the processing target and the arrangement of adjacent partitions under the boundary comparison condition “pb> wx and pa> hx”. As shown in FIG. 4, since the size and shape of the partition inside the prediction block are defined, the size and shape of the partition meeting the boundary comparison condition “pb> wx and pa> hx” shown in FIG. Limited. In this case, the partition to be processed is limited to N × N regardless of the shape of adjacent partitions. Therefore, the adjacent side lengths L (A) and L (B) of adjacent partitions A and B are both N. L (A) = L (B) = N, and it cannot be determined from the comparison by the adjacent side length, so it is determined as indefinite.
 こうして、判定された参照先パーティションの予測結果を、符号化及び復号過程において予測された参照先パーティションを示すフラグpred_direc_flagとして出力する。pred_direc_flagは、merge_direc_flagと同様に、参照先の隣接するパーティションとして選択される場合は”1(左)”或いは”0(上)”で参照先を表し、予測によって参照先パーティションが選択出来なかった場合は”2(不定)”と設定される。 Thus, the prediction result of the determined reference destination partition is output as a flag pred_dirc_flag indicating the reference destination partition predicted in the encoding and decoding processes. pred_direc_flag is the same as merge_direc_flag, when it is selected as a reference destination adjacent partition, “1 (left)” or “0 (upper)” indicates the reference destination, and the reference destination partition cannot be selected by prediction. Is set to “2 (undefined)”.
 以上述べたように、第1の実施例によれば、マージ処理における参照先ブロックを予測することで、参照先の指定に使用するフラグmerge_probable_flagで表されるビットの発生確率を偏らせ、発生する符号量を削減することが可能となり、符号化効率を向上させることができる。従来は左或いは上の隣接ブロックを参照先として指定する場合、参照先フラグは1/2の頻度になり算術符号化を活かせないでいたが、第1の実施例によれば、予測された参照先の確からしさを表すフラグmerge_probable_flagを用いることで、予測が的中すればフラグのビット頻度を偏らせることが可能であり、算術符号化においては発生符号量を削減することができる。また、符号化対象ブロックのサイズと一致する隣接ブロックや、隣接している辺の長さが長い隣接ブロックの方が符号化対象ブロックとの相関性が高いことを使用して、参照先のブロックを予測するようにしたことで予測の精度を高めることができる。 As described above, according to the first embodiment, the occurrence of the bit represented by the flag merge_proble_flag used for designating the reference destination is biased and generated by predicting the reference block in the merge process. The amount of codes can be reduced, and the encoding efficiency can be improved. Conventionally, when the left or upper adjacent block is designated as the reference destination, the reference destination flag has a frequency of ½ and the arithmetic coding cannot be used. However, according to the first embodiment, the predicted reference is By using the flag merge_proble_flag indicating the probability of the previous, if the prediction is correct, the bit frequency of the flag can be biased, and the amount of generated codes can be reduced in arithmetic coding. In addition, the adjacent block that matches the size of the encoding target block, or the adjacent block having a longer side length is highly correlated with the encoding target block, so that the reference block It is possible to improve the accuracy of prediction by predicting.
[実施例2]
 本発明の実施の形態による動き情報の参照先予測方法の第2の実施例について説明する。第1の実施例との違いは、参照先予測の判定条件として、処理対象のパーティションに隣接する左或いは上のパーティションの残差信号に基づく判定処理が追加される点である。
[Example 2]
A second example of the motion information reference destination prediction method according to the embodiment of the present invention will be described. The difference from the first embodiment is that a determination process based on the residual signal of the left or upper partition adjacent to the processing target partition is added as a determination condition for the reference destination prediction.
 図22は、動き情報の参照先予測方法の第2の実施例の動作を説明するフローチャートである。図22に示されるステップS600及びS610の処理は、図14に示される第1の実施例のステップS500及びS510の処理と同一である。処理対象のパーティションの左或いは上に隣接する周囲のパーティションとの境界判定を行い(S600)、処理対象のパーティションと隣接する左或いは上のパーティションの互いに隣接する辺の共通部分の長さ(「隣接辺長」)による判定を行う(S610)。 FIG. 22 is a flowchart for explaining the operation of the second embodiment of the motion information reference destination prediction method. The processes of steps S600 and S610 shown in FIG. 22 are the same as the processes of steps S500 and S510 of the first embodiment shown in FIG. A boundary determination is made with the surrounding partition adjacent to the left or upper side of the partition to be processed (S600), and the length of the common part of adjacent sides of the left or upper partition adjacent to the processing target partition ("adjacent" The side length ") is determined (S610).
 処理対象のパーティションと隣接する左或いは上のパーティションとの隣接辺長による判定(S610)の後、参照先としてパーティションが決定されなかった場合、残差信号に基づく判定(S620)に進む。 After the determination based on the adjacent side length between the left or upper partition adjacent to the partition to be processed (S610), if the partition is not determined as the reference destination, the process proceeds to the determination based on the residual signal (S620).
 ここで、残差信号に基づく判定処理を実施する際のパラメータについて説明する。図23は、隣接するパーティションの境界残差を算出する画素領域を示す図である。図23に示されるように、処理対象のパーティションXに対して隣接する左のパーティションをA、上のパーティションをBとする。それぞれのパーティションA,Bは、符号化過程における動き補償予測後の画素の残差信号、或いは復号過程における逆量子化・逆直交変換後の画素の残差信号で構成されているものとする。 Here, parameters when performing the determination process based on the residual signal will be described. FIG. 23 is a diagram illustrating a pixel region for calculating a boundary residual between adjacent partitions. As shown in FIG. 23, let A be the left partition adjacent to partition X to be processed, and B be the upper partition. Each partition A, B is assumed to be composed of a pixel residual signal after motion compensation prediction in the encoding process or a pixel residual signal after inverse quantization and inverse orthogonal transformation in the decoding process.
 パーティションAの画素でパーティションXに接している画素領域は、パーティションXの左側の境界に位置する高さがmin(ha,hx)、幅が1ピクセルの領域で表され、その画素の残差信号の絶対値の総和をδ(A)とすると、δ(A)は次式で表される。 The pixel area of the partition A that is in contact with the partition X is represented by an area having a height of min (ha, hx) and a width of 1 pixel located at the left boundary of the partition X, and a residual signal of the pixel. Δ (A) is expressed by the following equation, where δ (A) is the sum of absolute values of.
Figure JPOXMLDOC01-appb-M000001
Figure JPOXMLDOC01-appb-M000001
 ここで、aiはパーティションXに隣接しているパーティションAの画素の残差信号を表す。パーティションXに隣接しているパーティションAの画素数は、パーティションAの高さを表す画素数haとパーティションXの高さhxとの小さい方であり、関数min(ha,hx)で表される。 Here, ai represents the residual signal of the pixel in partition A adjacent to partition X. The number of pixels in the partition A adjacent to the partition X is the smaller of the number of pixels ha representing the height of the partition A and the height hx of the partition X, and is represented by a function min (ha, hx).
 同様に、パーティションBの画素でパーティションXに接している画素領域は、パーティションXの上側の境界に位置する幅がmin(wb,wx)、高さが1ピクセルの領域で表され、その画素の残差信号絶対値の総和をδ(B)とすると、δ(B)は次式で表される。 Similarly, the pixel area of the partition B that is in contact with the partition X is represented by an area having a width of min (wb, wx) and a height of 1 pixel located at the upper boundary of the partition X. If the sum of absolute values of residual signals is δ (B), δ (B) is expressed by the following equation.
Figure JPOXMLDOC01-appb-M000002
Figure JPOXMLDOC01-appb-M000002
 ここで、biはパーティションXに隣接しているパーティションBの画素の残差信号を表す。パーティションXに隣接しているパーティションBの画素数は、パーティションBの幅を表す画素数wbとパーティションXの幅wxとの小さい方であり、関数min(wb,wx)で表される。 Here, bi represents the residual signal of the pixel in partition B adjacent to partition X. The number of pixels in the partition B adjacent to the partition X is the smaller of the number of pixels wb representing the width of the partition B and the width wx of the partition X, and is represented by a function min (wb, wx).
 図24は、隣接するパーティションの境界残差に基づく参照先パーティションの判定の動作を説明するフローチャートである。まず、処理対象のパーティションXに隣接する左或いは上のパーティションのパーティションXとの境界残差として、上述した画素の残差信号絶対値の総和が算出される(S621)。 FIG. 24 is a flowchart for explaining the operation of determining the reference destination partition based on the boundary residual between adjacent partitions. First, as a boundary residual with the partition X of the left or upper partition adjacent to the partition X to be processed, the sum of the absolute value of the residual signal of the pixel is calculated (S621).
 次に算出された境界残差の正規化を行う(S622)。図23に示されるように、パーティションXに隣接しているパーティションA及びBの隣接辺長が異なる場合、パーティションAの境界残差とパーティションBの境界残差を直接比較することは適切ではないため、次式のように、パーティションA、Bの境界残差をそれぞれパーティションA、Bの隣接辺長で割ることにより正規化する。 Next, normalization of the calculated boundary residual is performed (S622). As shown in FIG. 23, when the adjacent side lengths of the partitions A and B adjacent to the partition X are different, it is not appropriate to directly compare the boundary residual of the partition A and the boundary residual of the partition B. As shown in the following equation, normalization is performed by dividing the boundary residuals of the partitions A and B by the adjacent side lengths of the partitions A and B, respectively.
Figure JPOXMLDOC01-appb-M000003
Figure JPOXMLDOC01-appb-M000003
Figure JPOXMLDOC01-appb-M000004
Figure JPOXMLDOC01-appb-M000004
 パーティションAおよびBの正規化された境界残差を比較する。まず、パーティションAの境界残差δ(A)とパーティションBの境界残差δ(B)とが等しい場合か否かを判定する(S623)。δ(A)とδ(B)とが等しくない場合はδ(A)とδ(B)とで大小関係が成立するので、大小比較を行う。ここでは、δ(A)がδ(B)より小さいか否かを判定する(S624)。処理対象のパーティションXに隣接する画素領域で動き補償予測で算出される残差が小さくなる動き情報を選択してパーティションXに適用した方が、パーティションXの予測誤差を小さくすることができると推測される。そこで、δ(A)がδ(B)より小さい場合はパーティションAを参照先として選択し(S625)、そうでない場合はパーティションBを参照先として選択する(S626)。 Compare the normalized boundary residuals of partitions A and B. First, it is determined whether or not the boundary residual δ (A) of the partition A is equal to the boundary residual δ (B) of the partition B (S623). If δ (A) and δ (B) are not equal, δ (A) and δ (B) have a magnitude relationship, so a magnitude comparison is performed. Here, it is determined whether or not δ (A) is smaller than δ (B) (S624). It is presumed that the prediction error of partition X can be reduced by selecting motion information in which the residual calculated by motion compensated prediction becomes smaller in the pixel region adjacent to the processing target partition X and applying it to partition X. Is done. Therefore, if δ (A) is smaller than δ (B), partition A is selected as a reference destination (S625), and if not, partition B is selected as a reference destination (S626).
 パーティションAの境界残差δ(A)とパーティションBの境界残差δ(B)とが等しい場合は、境界残差による参照先の判定はできないので、残差比の算出に進む(S627)。ここで、パーティションA、Bの残差比Δ(A),Δ(B)とは、ステップS621にて算出されたパーティションA、Bの正規化前の境界残差δ(A)、δ(B)のパーティションA、Bの総残差に対する比として次式で表される。 If the boundary residual δ (A) of the partition A is equal to the boundary residual δ (B) of the partition B, the reference destination cannot be determined by the boundary residual, and the process proceeds to calculation of the residual ratio (S627). Here, the residual ratios Δ (A) and Δ (B) of the partitions A and B are the boundary residuals δ (A) and δ (B) before normalization of the partitions A and B calculated in step S621. ) To the total residual of partitions A and B is expressed by the following equation.
Figure JPOXMLDOC01-appb-M000005
Figure JPOXMLDOC01-appb-M000005
Figure JPOXMLDOC01-appb-M000006
Figure JPOXMLDOC01-appb-M000006
 ここで、Ai及びBiはパーティションXに隣接しているパーティションA及びBの画素の残差信号であり、nA及びnBはパーティションA及びBの画素の総数である。但し、処理対象のパーティションXに隣接していない領域については対象とはならない。例えば、図23のパーティションAにおいて、ha>hxの場合、隣接辺長はmin(ha,hx)=hxであるから、パーティションAの高さhxまでの上側の矩形領域の画素が計算の対象であり、それを超える高さ(ha-hx)のパーティションAの下側の矩形領域の画素は対象外とする。更に、パーティションAの幅を表す画素数waとパーティションBの高さを表す画素数hbとを比較し、小さい方を対象領域の幅として定義する。図23で示される例では、対象領域の幅はmin(wa,hb)=hbとなる。したがって、対象領域は、幅hb、高さhxの斜線で示される領域であり、パーティションAの対象領域の画素数nAは次式で与えられる。 Here, Ai and Bi are residual signals of pixels in partitions A and B adjacent to partition X, and nA and nB are the total number of pixels in partitions A and B. However, an area that is not adjacent to the partition X to be processed is not a target. For example, in the partition A of FIG. 23, when ha> hx, the adjacent side length is min (ha, hx) = hx, and therefore the pixels in the upper rectangular area up to the height hx of the partition A are the objects of calculation. Yes, pixels in the rectangular area below the partition A having a height (ha-hx) exceeding the above are excluded. Further, the number of pixels wa representing the width of the partition A is compared with the number of pixels hb representing the height of the partition B, and the smaller one is defined as the width of the target region. In the example shown in FIG. 23, the width of the target region is min (wa, hb) = hb. Therefore, the target area is an area indicated by oblique lines having a width hb and a height hx, and the number of pixels nA of the target area of partition A is given by the following equation.
Figure JPOXMLDOC01-appb-M000007
Figure JPOXMLDOC01-appb-M000007
 同様に、パーティションBの対象領域の幅はパーティションBの隣接辺長min(wb,wx)=wbであり、高さはパーティションAの幅を表す画素数waとパーティションBの高さを表す画素数hbの小さい方であるからmin(wa,hb)=hbである。パーティションBの対象領域の画素数nBは次式で与えられる。 Similarly, the width of the target area of the partition B is the adjacent side length min (wb, wx) = wb of the partition B, and the height is the number of pixels wa indicating the width of the partition A and the number of pixels indicating the height of the partition B. Since hb is smaller, min (wa, hb) = hb. The number of pixels nB in the target area of partition B is given by the following equation.
Figure JPOXMLDOC01-appb-M000008
Figure JPOXMLDOC01-appb-M000008
 nAは、図23のパーティションA内の太点線内の薄灰色で示される矩形領域が対象となり、Δ(A)の算出式の分母は、この矩形領域内の画素の残差信号の絶対値の総和として表される。一方、nBは図23のパーティションBの全領域が対象となり、Δ(B)の算出式の分母部はパーティションB内の画素の残差信号の絶対値の総和として表される。正規化前の境界残差δ(A)及びδ(B)はそれぞれパーティションXに隣接する画素の残差信号の絶対値の総和であり、それぞれパーティションXに隣接する画素数はmin(ha,hx)=hx、min(wb,wx)=wbであるので、残差比Δ(A)及びΔ(B)は同率の画素数比で算出され、Δ(A)及びΔ(B)はパーティションXと隣接する画素の残差信号がパーティションA及びBの画素の残差信号に占める比率を表すことになる。この比率が高いということは、パーティションXと隣接する画素付近の動き補償予測が当たっていない、即ち、動き補償予測に使用した動き情報がパーティションXと隣接する画素付近の予測に用いるには適切でないことを意味する。即ち、残差比が高い方の隣接するパーティションの動き情報を参照しないと判定することになる。 nA is a rectangular area indicated by light gray in the thick dotted line in the partition A of FIG. 23, and the denominator of the calculation formula of Δ (A) is the absolute value of the residual signal of the pixel in this rectangular area. Expressed as the sum. On the other hand, nB covers the entire area of partition B in FIG. 23, and the denominator part of the calculation formula of Δ (B) is expressed as the sum of absolute values of residual signals of pixels in partition B. The boundary residuals δ (A) and δ (B) before normalization are sums of absolute values of residual signals of pixels adjacent to the partition X, respectively, and the number of pixels adjacent to the partition X is min (ha, hx). ) = Hx, min (wb, wx) = wb, the residual ratios Δ (A) and Δ (B) are calculated with the same number of pixel ratios, and Δ (A) and Δ (B) are the partition X Represents the ratio of the residual signals of adjacent pixels to the residual signals of the pixels in partitions A and B. If this ratio is high, the motion compensation prediction near the pixel adjacent to the partition X is not applied, that is, the motion information used for the motion compensation prediction is not appropriate for the prediction near the pixel adjacent to the partition X. Means that. That is, it is determined that the motion information of the adjacent partition having the higher residual ratio is not referred to.
 こうして算出された残差比の比較を行う。Δ(A)とΔ(B)とが等しい場合か否かを判定する(S628)。Δ(A)とΔ(B)とが等しくない場合はΔ(A)とΔ(B)とで大小関係が成立するので、大小比較を行う。ここでは、Δ(A)がΔ(B)より小さいか否かを判定する(S629)。Δ(A)がΔ(B)より小さい場合はAを選択し(S630)、そうでない場合はBを選択する(S631)。Δ(A)とΔ(B)とが等しい場合は、どちらを選択しても発生する残差に差が無いので、Aを選択することとして(S632)、終了する。 The residual ratio calculated in this way is compared. It is determined whether or not Δ (A) and Δ (B) are equal (S628). If Δ (A) and Δ (B) are not equal, Δ (A) and Δ (B) have a magnitude relationship, so a magnitude comparison is performed. Here, it is determined whether or not Δ (A) is smaller than Δ (B) (S629). If Δ (A) is smaller than Δ (B), A is selected (S630). Otherwise, B is selected (S631). If Δ (A) and Δ (B) are equal, there is no difference in the residual that occurs regardless of which one is selected, so A is selected (S632), and the process ends.
 以上のようにして、判定された参照先パーティションの予測結果を、符号化及び復号過程にフラグpred_direc_flagとして出力する。pred_direc_flagはmerge_direc_flagと同様に、参照先の隣接するパーティションとして選択される場合は”1(左)”或いは”0(上)”で参照先を表す。ステップS632での選択をAとしたのは、本実施例にてA(左)がビット”1”に割り当てられたためである。これは、動き情報の参照先パーティションの予測として、pred_direc_flagの発生頻度を”1”に偏らせることで、符号化において発生符号量を削減するためである。 As described above, the prediction result of the determined reference destination partition is output as the flag pred_direc_flag to the encoding and decoding processes. Similarly to merge_direc_flag, pred_direc_flag indicates a reference destination with “1 (left)” or “0 (upper)” when it is selected as an adjacent partition to be referred to. The reason for selecting A in step S632 is that A (left) is assigned to bit "1" in this embodiment. This is to reduce the amount of generated code in encoding by biasing the occurrence frequency of pred_direc_flag to “1” as prediction of a reference destination partition of motion information.
 尚、実施例2にて、図24で示されるステップS623において、境界残差δ(A)とδ(B)とが等しい場合か否かの厳密な判定を、δ(A)とδ(B)との絶対値差分|δ(A)-δ(B)|が閾値ε未満か否かとして判定を変更してもよい。δ(A)及びδ(B)の値が大きい場合の絶対値差分と小さい場合の絶対値差分とでは同じ差分値になったとしても、その意味合いは異なる。例えば、δ(A)=1000、δ(B)=1001の場合とδ(A)=8、δ(B)=7の場合とでは、共に絶対値差分は1となるが、前者の個々の境界残差の値から見れば誤差に等しい値を意味することになる。そこで、境界残差δ(A)とδ(B)の値の比較において、δ(A)とδ(B)との絶対値差分|δ(A)-δ(B)|と閾値εとの比較を行うことで、従来の境界残差δ(A)とδ(B)とが等しいか否かの場合(ε=0と等価)を含めるようにして、判定を拡張することが可能である。ここで、εは固定値としてもよいし、例えば量子化パラメータQpの値に応じて変更される可変値としてもよい。 In Example 2, in step S623 shown in FIG. 24, strict determination as to whether or not the boundary residuals δ (A) and δ (B) are equal is performed using δ (A) and δ (B The determination may be changed based on whether or not the absolute value difference | δ (A) −δ (B) | Even if the absolute value difference when the values of δ (A) and δ (B) are large and the absolute value difference when the values are small, the meanings are different. For example, in the case of δ (A) = 1000 and δ (B) = 1001 and in the case of δ (A) = 8 and δ (B) = 7, the absolute value difference is 1, but the former individual In terms of the boundary residual value, this means a value equal to the error. Therefore, in the comparison of the values of the boundary residuals δ (A) and δ (B), the absolute value difference | δ (A) −δ (B) | between δ (A) and δ (B) and the threshold value ε By performing the comparison, it is possible to extend the determination by including the case where the conventional boundary residuals δ (A) and δ (B) are equal (equivalent to ε = 0). . Here, ε may be a fixed value or may be a variable value that is changed according to the value of the quantization parameter Qp, for example.
 更に、実施例2では境界残差及び残差比の2つの差分判定を順に実施したが、境界残差と残差比をそれぞれで判定を行い、それらの結果から最終的な参照先パーティションの選択をしても良い。また、境界残差及び残差比を纏めて新たな評価基準を生成し、判定しても良い。例えば、次式のように境界残差と残差比との組み合わせによって生成される評価値に基づいて判定してもよい。 Furthermore, in the second embodiment, the difference determination of the boundary residual and the residual ratio is performed in order, but the boundary residual and the residual ratio are determined separately, and the final reference destination partition is selected from the results. You may do it. Further, the boundary residual and the residual ratio may be collected and a new evaluation criterion may be generated and determined. For example, the determination may be made based on the evaluation value generated by the combination of the boundary residual and the residual ratio as in the following equation.
Figure JPOXMLDOC01-appb-M000009
Figure JPOXMLDOC01-appb-M000009
 ここで、変数xはA若しくはBであり、αは重み付け係数である。αは固定値としてもよいし、パーティションやスライス単位、或いはスライスタイプ等に応じて可変値として設定してもよい。 Here, the variable x is A or B, and α is a weighting coefficient. α may be a fixed value, or may be set as a variable value according to a partition, a slice unit, a slice type, or the like.
 以上述べたように、第2の実施例によれば、符号化対象のブロックの周囲に隣接するブロックの残差信号の大きさに基づいて、参照先のブロックを予測することで参照先の予測の精度をさらに高めることができる。 As described above, according to the second embodiment, prediction of a reference destination is performed by predicting a reference block based on the magnitude of a residual signal of a block adjacent to the block to be encoded. Accuracy can be further increased.
[実施例3]
 本発明の実施の形態による動き情報の参照先予測方法の第3の実施例について説明する。第1の実施例との違いは、参照先パーティションの予測により判定され出力されるフラグpred_direc_flagで表されるパーティションを参照先パーティションとして選定する点である。参照先パーティションの予測によるフラグpred_direc_flagが0或いは1の場合は、参照先の隣接するパーティションとして左(pred_direc_flagが”1”)或いは上(pred_direc_flagが”0”)が選択される。
[Example 3]
A third example of the motion information reference destination prediction method according to the embodiment of the present invention will be described. The difference from the first embodiment is that the partition represented by the flag pred_direc_flag determined and output by prediction of the reference destination partition is selected as the reference destination partition. When the flag pred_direc_flag based on prediction of the reference destination partition is 0 or 1, the left (pred_direc_flag is “1”) or the top (pred_dirc_flag is “0”) is selected as the reference destination adjacent partition.
 参照先パーティションの予測において不定として判定された場合、即ちpred_direc_flagが”2”の場合は、実際に発生符号量が最小となる予測方法として選択されたマージ処理による予測で用いた動き情報の参照先パーティションを表すフラグmerge_direc_flagを、参照先のパーティションを表すものとして用いる。この条件をシンタックスとして表現すると図25となる。パーティション単位にマージするか否かを判定するmerge_flagの後に続いて、『マージ処理を実行し、且つ周囲に隣接する有効なパーティションの数NumMergeCandidatesが1よりも大きく、且つ参照先パーティションの予測で不定』である場合に、参照先パーティションを表すフラグmerge_direc_flagが設定される。マージ処理による予測が選択された場合、上述したシンタックスで定義されるマージ処理用のフラグが設定される。図26のフローチャートを使って、符号化における動き情報の参照先予測方法の第3の実施例の動作を説明する。 When it is determined as indefinite in the prediction of the reference destination partition, that is, when pred_direc_flag is “2”, the reference destination of the motion information used in the prediction by the merge processing selected as the prediction method that actually minimizes the generated code amount A flag merge_direc_flag representing a partition is used to represent a partition to be referred to. When this condition is expressed as syntax, FIG. 25 is obtained. Subsequent to merge_flag for determining whether or not to merge in units of partitions, “the number of valid partitions adjacent to the execution of merge processing and NumMergeCandidates is greater than 1 and undefined in prediction of reference destination partition” Is set, a flag merge_direc_flag indicating a reference destination partition is set. When prediction by merge processing is selected, a flag for merge processing defined by the syntax described above is set. The operation of the third embodiment of the motion information reference destination prediction method in encoding will be described using the flowchart of FIG.
 図26はマージ処理用のフラグを設定する手順を説明するフローチャートである。最初に予測選択部112により選択された予測方法がマージ処理で選択された動き情報に基づく予測であるか否かの判定が行われる(S701)。画像内予測や通常の動きベクトル検出を用いた画面間予測の場合はmerge_flagを0に設定して、終了する(S703)。マージ処理で選択された動き情報に基づく予測であれば、merge_flagを1に設定する(S702)。 FIG. 26 is a flowchart for explaining a procedure for setting a flag for merge processing. First, it is determined whether or not the prediction method selected by the prediction selection unit 112 is prediction based on motion information selected in the merge process (S701). In the case of inter-screen prediction using intra-picture prediction or normal motion vector detection, merge_flag is set to 0 and the process ends (S703). If the prediction is based on the motion information selected in the merge process, merge_flag is set to 1 (S702).
 次に、マージ処理で選択された動き情報の数を調べる(S704)。参照先指定部113が指定する参照先となる隣接するパーティションの候補数NumMergeCandidatesが1より大きいか否かの判定が行われる。候補数NumMergeCandidatesが1以下の場合、マージ処理を選択していることが前提であり、必ず有効な参照先の隣接するパーティションが1つ存在し、左或いは上のどちらか一方を特定することができるので、merge_flagのみを設定し、終了する。 Next, the number of motion information selected by the merge process is checked (S704). It is determined whether or not the number of adjacent partition candidates NumMergeCandidates to be specified by the reference destination specifying unit 113 is greater than one. When the number of candidates NumMergeCandidates is 1 or less, it is assumed that merge processing is selected, and there is always one adjacent partition of a valid reference destination, and either the left or the upper can be specified. Therefore, only merge_flag is set and the process ends.
 候補数NumMergeCandidatesが1より大きい場合、即ち2個の隣接するパーティションの動き情報の候補があった場合は参照先パーティションの予測に進む(S705)。ここで、参照先パーティションの予測では、第1の実施例にて説明した動き情報の参照先予測方法が使用される。処理対象のパーティションの左及び上に隣接するパーティションのサイズや、処理対象パーティションと隣接パーティションが互いに接している辺の長さ等の情報に基づいて、参照先パーティションの予測が行われ、フラグpred_direc_flagが出力される。 If the number of candidates NumMergeCandidates is larger than 1, that is, if there are motion information candidates of two adjacent partitions, the process proceeds to prediction of a reference destination partition (S705). Here, in the prediction of the reference destination partition, the motion information reference destination prediction method described in the first embodiment is used. Based on information such as the size of the partition adjacent to the left and top of the target partition and the length of the side where the target partition and the adjacent partition are in contact with each other, the reference destination partition is predicted, and the flag pred_dirc_flag is set. Is output.
 予測された参照先パーティションを示すフラグpred_direc_flagはmerge_direc_flagと同様に、参照先の隣接するパーティションとして左(1)或いは上(0)で選択先を表されるとともに、予測によって参照先パーティションが選択できなかった場合の不定(2)が設定される。 Like the merge_direc_flag, the flag pred_dirc_flag indicating the predicted reference destination partition represents the selection destination on the left (1) or above (0) as the adjacent partition of the reference destination, and the reference destination partition cannot be selected by prediction. Indeterminate (2) is set.
 次に、このようにして設定された、予測された参照先パーティションを示すフラグpred_direc_flagの判定を行う(S706)。予測された参照先パーティションを示すフラグpred_direc_flagが2となるのは参照先パーティションの予測において予測不可の場合であり、この時、実際に発生符号量が最小となる予測方法として選択されたマージ処理による動き補償予測で用いた動き情報の参照先パーティションを表すmerge_direc_flagを設定して、終了する(S707)。一方、pred_direc_flagが2でない場合は、pred_direc_flagが参照先のパーティションを直接示すことになり、”0”の場合は上を、”1”の場合は左に隣接するパーティションを選択する。 Next, the flag pred_direc_flag indicating the predicted reference destination partition set in this way is determined (S706). The flag pred_dirc_flag indicating the predicted reference destination partition is 2 when prediction is impossible in the prediction of the reference destination partition, and at this time, it depends on the merge process selected as the prediction method that actually minimizes the generated code amount. The process sets the merge_dir_flag indicating the reference partition of the motion information used in the motion compensation prediction, and the process ends (S707). On the other hand, when pred_direc_flag is not 2, pred_direc_flag directly indicates a reference destination partition, and when “0” is selected, a partition adjacent to the left is selected.
 参照先パーティションの予測の成否を示すフラグpred_direc_flagにより、参照先パーティションの予測が当たり、参照先として隣接するパーティションが選択される確率が高い程、参照先パーティションを示すフラグmerge_direc_flagを伝送する必要性がなくなるので、符号化効率が向上することになる。 With the flag pred_dirc_flag indicating the success or failure of the prediction of the reference destination partition, it is unnecessary to transmit the flag merge_direc_flag indicating the reference destination partition as the probability that the adjacent partition is selected as the reference destination is higher. Therefore, the encoding efficiency is improved.
 以上のようにして、パーティション毎にマージ処理により選択された隣接するパーティションの動き情報に基づく動き補償予測が選択された場合、SPSのヘッダーに記載される第1フラグinter_merge_flag、予測ブロックに記載される第2フラグmerge_flag及び第3フラグmerge_direc_flagが設定され、符号化される。 As described above, when motion-compensated prediction based on the motion information of adjacent partitions selected by the merge process for each partition is selected, the first flag inter_merge_flag described in the header of the SPS is described in the prediction block. The second flag merge_flag and the third flag merge_direc_flag are set and encoded.
 次に、上述のシンタックスに基づき、符号化された動画像のビットストリームを復号する場合の参照先予測方法の動作を説明する。 Next, the operation of the reference destination prediction method in the case of decoding the encoded moving image bit stream based on the above-described syntax will be described.
 最初に、可変長復号部201により復号されるビットストリームのSPSに記述されているフラグinter_merge_flagから、ビットストリームのシーケンス全体に対して、本実施の形態に係る動き情報の参照先予測方法を使用したマージ処理による画面間予測が適用されるか否かを判定する。inter_merge_flagが真(1)の場合、マージ処理が適用され、inter_merge_flagが偽(0)の場合、マージ処理が無視され、従来通りのビットストリームから復号された動き情報に基づく画面間予測が実施される。以下、inter_merge_flagが真(1)、即ち、本実施の形態に係る動き情報の参照先予測方法を使用したマージ処理が適用されているものとして説明する。 First, from the flag inter_merge_flag described in the SPS of the bitstream decoded by the variable length decoding unit 201, the motion information reference destination prediction method according to the present embodiment is used for the entire bitstream sequence. It is determined whether inter-screen prediction by merge processing is applied. When inter_merge_flag is true (1), merge processing is applied. When inter_merge_flag is false (0), merge processing is ignored and inter-screen prediction based on motion information decoded from a conventional bitstream is performed. . In the following description, it is assumed that inter_merge_flag is true (1), that is, the merge processing using the motion information reference destination prediction method according to the present embodiment is applied.
 次に、画面内/画面間予測選択部209により、スライスの中の予測ブロック毎に予測モードを参照して画面内または画面間予測の何れを選択するかが判定される。画面内予測(MODE_INTRA)が選択された場合は画面内予測部208の処理に進み、画面間予測(MODE_INTER)が選択された場合は動き情報選択部210の処理に進む。 Next, the intra-screen / inter-screen prediction selection unit 209 determines whether to select intra-screen or inter-screen prediction with reference to the prediction mode for each prediction block in the slice. When the intra prediction (MODE_INTRA) is selected, the process proceeds to the intra prediction unit 208, and when the inter prediction (MODE_INTER) is selected, the process proceeds to the motion information selection unit 210.
 ここで、動き情報選択部210は、予測ブロックのパーティション毎にビットストリームから復号された情報として、符号化の動きベクトル検出にて検出された動き情報、或いはマージ処理により選択された参照先パーティションの動き情報のどちらか一方を選択し、選択された動き情報を動き補償部203に出力する。 Here, the motion information selection unit 210 uses, as information decoded from the bit stream for each partition of the prediction block, the motion information detected by the motion vector detection of encoding or the reference destination partition selected by the merge process. Either one of the motion information is selected, and the selected motion information is output to the motion compensation unit 203.
 図27は、動き情報選択部210の選択手順を示すフローチャートであり、この図を参照して説明する。まず、処理対象のパーティションに隣接する周辺のパーティションの候補数NumMergeCandidatesが算出される(S801)。算出手順は符号化におけるS701と同様であるので割愛する。 FIG. 27 is a flowchart showing the selection procedure of the motion information selection unit 210, which will be described with reference to this figure. First, the number NumMergeCandidates of neighboring partitions adjacent to the partition to be processed is calculated (S801). Since the calculation procedure is the same as S701 in encoding, it is omitted.
 算出された候補数NumMergeCandidatesが0個より大きいか否かが判定される(S802)。0個の場合、マージ処理の参照先となる隣接するパーティションが存在しないので、動き情報選択部210は、従来同様ビットストリームから復号された動き情報を出力する。0個より大きい場合は、可変長復号部201により復号されたmerge_flagが読み込まれ(S803)、merge_flagに基づき動き情報の選択を判定する(S804)。merge_flagが偽(0)の場合、動き情報選択部210は、従来同様ビットストリームから復号された動き情報を出力する。merge_flagが真(1)の場合、マージ処理での参照先パーティションの動き情報を選択する為の処理に移る。 It is determined whether or not the calculated number of candidates NumMergeCandidates is greater than 0 (S802). In the case of 0, since there is no adjacent partition that becomes the reference destination of the merge process, the motion information selection unit 210 outputs the motion information decoded from the bitstream as in the conventional case. When the number is larger than 0, merge_flag decoded by the variable length decoding unit 201 is read (S803), and the selection of motion information is determined based on the merge_flag (S804). When merge_flag is false (0), the motion information selection unit 210 outputs motion information decoded from the bitstream as in the conventional case. When merge_flag is true (1), the process proceeds to processing for selecting motion information of the reference partition in the merge processing.
 再度、候補数NumMergeCandidatesが1個より大きいか否かの判定が行われる(S805)。候補数NumMergeCandidatesが1個の場合、その1個がマージ処理の参照先パーティションとなるので、その参照先パーティションの動き情報を、復号済み情報が記録されているメモリ207から取得する(S807)。取得された動き情報は動き補償部203に供給され、その動き情報にもとづいて動き補償が行われる。 Again, it is determined whether or not the number of candidates NumMergeCandidates is greater than one (S805). If the number of candidates NumMergeCandidates is one, one of them becomes a reference destination partition for the merge process, so the motion information of the reference destination partition is acquired from the memory 207 in which the decoded information is recorded (S807). The acquired motion information is supplied to the motion compensation unit 203, and motion compensation is performed based on the motion information.
 候補数NumMergeCandidatesが1個より大きい場合、即ち2個の場合、処理対象のパーティションの左と上に隣接するパーティションの動き情報をメモリ207から読み出し、取得する(S806)。取得された左と上に隣接するパーティションの動き情報を比較する(S808)。具体的には、動き情報を構成する予測方向、参照ピクチャ番号及び動きベクトルを比較する。動き情報が全く同一の場合、左或いは上のどちらか一方の隣接するパーティションの動き情報を取得して、動き補償部203に出力する(S810)。ここでは、符号化と同様に左に隣接するパーティションの動き情報を選択することにする。 If the number of candidates NumMergeCandidates is larger than 1, that is, two, the motion information of the partitions adjacent to the left and above the processing target partition is read from the memory 207 and acquired (S806). The obtained motion information of the left and upper adjacent partitions is compared (S808). Specifically, the prediction direction, the reference picture number, and the motion vector constituting the motion information are compared. If the motion information is exactly the same, the motion information of either the left or upper adjacent partition is acquired and output to the motion compensation unit 203 (S810). Here, the motion information of the partition adjacent to the left is selected as in the encoding.
 動き情報が異なる場合は参照先パーティションの予測に進む(S809)。参照先パーティションの予測は、符号化と同様に、実施例1にて説明された動き情報の参照先予測方法が使用される。処理対象のパーティションと周辺の左及び上に隣接するパーティションのサイズや、処理対象パーティションと隣接パーティションが互いに接している辺の長さ等の情報に基づいて、参照先パーティションの予測が行われ、予測された参照先パーティションを示すフラグpred_direc_flagが出力される。 If the motion information is different, the process proceeds to prediction of the reference destination partition (S809). For the prediction of the reference destination partition, the motion information reference destination prediction method described in the first embodiment is used, similarly to the encoding. The target partition is predicted based on information such as the size of the partition adjacent to the processing target partition and the left and top of the periphery, and the length of the side where the processing target partition and the adjacent partition are in contact with each other. The flag pred_dirc_flag indicating the referred reference partition is output.
 予測された参照先パーティションを示すフラグpred_direc_flagは、符号化と同様に参照先の隣接するパーティションが選択される場合は左(1)或いは上(0)で表され、予測によって参照先パーティションが選択出来なかった場合は不定(2)が設定される。 The flag pred_dirc_flag indicating the predicted reference destination partition is represented by left (1) or upper (0) when the adjacent reference destination partition is selected as in the encoding, and the reference destination partition can be selected by prediction. If not, undefined (2) is set.
 次に出力されたpred_direc_flagの判定を行う(S811)。pred_direc_flagが2となるのは参照先パーティションの予測で予測不可の場合であり、参照先パーティションの判定は行えない。そこで、可変長復号部201により復号されたmerge_direc_flagが読み込まれる(S812)。merge_direc_flagに基づき、参照先パーティションの選択を判定する(S813)。merge_direc_flagが1の場合、左に隣接するパーティション(S814)、merge_direc_flagが0の場合、上に隣接するパーティションを参照先パーティションとして選択する(S815)。 Next, the pred_direc_flag output is determined (S811). The pred_dirc_flag is 2 when prediction is impossible due to prediction of the reference destination partition, and the reference destination partition cannot be determined. Therefore, merge_direc_flag decoded by the variable length decoding unit 201 is read (S812). Based on merge_direc_flag, the selection of the reference destination partition is determined (S813). If merge_dirc_flag is 1, the left adjacent partition (S814) is selected. If merge_dirc_flag is 0, the upper adjacent partition is selected as the reference destination partition (S815).
 一方、pred_direc_flagが2でない場合はpred_direc_flagで示される値が直接参照先の隣接パーティションを示す。即ち、pred_direc_flagが1の場合、左に隣接するパーティション、merge_probable_flagが0の場合、上に隣接するパーティションを参照先パーティションとして選択する。こうして選択された参照先パーティションの動き情報が取得され、動き補償部203に供給され、終了する。 On the other hand, when pred_direc_flag is not 2, the value indicated by pred_direc_flag directly indicates the adjacent partition of the reference destination. That is, if pred_dirc_flag is 1, the left adjacent partition is selected, and if merge_proble_flag is 0, the upper adjacent partition is selected as a reference destination partition. The motion information of the reference destination partition thus selected is acquired, supplied to the motion compensation unit 203, and the process ends.
 以上のように、可変長復号部201により復号された各フラグと復号過程での演算結果に応じてマージ処理による画面間予測が選択された場合、前述した動き情報の参照先予測方法によって、参照先パーティションを選択する。選択された参照先パーティションの動き情報から動き補償により生成される予測画像とビットストリームから復号された残差信号と加算されることで復号画像が生成される。 As described above, when inter-screen prediction by merge processing is selected according to each flag decoded by the variable-length decoding unit 201 and the calculation result in the decoding process, the reference is made by the above-described motion information reference destination prediction method. Select the destination partition. A decoded image is generated by adding the predicted image generated by motion compensation from the motion information of the selected reference destination partition and the residual signal decoded from the bitstream.
 更に、第2の実施例にて説明した処理対象のパーティションに隣接する左或いは上のパーティションの残差信号に基づく判定処理を同様に適用することも可能である。 Furthermore, the determination processing based on the residual signal of the left or upper partition adjacent to the partition to be processed described in the second embodiment can be applied in the same manner.
 以上述べたように、第3の実施例によれば、参照先の予測による参照先となる隣接ブロックの判定の一致精度が高いものとして、merge_probable_flagそのものを定義せず、参照先の予測により得られた判定結果で参照先の隣接ブロックを示し、参照先の予測が不定の場合にのみフラグmerge_direc_flagで参照先の隣接ブロックを示してもよい。この場合、merge_direc_flagによる符号量の増加をmerge_probable_flagに要する符号量の削減で相殺し、相対的に符号化効率を向上させることができる。 As described above, according to the third embodiment, it is obtained by prediction of reference destination without defining merge_proble_flag itself, assuming that the matching accuracy of the determination of the adjacent block that is the reference destination by the prediction of the reference destination is high. The reference destination neighboring block may be indicated by the determination result, and the reference destination neighboring block may be indicated by the flag merge_dirc_flag only when the prediction of the reference destination is indefinite. In this case, the increase in the code amount due to merge_direct_flag can be offset by the reduction in the code amount required for merge_proble_flag, and the encoding efficiency can be relatively improved.
 更に、上述した第3の実施例において、参照先パーティションの予測で必ず左或いは上に隣接するパーティションを参照先パーティションとして指定するようにして、参照先パーティションの予測により不定と判定されることを無くしてもよい。参照先パーティションの予測で不定として判定されていた場合に、左或いは上の隣接パーティションを一意に指定することを予め既定しておくことで、符号化における動き情報の参照先予測方法では、参照先パーティションの予測により不定判定された場合に参照先パーティションを表すフラグmerge_direc_flagを符号化、伝送する必要性が無くなるので、符号化効率を向上させることができる。符号化に対応した復号における動き情報の参照先予測方法では、符号化側と同じ参照先パーティションの予測を行い、参照先パーティションの予測で判定された左或いは上に隣接するパーティションを参照先パーティションとして選択する。 Furthermore, in the third embodiment described above, the left or upper adjacent partition is always designated as the reference destination partition in the prediction of the reference destination partition, so that it is not determined that the reference destination partition is indeterminate by the prediction of the reference destination partition. May be. When it is determined as indefinite in the prediction of the reference destination partition, it is predetermined in advance that the adjacent partition on the left or the upper side is uniquely specified. Since it is not necessary to encode and transmit the flag merge_dirc_flag indicating the reference destination partition when it is determined indefinitely by partition prediction, it is possible to improve the encoding efficiency. In the reference destination prediction method of motion information in decoding corresponding to encoding, the same reference destination partition as that on the encoding side is predicted, and the left or upper adjacent partition determined by prediction of the reference destination partition is used as the reference destination partition. select.
 以上述べた実施の形態の動画像符号化装置が出力する動画像の符号化ストリームは、実施の形態で用いられた符号化方法に応じて復号することができるように特定のデータフォーマットを有しており、動画像符号化装置に対応する動画像復号装置がこの特定のデータフォーマットの符号化ストリームを復号することができる。 The moving image encoded stream output from the moving image encoding apparatus of the embodiment described above has a specific data format so that it can be decoded according to the encoding method used in the embodiment. Therefore, the moving picture decoding apparatus corresponding to the moving picture encoding apparatus can decode the encoded stream of this specific data format.
 動画像符号化装置と動画像復号装置の間で符号化ストリームをやりとりするために、有線または無線のネットワークが用いられる場合、符号化ストリームを通信路の伝送形態に適したデータ形式に変換して伝送してもよい。その場合、動画像符号化装置が出力する符号化ストリームを通信路の伝送形態に適したデータ形式の符号化データに変換してネットワークに送信する動画像送信装置と、ネットワークから符号化データを受信して符号化ストリームに復元して動画像復号装置に供給する動画像受信装置とが設けられる。 When a wired or wireless network is used to exchange an encoded stream between a moving image encoding device and a moving image decoding device, the encoded stream is converted into a data format suitable for the transmission form of the communication path. It may be transmitted. In that case, a video transmission apparatus that converts the encoded stream output from the video encoding apparatus into encoded data in a data format suitable for the transmission form of the communication channel and transmits the encoded data to the network, and receives the encoded data from the network Then, a moving image receiving apparatus that restores the encoded stream and supplies the encoded stream to the moving image decoding apparatus is provided.
 動画像送信装置は、動画像符号化装置が出力する符号化ストリームをバッファするメモリと、符号化ストリームをパケット化するパケット処理部と、パケット化された符号化データをネットワークを介して送信する送信部とを含む。動画像受信装置は、パケット化された符号化データをネットワークを介して受信する受信部と、受信された符号化データをバッファするメモリと、符号化データをパケット処理して符号化ストリームを生成し、動画像復号装置に提供するパケット処理部とを含む。 The moving image transmitting apparatus is a memory that buffers the encoded stream output from the moving image encoding apparatus, a packet processing unit that packetizes the encoded stream, and transmission that transmits the packetized encoded data via the network. Part. The moving image receiving apparatus generates a coded stream by packetizing the received data, a receiving unit that receives the packetized coded data via a network, a memory that buffers the received coded data, and packet processing. And a packet processing unit provided to the video decoding device.
 以上の符号化及び復号に関する処理は、ハードウェアを用いた伝送、蓄積、受信装置として実現することができるのは勿論のこと、ROM(リード・オンリ・メモリ)やフラッシュメモリ等に記憶されているファームウェアや、コンピュータ等のソフトウェアによっても実現することができる。そのファームウェアプログラム、ソフトウェアプログラムをコンピュータ等で読み取り可能な記録媒体に記録して提供することも、有線あるいは無線のネットワークを通してサーバから提供することも、地上波あるいは衛星ディジタル放送のデータ放送として提供することも可能である。 The above processing relating to encoding and decoding can be realized as a transmission, storage, and reception device using hardware, and is also stored in a ROM (Read Only Memory), a flash memory, or the like. It can also be realized by firmware or software such as a computer. The firmware program and software program can be provided by recording them on a computer-readable recording medium, provided from a server through a wired or wireless network, or provided as a data broadcast of terrestrial or satellite digital broadcasting. Is also possible.
 以上、本発明を実施の形態をもとに説明した。実施の形態は例示であり、それらの各構成要素や各処理プロセスの組合せにいろいろな変形例が可能なこと、またそうした変形例も本発明の範囲にあることは当業者に理解されるところである。 The present invention has been described based on the embodiments. The embodiments are exemplifications, and it will be understood by those skilled in the art that various modifications can be made to combinations of the respective constituent elements and processing processes, and such modifications are within the scope of the present invention. .
 101 動きベクトル検出部、 102 減算部、 103 直交変換・量子化部、 104 可変長符号化部、 105 逆量子化・逆直交変換部、 106 動き補償部、 107 重み付き予測部、 108 加算部、 109 デブロッキング・フィルタ部、 110 メモリ、 111 画面内予測部、 112 予測選択部、 113 参照先指定部、 201 可変長復号部、 202 逆量子化・逆直交変換部、 203 動き補償部、 204 重み付き予測部、 205 加算部、 206 デブロッキング・フィルタ部、 207 メモリ、 208 画面内予測部、 209 画面内/画面間予測選択部、 210 動き情報選択部。 101 motion vector detection unit, 102 subtraction unit, 103 orthogonal transform / quantization unit, 104 variable length coding unit, 105 inverse quantization / inverse orthogonal transform unit, 106 motion compensation unit, 107 weighted prediction unit, 108 addition unit, 109 deblocking filter unit, 110 memory, 111 intra prediction unit, 112 prediction selection unit, 113 reference destination designation unit, 201 variable length decoding unit, 202 inverse quantization / inverse orthogonal transform unit, 203 motion compensation unit, 204 weight Attached prediction unit, 205 addition unit, 206 deblocking filter unit, 207 memory, 208 intra-screen prediction unit, 209 intra-screen / inter-screen prediction selection unit, 210 motion information selection unit.
 本発明は、動画像の符号化及び復号技術に利用できる。 The present invention can be used for moving picture encoding and decoding techniques.

Claims (20)

  1.  動画像の各ピクチャを分割したブロック単位で動きベクトルを用いて前記動画像を符号化する動画像符号化装置であって、
     符号化対象ブロックに隣接する複数の符号化済みの隣接ブロックの中から、前記符号化対象ブロックと前記隣接ブロックとの隣接条件に基づいて判定されるいずれか一つの隣接ブロックを、動き情報を参照するための参照先ブロックとして指定する参照先指定部と、
     画像内予測、符号化対象ブロックの動きベクトルを利用した画像間予測、および隣接ブロックの動きベクトルを利用した画像間予測の中から、予測方法として、隣接ブロックの動きベクトルを利用した画像間予測が選択された場合、選択された予測方法により参照される隣接ブロックと前記参照先指定部により指定された前記参照先ブロックとが一致するか否かを表す参照先有効情報と選択された予測方法を示す情報とを動き情報に代わって生成する予測選択部とを含むことを特徴とする動画像符号化装置。
    A moving image encoding apparatus that encodes the moving image using a motion vector in units of blocks obtained by dividing each picture of the moving image,
    Refer to the motion information for any one of the neighboring blocks determined based on the neighboring condition between the coding target block and the neighboring block from among the plurality of coded neighboring blocks adjacent to the coding target block. A reference destination designating part that is designated as a reference destination block for
    Among intra prediction, inter prediction using motion vector of the encoding target block, and inter prediction using motion vector of adjacent block, inter prediction using motion vector of adjacent block is used as a prediction method. When selected, the reference destination valid information indicating whether or not the adjacent block referred to by the selected prediction method matches the reference destination block specified by the reference destination specifying unit and the selected prediction method And a prediction selection unit that generates information to be displayed instead of motion information.
  2.  前記参照先指定部は、前記隣接条件として、前記隣接ブロックの隣接する辺の長さが前記符号化対象ブロックの高さまたは幅と同じであるか否かを判定することにより、隣接する辺の長さが前記符号化対象ブロックの高さまたは幅と同じである隣接ブロックを選択し、選択された隣接ブロックを参照先ブロックとして指定することを特徴とする請求項1に記載の動画像符号化装置。 The reference destination designating unit determines, as the adjacency condition, whether or not the length of the adjacent side of the adjacent block is the same as the height or the width of the encoding target block. 2. The moving picture coding according to claim 1, wherein an adjacent block whose length is the same as a height or a width of the encoding target block is selected, and the selected adjacent block is designated as a reference block. apparatus.
  3.  前記参照先指定部は、前記隣接条件として、前記符号化対象ブロックと前記隣接ブロックの互いに隣接する辺の共通部分の長さの大小関係を判定することにより、隣接する辺の共通部分の長さが大きい方の隣接ブロックを選択し、選択された隣接ブロックを参照先ブロックとして指定することを特徴とする請求項1に記載の動画像符号化装置。 The reference destination designating unit determines, as the adjacency condition, the size of the common part of adjacent sides of the encoding target block and the adjacent block by determining the size relationship of the common parts of adjacent sides. The moving picture encoding apparatus according to claim 1, wherein an adjacent block having a larger value is selected, and the selected adjacent block is designated as a reference block.
  4.  前記参照先指定部は、前記隣接ブロックの残差信号にもとづいて、前記符号化対象ブロックに接している前記隣接ブロックの領域の残差信号の絶対値で表される境界残差を求め、境界残差の大小関係を判定することにより、境界残差が小さい方の隣接ブロックを選択し、選択された隣接ブロックを参照先ブロックとして指定することを特徴とする請求項1に記載の動画像符号化装置。 The reference destination designating unit obtains a boundary residual represented by an absolute value of a residual signal in an area of the adjacent block that is in contact with the encoding target block based on the residual signal of the adjacent block; 2. The moving picture code according to claim 1, wherein an adjacent block having a smaller boundary residual is selected by determining a magnitude relation of the residual, and the selected adjacent block is designated as a reference block. Device.
  5.  前記参照先指定部は、前記隣接ブロックの残差信号にもとづいて、前記符号化対象ブロックに接している前記隣接ブロックの領域の残差信号の絶対値で表される境界残差と、前記符号化対象ブロックに接している前記隣接ブロックの残差信号の絶対値で表される総残差から、総残差に対する境界残差の残差比を求め、残差比の大小関係を判定することにより、残差比が小さい方の隣接ブロックを選択し、選択された隣接ブロックを参照先ブロックとして指定することを特徴とする請求項1に記載の動画像符号化装置。 The reference destination specifying unit, based on the residual signal of the adjacent block, a boundary residual represented by an absolute value of a residual signal of the area of the adjacent block that is in contact with the encoding target block, and the code The residual ratio of the boundary residual to the total residual is obtained from the total residual represented by the absolute value of the residual signal of the adjacent block in contact with the block to be converted, and the magnitude relationship of the residual ratio is determined. The moving picture encoding apparatus according to claim 1, wherein an adjacent block having a smaller residual ratio is selected, and the selected adjacent block is designated as a reference block.
  6.  前記参照先指定部は、前記符号化対象ブロックのいずれかの辺に動き情報が同一である複数の前記隣接ブロックが連続して並んでいる場合、前記隣接する辺の長さは、これらの連続する前記隣接ブロックの辺の長さを加算したものとして、前記隣接条件を判定することを特徴とする請求項2または3に記載の動画像符号化装置。 When the plurality of adjacent blocks having the same motion information are continuously arranged on any side of the encoding target block, the reference destination designating unit determines the length of the adjacent sides The moving image encoding apparatus according to claim 2 or 3, wherein the adjacency condition is determined as an addition of lengths of sides of the adjacent blocks.
  7.  前記予測選択部は、予測方法として、隣接ブロックの動きベクトルを利用した画像間予測が選択された場合、前記参照先指定部により指定される前記参照先ブロックが不定であるときは、前記参照先有効情報の代わりに、選択された予測方法により参照される隣接ブロックを指定する参照先情報を生成することを特徴とする請求項1から6のいずれかに記載の動画像符号化装置。 When the inter-picture prediction using a motion vector of an adjacent block is selected as a prediction method, the prediction selection unit is configured to select the reference destination block when the reference destination block designated by the reference destination designation unit is indefinite. 7. The moving picture encoding apparatus according to claim 1, wherein reference destination information for specifying an adjacent block referred to by the selected prediction method is generated instead of the valid information.
  8.  前記予測選択部は、予測方法として、隣接ブロックの動きベクトルを利用した画像間予測が選択された場合、前記参照先指定部により指定される前記参照先ブロックが不定であるときは、予め既定値として設定した隣接ブロックを参照するように指定することを特徴とする請求項1から6のいずれかに記載の動画像符号化装置。 When the inter-picture prediction using the motion vector of the adjacent block is selected as the prediction method, the prediction selection unit is preset in advance when the reference block specified by the reference destination specifying unit is indefinite. 7. The moving picture coding apparatus according to claim 1, wherein the moving picture coding apparatus is designated to refer to an adjacent block set as.
  9.  動画像の各ピクチャを分割したブロック単位で動きベクトルを用いて前記動画像を符号化する動画像符号化方法であって、
     符号化対象ブロックに隣接する複数の符号化済みの隣接ブロックの中から、前記符号化対象ブロックと前記隣接ブロックとの隣接条件に基づいて判定されるいずれか一つの隣接ブロックを、動き情報を参照するための参照先ブロックとして指定する参照先指定ステップと、
     画像内予測、符号化対象ブロックの動きベクトルを利用した画像間予測、および隣接ブロックの動きベクトルを利用した画像間予測の中から、予測方法として、隣接ブロックの動きベクトルを利用した画像間予測が選択された場合、選択された予測方法により参照される隣接ブロックと前記参照先指定ステップにより指定された前記参照先ブロックとが一致するか否かを表す参照先有効情報と選択された予測方法を示す情報とを動き情報に代わって生成する予測選択ステップとを含むことを特徴とする動画像符号化方法。
    A moving image encoding method for encoding the moving image using a motion vector in units of blocks obtained by dividing each picture of the moving image,
    Refer to the motion information for any one of the neighboring blocks determined based on the neighboring condition between the coding target block and the neighboring block from among the plurality of coded neighboring blocks adjacent to the coding target block. A reference destination specifying step that is designated as a reference destination block for
    Among intra prediction, inter prediction using motion vector of the encoding target block, and inter prediction using motion vector of adjacent block, inter prediction using motion vector of adjacent block is used as a prediction method. If selected, the reference destination valid information indicating whether or not the adjacent block referred to by the selected prediction method matches the reference destination block specified by the reference destination specifying step, and the selected prediction method And a prediction selection step of generating information indicating the motion information instead of the motion information.
  10.  動画像の各ピクチャを分割したブロック単位で動きベクトルを用いて前記動画像を符号化する動画像符号化プログラムであって、
     符号化対象ブロックに隣接する複数の符号化済みの隣接ブロックの中から、前記符号化対象ブロックと前記隣接ブロックとの隣接条件に基づいて判定されるいずれか一つの隣接ブロックを、動き情報を参照するための参照先ブロックとして指定する参照先指定ステップと、
     画像内予測、符号化対象ブロックの動きベクトルを利用した画像間予測、および隣接ブロックの動きベクトルを利用した画像間予測の中から、予測方法として、隣接ブロックの動きベクトルを利用した画像間予測が選択された場合、選択された予測方法により参照される隣接ブロックと前記参照先指定ステップにより指定された前記参照先ブロックとが一致するか否かを表す参照先有効情報と選択された予測方法を示す情報とを動き情報に代わって生成する予測選択ステップとをコンピュータに実行させることを特徴とする動画像符号化プログラム。
    A moving image encoding program for encoding the moving image using a motion vector in units of blocks obtained by dividing each picture of the moving image,
    Refer to the motion information for any one of the neighboring blocks determined based on the neighboring condition between the coding target block and the neighboring block from among the plurality of coded neighboring blocks adjacent to the coding target block. A reference destination specifying step that is designated as a reference destination block for
    Among intra prediction, inter prediction using motion vector of the encoding target block, and inter prediction using motion vector of adjacent block, inter prediction using motion vector of adjacent block is used as a prediction method. If selected, the reference destination valid information indicating whether or not the adjacent block referred to by the selected prediction method matches the reference destination block specified by the reference destination specifying step, and the selected prediction method A motion picture encoding program that causes a computer to execute a prediction selection step of generating information to be displayed instead of motion information.
  11.  動画像の各ピクチャを分割したブロック単位で動きベクトルを用いて前記動画像が符号化されたビットストリームを復号する動画像復号装置であって、
     予測方法を示す情報と選択された予測方法により参照される隣接ブロックの有効性を示す参照先有効情報とが含まれる前記ビットストリームから前記予測方法を示す情報と前記参照先有効情報とを復号する可変長復号部と、
     画像内予測、復号対象ブロックの動きベクトルを利用した画像間予測、および前記復号対象ブロックに隣接する複数の復号済みの隣接ブロックのいずれか一つの動きベクトルを利用した画像間予測の中から、前記予測方法を示す情報に基づいていずれかの予測方法を選択する予測選択部と、
     前記予測選択部により、予測方法として、隣接ブロックの動きベクトルを利用した画像間予測が選択された場合、前記復号対象ブロックと前記隣接ブロックとの隣接条件に基づいて判定されるいずれか一つの隣接ブロックを、動き情報を参照するための参照先ブロックとして指定できるか否かを前記参照先有効情報に基づいて判定する動き情報選択部とを含むことを特徴とする動画像復号装置。
    A moving picture decoding apparatus for decoding a bitstream in which the moving picture is encoded using a motion vector in units of blocks obtained by dividing each picture of the moving picture,
    The information indicating the prediction method and the reference destination valid information are decoded from the bitstream including the information indicating the prediction method and the reference destination valid information indicating the validity of the adjacent block referred to by the selected prediction method. A variable length decoding unit;
    From among intra-picture prediction, inter-picture prediction using a motion vector of a decoding target block, and inter-picture prediction using any one of a plurality of decoded adjacent blocks adjacent to the decoding target block, A prediction selector that selects one of the prediction methods based on the information indicating the prediction method;
    When inter prediction using motion vectors of adjacent blocks is selected as a prediction method by the prediction selection unit, any one of the adjacent determinations determined based on an adjacent condition between the decoding target block and the adjacent block A motion image decoding apparatus comprising: a motion information selection unit that determines whether a block can be designated as a reference block for referring to motion information based on the reference destination valid information.
  12.  前記動き情報選択部は、前記隣接条件として、前記隣接ブロックの隣接する辺の長さが前記復号対象ブロックの高さまたは幅と同じであるか否かを判定することにより、隣接する辺の長さが前記復号対象ブロックの高さまたは幅と同じである隣接ブロックを選択し、選択された隣接ブロックを参照先ブロックとして指定できるか否かを判定することを特徴とする請求項11に記載の動画像復号装置。 The motion information selection unit, as the adjacency condition, determines whether the length of the adjacent side of the adjacent block is the same as the height or width of the decoding target block, thereby determining the length of the adjacent side. The adjacent block whose length is the same as the height or width of the decoding target block is selected, and it is determined whether or not the selected adjacent block can be designated as a reference block. Video decoding device.
  13.  前記動き情報選択部は、前記隣接条件として、前記復号対象ブロックと前記隣接ブロックの互いに隣接する辺の共通部分の長さの大小関係を判定することにより、隣接する辺の共通部分の長さが大きい方の隣接ブロックを選択し、選択された隣接ブロックを参照先ブロックとして指定できるか否かを判定することを特徴とする請求項11に記載の動画像復号装置。 The motion information selection unit determines, as the adjacency condition, the size relationship between the lengths of the common parts of the adjacent sides of the decoding target block and the adjacent block, thereby determining the length of the common part of the adjacent sides. 12. The moving picture decoding apparatus according to claim 11, wherein a larger adjacent block is selected, and it is determined whether or not the selected adjacent block can be designated as a reference block.
  14.  前記動き情報選択部は、前記隣接ブロックの残差信号にもとづいて、前記復号対象ブロックに接している前記隣接ブロックの領域の残差信号の絶対値で表される境界残差を求め、境界残差の大小関係を判定することにより、境界残差が小さい方の隣接ブロックを選択し、選択された隣接ブロックを参照先ブロックとして指定できるか否かを判定することを特徴とする請求項11に記載の動画像復号装置。 The motion information selection unit obtains a boundary residual represented by an absolute value of a residual signal in a region of the adjacent block that is in contact with the decoding target block based on the residual signal of the adjacent block, and obtains a boundary residual. 12. The method according to claim 11, further comprising: selecting an adjacent block having a smaller boundary residual by determining a magnitude relationship of the differences, and determining whether the selected adjacent block can be designated as a reference block. The moving picture decoding apparatus described.
  15.  前記動き情報選択部は、前記隣接ブロックの残差信号にもとづいて、前記復号対象ブロックに接している前記隣接ブロックの領域の残差信号の絶対値で表される境界残差と、前記復号対象ブロックに接している前記隣接ブロックの残差信号の絶対値で表される総残差から、総残差に対する境界残差の残差比を求め、残差比の大小関係を判定することにより、残差比が小さい方の隣接ブロックを選択し、選択された隣接ブロックを参照先ブロックとして指定できるか否かを判定することを特徴とする請求項11に記載の動画像復号装置。 The motion information selection unit, based on the residual signal of the adjacent block, a boundary residual represented by an absolute value of a residual signal in the area of the adjacent block in contact with the decoding target block, and the decoding target By calculating the residual ratio of the boundary residual with respect to the total residual from the total residual represented by the absolute value of the residual signal of the adjacent block in contact with the block, and determining the magnitude relationship of the residual ratio, The moving picture decoding apparatus according to claim 11, wherein an adjacent block with a smaller residual ratio is selected, and it is determined whether or not the selected adjacent block can be designated as a reference block.
  16.  前記動き情報選択部は、前記復号対象ブロックのいずれかの辺に動き情報が同一である複数の前記隣接ブロックが連続して並んでいる場合、前記隣接する辺の長さは、これらの連続する前記隣接ブロックの辺の長さを加算したものとして、前記隣接条件を判定することを特徴とする請求項12または13に記載の動画像復号装置。 When the plurality of adjacent blocks having the same motion information are continuously arranged on any side of the decoding target block, the motion information selection unit determines that the lengths of the adjacent sides are continuous. The moving image decoding apparatus according to claim 12 or 13, wherein the adjacency condition is determined by adding the lengths of the sides of the adjacent blocks.
  17.  前記動き情報選択部は、予測方法として、隣接ブロックの動きベクトルを利用した画像間予測が選択された場合、前記復号対象ブロックと前記隣接ブロックとの隣接条件に基づいて判定される前記参照先ブロックが不定であるときは、前記参照先有効情報の代わりに、選択された予測方法により参照される隣接ブロックを指定する参照先情報に基づいて参照先ブロックを指定することを特徴とする請求項11から16のいずれかに記載の動画像復号装置。 When the inter-picture prediction using a motion vector of an adjacent block is selected as a prediction method, the motion information selection unit is determined based on an adjacent condition between the decoding target block and the adjacent block The reference block is specified based on reference destination information that specifies an adjacent block to be referred to by the selected prediction method, instead of the reference destination valid information. To 16. The moving picture decoding apparatus according to any of claims 16 to 16.
  18.  前記動き情報選択部は、予測方法として、隣接ブロックの動きベクトルを利用した画像間予測が選択された場合、前記復号対象ブロックと前記隣接ブロックとの隣接条件に基づいて判定される前記参照先ブロックが不定であるときは、予め既定値として設定した隣接ブロックを指定する参照先情報に基づいて参照先ブロックを指定することを特徴とする請求項11から16のいずれかに記載の動画像復号装置。 When the inter-picture prediction using a motion vector of an adjacent block is selected as a prediction method, the motion information selection unit is determined based on an adjacent condition between the decoding target block and the adjacent block The moving picture decoding apparatus according to any one of claims 11 to 16, wherein the reference destination block is designated based on reference destination information for designating an adjacent block set in advance as a default value when is undefined. .
  19.  動画像の各ピクチャを分割したブロック単位で動きベクトルを用いて前記動画像が符号化されたビットストリームを復号する動画像復号方法であって、
     予測方法を示す情報と選択された予測方法により参照される隣接ブロックの有効性を示す参照先有効情報とが含まれる前記ビットストリームから前記予測方法を示す情報と前記参照先有効情報とを復号する可変長復号ステップと、
     画像内予測、復号対象ブロックの動きベクトルを利用した画像間予測、および前記復号対象ブロックに隣接する複数の復号済みの隣接ブロックのいずれか一つの動きベクトルを利用した画像間予測の中から、前記予測方法を示す情報に基づいていずれかの予測方法を選択する予測選択ステップと、
     前記予測選択ステップにより、予測方法として、隣接ブロックの動きベクトルを利用した画像間予測が選択された場合、前記復号対象ブロックと前記隣接ブロックとの隣接条件に基づいて判定されるいずれか一つの隣接ブロックを、動き情報を参照するための参照先ブロックとして指定できるか否かを前記参照先有効情報に基づいて判定する動き情報選択ステップとを含むことを特徴とする動画像復号方法。
    A moving picture decoding method for decoding a bitstream in which the moving picture is encoded using a motion vector in units of blocks obtained by dividing each picture of the moving picture,
    The information indicating the prediction method and the reference destination valid information are decoded from the bitstream including the information indicating the prediction method and the reference destination valid information indicating the validity of the adjacent block referred to by the selected prediction method. A variable length decoding step;
    From among intra-picture prediction, inter-picture prediction using a motion vector of a decoding target block, and inter-picture prediction using any one of a plurality of decoded adjacent blocks adjacent to the decoding target block, A prediction selection step of selecting one of the prediction methods based on the information indicating the prediction method;
    When inter prediction using motion vectors of adjacent blocks is selected as a prediction method by the prediction selection step, any one of the adjacent determinations determined based on an adjacent condition between the decoding target block and the adjacent block A motion information decoding method comprising: a motion information selection step for determining whether a block can be designated as a reference block for referring to motion information based on the reference destination valid information.
  20.  動画像の各ピクチャを分割したブロック単位で動きベクトルを用いて前記動画像が符号化されたビットストリームを復号する動画像復号プログラムであって、
     予測方法を示す情報と選択された予測方法により参照される隣接ブロックの有効性を示す参照先有効情報とが含まれる前記ビットストリームから前記予測方法を示す情報と前記参照先有効情報とを復号する可変長復号ステップと、
     画像内予測、復号対象ブロックの動きベクトルを利用した画像間予測、および前記復号対象ブロックに隣接する複数の復号済みの隣接ブロックのいずれか一つの動きベクトルを利用した画像間予測の中から、前記予測方法を示す情報に基づいていずれかの予測方法を選択する予測選択ステップと、
     前記予測選択ステップにより、予測方法として、隣接ブロックの動きベクトルを利用した画像間予測が選択された場合、前記復号対象ブロックと前記隣接ブロックとの隣接条件に基づいて判定されるいずれか一つの隣接ブロックを、動き情報を参照するための参照先ブロックとして指定できるか否かを前記参照先有効情報に基づいて判定する動き情報選択ステップとをコンピュータに実行させることを特徴とする動画像復号プログラム。
    A moving picture decoding program for decoding a bitstream in which the moving picture is encoded using a motion vector in units of blocks obtained by dividing each picture of the moving picture,
    The information indicating the prediction method and the reference destination valid information are decoded from the bitstream including the information indicating the prediction method and the reference destination valid information indicating the validity of the adjacent block referred to by the selected prediction method. A variable length decoding step;
    From among intra-picture prediction, inter-picture prediction using a motion vector of a decoding target block, and inter-picture prediction using any one of a plurality of decoded adjacent blocks adjacent to the decoding target block, A prediction selection step of selecting one of the prediction methods based on the information indicating the prediction method;
    When inter prediction using motion vectors of adjacent blocks is selected as a prediction method by the prediction selection step, any one of the adjacent determinations determined based on an adjacent condition between the decoding target block and the adjacent block A moving picture decoding program that causes a computer to execute a motion information selection step of determining whether a block can be designated as a reference block for referring to motion information based on the reference destination valid information.
PCT/JP2012/002649 2011-04-28 2012-04-17 Moving picture encoding device, moving picture encoding method, moving picture encoding program, moving picture decoding device, moving picture decoding method, and moving picture decoding program WO2012147306A1 (en)

Applications Claiming Priority (4)

Application Number Priority Date Filing Date Title
JP2011101951A JP5703942B2 (en) 2011-04-28 2011-04-28 Moving picture coding apparatus, moving picture coding method, and moving picture coding program
JP2011-101951 2011-04-28
JP2011-101952 2011-04-28
JP2011101952A JP5703943B2 (en) 2011-04-28 2011-04-28 Moving picture decoding apparatus, moving picture decoding method, and moving picture decoding program

Publications (1)

Publication Number Publication Date
WO2012147306A1 true WO2012147306A1 (en) 2012-11-01

Family

ID=47071839

Family Applications (1)

Application Number Title Priority Date Filing Date
PCT/JP2012/002649 WO2012147306A1 (en) 2011-04-28 2012-04-17 Moving picture encoding device, moving picture encoding method, moving picture encoding program, moving picture decoding device, moving picture decoding method, and moving picture decoding program

Country Status (1)

Country Link
WO (1) WO2012147306A1 (en)

Cited By (1)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN107431806A (en) * 2015-03-19 2017-12-01 Lg 电子株式会社 For handling the method and its equipment of vision signal

Citations (2)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
JP2010524397A (en) * 2007-04-12 2010-07-15 トムソン ライセンシング Method and apparatus for video encoding and merging depending on the context of skip / direct mode for decoding
JP2010529811A (en) * 2007-06-12 2010-08-26 トムソン ライセンシング Method and apparatus for supporting multi-pass video syntax structure for slice data

Patent Citations (2)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
JP2010524397A (en) * 2007-04-12 2010-07-15 トムソン ライセンシング Method and apparatus for video encoding and merging depending on the context of skip / direct mode for decoding
JP2010529811A (en) * 2007-06-12 2010-08-26 トムソン ライセンシング Method and apparatus for supporting multi-pass video syntax structure for slice data

Non-Patent Citations (2)

* Cited by examiner, † Cited by third party
Title
MARTIN WINKEN ET AL.: "Description of video coding technology proposal by Fraunhofer HHI", JOINT COLLABORATIVE TEAM ON VIDEO CODING (JCT-VC) OF ITU-T SG16 WP3 AND ISO/IEC JTC1/ SC29/WG11 1ST MEETING, 15 April 2010 (2010-04-15), DRESDEN, DE, pages 9 - 10 *
SHAO LI ET AL.: "Simplification and improvement of merge mode coding", JOINT COLLABORATIVE TEAM ON VIDEO CODING (JCT-VC) OF ITU-T SG16 WP3 AND ISO/IEC JTC1/SC29/WG11 5TH MEETING, 16 March 2011 (2011-03-16), GENEVA, CH *

Cited By (2)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN107431806A (en) * 2015-03-19 2017-12-01 Lg 电子株式会社 For handling the method and its equipment of vision signal
CN107431806B (en) * 2015-03-19 2020-07-31 Lg 电子株式会社 Method for processing video signal and apparatus therefor

Similar Documents

Publication Publication Date Title
KR101949818B1 (en) Method for video decoding and computer readable redording meduim thereof
US9053544B2 (en) Methods and apparatuses for encoding/decoding high resolution images
KR101336475B1 (en) Dynamic image encoding device and dynamic image decoding device
KR101641400B1 (en) Decoding apparatus and method
WO2010001917A1 (en) Image processing device and method
EP2863637B1 (en) Video decoding device, video decoding method and video decoding program
JP2010135864A (en) Image encoding method, device, image decoding method, and device
KR20110073263A (en) Method and apparatus for encoding and decoding for intra prediction
JP2012142702A (en) Image processing device, method, and program
WO2012042654A1 (en) Image decoding method, image encoding method, image decoding device, image encoding device, image decoding program, and image encoding program
JP2023118926A (en) Decryption method, decryption device, decryption-side apparatus, electronic apparatus, and non-volatile storage medium
JP6825506B2 (en) Moving image coding device, moving image coding method, moving image coding computer program, moving image decoding device and moving image decoding method, and moving image decoding computer program
JP2023138826A (en) Intra prediction device, image coding device, image decoding device, and program
KR101603887B1 (en) Method for coding data by fast coding unit decision in high efficiency video coding
JP5703943B2 (en) Moving picture decoding apparatus, moving picture decoding method, and moving picture decoding program
WO2012147306A1 (en) Moving picture encoding device, moving picture encoding method, moving picture encoding program, moving picture decoding device, moving picture decoding method, and moving picture decoding program
JP5703942B2 (en) Moving picture coding apparatus, moving picture coding method, and moving picture coding program
JP2009049969A (en) Device and method of coding moving image and device and method of decoding moving image
KR20140124920A (en) Method for encoding and decoding video using motion vector prediction, and apparatus thereof
JP2013168932A (en) Image decoding device, image decoding method, and image decoding program
US20240146959A1 (en) Image encoding method using a skip mode, and a device using the method
JP2013168931A (en) Image encoding device, image encoding method, and image encoding program
JP2010034946A (en) Moving image encoding apparatus and decoding apparatus
WO2013108634A1 (en) Image encoding device, image encoding method, image encoding program, image decoding device, image decoding method, and image decoding program
KR20140124437A (en) Method for encoding and decoding motion information and an appratus using it

Legal Events

Date Code Title Description
121 Ep: the epo has been informed by wipo that ep was designated in this application

Ref document number: 12776108

Country of ref document: EP

Kind code of ref document: A1

NENP Non-entry into the national phase

Ref country code: DE

122 Ep: pct application non-entry in european phase

Ref document number: 12776108

Country of ref document: EP

Kind code of ref document: A1