WO2013157251A1 - Video encoding device, video encoding method, video encoding program, transmission device, transmission method, transmission program, video decoding device, video decoding method, video decoding program, reception device, reception method, and reception program - Google Patents

Video encoding device, video encoding method, video encoding program, transmission device, transmission method, transmission program, video decoding device, video decoding method, video decoding program, reception device, reception method, and reception program Download PDF

Info

Publication number
WO2013157251A1
WO2013157251A1 PCT/JP2013/002565 JP2013002565W WO2013157251A1 WO 2013157251 A1 WO2013157251 A1 WO 2013157251A1 JP 2013002565 W JP2013002565 W JP 2013002565W WO 2013157251 A1 WO2013157251 A1 WO 2013157251A1
Authority
WO
WIPO (PCT)
Prior art keywords
prediction
motion information
block
motion
information
Prior art date
Application number
PCT/JP2013/002565
Other languages
French (fr)
Japanese (ja)
Inventor
上田 基晴
福島 茂
英樹 竹原
Original Assignee
株式会社Jvcケンウッド
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Application filed by 株式会社Jvcケンウッド filed Critical 株式会社Jvcケンウッド
Priority claimed from JP2013085474A external-priority patent/JP5987768B2/en
Priority claimed from JP2013085473A external-priority patent/JP5987767B2/en
Publication of WO2013157251A1 publication Critical patent/WO2013157251A1/en

Links

Images

Classifications

    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04NPICTORIAL COMMUNICATION, e.g. TELEVISION
    • H04N19/00Methods or arrangements for coding, decoding, compressing or decompressing digital video signals
    • H04N19/50Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using predictive coding
    • H04N19/503Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using predictive coding involving temporal prediction
    • H04N19/51Motion estimation or motion compensation
    • H04N19/577Motion compensation with bidirectional frame interpolation, i.e. using B-pictures
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04NPICTORIAL COMMUNICATION, e.g. TELEVISION
    • H04N19/00Methods or arrangements for coding, decoding, compressing or decompressing digital video signals
    • H04N19/10Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using adaptive coding
    • H04N19/102Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using adaptive coding characterised by the element, parameter or selection affected or controlled by the adaptive coding
    • H04N19/103Selection of coding mode or of prediction mode
    • H04N19/105Selection of the reference unit for prediction within a chosen coding or prediction mode, e.g. adaptive choice of position and number of pixels used for prediction
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04NPICTORIAL COMMUNICATION, e.g. TELEVISION
    • H04N19/00Methods or arrangements for coding, decoding, compressing or decompressing digital video signals
    • H04N19/10Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using adaptive coding
    • H04N19/102Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using adaptive coding characterised by the element, parameter or selection affected or controlled by the adaptive coding
    • H04N19/103Selection of coding mode or of prediction mode
    • H04N19/109Selection of coding mode or of prediction mode among a plurality of temporal predictive coding modes
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04NPICTORIAL COMMUNICATION, e.g. TELEVISION
    • H04N19/00Methods or arrangements for coding, decoding, compressing or decompressing digital video signals
    • H04N19/10Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using adaptive coding
    • H04N19/134Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using adaptive coding characterised by the element, parameter or criterion affecting or controlling the adaptive coding
    • H04N19/157Assigned coding mode, i.e. the coding mode being predefined or preselected to be further used for selection of another element or parameter
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04NPICTORIAL COMMUNICATION, e.g. TELEVISION
    • H04N19/00Methods or arrangements for coding, decoding, compressing or decompressing digital video signals
    • H04N19/10Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using adaptive coding
    • H04N19/169Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using adaptive coding characterised by the coding unit, i.e. the structural portion or semantic portion of the video signal being the object or the subject of the adaptive coding
    • H04N19/17Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using adaptive coding characterised by the coding unit, i.e. the structural portion or semantic portion of the video signal being the object or the subject of the adaptive coding the unit being an image region, e.g. an object
    • H04N19/176Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using adaptive coding characterised by the coding unit, i.e. the structural portion or semantic portion of the video signal being the object or the subject of the adaptive coding the unit being an image region, e.g. an object the region being a block, e.g. a macroblock

Definitions

  • the present invention relates to a video signal encoding and decoding technique, and more particularly to a video encoding and decoding technique used for motion compensation prediction.
  • moving picture coding represented by H.264 (hereinafter referred to as AVC) and the like
  • AVC moving picture coding represented by H.264
  • AVC moving picture coding
  • a picture to be coded which is a picture signal to be coded is already coded and decoded.
  • the detected local decoded signal is used as a reference picture, and a motion amount (hereinafter referred to as a motion vector) between the target picture and the reference picture is detected and predicted in a predetermined encoding processing unit (hereinafter referred to as an encoding target block).
  • Motion compensated prediction that generates a signal is used.
  • AVC single prediction for generating a prediction signal in a single direction using one motion vector from one reference picture in motion compensation prediction, and a prediction signal using two motion vectors from two reference pictures Bi-prediction is used to generate A method of changing the size (hereinafter referred to as prediction block size) of a block (hereinafter referred to as prediction target block) that is a prediction processing target within a 16 ⁇ 16 pixel two-dimensional block that is an encoding target block.
  • prediction target block a block that is a prediction processing target within a 16 ⁇ 16 pixel two-dimensional block that is an encoding target block.
  • it is applied to a method for selecting a reference picture used for prediction from a plurality of reference pictures, and the accuracy of a motion vector is expressed with 1/4 pixel accuracy, thereby improving the accuracy of a prediction signal and transmitting
  • the information amount of the difference hereinafter, prediction error
  • information specifying the prediction mode information and the reference image is selected and transmitted together with the motion vector information.
  • the information specifying the transmitted prediction mode information and the reference image and the decoded motion vector information are transmitted. In accordance with the motion compensation prediction process.
  • a motion vector of an encoded block adjacent to the processing target block is set as a prediction motion vector (hereinafter referred to as a prediction vector), and a difference between the motion vector of the processing target block and the prediction vector is obtained. Is transmitted as an encoded vector to improve the compression efficiency.
  • AVC uses a motion vector used for coding a block of a reference picture at the same position as a prediction target block, It is possible to use direct motion compensated prediction that realizes motion compensated prediction without transmitting.
  • Another solution is to reduce the number of motion vectors to be encoded by prohibiting bi-prediction and using only uni-prediction when the prediction block size is small in the encoding device as in Patent Document 1.
  • a technique for preventing an increase in the code amount of a motion vector is known.
  • the direct motion compensated prediction described above pays attention to the continuity of motion in the temporal direction in the block of the reference picture located at the same position as the prediction target block, and uses the motion information of other blocks as they are.
  • the motion compensation prediction process is performed without encoding the difference vector as an encoded vector.
  • an interpolation filter using a plurality of adjacent pixels is used to generate a 1/4 pixel accuracy specified by the motion vector.
  • a reference picture of an area corresponding to the number of pixels corresponding to the number of taps of the interpolation filter horizontally and vertically with respect to the prediction block size It is necessary to acquire an image signal.
  • the prediction block size is reduced, there is a problem that the memory access amount of the reference picture increases, and the same problem remains when direct motion compensation prediction is used.
  • the memory access amount of the reference picture in the encoding device can be reduced together with the number of motion vectors, but in the decoding device, it is encoded. Since the restriction on the number of motion vectors to be recognized cannot be recognized, a decoding processing capability assuming a case where bi-prediction is performed is necessary to realize a real-time decoding process. In addition, when a prediction method that does not transmit an encoded vector, such as direct motion compensation prediction, is used under conditions where implicit bi-prediction is used, bi-prediction prediction signal generation is required, which is required by the decoding device. The maximum memory access amount cannot be reduced, and the problem is not solved.
  • the present invention has been made in view of such circumstances, and an object of the present invention is to provide a technique for improving the coding efficiency while limiting the memory access amount of a reference picture to a predetermined amount or less when using motion compensated prediction. There is to do.
  • a moving picture encoding apparatus specifies a prediction block from a block in which a picture is divided into a plurality of blocks step by step, and the specified prediction block unit includes: A moving image encoding device for generating an encoded stream, wherein motion information is derived from at least one of a block spatially close to a prediction block to be encoded and a block close in time, and the encoding A candidate list construction unit (1506) for registering predetermined motion information from the derived motion information as a motion information candidate of a target prediction block and constructing a motion information candidate list, and the encoding target An encoding unit (118) for encoding index information for designating motion information candidates in the motion information candidate list used for the prediction block; A motion information conversion unit (1507) that converts a motion information candidate, and based on the motion information candidate, performs motion compensation prediction by either uni-prediction or bi-prediction, and generates a prediction signal of the prediction block to
  • the motion information conversion unit (1507) performs prediction conversion for converting the prediction type information indicating the bi-prediction among the motion information candidates into the prediction type information indicating the single prediction, and the motion compensation prediction unit (112) ) Is a case where the block size of the prediction block to be encoded is a predetermined first size, and when the prediction type information indicates the bi-prediction, based on the motion information converted by the prediction conversion The motion compensation prediction is performed.
  • This apparatus is a moving picture encoding apparatus that encodes the moving picture using motion compensation prediction in units of blocks obtained by dividing each picture of the moving picture, and is to be encoded by motion compensation using the derived motion information.
  • a motion compensation prediction unit (112) that generates a prediction signal of a prediction block, and a first control parameter (inter_4x4_enable) that specifies whether or not motion compensation prediction is permitted in the prediction block size of the specified first size
  • a coding block control parameter generation unit that generates a second control parameter (inter_bipred_restriction_idc) that specifies the second size and prohibits bi-prediction motion compensation in a prediction block size that is equal to or smaller than the specified second size.
  • (122) and an encoding unit (118) that encodes information used for motion compensation prediction, including the first and second control parameters.
  • Still another aspect of the present invention is a video encoding method.
  • This method is a moving picture coding method in which a prediction block is identified from a block in which a picture is divided into a plurality of blocks and a coded stream is generated in units of the identified prediction block.
  • Motion information is derived from at least one of a spatially close block and a temporally close block to the target prediction block, and the derived motion information is used as a motion information candidate for the prediction block to be encoded.
  • An encoding step for encoding the motion information, a motion information conversion step for converting the motion information candidate, and the motion information candidate Zui by, and a motion compensated prediction step of generating a prediction signal of the prediction block to be the encoding target performs motion compensation prediction by either single prediction or bi-prediction.
  • the motion information conversion step performs prediction conversion for converting the prediction type information indicating the bi-prediction among the motion information candidates into prediction type information indicating the uni-prediction, and the motion compensation prediction step includes the encoding
  • the motion compensation prediction is performed based on the motion information converted by the prediction conversion. .
  • Still another aspect of the present invention is a transmission device.
  • This apparatus identifies a prediction block from a block obtained by dividing a picture into a plurality of blocks in stages, and is encoded by the moving picture coding method for generating an encoded stream in the identified prediction block unit.
  • a packet processing unit that packetizes the encoded stream to obtain encoded data, and a transmission unit that transmits the packetized encoded data.
  • the video encoding method derives motion information from at least one of a block spatially adjacent to a prediction block to be encoded and a block adjacent to temporally, and the motion of the prediction block to be encoded
  • a motion compensation prediction step of performing compensation prediction and generating a prediction signal of the prediction block to be encoded.
  • the motion information conversion step performs prediction conversion for converting the prediction type information indicating the bi-prediction among the motion information candidates into prediction type information indicating the uni-prediction, and the motion compensation prediction step includes the encoding
  • the motion compensation prediction is performed based on the motion information converted by the prediction conversion. .
  • Still another aspect of the present invention is a transmission method.
  • a prediction block is identified from a block in which a picture is divided into a plurality of blocks in stages, and the coded image is encoded by the moving image coding method for generating an encoded stream in the identified prediction block unit.
  • the video encoding method derives motion information from at least one of a block spatially adjacent to a prediction block to be encoded and a block adjacent to temporally, and the motion of the prediction block to be encoded
  • a motion compensation prediction step of performing compensation prediction and generating a prediction signal of the prediction block to be encoded.
  • the motion information conversion step performs prediction conversion for converting the prediction type information indicating the bi-prediction among the motion information candidates into prediction type information indicating the uni-prediction, and the motion compensation prediction step includes the encoding
  • the motion compensation prediction is performed based on the motion information converted by the prediction conversion. .
  • a video decoding device specifies a prediction block from a block in which a picture is divided into a plurality of blocks in stages, and decodes an encoded stream in units of the specified prediction block
  • a decoding unit (1108) that decodes index information specifying motion information of the prediction block to be decoded from the encoded stream, and a block spatially adjacent to the prediction block to be decoded
  • Motion information is derived from at least one of temporally adjacent blocks, and predetermined motion information is registered from among the derived motion information as motion information candidates of the prediction block to be decoded.
  • the motion information conversion unit (3605) performs prediction conversion for converting the prediction type information indicating the bi-prediction among the motion information candidates into the prediction type information indicating the single prediction, and the motion compensation prediction unit (1114).
  • the motion compensation prediction is performed based on the motion information.
  • the apparatus is a moving picture decoding apparatus that decodes a coded stream obtained by coding the moving picture using motion compensated prediction in units of blocks obtained by dividing each picture of the moving picture, and the motion compensated prediction is performed from the coded stream. And a first control parameter for specifying whether or not motion compensated prediction is permitted in the prediction block size of the designated first size from the information used for the motion compensated prediction that has been decoded.
  • inter_4x4_enable and a second control parameter (inter_bipred_restriction_idc) that specifies the second size and prohibits bi-prediction motion compensation in a prediction block size equal to or smaller than the specified second size (1108)
  • a motion compensated prediction unit (1114) that generates a prediction signal of a decoding target prediction block using information used for the motion compensated prediction Equipped with a.
  • the motion compensation prediction unit (1114) performs motion compensation prediction based on the first and second control parameters.
  • Still another aspect of the present invention is a moving picture decoding method.
  • This method is a moving picture decoding method in which a prediction block is specified from a block in which a picture is divided into a plurality of blocks in stages, and an encoded stream is decoded in the specified prediction block unit.
  • a decoding step of decoding index information specifying motion information of the prediction block to be decoded from the stream, and at least one of a block spatially adjacent to the prediction block to be decoded and a block adjacent in time A candidate list construction step of deriving motion information from the above, and as a motion information candidate of the prediction block to be decoded, registering predetermined motion information from the derived motion information and constructing a motion information candidate list; A motion information conversion step for converting the motion information candidates; and the index of the motion information candidates.
  • the motion information conversion step performs prediction conversion for converting prediction type information indicating the bi-prediction among the motion information candidates into prediction type information indicating the single prediction, and the motion compensation prediction step includes the decoding target Based on the motion information converted by the prediction conversion, when the block size of the prediction block to be is a predetermined first size and the prediction type information of the specified motion information indicates the bi-prediction The motion compensation prediction is performed.
  • Still another aspect of the present invention is a receiving device.
  • This apparatus specifies a prediction block from a block in which a picture is divided into a plurality of blocks in stages, and receives and decodes an encoded stream in which a moving image is encoded in the specified prediction block unit.
  • a receiving unit that receives encoded data in which the encoded stream is packetized; a recovery unit that processes the received encoded stream to recover the original encoded stream; A decoding unit that decodes index information specifying motion information of the prediction block to be decoded from the encoded stream, a block that is spatially close to the prediction block that is to be decoded, and a block that is temporally close to the prediction block
  • Motion information is derived from at least one of the above, and the motion information candidate of the prediction block to be decoded is derived as the motion information candidate
  • a candidate list construction unit for registering predetermined motion information from among the motion information and constructing a motion information candidate list, a motion information conversion unit for converting the motion information candidates, and the index information of the motion information candidates.
  • a motion-compensated prediction unit that performs motion-compensated prediction based on specified motion information by either uni-prediction or bi-prediction and generates a prediction signal of a prediction block to be decoded.
  • the motion information conversion unit converts prediction type information indicating that the motion compensation prediction is performed by the bi-prediction among the motion information candidates into prediction type information indicating that the motion compensation prediction is performed by the single prediction.
  • the motion compensation prediction unit is a case where the block size of the prediction block to be decoded is a predetermined first size, and the prediction type information of the designated motion information is the bi-prediction Indicates that the motion compensation prediction is performed, the motion compensation prediction is performed based on the prediction type information converted by the prediction conversion.
  • Still another aspect of the present invention is a receiving method.
  • This method specifies a prediction block from a block in which a picture is divided into a plurality of blocks in stages, and receives and decodes an encoded stream in which a moving image is encoded in the specified prediction block unit.
  • a receiving step for receiving encoded data in which the encoded stream is packetized; a restoring step for packetizing the received encoded stream to restore the original encoded stream;
  • a decoding step for decoding the index information specifying the motion information of the prediction block to be decoded from the encoded stream, a block spatially adjacent to the prediction block to be decoded, and a block adjacent in time Motion information is derived from at least one of the prediction block motion information candidates of the prediction block to be decoded.
  • prediction type information indicating that the motion compensation prediction is performed by the bi-prediction among the motion information candidates is converted into prediction type information indicating that the motion compensation prediction is performed by the single prediction.
  • the motion compensation prediction step is performed when the block size of the prediction block to be decoded is a predetermined first size, and the prediction type information of the designated motion information is the bi-prediction. Indicates that the motion compensation prediction is performed, the motion compensation prediction is performed based on the prediction type information converted by the prediction conversion.
  • the present invention it is possible to improve the encoding efficiency while limiting the memory access amount of the reference picture to a predetermined amount or less.
  • FIG. 1 It is a figure which shows the structure of the moving image encoder which concerns on Embodiment 1 of this invention. It is a figure which shows an example of the division
  • FIGS. 8A and 8B are diagrams for explaining two prediction modes for encoding motion information used in motion compensated prediction according to Embodiment 1 of the present invention. It is a figure which shows the rough value of the reference image memory amount required for motion compensation prediction at the time of using a horizontal and vertical 7 tap filter for motion compensation prediction. It is a figure for demonstrating the control parameter which controls the block size and prediction process of motion compensation prediction based on Embodiment 1 of this invention. It is a figure which shows the structure of the moving image decoding apparatus which concerns on Embodiment 1 of this invention.
  • 16 is a flowchart for explaining an operation of motion compensation prediction mode / prediction signal generation, which is steps S701, S702, S703, and S705 of FIG. 7, which operates via the motion compensation prediction block structure selection unit of FIG. 15. It is a flowchart for demonstrating the detailed operation
  • FIGS. 23A and 23B are diagrams illustrating an example of comparison contents of combined motion information candidates. It is a figure which shows the time candidate block group used for a time joint movement information candidate list generation. It is a flowchart for demonstrating the detailed operation
  • FIG. 39 is a flowchart for explaining detailed operations of a predicted motion information decoding process in step S3805 of FIG. 38.
  • FIG. It is an example which shows the restriction
  • Embodiment 1 of this invention it is an example which integrated two control parameters which control the block size of a motion compensation prediction, and a prediction process as one encoding transmission parameter. It is a figure for demonstrating the control parameter which controls the block size and prediction process of motion compensation prediction based on Embodiment 2 of this invention. It is a figure which shows the relationship between the control parameter which controls bi-prediction, and prediction block size based on Embodiment 2 of this invention. It is an example of the syntax regarding the parameter which controls the restriction
  • FIG. 3 It is a figure which shows an example of the definition of a space periphery prediction block in combined motion information candidate generation in Embodiment 3 of this invention. It is a flowchart for demonstrating the detailed operation
  • FIG. 1 is a diagram showing a configuration of a moving picture coding apparatus according to Embodiment 1 of the present invention. Hereinafter, the operation of each unit will be described.
  • the moving picture coding apparatus according to Embodiment 1 includes an input terminal 100, an input picture memory 101, a coding block acquisition unit 102, a subtraction unit 103, an orthogonal transform / quantization unit 104, a prediction error coding unit 105, an inverse quantum.
  • Inverse conversion unit 106 addition unit 107, intra-frame decoded image buffer 108, loop filter unit 109, decoded image memory 110, motion vector detection unit 111, motion compensation prediction unit 112, motion compensation prediction block structure selection unit 113, intra Prediction unit 114, intra prediction block structure selection unit 115, prediction mode selection unit 116, coding block structure selection unit 117, block structure / prediction mode information additional information coding unit 118, prediction mode information memory 119, multiplexing unit 120, An output terminal 121 and a coding block control parameter generation unit 122 are provided.
  • the image signal input from the input terminal 100 is stored in the input image memory 101, and the image signal to be processed for the encoding target picture is input from the input image memory 101 to the encoding block acquisition unit 102.
  • the image signal of the encoding target block extracted based on the position information of the encoding target block by the encoding block acquisition unit 102 is a subtraction unit 103, a motion vector detection unit 111, a motion compensation prediction unit 112, and an intra prediction unit 114. To be supplied.
  • FIG. 2 is a diagram illustrating an example of an encoding target image.
  • the encoding target image is encoded in units of 64 ⁇ 64 pixel encoding blocks, and the prediction block is configured based on the encoding block. .
  • the maximum prediction block size is 64 ⁇ 64 pixels, which is the same as the encoded block, and the minimum prediction block size is 4 ⁇ 4 pixels.
  • the division configuration of a CU into prediction blocks includes non-division (2N ⁇ 2N), horizontal / vertical division (N ⁇ N), horizontal division only (2N ⁇ N), and vertical division only (N ⁇ 2N) is possible.
  • the prediction block further divided horizontally and vertically can be hierarchically divided into prediction blocks as coding blocks (CU), and the hierarchy is expressed by the number of CU divisions.
  • the divided areas viewed from the upper hierarchy CU of the four divided CUs are defined as division 1, division 2, division 3, and division 4.
  • FIG. 3 is a diagram illustrating an example of a detailed definition of the predicted block size.
  • the prediction block size in the case of performing motion compensation prediction that performs prediction using correlation between screens is divided only in the horizontal direction (2N ⁇ N), only in the vertical direction, with respect to the division configuration of the CU into prediction blocks. Can be defined and a total of 13 types of prediction block sizes can be defined. However, the prediction block size in the case of intra prediction in which prediction is performed using correlation in the screen is only in the horizontal direction. Since the division into two (2N ⁇ N) and the division only into the vertical direction (N ⁇ 2N) are not possible, a total of five types of prediction block sizes are defined.
  • the partition configuration of the prediction block according to Embodiment 1 of the present invention is not limited to this combination.
  • the encoding block size that can be defined can be changed by setting the maximum CU size and the minimum CU size using control parameters such as Maximum_cu_size and Minimum_cu_size shown in FIG. 3, and encoding and decoding these control parameters. Is possible.
  • the subtraction unit 103 calculates a prediction error signal by subtracting the image signal supplied from the coding block acquisition unit 102 and the prediction signal supplied from the coding block structure selection unit 117, and outputs a prediction error signal. Is supplied to the orthogonal transform / quantization unit 104.
  • the orthogonal transform / quantization unit 104 performs orthogonal transform and quantization on the prediction error signal supplied from the subtraction unit 103, and the quantized prediction error signal is subjected to a prediction error encoding unit 105 and an inverse quantization / inverse conversion unit. 106.
  • the prediction error encoding unit 105 entropy-encodes the quantized prediction error signal supplied from the orthogonal transform / quantization unit 104, generates a code string for the prediction error signal, and supplies the code sequence to the multiplexing unit 120. .
  • the inverse quantization / inverse transform unit 106 performs a process such as inverse quantization or inverse orthogonal transform on the quantized prediction error signal supplied from the orthogonal transform / quantization unit 104 to generate a decoded prediction error signal. Generated and supplied to the adder 107.
  • the addition unit 107 adds the decoded prediction error signal supplied from the inverse quantization / inverse conversion unit 106 and the prediction signal supplied from the coding block structure selection unit 117 to generate a decoded image signal, and generates a decoded image signal.
  • the signal is supplied to the intra-frame decoded image buffer 108 and the loop filter unit 109.
  • the intra-frame decoded image buffer 108 supplies the decoded image in the same frame in the region adjacent to the encoding target block to the intra prediction unit 114 and also stores the decoded image signal supplied from the addition unit 107.
  • the loop filter unit 109 performs a filtering process on the decoded image signal supplied from the adding unit 107 by applying a filter to remove distortion caused by encoding and to restore the image to a pre-encoded image.
  • the resulting decoded image is supplied to the decoded image memory 110.
  • the decoded image memory 110 stores the decoded image signal subjected to the filtering process supplied from the loop filter unit 109.
  • a decoded image for which decoding of the entire image has been completed is stored as a reference image by a predetermined number of images, and the reference image signal is supplied to the motion vector detection unit 111 and the motion compensation prediction unit 112.
  • the motion vector detection unit 111 receives the input of the image signal of the encoding target block supplied from the encoding block acquisition unit 102 and the reference image signal stored in the decoded image memory 110, and obtains a motion vector for each reference image.
  • the motion vector value is detected and supplied to the motion compensation prediction unit 112 and the motion compensation prediction block structure selection unit 113.
  • a general motion vector detection method calculates an error evaluation value for an image signal corresponding to a reference image moved by a predetermined movement amount from the same position as the image signal, and moves the movement amount that minimizes the error evaluation value. Let it be a vector.
  • the error evaluation value a sum of absolute differences SAD (Sum of Absolute Difference) for each pixel, a sum of squared error values SSE (Sum of Square Error) for each pixel, or the like is used.
  • the code amount related to the coding of the motion vector can also be included in the error evaluation value.
  • the motion compensated prediction unit 112 is configured to decode the decoded image memory according to the information specifying the prediction block structure specified by the motion compensated prediction block structure selecting unit 113, the reference image specifying information, and the motion vector value input from the motion vector detecting unit 111.
  • a prediction signal is generated by acquiring an image signal at a position obtained by moving the reference image indicated by the reference image designation information in 110 from the same position as the image signal of the prediction block by a motion vector value.
  • the prediction mode specified by the motion compensated prediction block structure selection unit 113 is prediction from a single reference image
  • a prediction signal acquired from one reference image is used as a motion compensation prediction signal
  • two prediction modes are referenced.
  • a weighted average of prediction signals acquired from two reference images is used as a motion compensation prediction signal
  • the motion compensation prediction signal is supplied to the prediction mode selection unit 116.
  • the ratio of the weighted average of bi-prediction is set to 1: 1.
  • 4 (a) to 4 (d) are diagrams for explaining the prediction type of motion compensation prediction.
  • a process for performing prediction from a single reference image is defined as single prediction, and in the case of single prediction, prediction using either one of two reference images registered in the reference image management list, that is, L0 prediction or L1 prediction. I do.
  • FIG. 4A shows a case in which the prediction image is uni-prediction and the reference image (RefL0Pic) for L0 prediction is at a time before the encoding target image (CurPic).
  • FIG. 4B shows a case in which the prediction image is a single prediction and the reference image of the L0 prediction is at a time after the encoding target image.
  • the L0 prediction reference image shown in FIGS. 4A and 4B can be replaced with the L1 prediction reference image (RefL1Pic) to perform single prediction.
  • FIG. 4C illustrates a case where bi-prediction is performed, and the reference image for L0 prediction is at a time before the encoding target image and the reference image for L1 prediction is at a time after the encoding target image.
  • FIG. 4D shows a case of bi-prediction, where the reference image for L0 prediction and the reference image for L1 prediction are at a time before the encoding target image.
  • the relationship between the prediction type of L0 / L1 and time can be used without being limited to L0 being the past direction and L1 being the future direction.
  • each of L0 prediction and L1 prediction may be performed using the same reference picture.
  • whether to perform motion compensation prediction by single prediction or bi-prediction is determined based on, for example, information (for example, a flag) indicating whether to use L0 prediction and whether to use L1 prediction.
  • Bi-prediction requires image information access to two reference image memories, and therefore may require twice or more memory bandwidth compared to single prediction.
  • bi-prediction when the prediction block size of motion compensation prediction is small becomes a bottleneck of the memory band, and the bottleneck of the memory band is suppressed in the embodiment of the present invention.
  • the motion compensated prediction block structure selection unit 113 detects the motion vector value detected for each reference image input from the motion vector detection unit 111 and the motion information stored in the prediction mode information memory 119 ( Based on the prediction type, motion vector value, and reference image designation information), the control parameters related to the prediction block size and motion compensation prediction mode defined in Embodiment 1 generated by the coding block control parameter generation unit 122 are The reference image designation information and the motion vector value used for each of the prediction block size and the motion compensation prediction mode that are input and determined based on the control parameter are set in the motion compensation prediction unit 112. Depending on the set value, the motion compensation prediction signal supplied from the motion compensation prediction unit 112 and the image signal of the target block to be encoded supplied from the coding block acquisition unit 102 are used to optimize the prediction block size and motion compensation prediction. Determine the mode.
  • the motion compensated prediction block structure selection unit 113 uses the determined prediction block size, motion compensation prediction mode, prediction type corresponding to the prediction mode, motion vector, and information specifying the reference image designation information as a motion compensation prediction signal and a prediction error. Is supplied to the prediction mode selection unit 116 together with an error evaluation value for.
  • the intra prediction unit 114 is adjacent to the encoding target block supplied from the intra-frame decoded image buffer 108 according to the intra prediction mode defined as the information specifying the prediction block structure specified by the intra prediction block structure selection unit 115.
  • An intra prediction signal is generated using the decoded image in the same frame, and is supplied to the intra prediction block structure selection unit 115.
  • Embodiment 1 Intra prediction block structure selection section 115 is generated by coding block control parameter generation section 122 according to intra prediction mode information stored in prediction mode information memory 119 and a plurality of defined intra prediction modes.
  • the control parameter related to the prediction block size defined in the step is input, and the intra prediction mode used for each of the prediction block sizes determined based on the control parameter is set in the intra prediction unit 114.
  • the optimal prediction block size and intra prediction mode are determined using the intra prediction signal supplied from the intra prediction unit 114 and the image signal of the encoding target block supplied from the coding block acquisition unit 102. To do.
  • the intra prediction block structure selection unit 115 supplies information specifying the determined prediction block size and intra prediction mode to the prediction mode selection unit 116 together with the intra prediction signal and the error evaluation value for the prediction error.
  • the prediction mode selection unit 116 is supplied from the motion compensated prediction block structure selection unit 113 and specifies the determined prediction block size, motion compensation prediction mode, prediction type according to the prediction mode, motion vector, and reference image designation information.
  • CU size units that are hierarchically configured from the error evaluation value for the prediction error, and the error prediction value for the predicted prediction block size, intra prediction mode, and prediction error supplied from the intra prediction block structure selection unit 115 The optimum prediction mode is selected by comparing the error evaluation values.
  • motion compensation prediction is selected as the optimal prediction mode information in units of CU size selected by the prediction mode selection unit 116 together with the sum of the prediction block size, the prediction signal, and the error evaluation value in units of CU size
  • the intra prediction is selected as the motion compensation prediction mode
  • the prediction type corresponding to the prediction mode, the motion vector, the information specifying the reference image designation information, and the motion compensation prediction signal, the intra prediction mode and the intra prediction signal are selected. Is supplied to the coding block structure selection unit 117.
  • the coding block structure selection unit 117 is defined in Embodiment 1 generated by the coding block control parameter generation unit 122 based on the optimal prediction mode information in CU size units supplied from the prediction mode selection unit 116.
  • the control parameter related to the encoded block size is input, the optimum CU_Depth configuration is selected in the encoded block size configuration determined based on the control parameter, the information for specifying the CU partition configuration, and the specified partition configuration.
  • the optimal prediction mode information in the CU size and additional information related to the prediction mode are supplied to the block structure / prediction mode information additional information encoding unit 118 and the selected prediction signal is subtracted. 103 and the adder 107.
  • the block structure / prediction mode information additional information encoding unit 118 is supplied from the encoding block structure selection unit 117, and specifies the CU partition configuration, and the optimal prediction mode information for the specified CU size for each partition configuration.
  • the coding block unit The CU partition configuration and mode information used for prediction are encoded and supplied to the multiplexing unit 120, and the information is stored in the prediction mode information memory 119.
  • the prediction mode information memory 119 predetermines the CU partition configuration of the coding block unit supplied from the block structure / prediction mode information additional information encoding unit 118 and the mode information used for prediction based on the minimum prediction block size unit. Memorize images. Since the first embodiment focuses on motion compensation prediction that is prediction between screens, motion information (prediction type, motion vector, and reference image index) that is information related to motion compensation prediction in mode information is used. And add a description.
  • the motion information of the adjacent block of the prediction block that is the processing target of motion compensation prediction is set as a spatial candidate block group, and the motion information of the block on ColPic and the surrounding blocks that are at the same position as the processing target prediction block is the time candidate block group. To do.
  • ColPic is a decoded image different from the prediction block to be processed, and is stored in the decoded image memory 110 as a reference image.
  • ColPic is a reference image decoded immediately before.
  • ColPic is the reference image decoded immediately before, but the reference image immediately before in display order or the reference image immediately after in display order may be used, and the reference image used for ColPic is included in the encoded stream. Direct specification is also possible.
  • the prediction mode information memory 119 supplies the motion information of the spatial candidate block group and the temporal candidate block group to the motion compensated prediction block structure selection unit 113 as motion information of the candidate block group, and intra prediction of adjacent blocks of the intra prediction block.
  • the mode information is supplied to the intra prediction block structure selection unit 115.
  • the multiplexing unit 120 includes a prediction error encoding sequence supplied from the prediction error encoding unit 105, a CU partitioning configuration in units of encoded blocks supplied from the block structure / prediction mode information additional information encoding unit 118, and prediction.
  • the encoded bit stream is generated by multiplexing the encoded sequence of the mode information and the additional information used in the above, and the encoded bit stream is output to the recording medium / transmission path via the output terminal 121.
  • the coding block control parameter generation unit 122 performs control parameters such as Maximum_cu_size and Minimum_cu_size shown in FIG. 3, which are parameters defining the coding block structure, and the block size and prediction processing of motion compensation prediction in the first embodiment. Generate parameters for defining a coding block structure or a prediction block structure, such as a control parameter to be limited, and a motion compensation prediction block structure selection unit 113, an intra prediction block structure selection unit 115, a coding block structure selection unit 117, And supplied to the block structure / prediction mode information additional information encoding unit 118. Details regarding the block size of motion compensation prediction and control parameters for limiting the prediction process will be described later.
  • the configuration of the moving picture encoding apparatus shown in FIG. 1 can also be realized by hardware such as an information processing apparatus including a CPU (Central Processing Unit), a frame memory, and a hard disk.
  • an information processing apparatus including a CPU (Central Processing Unit), a frame memory, and a hard disk.
  • FIG. 5 is a flowchart showing the flow of the encoding process in the video encoding apparatus according to Embodiment 1 of the present invention.
  • CU_Depth which is a control parameter for CU partitioning
  • a coding process target block image is obtained from the coding block acquisition unit 102 (S501).
  • the motion vector detection unit 111 uses a motion vector value for each reference image according to CU partitioning from a block image to be predicted according to CU partitioning from the encoding target block image and a plurality of reference images stored in the decoded image memory 110. Is calculated (S502).
  • the motion compensation prediction block structure selection unit 113 uses the motion vector supplied from the motion vector detection unit 111, the motion information and the intra prediction mode information stored in the prediction mode information memory 119, and performs the first embodiment.
  • the prediction signal for each of the prediction block size and the motion compensation prediction mode defined in (1) is acquired using the motion compensation prediction unit 112, and the result of selecting the optimal prediction block size and prediction mode in CU units is output.
  • the intra prediction block structure selection unit 115 acquires prediction signals for each of the prediction block size and the intra prediction mode using the intra prediction unit 114, and selects the optimal prediction block size and prediction mode in CU units. Is output.
  • the coding block structure selection unit 117 generates a prediction mode and a prediction signal in the optimum coding block structure using these results (S503). Details of the processing in step S503 will be described later.
  • the subtraction unit 103 calculates a difference between the encoded block image supplied from the encoded block acquisition unit 102 and the prediction signal supplied from the encoded block structure selection unit 117 as a prediction error signal (S504). ).
  • the block structure / prediction mode information additional information encoding unit 118 includes a coding type, a prediction mode, a prediction type according to a prediction mode in the case of motion compensation prediction, a motion vector, and a coding structure supplied from the coding block structure selection unit 117.
  • Information for specifying the reference image designation information and intra prediction mode information in the case of intra prediction are encoded according to a predetermined syntax structure, and encoded data of additional information related to the encoding structure and the prediction mode information is generated (S505). ).
  • the prediction error encoding unit 105 entropy encodes the quantized prediction error signal generated by the orthogonal transform / quantization unit 104 to generate encoded data of the prediction error (S506).
  • the multiplexing unit 120 is supplied from the coding structure supplied from the block structure / prediction mode information additional information encoding unit 118, encoded data of additional information related to the prediction mode information, and the prediction error encoding unit 105.
  • the encoded data of the prediction error is multiplexed to generate an encoded bit stream (S507).
  • the addition unit 107 adds the decoded prediction error signal supplied from the inverse quantization / inverse conversion unit 106 and the prediction signal supplied from the coding block structure selection unit 117 to generate a decoded image signal (S508).
  • the prediction mode information memory 119 includes motion information (prediction when motion compensation prediction is used as additional information related to the coding structure and prediction mode information supplied from the block structure / prediction mode information additional information encoding unit 118. Type, motion vector, and reference image designation information) and intra prediction mode information when intra prediction is used are stored in units of the smallest prediction block size (S509).
  • the decoded image signal generated by the addition unit 107 is stored in the intra-frame decoded image buffer 108, and the loop filter unit 109 performs a loop filter process for distortion removal (S510) and performs the filter.
  • the decoded image signal is supplied to and stored in the decoded image memory 110 and used for motion compensation prediction processing of an encoded image to be encoded thereafter (S511).
  • Max_CU_Depth a value indicating the number of hierarchies between the set maximum CU size and minimum CU size is set as Max_CU_Depth, and it is determined whether or not the CU_Depth of the target CU is smaller than Max_CU_Depth (S600).
  • S600 Max_CU_Depth
  • intra prediction block structure selection unit 115 and intra prediction unit 114 in FIG. 1 calculate intra prediction modes and generate prediction signals (S607). Intra prediction mode information, a prediction signal, and an error evaluation value in the target CU are calculated.
  • the motion compensation prediction block structure selection unit 113 and the motion compensation prediction unit 112 select a motion compensation prediction block size, and generate a motion compensation prediction mode and a prediction signal for each selected prediction block (S608).
  • a prediction block size, mode information, motion information, a prediction signal, and an error evaluation value for motion compensation prediction in the target CU are calculated. Details of step S608 will be described later.
  • the coding block structure selecting unit 117 compares the error evaluation value of the intra prediction in the target CU with the error evaluation value of the motion compensated prediction, selects a prediction method with a small error, and selects intra / inter (motion compensated prediction). ) Is determined (S609).
  • the lowest CU (Depth Max_CU_Depth) CU is sequentially compared with the upper CU, and the optimal CU_Depth for each divided region of the CU Prediction mode can be selected.
  • an encoded block image to be predicted is acquired for the target CU (S700).
  • motion compensation prediction mode / prediction signal generation processing is performed for each intra-CU division mode (S701 to S705).
  • the motion compensation prediction mode / prediction signal generation processing when the intra-CU division mode is 2N ⁇ 2N is performed by setting NumPart which is a value indicating the number of divisions to 1 (S701). Subsequently, NumPart is set to 2, and motion compensation prediction mode / prediction signal generation processing is performed in the case of 2N ⁇ N (S702) and N ⁇ 2N (S703).
  • step S704 when CU_Depth is equal to Max_CU_Depth, the target CU size is 8 ⁇ 8, and an inter_4x4_enable flag (to be described later) is 1 (S704: YES), NumPart is set to 4 and motion compensation is performed when N ⁇ N Prediction mode / prediction signal generation processing is performed (S705). Details of the motion compensation prediction mode / prediction signal generation processing performed in steps S701, S702, S703, and S705 will be described later. If the condition of step S704 is not satisfied (S704: NO), step S705 is skipped and the subsequent step is performed.
  • Embodiment 1 motion compensated prediction / prediction signal generation in intra-CU division in the order of 2N ⁇ 2N (S701), 2N ⁇ N (S702), N ⁇ 2N (S703), and N ⁇ N (S705)
  • the processing order of the steps of each of the CU divisions may be changed, and when processing is performed by a CPU or the like that can perform parallel processing, S701, S702, S703, and S705 are performed. Can also be performed in parallel.
  • an error evaluation value for each intra-CU partition mode for which motion compensation prediction mode / prediction signal generation has been performed is compared, and an optimal prediction block size (PU) that is an optimal intra-CU partition mode is selected (S706).
  • Prediction mode information / error evaluation value / prediction signal for the selected PU is stored (S707), and the process of step S608 in the flowchart of FIG. 6 ends.
  • FIGS. 8A and 8B are diagrams for explaining two prediction modes for encoding motion information used in motion compensated prediction according to Embodiment 1 of the present invention.
  • the prediction target block directly encodes its own motion information using the continuity of motion in the temporal direction and the spatial direction in the prediction target block and the encoded block adjacent to the prediction target block.
  • the motion information of spatially and temporally adjacent blocks is used for encoding, which is called a joint prediction mode (merge mode).
  • the spatially adjacent block refers to a block adjacent to the prediction target block among encoded blocks belonging to the same image as the prediction target block.
  • the temporally adjacent blocks indicate blocks in the same spatial position as the prediction target block and in the vicinity thereof among blocks belonging to an encoded image different from the prediction target block.
  • motion information that can be selectively combined from a plurality of adjacent block candidates can be defined, and the motion information is specified by encoding information (joined motion information index) that specifies the adjacent block to be used.
  • the motion information acquired based on the information is used as it is for motion compensation prediction.
  • a Skip mode is defined in which the prediction signal predicted in the joint prediction mode is a decoded picture without encoding prediction transmission of the prediction difference information, and a decoded image is obtained with information having only the combined motion information. Can be reproduced.
  • the Skip mode can be used when the intra-CU division mode is 2N ⁇ 2N, and the motion information transmitted in the Skip mode is the designation information that defines the adjacent block as in the combined prediction mode.
  • the second prediction mode is a technique for coding all the components of motion information individually and transmitting motion information with little prediction error to the prediction block, and is called a motion detection prediction mode.
  • the motion detection prediction mode includes a prediction type indicating whether the prediction is bi-prediction or uni-prediction, information for identifying a reference image (reference image index), and encoding of motion information in the conventional motion compensation prediction.
  • the information for specifying the motion vector is encoded separately.
  • the prediction mode indicates whether to use single prediction or bi-prediction.
  • single prediction single prediction information for specifying a reference image for one reference image, and a motion vector prediction vector
  • the difference vector is encoded.
  • bi-prediction information for specifying reference images for two reference images and a motion vector are individually encoded.
  • the prediction vector for the motion vector is generated from the motion information of the adjacent block similarly to the AVC.
  • the motion vector used for the prediction vector can be selected from a plurality of adjacent block candidates, and the motion vector is the prediction vector. Is transmitted by encoding two pieces of information (predicted vector index) for designating adjacent blocks to be used for and a difference vector.
  • an integer motion existing in the reference image is generated.
  • a pixel of the reference image at the motion position with a 1/4 pixel accuracy is calculated by an interpolation filter.
  • a 7-tap FIR filter is used as an interpolation filter.
  • FIG. 9 shows a case where a 7-tap filter is applied and a memory band is secured when performing uni-prediction and bi-prediction in each of the predictable block sizes definable for motion compensated prediction shown in FIG. 3 in the first embodiment.
  • various configurations such as a configuration in which memory access is possible in units of horizontal 4 pixels and a configuration in which units of horizontal and vertical 2 ⁇ 2 pixels are possible can be taken.
  • the memory access amount indicates the maximum value of the memory access amount that needs to be obtained at the minimum regardless of the configuration of the reference image memory.
  • the 4 ⁇ 4 pixel size is the most encoded block size (LCU) unit.
  • the memory access amount becomes larger, and access of nearly 6 times the size of 64 ⁇ 64 pixels is required.
  • two prediction signals are acquired from reference images at different positions, so that twice as many memory accesses as in single prediction are required.
  • a motion compensation prediction limiting method and a control parameter definition and setting method for limiting in which the memory access maximum amount of the reference image can be controlled step by step to limit the memory bandwidth, It is possible to achieve both the feasibility and encoding efficiency of a moving image encoding apparatus for fine images.
  • FIG. 10 shows an example of the motion compensation prediction block size and control parameters for limiting the prediction processing, which are generated by the coding block control parameter generation unit 122 of FIG. 1 according to Embodiment 1 of the present invention. To do.
  • inter_4x4_enable is a parameter for controlling the validity / invalidity of motion compensated prediction of 4 ⁇ 4 pixels, which is the smallest motion compensated prediction block size, and only prediction processing for which bi-prediction is performed among motion compensated predictions It consists of two parameters, inter_bipred_restriction_idc, which defines the block size that prohibits
  • 4 ⁇ 4 bi-prediction, 4 ⁇ 8/8 ⁇ 4 bi-prediction, 4 ⁇ 4 mono-prediction, 8 ⁇ 8 bi-prediction, 8 ⁇ 16/16 ⁇ 8 bi-prediction, 4 ⁇ 8/8 ⁇ 4 single prediction, and 16 ⁇ 16 bi-prediction are in this order. Relatively accessed except for the minimum prediction block size of 4 ⁇ 4 pixels. The amount is small.
  • inter_4x4_enable which is a control parameter for prohibiting the motion compensation prediction process itself, is prepared, and inter_bipred_restriction_idc that further restricts bi-prediction is prepared as a control parameter for each block size. Can control memory access amount explicitly.
  • the amount of memory access is larger than that of 16 ⁇ 16 bi-prediction.
  • the intra-CU partitioning mode is Since the entire motion compensated prediction with a prediction block size smaller than an N ⁇ N 8 ⁇ 8 block can be prohibited, the motion compensated prediction process itself is prohibited in a configuration having a restriction on a fixed minimum prediction block size. It is possible to control the memory access amount.
  • the memory access amount is controlled by combining the minimum CU size value in addition to inter_4x4_enable and inter_bipred_restriction_idc.
  • inter_bipred_restriction_idc defines a value from 0 to 5 as shown in FIG. 10, and from a state where there is no restriction on bi-prediction to a state where bi-prediction with a size of 16 ⁇ 16 blocks or less is restricted.
  • the range of definition is an example, and it is also possible to define a control value that is smaller or more than this value as another configuration of the embodiment of the present invention.
  • Control that disables the entire motion compensated prediction of a given size and a control parameter that restricts bi-prediction of motion compensated prediction of less than a given size, and controls the maximum memory access amount to be within the prescribed range
  • FIG. 11 is a diagram showing a configuration of the moving picture decoding apparatus according to Embodiment 1 of the present invention. Hereinafter, the operation of each unit will be described.
  • the video decoding apparatus according to Embodiment 1 includes an input terminal 1100, a demultiplexing unit 1101, a prediction difference information decoding unit 1102, an inverse quantization / inverse transform unit 1103, an addition unit 1104, an intra-frame decoded image buffer 1105, a loop filter.
  • Unit 1106 decoded image memory 1107, prediction mode / block structure decoding unit 1108, prediction mode / block structure selection unit 1109, intra prediction information decoding unit 1110, motion information decoding unit 1111, prediction mode information memory 1112, intra prediction unit 1113, A motion compensation prediction unit 1114 and an output terminal 1115 are provided.
  • the encoded bit stream is supplied from the input terminal 1100 to the demultiplexing unit 1101.
  • the demultiplexing unit 1101 is used for the code string of the supplied coded bitstream, the coded string of prediction error information, the control parameters related to the coding block and the prediction block structure, the CU partition configuration and coding block unit, and the prediction.
  • Mode information prediction mode according to the prediction mode in the case of motion compensated prediction, motion vector that is information specifying the motion vector, and reference image designation information, and intra prediction mode information in the case of intra prediction. Separate into coded sequences to be constructed.
  • the coding sequence of the prediction error information is supplied to the prediction difference information decoding unit 1102, and the control parameter, the CU partition configuration of the coding block unit, and the coding sequence of the mode information used for prediction are predicted mode / block structure It supplies to the decoding part 1108.
  • the prediction difference information decoding unit 1102 decodes the encoded sequence of the prediction error information supplied from the demultiplexing unit 1101, and generates a quantized prediction error signal.
  • the prediction difference information decoding unit 1102 supplies the generated quantized prediction error signal to the inverse quantization / inverse transform unit 1103.
  • the inverse quantization / inverse transform unit 1103 performs a process such as inverse quantization or inverse orthogonal transform on the quantized prediction error signal supplied from the prediction difference information decoding unit 1102 to generate a prediction error signal,
  • the decoded prediction error signal is supplied to the adding unit 1104.
  • the adder 1104 adds the decoded prediction error signal supplied from the inverse quantization / inverse transform unit 1103 and the prediction signal supplied from the prediction mode / block structure selection unit 1109 to generate a decoded image signal, and generates a decoded image signal.
  • the signal is supplied to the intra-frame decoded image buffer 1105 and the loop filter unit 1106.
  • the intra-frame decoded image buffer 1105 has the same function as the intra-frame decoded image buffer 108 in the moving picture encoding apparatus in FIG. 1, and supplies a decoded image signal in the same frame to the intra prediction unit 1113 as a reference image for intra prediction. At the same time, the decoded image signal supplied from the adding unit 1104 is stored.
  • the loop filter unit 1106 has the same function as the loop filter unit 109 in the moving picture coding apparatus in FIG. 1, performs a distortion removal filter on the decoded image signal supplied from the addition unit 1104, and performs filter processing.
  • the decoded image obtained as a result of the execution is supplied to the decoded image memory 1107.
  • the decoded image memory 1107 has the same function as the decoded image memory 110 in the moving image encoding apparatus in FIG. 1, stores the decoded image signal supplied from the loop filter unit 1106, and uses the reference image signal as the motion compensation prediction unit 1114. To supply.
  • the decoded image memory 1107 supplies the stored decoded image signal to the output terminal 1115 in accordance with the display order of images in accordance with the reproduction time.
  • the prediction mode / block structure decoding unit 1108 is a control parameter that defines the CU structure shown in FIG. 3 based on the control parameters related to the coding block and the prediction block structure supplied from the demultiplexing unit 1101, and the control parameter shown in FIG. Such a motion compensation prediction block configuration and control parameters for limiting the prediction process are generated.
  • the prediction mode / block structure decoding unit 1108 uses the CU division configuration for each coding block unit supplied from the demultiplexing unit 1101 and the coding sequence of the mode information used for prediction to determine the CU for each coding block unit.
  • the mode information used for the division configuration and prediction is decoded to generate a prediction block size and a prediction mode, and the prediction type, motion vector, and reference image designation information corresponding to the prediction mode in the case of motion compensated prediction are specified.
  • the motion information, which is information, and the intra prediction mode information in the case of intra prediction are separated, and the CU partition configuration and the prediction mode information for each coding block are supplied to the prediction mode / block structure selection unit 1109.
  • the prediction mode / block structure decoding unit 1108 supplies intra prediction mode information to the intra prediction information decoding unit 1110 together with the prediction block size, and motion compensated prediction is used. If so, the motion information decoding unit 1111 is supplied with information for specifying the motion compensation prediction mode, the prediction type corresponding to the prediction mode, the motion vector, and the reference image designation information together with the prediction block size.
  • the intra prediction information decoding unit 1110 decodes the prediction block size and intra prediction mode information supplied from the prediction mode / block structure decoding unit 1108, and reproduces the prediction block structure for the encoding target block and the intra prediction mode in each prediction block. To do.
  • the intra prediction information decoding unit 1110 supplies the reproduced intra prediction mode to the intra prediction unit 1113 and also supplies it to the prediction mode information memory 1112.
  • the motion information decoding unit 1111 is supplied from the prediction mode / block structure decoding unit 1108 and specifies the prediction block size, the motion compensation prediction mode, and the prediction type, motion vector, and reference image designation information corresponding to the prediction mode. From the decoded motion information and the motion information of the candidate block group supplied from the prediction mode information memory 1112 to reproduce the prediction type, motion vector, and reference image designation information used for motion compensation prediction, This is supplied to the compensation prediction unit 1114. The motion information decoding unit 1111 also supplies the reproduced motion information to the prediction mode information memory 1112. A detailed configuration of the motion information decoding unit 1111 will be described later.
  • the prediction mode information memory 1112 has the same function as the prediction mode information memory 119 in the moving picture encoding apparatus in FIG. 1 and is supplied from the reproduced motion information supplied from the motion information decoding unit 1111 and the intra prediction information decoding unit 1110.
  • the intra prediction mode to be performed is stored for a predetermined image on the basis of the minimum prediction block size unit.
  • the prediction mode information memory 1112 supplies motion information of the spatial candidate block group and the temporal candidate block group to the motion information decoding unit 1111 as motion information of the candidate block group, and also intra of the decoded adjacent block in the same frame.
  • the prediction mode information is supplied to the intra prediction information decoding unit 1110 as a prediction candidate of the mode information of the target prediction block.
  • the intra prediction unit 1113 has the same function as the intra prediction unit 114 in the moving picture coding apparatus in FIG. 1, and performs intra prediction from the intra-frame decoded image buffer 1105 according to the intra prediction mode supplied from the intra prediction information decoding unit 1110. A reference image is input, an intra prediction signal is generated, and supplied to the prediction mode / block structure selection unit 1109.
  • the motion compensation prediction unit 1114 has the same function as the motion compensation prediction unit 112 in the video encoding device of FIG. 1, and based on the motion information supplied from the motion information decoding unit 1111, the reference image in the decoded image memory 1107.
  • a prediction signal is generated by acquiring an image signal at a position obtained by moving the reference image indicated by the designation information from the same position as the image signal of the prediction block by the motion vector value. If the prediction type of motion compensation prediction is bi-prediction, an average of the prediction signals of each prediction type is generated as a prediction signal, and the prediction signal is supplied to the prediction mode / block structure selection unit 1109.
  • the prediction mode / block structure selection unit 1109 performs CU partitioning based on the CU partitioning configuration for each coding block supplied from the prediction mode / block structure decoding unit 1108 and the prediction mode information, and reproduces the predicted Depending on the prediction mode of the block structure unit, in the case of motion compensation prediction, a motion compensation prediction signal is input from the motion compensation prediction unit 1114, and in the case of intra prediction, an intra prediction signal is input from the intra prediction unit 1113 and reproduced. The predicted signal is supplied to the adding unit 1104.
  • the output terminal 1115 outputs the decoded image signal supplied from the decoded image memory 1107 to a display medium such as a display, thereby reproducing the decoded image signal.
  • the configuration of the video decoding device shown in FIG. 11 can also be realized by hardware such as an information processing device including a CPU, a frame memory, a hard disk, and the like, similarly to the configuration of the video encoding device shown in FIG. is there.
  • FIG. 12 is a flowchart showing a flow of operation in units of coding blocks of decoding processing in the video decoding apparatus according to Embodiment 1 of the present invention.
  • CU_Depth which is a control parameter for CU partitioning, is initialized to 0 (S1200), and the demultiplexing unit 1101 converts the coded bitstream supplied from the input terminal 1100 into the coded sequence of prediction error information and the coded data.
  • the block is divided into a CU partition configuration and a coded sequence of mode information used for prediction (S1201).
  • the encoded sequence of prediction error information in units of encoded blocks, the CU partition configuration in units of the encoded blocks, and the encoded sequence of mode information used for prediction are the prediction difference information decoding unit 1102, the prediction mode / It is supplied to the block structure decoding unit 1108 and subjected to decoding processing in units of CUs based on the CU partition structure (S1202). The detailed operation of step S1202 will be described later.
  • the CU partitioning configuration for each coding block is decoded by the prediction mode / block structure decoding unit 1108 in step S1202, and the decoded coding structure information is stored in the prediction mode information memory 1112 (S1203).
  • the decoded image signal decoded by the decoding process in units of CU is subjected to loop filter processing in the loop filter unit 1106 (S1204), stored in the decoded image memory 1107 (S1205), and decoded in units of coding blocks. Ends.
  • the loop filter is applied in the process of the coding block unit, but the decoded image signal subjected to the loop filter is not referred to in the decoding process of the same frame, and in the motion compensation prediction of the subsequent frame Since it is referred to, it is possible to perform the process on the entire frame after the decoding process for the entire frame is completed without performing the process for each coding block.
  • Max_CU_Depth indicating the number of layers between the set maximum CU size and the minimum CU size (S1300). Since the control parameters relating to the maximum CU size and the minimum CU size in FIG. 3 are encoded and transmitted, Max_CU_Depth at the time of encoding is decoded by decoding the control parameters in the decoding process. An example of the encoding information that defines Max_CU_Depth will be described later.
  • CU partition information is acquired (S1301).
  • 1-bit flag information (cu_split_flag) is encoded and transmitted according to the selection of whether or not to divide a CU, and whether or not the CU is divided by decoding this flag information. Recognize
  • CU_Depth is greater than or equal to Max_CU_Depth (S1300: NO), and when the CU is not divided (S1302: NO), the size of the CU to be decoded is determined, and it corresponds to the prediction mode in the determined CU. Decoding processing is performed.
  • skip_flag skip flag information indicating whether or not the skip mode is in CU units
  • prediction indicating whether the prediction is intra prediction or motion compensation prediction when the CU is not in skip mode
  • Mode flag information is encoded as prediction mode information in units of CU at the time of encoding, and information indicating whether it is intra prediction or motion compensated prediction (including skip mode) by decoding these. Can be obtained.
  • intra prediction decoding processing for each CU is performed by the intra prediction information decoding unit 1110 and the intra prediction unit 1113 in FIG. 11 (S1311), and the target An intra prediction signal in the CU is generated and added to the decoding error signal to generate a decoded image signal (S1312), and the decoding process in units of CUs is completed.
  • step S1310 When the CU is not intra prediction (S1309: NO), motion compensation prediction decoding processing for each CU is performed by the motion information decoding unit 1111 and the motion compensation prediction unit 1114 in FIG. 11 (S1310), and motion in the target CU.
  • a compensated prediction signal is generated and added to the decoded error signal to generate a decoded image signal (S1312), and the decoding process for each CU is completed. Details of the operation in step S1310 will be described later.
  • a decoded skip flag is acquired as information indicating the prediction mode in units of CUs (S1400).
  • the skip flag is 1, that is, in the skip mode (S1401: YES)
  • prediction block partitioning within the CU is performed.
  • the mode is 2N ⁇ 2N
  • NumPart is set to 1
  • prediction block unit decoding of a 2N ⁇ 2N prediction block is performed (S1402).
  • the CU partition (PU) mode is set as the CU partition mode value that is the type of the motion compensated prediction block size selected by the CU at the time of encoding.
  • the PU mode is 2N ⁇ 2N (S1404: YES)
  • NumPart is set to 1 and prediction block unit decoding of the 2N ⁇ 2N prediction block is performed (S1402). .
  • the PU mode is N ⁇ 2N (S1409: NO)
  • the PU mode is N ⁇ N
  • NumPart is set to 4
  • prediction block unit decoding of the N ⁇ N prediction block is performed (S1410).
  • step S1407 When the condition of step S1407 is not satisfied (S1407: NO), since the N ⁇ N prediction block is not applied in the CU, NumPart is set to 2, and prediction block unit decoding of the N ⁇ 2N prediction block is performed. (S1408). Details of the prediction block unit decoding process for each PU mode performed in steps S1402, S1406, S1408, and S1410 will be described later.
  • the processes are performed in the order shown in steps S1404 to S1409 as shown in the flowchart of FIG.
  • the decoding process is performed in units of prediction blocks in accordance with the decoded PU mode, it is possible to implement a different configuration regarding the order of conditional branches.
  • mode information such as the PU mode and motion information for each prediction block is stored in the prediction mode information memory 1112 in FIG. 11 (S1411), and motion compensation for the CU is performed.
  • the predictive decoding process ends.
  • FIG. 15 is a diagram illustrating a detailed configuration of the motion compensated prediction block structure selection unit 113 in the video encoding device according to the first embodiment.
  • the motion compensation prediction block structure selection unit 113 has a function of determining an optimal motion compensation prediction mode and a prediction block structure.
  • the motion compensation prediction block structure selection unit 113 includes a motion compensation prediction generation unit 1500, a prediction error calculation unit 1501, a prediction vector calculation unit 1502, a difference vector calculation unit 1503, a motion information code amount calculation unit 1504, and a prediction mode / block structure evaluation unit. 1505, a combined motion information calculation unit 1506, a combined motion information single prediction conversion unit 1507, and a combined motion compensation prediction generation unit 1508 are included.
  • the motion vector value input from the motion vector detection unit 111 to the motion compensation prediction block structure selection unit 113 in FIG. 1 is supplied to the motion compensation prediction generation unit 1500 and input from the prediction mode information memory 119. Is supplied to the prediction vector calculation unit 1502 and the combined motion information calculation unit 1506.
  • reference image designation information and motion vectors used for motion compensation prediction are output from the motion compensation prediction generation unit 1500 and the combined motion compensation prediction generation unit 1508 to the motion compensation prediction unit 112.
  • the generated motion compensated prediction image is supplied to the prediction error calculation unit 1501.
  • the prediction error calculation unit 1501 is further supplied with an image signal of a prediction block to be encoded from the encoding block acquisition unit 102.
  • the prediction mode / block structure evaluation unit 1505 supplies the prediction block structure, the motion information to be encoded and the determined prediction mode information, and the motion compensated prediction signal to the prediction mode selection unit 116.
  • the motion compensation prediction generation unit 1500 receives the motion vector value calculated for each reference image usable for prediction in each prediction block structure, and performs motion compensation prediction according to the bi-prediction restriction information shown in FIG.
  • the reference image designation information is supplied to the prediction vector calculation unit 1502, and the reference image designation information and the motion vector are output.
  • the prediction error calculation unit 1501 calculates a prediction error evaluation value from the input motion compensated prediction image and the prediction block image to be processed.
  • the sum SAD of the absolute difference value for each pixel, the sum SSE of the square error value for each pixel, and the like can be used as in the error evaluation value in motion vector detection.
  • a more accurate error evaluation value can be calculated by taking into account the amount of distortion components generated in the decoded image by performing orthogonal transform / quantization performed when encoding the prediction residual.
  • the prediction error calculation unit 1501 can be realized by having the functions of the subtraction unit 103, the orthogonal transformation / quantization unit 104, the inverse quantization / inverse transformation unit 106, and the addition unit 107 in FIG.
  • the prediction error calculation unit 1501 supplies the prediction error evaluation value calculated in each prediction mode and each prediction block structure and the motion compensation prediction signal to the prediction mode / block structure evaluation unit 1505.
  • the prediction vector calculation unit 1502 is supplied with the reference image designation information from the motion compensation prediction generation unit 1500, and the motion vector for the designated reference image from the candidate block group in the adjacent block motion information supplied from the prediction mode information memory 119. A value is input, a plurality of prediction vectors are generated together with a prediction vector candidate list, and supplied to the difference vector calculation unit 1503 together with reference image designation information. The prediction vector calculation unit 1502 creates prediction vector candidates and registers them as prediction vector candidates.
  • the difference vector calculation unit 1503 calculates the difference between each of the prediction vector candidates supplied from the prediction vector calculation unit 1502 and the motion vector value supplied from the motion compensated prediction generation unit 1500, and calculates the difference vector value. calculate.
  • the prediction vector index which is the designation information for the calculated difference vector value and the prediction vector candidate is encoded
  • the code amount is the smallest.
  • the difference vector calculation unit 1503 supplies the prediction vector index and the difference vector value for the prediction vector having the smallest information amount, together with the reference image designation information, to the motion information code amount calculation unit 1504.
  • the motion information code amount calculation unit 1504 requires motion information in each prediction block structure and each prediction mode from the difference vector value, reference image designation information, prediction vector index, and prediction mode supplied from the difference vector calculation unit 1503. The code amount is calculated. Also, the motion information code amount calculation unit 1504 receives from the combined motion compensation prediction generation unit 1508 information indicating the combined motion information index and the prediction mode that needs to be transmitted in the combined prediction mode, and moves in the combined prediction mode. The amount of code required for information is calculated.
  • the motion information code amount calculation unit 1504 supplies the motion information calculated in each prediction block structure and each prediction mode and the code amount required for the motion information to the prediction mode / block structure evaluation unit 1505.
  • the prediction mode / block structure evaluation unit 1505 uses the prediction error evaluation value of each prediction mode supplied from the prediction error calculation unit 1501 and the motion information code amount of each prediction mode supplied from the motion information code amount calculation unit 1504. Calculating a total motion compensation prediction error evaluation value of each prediction mode, selecting a prediction mode and a prediction block size which are the smallest evaluation values, and selecting a prediction mode, a prediction block size and motion information for the selected prediction mode. To the prediction mode selection unit 116. Similarly, the prediction mode / block structure evaluation unit 1505 selects a prediction signal in the selected prediction mode and prediction block size with respect to the motion compensation prediction signal supplied from the prediction error calculation unit 1501, and selects a prediction mode selection unit. To 116.
  • the combined motion information calculation unit 1506 uses the candidate block group in the motion information of the adjacent blocks supplied from the prediction mode information memory 119, a prediction type indicating whether the prediction is uni-prediction or bi-prediction, reference image designation information, A plurality of pieces of motion information are generated together with a combined motion information candidate list as motion information composed of motion vector values, and supplied to the combined motion information single prediction conversion unit 1507.
  • FIG. 16 is a diagram illustrating a configuration of the combined motion information calculation unit 1506.
  • the combined motion information calculation unit 1506 includes a spatial combined motion information candidate list generation unit 1600, a combined motion information candidate list deletion unit 1601, a temporal combined motion information candidate list generation unit 1602, a first combined motion information candidate list addition unit 1603, and a second.
  • a combined motion information candidate list adding unit 1604 is included.
  • the combined motion information calculation unit 1506 creates motion information candidates in a predetermined order from spatially adjacent candidate block groups, deletes candidates having the same motion information from the candidates, and then temporally adjacent. By adding motion information candidates created from the candidate block group, only valid motion information is registered as combined motion information candidates.
  • this temporally combined motion information candidate list generation unit is arranged after the combined motion information candidate list deletion unit is a characteristic configuration of the present embodiment, and deletes the same motion information from temporally combined motion information candidates. By eliminating the processing target, it is possible to reduce the amount of calculation without reducing the encoding efficiency.
  • the detailed operation of the combined motion information calculation unit 1506 will be described later.
  • the combined motion information single prediction conversion unit 1507 performs the bi-directional processing shown in FIG. 10 on the combined motion information candidate list supplied from the combined motion information calculation unit 1506 and the motion information registered in the candidate list.
  • motion information whose prediction type is bi-prediction is converted into uni-prediction motion information and supplied to the combined motion compensation prediction generation unit 1508.
  • the combined motion compensated prediction generation unit 1508 corresponds to each registered combined motion information candidate from the combined motion information candidate list supplied from the combined motion information single prediction conversion unit 1507 according to the prediction type based on the motion information.
  • the reference image designation information and motion vector value of one reference image (uni-prediction) or two different reference images (bi-prediction) are designated to the motion compensation prediction unit 112 to generate a motion compensated prediction image, and the respective combinations
  • the motion information index is supplied to the motion information code amount calculation unit 1504.
  • the prediction mode evaluation in each combined motion information index is performed by the prediction mode / block structure evaluation unit 1505.
  • the prediction error evaluation value and the motion information code amount are used as the prediction error calculation unit 1501 and the motion information.
  • FIG. 17 is a flowchart for explaining detailed operations of the motion compensation prediction mode / prediction signal generation processing in steps S701, S702, S703, and S705 in the flowchart of FIG. This operation represents a detailed operation in the motion compensated prediction block structure selection unit 113 in FIG.
  • step S1701 based on the NumPart set according to the predicted block size division mode (PU) in the defined CU, the steps from step S1701 to step S1708 are performed for each prediction block size obtained by PU division in the target CU (S1700). It is executed (S1709). First, a combined motion information candidate list is generated (S1701).
  • the prediction block size is generated.
  • the combined motion information candidate uni-prediction conversion is performed in which the bi-prediction motion information in each candidate in the combined motion information candidate list is replaced with the single-prediction motion information (S1703). If the predicted block size is not less than or equal to bipred_restriction_size (S1702: NO), the process proceeds to subsequent step S1704.
  • a combined prediction mode evaluation value is generated based on the motion information in the combined motion information candidate list generated or replaced (S1704). Subsequently, a prediction mode evaluation value is generated (S1705), and an optimal prediction mode is selected by comparing the generated evaluation values (S1706).
  • the order of evaluation value generation in steps S1704 and S1705 is not limited to this.
  • the prediction signal is output according to the selected prediction mode (S1707), and the motion information is output according to the selected prediction mode (S1708), thereby completing the motion compensation prediction mode / prediction signal generation processing for each prediction block.
  • steps S1701, S1703, S1704, and S1705 will be described later.
  • FIG. 18 is a flowchart for explaining the detailed operation of generating the combined motion information candidate list in step S1701 of FIG. This operation shows the detailed operation of the configuration in the combined motion information calculation unit 1506 in FIG.
  • the spatial combination motion information candidate list generation unit 1600 in FIG. 16 performs spatial combination from candidate blocks excluding candidate blocks outside the region or candidate blocks in the intra mode from the spatial candidate block group supplied from the prediction mode information memory 119.
  • a motion information candidate list is generated (S1800). Detailed operations for generating the spatially coupled motion information candidate list will be described later.
  • the combined motion information candidate list deletion unit 1601 deletes the combined motion information candidates having the same motion information from the generated spatial combined motion information candidate list and updates the motion information candidate list (S1801). Detailed operation of the combined motion information candidate deletion will be described later.
  • the temporally combined motion information candidate list generation unit 1602 subsequently performs temporally combined motion from candidate blocks excluding candidate blocks outside the region or candidate blocks in the intra mode from the temporal candidate block group supplied from the prediction mode information memory 119.
  • An information candidate list is generated (S1802) and combined with the temporally combined motion information candidate list to form a combined motion information candidate list. Detailed operation of the time combination motion information candidate list generation will be described later.
  • the first combined motion information candidate list adding unit 1603 generates 0 to 2 first combined motion information candidates from the combined motion information candidate registered in the combined motion information candidate list generated by the temporal combined motion information candidate list generating unit 1602.
  • a combined motion information candidate is generated and added to the combined motion information candidate list (S1803), and the combined motion information candidate list is supplied to the second combined motion information candidate list adding unit 1604. The detailed operation of adding the first combined motion information candidate list will be described later.
  • the second combined motion information candidate list adding unit 1604 selects 0 to 4 second combined motion information candidates that do not depend on the combined motion information candidate list supplied from the first combined motion information candidate list adding unit 1603. Generated and added to the combined motion information candidate list supplied from the first combined motion information candidate list adding unit 1603 (S1804), and the process ends. Detailed operations for adding the second combined motion information candidate list will be described later.
  • the candidate block group of motion information supplied from the prediction mode information memory 119 to the combined motion information calculation unit 1506 includes a spatial candidate block group and a temporal candidate block group. First, generation of a spatially coupled motion information candidate list will be described.
  • FIG. 19 is a diagram showing a spatial candidate block group used for generating a spatially coupled motion information candidate list.
  • the spatial candidate block group indicates a block of the same image adjacent to the prediction target block of the encoding target image.
  • the block group is managed in units of the minimum prediction block size, and the position of the candidate block is managed in units of the minimum prediction block size, but when the prediction block size of the adjacent block is larger than the minimum prediction block size
  • the same motion information is stored in all candidate blocks within the predicted block size.
  • FIG. 20 is a flowchart for explaining the detailed operation of generating the spatially coupled motion information candidate list.
  • the following processing is repeated for block A0, block A1, block B0, block B1, and block B2 in the order of block A1, block B1, block B0, and block A0. (S2000 to S2003).
  • the validity of the candidate block is checked (S2001). If the candidate block is not out of the region and not in the intra mode, the candidate block is valid. If the candidate block is valid (S2001: YES), the motion information of the candidate block is added to the spatially combined motion information candidate list (S2002).
  • the spatial combination motion information candidate list includes motion information of four or less candidate blocks, but the spatial candidate block group is at least one or more processed blocks adjacent to the prediction block to be processed.
  • the number of spatially coupled motion information candidate lists may be changed depending on the effectiveness of the candidate block, and the present invention is not limited to this.
  • step S2100 to S2105 Subsequent to the repetitive processing from step S2100 to S2105, 1 is subtracted from i, and the processing for candidate (i) is repeated (S2100 to S2106).
  • FIG. 22 shows a comparison relationship of candidates in the list when there are four combined motion information candidates. That is, four spatially combined motion information candidates that do not include temporally combined motion information candidates are compared by brute force to determine identity, and duplicate candidates are deleted.
  • the joint prediction mode uses temporal and spatial continuity of motion
  • the prediction target block encodes motion information of spatially and temporally adjacent blocks without directly encoding its own motion information.
  • the spatially coupled motion information candidate is based on continuity in the spatial direction
  • the temporally coupled motion information candidate is generated by the method described later based on the temporal direction continuity. These properties are different. Therefore, it is rare that the same motion information is included in the temporally combined motion information candidate and the spatially combined motion information candidate, and the temporally combined motion information candidate is removed from the target of the combined motion information candidate deletion process for deleting the same motion information. Even if they are excluded, it is rare that the same motion information is included in the finally obtained combined motion information candidate list.
  • temporally combined motion information candidate blocks are managed in units of minimum time prediction blocks that are larger in size than the minimum prediction block, so that the size of prediction blocks that are temporally adjacent are larger than the minimum time prediction block If it is small, motion information at a position deviating from the original position is used, and as a result, the motion information often includes an error. Therefore, the motion information is often different from the motion information of the spatially coupled motion information candidate, and there is little influence even if the motion information is excluded from the target of the combined motion information candidate deletion process for deleting the same motion information.
  • FIG. 23 is an example of comparison contents of candidates in deletion of combined motion information candidates when the maximum number of spatially combined motion information candidates is four.
  • FIG. 23A shows the comparison contents when only the spatially coupled motion information candidate is the target of the coupled motion information candidate deletion process
  • FIG. 23B is the target of processing the spatially coupled motion information candidate and the temporally coupled motion information. It is a comparison content in the case of.
  • the number of motion information comparisons is reduced from 10 to 6 while appropriately deleting the same motion information. Is possible.
  • the combined motion information calculated from the B1 position in FIG. 19 is compared with the combined motion information at the A1 position
  • the combined motion information calculated from the B0 position is compared with only the combined motion information at the B1 position
  • A0 By comparing the combined motion information calculated from the position with only A1 and the combined motion information calculated from the B2 position with only A1 and B1, the number of motion information comparisons can be limited to a maximum of five.
  • FIG. 24 is a diagram illustrating the definition of the temporal direction peripheral prediction block used for generating the temporally combined motion information candidate list.
  • the temporal candidate block group indicates blocks in the same position as and around the prediction target block among the blocks belonging to the decoded image ColPic different from the image to which the prediction target block belongs.
  • the block group is managed in units of minimum time prediction block size, and the positions of candidate blocks are managed in units of minimum time prediction block size.
  • the minimum temporal prediction block size is set to a size obtained by doubling the minimum prediction block size in the vertical and horizontal directions.
  • FIG. 24B shows motion information of the temporal direction neighboring prediction block when the prediction block size is smaller than the minimum temporal prediction block size.
  • blocks at positions A1 to A4, B1 to B4, C, D, E, F1 to F4, G1 to G4, H, and I1 to I16 are temporally adjacent block groups.
  • the temporal candidate block group is assumed to be two blocks, block H and block I6.
  • FIG. 25 is a flowchart for explaining the detailed operation of generating the time combination motion information candidate list.
  • the validity of the candidate block is checked in the order of the block H and the block I11 (S2501). If the candidate block is valid (S2501: YES), the processing from step S2502 to step S2504 is performed, the generated motion information is registered in the time combination motion information candidate list, and the processing ends.
  • the candidate block indicates a position outside the screen area, or when the candidate block is an intra prediction block (S2501: NO)
  • the candidate block is not valid, and valid / invalid determination of the next candidate block is performed.
  • the reference image selection candidate to be registered in the combined motion information candidate is determined based on the motion information of the candidate block (S2502).
  • the L0 prediction reference image is the reference image that is the closest to the processing target image among the L0 prediction reference images
  • the L1 prediction reference image is the processing target image among the L1 prediction reference images.
  • the reference image is the closest distance.
  • the method for determining the reference image selection candidate here is not limited to this as long as the reference image for L0 prediction and the reference image for L1 prediction can be determined.
  • the reference image intended at the time of encoding can be determined by determining the reference image by the same method in the encoding process and the decoding process.
  • a method of selecting a reference image having a reference image index of 0 for a reference image for L0 prediction and a reference image for L1 prediction, or a L0 reference image and a L1 reference image used by spatially neighboring blocks. can be used, and a method of specifying a reference image of each prediction type in the encoded stream can be used.
  • the motion vector value to be registered in the combined motion information candidate is determined based on the motion information of the candidate block (S2503).
  • the temporally coupled motion information calculates bi-prediction motion information based on motion vector values that are effective prediction types in motion information of candidate blocks.
  • the prediction type of the candidate block is L0 prediction or L1 prediction single prediction
  • motion information of the prediction type (L0 prediction or L1 prediction) used for prediction is selected, and its reference image designation information and motion vector value are selected. Is a reference value for generating bi-predictive motion information.
  • L0 prediction or L1 prediction motion information is selected as a reference value.
  • the reference value selection method selects, for example, motion information existing in the same prediction type as ColPic, and selects a reference image having a shorter inter-image distance from ColPic in each of L0 prediction and L1 prediction of a candidate block. For example, it is possible to select the transmission side and explicitly transmit the syntax.
  • the motion vector value used as the reference for bi-predictive motion information generation is determined, the motion vector value to be registered in the combined motion information candidate is calculated.
  • FIG. 26 is a diagram for explaining a calculation method of motion vector values mvL0t and mvL1t registered for L0 prediction and L1 prediction with respect to the reference motion vector value ColMv for temporally coupled motion information.
  • the distance between images between ColPic for the reference motion vector value ColMv and the reference image that is the target of the motion vector used as a reference for the candidate block is referred to as ColDist.
  • the inter-image distance between each reference image of L0 prediction and L1 prediction and the processing target image is set to CurrL0Dist and CurrL1Dist.
  • a motion vector obtained by scaling ColMv with a distance ratio of ColDist to CurrL0Dist and CurrL1Dist is set as a motion vector to be registered.
  • the motion vector values mvL0t and mvL1t to be registered are calculated by the following formulas 1 and 2.
  • mvL0t mvCol ⁇ CurrL0Dist / ColDist (Formula 1)
  • mvL1t mvCol ⁇ CurrL1Dist / ColDist (Formula 2) It becomes.
  • the bi-predicted reference image selection information (index) and the motion vector value generated in this way are added to the combined motion information candidates (S2504), and the temporal combined motion information candidate list creation process ends. To do.
  • FIG. 27 is a flowchart for explaining the operation of the first combined motion information candidate list adding unit 1603.
  • NumCandList the number of combined motion information candidates
  • MaxNumMergeCand the maximum number of combined motion information candidates registered in the combined motion information candidate list supplied from the temporally combined motion information candidate list generation unit 1602
  • MaxNumGenCand which is the maximum number for generating motion information candidates, is calculated from Equation 3 (S2700).
  • MaxNumGenCand MaxNumMergeCand-NumCandList; (NumCandList> 1)
  • MaxNumGenCand is larger than 0 (S2701). If MaxNumGenCand is not greater than 0 (S2701: NO), the process ends. If MaxNumGenCand is greater than 0 (S2701: YES), the following processing is performed. First, loopTimes that is the number of combination inspections is determined. LoopTimes is set to NumCandList ⁇ NumCandList. However, if loopTimes exceeds 8, loopTimes is limited to 8 (S2702). Here, loopTimes is an integer from 0 to 7. The following processing is repeatedly performed for loopTimes (S2702 to S2708).
  • the combination of the combined motion information candidate M and the combined motion information candidate N is determined (S2703).
  • the relationship between the number of combination inspections, the combined motion information candidate M, and the combined motion information candidate N will be described.
  • FIG. 28 is a diagram for explaining the relationship between the number of combination inspections, the combined motion information candidate M, and the combined motion information candidate N.
  • M and N are different values. First, M is fixed to 0, the value of N is changed to 1 to 4 (the maximum value is NumCandList), and then the value of N is fixed to 0. The value of M is changed to 1 to 4 (the maximum value is NumCandList).
  • Such a combination definition makes effective use of the first motion information in the combined motion information candidate list, which is the motion information with the highest probability of being selected, and actually calculates the combination pattern without having a combination table. There is an effect that can be calculated.
  • the combined motion information candidate M uses the motion vector of the L0 prediction of the combined motion information candidate M and the reference image.
  • a combined motion information candidate is generated by combining the motion vector of N L1 predictions and the reference image (S2705). If the L0 prediction of the combined motion information candidate M is not valid and the L1 prediction of the combined motion information candidate N is not valid (S2704: NO), the next combination is processed.
  • the motion information of the L0 prediction and the L1 prediction may be the same, and even if motion compensation is performed by bi-prediction, the same result as the single prediction of the L0 prediction or the L1 prediction is obtained. Therefore, the additional combined motion information candidate generation in which the motion information of the L0 prediction and the motion information of the L1 prediction are the same is a factor that increases the amount of calculation of the motion compensation prediction. For this reason, normally, whether or not the motion information of the L0 prediction and the motion information of the L1 prediction are the same is compared, and only when they are not the same, the first additional combined motion information candidate is set.
  • step S2705 a double-coupled motion information candidate is added to the combined motion information candidate list (S2706). Subsequent to step S2706, it is checked whether the number of generated double-coupled motion information is MaxNumGenCand (S2707). If the number of generated double coupled motion information is MaxNumGenCand (YES in S2707), the process ends. If the number of generated double coupled motion information is not MaxNumGenCand (NO in S2707), the next combination is processed.
  • the first additional combined motion information candidate is a combined motion information candidate when there is a slight difference between the motion information of the combined motion information candidate registered in the combined motion information candidate list and the motion information candidate motion to be processed. Coding efficiency can be improved by correcting the motion information of the combined motion information candidates registered in the list to generate effective combined motion information candidates.
  • FIG. 29 is a flowchart for explaining the operation of the second combined motion information candidate list adding unit 1604.
  • the first addition is performed based on the number of combined motion information candidates (NumCandList) and the maximum number of combined motion information candidates (MaxNumMergeCand) registered in the combined motion information candidate list supplied from the first combined motion information candidate list adding unit 1603.
  • MaxNumGenCand which is the maximum number for generating combined motion information candidates, is calculated from Equation 4 (S2900).
  • MaxNumGenCand MaxNumMergeCand-NumCandList; (Formula 4)
  • i is an integer from 0 to MaxNumGenCand-1.
  • the second additional combined motion in which the motion vector for L0 prediction is (0,0), the reference index is i, the motion vector for L1 prediction is (0,0), and the prediction type is i for the reference index is bi-prediction.
  • Information candidates are generated (S2902).
  • the second additional combined motion information candidate is added to the combined motion information candidate list (S2903).
  • the next i is processed (S2904).
  • the second additional combined motion information candidate has a motion vector for L0 prediction of (0, 0), a reference index of i, a motion vector of L1 prediction of (0, 0), and a reference index of i.
  • the combined motion information candidate whose prediction type is bi-prediction was used. This is because, in a general moving image, the frequency of occurrence of combined motion information candidates in which the motion vector for L0 prediction and the motion vector for L1 prediction are (0, 0) is statistically high.
  • the present invention is not limited to this as long as it is a combined motion information candidate that is statistically frequently used without depending on the motion information of the combined motion information candidate registered in the combined motion information candidate list.
  • the motion vectors of L0 prediction and L1 prediction may be vector values other than (0, 0), respectively, and may be set so that the reference indexes of L0 prediction and L1 prediction are different.
  • the second additional combined motion information candidate can be set as an encoded image or motion information with a high occurrence frequency of a part of the encoded image, encoded in an encoded stream, and transmitted.
  • the motion compensation unit to be described later performs a process of collectively converting bi-prediction into single prediction, so that the motion information and L1 of the L0 prediction in the second additional combined motion information candidate list adding unit It is not necessary to determine the identity of predicted motion information, and the amount of calculation can be reduced.
  • the combined motion information candidate registered in the combined motion information candidate list When the number is zero, it is possible to use the joint prediction mode and improve the encoding efficiency.
  • the motion information of the combined motion information candidate registered in the combined motion information candidate list and the motion information candidate motion to be processed are different, by generating a new combined motion information candidate and expanding the range of options, Encoding efficiency can be improved.
  • the motion information stored in the index i is acquired from the combined motion information candidate list (S3001). Subsequently, when the prediction type of the motion information is single prediction (S3002: YES), the process for the motion information stored in the index i is terminated as it is, and the process proceeds to the next index (S3005).
  • the motion information is not uni-prediction, that is, when the motion information is bi-prediction (S3002: NO)
  • the L1 information of the motion information stored in the index i is used to convert the bi-prediction motion information into uni-prediction. Is invalidated (S3003).
  • bi-prediction motion information is converted into L0 prediction single prediction by invalidating L1 information in this way, but conversely L0 information is invalidated and bi-prediction motion information.
  • Can be converted to single prediction of L1 prediction and can be realized by defining a prediction type to be invalidated when implicitly converting to single prediction.
  • the motion information of the index i converted to single prediction is stored (S3004), and the process proceeds to the next index (S3005).
  • the combined motion information bi-prediction restriction based on the prediction block size is performed after the combined motion information candidate list is generated once and then the combined motion information candidate single prediction conversion process shown in the flowchart of FIG. 30 is performed.
  • the combined motion information single prediction conversion process a determination is made for each candidate generation within the process shown in the flowchart of FIG. 18 which is a combined motion information candidate generation process, and a combined motion information candidate list for single prediction is generated.
  • the condition judgment based on the predicted block size enters each process, which complicates the process and increases the load of the list construction process.
  • the first embodiment has an effect of realizing a bi-prediction restriction process that prevents an increase in the load of the list building process by performing a process of converting motion information into a single prediction after building a list once.
  • FIG. 31 is a flowchart for explaining the detailed operation of the combined prediction mode evaluation value generation process in step S1702 of FIG. This operation shows a detailed operation of the configuration using the combined motion compensation prediction generation unit 1508 of FIG.
  • the motion information stored in the index i is acquired from the combined motion information candidate list (S3102). Subsequently, a motion information code amount is calculated (S3103). In the joint prediction mode, since only the joint motion information index is encoded, only the joint motion information index becomes the motion information code amount.
  • a Truncated Unary code string is used as the code string of the combined motion information index.
  • FIG. 32 is a diagram showing a Trunked Unary code string when the number of combined motion information candidates is five.
  • the value of the combined motion information index is encoded using the Truncated Unary code string, the smaller the combined motion information index, the smaller the code bits assigned to the combined motion information index.
  • the number of combined motion information candidates is 5, if the combined motion information index is 1, it is represented by 2 bits of “10”, but if the combined motion information index is 3, 4 bits of “1110”. It is expressed by
  • the Truncated Unary code string is used to encode the combined motion information index, but other code string generation methods can be used, and the present invention is not limited to this.
  • the motion information prediction type is single prediction (S3104: YES)
  • the reference image designation information and the motion vector for one reference image are set in the motion compensation prediction unit 112 in FIG.
  • a prediction block is generated (S3105).
  • the motion information is not uni-prediction, that is, when the motion information is bi-prediction (S3104: NO)
  • reference image designation information and motion vectors for two reference images are set in the motion compensation prediction unit 112, and motion compensation is performed.
  • a bi-prediction block is generated (S3105).
  • a prediction error evaluation value is calculated from the prediction error and the motion information code amount of the motion compensated prediction block and the prediction target block (S3107), and when the prediction error evaluation value is the minimum value, the evaluation value is updated.
  • the prediction error minimum index is updated (S3108).
  • the selected prediction error minimum index is output together with the prediction error minimum value and the motion compensated prediction block as a combined motion information index used in the combined prediction mode. (S3109), the combined prediction mode evaluation value generation process is terminated.
  • FIG. 33 is a flowchart for explaining the detailed operation of the prediction mode evaluation value generation processing in step S1703 of FIG.
  • FIG. 34 shows a syntax regarding motion information of a prediction block.
  • merge_flag indicates whether or not the mode is a joint prediction mode, and when merge_flag is 0, the motion detection prediction mode is indicated.
  • a flag inter_pred_flag indicating whether the prediction type is uni-prediction or bi-prediction is transmitted.
  • bi_pred_flag is transmitted without prohibiting bi-prediction.
  • conditional branch is necessary for entropy coding / decoding when switching whether to transmit inter_pred_flag depending on whether the size of the prediction block is less than or equal to the bi-prediction restricted block size. This is to prevent this from happening.
  • the reference image list (LX) to be processed is set as the reference image list used for prediction (S3301). If it is not uni-prediction, it is bi-prediction, so LX is set to L0 in this case (S3302).
  • reference image designation information (index) and motion vector values for LX prediction are acquired (S3303).
  • a prediction vector candidate list is generated (S3304), an optimal prediction vector is selected from the prediction vectors, and a difference vector is generated (S3305). It is desirable to select the optimal prediction vector with the least amount of code when the difference vector between the prediction vector and the motion vector to be transmitted is actually encoded. However, the horizontal and vertical components of the difference vector are simply selected. The calculation may be simplified by a method such as selecting one having a small absolute sum.
  • step S3306 it is determined again whether or not the prediction mode is single prediction (S3306). If the prediction mode is single prediction, the process proceeds to step S3311. If it is not uni-prediction, that is, if it is bi-prediction, it is determined whether or not the reference list LX to be processed is L1 (S3307). If the reference list LX is L1, the process proceeds to step S3311, and if it is not L1, that is, if it is L0, if the predicted block size is equal to or smaller than bipred_restriction_size (S3308: YES), information for L1 prediction is not calculated, The prediction mode is converted to single prediction (S3310), and the process proceeds to step S3311.
  • LX is set to L1 (S3309), and the same processing as the processing from step S3303 to step S3306 is performed.
  • bi-prediction is performed with the target prediction block size when bi-prediction restriction is performed on the prediction block size.
  • the process of steps S3308 and S3310 is used to limit the bi-prediction in the prediction mode evaluation value generation process.
  • motion vector information used in single prediction and motion vector information of single prediction generated by restricting bi-prediction in the above step are Since there may be different cases, by registering a new motion information candidate for uni-prediction, it is possible to improve the encoding efficiency as compared with the case where the motion information for bi-prediction is simply not used.
  • a motion information code amount is calculated (S3311).
  • the motion information to be encoded includes three elements of reference image designation information, a difference vector value, and a prediction vector index for one reference image.
  • L0 and L1 The reference image designation information, the difference vector value, and the prediction vector index for the two reference images are a total of six elements, and the total amount of the encoded amount is calculated as the motion information code amount.
  • a prediction vector index code string generation method a Truncated Unary code string is used in the same manner as the combined motion information index code string.
  • reference image designation information and a motion vector for the reference image are set in the motion compensated prediction unit 112 in FIG. 1 to generate a motion compensated prediction block (S3312).
  • a prediction error evaluation value is calculated from the prediction error and the motion information code amount of the motion compensated prediction block and the prediction target block (S3313), the prediction error evaluation value, and reference image designation information that is motion information for the reference image;
  • the difference vector value and the prediction vector index are output together with the motion compensated prediction block (S3314), and the prediction mode evaluation value generation process ends.
  • the above processing is the detailed operation of the motion compensated prediction block structure selection unit 113 in the video encoding apparatus in the first embodiment.
  • Embodiment 1 of the present invention an example of syntax that is transmitted in order to recognize the inter_4x4_enable and inter_bipred_restriction_idc shown in FIG. 10, which are control parameters for limiting the memory access amount in motion compensation prediction, in the decoding apparatus Is shown in FIG. 10, which are control parameters for limiting the memory access amount in motion compensation prediction, in the decoding apparatus Is shown in FIG.
  • control parameter values shown in FIG. 10 are transmitted as they are as part of the header information set for each sequence or image.
  • it is transmitted inside seq_parameter_set_rbsp () that transmits parameters in sequence units, and the information for the minimum CU size shown in FIG. 3 is defined by a power of 2 based on 8 (indicating 8 ⁇ 8) in log2_min_coding_block_size_minus3
  • the maximum CU size (encoded block size in the first embodiment) is transmitted as log2_diff_max_min_coding_block_size having a value indicating the maximum number of CU divisions (Max_CU_Depth).
  • inter_4x4_enable is transmitted only when log2_min_coding_block_size_minus3 is 0, that is, when the minimum CU size is 8 ⁇ 8, as inter_4x4_enable_flag, and by sending control parameters only when the control by inter_4x4_enable is valid, transmission of invalid control information Can be prevented.
  • inter_bipred_restriction_idc is necessary for control even when the minimum CU size is 16 ⁇ 16, a configuration in which it is always transmitted is adopted.
  • control parameter values are encoded and transmitted using parameters in sequence units.
  • the configuration of the first embodiment is not limited to the control parameter configuration in sequence units, and the decoding apparatus can acquire the control parameters in predetermined units.
  • FIG. 36 is a diagram showing a detailed configuration of the motion information decoding unit 1111 in the video decoding device according to Embodiment 1 shown in FIG.
  • the motion information decoding unit 1111 includes a motion information bitstream decoding unit 3600, a prediction vector calculation unit 3601, a vector addition unit 3602, a motion compensation prediction decoding unit 3603, a combined motion information calculation unit 3604, a combined motion information single prediction conversion unit 3605, and A combined motion compensated prediction decoding unit 3606 is included.
  • a bit stream related to motion information input from the prediction mode / block structure decoding unit 1108 is supplied to the motion information bit stream decoding unit 3600 and input from the prediction mode information memory 1112 to the motion information decoding unit 1111 in FIG.
  • the obtained motion information is supplied to the prediction vector calculation unit 3601 and the combined motion information calculation unit 3604.
  • reference image designation information and motion vectors used for motion compensation prediction are output from the motion compensation prediction decoding unit 3603 and the joint motion compensation prediction decoding unit 3606 to the motion information decoding unit 1111 and information indicating the prediction type is output.
  • the included decoded motion information is supplied to the motion compensation prediction unit 1114 and the prediction mode information memory 1112.
  • the motion information bitstream decoding unit 3600 decodes the input motion information bitstream according to the encoding syntax, thereby generating the transmitted prediction mode and motion information corresponding to the prediction mode.
  • the combined motion information index is supplied to the combined motion compensated prediction decoding unit 3606, the reference image designation information is supplied to the prediction vector calculation unit 3601, and the prediction vector index is supplied to the vector addition unit 3602.
  • the difference vector value is supplied to the vector addition unit 3602.
  • the prediction vector calculation unit 3601 applies a motion compensation prediction target reference image based on the motion information of adjacent blocks supplied from the prediction mode information memory 1112 and the reference image designation information supplied from the motion information bitstream decoding unit 3600.
  • a prediction vector candidate list is generated and supplied to the vector addition unit 3602 together with the reference image designation information.
  • the same operation as the prediction vector calculation unit 1502 of FIG. 15 in the moving image encoding apparatus is performed, and the same candidate list as the prediction vector candidate list at the time of encoding is generated.
  • the vector addition unit 3602 indicates a prediction vector index from the prediction vector candidate list and reference image designation information supplied from the prediction vector calculation unit 3601, and the prediction vector index and difference vector supplied from the motion information bitstream decoding unit 3600. By adding the prediction vector value and the difference vector value registered at the set position, the motion vector value for the reference image to be motion compensated prediction is reproduced. The reproduced motion vector value is supplied to the motion compensated prediction decoding unit 3603 together with the reference image designation information.
  • the motion compensation prediction decoding unit 3603 is supplied with the reproduced motion vector value and reference image designation information for the reference image from the vector addition unit 2602, and sets the motion vector value and the reference image designation information in the motion compensation prediction unit 1114. Thus, a motion compensated prediction signal is generated.
  • the combined motion information calculation unit 3604 generates a combined motion information candidate list from the motion information of adjacent blocks supplied from the prediction mode information memory 1112 and combines the combined motion information candidate list and the combined motion information candidate that is a component in the list.
  • the reference image designation information and the motion vector value are supplied to the combined motion information single prediction conversion unit 3605.
  • the same operation as the combined motion information calculation unit 1506 in FIG. 15 in the moving image encoding apparatus is performed, and the same candidate list as the combined motion information candidate list at the time of encoding is generated. Is done.
  • the combined motion information single prediction conversion unit 3605 performs the same operation as the combined motion information single prediction conversion unit 1507 of FIG. 15 in the moving image encoding device, and is supplied from the combined motion information calculation unit 3604. And, for the motion information registered in the candidate list, the motion information whose prediction type is bi-prediction is converted into motion information of uni-prediction according to the bi-prediction restriction information shown in FIG. 3606.
  • the combined motion compensated prediction decoding unit 3606 includes a combined motion information candidate list supplied from the combined motion information single prediction conversion unit 3605, reference image designation information of a combined motion information candidate that is a component in the list, a motion vector value, and motion. Based on the combined motion information index supplied from the information bitstream decoding unit 3600, the reference image designation information and the motion vector value in the combined motion information candidate list indicated by the combined motion information index are reproduced and set in the motion compensation prediction unit 1114. Thus, a motion compensated prediction signal is generated.
  • FIG. 37 is a flowchart for explaining the detailed operation of the prediction block unit decoding processing in steps S1402, S1405, S1408, and S1410 of FIG.
  • an encoded stream of a CU unit is acquired (S3700), and for each prediction block size obtained by performing PU division on the target CU based on NumPart set according to the prediction block size division mode (PU) in the CU (S3701). Steps S3702 to S3706 are executed (S3707).
  • the encoded sequence of motion information separated from the encoded stream of the CU unit is supplied to the motion information decoding unit 1111 from the prediction mode / block structure decoding unit 1108 of FIG. 11 and is supplied from the prediction mode information memory 1112.
  • the motion information of the decoding target block is decoded using the group motion information (S3702). Details of the processing in step S3702 will be described later.
  • the separated coded sequence of prediction error information is supplied to the prediction difference information decoding unit 1102 and decoded as a quantized prediction error signal, and the inverse quantization / inverse transformation unit 1103 performs inverse quantization, inverse orthogonal transform, etc.
  • a decoded prediction error signal is generated (S3703).
  • the motion information decoding unit 1111 supplies the motion information of the decoding target block to the motion compensation prediction unit 1114, and the motion compensation prediction unit 1114 performs motion compensation prediction according to the motion information and calculates a prediction signal (S3704).
  • the adder 1104 supplies the decoded prediction error signal supplied from the inverse quantization / inverse transform unit 1103 and the motion compensation prediction unit 1114 to the prediction mode / block structure selection unit 1109, and further selects motion compensation prediction in the prediction mode. As a result, the prediction signal supplied to the addition unit 1104 is added to generate a decoded image signal (S3705).
  • the decoded image signal supplied from the adding unit 1104 is stored in the intra-frame decoded image buffer 1105 and also supplied to the loop filter unit 1106. Also, the motion information of the decoding target block supplied from the motion information decoding unit 1111 is stored in the prediction mode information memory 1112 (S3706). This is applied to all the prediction blocks in the target CU, thereby completing the decoding process for each prediction block.
  • FIG. 38 is a flowchart for explaining the detailed operation of the motion information decoding process in step S3702 of FIG.
  • the motion information decoding process of step S3702 of FIG. 37 is performed by the motion information bitstream decoding unit 3600, the prediction vector calculation unit 3601, and the combined motion information calculation unit 3604.
  • the motion information decoding process is a process of decoding motion information from an encoded bit stream encoded with a specific syntax structure.
  • the Skip flag decoded first in the CU unit of the encoded block indicates the Skip mode (S3800: YES)
  • joint prediction motion information decoding is performed (S3801). Detailed processing in step S3801 will be described later.
  • step S3805 Detailed operation of step S3805 will be described later.
  • FIG. 39 is a flowchart for explaining the detailed operation of the joint prediction motion information decoding process in step S3801 of FIG.
  • the combined prediction mode is set as the prediction mode (S3900), and a combined motion information candidate list is generated (S3901).
  • the process of step S3901 is the same process as the combined motion information candidate list generation process of step S1701 of FIG. 17 in the video encoding device.
  • bipred_restriction_size a prediction block size that restricts bi-prediction set by the control parameter inter_bipred_restriction_idc that restricts bi-prediction shown in FIG. 10 (S3902: YES)
  • the combined motion information candidate single prediction conversion is performed in which the bi-prediction motion information in each candidate in the combined motion information candidate list is replaced with the single prediction motion information (S3903).
  • the same process as the combined motion information single prediction conversion process in the encoding apparatus shown in the flowchart of FIG. 30 is performed. If the predicted block size is not less than or equal to bipred_restriction_size (S3902: NO), the process proceeds to step S3904.
  • the motion information to be acquired includes a prediction type indicating single prediction / bi-prediction, reference image designation information, and a motion vector value.
  • step S3904 and step S3903 for restricting bi-prediction based on the prediction block size are performed after performing step S3904 and step S3905 in FIG.
  • the generated motion information is stored as motion information in the joint prediction mode (S3906), and is supplied to the joint motion compensation prediction decoding unit 3606.
  • FIG. 40 is a flowchart for explaining the detailed operation of the predicted motion information decoding process in step S3805 of FIG.
  • the prediction type is simple prediction (S4000). If it is single prediction, the reference image list (LX) to be processed is set as the reference image list used for prediction (S4001). If it is not uni-prediction, it is bi-prediction, so LX is set to L0 in this case (S4002).
  • the reference image designation information is decoded (S4003), and the difference vector value is decoded (S4004).
  • a prediction vector candidate list is generated (S4005).
  • the prediction vector index is decoded (S4007), and when the prediction vector candidate list is 1 (S4006: NO), 0 is set to the prediction vector index (S4008).
  • step S4005 processing similar to that in step S3304 in the flowchart of FIG. 33 in the video encoding device is performed.
  • the motion vector value stored at the position indicated by the prediction vector index is acquired from the prediction vector candidate list (S4009).
  • a motion vector is reproduced by adding the decoded difference vector value and motion vector value (S4010).
  • step S4011 it is determined again whether or not the prediction type is single prediction (S4011). If the prediction type is single prediction, the process proceeds to step S4014. If it is not uni-prediction, that is, if it is bi-prediction, it is determined whether or not the reference list LX to be processed is L1 (S4012). If the reference list LX is L1, the process proceeds to step S4014. If it is not L1, that is, if it is L0, the predicted block size is equal to or smaller than bipred_restrcition_size (S4013: YES), the process proceeds to step S4016, and the predicted block size is bipred_restriction_size. If larger (S4013: NO), LX is set to L1 (S4015), and the same processing as the processing from step S4003 to step S4011 is performed.
  • the generated motion information in the case of single prediction, reference image designation information and motion vector values for one reference image, and in the case of bi-prediction, reference image designation information and motion for two reference images.
  • the vector value is stored as motion information (S4014) and supplied to the motion compensated prediction decoding unit 3603.
  • the motion information transmitted at the time of encoding is decoded according to the syntax. Therefore, in the prediction mode evaluation value generation process of FIG. Although it is possible to implement the conditional branching regarding the bi-prediction restriction to ensure the restriction of the memory access amount as in the case where the condition determination in step S4013 and the process in step S4016 are omitted, In the first embodiment, the prediction motion information decoding process according to the flowchart of FIG. 40 is adopted as a configuration that ensures the limitation of the memory band even in the decoding device.
  • FIG. 41 is a seq_parameter_set_rbsp () etc. that transmits parameters in sequence as shown in FIG. 35, and a level_idc that defines the maximum image size of encoding / decoding processing or the maximum number of pixels for a predetermined time unit is transmitted.
  • a level_idc that defines the maximum image size of encoding / decoding processing or the maximum number of pixels for a predetermined time unit is transmitted.
  • the prediction block size and the bi-prediction of motion compensation prediction are linked to the maximum number of usable processing pixels. It is an example of the structure which adds a restriction
  • the memory access is limited according to the assumed image size of the encoding device / decoding device. Therefore, according to the use of the encoding device and the decoding device, it is possible to realize an encoding device and a decoding device that can secure the necessary memory bandwidth and can maintain the encoding efficiency while reducing the processing load and the scale of the device.
  • inter_4x4_enable when level_idc is set to 6 levels, inter_4x4_enable is not restricted (in the case of 0 and 1 can be set) under the condition that encoding with a small number of pixels is assumed, inter_bipred_restriction_idc All of the defined values can be set, but with the increase of level_idc, the prediction block size and bi-prediction restrictions are added step by step from the prediction process with a large memory access amount shown in FIG. , Inter_4x4_enable (always set to only 0) and inter_bipred_restriction_idc (increase the minimum value of possible values) can be controlled in conjunction with the maximum image size and the maximum number of processed pixels.
  • inter_4x4_enable and inter_bipred_restriction_idc are implicitly set to a fixed value under restriction without being transmitted in conjunction with the maximum image size and the maximum number of processed pixels with reference to level_idc.
  • control parameter prohibiting motion compensated prediction of 4 ⁇ 4 prediction block size called inter_4x4_enable is used, but the prediction block restriction of motion compensated prediction is the same as the inter_bipred_restriction_idc specified prediction block. It is also possible to use a control parameter that prohibits motion compensation prediction of a block size equal to or smaller than the size, which makes it possible to control the memory access amount more finely.
  • bi-prediction restriction is performed on the same basis when the area of the prediction block size is the same and the number of horizontal and vertical pixels is different, such as 4 ⁇ 8 pixels and 8 ⁇ 4 pixels.
  • the access unit of the reference image memory is generally composed of a plurality of pixels such as 4 pixels or 8 pixels in the horizontal direction, 4 ⁇ 8 pixels having a small number of pixels in the horizontal direction are more
  • motion block prediction and bi-prediction by defining a prediction block size with a large access amount, and it is possible to control the memory access amount more suitable for the configuration of the decoding device.
  • the division configuration of the CU into prediction blocks is non-division (2N ⁇ 2N), horizontal / vertical division (N ⁇ N), and division only in the horizontal direction.
  • (2N ⁇ N) vertical division only (N ⁇ 2N), horizontal only upper 1/4, lower 3/4 asymmetric division (2N ⁇ nU), horizontal upper only 3/4, lower 1/4 asymmetric division (2N ⁇ nD), left 1/4 to vertical only, right 3/4 asymmetric division (nL ⁇ 2N), left 3/4 to vertical only,
  • a split configuration is applicable.
  • FIG. 43 shows an example of the block size of motion compensation prediction and control parameters for limiting the prediction processing in the prediction block configuration of FIG.
  • the control parameter is a parameter for controlling validity / invalidity of motion compensated prediction of 4 ⁇ 4, 4 ⁇ 8, and 8 ⁇ 4 prediction blocks, which is a configuration that divides an 8 ⁇ 8 block that is the smallest CU size, and inter_pred_enable_idc And inter_bipred_restriction_idc that defines a block size that prohibits only the prediction processing in which bi-prediction is performed in motion compensation prediction.
  • the order of the size of the prediction block size of 16 ⁇ 16 pixels or less which takes into account the influence on the memory access of the horizontal and vertical number of pixels with respect to inter_bipred_restriction_idc, ⁇ 4, 4 ⁇ 8, 8 ⁇ 4, 8 ⁇ 8, 4 ⁇ 16/12 ⁇ 16 (nL ⁇ 2N / nR ⁇ 2N), 8 ⁇ 16, 16 ⁇ 12/16 ⁇ 4 (2N ⁇ nU / 2N ⁇ nD), 16 ⁇ 8, and 16 ⁇ 16, and sets a prediction block size value that restricts bi-prediction.
  • the memory access amount can be controlled in a fine unit even for a prediction block having an asymmetric configuration in which the efficiency of motion compensation prediction is improved, as in the configuration using the control parameters shown in FIG.
  • the efficiency of motion compensation prediction is improved, and the memory access amount can be controlled according to the allowable memory bandwidth.
  • bi-prediction restriction is applied to a prediction block having a size equal to or smaller than the defined size on the basis of the prediction block size defined by inter_bipred_restriction_idc.
  • Limiting bi-prediction to a block is also possible, and bi-prediction is limited to a prediction block of a defined size if motion compensated prediction is not performed at a prediction block size that is smaller than the prediction block size to which bi-prediction restriction is applied. It is also possible to add the above as a configuration for realizing the present invention.
  • step S3902 shown in the flowchart of FIG. 39 and step S4013 shown in the flowchart of FIG. 40 is less than bipred_restriction_size, and the prediction block size defined by inter_bipred_restriction_idc This is realized by setting the value as a prediction block size one larger.
  • inter_4x4_enable and inter_bipred_restriction_idc which are control parameters for limiting the memory access amount in motion compensation prediction, are encoded and transmitted as individual parameters, respectively.
  • the control parameter information can be transmitted as a parameter for controlling the memory access amount restriction of the video encoding device and the video decoding device
  • a configuration in which the information to be defined (inter_mc_restrcution_idc) is encoded and transmitted is also possible.
  • the first embodiment as a means for prohibiting bi-prediction used for joint motion compensation prediction to limit the memory access amount with respect to motion compensation prediction, after being stored in the joint motion information candidate index Motion information is converted from bi-prediction motion information to uni-prediction motion information according to conditions, stored, and used for prediction processing. As a result, the prediction accuracy of motion compensated prediction in the prediction block size under the condition prohibiting bi-prediction is improved, and the coding efficiency is improved.
  • the configuration for limiting the maximum memory access amount is the same by combining the motion compensation prediction limitation based on the prediction block size and the bi-prediction limitation equal to or less than the prediction block size.
  • the configuration for limiting the maximum memory access amount is the same by combining the motion compensation prediction limitation based on the prediction block size and the bi-prediction limitation equal to or less than the prediction block size.
  • FIG. 45 shows an example of the motion compensated prediction block size and control parameters for limiting the prediction processing in Embodiment 2 of the present invention.
  • the control parameters are inter_4x4_enable, which is a parameter for controlling the validity / invalidity of motion compensated prediction of 4 ⁇ 4 pixels, which is the smallest motion compensated prediction block size, and only prediction processing for which bi-prediction is performed among motion compensated predictions It consists of two parameters of inter_bipred_restriction_for_mincb_idc that define the CU partition structure in the minimum CU size that prohibits.
  • Inter_bipred_restriction_for_mincb_idc defines four values, and controls four states: no limit, N ⁇ N limit, N ⁇ 2N / 2N ⁇ N limit or less, and all divisions (PUs) in the CU.
  • the minimum CU size is defined as a power of 2 with log2_min_coding_block_size_minus3 as a reference (indicating 8 ⁇ 8) as shown in the syntax of FIG. 35 in Embodiment 1, and the value of inter_bipred_restriction_for_mincb_idc and the minimum CU size As a result, the block size bipred_restriction_size for limiting bi-prediction is set.
  • bipred_restriction_size in the first embodiment is defined by a combination of the above log2_min_coding_block_size_minus3 and inter_bipred_restriction_for_mincb_idc. , Has a different configuration.
  • a specific definition of bipred_restriction_size is shown in FIG.
  • inter_bipred_restriction_for_mincb_idc is configured in the same way as the syntax of FIG. 35 in Embodiment 1, and is transmitted as a sequence unit parameter by seq_parameter_set_rbsp (), and instead of inter_bipred_restriction_idc Is the value to be transmitted.
  • the configuration for limiting the bi-prediction in conjunction with the minimum CU size is managed and transmitted.
  • the bi-prediction limitation at a larger size can be defined with a small control parameter value.
  • the size restriction for each block size is set for each CU. Even if it is not added in the hierarchy, it is sufficient to add only the definition in the minimum CU size. Therefore, the expandability is high, and the encoding block size of the high-definition image that finishes high-definition is large. It has an effect that can easily realize the restriction of sheath prediction.
  • Embodiment 3 Next, the video encoding device and video decoding device according to Embodiment 3 of the present invention will be described.
  • Embodiment 3 in addition to motion compensation prediction and bi-prediction limitations for limiting the memory access amount, by limiting the number of operations of combined motion prediction candidate generation processing when the prediction block size is reduced, The configuration is such that the processing load required for generating the combined motion prediction candidate is reduced.
  • the same combined motion information candidate generation process is performed using the motion information of the same adjacent block in each prediction block in a prediction block size equal to or smaller than a predetermined CU size.
  • the configuration having the above configuration is adopted for the prediction block of 8 ⁇ 8 CU size that is the minimum CU size, and the spatial peripheral prediction block in the combined motion information candidate generation of 8 ⁇ 8 CU size of Embodiment 3 Will be described with reference to FIG.
  • the positions of the five blocks of the block A0, the block A1, the block B0, the block B1, and the block B2 of the spatial candidate block group for the prediction block (2N ⁇ 2N) of 8 ⁇ 8 pixels are shown in FIG. As shown, the same position as the definition of the space candidate block group in the first embodiment shown in FIG. 19 is shown.
  • the same combined motion information candidate is used in all configured prediction block structures, and the combined motion information generation process in the encoding device and the decoding device is performed once. It can be realized by the generation process.
  • Embodiment 3 an encoding process for each encoding block of the moving image encoding apparatus according to Embodiment 3 will be described.
  • FIG. 49 shows a flowchart of motion compensation prediction block size selection / prediction signal generation processing in the third embodiment.
  • the same numbers are assigned and new step numbers are assigned only to different portions.
  • an encoded block image to be predicted is acquired for the target CU (S700).
  • the CU size of the target CU is 8 ⁇ 8 (S4908: YES)
  • combined motion information candidate list generation processing is performed (S4909). If the CU size of the target CU is not 8 ⁇ 8 (S4908: NO), the process proceeds to step S701. Regarding the details of step S4909, the same process as the combined motion information candidate list generation process of FIG. 18 in the first embodiment is performed.
  • step S4909 when the minimum prediction block size in the target CU is equal to or smaller than bipred_restriction_size (S4910: YES), combined motion information candidate single prediction conversion processing is performed (S4911). If the minimum predicted block size in the target CU is not less than or equal to bipred_restriction_size (S4910: NO), the process proceeds to step S701. Regarding the details of step S4911, the same process as the combined motion information candidate single prediction conversion process of FIG. 30 in the first embodiment is performed.
  • Embodiment 3 when the combined motion information candidate generation process in bipred_restriction_size, which is a prediction block size that restricts bi-prediction, is the prediction block size used in the target CU (when inter_4x4_enable is 1, 4 * 4/4 * 8/8 * 4/8 * 8 prediction block, and when inter_4x4_enable is 0, 4 * 8/8 * 4/8 * 8 prediction block) is the same for the target CU
  • the bi-prediction motion information is converted into a single prediction for the combined motion information candidate list generated at the same time. That is, processing in which bipred_restriction_size is expanded to 3 (8 ⁇ 8 or less restriction) is performed.
  • step S701 After performing the combined motion information candidate single prediction conversion process of step S4911, the process proceeds to step S701.
  • the processing from step S701 to step S707 is the same as the processing from step S701 to step S707 in the flowchart of FIG. 7 in the first embodiment.
  • the combined motion information candidate list generation process and the combined motion information candidate uni-prediction conversion process for the 8 ⁇ 8 CU size are performed by the same operation, and the encoding device performs 8 generations by one generation process. There is an effect that it becomes possible to generate all combined motion information candidates within the ⁇ 8 CU size.
  • the combined motion information candidate list generation process is performed with the same operation for the 8 ⁇ 8 CU size, and bipred_restriction_size is expanded.
  • the combined motion information candidate uni-prediction conversion process in a state where the prediction is not performed can be performed, but the encoding apparatus needs the combined motion information candidate uni-prediction conversion process for each prediction block size within the 8 ⁇ 8 CU size.
  • FIG. 50 shows a flowchart of the motion compensation prediction mode / prediction signal generation processing in the third embodiment.
  • the same numbers are assigned and new step numbers are assigned only to different portions.
  • Step S1701 to Step S1708 are executed (S1709).
  • the same processing as the flowchart of FIG. 17 in the first embodiment is performed.
  • step S5010 If the target CU size is 8 ⁇ 8 (S5010: YES), the process proceeds from step S1701 to step S1703 and proceeds to step S1704. That is, in the case of the prediction block size where the target CU size is 8 ⁇ 8, the combined motion information candidate generated by the process in the flowchart of the motion compensation prediction block size selection / prediction signal generation process shown in FIG. 49 is selected. It is configured to perform motion compensation prediction in the combined prediction mode by using it as it is.
  • the decoding process in units of coding blocks of the video decoding device in the third embodiment performs the same process as in the first embodiment, and the candidate blocks used for generating the combined motion information candidate list in the combined motion prediction.
  • the candidate blocks at the same position are obtained in all the prediction blocks as shown in FIG. 48, and in the combined motion information decoding process shown in the flowchart of FIG.
  • the CU size is 8 ⁇ 8
  • the combined motion information candidate list for the 8 ⁇ 8 CU size is provided. This has the effect of enabling the combined motion information candidate single prediction conversion processing in a state where the generation processing is performed with the same operation and bipred_restriction_size is not expanded.
  • the decoding apparatus since the prediction block size for the decoding target block is specified by decoding the encoded stream, a single combined motion information candidate single prediction conversion process is performed for the specified prediction block size.
  • the combined motion information candidate generation process can be realized with fewer processes. It is possible to replace the motion information decoding process with the process of the flowchart shown in FIG. 51, and the operation will be described. With respect to the same steps as those in the flowchart of FIG. 39, the same numbers are assigned and new step numbers are assigned only to different portions.
  • the combined prediction mode is set as the prediction mode (S3900)
  • step S5109 as shown in FIG. 48, the same processing as that in step S3901 is performed with a configuration in which candidate blocks at the same position are acquired for all prediction blocks in the CU.
  • step S5109 it is determined whether or not the minimum predicted block size definable in the CU is equal to or smaller than bipred_restriction_size (S5110). If the minimum predicted block size is equal to or smaller than bipred_restriction_size (S5110: YES). Then, the combined motion information candidate single prediction conversion process is performed (S3903), and if the minimum prediction block size is larger than bipred_restriction_size (S5110: NO), the process proceeds to step S3904.
  • step S3904 to step S3906 the same processing as the processing in the flowchart of FIG. 39 in the first embodiment is performed, and the motion information in the joint prediction mode is decoded and stored.
  • the moving picture coding apparatus and the moving picture decoding apparatus in the third embodiment combined motion prediction candidate generation when the motion compensation prediction for limiting the memory access amount, the restriction of bi-prediction, and the prediction block size are reduced It is possible to realize processing reduction with a configuration that is consistent with each restriction, and to improve the coding efficiency while simultaneously reducing the memory bandwidth and reducing the combined motion information candidate generation process.
  • the unit constituting the same combined motion information candidate list in Embodiment 3 has been described as an 8 ⁇ 8 size, but is not limited to the 8 ⁇ 8 size, and is a predetermined unit such as a picture unit or a sequence unit.
  • the unit can be changed by transmitting parameter information defining the maximum predicted block size for generating the same list.
  • log2_parallel_merge_level_minus2 can be defined as a value corresponding to a power of 2 that serves as a reference for the horizontal / vertical size of the predicted block size for generating the same list.
  • Embodiment 4 Next, the video encoding device and video decoding device according to Embodiment 4 of the present invention will be described.
  • the fourth embodiment as in the third embodiment, in addition to motion compensation prediction and bi-prediction restriction for restricting the memory access amount, combined motion prediction candidate generation processing when the prediction block size is reduced By limiting the number of operations, the processing load required for generating the combined motion prediction candidate is reduced.
  • the combined motion information single prediction is performed in the motion compensated prediction block structure selection unit 113 shown in FIG. 15 in contrast to the moving picture coding apparatus shown in the first embodiment.
  • the conversion unit 1507 is eliminated, and the motion vector, the reference image designation information, and the combined motion information candidate list output from the combined motion information calculation unit 1506 are directly supplied to the combined motion compensation prediction generation unit 1508.
  • the combined motion information single prediction conversion unit 3605 in the motion information decoding unit 1111 shown in FIG. 36 is compared with the video decoding device shown in the first embodiment.
  • the motion vector, the reference image designation information, and the combined motion information candidate list output from the combined motion information calculation unit 3604 are directly supplied to the combined motion compensation prediction decoding unit 3606.
  • bi-prediction is performed when the prediction block size is equal to or less than bipred_restriction_size during motion compensation prediction.
  • step S1702 and step S1703 are eliminated in the motion compensation prediction mode / predicted signal generation process shown in the flowchart of FIG. 17 in the first embodiment.
  • a process for limiting to single prediction is performed.
  • the motion compensated prediction block generation operation performed in steps S3105 and S3106 of the flowchart of FIG. 31 and step S3312 of the flowchart of FIG. 33 is shown in the flowchart of FIG.
  • the flowchart of FIG. 52 is the detailed operation of the motion compensation prediction unit 112 in the moving picture encoding apparatus shown in FIG. 1 in the fourth embodiment, and performs the following operation.
  • a motion compensated single prediction block is generated using reference image designation information and a motion vector for one reference image (S5203).
  • the supplied motion information is not uni-prediction, that is, if the motion information is bi-prediction (S5200: NO), whether the motion information for L0 prediction and the motion information for L1 prediction (reference image information and motion vector) are the same. If the motion information of the L0 prediction and the motion information of the L1 prediction are the same (S5201: YES), the L0 single prediction motion compensation prediction is performed using only the motion information of the L0 prediction (S5204). However, the motion information of bi-prediction is maintained and the motion information of L1 prediction is not changed.
  • the prediction block size is equal to or smaller than bipred_restriction_size, and when the predicted block size is equal to or smaller than bipred_restriction_size (S5202: If the motion information for L0 prediction and the motion information for L1 prediction are the same (S5201: YES), L0 single prediction motion compensation prediction is performed using only the motion information for L0 prediction (S5204). However, the motion information of bi-prediction is maintained and the motion information of L1 prediction is not changed.
  • the purpose of the bi-prediction restriction is to limit the memory band of motion compensation prediction by restricting the bi-prediction to the single prediction. Therefore, the prediction list (L0 / L1) restricted by the bi-prediction restriction is set to the L1 single prediction. May be.
  • a motion-compensated bi-prediction block is generated using reference image designation information and motion vectors for two reference images (S5205).
  • the processes in steps S3902 and S3903 are eliminated, and the prediction shown in the flowchart of FIG.
  • the process for limiting to single prediction is performed by the process shown in the flowchart of FIG. 52, similarly to the encoding process.
  • one of the L0 prediction and the L1 prediction is included in the motion information of the bi-prediction at the time of motion compensation prediction without using the configuration for converting the combined motion information candidate list into the single prediction for the restriction process of the bi-prediction.
  • the prediction information is the same as the single prediction, but the motion information can maintain a combined motion information candidate that is bi-predicted.
  • the motion information is stored for both L0 prediction and L1 prediction, so that information of bi-prediction is used as it is as adjacent reference motion information of a prediction block to be encoded and decoded thereafter.
  • the same motion information is used as the combined motion information candidate list, and the prediction block size is small because the memory access amount can be limited by the bi-prediction restriction at the time of motion compensation prediction in the prediction block sizes of different sizes.
  • the combined motion information candidate list is obtained by adopting the configuration of the fourth embodiment.
  • bi-prediction restriction in the configuration in which bi-prediction restriction is performed at the time of motion compensation prediction, both the bi-prediction restrictions of two prediction modes (joint prediction mode and motion detection prediction mode) for encoding motion information can be handled in a lump.
  • Bi-prediction restriction can be realized with a minimum configuration.
  • Embodiment 5 Next, the video encoding device and video decoding device according to Embodiment 5 of the present invention will be described.
  • the motion compensation prediction restriction based on the prediction block size and the bi-predictive motion compensation restriction for restricting the memory access amount are performed.
  • the bi-prediction to uni-prediction conversion method for motion information in the combined motion information candidate list is different.
  • the same configuration and processing as in the first embodiment are performed, but the combined motion information candidate list generation processing shown in the flowchart of FIG. 18 and the flowchart of FIG. 30 in the first embodiment.
  • the combined motion information candidate single prediction conversion process is different.
  • step S1701 in the flowchart of FIG. 17 for the encoding process
  • step S3901 in the flowchart of FIG. 39 for the decoding process.
  • the same numbers are assigned and new step numbers are assigned only to different portions.
  • step S1800 to step S1802 spatially combined motion information candidates and temporally combined motion information candidates obtained by deleting the same information from the spatial candidate block group, which are candidates for combined motion information, are calculated, and candidate blocks The combined motion information calculated from the motion information is generated.
  • num_list_before_combined_merge which is the number of combined motion information generated up to step S1802, is stored (S5305). This value is used in the combined motion information candidate single prediction conversion process described later.
  • the first combined motion information candidate and the combined motion information candidate list generated by combining the motion information of a plurality of combined motion information candidates registered in the combined motion information candidate list by the processing from step S1803 to step S1804.
  • the second combined motion information candidate generated without depending on the motion information registered in the above is added as necessary, and the combined motion information candidate list generation process is terminated.
  • the processing different from the first embodiment in the combined motion information candidate list generation processing in the fifth embodiment is a storage process of num_list_before_combined_merge, and the combined motion information in which the motion information of a candidate block group defined by adjacent blocks is registered.
  • the boundary list number of the combined motion information in which the motion information combination of the candidate block group and the motion information not depending on the motion information of the candidate block is registered is stored.
  • the processing shown in FIG. 54 is performed in step S1703 in the flowchart of FIG. 17 for the encoding process and in step S3903 in the flowchart of FIG. 39 for the decoding process.
  • the same numbers are assigned and new step numbers are assigned only to different portions.
  • the combined motion information candidate single prediction conversion process shown in the flowchart of FIG. 52 differs from the flowchart of FIG. 30 in the case where the motion information is not a single prediction (S3002: NO), and the index i of the combined motion information candidate list is When smaller than num_list_before_combined_merge (S5407: YES), the L1 information of the motion information stored in the index i is invalidated in order to convert the motion information of bi-prediction into single prediction (S3003).
  • the index i is greater than or equal to num_list_before_combined_merge (S5407: NO)
  • the L0 information of the motion information stored in the index i is invalidated in order to convert the motion information of bi-prediction into single prediction (S5408).
  • candidate motion information in the combined motion information candidate list includes motion information calculated from motion information of adjacent candidate blocks and a plurality of registered motion information.
  • the motion information to be invalidated at the time of single prediction conversion is switched according to the prediction type (L0 prediction / L1 prediction).
  • L0 prediction and L1 prediction are used as candidates without being biased. Therefore, even in motion information stored as motion information used at the time of encoding / decoding, L0 prediction and L1 Forecast bias is reduced. Therefore, the accuracy of the bi-prediction motion information that can be generated by the first combined motion information candidate list adding unit at the time of generating the combined motion information candidate of the subsequent prediction block can be improved, and the encoding efficiency can be improved.
  • the L1 information is invalidated when the index i is smaller than num_list_before_combined_merge, and the L0 information is invalidated when the index i is equal to or larger than num_list_before_combined_merge.
  • a feature of this embodiment is that the L0 information is invalidated when the index i is smaller than num_list_before_combined_merge, and the L1 information is invalidated when the index i is greater than or equal to num_list_before_combined_merge.
  • Embodiment 6 Next, the video encoding device and video decoding device according to Embodiment 6 of the present invention will be described.
  • the feature of the sixth embodiment is that it has the same configuration as that of the fifth embodiment, and switches the prediction type (L0 prediction / L1 prediction) to be invalidated in the combined motion information candidate single prediction conversion. It takes a configuration that switches based on the position.
  • the same configuration and processing as in the fifth embodiment are performed, but the combined motion information candidate list generation processing shown in the flowchart of FIG. 53 in the fifth embodiment is not performed, and the first embodiment is performed.
  • the combined motion information candidate list generation process shown in the flowchart of FIG. 18 is performed.
  • the combined motion information candidate single prediction conversion process shown in the flowchart of FIG. 54 in the fifth embodiment is replaced with the process shown in the flowchart of FIG.
  • the process shown in FIG. 55 is performed in step S1703 in the flowchart of FIG. 17 for the encoding process and in step S3903 in the flowchart of FIG. 39 for the decoding process.
  • the combined motion information candidate single prediction conversion process shown in the flowchart of FIG. 55 is different from the flowchart of FIG. 54 in the case where the motion information is not a single prediction (S3002: NO), and the index i of the combined motion information candidate list is If it is smaller than 2 (S5507: YES), the L1 information of the motion information stored in the index i is invalidated in order to convert the motion information of bi-prediction into single prediction (S3003).
  • the index i is 2 or more (S5507: NO)
  • the L0 information of the motion information stored in the index i is invalidated in order to convert the motion information of bi-prediction into single prediction (S5408). .
  • the candidate motion information in the combined motion information candidate list is added to the first combined motion information candidate list by motion information calculated from the motion information of adjacent candidate blocks.
  • Two motion information which are the minimum motion information necessary for generating additional motion information for bi-prediction, and a first combined motion information candidate list adding unit and a second combined motion registered in the latter half of the list
  • the prediction type L0 prediction / L1 prediction
  • the combined motion information candidate uni-predictive conversion in the sixth embodiment combined motion information in which the motion information of the candidate block group defined by the adjacent block is registered and the motion information of the candidate block group are compared to the fifth embodiment. Since the processing for saving the boundary list number of the combined motion information in which the motion information that does not depend on the motion information of the candidate block and the candidate block is registered can be eliminated, the processing load can be reduced and the same as in the fifth embodiment.
  • the motion information added by the first combined motion information candidate list adding unit the motion information of the prediction type enabled at the time of the single prediction conversion is left, leaving the motion information of the prediction type disabled at the time of the single prediction conversion. It is possible to invalidate the information, and it is possible to leave a lot of effective motion information as the combined motion information, thereby improving the encoding efficiency.
  • Embodiment 6 since the prediction type that is invalidated not only for the first combined motion information candidate and the second combined motion information candidate but also for the spatial prediction candidate and the temporal prediction candidate can be switched, the prediction type However, when the same motion information is registered in bi-prediction, since the motion information of L0 uni-prediction and L1 uni-prediction can be used as combined motion information, the encoding efficiency can be improved.
  • the position of the index for switching the prediction type (L0 prediction / L1 prediction) to be invalidated is fixed to 2, but the prediction type is switched with a fixed index.
  • the features in Embodiment 6 and the number of pieces of motion information that can be registered as a spatially combined motion information candidate, a temporally combined motion information candidate, a first combined motion information candidate, and a second combined motion information candidate, and can be registered at the maximum It is also possible to set the index value of the switching position to be fixed according to the number of combined motion information candidates.
  • the moving image encoded stream output from the moving image encoding apparatus of the embodiment described above has a specific data format so that it can be decoded according to the encoding method used in the embodiment. Therefore, the moving picture decoding apparatus corresponding to the moving picture encoding apparatus can decode the encoded stream of this specific data format.
  • the encoded stream When a wired or wireless network is used to exchange an encoded stream between a moving image encoding device and a moving image decoding device, the encoded stream is converted into a data format suitable for the transmission form of the communication path. It may be transmitted.
  • a video transmission apparatus that converts the encoded stream output from the video encoding apparatus into encoded data in a data format suitable for the transmission form of the communication channel and transmits the encoded data to the network, and receives the encoded data from the network Then, a moving image receiving apparatus that restores the encoded stream and supplies the encoded stream to the moving image decoding apparatus is provided.
  • the moving image transmitting apparatus is a memory that buffers the encoded stream output from the moving image encoding apparatus, a packet processing unit that packetizes the encoded stream, and transmission that transmits the packetized encoded data via the network.
  • the moving image receiving apparatus generates a coded stream by packetizing the received data, a receiving unit that receives the packetized coded data via a network, a memory that buffers the received coded data, and packet processing. And a packet processing unit provided to the video decoding device.
  • the above-described processing related to encoding and decoding can be realized as a transmission, storage, and reception device using hardware, and is stored in a ROM (Read Only Memory), a flash memory, or the like. It can also be realized by firmware or software such as a computer.
  • the firmware program and software program can be recorded on a computer-readable recording medium, provided from a server through a wired or wireless network, or provided as a data broadcast of terrestrial or satellite digital broadcasting Is also possible.
  • the present invention can be used for encoding and decoding techniques of moving image signals.

Landscapes

  • Engineering & Computer Science (AREA)
  • Multimedia (AREA)
  • Signal Processing (AREA)
  • Compression Or Coding Systems Of Tv Signals (AREA)

Abstract

A motion compensation predicting unit (112) generates a predicted signal, for a predicted block to be encoded, by means of motion compensation using derived motion information. An encoding block control parameter generating unit (122) generates a first control parameter indicating whether to permit motion compensation prediction for a predicted block size of a first size, as well as a second control parameter indicating a second size that forbids bi-predictive motion compensation for a predicted block size less than or equal to the second size. A block structure/prediction mode information supplementary information encoding unit (118) encodes information used in motion compensation prediction, including the first and second control parameters. The motion compensation prediction unit (112) performs motion compensation prediction on the basis of the first and second control parameters.

Description

動画像符号化装置、動画像符号化方法、動画像符号化プログラム、送信装置、送信方法及び送信プログラム、並びに動画像復号装置、動画像復号方法、動画像復号プログラム、受信装置、受信方法及び受信プログラムVideo encoding device, video encoding method, video encoding program, transmission device, transmission method and transmission program, and video decoding device, video decoding method, video decoding program, reception device, reception method and reception program
 本発明は、動画像信号の符号化及び復号技術に関し、特に動き補償予測に利用する動画像符号化及び復号技術に関する。 The present invention relates to a video signal encoding and decoding technique, and more particularly to a video encoding and decoding technique used for motion compensation prediction.
 MPEG-4 AVC/H.264(以後、AVC)等に代表される動画像符号化では、時間方向の相関を利用した情報圧縮として、符号化対象となる画像信号である符号化対象ピクチャに対して、既に符号化され復号された局部復号信号を参照ピクチャとして用い、所定の符号化処理単位(以後、符号化対象ブロック)で、対象ピクチャと参照ピクチャとの間での動き量(以後、動きベクトル)を検出し、予測信号を生成する動き補償予測が用いられる。 MPEG-4 AVC / H. In moving picture coding represented by H.264 (hereinafter referred to as AVC) and the like, as information compression using correlation in the time direction, a picture to be coded which is a picture signal to be coded is already coded and decoded. The detected local decoded signal is used as a reference picture, and a motion amount (hereinafter referred to as a motion vector) between the target picture and the reference picture is detected and predicted in a predetermined encoding processing unit (hereinafter referred to as an encoding target block). Motion compensated prediction that generates a signal is used.
 AVCでは、動き補償予測において1つの参照ピクチャから1本の動きベクトルを利用して単一方向に予測信号を生成する単予測と、2つの参照ピクチャから2本の動きベクトルを利用して予測信号を生成する双予測が用いられる。これらを、符号化対象ブロックとなる16×16画素の2次元ブロック内で、予測処理対象となっているブロック(以後、予測対象ブロック)の大きさ(以後、予測ブロックサイズ)を可変にする手法や、複数の参照ピクチャの中から予測に用いる参照ピクチャを選択する手法に適用すること、また動きベクトルの精度を1/4画素精度で表現することで、予測信号の精度を向上し、伝送する差分(以後、予測誤差)の情報量を削減している。符号化側では、予測モード情報や参照画像を指定する情報を選択して動きベクトル情報と共に伝送し、復号側では、伝送された予測モード情報や参照画像を指定する情報と復号された動きベクトル情報に従って動き補償予測処理が施される。 In AVC, single prediction for generating a prediction signal in a single direction using one motion vector from one reference picture in motion compensation prediction, and a prediction signal using two motion vectors from two reference pictures Bi-prediction is used to generate A method of changing the size (hereinafter referred to as prediction block size) of a block (hereinafter referred to as prediction target block) that is a prediction processing target within a 16 × 16 pixel two-dimensional block that is an encoding target block. In addition, it is applied to a method for selecting a reference picture used for prediction from a plurality of reference pictures, and the accuracy of a motion vector is expressed with 1/4 pixel accuracy, thereby improving the accuracy of a prediction signal and transmitting The information amount of the difference (hereinafter, prediction error) is reduced. On the encoding side, information specifying the prediction mode information and the reference image is selected and transmitted together with the motion vector information. On the decoding side, the information specifying the transmitted prediction mode information and the reference image and the decoded motion vector information are transmitted. In accordance with the motion compensation prediction process.
 動きベクトルの伝送については、処理対象ブロックに隣接する符号化済みのブロックの動きベクトルを予測動きベクトル(以降、予測ベクトル)とし、処理対象ブロックの動きベクトルと予測ベクトルとの差分を求め、差分ベクトルを符号化ベクトルとして伝送することで圧縮効率を向上させている。 For motion vector transmission, a motion vector of an encoded block adjacent to the processing target block is set as a prediction motion vector (hereinafter referred to as a prediction vector), and a difference between the motion vector of the processing target block and the prediction vector is obtained. Is transmitted as an encoded vector to improve the compression efficiency.
 しかしながら、AVCにおいては、予測ブロックサイズを小さくした場合に、符号化対象ブロックの画素に対して、符号化する際に必要となる動きベクトルの数が増大し、予測誤差を符号化した際に要する符号量に対して、動きベクトルの符号化に要する符号量が増大し、予測誤差を十分な精度で符号化出来ず、符号化された画像信号の品質が低下する課題があった。 However, in AVC, when the prediction block size is reduced, the number of motion vectors required for encoding increases with respect to the pixels of the encoding target block, which is required when encoding a prediction error. There is a problem that the amount of code required for coding a motion vector increases with respect to the amount of code, the prediction error cannot be coded with sufficient accuracy, and the quality of the coded image signal is lowered.
 動きベクトルの符号化に要する符号量が増大する課題を解決するために、AVCでは予測対象ブロックと同一位置にある参照ピクチャのブロックの符号化に用いられた動きベクトルを利用して、符号化ベクトルを伝送することなく動き補償予測を実現するダイレクト動き補償予測を用いることが可能になっている。 In order to solve the problem of increasing the amount of code required for coding a motion vector, AVC uses a motion vector used for coding a block of a reference picture at the same position as a prediction target block, It is possible to use direct motion compensated prediction that realizes motion compensated prediction without transmitting.
 また、他の解決手段としては、特許文献1のように、符号化装置において予測ブロックサイズが小さい場合に、双予測を禁止し単予測のみを用いることで、符号化する動きベクトル数を少なくして、動きベクトルの符号量増大を防ぐ手法が知られている。 Another solution is to reduce the number of motion vectors to be encoded by prohibiting bi-prediction and using only uni-prediction when the prediction block size is small in the encoding device as in Patent Document 1. Thus, a technique for preventing an increase in the code amount of a motion vector is known.
WO2006/082690号公報WO2006 / 082690
 上述したダイレクト動き補償予測は、予測対象ブロックと同一位置にある参照ピクチャのブロックにおける時間方向の動きの連続性に着目し、他のブロックの動き情報をそのまま利用する。これにより、差分ベクトルを符号化ベクトルとして符号化せずに動き補償予測処理を行う。 The direct motion compensated prediction described above pays attention to the continuity of motion in the temporal direction in the block of the reference picture located at the same position as the prediction target block, and uses the motion information of other blocks as they are. Thus, the motion compensation prediction process is performed without encoding the difference vector as an encoded vector.
 しかしながら、動きの連続性が十分保たれていない場合や、他ブロックの動き情報における動きベクトルが正確な動きを示していない場合などに、他のブロックの動き情報をそのまま利用する方式を用いた際には、ずれの生じた動き情報を用いた予測画像が生成される。その場合、精度良い動き補償予測画像を生成できず、符号化効率が向上しないという難しい面がある。 However, when the continuity of motion is not sufficiently maintained or when the motion vector in the motion information of other blocks does not indicate an accurate motion, etc. In this case, a predicted image using the motion information in which the shift has occurred is generated. In that case, there is a difficult aspect that a motion compensated prediction image with high accuracy cannot be generated and the encoding efficiency is not improved.
 更に、1/4画素の動きベクトル精度で表現された動きベクトルより、予測信号を生成する際に、隣接する複数画素を用いた補間フィルタを用いて、動きベクトルにより指定された1/4画素精度の位置に対する予測画素を生成する為に、動き補償予測時に予測信号を生成するために、予測ブロックサイズに対して水平・垂直それぞれに補間フィルタのタップ数に相当する画素分の領域の参照ピクチャの画像信号を取得する必要があり、特に予測ブロックサイズを小さくした場合に、参照ピクチャのメモリアクセス量が増大する課題があり、ダイレクト動き補償予測を用いた際にも同様の課題が残る。 Furthermore, when generating a prediction signal from a motion vector expressed with a motion vector accuracy of 1/4 pixel, an interpolation filter using a plurality of adjacent pixels is used to generate a 1/4 pixel accuracy specified by the motion vector. In order to generate a prediction pixel for the position of the image, in order to generate a prediction signal at the time of motion compensation prediction, a reference picture of an area corresponding to the number of pixels corresponding to the number of taps of the interpolation filter horizontally and vertically with respect to the prediction block size It is necessary to acquire an image signal. In particular, when the prediction block size is reduced, there is a problem that the memory access amount of the reference picture increases, and the same problem remains when direct motion compensation prediction is used.
 特許文献1の手法によれば、単予測に予測手法を制限することにより、動きベクトルの数と共に、符号化装置における参照ピクチャのメモリアクセス量に関しては削減できるが、復号装置においては、符号化される動きベクトルの数に対する制限を認識できないため、実時間の復号処理を実現するためには、双予測が施される場合を想定した復号処理能力が必要となる。また、ダイレクト動き補償予測などの、符号化ベクトルを伝送しない予測手法を用いた場合に、暗黙で双予測が用いられる条件の場合には、双予測の予測信号生成が必要となり、復号装置に要求される最大メモリアクセス量を削減する事が出来ず、課題は解決しない。 According to the method of Patent Literature 1, by limiting the prediction method to single prediction, the memory access amount of the reference picture in the encoding device can be reduced together with the number of motion vectors, but in the decoding device, it is encoded. Since the restriction on the number of motion vectors to be recognized cannot be recognized, a decoding processing capability assuming a case where bi-prediction is performed is necessary to realize a real-time decoding process. In addition, when a prediction method that does not transmit an encoded vector, such as direct motion compensation prediction, is used under conditions where implicit bi-prediction is used, bi-prediction prediction signal generation is required, which is required by the decoding device. The maximum memory access amount cannot be reduced, and the problem is not solved.
 本発明はこうした状況に鑑みてなされたものであり、その目的は、動き補償予測を使用する際の参照ピクチャのメモリアクセス量を所定量以下に制限しつつ、符号化効率を向上させる技術を提供することにある。 The present invention has been made in view of such circumstances, and an object of the present invention is to provide a technique for improving the coding efficiency while limiting the memory access amount of a reference picture to a predetermined amount or less when using motion compensated prediction. There is to do.
 上記課題を解決するために、本発明のある態様の動画像符号化装置は、ピクチャが複数のブロックに段階的に分割されたブロックから予測ブロックを特定し、その特定された予測ブロック単位で、符号化ストリームを生成する動画像符号化装置であって、符号化対象となる予測ブロックに空間的に近接するブロック及び時間的に近接するブロックの少なくとも何れかから動き情報を導出し、前記符号化対象となる予測ブロックの動き情報候補として、その導出された動き情報の中から所定の動き情報を登録して動き情報候補リストを構築する候補リスト構築部(1506)と、前記符号化対象となる前記予測ブロックに用いる前記動き情報候補リスト内の動き情報候補を指定するインデックス情報を符号化する符号化部(118)と、前記動き情報候補を変換する動き情報変換部(1507)と、前記動き情報候補に基づいて、単予測又は双予測の何れかにより動き補償予測を行い前記符号化対象となる予測ブロックの予測信号を生成する動き補償予測部(112)とを備える。前記動き情報変換部(1507)は、前記動き情報候補のうち、前記双予測を示す予測種別情報を、前記単予測を示す予測種別情報に変換する予測変換を行い、前記動き補償予測部(112)は、前記符号化対象となる予測ブロックのブロックサイズが所定の第1サイズの場合であり、かつ予測種別情報が前記双予測を示す場合に、前記予測変換により変換された動き情報に基づいて前記動き補償予測を行う。 In order to solve the above-described problem, a moving picture encoding apparatus according to an aspect of the present invention specifies a prediction block from a block in which a picture is divided into a plurality of blocks step by step, and the specified prediction block unit includes: A moving image encoding device for generating an encoded stream, wherein motion information is derived from at least one of a block spatially close to a prediction block to be encoded and a block close in time, and the encoding A candidate list construction unit (1506) for registering predetermined motion information from the derived motion information as a motion information candidate of a target prediction block and constructing a motion information candidate list, and the encoding target An encoding unit (118) for encoding index information for designating motion information candidates in the motion information candidate list used for the prediction block; A motion information conversion unit (1507) that converts a motion information candidate, and based on the motion information candidate, performs motion compensation prediction by either uni-prediction or bi-prediction, and generates a prediction signal of the prediction block to be encoded A motion compensation prediction unit (112). The motion information conversion unit (1507) performs prediction conversion for converting the prediction type information indicating the bi-prediction among the motion information candidates into the prediction type information indicating the single prediction, and the motion compensation prediction unit (112) ) Is a case where the block size of the prediction block to be encoded is a predetermined first size, and when the prediction type information indicates the bi-prediction, based on the motion information converted by the prediction conversion The motion compensation prediction is performed.
 本発明の別の態様もまた、動画像符号化装置である。この装置は、動画像の各ピクチャを分割したブロック単位で動き補償予測を用いて前記動画像を符号化する動画像符号化装置であって、導出した動き情報を用いた動き補償により符号化対象予測ブロックの予測信号を生成する動き補償予測部(112)と、指定された第1のサイズの予測ブロックサイズにおける、動き補償予測を許可するか否かを指定する第1の制御パラメータ(inter_4x4_enable)と、指定された第2のサイズ以下の予測ブロックサイズにおける双予測の動き補償を禁止する、前記第2のサイズを指定する第2の制御パラメータ(inter_bipred_restriction_idc)を生成する符号化ブロック制御パラメータ生成部(122)と、前記第1及び第2の制御パラメータを含む、動き補償予測に用いる情報を符号化する符号化部(118)とを備える。前記動き補償予測部(112)は、前記第1及び第2の制御パラメータに基づき、動き補償予測を行う。 Another aspect of the present invention is also a moving picture coding apparatus. This apparatus is a moving picture encoding apparatus that encodes the moving picture using motion compensation prediction in units of blocks obtained by dividing each picture of the moving picture, and is to be encoded by motion compensation using the derived motion information. A motion compensation prediction unit (112) that generates a prediction signal of a prediction block, and a first control parameter (inter_4x4_enable) that specifies whether or not motion compensation prediction is permitted in the prediction block size of the specified first size And a coding block control parameter generation unit that generates a second control parameter (inter_bipred_restriction_idc) that specifies the second size and prohibits bi-prediction motion compensation in a prediction block size that is equal to or smaller than the specified second size. (122) and an encoding unit (118) that encodes information used for motion compensation prediction, including the first and second control parameters. Obtain. The motion compensation prediction unit (112) performs motion compensation prediction based on the first and second control parameters.
 本発明のさらに別の態様は、動画像符号化方法である。この方法は、ピクチャが複数のブロックに段階的に分割されたブロックから予測ブロックを特定し、その特定された予測ブロック単位で、符号化ストリームを生成する動画像符号化方法であって、符号化対象となる予測ブロックに空間的に近接するブロック及び時間的に近接するブロックの少なくとも何れかから動き情報を導出し、前記符号化対象となる予測ブロックの動き情報候補として、その導出された動き情報の中から所定の動き情報を登録して動き情報候補リストを構築する候補リスト構築ステップと、前記符号化対象となる前記予測ブロックに用いる前記動き情報候補リスト内の動き情報候補を指定するインデックス情報を符号化する符号化ステップと、前記動き情報候補を変換する動き情報変換ステップと、前記動き情報候補に基づいて、単予測又は双予測の何れかにより動き補償予測を行い前記符号化対象となる予測ブロックの予測信号を生成する動き補償予測ステップとを備える。前記動き情報変換ステップは、前記動き情報候補のうち、前記双予測を示す予測種別情報を、前記単予測を示す予測種別情報に変換する予測変換を行い、前記動き補償予測ステップは、前記符号化対象となる予測ブロックのブロックサイズが所定の第1サイズの場合であり、かつ予測種別情報が前記双予測を示す場合に、前記予測変換により変換された動き情報に基づいて前記動き補償予測を行う。 Still another aspect of the present invention is a video encoding method. This method is a moving picture coding method in which a prediction block is identified from a block in which a picture is divided into a plurality of blocks and a coded stream is generated in units of the identified prediction block. Motion information is derived from at least one of a spatially close block and a temporally close block to the target prediction block, and the derived motion information is used as a motion information candidate for the prediction block to be encoded. A candidate list construction step of registering predetermined motion information from among them to construct a motion information candidate list, and index information for designating motion information candidates in the motion information candidate list used for the prediction block to be encoded An encoding step for encoding the motion information, a motion information conversion step for converting the motion information candidate, and the motion information candidate Zui by, and a motion compensated prediction step of generating a prediction signal of the prediction block to be the encoding target performs motion compensation prediction by either single prediction or bi-prediction. The motion information conversion step performs prediction conversion for converting the prediction type information indicating the bi-prediction among the motion information candidates into prediction type information indicating the uni-prediction, and the motion compensation prediction step includes the encoding When the block size of the target prediction block is a predetermined first size and the prediction type information indicates the bi-prediction, the motion compensation prediction is performed based on the motion information converted by the prediction conversion. .
 本発明のさらに別の態様は、送信装置である。この装置は、ピクチャが複数のブロックに段階的に分割されたブロックから予測ブロックを特定し、その特定された予測ブロック単位で、符号化ストリームを生成する動画像符号化方法により符号化された前記符号化ストリームをパケット化して符号化データを得るパケット処理部と、パケット化された前記符号化データを送信する送信部とを備える。前記動画像符号化方法は、符号化対象となる予測ブロックに空間的に近接するブロック及び時間的に近接するブロックの少なくとも何れかから動き情報を導出し、前記符号化対象となる予測ブロックの動き情報候補として、その導出された動き情報の中から所定の動き情報を登録して動き情報候補リストを構築する候補リスト構築ステップと、前記符号化対象となる前記予測ブロックに用いる前記動き情報候補リスト内の動き情報候補を指定するインデックス情報を符号化する符号化ステップと、前記動き情報候補を変換する動き情報変換ステップと、前記動き情報候補に基づいて、単予測又は双予測の何れかにより動き補償予測を行い前記符号化対象となる予測ブロックの予測信号を生成する動き補償予測ステップとを備える。前記動き情報変換ステップは、前記動き情報候補のうち、前記双予測を示す予測種別情報を、前記単予測を示す予測種別情報に変換する予測変換を行い、前記動き補償予測ステップは、前記符号化対象となる予測ブロックのブロックサイズが所定の第1サイズの場合であり、かつ予測種別情報が前記双予測を示す場合に、前記予測変換により変換された動き情報に基づいて前記動き補償予測を行う。 Still another aspect of the present invention is a transmission device. This apparatus identifies a prediction block from a block obtained by dividing a picture into a plurality of blocks in stages, and is encoded by the moving picture coding method for generating an encoded stream in the identified prediction block unit. A packet processing unit that packetizes the encoded stream to obtain encoded data, and a transmission unit that transmits the packetized encoded data. The video encoding method derives motion information from at least one of a block spatially adjacent to a prediction block to be encoded and a block adjacent to temporally, and the motion of the prediction block to be encoded A candidate list construction step of constructing a motion information candidate list by registering predetermined motion information from the derived motion information as information candidates, and the motion information candidate list used for the prediction block to be encoded An encoding step for encoding index information for specifying a motion information candidate, a motion information conversion step for converting the motion information candidate, and motion based on either the single prediction or the bi-prediction based on the motion information candidate. A motion compensation prediction step of performing compensation prediction and generating a prediction signal of the prediction block to be encoded. The motion information conversion step performs prediction conversion for converting the prediction type information indicating the bi-prediction among the motion information candidates into prediction type information indicating the uni-prediction, and the motion compensation prediction step includes the encoding When the block size of the target prediction block is a predetermined first size and the prediction type information indicates the bi-prediction, the motion compensation prediction is performed based on the motion information converted by the prediction conversion. .
 本発明のさらに別の態様は、送信方法である。この方法は、ピクチャが複数のブロックに段階的に分割されたブロックから予測ブロックを特定し、その特定された予測ブロック単位で、符号化ストリームを生成する動画像符号化方法により符号化された前記符号化ストリームをパケット化して符号化データを得るパケット処理ステップと、パケット化された前記符号化データを送信する送信ステップとを備える。前記動画像符号化方法は、符号化対象となる予測ブロックに空間的に近接するブロック及び時間的に近接するブロックの少なくとも何れかから動き情報を導出し、前記符号化対象となる予測ブロックの動き情報候補として、その導出された動き情報の中から所定の動き情報を登録して動き情報候補リストを構築する候補リスト構築ステップと、前記符号化対象となる前記予測ブロックに用いる前記動き情報候補リスト内の動き情報候補を指定するインデックス情報を符号化する符号化ステップと、前記動き情報候補を変換する動き情報変換ステップと、前記動き情報候補に基づいて、単予測又は双予測の何れかにより動き補償予測を行い前記符号化対象となる予測ブロックの予測信号を生成する動き補償予測ステップとを備える。前記動き情報変換ステップは、前記動き情報候補のうち、前記双予測を示す予測種別情報を、前記単予測を示す予測種別情報に変換する予測変換を行い、前記動き補償予測ステップは、前記符号化対象となる予測ブロックのブロックサイズが所定の第1サイズの場合であり、かつ予測種別情報が前記双予測を示す場合に、前記予測変換により変換された動き情報に基づいて前記動き補償予測を行う。 Still another aspect of the present invention is a transmission method. In this method, a prediction block is identified from a block in which a picture is divided into a plurality of blocks in stages, and the coded image is encoded by the moving image coding method for generating an encoded stream in the identified prediction block unit. A packet processing step of packetizing the encoded stream to obtain encoded data; and a transmitting step of transmitting the packetized encoded data. The video encoding method derives motion information from at least one of a block spatially adjacent to a prediction block to be encoded and a block adjacent to temporally, and the motion of the prediction block to be encoded A candidate list construction step of constructing a motion information candidate list by registering predetermined motion information from the derived motion information as information candidates, and the motion information candidate list used for the prediction block to be encoded An encoding step for encoding index information for specifying a motion information candidate, a motion information conversion step for converting the motion information candidate, and motion based on either the single prediction or the bi-prediction based on the motion information candidate. A motion compensation prediction step of performing compensation prediction and generating a prediction signal of the prediction block to be encoded. The motion information conversion step performs prediction conversion for converting the prediction type information indicating the bi-prediction among the motion information candidates into prediction type information indicating the uni-prediction, and the motion compensation prediction step includes the encoding When the block size of the target prediction block is a predetermined first size and the prediction type information indicates the bi-prediction, the motion compensation prediction is performed based on the motion information converted by the prediction conversion. .
 本発明のある態様の動画像復号装置は、ピクチャが複数のブロックに段階的に分割されたブロックから予測ブロックを特定し、その特定された予測ブロック単位で、符号化ストリームを復号する動画像復号装置であって、前記符号化ストリームから、復号対象となる前記予測ブロックの動き情報を指定したインデックス情報を復号する復号部(1108)と、前記復号対象となる予測ブロックに空間的に近接するブロック及び時間的に近接するブロックの少なくとも何れかから動き情報を導出し、前記復号対象となる予測ブロックの動き情報候補として、その導出された動き情報の中から所定の動き情報を登録して動き情報候補リストを構築する候補リスト構築部(3604)と、前記動き情報候補を変換する動き情報変換部(3605)と、前記動き情報候補のうちの前記インデックス情報により指定された動き情報に基づいて、単予測又は双予測の何れかにより動き補償予測を行い前記復号対象となる予測ブロックの予測信号を生成する動き補償予測部(1114)とを備える。前記動き情報変換部(3605)は、前記動き情報候補のうち、前記双予測を示す予測種別情報を、前記単予測を示す予測種別情報に変換する予測変換を行い、前記動き補償予測部(1114)は、前記復号対象となる予測ブロックのブロックサイズが所定の第1サイズの場合であり、かつ前記指定された動き情報の予測種別情報が前記双予測を示す場合に、前記予測変換により変換された動き情報に基づいて前記動き補償予測を行う。 A video decoding device according to an aspect of the present invention specifies a prediction block from a block in which a picture is divided into a plurality of blocks in stages, and decodes an encoded stream in units of the specified prediction block A decoding unit (1108) that decodes index information specifying motion information of the prediction block to be decoded from the encoded stream, and a block spatially adjacent to the prediction block to be decoded Motion information is derived from at least one of temporally adjacent blocks, and predetermined motion information is registered from among the derived motion information as motion information candidates of the prediction block to be decoded. A candidate list construction unit (3604) for constructing a candidate list, and a motion information conversion unit (3605) for transforming the motion information candidates Motion compensation for generating a prediction signal of a prediction block to be decoded by performing motion compensation prediction by either uni-prediction or bi-prediction based on the motion information specified by the index information of the motion information candidates A prediction unit (1114). The motion information conversion unit (3605) performs prediction conversion for converting the prediction type information indicating the bi-prediction among the motion information candidates into the prediction type information indicating the single prediction, and the motion compensation prediction unit (1114). ) Is a case where the block size of the prediction block to be decoded is a predetermined first size, and when the prediction type information of the designated motion information indicates the bi-prediction, is converted by the prediction conversion. The motion compensation prediction is performed based on the motion information.
 本発明の別の態様もまた、動画像復号装置である。この装置は、動画像の各ピクチャを分割したブロック単位で動き補償予測を用いて前記動画像を符号化した符号化ストリームを復号する動画像復号装置であって、前記符号化ストリームから動き補償予測に用いる情報を復号すると共に、その復号した動き補償予測に用いる情報から、指定された第1のサイズの予測ブロックサイズにおける、動き補償予測を許可するか否かを指定する第1の制御パラメータ(inter_4x4_enable)と、指定された第2のサイズ以下の予測ブロックサイズにおける双予測の動き補償を禁止する、前記第2のサイズを指定する第2の制御パラメータ(inter_bipred_restriction_idc)とを得る復号部(1108)と、前記動き補償予測に用いる情報を用いて復号対象予測ブロックの予測信号を生成する動き補償予測部(1114)とを備える。前記動き補償予測部(1114)は、前記第1及び第2の制御パラメータに基づき、動き補償予測を行う。 Another aspect of the present invention is also a video decoding device. The apparatus is a moving picture decoding apparatus that decodes a coded stream obtained by coding the moving picture using motion compensated prediction in units of blocks obtained by dividing each picture of the moving picture, and the motion compensated prediction is performed from the coded stream. And a first control parameter for specifying whether or not motion compensated prediction is permitted in the prediction block size of the designated first size from the information used for the motion compensated prediction that has been decoded. inter_4x4_enable) and a second control parameter (inter_bipred_restriction_idc) that specifies the second size and prohibits bi-prediction motion compensation in a prediction block size equal to or smaller than the specified second size (1108) And a motion compensated prediction unit (1114) that generates a prediction signal of a decoding target prediction block using information used for the motion compensated prediction Equipped with a. The motion compensation prediction unit (1114) performs motion compensation prediction based on the first and second control parameters.
 本発明のさらに別の態様は、動画像復号方法である。この方法は、ピクチャが複数のブロックに段階的に分割されたブロックから予測ブロックを特定し、その特定された予測ブロック単位で、符号化ストリームを復号する動画像復号方法であって、前記符号化ストリームから、復号対象となる前記予測ブロックの動き情報を指定したインデックス情報を復号する復号ステップと、前記復号対象となる予測ブロックに空間的に近接するブロック及び時間的に近接するブロックの少なくとも何れかから動き情報を導出し、前記復号対象となる予測ブロックの動き情報候補として、その導出された動き情報の中から所定の動き情報を登録して動き情報候補リストを構築する候補リスト構築ステップと、前記動き情報候補を変換する動き情報変換ステップと、前記動き情報候補のうちの前記インデックス情報により指定された動き情報に基づいて、単予測又は双予測の何れかにより動き補償予測を行い前記復号対象となる予測ブロックの予測信号を生成する動き補償予測ステップとを備える。前記動き情報変換ステップは、前記動き情報候補のうち、前記双予測を示す予測種別情報を、前記単予測を示す予測種別情報に変換する予測変換を行い、前記動き補償予測ステップは、前記復号対象となる予測ブロックのブロックサイズが所定の第1サイズの場合であり、かつ前記指定された動き情報の予測種別情報が前記双予測を示す場合に、前記予測変換により変換された動き情報に基づいて前記動き補償予測を行う。 Still another aspect of the present invention is a moving picture decoding method. This method is a moving picture decoding method in which a prediction block is specified from a block in which a picture is divided into a plurality of blocks in stages, and an encoded stream is decoded in the specified prediction block unit. A decoding step of decoding index information specifying motion information of the prediction block to be decoded from the stream, and at least one of a block spatially adjacent to the prediction block to be decoded and a block adjacent in time A candidate list construction step of deriving motion information from the above, and as a motion information candidate of the prediction block to be decoded, registering predetermined motion information from the derived motion information and constructing a motion information candidate list; A motion information conversion step for converting the motion information candidates; and the index of the motion information candidates. Based on the specified motion information by broadcast, and a motion compensated prediction step of generating a prediction signal of a prediction block to be the decoding target subjected to the motion compensation prediction by either single prediction or bi-prediction. The motion information conversion step performs prediction conversion for converting prediction type information indicating the bi-prediction among the motion information candidates into prediction type information indicating the single prediction, and the motion compensation prediction step includes the decoding target Based on the motion information converted by the prediction conversion, when the block size of the prediction block to be is a predetermined first size and the prediction type information of the specified motion information indicates the bi-prediction The motion compensation prediction is performed.
 本発明のさらに別の態様は、受信装置である。この装置は、ピクチャが複数のブロックに段階的に分割されたブロックから予測ブロックを特定し、その特定された予測ブロック単位で、動画像が符号化された符号化ストリームを受信して復号する受信装置であって、前記符号化ストリームがパケット化された符号化データを受信する受信部と、受信された前記符号化ストリームをパケット処理して元の符号化ストリームを復元する復元部と、復元された前記符号化ストリームから、復号対象となる前記予測ブロックの動き情報を指定したインデックス情報を復号する復号部と、前記復号対象となる予測ブロックに空間的に近接するブロック及び時間的に近接するブロックの少なくとも何れかから動き情報を導出し、前記復号対象となる予測ブロックの動き情報候補として、その導出された動き情報の中から所定の動き情報を登録して動き情報候補リストを構築する候補リスト構築部と、前記動き情報候補を変換する動き情報変換部と、前記動き情報候補のうちの前記インデックス情報により指定された動き情報に基づいて、単予測又は双予測の何れかにより動き補償予測を行い前記復号対象となる予測ブロックの予測信号を生成する動き補償予測部とを備える。前記動き情報変換部は、前記動き情報候補のうち、前記双予測により前記動き補償予測を行うことを示す予測種別情報を、前記単予測により前記動き補償予測を行うことを示す予測種別情報に変換する予測変換を行い、前記動き補償予測部は、前記復号対象となる予測ブロックのブロックサイズが所定の第1サイズの場合であり、かつ前記指定された動き情報の前記予測種別情報が前記双予測により前記動き補償予測を行うことを示す場合に、前記予測変換により変換された前記予測種別情報により前記動き補償予測を行う。 Still another aspect of the present invention is a receiving device. This apparatus specifies a prediction block from a block in which a picture is divided into a plurality of blocks in stages, and receives and decodes an encoded stream in which a moving image is encoded in the specified prediction block unit. A receiving unit that receives encoded data in which the encoded stream is packetized; a recovery unit that processes the received encoded stream to recover the original encoded stream; A decoding unit that decodes index information specifying motion information of the prediction block to be decoded from the encoded stream, a block that is spatially close to the prediction block that is to be decoded, and a block that is temporally close to the prediction block Motion information is derived from at least one of the above, and the motion information candidate of the prediction block to be decoded is derived as the motion information candidate A candidate list construction unit for registering predetermined motion information from among the motion information and constructing a motion information candidate list, a motion information conversion unit for converting the motion information candidates, and the index information of the motion information candidates. A motion-compensated prediction unit that performs motion-compensated prediction based on specified motion information by either uni-prediction or bi-prediction and generates a prediction signal of a prediction block to be decoded. The motion information conversion unit converts prediction type information indicating that the motion compensation prediction is performed by the bi-prediction among the motion information candidates into prediction type information indicating that the motion compensation prediction is performed by the single prediction. The motion compensation prediction unit is a case where the block size of the prediction block to be decoded is a predetermined first size, and the prediction type information of the designated motion information is the bi-prediction Indicates that the motion compensation prediction is performed, the motion compensation prediction is performed based on the prediction type information converted by the prediction conversion.
 本発明のさらに別の態様は、受信方法である。この方法は、ピクチャが複数のブロックに段階的に分割されたブロックから予測ブロックを特定し、その特定された予測ブロック単位で、動画像が符号化された符号化ストリームを受信して復号する受信方法であって、前記符号化ストリームがパケット化された符号化データを受信する受信ステップと、受信された前記符号化ストリームをパケット処理して元の符号化ストリームを復元する復元ステップと、復元された前記符号化ストリームから、復号対象となる前記予測ブロックの動き情報を指定したインデックス情報を復号する復号ステップと、前記復号対象となる予測ブロックに空間的に近接するブロック及び時間的に近接するブロックの少なくとも何れかから動き情報を導出し、前記復号対象となる予測ブロックの動き情報候補として、その導出された動き情報の中から所定の動き情報を登録して動き情報候補リストを構築する候補リスト構築ステップと、前記動き情報候補を変換する動き情報変換ステップと、前記動き情報候補のうちの前記インデックス情報により指定された動き情報に基づいて、単予測又は双予測の何れかにより動き補償予測を行い前記復号対象となる予測ブロックの予測信号を生成する動き補償予測ステップとを備える。前記動き情報変換ステップは、前記動き情報候補のうち、前記双予測により前記動き補償予測を行うことを示す予測種別情報を、前記単予測により前記動き補償予測を行うことを示す予測種別情報に変換する予測変換を行い、前記動き補償予測ステップは、前記復号対象となる予測ブロックのブロックサイズが所定の第1サイズの場合であり、かつ前記指定された動き情報の前記予測種別情報が前記双予測により前記動き補償予測を行うことを示す場合に、前記予測変換により変換された前記予測種別情報により前記動き補償予測を行う。 Still another aspect of the present invention is a receiving method. This method specifies a prediction block from a block in which a picture is divided into a plurality of blocks in stages, and receives and decodes an encoded stream in which a moving image is encoded in the specified prediction block unit. A receiving step for receiving encoded data in which the encoded stream is packetized; a restoring step for packetizing the received encoded stream to restore the original encoded stream; A decoding step for decoding the index information specifying the motion information of the prediction block to be decoded from the encoded stream, a block spatially adjacent to the prediction block to be decoded, and a block adjacent in time Motion information is derived from at least one of the prediction block motion information candidates of the prediction block to be decoded. A candidate list construction step for registering predetermined motion information from the derived motion information to construct a motion information candidate list, a motion information conversion step for converting the motion information candidates, and among the motion information candidates A motion compensation prediction step of performing motion compensation prediction by either uni-prediction or bi-prediction based on the motion information specified by the index information, and generating a prediction signal of the prediction block to be decoded. In the motion information conversion step, prediction type information indicating that the motion compensation prediction is performed by the bi-prediction among the motion information candidates is converted into prediction type information indicating that the motion compensation prediction is performed by the single prediction. The motion compensation prediction step is performed when the block size of the prediction block to be decoded is a predetermined first size, and the prediction type information of the designated motion information is the bi-prediction. Indicates that the motion compensation prediction is performed, the motion compensation prediction is performed based on the prediction type information converted by the prediction conversion.
 なお、以上の構成要素の任意の組み合わせ、本発明の表現を方法、装置、システム、記録媒体、コンピュータプログラムなどの間で変換したものもまた、本発明の態様として有効である。 It should be noted that any combination of the above-described constituent elements, and a conversion of the expression of the present invention between a method, an apparatus, a system, a recording medium, a computer program, etc. are also effective as an aspect of the present invention.
 本発明によれば、参照ピクチャのメモリアクセス量を所定量以下に制限しつつ、符号化効率を向上させることができる。 According to the present invention, it is possible to improve the encoding efficiency while limiting the memory access amount of the reference picture to a predetermined amount or less.
本発明の実施の形態1に係る動画像符号化装置の構成を示す図である。It is a figure which shows the structure of the moving image encoder which concerns on Embodiment 1 of this invention. 符号化対象画像の分割構造の一例を示す図である。It is a figure which shows an example of the division | segmentation structure of an encoding object image. CU/予測ブロックサイズの詳細な定義を示す図である。It is a figure which shows the detailed definition of CU / prediction block size. 図4(a)~(d)は、動き補償予測の予測種別について説明するための図である。4A to 4D are diagrams for explaining the prediction type of motion compensation prediction. 本発明の実施の形態1に係る動画像符号化装置における符号化ブロック単位の符号化処理の動作の流れを示すフローチャートである。It is a flowchart which shows the flow of an operation | movement of the encoding process per encoding block in the moving image encoder which concerns on Embodiment 1 of this invention. 図5のステップS503におけるCU予測モード/予測信号生成処理の詳細動作を説明するためのフローチャートである。6 is a flowchart for explaining a detailed operation of a CU prediction mode / prediction signal generation process in step S503 of FIG. 図6のステップS608における動き補償予測ブロック(PU)サイズ選択/予測信号生成処理の詳細動作を説明するためのフローチャートである。7 is a flowchart for explaining detailed operation of motion compensated prediction block (PU) size selection / prediction signal generation processing in step S608 of FIG. 6. 図8(a)(b)は、本発明の実施の形態1における動き補償予測において使用される動き情報を符号化するための2つの予測モードを説明するための図である。FIGS. 8A and 8B are diagrams for explaining two prediction modes for encoding motion information used in motion compensated prediction according to Embodiment 1 of the present invention. 動き補償予測に水平・垂直7タップフィルタを用いた場合の動き補償予測に必要な参照画像メモリ量の概算値を示す図である。It is a figure which shows the rough value of the reference image memory amount required for motion compensation prediction at the time of using a horizontal and vertical 7 tap filter for motion compensation prediction. 本発明の実施の形態1に係る、動き補償予測のブロックサイズ及び予測処理を制御する制御パラメータを説明するための図である。It is a figure for demonstrating the control parameter which controls the block size and prediction process of motion compensation prediction based on Embodiment 1 of this invention. 本発明の実施の形態1に係る動画像復号装置の構成を示す図である。It is a figure which shows the structure of the moving image decoding apparatus which concerns on Embodiment 1 of this invention. 本発明の実施の形態1に係る動画像復号装置における符号化ブロック単位の復号処理の動作の流れを示すフローチャートである。It is a flowchart which shows the flow of operation | movement of the decoding process of the encoding block unit in the moving image decoding apparatus which concerns on Embodiment 1 of this invention. 図12のステップS1202におけるCU単位復号処理の詳細動作を説明するためのフローチャートである。It is a flowchart for demonstrating the detailed operation | movement of the CU unit decoding process in step S1202 of FIG. 図13のステップS1310におけるCU単位動き補償予測復号処理の詳細動作を説明するためのフローチャートである。It is a flowchart for demonstrating the detailed operation | movement of the CU unit motion compensation prediction decoding process in step S1310 of FIG. 本発明の実施の形態1の動画像符号化装置における動き補償予測ブロック構造選択部の詳細な構成を示す図である。It is a figure which shows the detailed structure of the motion compensation prediction block structure selection part in the moving image encoder of Embodiment 1 of this invention. 結合動き情報算出部の構成を示す図である。It is a figure which shows the structure of a joint motion information calculation part. 図15の動き補償予測ブロック構造選択部を介して動作する、図7のステップS701、S702、S703、S705である動き補償予測モード/予測信号生成の動作について説明するためのフローチャートである。16 is a flowchart for explaining an operation of motion compensation prediction mode / prediction signal generation, which is steps S701, S702, S703, and S705 of FIG. 7, which operates via the motion compensation prediction block structure selection unit of FIG. 15. 図17のステップS1701における結合動き情報候補リスト生成の詳細動作を説明するためのフローチャートである。It is a flowchart for demonstrating the detailed operation | movement of combined motion information candidate list production | generation in step S1701 of FIG. 空間結合動き情報候補リスト生成に用いる空間候補ブロック群を示す図である。It is a figure which shows the space candidate block group used for a space joint motion information candidate list generation. 図18のステップS1800における空間結合動き情報候補リスト生成処理の詳細動作を説明するためのフローチャートである。It is a flowchart for demonstrating the detailed operation | movement of the space joint motion information candidate list production | generation process in step S1800 of FIG. 図18のステップS1801における結合動き情報候補削除処理の詳細動作を説明するためのフローチャートである。It is a flowchart for demonstrating the detailed operation | movement of the combined motion information candidate deletion process in step S1801 of FIG. 結合動き情報候補が4つの場合のリスト中の候補の比較関係を示す図である。It is a figure which shows the comparison relationship of the candidate in the list | wrist in the case of four combined motion information candidates. 図23(a)(b)は、結合動き情報候補の比較内容の一例を示す図である。FIGS. 23A and 23B are diagrams illustrating an example of comparison contents of combined motion information candidates. 時間結合動き情報候補リスト生成に用いる時間候補ブロック群を示す図である。It is a figure which shows the time candidate block group used for a time joint movement information candidate list generation. 図18のステップS1802における時間結合動き情報候補リスト生成処理の詳細動作を説明するためのフローチャートである。It is a flowchart for demonstrating the detailed operation | movement of the time combination motion information candidate list production | generation process in FIG.18 S1802. 時間結合動き情報に対する基準動きベクトル値ColMvに対する、L0予測、L1予測に対して登録する動きベクトル値mvL0t、mvL1tの算出手法を説明するための図である。It is a figure for demonstrating the calculation method of the motion vector value mvL0t and mvL1t registered with respect to the reference | standard motion vector value ColMv with respect to time coupling | bonding motion information with respect to L0 prediction and L1 prediction. 図18のステップS1803における第1結合動き情報候補リスト追加処理の詳細動作を説明するためのフローチャートである。It is a flowchart for demonstrating the detailed operation | movement of the 1st joint motion information candidate list addition process in step S1803 of FIG. 第1結合動き情報候補リスト追加処理における、組み合わせ検査回数と結合動き情報候補Mと結合動き情報候補Nの関係を説明するための図であるIt is a figure for demonstrating the relationship between the frequency | count of combination inspection, the combined motion information candidate M, and the combined motion information candidate N in the 1st combined motion information candidate list addition process. 図18のステップS1804における第2結合動き情報候補リスト追加処理の詳細動作を説明するためのフローチャートである。It is a flowchart for demonstrating the detailed operation | movement of the 2nd joint motion information candidate list addition process in step S1804 of FIG. 図17のステップS1703における結合動き情報候補単予測変換処理の詳細動作を説明するためのフローチャートである。It is a flowchart for demonstrating the detailed operation | movement of the joint motion information candidate single prediction conversion process in step S1703 of FIG. 図17のステップS1704における結合予測モード評価値生成処理の詳細動作を説明するためのフローチャートである。It is a flowchart for demonstrating the detailed operation | movement of the joint prediction mode evaluation value production | generation process in step S1704 of FIG. 結合動き情報候補数が5の場合のTruncated Unary符号列を示す図である。It is a figure which shows the Truncated Unary code sequence when the number of combined motion information candidates is 5. 図17のステップS1705における予測モード評価値生成処理の詳細動作を説明するためのフローチャートである。It is a flowchart for demonstrating the detailed operation | movement of the prediction mode evaluation value production | generation process in step S1705 of FIG. 予測ブロックの動き情報に関するシンタックスである。It is the syntax regarding the motion information of a prediction block. 予測ブロックサイズによる双予測及び予測処理の制限を制御するパラメータに関するシンタックスである。It is the syntax regarding the parameter which controls the restriction | limiting of the bi-prediction by prediction block size, and a prediction process. 本発明の実施の形態1の動画像復号装置における動き情報復号部の詳細な構成を示す図である。It is a figure which shows the detailed structure of the motion information decoding part in the moving image decoding apparatus of Embodiment 1 of this invention. 図14のステップS1402、S1406、S1408、S1410における予測ブロック単位復号処理の詳細動作を説明するためのフローチャートである。15 is a flowchart for explaining detailed operations of prediction block unit decoding processing in steps S1402, S1406, S1408, and S1410 of FIG. 図37のステップS3702における動き情報復号処理の詳細動作を説明するためのフローチャートである。It is a flowchart for demonstrating the detailed operation | movement of the motion information decoding process in step S3702 of FIG. 図38のステップS3801における結合予測動き情報復号処理の詳細動作を説明するためのフローチャートである。It is a flowchart for demonstrating the detailed operation | movement of the joint prediction motion information decoding process in FIG.38 S3801. 図38のステップS3805における予測動き情報復号処理の詳細動作を説明するためのフローチャートである。FIG. 39 is a flowchart for explaining detailed operations of a predicted motion information decoding process in step S3805 of FIG. 38. FIG. レベルと連動した予測ブロックサイズ制御パラメータの制限を示す一例である。It is an example which shows the restriction | limiting of the prediction block size control parameter linked | linked with the level. 本発明の実施の形態1の別構成における、予測ブロックサイズの分割構造を示す図である。It is a figure which shows the division | segmentation structure of the prediction block size in another structure of Embodiment 1 of this invention. 本発明の実施の形態1の別構成における、動き補償予測のブロックサイズ及び予測処理を制御する制御パラメータを説明するための図である。It is a figure for demonstrating the control parameter which controls the block size and prediction process of motion compensation prediction in another structure of Embodiment 1 of this invention. 本発明の実施の形態1において、動き補償予測のブロックサイズ及び予測処理を制御する2つの制御パラメータを1つの符号化伝送パラメータとして統合した一例である。In Embodiment 1 of this invention, it is an example which integrated two control parameters which control the block size of a motion compensation prediction, and a prediction process as one encoding transmission parameter. 本発明の実施の形態2に係る、動き補償予測のブロックサイズ及び予測処理を制御する制御パラメータを説明するための図であるIt is a figure for demonstrating the control parameter which controls the block size and prediction process of motion compensation prediction based on Embodiment 2 of this invention. 本発明の実施の形態2に係る、双予測を制御する制御パラメータと予測ブロックサイズの関係を示す図であるIt is a figure which shows the relationship between the control parameter which controls bi-prediction, and prediction block size based on Embodiment 2 of this invention. 本発明の実施の形態2における、予測ブロックサイズによる双予測及び予測処理の制限を制御するパラメータに関するシンタックスの一例である。It is an example of the syntax regarding the parameter which controls the restriction | limiting of the bi-prediction by prediction block size and prediction processing in Embodiment 2 of this invention. 本発明の実施の形態3における、結合動き情報候補生成における空間周辺予測ブロックの定義の一例を示す図である。It is a figure which shows an example of the definition of a space periphery prediction block in combined motion information candidate generation in Embodiment 3 of this invention. 本発明の実施の形態3における、動き補償予測ブロック(PU)サイズ選択/予測信号生成処理の詳細動作を説明するためのフローチャートである。It is a flowchart for demonstrating the detailed operation | movement of the motion compensation prediction block (PU) size selection / prediction signal generation process in Embodiment 3 of this invention. 本発明の実施の形態3における、動き補償予測モード/予測信号生成処理の詳細動作を説明するためのフローチャートである。It is a flowchart for demonstrating the detailed operation | movement of the motion compensation prediction mode / prediction signal generation process in Embodiment 3 of this invention. 本発明の実施の形態3における、結合予測動き情報復号処理の一例の詳細動作を説明するためのフローチャートである。It is a flowchart for demonstrating the detailed operation | movement of an example of a joint prediction motion information decoding process in Embodiment 3 of this invention. 本発明の実施の形態4における、動き補償予測ブロック生成処理の詳細動作を説明するためのフローチャートである。It is a flowchart for demonstrating the detailed operation | movement of the motion compensation prediction block generation process in Embodiment 4 of this invention. 本発明の実施の形態5における、結合動き情報候補リスト生成処理の詳細動作を説明するためのフローチャートである。It is a flowchart for demonstrating the detailed operation | movement of the joint motion information candidate list production | generation process in Embodiment 5 of this invention. 本発明の実施の形態5における、結合動き情報候補単予測変換処理の詳細動作を説明するためのフローチャートである。It is a flowchart for demonstrating the detailed operation | movement of the joint motion information candidate single prediction conversion process in Embodiment 5 of this invention. 本発明の実施の形態6における、結合動き情報候補単予測変換処理の詳細動作を説明するためのフローチャートである。It is a flowchart for demonstrating the detailed operation | movement of the joint motion information candidate single prediction conversion process in Embodiment 6 of this invention.
 以下、図面とともに本発明の実施の形態に係る動画像符号化装置、動画像符号化方法、動画像符号化プログラム、並びに動画像復号装置、動画像復号方法、動画像復号プログラムの好適な実施の形態について詳細に説明する。なお、図面の説明には同一要素には同一符号を付与して重複する説明を省略する。 DESCRIPTION OF EXEMPLARY EMBODIMENTS Preferred embodiments of a moving image encoding apparatus, a moving image encoding method, a moving image encoding program, a moving image decoding apparatus, a moving image decoding method, and a moving image decoding program according to an embodiment of the present invention will be described with reference to the drawings. A form is demonstrated in detail. In the description of the drawings, the same elements are denoted by the same reference numerals, and redundant description is omitted.
 (実施の形態1)
 [動画像符号化装置全体構成]
 図1は、本発明の実施の形態1に係る動画像符号化装置の構成を示す図である。以下、各部の動作について説明する。実施の形態1に係る動画像符号化装置は、入力端子100、入力画像メモリ101、符号化ブロック取得部102、減算部103、直交変換・量子化部104、予測誤差符号化部105、逆量子化・逆変換部106、加算部107、フレーム内復号画像バッファ108、ループフィルタ部109、復号画像メモリ110、動きベクトル検出部111、動き補償予測部112、動き補償予測ブロック構造選択部113、イントラ予測部114、イントラ予測ブロック構造選択部115、予測モード選択部116、符号化ブロック構造選択部117、ブロック構造/予測モード情報付加情報符号化部118、予測モード情報メモリ119、多重化部120、出力端子121、及び符号化ブロック制御パラメータ生成部122を備える。
(Embodiment 1)
[Overall configuration of video encoding apparatus]
FIG. 1 is a diagram showing a configuration of a moving picture coding apparatus according to Embodiment 1 of the present invention. Hereinafter, the operation of each unit will be described. The moving picture coding apparatus according to Embodiment 1 includes an input terminal 100, an input picture memory 101, a coding block acquisition unit 102, a subtraction unit 103, an orthogonal transform / quantization unit 104, a prediction error coding unit 105, an inverse quantum. / Inverse conversion unit 106, addition unit 107, intra-frame decoded image buffer 108, loop filter unit 109, decoded image memory 110, motion vector detection unit 111, motion compensation prediction unit 112, motion compensation prediction block structure selection unit 113, intra Prediction unit 114, intra prediction block structure selection unit 115, prediction mode selection unit 116, coding block structure selection unit 117, block structure / prediction mode information additional information coding unit 118, prediction mode information memory 119, multiplexing unit 120, An output terminal 121 and a coding block control parameter generation unit 122 are provided.
 入力端子100より入力される画像信号は、入力画像メモリ101に格納され、入力画像メモリ101より、符号化対象ピクチャに対する処理対象の画像信号が、符号化ブロック取得部102に入力される。符号化ブロック取得部102により符号化対象ブロックの位置情報に基づいて切り出された符号化対象ブロックの画像信号は、減算部103、動きベクトル検出部111、動き補償予測部112、及びイントラ予測部114に供給される。 The image signal input from the input terminal 100 is stored in the input image memory 101, and the image signal to be processed for the encoding target picture is input from the input image memory 101 to the encoding block acquisition unit 102. The image signal of the encoding target block extracted based on the position information of the encoding target block by the encoding block acquisition unit 102 is a subtraction unit 103, a motion vector detection unit 111, a motion compensation prediction unit 112, and an intra prediction unit 114. To be supplied.
 図2は、符号化対象画像の一例を示す図である。実施の形態1に係る予測ブロックサイズに関しては、図2に示すように符号化対象画像が64×64画素の符号化ブロック単位で符号化処理され、予測ブロックは符号化ブロックを基準に構成される。最大予測ブロックサイズは符号化ブロックと同じ64×64画素で、最小予測ブロックサイズは4×4画素である。CUの予測ブロックへの分割構成は、非分割(2N×2N)、水平・垂直への分割(N×N)、水平方向のみへの分割(2N×N)、垂直方向のみへの分割(N×2N)が可能である。水平・垂直への分割の場合は、更に水平・垂直に分割された予測ブロックを符号化ブロック(CU)として階層的に予測ブロックに分割でき、その階層をCU分割数で表現する。4分割されたCUの上位階層CUから見た分割領域をここでは分割1、分割2、分割3、分割4と定義する。 FIG. 2 is a diagram illustrating an example of an encoding target image. Regarding the prediction block size according to the first embodiment, as shown in FIG. 2, the encoding target image is encoded in units of 64 × 64 pixel encoding blocks, and the prediction block is configured based on the encoding block. . The maximum prediction block size is 64 × 64 pixels, which is the same as the encoded block, and the minimum prediction block size is 4 × 4 pixels. The division configuration of a CU into prediction blocks includes non-division (2N × 2N), horizontal / vertical division (N × N), horizontal division only (2N × N), and vertical division only (N × 2N) is possible. In the case of horizontal / vertical division, the prediction block further divided horizontally and vertically can be hierarchically divided into prediction blocks as coding blocks (CU), and the hierarchy is expressed by the number of CU divisions. Here, the divided areas viewed from the upper hierarchy CU of the four divided CUs are defined as division 1, division 2, division 3, and division 4.
 図3は、予測ブロックサイズの詳細な定義の一例を示す図である。CUのブロックサイズ(CUサイズ)は、CU分割数(CU_Depth)が0である64画素×64画素から、CU分割数が3である8×8画素まで定義され、最大の予測ブロックサイズはCU_Depth=0で非分割(2N×2N)の64画素×64画素、最小の予測ブロックサイズはCU_Depth=3で水平・垂直への分割(N×N)の4画素×4画素までの予測ブロックサイズが存在することになる。 FIG. 3 is a diagram illustrating an example of a detailed definition of the predicted block size. The block size (CU size) of the CU is defined from 64 pixels × 64 pixels in which the CU division number (CU_Depth) is 0 to 8 × 8 pixels in which the CU division number is 3, and the maximum prediction block size is CU_Depth = Predicted block size of 64 pixels x 64 pixels of 0, non-divided (2N x 2N), minimum prediction block size up to 4 pixels x 4 pixels of horizontal / vertical division (N x N) with CU_Depth = 3 Will do.
 画面間の相関を用いて予測を行う、動き補償予測を行う場合の予測ブロックサイズは、CUの予測ブロックへの分割構成に対して、水平方向のみへの分割(2N×N)、垂直方向のみへの分割(N×2N)を可能としており、計13種類の予測ブロックサイズが定義可能であるが、画面内の相関を用いて予測を行うイントラ予測の場合の予測ブロックサイズは、水平方向のみへの分割(2N×N)、垂直方向のみへの分割(N×2N)を可能としないため、計5種類の予測ブロックサイズが定義される。 The prediction block size in the case of performing motion compensation prediction that performs prediction using correlation between screens is divided only in the horizontal direction (2N × N), only in the vertical direction, with respect to the division configuration of the CU into prediction blocks. Can be defined and a total of 13 types of prediction block sizes can be defined. However, the prediction block size in the case of intra prediction in which prediction is performed using correlation in the screen is only in the horizontal direction. Since the division into two (2N × N) and the division only into the vertical direction (N × 2N) are not possible, a total of five types of prediction block sizes are defined.
 本発明の実施の形態1に係る予測ブロックの分割構成に関して、この組み合わせに限定されない。定義できる符号化ブロックサイズは、図3に示すMaximum_cu_sizeやMinimum_cu_sizeなどの制御パラメータを用いて最大CUサイズや最小CUサイズを設定し、これらの制御パラメータを符号化・復号することで、変化させることが可能である。 The partition configuration of the prediction block according to Embodiment 1 of the present invention is not limited to this combination. The encoding block size that can be defined can be changed by setting the maximum CU size and the minimum CU size using control parameters such as Maximum_cu_size and Minimum_cu_size shown in FIG. 3, and encoding and decoding these control parameters. Is possible.
 図1に戻り、減算部103は、符号化ブロック取得部102より供給される画像信号と符号化ブロック構造選択部117より供給される予測信号を減算して予測誤差信号を算出し、予測誤差信号を直交変換・量子化部104に供給する。 Returning to FIG. 1, the subtraction unit 103 calculates a prediction error signal by subtracting the image signal supplied from the coding block acquisition unit 102 and the prediction signal supplied from the coding block structure selection unit 117, and outputs a prediction error signal. Is supplied to the orthogonal transform / quantization unit 104.
 直交変換・量子化部104は、減算部103より供給される予測誤差信号に直交変換及び量子化を施し、量子化された予測誤差信号を予測誤差符号化部105及び逆量子化・逆変換部106に供給する。 The orthogonal transform / quantization unit 104 performs orthogonal transform and quantization on the prediction error signal supplied from the subtraction unit 103, and the quantized prediction error signal is subjected to a prediction error encoding unit 105 and an inverse quantization / inverse conversion unit. 106.
 予測誤差符号化部105は、直交変換・量子化部104より供給される量子化された予測誤差信号をエントロピー符号化して、予測誤差信号に対する符号列を生成して、多重化部120に供給する。 The prediction error encoding unit 105 entropy-encodes the quantized prediction error signal supplied from the orthogonal transform / quantization unit 104, generates a code string for the prediction error signal, and supplies the code sequence to the multiplexing unit 120. .
 逆量子化・逆変換部106は、直交変換・量子化部104より供給される量子化された予測誤差信号に対して、逆量子化や逆直交変換などの処理を行い、復号予測誤差信号を生成し加算部107に供給する。 The inverse quantization / inverse transform unit 106 performs a process such as inverse quantization or inverse orthogonal transform on the quantized prediction error signal supplied from the orthogonal transform / quantization unit 104 to generate a decoded prediction error signal. Generated and supplied to the adder 107.
 加算部107は、逆量子化・逆変換部106より供給される復号予測誤差信号と、符号化ブロック構造選択部117より供給される予測信号を加算して、復号画像信号を生成し、復号画像信号をフレーム内復号画像バッファ108及びループフィルタ部109に供給する。 The addition unit 107 adds the decoded prediction error signal supplied from the inverse quantization / inverse conversion unit 106 and the prediction signal supplied from the coding block structure selection unit 117 to generate a decoded image signal, and generates a decoded image signal. The signal is supplied to the intra-frame decoded image buffer 108 and the loop filter unit 109.
 フレーム内復号画像バッファ108は、符号化対象ブロックに隣接した領域の同一フレーム内の復号画像をイントラ予測部114に供給すると共に、加算部107より供給される復号画像信号を格納する。 The intra-frame decoded image buffer 108 supplies the decoded image in the same frame in the region adjacent to the encoding target block to the intra prediction unit 114 and also stores the decoded image signal supplied from the addition unit 107.
 ループフィルタ部109は、加算部107より供給される復号画像信号に対して、符号化によって生じる歪の除去や符号化前画像に近づける復元処理を、フィルタを施すことにより行い、フィルタ処理を行った結果の復号画像を復号画像メモリ110に供給する。 The loop filter unit 109 performs a filtering process on the decoded image signal supplied from the adding unit 107 by applying a filter to remove distortion caused by encoding and to restore the image to a pre-encoded image. The resulting decoded image is supplied to the decoded image memory 110.
 復号画像メモリ110は、ループフィルタ部109より供給されるフィルタ処理を行った復号画像信号を格納する。また、画像全体の復号が完了した復号画像については参照画像として1以上の所定画像数記憶し、参照画像信号を動きベクトル検出部111と動き補償予測部112に供給する。 The decoded image memory 110 stores the decoded image signal subjected to the filtering process supplied from the loop filter unit 109. In addition, a decoded image for which decoding of the entire image has been completed is stored as a reference image by a predetermined number of images, and the reference image signal is supplied to the motion vector detection unit 111 and the motion compensation prediction unit 112.
 動きベクトル検出部111は、符号化ブロック取得部102より供給される符号化対象ブロックの画像信号と、復号画像メモリ110に記憶された参照画像信号の入力を受けて、各参照画像に対する動きベクトルを検出し、動きベクトル値を動き補償予測部112及び動き補償予測ブロック構造選択部113に供給する。 The motion vector detection unit 111 receives the input of the image signal of the encoding target block supplied from the encoding block acquisition unit 102 and the reference image signal stored in the decoded image memory 110, and obtains a motion vector for each reference image. The motion vector value is detected and supplied to the motion compensation prediction unit 112 and the motion compensation prediction block structure selection unit 113.
 一般的な動きベクトルの検出方法は、画像信号と同一位置より所定の移動量だけ移動させた参照画像に相当する画像信号について誤差評価値を算出し、誤差評価値が最小となる移動量を動きベクトルとする。誤差評価値としては、画素毎の差分絶対値の総和SAD(Sum of Absolute Difference)や、画素毎の二乗誤差値の総和SSE(Sum of Square Error)等を利用する。さらに、動きベクトルの符号化に関わる符号量も誤差評価値に含めることも可能である。 A general motion vector detection method calculates an error evaluation value for an image signal corresponding to a reference image moved by a predetermined movement amount from the same position as the image signal, and moves the movement amount that minimizes the error evaluation value. Let it be a vector. As the error evaluation value, a sum of absolute differences SAD (Sum of Absolute Difference) for each pixel, a sum of squared error values SSE (Sum of Square Error) for each pixel, or the like is used. Furthermore, the code amount related to the coding of the motion vector can also be included in the error evaluation value.
 動き補償予測部112は、動き補償予測ブロック構造選択部113より指定される予測ブロック構造を指定する情報及び参照画像指定情報と、動きベクトル検出部111より入力される動きベクトル値に従って、復号画像メモリ110内の参照画像指定情報で示される参照画像を、動きベクトル値だけ予測ブロックの画像信号と同一位置より移動させた位置の画像信号を取得して予測信号を生成する。 The motion compensated prediction unit 112 is configured to decode the decoded image memory according to the information specifying the prediction block structure specified by the motion compensated prediction block structure selecting unit 113, the reference image specifying information, and the motion vector value input from the motion vector detecting unit 111. A prediction signal is generated by acquiring an image signal at a position obtained by moving the reference image indicated by the reference image designation information in 110 from the same position as the image signal of the prediction block by a motion vector value.
 動き補償予測ブロック構造選択部113より指定される予測モードが単一の参照画像からの予測の場合には、1つの参照画像から取得した予測信号を動き補償予測信号とし、予測モードが2つの参照画像からの予測の場合には、2つの参照画像から取得した予測信号を重みづけ平均したものを動き補償予測信号とし、動き補償予測信号を予測モード選択部116に供給する。ここでは双予測の重みづけ平均の比率を1:1とする。 When the prediction mode specified by the motion compensated prediction block structure selection unit 113 is prediction from a single reference image, a prediction signal acquired from one reference image is used as a motion compensation prediction signal, and two prediction modes are referenced. In the case of prediction from an image, a weighted average of prediction signals acquired from two reference images is used as a motion compensation prediction signal, and the motion compensation prediction signal is supplied to the prediction mode selection unit 116. Here, the ratio of the weighted average of bi-prediction is set to 1: 1.
 図4(a)~(d)は、動き補償予測の予測種別について説明するための図である。単一の参照画像からの予測を行う処理を単予測と定義し、単予測の場合はL0予測またはL1予測という、参照画像管理リストに登録された2つの参照画像のいずれか一方を利用した予測を行う。 4 (a) to 4 (d) are diagrams for explaining the prediction type of motion compensation prediction. A process for performing prediction from a single reference image is defined as single prediction, and in the case of single prediction, prediction using either one of two reference images registered in the reference image management list, that is, L0 prediction or L1 prediction. I do.
 図4(a)は単予測であってL0予測の参照画像(RefL0Pic)が符号化対象画像(CurPic)より前の時刻にある場合を示している。図4(b)は単予測であってL0予測の参照画像が符号化対象画像より後の時刻にある場合を示している。同様に、図4(a)および図4(b)のL0予測の参照画像をL1予測の参照画像(RefL1Pic)に置き換えて単予測を行うこともできる。 FIG. 4A shows a case in which the prediction image is uni-prediction and the reference image (RefL0Pic) for L0 prediction is at a time before the encoding target image (CurPic). FIG. 4B shows a case in which the prediction image is a single prediction and the reference image of the L0 prediction is at a time after the encoding target image. Similarly, the L0 prediction reference image shown in FIGS. 4A and 4B can be replaced with the L1 prediction reference image (RefL1Pic) to perform single prediction.
 2つの参照画像からの予測を行う処理を双予測と定義し、双予測の場合はL0予測とL1予測の双方を利用してBI予測と表現する。図4(c)は双予測であってL0予測の参照画像が符号化対象画像より前の時刻にあって、L1予測の参照画像が符号化対象画像より後の時刻にある場合を示している。図4(d)は双予測であってL0予測の参照画像とL1予測の参照画像が符号化対象画像より前の時刻にある場合を示している。このように、L0/L1の予測種別と時間の関係は、L0が過去方向、L1が未来方向とは限定されずに用いることが可能である。また双予測の場合に、同一の参照ピクチャを用いてL0予測及びL1予測のそれぞれを行ってもよい。なお、動き補償予測を単予測で行うか双予測で行うかの判断は、例えばL0予測を利用するか否か及びL1予測を利用するか否かを示す情報(例えば、フラグ)に基づき判断される。 The process of performing prediction from two reference images is defined as bi-prediction, and bi-prediction is expressed as BI prediction using both L0 prediction and L1 prediction. FIG. 4C illustrates a case where bi-prediction is performed, and the reference image for L0 prediction is at a time before the encoding target image and the reference image for L1 prediction is at a time after the encoding target image. . FIG. 4D shows a case of bi-prediction, where the reference image for L0 prediction and the reference image for L1 prediction are at a time before the encoding target image. As described above, the relationship between the prediction type of L0 / L1 and time can be used without being limited to L0 being the past direction and L1 being the future direction. In the case of bi-prediction, each of L0 prediction and L1 prediction may be performed using the same reference picture. Note that whether to perform motion compensation prediction by single prediction or bi-prediction is determined based on, for example, information (for example, a flag) indicating whether to use L0 prediction and whether to use L1 prediction. The
 双予測は2つの参照画像メモリへの画像情報アクセスが必要となるため、単予測と比較して2倍以上のメモリ帯域を必要とする場合がある。ハードウェアを構成する場合、動き補償予測の予測ブロックサイズが小さい場合の双予測がメモリ帯域のボトルネックとなり、本発明の実施の形態ではメモリ帯域のボトルネックを抑制する。 Bi-prediction requires image information access to two reference image memories, and therefore may require twice or more memory bandwidth compared to single prediction. When hardware is configured, bi-prediction when the prediction block size of motion compensation prediction is small becomes a bottleneck of the memory band, and the bottleneck of the memory band is suppressed in the embodiment of the present invention.
 図1に戻り、動き補償予測ブロック構造選択部113は、動きベクトル検出部111より入力される各参照画像に対して検出された動きベクトル値と、予測モード情報メモリ119に格納された動き情報(予測種別、動きベクトル値、及び参照画像指定情報)をもとに、符号化ブロック制御パラメータ生成部122で生成された、実施の形態1において定義した予測ブロックサイズ及び動き補償予測モードに関する制御パラメータが入力され、制御パラメータに基づいて決められた、予測ブロックサイズ及び動き補償予測モードのそれぞれに対して用いられる、参照画像指定情報と動きベクトル値を動き補償予測部112に設定する。設定した値によって、動き補償予測部112から供給される動き補償予測信号と、符号化ブロック取得部102より供給される符号化対象ブロックの画像信号を用いて、最適な予測ブロックサイズと動き補償予測モードを決定する。 Returning to FIG. 1, the motion compensated prediction block structure selection unit 113 detects the motion vector value detected for each reference image input from the motion vector detection unit 111 and the motion information stored in the prediction mode information memory 119 ( Based on the prediction type, motion vector value, and reference image designation information), the control parameters related to the prediction block size and motion compensation prediction mode defined in Embodiment 1 generated by the coding block control parameter generation unit 122 are The reference image designation information and the motion vector value used for each of the prediction block size and the motion compensation prediction mode that are input and determined based on the control parameter are set in the motion compensation prediction unit 112. Depending on the set value, the motion compensation prediction signal supplied from the motion compensation prediction unit 112 and the image signal of the target block to be encoded supplied from the coding block acquisition unit 102 are used to optimize the prediction block size and motion compensation prediction. Determine the mode.
 動き補償予測ブロック構造選択部113は、決定した予測ブロックサイズ、動き補償予測モード、予測モードに応じた予測種別、動きベクトル、及び参照画像指定情報を特定する情報を、動き補償予測信号及び予測誤差に対する誤差評価値と共に予測モード選択部116に供給する。 The motion compensated prediction block structure selection unit 113 uses the determined prediction block size, motion compensation prediction mode, prediction type corresponding to the prediction mode, motion vector, and information specifying the reference image designation information as a motion compensation prediction signal and a prediction error. Is supplied to the prediction mode selection unit 116 together with an error evaluation value for.
 イントラ予測部114は、イントラ予測ブロック構造選択部115より指定される予測ブロック構造を指定する情報と定義されたイントラ予測モードに従って、フレーム内復号画像バッファ108より供給される符号化対象ブロックに隣接した同一フレーム内の復号画像を用いて、イントラ予測信号を生成しイントラ予測ブロック構造選択部115に供給する。 The intra prediction unit 114 is adjacent to the encoding target block supplied from the intra-frame decoded image buffer 108 according to the intra prediction mode defined as the information specifying the prediction block structure specified by the intra prediction block structure selection unit 115. An intra prediction signal is generated using the decoded image in the same frame, and is supplied to the intra prediction block structure selection unit 115.
 イントラ予測ブロック構造選択部115は、予測モード情報メモリ119に格納されたイントラ予測モード情報と複数の定義されたイントラ予測モードに従って、符号化ブロック制御パラメータ生成部122で生成された、実施の形態1において定義した予測ブロックサイズに関する制御パラメータが入力され、制御パラメータに基づいて決められた予測ブロックサイズのそれぞれに対して用いられるイントラ予測モードをイントラ予測部114に設定する。設定した値によって、イントラ予測部114から供給されるイントラ予測信号と、符号化ブロック取得部102より供給される符号化対象ブロックの画像信号を用いて、最適な予測ブロックサイズとイントラ予測モードを決定する。 Embodiment 1 Intra prediction block structure selection section 115 is generated by coding block control parameter generation section 122 according to intra prediction mode information stored in prediction mode information memory 119 and a plurality of defined intra prediction modes. The control parameter related to the prediction block size defined in the step is input, and the intra prediction mode used for each of the prediction block sizes determined based on the control parameter is set in the intra prediction unit 114. Based on the set value, the optimal prediction block size and intra prediction mode are determined using the intra prediction signal supplied from the intra prediction unit 114 and the image signal of the encoding target block supplied from the coding block acquisition unit 102. To do.
 また、イントラ予測ブロック構造選択部115は、決定した予測ブロックサイズ、イントラ予測モードを特定する情報を、イントラ予測信号及び予測誤差に対する誤差評価値と共に予測モード選択部116に供給する。 Also, the intra prediction block structure selection unit 115 supplies information specifying the determined prediction block size and intra prediction mode to the prediction mode selection unit 116 together with the intra prediction signal and the error evaluation value for the prediction error.
 予測モード選択部116は、動き補償予測ブロック構造選択部113より供給される、決定した予測ブロックサイズ、動き補償予測モード、予測モードに応じた予測種別、動きベクトル、参照画像指定情報を特定する情報、及び予測誤差に対する誤差評価値と、イントラ予測ブロック構造選択部115より供給される、決定した予測ブロックサイズ、イントラ予測モード、及び予測誤差に対する誤差評価値より、階層的に構成されるCUサイズ単位の最適な予測モードを、誤差評価値を比較し選択する。 The prediction mode selection unit 116 is supplied from the motion compensated prediction block structure selection unit 113 and specifies the determined prediction block size, motion compensation prediction mode, prediction type according to the prediction mode, motion vector, and reference image designation information. CU size units that are hierarchically configured from the error evaluation value for the prediction error, and the error prediction value for the predicted prediction block size, intra prediction mode, and prediction error supplied from the intra prediction block structure selection unit 115 The optimum prediction mode is selected by comparing the error evaluation values.
 予測モード選択部116により選択された、CUサイズ単位の最適な予測モード情報として、予測ブロックサイズ、予測信号、誤差評価値のCUサイズ単位の総和と共に、動き補償予測が選択された場合には、動き補償予測モード、予測モードに応じた予測種別、動きベクトル、参照画像指定情報を特定する情報、及び動き補償予測信号が、イントラ予測が選択された場合には、イントラ予測モード、及びイントラ予測信号が、符号化ブロック構造選択部117に供給される。 When motion compensation prediction is selected as the optimal prediction mode information in units of CU size selected by the prediction mode selection unit 116 together with the sum of the prediction block size, the prediction signal, and the error evaluation value in units of CU size, When the intra prediction is selected as the motion compensation prediction mode, the prediction type corresponding to the prediction mode, the motion vector, the information specifying the reference image designation information, and the motion compensation prediction signal, the intra prediction mode and the intra prediction signal are selected. Is supplied to the coding block structure selection unit 117.
 符号化ブロック構造選択部117は、予測モード選択部116より供給されたCUサイズ単位の最適な予測モード情報を元に、符号化ブロック制御パラメータ生成部122で生成された、実施の形態1において定義した符号化ブロックサイズに関する制御パラメータが入力され、制御パラメータに基づいて決められた符号化ブロックサイズ構成において最適なCU_Depth構成を選択し、CU分割構成を指定する情報と、指定された分割構成毎のCUサイズにおける最適な予測モード情報と予測モードに関連する付加情報(動き情報、イントラ予測モード)を、ブロック構造/予測モード情報付加情報符号化部118に供給すると共に、選択した予測信号を減算部103及び加算部107に供給する。 The coding block structure selection unit 117 is defined in Embodiment 1 generated by the coding block control parameter generation unit 122 based on the optimal prediction mode information in CU size units supplied from the prediction mode selection unit 116. The control parameter related to the encoded block size is input, the optimum CU_Depth configuration is selected in the encoded block size configuration determined based on the control parameter, the information for specifying the CU partition configuration, and the specified partition configuration The optimal prediction mode information in the CU size and additional information related to the prediction mode (motion information, intra prediction mode) are supplied to the block structure / prediction mode information additional information encoding unit 118 and the selected prediction signal is subtracted. 103 and the adder 107.
 ブロック構造/予測モード情報付加情報符号化部118は、符号化ブロック構造選択部117より供給された、CU分割構成を指定する情報と、指定した分割構成毎のCUサイズにおける最適な予測モード情報と予測モードに関連する付加情報と、符号化ブロック制御パラメータ生成部122より供給された、符号化ブロック及び予測ブロック構造に関する制御パラメータを、所定のシンタックス構造に従って符号化することで、符号化ブロック単位のCU分割構成と予測に用いられたモード情報を符号化し、多重化部120に供給すると共に、これらの情報を予測モード情報メモリ119に格納する。 The block structure / prediction mode information additional information encoding unit 118 is supplied from the encoding block structure selection unit 117, and specifies the CU partition configuration, and the optimal prediction mode information for the specified CU size for each partition configuration. By coding the additional information related to the prediction mode and the control parameters related to the coding block and the prediction block structure supplied from the coding block control parameter generation unit 122 according to a predetermined syntax structure, the coding block unit The CU partition configuration and mode information used for prediction are encoded and supplied to the multiplexing unit 120, and the information is stored in the prediction mode information memory 119.
 予測モード情報メモリ119は、ブロック構造/予測モード情報付加情報符号化部118より供給される符号化ブロック単位のCU分割構成と予測に用いられたモード情報を、最小予測ブロックサイズ単位を基準として所定画像分記憶する。実施の形態1は画面間の予測である動き補償予測に注目したものであるため、モード情報における動き補償予測に関連する情報である動き情報(予測種別、動きベクトル、及び参照画像インデックス)に対して、説明を加える。 The prediction mode information memory 119 predetermines the CU partition configuration of the coding block unit supplied from the block structure / prediction mode information additional information encoding unit 118 and the mode information used for prediction based on the minimum prediction block size unit. Memorize images. Since the first embodiment focuses on motion compensation prediction that is prediction between screens, motion information (prediction type, motion vector, and reference image index) that is information related to motion compensation prediction in mode information is used. And add a description.
 動き補償予測の処理対象である予測ブロックの隣接ブロックの動き情報を空間候補ブロック群とし、処理対象の予測ブロックと同一位置にあるColPic上のブロックとその周辺ブロックの動き情報を時間候補ブロック群とする。 The motion information of the adjacent block of the prediction block that is the processing target of motion compensation prediction is set as a spatial candidate block group, and the motion information of the block on ColPic and the surrounding blocks that are at the same position as the processing target prediction block is the time candidate block group. To do.
 ColPicとは、処理対象の予測ブロックとは別の復号済みの画像であって、復号画像メモリ110に参照画像として記憶されている。実施の形態1では、ColPicは直前に復号した参照画像とする。なお、実施の形態1では、ColPicは直前に復号した参照画像としたが、表示順で直前の参照画像や表示順で直後の参照画像でもよく、符号化ストリーム中に、ColPicに用いる参照画像を直接指定することも可能である。 ColPic is a decoded image different from the prediction block to be processed, and is stored in the decoded image memory 110 as a reference image. In Embodiment 1, ColPic is a reference image decoded immediately before. In the first embodiment, ColPic is the reference image decoded immediately before, but the reference image immediately before in display order or the reference image immediately after in display order may be used, and the reference image used for ColPic is included in the encoded stream. Direct specification is also possible.
 予測モード情報メモリ119は、空間候補ブロック群と時間候補ブロック群の動き情報を、候補ブロック群の動き情報として動き補償予測ブロック構造選択部113に供給すると共に、イントラ予測ブロックの隣接ブロックのイントラ予測モード情報をイントラ予測ブロック構造選択部115に供給する。 The prediction mode information memory 119 supplies the motion information of the spatial candidate block group and the temporal candidate block group to the motion compensated prediction block structure selection unit 113 as motion information of the candidate block group, and intra prediction of adjacent blocks of the intra prediction block. The mode information is supplied to the intra prediction block structure selection unit 115.
 多重化部120は、予測誤差符号化部105から供給される予測誤差の符号化列と、ブロック構造/予測モード情報付加情報符号化部118から供給される符号化ブロック単位のCU分割構成と予測に用いられたモード情報及び付加情報の符号化列を多重化することで符号化ビットストリームを生成し、出力端子121経由で、記録媒体・伝送路等に当該符号化ビットストリームを出力する。 The multiplexing unit 120 includes a prediction error encoding sequence supplied from the prediction error encoding unit 105, a CU partitioning configuration in units of encoded blocks supplied from the block structure / prediction mode information additional information encoding unit 118, and prediction. The encoded bit stream is generated by multiplexing the encoded sequence of the mode information and the additional information used in the above, and the encoded bit stream is output to the recording medium / transmission path via the output terminal 121.
 符号化ブロック制御パラメータ生成部122は、実施の形態1における、符号化ブロック構造を定義するパラメータである、図3に示すMaximum_cu_sizeやMinimum_cu_size等の制御パラメータや、動き補償予測のブロックサイズ及び予測処理を制限する制御パラメータ等の、符号化ブロック構造や予測ブロック構造を定義するためのパラメータを生成し、動き補償予測ブロック構造選択部113、イントラ予測ブロック構造選択部115、符号化ブロック構造選択部117、及びブロック構造/予測モード情報付加情報符号化部118に供給する。動き補償予測のブロックサイズ及び予測処理を制限する制御パラメータに関する詳細については後述する。 The coding block control parameter generation unit 122 performs control parameters such as Maximum_cu_size and Minimum_cu_size shown in FIG. 3, which are parameters defining the coding block structure, and the block size and prediction processing of motion compensation prediction in the first embodiment. Generate parameters for defining a coding block structure or a prediction block structure, such as a control parameter to be limited, and a motion compensation prediction block structure selection unit 113, an intra prediction block structure selection unit 115, a coding block structure selection unit 117, And supplied to the block structure / prediction mode information additional information encoding unit 118. Details regarding the block size of motion compensation prediction and control parameters for limiting the prediction process will be described later.
 図1に示した動画像符号化装置の構成は、CPU(Central Processing Unit)、フレームメモリ、ハードディスクなどを備える情報処理装置等のハードウェアによっても実現可能である。 The configuration of the moving picture encoding apparatus shown in FIG. 1 can also be realized by hardware such as an information processing apparatus including a CPU (Central Processing Unit), a frame memory, and a hard disk.
 図5は、本発明の実施の形態1に係る動画像符号化装置における符号化処理の動作の流れを示すフローチャートである。符号化ブロック単位毎に、CU分割の制御パラメータであるCU_Depthを0に初期化(S500)し、符号化ブロック取得部102より符号化処理対象ブロック画像を取得する(S501)。動きベクトル検出部111は、符号化対象ブロック画像よりCU分割に応じた予測対象のブロック画像と復号画像メモリ110に格納された複数の参照画像より、CU分割に応じた参照画像毎の動きベクトル値を算出する(S502)。 FIG. 5 is a flowchart showing the flow of the encoding process in the video encoding apparatus according to Embodiment 1 of the present invention. For each coding block unit, CU_Depth, which is a control parameter for CU partitioning, is initialized to 0 (S500), and a coding process target block image is obtained from the coding block acquisition unit 102 (S501). The motion vector detection unit 111 uses a motion vector value for each reference image according to CU partitioning from a block image to be predicted according to CU partitioning from the encoding target block image and a plurality of reference images stored in the decoded image memory 110. Is calculated (S502).
 続いて、動き補償予測ブロック構造選択部113は、動きベクトル検出部111より供給される動きベクトルと、予測モード情報メモリ119に格納された動き情報及びイントラ予測モード情報を用いて、実施の形態1において定義した予測ブロックサイズ、動き補償予測モードのそれぞれに対する予測信号を、動き補償予測部112を用いて取得し、最適なCU単位の予測ブロックサイズ及び予測モードを選択した結果を出力する。また、イントラ予測ブロック構造選択部115は、予測ブロックサイズ、イントラ予測モードのそれぞれに対する予測信号を、イントラ予測部114を用いて取得し、最適なCU単位の予測ブロックサイズ及び予測モードを選択した結果を出力する。符号化ブロック構造選択部117は、これらの結果を用いて最適な符号化ブロック構造における予測モードと予測信号を生成する(S503)。ステップS503の処理の詳細については後述する。 Subsequently, the motion compensation prediction block structure selection unit 113 uses the motion vector supplied from the motion vector detection unit 111, the motion information and the intra prediction mode information stored in the prediction mode information memory 119, and performs the first embodiment. The prediction signal for each of the prediction block size and the motion compensation prediction mode defined in (1) is acquired using the motion compensation prediction unit 112, and the result of selecting the optimal prediction block size and prediction mode in CU units is output. In addition, the intra prediction block structure selection unit 115 acquires prediction signals for each of the prediction block size and the intra prediction mode using the intra prediction unit 114, and selects the optimal prediction block size and prediction mode in CU units. Is output. The coding block structure selection unit 117 generates a prediction mode and a prediction signal in the optimum coding block structure using these results (S503). Details of the processing in step S503 will be described later.
 続いて、減算部103は、符号化ブロック取得部102より供給された符号化処理ブロック画像と、符号化ブロック構造選択部117より供給された予測信号との差分を予測誤差信号として算出する(S504)。ブロック構造/予測モード情報付加情報符号化部118は、符号化ブロック構造選択部117より供給される符号化構造、予測モード、動き補償予測の場合の予測モードに応じた予測種別、動きベクトル、及び参照画像指定情報を特定する情報、イントラ予測の場合のイントラ予測モード情報を、所定のシンタックス構造に従って符号化し、符号化構造と予測モード情報に関連する付加情報の符号化データを生成する(S505)。 Subsequently, the subtraction unit 103 calculates a difference between the encoded block image supplied from the encoded block acquisition unit 102 and the prediction signal supplied from the encoded block structure selection unit 117 as a prediction error signal (S504). ). The block structure / prediction mode information additional information encoding unit 118 includes a coding type, a prediction mode, a prediction type according to a prediction mode in the case of motion compensation prediction, a motion vector, and a coding structure supplied from the coding block structure selection unit 117. Information for specifying the reference image designation information and intra prediction mode information in the case of intra prediction are encoded according to a predetermined syntax structure, and encoded data of additional information related to the encoding structure and the prediction mode information is generated (S505). ).
 続いて、予測誤差符号化部105は、直交変換・量子化部104で生成された量子化された予測誤差信号をエントロピー符号化して、予測誤差の符号化データを生成する(S506)。多重化部120は、ブロック構造/予測モード情報付加情報符号化部118から供給される符号化構造と予測モード情報に関連する付加情報の符号化データと、予測誤差符号化部105から供給される予測誤差の符号化データを多重化し、符号化ビットストリームを生成する(S507)。 Subsequently, the prediction error encoding unit 105 entropy encodes the quantized prediction error signal generated by the orthogonal transform / quantization unit 104 to generate encoded data of the prediction error (S506). The multiplexing unit 120 is supplied from the coding structure supplied from the block structure / prediction mode information additional information encoding unit 118, encoded data of additional information related to the prediction mode information, and the prediction error encoding unit 105. The encoded data of the prediction error is multiplexed to generate an encoded bit stream (S507).
 加算部107は、逆量子化・逆変換部106より供給される復号予測誤差信号と、符号化ブロック構造選択部117より供給される予測信号を加算して、復号画像信号を生成する(S508)。予測モード情報メモリ119は、ブロック構造/予測モード情報付加情報符号化部118より供給される符号化構造と予測モード情報に関連する付加情報として、動き補償予測が用いられた場合の動き情報(予測種別、動きベクトル、及び参照画像指定情報)と、イントラ予測が用いられた場合のイントラ予測モード情報を、最小の予測ブロックサイズ単位で格納する(S509)。 The addition unit 107 adds the decoded prediction error signal supplied from the inverse quantization / inverse conversion unit 106 and the prediction signal supplied from the coding block structure selection unit 117 to generate a decoded image signal (S508). . The prediction mode information memory 119 includes motion information (prediction when motion compensation prediction is used as additional information related to the coding structure and prediction mode information supplied from the block structure / prediction mode information additional information encoding unit 118. Type, motion vector, and reference image designation information) and intra prediction mode information when intra prediction is used are stored in units of the smallest prediction block size (S509).
 加算部107によって、生成された復号画像信号はフレーム内復号画像バッファ108に格納されると共に、ループフィルタ部109において、歪除去のためのループフィルタ処理が施され(S510)、フィルタを施された復号画像信号が復号画像メモリ110に供給、格納され、以降に符号化する符号化画像の動き補償予測処理に用いられる(S511)。 The decoded image signal generated by the addition unit 107 is stored in the intra-frame decoded image buffer 108, and the loop filter unit 109 performs a loop filter process for distortion removal (S510) and performs the filter. The decoded image signal is supplied to and stored in the decoded image memory 110 and used for motion compensation prediction processing of an encoded image to be encoded thereafter (S511).
 [CU単位の予測モード/予測信号生成処理の詳細]
 次に、図5のフローチャートにおけるステップS503であるCU単位の予測モード/予測信号生成処理の詳細について、図6のフローチャートを用いて説明する。
[Details of prediction mode / prediction signal generation processing per CU]
Next, details of the prediction mode / prediction signal generation processing in units of CUs, which is step S503 in the flowchart of FIG. 5, will be described using the flowchart of FIG.
 最初に、設定された最大CUサイズと最小CUサイズの間の階層数を示す値をMax_CU_Depthとして、対象CUのCU_DepthがMax_CU_Depthより小さいか否かを判定する(S600)。実施の形態1においては、図3に示すCU分割構成を取るものとし、Max_CU_Depth=3とする。 First, a value indicating the number of hierarchies between the set maximum CU size and minimum CU size is set as Max_CU_Depth, and it is determined whether or not the CU_Depth of the target CU is smaller than Max_CU_Depth (S600). In the first embodiment, the CU division configuration shown in FIG. 3 is assumed, and Max_CU_Depth = 3.
 CU_DepthがMax_CU_Depthより小さい場合には(S600:YES)、CU_Depthを1加算して(S601)、現在の対象CUを4分割した一階層下のCUに対する、CU単位の予測モード/予測信号生成処理を行う(S602-S605)。図2で示したCUの分割領域に対して、分割1領域の処理(S602)、分割2領域の処理(S603)、分割3領域の処理(S604)、分割4領域の処理(S605)の順で、再帰的に図6のフローチャートで説明されるCU単位の予測モード/予測信号生成処理が行われる。 When CU_Depth is smaller than Max_CU_Depth (S600: YES), 1 is added to CU_Depth (S601), and the prediction mode / prediction signal generation processing in units of CUs is performed for the CU one layer below the current target CU divided into four. Perform (S602-S605). In the order of the division 1 area processing (S602), the division 2 area processing (S603), the division 3 area processing (S604), and the division 4 area processing (S605) with respect to the CU division area shown in FIG. Thus, the prediction mode / prediction signal generation processing for each CU described recursively in the flowchart of FIG. 6 is performed.
 各CU分割領域の予測モード算出結果の内、誤差評価値が積算され、4つの分割CUの誤差評価値総和が算出される(S606)。 Among the prediction mode calculation results of each CU divided region, error evaluation values are integrated, and a total error evaluation value of four divided CUs is calculated (S606).
 一方、CU_DepthがMax_CU_Depth以上の場合(S600:NO)には、図1のイントラ予測ブロック構造選択部115及びイントラ予測部114において、イントラ予測モードの算出と予測信号の生成が行われ(S607)、対象CUにおけるイントラ予測のモード情報、予測信号と誤差評価値が算出される。 On the other hand, when CU_Depth is equal to or greater than Max_CU_Depth (S600: NO), intra prediction block structure selection unit 115 and intra prediction unit 114 in FIG. 1 calculate intra prediction modes and generate prediction signals (S607). Intra prediction mode information, a prediction signal, and an error evaluation value in the target CU are calculated.
 続いて、動き補償予測ブロック構造選択部113及び動き補償予測部112において、動き補償予測ブロックサイズの選択と、選択された予測ブロック単位の動き補償予測モード及び予測信号生成が行われ(S608)、対象CUにおける動き補償予測の予測ブロックサイズ、モード情報、動き情報、予測信号と誤差評価値が算出される。ステップS608の詳細に関しては、後述する。 Subsequently, the motion compensation prediction block structure selection unit 113 and the motion compensation prediction unit 112 select a motion compensation prediction block size, and generate a motion compensation prediction mode and a prediction signal for each selected prediction block (S608). A prediction block size, mode information, motion information, a prediction signal, and an error evaluation value for motion compensation prediction in the target CU are calculated. Details of step S608 will be described later.
 続いて、符号化ブロック構造選択部117は、対象CUにおけるイントラ予測の誤差評価値と、動き補償予測の誤差評価値を比較して、誤差の少ない予測手法を選択しイントラ/インター(動き補償予測)の判定を行う(S609)。 Subsequently, the coding block structure selecting unit 117 compares the error evaluation value of the intra prediction in the target CU with the error evaluation value of the motion compensated prediction, selects a prediction method with a small error, and selects intra / inter (motion compensated prediction). ) Is determined (S609).
 次に、再帰的に施された図6のフローチャートの処理(図6のS602-S605)と誤差評価値の総和算出(S606)によって生成された、対象CUよりも下位階層の(CU_Depthが大きい)CUに対する誤差評価値と、対象CUの誤差評価値を比較し、予測に適用するCU_Depthの判定を行う(S610)。 Next, the recursively performed processing of the flowchart of FIG. 6 (S602 to S605 of FIG. 6) and the error evaluation value summation calculation (S606), which is lower than the target CU (CU_Depth is greater) The error evaluation value for the CU is compared with the error evaluation value of the target CU, and CU_Depth applied to prediction is determined (S610).
 再帰的に図6のフローチャートで示される処理が呼び出されるために、最も下位(CU_Depth=Max_CU_Depth)のCUから上位のCUに対して、順次比較が行われ、CUの分割領域毎の最適なCU_Depthと予測モードが選択できる。 Since the processing shown in the flowchart of FIG. 6 is recursively called, the lowest CU (Depth = Max_CU_Depth) CU is sequentially compared with the upper CU, and the optimal CU_Depth for each divided region of the CU Prediction mode can be selected.
 最後に、選択された、対象CUと対象CUより下位のCUの間での最適CU_Depth、予測モード及び、選択されたイントラ予測または動き補償予測に関する付加情報と、誤差評価値及び予測信号が格納され(S611)、対象CUにおける予測モード/予測信号生成処理が終了する。 Finally, additional information regarding the selected CU_Depth, prediction mode, and selected intra prediction or motion compensated prediction, the error evaluation value, and the prediction signal between the target CU and the CU lower than the target CU are stored. (S611) The prediction mode / prediction signal generation process in the target CU ends.
 [動き補償予測ブロックサイズ選択/予測信号生成処理の詳細]
 次に、図6のフローチャートにおけるステップS608である対象CUにおける動き補償予測ブロックサイズ選択及び、予測ブロック単位の動き補償予測モード/予測信号生成処理の詳細について、図7のフローチャートを用いて説明する。
[Details of motion compensation prediction block size selection / prediction signal generation processing]
Next, details of the motion compensated prediction block size selection in step 608 in the flowchart of FIG. 6 and the motion compensation prediction mode / prediction signal generation processing for each prediction block will be described with reference to the flowchart of FIG.
 最初に、対象CUに対して予測対象となる符号化ブロック画像を取得する(S700)。次に、図3に示した構成により、CU内分割モード毎の動き補償予測モード/予測信号生成処理を行う(S701~S705)。 First, an encoded block image to be predicted is acquired for the target CU (S700). Next, with the configuration shown in FIG. 3, motion compensation prediction mode / prediction signal generation processing is performed for each intra-CU division mode (S701 to S705).
 先ず、CU内分割モードが2N×2Nの場合の動き補償予測モード/予測信号生成処理を、分割数を示す値であるNumPartを1に設定して行う(S701)。続いて、NumPartを2に設定して、2N×Nの場合(S702)、N×2Nの場合(S703)の動き補償予測モード/予測信号生成処理を行う。 First, the motion compensation prediction mode / prediction signal generation processing when the intra-CU division mode is 2N × 2N is performed by setting NumPart which is a value indicating the number of divisions to 1 (S701). Subsequently, NumPart is set to 2, and motion compensation prediction mode / prediction signal generation processing is performed in the case of 2N × N (S702) and N × 2N (S703).
 次に、CU_DepthがMax_CU_Depthと等しく且つ、対象CUサイズが8×8で、後述するinter_4x4_enableフラグが1である場合(S704:YES)、NumPartを4に設定して、N×Nの場合の動き補償予測モード/予測信号生成処理を行う(S705)。ステップS701、S702、S703、S705で施される、動き補償予測モード/予測信号生成処理の詳細は後述する。ステップS704の条件を満たさない場合(S704:NO)には、ステップS705をスキップし後続するステップが施される。 Next, when CU_Depth is equal to Max_CU_Depth, the target CU size is 8 × 8, and an inter_4x4_enable flag (to be described later) is 1 (S704: YES), NumPart is set to 4 and motion compensation is performed when N × N Prediction mode / prediction signal generation processing is performed (S705). Details of the motion compensation prediction mode / prediction signal generation processing performed in steps S701, S702, S703, and S705 will be described later. If the condition of step S704 is not satisfied (S704: NO), step S705 is skipped and the subsequent step is performed.
 実施の形態1においては、2N×2N(S701)、2N×N(S702)、N×2N(S703)、及びN×N(S705)の順で、CU内分割における動き補償予測/予測信号生成を行っているが、上記CU分割それぞれのステップの処理順に関しては、順序が変更されても構わず、また並列処理を可能とするCPU等で処理を施す場合に、S701、S702、S703及びS705を並列に行うことも可能である。 In Embodiment 1, motion compensated prediction / prediction signal generation in intra-CU division in the order of 2N × 2N (S701), 2N × N (S702), N × 2N (S703), and N × N (S705) However, the processing order of the steps of each of the CU divisions may be changed, and when processing is performed by a CPU or the like that can perform parallel processing, S701, S702, S703, and S705 are performed. Can also be performed in parallel.
 続いて、動き補償予測モード/予測信号生成を行った、CU内分割モード毎の誤差評価値を比較し、最適なCU内分割モードである最適予測ブロックサイズ(PU)を選択する(S706)。選択されたPUに対する予測モード情報/誤差評価値/予測信号が格納され(S707)、図6のフローチャートにおけるステップS608の処理が終了する。 Subsequently, an error evaluation value for each intra-CU partition mode for which motion compensation prediction mode / prediction signal generation has been performed is compared, and an optimal prediction block size (PU) that is an optimal intra-CU partition mode is selected (S706). Prediction mode information / error evaluation value / prediction signal for the selected PU is stored (S707), and the process of step S608 in the flowchart of FIG. 6 ends.
 [実施の形態1における動き補償予測モードの定義]
 図8(a)、(b)は、本発明の実施の形態1における動き補償予測において使用される動き情報を符号化するための2つの予測モードを説明するための図である。
[Definition of Motion Compensation Prediction Mode in Embodiment 1]
FIGS. 8A and 8B are diagrams for explaining two prediction modes for encoding motion information used in motion compensated prediction according to Embodiment 1 of the present invention.
 第一の予測モードは、予測対象ブロックと当該予測対象ブロックに隣接する符号化済ブロックにおける時間方向や空間方向の動きの連続性を用いて、当該予測対象ブロックは自身の動き情報を直接符号化せずに、空間及び時間的に隣接するブロックの動き情報を符号化に使用する手法であり、結合予測モード(マージモード)と呼ぶ。 In the first prediction mode, the prediction target block directly encodes its own motion information using the continuity of motion in the temporal direction and the spatial direction in the prediction target block and the encoded block adjacent to the prediction target block. In this method, the motion information of spatially and temporally adjacent blocks is used for encoding, which is called a joint prediction mode (merge mode).
 ここで、空間的に隣接するブロックとは予測対象ブロックと同じ画像に属する符号化済みブロックの中で、予測対象ブロックに隣接するブロックを指す。ここで、時間的に隣接するブロックとは予測対象ブロックとは別の符号化済みの画像に属するブロックの中で、予測対象ブロックと同一空間位置及びその周辺にあるブロックを指す。 Here, the spatially adjacent block refers to a block adjacent to the prediction target block among encoded blocks belonging to the same image as the prediction target block. Here, the temporally adjacent blocks indicate blocks in the same spatial position as the prediction target block and in the vicinity thereof among blocks belonging to an encoded image different from the prediction target block.
 結合予測モードの場合には、複数の隣接ブロック候補より選択的に結合する動き情報が定義でき、動き情報は使用する隣接ブロックを指定する情報(結合動き情報インデックス)を符号化することで、指定情報をもとに取得した動き情報をそのまま動き補償予測に用いる。更に、結合予測モードにおいては、予測差分情報を符号化伝送せずに、結合予測モードで予測された予測信号を復号ピクチャとするSkipモードを定義し、結合した動き情報のみの少ない情報で復号画像が再生できる構成を有する。SkipモードはCU内分割モードが2N×2Nの場合に用いることが可能であり、Skipモードにおいて伝送する動き情報は、結合予測モードと同様に隣接ブロックを定義する指定情報となる。 In the combined prediction mode, motion information that can be selectively combined from a plurality of adjacent block candidates can be defined, and the motion information is specified by encoding information (joined motion information index) that specifies the adjacent block to be used. The motion information acquired based on the information is used as it is for motion compensation prediction. Further, in the joint prediction mode, a Skip mode is defined in which the prediction signal predicted in the joint prediction mode is a decoded picture without encoding prediction transmission of the prediction difference information, and a decoded image is obtained with information having only the combined motion information. Can be reproduced. The Skip mode can be used when the intra-CU division mode is 2N × 2N, and the motion information transmitted in the Skip mode is the designation information that defines the adjacent block as in the combined prediction mode.
 第二の予測モードは、動き情報の構成要素を個別にすべて符号化し、予測ブロックに対して予測誤差の少ない動き情報を伝送する手法であり、動き検出予測モードと呼ぶ。動き検出予測モードは、従来の動き補償予測の動き情報の符号化と同様に、双予測であるか単予測であるかを示す予測種別、参照画像を特定するための情報(参照画像インデックス)と、動きベクトルを特定するための情報が別々に符号化される。 The second prediction mode is a technique for coding all the components of motion information individually and transmitting motion information with little prediction error to the prediction block, and is called a motion detection prediction mode. The motion detection prediction mode includes a prediction type indicating whether the prediction is bi-prediction or uni-prediction, information for identifying a reference image (reference image index), and encoding of motion information in the conventional motion compensation prediction. The information for specifying the motion vector is encoded separately.
 動き検出予測モードには、単予測と双予測のどちらを使用するか予測モードで指示し、単予測単予測の場合には1つの参照画像に対する参照画像を特定する情報と、動きベクトルの予測ベクトルとの差分ベクトルを符号化する。双予測の場合には2つの参照画像に対する参照画像を特定する情報と、動きベクトルがそれぞれ個別に符号化される。動きベクトルに対する予測ベクトルは、AVCと同様に隣接ブロックの動き情報から生成されるが、結合予測モードと同様に、複数の隣接ブロック候補より予測ベクトルに用いる動きベクトルを選択でき、動きベクトルは予測ベクトルに使用する隣接ブロックを指定する情報(予測ベクトルインデックス)と差分ベクトルの2つを符号化することで伝送される。 In the motion detection prediction mode, the prediction mode indicates whether to use single prediction or bi-prediction. In the case of single prediction single prediction, information for specifying a reference image for one reference image, and a motion vector prediction vector The difference vector is encoded. In the case of bi-prediction, information for specifying reference images for two reference images and a motion vector are individually encoded. The prediction vector for the motion vector is generated from the motion information of the adjacent block similarly to the AVC. However, similarly to the combined prediction mode, the motion vector used for the prediction vector can be selected from a plurality of adjacent block candidates, and the motion vector is the prediction vector. Is transmitted by encoding two pieces of information (predicted vector index) for designating adjacent blocks to be used for and a difference vector.
 [実施の形態1における、動き補償予測のブロックサイズ及び予測処理を制限する手法に関する説明]
 次に、動き補償予測における予測時に必要な参照画像メモリ量に関する概算値を図9に示し、実施の形態1における予測ブロックサイズと予測処理の制限手法の説明を行う。動き補償予測においては、動きの精度を細かくすることで予測精度を向上させており、AVCを例にとると、1/4画素精度で動きベクトルを検出・伝送することが可能となっている。
[Explanation Regarding Block Size of Motion Compensated Prediction and Method for Limiting Prediction Processing in Embodiment 1]
Next, an approximate value related to the reference image memory amount required at the time of prediction in motion compensation prediction is shown in FIG. 9, and the prediction block size and the prediction processing restriction method in Embodiment 1 will be described. In motion compensation prediction, prediction accuracy is improved by making the accuracy of motion fine. Taking AVC as an example, motion vectors can be detected and transmitted with 1/4 pixel accuracy.
 実施の形態1においても、1/4画素精度で動きベクトルを検出・伝送する構成を取り、1/4画素精度の動きに対する動き補償予測信号を生成する際には、参照画像に存在する整数動き位置の画素を複数画素用いて、補間フィルタにより1/4画素精度の動き位置の参照画像の画素を算出する。実施の形態1における動画像符号化装置、動画像復号装置においては、補間フィルタとして7タップのFIRフィルタを用いる。 Also in the first embodiment, when a motion vector is detected and transmitted with ¼ pixel accuracy and a motion compensated prediction signal for a ¼ pixel accuracy motion is generated, an integer motion existing in the reference image is generated. Using a plurality of pixels at the position, a pixel of the reference image at the motion position with a 1/4 pixel accuracy is calculated by an interpolation filter. In the moving picture encoding apparatus and moving picture decoding apparatus according to Embodiment 1, a 7-tap FIR filter is used as an interpolation filter.
 7タップのフィルタを施す為には、対象となる位置に最も近い水平・垂直の整数動き位置の画素に対して、水平および垂直にプラスマイナス6画素の画素を取得する必要がある。予測ブロックの右境界部分において3/4画素離れた動き位置の予測画像を取得する際には、対象となる位置に最も近い整数動き位置の画素が、予測ブロックの1画素外に属する画素となるため、更に1画素取得する画素が増加し、予測ブロックサイズに対して、水平および垂直にタップ数と同じ7画素分のフィルタ処理に必要な参照画像取得が必要となる。 In order to apply a 7-tap filter, it is necessary to acquire 6 pixels in the horizontal and vertical directions with respect to the pixel at the horizontal / vertical integer motion position closest to the target position. When a predicted image at a motion position that is 3/4 pixels away from the right boundary portion of the prediction block is acquired, a pixel at an integer motion position that is closest to the target position is a pixel that is outside one pixel of the prediction block. Therefore, the number of pixels to be acquired further increases, and it is necessary to acquire a reference image necessary for filtering processing for the same 7 pixels as the number of taps horizontally and vertically with respect to the predicted block size.
 図9は、7タップフィルタを施す場合の、実施の形態1における図3で示した動き補償予測の定義可能なそれぞれの予測ブロックサイズにおいて、単予測及び双予測を行う際にメモリ帯域として確保が必要な、参照画像のメモリアクセス量を示す。符号化装置及び復号装置の参照画像メモリの構成によっては、メモリアクセスが水平4画素単位で可能な構成や、水平・垂直2×2画素単位で可能な構成など、様々な構成を取ることができるが、上記メモリアクセス量は、参照画像メモリの構成に関わらず最小限取得が必要なメモリアクセス量の最大値を示している。 FIG. 9 shows a case where a 7-tap filter is applied and a memory band is secured when performing uni-prediction and bi-prediction in each of the predictable block sizes definable for motion compensated prediction shown in FIG. 3 in the first embodiment. Indicates the required memory access amount of the reference image. Depending on the configuration of the reference image memory of the encoding device and the decoding device, various configurations such as a configuration in which memory access is possible in units of horizontal 4 pixels and a configuration in which units of horizontal and vertical 2 × 2 pixels are possible can be taken. However, the memory access amount indicates the maximum value of the memory access amount that needs to be obtained at the minimum regardless of the configuration of the reference image memory.
 予測ブロックサイズの大きさに関わらず、フィルタ処理の為に追加で取得が必要となる水平・垂直のサイズは変わらないため、4×4画素サイズの場合が最も符号化ブロックサイズ(LCU)単位でのメモリアクセス量は大きくなり、64×64画素サイズの6倍近くのアクセスが必要となる。また、双予測の動き補償予測の場合には、異なる位置の参照画像から2つの予測信号を取得するため、単予測の2倍のメモリアクセスが必要となる。 Regardless of the size of the predicted block size, the horizontal and vertical sizes that need to be additionally acquired for the filtering process do not change, so the 4 × 4 pixel size is the most encoded block size (LCU) unit. The memory access amount becomes larger, and access of nearly 6 times the size of 64 × 64 pixels is required. In addition, in the case of bi-predictive motion compensated prediction, two prediction signals are acquired from reference images at different positions, so that twice as many memory accesses as in single prediction are required.
 動き補償予測のブロックサイズが小さい場合や、双予測の動き補償の場合に確保する必要があるメモリ帯域は、特に符号化する画像サイズが大きくなりハイビジョン以上の高精細画像になる場合に大きくなり、符号化装置及び復号装置の実現性が困難になる課題がある。本発明においては、メモリ帯域を制限するための、参照画像のメモリアクセス最大量が段階的に制御できる、動き補償予測の制限手法と制限するための制御パラメータの定義及び設定手法を提供し、高精細画像における動画像符号化装置の実現性と符号化効率を両立することを可能とする。 The memory bandwidth that needs to be secured when the block size for motion compensation prediction is small or when motion compensation for bi-prediction is large, especially when the image size to be encoded becomes high definition images that are higher than HD. There is a problem that the feasibility of the encoding device and the decoding device becomes difficult. In the present invention, there are provided a motion compensation prediction limiting method and a control parameter definition and setting method for limiting, in which the memory access maximum amount of the reference image can be controlled step by step to limit the memory bandwidth, It is possible to achieve both the feasibility and encoding efficiency of a moving image encoding apparatus for fine images.
 続いて、図10に本発明の実施の形態1における、図1の符号化ブロック制御パラメータ生成部122において生成する、動き補償予測のブロックサイズ及び予測処理を制限する制御パラメータの一例を示し、説明する。 Next, FIG. 10 shows an example of the motion compensation prediction block size and control parameters for limiting the prediction processing, which are generated by the coding block control parameter generation unit 122 of FIG. 1 according to Embodiment 1 of the present invention. To do.
 制御パラメータは、最も小さな動き補償予測ブロックサイズである4×4画素の動き補償予測の有効・無効を制御するパラメータである、inter_4x4_enableと、動き補償予測の内、双予測が施される予測処理のみを禁止するブロックサイズを定義する、inter_bipred_restriction_idcの2つのパラメータで構成される。 The control parameters are inter_4x4_enable, which is a parameter for controlling the validity / invalidity of motion compensated prediction of 4 × 4 pixels, which is the smallest motion compensated prediction block size, and only prediction processing for which bi-prediction is performed among motion compensated predictions It consists of two parameters, inter_bipred_restriction_idc, which defines the block size that prohibits
 図9の必要な参照画像メモリ量を比較すると、最もアクセス量が大きな条件から、4×4双予測、4×8/8×4双予測、4×4単予測、8×8双予測、8×16/16×8双予測、4×8/8×4単予測、16×16双予測の順となっており、単予測に関しては、4×4画素の最小予測ブロックサイズ以外は比較的アクセス量が少ない。 When the necessary reference image memory amounts in FIG. 9 are compared, 4 × 4 bi-prediction, 4 × 8/8 × 4 bi-prediction, 4 × 4 mono-prediction, 8 × 8 bi-prediction, 8 × 16/16 × 8 bi-prediction, 4 × 8/8 × 4 single prediction, and 16 × 16 bi-prediction are in this order. Relatively accessed except for the minimum prediction block size of 4 × 4 pixels. The amount is small.
 そのため、最小予測ブロックサイズに関しては、動き補償予測処理そのものを禁止する制御パラメータであるinter_4x4_enableを用意し、各ブロックサイズに関して、更に双予測に対する制限を加えるinter_bipred_restriction_idcを制御パラメータとして用意することで、段階的なメモリアクセス量の制御を明示的に実現できる。 Therefore, with regard to the minimum prediction block size, inter_4x4_enable, which is a control parameter for prohibiting the motion compensation prediction process itself, is prepared, and inter_bipred_restriction_idc that further restricts bi-prediction is prepared as a control parameter for each block size. Can control memory access amount explicitly.
 ちなみに、4×8/8×4単予測に関しては、16×16双予測よりはメモリアクセス量が多くなるが、4×8/8×4単予測に対して制限を加える場合には、それよりもメモリアクセス量の大きな4×4及び4×8/8×4双予測にも制限をかける必要があり、その場合には最小CUサイズを16×16に設定することで、CU内分割モードがN×Nである8×8ブロックより小さな予測ブロックサイズの動き補償予測全体を禁止できるため、動き補償予測処理そのものを禁止に関しては、固定的な最小予測ブロックサイズに対する制限を有する構成で、段階的なメモリアクセス量の制御が可能である。 By the way, for 4 × 8/8 × 4 single prediction, the amount of memory access is larger than that of 16 × 16 bi-prediction. However, it is also necessary to limit 4 × 4 and 4 × 8/8 × 4 bi-prediction with a large memory access amount. In this case, by setting the minimum CU size to 16 × 16, the intra-CU partitioning mode is Since the entire motion compensated prediction with a prediction block size smaller than an N × N 8 × 8 block can be prohibited, the motion compensated prediction process itself is prohibited in a configuration having a restriction on a fixed minimum prediction block size. It is possible to control the memory access amount.
 上記制御を行う場合には、inter_4x4_enableとinter_bipred_restriction_idcに加えて、最小CUサイズ値を組合せて、メモリアクセス量の制御を行う構成となる。 When performing the above control, the memory access amount is controlled by combining the minimum CU size value in addition to inter_4x4_enable and inter_bipred_restriction_idc.
 実施の形態1においてinter_bipred_restriction_idcは、図10に示すように0から5までの値を定義し、双予測に対する制限なしの状態から、16×16ブロック以下のサイズの双予測を制限する状態までを、制御可能としているが、定義の範囲は一例であり、この値よりも少なくまたは多くの制御値を定義することも、本発明の実施の形態の別構成として実現できる。 In Embodiment 1, inter_bipred_restriction_idc defines a value from 0 to 5 as shown in FIG. 10, and from a state where there is no restriction on bi-prediction to a state where bi-prediction with a size of 16 × 16 blocks or less is restricted. Although controllable, the range of definition is an example, and it is also possible to define a control value that is smaller or more than this value as another configuration of the embodiment of the present invention.
 所定サイズの動き補償予測全体の無効化を制御するパラメータと、所定サイズ以下の動き補償予測の双予測を制限する制御パラメータを組合せて、メモリアクセス量の最大値が所定範囲内に収まるように制御する手法が、本発明の実施の形態1における構成である。 Control that disables the entire motion compensated prediction of a given size and a control parameter that restricts bi-prediction of motion compensated prediction of less than a given size, and controls the maximum memory access amount to be within the prescribed range This technique is the configuration in the first embodiment of the present invention.
 [動画像復号装置全体構成]
 図11は、本発明の実施の形態1に係る動画像復号装置の構成を示す図である。以下、各部の動作について説明する。実施の形態1に係る動画像復号装置は、入力端子1100、多重分離部1101、予測差分情報復号部1102、逆量子化・逆変換部1103、加算部1104、フレーム内復号画像バッファ1105、ループフィルタ部1106、復号画像メモリ1107、予測モード/ブロック構造復号部1108、予測モード/ブロック構造選択部1109、イントラ予測情報復号部1110、動き情報復号部1111、予測モード情報メモリ1112、イントラ予測部1113、動き補償予測部1114、及び出力端子1115を備える。
[Overall configuration of video decoding apparatus]
FIG. 11 is a diagram showing a configuration of the moving picture decoding apparatus according to Embodiment 1 of the present invention. Hereinafter, the operation of each unit will be described. The video decoding apparatus according to Embodiment 1 includes an input terminal 1100, a demultiplexing unit 1101, a prediction difference information decoding unit 1102, an inverse quantization / inverse transform unit 1103, an addition unit 1104, an intra-frame decoded image buffer 1105, a loop filter. Unit 1106, decoded image memory 1107, prediction mode / block structure decoding unit 1108, prediction mode / block structure selection unit 1109, intra prediction information decoding unit 1110, motion information decoding unit 1111, prediction mode information memory 1112, intra prediction unit 1113, A motion compensation prediction unit 1114 and an output terminal 1115 are provided.
 入力端子1100より符号化ビットストリームが多重分離部1101に供給される。多重分離部1101は、供給された符号化ビットストリームの符号列を予測誤差情報の符号化列と、符号化ブロック及び予測ブロック構造に関する制御パラメータ、符号化ブロック単位のCU分割構成と予測に用いられたモード情報である、予測モード、動き補償予測の場合の予測モードに応じた予測種別、動きベクトル、及び参照画像指定情報を特定する情報である動き情報、イントラ予測の場合のイントラ予測モード情報で構成される符号化列に分離する。当該予測誤差情報の符号化列を予測差分情報復号部1102に供給し、制御パラメータ、及び当該符号化ブロック単位のCU分割構成と予測に用いられたモード情報の符号化列を予測モード/ブロック構造復号部1108に供給する。 The encoded bit stream is supplied from the input terminal 1100 to the demultiplexing unit 1101. The demultiplexing unit 1101 is used for the code string of the supplied coded bitstream, the coded string of prediction error information, the control parameters related to the coding block and the prediction block structure, the CU partition configuration and coding block unit, and the prediction. Mode information, prediction mode according to the prediction mode in the case of motion compensated prediction, motion vector that is information specifying the motion vector, and reference image designation information, and intra prediction mode information in the case of intra prediction. Separate into coded sequences to be constructed. The coding sequence of the prediction error information is supplied to the prediction difference information decoding unit 1102, and the control parameter, the CU partition configuration of the coding block unit, and the coding sequence of the mode information used for prediction are predicted mode / block structure It supplies to the decoding part 1108.
 予測差分情報復号部1102は、多重分離部1101より供給された予測誤差情報の符号化列を復号し、量子化された予測誤差信号を生成する。予測差分情報復号部1102は、生成した量子化された予測誤差信号を逆量子化・逆変換部1103に供給する。 The prediction difference information decoding unit 1102 decodes the encoded sequence of the prediction error information supplied from the demultiplexing unit 1101, and generates a quantized prediction error signal. The prediction difference information decoding unit 1102 supplies the generated quantized prediction error signal to the inverse quantization / inverse transform unit 1103.
 逆量子化・逆変換部1103は、予測差分情報復号部1102より供給される、量子化された予測誤差信号を、逆量子化や逆直交変換などの処理を行って予測誤差信号を生成し、復号予測誤差信号を加算部1104に供給する。 The inverse quantization / inverse transform unit 1103 performs a process such as inverse quantization or inverse orthogonal transform on the quantized prediction error signal supplied from the prediction difference information decoding unit 1102 to generate a prediction error signal, The decoded prediction error signal is supplied to the adding unit 1104.
 加算部1104は、逆量子化・逆変換部1103より供給される復号予測誤差信号と、予測モード/ブロック構造選択部1109より供給される予測信号を加算して復号画像信号を生成し、復号画像信号をフレーム内復号画像バッファ1105及びループフィルタ部1106に供給する。 The adder 1104 adds the decoded prediction error signal supplied from the inverse quantization / inverse transform unit 1103 and the prediction signal supplied from the prediction mode / block structure selection unit 1109 to generate a decoded image signal, and generates a decoded image signal. The signal is supplied to the intra-frame decoded image buffer 1105 and the loop filter unit 1106.
 フレーム内復号画像バッファ1105は、図1の動画像符号化装置におけるフレーム内復号画像バッファ108と同じ機能を有し、イントラ予測の参照画像としてイントラ予測部1113に同一フレーム内の復号画像信号を供給すると共に、加算部1104から供給された復号画像信号を格納する。 The intra-frame decoded image buffer 1105 has the same function as the intra-frame decoded image buffer 108 in the moving picture encoding apparatus in FIG. 1, and supplies a decoded image signal in the same frame to the intra prediction unit 1113 as a reference image for intra prediction. At the same time, the decoded image signal supplied from the adding unit 1104 is stored.
 ループフィルタ部1106は、図1の動画像符号化装置におけるループフィルタ部109と同じ機能を有し、加算部1104より供給される復号画像信号に対して、歪除去のフィルタを施し、フィルタ処理を行った結果の復号画像を復号画像メモリ1107に供給する。 The loop filter unit 1106 has the same function as the loop filter unit 109 in the moving picture coding apparatus in FIG. 1, performs a distortion removal filter on the decoded image signal supplied from the addition unit 1104, and performs filter processing. The decoded image obtained as a result of the execution is supplied to the decoded image memory 1107.
 復号画像メモリ1107は、図1の動画像符号化装置における復号画像メモリ110と同じ機能を有し、ループフィルタ部1106から供給された復号画像信号を格納し、参照画像信号を動き補償予測部1114に供給する。また、復号画像メモリ1107は、格納された復号画像信号を再生時刻に合わせて、画像の表示順序に従い出力端子1115に供給する。 The decoded image memory 1107 has the same function as the decoded image memory 110 in the moving image encoding apparatus in FIG. 1, stores the decoded image signal supplied from the loop filter unit 1106, and uses the reference image signal as the motion compensation prediction unit 1114. To supply. The decoded image memory 1107 supplies the stored decoded image signal to the output terminal 1115 in accordance with the display order of images in accordance with the reproduction time.
 予測モード/ブロック構造復号部1108は、多重分離部1101より供給される、符号化ブロック及び予測ブロック構造に関する制御パラメータより、図3で示されるCU構造を定義する制御パラメータや、図10で示されるような動き補償予測のブロック構成及び予測処理を制限する制御パラメータを生成する。 The prediction mode / block structure decoding unit 1108 is a control parameter that defines the CU structure shown in FIG. 3 based on the control parameters related to the coding block and the prediction block structure supplied from the demultiplexing unit 1101, and the control parameter shown in FIG. Such a motion compensation prediction block configuration and control parameters for limiting the prediction process are generated.
 また、予測モード/ブロック構造復号部1108は、多重分離部1101より供給される、当該符号化ブロック単位のCU分割構成と予測に用いられたモード情報の符号化列より、符号化ブロック単位のCU分割構成と予測に用いられたモード情報を復号し、予測ブロックサイズ及び予測モードを生成すると共に、動き補償予測の場合の予測モードに応じた予測種別、動きベクトル、及び参照画像指定情報を特定する情報である動き情報、イントラ予測の場合のイントラ予測モード情報を分離し、当該符号化ブロック単位のCU分割構成と、予測モード情報を予測モード/ブロック構造選択部1109に供給する。 Also, the prediction mode / block structure decoding unit 1108 uses the CU division configuration for each coding block unit supplied from the demultiplexing unit 1101 and the coding sequence of the mode information used for prediction to determine the CU for each coding block unit. The mode information used for the division configuration and prediction is decoded to generate a prediction block size and a prediction mode, and the prediction type, motion vector, and reference image designation information corresponding to the prediction mode in the case of motion compensated prediction are specified. The motion information, which is information, and the intra prediction mode information in the case of intra prediction are separated, and the CU partition configuration and the prediction mode information for each coding block are supplied to the prediction mode / block structure selection unit 1109.
 予測モード/ブロック構造復号部1108は、予測ブロックにイントラ予測が用いられている場合には、イントラ予測情報復号部1110に予測ブロックサイズと共に、イントラ予測モード情報を供給し、動き補償予測が用いられている場合には、動き情報復号部1111に予測ブロックサイズと共に、動き補償予測モード、並びに予測モードに応じた予測種別、動きベクトル、及び参照画像指定情報を特定する情報を供給する。 When intra prediction is used for the prediction block, the prediction mode / block structure decoding unit 1108 supplies intra prediction mode information to the intra prediction information decoding unit 1110 together with the prediction block size, and motion compensated prediction is used. If so, the motion information decoding unit 1111 is supplied with information for specifying the motion compensation prediction mode, the prediction type corresponding to the prediction mode, the motion vector, and the reference image designation information together with the prediction block size.
 イントラ予測情報復号部1110は、予測モード/ブロック構造復号部1108より供給された予測ブロックサイズ、イントラ予測モード情報を復号し、符号化対象ブロックに対する予測ブロック構造と各予測ブロックにおけるイントラ予測モードを再生する。イントラ予測情報復号部1110は、再生したイントラ予測モードをイントラ予測部1113に供給すると共に、予測モード情報メモリ1112に対しても供給する。 The intra prediction information decoding unit 1110 decodes the prediction block size and intra prediction mode information supplied from the prediction mode / block structure decoding unit 1108, and reproduces the prediction block structure for the encoding target block and the intra prediction mode in each prediction block. To do. The intra prediction information decoding unit 1110 supplies the reproduced intra prediction mode to the intra prediction unit 1113 and also supplies it to the prediction mode information memory 1112.
 動き情報復号部1111は、予測モード/ブロック構造復号部1108より供給された、予測ブロックサイズ、動き補償予測モード、並びに予測モードに応じた予測種別、動きベクトル、及び参照画像指定情報を特定する情報を動き情報として復号し、復号した動き情報と、予測モード情報メモリ1112より供給される候補ブロック群の動き情報より、動き補償予測に用いる予測種別、動きベクトル及び参照画像指定情報を再生し、動き補償予測部1114に供給する。また、動き情報復号部1111は再生した動き情報を、予測モード情報メモリ1112に対しても供給する。動き情報復号部1111の詳細な構成については後述する。 The motion information decoding unit 1111 is supplied from the prediction mode / block structure decoding unit 1108 and specifies the prediction block size, the motion compensation prediction mode, and the prediction type, motion vector, and reference image designation information corresponding to the prediction mode. From the decoded motion information and the motion information of the candidate block group supplied from the prediction mode information memory 1112 to reproduce the prediction type, motion vector, and reference image designation information used for motion compensation prediction, This is supplied to the compensation prediction unit 1114. The motion information decoding unit 1111 also supplies the reproduced motion information to the prediction mode information memory 1112. A detailed configuration of the motion information decoding unit 1111 will be described later.
 予測モード情報メモリ1112は、図1の動画像符号化装置における予測モード情報メモリ119と同じ機能を持ち、動き情報復号部1111より供給される再生した動き情報、及びイントラ予測情報復号部1110から供給されるイントラ予測モードを、最小予測ブロックサイズ単位を基準として所定画像分記憶する。また、予測モード情報メモリ1112は、空間候補ブロック群と時間候補ブロック群の動き情報を、候補ブロック群の動き情報として動き情報復号部1111に供給すると共に、同一フレーム内の復号済隣接ブロックのイントラ予測モード情報を、対象予測ブロックのモード情報の予測候補として、イントラ予測情報復号部1110に供給する。 The prediction mode information memory 1112 has the same function as the prediction mode information memory 119 in the moving picture encoding apparatus in FIG. 1 and is supplied from the reproduced motion information supplied from the motion information decoding unit 1111 and the intra prediction information decoding unit 1110. The intra prediction mode to be performed is stored for a predetermined image on the basis of the minimum prediction block size unit. In addition, the prediction mode information memory 1112 supplies motion information of the spatial candidate block group and the temporal candidate block group to the motion information decoding unit 1111 as motion information of the candidate block group, and also intra of the decoded adjacent block in the same frame. The prediction mode information is supplied to the intra prediction information decoding unit 1110 as a prediction candidate of the mode information of the target prediction block.
 イントラ予測部1113は、図1の動画像符号化装置におけるイントラ予測部114と同じ機能を持ち、イントラ予測情報復号部1110より供給されるイントラ予測モードに従って、フレーム内復号画像バッファ1105よりイントラ予測の参照画像を入力し、イントラ予測信号を生成して、予測モード/ブロック構造選択部1109に供給する。 The intra prediction unit 1113 has the same function as the intra prediction unit 114 in the moving picture coding apparatus in FIG. 1, and performs intra prediction from the intra-frame decoded image buffer 1105 according to the intra prediction mode supplied from the intra prediction information decoding unit 1110. A reference image is input, an intra prediction signal is generated, and supplied to the prediction mode / block structure selection unit 1109.
 動き補償予測部1114は、図1の動画像符号化装置における動き補償予測部112と同じ機能を持ち、動き情報復号部1111より供給される動き情報に基づいて、復号画像メモリ1107内の参照画像指定情報が示す参照画像を、動きベクトル値だけ予測ブロックの画像信号と同一位置より移動させた位置の画像信号を取得して予測信号を生成する。動き補償予測の予測種別が双予測であれば、各予測種別の予測信号を平均したものを予測信号として生成し、予測信号を予測モード/ブロック構造選択部1109に供給する。 The motion compensation prediction unit 1114 has the same function as the motion compensation prediction unit 112 in the video encoding device of FIG. 1, and based on the motion information supplied from the motion information decoding unit 1111, the reference image in the decoded image memory 1107. A prediction signal is generated by acquiring an image signal at a position obtained by moving the reference image indicated by the designation information from the same position as the image signal of the prediction block by the motion vector value. If the prediction type of motion compensation prediction is bi-prediction, an average of the prediction signals of each prediction type is generated as a prediction signal, and the prediction signal is supplied to the prediction mode / block structure selection unit 1109.
 予測モード/ブロック構造選択部1109は、予測モード/ブロック構造復号部1108より供給された、当該符号化ブロック単位のCU分割構成と、予測モード情報を元に、CU分割を行い、再生された予測ブロック構造単位の予測モードによって、動き補償予測の場合には、動き補償予測部1114より動き補償予測信号を入力し、イントラ予測の場合には、イントラ予測部1113よりイントラ予測信号を入力し、再生された予測信号を加算部1104に供給する。 The prediction mode / block structure selection unit 1109 performs CU partitioning based on the CU partitioning configuration for each coding block supplied from the prediction mode / block structure decoding unit 1108 and the prediction mode information, and reproduces the predicted Depending on the prediction mode of the block structure unit, in the case of motion compensation prediction, a motion compensation prediction signal is input from the motion compensation prediction unit 1114, and in the case of intra prediction, an intra prediction signal is input from the intra prediction unit 1113 and reproduced. The predicted signal is supplied to the adding unit 1104.
 出力端子1115は、復号画像メモリ1107より供給された復号画像信号を、ディスプレイなどの表示媒体に出力することで、復号画像信号が再生される。 The output terminal 1115 outputs the decoded image signal supplied from the decoded image memory 1107 to a display medium such as a display, thereby reproducing the decoded image signal.
 図11に示した動画像復号装置の構成も、図1に示した動画像符号化装置の構成と同様に、CPU、フレームメモリ、ハードディスクなどを備える情報処理装置等のハードウェアによっても実現可能である。 The configuration of the video decoding device shown in FIG. 11 can also be realized by hardware such as an information processing device including a CPU, a frame memory, a hard disk, and the like, similarly to the configuration of the video encoding device shown in FIG. is there.
 図12は、本発明の実施の形態1に係る動画像復号装置における復号処理の符号化ブロック単位の動作の流れを示すフローチャートである。最初にCU分割の制御パラメータであるCU_Depthを0に初期化(S1200)し、多重分離部1101は、入力端子1100より供給された符号化ビットストリームを予測誤差情報の符号化列と、当該符号化ブロック単位のCU分割構成と予測に用いられたモード情報の符号化列に分離する(S1201)。分離された符号化ブロック単位の予測誤差情報の符号化列と、当該符号化ブロック単位のCU分割構成と予測に用いられたモード情報の符号化列が予測差分情報復号部1102及び、予測モード/ブロック構造復号部1108に供給され、CU分割構造を元にしたCU単位の復号処理が施される(S1202)。ステップS1202の詳細動作に関しては、後述する。 FIG. 12 is a flowchart showing a flow of operation in units of coding blocks of decoding processing in the video decoding apparatus according to Embodiment 1 of the present invention. First, CU_Depth, which is a control parameter for CU partitioning, is initialized to 0 (S1200), and the demultiplexing unit 1101 converts the coded bitstream supplied from the input terminal 1100 into the coded sequence of prediction error information and the coded data. The block is divided into a CU partition configuration and a coded sequence of mode information used for prediction (S1201). The encoded sequence of prediction error information in units of encoded blocks, the CU partition configuration in units of the encoded blocks, and the encoded sequence of mode information used for prediction are the prediction difference information decoding unit 1102, the prediction mode / It is supplied to the block structure decoding unit 1108 and subjected to decoding processing in units of CUs based on the CU partition structure (S1202). The detailed operation of step S1202 will be described later.
 続いて、当該符号化ブロック単位のCU分割構成は、ステップS1202で予測モード/ブロック構造復号部1108において復号され、復号された符号化構造情報が予測モード情報メモリ1112に格納される(S1203)。 Subsequently, the CU partitioning configuration for each coding block is decoded by the prediction mode / block structure decoding unit 1108 in step S1202, and the decoded coding structure information is stored in the prediction mode information memory 1112 (S1203).
 CU単位の復号処理(S1202)により復号された復号画像信号は、ループフィルタ部1106においてループフィルタ処理が施され(S1204)、復号画像メモリ1107に格納され(S1205)、符号化ブロック単位の復号処理が終了する。実施の形態1において、符号化ブロック単位の処理でループフィルタを施しているが、ループフィルタを施した復号画像信号は、同一フレームの復号処理には参照されず、後続するフレームの動き補償予測において参照されるため、符号化ブロック単位の処理を行わずに、フレーム全体の復号処理完了後に、フレーム全体に対して施すことも可能である。 The decoded image signal decoded by the decoding process in units of CU (S1202) is subjected to loop filter processing in the loop filter unit 1106 (S1204), stored in the decoded image memory 1107 (S1205), and decoded in units of coding blocks. Ends. In the first embodiment, the loop filter is applied in the process of the coding block unit, but the decoded image signal subjected to the loop filter is not referred to in the decoding process of the same frame, and in the motion compensation prediction of the subsequent frame Since it is referred to, it is possible to perform the process on the entire frame after the decoding process for the entire frame is completed without performing the process for each coding block.
 [CU単位の復号処理の詳細]
 続いて、図12のフローチャートにおけるステップS1202であるCU単位の復号処理の詳細について、図13のフローチャートを用いて説明する。
[Details of decryption processing in CU units]
Next, details of the decoding process in units of CUs, which is step S1202 in the flowchart of FIG. 12, will be described using the flowchart of FIG.
 最初に、設定された最大CUサイズと最小CUサイズの間の階層数を示す値Max_CU_Depthに対して、対象CUのCU_Depthが小さいか否かを判定する(S1300)。図3における最大CUサイズ及び最小CUサイズに関する制御パラメータが符号化・伝送されるため、復号処理において制御パラメータを復号することで、符号化時のMax_CU_Depthは復号される。Max_CU_Depthを定義する符号化情報の一例に関しては後述する。 First, it is determined whether or not the CU_Depth of the target CU is smaller than the value Max_CU_Depth indicating the number of layers between the set maximum CU size and the minimum CU size (S1300). Since the control parameters relating to the maximum CU size and the minimum CU size in FIG. 3 are encoded and transmitted, Max_CU_Depth at the time of encoding is decoded by decoding the control parameters in the decoding process. An example of the encoding information that defines Max_CU_Depth will be described later.
 CU_DepthがMax_CU_Depthより小さい場合には(S1300:YES)、CU分割情報を取得する(S1301)。一例としては、1ビットのフラグ情報(cu_split_flag)が、CUを分割するか否かの選択に合わせて、符号化・伝送され、このフラグ情報を復号することで、CUが分割されているか否かを認識する。 If CU_Depth is smaller than Max_CU_Depth (S1300: YES), CU partition information is acquired (S1301). As an example, 1-bit flag information (cu_split_flag) is encoded and transmitted according to the selection of whether or not to divide a CU, and whether or not the CU is divided by decoding this flag information. Recognize
 CUが分割されている場合(S1302:YES)には、CUを分割して復号するため、CU分割CU_Depthを1加算して(S1303)一階層下のCUに対するCU単位の復号処理を行う(S1304-S1307)、CUの分割領域に対して分割1領域の処理(S1304)、分割2領域の処理(S1305)、分割3領域の処理(S1306)、分割4領域の処理(S1307)の順で再帰的に図13のフローチャートで説明される処理が行われる。 When the CU is divided (S1302: YES), in order to divide and decode the CU, 1 is added to the CU division CU_Depth (S1303), and the CU unit decoding process is performed on the CU one layer below (S1304). -S1307), processing of division 1 area (S1304), division 2 area processing (S1305), division 3 area processing (S1306), division 4 area processing (S1307) with respect to the CU division area Specifically, the processing described in the flowchart of FIG. 13 is performed.
 CU_DepthがMax_CU_Depth以上の場合(S1300:NO)及び、CUが分割されていない場合(S1302:NO)には、復号対象となるCUの大きさが確定し、確定したCU内の予測モードに応じた復号処理が施される。 When CU_Depth is greater than or equal to Max_CU_Depth (S1300: NO), and when the CU is not divided (S1302: NO), the size of the CU to be decoded is determined, and it corresponds to the prediction mode in the determined CU. Decoding processing is performed.
 最初に、CU内の予測にイントラ予測が用いられているか、動き補償予測が用いられているかを示す情報を取得する。(S1308)。実施の形態1においては、CU単位でskipモードであるか否かを示すskipフラグ情報(skip_flag)、また当該CUがskipモードで無い場合にイントラ予測であるか動き補償予測であるかを示す予測モードフラグ情報(pred_mode_flag)が符号化時にCU単位の予測モード情報として符号化されており、これらを復号する事により、イントラ予測であるか、動き補償予測(skipモードを含む)であるかの情報が取得できる。 First, information indicating whether intra prediction or motion compensation prediction is used for prediction within the CU is acquired. (S1308). In the first embodiment, skip flag information (skip_flag) indicating whether or not the skip mode is in CU units, and prediction indicating whether the prediction is intra prediction or motion compensation prediction when the CU is not in skip mode. Mode flag information (pred_mode_flag) is encoded as prediction mode information in units of CU at the time of encoding, and information indicating whether it is intra prediction or motion compensated prediction (including skip mode) by decoding these. Can be obtained.
 続いて、当該CUがイントラ予測である場合には(S1309:YES)、CU単位のイントラ予測復号処理が、図11のイントラ予測情報復号部1110及びイントラ予測部1113で行われ(S1311)、対象CUにおけるイントラ予測信号を生成し、復号誤差信号と加算されることにより、復号画像信号を生成し(S1312)、CU単位の復号処理を終了する。 Subsequently, when the CU is intra prediction (S1309: YES), intra prediction decoding processing for each CU is performed by the intra prediction information decoding unit 1110 and the intra prediction unit 1113 in FIG. 11 (S1311), and the target An intra prediction signal in the CU is generated and added to the decoding error signal to generate a decoded image signal (S1312), and the decoding process in units of CUs is completed.
 当該CUがイントラ予測でない場合には(S1309:NO)、CU単位の動き補償予測復号処理が、図11の動き情報復号部1111及び動き補償予測部1114で行われ(S1310)、対象CUにおける動き補償予測信号を生成し、復号誤差信号と加算されることにより、復号画像信号を生成し(S1312)、CU単位の復号処理を終了する。ステップS1310の動作の詳細に関しては後述する。 When the CU is not intra prediction (S1309: NO), motion compensation prediction decoding processing for each CU is performed by the motion information decoding unit 1111 and the motion compensation prediction unit 1114 in FIG. 11 (S1310), and motion in the target CU. A compensated prediction signal is generated and added to the decoded error signal to generate a decoded image signal (S1312), and the decoding process for each CU is completed. Details of the operation in step S1310 will be described later.
 続いて、図13のフローチャートにおけるステップS1310である対象CUにおける動き補償予測復号処理の詳細について、図14のフローチャートを用いて説明する。最初に、CU単位の予測モードを示す情報として復号したskipフラグを取得し(S1400)、skipフラグが1である、すなわちskipモードである場合(S1401:YES)には、CU内の予測ブロック分割モードは2N×2Nとなり、NumPartが1に設定され2N×2N予測ブロックの予測ブロック単位復号が施される(S1402)。 Next, details of the motion compensated prediction decoding process in the target CU, which is step S1310 in the flowchart of FIG. 13, will be described using the flowchart of FIG. First, a decoded skip flag is acquired as information indicating the prediction mode in units of CUs (S1400). When the skip flag is 1, that is, in the skip mode (S1401: YES), prediction block partitioning within the CU is performed. The mode is 2N × 2N, NumPart is set to 1, and prediction block unit decoding of a 2N × 2N prediction block is performed (S1402).
 skip_flagが0である、すなわちskipモードでない場合(S1401:NO)には、CU分割(PU)モードとして、符号化時に当該CUで選択した動き補償予測ブロックサイズの種別であるCU内分割モード値を、予測モード情報より取得し(S1403)、PUモードが2N×2Nの場合(S1404:YES)には、NumPartが1に設定され2N×2N予測ブロックの予測ブロック単位復号が施される(S1402)。 When skip_flag is 0, that is, when the skip mode is not set (S1401: NO), the CU partition (PU) mode is set as the CU partition mode value that is the type of the motion compensated prediction block size selected by the CU at the time of encoding. When the PU mode is 2N × 2N (S1404: YES), NumPart is set to 1 and prediction block unit decoding of the 2N × 2N prediction block is performed (S1402). .
 PUモードが2N×2Nでない場合(S1404:NO)、PUモードが2N×Nの場合(S1405:YES)には、NumPartが2に設定され2N×N予測ブロックの予測ブロック単位復号が施される(S1406)。 When the PU mode is not 2N × 2N (S1404: NO) and when the PU mode is 2N × N (S1405: YES), NumPart is set to 2 and prediction block unit decoding of 2N × N prediction blocks is performed. (S1406).
 続いて、CU_DepthがMax_CU_Depthと等しく且つ、対象CUサイズが8×8で、後述するinter_4x4_enableフラグが1である場合(S1407:YES)には、更にPUモードがN×2Nであるか否かを判定し(S1409)、PUモードがN×2Nである場合(S1409:YES)には、NumPartが2に設定され、N×2N予測ブロックの予測ブロック単位復号が施される(S1408)。 Subsequently, when CU_Depth is equal to Max_CU_Depth, the target CU size is 8 × 8, and an inter_4x4_enable flag described later is 1 (S1407: YES), it is further determined whether or not the PU mode is N × 2N. However, when the PU mode is N × 2N (S1409: YES), NumPart is set to 2 and prediction block unit decoding of the N × 2N prediction block is performed (S1408).
 PUモードがN×2Nでない場合(S1409:NO)には、PUモードはN×Nとなり、NumPartを4に設定して、N×N予測ブロックの予測ブロック単位復号が施される(S1410)。 When the PU mode is not N × 2N (S1409: NO), the PU mode is N × N, NumPart is set to 4, and prediction block unit decoding of the N × N prediction block is performed (S1410).
 ステップS1407の条件を満たさない場合(S1407:NO)には、当該CUにおいてN×N予測ブロックは適用されないため、NumPartが2に設定され、N×2N予測ブロックの予測ブロック単位復号が施される(S1408)。ステップS1402、S1406、S1408、S1410で施される、PUモード毎の予測ブロック単位復号処理の詳細は後述する。 When the condition of step S1407 is not satisfied (S1407: NO), since the N × N prediction block is not applied in the CU, NumPart is set to 2, and prediction block unit decoding of the N × 2N prediction block is performed. (S1408). Details of the prediction block unit decoding process for each PU mode performed in steps S1402, S1406, S1408, and S1410 will be described later.
 実施の形態1においては、復号したPUモードに対する予測ブロック単位の復号処理を選択するための条件判断に関しては、図14のフローチャートに示すように、ステップS1404からS1409までに示した順番で処理を行っているが、復号したPUモードに従って、予測ブロック単位の復号処理が施される構成であれば、条件分岐の順番に関しては異なる構成でも実現可能である。 In the first embodiment, regarding the condition judgment for selecting the prediction block unit decoding process for the decoded PU mode, the processes are performed in the order shown in steps S1404 to S1409 as shown in the flowchart of FIG. However, if the decoding process is performed in units of prediction blocks in accordance with the decoded PU mode, it is possible to implement a different configuration regarding the order of conditional branches.
 PUモード毎の予測ブロック単位復号処理が施された後、PUモード及び予測ブロック単位の動き情報等のモード情報が、図11における予測モード情報メモリ1112に格納され(S1411)、当該CUに対する動き補償予測復号処理が終了する。 After the prediction block unit decoding process for each PU mode is performed, mode information such as the PU mode and motion information for each prediction block is stored in the prediction mode information memory 1112 in FIG. 11 (S1411), and motion compensation for the CU is performed. The predictive decoding process ends.
 [実施の形態1の詳細機能説明]
 続いて、本発明の実施の形態1に係る動画像符号化装置の動き補償予測ブロック構造選択部113の動作、図7のフローチャートにおけるステップS701、S702、S703、S705の処理の詳細動作を、以下説明する。
[Detailed Function Description of Embodiment 1]
Next, the operation of the motion compensated prediction block structure selection unit 113 of the video encoding apparatus according to Embodiment 1 of the present invention, the detailed operation of the processes in steps S701, S702, S703, and S705 in the flowchart of FIG. explain.
 [実施の形態1における動画像符号化装置における動き補償予測ブロック構造選択部の詳細動作説明]
 図15は、実施の形態1の動画像符号化装置における動き補償予測ブロック構造選択部113の詳細な構成を示す図である。動き補償予測ブロック構造選択部113は、最適な動き補償予測モード及び予測ブロック構造を決定する機能を有する。
[Detailed Operation Explanation of Motion Compensated Prediction Block Structure Selection Unit in Moving Picture Encoding Device in Embodiment 1]
FIG. 15 is a diagram illustrating a detailed configuration of the motion compensated prediction block structure selection unit 113 in the video encoding device according to the first embodiment. The motion compensation prediction block structure selection unit 113 has a function of determining an optimal motion compensation prediction mode and a prediction block structure.
 動き補償予測ブロック構造選択部113は、動き補償予測生成部1500、予測誤差算出部1501、予測ベクトル算出部1502、差分ベクトル算出部1503、動き情報符号量算出部1504、予測モード/ブロック構造評価部1505、結合動き情報算出部1506、結合動き情報単予測変換部1507、及び結合動き補償予測生成部1508を含む。 The motion compensation prediction block structure selection unit 113 includes a motion compensation prediction generation unit 1500, a prediction error calculation unit 1501, a prediction vector calculation unit 1502, a difference vector calculation unit 1503, a motion information code amount calculation unit 1504, and a prediction mode / block structure evaluation unit. 1505, a combined motion information calculation unit 1506, a combined motion information single prediction conversion unit 1507, and a combined motion compensation prediction generation unit 1508 are included.
 図1における動き補償予測ブロック構造選択部113に対して、動きベクトル検出部111より入力された動きベクトル値が、動き補償予測生成部1500に供給され、予測モード情報メモリ119より入力された動き情報が、予測ベクトル算出部1502、及び結合動き情報算出部1506に供給される。 The motion vector value input from the motion vector detection unit 111 to the motion compensation prediction block structure selection unit 113 in FIG. 1 is supplied to the motion compensation prediction generation unit 1500 and input from the prediction mode information memory 119. Is supplied to the prediction vector calculation unit 1502 and the combined motion information calculation unit 1506.
 また、動き補償予測部112に対して、動き補償予測生成部1500、及び結合動き補償予測生成部1508から、動き補償予測に用いる参照画像指定情報と動きベクトルが出力され、動き補償予測部112より、生成された動き補償予測画像が予測誤差算出部1501に供給される。予測誤差算出部1501には更に、符号化ブロック取得部102より符号化対象となる予測ブロックの画像信号が供給される。 In addition, reference image designation information and motion vectors used for motion compensation prediction are output from the motion compensation prediction generation unit 1500 and the combined motion compensation prediction generation unit 1508 to the motion compensation prediction unit 112. The generated motion compensated prediction image is supplied to the prediction error calculation unit 1501. The prediction error calculation unit 1501 is further supplied with an image signal of a prediction block to be encoded from the encoding block acquisition unit 102.
 また、予測モード/ブロック構造評価部1505から、予測モード選択部116に対して、予測ブロック構造、符号化する動き情報と確定した予測モード情報、及び動き補償予測信号を供給する。 Also, the prediction mode / block structure evaluation unit 1505 supplies the prediction block structure, the motion information to be encoded and the determined prediction mode information, and the motion compensated prediction signal to the prediction mode selection unit 116.
 動き補償予測生成部1500は、各予測ブロック構造において、予測に使用可能な各参照画像に対して算出された動きベクトル値を受信し、図10で示される双予測制限情報に従って動き補償予測を行い、参照画像指定情報を予測ベクトル算出部1502に供給し、参照画像指定情報と動きベクトルを出力する。 The motion compensation prediction generation unit 1500 receives the motion vector value calculated for each reference image usable for prediction in each prediction block structure, and performs motion compensation prediction according to the bi-prediction restriction information shown in FIG. The reference image designation information is supplied to the prediction vector calculation unit 1502, and the reference image designation information and the motion vector are output.
 予測誤差算出部1501は、入力された動き補償予測画像と処理対象の予測ブロック画像より、予測誤差評価値を算出する。誤差評価値を算出するための演算としては、動きベクトル検出における誤差評価値と同様に、画素毎の差分絶対値の総和SADや、画素毎の二乗誤差値の総和SSE等を使用できる。更に、予測残差の符号化を行う際に施される、直交変換・量子化を行うことによって復号画像に生じる歪成分の量を加味することで、より正確な誤差評価値が算出可能である。この場合には、予測誤差算出部1501内に、図1における減算部103、直交変換・量子化部104、逆量子化・逆変換部106、加算部107の機能を有することで実現できる。 The prediction error calculation unit 1501 calculates a prediction error evaluation value from the input motion compensated prediction image and the prediction block image to be processed. As the calculation for calculating the error evaluation value, the sum SAD of the absolute difference value for each pixel, the sum SSE of the square error value for each pixel, and the like can be used as in the error evaluation value in motion vector detection. Furthermore, a more accurate error evaluation value can be calculated by taking into account the amount of distortion components generated in the decoded image by performing orthogonal transform / quantization performed when encoding the prediction residual. . In this case, the prediction error calculation unit 1501 can be realized by having the functions of the subtraction unit 103, the orthogonal transformation / quantization unit 104, the inverse quantization / inverse transformation unit 106, and the addition unit 107 in FIG.
 予測誤差算出部1501は、各予測モード及び各予測ブロック構造において算出された予測誤差評価値と、動き補償予測信号を予測モード/ブロック構造評価部1505に供給する。 The prediction error calculation unit 1501 supplies the prediction error evaluation value calculated in each prediction mode and each prediction block structure and the motion compensation prediction signal to the prediction mode / block structure evaluation unit 1505.
 予測ベクトル算出部1502は、動き補償予測生成部1500より参照画像指定情報を供給され、予測モード情報メモリ119から供給される隣接ブロックの動き情報における候補ブロック群より、指定された参照画像に対する動きベクトル値を入力し、複数の予測ベクトルを予測ベクトル候補リストと共に生成し、差分ベクトル算出部1503に、参照画像指定情報と共に供給する。予測ベクトル算出部1502は、予測ベクトルの候補を作成し、予測ベクトル候補として登録する。 The prediction vector calculation unit 1502 is supplied with the reference image designation information from the motion compensation prediction generation unit 1500, and the motion vector for the designated reference image from the candidate block group in the adjacent block motion information supplied from the prediction mode information memory 119. A value is input, a plurality of prediction vectors are generated together with a prediction vector candidate list, and supplied to the difference vector calculation unit 1503 together with reference image designation information. The prediction vector calculation unit 1502 creates prediction vector candidates and registers them as prediction vector candidates.
 差分ベクトル算出部1503は、予測ベクトル算出部1502より供給された、予測ベクトル候補のそれぞれに対して、動き補償予測生成部1500から供給される動きベクトル値との差分を計算し、差分ベクトル値を算出する。算出された差分ベクトル値と予測ベクトル候補に対する指定情報である予測ベクトルインデックスを符号化した際、符号量が最も少ない。差分ベクトル算出部1503は、最も少ない情報量である予測ベクトルに対する予測ベクトルインデックスと差分ベクトル値を参照画像指定情報と共に、動き情報符号量算出部1504に供給する。 The difference vector calculation unit 1503 calculates the difference between each of the prediction vector candidates supplied from the prediction vector calculation unit 1502 and the motion vector value supplied from the motion compensated prediction generation unit 1500, and calculates the difference vector value. calculate. When the prediction vector index which is the designation information for the calculated difference vector value and the prediction vector candidate is encoded, the code amount is the smallest. The difference vector calculation unit 1503 supplies the prediction vector index and the difference vector value for the prediction vector having the smallest information amount, together with the reference image designation information, to the motion information code amount calculation unit 1504.
 動き情報符号量算出部1504は、差分ベクトル算出部1503より供給される、差分ベクトル値、参照画像指定情報、予測ベクトルインデックス、および予測モードより、各予測ブロック構造及び各予測モードにおける動き情報に要する符号量を算出する。また、動き情報符号量算出部1504は、結合動き補償予測生成部1508より、結合予測モードにおいて伝送する必要がある、結合動き情報インデックスと予測モードを示すための情報を受け取り、結合予測モードにおける動き情報に要する符号量を算出する。 The motion information code amount calculation unit 1504 requires motion information in each prediction block structure and each prediction mode from the difference vector value, reference image designation information, prediction vector index, and prediction mode supplied from the difference vector calculation unit 1503. The code amount is calculated. Also, the motion information code amount calculation unit 1504 receives from the combined motion compensation prediction generation unit 1508 information indicating the combined motion information index and the prediction mode that needs to be transmitted in the combined prediction mode, and moves in the combined prediction mode. The amount of code required for information is calculated.
 動き情報符号量算出部1504は、各予測ブロック構造及び各予測モードにおいて算出された動き情報及び動き情報に要する符号量を予測モード/ブロック構造評価部1505に供給する。 The motion information code amount calculation unit 1504 supplies the motion information calculated in each prediction block structure and each prediction mode and the code amount required for the motion information to the prediction mode / block structure evaluation unit 1505.
 予測モード/ブロック構造評価部1505は、予測誤差算出部1501より供給された各予測モードの予測誤差評価値と、動き情報符号量算出部1504から供給された各予測モードの動き情報符号量を用いて、各予測モードの総合動き補償予測誤差評価値を算出し、最も少ない評価値である予測モード及び予測ブロックサイズを選択し、選択した予測モード、予測ブロックサイズと選択した予測モードに対する動き情報を、予測モード選択部116に出力する。また、予測モード/ブロック構造評価部1505は同様に、予測誤差算出部1501より供給された動き補償予測信号に対して、選択した予測モード、予測ブロックサイズにおける予測信号を選択して予測モード選択部116に出力する。 The prediction mode / block structure evaluation unit 1505 uses the prediction error evaluation value of each prediction mode supplied from the prediction error calculation unit 1501 and the motion information code amount of each prediction mode supplied from the motion information code amount calculation unit 1504. Calculating a total motion compensation prediction error evaluation value of each prediction mode, selecting a prediction mode and a prediction block size which are the smallest evaluation values, and selecting a prediction mode, a prediction block size and motion information for the selected prediction mode. To the prediction mode selection unit 116. Similarly, the prediction mode / block structure evaluation unit 1505 selects a prediction signal in the selected prediction mode and prediction block size with respect to the motion compensation prediction signal supplied from the prediction error calculation unit 1501, and selects a prediction mode selection unit. To 116.
 結合動き情報算出部1506は、予測モード情報メモリ119より供給される隣接ブロックの動き情報における候補ブロック群を用いて、単予測であるか双予測であるかを示す予測種別、参照画像指定情報、動きベクトル値で構成される動き情報として、複数の動き情報を結合動き情報候補リストと共に生成し、結合動き情報単予測変換部1507に供給する。 The combined motion information calculation unit 1506 uses the candidate block group in the motion information of the adjacent blocks supplied from the prediction mode information memory 119, a prediction type indicating whether the prediction is uni-prediction or bi-prediction, reference image designation information, A plurality of pieces of motion information are generated together with a combined motion information candidate list as motion information composed of motion vector values, and supplied to the combined motion information single prediction conversion unit 1507.
 図16は、結合動き情報算出部1506の構成を示す図である。結合動き情報算出部1506は、空間結合動き情報候補リスト生成部1600、結合動き情報候補リスト削除部1601、時間結合動き情報候補リスト生成部1602、第1結合動き情報候補リスト追加部1603および第2結合動き情報候補リスト追加部1604を含む。結合動き情報算出部1506は、空間的に隣接する候補ブロック群より所定の順番で動き情報の候補を作成し、その中から、同一の動き情報を持つ候補を削除したのち、時間的に隣接する候補ブロック群より作成した動き情報の候補を追加することで、有効な動き情報のみを結合動き情報候補として登録する。この時間結合動き情報候補リスト生成部を結合動き情報候補リスト削除部より後段に配置した点が、本実施の形態の特徴的な構成であり、時間結合動き情報候補を同一の動き情報を削除する処理の対象からはずすことにより、符号化効率を落とすことなく演算量を削減することが可能である。結合動き情報算出部1506の詳細動作に関しては、後述する。 FIG. 16 is a diagram illustrating a configuration of the combined motion information calculation unit 1506. The combined motion information calculation unit 1506 includes a spatial combined motion information candidate list generation unit 1600, a combined motion information candidate list deletion unit 1601, a temporal combined motion information candidate list generation unit 1602, a first combined motion information candidate list addition unit 1603, and a second. A combined motion information candidate list adding unit 1604 is included. The combined motion information calculation unit 1506 creates motion information candidates in a predetermined order from spatially adjacent candidate block groups, deletes candidates having the same motion information from the candidates, and then temporally adjacent. By adding motion information candidates created from the candidate block group, only valid motion information is registered as combined motion information candidates. The point that this temporally combined motion information candidate list generation unit is arranged after the combined motion information candidate list deletion unit is a characteristic configuration of the present embodiment, and deletes the same motion information from temporally combined motion information candidates. By eliminating the processing target, it is possible to reduce the amount of calculation without reducing the encoding efficiency. The detailed operation of the combined motion information calculation unit 1506 will be described later.
 図15に戻り、結合動き情報単予測変換部1507は、結合動き情報算出部1506より供給される結合動き情報候補リスト及び、候補リストに登録される動き情報に対して、図10で示される双予測制限情報に従って、予測種別が双予測である動き情報を単予測の動き情報に変換し、結合動き補償予測生成部1508に供給する。 Returning to FIG. 15, the combined motion information single prediction conversion unit 1507 performs the bi-directional processing shown in FIG. 10 on the combined motion information candidate list supplied from the combined motion information calculation unit 1506 and the motion information registered in the candidate list. In accordance with the prediction restriction information, motion information whose prediction type is bi-prediction is converted into uni-prediction motion information and supplied to the combined motion compensation prediction generation unit 1508.
 結合動き補償予測生成部1508は、結合動き情報単予測変換部1507より供給された結合動き情報候補リストより、登録された結合動き情報候補のそれぞれに対して、動き情報より、予測種別に応じて1つの参照画像(単予測)もしくは異なる2つの参照画像(双予測)の参照画像指定情報と動きベクトル値を動き補償予測部112に指定して、動き補償予測画像を生成すると共に、それぞれの結合動き情報インデックスを動き情報符号量算出部1504に供給する。 The combined motion compensated prediction generation unit 1508 corresponds to each registered combined motion information candidate from the combined motion information candidate list supplied from the combined motion information single prediction conversion unit 1507 according to the prediction type based on the motion information. The reference image designation information and motion vector value of one reference image (uni-prediction) or two different reference images (bi-prediction) are designated to the motion compensation prediction unit 112 to generate a motion compensated prediction image, and the respective combinations The motion information index is supplied to the motion information code amount calculation unit 1504.
 図15の構成では、それぞれの結合動き情報インデックスにおける予測モード評価は、予測モード/ブロック構造評価部1505で施されるが、予測誤差評価値及び動き情報符号量を予測誤差算出部1501及び動き情報符号量算出部1504より受け取り、結合動き補償予測生成部1508内で、最適な結合動き補償予測の結合動きインデックスを確定させた後に、他の予測モードを含めた最適予測モードの評価を行う構成を取ることも可能である。 In the configuration of FIG. 15, the prediction mode evaluation in each combined motion information index is performed by the prediction mode / block structure evaluation unit 1505. The prediction error evaluation value and the motion information code amount are used as the prediction error calculation unit 1501 and the motion information. A configuration in which the combined motion compensation prediction generation unit 1508 receives the code amount calculation unit 1504 and determines the combined motion index of the optimal combined motion compensation prediction, and then evaluates the optimal prediction mode including other prediction modes. It is also possible to take.
 図17は、図7のフローチャートにおけるステップS701、S702、S703、S705ステップの動き補償予測モード/予測信号生成処理の詳細動作を説明するためのフローチャートである。この動作は、図15の動き補償予測ブロック構造選択部113における詳細動作を示している。 FIG. 17 is a flowchart for explaining detailed operations of the motion compensation prediction mode / prediction signal generation processing in steps S701, S702, S703, and S705 in the flowchart of FIG. This operation represents a detailed operation in the motion compensated prediction block structure selection unit 113 in FIG.
 最初に、定義されたCU内の予測ブロックサイズ分割モード(PU)に従って設定されたNumPartに基づき、対象CU内をPU分割した予測ブロックサイズ毎に(S1700)、ステップS1701からステップS1708までのステップが実行される(S1709)。先ず、結合動き情報候補リスト生成を行う(S1701)。 First, based on the NumPart set according to the predicted block size division mode (PU) in the defined CU, the steps from step S1701 to step S1708 are performed for each prediction block size obtained by PU division in the target CU (S1700). It is executed (S1709). First, a combined motion information candidate list is generated (S1701).
 続いて、予測ブロックサイズが、図10に示した双予測を制限する制御パラメータinter_bipred_restriction_idcにより設定される双予測を制限する予測ブロックサイズであるbipred_restriction_size以下である場合(S1702:YES)には、生成された結合動き情報候補リスト内の各候補における双予測の動き情報を単予測の動き情報に置き換える、結合動き情報候補単予測変換を行う(S1703)。予測ブロックサイズが、bipred_restriction_size以下でない場合(S1702:NO)には、続くステップS1704に進む。 Subsequently, when the prediction block size is equal to or smaller than bipred_restriction_size that is a prediction block size that restricts bi-prediction set by the control parameter inter_bipred_restriction_idc that restricts bi-prediction illustrated in FIG. 10 (S1702: YES), the prediction block size is generated. The combined motion information candidate uni-prediction conversion is performed in which the bi-prediction motion information in each candidate in the combined motion information candidate list is replaced with the single-prediction motion information (S1703). If the predicted block size is not less than or equal to bipred_restriction_size (S1702: NO), the process proceeds to subsequent step S1704.
 次に、生成または置き換えられた結合動き情報候補リストの動き情報を元に、結合予測モード評価値を生成する(S1704)。続いて、予測モード評価値を生成し(S1705)、生成した評価値を比較することで最適な予測モードを選択する(S1706)。ただし、ステップS1704及びS1705の評価値生成の順序はこれに限らない。 Next, a combined prediction mode evaluation value is generated based on the motion information in the combined motion information candidate list generated or replaced (S1704). Subsequently, a prediction mode evaluation value is generated (S1705), and an optimal prediction mode is selected by comparing the generated evaluation values (S1706). However, the order of evaluation value generation in steps S1704 and S1705 is not limited to this.
 選択された予測モードに従い予測信号を出力し(S1707)、選択された予測モードに従い動き情報を出力する(S1708)ことで、予測ブロック単位の動き補償予測モード/予測信号生成処理が終了する。ステップS1701、S1703、S1704及びS1705の詳細動作に関しては後述する。 The prediction signal is output according to the selected prediction mode (S1707), and the motion information is output according to the selected prediction mode (S1708), thereby completing the motion compensation prediction mode / prediction signal generation processing for each prediction block. Detailed operations of steps S1701, S1703, S1704, and S1705 will be described later.
 図18は、図17のステップS1701の結合動き情報候補リスト生成の詳細動作を説明するためのフローチャートである。この動作は、図15の結合動き情報算出部1506における構成の詳細動作を示している。 FIG. 18 is a flowchart for explaining the detailed operation of generating the combined motion information candidate list in step S1701 of FIG. This operation shows the detailed operation of the configuration in the combined motion information calculation unit 1506 in FIG.
 図16の空間結合動き情報候補リスト生成部1600は、予測モード情報メモリ119より供給される空間候補ブロック群から領域外である候補ブロックや、イントラモードである候補ブロックを除いた候補ブロックから空間結合動き情報候補リストを生成する(S1800)。空間結合動き情報候補リスト生成の詳細な動作は後述する。 The spatial combination motion information candidate list generation unit 1600 in FIG. 16 performs spatial combination from candidate blocks excluding candidate blocks outside the region or candidate blocks in the intra mode from the spatial candidate block group supplied from the prediction mode information memory 119. A motion information candidate list is generated (S1800). Detailed operations for generating the spatially coupled motion information candidate list will be described later.
 続いて、結合動き情報候補リスト削除部1601において、生成された空間結合動き情報候補リストより、同一の動き情報を持つ結合動き情報候補を削除して動き情報候補リストを更新する(S1801)。結合動き情報候補削除の詳細な動作は後述する。 Subsequently, the combined motion information candidate list deletion unit 1601 deletes the combined motion information candidates having the same motion information from the generated spatial combined motion information candidate list and updates the motion information candidate list (S1801). Detailed operation of the combined motion information candidate deletion will be described later.
 時間結合動き情報候補リスト生成部1602は、続いて予測モード情報メモリ119より供給される時間候補ブロック群から領域外である候補ブロックや、イントラモードである候補ブロックを除いた候補ブロックから時間結合動き情報候補リストを生成し(S1802)、時間結合動き情報候補リストと結合して結合動き情報候補リストとする。時間結合動き情報候補リスト生成の詳細な動作は後述する。 The temporally combined motion information candidate list generation unit 1602 subsequently performs temporally combined motion from candidate blocks excluding candidate blocks outside the region or candidate blocks in the intra mode from the temporal candidate block group supplied from the prediction mode information memory 119. An information candidate list is generated (S1802) and combined with the temporally combined motion information candidate list to form a combined motion information candidate list. Detailed operation of the time combination motion information candidate list generation will be described later.
 次に、第1結合動き情報候補リスト追加部1603は、時間結合動き情報候補リスト生成部1602で生成された結合動き情報候補リストに登録された結合動き情報候補から0個から2個の第1結合動き情報候補を生成して結合動き情報候補リストに追加し(S1803)、当該結合動き情報候補リストを第2結合動き情報候補リスト追加部1604に供給する。第1結合動き情報候補リスト追加の詳細な動作は後述する。 Next, the first combined motion information candidate list adding unit 1603 generates 0 to 2 first combined motion information candidates from the combined motion information candidate registered in the combined motion information candidate list generated by the temporal combined motion information candidate list generating unit 1602. A combined motion information candidate is generated and added to the combined motion information candidate list (S1803), and the combined motion information candidate list is supplied to the second combined motion information candidate list adding unit 1604. The detailed operation of adding the first combined motion information candidate list will be described later.
 次に、第2結合動き情報候補リスト追加部1604は、第1結合動き情報候補リスト追加部1603より供給される結合動き情報候補リストに依存しない0個から4個の第2結合動き情報候補を生成して第1結合動き情報候補リスト追加部1603より供給される結合動き情報候補リストに追加し(S1804)、処理を終了する。第2結合動き情報候補リスト追加の詳細な動作は後述する。 Next, the second combined motion information candidate list adding unit 1604 selects 0 to 4 second combined motion information candidates that do not depend on the combined motion information candidate list supplied from the first combined motion information candidate list adding unit 1603. Generated and added to the combined motion information candidate list supplied from the first combined motion information candidate list adding unit 1603 (S1804), and the process ends. Detailed operations for adding the second combined motion information candidate list will be described later.
 予測モード情報メモリ119より、結合動き情報算出部1506に供給される動き情報の候補ブロック群には、空間候補ブロック群と時間候補ブロック群が含まれる。まず、空間結合動き情報候補リスト生成について説明する。 The candidate block group of motion information supplied from the prediction mode information memory 119 to the combined motion information calculation unit 1506 includes a spatial candidate block group and a temporal candidate block group. First, generation of a spatially coupled motion information candidate list will be described.
 図19は、空間結合動き情報候補リスト生成に用いる空間候補ブロック群を示す図である。空間候補ブロック群は、符号化対象画像の予測対象ブロックに隣接している同一画像のブロックを示す。ブロック群は、その管理が最小予測ブロックサイズ単位で行われ、候補ブロックの位置は、最小予測ブロックサイズの単位で管理されるが、隣接ブロックの予測ブロックサイズが最小予測ブロックサイズよりも大きな場合には、予測ブロックサイズ内の全ての候補ブロックに同一の動き情報が格納される。実施の形態1においては、隣接するブロック群の内、図19に示すようなブロックA0、ブロックA1、ブロックB0、ブロックB1、ブロックB2の5ブロックを空間候補ブロック群とする。 FIG. 19 is a diagram showing a spatial candidate block group used for generating a spatially coupled motion information candidate list. The spatial candidate block group indicates a block of the same image adjacent to the prediction target block of the encoding target image. The block group is managed in units of the minimum prediction block size, and the position of the candidate block is managed in units of the minimum prediction block size, but when the prediction block size of the adjacent block is larger than the minimum prediction block size The same motion information is stored in all candidate blocks within the predicted block size. In the first embodiment, among the adjacent block groups, five blocks of block A0, block A1, block B0, block B1, and block B2 as shown in FIG.
 図20は、空間結合動き情報候補リスト生成の詳細動作を説明するためのフローチャートである。空間候補ブロック群に含まれる5つの候補ブロックのうち、ブロックA0、ブロックA1、ブロックB0、ブロックB1、ブロックB2について、ブロックA1、ブロックB1、ブロックB0、ブロックA0の順序で以下の処理を繰り返し行う(S2000~S2003)。 FIG. 20 is a flowchart for explaining the detailed operation of generating the spatially coupled motion information candidate list. Of the five candidate blocks included in the spatial candidate block group, the following processing is repeated for block A0, block A1, block B0, block B1, and block B2 in the order of block A1, block B1, block B0, and block A0. (S2000 to S2003).
 最初に候補ブロックの有効性を検査する(S2001)。候補ブロックが領域外でなくイントラモードでない場合、候補ブロックは有効となる。候補ブロックが有効であれば(S2001:YES)、候補ブロックの動き情報を空間結合動き情報候補リストに追加する(S2002)。 First, the validity of the candidate block is checked (S2001). If the candidate block is not out of the region and not in the intra mode, the candidate block is valid. If the candidate block is valid (S2001: YES), the motion information of the candidate block is added to the spatially combined motion information candidate list (S2002).
 ステップS2000からS2003までの繰り返し処理に続いて、空間結合動き情報候補リストに追加された候補数が4未満だった場合(S2004:YES)、候補ブロックB2の有効性を検査する(S2005)。ブロックB2が領域外でなくイントラモードでない場合(S2005:YES)、ブロックB2の動き情報を空間結合動き情報候補リストに追加する(S2006)。 When the number of candidates added to the spatially coupled motion information candidate list is less than 4 following the repetitive processing from steps S2000 to S2003 (S2004: YES), the validity of the candidate block B2 is checked (S2005). When the block B2 is not out of the region and is not in the intra mode (S2005: YES), the motion information of the block B2 is added to the spatially coupled motion information candidate list (S2006).
 ここでは、空間結合動き情報候補リストには4以下の候補ブロックの動き情報が含まれるとしたが、空間候補ブロック群は、処理対象の予測ブロックに隣接する少なくとも1以上の処理済みのブロックであり、候補ブロックの有効性によって空間結合動き情報候補リストの数が変動すればよく、これに限定されない。 Here, it is assumed that the spatial combination motion information candidate list includes motion information of four or less candidate blocks, but the spatial candidate block group is at least one or more processed blocks adjacent to the prediction block to be processed. The number of spatially coupled motion information candidate lists may be changed depending on the effectiveness of the candidate block, and the present invention is not limited to this.
 図21は、結合動き情報候補削除の詳細動作を説明するためのフローチャートである。空間結合動き情報候補リスト作成処理により、生成される結合動き情報候補の最大数をMaxSpatialCandとすると、i=MaxSpatialCand-1からi>0までの結合動き情報候補(候補(i))に対して以下の処理を繰り返し行う(S2100~S2106)。 FIG. 21 is a flowchart for explaining the detailed operation of combined motion information candidate deletion. Assuming that the maximum number of combined motion information candidates generated by the spatial combined motion information candidate list creation process is Max Spatial Cand, the following is performed for combined motion information candidates (candidate (i)) from i = Max Spatial Cand-1 to i> 0. This process is repeated (S2100 to S2106).
 候補(i)が存在していれば(S2101のYES)、ii=i-1からii>=0までの結合動き情報候補(候補(ii))に対して以下の処理を繰り返し行い(S2102~S2105)、候補(i)が存在しない場合(S2101のNO)はステップS2102からS2105までの候補(ii)についての繰り返し処理をスキップする。 If the candidate (i) exists (YES in S2101), the following processing is repeated for the combined motion information candidates (candidate (ii)) from ii = i−1 to ii> = 0 (S2102˜). In S2105), when there is no candidate (i) (NO in S2101), the iterative process for candidate (ii) from step S2102 to S2105 is skipped.
 まず、候補(i)の動き情報(動き情報(i))と候補(ii)の動き情報(動き情報(ii))が同一であるか検査し(S2103)、同じである場合(S2103のYES)、候補(i)を結合動き情報候補リストから削除し(S2104)、候補(ii)についての繰り返し処理を終了する。 First, whether the motion information (motion information (i)) of the candidate (i) and the motion information (motion information (ii)) of the candidate (ii) are the same is checked (S2103), and if they are the same (YES in S2103) ), Candidate (i) is deleted from the combined motion information candidate list (S2104), and the iterative process for candidate (ii) ends.
 動き情報(i)と動き情報(ii)が同一でない場合(S2103のNO)、iiから1を減算し、候補(ii)についての処理を繰り返す(S2102~S2105)。 If the motion information (i) and the motion information (ii) are not the same (NO in S2103), 1 is subtracted from ii, and the process for the candidate (ii) is repeated (S2102 to S2105).
 ステップS2100からS2105までの繰り返し処理に続いて、iから1を減算し、候補(i)についての処理を繰り返す(S2100~S2106)。 Subsequent to the repetitive processing from step S2100 to S2105, 1 is subtracted from i, and the processing for candidate (i) is repeated (S2100 to S2106).
 図22に結合動き情報候補が4つの場合のリスト中の候補の比較関係を示す。すなわち時間結合動き情報候補を含まない4つの空間結合動き情報候補について総当たりで比較して同一性を判定し、重複する候補を削除する。 FIG. 22 shows a comparison relationship of candidates in the list when there are four combined motion information candidates. That is, four spatially combined motion information candidates that do not include temporally combined motion information candidates are compared by brute force to determine identity, and duplicate candidates are deleted.
 ここで、結合予測モードは時間方向や空間方向の動きの連続性を用いて、予測対象ブロックは自身の動き情報を直接符号化せずに、空間及び時間的に隣接するブロックの動き情報を符号化に使用する手法であるが、空間結合動き情報候補が空間方向の連続性を基づいているのに対し、時間結合動き情報候補は時間方向の連続性に基づいて後述する方法で生成されており、これらの性質は異なるものである。よって時間結合動き情報候補と空間結合動き情報候補に同一の動き情報が含まれることは稀であり、同一の動き情報を削除するための結合動き情報候補削除処理の対象から時間結合動き情報候補を除いても、最終的に得られる結合動き情報候補リストに同一の動き情報が含まれることは稀である。 Here, the joint prediction mode uses temporal and spatial continuity of motion, and the prediction target block encodes motion information of spatially and temporally adjacent blocks without directly encoding its own motion information. Although the spatially coupled motion information candidate is based on continuity in the spatial direction, the temporally coupled motion information candidate is generated by the method described later based on the temporal direction continuity. These properties are different. Therefore, it is rare that the same motion information is included in the temporally combined motion information candidate and the spatially combined motion information candidate, and the temporally combined motion information candidate is removed from the target of the combined motion information candidate deletion process for deleting the same motion information. Even if they are excluded, it is rare that the same motion information is included in the finally obtained combined motion information candidate list.
 また、後述するように時間結合動き情報候補ブロックは最小予測ブロックよりも大きなサイズである最小時間予測ブロック単位で管理されるため、時間的に隣接する予測ブロックの大きさが最小時間予測ブロックよりも小さな場合には、本来の位置とはずれた位置の動き情報が用いられることとなり、その結果、動き情報に誤差を含む場合が多い。そのため、空間結合動き情報候補の動き情報とは異なる動き情報となることが多く、同一の動き情報を削除するための結合動き情報候補削除処理の対象から除いても影響が少ない。 Also, as will be described later, temporally combined motion information candidate blocks are managed in units of minimum time prediction blocks that are larger in size than the minimum prediction block, so that the size of prediction blocks that are temporally adjacent are larger than the minimum time prediction block If it is small, motion information at a position deviating from the original position is used, and as a result, the motion information often includes an error. Therefore, the motion information is often different from the motion information of the spatially coupled motion information candidate, and there is little influence even if the motion information is excluded from the target of the combined motion information candidate deletion process for deleting the same motion information.
 図23は、空間結合動き情報候補の最大数が4である場合の結合動き情報候補削除における候補の比較内容の一例である。図23(a)は空間結合動き情報候補のみを結合動き情報候補削除処理の対象とした場合の比較内容であり、図23(b)は空間結合動き情報候補と時間結合動き情報を処理の対象とした場合の比較内容である。空間結合動き情報候補のみを結合動き情報候補削除処理の対象とすることにより、動き情報を比較の回数が10回から6回に減少している。 FIG. 23 is an example of comparison contents of candidates in deletion of combined motion information candidates when the maximum number of spatially combined motion information candidates is four. FIG. 23A shows the comparison contents when only the spatially coupled motion information candidate is the target of the coupled motion information candidate deletion process, and FIG. 23B is the target of processing the spatially coupled motion information candidate and the temporally coupled motion information. It is a comparison content in the case of. By using only the spatially combined motion information candidates as the target of the combined motion information candidate deletion process, the number of comparisons of motion information is reduced from 10 times to 6 times.
 このように、時間結合動き情報候補を結合動き情報候補削除処理の対象にしないことにより、同一の動き情報を適切に削除しながら、動き情報の比較の回数を10回から6回に削減することが可能である。 Thus, by not deleting the temporally combined motion information candidate as the target of the combined motion information candidate deletion process, the number of motion information comparisons is reduced from 10 to 6 while appropriately deleting the same motion information. Is possible.
 また、すべての空間予測候補の同一を比較せず、空間的な位置が近い候補間同士の比較のみ行うことで 結合動き情報候補削除処理の回数を削減することも可能である。具体的には、図19のB1位置から算出された結合動き情報はA1位置の結合動き情報と比較し、B0位置から算出された結合動き情報はB1位置の結合動き情報のみと比較し、A0位置から算出された結合動き情報はA1のみと比較し、B2位置から算出された結合動き情報はA1、B1のみと比較することで、動き情報の比較回数を最大5回に制限できる。 Also, it is possible to reduce the number of combined motion information candidate deletion processes by comparing only candidates that are close in spatial position without comparing the same of all spatial prediction candidates. Specifically, the combined motion information calculated from the B1 position in FIG. 19 is compared with the combined motion information at the A1 position, the combined motion information calculated from the B0 position is compared with only the combined motion information at the B1 position, and A0 By comparing the combined motion information calculated from the position with only A1 and the combined motion information calculated from the B2 position with only A1 and B1, the number of motion information comparisons can be limited to a maximum of five.
 上記のように特定の空間予測候補のみ結合動き情報の同一比較を行う場合、空間結合動き情報候補リスト生成中に(S1800)、結合動き情報候補削減処理(S1801)を行った方が、同一結合動き情報が残存してしまうことによる符号化効率低下の影響が少ない。つまり、空間結合動き情報候補リスト生成時に結合動き情報の同一比較を行うことで、不要な結合動き情報を追加せずにすむため、図20のステップS2004の最大空間予測候補数を4つに制限する場合に、B2位置から算出される結合動き情報が追加できる可能性が高まるためである。 When the same comparison of combined motion information is performed only for a specific spatial prediction candidate as described above, it is better to perform the combined motion information candidate reduction process (S1801) during the generation of the spatial combined motion information candidate list (S1800). There is little influence of a decrease in coding efficiency due to motion information remaining. That is, by performing the same comparison of the combined motion information when generating the spatially combined motion information candidate list, it is possible to avoid adding unnecessary combined motion information, so the maximum number of spatial prediction candidates in step S2004 in FIG. 20 is limited to four. This is because the possibility that the combined motion information calculated from the B2 position can be added increases.
 続いて、時間結合動き情報候補リスト生成について説明する。図24は、時間結合動き情報候補リスト生成に用いる時間方向周辺予測ブロックの定義を説明する図である。時間候補ブロック群は、予測対象ブロックが属する画像とは別の復号済みの画像ColPicに属するブロックの中で、予測対象ブロックと同位置及びその周辺にあるブロックを示す。ブロック群は、その管理が最小時間予測ブロックサイズ単位で行われ、候補ブロックの位置は、最小時間予測ブロックサイズの単位で管理される。本発明の実施の形態1においては、最小時間予測ブロックサイズは最小予測ブロックサイズを垂直方向、水平方向にそれぞれ2倍した大きさとする。時間的に隣接するブロックの予測ブロックのサイズが最小時間予測ブロックサイズよりも大きな場合には、予測ブロックサイズ内のすべての候補ブロックに同一の動きの情報が格納される。一方、予測ブロックのサイズが最小時間予測ブロックサイズよりも小さな場合には、時間方向周辺予測ブロックの左上に位置する予測ブロックの動きの情報を時間方向周辺予測ブロックの情報とする。図24(b)に予測ブロックサイズが最小時間予測ブロックサイズより小さい場合の時間方向周辺予測ブロックの動き情報を示す。 Subsequently, generation of a time-coupled motion information candidate list will be described. FIG. 24 is a diagram illustrating the definition of the temporal direction peripheral prediction block used for generating the temporally combined motion information candidate list. The temporal candidate block group indicates blocks in the same position as and around the prediction target block among the blocks belonging to the decoded image ColPic different from the image to which the prediction target block belongs. The block group is managed in units of minimum time prediction block size, and the positions of candidate blocks are managed in units of minimum time prediction block size. In Embodiment 1 of the present invention, the minimum temporal prediction block size is set to a size obtained by doubling the minimum prediction block size in the vertical and horizontal directions. When the size of the prediction block of temporally adjacent blocks is larger than the minimum temporal prediction block size, the same motion information is stored in all candidate blocks within the prediction block size. On the other hand, when the size of the prediction block is smaller than the minimum temporal prediction block size, the motion information of the prediction block located at the upper left of the temporal direction peripheral prediction block is used as the temporal direction peripheral prediction block information. FIG. 24B shows motion information of the temporal direction neighboring prediction block when the prediction block size is smaller than the minimum temporal prediction block size.
 図24(a)におけるA1~A4、B1~B4、C、D、E、F1~F4、G1~G4、H、I1~I16の位置のブロックが、時間的に隣接するブロック群となる。実施の形態1においては、これら時間的に隣接するブロック群の内、時間候補ブロック群をブロックHとブロックI6の2ブロックとする。 In FIG. 24A, blocks at positions A1 to A4, B1 to B4, C, D, E, F1 to F4, G1 to G4, H, and I1 to I16 are temporally adjacent block groups. In the first embodiment, among these temporally adjacent block groups, the temporal candidate block group is assumed to be two blocks, block H and block I6.
 図25は、時間結合動き情報候補リスト生成の詳細動作を説明するためのフローチャートである。時間候補ブロック群に含まれる2つの候補ブロックであるブロックHとブロックI11について(S2500、S2505)、ブロックH、ブロックI11の順序で候補ブロックの有効性を検査する(S2501)。候補ブロックが有効である場合(S2501:YES)、ステップS2502~ステップS2504の処理が行われ、生成された動き情報が時間結合動き情報候補リストに登録され、処理が終了する。候補ブロックが画面領域外の位置を示す場合や、候補ブロックがイントラ予測ブロックである場合(S2501:NO)、候補ブロックが有効でなく、次の候補ブロックの有効/無効判定が行われる。 FIG. 25 is a flowchart for explaining the detailed operation of generating the time combination motion information candidate list. Regarding the block H and the block I11 that are two candidate blocks included in the time candidate block group (S2500, S2505), the validity of the candidate block is checked in the order of the block H and the block I11 (S2501). If the candidate block is valid (S2501: YES), the processing from step S2502 to step S2504 is performed, the generated motion information is registered in the time combination motion information candidate list, and the processing ends. When the candidate block indicates a position outside the screen area, or when the candidate block is an intra prediction block (S2501: NO), the candidate block is not valid, and valid / invalid determination of the next candidate block is performed.
 候補ブロックが有効である場合(S2501:YES)、候補ブロックの動き情報をもとに結合動き情報候補に登録する参照画像選択候補を確定する(S2502)。実施の形態1ではL0予測の参照画像を、L0予測の参照画像のうち処理対象画像に最も近い距離にある参照画像とし、L1予測の参照画像を、L1予測の参照画像のうち処理対象画像に最も近い距離にある参照画像とする。 If the candidate block is valid (S2501: YES), the reference image selection candidate to be registered in the combined motion information candidate is determined based on the motion information of the candidate block (S2502). In the first embodiment, the L0 prediction reference image is the reference image that is the closest to the processing target image among the L0 prediction reference images, and the L1 prediction reference image is the processing target image among the L1 prediction reference images. The reference image is the closest distance.
 ここでの参照画像選択候補の確定手法は、L0予測の参照画像とL1予測の参照画像が決定できればよく、これに限定されない。符号化処理と復号処理で同一の手法で参照画像を確定することで、符号化時に意図した参照画像を確定できる。他の確定手法としては、例えばL0予測の参照画像及びL1予測の参照画像の参照画像インデックスが0である参照画像を選択する手法や、空間隣接ブロックが使用しているL0参照画像及びL1参照画像を選択する手法や、符号化ストリーム中で各予測種別の参照画像を指定する手法を用いることが可能である。 The method for determining the reference image selection candidate here is not limited to this as long as the reference image for L0 prediction and the reference image for L1 prediction can be determined. The reference image intended at the time of encoding can be determined by determining the reference image by the same method in the encoding process and the decoding process. As another determination method, for example, a method of selecting a reference image having a reference image index of 0 for a reference image for L0 prediction and a reference image for L1 prediction, or a L0 reference image and a L1 reference image used by spatially neighboring blocks. Can be used, and a method of specifying a reference image of each prediction type in the encoded stream can be used.
 次に、候補ブロックの動き情報をもとに結合動き情報候補に登録する動きベクトル値を確定する(S2503)。実施の形態1における、時間結合動き情報は、候補ブロックの動き情報で有効な予測種別である動きベクトル値をもとに、双予測の動き情報を算出する。候補ブロックの予測種別がL0予測もしくはL1予測の単予測の場合には、予測に用いられている予測種別(L0予測或いはL1予測)の動き情報を選択し、その参照画像指定情報と動きベクトル値を双予測動き情報生成の基準値とする。 Next, the motion vector value to be registered in the combined motion information candidate is determined based on the motion information of the candidate block (S2503). In the first embodiment, the temporally coupled motion information calculates bi-prediction motion information based on motion vector values that are effective prediction types in motion information of candidate blocks. When the prediction type of the candidate block is L0 prediction or L1 prediction single prediction, motion information of the prediction type (L0 prediction or L1 prediction) used for prediction is selected, and its reference image designation information and motion vector value are selected. Is a reference value for generating bi-predictive motion information.
 候補ブロックの予測種別が双予測である場合には、L0予測或いはL1予測のどちらか一方の動き情報を基準値として選択する。基準値の選択方法は、例えばColPicと同じ予測種別に存在する動き情報を選択する、候補ブロックのL0予測、L1予測のそれぞれの参照画像でColPicとの画像間距離が近い方を選択する、符号化側で選択してシンタックスで明示的に伝送する等が挙げられる。 If the prediction type of the candidate block is bi-prediction, either L0 prediction or L1 prediction motion information is selected as a reference value. The reference value selection method selects, for example, motion information existing in the same prediction type as ColPic, and selects a reference image having a shorter inter-image distance from ColPic in each of L0 prediction and L1 prediction of a candidate block. For example, it is possible to select the transmission side and explicitly transmit the syntax.
 双予測動き情報生成の基準とする動きベクトル値が確定したら、結合動き情報候補に登録する動きベクトル値を算出する。 When the motion vector value used as the reference for bi-predictive motion information generation is determined, the motion vector value to be registered in the combined motion information candidate is calculated.
 図26は、時間結合動き情報に対する基準動きベクトル値ColMvに対する、L0予測、L1予測に対して登録する動きベクトル値mvL0t、mvL1tの算出手法を説明するための図である。 FIG. 26 is a diagram for explaining a calculation method of motion vector values mvL0t and mvL1t registered for L0 prediction and L1 prediction with respect to the reference motion vector value ColMv for temporally coupled motion information.
 基準動きベクトル値ColMvに対するColPicと候補ブロックの基準とする動きベクトルの対象となる参照画像との画像間距離をColDistとする。L0予測、L1予測の各参照画像と処理対象画像との画像間距離をCurrL0Dist、CurrL1Distとする。ColMvを、ColDistとCurrL0Dist、CurrL1Distの距離比率でスケーリングした動きベクトルを、それぞれに登録する動きベクトルとする。具体的には、登録する動きベクトル値mvL0t、mvL1tは、下記式1、2で算出される。
 mvL0t=mvCol×CurrL0Dist/ColDist・・・(式1)
 mvL1t=mvCol×CurrL1Dist/ColDist・・・(式2)
となる。
The distance between images between ColPic for the reference motion vector value ColMv and the reference image that is the target of the motion vector used as a reference for the candidate block is referred to as ColDist. The inter-image distance between each reference image of L0 prediction and L1 prediction and the processing target image is set to CurrL0Dist and CurrL1Dist. A motion vector obtained by scaling ColMv with a distance ratio of ColDist to CurrL0Dist and CurrL1Dist is set as a motion vector to be registered. Specifically, the motion vector values mvL0t and mvL1t to be registered are calculated by the following formulas 1 and 2.
mvL0t = mvCol × CurrL0Dist / ColDist (Formula 1)
mvL1t = mvCol × CurrL1Dist / ColDist (Formula 2)
It becomes.
 図25に戻り、このようにして生成された、双予測の参照画像選択情報(インデックス)と、動きベクトル値を結合動き情報候補に追加し(S2504)、時間結合動き情報候補リスト作成処理が終了する。 Returning to FIG. 25, the bi-predicted reference image selection information (index) and the motion vector value generated in this way are added to the combined motion information candidates (S2504), and the temporal combined motion information candidate list creation process ends. To do.
 続いて、第1結合動き情報候補リスト追加部1603の詳細な動作について説明する。図27は、第1結合動き情報候補リスト追加部1603の動作を説明するためのフローチャートである。最初に、時間結合動き情報候補リスト生成部1602より供給される結合動き情報候補リストに登録された結合動き情報候補の数(NumCandList)と結合動き情報候補最大数(MaxNumMergeCand)から、第1追加結合動き情報候補を生成する最大数であるMaxNumGenCandを式3より算出する(S2700)。
 MaxNumGenCand=MaxNumMergeCand-NumCandList; (NumCandList>1)
 MaxNumGenCand=0; (NumCandList<=1)             (式3)
Next, a detailed operation of the first combined motion information candidate list adding unit 1603 will be described. FIG. 27 is a flowchart for explaining the operation of the first combined motion information candidate list adding unit 1603. First, based on the number of combined motion information candidates (NumCandList) and the maximum number of combined motion information candidates (MaxNumMergeCand) registered in the combined motion information candidate list supplied from the temporally combined motion information candidate list generation unit 1602, the first additional combination is performed. MaxNumGenCand, which is the maximum number for generating motion information candidates, is calculated from Equation 3 (S2700).
MaxNumGenCand = MaxNumMergeCand-NumCandList; (NumCandList> 1)
MaxNumGenCand = 0; (NumCandList <= 1) (Formula 3)
 次に、MaxNumGenCandが0より大きいか検査する(S2701)。MaxNumGenCandが0より大きくなければ(S2701:NO)、処理を終了する。MaxNumGenCandが0より大きければ(S2701:YES)、以下の処理を行う。まず、組み合わせ検査回数であるloopTimesを決定する。loopTimesはNumCandList×NumCandListに設定する。ただし、loopTimesが8を超える場合にはloopTimesは8に制限する(S2702)。ここで、loopTimesは0から7までの整数となる。loopTimesだけ以下の処理を繰り返し行う(S2702からS2708)。 Next, it is inspected whether MaxNumGenCand is larger than 0 (S2701). If MaxNumGenCand is not greater than 0 (S2701: NO), the process ends. If MaxNumGenCand is greater than 0 (S2701: YES), the following processing is performed. First, loopTimes that is the number of combination inspections is determined. loopTimes is set to NumCandList × NumCandList. However, if loopTimes exceeds 8, loopTimes is limited to 8 (S2702). Here, loopTimes is an integer from 0 to 7. The following processing is repeatedly performed for loopTimes (S2702 to S2708).
 結合動き情報候補Mと結合動き情報候補Nの組み合わせを決定する(S2703)。ここで、組み合わせ検査回数と結合動き情報候補Mと結合動き情報候補Nの関係について説明する。 The combination of the combined motion information candidate M and the combined motion information candidate N is determined (S2703). Here, the relationship between the number of combination inspections, the combined motion information candidate M, and the combined motion information candidate N will be described.
 図28は組み合わせ検査回数と結合動き情報候補Mと結合動き情報候補Nの関係を説明するための図である。図28のようにMとNは異なる値であって、まずMを0に固定してNの値を1~4(最大値はNumCandList)に変化させ、その後、Nの値を0に固定してMの値を1~4(最大値はNumCandList)に変化させる。このような組み合わせ定義は、最も選択される確率の高い動き情報である結合動き情報候補リスト内で最初の動き情報を有効に活用しつつ、実際には組み合わせテーブルを持たずに計算で組み合わせパターンを算出できる効果がある。 FIG. 28 is a diagram for explaining the relationship between the number of combination inspections, the combined motion information candidate M, and the combined motion information candidate N. As shown in FIG. 28, M and N are different values. First, M is fixed to 0, the value of N is changed to 1 to 4 (the maximum value is NumCandList), and then the value of N is fixed to 0. The value of M is changed to 1 to 4 (the maximum value is NumCandList). Such a combination definition makes effective use of the first motion information in the combined motion information candidate list, which is the motion information with the highest probability of being selected, and actually calculates the combination pattern without having a combination table. There is an effect that can be calculated.
 結合動き情報候補MのL0予測が有効で且つ結合動き情報候補NのL1予測が有効であるか検査する(S2704)。結合動き情報候補MのL0予測が有効で且つ結合動き情報候補NのL1予測が有効であれば(S2704:YES)、結合動き情報候補MのL0予測の動きベクトルと参照画像を結合動き情報候補NのL1予測の動きベクトルと参照画像と組み合わせて双結合動き情報候補を生成する(S2705)。結合動き情報候補MのL0予測が有効で且つ結合動き情報候補NのL1予測が有効でなければ(S2704:NO)、次の組み合わせを処理する。ここで、第1追加結合動き情報候補として、L0予測の動き情報とL1予測が同一である場合があり、動き補償を双予測で行ってもL0予測もしくはL1予測の単予測と同じ結果が得られるため、L0予測の動き情報とL1予測の動き情報が同一である追加結合動き情報候補生成は、動き補償予測の演算量を増加させる要因となる。そのため、通常は、L0予測の動き情報とL1予測の動き情報が同一であるかどうか比較し、同一でない場合にのみ、第1追加結合動き情報候補とする。 It is checked whether the L0 prediction of the combined motion information candidate M is valid and the L1 prediction of the combined motion information candidate N is valid (S2704). If the L0 prediction of the combined motion information candidate M is valid and the L1 prediction of the combined motion information candidate N is valid (S2704: YES), the combined motion information candidate uses the motion vector of the L0 prediction of the combined motion information candidate M and the reference image. A combined motion information candidate is generated by combining the motion vector of N L1 predictions and the reference image (S2705). If the L0 prediction of the combined motion information candidate M is not valid and the L1 prediction of the combined motion information candidate N is not valid (S2704: NO), the next combination is processed. Here, as the first additional combined motion information candidate, the motion information of the L0 prediction and the L1 prediction may be the same, and even if motion compensation is performed by bi-prediction, the same result as the single prediction of the L0 prediction or the L1 prediction is obtained. Therefore, the additional combined motion information candidate generation in which the motion information of the L0 prediction and the motion information of the L1 prediction are the same is a factor that increases the amount of calculation of the motion compensation prediction. For this reason, normally, whether or not the motion information of the L0 prediction and the motion information of the L1 prediction are the same is compared, and only when they are not the same, the first additional combined motion information candidate is set.
 ステップS2705に続いて、双結合動き情報候補を結合動き情報候補リストに追加する(S2706)。ステップS2706に続いて、生成した双結合動き情報の数がMaxNumGenCandであるか検査する(S2707)。生成された双結合動き情報の数がMaxNumGenCandであれば(S2707のYES)、処理を終了する。生成された双結合動き情報の数がMaxNumGenCandでなければ(S2707のNO)、次の組み合わせを処理する。 Subsequent to step S2705, a double-coupled motion information candidate is added to the combined motion information candidate list (S2706). Subsequent to step S2706, it is checked whether the number of generated double-coupled motion information is MaxNumGenCand (S2707). If the number of generated double coupled motion information is MaxNumGenCand (YES in S2707), the process ends. If the number of generated double coupled motion information is not MaxNumGenCand (NO in S2707), the next combination is processed.
 ここで、第1追加結合動き情報候補は、結合動き情報候補リストに登録された結合動き情報候補の動き情報と処理対象の動き情報候補の動きに微妙にずれがある場合に、結合動き情報候補リストに登録された結合動き情報候補の動き情報を修正して有効な結合動き情報候補を生成することで、符号化効率を高めることができる。 Here, the first additional combined motion information candidate is a combined motion information candidate when there is a slight difference between the motion information of the combined motion information candidate registered in the combined motion information candidate list and the motion information candidate motion to be processed. Coding efficiency can be improved by correcting the motion information of the combined motion information candidates registered in the list to generate effective combined motion information candidates.
 続いて、第2結合動き情報候補リスト追加部1604の詳細な動作について説明する。図29は、第2結合動き情報候補リスト追加部1604の動作を説明するためのフローチャートである。最初に、第1結合動き情報候補リスト追加部1603より供給される結合動き情報候補リストに登録された結合動き情報候補の数(NumCandList)と結合動き情報候補最大数(MaxNumMergeCand)から、第1追加結合動き情報候補を生成する最大数であるMaxNumGenCandを式4より算出する(S2900)。
 MaxNumGenCand=MaxNumMergeCand-NumCandList;   (式4)
Next, the detailed operation of the second combined motion information candidate list adding unit 1604 will be described. FIG. 29 is a flowchart for explaining the operation of the second combined motion information candidate list adding unit 1604. First, the first addition is performed based on the number of combined motion information candidates (NumCandList) and the maximum number of combined motion information candidates (MaxNumMergeCand) registered in the combined motion information candidate list supplied from the first combined motion information candidate list adding unit 1603. MaxNumGenCand, which is the maximum number for generating combined motion information candidates, is calculated from Equation 4 (S2900).
MaxNumGenCand = MaxNumMergeCand-NumCandList; (Formula 4)
 次に、以下の処理をiについてMaxNumGenCand回繰り返し行う(S2901からS2905)。ここで、iは0からMaxNumGenCand-1の整数となる。L0予測の動きベクトルが(0,0)、参照インデックスがiであって、L1予測の動きベクトルが(0,0)、参照インデックスがiである予測種別が双予測である第2追加結合動き情報候補を生成する(S2902)。第2追加結合動き情報候補を結合動き情報候補リストに追加する(S2903)。次のiについて処理する(S2904)。 Next, the following process is repeated MaxNumCand times for i (S2901 to S2905). Here, i is an integer from 0 to MaxNumGenCand-1. The second additional combined motion in which the motion vector for L0 prediction is (0,0), the reference index is i, the motion vector for L1 prediction is (0,0), and the prediction type is i for the reference index is bi-prediction. Information candidates are generated (S2902). The second additional combined motion information candidate is added to the combined motion information candidate list (S2903). The next i is processed (S2904).
 ここでは、第2追加結合動き情報候補を、L0予測の動きベクトルが(0,0)、参照インデックスがiであって、L1予測の動きベクトルが(0,0)、参照インデックスがiである予測種別が双予測である結合動き情報候補とした。これは、一般的な動画像において、L0予測の動きベクトルとL1予測の動きベクトルが(0,0)である結合動き情報候補の発生頻度が統計的に高いためである。結合動き情報候補リストに登録された結合動き情報候補の動き情報に依存せず、統計的に利用頻度が高い結合動き情報候補であれば、これに限定されない。例えば、L0予測やL1予測の動きベクトルはそれぞれ(0,0)以外のベクトル値でもよく、L0予測とL1予測の参照インデックスが異なるように設定してもよい。また、第2追加結合動き情報候補を符号化済みの画像や符号化済みの画像の一部の発生頻度の高い動き情報とし、符号化ストリームに符号化して伝送して設定することもできる。なお、ここではBピクチャ(Bスライス)について説明したが、Pピクチャ(Pスライス)の場合は、L0予測の動きベクトルが(0,0)で、予測種別がL0予測である第2追加結合動き情報候補を生成する。 Here, the second additional combined motion information candidate has a motion vector for L0 prediction of (0, 0), a reference index of i, a motion vector of L1 prediction of (0, 0), and a reference index of i. The combined motion information candidate whose prediction type is bi-prediction was used. This is because, in a general moving image, the frequency of occurrence of combined motion information candidates in which the motion vector for L0 prediction and the motion vector for L1 prediction are (0, 0) is statistically high. The present invention is not limited to this as long as it is a combined motion information candidate that is statistically frequently used without depending on the motion information of the combined motion information candidate registered in the combined motion information candidate list. For example, the motion vectors of L0 prediction and L1 prediction may be vector values other than (0, 0), respectively, and may be set so that the reference indexes of L0 prediction and L1 prediction are different. Alternatively, the second additional combined motion information candidate can be set as an encoded image or motion information with a high occurrence frequency of a part of the encoded image, encoded in an encoded stream, and transmitted. Although the B picture (B slice) has been described here, in the case of the P picture (P slice), the second additional combined motion in which the motion vector for L0 prediction is (0, 0) and the prediction type is L0 prediction. Generate information candidates.
 ここで、第2追加結合動き情報候補として、L0予測の参照画像とL1予測の参照画像が同一である場合、第1追加結合動き情報候補リスト生成部と同様に、動き補償を双予測で行ってもL0予測もしくはL1予測の単予測と同じ結果が得られため、L0予測の参照画像とL1予測の参照画像が同一である追加結合動き情報候補生成は、動き補償予測の演算量を増加させる要因となる。しかし、本発明の実施の形態では後述する動き補償部で一括して双予測を単予測に変換する処理を行うため、第2追加結合動き情報候補リスト追加部でのL0予測の動き情報とL1予測の動き情報の同一性判断を行う必要はなく演算量が削減できる。 Here, when the reference image for L0 prediction and the reference image for L1 prediction are the same as the second additional combined motion information candidate, motion compensation is performed by bi-prediction in the same manner as the first additional combined motion information candidate list generation unit. However, since the same result as the single prediction of the L0 prediction or the L1 prediction is obtained, the additional combined motion information candidate generation in which the reference image of the L0 prediction and the reference image of the L1 prediction are the same increases the calculation amount of the motion compensation prediction. It becomes a factor. However, in the embodiment of the present invention, the motion compensation unit to be described later performs a process of collectively converting bi-prediction into single prediction, so that the motion information and L1 of the L0 prediction in the second additional combined motion information candidate list adding unit It is not necessary to determine the identity of predicted motion information, and the amount of calculation can be reduced.
 ここで、第2追加結合動き情報候補として結合動き情報候補リストに登録された結合動き情報候補に依存しない結合動き情報候補を設定することで、結合動き情報候補リストに登録された結合動き情報候補が0個である場合に、結合予測モードを利用することを可能とし、符号化効率を向上させることができる。また、結合動き情報候補リストに登録された結合動き情報候補の動き情報と処理対象の動き情報候補の動きが異なる場合に、新たな結合動き情報候補を生成して選択肢の幅を広げることで、符号化効率を向上させることができる。 Here, by setting a combined motion information candidate that does not depend on the combined motion information candidate registered in the combined motion information candidate list as the second additional combined motion information candidate, the combined motion information candidate registered in the combined motion information candidate list When the number is zero, it is possible to use the joint prediction mode and improve the encoding efficiency. In addition, when the motion information of the combined motion information candidate registered in the combined motion information candidate list and the motion information candidate motion to be processed are different, by generating a new combined motion information candidate and expanding the range of options, Encoding efficiency can be improved.
 図30は、図17のステップS1703における結合動き情報候補単予測変換処理の詳細動作を説明するためのフローチャートである。最初に、結合動き情報候補リスト生成処理により、生成された結合動き情報候補リストの数をnum_of_indexとすると、i=0からnum_of_index-1までの結合動き情報候補について以下の処理を繰り返し行う(S3000からS3005)。 FIG. 30 is a flowchart for explaining the detailed operation of the combined motion information candidate single prediction conversion process in step S1703 of FIG. First, assuming that the number of combined motion information candidate lists generated by the combined motion information candidate list generation process is num_of_index, the following processing is repeated for combined motion information candidates from i = 0 to num_of_index-1 (from S3000) S3005).
 先ず、結合動き情報候補リストよりインデックスiに格納された動き情報を取得する(S3001)。続いて、動き情報の予測種別が単予測である場合(S3002:YES)、そのままインデックスiに格納された動き情報に対しての処理を終了し、次のインデックスに進む(S3005)。 First, the motion information stored in the index i is acquired from the combined motion information candidate list (S3001). Subsequently, when the prediction type of the motion information is single prediction (S3002: YES), the process for the motion information stored in the index i is terminated as it is, and the process proceeds to the next index (S3005).
 動き情報が単予測でない場合、つまり動き情報が双予測である場合(S3002:NO)には、双予測の動き情報を単予測に変換するために、インデックスiに格納された動き情報のL1情報を無効にする(S3003)。実施の形態1においては、このようにL1情報を無効にすることで双予測の動き情報をL0予測の単予測に変換させているが、逆にL0情報を無効にして、双予測の動き情報をL1予測の単予測に変換させることも可能であり、暗黙に単予測に変換する際の無効にする予測種別を定義することで実現できる。 When the motion information is not uni-prediction, that is, when the motion information is bi-prediction (S3002: NO), the L1 information of the motion information stored in the index i is used to convert the bi-prediction motion information into uni-prediction. Is invalidated (S3003). In Embodiment 1, bi-prediction motion information is converted into L0 prediction single prediction by invalidating L1 information in this way, but conversely L0 information is invalidated and bi-prediction motion information. Can be converted to single prediction of L1 prediction, and can be realized by defining a prediction type to be invalidated when implicitly converting to single prediction.
 次に、単予測に変換したインデックスiの動き情報を格納し(S3004)、次のインデックスに進む(S3005)。i=0からnum_of_index-1までの結合動き情報候補について処理を行い、結合動き情報候補単予測変換処理が終了する。 Next, the motion information of the index i converted to single prediction is stored (S3004), and the process proceeds to the next index (S3005). The combined motion information candidates from i = 0 to num_of_index-1 are processed, and the combined motion information candidate single prediction conversion process is completed.
 実施の形態1において、予測ブロックサイズによる結合動き情報の双予測制限は、結合動き情報候補リストを一旦生成した後に、図30のフローチャートに示す結合動き情報候補の単予測変換処理を行っている。結合動き情報の単予測変換処理に関しては、結合動き情報候補生成処理である図18のフローチャートで示される処理の内部で、候補生成毎に判断を加えて、単予測の結合動き情報候補リストを生成することも可能であるが、その場合には、予測ブロックサイズによる条件判断が各処理に入ることになり、処理が複雑になり、リスト構築処理の負荷が増大する。実施の形態1では、一旦リストを構築した後に単予測への動き情報の変換処理を施すことで、リスト構築処理の負荷増大を防いだ、双予測の制限処理を実現できる効果を有する。 In the first embodiment, the combined motion information bi-prediction restriction based on the prediction block size is performed after the combined motion information candidate list is generated once and then the combined motion information candidate single prediction conversion process shown in the flowchart of FIG. 30 is performed. With regard to the combined motion information single prediction conversion process, a determination is made for each candidate generation within the process shown in the flowchart of FIG. 18 which is a combined motion information candidate generation process, and a combined motion information candidate list for single prediction is generated. However, in this case, the condition judgment based on the predicted block size enters each process, which complicates the process and increases the load of the list construction process. The first embodiment has an effect of realizing a bi-prediction restriction process that prevents an increase in the load of the list building process by performing a process of converting motion information into a single prediction after building a list once.
 図31は、図11のステップS1702における結合予測モード評価値生成処理の詳細動作を説明するためのフローチャートである。この動作は、図15の結合動き補償予測生成部1508を用いた構成の詳細動作を示している。 FIG. 31 is a flowchart for explaining the detailed operation of the combined prediction mode evaluation value generation process in step S1702 of FIG. This operation shows a detailed operation of the configuration using the combined motion compensation prediction generation unit 1508 of FIG.
 最初に予測誤差評価値を最大値に設定し、予測誤差が最小となる結合動き情報インデックスを初期化(例えば、-1などのリスト外の値)する(S3100)。結合動き情報候補リスト生成処理により、生成された結合動き情報候補リストの数をnum_of_indexとすると、i=0からnum_of_index-1までの結合動き情報候補について以下の処理を繰り返し行う(S3101からS3109)。 First, the prediction error evaluation value is set to the maximum value, and the combined motion information index that minimizes the prediction error is initialized (for example, a value outside the list such as −1) (S3100). If the number of combined motion information candidate lists generated by the combined motion information candidate list generation process is num_of_index, the following processing is repeated for combined motion information candidates from i = 0 to num_of_index-1 (S3101 to S3109).
 先ず、結合動き情報候補リストよりインデックスiに格納された動き情報を取得する(S3102)。続いて動き情報符号量を算出する(S3103)。結合予測モードにおいては、結合動き情報インデックスのみが符号化されるので、結合動き情報インデックスのみが動き情報符号量となる。 First, the motion information stored in the index i is acquired from the combined motion information candidate list (S3102). Subsequently, a motion information code amount is calculated (S3103). In the joint prediction mode, since only the joint motion information index is encoded, only the joint motion information index becomes the motion information code amount.
 結合動き情報インデックスの符号列としては、実施の形態1では、Truncated Unary符号列を用いる。図32は、結合動き情報候補数が5の場合のTruncated Unary符号列を示す図である。Truncated Unary符号列を用いて結合動き情報インデックスの値を符号化する場合には、結合動き情報インデックスが小さいほど、結合動き情報インデックスに割り当てられる符号ビットが小さくなる。例えば、結合動き情報候補数が5個である場合、結合動き情報インデックスが1であれば'10'の2ビットで表現されるが、結合動き情報インデックスが3であれば'1110'の4ビットで表現される。なお、ここでは上記のように結合動き情報インデックスの符号化にTruncated Unary符号列を利用しているが、他の符号列生成手法を用いることも可能であり、これに限定されない。 In the first embodiment, a Truncated Unary code string is used as the code string of the combined motion information index. FIG. 32 is a diagram showing a Trunked Unary code string when the number of combined motion information candidates is five. When the value of the combined motion information index is encoded using the Truncated Unary code string, the smaller the combined motion information index, the smaller the code bits assigned to the combined motion information index. For example, when the number of combined motion information candidates is 5, if the combined motion information index is 1, it is represented by 2 bits of “10”, but if the combined motion information index is 3, 4 bits of “1110”. It is expressed by Here, as described above, the Truncated Unary code string is used to encode the combined motion information index, but other code string generation methods can be used, and the present invention is not limited to this.
 続いて、動き情報の予測種別が単予測である場合(S3104:YES)、1つの参照画像に対する参照画像指定情報と動きベクトルを、図1における動き補償予測部112に設定して、動き補償単予測ブロックを生成する(S3105)。動き情報が単予測でない場合、つまり動き情報が双予測である場合(S3104:NO)には、2つの参照画像に対する参照画像指定情報と動きベクトルを動き補償予測部112に設定して、動き補償双予測ブロックを生成する(S3105)。 Subsequently, when the motion information prediction type is single prediction (S3104: YES), the reference image designation information and the motion vector for one reference image are set in the motion compensation prediction unit 112 in FIG. A prediction block is generated (S3105). When the motion information is not uni-prediction, that is, when the motion information is bi-prediction (S3104: NO), reference image designation information and motion vectors for two reference images are set in the motion compensation prediction unit 112, and motion compensation is performed. A bi-prediction block is generated (S3105).
 続いて、動き補償予測ブロックと予測対象ブロックの予測誤差と動き情報符号量より、予測誤差評価値を算出し(S3107)、予測誤差評価値が最小値である場合には評価値を更新すると共に、予測誤差最小インデックスを更新する(S3108)。 Subsequently, a prediction error evaluation value is calculated from the prediction error and the motion information code amount of the motion compensated prediction block and the prediction target block (S3107), and when the prediction error evaluation value is the minimum value, the evaluation value is updated. The prediction error minimum index is updated (S3108).
 全ての結合動き情報候補についての予測誤差評価値が比較された結果、選択された予測誤差最小インデックスを、結合予測モードで用いる結合動き情報インデックスとして、予測誤差最小値、動き補償予測ブロックと共に出力し(S3109)、結合予測モード評価値生成処理を終了する。 As a result of comparison of the prediction error evaluation values for all the combined motion information candidates, the selected prediction error minimum index is output together with the prediction error minimum value and the motion compensated prediction block as a combined motion information index used in the combined prediction mode. (S3109), the combined prediction mode evaluation value generation process is terminated.
 図33は、図17のステップS1703の予測モード評価値生成処理の詳細動作を説明するためのフローチャートである。 FIG. 33 is a flowchart for explaining the detailed operation of the prediction mode evaluation value generation processing in step S1703 of FIG.
 最初に、予測モードが単予測であるか否かを判別する(S3300)。図34に予測ブロックの動き情報に関するシンタックスを示す。図34中のmerge_flagは結合予測モードであるかどうかを示し、merge_flagが0の場合が動き検出予測モードを示す。動き検出予測モードである場合で双予測を用いることが可能なBスライスの場合、予測種別が単予測であるか双予測であるかを示すフラグinter_pred_flagを伝送する。ここで、予測ブロックのサイズが双予測制限ブロックサイズ以下である場合にも双予測を禁止せずinter_pred_flagを伝送する。これは、予測ブロックのサイズが双予測制限ブロックサイズ以下であるかどうかに応じてinter_pred_flagを伝送するかどうかを切り替えると、エントロピー符号化・復号化に条件分岐が必要になるため、処理が複雑になるのを防止するためである。 First, it is determined whether or not the prediction mode is single prediction (S3300). FIG. 34 shows a syntax regarding motion information of a prediction block. In FIG. 34, merge_flag indicates whether or not the mode is a joint prediction mode, and when merge_flag is 0, the motion detection prediction mode is indicated. In the case of a B slice in which bi-prediction can be used in the motion detection prediction mode, a flag inter_pred_flag indicating whether the prediction type is uni-prediction or bi-prediction is transmitted. Here, even when the size of the prediction block is equal to or smaller than the bi-prediction restricted block size, bi_pred_flag is transmitted without prohibiting bi-prediction. This is because the conditional branch is necessary for entropy coding / decoding when switching whether to transmit inter_pred_flag depending on whether the size of the prediction block is less than or equal to the bi-prediction restricted block size. This is to prevent this from happening.
 図33のフローチャートに戻り、単予測であれば(S3300:YES)、処理対象とする参照画像リスト(LX)を予測に用いている参照画像リストに設定する(S3301)。単予測でなければ、双予測であるから、この場合はLXをL0とする(S3302)。 33, if it is simple prediction (S3300: YES), the reference image list (LX) to be processed is set as the reference image list used for prediction (S3301). If it is not uni-prediction, it is bi-prediction, so LX is set to L0 in this case (S3302).
 次に、LX予測に対する参照画像指定情報(インデックス)と動きベクトル値を取得する(S3303)。続いて、予測ベクトル候補リストを生成し(S3304)、予測ベクトルの中から最適予測ベクトルを選択し、差分ベクトルを生成する(S3305)。最適予測ベクトルの選択手法は、予測ベクトルと伝送する動きベクトルの差分ベクトルを実際に符号化する際の符号量が最も少ないものを選択することが望ましいが、単純に差分ベクトルの水平・垂直成分の絶対値の総和が小さいものを選択するなどの手法で、簡易的に算出してもよい。 Next, reference image designation information (index) and motion vector values for LX prediction are acquired (S3303). Subsequently, a prediction vector candidate list is generated (S3304), an optimal prediction vector is selected from the prediction vectors, and a difference vector is generated (S3305). It is desirable to select the optimal prediction vector with the least amount of code when the difference vector between the prediction vector and the motion vector to be transmitted is actually encoded. However, the horizontal and vertical components of the difference vector are simply selected. The calculation may be simplified by a method such as selecting one having a small absolute sum.
 続いて、再度予測モードが単予測である否かを判別し(S3306)、予測モードが単予測であればステップS3311に進む。単予測でない、すなわち、双予測であれば、処理対象の参照リストLXがL1か否かを判定する(S3307)。参照リストLXがL1であれば、ステップS3311に進み、L1でない、すなわちL0の場合には、予測ブロックサイズがbipred_restriction_size以下である場合(S3308:YES)には、L1予測に対する情報を算出せず、予測モードを単予測に変換し(S3310)、ステップS3311に進む。 Subsequently, it is determined again whether or not the prediction mode is single prediction (S3306). If the prediction mode is single prediction, the process proceeds to step S3311. If it is not uni-prediction, that is, if it is bi-prediction, it is determined whether or not the reference list LX to be processed is L1 (S3307). If the reference list LX is L1, the process proceeds to step S3311, and if it is not L1, that is, if it is L0, if the predicted block size is equal to or smaller than bipred_restriction_size (S3308: YES), information for L1 prediction is not calculated, The prediction mode is converted to single prediction (S3310), and the process proceeds to step S3311.
 予測ブロックサイズがbipred_restriction_sizeより大きい場合(S3308:NO)には、LXをL1として(S3309)、ステップS3303からステップS3306までの処理と同じ処理が行われる。 When the predicted block size is larger than bipred_restriction_size (S3308: NO), LX is set to L1 (S3309), and the same processing as the processing from step S3303 to step S3306 is performed.
 実施の形態1においては、復号装置において、図34に示された予測ブロックの動き情報に関するシンタックスに従い復号を行った場合に、予測ブロックサイズに対する双予測制限時に対象となる予測ブロックサイズで双予測の動き情報が復号されないように、ステップS3308及びステップS3310の処理を用いて、予測モード評価値生成処理で双予測を制限する構成を取っている。 In the first embodiment, when decoding is performed according to the syntax related to motion information of the prediction block shown in FIG. 34 in the decoding apparatus, bi-prediction is performed with the target prediction block size when bi-prediction restriction is performed on the prediction block size. In order to prevent the motion information from being decoded, the process of steps S3308 and S3310 is used to limit the bi-prediction in the prediction mode evaluation value generation process.
 動きベクトル検出時に、双予測を想定した動きベクトル検出を行った場合には、単予測で用いられる動きベクトル情報と、上記ステップにおいて双予測を制限することで生成された単予測の動きベクトル情報が異なる場合があるため、単予測の新たな動き情報候補が登録されることで、単純に双予測の動き情報を使用しないように制限した場合よりも、符号化効率を向上させることができる。 When motion vector detection assuming bi-prediction is performed at the time of motion vector detection, motion vector information used in single prediction and motion vector information of single prediction generated by restricting bi-prediction in the above step are Since there may be different cases, by registering a new motion information candidate for uni-prediction, it is possible to improve the encoding efficiency as compared with the case where the motion information for bi-prediction is simply not used.
 続いて、動き情報符号量を算出する(S3311)。単予測モードの場合に、符号化する動き情報としては、1つの参照画像に対する、参照画像指定情報、差分ベクトル値、及び予測ベクトルインデックスの3要素、双予測モードの場合には、L0とL1の2つの参照画像に対する、参照画像指定情報、差分ベクトル値、及び予測ベクトルインデックスの計6要素となり、各々符号化された符号量の総量が動き情報符号量として算出される。本実施の形態における予測ベクトルインデックスの符号列生成手法としては、結合動き情報インデックスの符号列と同様にTruncated Unary符号列を用いるものとする。 Subsequently, a motion information code amount is calculated (S3311). In the case of the uni-prediction mode, the motion information to be encoded includes three elements of reference image designation information, a difference vector value, and a prediction vector index for one reference image. In the case of the bi-prediction mode, L0 and L1 The reference image designation information, the difference vector value, and the prediction vector index for the two reference images are a total of six elements, and the total amount of the encoded amount is calculated as the motion information code amount. As a prediction vector index code string generation method according to the present embodiment, a Truncated Unary code string is used in the same manner as the combined motion information index code string.
 続いて、参照画像に対する参照画像指定情報と動きベクトルを図1における動き補償予測部112に設定して、動き補償予測ブロックを生成する(S3312)。 Subsequently, reference image designation information and a motion vector for the reference image are set in the motion compensated prediction unit 112 in FIG. 1 to generate a motion compensated prediction block (S3312).
 さらに、動き補償予測ブロックと予測対象ブロックの予測誤差と動き情報符号量より、予測誤差評価値を算出し(S3313)、予測誤差評価値と、参照画像に対する動き情報である、参照画像指定情報と差分ベクトル値と予測ベクトルインデックスを動き補償予測ブロックと共に出力し(S3314)、予測モード評価値生成処理を終了する。 Further, a prediction error evaluation value is calculated from the prediction error and the motion information code amount of the motion compensated prediction block and the prediction target block (S3313), the prediction error evaluation value, and reference image designation information that is motion information for the reference image; The difference vector value and the prediction vector index are output together with the motion compensated prediction block (S3314), and the prediction mode evaluation value generation process ends.
 以上の処理が、実施の形態1における動画像符号化装置における動き補償予測ブロック構造選択部113の詳細動作となる。 The above processing is the detailed operation of the motion compensated prediction block structure selection unit 113 in the video encoding apparatus in the first embodiment.
 本発明の実施の形態1において、動き補償予測におけるメモリアクセス量を制限するための制御パラメータである、図10に示した、inter_4x4_enable及びinter_bipred_restriction_idcを、復号装置で認識するために伝送するシンタックスの一例を図35に示す。 In Embodiment 1 of the present invention, an example of syntax that is transmitted in order to recognize the inter_4x4_enable and inter_bipred_restriction_idc shown in FIG. 10, which are control parameters for limiting the memory access amount in motion compensation prediction, in the decoding apparatus Is shown in FIG.
 図35においては、図10に示した制御パラメータ値をそのまま、シーケンスや画像単位で設定するヘッダ情報の一部として伝送する構成を取っている。一例では、シーケンス単位のパラメータを伝送するseq_parameter_set_rbsp()の内部において伝送され、図3に示す最小CUサイズに対する情報がlog2_min_coding_block_size_minus3で8を基準とした(8×8を示す)2のべき乗値で定義され、最大CUサイズ(実施の形態1における符号化ブロックサイズ)が最大CU分割数(Max_CU_Depth)を示す値となるlog2_diff_max_min_coding_block_sizeとして、伝送される。 In FIG. 35, the control parameter values shown in FIG. 10 are transmitted as they are as part of the header information set for each sequence or image. In one example, it is transmitted inside seq_parameter_set_rbsp () that transmits parameters in sequence units, and the information for the minimum CU size shown in FIG. 3 is defined by a power of 2 based on 8 (indicating 8 × 8) in log2_min_coding_block_size_minus3 The maximum CU size (encoded block size in the first embodiment) is transmitted as log2_diff_max_min_coding_block_size having a value indicating the maximum number of CU divisions (Max_CU_Depth).
 inter_4x4_enableは、inter_4x4_enable_flagとして、log2_min_coding_block_size_minus3が0すなわち最小CUサイズが8×8の場合にのみ伝送されることで、inter_4x4_enableによる制御が有効な条件の時にのみ制御パラメータを送ることで、無効な制御情報の伝送を防ぐことができる。一方、inter_bipred_restriction_idcに関しては、最小CUサイズが16×16の場合においても制御に必要となるため、常に伝送される構成を取る。 inter_4x4_enable is transmitted only when log2_min_coding_block_size_minus3 is 0, that is, when the minimum CU size is 8 × 8, as inter_4x4_enable_flag, and by sending control parameters only when the control by inter_4x4_enable is valid, transmission of invalid control information Can be prevented. On the other hand, since inter_bipred_restriction_idc is necessary for control even when the minimum CU size is 16 × 16, a configuration in which it is always transmitted is adopted.
 実施の形態1においては、これらの制御パラメータ値をシーケンス単位のパラメータで符号化伝送する構成を示したが、フレーム単位等の所定符号化ブロック単位以上の間隔で設定を変更させることも可能であり、シーケンス単位での制御パラメータ構成に制限されることはなく、所定単位に制御パラメータを復号装置が取得できることが、実施の形態1における構成の特徴である。 In the first embodiment, a configuration is shown in which these control parameter values are encoded and transmitted using parameters in sequence units. However, it is also possible to change the setting at intervals of a predetermined encoded block unit or more such as a frame unit. The configuration of the first embodiment is not limited to the control parameter configuration in sequence units, and the decoding apparatus can acquire the control parameters in predetermined units.
 [実施の形態1における動画像復号装置における動き情報復号部の詳細動作説明]
 図36は、図11に示した実施の形態1の動画像復号装置における動き情報復号部1111の詳細な構成を示す図である。動き情報復号部1111は、動き情報ビットストリーム復号部3600、予測ベクトル算出部3601、ベクトル加算部3602、動き補償予測復号部3603、結合動き情報算出部3604、結合動き情報単予測変換部3605、及び結合動き補償予測復号部3606を含む。
[Detailed Operation Description of Motion Information Decoding Unit in Moving Picture Decoding Device in Embodiment 1]
FIG. 36 is a diagram showing a detailed configuration of the motion information decoding unit 1111 in the video decoding device according to Embodiment 1 shown in FIG. The motion information decoding unit 1111 includes a motion information bitstream decoding unit 3600, a prediction vector calculation unit 3601, a vector addition unit 3602, a motion compensation prediction decoding unit 3603, a combined motion information calculation unit 3604, a combined motion information single prediction conversion unit 3605, and A combined motion compensated prediction decoding unit 3606 is included.
 図11における動き情報復号部1111に対して、予測モード/ブロック構造復号部1108より入力された動き情報に関するビットストリームが、動き情報ビットストリーム復号部3600に供給され、予測モード情報メモリ1112より入力された動き情報が、予測ベクトル算出部3601、及び結合動き情報算出部3604に供給される。 A bit stream related to motion information input from the prediction mode / block structure decoding unit 1108 is supplied to the motion information bit stream decoding unit 3600 and input from the prediction mode information memory 1112 to the motion information decoding unit 1111 in FIG. The obtained motion information is supplied to the prediction vector calculation unit 3601 and the combined motion information calculation unit 3604.
 また、動き情報復号部1111に対して、動き補償予測復号部3603、及び結合動き補償予測復号部3606から、動き補償予測に用いる参照画像指定情報と動きベクトルが出力され、予測種別を示す情報を含めた復号された動き情報が、動き補償予測部1114及び予測モード情報メモリ1112に供給される。 Also, reference image designation information and motion vectors used for motion compensation prediction are output from the motion compensation prediction decoding unit 3603 and the joint motion compensation prediction decoding unit 3606 to the motion information decoding unit 1111 and information indicating the prediction type is output. The included decoded motion information is supplied to the motion compensation prediction unit 1114 and the prediction mode information memory 1112.
 動き情報ビットストリーム復号部3600は、入力された動き情報ビットストリームを符号化シンタックスに従って復号していくことで、伝送された予測モードと、予測モードに応じた動き情報を生成する。生成した動き情報の中で、結合動き情報インデックスは、結合動き補償予測復号部3606に供給され、参照画像指定情報が予測ベクトル算出部3601に供給され、予測ベクトルインデックスがベクトル加算部3602に供給され、差分ベクトル値がベクトル加算部3602に供給される。 The motion information bitstream decoding unit 3600 decodes the input motion information bitstream according to the encoding syntax, thereby generating the transmitted prediction mode and motion information corresponding to the prediction mode. Among the generated motion information, the combined motion information index is supplied to the combined motion compensated prediction decoding unit 3606, the reference image designation information is supplied to the prediction vector calculation unit 3601, and the prediction vector index is supplied to the vector addition unit 3602. The difference vector value is supplied to the vector addition unit 3602.
 予測ベクトル算出部3601は、予測モード情報メモリ1112より供給された隣接ブロックの動き情報と、動き情報ビットストリーム復号部3600より供給された参照画像指定情報から、動き補償予測の対象となる参照画像に対する予測ベクトル候補リストを生成し、参照画像指定情報と共にベクトル加算部3602に供給する。予測ベクトル算出部3601の動作に関しては、動画像符号化装置における図15の予測ベクトル算出部1502と同一の動作が行われ、符号化時の予測ベクトル候補リストと同一の候補リストが生成される。 The prediction vector calculation unit 3601 applies a motion compensation prediction target reference image based on the motion information of adjacent blocks supplied from the prediction mode information memory 1112 and the reference image designation information supplied from the motion information bitstream decoding unit 3600. A prediction vector candidate list is generated and supplied to the vector addition unit 3602 together with the reference image designation information. Regarding the operation of the prediction vector calculation unit 3601, the same operation as the prediction vector calculation unit 1502 of FIG. 15 in the moving image encoding apparatus is performed, and the same candidate list as the prediction vector candidate list at the time of encoding is generated.
 ベクトル加算部3602は、予測ベクトル算出部3601より供給された予測ベクトル候補リスト及び参照画像指定情報と、動き情報ビットストリーム復号部3600から供給された予測ベクトルインデックス及び差分ベクトルより、予測ベクトルインデックスで示された位置に登録された予測ベクトル値と差分ベクトル値を加算することで、動き補償予測対象となる参照画像に対しての動きベクトル値が再生される。再生された動きベクトル値は、参照画像指定情報と共に、動き補償予測復号部3603に供給される。 The vector addition unit 3602 indicates a prediction vector index from the prediction vector candidate list and reference image designation information supplied from the prediction vector calculation unit 3601, and the prediction vector index and difference vector supplied from the motion information bitstream decoding unit 3600. By adding the prediction vector value and the difference vector value registered at the set position, the motion vector value for the reference image to be motion compensated prediction is reproduced. The reproduced motion vector value is supplied to the motion compensated prediction decoding unit 3603 together with the reference image designation information.
 動き補償予測復号部3603は、ベクトル加算部2602より参照画像に対する、再生された動きベクトル値と参照画像指定情報が供給され、動きベクトル値と参照画像指定情報を動き補償予測部1114に設定することで、動き補償予測信号を生成する。 The motion compensation prediction decoding unit 3603 is supplied with the reproduced motion vector value and reference image designation information for the reference image from the vector addition unit 2602, and sets the motion vector value and the reference image designation information in the motion compensation prediction unit 1114. Thus, a motion compensated prediction signal is generated.
 結合動き情報算出部3604は、予測モード情報メモリ1112から供給される隣接ブロックの動き情報より、結合動き情報候補リストを生成し、結合動き情報候補リストとリスト内の構成要素である結合動き情報候補の参照画像指定情報と動きベクトル値を、結合動き情報単予測変換部3605に供給する。 The combined motion information calculation unit 3604 generates a combined motion information candidate list from the motion information of adjacent blocks supplied from the prediction mode information memory 1112 and combines the combined motion information candidate list and the combined motion information candidate that is a component in the list. The reference image designation information and the motion vector value are supplied to the combined motion information single prediction conversion unit 3605.
 結合動き情報算出部3604の動作に関しては、動画像符号化装置における図15の結合動き情報算出部1506と同一の動作が行われ、符号化時の結合動き情報候補リストと同一の候補リストが生成される。 Regarding the operation of the combined motion information calculation unit 3604, the same operation as the combined motion information calculation unit 1506 in FIG. 15 in the moving image encoding apparatus is performed, and the same candidate list as the combined motion information candidate list at the time of encoding is generated. Is done.
 結合動き情報単予測変換部3605では、動画像符号化装置における図15の結合動き情報単予測変換部1507と同一の動作が行われ、結合動き情報算出部3604より供給される結合動き情報候補リスト及び、候補リストに登録される動き情報に対して、図10で示される双予測制限情報に従って、予測種別が双予測である動き情報を単予測の動き情報に変換し、結合動き補償予測復号部3606に供給する。 The combined motion information single prediction conversion unit 3605 performs the same operation as the combined motion information single prediction conversion unit 1507 of FIG. 15 in the moving image encoding device, and is supplied from the combined motion information calculation unit 3604. And, for the motion information registered in the candidate list, the motion information whose prediction type is bi-prediction is converted into motion information of uni-prediction according to the bi-prediction restriction information shown in FIG. 3606.
 結合動き補償予測復号部3606は、結合動き情報単予測変換部3605より供給される結合動き情報候補リストとリスト内の構成要素である結合動き情報候補の参照画像指定情報と動きベクトル値と、動き情報ビットストリーム復号部3600より供給される結合動き情報インデックスより、結合動き情報インデックスで示された結合動き情報候補リストにおける参照画像指定情報と動きベクトル値を再生し、動き補償予測部1114に設定することで、動き補償予測信号を生成する。 The combined motion compensated prediction decoding unit 3606 includes a combined motion information candidate list supplied from the combined motion information single prediction conversion unit 3605, reference image designation information of a combined motion information candidate that is a component in the list, a motion vector value, and motion. Based on the combined motion information index supplied from the information bitstream decoding unit 3600, the reference image designation information and the motion vector value in the combined motion information candidate list indicated by the combined motion information index are reproduced and set in the motion compensation prediction unit 1114. Thus, a motion compensated prediction signal is generated.
 図37は、図14のステップS1402、S1405、S1408、S1410の予測ブロック単位復号処理の詳細動作を説明するためのフローチャートである。最初に、CU単位の符号化ストリームを取得し(S3700)、CU内の予測ブロックサイズ分割モード(PU)に従って設定されたNumPartに基づき、対象CU内をPU分割した予測ブロックサイズ毎に(S3701)、ステップS3702からステップS3706までのステップが実行される(S3707)。 FIG. 37 is a flowchart for explaining the detailed operation of the prediction block unit decoding processing in steps S1402, S1405, S1408, and S1410 of FIG. First, an encoded stream of a CU unit is acquired (S3700), and for each prediction block size obtained by performing PU division on the target CU based on NumPart set according to the prediction block size division mode (PU) in the CU (S3701). Steps S3702 to S3706 are executed (S3707).
 CU単位の符号化ストリームより分離された動き情報の符号化列は、図11の予測モード/ブロック構造復号部1108より動き情報復号部1111に供給され、予測モード情報メモリ1112より供給される候補ブロック群の動き情報を用いて、復号対象ブロックの動き情報を復号する(S3702)。ステップS3702の処理の詳細については後述する。 The encoded sequence of motion information separated from the encoded stream of the CU unit is supplied to the motion information decoding unit 1111 from the prediction mode / block structure decoding unit 1108 of FIG. 11 and is supplied from the prediction mode information memory 1112. The motion information of the decoding target block is decoded using the group motion information (S3702). Details of the processing in step S3702 will be described later.
 分離された予測誤差情報の符号化列は、予測差分情報復号部1102に供給され、量子化された予測誤差信号として復号され、逆量子化・逆変換部1103で逆量子化や逆直交変換などの処理を施されることで、復号予測誤差信号が生成される(S3703)。 The separated coded sequence of prediction error information is supplied to the prediction difference information decoding unit 1102 and decoded as a quantized prediction error signal, and the inverse quantization / inverse transformation unit 1103 performs inverse quantization, inverse orthogonal transform, etc. By performing the above process, a decoded prediction error signal is generated (S3703).
 動き情報復号部1111より、復号対象ブロックの動き情報が動き補償予測部1114に供給され、動き補償予測部1114は動き情報にしたがって動き補償予測を行い、予測信号を算出する(S3704)。加算部1104は、逆量子化・逆変換部1103から供給された復号予測誤差信号と、動き補償予測部1114から予測モード/ブロック構造選択部1109に供給され、更に予測モードで動き補償予測が選択されることで加算部1104に供給された予測信号とを加算し、復号画像信号を生成する(S3705)。 The motion information decoding unit 1111 supplies the motion information of the decoding target block to the motion compensation prediction unit 1114, and the motion compensation prediction unit 1114 performs motion compensation prediction according to the motion information and calculates a prediction signal (S3704). The adder 1104 supplies the decoded prediction error signal supplied from the inverse quantization / inverse transform unit 1103 and the motion compensation prediction unit 1114 to the prediction mode / block structure selection unit 1109, and further selects motion compensation prediction in the prediction mode. As a result, the prediction signal supplied to the addition unit 1104 is added to generate a decoded image signal (S3705).
 加算部1104より供給された復号画像信号は、フレーム内復号画像バッファ1105に格納されると共に、ループフィルタ部1106に供給される。また、動き情報復号部1111より供給された復号対象ブロックの動き情報が予測モード情報メモリ1112に格納される(S3706)。これが、対象CU内のすべての予測ブロックに対して施されることにより、予測ブロック単位の復号処理が終了する。 The decoded image signal supplied from the adding unit 1104 is stored in the intra-frame decoded image buffer 1105 and also supplied to the loop filter unit 1106. Also, the motion information of the decoding target block supplied from the motion information decoding unit 1111 is stored in the prediction mode information memory 1112 (S3706). This is applied to all the prediction blocks in the target CU, thereby completing the decoding process for each prediction block.
 図38は、図37のステップS3702の動き情報復号処理の詳細動作を説明するためのフローチャートである。動き情報ビットストリーム復号部3600と予測ベクトル算出部3601、及び結合動き情報算出部3604により、図37のステップS3702の動き情報復号処理が行われる。 FIG. 38 is a flowchart for explaining the detailed operation of the motion information decoding process in step S3702 of FIG. The motion information decoding process of step S3702 of FIG. 37 is performed by the motion information bitstream decoding unit 3600, the prediction vector calculation unit 3601, and the combined motion information calculation unit 3604.
 動き情報復号処理は、特定のシンタックス構造で符号化された符号化ビットストリームより動き情報を復号する処理である。最初に符号化ブロックのCU単位で復号されたSkipフラグがSkipモードを示している場合(S3800:YES)、結合予測動き情報復号を行う(S3801)。ステップS3801の詳細処理については、後述する。 The motion information decoding process is a process of decoding motion information from an encoded bit stream encoded with a specific syntax structure. When the Skip flag decoded first in the CU unit of the encoded block indicates the Skip mode (S3800: YES), joint prediction motion information decoding is performed (S3801). Detailed processing in step S3801 will be described later.
 一方、Skipモードでない場合には(S3800:NO)、マージフラグを復号する(S3802)。マージフラグが1を示している場合(S3803:YES)には、ステップS3801の結合予測動き情報復号に進む。 On the other hand, if it is not the Skip mode (S3800: NO), the merge flag is decoded (S3802). When the merge flag indicates 1 (S3803: YES), the process proceeds to joint prediction motion information decoding in step S3801.
 マージフラグが1でない場合(S3803:NO)、動き予測フラグを復号し(S3804)、予測動き情報復号を行い(S3805)、処理を終了する。ステップS3805の詳細動作については、後述する。 If the merge flag is not 1 (S3803: NO), the motion prediction flag is decoded (S3804), the prediction motion information is decoded (S3805), and the process is terminated. Detailed operation of step S3805 will be described later.
 図39は、図38のステップS3801の結合予測動き情報復号処理の詳細動作を説明するためのフローチャートである。 FIG. 39 is a flowchart for explaining the detailed operation of the joint prediction motion information decoding process in step S3801 of FIG.
 最初に予測モードに結合予測モードを設定し(S3900)、結合動き情報候補リストを生成する(S3901)。ステップS3901の処理は、動画像符号化装置における図17のステップS1701の結合動き情報候補リスト生成処理と同一の処理である。 First, the combined prediction mode is set as the prediction mode (S3900), and a combined motion information candidate list is generated (S3901). The process of step S3901 is the same process as the combined motion information candidate list generation process of step S1701 of FIG. 17 in the video encoding device.
 続いて、予測ブロックサイズが、図10に示した双予測を制限する制御パラメータinter_bipred_restriction_idcにより設定される双予測を制限する予測ブロックサイズであるbipred_restriction_size以下である場合(S3902:YES)には、格納された結合動き情報候補リスト内の各候補における双予測の動き情報を単予測の動き情報に置き換える、結合動き情報候補単予測変換を行う(S3903)。この処理では、図30のフローチャートで示した符号化装置における結合動き情報単予測変換処理と同じ処理が施される。予測ブロックサイズが、bipred_restriction_size以下でない場合(S3902:NO)には、ステップS3904に進む。 Subsequently, when the prediction block size is equal to or smaller than bipred_restriction_size that is a prediction block size that restricts bi-prediction set by the control parameter inter_bipred_restriction_idc that restricts bi-prediction shown in FIG. 10 (S3902: YES), it is stored. The combined motion information candidate single prediction conversion is performed in which the bi-prediction motion information in each candidate in the combined motion information candidate list is replaced with the single prediction motion information (S3903). In this process, the same process as the combined motion information single prediction conversion process in the encoding apparatus shown in the flowchart of FIG. 30 is performed. If the predicted block size is not less than or equal to bipred_restriction_size (S3902: NO), the process proceeds to step S3904.
 つぎに、結合動き情報インデックスを復号し(S3904)、続いて、結合動き情報候補リストより、結合動き情報インデックスで示す位置に格納されている動き情報を取得する(S3905)。取得する動き情報としては、単予測/双予測を示す予測種別、参照画像指定情報、動きベクトル値となる。 Next, the combined motion information index is decoded (S3904), and then the motion information stored at the position indicated by the combined motion information index is acquired from the combined motion information candidate list (S3905). The motion information to be acquired includes a prediction type indicating single prediction / bi-prediction, reference image designation information, and a motion vector value.
 実施の形態1においては、結合動き情報の双予測から単予測への変換処理は、結合動き情報のインデックスの値を変更することがないため、復号装置においては復号に必要なインデックスの結合動き情報にのみ変換処理を行うことも可能であり、その場合には図39のステップS3904及びステップS3905を行った後に、予測ブロックサイズによる双予測の制限を行うステップS3902及びステップS3903が行われる。 In Embodiment 1, since the conversion process from the bi-prediction to the single prediction of the combined motion information does not change the index value of the combined motion information, the combined motion information of the index necessary for decoding in the decoding device In this case, step S3904 and step S3903 for restricting bi-prediction based on the prediction block size are performed after performing step S3904 and step S3905 in FIG.
 生成された動き情報は、結合予測モードの動き情報として格納され(S3906)、結合動き補償予測復号部3606に供給される。 The generated motion information is stored as motion information in the joint prediction mode (S3906), and is supplied to the joint motion compensation prediction decoding unit 3606.
 図40は、図38のステップS3805の予測動き情報復号処理の詳細動作を説明するためのフローチャートである。 FIG. 40 is a flowchart for explaining the detailed operation of the predicted motion information decoding process in step S3805 of FIG.
 最初に、予測種別が単予測であるか否かを判別する(S4000)。単予測であれば、処理対象とする参照画像リスト(LX)を予測に用いている参照画像リストに設定する(S4001)。単予測でなければ、双予測であるから、この場合はLXをL0とする(S4002)。 First, it is determined whether or not the prediction type is simple prediction (S4000). If it is single prediction, the reference image list (LX) to be processed is set as the reference image list used for prediction (S4001). If it is not uni-prediction, it is bi-prediction, so LX is set to L0 in this case (S4002).
 次に、参照画像指定情報を復号し(S4003)、差分ベクトル値を復号する(S4004)。次に、予測ベクトル候補リストを生成し(S4005)、予測ベクトル候補リストが1より大きな場合(S4006:YES)、予測ベクトルインデックスを復号し(S4007)、予測ベクトル候補リストが1の場合(S4006:NO)、予測ベクトルインデックスに0を設定する(S4008)。 Next, the reference image designation information is decoded (S4003), and the difference vector value is decoded (S4004). Next, a prediction vector candidate list is generated (S4005). When the prediction vector candidate list is larger than 1 (S4006: YES), the prediction vector index is decoded (S4007), and when the prediction vector candidate list is 1 (S4006: NO), 0 is set to the prediction vector index (S4008).
 ここで、ステップS4005では、動画像符号化装置における図33のフローチャートのステップS3304と同様の処理が行われる。 Here, in step S4005, processing similar to that in step S3304 in the flowchart of FIG. 33 in the video encoding device is performed.
 次に、予測ベクトル候補リストより、予測ベクトルインデックスで示す位置に格納されている動きベクトル値を取得する(S4009)。復号した差分ベクトル値と動きベクトル値を加算することで動きベクトルを再生する(S4010)。 Next, the motion vector value stored at the position indicated by the prediction vector index is acquired from the prediction vector candidate list (S4009). A motion vector is reproduced by adding the decoded difference vector value and motion vector value (S4010).
 続いて、再度予測種別が単予測である否かを判別し(S4011)、予測種別が単予測であればステップS4014に進む。単予測でない、すなわち、双予測であれば、処理対象の参照リストLXがL1か否かを判定する(S4012)。参照リストLXがL1であれば、ステップS4014に進み、L1でない、すなわち、L0であれば予測ブロックサイズがbipred_restrcition_size以下である場合には(S4013:YES)、ステップS4016に進み、予測ブロックサイズがbipred_restriction_sizeより大きい場合には(S4013:NO)、LXをL1として(S4015)、ステップS4003からステップS4011までの処理と同じ処理が行われる。 Subsequently, it is determined again whether or not the prediction type is single prediction (S4011). If the prediction type is single prediction, the process proceeds to step S4014. If it is not uni-prediction, that is, if it is bi-prediction, it is determined whether or not the reference list LX to be processed is L1 (S4012). If the reference list LX is L1, the process proceeds to step S4014. If it is not L1, that is, if it is L0, the predicted block size is equal to or smaller than bipred_restrcition_size (S4013: YES), the process proceeds to step S4016, and the predicted block size is bipred_restriction_size. If larger (S4013: NO), LX is set to L1 (S4015), and the same processing as the processing from step S4003 to step S4011 is performed.
 予測ブロックサイズがbipred_restrcition_size以下である場合には、双予測の動き補償が禁止されるため、復号装置におけるメモリアクセス量の制限を確実にするために、伝送された動き情報を単予測に変換して(S4016)、ステップS4014に進む。 When the prediction block size is less than or equal to bipred_restrcition_size, bi-prediction motion compensation is prohibited, so the transmitted motion information is converted to single prediction to ensure the memory access amount limit in the decoding device. (S4016), the process proceeds to step S4014.
 続いて、生成された動き情報として、単予測の場合には、1つの参照画像に対する参照画像指定情報と動きベクトル値が、双予測の場合には、2つの参照画像に対する参照画像指定情報と動きベクトル値が、動き情報として格納され(S4014)、動き補償予測復号部3603に供給される。 Subsequently, as the generated motion information, in the case of single prediction, reference image designation information and motion vector values for one reference image, and in the case of bi-prediction, reference image designation information and motion for two reference images. The vector value is stored as motion information (S4014) and supplied to the motion compensated prediction decoding unit 3603.
 実施の形態1における、予測動き情報復号処理においては、符号化時に伝送された動き情報に対してシンタックスに従って復号を行うため、動画像符号化装置における、図33の予測モード評価値生成処理において施しているような、メモリアクセス量の制限を確実にするために双予測制限に関する条件分岐を行う、ステップS4013の条件判断及びステップS4016の処理を省略する形態においても実現可能であるが、実施の形態1においては、メモリ帯域の制限を復号装置においても確実にする構成として、図40のフローチャートによる予測動き情報復号処理をとる。 In the prediction motion information decoding process according to the first embodiment, the motion information transmitted at the time of encoding is decoded according to the syntax. Therefore, in the prediction mode evaluation value generation process of FIG. Although it is possible to implement the conditional branching regarding the bi-prediction restriction to ensure the restriction of the memory access amount as in the case where the condition determination in step S4013 and the process in step S4016 are omitted, In the first embodiment, the prediction motion information decoding process according to the flowchart of FIG. 40 is adopted as a configuration that ensures the limitation of the memory band even in the decoding device.
 図41は、図35に示したような、シーケンス単位のパラメータを伝送するseq_parameter_set_rbsp()等で、符号化処理/復号処理の最大画像サイズもしくは所定時間単位の最大処理画素数を定義するlevel_idcが伝送される場合に、参照画像のメモリアクセス量の負荷は、最大処理画素数に比例して増大するため、使用可能な最大処理画素数に連動させて、動き補償予測の予測ブロックサイズと双予測の制限を加える構成の一例である。符号化装置で定義され伝送されるlevel_idcに応じて、inter_4x4_enable及びinter_bipred_restriction_idcの取り得る値に制限を加えることで、符号化装置・復号装置の想定される画像サイズに応じたメモリアクセスの制限をかけることが可能となり、符号化装置及び復号装置の用途に応じて、必要なメモリ帯域を確保できる、処理負荷及び装置の規模を削減しつつ符号化効率を維持できる符号化装置及び復号装置が実現できる。 FIG. 41 is a seq_parameter_set_rbsp () etc. that transmits parameters in sequence as shown in FIG. 35, and a level_idc that defines the maximum image size of encoding / decoding processing or the maximum number of pixels for a predetermined time unit is transmitted. In this case, since the load of the memory access amount of the reference image increases in proportion to the maximum number of processing pixels, the prediction block size and the bi-prediction of motion compensation prediction are linked to the maximum number of usable processing pixels. It is an example of the structure which adds a restriction | limiting. By limiting the possible values of inter_4x4_enable and inter_bipred_restriction_idc according to the level_idc defined and transmitted by the encoding device, the memory access is limited according to the assumed image size of the encoding device / decoding device. Therefore, according to the use of the encoding device and the decoding device, it is possible to realize an encoding device and a decoding device that can secure the necessary memory bandwidth and can maintain the encoding efficiency while reducing the processing load and the scale of the device.
 図41は、一例としてlevel_idcが6段階に設定されている場合に、少ない画素数の符号化を想定した条件の場合には、inter_4x4_enableは制約なく(0と1のどちらも設定可能)、inter_bipred_restriction_idcに対しても定義された値のすべてを設定可能としているが、level_idcの増加に伴い、図9で示したメモリアクセス量の大きな予測処理から段階的に予測ブロックサイズ及び双予測の制限を加えることで、inter_4x4_enable(常に0のみにする)及びinter_bipred_restriction_idc(取り得る値の最小値を大きくする)を、最大画像サイズや最大処理画素数と連動して制御することができる。 In FIG. 41, as an example, when level_idc is set to 6 levels, inter_4x4_enable is not restricted (in the case of 0 and 1 can be set) under the condition that encoding with a small number of pixels is assumed, inter_bipred_restriction_idc All of the defined values can be set, but with the increase of level_idc, the prediction block size and bi-prediction restrictions are added step by step from the prediction process with a large memory access amount shown in FIG. , Inter_4x4_enable (always set to only 0) and inter_bipred_restriction_idc (increase the minimum value of possible values) can be controlled in conjunction with the maximum image size and the maximum number of processed pixels.
 また、図41のように、level_idcを基準に最大画像サイズや最大処理画素数と連動してinter_4x4_enableやinter_bipred_restriction_idcの値を、伝送せずに暗黙で制限下の固定値に設定し、符号化装置・復号装置にて、設定された制限による動き補償予測及び双予測の制限をおこなうことも可能であり、その場合にはlevel_idcが伝送されることで、対応したinter_4x4_enableやinter_bipred_restriction_idcの値が復号できる構成になる。 Further, as shown in FIG. 41, the value of inter_4x4_enable and inter_bipred_restriction_idc is implicitly set to a fixed value under restriction without being transmitted in conjunction with the maximum image size and the maximum number of processed pixels with reference to level_idc. In the decoding device, it is also possible to restrict motion compensation prediction and bi-prediction according to the set restriction.In that case, the level_idc is transmitted, so that the corresponding inter_4x4_enable and inter_bipred_restriction_idc values can be decoded. Become.
 実施の形態1においては、inter_4x4_enableという4×4予測ブロックサイズの動き補償予測を禁止する制御パラメータを用いているが、動き補償予測の予測ブロック制限に関しても、inter_bipred_restriction_idcと同様に、指定された予測ブロックサイズ以下のブロックサイズの動き補償予測を禁止する制御パラメータを用いることも可能であり、よりメモリアクセス量を細かく制御することを可能にする。 In Embodiment 1, the control parameter prohibiting motion compensated prediction of 4 × 4 prediction block size called inter_4x4_enable is used, but the prediction block restriction of motion compensated prediction is the same as the inter_bipred_restriction_idc specified prediction block. It is also possible to use a control parameter that prohibits motion compensation prediction of a block size equal to or smaller than the size, which makes it possible to control the memory access amount more finely.
 実施の形態1において、4×8画素と8×4画素のように、予測ブロックサイズの面積は同一で水平・垂直の画素数が異なる場合の双予測の制限を、同一基準で行っているが、一般的に参照画像メモリのアクセス単位が水平方向に4画素や8画素などの複数画素で構成される場合が多いことを想定し、水平方向の画素数が少ない4×8画素を、よりメモリアクセス量が多い予測ブロックサイズであると定義して、動き補償予測や双予測の制限を加えることも可能であり、より復号装置の構成に適したメモリアクセス量の制御が可能となる。 In the first embodiment, bi-prediction restriction is performed on the same basis when the area of the prediction block size is the same and the number of horizontal and vertical pixels is different, such as 4 × 8 pixels and 8 × 4 pixels. Assuming that the access unit of the reference image memory is generally composed of a plurality of pixels such as 4 pixels or 8 pixels in the horizontal direction, 4 × 8 pixels having a small number of pixels in the horizontal direction are more It is also possible to define motion block prediction and bi-prediction by defining a prediction block size with a large access amount, and it is possible to control the memory access amount more suitable for the configuration of the decoding device.
 また、実施の形態1において、動き補償予測の効率を向上させるために、図42に示すように、CU内の分割を更に細かくて左右や上下が非対称の予測ブロックを定義する場合においても、非対称のブロックに対する、予測ブロックサイズの制限を加えることにより、段階的なメモリアクセス量の制御が可能である。 Further, in the first embodiment, in order to improve the efficiency of motion compensation prediction, as shown in FIG. 42, even when a prediction block having a more detailed division within the CU and asymmetric left and right and up and down is defined. By adding a restriction on the predicted block size for these blocks, it is possible to control the memory access amount in stages.
 図42に示す、実施の形態1の別構成における、CUの予測ブロックへの分割構成は、非分割(2N×2N)、水平・垂直への分割(N×N)、水平方向のみへの分割(2N×N)、垂直方向のみへの分割(N×2N)に加えて、水平方向のみへの上1/4、下3/4の非対称分割(2N×nU)、水平方向のみへの上3/4、下1/4の非対称分割(2N×nD)、垂直方向のみへの左1/4、右3/4の非対称分割(nL×2N)、垂直方向のみへの左3/4、右1/4の非対称分割(nR×2N)となっており、水平4画素、垂直4画素未満の予測ブロックサイズを適用しないように、CUサイズが16×16以上のCUにのみ、非対称分割の分割構成が適用可能になっている。 In the other configuration of the first embodiment shown in FIG. 42, the division configuration of the CU into prediction blocks is non-division (2N × 2N), horizontal / vertical division (N × N), and division only in the horizontal direction. (2N × N), vertical division only (N × 2N), horizontal only upper 1/4, lower 3/4 asymmetric division (2N × nU), horizontal upper only 3/4, lower 1/4 asymmetric division (2N × nD), left 1/4 to vertical only, right 3/4 asymmetric division (nL × 2N), left 3/4 to vertical only, Asymmetrical division (nR × 2N) of the right 1/4, so that a prediction block size of less than 4 horizontal pixels and 4 vertical pixels is not applied, only a CU with a CU size of 16 × 16 or larger is used. A split configuration is applicable.
 続いて、図43に図42の予測ブロック構成における、動き補償予測のブロックサイズ及び予測処理を制限する制御パラメータの一例を示し、説明する。制御パラメータは、最も小さなCUサイズである8×8ブロックを分割する構成である、4×4、4×8及び8×4予測ブロックの動き補償予測の有効・無効を制御するパラメータである、inter_pred_enable_idcと、動き補償予測の内、双予測が施される予測処理のみを禁止するブロックサイズを定義する、inter_bipred_restriction_idcの2つのパラメータで構成される。 Subsequently, FIG. 43 shows an example of the block size of motion compensation prediction and control parameters for limiting the prediction processing in the prediction block configuration of FIG. The control parameter is a parameter for controlling validity / invalidity of motion compensated prediction of 4 × 4, 4 × 8, and 8 × 4 prediction blocks, which is a configuration that divides an 8 × 8 block that is the smallest CU size, and inter_pred_enable_idc And inter_bipred_restriction_idc that defines a block size that prohibits only the prediction processing in which bi-prediction is performed in motion compensation prediction.
 図43の制御パラメータの構成では、inter_bipred_restriction_idcに対して、水平・垂直の画素数のメモリアクセスに与える影響を加味した、16×16画素以下の予測ブロックサイズの大きさの順番を小さい方から、4×4、4×8、8×4、8×8、4×16/12×16(nL×2N/nR×2N)、8×16、16×12/16×4(2N×nU/2N×nD)、16×8、16×16と定義し、双予測を制限する予測ブロックサイズの値を設定する。これにより、動き補償予測の効率を向上させた非対称構成の予測ブロックに対しても、図10で示した制御パラメータを用いた構成と同様に、細かい単位でのメモリアクセス量の制御を行うことが可能となり、動き補償予測の効率を向上させたうえで、許容されるメモリ帯域に応じた、メモリアクセス量の制御を可能とする。 In the configuration of the control parameters in FIG. 43, the order of the size of the prediction block size of 16 × 16 pixels or less, which takes into account the influence on the memory access of the horizontal and vertical number of pixels with respect to inter_bipred_restriction_idc, × 4, 4 × 8, 8 × 4, 8 × 8, 4 × 16/12 × 16 (nL × 2N / nR × 2N), 8 × 16, 16 × 12/16 × 4 (2N × nU / 2N × nD), 16 × 8, and 16 × 16, and sets a prediction block size value that restricts bi-prediction. As a result, the memory access amount can be controlled in a fine unit even for a prediction block having an asymmetric configuration in which the efficiency of motion compensation prediction is improved, as in the configuration using the control parameters shown in FIG. Thus, the efficiency of motion compensation prediction is improved, and the memory access amount can be controlled according to the allowable memory bandwidth.
 実施の形態1においては、inter_bipred_restriction_idcで定義された予測ブロックサイズを基準に、定義されたサイズ以下の予測ブロックに双予測の制限を加えているが、値による制限を、定義されたサイズ未満の予測ブロックに双予測の制限を加えるようにすることも、双予測の制限を加える予測ブロックサイズ未満の予測ブロックサイズにおいて動き補償予測を行わない場合に、定義されたサイズの予測ブロックに双予測の制限を加えるようにすることも、本発明を実現する構成として可能である。定義されたサイズ未満の予測ブロックに双予測の制限を加えるようにする場合には、実施の形態1の符号化装置における、図17のフローチャートで示すステップS1702、図33のフローチャートで示すステップS3308、実施の形態1の復号装置における、図39のフローチャートで示すステップS3902、図40のフローチャートで示すステップS4013における条件判断が、bipred_restriction_size未満で有るか否かになると共に、inter_bipred_restriction_idcで定義する予測ブロックサイズの値が1つ大きい予測ブロックサイズとして設定されることで、実現される。 In Embodiment 1, bi-prediction restriction is applied to a prediction block having a size equal to or smaller than the defined size on the basis of the prediction block size defined by inter_bipred_restriction_idc. Limiting bi-prediction to a block is also possible, and bi-prediction is limited to a prediction block of a defined size if motion compensated prediction is not performed at a prediction block size that is smaller than the prediction block size to which bi-prediction restriction is applied. It is also possible to add the above as a configuration for realizing the present invention. When the restriction of bi-prediction is added to a prediction block having a size smaller than the defined size, step S1702 shown in the flowchart of FIG. 17 and step S3308 shown in the flowchart of FIG. 33 in the encoding apparatus of the first embodiment. In the decoding apparatus of the first embodiment, whether or not the condition judgment in step S3902 shown in the flowchart of FIG. 39 and step S4013 shown in the flowchart of FIG. 40 is less than bipred_restriction_size, and the prediction block size defined by inter_bipred_restriction_idc This is realized by setting the value as a prediction block size one larger.
 実施の形態1においては、図35に示したように、動き補償予測におけるメモリアクセス量を制限するための制御パラメータである、inter_4x4_enable及びinter_bipred_restriction_idcを、それぞれ個別のパラメータとして符号化伝送する構成を一例として示したが、これらの制御パラメータ情報が、動画像符号化装置及び動画像復号装置のメモリアクセス量制限を制御するパラメータとして伝送できる構成であれば、図44に示すようなinter_4x4_enableとinter_bipred_restriction_idcの組合せで定義する情報(inter_mc_restrcution_idc)を符号化伝送する構成も可能であり、所定もしくは所定以下の予測ブロックサイズでの動き補償予測処理を行わないように制御する情報と、所定以下の予測ブロックサイズでの双予測を行わないように制御する情報により、更に動き補償予測及び結合動き情報候補の単予測制限を行う処理を1つの指示情報に統合して符号化伝送及び復号できる効果が生じる。 In the first embodiment, as shown in FIG. 35, as an example, a configuration in which inter_4x4_enable and inter_bipred_restriction_idc, which are control parameters for limiting the memory access amount in motion compensation prediction, are encoded and transmitted as individual parameters, respectively. As shown in FIG. 44, if the control parameter information can be transmitted as a parameter for controlling the memory access amount restriction of the video encoding device and the video decoding device, the combination of inter_4x4_enable and inter_bipred_restriction_idc as shown in FIG. A configuration in which the information to be defined (inter_mc_restrcution_idc) is encoded and transmitted is also possible. Information that is controlled so as not to perform motion compensation prediction processing with a prediction block size that is predetermined or less than a predetermined value and dual information with a prediction block size that is less than a predetermined value The motion compensation is further improved by the information that controls so as not to perform the prediction There is an effect that the processing for performing compensation prediction and single prediction restriction of combined motion information candidates can be integrated into one instruction information and encoded transmission and decoding can be performed.
 また、実施の形態1においては、動き補償予測に対してメモリアクセス量を制限するための、結合動き補償予測に用いられる双予測を禁止する手段として、結合動き情報候補インデックスに格納された後の動き情報を、条件に応じて双予測の動き情報から単予測の動き情報に変換し格納し、予測処理に使用するため、双予測の結合動き情報候補を禁止するのでなく、単予測の動き情報として用いることが可能となり、双予測を禁止する条件の予測ブロックサイズにおける動き補償予測の予測精度が向上し、符号化効率が向上する効果を有する。 Further, in the first embodiment, as a means for prohibiting bi-prediction used for joint motion compensation prediction to limit the memory access amount with respect to motion compensation prediction, after being stored in the joint motion information candidate index Motion information is converted from bi-prediction motion information to uni-prediction motion information according to conditions, stored, and used for prediction processing. As a result, the prediction accuracy of motion compensated prediction in the prediction block size under the condition prohibiting bi-prediction is improved, and the coding efficiency is improved.
 (実施の形態2)
 続いて、本発明の実施の形態2に係る動画像符号化装置及び動画像復号装置の説明を行う。実施の形態2においては、実施の形態1と同様に、予測ブロックサイズによる動き補償予測の制限と、予測ブロックサイズ以下の双予測の制限の組合せで、最大メモリアクセス量を制限する構成は同一であるが、双予測の制限を定義するパラメータを制限する予測ブロックサイズを示す情報ではなく、最小CUサイズにおけるCU分割構造に対する双予測の制限を加える構造を取る。
(Embodiment 2)
Next, the video encoding device and video decoding device according to Embodiment 2 of the present invention will be described. In the second embodiment, as in the first embodiment, the configuration for limiting the maximum memory access amount is the same by combining the motion compensation prediction limitation based on the prediction block size and the bi-prediction limitation equal to or less than the prediction block size. There is a structure that adds a restriction of bi-prediction to a CU partition structure at the minimum CU size, not information indicating a prediction block size that restricts a parameter that defines a restriction of bi-prediction.
 図45に本発明の実施の形態2における、動き補償予測のブロックサイズ及び予測処理を制限する制御パラメータの一例を示し、説明する。 FIG. 45 shows an example of the motion compensated prediction block size and control parameters for limiting the prediction processing in Embodiment 2 of the present invention.
 制御パラメータは、最も小さな動き補償予測ブロックサイズである4×4画素の動き補償予測の有効・無効を制御するパラメータである、inter_4x4_enableと、動き補償予測の内、双予測が施される予測処理のみを禁止する最小CUサイズにおけるCU分割構造を定義する、inter_bipred_restriction_for_mincb_idcの2つのパラメータで構成される。 The control parameters are inter_4x4_enable, which is a parameter for controlling the validity / invalidity of motion compensated prediction of 4 × 4 pixels, which is the smallest motion compensated prediction block size, and only prediction processing for which bi-prediction is performed among motion compensated predictions It consists of two parameters of inter_bipred_restriction_for_mincb_idc that define the CU partition structure in the minimum CU size that prohibits.
 inter_bipred_restriction_for_mincb_idcは、4つの値を定義し、制限なし、N×N制限、N×2N/2N×N以下制限、CU内すべての分割(PU)で制限の4つの状態を制御する。最小CUサイズは、実施の形態1における、図35のシンタックスで示されるようにlog2_min_coding_block_size_minus3で8を基準とした(8×8を示す)2のべき乗値で定義され、inter_bipred_restriction_for_mincb_idcの値と最小CUサイズの連動により、双予測を制限するブロックサイズbipred_restriction_sizeが設定される。 Inter_bipred_restriction_for_mincb_idc defines four values, and controls four states: no limit, N × N limit, N × 2N / 2N × N limit or less, and all divisions (PUs) in the CU. The minimum CU size is defined as a power of 2 with log2_min_coding_block_size_minus3 as a reference (indicating 8 × 8) as shown in the syntax of FIG. 35 in Embodiment 1, and the value of inter_bipred_restriction_for_mincb_idc and the minimum CU size As a result, the block size bipred_restriction_size for limiting bi-prediction is set.
 実施の形態2における、符号化装置及び復号装置の構成は、実施の形態1と同様の構成をとることができ、実施の形態1におけるbipred_restriction_sizeが、上記log2_min_coding_block_size_minus3とinter_bipred_restriction_for_mincb_idcの組合せで定義される点が、異なる構成となっている。具体的なbipred_restriction_sizeの定義を図46に示す。 The configuration of the encoding device and the decoding device in the second embodiment can be the same as that in the first embodiment, and the point that bipred_restriction_size in the first embodiment is defined by a combination of the above log2_min_coding_block_size_minus3 and inter_bipred_restriction_for_mincb_idc. , Has a different configuration. A specific definition of bipred_restriction_size is shown in FIG.
 inter_bipred_restriction_for_mincb_idcは、図47に示すシンタックスの一例に示されるように、実施の形態1における図35のシンタックスと同様に構成され、シーケンス単位のパラメータとしてseq_parameter_set_rbsp()で伝送され、inter_bipred_restriction_idcに代わってinter_bipred_restriction_for_mincb_idcが伝送する値となる。 As shown in an example of the syntax shown in FIG. 47, inter_bipred_restriction_for_mincb_idc is configured in the same way as the syntax of FIG. 35 in Embodiment 1, and is transmitted as a sequence unit parameter by seq_parameter_set_rbsp (), and instead of inter_bipred_restriction_idc Is the value to be transmitted.
 メモリアクセス量が大きくなり、メモリ帯域の制限が必要になる状態は、符号化時の最小CUサイズに対して生じるため、最小CUサイズと連動して双予測の制限を加える構成は、管理・伝送するパラメータの無駄が少なく、なお且つ符号化装置においてメモリアクセス量の制限を加えたい場合に、より大きなサイズでの双予測制限を少ない制御パラメータ値で定義できる効果を有する。 Since the memory access amount becomes large and the state where the memory bandwidth needs to be limited occurs with respect to the minimum CU size at the time of encoding, the configuration for limiting the bi-prediction in conjunction with the minimum CU size is managed and transmitted. When there is little wasted parameter, and it is desired to limit the memory access amount in the encoding apparatus, the bi-prediction limitation at a larger size can be defined with a small control parameter value.
 さらに、実施の形態2においては、CU内の分割を更に細かくて左右や上下が非対称の予測ブロックを定義して、動き補償予測効率を向上させる場合においても、ブロックサイズ毎のサイズ制限を各CU階層で追加しなくても、最小CUサイズにおける定義のみを追加すれば良いため、拡張性も高いと共に、ハイビジョンを終える超高精細画像の符号化・復号処理を行う際に、予測ブロックサイズの大きさや双予測の制限を明示的に行うことが容易に実現できる効果を有する。 Furthermore, in the second embodiment, even when the prediction within the CU is further finely divided and the left and right and up and down asymmetric prediction blocks are defined to improve the motion compensation prediction efficiency, the size restriction for each block size is set for each CU. Even if it is not added in the hierarchy, it is sufficient to add only the definition in the minimum CU size. Therefore, the expandability is high, and the encoding block size of the high-definition image that finishes high-definition is large. It has an effect that can easily realize the restriction of sheath prediction.
 (実施の形態3)
 続いて、本発明の実施の形態3に係る動画像符号化装置及び動画像復号装置の説明を行う。実施の形態3においては、メモリアクセス量を制限する為の動き補償予測や双予測の制限に加えて、予測ブロックサイズが小さくなった際の結合動き予測候補生成処理の動作回数を制限することで、結合動き予測候補生成に要する処理負荷を軽減させる構成を取る。
(Embodiment 3)
Next, the video encoding device and video decoding device according to Embodiment 3 of the present invention will be described. In Embodiment 3, in addition to motion compensation prediction and bi-prediction limitations for limiting the memory access amount, by limiting the number of operations of combined motion prediction candidate generation processing when the prediction block size is reduced, The configuration is such that the processing load required for generating the combined motion prediction candidate is reduced.
 具体的には、所定のCUサイズ以下予測ブロックサイズにおける、各予測ブロックにおいて、同一の隣接ブロックの動き情報を用いて同一の結合動き情報候補生成処理を行う構成となる。実施の形態3では、最小CUサイズである8×8CUサイズの予測ブロックに対して、上記構成を取る構成を取り、実施の形態3の8×8CUサイズの結合動き情報候補生成における空間周辺予測ブロックの位置を、図48を用いて説明する。 Specifically, the same combined motion information candidate generation process is performed using the motion information of the same adjacent block in each prediction block in a prediction block size equal to or smaller than a predetermined CU size. In Embodiment 3, the configuration having the above configuration is adopted for the prediction block of 8 × 8 CU size that is the minimum CU size, and the spatial peripheral prediction block in the combined motion information candidate generation of 8 × 8 CU size of Embodiment 3 Will be described with reference to FIG.
 8×8CUにおいて、8×8画素の予測ブロック(2N×2N)に対する空間候補ブロック群のブロックA0、ブロックA1、ブロックB0、ブロックB1、ブロックB2の5ブロックの位置は、図48(a)に示すように、図19に示した、実施の形態1における空間候補ブロック群の定義と同一の位置を示す。 In the 8 × 8 CU, the positions of the five blocks of the block A0, the block A1, the block B0, the block B1, and the block B2 of the spatial candidate block group for the prediction block (2N × 2N) of 8 × 8 pixels are shown in FIG. As shown, the same position as the definition of the space candidate block group in the first embodiment shown in FIG. 19 is shown.
 これに対して、4×8画素の予測ブロック(N×2N)、8×4画素の予測ブロック(2N×N)、4×4画素の予測ブロック(N×N)に対する空間候補ブロック群の位置に関しては、図48(b)、(c)、(d)に示すように、図19に示した、実施の形態1における空間候補ブロック群の定義で示される、対象予測ブロックの隣接位置のブロックではなく、8×8画素に対する空間候補ブロック群と同じ位置が、すべての予測ブロックに対して用いられる。時間候補ブロック群の位置に関しても同様に、8×8画素の予測ブロックと同一の位置が、4×8画素、8×4画素、4×4画素すべての予測ブロックに対して用いられる。 On the other hand, the position of the spatial candidate block group with respect to the 4 × 8 pixel prediction block (N × 2N), the 8 × 4 pixel prediction block (2N × N), and the 4 × 4 pixel prediction block (N × N). 48 (b), (c), and (d), as shown in FIG. 19, the block at the adjacent position of the target prediction block shown in the definition of the spatial candidate block group in the first embodiment shown in FIG. Instead, the same position as the spatial candidate block group for 8 × 8 pixels is used for all prediction blocks. Similarly, regarding the position of the temporal candidate block group, the same position as the prediction block of 8 × 8 pixels is used for all prediction blocks of 4 × 8 pixels, 8 × 4 pixels, and 4 × 4 pixels.
 すなわち、対象となる8×8CUに対して、構成されるすべての予測ブロック構造において、同一の結合動き情報候補が用いられることになり、符号化装置及び復号装置における結合動き情報生成処理は1回の生成処理で実現できる。 That is, for the target 8 × 8 CU, the same combined motion information candidate is used in all configured prediction block structures, and the combined motion information generation process in the encoding device and the decoding device is performed once. It can be realized by the generation process.
 続いて、実施の形態3における、動画像符号化装置の符号化ブロック単位の符号化処理の説明を行う。実施の形態1における符号化ブロック単位の符号化処理に対して、図7のフローチャートで示される動き補償予測ブロックサイズ選択/予測信号生成処理と、図17のフローチャートで示される動き補償予測モード/予測信号生成処理のみが異なるため、これらの処理に関して説明する。 Subsequently, an encoding process for each encoding block of the moving image encoding apparatus according to Embodiment 3 will be described. In contrast to the encoding process in units of encoding blocks in Embodiment 1, the motion compensated prediction block size selection / prediction signal generation process shown in the flowchart of FIG. 7 and the motion compensated prediction mode / prediction shown in the flowchart of FIG. Since only the signal generation process is different, these processes will be described.
 図49に実施の形態3における動き補償予測ブロックサイズ選択/予測信号生成処理のフローチャートを示す。実施の形態1の図7のフローチャートと同一のステップに関しては、同一番号をつけ異なる部分にのみ新しいステップ番号をつけている。 FIG. 49 shows a flowchart of motion compensation prediction block size selection / prediction signal generation processing in the third embodiment. Regarding the same steps as those in the flowchart of FIG. 7 of the first embodiment, the same numbers are assigned and new step numbers are assigned only to different portions.
 最初に、対象CUに対して予測対象となる符号化ブロック画像を取得する(S700)。次に、対象CUのCUサイズが8×8であるか否かを判定する(S4908)。対象CUのCUサイズが8×8である場合(S4908:YES)には、結合動き情報候補リスト生成処理が行われる(S4909)。対象CUのCUサイズが8×8でない場合(S4908:NO)には、ステップS701に進む。ステップS4909の詳細に関しては、実施の形態1における図18の結合動き情報候補リスト生成処理と同一の処理が行われる。 First, an encoded block image to be predicted is acquired for the target CU (S700). Next, it is determined whether or not the CU size of the target CU is 8 × 8 (S4908). When the CU size of the target CU is 8 × 8 (S4908: YES), combined motion information candidate list generation processing is performed (S4909). If the CU size of the target CU is not 8 × 8 (S4908: NO), the process proceeds to step S701. Regarding the details of step S4909, the same process as the combined motion information candidate list generation process of FIG. 18 in the first embodiment is performed.
 ステップS4909を行った後、対象CU内の最小予測ブロックサイズがbipred_restriction_size以下である場合(S4910:YES)、結合動き情報候補単予測変換処理が行われる(S4911)。対象CU内の最小予測ブロックサイズがbipred_restriction_size以下でない場合には(S4910:NO)、ステップS701に進む。ステップS4911の詳細に関しては、実施の形態1における図30の結合動き情報候補単予測変換処理と同一の処理が行われる。 After performing step S4909, when the minimum prediction block size in the target CU is equal to or smaller than bipred_restriction_size (S4910: YES), combined motion information candidate single prediction conversion processing is performed (S4911). If the minimum predicted block size in the target CU is not less than or equal to bipred_restriction_size (S4910: NO), the process proceeds to step S701. Regarding the details of step S4911, the same process as the combined motion information candidate single prediction conversion process of FIG. 30 in the first embodiment is performed.
 実施の形態3においては、双予測を制限する予測ブロックサイズであるbipred_restriction_sizeにおける結合動き情報候補生成処理が、対象となるCUにおいて用いられる予測ブロックサイズである場合(inter_4x4_enableが1の場合には、4×4/4×8/8×4/8×8の予測ブロック、inter_4x4_enableが0の場合には、4×8/8×4/8×8の予測ブロック)には、対象CUに対して同一に生成する結合動き情報候補リストに対して双予測の動き情報を単予測に変換する処理を行う。つまり、bipred_restriction_sizeが3(8×8以下制限)に拡張された処理が行われることになる。 In Embodiment 3, when the combined motion information candidate generation process in bipred_restriction_size, which is a prediction block size that restricts bi-prediction, is the prediction block size used in the target CU (when inter_4x4_enable is 1, 4 * 4/4 * 8/8 * 4/8 * 8 prediction block, and when inter_4x4_enable is 0, 4 * 8/8 * 4/8 * 8 prediction block) is the same for the target CU The bi-prediction motion information is converted into a single prediction for the combined motion information candidate list generated at the same time. That is, processing in which bipred_restriction_size is expanded to 3 (8 × 8 or less restriction) is performed.
 図49の説明に戻り、ステップS4911の結合動き情報候補単予測変換処理を行った後、ステップS701に進む。ステップS701以降、ステップS707までの処理に関しては、実施の形態1における図7のフローチャートにおけるステップS701からステップS707までの処理と同一の処理が行われる。 49, after performing the combined motion information candidate single prediction conversion process of step S4911, the process proceeds to step S701. The processing from step S701 to step S707 is the same as the processing from step S701 to step S707 in the flowchart of FIG. 7 in the first embodiment.
 実施の形態3においては、8×8CUサイズに対する、結合動き情報候補リスト生成処理と結合動き情報候補単予測変換処理が同一の動作で行われ、符号化装置においては1回の生成処理によって、8×8CUサイズ内のすべての結合動き情報候補を生成することが可能となる効果を有する。また、実施の形態3において、図49のフローチャートのステップS4910の処理を行わない構成においては、8×8CUサイズに対して結合動き情報候補リスト生成処理が同一の動作で行われ、且つbipred_restriction_sizeを拡張しない状態での結合動き情報候補単予測変換処理を可能とする効果を有するが、符号化装置においては8×8CUサイズ内の予測ブロックサイズ毎の結合動き情報候補単予測変換処理が必要となる。 In the third embodiment, the combined motion information candidate list generation process and the combined motion information candidate uni-prediction conversion process for the 8 × 8 CU size are performed by the same operation, and the encoding device performs 8 generations by one generation process. There is an effect that it becomes possible to generate all combined motion information candidates within the × 8 CU size. In Embodiment 3, in the configuration in which the process of step S4910 in the flowchart of FIG. 49 is not performed, the combined motion information candidate list generation process is performed with the same operation for the 8 × 8 CU size, and bipred_restriction_size is expanded. The combined motion information candidate uni-prediction conversion process in a state where the prediction is not performed can be performed, but the encoding apparatus needs the combined motion information candidate uni-prediction conversion process for each prediction block size within the 8 × 8 CU size.
 次に、図50に実施の形態3における動き補償予測モード/予測信号生成処理のフローチャートを示し説明する。実施の形態1の図17のフローチャートと同一のステップに関しては、同一番号をつけ異なる部分にのみ新しいステップ番号をつけている。 Next, FIG. 50 shows a flowchart of the motion compensation prediction mode / prediction signal generation processing in the third embodiment. With respect to the same steps as those in the flowchart of FIG. 17 of the first embodiment, the same numbers are assigned and new step numbers are assigned only to different portions.
 定義されたCU内の予測ブロックサイズ分割モード(PU)に従って設定されたNumPartに基づき、対象CU内をPU分割した予測ブロックサイズ毎に(S1700)、対象CUサイズが8×8でない場合(S5010:NO)には、ステップS1701からステップS1708までのステップが実行される(S1709)。ステップS1701からステップS1708までの処理に関しては、実施の形態1における図17のフローチャートと同一の処理が行われる。 When the target CU size is not 8 × 8 for each predicted block size obtained by PU partitioning within the target CU based on the NumPart set according to the predicted block size partitioning mode (PU) within the defined CU (S5010: NO), the steps from Step S1701 to Step S1708 are executed (S1709). Regarding the processing from step S1701 to step S1708, the same processing as the flowchart of FIG. 17 in the first embodiment is performed.
 対象CUサイズが8×8である場合(S5010:YES)には、ステップS1701からステップS1703までの処理を行わずに、ステップS1704に進む。すなわち、対象CUサイズが8×8である、予測ブロックサイズの場合には、図49で示した動き補償予測ブロックサイズ選択/予測信号生成処理のフローチャート内の処理で生成された結合動き情報候補をそのまま使用して、結合予測モードの動き補償予測を行う構成となっている。 If the target CU size is 8 × 8 (S5010: YES), the process proceeds from step S1701 to step S1703 and proceeds to step S1704. That is, in the case of the prediction block size where the target CU size is 8 × 8, the combined motion information candidate generated by the process in the flowchart of the motion compensation prediction block size selection / prediction signal generation process shown in FIG. 49 is selected. It is configured to perform motion compensation prediction in the combined prediction mode by using it as it is.
 続いて、実施の形態3における、動画像復号装置の符号化ブロック単位の復号処理は実施の形態1と同一の処理を行い、結合動き予測における結合動き情報候補リスト生成のために用いる候補ブロックの位置のみを対象となるCUが8×8の場合に、図48で示したようにすべての予測ブロックで同一位置の候補ブロックを取得し、図39のフローチャートで示された結合動き情報復号処理におけるステップS3902の判断条件として、CUサイズが8×8の場合には、CU内で定義可能な最小予測ブロックサイズがbipred_restriction_size以下であるか否かの条件に置き換わる構成で実現できる。 Subsequently, the decoding process in units of coding blocks of the video decoding device in the third embodiment performs the same process as in the first embodiment, and the candidate blocks used for generating the combined motion information candidate list in the combined motion prediction. In the case where the CU targeted for only the position is 8 × 8, candidate blocks at the same position are obtained in all the prediction blocks as shown in FIG. 48, and in the combined motion information decoding process shown in the flowchart of FIG. As a determination condition in step S3902, when the CU size is 8 × 8, it can be realized by replacing the minimum prediction block size definable in the CU with bipred_restriction_size.
 なお、実施の形態3において、図39のフローチャートで示された結合動き情報復号処理におけるステップS3902の判断条件を変更しない構成で実現した場合には、8×8CUサイズに対して結合動き情報候補リスト生成処理を同一の動作で行い、且つbipred_restriction_sizeを拡張しない状態での結合動き情報候補単予測変換処理を可能とする効果を有する。復号装置においては、符号化ストリームを復号することで、復号対象ブロックに対する予測ブロックサイズは特定されるため、特定された予測ブロックサイズに対する単一の結合動き情報候補単予測変換処理が行われる。 In the third embodiment, when the configuration in which the determination condition in step S3902 in the combined motion information decoding process shown in the flowchart of FIG. 39 is not changed is realized, the combined motion information candidate list for the 8 × 8 CU size is provided. This has the effect of enabling the combined motion information candidate single prediction conversion processing in a state where the generation processing is performed with the same operation and bipred_restriction_size is not expanded. In the decoding apparatus, since the prediction block size for the decoding target block is specified by decoding the encoded stream, a single combined motion information candidate single prediction conversion process is performed for the specified prediction block size.
 また、実施の形態3における、動画像復号装置において、更に結合動き情報候補生成処理を少ない処理で実現できる構成として、実施の形態1における符号化ブロック単位の復号処理に対して、図39の結合動き情報復号処理を図51で示すフローチャートの処理に置き換える構成をとることが可能であり、その動作を説明する。図39のフローチャートと同一のステップに関しては、同一番号をつけ異なる部分にのみ新しいステップ番号をつけている。 In addition, in the moving picture decoding apparatus according to the third embodiment, the combined motion information candidate generation process can be realized with fewer processes. It is possible to replace the motion information decoding process with the process of the flowchart shown in FIG. 51, and the operation will be described. With respect to the same steps as those in the flowchart of FIG. 39, the same numbers are assigned and new step numbers are assigned only to different portions.
 予測モードに結合予測モードを設定した後(S3900)、対象となっている予測ブロックのCUサイズが8×8であるか否かを判断する(S5107)。CUサイズが8×8でない場合(S5107:NO)には、ステップS3901に進み、実施の形態1と同様の結合予測動き情報復号処理が行われる。 After the combined prediction mode is set as the prediction mode (S3900), it is determined whether or not the CU size of the target prediction block is 8 × 8 (S5107). If the CU size is not 8 × 8 (S5107: NO), the process proceeds to step S3901, and the joint prediction motion information decoding process similar to that of the first embodiment is performed.
 一方、CUサイズが8×8である場合には(S5107:YES)、対象となっている予測ブロックが対象CU内の最初の結合予測モードであるか否かを判断する(S5108)。最初の結合予測モードである場合には(S5108:YES)、結合動き情報候補リスト生成処理が行われる(S5109)。ステップS5109においては、図48で示したように、CU内の全予測ブロックで同一位置の候補ブロックを取得する構成で、ステップS3901と同じ処理が行われる。 On the other hand, if the CU size is 8 × 8 (S5107: YES), it is determined whether the target prediction block is the first combined prediction mode in the target CU (S5108). If it is the first combined prediction mode (S5108: YES), combined motion information candidate list generation processing is performed (S5109). In step S5109, as shown in FIG. 48, the same processing as that in step S3901 is performed with a configuration in which candidate blocks at the same position are acquired for all prediction blocks in the CU.
 最初の結合予測モードでない場合には(S5108:NO)、既に対象CUで同一に生成される結合動き情報候補リストが生成済であるため、結合動き情報候補リスト生成処理を行わず、ステップS3904に進む。対象CUにおいて一度の結合動き情報候補リスト生成で復号処理が可能となるため、8×8CU内に複数の結合予測モードが存在する場合の結合動き情報候補リスト生成処理が削減される。 If it is not the first combined prediction mode (S5108: NO), since the combined motion information candidate list generated in the same way in the target CU has already been generated, the combined motion information candidate list generation processing is not performed, and the process proceeds to step S3904. move on. Since decoding processing can be performed by generating a combined motion information candidate list once in the target CU, combined motion information candidate list generation processing when a plurality of combined prediction modes exist in the 8 × 8 CU is reduced.
 ステップS5109を行った後、CU内で定義可能な最小予測ブロックサイズがbipred_restriction_size以下であるか否かの判断を行い(S5110)、最小予測ブロックサイズがbipred_restriction_size以下である場合(S5110:YES)には、結合動き情報候補単予測変換処理を行い(S3903)、最小予測ブロックサイズがbipred_restriction_sizeより大きい場合(S5110:NO)には、ステップS3904に進む。 After performing step S5109, it is determined whether or not the minimum predicted block size definable in the CU is equal to or smaller than bipred_restriction_size (S5110). If the minimum predicted block size is equal to or smaller than bipred_restriction_size (S5110: YES). Then, the combined motion information candidate single prediction conversion process is performed (S3903), and if the minimum prediction block size is larger than bipred_restriction_size (S5110: NO), the process proceeds to step S3904.
 ステップS3904からステップS3906の処理に関しては、実施の形態1における図39のフローチャートの処理と同一の処理が行われ、結合予測モードの動き情報が復号され格納される。 Regarding the processing from step S3904 to step S3906, the same processing as the processing in the flowchart of FIG. 39 in the first embodiment is performed, and the motion information in the joint prediction mode is decoded and stored.
 実施の形態3における動画像符号化装置及び動画像復号装置によれば、メモリアクセス量を制限する為の動き補償予測や双予測の制限と予測ブロックサイズが小さくなった際の結合動き予測候補生成処理の処理削減を、それぞれの制限に対して整合がとれた構成で実現可能とし、メモリ帯域の制限及び結合動き情報候補生成処理削減を両立しつつ符号化効率を向上させることができる。 According to the moving picture coding apparatus and the moving picture decoding apparatus in the third embodiment, combined motion prediction candidate generation when the motion compensation prediction for limiting the memory access amount, the restriction of bi-prediction, and the prediction block size are reduced It is possible to realize processing reduction with a configuration that is consistent with each restriction, and to improve the coding efficiency while simultaneously reducing the memory bandwidth and reducing the combined motion information candidate generation process.
 実施の形態3における同一結合動き情報候補リストを構成する単位は、8×8サイズとして説明を行ったが、8×8サイズに限定される必要はなく、ピクチャ単位やシーケンス単位などの所定単位で、同一リストを生成する最大予測ブロックサイズを定義するパラメータ情報を伝送することで、その単位を変化されることが可能である。パラメータとしては、例えばlog2_parallel_merge_level_minus2として、同一リストを生成する予測ブロックサイズの水平・垂直サイズの基準となる、2のべき乗値に対応する値が定義できる。 The unit constituting the same combined motion information candidate list in Embodiment 3 has been described as an 8 × 8 size, but is not limited to the 8 × 8 size, and is a predetermined unit such as a picture unit or a sequence unit. The unit can be changed by transmitting parameter information defining the maximum predicted block size for generating the same list. As a parameter, for example, log2_parallel_merge_level_minus2 can be defined as a value corresponding to a power of 2 that serves as a reference for the horizontal / vertical size of the predicted block size for generating the same list.
 (実施の形態4)
 続いて、本発明の実施の形態4に係る動画像符号化装置及び動画像復号装置の説明を行う。実施の形態4においては、実施の形態3と同様に、メモリアクセス量を制限する為の動き補償予測や双予測の制限に加えて、予測ブロックサイズが小さくなった際の結合動き予測候補生成処理の動作回数を制限することで、結合動き予測候補生成に要する処理負荷を軽減させる構成を取る。
(Embodiment 4)
Next, the video encoding device and video decoding device according to Embodiment 4 of the present invention will be described. In the fourth embodiment, as in the third embodiment, in addition to motion compensation prediction and bi-prediction restriction for restricting the memory access amount, combined motion prediction candidate generation processing when the prediction block size is reduced By limiting the number of operations, the processing load required for generating the combined motion prediction candidate is reduced.
 実施の形態4における、動画像符号化装置においては、実施の形態1に示した動画像符号化装置に対して、図15に示す、動き補償予測ブロック構造選択部113において、結合動き情報単予測変換部1507が無くなり、結合動き情報算出部1506より出力された、動きベクトル、参照画像指定情報、結合動き情報候補リストが直接、結合動き補償予測生成部1508に供給される構成を取る。 In the moving picture coding apparatus according to the fourth embodiment, the combined motion information single prediction is performed in the motion compensated prediction block structure selection unit 113 shown in FIG. 15 in contrast to the moving picture coding apparatus shown in the first embodiment. The conversion unit 1507 is eliminated, and the motion vector, the reference image designation information, and the combined motion information candidate list output from the combined motion information calculation unit 1506 are directly supplied to the combined motion compensation prediction generation unit 1508.
 また、実施の形態4における、動画像復号装置においては、実施の形態1に示した動画像復号装置に対して、図36に示す、動き情報復号部1111において、結合動き情報単予測変換部3605が無くなり、結合動き情報算出部3604より出力された、動きベクトル、参照画像指定情報、結合動き情報候補リストが直接、結合動き補償予測復号部3606に供給される構成を取る。 In addition, in the video decoding device in the fourth embodiment, the combined motion information single prediction conversion unit 3605 in the motion information decoding unit 1111 shown in FIG. 36 is compared with the video decoding device shown in the first embodiment. The motion vector, the reference image designation information, and the combined motion information candidate list output from the combined motion information calculation unit 3604 are directly supplied to the combined motion compensation prediction decoding unit 3606.
 実施の形態4においては、結合動き情報単予測変換部で行われていた動き情報の双予測から単予測への変換処理の変わりに、動き補償予測時に予測ブロックサイズがbipred_restriction_size以下の場合に双予測の動き情報のL0予測もしくはL1予測の一方の動き情報のみを用いた単予測の動き補償を行うことで、メモリアクセス量に制限を加えた動き補償予測をおこなう構造を取る。 In the fourth embodiment, instead of the motion information bi-prediction to uni-prediction conversion processing performed by the combined motion information uni-prediction conversion unit, bi-prediction is performed when the prediction block size is equal to or less than bipred_restriction_size during motion compensation prediction. By performing motion compensation for single prediction using only one of the motion information of the L0 prediction or the L1 prediction of the motion information, motion compensation prediction with a limited memory access amount is performed.
 具体的には、実施の形態4における符号化処理では、実施の形態1における、図17のフローチャートで示される動き補償予測モード/予測信号生成処理において、ステップS1702及びステップS1703の処理が無くなり、図31のフローチャートで示される結合予測モード評価値生成処理におけるステップS3105及びステップS3106で行われる、動き補償(単/双)予測ブロック生成処理の内部、及び図33のフローチャートで示される予測モード評価値生成処理のステップS3312で行われる動き補償予測ブロック生成処理の内部で、単予測への制限処理が行われる。 Specifically, in the encoding process in the fourth embodiment, the processes in step S1702 and step S1703 are eliminated in the motion compensation prediction mode / predicted signal generation process shown in the flowchart of FIG. 17 in the first embodiment. The prediction mode evaluation value generation shown in the flowchart of FIG. 33 and the inside of the motion compensation (single / bi) prediction block generation processing performed in step S3105 and step S3106 in the combined prediction mode evaluation value generation processing shown in the flowchart of 31. Within the motion compensated prediction block generation process performed in step S3312 of the process, a process for limiting to single prediction is performed.
 実施の形態4において、図31のフローチャートのステップS3105、ステップS3106及び、図33のフローチャートのステップS3312で施される動き補償予測ブロック生成の動作を、図52のフローチャートに示し、説明する。図52のフローチャートは実施の形態4において、図1に示す動画像符号化装置における動き補償予測部112の詳細動作となり、以下の動作を行う。 In the fourth embodiment, the motion compensated prediction block generation operation performed in steps S3105 and S3106 of the flowchart of FIG. 31 and step S3312 of the flowchart of FIG. 33 is shown in the flowchart of FIG. The flowchart of FIG. 52 is the detailed operation of the motion compensation prediction unit 112 in the moving picture encoding apparatus shown in FIG. 1 in the fourth embodiment, and performs the following operation.
 供給された動き情報の予測種別が単予測である場合(S5200:YES)、1つの参照画像に対する参照画像指定情報と動きベクトルを用いて動き補償単予測ブロックを生成する(S5203)。 When the prediction type of the supplied motion information is single prediction (S5200: YES), a motion compensated single prediction block is generated using reference image designation information and a motion vector for one reference image (S5203).
 供給された動き情報が単予測でない場合、つまり動き情報が双予測である場合(S5200:NO)、L0予測の動き情報とL1予測の動き情報(参照画像情報及び動きベクトル)が同一であるかどうか判定し、L0予測の動き情報とL1予測の動き情報が同一である場合(S5201:YES)、L0予測の動き情報のみを用いてL0単予測動き補償予測を行う(S5204)。ただし、双予測の動き情報は維持しL1予測の動き情報は変更しない。 If the supplied motion information is not uni-prediction, that is, if the motion information is bi-prediction (S5200: NO), whether the motion information for L0 prediction and the motion information for L1 prediction (reference image information and motion vector) are the same. If the motion information of the L0 prediction and the motion information of the L1 prediction are the same (S5201: YES), the L0 single prediction motion compensation prediction is performed using only the motion information of the L0 prediction (S5204). However, the motion information of bi-prediction is maintained and the motion information of L1 prediction is not changed.
 供給されたL0予測の動き情報とL1予測の動き情報が同一でない場合(S5201:NO)、予測ブロックサイズがbipred_restriction_size以下であるかどうかを判定し、予測ブロックサイズがbipred_restriction_size以下である場合(S5202:YES)、L0予測の動き情報とL1予測の動き情報が同一である場合(S5201:YES)と同様に、L0予測の動き情報のみを用いてL0単予測動き補償予測を行う(S5204)。ただし、双予測の動き情報は維持しL1予測の動き情報は変更しない。双予測制限は双予測を単予測に制限することで動き補償予測のメモリ帯域を抑制することが目的であるため、双予測制限によって制限される予測リスト(L0/L1)は、L1単予測にしても良い。 When the supplied motion information of the L0 prediction and the motion information of the L1 prediction are not the same (S5201: NO), it is determined whether or not the prediction block size is equal to or smaller than bipred_restriction_size, and when the predicted block size is equal to or smaller than bipred_restriction_size (S5202: If the motion information for L0 prediction and the motion information for L1 prediction are the same (S5201: YES), L0 single prediction motion compensation prediction is performed using only the motion information for L0 prediction (S5204). However, the motion information of bi-prediction is maintained and the motion information of L1 prediction is not changed. The purpose of the bi-prediction restriction is to limit the memory band of motion compensation prediction by restricting the bi-prediction to the single prediction. Therefore, the prediction list (L0 / L1) restricted by the bi-prediction restriction is set to the L1 single prediction. May be.
 供給された予測ブロックサイズがbipred_restriction_sizeより大きい場合(S5202:NO)、2つの参照画像に対する参照画像指定情報と動きベクトルを用いて動き補償双予測ブロックを生成する(S5205)。 When the supplied prediction block size is larger than bipred_restriction_size (S5202: NO), a motion-compensated bi-prediction block is generated using reference image designation information and motion vectors for two reference images (S5205).
 また、実施の形態4における復号処理では、実施の形態1における、図39のフローチャートで示される結合予測動き情報復号処理において、ステップS3902及びステップS3903の処理が無くなり、図37のフローチャートで示される予測ブロック単位復号処理におけるステップS3704で行われる、動き補償予測信号算出処理の内部で、単予測への制限処理が、符号化処理と同様に、図52のフローチャートで示した処理で行われる。 Further, in the decoding process in the fourth embodiment, in the joint prediction motion information decoding process shown in the flowchart of FIG. 39 in the first embodiment, the processes in steps S3902 and S3903 are eliminated, and the prediction shown in the flowchart of FIG. Within the motion compensated prediction signal calculation process performed in step S3704 in the block unit decoding process, the process for limiting to single prediction is performed by the process shown in the flowchart of FIG. 52, similarly to the encoding process.
 実施の形態4においては、双予測の制限処理を結合動き情報候補リストを単予測に変換する構成を用いずに、動き補償予測時に双予測の動き情報のうち、L0予測もしくはL1予測の一方の動き情報のみを用いて、単予測の動き補償予測を行う構成を用いることで、メモリアクセス量の制限を実現する。 In the fourth embodiment, one of the L0 prediction and the L1 prediction is included in the motion information of the bi-prediction at the time of motion compensation prediction without using the configuration for converting the combined motion information candidate list into the single prediction for the restriction process of the bi-prediction. By using a configuration that performs motion compensation prediction of single prediction using only motion information, the memory access amount is limited.
 また、双予測の制限処理において、予測信号としては単予測と同一でありながら、動き情報は双予測となる結合動き情報候補を維持することができる。これにより、bipred_restriction_size以下の予測ブロックであっても、L0予測、L1予測ともに動き情報が保存されるため、以降に符号化・復号される予測ブロックの隣接参照動き情報として双予測の情報がそのまま利用でき、以降に符号化・復号される予測ブロックの動き予測処理の予測効率を向上させることができる。 Also, in the bi-prediction restriction process, the prediction information is the same as the single prediction, but the motion information can maintain a combined motion information candidate that is bi-predicted. As a result, even in a prediction block of bipred_restriction_size or less, motion information is stored for both L0 prediction and L1 prediction, so that information of bi-prediction is used as it is as adjacent reference motion information of a prediction block to be encoded and decoded thereafter. In addition, it is possible to improve the prediction efficiency of the motion prediction process of the prediction block that is encoded / decoded thereafter.
 また、結合動き情報候補リストとしては同じ動き情報を用いる、異なる大きさの予測ブロックサイズにおいて、動き補償予測時の双予測制限で、メモリアクセス量の制限を可能とするため、予測ブロックサイズが小さくなった際に、同一の結合動き予測候補リスト生成を行う場合に、同一リストを構成する基準の予測ブロックサイズとbipred_restriction_sizeが異なる場合でも、実施の形態4の構成をとることにより、結合動き情報候補リスト生成時に同一リスト構成と双予測制限の両方を加味した条件判断を加える必要がなく、動き補償予測時の双予測制限だけで機能を実現できる効果を有すると共に、メモリアクセスの制限を行うために制御するbipred_restriction_sizeより大きな予測ブロックサイズにおいて、結合動き情報に双予測の制限を加える必要がなくなるため、符号化効率を向上させる効果を有する。 In addition, the same motion information is used as the combined motion information candidate list, and the prediction block size is small because the memory access amount can be limited by the bi-prediction restriction at the time of motion compensation prediction in the prediction block sizes of different sizes. When the same combined motion prediction candidate list is generated, even if the reference prediction block size and bipred_restriction_size constituting the same list are different, the combined motion information candidate is obtained by adopting the configuration of the fourth embodiment. In order to limit the memory access as well as having the effect of realizing the function only by the bi-prediction restriction at the time of motion compensation prediction, without having to add a condition judgment that takes into account both the same list configuration and the bi-prediction restriction at the time of list generation Add bi-prediction restriction to joint motion information at a prediction block size larger than the controlled bipred_restriction_size Since it becomes unnecessary, it has the effect of improving the encoding efficiency.
 また、動き補償予測時に双予測制限する構成では、動き情報を符号化するための2つの予測モード(結合予測モード、動き検出予測モード)の双方の双予測制限を一括して対応可能であるため、最小の構成で双予測制限を実現できる。 In addition, in the configuration in which bi-prediction restriction is performed at the time of motion compensation prediction, both the bi-prediction restrictions of two prediction modes (joint prediction mode and motion detection prediction mode) for encoding motion information can be handled in a lump. Bi-prediction restriction can be realized with a minimum configuration.
 (実施の形態5)
 続いて、本発明の実施の形態5に係る動画像符号化装置及び動画像復号装置の説明を行う。実施の形態5においては、実施の形態1と同様に、メモリアクセス量の制限を行うための、予測ブロックサイズによる動き補償予測制限と、双予測動き補償の制限を行うが、双予測の制限を行うための、結合動き情報候補リストの動き情報に対する双予測から単予測への変換手法が異なる構成を取る。
(Embodiment 5)
Next, the video encoding device and video decoding device according to Embodiment 5 of the present invention will be described. In the fifth embodiment, as in the first embodiment, the motion compensation prediction restriction based on the prediction block size and the bi-predictive motion compensation restriction for restricting the memory access amount are performed. For this purpose, the bi-prediction to uni-prediction conversion method for motion information in the combined motion information candidate list is different.
 実施の形態5においては、実施の形態1と同様の構成及び処理を行うが、実施の形態1における、図18のフローチャートで示される、結合動き情報候補リスト生成処理、及び図30のフローチャートで示される、結合動き情報候補単予測変換処理が異なる構成を取る。 In the fifth embodiment, the same configuration and processing as in the first embodiment are performed, but the combined motion information candidate list generation processing shown in the flowchart of FIG. 18 and the flowchart of FIG. 30 in the first embodiment. The combined motion information candidate single prediction conversion process is different.
 図51のフローチャートで、実施の形態5における結合動き情報候補リスト生成処理の説明を行う。実施の形態5においては、符号化処理に対する図17のフローチャートにおけるステップS1701及び、復号処理に対する図39のフローチャートにおけるステップS3901において、図53に示される処理が施される。実施の形態1の図18のフローチャートと同一のステップに関しては、同一番号をつけ異なる部分にのみ新しいステップ番号をつけている。 51, the combined motion information candidate list generation process in the fifth embodiment will be described. In the fifth embodiment, the process shown in FIG. 53 is performed in step S1701 in the flowchart of FIG. 17 for the encoding process and in step S3901 in the flowchart of FIG. 39 for the decoding process. With respect to the same steps as those in the flowchart of FIG. 18 of the first embodiment, the same numbers are assigned and new step numbers are assigned only to different portions.
 ステップS1800からステップS1802までの処理によって、結合動き情報の候補となる、空間候補ブロック群から同一の情報を削除した形での空間結合動き情報候補と、時間結合動き情報候補が算出され、候補ブロックの動き情報から算出された結合動き情報が生成される。次に、ステップS1802までに生成された結合動き情報の数である、num_list_before_combined_mergeを格納する(S5305)。この値は、後述する結合動き情報候補単予測変換処理において用いられる。 Through the processing from step S1800 to step S1802, spatially combined motion information candidates and temporally combined motion information candidates obtained by deleting the same information from the spatial candidate block group, which are candidates for combined motion information, are calculated, and candidate blocks The combined motion information calculated from the motion information is generated. Next, num_list_before_combined_merge, which is the number of combined motion information generated up to step S1802, is stored (S5305). This value is used in the combined motion information candidate single prediction conversion process described later.
 続いて、ステップS1803からステップS1804までの処理によって、結合動き情報候補リストに登録された複数の結合動き情報候補の動き情報を組合せて生成した、第1結合動き情報候補と、結合動き情報候補リストに登録された動き情報に依存せずに生成した、第2結合動き情報候補が必要に応じて追加され、結合動き情報候補リスト生成処理を終了する。 Subsequently, the first combined motion information candidate and the combined motion information candidate list generated by combining the motion information of a plurality of combined motion information candidates registered in the combined motion information candidate list by the processing from step S1803 to step S1804. The second combined motion information candidate generated without depending on the motion information registered in the above is added as necessary, and the combined motion information candidate list generation process is terminated.
 実施の形態5における、結合動き情報候補リスト生成処理においての実施の形態1と異なる処理は、num_list_before_combined_mergeの格納処理であり、隣接ブロックより定義される候補ブロック群の動き情報を登録した結合動き情報と、候補ブロック群の動き情報の組合せや、候補ブロックの動き情報に依存しない動き情報を登録した結合動き情報の、境界のリスト番号を保存している。 The processing different from the first embodiment in the combined motion information candidate list generation processing in the fifth embodiment is a storage process of num_list_before_combined_merge, and the combined motion information in which the motion information of a candidate block group defined by adjacent blocks is registered. The boundary list number of the combined motion information in which the motion information combination of the candidate block group and the motion information not depending on the motion information of the candidate block is registered is stored.
 続いて、図54のフローチャートで、実施の形態5における結合動き情報候補単予測変換処理の説明を行う。実施の形態5においては、符号化処理に対する図17のフローチャートにおけるステップS1703及び、復号処理に対する図39のフローチャートにおけるステップS3903において、図54に示される処理が施される。実施の形態1の図30のフローチャートと同一のステップに関しては、同一番号をつけ異なる部分にのみ新しいステップ番号をつけている。 Subsequently, the combined motion information candidate single prediction conversion process in Embodiment 5 will be described with reference to the flowchart of FIG. In the fifth embodiment, the processing shown in FIG. 54 is performed in step S1703 in the flowchart of FIG. 17 for the encoding process and in step S3903 in the flowchart of FIG. 39 for the decoding process. With respect to the same steps as those in the flowchart of FIG. 30 of the first embodiment, the same numbers are assigned and new step numbers are assigned only to different portions.
 図52のフローチャートで示す結合動き情報候補単予測変換処理は、図30のフローチャートに対して、動き情報が単予測でない場合(S3002:NO)の処理が異なり、結合動き情報候補リストのインデックスiがnum_list_before_combined_mergeより小さい場合(S5407:YES)には、双予測の動き情報を単予測に変換するために、インデックスiに格納された動き情報のL1情報を無効にする(S3003)。 The combined motion information candidate single prediction conversion process shown in the flowchart of FIG. 52 differs from the flowchart of FIG. 30 in the case where the motion information is not a single prediction (S3002: NO), and the index i of the combined motion information candidate list is When smaller than num_list_before_combined_merge (S5407: YES), the L1 information of the motion information stored in the index i is invalidated in order to convert the motion information of bi-prediction into single prediction (S3003).
 一方、インデックスiがnum_list_before_combined_merge以上である場合(S5407:NO)には、双予測の動き情報を単予測に変換するために、インデックスiに格納された動き情報のL0情報を無効にする(S5408)。 On the other hand, when the index i is greater than or equal to num_list_before_combined_merge (S5407: NO), the L0 information of the motion information stored in the index i is invalidated in order to convert the motion information of bi-prediction into single prediction (S5408). .
 ステップS3003及びステップS5408で、単予測に変換されたインデックスiの動き情報が結合動き情報候補リストに格納され(S3004)、次のインデックスに進む(S3005)。 In step S3003 and step S5408, the motion information of index i converted to single prediction is stored in the combined motion information candidate list (S3004), and the process proceeds to the next index (S3005).
 実施の形態5における、結合動き情報候補単予測変換においては、結合動き情報候補リスト内の候補動き情報を、隣接する候補ブロックの動き情報から算出された動き情報と、登録された複数の動き情報の組合せや、候補ブロックの動き情報に依存せずに生成した動き情報に対して、単予測変換時に無効にする動き情報を予測種別(L0予測/L1予測)で切り替える。これにより、特に第1結合動き情報候補リスト追加部で追加された動き情報に対して、単予測変換時に無効にされた予測種別の動き情報を残して、単予測変換時に有効にされた予測種別の動き情報を無効にすることができ、結合動き情報として有効な動き情報を多く残すことが可能となり、符号化効率を向上させることができる。 In combined motion information candidate single prediction conversion in Embodiment 5, candidate motion information in the combined motion information candidate list includes motion information calculated from motion information of adjacent candidate blocks and a plurality of registered motion information. For the motion information generated without depending on the motion information of the candidate block or the motion information of the candidate block, the motion information to be invalidated at the time of single prediction conversion is switched according to the prediction type (L0 prediction / L1 prediction). Thereby, especially with respect to the motion information added by the first combined motion information candidate list adding unit, the motion information of the prediction type disabled during the single prediction conversion is left, and the prediction type enabled during the single prediction conversion Motion information can be invalidated, and it is possible to leave a lot of valid motion information as combined motion information, thereby improving the coding efficiency.
 また、双予測が制限された予測ブロックサイズにおいて、L0予測とL1予測が偏らずに候補として用いられるため、符号化・復号時に用いられた動き情報として保存される動き情報においてもL0予測とL1予測の偏りが少なくなる。そのため、後続する予測ブロックの結合動き情報候補生成時に、第1結合動き情報候補リスト追加部で生成できる双予測の動き情報の精度が向上し、符号化効率を向上させることができる。 In addition, in the prediction block size in which bi-prediction is limited, L0 prediction and L1 prediction are used as candidates without being biased. Therefore, even in motion information stored as motion information used at the time of encoding / decoding, L0 prediction and L1 Forecast bias is reduced. Therefore, the accuracy of the bi-prediction motion information that can be generated by the first combined motion information candidate list adding unit at the time of generating the combined motion information candidate of the subsequent prediction block can be improved, and the encoding efficiency can be improved.
 実施の形態5においては、インデックスiがnum_list_before_combined_mergeより小さい時にL1情報を無効にし、インデックスiがnum_list_before_combined_merge以上の場合にL0情報を無効にしたが、num_list_before_combined_mergeを基準に無効にする予測種別を切り替えることが、この実施の形態における特徴であり、インデックスiがnum_list_before_combined_mergeより小さい時にL0情報を無効にし、インデックスiがnum_list_before_combined_merge以上の場合にL1情報を無効にする構成を取ることも可能である。 In the fifth embodiment, the L1 information is invalidated when the index i is smaller than num_list_before_combined_merge, and the L0 information is invalidated when the index i is equal to or larger than num_list_before_combined_merge. A feature of this embodiment is that the L0 information is invalidated when the index i is smaller than num_list_before_combined_merge, and the L1 information is invalidated when the index i is greater than or equal to num_list_before_combined_merge.
 (実施の形態6)
 続いて、本発明の実施の形態6に係る動画像符号化装置及び動画像復号装置の説明を行う。実施の形態6においては、実施の形態5と同様の構成をとり、結合動き情報候補単予測変換における無効にする予測種別(L0予測/L1予測)を切り替える点が特徴であるが、インデックスの固定位置を基準に切り替える構成を取る。
(Embodiment 6)
Next, the video encoding device and video decoding device according to Embodiment 6 of the present invention will be described. The feature of the sixth embodiment is that it has the same configuration as that of the fifth embodiment, and switches the prediction type (L0 prediction / L1 prediction) to be invalidated in the combined motion information candidate single prediction conversion. It takes a configuration that switches based on the position.
 実施の形態6においては、実施の形態5と同様の構成及び処理を行うが、実施の形態5における、図53のフローチャートで示される、結合動き情報候補リスト生成処理は行わず、実施の形態1と同一の、図18のフローチャートで示される、結合動き情報候補リスト生成処理が行われる。 In the sixth embodiment, the same configuration and processing as in the fifth embodiment are performed, but the combined motion information candidate list generation processing shown in the flowchart of FIG. 53 in the fifth embodiment is not performed, and the first embodiment is performed. The combined motion information candidate list generation process shown in the flowchart of FIG. 18 is performed.
 また、実施の形態6においては、実施の形態5における、図54のフローチャートで示される、結合動き情報候補単予測変換処理が、図55のフローチャートで示される処理に置き換わる構成を取る。実施の形態6においては、符号化処理に対する図17のフローチャートにおけるステップS1703及び、復号処理に対する図39のフローチャートにおけるステップS3903において、図55に示される処理が施される。 In the sixth embodiment, the combined motion information candidate single prediction conversion process shown in the flowchart of FIG. 54 in the fifth embodiment is replaced with the process shown in the flowchart of FIG. In the sixth embodiment, the process shown in FIG. 55 is performed in step S1703 in the flowchart of FIG. 17 for the encoding process and in step S3903 in the flowchart of FIG. 39 for the decoding process.
 以下、図55のフローチャートの説明を行う。図54のフローチャートと同一のステップに関しては、同一番号をつけ異なる部分にのみ新しいステップ番号をつけている。 Hereinafter, the flowchart of FIG. 55 will be described. Regarding the same steps as those in the flowchart of FIG. 54, the same numbers are assigned and new step numbers are assigned only to different portions.
 図55のフローチャートで示す結合動き情報候補単予測変換処理は、図54のフローチャートに対して、動き情報が単予測でない場合(S3002:NO)の処理が異なり、結合動き情報候補リストのインデックスiが2より小さい場合(S5507:YES)には、双予測の動き情報を単予測に変換するために、インデックスiに格納された動き情報のL1情報を無効にする(S3003)。 The combined motion information candidate single prediction conversion process shown in the flowchart of FIG. 55 is different from the flowchart of FIG. 54 in the case where the motion information is not a single prediction (S3002: NO), and the index i of the combined motion information candidate list is If it is smaller than 2 (S5507: YES), the L1 information of the motion information stored in the index i is invalidated in order to convert the motion information of bi-prediction into single prediction (S3003).
 一方、インデックスiが2以上である場合(S5507:NO)には、双予測の動き情報を単予測に変換するために、インデックスiに格納された動き情報のL0情報を無効にする(S5408)。 On the other hand, when the index i is 2 or more (S5507: NO), the L0 information of the motion information stored in the index i is invalidated in order to convert the motion information of bi-prediction into single prediction (S5408). .
ステップS3003及びステップS5408で、単予測に変換されたインデックスiの動き情報が結合動き情報候補リストに格納され(S3004)、次のインデックスに進む(S3005)。 In step S3003 and step S5408, the motion information of index i converted to single prediction is stored in the combined motion information candidate list (S3004), and the process proceeds to the next index (S3005).
 実施の形態6における、結合動き情報候補単予測変換においては、結合動き情報候補リスト内の候補動き情報を、隣接する候補ブロックの動き情報から算出された動き情報によって第1結合動き情報候補リスト追加部で、双予測の追加動き情報を生成するために必要な最小限の動き情報である2つの動き情報と、リストの後半で登録される第1結合動き情報候補リスト追加部及び第2結合動き情報候補リスト追加部で追加された動き情報に対して、インデックスの位置で固定的に単予測変換時に無効にする動き情報を予測種別(L0予測/L1予測)を切り替える構成を取る。 In combined motion information candidate single prediction conversion in Embodiment 6, the candidate motion information in the combined motion information candidate list is added to the first combined motion information candidate list by motion information calculated from the motion information of adjacent candidate blocks. Two motion information, which are the minimum motion information necessary for generating additional motion information for bi-prediction, and a first combined motion information candidate list adding unit and a second combined motion registered in the latter half of the list For the motion information added by the information candidate list adding unit, a configuration is adopted in which the prediction type (L0 prediction / L1 prediction) is switched for motion information that is fixedly invalidated at the time of single prediction conversion at the index position.
 実施の形態6における、結合動き情報候補単予測変換においては、実施の形態5に対して、隣接ブロックより定義される候補ブロック群の動き情報を登録した結合動き情報と、候補ブロック群の動き情報の組合せや、候補ブロックの動き情報に依存しない動き情報を登録した結合動き情報の、境界のリスト番号を保存する処理を無くすことができるため、処理負荷が軽減できると共に、実施の形態5と同様に、第1結合動き情報候補リスト追加部で追加された動き情報に対して、単予測変換時に無効にされた予測種別の動き情報を残して、単予測変換時に有効にされた予測種別の動き情報を無効にすることができ、結合動き情報として有効な動き情報を多く残すことが可能となり、符号化効率を向上させることができる。 In the combined motion information candidate uni-predictive conversion in the sixth embodiment, combined motion information in which the motion information of the candidate block group defined by the adjacent block is registered and the motion information of the candidate block group are compared to the fifth embodiment. Since the processing for saving the boundary list number of the combined motion information in which the motion information that does not depend on the motion information of the candidate block and the candidate block is registered can be eliminated, the processing load can be reduced and the same as in the fifth embodiment In addition, for the motion information added by the first combined motion information candidate list adding unit, the motion information of the prediction type enabled at the time of the single prediction conversion is left, leaving the motion information of the prediction type disabled at the time of the single prediction conversion. It is possible to invalidate the information, and it is possible to leave a lot of effective motion information as the combined motion information, thereby improving the encoding efficiency.
 また、実施の形態6では、第1結合動き情報候補、第2結合動き情報候補だけでなく、空間予測候補や時間予測候補に対しても無効にする予測種別を切り替えることができるため、予測種別が双予測で同一の動き情報が登録されている場合に、結合動き情報としてL0単予測、L1単予測の動き情報をそれぞれ利用可能になるため、符号化効率を向上させることができる。 Further, in Embodiment 6, since the prediction type that is invalidated not only for the first combined motion information candidate and the second combined motion information candidate but also for the spatial prediction candidate and the temporal prediction candidate can be switched, the prediction type However, when the same motion information is registered in bi-prediction, since the motion information of L0 uni-prediction and L1 uni-prediction can be used as combined motion information, the encoding efficiency can be improved.
 実施の形態6における、結合動き情報候補単予測変換において、無効にする予測種別(L0予測/L1予測)を切り替えるインデックスの位置を2に固定しているが、固定のインデックスで予測種別を切り替えることが実施の形態6における特徴であり、空間結合動き情報候補、時間結合動き情報候補、第1結合動き情報候補、第2結合動き情報候補として登録可能な動き情報の数と、最大で登録可能な結合動き情報候補の数に応じて、固定にする切り替え位置のインデックスの値を設定することも可能である。 In the combined motion information candidate single prediction conversion in Embodiment 6, the position of the index for switching the prediction type (L0 prediction / L1 prediction) to be invalidated is fixed to 2, but the prediction type is switched with a fixed index. Are the features in Embodiment 6, and the number of pieces of motion information that can be registered as a spatially combined motion information candidate, a temporally combined motion information candidate, a first combined motion information candidate, and a second combined motion information candidate, and can be registered at the maximum It is also possible to set the index value of the switching position to be fixed according to the number of combined motion information candidates.
 以上述べた実施の形態の動画像符号化装置が出力する動画像の符号化ストリームは、実施の形態で用いられた符号化方法に応じて復号することができるように特定のデータフォーマットを有しており、動画像符号化装置に対応する動画像復号装置がこの特定のデータフォーマットの符号化ストリームを復号することができる。 The moving image encoded stream output from the moving image encoding apparatus of the embodiment described above has a specific data format so that it can be decoded according to the encoding method used in the embodiment. Therefore, the moving picture decoding apparatus corresponding to the moving picture encoding apparatus can decode the encoded stream of this specific data format.
 動画像符号化装置と動画像復号装置の間で符号化ストリームをやりとりするために、有線または無線のネットワークが用いられる場合、符号化ストリームを通信路の伝送形態に適したデータ形式に変換して伝送してもよい。その場合、動画像符号化装置が出力する符号化ストリームを通信路の伝送形態に適したデータ形式の符号化データに変換してネットワークに送信する動画像送信装置と、ネットワークから符号化データを受信して符号化ストリームに復元して動画像復号装置に供給する動画像受信装置とが設けられる。 When a wired or wireless network is used to exchange an encoded stream between a moving image encoding device and a moving image decoding device, the encoded stream is converted into a data format suitable for the transmission form of the communication path. It may be transmitted. In that case, a video transmission apparatus that converts the encoded stream output from the video encoding apparatus into encoded data in a data format suitable for the transmission form of the communication channel and transmits the encoded data to the network, and receives the encoded data from the network Then, a moving image receiving apparatus that restores the encoded stream and supplies the encoded stream to the moving image decoding apparatus is provided.
 動画像送信装置は、動画像符号化装置が出力する符号化ストリームをバッファするメモリと、符号化ストリームをパケット化するパケット処理部と、パケット化された符号化データをネットワークを介して送信する送信部とを含む。動画像受信装置は、パケット化された符号化データをネットワークを介して受信する受信部と、受信された符号化データをバッファするメモリと、符号化データをパケット処理して符号化ストリームを生成し、動画像復号装置に提供するパケット処理部とを含む。 The moving image transmitting apparatus is a memory that buffers the encoded stream output from the moving image encoding apparatus, a packet processing unit that packetizes the encoded stream, and transmission that transmits the packetized encoded data via the network. Part. The moving image receiving apparatus generates a coded stream by packetizing the received data, a receiving unit that receives the packetized coded data via a network, a memory that buffers the received coded data, and packet processing. And a packet processing unit provided to the video decoding device.
 また、以上の符号化及び復号に関する処理は、ハードウェアを用いた伝送、蓄積、受信装置として実現することができるのは勿論のこと、ROM(Read Only Memory)やフラッシュメモリ等に記憶されているファームウェアや、コンピュータ等のソフトウェアによっても実現することができる。そのファームウェアプログラム、ソフトウェアプログラムをコンピュータ等で読み取り可能な記録媒体に記録して提供することも、有線あるいは無線のネットワークを通してサーバから提供することも、地上波あるいは衛星ディジタル放送のデータ放送として提供することも可能である。 In addition, the above-described processing related to encoding and decoding can be realized as a transmission, storage, and reception device using hardware, and is stored in a ROM (Read Only Memory), a flash memory, or the like. It can also be realized by firmware or software such as a computer. The firmware program and software program can be recorded on a computer-readable recording medium, provided from a server through a wired or wireless network, or provided as a data broadcast of terrestrial or satellite digital broadcasting Is also possible.
 以上、本発明を実施の形態をもとに説明した。実施の形態は例示であり、それらの各構成要素や各処理プロセスの組み合わせにいろいろな変形例が可能なこと、またそうした変形例も本発明の範囲にあることは当業者に理解されるところである。 The present invention has been described based on the embodiments. The embodiments are exemplifications, and it will be understood by those skilled in the art that various modifications can be made to combinations of the respective constituent elements and processing processes, and such modifications are within the scope of the present invention. .
 100 入力端子、 101 入力画像メモリ、 102 符号化ブロック取得部、 103 減算部、 104 直交変換・量子化部、 105 予測誤差符号化部、 106 逆量子化・逆変換部、 107 加算部、 108 フレーム内復号画像バッファ、 109 ループフィルタ部、 110 復号画像メモリ、 111 動きベクトル検出部、 112 動き補償予測部、 113 動き補償予測ブロック構造選択部、 114 イントラ予測部、 115 イントラ予測ブロック構造選択部、 116 予測モード選択部、 117 符号化ブロック構造選択部、 118 ブロック構造/予測モード情報付加情報符号化部、 119 予測モード情報メモリ、 120 多重化部、 121 出力端子、 122 符号化ブロック制御パラメータ生成部、 1100 入力端子、 1101 多重分離部、 1102 予測差分情報復号部、 1103 逆量子化・逆変換部、 1104 加算部、 1105 フレーム内復号画像バッファ、 1106 ループフィルタ部、 1107 復号画像メモリ、 1108 予測モード/ブロック構造復号部、 1109 予測モード/ブロック構造選択部、 1110 イントラ予測情報復号部、 1111 動き情報復号部、 1112 予測モード情報メモリ、 1113 イントラ予測部、 1114 動き補償予測部、 1115 出力端子、 1500 動き補償予測生成部、 1501 予測誤差算出部、 1502 予測ベクトル算出部、 1503 差分ベクトル算出部、 1504 動き情報符号量算出部、 1505 予測モード/ブロック構造評価部、 1506 結合動き情報算出部、 1507 結合動き情報単予測変換部、 1508 結合動き補償予測生成部、 1600 空間結合動き情報候補リスト生成部、 1601 結合動き情報候補リスト削除部、 1602 時間結合動き情報候補リスト生成部、 1603 第1結合動き情報候補リスト追加部、 1604 第2結合動き情報候補リスト追加部、 3600 動き情報ビットストリーム復号部、 3601 予測ベクトル算出部、 3602 ベクトル加算部、 3603 動き補償予測復号部、 3604 結合動き情報算出部、 3605 結合動き情報単予測変換部、 3606 結合動き補償予測復号部。 100 input terminal, 101 input image memory, 102 encoding block acquisition unit, 103 subtraction unit, 104 orthogonal transform / quantization unit, 105 prediction error encoding unit, 106 dequantization / inverse transform unit, 107 addition unit, 108 frame Inner decoded image buffer, 109 loop filter unit, 110 decoded image memory, 111 motion vector detection unit, 112 motion compensation prediction unit, 113 motion compensation prediction block structure selection unit, 114 intra prediction unit, 115 intra prediction block structure selection unit, 116 Prediction mode selection unit, 117 encoding block structure selection unit, 118 block structure / prediction mode information additional information encoding unit, 119 prediction mode information memory, 120 multiplexing unit, 121 output terminal, 122 code Block control parameter generation unit, 1100 input terminal, 1101 demultiplexing unit, 1102 prediction difference information decoding unit, 1103 inverse quantization / inverse transformation unit, 1104 addition unit, 1105 intra-frame decoded image buffer, 1106 loop filter unit, 1107 decoded image Memory, 1108 prediction mode / block structure decoding unit, 1109 prediction mode / block structure selection unit, 1110 intra prediction information decoding unit, 1111 motion information decoding unit, 1112 prediction mode information memory, 1113 intra prediction unit, 1114 motion compensation prediction unit, 1115 output terminal, 1500 motion compensation prediction generation unit, 1501 prediction error calculation unit, 1502 prediction vector calculation unit, 1503 difference vector calculation unit, 1504 motion information Code amount calculation unit, 1505 prediction mode / block structure evaluation unit, 1506 combined motion information calculation unit, 1507 combined motion information single prediction conversion unit, 1508 combined motion compensation prediction generation unit, 1600 spatial combined motion information candidate list generation unit, 1601 combination Motion information candidate list deletion unit, 1602 Time combined motion information candidate list generation unit, 1603 First combined motion information candidate list addition unit, 1604 Second combined motion information candidate list addition unit, 3600 motion information bitstream decoding unit, 3601 prediction vector Calculation unit, 3602 vector addition unit, 3603 motion compensation prediction decoding unit, 3604 combined motion information calculation unit, 3605 combined motion information single prediction conversion unit, 3606 combined motion compensation prediction decoding unit.
 本発明は、動画像信号の符号化及び復号技術に利用できる。 The present invention can be used for encoding and decoding techniques of moving image signals.

Claims (24)

  1.  ピクチャが複数のブロックに段階的に分割されたブロックから予測ブロックを特定し、その特定された予測ブロック単位で、符号化ストリームを生成する動画像符号化装置であって、
     符号化対象となる予測ブロックに空間的に近接するブロック及び時間的に近接するブロックの少なくとも何れかから動き情報を導出し、前記符号化対象となる予測ブロックの動き情報候補として、その導出された動き情報の中から所定の動き情報を登録して動き情報候補リストを構築する候補リスト構築部と、
     前記符号化対象となる前記予測ブロックに用いる前記動き情報候補リスト内の動き情報候補を指定するインデックス情報を符号化する符号化部と、
     前記動き情報候補を変換する動き情報変換部と、
     前記動き情報候補に基づいて、単予測又は双予測の何れかにより動き補償予測を行い前記符号化対象となる予測ブロックの予測信号を生成する動き補償予測部とを備え、
     前記動き情報変換部は、前記動き情報候補のうち、前記双予測を示す予測種別情報を、前記単予測を示す予測種別情報に変換する予測変換を行い、
     前記動き補償予測部は、前記符号化対象となる予測ブロックのブロックサイズが所定の第1サイズの場合であり、かつ予測種別情報が前記双予測を示す場合に、前記予測変換により変換された動き情報に基づいて前記動き補償予測を行うことを特徴とする動画像符号化装置。
    A video encoding device that identifies a prediction block from a block in which a picture is divided into a plurality of blocks in stages, and generates an encoded stream in units of the identified prediction block,
    Motion information is derived from at least one of a block spatially adjacent to the prediction block to be encoded and a block temporally adjacent to the prediction block, and the motion information is derived as a motion information candidate of the prediction block to be encoded. A candidate list construction unit for registering predetermined motion information from the motion information and constructing a motion information candidate list;
    An encoding unit that encodes index information that specifies motion information candidates in the motion information candidate list used for the prediction block to be encoded;
    A motion information conversion unit for converting the motion information candidates;
    A motion-compensated prediction unit that performs motion-compensated prediction by either uni-prediction or bi-prediction based on the motion information candidates and generates a prediction signal of the prediction block to be encoded,
    The motion information conversion unit performs prediction conversion to convert prediction type information indicating the bi-prediction into prediction type information indicating the single prediction among the motion information candidates,
    The motion compensated prediction unit is a motion converted by the prediction conversion when the block size of the prediction block to be encoded is a predetermined first size and the prediction type information indicates the bi-prediction. A moving picture coding apparatus that performs the motion compensation prediction based on information.
  2.  前記予測種別情報は、第1の参照ピクチャを用いるか否かを示す第1情報、及び第2の参照ピクチャを用いるか否かを示す第2情報からなり、
     前記動き情報変換部は、前記第1情報が前記第1の参照ピクチャを用いる旨を示し、かつ前記第2情報が前記第2の参照ピクチャを用いる旨を示す前記予測種別情報について、前記第1の参照ピクチャを用いる旨を無効とする、又は前記第2の参照ピクチャを用いる旨を無効とすることで、前記予測変換を行うことを特徴とする請求項1に記載の動画像符号化装置。
    The prediction type information includes first information indicating whether or not to use a first reference picture, and second information indicating whether or not to use a second reference picture,
    The motion information conversion unit indicates the first type of prediction type information indicating that the first reference picture is used and the second information is that the second reference picture is used. The video encoding apparatus according to claim 1, wherein the prediction conversion is performed by invalidating the use of the reference picture or invalidating the use of the second reference picture.
  3.  前記候補リスト構築部は、
     少なくとも、前記符号化対象となる予測ブロックに前記空間的に近接するブロックから導出された前記動き情報候補を、前記予測種別情報に応じて組み合わせ、第1の新たな動き情報を導出し、その導出された第1の新たな動き情報を前記動き情報候補リストに追加する第1リスト追加部と、
     少なくとも、前記符号化対象となる予測ブロックに前記空間的に近接するブロックから前記導出された前記動き情報候補のいずれにも依存しない第2の新たな動き情報を導出し、その導出された第2の新たな動き情報を前記動き情報候補リストに追加する第2リスト追加部とを備え、
     少なくとも前記第1リスト追加部が前記第1の新たな動き情報を追加して前記動き情報候補リストの構築を行うことを特徴とする請求項1または2に記載の動画像符号化装置。
    The candidate list construction unit
    Combining at least the motion information candidates derived from the block spatially adjacent to the prediction block to be encoded according to the prediction type information, deriving first new motion information, and deriving the first new motion information A first list adding unit that adds the first new motion information that has been added to the motion information candidate list;
    Deriving at least second new motion information that does not depend on any of the derived motion information candidates from a block that is spatially adjacent to the prediction block to be encoded, and the second A second list adding unit for adding the new motion information to the motion information candidate list,
    3. The moving picture encoding apparatus according to claim 1, wherein at least the first list adding unit adds the first new motion information to construct the motion information candidate list. 4.
  4.  前記動き情報変換部は、候補リスト構築部が前記動き情報候補リストを構築した後に、前記予測変換を行うことを特徴とする請求項1から3のいずれかに記載の動画像符号化装置。 4. The moving picture encoding apparatus according to claim 1, wherein the motion information conversion unit performs the prediction conversion after the candidate list construction unit constructs the motion information candidate list.
  5.  前記候補リスト構築部は、サイズが所定の第2サイズの前記分割されたブロック内に存在する前記符号化対象となる予測ブロックに対しては、前記第2サイズの候補ブロックに空間的に近接する前記分割されたブロック及び時間的に近接する前記分割されたブロックの少なくとも何れかから導出された動き情報を、前記動き情報候補とすることを特徴とする請求項1から4のいずれかに記載の動画像符号化装置。 The candidate list construction unit is spatially close to the candidate block of the second size for the prediction block to be encoded that exists in the divided block having a predetermined second size. 5. The motion information derived from at least one of the divided blocks and the divided blocks that are temporally adjacent to each other is used as the motion information candidate. Video encoding device.
  6.  前記動き補償予測部はさらに、前記符号化対象となる予測ブロックのブロックサイズが前記第1サイズ未満の第3サイズの場合には、前記動き補償予測を禁止することを特徴とする請求項1から5のいずれかに記載の動画像符号化装置。 The motion compensated prediction unit further prohibits the motion compensated prediction when the block size of the prediction block to be encoded is a third size less than the first size. 6. The moving image encoding device according to any one of 5).
  7.  動画像の各ピクチャを分割したブロック単位で動き補償予測を用いて前記動画像を符号化する動画像符号化装置であって、
     導出した動き情報を用いた動き補償により符号化対象予測ブロックの予測信号を生成する動き補償予測部と、
     指定された第1のサイズの予測ブロックサイズにおける、動き補償予測を許可するか否かを指定する第1の制御パラメータと、指定された第2のサイズ以下の予測ブロックサイズにおける双予測の動き補償を禁止する、前記第2のサイズを指定する第2の制御パラメータを生成する符号化ブロック制御パラメータ生成部と、
     前記第1及び第2の制御パラメータを含む、動き補償予測に用いる情報を符号化する符号化部とを備え、
     前記動き補償予測部は、前記第1及び第2の制御パラメータに基づき、動き補償予測を行うことを特徴とする動画像符号化装置。
    A moving image encoding apparatus that encodes the moving image using motion compensated prediction in units of blocks obtained by dividing each picture of the moving image,
    A motion compensation prediction unit that generates a prediction signal of the prediction block to be encoded by motion compensation using the derived motion information;
    A first control parameter that specifies whether or not motion compensated prediction is permitted in a prediction block size of a specified first size, and bi-prediction motion compensation in a prediction block size that is less than or equal to a specified second size An encoded block control parameter generation unit that generates a second control parameter that specifies the second size,
    An encoding unit that encodes information used for motion compensated prediction, including the first and second control parameters;
    The moving image encoding apparatus, wherein the motion compensation prediction unit performs motion compensation prediction based on the first and second control parameters.
  8.  ピクチャが複数のブロックに段階的に分割されたブロックから予測ブロックを特定し、その特定された予測ブロック単位で、符号化ストリームを生成する動画像符号化方法であって、
     符号化対象となる予測ブロックに空間的に近接するブロック及び時間的に近接するブロックの少なくとも何れかから動き情報を導出し、前記符号化対象となる予測ブロックの動き情報候補として、その導出された動き情報の中から所定の動き情報を登録して動き情報候補リストを構築する候補リスト構築ステップと、
     前記符号化対象となる前記予測ブロックに用いる前記動き情報候補リスト内の動き情報候補を指定するインデックス情報を符号化する符号化ステップと、
     前記動き情報候補を変換する動き情報変換ステップと、
     前記動き情報候補に基づいて、単予測又は双予測の何れかにより動き補償予測を行い前記符号化対象となる予測ブロックの予測信号を生成する動き補償予測ステップとを備え、
     前記動き情報変換ステップは、前記動き情報候補のうち、前記双予測を示す予測種別情報を、前記単予測を示す予測種別情報に変換する予測変換を行い、
     前記動き補償予測ステップは、前記符号化対象となる予測ブロックのブロックサイズが所定の第1サイズの場合であり、かつ予測種別情報が前記双予測を示す場合に、前記予測変換により変換された動き情報に基づいて前記動き補償予測を行うことを特徴とする動画像符号化方法。
    A moving picture encoding method for specifying a prediction block from a block in which a picture is divided into a plurality of blocks and generating an encoded stream in the specified prediction block unit,
    Motion information is derived from at least one of a block spatially adjacent to the prediction block to be encoded and a block temporally adjacent to the prediction block, and the motion information is derived as a motion information candidate of the prediction block to be encoded. A candidate list construction step of registering predetermined motion information from the motion information and constructing a motion information candidate list;
    An encoding step for encoding index information for specifying motion information candidates in the motion information candidate list used for the prediction block to be encoded;
    A motion information conversion step for converting the motion information candidates;
    A motion-compensated prediction step of performing motion-compensated prediction based on either the single prediction or the bi-prediction based on the motion information candidate and generating a prediction signal of the prediction block to be encoded,
    The motion information conversion step performs prediction conversion for converting prediction type information indicating the bi-prediction into prediction type information indicating the single prediction among the motion information candidates,
    The motion-compensated prediction step is a motion converted by the prediction conversion when the block size of the prediction block to be encoded is a predetermined first size and the prediction type information indicates the bi-prediction. A moving picture coding method, wherein the motion compensation prediction is performed based on information.
  9.  ピクチャが複数のブロックに段階的に分割されたブロックから予測ブロックを特定し、その特定された予測ブロック単位で、符号化ストリームを生成する動画像符号化プログラムであって、
     符号化対象となる予測ブロックに空間的に近接するブロック及び時間的に近接するブロックの少なくとも何れかから動き情報を導出し、前記符号化対象となる予測ブロックの動き情報候補として、その導出された動き情報の中から所定の動き情報を登録して動き情報候補リストを構築する候補リスト構築ステップと、
     前記符号化対象となる前記予測ブロックに用いる前記動き情報候補リスト内の動き情報候補を指定するインデックス情報を符号化する符号化ステップと、
     前記動き情報候補を変換する動き情報変換ステップと、
     前記動き情報候補に基づいて、単予測又は双予測の何れかにより動き補償予測を行い前記符号化対象となる予測ブロックの予測信号を生成する動き補償予測ステップとをコンピュータに実行させ、
     前記動き情報変換ステップは、前記動き情報候補のうち、前記双予測を示す予測種別情報を、前記単予測を示す予測種別情報に変換する予測変換を行い、
     前記動き補償予測ステップは、前記符号化対象となる予測ブロックのブロックサイズが所定の第1サイズの場合であり、かつ予測種別情報が前記双予測を示す場合に、前記予測変換により変換された動き情報に基づいて前記動き補償予測を行うことを特徴とする動画像符号化プログラム。
    A moving picture encoding program for specifying a prediction block from a block in which a picture is divided into a plurality of blocks in stages, and generating an encoded stream in the specified prediction block unit,
    Motion information is derived from at least one of a block spatially adjacent to the prediction block to be encoded and a block temporally adjacent to the prediction block, and the motion information is derived as a motion information candidate of the prediction block to be encoded. A candidate list construction step of registering predetermined motion information from the motion information and constructing a motion information candidate list;
    An encoding step for encoding index information for specifying motion information candidates in the motion information candidate list used for the prediction block to be encoded;
    A motion information conversion step for converting the motion information candidates;
    Based on the motion information candidates, a computer performs a motion compensation prediction step of performing motion compensation prediction by either uni-prediction or bi-prediction and generating a prediction signal of the prediction block to be encoded,
    The motion information conversion step performs prediction conversion for converting prediction type information indicating the bi-prediction into prediction type information indicating the single prediction among the motion information candidates,
    The motion-compensated prediction step is a motion converted by the prediction conversion when the block size of the prediction block to be encoded is a predetermined first size and the prediction type information indicates the bi-prediction. A moving picture coding program which performs the motion compensation prediction based on information.
  10.  ピクチャが複数のブロックに段階的に分割されたブロックから予測ブロックを特定し、その特定された予測ブロック単位で、符号化ストリームを生成する動画像符号化方法により符号化された前記符号化ストリームをパケット化して符号化データを得るパケット処理部と、
     パケット化された前記符号化データを送信する送信部とを備え、
     前記動画像符号化方法は、
     符号化対象となる予測ブロックに空間的に近接するブロック及び時間的に近接するブロックの少なくとも何れかから動き情報を導出し、前記符号化対象となる予測ブロックの動き情報候補として、その導出された動き情報の中から所定の動き情報を登録して動き情報候補リストを構築する候補リスト構築ステップと、
     前記符号化対象となる前記予測ブロックに用いる前記動き情報候補リスト内の動き情報候補を指定するインデックス情報を符号化する符号化ステップと、
     前記動き情報候補を変換する動き情報変換ステップと、
     前記動き情報候補に基づいて、単予測又は双予測の何れかにより動き補償予測を行い前記符号化対象となる予測ブロックの予測信号を生成する動き補償予測ステップとを備え、
     前記動き情報変換ステップは、前記動き情報候補のうち、前記双予測を示す予測種別情報を、前記単予測を示す予測種別情報に変換する予測変換を行い、
     前記動き補償予測ステップは、前記符号化対象となる予測ブロックのブロックサイズが所定の第1サイズの場合であり、かつ予測種別情報が前記双予測を示す場合に、前記予測変換により変換された動き情報に基づいて前記動き補償予測を行うことを特徴とする送信装置。
    A predicted block is identified from a block obtained by dividing a picture into a plurality of blocks in stages, and the coded stream encoded by a moving picture coding method for generating a coded stream in the identified predicted block unit A packet processing unit that packetizes to obtain encoded data;
    A transmission unit for transmitting the packetized encoded data,
    The moving image encoding method includes:
    Motion information is derived from at least one of a block spatially adjacent to the prediction block to be encoded and a block temporally adjacent to the prediction block, and the motion information is derived as a motion information candidate of the prediction block to be encoded. A candidate list construction step of registering predetermined motion information from the motion information and constructing a motion information candidate list;
    An encoding step for encoding index information for specifying motion information candidates in the motion information candidate list used for the prediction block to be encoded;
    A motion information conversion step for converting the motion information candidates;
    A motion-compensated prediction step of performing motion-compensated prediction based on either the single prediction or the bi-prediction based on the motion information candidate and generating a prediction signal of the prediction block to be encoded,
    The motion information conversion step performs prediction conversion for converting prediction type information indicating the bi-prediction into prediction type information indicating the single prediction among the motion information candidates,
    The motion-compensated prediction step is a motion converted by the prediction conversion when the block size of the prediction block to be encoded is a predetermined first size and the prediction type information indicates the bi-prediction. A transmission apparatus that performs the motion compensation prediction based on information.
  11.  ピクチャが複数のブロックに段階的に分割されたブロックから予測ブロックを特定し、その特定された予測ブロック単位で、符号化ストリームを生成する動画像符号化方法により符号化された前記符号化ストリームをパケット化して符号化データを得るパケット処理ステップと、
     パケット化された前記符号化データを送信する送信ステップとを備え、
     前記動画像符号化方法は、
     符号化対象となる予測ブロックに空間的に近接するブロック及び時間的に近接するブロックの少なくとも何れかから動き情報を導出し、前記符号化対象となる予測ブロックの動き情報候補として、その導出された動き情報の中から所定の動き情報を登録して動き情報候補リストを構築する候補リスト構築ステップと、
     前記符号化対象となる前記予測ブロックに用いる前記動き情報候補リスト内の動き情報候補を指定するインデックス情報を符号化する符号化ステップと、
     前記動き情報候補を変換する動き情報変換ステップと、
     前記動き情報候補に基づいて、単予測又は双予測の何れかにより動き補償予測を行い前記符号化対象となる予測ブロックの予測信号を生成する動き補償予測ステップとを備え、
     前記動き情報変換ステップは、前記動き情報候補のうち、前記双予測を示す予測種別情報を、前記単予測を示す予測種別情報に変換する予測変換を行い、
     前記動き補償予測ステップは、前記符号化対象となる予測ブロックのブロックサイズが所定の第1サイズの場合であり、かつ予測種別情報が前記双予測を示す場合に、前記予測変換により変換された動き情報に基づいて前記動き補償予測を行うことを特徴とする送信方法。
    A predicted block is identified from a block obtained by dividing a picture into a plurality of blocks in stages, and the coded stream encoded by a moving picture coding method for generating a coded stream in the identified predicted block unit A packet processing step of packetizing to obtain encoded data;
    Transmitting the packetized encoded data, and
    The moving image encoding method includes:
    Motion information is derived from at least one of a block spatially adjacent to the prediction block to be encoded and a block temporally adjacent to the prediction block, and the motion information is derived as a motion information candidate of the prediction block to be encoded. A candidate list construction step of registering predetermined motion information from the motion information and constructing a motion information candidate list;
    An encoding step for encoding index information for specifying motion information candidates in the motion information candidate list used for the prediction block to be encoded;
    A motion information conversion step for converting the motion information candidates;
    A motion-compensated prediction step of performing motion-compensated prediction based on either the single prediction or the bi-prediction based on the motion information candidate and generating a prediction signal of the prediction block to be encoded,
    The motion information conversion step performs prediction conversion for converting prediction type information indicating the bi-prediction into prediction type information indicating the single prediction among the motion information candidates,
    The motion-compensated prediction step is a motion converted by the prediction conversion when the block size of the prediction block to be encoded is a predetermined first size and the prediction type information indicates the bi-prediction. A transmission method characterized in that the motion compensation prediction is performed based on information.
  12.  ピクチャが複数のブロックに段階的に分割されたブロックから予測ブロックを特定し、その特定された予測ブロック単位で、符号化ストリームを生成する動画像符号化方法により符号化された前記符号化ストリームをパケット化して符号化データを得るパケット処理ステップと、
     パケット化された前記符号化データを送信する送信ステップとをコンピュータに実行させ、
     前記動画像符号化方法は、
     符号化対象となる予測ブロックに空間的に近接するブロック及び時間的に近接するブロックの少なくとも何れかから動き情報を導出し、前記符号化対象となる予測ブロックの動き情報候補として、その導出された動き情報の中から所定の動き情報を登録して動き情報候補リストを構築する候補リスト構築ステップと、
     前記符号化対象となる前記予測ブロックに用いる前記動き情報候補リスト内の動き情報候補を指定するインデックス情報を符号化する符号化ステップと、
     前記動き情報候補を変換する動き情報変換ステップと、
     前記動き情報候補に基づいて、単予測又は双予測の何れかにより動き補償予測を行い前記符号化対象となる予測ブロックの予測信号を生成する動き補償予測ステップとを備え、
     前記動き情報変換ステップは、前記動き情報候補のうち、前記双予測を示す予測種別情報を、前記単予測を示す予測種別情報に変換する予測変換を行い、
     前記動き補償予測ステップは、前記符号化対象となる予測ブロックのブロックサイズが所定の第1サイズの場合であり、かつ予測種別情報が前記双予測を示す場合に、前記予測変換により変換された動き情報に基づいて前記動き補償予測を行うことを特徴とする送信プログラム。
    A predicted block is identified from a block obtained by dividing a picture into a plurality of blocks in stages, and the coded stream encoded by a moving picture coding method for generating a coded stream in the identified predicted block unit A packet processing step of packetizing to obtain encoded data;
    Transmitting the packetized encoded data to the computer, and
    The moving image encoding method includes:
    Motion information is derived from at least one of a block spatially adjacent to the prediction block to be encoded and a block temporally adjacent to the prediction block, and the motion information is derived as a motion information candidate of the prediction block to be encoded. A candidate list construction step of registering predetermined motion information from the motion information and constructing a motion information candidate list;
    An encoding step for encoding index information for specifying motion information candidates in the motion information candidate list used for the prediction block to be encoded;
    A motion information conversion step for converting the motion information candidates;
    A motion-compensated prediction step of performing motion-compensated prediction based on either the single prediction or the bi-prediction based on the motion information candidate and generating a prediction signal of the prediction block to be encoded,
    The motion information conversion step performs prediction conversion for converting prediction type information indicating the bi-prediction into prediction type information indicating the single prediction among the motion information candidates,
    The motion-compensated prediction step is a motion converted by the prediction conversion when the block size of the prediction block to be encoded is a predetermined first size and the prediction type information indicates the bi-prediction. A transmission program that performs the motion compensation prediction based on information.
  13.  ピクチャが複数のブロックに段階的に分割されたブロックから予測ブロックを特定し、その特定された予測ブロック単位で、符号化ストリームを復号する動画像復号装置であって、
     前記符号化ストリームから、復号対象となる前記予測ブロックの動き情報を指定したインデックス情報を復号する復号部と、
     前記復号対象となる予測ブロックに空間的に近接するブロック及び時間的に近接するブロックの少なくとも何れかから動き情報を導出し、前記復号対象となる予測ブロックの動き情報候補として、その導出された動き情報の中から所定の動き情報を登録して動き情報候補リストを構築する候補リスト構築部と、
     前記動き情報候補を変換する動き情報変換部と、
     前記動き情報候補のうちの前記インデックス情報により指定された動き情報に基づいて、単予測又は双予測の何れかにより動き補償予測を行い前記復号対象となる予測ブロックの予測信号を生成する動き補償予測部とを備え、
     前記動き情報変換部は、前記動き情報候補のうち、前記双予測を示す予測種別情報を、前記単予測を示す予測種別情報に変換する予測変換を行い、
     前記動き補償予測部は、前記復号対象となる予測ブロックのブロックサイズが所定の第1サイズの場合であり、かつ前記指定された動き情報の予測種別情報が前記双予測を示す場合に、前記予測変換により変換された動き情報に基づいて前記動き補償予測を行うことを特徴とする動画像復号装置。
    A video decoding device that identifies a prediction block from a block in which a picture is divided into a plurality of blocks and decodes an encoded stream in units of the identified prediction block,
    A decoding unit that decodes index information that specifies motion information of the prediction block to be decoded from the encoded stream;
    Motion information is derived from at least one of a block spatially adjacent to the prediction block to be decoded and a block temporally adjacent to the prediction block, and the derived motion is used as a motion information candidate of the prediction block to be decoded. A candidate list construction unit for registering predetermined motion information from the information and constructing a motion information candidate list;
    A motion information conversion unit for converting the motion information candidates;
    Motion compensated prediction based on the motion information specified by the index information of the motion information candidates, and performing motion compensation prediction by either uni-prediction or bi-prediction to generate a prediction signal of the prediction block to be decoded With
    The motion information conversion unit performs prediction conversion to convert prediction type information indicating the bi-prediction into prediction type information indicating the single prediction among the motion information candidates,
    The motion compensation prediction unit is configured when the block size of the prediction block to be decoded is a predetermined first size and the prediction type information of the designated motion information indicates the bi-prediction. A moving picture decoding apparatus that performs the motion compensation prediction based on motion information converted by conversion.
  14.  前記予測種別情報は、第1の参照ピクチャを用いるか否かを示す第1情報、及び第2の参照ピクチャを用いるか否かを示す第2情報からなり、
     前記動き情報変換部は、前記第1情報が前記第1の参照ピクチャを用いる旨を示し、かつ前記第2情報が前記第2の参照ピクチャを用いる旨を示す前記予測種別情報について、前記第1の参照ピクチャを用いる旨を無効とする、又は前記第2の参照ピクチャを用いる旨を無効とすることで、前記予測変換を行うことを特徴とする請求項13に記載の動画像復号装置。
    The prediction type information includes first information indicating whether or not to use a first reference picture, and second information indicating whether or not to use a second reference picture,
    The motion information conversion unit indicates the first type of prediction type information indicating that the first reference picture is used and the second information is that the second reference picture is used. 14. The moving picture decoding apparatus according to claim 13, wherein the predictive conversion is performed by invalidating the use of the reference picture or invalidating the use of the second reference picture.
  15.  前記候補リスト構築部は、
     少なくとも、前記復号対象となる予測ブロックに前記空間的に近接するブロックから導出された前記動き情報候補を、前記予測種別情報に応じて組み合わせ、第1の新たな動き情報を導出し、その導出された第1の新たな動き情報を前記動き情報候補リストに追加する第1リスト追加部と、
     少なくとも、前記復号対象となる予測ブロックに前記空間的に近接するブロックから前記導出された前記動き情報候補のいずれにも依存しない第2の新たな動き情報を導出し、その導出された第2の新たな動き情報を前記動き情報候補リストに追加する第2リスト追加部とを備え、
     少なくとも前記第1リスト追加部が前記第1の新たな動き情報を追加して前記動き情報候補リストの構築を行うことを特徴とする請求項13または14に記載の動画像復号装置。
    The candidate list construction unit
    At least the motion information candidates derived from the spatially adjacent blocks to the decoding target block are combined according to the prediction type information to derive first new motion information, which is derived A first list adding unit for adding the first new motion information to the motion information candidate list;
    Deriving at least second new motion information that does not depend on any of the derived motion information candidates from a block that is spatially close to the prediction block to be decoded; A second list adding unit for adding new motion information to the motion information candidate list,
    The moving picture decoding apparatus according to claim 13 or 14, wherein at least the first list adding unit adds the first new motion information to construct the motion information candidate list.
  16.  前記動き情報変換部は、候補リスト構築部が前記動き情報候補リストを構築した後に、前記予測変換を行うことを特徴とする請求項13から15のいずれかに記載の動画像復号装置。 16. The moving picture decoding apparatus according to claim 13, wherein the motion information conversion unit performs the prediction conversion after the candidate list construction unit constructs the motion information candidate list.
  17.  前記候補リスト構築部は、サイズが所定の第2サイズの前記分割されたブロック内に存在する前記復号対象となる予測ブロックに対しては、前記第2サイズの候補ブロックに空間的に近接する前記分割されたブロック及び時間的に近接する前記分割されたブロックの少なくとも何れかから導出された動き情報を、前記動き情報候補とすることを特徴とする請求項13から16のいずれかに記載の動画像復号装置。 The candidate list construction unit is spatially close to the candidate block of the second size for the prediction block to be decoded that exists in the divided block of a predetermined second size. The moving image according to claim 13, wherein motion information derived from at least one of the divided blocks and the divided blocks that are temporally adjacent to each other is set as the motion information candidate. Image decoding device.
  18.  前記動き補償予測部はさらに、前記復号対象となる予測ブロックのブロックサイズが前記第1サイズ未満の第3サイズの場合には、前記動き補償予測を禁止することを特徴とする請求項13から17のいずれかに記載の動画像復号装置。 The motion compensated prediction unit further prohibits the motion compensated prediction when a block size of the prediction block to be decoded is a third size less than the first size. Any one of the moving image decoding apparatuses.
  19.  動画像の各ピクチャを分割したブロック単位で動き補償予測を用いて前記動画像を符号化した符号化ストリームを復号する動画像復号装置であって、
     前記符号化ストリームから動き補償予測に用いる情報を復号すると共に、その復号した動き補償予測に用いる情報から、指定された第1のサイズの予測ブロックサイズにおける、動き補償予測を許可するか否かを指定する第1の制御パラメータと、指定された第2のサイズ以下の予測ブロックサイズにおける双予測の動き補償を禁止する、前記第2のサイズを指定する第2の制御パラメータとを得る復号部と、
     前記動き補償予測に用いる情報を用いて復号対象予測ブロックの予測信号を生成する動き補償予測部とを備え、
     前記動き補償予測部は、前記第1及び第2の制御パラメータに基づき、動き補償予測を行うことを特徴とする動画像復号装置。
    A moving picture decoding apparatus for decoding a coded stream obtained by coding the moving picture using motion compensated prediction in units of blocks obtained by dividing each picture of the moving picture,
    Whether to decode motion-compensated prediction from the encoded stream and whether to permit motion-compensated prediction in the designated prediction block size of the first size from the decoded information used for motion-compensated prediction. A decoding unit that obtains a first control parameter to be designated, and a second control parameter to designate the second size, which prohibits bi-prediction motion compensation in a prediction block size equal to or smaller than the designated second size; ,
    A motion compensation prediction unit that generates a prediction signal of a decoding target prediction block using information used for the motion compensation prediction,
    The video decoding apparatus, wherein the motion compensation prediction unit performs motion compensation prediction based on the first and second control parameters.
  20.  ピクチャが複数のブロックに段階的に分割されたブロックから予測ブロックを特定し、その特定された予測ブロック単位で、符号化ストリームを復号する動画像復号方法であって、
     前記符号化ストリームから、復号対象となる前記予測ブロックの動き情報を指定したインデックス情報を復号する復号ステップと、
     前記復号対象となる予測ブロックに空間的に近接するブロック及び時間的に近接するブロックの少なくとも何れかから動き情報を導出し、前記復号対象となる予測ブロックの動き情報候補として、その導出された動き情報の中から所定の動き情報を登録して動き情報候補リストを構築する候補リスト構築ステップと、
     前記動き情報候補を変換する動き情報変換ステップと、
     前記動き情報候補のうちの前記インデックス情報により指定された動き情報に基づいて、単予測又は双予測の何れかにより動き補償予測を行い前記復号対象となる予測ブロックの予測信号を生成する動き補償予測ステップとを備え、
     前記動き情報変換ステップは、前記動き情報候補のうち、前記双予測を示す予測種別情報を、前記単予測を示す予測種別情報に変換する予測変換を行い、
     前記動き補償予測ステップは、前記復号対象となる予測ブロックのブロックサイズが所定の第1サイズの場合であり、かつ前記指定された動き情報の予測種別情報が前記双予測を示す場合に、前記予測変換により変換された動き情報に基づいて前記動き補償予測を行うことを特徴とする動画像復号方法。
    A video decoding method for identifying a prediction block from a block in which a picture is divided into a plurality of blocks in stages, and decoding an encoded stream in units of the identified prediction block,
    A decoding step of decoding index information specifying motion information of the prediction block to be decoded from the encoded stream;
    Motion information is derived from at least one of a block spatially adjacent to the prediction block to be decoded and a block temporally adjacent to the prediction block, and the derived motion is used as a motion information candidate of the prediction block to be decoded. A candidate list construction step of registering predetermined motion information from the information and constructing a motion information candidate list;
    A motion information conversion step for converting the motion information candidates;
    Motion compensated prediction based on the motion information specified by the index information of the motion information candidates, and performing motion compensation prediction by either uni-prediction or bi-prediction to generate a prediction signal of the prediction block to be decoded With steps,
    The motion information conversion step performs prediction conversion for converting prediction type information indicating the bi-prediction into prediction type information indicating the single prediction among the motion information candidates,
    The motion compensation prediction step is performed when the block size of the prediction block to be decoded is a predetermined first size and the prediction type information of the designated motion information indicates the bi-prediction. A moving picture decoding method, wherein the motion compensation prediction is performed based on motion information converted by conversion.
  21.  ピクチャが複数のブロックに段階的に分割されたブロックから予測ブロックを特定し、その特定された予測ブロック単位で、符号化ストリームを復号する動画像復号プログラムであって、
     前記符号化ストリームから、復号対象となる前記予測ブロックの動き情報を指定したインデックス情報を復号する復号ステップと、
     前記復号対象となる予測ブロックに空間的に近接するブロック及び時間的に近接するブロックの少なくとも何れかから動き情報を導出し、前記復号対象となる予測ブロックの動き情報候補として、その導出された動き情報の中から所定の動き情報を登録して動き情報候補リストを構築する候補リスト構築ステップと、
     前記動き情報候補を変換する動き情報変換ステップと、
     前記動き情報候補のうちの前記インデックス情報により指定された動き情報に基づいて、単予測又は双予測の何れかにより動き補償予測を行い前記復号対象となる予測ブロックの予測信号を生成する動き補償予測ステップとをコンピュータに実行させ、
     前記動き情報変換ステップは、前記動き情報候補のうち、前記双予測を示す予測種別情報を、前記単予測を示す予測種別情報に変換する予測変換を行い、
     前記動き補償予測ステップは、前記復号対象となる予測ブロックのブロックサイズが所定の第1サイズの場合であり、かつ前記指定された動き情報の予測種別情報が前記双予測を示す場合に、前記予測変換により変換された動き情報に基づいて前記動き補償予測を行うことを特徴とする動画像復号プログラム。
    A moving picture decoding program that identifies a prediction block from a block in which a picture is divided into a plurality of blocks and decodes an encoded stream in units of the identified prediction block,
    A decoding step of decoding index information specifying motion information of the prediction block to be decoded from the encoded stream;
    Motion information is derived from at least one of a block spatially adjacent to the prediction block to be decoded and a block temporally adjacent to the prediction block, and the derived motion is used as a motion information candidate of the prediction block to be decoded. A candidate list construction step of registering predetermined motion information from the information and constructing a motion information candidate list;
    A motion information conversion step for converting the motion information candidates;
    Motion compensated prediction based on the motion information specified by the index information of the motion information candidates, and performing motion compensation prediction by either uni-prediction or bi-prediction to generate a prediction signal of the prediction block to be decoded Step to the computer,
    The motion information conversion step performs prediction conversion for converting prediction type information indicating the bi-prediction into prediction type information indicating the single prediction among the motion information candidates,
    The motion compensation prediction step is performed when the block size of the prediction block to be decoded is a predetermined first size and the prediction type information of the designated motion information indicates the bi-prediction. A moving picture decoding program that performs the motion compensation prediction based on motion information converted by conversion.
  22.  ピクチャが複数のブロックに段階的に分割されたブロックから予測ブロックを特定し、その特定された予測ブロック単位で、動画像が符号化された符号化ストリームを受信して復号する受信装置であって、
     前記符号化ストリームがパケット化された符号化データを受信する受信部と、
     受信された前記符号化ストリームをパケット処理して元の符号化ストリームを復元する復元部と、
     復元された前記符号化ストリームから、復号対象となる前記予測ブロックの動き情報を指定したインデックス情報を復号する復号部と、
     前記復号対象となる予測ブロックに空間的に近接するブロック及び時間的に近接するブロックの少なくとも何れかから動き情報を導出し、前記復号対象となる予測ブロックの動き情報候補として、その導出された動き情報の中から所定の動き情報を登録して動き情報候補リストを構築する候補リスト構築部と、
     前記動き情報候補を変換する動き情報変換部と、
     前記動き情報候補のうちの前記インデックス情報により指定された動き情報に基づいて、単予測又は双予測の何れかにより動き補償予測を行い前記復号対象となる予測ブロックの予測信号を生成する動き補償予測部とを備え、
     前記動き情報変換部は、前記動き情報候補のうち、前記双予測により前記動き補償予測を行うことを示す予測種別情報を、前記単予測により前記動き補償予測を行うことを示す予測種別情報に変換する予測変換を行い、
     前記動き補償予測部は、前記復号対象となる予測ブロックのブロックサイズが所定の第1サイズの場合であり、かつ前記指定された動き情報の前記予測種別情報が前記双予測により前記動き補償予測を行うことを示す場合に、前記予測変換により変換された前記予測種別情報により前記動き補償予測を行うことを特徴とする受信装置。
    A receiving device that identifies a prediction block from a block in which a picture is divided into a plurality of blocks in stages, and receives and decodes an encoded stream in which a moving image is encoded in the specified prediction block unit. ,
    A receiving unit that receives encoded data in which the encoded stream is packetized;
    A restoration unit that packet-processes the received encoded stream to restore the original encoded stream;
    A decoding unit that decodes index information that specifies motion information of the prediction block to be decoded from the restored encoded stream;
    Motion information is derived from at least one of a block spatially adjacent to the prediction block to be decoded and a block temporally adjacent to the prediction block, and the derived motion is used as a motion information candidate of the prediction block to be decoded. A candidate list construction unit for registering predetermined motion information from the information and constructing a motion information candidate list;
    A motion information conversion unit for converting the motion information candidates;
    Motion compensated prediction based on the motion information specified by the index information of the motion information candidates, and performing motion compensation prediction by either uni-prediction or bi-prediction to generate a prediction signal of the prediction block to be decoded With
    The motion information conversion unit converts prediction type information indicating that the motion compensation prediction is performed by the bi-prediction among the motion information candidates into prediction type information indicating that the motion compensation prediction is performed by the single prediction. Perform predictive transformations,
    The motion compensation prediction unit is a case where the block size of the prediction block to be decoded is a predetermined first size, and the prediction type information of the designated motion information is the motion compensated prediction by the bi-prediction. A receiving apparatus that performs the motion-compensated prediction based on the prediction type information converted by the prediction conversion when indicating to perform.
  23.  ピクチャが複数のブロックに段階的に分割されたブロックから予測ブロックを特定し、その特定された予測ブロック単位で、動画像が符号化された符号化ストリームを受信して復号する受信方法であって、
     前記符号化ストリームがパケット化された符号化データを受信する受信ステップと、
     受信された前記符号化ストリームをパケット処理して元の符号化ストリームを復元する復元ステップと、
     復元された前記符号化ストリームから、復号対象となる前記予測ブロックの動き情報を指定したインデックス情報を復号する復号ステップと、
     前記復号対象となる予測ブロックに空間的に近接するブロック及び時間的に近接するブロックの少なくとも何れかから動き情報を導出し、前記復号対象となる予測ブロックの動き情報候補として、その導出された動き情報の中から所定の動き情報を登録して動き情報候補リストを構築する候補リスト構築ステップと、
     前記動き情報候補を変換する動き情報変換ステップと、
     前記動き情報候補のうちの前記インデックス情報により指定された動き情報に基づいて、単予測又は双予測の何れかにより動き補償予測を行い前記復号対象となる予測ブロックの予測信号を生成する動き補償予測ステップとを備え、
     前記動き情報変換ステップは、前記動き情報候補のうち、前記双予測により前記動き補償予測を行うことを示す予測種別情報を、前記単予測により前記動き補償予測を行うことを示す予測種別情報に変換する予測変換を行い、
     前記動き補償予測ステップは、前記復号対象となる予測ブロックのブロックサイズが所定の第1サイズの場合であり、かつ前記指定された動き情報の前記予測種別情報が前記双予測により前記動き補償予測を行うことを示す場合に、前記予測変換により変換された前記予測種別情報により前記動き補償予測を行うことを特徴とする受信方法。
    A reception method for specifying a prediction block from a block in which a picture is divided into a plurality of blocks in stages, and receiving and decoding an encoded stream in which a moving image is encoded in the specified prediction block unit. ,
    A reception step of receiving encoded data in which the encoded stream is packetized;
    A restoration step of packetizing the received encoded stream to restore the original encoded stream;
    A decoding step of decoding index information specifying motion information of the prediction block to be decoded from the restored encoded stream;
    Motion information is derived from at least one of a block spatially adjacent to the prediction block to be decoded and a block temporally adjacent to the prediction block, and the derived motion is used as a motion information candidate of the prediction block to be decoded. A candidate list construction step of registering predetermined motion information from the information and constructing a motion information candidate list;
    A motion information conversion step for converting the motion information candidates;
    Motion compensated prediction based on the motion information specified by the index information of the motion information candidates, and performing motion compensation prediction by either uni-prediction or bi-prediction to generate a prediction signal of the prediction block to be decoded With steps,
    In the motion information conversion step, prediction type information indicating that the motion compensation prediction is performed by the bi-prediction among the motion information candidates is converted into prediction type information indicating that the motion compensation prediction is performed by the single prediction. Perform predictive transformations,
    The motion compensation prediction step is a case where the block size of the prediction block to be decoded is a predetermined first size, and the prediction type information of the designated motion information is the motion compensated prediction by the bi-prediction. A reception method, wherein the motion compensation prediction is performed based on the prediction type information converted by the prediction conversion.
  24.  ピクチャが複数のブロックに段階的に分割されたブロックから予測ブロックを特定し、その特定された予測ブロック単位で、動画像が符号化された符号化ストリームを受信して復号する受信プログラムであって、
     前記符号化ストリームがパケット化された符号化データを受信する受信ステップと、
     受信された前記符号化ストリームをパケット処理して元の符号化ストリームを復元する復元ステップと、
     復元された前記符号化ストリームから、復号対象となる前記予測ブロックの動き情報を指定したインデックス情報を復号する復号ステップと、
     前記復号対象となる予測ブロックに空間的に近接するブロック及び時間的に近接するブロックの少なくとも何れかから動き情報を導出し、前記復号対象となる予測ブロックの動き情報候補として、その導出された動き情報の中から所定の動き情報を登録して動き情報候補リストを構築する候補リスト構築ステップと、
     前記動き情報候補を変換する動き情報変換ステップと、
     前記動き情報候補のうちの前記インデックス情報により指定された動き情報に基づいて、単予測又は双予測の何れかにより動き補償予測を行い前記復号対象となる予測ブロックの予測信号を生成する動き補償予測ステップとをコンピュータに実行させ、
     前記動き情報変換ステップは、前記動き情報候補のうち、前記双予測により前記動き補償予測を行うことを示す予測種別情報を、前記単予測により前記動き補償予測を行うことを示す予測種別情報に変換する予測変換を行い、
     前記動き補償予測ステップは、前記復号対象となる予測ブロックのブロックサイズが所定の第1サイズの場合であり、かつ前記指定された動き情報の前記予測種別情報が前記双予測により前記動き補償予測を行うことを示す場合に、前記予測変換により変換された前記予測種別情報により前記動き補償予測を行うことを特徴とする受信プログラム。
    A reception program that identifies a prediction block from a block in which a picture is divided into a plurality of blocks in stages, and receives and decodes an encoded stream in which a moving image is encoded in the specified prediction block unit. ,
    A reception step of receiving encoded data in which the encoded stream is packetized;
    A restoration step of packetizing the received encoded stream to restore the original encoded stream;
    A decoding step of decoding index information specifying motion information of the prediction block to be decoded from the restored encoded stream;
    Motion information is derived from at least one of a block spatially adjacent to the prediction block to be decoded and a block temporally adjacent to the prediction block, and the derived motion is used as a motion information candidate of the prediction block to be decoded. A candidate list construction step of registering predetermined motion information from the information and constructing a motion information candidate list;
    A motion information conversion step for converting the motion information candidates;
    Motion compensated prediction based on the motion information specified by the index information of the motion information candidates, and performing motion compensation prediction by either uni-prediction or bi-prediction to generate a prediction signal of the prediction block to be decoded Step to the computer,
    In the motion information conversion step, prediction type information indicating that the motion compensation prediction is performed by the bi-prediction among the motion information candidates is converted into prediction type information indicating that the motion compensation prediction is performed by the single prediction. Perform predictive transformations,
    The motion compensation prediction step is a case where the block size of the prediction block to be decoded is a predetermined first size, and the prediction type information of the designated motion information is the motion compensated prediction by the bi-prediction. A reception program that performs the motion-compensated prediction based on the prediction type information converted by the prediction conversion when indicating to perform.
PCT/JP2013/002565 2012-04-16 2013-04-16 Video encoding device, video encoding method, video encoding program, transmission device, transmission method, transmission program, video decoding device, video decoding method, video decoding program, reception device, reception method, and reception program WO2013157251A1 (en)

Applications Claiming Priority (8)

Application Number Priority Date Filing Date Title
JP2012-093091 2012-04-16
JP2012-093092 2012-04-16
JP2012093092 2012-04-16
JP2012093091 2012-04-16
JP2013-085474 2013-04-16
JP2013-085473 2013-04-16
JP2013085474A JP5987768B2 (en) 2012-04-16 2013-04-16 Moving picture coding apparatus, moving picture coding method, moving picture coding program, transmission apparatus, transmission method, and transmission program
JP2013085473A JP5987767B2 (en) 2012-04-16 2013-04-16 Moving picture decoding apparatus, moving picture decoding method, moving picture decoding program, receiving apparatus, receiving method, and receiving program

Publications (1)

Publication Number Publication Date
WO2013157251A1 true WO2013157251A1 (en) 2013-10-24

Family

ID=49383222

Family Applications (1)

Application Number Title Priority Date Filing Date
PCT/JP2013/002565 WO2013157251A1 (en) 2012-04-16 2013-04-16 Video encoding device, video encoding method, video encoding program, transmission device, transmission method, transmission program, video decoding device, video decoding method, video decoding program, reception device, reception method, and reception program

Country Status (1)

Country Link
WO (1) WO2013157251A1 (en)

Cited By (13)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN110662054A (en) * 2018-06-29 2020-01-07 北京字节跳动网络技术有限公司 Partial/full pruning when adding HMVP candidates to Merge/AMVP
US20210297659A1 (en) 2018-09-12 2021-09-23 Beijing Bytedance Network Technology Co., Ltd. Conditions for starting checking hmvp candidates depend on total number minus k
US11463685B2 (en) 2018-07-02 2022-10-04 Beijing Bytedance Network Technology Co., Ltd. LUTS with intra prediction modes and intra mode prediction from non-adjacent blocks
US11528501B2 (en) 2018-06-29 2022-12-13 Beijing Bytedance Network Technology Co., Ltd. Interaction between LUT and AMVP
US11589071B2 (en) 2019-01-10 2023-02-21 Beijing Bytedance Network Technology Co., Ltd. Invoke of LUT updating
US11641483B2 (en) 2019-03-22 2023-05-02 Beijing Bytedance Network Technology Co., Ltd. Interaction between merge list construction and other tools
US11695921B2 (en) 2018-06-29 2023-07-04 Beijing Bytedance Network Technology Co., Ltd Selection of coded motion information for LUT updating
US11877002B2 (en) 2018-06-29 2024-01-16 Beijing Bytedance Network Technology Co., Ltd Update of look up table: FIFO, constrained FIFO
US11895318B2 (en) 2018-06-29 2024-02-06 Beijing Bytedance Network Technology Co., Ltd Concept of using one or multiple look up tables to store motion information of previously coded in order and use them to code following blocks
US11909951B2 (en) 2019-01-13 2024-02-20 Beijing Bytedance Network Technology Co., Ltd Interaction between lut and shared merge list
US11909989B2 (en) 2018-06-29 2024-02-20 Beijing Bytedance Network Technology Co., Ltd Number of motion candidates in a look up table to be checked according to mode
US11956464B2 (en) 2019-01-16 2024-04-09 Beijing Bytedance Network Technology Co., Ltd Inserting order of motion candidates in LUT
US11973971B2 (en) 2018-06-29 2024-04-30 Beijing Bytedance Network Technology Co., Ltd Conditions for updating LUTs

Non-Patent Citations (3)

* Cited by examiner, † Cited by third party
Title
KENJI KONDO ET AL.: "AHG7: Modification of merge candidate derivation to reduce MC memory bandwidth [JCTVC-H0221]", JOINT COLLABORATIVE TEAM ON VIDEO CODING (JCT-VC) OF ITU-T SG16 WP3 AND ISO/IEC JTC1/SC29/WG11 8TH MEETING, 1 February 2012 (2012-02-01), SAN JOSE, CA, USA *
TOMOHIRO IKAI: "Bi-prediction restriction in small PU [JCTVC-G307_r1]", JOINT COLLABORATIVE TEAM ON VIDEO CODING (JCT-VC) OF ITU-T SG16 WP3 AND ISO/IEC JTC1/SC29/WG11 7TH MEETING, 21 November 2011 (2011-11-21) - 30 November 2011 (2011-11-30), GENEVA, CH *
TOSHIYASU SUGIO ET AL.: "Parsing Robustness for Merge/AMVP [JCTVC-F470]", JOINT COLLABORATIVE TEAM ON VIDEO CODING (JCT-VC) OF ITU-T SG16 WP3 AND ISO/IEC JTC1/SC29/WG11 6TH MEETING, 14 July 2011 (2011-07-14) - 22 July 2011 (2011-07-22), TORINO, IT *

Cited By (17)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US11706406B2 (en) 2018-06-29 2023-07-18 Beijing Bytedance Network Technology Co., Ltd Selection of coded motion information for LUT updating
US11909989B2 (en) 2018-06-29 2024-02-20 Beijing Bytedance Network Technology Co., Ltd Number of motion candidates in a look up table to be checked according to mode
CN110662054A (en) * 2018-06-29 2020-01-07 北京字节跳动网络技术有限公司 Partial/full pruning when adding HMVP candidates to Merge/AMVP
US11528501B2 (en) 2018-06-29 2022-12-13 Beijing Bytedance Network Technology Co., Ltd. Interaction between LUT and AMVP
US11877002B2 (en) 2018-06-29 2024-01-16 Beijing Bytedance Network Technology Co., Ltd Update of look up table: FIFO, constrained FIFO
US11895318B2 (en) 2018-06-29 2024-02-06 Beijing Bytedance Network Technology Co., Ltd Concept of using one or multiple look up tables to store motion information of previously coded in order and use them to code following blocks
US11528500B2 (en) 2018-06-29 2022-12-13 Beijing Bytedance Network Technology Co., Ltd. Partial/full pruning when adding a HMVP candidate to merge/AMVP
US11695921B2 (en) 2018-06-29 2023-07-04 Beijing Bytedance Network Technology Co., Ltd Selection of coded motion information for LUT updating
US11973971B2 (en) 2018-06-29 2024-04-30 Beijing Bytedance Network Technology Co., Ltd Conditions for updating LUTs
US11463685B2 (en) 2018-07-02 2022-10-04 Beijing Bytedance Network Technology Co., Ltd. LUTS with intra prediction modes and intra mode prediction from non-adjacent blocks
US20210297659A1 (en) 2018-09-12 2021-09-23 Beijing Bytedance Network Technology Co., Ltd. Conditions for starting checking hmvp candidates depend on total number minus k
US11997253B2 (en) 2018-09-12 2024-05-28 Beijing Bytedance Network Technology Co., Ltd Conditions for starting checking HMVP candidates depend on total number minus K
US11589071B2 (en) 2019-01-10 2023-02-21 Beijing Bytedance Network Technology Co., Ltd. Invoke of LUT updating
US11909951B2 (en) 2019-01-13 2024-02-20 Beijing Bytedance Network Technology Co., Ltd Interaction between lut and shared merge list
US11956464B2 (en) 2019-01-16 2024-04-09 Beijing Bytedance Network Technology Co., Ltd Inserting order of motion candidates in LUT
US11962799B2 (en) 2019-01-16 2024-04-16 Beijing Bytedance Network Technology Co., Ltd Motion candidates derivation
US11641483B2 (en) 2019-03-22 2023-05-02 Beijing Bytedance Network Technology Co., Ltd. Interaction between merge list construction and other tools

Similar Documents

Publication Publication Date Title
WO2013157251A1 (en) Video encoding device, video encoding method, video encoding program, transmission device, transmission method, transmission program, video decoding device, video decoding method, video decoding program, reception device, reception method, and reception program
JP6004136B1 (en) Moving picture decoding apparatus, moving picture decoding method, moving picture decoding program, receiving apparatus, receiving method, and receiving program
JP6183511B2 (en) Moving picture coding apparatus, moving picture coding method, moving picture coding program, transmission apparatus, transmission method, and transmission program
JP5786498B2 (en) Image coding apparatus, image coding method, and image coding program
TWI657696B (en) Motion image encoding device, motion image encoding method, and recording medium for recording motion image encoding program
JP5786499B2 (en) Image decoding apparatus, image decoding method, and image decoding program
JP6172326B2 (en) Image encoding device, image encoding method, image encoding program, transmission device, transmission method, and transmission program
JP6172324B2 (en) Image decoding apparatus, image decoding method, image decoding program, receiving apparatus, receiving method, and receiving program
JP2016167858A (en) Image decoder, image decoding method, image decoding program, receiver, receiving method and receiving program
JP6142943B2 (en) Image decoding apparatus, image decoding method, image decoding program, receiving apparatus, receiving method, and receiving program
JP6172327B2 (en) Image encoding device, image encoding method, image encoding program, transmission device, transmission method, and transmission program
JP6172328B2 (en) Image encoding device, image encoding method, image encoding program, transmission device, transmission method, and transmission program
KR20230108215A (en) Method for Decoder-side Motion Vector List Modification in Inter Prediction

Legal Events

Date Code Title Description
121 Ep: the epo has been informed by wipo that ep was designated in this application

Ref document number: 13778276

Country of ref document: EP

Kind code of ref document: A1

NENP Non-entry into the national phase

Ref country code: DE

122 Ep: pct application non-entry in european phase

Ref document number: 13778276

Country of ref document: EP

Kind code of ref document: A1