WO2013153823A1 - 動画像符号化装置、動画像符号化方法、動画像符号化プログラム、送信装置、送信方法及び送信プログラム、並びに動画像復号装置、動画像復号方法、動画像復号プログラム、受信装置、受信方法及び受信プログラム - Google Patents
動画像符号化装置、動画像符号化方法、動画像符号化プログラム、送信装置、送信方法及び送信プログラム、並びに動画像復号装置、動画像復号方法、動画像復号プログラム、受信装置、受信方法及び受信プログラム Download PDFInfo
- Publication number
- WO2013153823A1 WO2013153823A1 PCT/JP2013/002513 JP2013002513W WO2013153823A1 WO 2013153823 A1 WO2013153823 A1 WO 2013153823A1 JP 2013002513 W JP2013002513 W JP 2013002513W WO 2013153823 A1 WO2013153823 A1 WO 2013153823A1
- Authority
- WO
- WIPO (PCT)
- Prior art keywords
- inter prediction
- prediction information
- candidates
- inter
- candidate list
- Prior art date
Links
- 238000000034 method Methods 0.000 title claims description 316
- 230000005540 biological transmission Effects 0.000 title claims description 16
- 239000013598 vector Substances 0.000 claims abstract description 286
- 239000013589 supplement Substances 0.000 claims abstract description 27
- 230000009469 supplementation Effects 0.000 claims abstract description 11
- 238000012545 processing Methods 0.000 claims description 133
- 238000009795 derivation Methods 0.000 claims description 119
- 238000010276 construction Methods 0.000 claims description 114
- 230000001502 supplementing effect Effects 0.000 claims description 36
- 238000005192 partition Methods 0.000 description 93
- 230000002123 temporal effect Effects 0.000 description 69
- 238000010586 diagram Methods 0.000 description 25
- 238000004364 calculation method Methods 0.000 description 21
- 238000013139 quantization Methods 0.000 description 21
- 230000006870 function Effects 0.000 description 8
- 238000001514 detection method Methods 0.000 description 7
- 230000007274 generation of a signal involved in cell-cell signaling Effects 0.000 description 5
- 238000005259 measurement Methods 0.000 description 4
- 230000011218 segmentation Effects 0.000 description 3
- 238000000926 separation method Methods 0.000 description 3
- 239000000872 buffer Substances 0.000 description 2
- 238000006243 chemical reaction Methods 0.000 description 2
- 238000004891 communication Methods 0.000 description 2
- 238000011156 evaluation Methods 0.000 description 2
- 238000001914 filtration Methods 0.000 description 2
- 238000012986 modification Methods 0.000 description 2
- 230000004048 modification Effects 0.000 description 2
- 230000000295 complement effect Effects 0.000 description 1
- 230000006835 compression Effects 0.000 description 1
- 238000007906 compression Methods 0.000 description 1
- 238000004590 computer program Methods 0.000 description 1
- 239000000470 constituent Substances 0.000 description 1
- 230000003111 delayed effect Effects 0.000 description 1
- 230000000694 effects Effects 0.000 description 1
- 238000005516 engineering process Methods 0.000 description 1
- 238000004088 simulation Methods 0.000 description 1
Images
Classifications
-
- H—ELECTRICITY
- H04—ELECTRIC COMMUNICATION TECHNIQUE
- H04N—PICTORIAL COMMUNICATION, e.g. TELEVISION
- H04N19/00—Methods or arrangements for coding, decoding, compressing or decompressing digital video signals
- H04N19/10—Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using adaptive coding
- H04N19/102—Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using adaptive coding characterised by the element, parameter or selection affected or controlled by the adaptive coding
- H04N19/103—Selection of coding mode or of prediction mode
- H04N19/105—Selection of the reference unit for prediction within a chosen coding or prediction mode, e.g. adaptive choice of position and number of pixels used for prediction
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06T—IMAGE DATA PROCESSING OR GENERATION, IN GENERAL
- G06T7/00—Image analysis
- G06T7/20—Analysis of motion
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06T—IMAGE DATA PROCESSING OR GENERATION, IN GENERAL
- G06T9/00—Image coding
-
- H—ELECTRICITY
- H04—ELECTRIC COMMUNICATION TECHNIQUE
- H04N—PICTORIAL COMMUNICATION, e.g. TELEVISION
- H04N19/00—Methods or arrangements for coding, decoding, compressing or decompressing digital video signals
- H04N19/10—Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using adaptive coding
- H04N19/102—Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using adaptive coding characterised by the element, parameter or selection affected or controlled by the adaptive coding
- H04N19/119—Adaptive subdivision aspects, e.g. subdivision of a picture into rectangular or non-rectangular coding blocks
-
- H—ELECTRICITY
- H04—ELECTRIC COMMUNICATION TECHNIQUE
- H04N—PICTORIAL COMMUNICATION, e.g. TELEVISION
- H04N19/00—Methods or arrangements for coding, decoding, compressing or decompressing digital video signals
- H04N19/10—Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using adaptive coding
- H04N19/134—Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using adaptive coding characterised by the element, parameter or criterion affecting or controlling the adaptive coding
- H04N19/157—Assigned coding mode, i.e. the coding mode being predefined or preselected to be further used for selection of another element or parameter
- H04N19/159—Prediction type, e.g. intra-frame, inter-frame or bidirectional frame prediction
-
- H—ELECTRICITY
- H04—ELECTRIC COMMUNICATION TECHNIQUE
- H04N—PICTORIAL COMMUNICATION, e.g. TELEVISION
- H04N19/00—Methods or arrangements for coding, decoding, compressing or decompressing digital video signals
- H04N19/10—Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using adaptive coding
- H04N19/169—Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using adaptive coding characterised by the coding unit, i.e. the structural portion or semantic portion of the video signal being the object or the subject of the adaptive coding
- H04N19/17—Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using adaptive coding characterised by the coding unit, i.e. the structural portion or semantic portion of the video signal being the object or the subject of the adaptive coding the unit being an image region, e.g. an object
- H04N19/172—Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using adaptive coding characterised by the coding unit, i.e. the structural portion or semantic portion of the video signal being the object or the subject of the adaptive coding the unit being an image region, e.g. an object the region being a picture, frame or field
-
- H—ELECTRICITY
- H04—ELECTRIC COMMUNICATION TECHNIQUE
- H04N—PICTORIAL COMMUNICATION, e.g. TELEVISION
- H04N19/00—Methods or arrangements for coding, decoding, compressing or decompressing digital video signals
- H04N19/10—Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using adaptive coding
- H04N19/169—Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using adaptive coding characterised by the coding unit, i.e. the structural portion or semantic portion of the video signal being the object or the subject of the adaptive coding
- H04N19/17—Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using adaptive coding characterised by the coding unit, i.e. the structural portion or semantic portion of the video signal being the object or the subject of the adaptive coding the unit being an image region, e.g. an object
- H04N19/174—Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using adaptive coding characterised by the coding unit, i.e. the structural portion or semantic portion of the video signal being the object or the subject of the adaptive coding the unit being an image region, e.g. an object the region being a slice, e.g. a line of blocks or a group of blocks
-
- H—ELECTRICITY
- H04—ELECTRIC COMMUNICATION TECHNIQUE
- H04N—PICTORIAL COMMUNICATION, e.g. TELEVISION
- H04N19/00—Methods or arrangements for coding, decoding, compressing or decompressing digital video signals
- H04N19/10—Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using adaptive coding
- H04N19/189—Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using adaptive coding characterised by the adaptation method, adaptation tool or adaptation type used for the adaptive coding
- H04N19/196—Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using adaptive coding characterised by the adaptation method, adaptation tool or adaptation type used for the adaptive coding being specially adapted for the computation of encoding parameters, e.g. by averaging previously computed encoding parameters
- H04N19/197—Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using adaptive coding characterised by the adaptation method, adaptation tool or adaptation type used for the adaptive coding being specially adapted for the computation of encoding parameters, e.g. by averaging previously computed encoding parameters including determination of the initial value of an encoding parameter
-
- H—ELECTRICITY
- H04—ELECTRIC COMMUNICATION TECHNIQUE
- H04N—PICTORIAL COMMUNICATION, e.g. TELEVISION
- H04N19/00—Methods or arrangements for coding, decoding, compressing or decompressing digital video signals
- H04N19/46—Embedding additional information in the video signal during the compression process
- H04N19/463—Embedding additional information in the video signal during the compression process by compressing encoding parameters before transmission
-
- H—ELECTRICITY
- H04—ELECTRIC COMMUNICATION TECHNIQUE
- H04N—PICTORIAL COMMUNICATION, e.g. TELEVISION
- H04N19/00—Methods or arrangements for coding, decoding, compressing or decompressing digital video signals
- H04N19/50—Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using predictive coding
-
- H—ELECTRICITY
- H04—ELECTRIC COMMUNICATION TECHNIQUE
- H04N—PICTORIAL COMMUNICATION, e.g. TELEVISION
- H04N19/00—Methods or arrangements for coding, decoding, compressing or decompressing digital video signals
- H04N19/50—Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using predictive coding
- H04N19/503—Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using predictive coding involving temporal prediction
- H04N19/51—Motion estimation or motion compensation
- H04N19/513—Processing of motion vectors
-
- H—ELECTRICITY
- H04—ELECTRIC COMMUNICATION TECHNIQUE
- H04N—PICTORIAL COMMUNICATION, e.g. TELEVISION
- H04N19/00—Methods or arrangements for coding, decoding, compressing or decompressing digital video signals
- H04N19/50—Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using predictive coding
- H04N19/503—Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using predictive coding involving temporal prediction
- H04N19/51—Motion estimation or motion compensation
- H04N19/513—Processing of motion vectors
- H04N19/517—Processing of motion vectors by encoding
- H04N19/52—Processing of motion vectors by encoding by predictive encoding
-
- H—ELECTRICITY
- H04—ELECTRIC COMMUNICATION TECHNIQUE
- H04N—PICTORIAL COMMUNICATION, e.g. TELEVISION
- H04N19/00—Methods or arrangements for coding, decoding, compressing or decompressing digital video signals
- H04N19/50—Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using predictive coding
- H04N19/503—Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using predictive coding involving temporal prediction
- H04N19/51—Motion estimation or motion compensation
- H04N19/573—Motion compensation with multiple frame prediction using two or more reference frames in a given prediction direction
-
- H—ELECTRICITY
- H04—ELECTRIC COMMUNICATION TECHNIQUE
- H04N—PICTORIAL COMMUNICATION, e.g. TELEVISION
- H04N19/00—Methods or arrangements for coding, decoding, compressing or decompressing digital video signals
- H04N19/70—Methods or arrangements for coding, decoding, compressing or decompressing digital video signals characterised by syntax aspects related to video coding, e.g. related to compression standards
-
- H—ELECTRICITY
- H04—ELECTRIC COMMUNICATION TECHNIQUE
- H04N—PICTORIAL COMMUNICATION, e.g. TELEVISION
- H04N19/00—Methods or arrangements for coding, decoding, compressing or decompressing digital video signals
- H04N19/10—Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using adaptive coding
- H04N19/134—Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using adaptive coding characterised by the element, parameter or criterion affecting or controlling the adaptive coding
- H04N19/136—Incoming video signal characteristics or properties
- H04N19/137—Motion inside a coding unit, e.g. average field, frame or block difference
- H04N19/139—Analysis of motion vectors, e.g. their magnitude, direction, variance or reliability
-
- H—ELECTRICITY
- H04—ELECTRIC COMMUNICATION TECHNIQUE
- H04N—PICTORIAL COMMUNICATION, e.g. TELEVISION
- H04N19/00—Methods or arrangements for coding, decoding, compressing or decompressing digital video signals
- H04N19/10—Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using adaptive coding
- H04N19/169—Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using adaptive coding characterised by the coding unit, i.e. the structural portion or semantic portion of the video signal being the object or the subject of the adaptive coding
- H04N19/17—Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using adaptive coding characterised by the coding unit, i.e. the structural portion or semantic portion of the video signal being the object or the subject of the adaptive coding the unit being an image region, e.g. an object
- H04N19/176—Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using adaptive coding characterised by the coding unit, i.e. the structural portion or semantic portion of the video signal being the object or the subject of the adaptive coding the unit being an image region, e.g. an object the region being a block, e.g. a macroblock
-
- H—ELECTRICITY
- H04—ELECTRIC COMMUNICATION TECHNIQUE
- H04N—PICTORIAL COMMUNICATION, e.g. TELEVISION
- H04N19/00—Methods or arrangements for coding, decoding, compressing or decompressing digital video signals
- H04N19/10—Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using adaptive coding
- H04N19/169—Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using adaptive coding characterised by the coding unit, i.e. the structural portion or semantic portion of the video signal being the object or the subject of the adaptive coding
- H04N19/184—Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using adaptive coding characterised by the coding unit, i.e. the structural portion or semantic portion of the video signal being the object or the subject of the adaptive coding the unit being bits, e.g. of the compressed video stream
Definitions
- the present invention relates to a moving image encoding and decoding technique, and more particularly to a moving image encoding and decoding technique using motion compensated prediction.
- MPEG-4 AVC / H.3 is a typical video compression encoding system.
- H.264 Motion compensation is used in which a picture is divided into a plurality of rectangular blocks, a picture that has already been encoded / decoded is used as a reference picture, and motion from the reference picture is predicted. This method of predicting motion by motion compensation is called inter prediction or motion compensated prediction.
- MPEG-4 AVC / H. In the inter prediction in H.264, a plurality of pictures can be used as reference pictures, and the most suitable reference picture is selected for each block from the plurality of reference pictures to perform motion compensation. Therefore, a reference index is assigned to each reference picture, and the reference picture is specified by this reference index.
- L0 prediction mainly used as forward prediction
- L1 prediction list 1 prediction
- bi-prediction using two inter predictions of L0 prediction and L1 prediction at the same time is also defined.
- bi-directional prediction is performed, the inter-predicted signals of L0 prediction and L1 prediction are multiplied by a weighting coefficient, an offset value is added and superimposed, and a final inter-predicted image signal is obtained.
- weighting coefficients and offset values used for weighted prediction representative values are set for each reference picture in each list and encoded.
- the encoding information related to inter prediction includes, for each block, a prediction mode for distinguishing between L0 prediction and L1 prediction and bi-prediction, a reference index for specifying a reference picture for each reference list for each block, and a moving direction and a moving amount of the block.
- MPEG-4 AVC / H. H.264 defines a direct mode for generating inter prediction information of a block to be encoded / decoded from inter prediction information of an encoded / decoded block.
- the direct mode encoding of inter prediction information is not necessary, so that encoding efficiency is improved.
- a picture in which the reference index of L1 is registered as 0 is defined as a reference picture colPic.
- a block at the same position as the encoding / decoding target block in the reference picture colPic is set as a reference block.
- the L0 motion vector of the reference block is set as the reference motion vector mvCol, and the reference block is not encoded using the L0 prediction and is encoded using the L1 prediction. If this is the case, the L1 motion vector of the reference block is set as the reference motion vector mvCol.
- the picture referred to by the reference motion vector mvCol is the L0 reference picture in the temporal direct mode
- the reference picture colPic is the L1 reference picture in the temporal direct mode.
- the L0 motion vector mvL0 and the L1 motion vector mvL1 in the temporal direct mode are derived from the reference motion vector mvCol by scaling calculation processing.
- the inter-picture distance td is derived by subtracting the POC of the L0 reference picture in the temporal direct mode from the POC of the base picture colPic.
- POC is a variable associated with a picture to be encoded, and is set to a value that increases by 1 in the picture output / display order. The difference in POC between two pictures indicates the inter-picture distance in the time axis direction.
- td POC of base picture colPic-POC of L0 reference picture in temporal direct mode
- the inter-picture distance tb is derived by subtracting the POC of the L0 reference picture in the temporal direct mode from the POC of the picture to be encoded / decoded.
- tb POC of picture to be encoded / decoded-POC of reference picture of L0 in temporal direct mode
- the motion vector mvL0 of L0 in the temporal direct mode is derived from the reference motion vector mvCol by scaling calculation processing.
- mvL0 tb / td * mvCol
- the motion vector mvL1 of L1 is derived by subtracting the reference motion vector mvCol from the motion vector mvL0 of L0 in the temporal direct mode.
- mvL1 mvL0-mvCol
- the present inventors have come to recognize the necessity of further compressing the encoded information and reducing the overall code amount in the moving image encoding method using motion compensated prediction.
- the present invention has been made in view of such a situation, and an object of the present invention is to derive a prediction information candidate used for motion compensation prediction according to the situation, thereby reducing the amount of coding information. It is an object of the present invention to provide a moving image encoding and decoding technique that improves efficiency.
- a video encoding device is a video encoding device that encodes the video using motion compensated prediction in units of blocks obtained by dividing each picture of the video.
- a prediction information encoding unit (110) that encodes information indicating the number of designated inter prediction information candidates, a prediction block adjacent to a prediction block to be encoded, or a prediction block to be encoded;
- a prediction information deriving unit (104) for deriving the inter prediction information candidate from inter prediction information of a prediction block existing at the same position or in the vicinity of the prediction block to be encoded in pictures that have been encoded differently in time;
- a candidate list construction unit (130) for constructing an inter prediction information candidate list from the derived inter prediction information candidates, and the constructed I When the number of inter prediction information candidates included in the inter prediction information candidate list is less than the number of specified inter prediction information candidates, the number of inter prediction information candidates included in the inter prediction information candidate list is specified.
- This apparatus is a moving picture encoding apparatus that encodes the moving picture using motion compensation prediction in units of blocks obtained by dividing each picture of the moving picture, and includes information indicating the number of designated inter prediction information candidates.
- the number of inter prediction information candidates included in the further added inter prediction information candidate list is the number of designated inter prediction information candidates.
- a candidate supplementing unit (135) for deriving inter prediction information candidates having predetermined values for each of the prediction mode, the reference index, and the motion vector, and further adding to the further added inter prediction information candidate list Motion compensation for selecting one inter prediction information candidate from the inter prediction information candidates included in the inter prediction information candidate list, and performing inter prediction on the prediction block to be encoded using the selected inter prediction information candidate A prediction unit (105).
- Still another embodiment of the present invention is also a moving image encoding device.
- This apparatus is a moving picture encoding apparatus that encodes a coded bit string obtained by coding the moving picture using motion compensated prediction in units of blocks obtained by dividing each picture of the moving picture.
- a prediction information encoding unit (110) that encodes information indicating the number of information candidates, and a prediction block adjacent to the prediction block to be encoded, or an encoded picture that is temporally different from the prediction block to be encoded
- a prediction information deriving unit (104) for deriving the inter prediction information candidate from the inter prediction information of the prediction block existing at the same position or in the vicinity of the encoding target prediction block, and the derived inter prediction information candidate
- a candidate list construction unit (130) for constructing an inter prediction information candidate list from the above, and included in the constructed inter prediction information candidate list If the number of inter prediction information candidates is less than the number of designated inter prediction information candidates, an inter prediction information candidate having a predetermined value for each of the prediction mode, the reference index, and the motion vector is derived, When the number of inter prediction information candidates included in the added inter prediction information candidate list is less than the number of designated inter prediction information candidates, the predetermined inter prediction information candidate list is added.
- inter-prediction information candidates having predetermined values for the prediction mode, the reference index, and the motion vector are derived, and further added to the further added inter-prediction information candidate list (135) ) And one inter prediction information candidate from the inter prediction information candidates included in the inter prediction information candidate list, and inter prediction of the prediction block to be encoded is performed using the selected inter prediction information candidate.
- a motion compensation prediction unit (105) is a motion compensation prediction unit (105).
- Still another embodiment of the present invention is also a moving image encoding device.
- This apparatus is a moving picture coding apparatus for coding the moving picture using motion compensated prediction in units of blocks obtained by dividing each picture of a moving picture, and stores a specified number of candidates for inter prediction information.
- inter prediction information candidate list to be stored inter prediction information candidates having the predetermined number of prediction modes, reference indexes and motion vectors for the designated number of candidates are stored in advance and initialized, and then the prediction block to be encoded
- the inter prediction information is obtained from inter prediction information of a prediction block that is present at or near the same position as the prediction block of the encoding target in an adjacent prediction block or a coded picture that is temporally different from the prediction block of the encoding target.
- a prediction information deriving unit (104) for deriving a candidate for the inter prediction information candidate included in the inter prediction information candidate list Select candidates for one inter prediction information from over prediction information candidates, and a motion compensation prediction unit (105) for performing inter prediction of the selected prediction block of the encoding target by the candidate inter prediction information.
- Still another embodiment of the present invention is also a moving image encoding device.
- This apparatus is a moving picture encoding apparatus that encodes the moving picture using motion compensated prediction in units of blocks obtained by dividing each picture of the moving picture, and indicates information indicating the number of designated inter prediction information candidates.
- a prediction information encoding unit (110) for encoding, and a prediction block adjacent to the prediction block to be encoded, or the prediction block to be encoded in an encoded picture that is temporally different from the prediction block to be encoded A prediction information deriving unit (104) for deriving inter prediction information candidates based on the number of candidates specified as the number of inter prediction information candidates from the inter prediction information of a prediction block existing at or near the same position, A candidate list construction unit (130) for constructing an inter prediction information candidate list from the derived inter prediction information candidates, and the designated When the complement is 1 or more, one inter prediction information candidate is selected from the inter prediction information candidates included in the inter prediction information candidate list, and the prediction block to be encoded is selected based on the selected inter prediction information candidate. And a motion compensated prediction unit (105) that performs inter prediction of the prediction block to be encoded by using a predetermined value of inter prediction information when the designated number of candidates is zero.
- Still another embodiment of the present invention is also a moving image encoding device.
- This apparatus is a moving picture encoding apparatus that encodes the moving picture using motion compensated prediction in units of blocks obtained by dividing each picture of a moving picture, and a prediction block adjacent to a prediction block to be encoded, or Prediction for deriving a candidate for inter prediction information from inter prediction information of a prediction block existing at the same position or in the vicinity of the prediction block of the encoding target in a coded picture that is temporally different from the prediction block of the encoding target.
- the prediction mode, the reference index, and the motion vector value of the same inter prediction information are the same until the specified number of candidates is reached.
- a candidate supplementing unit (135) for supplementing candidates and selecting one inter prediction information candidate from the inter prediction information candidates.
- a motion compensation prediction unit (105) for performing inter prediction of the prediction block of the encoding target by the candidate of the selected inter prediction information.
- Still another aspect of the present invention is a video encoding method.
- This method is a moving image coding method for coding the moving image using motion compensated prediction in units of blocks obtained by dividing each picture of the moving image, and includes information indicating the number of designated inter prediction information candidates.
- Prediction information encoding step to be encoded and the same position as the prediction block to be encoded in a prediction block adjacent to the prediction block to be encoded or an encoded picture temporally different from the prediction block to be encoded
- a prediction information deriving step for deriving the inter prediction information candidate from inter prediction information of a prediction block existing in the vicinity thereof, and a candidate list building step for constructing an inter prediction information candidate list from the derived inter prediction information candidate;
- the number of inter prediction information candidates included in the constructed inter prediction information candidate list is the index.
- the prediction mode When the number of inter prediction information candidates included in the inter prediction information candidate list reaches the designated number of inter prediction information candidates, the prediction mode, the reference index, And a candidate supplement step for deriving an inter prediction information candidate having a predetermined value for each of the motion vectors and adding it to the constructed inter prediction information candidate list, and from the inter prediction information candidates included in the inter prediction information candidate list
- Still another aspect of the present invention is also a moving image encoding method.
- This method is a moving image encoding method for encoding the moving image using motion compensated prediction in units of blocks obtained by dividing each picture of the moving image, and includes information indicating the number of designated inter prediction information candidates.
- the number of prediction information to be encoded is the same as the prediction block to be encoded in the prediction block adjacent to the prediction block to be encoded, or in the encoded picture that is temporally different from the prediction block to be encoded
- Prediction information deriving step for deriving candidates for the inter prediction information from inter prediction information of a prediction block existing at or near the position, and candidate list construction for building an inter prediction information candidate list from the derived inter prediction information candidates Step and the number of inter prediction information candidates included in the constructed inter prediction information candidate list If it is less than the number of designated inter prediction information candidates, an inter prediction information candidate having a predetermined value for each of the prediction mode, the reference index, and the motion vector is derived, and the constructed inter prediction information candidate list When the number of inter prediction information candidates included in the added inter prediction information candidate list is less than the number of designated inter prediction information candidates, the inter prediction information candidates having the predetermined value
- Still another aspect of the present invention is also a moving image encoding method.
- This method is a moving picture coding method for coding a coded bit string in which the moving picture is coded by using motion compensated prediction in units of blocks obtained by dividing each picture of the moving picture, and the specified inter prediction
- a prediction information number encoding step for encoding information indicating the number of information candidates, and a prediction block adjacent to the prediction block to be encoded, or an encoded picture that is temporally different from the prediction block to be encoded.
- a prediction information deriving step for deriving the inter prediction information candidate from the inter prediction information of a prediction block existing at the same position as or near the encoding target prediction block, and an inter prediction information candidate from the derived inter prediction information candidate A candidate list construction step of constructing a list, and an interface included in the constructed inter prediction information candidate list.
- the predetermined value is One or a plurality of inter prediction information candidates having the same value for the prediction mode and the motion vector with respect to the inter prediction information candidate having, and having the value changed for the reference index, and the added inter prediction information Further added to the candidate list, the inter prediction information candidate list added to the further added inter prediction information candidate list. If the number of measurement information candidates is less than the number of designated inter prediction information candidates, the number of inter prediction information candidates included in the further added inter prediction information candidate list is the number of specified inter prediction information candidates.
- Still another aspect of the present invention is also a moving image encoding method.
- This method is a moving picture coding method for coding the moving picture using motion compensated prediction in units of blocks obtained by dividing each picture of the moving picture, and stores a specified number of candidates for inter prediction information.
- the inter prediction information candidate list the inter prediction information candidates having the predetermined number of prediction modes, reference indexes, and motion vectors for the designated number of candidates are stored in advance and initialized, and then the prediction block to be encoded is stored.
- inter prediction information From inter prediction information of an adjacent prediction block or a prediction block existing at the same position or in the vicinity of the prediction block to be encoded in an encoded picture that is temporally different from the prediction block to be encoded, inter prediction information
- a prediction information deriving step for deriving a candidate for the inter prediction, and an inter prediction information candidate list Select candidates for one inter prediction information from the measurement information candidates, and a motion compensated prediction step of performing the inter prediction of the selected prediction block of the encoding target by the candidate inter prediction information.
- Still another aspect of the present invention is also a moving image encoding method.
- This method is a moving image coding method for coding the moving image using motion compensated prediction in units of blocks obtained by dividing each picture of the moving image, and includes information indicating the number of designated inter prediction information candidates.
- Still another aspect of the present invention is a transmission device.
- This apparatus packetizes an encoded bit string encoded by a moving image encoding method that encodes the moving image using motion compensated prediction in units of blocks obtained by dividing each picture of the moving image to obtain an encoded stream.
- the moving image encoding method includes: a prediction information number encoding step for encoding information indicating the number of designated inter prediction information candidates; a prediction block adjacent to a prediction block to be encoded; A prediction information deriving step for deriving candidates for the inter prediction information from inter prediction information of a prediction block existing at the same position or in the vicinity of the prediction block to be encoded in a coded picture that is temporally different from the block; A candidate list construction step of constructing an inter prediction information candidate list from the derived inter prediction information candidates, and the number of inter prediction information candidates included in the constructed inter prediction information candidate list is the designated inter prediction information If less than the number of candidates, prediction mode, reference index, and motion vector Inter prediction information candidates having a predetermined value for each of them, are added to the constructed inter prediction information candidate list, and the number of inter prediction information candidates included in the added inter prediction information candidate list is If the number of specified inter prediction information candidates is less than the number of inter prediction information candidates, one or a plurality of prediction modes, reference indices, and motion vectors
- a candidate addition step of deriving inter prediction information candidates and further adding the candidate to the added inter prediction information candidate list, and the number of inter prediction information candidates included in the further added inter prediction information candidate list is specified.
- the number of inter prediction information candidates is less than the number of inter prediction information candidates, Until the number of inter prediction information candidates included in the target reaches the number of designated inter prediction information candidates, deriving inter prediction information candidates having predetermined values for each of the prediction mode, the reference index, and the motion vector,
- a candidate supplement step for further adding to the further added inter prediction information candidate list, selecting one inter prediction information candidate from the inter prediction information candidates included in the inter prediction information candidate list, and selecting the selected inter prediction A motion compensated prediction step of performing inter prediction of the prediction block to be encoded according to information candidates.
- Still another aspect of the present invention is a transmission method.
- an encoded stream is obtained by packetizing an encoded bit sequence encoded by a moving image encoding method that encodes the moving image using motion compensated prediction in units of blocks obtained by dividing each picture of a moving image.
- the moving image encoding method includes: a prediction information number encoding step for encoding information indicating the number of designated inter prediction information candidates; a prediction block adjacent to a prediction block to be encoded; A prediction information deriving step for deriving candidates for the inter prediction information from inter prediction information of a prediction block existing at the same position or in the vicinity of the prediction block to be encoded in a coded picture that is temporally different from the block; A candidate list construction step of constructing an inter prediction information candidate list from the derived inter prediction information candidates, and the number of inter prediction information candidates included in the constructed inter prediction information candidate list is the designated inter prediction information If less than the number of candidates, prediction mode, reference index, and motion vector Inter prediction information candidates having a predetermined value for each of them, are added to the constructed inter prediction information candidate list, and the number of inter prediction information candidates included in the added inter prediction information candidate list is If the number of specified inter prediction information candidates is less than the number of inter prediction information candidates, one or a plurality of prediction modes, reference indices, and motion vectors
- a candidate addition step of deriving inter prediction information candidates and further adding the candidate to the added inter prediction information candidate list, and the number of inter prediction information candidates included in the further added inter prediction information candidate list is specified.
- the number of inter prediction information candidates is less than the number of inter prediction information candidates, Until the number of inter prediction information candidates included in the target reaches the number of designated inter prediction information candidates, deriving inter prediction information candidates having predetermined values for each of the prediction mode, the reference index, and the motion vector,
- a candidate supplement step for further adding to the further added inter prediction information candidate list, selecting one inter prediction information candidate from the inter prediction information candidates included in the inter prediction information candidate list, and selecting the selected inter prediction A motion compensated prediction step of performing inter prediction of the prediction block to be encoded according to information candidates.
- a moving picture decoding apparatus is a moving picture decoding apparatus that decodes a coded bit string obtained by coding the moving picture using motion compensation prediction in units of blocks obtained by dividing each picture of the moving picture.
- a prediction information decoding unit (202) that decodes information indicating the number of inter prediction information candidates designated in advance, and a prediction block adjacent to the prediction block to be decoded or a decoding block that is temporally different from the prediction block to be decoded
- a prediction information deriving unit (205) for deriving the inter prediction information candidate from the inter prediction information of a prediction block existing at the same position or in the vicinity of the decoding target prediction block in the current picture, and the derived inter prediction information
- a candidate list construction unit (230) for constructing an inter prediction information candidate list from candidates, and the constructed inter prediction information candidate When the number of inter prediction information candidates included in the list is less than the number of inter prediction information candidates specified in advance, the number of inter prediction information candidates included in the inter prediction information candidate list is determined as A candidate supplementing unit that derives
- This apparatus is a moving picture decoding apparatus that decodes a coded bit string in which the moving picture is coded using motion compensated prediction in units of blocks obtained by dividing each picture of the moving picture, and includes pre-specified inter prediction information
- a prediction information decoding unit (202) for decoding information indicating the number of candidates, and the prediction of the decoding target in a prediction block adjacent to the decoding target prediction block or a decoded picture temporally different from the prediction block of the decoding target
- a prediction information deriving unit (205) for deriving the inter prediction information candidate from the inter prediction information of the prediction block existing at the same position as or near the block, and an inter prediction information candidate list from the derived inter prediction information candidate
- a candidate list construction unit (230) to construct, and an inter prediction information candidate list constructed in the constructed inter prediction information candidate list
- an inter prediction information candidate having a predetermined value for each of the prediction mode, the reference index
- Still another aspect of the present invention is also a video decoding device.
- This apparatus is a moving picture decoding apparatus that decodes a coded bit string in which the moving picture is coded using motion compensated prediction in units of blocks obtained by dividing each picture of the moving picture, and includes pre-specified inter prediction information
- a prediction information decoding unit (202) for decoding information indicating the number of candidates, and the prediction of the decoding target in a prediction block adjacent to the decoding target prediction block or a decoded picture temporally different from the prediction block of the decoding target
- a prediction information deriving unit (205) for deriving the inter prediction information candidate from the inter prediction information of the prediction block existing at the same position as or near the block, and an inter prediction information candidate list from the derived inter prediction information candidate
- a candidate list construction unit (230) to construct, and an inter prediction information candidate list constructed in the constructed inter prediction information candidate list
- the number of inter prediction information candidates included in the further added inter prediction information candidate list is determined as the inter prediction information specified in advance.
- Still another aspect of the present invention is also a video decoding device.
- This apparatus is a moving picture decoding apparatus that decodes a coded bit sequence in which the moving picture is coded using motion compensated prediction in units of blocks obtained by dividing each picture of the moving picture, and inter prediction information is designated.
- the candidate for inter prediction information having a motion vector is stored in advance and initialized, and then the decoding target in a prediction block adjacent to the prediction block to be decoded or a decoded picture temporally different from the prediction block to be decoded
- Still another aspect of the present invention is also a video decoding device.
- This apparatus is a moving picture encoding apparatus that encodes the moving picture using motion compensated prediction in units of blocks obtained by dividing each picture of the moving picture, and indicates information indicating the number of inter prediction information candidates designated in advance.
- Still another aspect of the present invention is also a video decoding device.
- This apparatus is a moving picture decoding apparatus that decodes a coded bit string obtained by coding the moving picture using motion compensated prediction in units of blocks obtained by dividing each picture of the moving picture, and is adjacent to a prediction block to be decoded.
- a candidate for inter prediction information is derived from inter prediction information of a prediction block existing at the same position or in the vicinity of the prediction block to be decoded in a decoded picture temporally different from the prediction block to be decoded or the prediction block to be decoded And when the number of inter prediction information candidates is less than the designated number of candidates, the inter prediction with the same prediction mode, reference index, and motion vector value until the designated number of candidates is reached.
- Prediction information decoding step for decoding information indicating the number of candidates, the same as the prediction block to be decoded in a prediction block adjacent to the prediction block to be decoded or a decoded picture temporally different from the prediction block to be decoded
- a prediction information deriving step for deriving the inter prediction information candidate from inter prediction information of a prediction block existing at or near the position, and a candidate list building step for constructing an inter prediction information candidate list from the derived inter prediction information candidate And inter prediction information candidates included in the constructed inter prediction information candidate list If the number is less than the number of inter prediction information candidates specified in advance, until the number of inter prediction information candidates included in the inter prediction information candidate list reaches the number of inter prediction information candidates specified in advance,
- a motion compensation prediction step of selecting an inter prediction information candidate
- Still another aspect of the present invention is also a moving image decoding method.
- This method is a moving picture decoding method for decoding a coded bit string in which the moving picture is coded using motion compensated prediction in units of blocks obtained by dividing each picture of the moving picture, and includes pre-designated inter prediction information.
- a prediction information number decoding step for decoding information indicating the number of candidates, a prediction block adjacent to the prediction block to be decoded, or the prediction block to be decoded in a decoded picture temporally different from the prediction block to be decoded;
- Prediction information deriving step for deriving the inter prediction information candidate from inter prediction information of a prediction block existing at or near the same position, and candidate list construction for constructing an inter prediction information candidate list from the derived inter prediction information candidate
- inter prediction information candidates included in the constructed inter prediction information candidate list Is less than the number of inter prediction information candidates designated in advance, an inter prediction information candidate having a predetermined value for each of the prediction mode, the reference index, and the motion vector is derived, and the constructed inter prediction information candidate is derived.
- the predetermined value is included.
- One or a plurality of inter prediction information candidates in which at least one of the prediction mode, the reference index, and the motion vector is changed with respect to the inter prediction information candidates are derived and further added to the added inter prediction information candidate list.
- the predetermined value is included.
- One or a plurality of inter prediction information candidates having the same value for the prediction mode and the motion vector with respect to the inter prediction information candidate, and having a changed value for the reference index, are derived, and the added inter prediction information candidate Inter predictions added to the list and included in the additional inter prediction information candidate list.
- the number of information prediction candidates is less than the number of inter prediction information candidates designated in advance
- the number of inter prediction information candidates included in the further added inter prediction information candidate list is the number of inter prediction information candidates designated in advance.
- a prediction information deriving step for deriving complements From the inter prediction information of the prediction block that exists at or near the same position as the block, A prediction information deriving step for deriving complements, selecting one inter prediction information candidate from the inter prediction information candidates included in the inter prediction information candidate list, and predicting the decoding target based on the selected inter prediction information candidate
- a motion compensation prediction step for performing inter prediction of the block From the inter prediction information of the prediction block that exists at or near the same position as the block, A prediction information deriving step for deriving complements, selecting one inter prediction information candidate from the inter prediction information candidates included in the inter prediction information candidate list, and predicting the decoding target based on the selected inter prediction information candidate.
- a candidate list construction step of constructing an inter prediction information candidate list from candidates, and the number of candidates specified in advance is In the above process, one inter prediction information candidate is selected from the inter prediction information candidates included in the inter prediction information candidate list, and the inter prediction of the prediction block to be encoded is selected based on the selected inter prediction information candidate.
- Still another aspect of the present invention is a receiving device.
- This apparatus is a receiving apparatus that receives and decodes a coded bit sequence in which a moving image is encoded, and the moving image is encoded using motion compensated prediction in units of blocks obtained by dividing each picture of the moving image.
- a receiving unit that receives an encoded stream in which the encoded bit string is packetized, a restoration unit that packet-processes the received encoded stream and restores the original encoded bit string, and the restored encoded bit string
- a prediction information decoding unit (202) that decodes information indicating the number of inter prediction information candidates designated in advance, and a temporally different decoding from a prediction block adjacent to a decoding target prediction block or a decoding target prediction block From the inter prediction information of a prediction block existing at the same position as or near the prediction block to be decoded in a completed picture, the inter prediction is performed.
- the number of inter prediction information candidates included in the information candidate list is less than the number of inter prediction information candidates specified in advance
- the number of inter prediction information candidates included in the further added inter prediction information candidate list Until the number of inter prediction information candidates specified in advance is reached, inter prediction information candidates having predetermined values for each of the prediction mode, the reference index, and the motion vector are derived, and the further added inter prediction information candidate list
- a candidate supplementing unit (235) to be added and one candidate from the inter prediction information candidates A motion compensation prediction unit (206) that selects an inter prediction information candidate and performs inter prediction on the prediction block to be decoded by the selected inter prediction information candidate.
- a reception step of receiving an encoded stream in which the encoded bit sequence is packetized, a restoration step of performing packet processing on the received encoded stream to restore the original encoded bit sequence, and the restored encoded bit sequence A prediction information number decoding step for decoding information indicating the number of inter prediction information candidates designated in advance, and a prediction block adjacent to the decoding target prediction block or a decoding target that is temporally different from the decoding target prediction block From the inter prediction information of a prediction block existing at the same position as or near the prediction block to be decoded in a picture, A prediction information derivation step for deriving an inter prediction information candidate, a candidate list construction step for constructing an inter prediction information candidate list from the derived inter prediction information candidates, and an inter prediction included in the constructed inter prediction information candidate list When the number of information candidates is less than the number of inter prediction information candidates designated in advance, an inter prediction information candidate having a predetermined value for each of the prediction mode, reference index, and motion vector is derived and constructed The predetermined value when the number of inter prediction information candidates included in
- the number of inter prediction information candidates included in the list is less than the number of inter prediction information candidates designated in advance
- the number of inter prediction information candidates included in the further added inter prediction information candidate list is designated in advance.
- inter prediction information candidates having predetermined values for the prediction mode, the reference index, and the motion vector are derived, and further added to the further added inter prediction information candidate list A candidate replenishment step, and one inter prediction from the inter prediction information candidates.
- the present invention by deriving prediction information candidates used for motion compensation prediction according to the situation, it is possible to reduce the amount of generated code of encoded information to be transmitted and improve the encoding efficiency.
- FIG. 8 is a block diagram showing a detailed configuration of an inter prediction information deriving unit of the moving picture coding apparatus of FIG. 1 in the second to seventh embodiments.
- FIG. 8 is a block diagram showing a detailed configuration of an inter prediction information deriving unit of the moving picture coding apparatus of FIG. 1 in the second to seventh embodiments.
- 10 is a block diagram illustrating a detailed configuration of an inter prediction information deriving unit of the video decoding device in FIG. 2 according to the second to seventh embodiments.
- 15 is a flowchart for explaining merge candidate derivation processing and merge candidate list construction processing procedures in merge modes according to the second to seventh embodiments. It is a flowchart explaining the effective merge candidate supplement processing procedure of the merge mode of 2nd Example. It is a flowchart explaining the effective merge candidate supplement processing procedure of the merge mode of 3rd Example. It is a flowchart explaining the effective merge candidate supplement processing procedure of the merge mode of 4th Example. It is a flowchart explaining the additional merge candidate derivation
- the present embodiment with regard to moving picture coding, in particular, to improve coding efficiency in moving picture coding in which a picture is divided into rectangular blocks of an arbitrary size and shape and motion compensation is performed in units of blocks between pictures.
- a plurality of predicted motion vectors are derived from the motion vectors of the block adjacent to the encoding target block or the block of the encoded picture, and the difference vector between the motion vector of the encoding target block and the selected prediction motion vector
- the amount of code is reduced by calculating and encoding.
- the coding amount is reduced by deriving the coding information of the coding target block by using the coding information of the block adjacent to the coding target block or the block of the coded picture.
- a plurality of predicted motion vectors are calculated from the motion vectors of a block adjacent to the decoding target block or a decoded picture block, and selected from the difference vector decoded from the encoded stream
- the motion vector of the decoding target block is calculated from the predicted motion vector and decoded.
- the encoding information of the decoding target block is derived by using the encoding information of the block adjacent to the decoding target block or the block of the decoded picture.
- the size of the tree block can be freely set to a power of 2 depending on the picture size and the texture in the picture.
- the tree block divides the luminance signal and chrominance signal in the tree block hierarchically into four parts (two parts vertically and horizontally) as necessary, The block can be made smaller in block size.
- Each block is defined as a coding block, and is a basic unit of processing when performing coding and decoding. Except for monochrome, the coding block is also composed of one luminance signal and two color difference signals.
- the maximum size of the coding block is the same as the size of the tree block.
- An encoded block having the minimum size of the encoded block is called a minimum encoded block, and can be freely set to a power of 2.
- the coding block A is a single coding block without dividing the tree block.
- the encoding block B is an encoding block formed by dividing a tree block into four.
- the coding block C is a coding block obtained by further dividing the block obtained by dividing the tree block into four.
- the coding block D is a coding block obtained by further dividing the block obtained by dividing the tree block into four parts and further dividing the block into four twice hierarchically, and is a coding block of the minimum size.
- (About prediction mode) Encoded / decoded in a picture to be encoded / decoded in units of coding blocks (in the encoding process, the encoded signal is used for a decoded picture, a prediction block, an image signal, etc., and the decoded picture is used in the decoding process) Used for prediction blocks, image signals, etc.
- Intra prediction in which prediction is performed from surrounding image signals, and image signals of encoded / decoded pictures
- Switch inter prediction (MODE_INTER) to perform prediction.
- a mode for identifying the intra prediction (MODE_INTRA) and the inter prediction (MODE_INTER) is defined as a prediction mode (PredMode).
- the prediction mode (PredMode) has intra prediction (MODE_INTRA) or inter prediction (MODE_INTER) as a value, and can be selected and encoded.
- PartMode A mode for identifying the division method of the luminance signal and the color difference signal of the coding block is defined as a division mode (PartMode). Furthermore, this divided block is defined as a prediction block. As shown in FIG. 4, eight types of partition modes (PartMode) are defined according to the method of dividing the luminance signal of the coding block.
- PartMode The partition mode (PartMode) that is regarded as one prediction block without dividing the luminance signal of the coding block shown in FIG. 4A is defined as 2N ⁇ 2N partition (PART_2Nx2N).
- the division modes (PartMode) for dividing the luminance signal of the coding block shown in FIGS. 4B, 4C, and 4D into two prediction blocks arranged vertically are 2N ⁇ N division (PART_2NxN) and 2N ⁇ nU, respectively. It is defined as division (PART_2NxnU) and 2N ⁇ nD division (PART_2NxnD).
- 2N ⁇ N division (PART_2NxN) is a division mode divided up and down at a ratio of 1: 1
- 2N ⁇ nU division (PART_2NxnU) is a division mode divided up and down at a ratio of 1: 3 and 2N ⁇
- the nD division (PART_2NxnD) is a division mode in which division is performed at a ratio of 3: 1 up and down.
- the division modes (PartMode) for dividing the luminance signals of the coding blocks shown in FIGS. 4 (e), (f), and (g) into two prediction blocks arranged on the left and right are divided into N ⁇ 2N divisions (PART_Nx2N) and nL ⁇ 2N, respectively.
- N ⁇ 2N division (PART_Nx2N) is a division mode in which left and right are divided at a ratio of 1: 1
- nL ⁇ 2N division (PART_nLx2N) is a division mode in which division is performed at a ratio of 1: 3 in the left and right
- nR ⁇ 2N division (PART_nRx2N) is a division mode in which the image is divided in the ratio of 3: 1 to the left and right.
- the division mode (PartMode) in which the luminance signal of the coding block shown in FIG. 4 (h) is divided into four parts in the vertical and horizontal directions and defined as four prediction blocks is defined as N ⁇ N division (PART_NxN).
- the color difference signal is also divided in the same manner as the vertical and horizontal division ratios of the luminance signal for each division mode (PartMode).
- a number starting from 0 is assigned to the prediction block existing inside the coding block in the coding order. This number is defined as a split index PartIdx.
- a number described in each prediction block of the encoded block in FIG. 4 represents a partition index PartIdx of the prediction block.
- the partition index PartIdx of the upper prediction block is set to 0.
- the partition index PartIdx of the lower prediction block is set to 1.
- the partition mode (PartMode) is 2N ⁇ 2N partition (PART_2Nx2N), 2N ⁇ N partition (PART_2NxN), 2N ⁇ nU partition (PART_2NxnU), 2N ⁇ nD partition (PART_2NxnD) , N ⁇ 2N partition (PART_Nx2N), nL ⁇ 2N partition (PART_nLx2N), and nR ⁇ 2N partition (PART_nRx2N).
- PartMode 2N ⁇ 2N partition
- PART_2Nx2N 2N ⁇ N partition
- PART_2NxN 2N ⁇ nU partition
- PART_2NxnU 2N ⁇ nD partition
- N ⁇ N division (PART_NxN) can be defined. It is assumed that the partition mode (PartMode) does not define N ⁇ N partition (PART_NxN).
- the partition mode (PartMode) When the prediction mode (PredMode) is intra prediction (MODE_INTRA), only the 2N ⁇ 2N partition (PART_2Nx2N) is defined as the partition mode (PartMode) except for the encoding block D which is the minimum encoding block, and the minimum encoding block
- the partition mode (PartMode) defines N ⁇ N partition (PART_NxN) in addition to 2N ⁇ 2N partition (PART_2Nx2N).
- N ⁇ N division (PART_NxN) is not defined other than the smallest coding block is that, except for the smallest coding block, the coding block can be divided into four to represent a small coding block.
- the position of each block including the tree block, the encoding block, the prediction block, and the transform block according to the present embodiment has the position of the pixel of the luminance signal at the upper left of the luminance signal screen as the origin (0, 0).
- the pixel position of the upper left luminance signal included in each block area is represented by two-dimensional coordinates (x, y).
- the direction of the coordinate axis is a right direction in the horizontal direction and a downward direction in the vertical direction, respectively, and the unit is one pixel unit of the luminance signal.
- the luminance signal and the color difference signal have the same image size (number of pixels) and the color difference format is 4: 4: 4.
- the position of each block of the color difference signal is represented by the coordinates of the pixel of the luminance signal included in the block area, and the unit is one pixel of the luminance signal. In this way, not only can the position of each block of the color difference signal be specified, but also the relationship between the positions of the luminance signal block and the color difference signal block can be clarified only by comparing the coordinate values.
- Inter prediction mode reference list
- a plurality of decoded pictures can be used as reference pictures in inter prediction in which prediction is performed from image signals of encoded / decoded pictures.
- a reference index is attached to each prediction block.
- any two reference pictures can be selected for each prediction block, and inter prediction can be performed.
- inter prediction modes there are L0 prediction (Pred_L0), L1 prediction (Pred_L1), and bi-prediction (Pred_BI).
- the reference picture is managed by L0 (reference list 0) and L1 (reference list 1) of the list structure, and the reference picture can be specified by specifying the reference index of L0 or L1.
- L0 prediction is inter prediction that refers to a reference picture managed in L0
- L1 prediction is inter prediction that refers to a reference picture managed in L1
- bi-prediction is This is inter prediction in which both L0 prediction and L1 prediction are performed and one reference picture managed by each of L0 and L1 is referred to. Only L0 prediction can be used in inter prediction of P slice, and L0 prediction, L1 prediction, and bi-prediction (Pred_BI) that averages or weights and adds L0 prediction and L1 prediction can be used in inter prediction of B slice. In the subsequent processing, it is assumed that the constants and variables with the subscript LX in the output are processed for each of L0 and L1.
- the merge mode does not encode / decode inter prediction information such as a prediction mode, a reference index, and a motion vector of a prediction block to be encoded / decoded, but within the same picture as a prediction block to be encoded / decoded.
- inter prediction is performed by deriving inter prediction information of a prediction block to be encoded / decoded from inter prediction information of a prediction block existing in the vicinity (neighboring position).
- the prediction block adjacent to the prediction block to be encoded / decoded in the same picture as the prediction block to be encoded / decoded and inter prediction information of the prediction block are spatial merge candidates, the prediction block to be encoded / decoded, and the time.
- Prediction information derived from a prediction block existing at the same position as or near (previously near) a prediction block to be encoded / decoded of a differently encoded / decoded picture and inter prediction information of the prediction block Are time merge candidates.
- Each merge candidate is registered in the merge candidate list, and the merge candidate used in the inter prediction is specified by the merge index.
- FIG. 9 shows a picture that has been encoded / decoded that is temporally different from the prediction block to be encoded / decoded that is referred to when the temporal merge candidate is derived, at the same position as or near the prediction block to be encoded / decoded. It is a figure explaining the prediction block which has already been encoded / decoded.
- a prediction block adjacent in the spatial direction of a prediction block to be encoded / decoded and a prediction block at the same position at different times will be described with reference to FIGS. 5, 6, 7, 8, and 9.
- a prediction block A adjacent to the left side of the prediction block to be encoded / decoded within the same picture as the prediction block to be encoded / decoded, a prediction block B adjacent to the upper side, A prediction block C adjacent to the upper right vertex, a prediction block D adjacent to the lower left vertex, and a prediction block E adjacent to the upper left vertex are defined as prediction blocks adjacent in the spatial direction.
- this embodiment when the size of the prediction block adjacent to the left side of the prediction block to be encoded / decoded is smaller than the prediction block to be encoded / decoded, and there are a plurality of prediction blocks, this embodiment In the embodiment, only the lowest prediction block A10 among the prediction blocks adjacent to the left side is set as the prediction block A adjacent to the left side.
- the left side in the present embodiment Only the rightmost prediction block B10 among the prediction blocks adjacent to is set as the prediction block B adjacent to the upper side.
- the prediction block F adjacent to the left side of the prediction block to be encoded / decoded is larger than the prediction block to be encoded / decoded, it is adjacent to the left side according to the above condition. If the prediction block F is adjacent to the left side of the prediction block to be encoded / decoded, the prediction block A is determined. If the prediction block F is adjacent to the lower left vertex of the prediction block to be encoded / decoded, the prediction block D is determined. If it is adjacent to the top left vertex of the prediction block to be encoded / decoded, the prediction block E is determined. In the example of FIG. 7, the prediction block A, the prediction block D, and the prediction block E are the same prediction block.
- the prediction block G adjacent to the upper side of the prediction block to be encoded / decoded is larger than the prediction block to be encoded / decoded, it is adjacent to the upper side according to the above condition.
- the prediction block B is determined.
- the prediction block G is adjacent to the upper right vertex of the prediction block to be encoded / decoded
- the prediction block C is determined.
- the prediction block E is determined. In the example of FIG. 8, the prediction block B, the prediction block C, and the prediction block E are the same prediction block.
- the decoded prediction blocks T0 and T1 are defined as prediction blocks at the same position at different times.
- (About POC) POC is a variable associated with the picture to be encoded, and is set to a value that increases by 1 in the picture output / display order. Based on the POC value, it is possible to determine whether they are the same picture, to determine the front-to-back relationship between pictures in the output / display order, or to derive the distance between pictures. For example, if the POCs of two pictures have the same value, it can be determined that they are the same picture. When the POCs of two pictures have different values, it can be determined that the picture with the smaller POC value is the picture that is output / displayed first, and the difference between the POCs of the two pictures is the time axis direction difference between the pictures. Indicates distance.
- FIG. 1 is a block diagram showing a configuration of a moving picture coding apparatus according to an embodiment of the present invention.
- the moving image encoding apparatus includes an image memory 101, a header information setting unit 117, a motion vector detection unit 102, a difference motion vector calculation unit 103, an inter prediction information derivation unit 104, a motion compensation prediction unit 105, and an intra prediction unit.
- a multiplexing unit 112 an inverse quantization / inverse orthogonal transform unit 113, a decoded image signal superimposing unit 114, an encoded information storage memory 115, and a decoded image memory 116 are provided.
- the header information setting unit 117 sets information in units of sequences, pictures, and slices.
- the set sequence, picture, and slice unit information is supplied to the inter prediction information deriving unit 104 and the first encoded bit string generating unit 118, and is also supplied to all blocks (not shown).
- the maximum merge candidate number maxNumMergeCand described later is also set by the header information setting unit 117.
- the image memory 101 temporarily stores the image signal of the encoding target picture supplied in the order of shooting / display time.
- the image memory 101 supplies the stored image signal of the picture to be encoded to the motion vector detection unit 102, the prediction method determination unit 107, and the residual signal generation unit 108 in units of predetermined pixel blocks.
- the image signals of the pictures stored in the order of shooting / display time are rearranged in the encoding order and output from the image memory 101 in units of pixel blocks.
- the difference motion vector calculation unit 103 calculates a plurality of motion vector predictor candidates by using the encoded information of the already encoded image signal stored in the encoded information storage memory 115, and generates a prediction motion vector list.
- the optimum motion vector predictor is selected from a plurality of motion vector predictor candidates registered and registered in the motion vector predictor list, and a motion vector difference is calculated from the motion vector detected by the motion vector detector 102 and the motion vector predictor. Then, the calculated difference motion vector is supplied to the prediction method determination unit 107.
- a prediction motion vector index that identifies a prediction motion vector selected from prediction motion vector candidates registered in the prediction motion vector list is supplied to the prediction method determination unit 107.
- the inter prediction information deriving unit 104 derives merge candidates in the merge mode.
- a plurality of merge candidates are derived using the encoding information of the already encoded prediction block stored in the encoding information storage memory 115 and registered in a merge candidate list described later, and registered in the merge candidate list Flags predFlagL0 [xP] [yP], predFlagL1 [xP indicating whether or not to use the L0 prediction and L1 prediction of each prediction block of the selected merge candidate from among a plurality of merge candidates ] [yP], reference index refIdxL0 [xP] [yP], refIdxL1 [xP] [yP], motion vector mvL0 [xP] [yP], mvL1 [xP] [yP], etc.
- a merge index for specifying the selected merge candidate is supplied to the prediction method determination unit 107.
- xP and yP are indexes indicating the position of the upper left pixel of the prediction block in the picture. The detailed configuration and operation of the inter prediction information deriving unit 104 will be described later.
- the motion compensated prediction unit 105 generates a predicted image signal by inter prediction (motion compensated prediction) from the reference picture using the motion vectors detected by the motion vector detection unit 102 and the inter prediction information deriving unit 104, and generates the predicted image signal. This is supplied to the prediction method determination unit 107.
- inter prediction motion compensated prediction
- L0 prediction and L1 prediction one-way prediction is performed.
- Pred_BI bi-prediction
- bi-directional prediction is performed, the inter-predicted signals of L0 prediction and L1 prediction are adaptively multiplied by weighting factors, offset values are added and superimposed, and finally A predicted image signal is generated.
- the intra prediction unit 106 performs intra prediction for each intra prediction mode.
- a prediction image signal is generated by intra prediction from the decoded image signal stored in the decoded image memory 116, a suitable intra prediction mode is selected from a plurality of intra prediction modes, the selected intra prediction mode, and A prediction image signal corresponding to the selected intra prediction mode is supplied to the prediction method determination unit 107.
- the prediction method determination unit 107 evaluates the encoding information and the code amount of the residual signal for each prediction method, the distortion amount between the prediction image signal and the image signal, and the like from among a plurality of prediction methods.
- the prediction mode PredMode and split mode PartMode are determined to determine whether inter prediction (PRED_INTER) or intra prediction (PRED_INTRA) is optimal for each coding block.
- inter prediction PRED_INTER
- the prediction method determination unit 107 stores information indicating the determined prediction method and encoded information including a motion vector corresponding to the determined prediction method in the encoded information storage memory 115.
- the encoding information stored here is a flag predFlagL0 [xP] [yP], predFlagL1 [that indicates whether to use the prediction mode PredMode, the partition mode PartMode, the L0 prediction of each prediction block, and the L1 prediction of each prediction block.
- xP and yP are indexes indicating the position of the upper left pixel of the prediction block in the picture.
- PredMode is intra prediction (MODE_INTRA)
- a flag predFlagL0 [xP] [yP] indicating whether to use L0 prediction and a flag predFlagL1 [xP] [yP] indicating whether to use L1 prediction are Both are zero.
- the flag predFlagL0 [xP] [yP] indicating whether or not to use L0 prediction uses 1, L1 prediction.
- the flag predFlagL1 [xP] [yP] indicating whether or not is zero.
- the flag predFlagL0 [xP] [yP] indicating whether to use L0 prediction is 0, and the flag predFlagL1 [xP] [yP] indicating whether to use L1 prediction is 1.
- the prediction method determination unit 107 supplies a prediction image signal corresponding to the determined prediction mode to the residual signal generation unit 108 and the decoded image signal superimposition unit 114.
- the residual signal generation unit 108 generates a residual signal by performing subtraction between the image signal to be encoded and the predicted image signal, and supplies the residual signal to the orthogonal transform / quantization unit 109.
- the orthogonal transform / quantization unit 109 performs orthogonal transform and quantization on the residual signal according to the quantization parameter to generate an orthogonal transform / quantized residual signal, and a third encoded bit string generation unit 111 And supplied to the inverse quantization / inverse orthogonal transform unit 113. Further, the orthogonal transform / quantization unit 109 stores the quantization parameter in the encoded information storage memory 115.
- the first encoded bit string generation unit 118 encodes the sequence, picture, and slice unit information set by the header information setting unit 117.
- a first encoded bit string is generated and supplied to the multiplexing unit 112.
- the maximum number of merge candidates maxNumMergeCand described later is also encoded by the first encoded bit string generation unit 118.
- the second encoded bit string generation unit 110 encodes encoding information corresponding to the prediction method determined by the prediction method determination unit 107 for each encoded block and prediction block. Specifically, in the case of prediction mode PredMode, split mode PartMode, and intra prediction (PRED_INTRA) that determine whether the prediction mode is skip mode for each coding block, inter prediction (PRED_INTER) or intra prediction (PRED_INTRA), For intra prediction mode and inter prediction (PRED_INTER), a flag that determines whether or not it is merge mode, merge index if merge mode, inter prediction mode, prediction motion vector index, information on difference motion vector, etc.
- the encoded information is encoded according to a prescribed syntax rule described later to generate a second encoded bit string, and the encoded information is supplied to the multiplexing unit 112.
- the prediction block PredMode value of the prediction block is inter prediction (MODE_INTER)
- the merge mode merge_flag [x0] [y0] is 1)
- the partition mode PartMode is 2N ⁇ 2N partition (PART_2Nx2N).
- the third encoded bit string generation unit 111 entropy-encodes the residual signal that has been orthogonally transformed and quantized according to a specified syntax rule to generate a third encoded bit string, and supplies the third encoded bit string to the multiplexing unit 112.
- the multiplexing unit 112 multiplexes the first encoded bit string, the second encoded bit string, and the third encoded bit string in accordance with a specified syntax rule, and outputs a bit stream.
- FIG. 2 is a block diagram showing the configuration of the moving picture decoding apparatus according to the embodiment of the present invention corresponding to the moving picture encoding apparatus of FIG.
- the moving picture decoding apparatus includes a separation unit 201, a first encoded bit string decoding unit 212, a second encoded bit string decoding unit 202, a third encoded bit string decoding unit 203, a motion vector calculation unit 204, and inter prediction information.
- a derivation unit 205, a motion compensation prediction unit 206, an intra prediction unit 207, an inverse quantization / inverse orthogonal transform unit 208, a decoded image signal superimposing unit 209, an encoded information storage memory 210, and a decoded image memory 211 are provided.
- the decoding process of the moving picture decoding apparatus in FIG. 2 corresponds to the decoding process provided in the moving picture encoding apparatus in FIG. 1, so the motion compensation prediction unit 206 in FIG.
- the configuration of the inverse orthogonal transform unit 208, the decoded image signal superimposing unit 209, the encoded information storage memory 210, and the decoded image memory 211 is the same as that of the motion compensation prediction unit 105, the inverse quantization / inverse of the moving image encoding device in FIG.
- the orthogonal transform unit 113, the decoded image signal superimposing unit 114, the encoded information storage memory 115, and the decoded image memory 116 have functions corresponding to the respective configurations.
- the bit stream supplied to the separation unit 201 is separated according to a rule of a prescribed syntax, and the separated encoded bit string is a first encoded bit string decoding unit 212, a second encoded bit string decoding unit 202, and a third encoded bit string.
- the data is supplied to the decoding unit 203.
- the first encoded bit string decoding unit 212 decodes the supplied encoded bit string to obtain information in units of sequences, pictures, and slices. Although the obtained sequence, picture, and slice unit information is not shown, it is supplied to all blocks. The maximum number of merge candidates maxNumMergeCand described later is also decoded by the first encoded bit string decoding unit 212.
- the second encoded bit string decoding unit 202 decodes the supplied encoded bit string to obtain encoded block unit information and predicted block unit encoded information. Specifically, in the case of prediction mode PredMode, split mode PartMode, and inter prediction (PRED_INTER) to determine whether it is skip mode for each coding block, inter prediction (PRED_INTER) or intra prediction (PRED_INTRA), A flag for determining whether or not the mode is a merge mode. In the case of the merge mode, the encoding information is decoded according to a specified syntax rule to be described later.
- the encoded information such as the decoded prediction mode PredMode and split mode PartMode is stored in the encoded information storage memory 210 and supplied to the motion vector calculation unit 204, the inter prediction information derivation unit 205, or the intra prediction unit 207.
- the prediction block PredMode value of the prediction block is inter prediction (MODE_INTER)
- the merge mode merge_flag [x0] [y0] is 1)
- the partition mode PartMode
- PartMode is 2N ⁇ 2N partition (PART_2Nx2N).
- the third encoded bit string decoding unit 203 calculates a residual signal that has been orthogonally transformed / quantized by decoding the supplied encoded bit string, and dequantized / inverted the residual signal that has been orthogonally transformed / quantized. This is supplied to the orthogonal transform unit 208.
- the motion vector calculation unit 204 stores the encoded information of the already decoded image signal stored in the encoded information storage memory 210.
- the encoding information of the prediction block to be supplied / stored here is a reference index of flags predFlagL0 [xP] [yP], predFlagL1 [xP] [yP], L0, L1 indicating whether to use L0 prediction and L1 prediction refIdxL0 [xP] [yP], RefIdxL1 [xP] [yP], L0, L1 motion vectors mvL0 [xP] [yP], mvL1 [xP] [yP], and the like.
- xP and yP are indexes indicating the position of the upper left pixel of the prediction block in the picture.
- the flag predFlagL0 indicating whether to use L0 prediction is 1, and the flag predFlagL1 indicating whether to use L1 prediction is 0 It is.
- the inter prediction mode is L1 prediction (Pred_L1)
- the flag predFlagL0 indicating whether to use L0 prediction is 0, and the flag predFlagL1 indicating whether to use L1 prediction is 1.
- the inter prediction mode is bi-prediction (Pred_BI)
- a flag predFlagL0 indicating whether to use L0 prediction and a flag predFlagL1 indicating whether to use L1 prediction are both 1.
- the inter prediction information deriving unit 205 derives merge candidates when the prediction mode PredMode of the prediction block to be decoded is inter prediction (PRED_INTER) and in the merge mode.
- a plurality of merge candidates are derived and registered in a merge candidate list, which will be described later, using the encoded information of already decoded prediction blocks stored in the encoded information storage memory 115, and registered in the merge candidate list Whether a merge candidate corresponding to a merge index decoded and supplied by the second encoded bit string decoding unit 202 is selected from among a plurality of merge candidates, and whether to use the L0 prediction and the L1 prediction of the selected merge candidate Flags predFlagL0 [xP] [yP], predFlagL1 [xP] [yP], L0, L1 reference indices refIdxL0 [xP] [yP], refIdxL1 [xP] [yP], L0, L1 motion vectors mvL0 [xP] Inter prediction information such as [y
- the motion compensation prediction unit 206 performs prediction by inter prediction (motion compensation prediction) from the reference picture stored in the decoded image memory 211 using the inter prediction information calculated by the motion vector calculation unit 204 or the inter prediction information deriving unit 205. An image signal is generated, and the predicted image signal is supplied to the decoded image signal superimposing unit 209.
- Pred_BI bi-prediction
- a weighted coefficient is adaptively multiplied and superimposed as necessary on the two motion compensated prediction image signals of L0 prediction and L1 prediction to generate a final prediction image signal.
- the inverse quantization / inverse orthogonal transform unit 208 performs inverse orthogonal transform and inverse quantization on the orthogonal transform / quantized residual signal decoded by the second encoded bit string decoding unit 202, and performs inverse orthogonal transform / An inverse quantized residual signal is obtained.
- the decoded image signal superimposing unit 209 performs the prediction image signal inter-predicted by the motion compensation prediction unit 206 or the prediction image signal intra-predicted by the intra prediction unit 207 and the inverse orthogonal transform by the inverse quantization / inverse orthogonal transform unit 208.
- the decoded image signal is decoded by superimposing the dequantized residual signal, and stored in the decoded image memory 211.
- the decoded image may be stored in the decoded image memory 211 after filtering processing that reduces block distortion or the like due to encoding is performed on the decoded image.
- the header information setting unit 117 sets the maximum number of merge candidates maxNumMergeCand registered in the merge candidate list mergeCandList in sequence units, picture units, or slice units, and uses the first number of the moving picture coding apparatus as syntax elements. It is encoded by one encoded bit string generation unit 118 and decoded by a first encoded bit string decoding unit 212 of the video decoding device. It is assumed that a value from 0 to 5 can be set in the maximum merge candidate number maxNumMergeCand, and a small value is set in the maximum merge candidate number maxNumMergeCand mainly when reducing the processing amount of the moving picture coding apparatus. When 0 is set to the maximum merge candidate number maxNumMergeCand, the prescribed inter prediction information is used as a merge candidate. In the description of the present embodiment, the maximum merge candidate number maxNumMergeCand is assumed to be 5.
- merge_flag [x0] [y0] When merge_flag [x0] [y0] is 1, this indicates merge mode, and when the value of the maximum number of merge candidates maxNumMergeCand is greater than 1, the syntax element of the index of the merge list that is a list of merge candidates to be referenced merge_idx [x0] [y0] is installed.
- skip_flag [x0] [y0] this indicates skip mode, and when the value of the maximum merge candidate number maxNumMergeCand is larger than 1, the syntax element of the index of the merge list that is a list of merge candidates to be referred to merge_idx [x0] [y0] is installed.
- merge_flag [x0] [y0] indicating whether it is merge mode is set.
- x0 and y0 are indices indicating the position of the upper left pixel of the prediction block in the picture of the luminance signal
- merge_flag [x0] [y0] is the prediction block located at (x0, y0) in the picture It is a flag indicating whether or not merge mode.
- merge_flag [x0] [y0] when merge_flag [x0] [y0] is 1, it indicates merge mode.
- the merge list index that is a list of merge candidates to be referred to
- the syntax element merge_idx [x0] [y0] is set.
- x0 and y0 are indices indicating the position of the upper left pixel of the prediction block in the picture
- merge_idx [x0] [y0] is the merge index of the prediction block located at (x0, y0) in the picture is there.
- FIG. 11 shows an example of the entropy code (code) of the merge index syntax element merge_idx [x0] [y0].
- code the code of the merge index syntax element merge_idx [x0] [y0] is “0” and “1”, respectively.
- the codes of the merge index syntax elements merge_idx [x0] [y0] are “0”, “10”, and “11”, respectively. .
- the codes of the merge index syntax elements merge_idx [x0] [y0] are '0', '10', and '110', respectively. , '111'.
- the signs of the merge index syntax elements merge_idx [x0] [y0] are '0', '10', ' 110 ',' 1110 ', and' 1111 '.
- the code amount of the merge index is reduced by switching the code indicating each value of the merge index in accordance with the number of merge candidates as shown in FIG.
- a merge index having a value equal to or greater than the maximum merge candidate number maxNumMergeCand is not encoded or decoded.
- the maximum merge candidate number maxNumMergeCand is 1, the merge index is not encoded / decoded and the merge index is 0. Further, when the maximum number of merge candidates is 0, the specified inter prediction information is used as a merge candidate, so a merge index is unnecessary.
- merge_flag [x0] [y0] when merge_flag [x0] [y0] is 0, it indicates that the mode is not merge mode.
- the slice type is B slice
- a syntax element inter_pred_flag [x0] [y0] for identifying the inter prediction mode is installed.
- the syntax element identifies L0 prediction (Pred_L0), L1 prediction (Pred_L1), and bi-prediction (Pred_BI).
- syntax elements ref_idx_l0 [x0] [y0] and ref_idx_l1 [x0] [y0] of the reference index for specifying the reference picture, the motion vector and prediction of the prediction block obtained by motion vector detection The differential motion vector syntax elements mvd_l0 [x0] [y0] [j] and mvd_l1 [x0] [y0] [j], which are the differences from the motion vector, are provided.
- x0 and y0 are indexes indicating the position of the upper left pixel of the prediction block in the picture
- ref_idx_l0 [x0] [y0] and mvd_l0 [x0] [y0] [j] are respectively (x0 , Y0) is the reference index of L0 of the prediction block and the differential motion vector
- ref_idx_l1 [x0] [y0] and mvd_l1 [x0] [y0] [j] are respectively (x0, y0) in the picture It is the reference index of L1 of the prediction block located, and a difference motion vector.
- j represents a differential motion vector component
- j represents 0 as an x component
- j represents 1 as a y component.
- syntax elements mvp_idx_l0 [x0] [y0] and mvp_idx_l1 [x0] [y0] of an index of a predicted motion vector list that is a list of predicted motion vector candidates to be referred to are set.
- x0 and y0 are indices indicating the position of the upper left pixel of the prediction block in the picture
- mvp_idx_l0 [x0] [y0] and mvp_idx_l1 [x0] [y0] are (x0, y0) in the picture It is the prediction motion vector index of L0 and L1 of the prediction block located.
- the value of the number of candidates is set to 2.
- the inter prediction information deriving method according to the embodiment is implemented in the inter prediction information deriving unit 104 of the video encoding device in FIG. 1 and the inter prediction information deriving unit 205 of the video decoding device in FIG.
- the inter prediction information deriving method is performed in any of the encoding and decoding processes for each prediction block constituting the encoding block.
- the prediction mode PredMode of the prediction block is inter prediction (MODE_INTER) and the merge mode includes the skip mode
- encoding is performed using the prediction mode, reference index, and motion vector of the encoded prediction block.
- the prediction mode, reference index, and motion vector of the target prediction block When deriving the prediction mode, reference index, and motion vector of the target prediction block, in the case of decoding, in the case of decoding, the prediction mode, reference index, and motion vector of the prediction block to be decoded using the prediction mode, reference index, and motion vector of the decoded prediction block This is performed when an index and a motion vector are derived.
- the merge mode includes the prediction block A adjacent to the left, the prediction block B adjacent to the upper side, the prediction block C adjacent to the upper right, and the prediction block D adjacent to the lower left described with reference to FIGS. 5, 6, 7, and 8.
- the prediction block Col either T0 or T1 existing at the same position at or near the different time described with reference to FIG. Derive merge candidates.
- the inter prediction information deriving unit 104 of the moving image encoding device and the inter prediction information deriving unit 205 of the moving image decoding device register these merge candidates in the merge candidate list in the order prescribed in common on the encoding side and the decoding side.
- the inter prediction information deriving unit 104 of the video encoding device determines a merge index for specifying an element of the merge candidate list, encodes it via the second encoded bit string generation unit 110, and performs inter prediction of the video decoding device.
- the information deriving unit 205 is supplied with the merge index decoded by the second encoded bit string decoding unit 202, selects a prediction block corresponding to the merge index from the merge candidate list, and selects the prediction mode of the selected merge candidate, Motion compensation prediction is performed using inter prediction information such as a reference index and a motion vector.
- FIG. 12 is a diagram illustrating a detailed configuration of the inter prediction information deriving unit 104 of the video encoding device in FIG. 1 of the first example according to the embodiment.
- FIG. 13 is a diagram illustrating a detailed configuration of the inter prediction information deriving unit 205 of the video decoding device in FIG. 2 of the first example according to the embodiment.
- the portions surrounded by the thick frame lines in FIGS. 12 and 13 indicate the inter prediction information deriving unit 104 and the inter prediction information deriving unit 205, respectively.
- the portions surrounded by the thick dotted lines inside them are the merge candidate list construction unit 120 of the moving picture coding apparatus that derives the respective merge candidates and constructs the merge candidate list, and the merge candidate list of the moving picture decoding apparatus.
- a construction unit 220 is shown, which is similarly installed in a moving picture decoding apparatus corresponding to the moving picture encoding apparatus of the embodiment, so that the same determination result that is consistent between encoding and decoding can be obtained.
- merge candidate derivation and merge candidate list construction processing in the merge candidate list construction unit 120 of the video encoding device and the merge candidate list construction unit 220 of the video decoding device Merging candidate derivation and merge candidate list construction processing of the prediction block to be processed are performed without referring to a prediction block included in the same encoding block as the encoding block including the prediction block to be processed.
- the coding block division mode (PartMode) is not 2N ⁇ 2N division (PART_2Nx2N), that is, when there are a plurality of prediction blocks in the coding block
- the coding side performs coding.
- the merge candidate derivation of each prediction block and the merge candidate list construction process can be performed in parallel in the conversion block.
- FIG. 14 is a diagram for explaining a prediction block adjacent to a prediction block to be processed for each division mode (PartMode) of a processing target coding block.
- A0, B0, C0, D0, and E0 are a prediction block A adjacent to the left side of the processing target prediction block whose partition index PartIdx is 0, a prediction block B adjacent to the upper side, and an apex on the upper right.
- a prediction block D adjacent to the lower left vertex, and a prediction block E adjacent to the upper left vertex, and A1, B1, C1, D1, and E1 are predictions to be processed with a partition index PartIdx of 1.
- Prediction block A adjacent to the left side of each block, prediction block B adjacent to the upper side, prediction block C adjacent to the upper right vertex, prediction block D adjacent to the lower left vertex, and adjacent to the upper left vertex Indicates the prediction block E, and A2, B2, C2, D2, and E2 are the left sides of the prediction block to be processed whose partition index PartIdx is 2, respectively.
- An adjacent prediction block A, a prediction block B adjacent to the upper side, a prediction block C adjacent to the upper right vertex, a prediction block D adjacent to the lower left vertex, and a prediction block E adjacent to the upper left vertex are shown, A3 , B3, C3, D3, and E3 are the prediction block A adjacent to the left side of the processing target prediction block having the partition index PartIdx of 3, the prediction block B adjacent to the upper side, and the prediction block adjacent to the upper right vertex, respectively.
- PartMode partition mode for dividing the processing target coding block into two prediction blocks arranged vertically is 2N ⁇ N partition (PART_2NxN), 2N ⁇ nU partition ( It is a figure which shows the prediction block which adjoins in PART_2NxnU) and 2NxnD division
- a prediction block B1 adjacent to the processing target prediction block with PartIdx of 1 is a prediction block with PartIdx of 0.
- the partition mode (PartMode) is 2N ⁇ N partition (PART_2NxN), 2N ⁇ nU partition (PART_2NxnU), 2N ⁇ nD partition (PART_2NxnD), and the prediction of the processing target
- the block's PartIdx is 1
- the prediction of the prediction block B1 that is the prediction block with the PartIdx of 0 is not referred to, and the merge candidate derivation of the prediction block of the PartIdx of 1 and the merge candidate list construction process are performed.
- the merge candidate derivation of the two prediction blocks and the merge candidate list construction process can be processed in parallel.
- a prediction block A1 adjacent to a processing target prediction block with PartIdx of 1 is a prediction block with PartIdx of 0.
- the partition mode (PartMode) is N ⁇ 2N partition (PART_Nx2N), nL ⁇ 2N partition (PART_nLx2N), and nR ⁇ 2N partition (PART_nRx2N),
- the prediction process A1 that is the prediction block with the PartIdx of 0 is not referred to, and the merge candidate derivation of the prediction block of the PartIdx of 1 and the merge candidate list construction process are performed.
- the merge candidate derivation of each prediction block and the merge candidate list construction process can be processed in parallel.
- the prediction mode is a prediction block in which the partition mode (PartMode) is N ⁇ N partition (PART_NxN), the processing target prediction block PartIdx is 1, and PartIdx is 0
- the merge candidate derivation and merge candidate list construction processing for each prediction block in the encoded block by performing merge candidate derivation and merge candidate list construction processing for a prediction block whose PartIdx is 1 without referring to the block A1 Can be processed in parallel.
- a prediction block B2 adjacent to a processing target prediction block with PartIdx of 2 is a prediction block with PartIdx of 0, and a prediction block C2 is a prediction block of PartIdx with 1. Therefore, in order to perform merge candidate derivation and merge candidate list construction processing for a prediction block whose PartIdx is 2 with reference to the prediction blocks B2 and C2, the PartIdx belonging to the same coding block that is the prediction blocks B2 and C2 is 0 and 1 After the prediction block merge candidate derivation and merge candidate list construction processing are completed and the merge candidates to be used are specified, the processing cannot be performed.
- the partition mode (PartMode) is N ⁇ N partition (PART_NxN)
- the prediction block to be processed is PartIdx of 2
- the partIdx is 0 and 1 prediction blocks.
- the prediction block E3 adjacent to the processing target prediction block whose PartIdx is 3 is a prediction block whose PartIdx is 0, the prediction block B3 is a prediction block whose PartIdx is 1, and the prediction block A3 is a prediction block whose PartIdx is 2. . Accordingly, in order to perform merge candidate derivation and merge candidate list construction processing for a prediction block with PartIdx of 3 with reference to the prediction blocks E3, B3, and A3, PartIdx belonging to the same encoded block that is the prediction blocks E3, B3, and A3 If the merge candidate derivation and merge candidate list construction process for the prediction blocks of 0, 1, and 2 are completed and the merge candidates to be used are specified, the process cannot be performed.
- the partition mode (PartMode) is N ⁇ N partition (PART_NxN)
- the PartIdx of the prediction block to be processed is 3
- the prediction of PartIdx is 0, 1 and 2
- the inter prediction information deriving unit 104 of the video encoding device in FIG. 12 includes a merge candidate list generating unit 130, a spatial merge candidate generating unit 131, a temporal merge candidate reference index deriving unit 132, a temporal merge candidate generating unit 133, and an additional merge.
- a candidate generation unit 134, a merge candidate restriction unit 136, and an inter prediction information selection unit 137 are included.
- the inter prediction information deriving unit 205 of the video decoding device in FIG. 13 includes a merge candidate list generating unit 230, a spatial merge candidate generating unit 231, a temporal merge candidate reference index deriving unit 232, a temporal merge candidate generating unit 233, and an additional merge candidate.
- a generation unit 234, a merge candidate restriction unit 236, and an inter prediction information selection unit 237 are included.
- FIG. 15 shows the merge candidate list construction unit 120 of the inter prediction information deriving unit 104 and the inter prediction information deriving unit 205 of the video decoding device of the video coding device according to the first embodiment of the present invention.
- 10 is a flowchart for explaining a procedure of merge candidate derivation processing and merge candidate list construction processing having functions common to the merge candidate list construction unit 220; Hereinafter, the processes will be described in order. In the following description, a case where the slice type slice_type is a B slice will be described unless otherwise specified, but the present invention can also be applied to a P slice.
- the slice type slice_type is a P slice
- the L0 prediction (Pred_L0) is provided as the inter prediction mode, and there is no L1 prediction (Pred_L1) and bi-prediction (Pred_BI). Therefore, the processing related to L1 can be omitted.
- the maximum merge candidate number maxNumMergeCand is 0, the merge candidate derivation process and the merge candidate list construction process in FIG. 15 are omitted. be able to.
- a merge candidate list mergeCandList is prepared in the merge candidate list generating unit 130 of the inter prediction information deriving unit 104 of the video encoding device and the merge candidate list generating unit 230 of the inter prediction information deriving unit 205 of the video decoding device ( Step S100 in FIG. 15).
- the merge candidate list mergeCandList has a list structure, and is provided with a merge index indicating the location in the merge candidate list and a storage area for storing merge candidates corresponding to the index as elements. The number of the merge index starts from 0, and merge candidates are stored in the storage area of the merge candidate list mergeCandList.
- a prediction block that is a merge candidate of the merge index i registered in the merge candidate list mergeCandList is represented by mergeCandList [i], and is distinguished from the merge candidate list mergeCandList by array notation.
- the merge candidate list mergeCandList can register at least five merge candidates (inter prediction information).
- a variable numMergeCand indicating the number of merge candidates registered in the merge candidate list mergeCandList is set to 0.
- the prepared merge candidate list mergeCandList is respectively supplied to the spatial merge candidate generation unit 131 of the inter prediction information deriving unit 104 of the video encoding device and the spatial merge candidate generation unit 231 of the inter prediction information deriving unit 205 of the video decoding device.
- the spatial merge candidate generation unit 131 of the inter prediction information deriving unit 104 of the video encoding device and the spatial merge candidate generation unit 231 of the inter prediction information deriving unit 205 of the video decoding device store encoding information of the video encoding device. Spatial merging from each prediction block A, B, C, D, E adjacent to the encoding / decoding target block from the encoding information stored in the memory 115 or the encoding information storage memory 210 of the video decoding device Candidates A, B, C, D, and E are derived, and the derived spatial merge candidates are registered in the merge candidate list mergeCandList (step S101 in FIG. 15).
- N indicating any one of A, B, C, D, E and the temporal merge candidate Col is defined.
- the flag predFlagL0N and the L1 prediction flag predFlagL1N indicating whether or not L1 prediction is performed, the L0 motion vector mvL0N, and the L1 motion vector mvL1N are derived.
- the merge candidate is derived without referring to the prediction block included in the same encoding block as the encoding block including the prediction block to be processed, so that the code including the prediction block to be processed Spatial merge candidates included in the same coding block as the coding block are not derived.
- the detailed processing procedure of step S101 will be described later in detail using the flowchart of FIG.
- the merge candidate list mergeCandList is supplied to the temporal merge candidate generation unit 133 of the inter prediction information deriving unit 104 of the video encoding device and the temporal merge candidate generation unit 233 of the inter prediction information deriving unit 205 of the video decoding device, respectively.
- the reference index of the temporal merge candidate is derived without referring to the prediction block included in the same encoded block as the encoded block including the prediction block to be processed.
- the inter prediction is performed using the inter prediction information of the temporal merge candidate when the slice type slice_type is P slice, only the L0 reference index is derived in order to perform the L0 prediction (Pred_L0), and the slice type slice_type is the B slice.
- Pred_BI bi-prediction
- temporal merge candidate generation unit 133 of the inter prediction information deriving unit 104 of the video encoding device and the temporal merge candidate generation unit 233 of the inter prediction information deriving unit 205 of the video decoding device time from pictures at different times A merge candidate is derived, and the derived time merge candidate is registered in the merge candidate list mergeCandList (step S103 in FIG. 15).
- the detailed processing procedure of step S103 will be described later in detail using the flowchart of FIG.
- the merge candidate list mergeCandList is supplied to the additional merge candidate generation unit 134 of the inter prediction information deriving unit 104 of the video encoding device and the additional merge candidate generation unit 234 of the inter prediction information deriving unit 205 of the video decoding device, respectively.
- the additional merge candidate generation unit 134 of the inter prediction information deriving unit 104 of the video encoding device and the additional merge candidate generation unit 234 of the inter prediction information deriving unit 205 of the video decoding device register the merge candidate list mergeCandList. If the number of merge candidates numMergeCand is smaller than the maximum merge candidate number maxNumMergeCand, the merge candidate number numMergeCand registered in the merge candidate list mergeCandList derives additional merge candidates up to the maximum merge candidate number maxNumMergeCand and merges Registration in the candidate list mergeCandList (step S104 in FIG. 15).
- the merge candidate list mergeCandList is supplied to the merge candidate restriction unit 136 of the inter prediction information deriving unit 104 of the video encoding device and the merge candidate restriction unit 236 of the inter prediction information deriving unit 205 of the video decoding device, respectively.
- the merge candidate list mergeCandList is supplied to the inter prediction information selection unit 137 of the inter prediction information derivation unit 104 of the video encoding device and the inter prediction information selection unit 237 of the inter prediction information derivation unit 205 of the video decoding device, respectively.
- the detailed processing procedure of step S106 is demonstrated using the flowchart of FIG.
- step S7101 If it is larger than the maximum merge candidate number maxNumMergeCand (YES in step S7101 in FIG. 27), the value of the merge candidate number numMergeCand registered in the merge candidate list mergeCandList is updated to the maximum merge candidate number maxNumMergeCand (step S7102 in FIG. 27). .
- the processing in step S7102 prohibits access to all merge candidates whose merge index in the merge candidate list mergeCandList is larger than (maxNumMergeCand-1), and sets the maximum number of merge candidates maxNumMergeCand to be registered in the merge candidate list mergeCandList. It means to limit to.
- the number of merge candidates registered in the merge candidate list mergeCandList in slice units is set to a fixed number.
- the reason why the number of merge candidates registered in the merge candidate list mergeCandList is fixed is that when the number of merge candidates registered in the merge candidate list mergeCandList fluctuates according to the build status of the merge candidate list, entropy decoding and the merge candidate list construction Dependency occurs, and on the decoding side, the merge candidate list must be constructed for each prediction block and the merge candidate number registered in the merge candidate list mergeCandList must be derived before entropy decoding of the merge index can be performed. This is because the decoding of the merge index is delayed and the entropy decoding becomes complicated.
- the current picture when an error occurs when decoding the encoded bit string of another picture The encoded bit string is also affected by the error, and the number of merge candidates registered in the normal merge candidate list mergeCandList cannot be derived, and entropy decoding cannot be continued normally.
- the number of merge candidates registered in the merge candidate list mergeCandList in slice units is set to a fixed number as in the present embodiment, the number of merge candidates registered in the merge candidate list mergeCandList in prediction block units is derived.
- FIG. 16 is a flowchart for explaining the spatial merge candidate derivation processing procedure in step S101 of FIG.
- N A (left side), B (upper side), C (upper right), D (lower left), or E (upper left) representing an area of an adjacent prediction block is entered.
- a maximum of four spatial merge candidates are derived from five adjacent prediction blocks.
- the encoding information of the prediction block A adjacent to the left side of the prediction block to be encoded / decoded is checked with the variable N as A, the merge candidate A is derived, and the prediction block adjacent to the upper side with the variable N as B
- the encoding information of B is examined to derive a merge candidate B, the variable N is set as C, the encoding information of the prediction block C adjacent on the upper right side is checked, the merge candidate C is derived, and the variable N is set as D to the lower left side.
- the encoding information of the adjacent prediction block D is checked to derive the merge candidate D, the variable N is set as E, the encoding information of the prediction block E adjacent to the upper left is checked, the merge candidate E is derived, and the merge candidate list is displayed. Registration is performed (steps S1101 to S1118 in FIG. 16).
- step S1102 when the variable N is E and the values of the flags availableFlagA, availableFlagB, availableFlagC, availableFlagD are added and the sum is 4 (YES in step S1102 in FIG. 16), that is, when four spatial merge candidates are derived, merge The flag availableFlagE of the candidate E is set to 0 (step S1107 in FIG. 16), the motion vectors mvL0E and mvL1E of the merge candidate E are both set to (0, 0) (step S1108 in FIG. 16), and the merge candidate E Both the values of the flags predFlagL0E and predFlagL1E are set to 0 (step S1109 in FIG. 16), the process proceeds to step S1118, and the space merge candidate derivation process is terminated.
- step S1102 if the variable N is not E, or the values of the flags availableFlagA, availableFlagB, availableFlagC, availableFlagD are added and the sum is not 4 (NO in step S1102 in FIG. 16), that is, four spatial merge candidates are not derived.
- the process proceeds to step S1103.
- the adjacent prediction block N is included in the same encoding block as the encoding block including the prediction block to be derived (YES in step S1103 in FIG. 16)
- the value of the flag availableFlagN of the merge candidate N is set to 0 (FIG. 16 step S1107)
- the values of motion vectors mvL0N and mvL1N of merge candidate N are both set to (0, 0) (step S1108 of FIG.
- step S1109 in FIG. 16 the process proceeds to step S1118.
- the prediction block N is included in the same encoding block as the encoding block including the prediction block to be derived (YES in step S1103 in FIG. 16)
- the prediction block is merged by not referring to the adjacent prediction block N. It enables parallel processing of candidate derivation and merge candidate list construction processing.
- the prediction block PartIdx is 1 and adjacent prediction
- the case of the block B is a case where the adjacent prediction block N is included in the same encoding block as the encoding block including the prediction block to be derived.
- the adjacent prediction block B is a prediction block whose PartIdx is 0, the prediction block merge candidate derivation and the merge candidate list construction process can be performed in parallel by not referring to the adjacent prediction block B. .
- the partition mode (PartMode) is N ⁇ 2N partition (PART_Nx2N), nL ⁇ 2N partition (PART_nLx2N), or nR ⁇ 2N partition (PART_nRx2N), and the processing target prediction block PartIdx is 1, and the adjacent prediction block A
- the adjacent prediction block N is included in the same encoded block as the encoded block including the prediction block to be derived.
- the adjacent prediction block A is a prediction block whose PartIdx is 0, it is possible to perform parallel processing of prediction block merge candidate derivation and merge candidate list construction processing without referring to the adjacent prediction block A. To do.
- PartMode partition mode
- Part_NxN partition (PART_NxN)
- PartIdx of the prediction block to be processed is 1, 2 or 3
- an encoding block in which the adjacent prediction block N includes the prediction block to be derived May be included in the same coding block.
- the adjacent prediction block N is not included in the same encoding block as the encoding block including the processing target prediction block (NO in step S1103 in FIG. 16)
- the prediction adjacent to the prediction block to be encoded / decoded is performed.
- the block N is specified, and when each prediction block N can be used, the encoding information of the prediction block N is acquired from the encoding information storage memory 115 or 210 (step S1104 in FIG. 16).
- step S1105 If the adjacent prediction block N cannot be used (NO in step S1105 in FIG. 16), or if the prediction mode PredMode of the prediction block N is intra prediction (MODE_INTRA) (NO in step S1106 in FIG. 16), the flag of the merge candidate N
- the value of availableFlagN is set to 0 (step S1107 in FIG. 16)
- the motion vectors mvL0N and mvL1N of the merge candidate N are both set to (0, 0) (step S1108 in FIG. 16)
- step S1109 the flag of the merge candidate N Both predFlagL0N and predFlagL1N are set to 0 (step S1109), and the process proceeds to step S1118.
- the adjacent prediction block N cannot be used, specifically, when the adjacent prediction block N is located outside the encoding / decoding target slice, or is still later in the encoding / decoding processing order. This corresponds to the case where the encoding / decoding process is not completed.
- the adjacent prediction block N when the adjacent prediction block N is outside the same encoding block as the encoding block of the prediction block to be derived (YES in step S1104 in FIG. 16), the adjacent prediction block N can be used (YES in step S1105 in FIG. 16). ), When the prediction mode PredMode of the prediction block N is not intra prediction (MODE_INTRA) (YES in step S1106 in FIG. 16), the inter prediction information of the prediction block N is set as inter prediction information of the merge candidate N. The value of the flag availableFlagN of the merge candidate N is set to 1 (step S1110 in FIG.
- step S1111 is set to the same value (step S1111 in FIG. 16), and the reference indexes refIdxL0N and refIdxL1N of the merge candidate N are set as the reference indexes refIdxL0 [xN] [yN] and refIdxL1 [xN] [yN] of the prediction block N, respectively.
- the same value is set (step S1112 in FIG.
- Step S1113 the flags predFlagL0N and predFlagL1N of the merge candidate N are set to the flags predFlagL0 [xN] [yN] and predFlagL1 [xN] [yN] of the prediction block N, respectively (FIG. 16).
- xN and yN are indexes indicating the position of the upper left pixel of the prediction block N in the picture.
- step S1114 the flags predFlagL0N and predFlagL1N of the merge candidate N, the reference indexes refIdxL0N and refIdxL1N of the merge candidate N, and the motion vectors mvL0N and mvL1N of the merge candidate N are compared with the already derived merge candidates (step S1114 in FIG. 16). If the same merge candidate does not exist (YES in step S1115 in FIG. 16), the merge candidate N is registered at a position where the merge index of the merge candidate list mergeCandList is indicated by the same value as numMergeCand (step S1116 in FIG. 16). 1 is added to the number numMergeCand (step S1117 in FIG. 16). On the other hand, when the same merge candidate exists (NO in step S1115 in FIG. 16), step S1116 and step S1117 are skipped, and the process proceeds to step S1118.
- the reference index of the temporal merge candidate is derived using the reference index of the spatial merge candidate, that is, the reference index used in the prediction block adjacent to the encoding / decoding target block. This is because when the temporal merge candidate is selected, the reference index of the prediction block to be encoded / decoded has a high correlation with the reference index of the prediction block adjacent to the encoding / decoding target block to be the spatial merge candidate. It is. In particular, in the present embodiment, only the reference index of the prediction block A adjacent to the left side of the prediction block to be encoded / decoded is used.
- prediction blocks A and B that are in contact with the sides of the prediction block to be encoded / decoded among the adjacent prediction blocks A, B, C, D, and E that are also spatial merge candidates are predicted to be encoded / decoded.
- the prediction block to be used is limited to the prediction block A without using the prediction blocks C, D, and E having relatively low correlation, thereby obtaining an effect of improving the coding efficiency by deriving the reference index of the temporal merge candidate.
- the calculation amount and memory access amount relating to the reference index derivation process of the temporal merge candidate are reduced.
- FIG. 17 is a diagram showing adjacent blocks that are referred to in the time merge candidate reference index derivation process according to the present embodiment.
- whether or not to refer to the prediction block adjacent to the left side of the prediction block to be derived is switched according to the prediction index PartIdx of the prediction block, regardless of the encoding block partition mode (PartMode).
- PartMode the prediction index PartIdx of the prediction block
- the prediction block adjacent to the left side is referred to
- the partition index PartIdx is other than 0, the adjacent prediction block is not referred to and is set as a default value.
- the prediction block adjacent to the left side is always outside the encoded block in any partition mode (PartMode), but when the partition index PartIdx of the prediction block is other than 0, the partition mode ( PartMode) depending on the coding block.
- PartMode partition mode
- the partition mode (PartMode) is 2N ⁇ 2N partition (PART_2Nx2N)
- the prediction block A0 adjacent to the left of the prediction block to be derived is referred to, and the LX reference index of the temporal merge candidate Is set to the value of the reference index of the LX of the prediction block A0.
- the partition mode (PartMode) for dividing the processing target coding block into two prediction blocks arranged vertically is 2N ⁇ N partition (PART_2NxN), 2N ⁇ nU partition (PART_2NxnU), 2N ⁇ nD partition (PART_2NxnD), and processing target
- the division mode (PartMode) for dividing the coded block of 2 into two prediction blocks arranged side by side is N ⁇ 2N division (PART_Nx2N), nL ⁇ 2N division (PART_nLx2N), and nR ⁇ 2N division (PART_nRx2N), FIG.
- the prediction block A0 that is the derivation target partition index PartIdx refers to the prediction block A0 adjacent to the left, and the time
- the LX reference index of the merge candidate is set to the value of the LX reference index of the prediction block A0.
- the prediction block having the division index PartIdx to be derived is not referred to the adjacent prediction block, and the LX reference index of the temporal merge candidate is set to 0 as the default value. Since the prediction block A0 to be referred to is outside the encoded block, the reference indexes of the temporal merge candidates of the two prediction blocks whose division indexes PartIdx are 0 and 1 can be derived in parallel.
- the division mode (PartMode) for dividing the coding block to be processed into four prediction blocks by dividing it into four prediction blocks is N ⁇ N division (PART_NxN), as shown in FIG.
- the prediction block with index PartIdx of 0 the prediction block A0 adjacent on the left is referred to, and the LX reference index of the temporal merge candidate is set to the value of the LX reference index of the prediction block A0.
- the prediction blocks whose partition indexes PartIdx to be derived are 1, 2, and 3
- the adjacent prediction blocks are not referred to, and the LX reference index of the temporal merge candidate is set to a default value of 0. Since the prediction block A0 to be referred to is outside the encoded block, the reference indexes of the temporal merge candidates of the four prediction blocks whose division indexes PartIdx are 0, 1, 2, and 3 can be derived in parallel.
- the value of the reference index of the LX that is a temporal merge candidate is set to 0 as a default value.
- the reason why the default value of the LX reference index of the temporal merge candidate is 0 when the adjacent prediction block A does not perform LX prediction or when the partition index PartIdx of the prediction block to be derived is 1 is referred to in inter prediction This is because there is a high probability that a reference picture corresponding to an index value of 0 is most selected.
- the present invention is not limited to this, and the default value of the reference index may be a value other than 0 (1, 2, etc.).
- the syntax element shown may be installed and transmitted so that it can be selected on the encoding side.
- FIG. 18 is a flow chart for explaining the procedure for deriving the reference index of the time merge candidate in step S102 of FIG. 15 according to this embodiment.
- the division index PartIdx is 0 (YES in step S2104)
- the encoding information of the prediction block A adjacent to the left of the prediction block to be derived is acquired from the encoding information storage memory 115 or 210 (step S2111). .
- step S2113 to step S2115 The subsequent processing from step S2113 to step S2115 is performed in each of L0 and L1 (steps S2112 to S2116).
- LX is set to L0 when deriving the L0 reference index of the temporal merge candidate
- LX is set to L1 when deriving the L1 reference index.
- the slice type slice_type is a P slice
- Pred_L0 the L0 prediction
- Pred_L1 Pred_L1
- Pred_BI bi-prediction
- the LX reference index refIdxLXCol of the temporal merge candidate is used as the LX reference index refIdxLX of the prediction block A.
- the same value as [xA] [yA] is set (step S2114).
- xA and yA are indexes indicating the position of the upper left pixel of the prediction block A in the picture.
- the predFlagL0 [xN] [yN] and the flag predFlagL1 [xN] [yN] indicating whether to use the L1 prediction of the prediction block N are both 0.
- xN and yN are indexes indicating the position of the upper left pixel of the prediction block N in the picture.
- PredMode of the prediction block N is inter prediction (MODE_INTER) and the inter prediction mode is L0 prediction (Pred_L0)
- the flag predFlagL0 [xN] [yN] indicating whether to use the L0 prediction of the prediction block N is 1
- the flag predFlagL1 [xN] [yN] indicating whether to use the L1 prediction is 0.
- a flag predFlagL0 [xN] [yN] indicating whether to use the L0 prediction of the prediction block N is a flag indicating whether to use 0, L1 prediction predFlagL1 [xN] [yN] is 1.
- a flag predFlagL0 [xN] [yN] indicating whether to use the L0 prediction of the prediction block N, and a flag predFlagL1 [indicating whether to use the L1 prediction xN] [yN] are both 1.
- the reference index refIdxLXCol of the LX that is a temporal merge candidate is set to a default value of 0 ( Step S2115).
- step S2113 to step S2115 is performed in each of L0 and L1 (steps S2112 to S2116), and this reference index derivation processing is terminated.
- step S2121 the processing in subsequent step S2121 is performed in each of L0 and L1 (steps S2118 to S2122).
- LX is set to L0 when deriving the L0 reference index of the temporal merge candidate
- LX is set to L1 when deriving the L1 reference index.
- the slice type slice_type is a P slice
- Pred_L0 the L0 prediction
- Pred_BI bi-prediction
- the time merge candidate LX reference index refIdxLXCol is set to a default value of 0 (step S2121).
- whether to refer to the prediction block adjacent to the left side of the prediction block to be derived is switched, but whether to refer to the prediction block adjacent to the upper side instead of the prediction block adjacent to the left side. May be switched.
- FIG. 19 is a flowchart illustrating the time merge candidate derivation processing procedure in step S103 of FIG.
- the picture type colPic used when deriving a slice type slice_type described in the slice header in slice units and a prediction motion vector candidate in the temporal direction, or a merge candidate a picture colPic of a picture including a prediction block to be processed is included.
- a picture colPic at a different time is derived based on a flag collocated_from_l0_flag indicating whether a reference picture registered in the reference list of L1 or the reference list of L1 is used (step S3101).
- FIG. 20 is a flowchart for explaining the procedure for deriving the picture colPic at different times in step S3101 of FIG.
- the slice type slice_type is a B slice and the flag collocated_from_l0_flag is 0 (YES in step S3201 and YES in step S3202)
- RefPicList1 [0] that is, the picture colPic of the reference list L1 with a reference index of 0 is a different time.
- RefPicList0 [0] that is, the picture colPic of the reference list L0 with the reference index 0 is a different time (step S3205).
- a prediction block located in the lower right (outside) of the same position as the processing target prediction block in the picture colPic at different times is set as a prediction block colPU at different times (step S3301).
- This prediction block corresponds to the prediction block T0 in FIG.
- step S3302 the encoding information of the prediction block colPU at different times is acquired (step S3302).
- PredMode of a prediction block colPU at a different time cannot be used, or when the prediction mode PredMode of a prediction block colPU at a different time is intra prediction (MODE_INTRA) (YES in step S3303, YES in step S3304)
- the picture colPic in a different time The prediction block located in the upper left center of the same position as the processing target prediction block is set as a prediction block colPU at a different time (step S3305). This prediction block corresponds to the prediction block T1 in FIG.
- FIG. 22 is a flowchart for explaining the process of deriving inter prediction information of the temporal merge candidate in step S3103 and step S3104 in FIG.
- L0 or L1 the list of time merge candidate derivation targets is LX
- LX prediction the prediction using LX is LX prediction.
- LX becomes L0 when called as step S3103, which is a time merge candidate L0 derivation process
- LX becomes L1, when called as step S3104, which is a time merge candidate L1 derivation process.
- the prediction mode PredMode of the prediction block colPU at different times is intra prediction (MODE_INTRA) or cannot be used (NO in step S3401 and NO in step S3402), it is assumed that there is no temporal merge candidate.
- the flag availableFlagLXCol and the flag predFlagLXCol are both set to 0 (step S3403), the motion vector mvLXCol is set to (0, 0) (step S3404), and the inter prediction information derivation process for the current time merge candidate ends.
- the prediction mode of the prediction block colPU is Pred_L1, and therefore the motion vector mvCol is predicted.
- MvL1 [xPCol] [yPCol] that is the L1 motion vector of the block colPU is set (step S3406), and the reference index refIdxCol is set to the same value as the reference index RefIdxL1 [xPCol] [yPCol] of L1 (step S3406).
- the list ListCol is set to L1 (step S3408).
- xPCol and yPCol are indexes indicating the position of the upper left pixel of the prediction block colPU in the picture colPic at different times.
- the L0 prediction flag PredFlagL0 [xPCol] [yPCol] of the prediction block colPU is not 0 (NO in step S3405 in FIG. 22)
- the motion vector mvCol becomes the same value as MvL0 [xPCol] [yPCol], which is the L0 motion vector of the prediction block colPU.
- the reference index refIdxCol is set to the same value as the reference index RefIdxL0 [xPCol] [yPCol] of L0 (step S3411), and the list ListCol is set to L0 (step S3412).
- FIG. 23 is a flowchart showing a procedure for deriving inter prediction information of temporal merge candidates when the inter prediction mode of the prediction block colPU is bi-prediction (Pred_BI).
- step S3501 it is determined whether or not the POC of all the pictures registered in all the reference lists is smaller than the POC of the current encoding / decoding target picture (step S3501), and L0 which is all the reference lists of the prediction block colPU.
- LX is L0, that is, the motion vector of L0 of the encoding / decoding target picture
- the prediction vector candidate is derived (YES in step S3502)
- the L0 inter prediction information of the prediction block colPU is selected, and LX is L1, that is, the prediction of the motion vector of L1 of the encoding / decoding target picture.
- L1 of the prediction block colPU When a vector candidate is derived (NO in step S3502), L1 of the prediction block colPU To select a rectangular inter prediction information.
- the flag collocated_from_l0_flag is 0. (YES in step S3503), the L0 inter prediction information of the prediction block colPU is selected. If the flag collocated_from_l0_flag is 1 (NO in step S3503), the L1 inter prediction information of the prediction block colPU is selected. To do.
- the motion vector mvCol is set to the same value as MvL1 [xPCol] [yPCol] (step S3507), see The index refIdxCol is set to the same value as RefIdxL1 [xPCol] [yPCol] (step S3508), and the list ListCol is set to L1 (step S3509).
- both the flag availableFlagLXCol and the flag predFlagLXCol are set to 1 (step S3414).
- the motion vector mvCol is scaled to be a time merge candidate LX motion vector mvLXCol (step S3415).
- This motion vector scaling calculation processing procedure will be described with reference to FIGS.
- the inter-picture distance tb is derived by subtracting the POC of the reference picture corresponding to the LX reference index of the temporal merge candidate derived in step S102 of FIG. 15 from the POC of the current encoding / decoding target picture (step S3602). .
- the inter-picture distance tb becomes a positive value
- the inter-picture distance tb is a negative value.
- step S3603 scaling calculation processing is performed by multiplying mvCol by the scale factor tb / td according to the following equation (step S3605), and the scaled time merge candidate Obtain the motion vector mvLXCol of LX.
- mvLXCol tb / td * mvCol
- the inter-picture distance td and the inter-picture distance tb are derived (steps S3601 and S3602).
- step S3603 the inter-picture distances td and tb are compared. If the inter-picture distances td and tb are equal (YES in step S3603), the LX motion vector mvLXCol as a temporal merge candidate is obtained as in the flowchart of FIG. Is set to the same value as the motion vector mvCol (step S3604), and this scaling calculation process is terminated.
- mvLXCol mvCol
- tx (16384 + Abs (td / 2)) / td
- DistScaleFactor (tb * tx + 32) >> 6
- mvLXCol ClipMv (Sign (DistScaleFactor * mvCol) * ((Abs (DistScaleFactor * mvCol) + 127) >> 8)
- step S3105 if there is a temporal merge candidate (YES in step S3105), the temporal merge candidate is registered at a position where the merge index of the merge candidate list mergeCandList is indicated by the same value as numMergeCand (step S3106). ), 1 is added to the number of merge candidates numMergeCand (step S3107), and this time merge candidate derivation process is terminated.
- step S3106 and step S3107 are skipped, and this time merge candidate derivation process ends.
- FIG. 26 is a flowchart for explaining the additional merge candidate derivation processing procedure in step S104 of FIG.
- a plurality of merge candidates having different values of inter prediction information are generated and registered in the merge candidate list.
- the prediction mode and the motion vector value are fixed, a plurality of merge candidates having different reference index values are generated, and registered in the merge candidate list (step S5101 in FIG. 26). To S5119).
- the slice type is P slice (YES in step S5101 in FIG. 26)
- the value of the reference index number of L0 is set in the variable numRefIdx indicating the reference index number (step S5102 in FIG. 26).
- the slice type is not P slice (NO in step S5101 in FIG. 26)
- the smaller of the reference index number of L0 and the reference index number of L1 in the variable numRefIdx indicating the reference index number Is set step S5103 in FIG. 26.
- 0 is set to the reference index i (step S5104 in FIG. 26).
- step S5106 the process proceeds to step S5107, and if the merge candidate number numMergeCand is not smaller than the maximum merge candidate number maxNumMergeCand (step S5106 in FIG. 26). NO), this additional merge candidate derivation process is terminated.
- step S5107 the process proceeds to step S5109.
- the reference index i is not smaller than the variable numRefIdx (NO in step S5107 in FIG. 26)
- this addition ends.
- step S5110 when the slice type is P slice (YES in step S5109 in FIG. 26), (0, 0) is set in the motion vectors mvL0Zero and mvL1Zero of the additional merge candidates (step S5110 in FIG. 26), and the additional merge candidate is selected.
- the value of the reference index i is set to the reference index refIdxL0Zero
- ⁇ 1 is set to refIdxL1Zero
- 1 is set to the flag predFlagL0Zero of the additional merge candidate
- 0 is set to predFlagL1Zero (FIG. 26).
- Step S5112 the process proceeds to step S5116.
- step S5113 the value of the reference index i is set in the reference indexes refIdxL0Zero and refIdxL1Zero of the additional merge candidates (step S5114 in FIG. 26), and 1 is set in the flags predFlagL0Zero and predFlagL1Zero of the additional merge candidates (step S5115 in FIG. 26). ), The process proceeds to step S5116.
- step S5116 an additional merge candidate is registered at a position where the merge index of the merge candidate list mergeCandList is indicated by the same value as numMergeCand (step S5116 in FIG. 26), and 1 is added to the number of merge candidates numMergeCand (step S5117 in FIG. 26). . Subsequently, 1 is added to the index i (step S5118 in FIG. 26), and the process proceeds to step S5119.
- steps S5106 to S5118 are repeated for each reference index i (steps S5105 to S5119 in FIG. 26).
- the prediction mode and the motion vector value are fixed, and a plurality of merge candidates with different reference index values are generated and registered in the merge candidate list. However, a plurality of merge candidates of different prediction modes are generated. Then, the merge candidates may be registered in the merge candidate list, or merge candidates having different motion vector values may be generated and registered in the merge candidate list.
- the motion vector is changed in the order of (0, 0), (1, 0), (-1, 0), (0, 1), (0, -1). Add while changing the value.
- FIG. 37 is a flowchart illustrating the processing procedure of the inter prediction information selection unit 137 of the inter prediction information deriving unit 104 of the video encoding device.
- the number of merge candidates numMergeCand is greater than 0 (YES in step S8101 of FIG. 37).
- a merge candidate is selected from the valid merge candidates whose merge index registered in the merge candidate list is in the range of 0 to (numMergeCand-1), and L0 of each prediction block of the selected merge candidate PredFlagL0 [xP] [yP], predFlagL1 [xP] [yP], reference index refIdxL0 [xP] [yP], refIdxL1 [xP] [yP], motion vector mvL0 indicating whether to use prediction and L1 prediction Inter prediction information such as [xP] [yP] and mvL1 [xP] [yP] is supplied to the motion compensation prediction unit 105, and a merge index that identifies the selected merge candidate is supplied to the prediction method determination unit 107 ( FIG.
- the merge index mergeIdx indicates a valid merge candidate registered in the merge candidate list mergeCandList.
- the merge index mergeIdx indicates an invalid merge candidate that is not registered in the merge candidate list mergeCandList. Note that by applying the rules described later to the encoding side, even when the merge index mergeIdx indicates an invalid merge candidate, it is possible to select a valid merge candidate.
- merge candidates the same method as the prediction method determination unit 107 can be used.
- the coding information and the coding amount of the residual signal and the coding distortion between the predicted image signal and the image signal are derived, and the merge candidate that produces the least generated code amount and coding distortion is determined.
- entropy coding is performed on the merge index syntax element merge_idx, which is coding information in the merge mode, and the code amount of the coding information is calculated.
- each merge candidate a predicted image signal motion-compensated according to the inter prediction information of each merge candidate by the same method as the motion compensation prediction unit 105, and an encoding target image signal supplied from the image memory 101, The amount of code of the prediction residual signal obtained by encoding the prediction residual signal is calculated. Coding information, that is, the total generated code amount obtained by adding the code amount of the merge index and the code amount of the prediction residual signal is calculated as an evaluation value.
- the prediction residual signal is decoded for distortion evaluation, and the encoding distortion is calculated as a ratio representing an error from the original image signal generated by the encoding.
- the encoding distortion is calculated as a ratio representing an error from the original image signal generated by the encoding.
- encoding information with a small generated code amount and encoding distortion is determined.
- a merge index corresponding to the determined encoding information is encoded as a flag merge_idx represented by a second syntax pattern in units of prediction blocks.
- the generated code amount calculated here is preferably a simulation of the encoding process, but can be approximated or approximated easily.
- the inter prediction information having a default value corresponding to the specified slice type is supplied to the motion compensation prediction unit 105 (steps S8103 to S8105).
- the default value of the inter prediction information is L0 prediction (Pred_L0) (the value of the flag predFlagL0 [xP] [yP] is 1, predFlagL1 [xP] [xP] [YES] when the slice type is P slice (YES in step S8103 in FIG. 37).
- the value of yP] is 0
- the reference index of L0 is 0
- the value of the reference index refIdxL0 [xP] [yP] is 0, the value of refIdxL1 [xP] [yP] is -1
- the vector value mvL0 [ xP] [yP] is set to (0, 0) (step S8104 in FIG. 37).
- the inter prediction mode is bi-prediction (Pred_BI) (flags predFlagL0 [xP] [yP], predFlagL1 [xP] [ yP] is 1), both reference indices are 0 (reference indices refIdxL0 [xP] [yP] and refIdxL1 [xP] [yP] are both 0), and vector values mvL0 [xP of L0 and L1 ] [yP] and mvL1 [xP] [yP] are both (0, 0) (step S8105).
- Pred_BI bi-prediction
- the default value of the inter prediction information is L0 prediction (Pred_L0) (the value of the flag predFlagL0 [xP] [yP] is 1, predFlagL1 [xP] [ The value of yP] is 0), the reference index of L0 is 0 (the value of the reference index refIdxL0 [xP] [yP] is 0, the value of refIdxL1 [xP] [yP] is -1), and the vector value mvL0 [ xP] [yP] may be (0, 0).
- FIG. 38 is a flowchart illustrating the processing procedure of the inter prediction information selection unit 237 of the inter prediction information derivation unit 205 of the video decoding device.
- the number of merge candidates numMergeCand is greater than 0 (YES in step S9101 of FIG. 38)
- a merge candidate corresponding to the merge index mergeIdx supplied from the second encoded bit string decoding unit 202 is selected from the merge candidates registered in the candidate list mergeCandList, and the L0 prediction and the L1 prediction of the selected merge candidate are selected.
- predFlagL0 [xP] [yP] [yP] [yP] L0, L1 reference indexes refIdxL0 [xP] [yP], refIdxL1 [xP] [yP], L0, L1 Inter prediction information such as motion vectors mvL0 [xP] [yP] and mvL1 [xP] [yP] is supplied to the motion compensation prediction unit 206 and stored in the encoded information storage memory 210 (step S91 in FIG. 38). 2).
- an invalid merge candidate is selected on the decoding side.
- inter prediction is performed with invalid inter prediction information, and an unexpected prediction signal may be obtained.
- the inter prediction mode indicates a non-standard value or indicates a reference picture for which no reference index exists, an error may occur and decoding may end abnormally.
- the value of the supplied merge index mergeIdx is equal to or greater than the value of the merge candidate number numMergeIdx, the value of the merge candidate number numMergeIdx is set in the merge index mergeIdx, and then Processing shall be performed.
- the merge index mergeIdx is equal to or greater than the number of merge candidates numMergeIdx, the merge index mergeIdx set on the encoding side indicates an invalid merge candidate that is not registered in the merge candidate list mergeCandList.
- the clipping process of the merge index mergeIdx the merge candidate registered at the end of the merge candidate list mergeCandList can be obtained.
- the value of yP] is 0
- the reference index of L0 is 0
- the value of the reference index refIdxL0 [xP] [yP] is 0, the value of refIdxL1 [xP] [yP] is -1
- the vector value mvL0 [ xP] [yP] is set to (0, 0) (step S9104 in FIG. 38).
- the slice type is not P slice (NO in step S9103 in FIG.
- the inter prediction mode is bi-prediction (Pred_BI) (flags predFlagL0 [xP] [yP], predFlagL1 [ xP] [yP] values are both 1), both reference indices are 0 (reference indices refIdxL0 [xP] [yP] and refIdxL1 [xP] [yP] are both 0), and L0 and L1 vector values Both mvL0 [xP] [yP] and mvL1 [xP] [yP] are set to (0, 0) (step S9105 in FIG. 38).
- Pred_BI bi-prediction
- FIG. 28 is a diagram illustrating a detailed configuration of the inter prediction information deriving unit 104 of the video encoding device in FIG. 1 of the second example according to the embodiment.
- FIG. 29 is a diagram illustrating a detailed configuration of the inter prediction information deriving unit 205 of the video decoding device in FIG. 2 of the second example according to the embodiment.
- the inter prediction information deriving unit 104 of FIG. 28 of the second embodiment is different from the inter prediction information deriving unit 104 of FIG. 12 of the first embodiment in that an effective merge candidate supplementing unit 135 is added.
- FIG. 30 illustrates the merge candidate list construction unit 120 of the inter prediction information deriving unit 104 and the inter prediction information deriving unit 205 of the video decoding device according to the second embodiment of the present invention.
- 10 is a flowchart for explaining a procedure of merge candidate derivation processing and merge candidate list construction processing having functions common to the merge candidate list construction unit 220;
- the flowchart of FIG. 30 of the second embodiment is different from the flowchart of FIG. 15 of the first embodiment in that a process for deriving a valid merge candidate in step S105 is added.
- a merge candidate list mergeCandList is prepared in the merge candidate list generating unit 130 of the inter prediction information deriving unit 104 of the moving image encoding device and the merge candidate list generating unit 230 of the inter prediction information deriving unit 205 of the moving image decoding device (FIG. 30).
- Step S100) the spatial merge candidate generating unit 131 of the inter prediction information deriving unit 104 of the video encoding device and the spatial merge candidate generating unit 231 of the inter prediction information deriving unit 205 of the video decoding device include the video encoding device.
- each prediction block A, B, C, D, adjacent to the encoding / decoding target block is detected.
- the candidate reference index deriving unit 232 derives the reference index of the temporal merge candidate from the prediction block adjacent to the encoding / decoding target block, and uses the derived reference index as the inter prediction information deriving unit 104 of the video encoding device.
- the time merge candidate generation unit 233 of the unit 205 derives time merge candidates from pictures at different times, registers the derived time merge candidates in the merge candidate list mergeCandList (step S103 in FIG.
- valid merge candidates are supplemented by the valid merge candidate supplement unit 135 and the valid merge candidate supplement unit 235, so that the merge index in the merge candidate list is changed from 0 to (maxNumMergeCand-1).
- invalid merge candidates are eliminated (step S105 in FIG. 30).
- the invalid merge candidates are not selected on the decoding side, and only valid merge candidates are selected. Being selected is compensated.
- FIG. 31 is a flowchart for explaining the effective merge candidate derivation process in step S105 of FIG. 30 in the second example of the present embodiment.
- a plurality of merge candidates having the same value of inter prediction information are registered in the merge candidate list.
- An effective merge candidate having a motion vector value (0, 0) in the inter prediction mode corresponding to the slice type is registered in the merge candidate list. (Steps S6101 to S6113 in FIG. 31).
- step S6102 the process proceeds to step S6103, and if the merge candidate number numMergeCand is not smaller than the maximum merge candidate number maxNumMergeCand (step S6102 in FIG. 31). NO), this valid merge candidate derivation process is terminated.
- merge candidate For prediction, (0, 0) is set to the motion vectors mvL0Zero and mvL1Zero of the effective merge candidates (step S6104 in FIG. 31), 0 is set to the reference index refIdxL0Zero of the effective merge candidates, and ⁇ 1 is set to refIdxL1Zero (FIG.
- the valid merge candidate flag predFlagL0Zero is set to 1 and predFlagL1Zero is set to 0 (step S6106 in FIG. 31), and the process proceeds to step S6110.
- step S6110 a valid merge candidate is registered at a position where the merge index of the merge candidate list mergeCandList is indicated by the same value as numMergeCand (step S6110 in FIG. 31), and 1 is added to the number of merge candidates numMergeCand (step S6112 in FIG. 31). ), And the process proceeds to step S6113.
- steps S6102 to S6112 are repeated until the merge candidate number numMergeCand reaches the maximum merge candidate number maxNumMergeCand (steps S6101 to S6113 in FIG. 31).
- invalid merge candidates are eliminated within a range in which the merge index in the merge candidate list is indicated by a value from 0 to (maxNumMergeCand-1).
- the inter prediction information selection unit 137 of the inter prediction information deriving unit 104 of the video encoding device similarly to the inter prediction information selection unit 137 of FIG. 12 of the first embodiment,
- a merge candidate is selected from the merge candidates registered in the merge candidate list, and L0 of each prediction block of the selected merge candidate.
- the inter prediction information selection unit 237 of the inter prediction information deriving unit 205 of the video decoding device is similar to the inter prediction information selection unit 237 of FIG. 13 of the first embodiment. If the merge candidate number numMergeCand is greater than 0 (YES in step S9101 of FIG. 38), the merge index mergeIdx supplied from the second encoded bit string decoding unit 202 is selected from the merge candidates registered in the merge candidate list mergeCandList.
- the inter prediction information such as the indices refIdxL0 [xP] [yP], refIdxL1 [xP] [yP], motion vectors mvL0 [xP] [yP], mvL1 [xP] [yP] of L0 and L1 is sent to the motion compensation prediction unit 206.
- Supply Moni and stores the encoded information storage memory 210.
- FIG. 28 is also a diagram illustrating a detailed configuration of the inter prediction information deriving unit 104 of the video encoding device in FIG. 1 of the third example according to the embodiment.
- FIG. 29 is also a diagram illustrating a detailed configuration of the inter prediction information deriving unit 205 of the video decoding device in FIG. 2 of the third example according to the embodiment.
- FIG. 30 illustrates the merge candidate list construction unit 120 of the inter prediction information deriving unit 104 and the inter prediction information deriving unit 205 of the video decoding device according to the third embodiment of the present invention.
- the merge candidate list is replenished by valid merge candidates supplemented by the valid merge candidate supplement unit 135 in FIG. 28 and the valid merge candidate supplement unit 235 in FIG.
- the invalid merge candidates are eliminated within the range indicated by the merge index in the range from 0 to (maxNumMergeCand-1) (step S105 in FIG. 30).
- the invalid merge candidates are not selected on the decoding side, and only valid merge candidates are selected. Being selected is compensated.
- a merge candidate in which the inter prediction mode is L0 prediction (Pred_L0), the reference index is 0, and the vector value is (0, 0) is set as an effective merge candidate.
- the configuration is the same as the merge candidate list construction unit 220 of the unit 205.
- the processing procedure of step S105 of FIG. 30 performed by the valid merge candidate supplementing unit 135 and the valid merge candidate supplementing unit 235 is different from that of the second embodiment.
- Derivation of valid merge candidates which is the processing procedure of step S105 in FIG. 30 performed by the valid merge candidate supplement unit 135 in FIG. 28 and the valid merge candidate supplement unit 235 in FIG. 29 of the third example according to the present embodiment.
- the method will be described in detail with reference to the flowchart of FIG. FIG. 32 is a flowchart for explaining the effective merge candidate derivation processing procedure in step S105 of FIG. 30 in the third example of the present embodiment.
- step S6102 the process proceeds to step S6103, and if the merge candidate number numMergeCand is not smaller than the maximum merge candidate number maxNumMergeCand (step S6102 in FIG. 32). NO), this valid merge candidate derivation process is terminated.
- a merge candidate having an inter prediction mode of L0 prediction (Pred_L0), a reference index of 0, and a vector value of (0, 0) is set as an effective merge candidate.
- (0, 0) is set to the motion vectors mvL0Zero and mvL1Zero of the effective merge candidate (step S6104 in FIG. 32)
- 0 is set to the reference index refIdxL0Zero of the effective merge candidate
- ⁇ 1 is set to refIdxL1Zero
- the valid merge candidate flag predFlagL0Zero is set to 1
- predFlagL1Zero is set to 0 (step S6106 in FIG. 32).
- a valid merge candidate is registered at a position where the merge index of the merge candidate list mergeCandList is indicated by the same value as numMergeCand (step S6110 in FIG. 32), and 1 is added to the number of merge candidates numMergeCand (step S6112 in FIG. 32). ), And the process proceeds to step S6113.
- steps S6102 to S6112 are repeated until the merge candidate number numMergeCand reaches the maximum merge candidate number maxNumMergeCand (steps S6101 to S6113 in FIG. 32).
- the merge candidate number numMergeCand reaches the maximum merge candidate number maxNumMergeCand (steps S6101 to S6113 in FIG. 32).
- invalid merge candidates are eliminated within a range in which the merge index in the merge candidate list is indicated by a value from 0 to (maxNumMergeCand-1).
- FIG. 28 is also a diagram illustrating a detailed configuration of the inter prediction information deriving unit 104 of the video encoding device in FIG. 1 of the fourth example according to the embodiment.
- FIG. 29 is also a diagram illustrating a detailed configuration of the inter prediction information deriving unit 205 of the video decoding device in FIG. 2 of the fourth example according to the embodiment.
- FIG. 30 shows the merge candidate list construction unit 120 of the inter prediction information deriving unit 104 and the inter prediction information deriving unit 205 of the video decoding device in the video encoding device according to the fourth embodiment of the present invention.
- the merge candidate registered last in the merge list is repeatedly registered in the merge candidate list as an effective merge candidate.
- the configuration is the same as the merge candidate list construction unit 220 of the inter prediction information deriving unit 205 of the apparatus.
- the processing procedure of step S105 in FIG. 30 performed by the valid merge candidate supplementing unit 135 and the valid merge candidate supplementing unit 235 is different from that of the second and third examples.
- the merge candidate list is obtained by a simple process similar to the effective merge candidate derivation process of FIG. 31 of the second embodiment and FIG. 32 of the third embodiment.
- Inter prediction information with the same value in order to register valid merge candidates in the merge candidate list until there are no invalid merge candidates within the range indicated by the merge index in the range 0 to (maxNumMergeCand-1) Register a plurality of merge candidates with the merge candidate list.
- the merge candidate last registered in the merge list is registered as an effective merge candidate in the repeated merge candidate list. (Steps S6101 to S6113 in FIG. 33).
- steps S6102 to S6112 are repeated until the merge candidate number numMergeCand reaches the maximum merge candidate number maxNumMergeCand (steps S6101 to S6113 in FIG. 33).
- invalid merge candidates are eliminated within a range in which the merge index in the merge candidate list is indicated by a value from 0 to (maxNumMergeCand-1).
- step S5106 in FIG. 34 the process proceeds to step S5107, and if the merge candidate number numMergeCand is not smaller than the maximum merge candidate number maxNumMergeCand (step S5106 in FIG. 34). NO), this additional merge candidate derivation process is terminated.
- step S5109 the process advances to step S5109 to perform additional merge candidate registration processing. If the reference index i is not smaller than the variable numRefIdx (NO in step S5107 in FIG. 34), (numRefIdx-1) is set in the reference index i (step S5108 in FIG. 34), and the process proceeds to step S5109, where valid merge candidates Registration process.
- step S5110 when the slice type is P slice (YES in step S5109 in FIG. 34), (0, 0) is set in the motion vectors mvL0Zero and mvL1Zero of the additional merge candidates (step S5110 in FIG. 34), and the additional merge candidates are selected.
- the value of the reference index i is set in the reference index refIdxL0Zero
- -1 is set in refIdxL1Zero
- 1 is set in the flag predFlagL0Zero of the additional merge candidate
- 0 is set in predFlagL1Zero (FIG. 34).
- Step S5112 the process proceeds to step S5116.
- steps S5106 to S5118 are repeated for each reference index i (steps S5105 to S5119 in FIG. 34).
- invalid merge candidates are eliminated within a range in which the merge index in the merge candidate list is indicated by a value from 0 to (maxNumMergeCand-1).
- FIG. 28 is also a diagram illustrating a detailed configuration of the inter prediction information deriving unit 104 of the video encoding device in FIG. 1 according to the sixth example of the embodiment.
- FIG. 29 is also a diagram illustrating a detailed configuration of the inter prediction information deriving unit 205 of the video decoding device in FIG. 2 of the sixth example according to the embodiment.
- FIG. 30 shows the merge candidate list construction unit 120 of the inter prediction information deriving unit 104 and the inter prediction information deriving unit 205 of the video decoding device in the moving picture coding apparatus according to the sixth embodiment of the present invention.
- the inter prediction mode is L0 prediction (Pred_L0) for all elements in the merge candidate list
- Initialization is performed by setting a reference index of 0 and a vector value of (0, 0).
- the slice type is not P slice, that is, when the slice type is B slice, the inter prediction mode is bi-prediction (Pred_BI)
- the reference indices are both 0, and the vector values are both for all elements in the merge candidate list. Both are initialized by setting a value of (0, 0).
- 0 is set to the number of merge candidates numMergeCand.
- step S6201 in FIG. 35 When the merge candidate number numMergeCand is smaller than the maximum merge candidate number maxNumMergeCand (step S6201 in FIG. 35), the value of the maximum merge candidate number maxNumMergeCand is set in the merge candidate number numMergeCand (step S6201 in FIG. 35).
- the inter prediction information initialized by the merge candidate list generation unit 130 in FIG. 29 and the merge candidate list generation unit 230 in FIG. 30 is validated and is set as an effective merge candidate.
- the seventh embodiment differs from the third embodiment in the mounting method, the same inter prediction information can be obtained on the decoding side. Also in the seventh embodiment, as in the sixth embodiment, all indexes in the merge candidate list or inter prediction information within a range indicated by a value from 0 to (maxMergeCand-1) is set to a specified value. After initialization, each merge candidate derivation and registration process is performed. However, in the seventh embodiment, the merge candidate list generation unit 130 in FIG. 28 and the merge candidate list generation unit 230 in FIG. 29 use the inter prediction mode for all elements in the merge candidate list regardless of the slice type. Is L0 prediction (Pred_L0), the reference index is 0, and the vector value is initialized by setting a value of (0, 0). The other processes are the same as in the sixth embodiment.
- the present embodiment has been described above.
- a merge index indicating an invalid merge candidate is encoded on the encoding side
- inter prediction is performed with invalid inter prediction information, and an unexpected prediction signal is generated. May be obtained.
- the inter prediction mode indicates a non-standard value or indicates a reference picture for which no reference index exists, an error may occur and decoding may end abnormally.
- the same inter prediction information can be obtained and the same prediction signal can be obtained, so that the same decoded picture can be obtained.
- the merge index indicating an invalid merge candidate on the encoding side is not selected and encoded, and the inter prediction of the invalid merge candidate on the decoding side is performed. It is compensated that inter prediction using information is not performed, and the same inter prediction information can be obtained and the same prediction signal can be obtained in the video decoding device, so that the same decoded image is obtained. Can do.
- the merge index in the merge candidate list is determined in the effective merge candidate supplementing unit 135 of the moving image encoding device and the effective merge candidate supplementing unit 235 of the moving image decoding device.
- Valid merge candidates have been registered until there are no invalid merge candidates within the range indicated by the value 0 to (maxNumMergeCand-1), but invalid merges at least within the range 0 to (maxNumMergeCand-1) It is sufficient if there are no candidates, and valid merge candidates up to a predetermined range of (maxNumMergeCand-1) or more may be registered.
- the merge candidate list construction unit 120 of the inter prediction information deriving unit 104 of the video encoding device and the inter prediction information deriving unit 205 of the video decoding device are merged.
- the inter prediction information within the range indicated by the value of the merge index in the merge candidate list from 0 to (maxNumMergeCand-1) is initialized with a specified value, but at least from 0 (maxNumMergeCand -1) and may be initialized to a predetermined range equal to or greater than (maxNumMergeCand-1).
- the explanation has been made assuming that the spatial merge candidate, the temporal merge candidate, and the additional merge candidate are derived, respectively.
- a form in which the process for deriving each merge candidate is also conceivable and is included in the present invention.
- a mode in which each merge candidate derivation process is changed or a new merge candidate derivation process is added is also conceivable and is included in the present invention.
- the second example of supplementing the effective merge candidate having the same inter prediction information value as the additional merge candidate in the B slice The methods of the third and seventh embodiments are more suitable than the method of the sixth embodiment, in which effective merge candidates for L0 prediction having different inter prediction information values from the additional merge candidates are supplemented.
- the methods of the third and seventh embodiments in which effective merge candidates for L0 prediction are supplemented at the time of B slices are obtained.
- the methods of the second embodiment and the sixth embodiment for supplementing bi-predictive effective merge candidates with high prediction efficiency are more suitable.
- the default inter prediction information is not specified and skip mode and merge mode are prohibited.
- skip mode and merge mode flags are transmitted, skip mode and merge mode flags are transmitted. Since the merge mode cannot be selected, the encoding efficiency is lowered.
- an error may occur on the decoding side, and the decoding process may end abnormally.
- the encoding side does not encode the skip mode or the merge mode as an invalid value, and the inter prediction of the default value defined on the decoding side Inter-prediction using information is compensated, and the same inter-prediction information can be obtained and the same prediction signal can be obtained in the video decoding device, so that the same decoded image can be obtained. it can. Furthermore, since the skip mode and the merge mode can be selected even when the maximum merge candidate number maxNumMergeCand is 0, the coding efficiency is improved as compared with the case where the skip mode or the merge mode is prohibited.
- the inter prediction information of the merge mode including the skip mode when the value of the maximum merge candidate number maxNumMergeCand is 0 uses a default value, unlike the case where the value of the maximum merge candidate number maxNumMergeCand is 1 or more, a merge candidate list is constructed. It is possible to realize an encoding apparatus that requires no processing and that does not perform merge candidate list construction processing and that has a small amount of calculation. In addition, since the decoding side process only sets a default value in the inter prediction information in the merge mode including the skip mode, the decoding apparatus that reduces the amount of calculation and suppresses the decrease in the encoding efficiency with the minimum correspondence on the decoding side Can be supported.
- the moving image encoded stream output from the moving image encoding apparatus of the embodiment described above has a specific data format so that it can be decoded according to the encoding method used in the embodiment. Therefore, the moving picture decoding apparatus corresponding to the moving picture encoding apparatus can decode the encoded stream of this specific data format.
- the encoded stream When a wired or wireless network is used to exchange an encoded stream between a moving image encoding device and a moving image decoding device, the encoded stream is converted into a data format suitable for the transmission form of the communication path. It may be transmitted.
- a video transmission apparatus that converts the encoded stream output from the video encoding apparatus into encoded data in a data format suitable for the transmission form of the communication channel and transmits the encoded data to the network, and receives the encoded data from the network Then, a moving image receiving apparatus that restores the encoded stream and supplies the encoded stream to the moving image decoding apparatus is provided.
- the moving image transmitting apparatus is a memory that buffers the encoded stream output from the moving image encoding apparatus, a packet processing unit that packetizes the encoded stream, and transmission that transmits the packetized encoded data via the network.
- the moving image receiving apparatus generates a coded stream by packetizing the received data, a receiving unit that receives the packetized coded data via a network, a memory that buffers the received coded data, and packet processing. And a packet processing unit provided to the video decoding device.
- the above processing relating to encoding and decoding can be realized as a transmission, storage, and reception device using hardware, and is stored in a ROM (Read Only Memory), a flash memory, or the like. It can also be realized by firmware or software such as a computer.
- the firmware program and software program can be recorded on a computer-readable recording medium, provided from a server through a wired or wireless network, or provided as a data broadcast of terrestrial or satellite digital broadcasting Is also possible.
Landscapes
- Engineering & Computer Science (AREA)
- Multimedia (AREA)
- Signal Processing (AREA)
- Theoretical Computer Science (AREA)
- Physics & Mathematics (AREA)
- General Physics & Mathematics (AREA)
- Computing Systems (AREA)
- Computer Vision & Pattern Recognition (AREA)
- Compression Or Coding Systems Of Tv Signals (AREA)
- Magnetic Resonance Imaging Apparatus (AREA)
- Ultra Sonic Daignosis Equipment (AREA)
- Liquid Crystal Substances (AREA)
- Measuring And Recording Apparatus For Diagnosis (AREA)
Abstract
Description
td=基準ピクチャcolPicのPOC-時間ダイレクトモードのL0の参照ピクチャのPOC
tb=符号化/復号対象ピクチャのPOC-時間ダイレクトモードのL0の参照ピクチャのPOC
mvL0=tb / td * mvCol
mvL1=mvL0 - mvCol
なお、動画像符号化装置、動画像復号装置の処理能力が低い場合、時間ダイレクトモードの処理を省略することもできる。
実施の形態では、ピクチャを1つまたは複数分割したスライスが符号化の基本単位となっており、スライス毎にスライスの種類を示す情報であるスライスタイプが設定される。図3に示されるように、スライス内を任意の同一サイズの正方の矩形の単位にて均等分割する。この単位をツリーブロックと定義し、スライス内での符号化/復号対象ブロック(符号化処理においては符号化対象ブロック、復号処理においては復号対象ブロックのことである。以下、断りのない限り、この意味で用いる。)を特定するためのアドレス管理の基本単位とする。モノクロを除きツリーブロックは1つの輝度信号と2つの色差信号で構成される。ツリーブロックのサイズはピクチャサイズやピクチャ内のテクスチャに応じて、2のべき乗のサイズで自由に設定することができる。ツリーブロックはピクチャ内のテクスチャに応じて、符号化処理を最適にすべく、必要に応じてツリーブロック内の輝度信号、及び色差信号を階層的に4分割(縦横に2分割ずつ)して、ブロックサイズの小さいブロックにすることができる。このブロックをそれぞれ符号化ブロックと定義し、符号化及び復号を行う際の処理の基本単位とする。モノクロを除き符号化ブロックも1つの輝度信号と2つの色差信号で構成される。符号化ブロックの最大サイズはツリーブロックのサイズと同一である。符号化ブロックの最小のサイズとなる符号化ブロックを最小符号化ブロックと呼び、2のべき乗のサイズで自由に設定することができる。
符号化ブロック単位で、符号化/復号対象ピクチャ内の符号化/復号済み(符号化処理においては符号化した信号を復号したピクチャ、予測ブロック、画像信号等に用い、復号処理においては復号したピクチャ、予測ブロック、画像信号等に用いる。以下、断りのない限り、この意味で用いる。)の周囲の画像信号から予測を行うイントラ予測(MODE_INTRA)、及び符号化/復号済みのピクチャの画像信号から予測を行うインター予測(MODE_INTER)を切り替える。このイントラ予測(MODE_INTRA)とインター予測(MODE_INTER)を識別するモードを予測モード(PredMode)と定義する。予測モード(PredMode)はイントラ予測(MODE_INTRA)、またはインター予測(MODE_INTER)を値として持ち、選択して符号化できる。
ピクチャ内をブロックに分割してイントラ予測(MODE_INTRA)及びインター予測(MODE_INTER)を行う場合、イントラ予測及びインター予測の方法を切り替える単位をより小さくするために、必要に応じて符号化ブロックを分割して予測を行う。この符号化ブロックの輝度信号と色差信号の分割方法を識別するモードを分割モード(PartMode)と定義する。さらに、この分割されたブロックを予測ブロックと定義する。図4に示すように、符号化ブロックの輝度信号の分割方法に応じて8種類の分割モード(PartMode)を定義する。
本実施の形態のツリーブロック、符号化ブロック、予測ブロック、変換ブロックを始めとする各ブロックの位置は、輝度信号の画面の一番左上の輝度信号の画素の位置を原点(0,0)とし、それぞれのブロックの領域に含まれる一番左上の輝度信号の画素の位置を(x,y)の二次元座標で表す。座標軸の向きは水平方向に右の方向、垂直方向に下の方向をそれぞれ正の向きとし、単位は輝度信号の1画素単位である。輝度信号と色差信号で画像サイズ(画素数)が同じである色差フォーマットが4:4:4の場合ではもちろんのこと、輝度信号と色差信号で画像サイズ(画素数)が異なる色差フォーマットが4:2:0、4:2:2の場合でも色差信号の各ブロックの位置をそのブロックの領域に含まれる輝度信号の画素の座標で表し、単位は輝度信号の1画素である。この様にすることで、色差信号の各ブロックの位置が特定できるのはもちろんのこと、座標の値を比較するだけで、輝度信号のブロックと色差信号のブロックの位置の関係も明確となる。
本発明の実施の形態においては、符号化/復号済みのピクチャの画像信号から予測を行うインター予測では、複数の復号済みのピクチャを参照ピクチャとして用いることができる。複数の参照ピクチャから選択された参照ピクチャを特定するため、予測ブロック毎に参照インデックスを付ける。Bスライスでは予測ブロック毎に任意の2枚の参照ピクチャを選択してインター予測を行うことができ、インター予測モードとしてL0予測(Pred_L0)、L1予測(Pred_L1)、双予測(Pred_BI)がある。参照ピクチャはリスト構造のL0(参照リスト0)とL1(参照リスト1)で管理され、L0またはL1の参照インデックスを指定することにより参照ピクチャを特定することができる。L0予測(Pred_L0)はL0で管理されている参照ピクチャを参照するインター予測であり、L1予測(Pred_L1)はL1で管理されている参照ピクチャを参照するインター予測であり、双予測(Pred_BI)はL0予測とL1予測が共に行われ、L0とL1のそれぞれで管理されている1つずつの参照ピクチャを参照するインター予測である。Pスライスのインター予測ではL0予測のみが使用でき、Bスライスのインター予測ではL0予測、L1予測、L0予測とL1予測を平均または重み付け加算する双予測(Pred_BI)が使用できる。以降の処理において出力に添え字LXが付いている定数、変数に関しては、L0、L1ごとに処理が行われることを前提とする。
マージモードとは、符号化/復号対象の予測ブロックの予測モード、参照インデックス、動きベクトル等のインター予測情報を符号化/復号するのではなく、符号化/復号対象の予測ブロックと同一ピクチャ内でその符号化/復号対象の予測ブロックに隣接する予測ブロック、あるいは符号化/復号対象の予測ブロックと時間的に異なる符号化/復号済みのピクチャの符号化/復号対象の予測ブロックと同一位置あるいはその付近(近傍の位置)に存在する予測ブロックのインター予測情報から符号化/復号対象の予測ブロックのインター予測情報を導出することによりインター予測を行うモードである。符号化/復号対象の予測ブロックと同一ピクチャ内でその符号化/復号対象の予測ブロックに隣接する予測ブロック及びその予測ブロックのインター予測情報を空間マージ候補、符号化/復号対象の予測ブロックと時間的に異なる符号化/復号済みのピクチャの符号化/復号対象の予測ブロックと同一位置あるいはその付近(近傍の位置)に存在する予測ブロック及びその予測ブロックのインター予測情報から導出されるインター予測情報を時間マージ候補とする。それぞれのマージ候補はマージ候補リストに登録され、マージインデックスによりインター予測で用いるマージ候補を特定する。
図5、図6、図7及び図8は空間マージ候補の導出、および時間マージ候補の参照インデックスの導出の際に参照する符号化/復号対象の予測ブロックと同一ピクチャ内でその符号化/復号対象の予測ブロックに隣接する予測ブロックを説明する図である。図9は時間マージ候補の導出の際に参照する符号化/復号対象の予測ブロックと時間的に異なる符号化/復号済みのピクチャにおいて、符号化/復号対象の予測ブロックと同一位置あるいはその付近に存在する既に符号化/復号済みの予測ブロックを説明する図である。図5、図6、図7、図8及び図9を用いて符号化/復号対象の予測ブロックの空間方向に隣接する予測ブロック、及び異なる時間の同一位置の予測ブロックについて説明する。
POCは符号化されるピクチャに関連付けられる変数とし、ピクチャの出力/表示順序で1ずつ増加する値が設定される。POCの値によって、同じピクチャであるかを判別したり、出力/表示順序でのピクチャ間の前後関係を判別したり、ピクチャ間の距離を導出したりすることができる。例えば、2つのピクチャのPOCが同じ値を持つ場合、同一のピクチャであると判断できる。2つのピクチャのPOCが違う値を持つ場合、POCの値が小さいピクチャのほうが、先に出力/表示されるピクチャであると判断でき、2つのピクチャのPOCの差が時間軸方向でのピクチャ間距離を示す。
直交変換・量子化部109は、残差信号に対して量子化パラメータに応じて直交変換及び量子化を行い直交変換・量子化された残差信号を生成し、第3符号化ビット列生成部111と逆量子化・逆直交変換部113に供給する。さらに、直交変換・量子化部109は、量子化パラメータを符号化情報格納メモリ115に格納する。
次に、本実施の形態に係る動きベクトルの予測方法を備える動画像符号化装置により符号化され、復号装置により復号される動画像のビットストリームの符号化および復号の共通規則であるシンタックスについて説明する。
以下、諸過程を順を追って説明する。なお、以下の説明においては特に断りのない限りスライスタイプslice_typeがBスライスの場合について説明するが、Pスライスの場合にも適用できる。ただし、スライスタイプslice_typeがPスライスの場合、インター予測モードとしてL0予測(Pred_L0)だけがあり、L1予測(Pred_L1)、双予測(Pred_BI)がないので、L1に纏わる処理を省略することができる。なお、本実施の形態において、動画像符号化装置および動画像復号装置において、最大マージ候補数maxNumMergeCandの値が0の場合、図15のマージ候補の導出処理及びマージ候補リストの構築処理を省略することができる。
td=異なる時間のピクチャcolPicのPOC-予測ブロックcolPUのリストListColで参照する参照ピクチャのPOC
tb=現在の符号化/復号対象ピクチャのPOC-時間マージ候補のLXの参照インデックスに対応する参照ピクチャのPOC
mvLXCol=mvCol
mvLXCol=tb / td * mvCol
mvLXCol=mvCol
tx = ( 16384 + Abs( td / 2 ) ) / td
DistScaleFactor = ( tb * tx + 32 ) >> 6
mvLXCol = ClipMv( Sign( DistScaleFactor * mvCol ) * ( (Abs( DistScaleFactor * mvCol ) + 127 ) >> 8 ) )
尚、ここで算出される発生符号量は、符号化過程をシミュレートしたものであることが望ましいが、簡便に近似したり、概算することも可能である。
Claims (46)
- 動画像の各ピクチャを分割したブロック単位で動き補償予測を用いて前記動画像を符号化する動画像符号化装置であって、
指定されたインター予測情報候補の数を示す情報を符号化する予測情報符号化部と、
符号化対象の予測ブロックに隣接する予測ブロック、または符号化対象の予測ブロックと時間的に異なる符号化済みのピクチャにおける前記符号化対象の予測ブロックと同一位置あるいはその付近に存在する予測ブロックのインター予測情報から、前記インター予測情報候補を導出する予測情報導出部と、
前記導出されたインター予測情報候補からインター予測情報候補リストを構築する候補リスト構築部と、
前記構築されたインター予測情報候補リストに含まれるインター予測情報候補の数が前記指定されたインター予測情報候補の数未満であった場合、前記インター予測情報候補リストに含まれるインター予測情報候補の数が前記指定されたインター予測情報候補の数に達するまで、予測モード、参照インデックス、および動きベクトルのそれぞれに所定の値を有するインター予測情報候補を導出し、前記構築されたインター予測情報候補リストに追加する候補補充部と、
前記インター予測情報候補リストに含まれるインター予測情報候補から1つのインター予測情報の候補を選択して、その選択されたインター予測情報の候補により前記符号化対象の予測ブロックのインター予測を行う動き補償予測部とを備えることを特徴とする動画像符号化装置。 - 動画像の各ピクチャを分割したブロック単位で動き補償予測を用いて前記動画像を符号化する動画像符号化装置であって、
指定されたインター予測情報候補の数を示す情報を符号化する予測情報符号化部と、
符号化対象の予測ブロックに隣接する予測ブロック、または符号化対象の予測ブロックと時間的に異なる符号化済みのピクチャにおける前記符号化対象の予測ブロックと同一位置あるいはその付近に存在する予測ブロックのインター予測情報から、前記インター予測情報の候補を導出する予測情報導出部と、
前記導出されたインター予測情報候補からインター予測情報候補リストを構築する候補リスト構築部と、
前記構築されたインター予測情報候補リストに含まれるインター予測情報候補の数が前記指定されたインター予測情報候補の数未満であった場合、予測モード、参照インデックス、および動きベクトルのそれぞれに所定の値を有するインター予測情報候補を導出し、前記構築されたインター予測情報候補リストに追加し、前記追加されたインター予測情報候補リストに含まれるインター予測情報候補の数が前記指定されたインター予測情報候補の数未満であった場合、前記所定の値を有するインター予測情報候補に対して予測モード、参照インデックス、および動きベクトルの少なくも一つが変更された1つまたは複数のインター予測情報候補を導出し、前記追加されたインター予測情報候補リストにさらに追加する候補追加部と、
前記さらに追加されたインター予測情報候補リストに含まれるインター予測情報候補の数が前記指定されたインター予測情報候補の数に満たない場合、前記さらに追加されたインター予測情報候補リストに含まれるインター予測情報候補の数が前記指定されたインター予測情報候補の数に達するまで、予測モード、参照インデックス、および動きベクトルのそれぞれに所定の値を有するインター予測情報候補を導出し、前記さらに追加されたインター予測情報候補リストにさらに追加する候補補充部と、
前記インター予測情報候補リストに含まれるインター予測情報候補から1つのインター予測情報の候補を選択して、その選択したインター予測情報の候補により前記符号化対象の予測ブロックのインター予測を行う動き補償予測部とを備えることを特徴とする動画像符号化装置。 - 動画像の各ピクチャを分割したブロック単位で動き補償予測を用いて前記動画像が符号化された符号化ビット列を符号化する動画像符号化装置であって、
指定されたインター予測情報候補の数を示す情報を符号化する予測情報符号化部と、
符号化対象の予測ブロックに隣接する予測ブロック、または符号化対象の予測ブロックと時間的に異なる符号化済みのピクチャにおける前記符号化対象の予測ブロックと同一位置又はその付近に存在する予測ブロックのインター予測情報から、前記インター予測情報候補を導出する予測情報導出部と、
前記導出されたインター予測情報候補からインター予測情報候補リストを構築する候補リスト構築部と、
前記構築されたインター予測情報候補リストに含まれるインター予測情報候補の数が前記指定されたインター予測情報候補の数未満であった場合、予測モード、参照インデックス、および動きベクトルのそれぞれに所定の値を有するインター予測情報候補を導出し、前記構築されたインター予測情報候補リストに追加し、前記追加されたインター予測情報候補リストに含まれるインター予測情報候補の数が前記指定されたインター予測情報候補の数未満であった場合、前記所定の値を有するインター予測情報候補に対して予測モードと動きベクトルについては同じ値を有し、参照インデックスについては値が変更された1つまたは複数のインター予測情報候補を導出し、前記追加されたインター予測情報候補リストにさらに追加し、前記さらに追加されたインター予測情報候補リストに含まれるインター予測情報候補の数が前記指定されたインター予測情報候補の数に満たない場合、前記さらに追加されたインター予測情報候補リストに含まれるインター予測情報候補の数が前記指定されたインター予測情報候補の数に達するまで、予測モード、参照インデックス、および動きベクトルのそれぞれに所定の値を有するインター予測情報候補を導出し、前記さらに追加されたインター予測情報候補リストにさらに追加する候補補充部と、
前記インター予測情報候補リストに含まれるインター予測情報候補から1つのインター予測情報の候補を選択して、その選択したインター予測情報の候補により前記符号化対象の予測ブロックのインター予測を行う動き補償予測部とを備えることを特徴とする動画像符号化装置。 - 前記候補補充部は、スライスタイプに応じた予測モードのインター予測情報の候補を追加することを特徴とする請求項1から3のいずれかに記載の動画像符号化装置。
- 前記候補補充部は、前記ピクチャにおけるBスライスに対しては予測モードが双予測、Pスライスに対しては予測モードがL0予測のインター予測情報候補を追加することを特徴とする請求項1から3のいずれかに記載の動画像符号化装置。
- 前記候補補充部は、参照インデックスの値が0のインター予測情報候補を追加することを特徴とする請求項1から3のいずれかに記載の動画像符号化装置。
- 前記候補補充部は、動きベクトルの値が(0,0)のインター予測情報の候補を追加することを特徴とする請求項1から3のいずれかに記載の動画像符号化装置。
- 動画像の各ピクチャを分割したブロック単位で動き補償予測を用いて前記動画像を符号化する動画像符号化装置であって、
指定される候補数のインター予測情報の候補を格納するインター予測情報候補リストに前記指定される候補数分の所定の予測モード、参照インデックスおよび動きベクトルをもつインター予測情報の候補をあらかじめ格納して初期化した上で、符号化対象の予測ブロックに隣接する予測ブロック、または符号化対象の予測ブロックと時間的に異なる符号化済みのピクチャにおける前記符号化対象の予測ブロックと同一位置あるいはその付近に存在する予測ブロックのインター予測情報から、インター予測情報の候補を導出する予測情報導出部と、
前記インター予測情報候補リストに含まれるインター予測情報候補から1つのインター予測情報の候補を選択して、その選択したインター予測情報の候補により前記符号化対象の予測ブロックのインター予測を行う動き補償予測部とを備えることを特徴とする動画像符号化装置。 - 動画像の各ピクチャを分割したブロック単位で動き補償予測を用いて前記動画像を符号化する動画像符号化装置であって、
指定されたインター予測情報候補の数を示す情報を符号化する予測情報符号化部と、
符号化対象の予測ブロックに隣接する予測ブロック、または符号化対象の予測ブロックと時間的に異なる符号化済みのピクチャにおける前記符号化対象の予測ブロックと同一位置あるいはその付近に存在する予測ブロックのインター予測情報から、インター予測情報の候補数として指定される候補数に基づいて、インター予測情報の候補を導出する予測情報導出部と、
前記導出されたインター予測情報候補からインター予測情報候補リストを構築する候補リスト構築部と、
前記指定される候補数が1以上の際に、前記インター予測情報候補リストに含まれるインター予測情報候補から1つのインター予測情報の候補を選択して、その選択したインター予測情報の候補により前記符号化対象の予測ブロックのインター予測を行うと共に、前記指定される候補数が0の際に、規定値のインター予測情報により前記符号化対象の予測ブロックのインター予測を行う動き補償予測部とを備えることを特徴とする動画像符号化装置。 - 動画像の各ピクチャを分割したブロック単位で動き補償予測を用いて前記動画像を符号化する動画像符号化装置であって、
符号化対象の予測ブロックに隣接する予測ブロック、または符号化対象の予測ブロックと時間的に異なる符号化済みのピクチャにおける前記符号化対象の予測ブロックと同一位置あるいはその付近に存在する予測ブロックのインター予測情報から、インター予測情報の候補を導出する予測情報導出部と、
インター予測情報の候補数が指定される候補数に満たない場合、前記指定される候補数になるまで、予測モード、参照インデックスおよび動きベクトル値が同じインター予測情報の候補を補充する候補補充部と、
前記インター予測情報の候補から1つのインター予測情報の候補を選択して、その選択したインター予測情報の候補により前記符号化対象の予測ブロックのインター予測を行う動き補償予測部とを備えることを特徴とする動画像符号化装置。 - 動画像の各ピクチャを分割したブロック単位で動き補償予測を用いて前記動画像を符号化する動画像符号化方法であって、
指定されたインター予測情報候補の数を示す情報を符号化する予測情報符号化ステップと、
符号化対象の予測ブロックに隣接する予測ブロック、または符号化対象の予測ブロックと時間的に異なる符号化済みのピクチャにおける前記符号化対象の予測ブロックと同一位置あるいはその付近に存在する予測ブロックのインター予測情報から、前記インター予測情報候補を導出する予測情報導出ステップと、
前記導出されたインター予測情報候補からインター予測情報候補リストを構築する候補リスト構築ステップと、
前記構築されたインター予測情報候補リストに含まれるインター予測情報候補の数が前記指定されたインター予測情報候補の数未満であった場合、前記インター予測情報候補リストに含まれるインター予測情報候補の数が前記指定されたインター予測情報候補の数に達するまで、予測モード、参照インデックス、および動きベクトルのそれぞれに所定の値を有するインター予測情報候補を導出し、前記構築されたインター予測情報候補リストに追加する候補補充ステップと、
前記インター予測情報候補リストに含まれるインター予測情報候補から1つのインター予測情報の候補を選択して、その選択されたインター予測情報の候補により前記符号化対象の予測ブロックのインター予測を行う動き補償予測ステップとを備えることを特徴とする動画像符号化方法。 - 動画像の各ピクチャを分割したブロック単位で動き補償予測を用いて前記動画像を符号化する動画像符号化方法であって、
指定されたインター予測情報候補の数を示す情報を符号化する予測情報数符号化ステップと、
符号化対象の予測ブロックに隣接する予測ブロック、または符号化対象の予測ブロックと時間的に異なる符号化済みのピクチャにおける前記符号化対象の予測ブロックと同一位置あるいはその付近に存在する予測ブロックのインター予測情報から、前記インター予測情報の候補を導出する予測情報導出ステップと、
前記導出されたインター予測情報候補からインター予測情報候補リストを構築する候補リスト構築ステップと、
前記構築されたインター予測情報候補リストに含まれるインター予測情報候補の数が前記指定されたインター予測情報候補の数未満であった場合、予測モード、参照インデックス、および動きベクトルのそれぞれに所定の値を有するインター予測情報候補を導出し、前記構築されたインター予測情報候補リストに追加し、前記追加されたインター予測情報候補リストに含まれるインター予測情報候補の数が前記指定されたインター予測情報候補の数未満であった場合、前記所定の値を有するインター予測情報候補に対して予測モード、参照インデックス、および動きベクトルの少なくも一つが変更された1つまたは複数のインター予測情報候補を導出し、前記追加されたインター予測情報候補リストにさらに追加する候補追加ステップと、
前記さらに追加されたインター予測情報候補リストに含まれるインター予測情報候補の数が前記指定されたインター予測情報候補の数に満たない場合、前記さらに追加されたインター予測情報候補リストに含まれるインター予測情報候補の数が前記指定されたインター予測情報候補の数に達するまで、予測モード、参照インデックス、および動きベクトルのそれぞれに所定の値を有するインター予測情報候補を導出し、前記さらに追加されたインター予測情報候補リストにさらに追加する候補補充ステップと、
前記インター予測情報候補リストに含まれるインター予測情報候補から1つのインター予測情報の候補を選択して、その選択したインター予測情報の候補により前記符号化対象の予測ブロックのインター予測を行う動き補償予測ステップとを備えることを特徴とする動画像符号化方法。 - 動画像の各ピクチャを分割したブロック単位で動き補償予測を用いて前記動画像が符号化された符号化ビット列を符号化する動画像符号化方法であって、
指定されたインター予測情報候補の数を示す情報を符号化する予測情報数符号化ステップと、
符号化対象の予測ブロックに隣接する予測ブロック、または符号化対象の予測ブロックと時間的に異なる符号化済みのピクチャにおける前記符号化対象の予測ブロックと同一位置又はその付近に存在する予測ブロックのインター予測情報から、前記インター予測情報候補を導出する予測情報導出ステップと、
前記導出されたインター予測情報候補からインター予測情報候補リストを構築する候補リスト構築ステップと、
前記構築されたインター予測情報候補リストに含まれるインター予測情報候補の数が前記指定されたインター予測情報候補の数未満であった場合、予測モード、参照インデックス、および動きベクトルのそれぞれに所定の値を有するインター予測情報候補を導出し、前記構築されたインター予測情報候補リストに追加し、前記追加されたインター予測情報候補リストに含まれるインター予測情報候補の数が前記指定されたインター予測情報候補の数未満であった場合、前記所定の値を有するインター予測情報候補に対して予測モードと動きベクトルについては同じ値を有し、参照インデックスについては値が変更された1つまたは複数のインター予測情報候補を導出し、前記追加されたインター予測情報候補リストにさらに追加し、前記さらに追加されたインター予測情報候補リストに含まれるインター予測情報候補の数が前記指定されたインター予測情報候補の数に満たない場合、前記さらに追加されたインター予測情報候補リストに含まれるインター予測情報候補の数が前記指定されたインター予測情報候補の数に達するまで、予測モード、参照インデックス、および動きベクトルのそれぞれに所定の値を有するインター予測情報候補を導出し、前記さらに追加されたインター予測情報候補リストにさらに追加する候補追加ステップと、
前記インター予測情報候補リストに含まれるインター予測情報候補から1つのインター予測情報の候補を選択して、その選択したインター予測情報の候補により前記符号化対象の予測ブロックのインター予測を行う動き補償予測ステップとを備えることを特徴とする動画像符号化方法。 - 動画像の各ピクチャを分割したブロック単位で動き補償予測を用いて前記動画像を符号化する動画像符号化方法であって、
指定される候補数のインター予測情報の候補を格納するインター予測情報候補リストに前記指定される候補数分の所定の予測モード、参照インデックスおよび動きベクトルをもつインター予測情報の候補をあらかじめ格納して初期化した上で、符号化対象の予測ブロックに隣接する予測ブロック、または符号化対象の予測ブロックと時間的に異なる符号化済みのピクチャにおける前記符号化対象の予測ブロックと同一位置あるいはその付近に存在する予測ブロックのインター予測情報から、インター予測情報の候補を導出する予測情報導出ステップと、
前記インター予測情報候補リストに含まれるインター予測情報候補から1つのインター予測情報の候補を選択して、その選択したインター予測情報の候補により前記符号化対象の予測ブロックのインター予測を行う動き補償予測ステップとを備えることを特徴とする動画像符号化方法。 - 動画像の各ピクチャを分割したブロック単位で動き補償予測を用いて前記動画像を符号化する動画像符号化方法であって、
指定されたインター予測情報候補の数を示す情報を符号化する予測情報数符号化ステップと、
符号化対象の予測ブロックに隣接する予測ブロック、または符号化対象の予測ブロックと時間的に異なる符号化済みのピクチャにおける前記符号化対象の予測ブロックと同一位置あるいはその付近に存在する予測ブロックのインター予測情報から、インター予測情報の候補数として指定される候補数に基づいて、インター予測情報の候補を導出する予測情報導出ステップと、
前記導出されたインター予測情報候補からインター予測情報候補リストを構築する候補リスト構築ステップと、
前記指定される候補数が1以上の際に、前記インター予測情報候補リストに含まれるインター予測情報候補から1つのインター予測情報の候補を選択して、その選択したインター予測情報の候補により前記符号化対象の予測ブロックのインター予測を行うと共に、前記指定される候補数が0の際に、規定値のインター予測情報により前記符号化対象の予測ブロックのインター予測を行う動き補償予測ステップとを備えることを特徴とする動画像符号化方法。 - 動画像の各ピクチャを分割したブロック単位で動き補償予測を用いて前記動画像を符号化する動画像符号化プログラムであって、
指定されたインター予測情報候補の数を示す情報を符号化する予測情報符号化ステップと、
符号化対象の予測ブロックに隣接する予測ブロック、または符号化対象の予測ブロックと時間的に異なる符号化済みのピクチャにおける前記符号化対象の予測ブロックと同一位置あるいはその付近に存在する予測ブロックのインター予測情報から、前記インター予測情報候補を導出する予測情報導出ステップと、
前記導出されたインター予測情報候補からインター予測情報候補リストを構築する候補リスト構築ステップと、
前記構築されたインター予測情報候補リストに含まれるインター予測情報候補の数が前記指定されたインター予測情報候補の数未満であった場合、前記インター予測情報候補リストに含まれるインター予測情報候補の数が前記指定されたインター予測情報候補の数に達するまで、予測モード、参照インデックス、および動きベクトルのそれぞれに所定の値を有するインター予測情報候補を導出し、前記構築されたインター予測情報候補リストに追加する候補補充ステップと、
前記インター予測情報候補リストに含まれるインター予測情報候補から1つのインター予測情報の候補を選択して、その選択されたインター予測情報の候補により前記符号化対象の予測ブロックのインター予測を行う動き補償予測ステップとをコンピュータに実行させることを特徴とする動画像符号化プログラム。 - 動画像の各ピクチャを分割したブロック単位で動き補償予測を用いて前記動画像を符号化する動画像符号化プログラムであって、
指定されたインター予測情報候補の数を示す情報を符号化する予測情報数符号化ステップと、
符号化対象の予測ブロックに隣接する予測ブロック、または符号化対象の予測ブロックと時間的に異なる符号化済みのピクチャにおける前記符号化対象の予測ブロックと同一位置あるいはその付近に存在する予測ブロックのインター予測情報から、前記インター予測情報の候補を導出する予測情報導出ステップと、
前記導出されたインター予測情報候補からインター予測情報候補リストを構築する候補リスト構築ステップと、
前記構築されたインター予測情報候補リストに含まれるインター予測情報候補の数が前記指定されたインター予測情報候補の数未満であった場合、予測モード、参照インデックス、および動きベクトルのそれぞれに所定の値を有するインター予測情報候補を導出し、前記構築されたインター予測情報候補リストに追加し、前記追加されたインター予測情報候補リストに含まれるインター予測情報候補の数が前記指定されたインター予測情報候補の数未満であった場合、前記所定の値を有するインター予測情報候補に対して予測モード、参照インデックス、および動きベクトルの少なくも一つが変更された1つまたは複数のインター予測情報候補を導出し、前記追加されたインター予測情報候補リストにさらに追加する候補追加ステップと、
前記さらに追加されたインター予測情報候補リストに含まれるインター予測情報候補の数が前記指定されたインター予測情報候補の数に満たない場合、前記さらに追加されたインター予測情報候補リストに含まれるインター予測情報候補の数が前記指定されたインター予測情報候補の数に達するまで、予測モード、参照インデックス、および動きベクトルのそれぞれに所定の値を有するインター予測情報候補を導出し、前記さらに追加されたインター予測情報候補リストにさらに追加する候補補充ステップと、
前記インター予測情報候補リストに含まれるインター予測情報候補から1つのインター予測情報の候補を選択して、その選択したインター予測情報の候補により前記符号化対象の予測ブロックのインター予測を行う動き補償予測ステップとをコンピュータに実行させることを特徴とする動画像符号化プログラム。 - 動画像の各ピクチャを分割したブロック単位で動き補償予測を用いて前記動画像が符号化された符号化ビット列を符号化する動画像符号化プログラムであって、
指定されたインター予測情報候補の数を示す情報を符号化する予測情報数符号化ステップと、
符号化対象の予測ブロックに隣接する予測ブロック、または符号化対象の予測ブロックと時間的に異なる符号化済みのピクチャにおける前記符号化対象の予測ブロックと同一位置又はその付近に存在する予測ブロックのインター予測情報から、前記インター予測情報候補を導出する予測情報導出ステップと、
前記導出されたインター予測情報候補からインター予測情報候補リストを構築する候補リスト構築ステップと、
前記構築されたインター予測情報候補リストに含まれるインター予測情報候補の数が前記指定されたインター予測情報候補の数未満であった場合、予測モード、参照インデックス、および動きベクトルのそれぞれに所定の値を有するインター予測情報候補を導出し、前記構築されたインター予測情報候補リストに追加し、前記追加されたインター予測情報候補リストに含まれるインター予測情報候補の数が前記指定されたインター予測情報候補の数未満であった場合、前記所定の値を有するインター予測情報候補に対して予測モードと動きベクトルについては同じ値を有し、参照インデックスについては値が変更された1つまたは複数のインター予測情報候補を導出し、前記追加されたインター予測情報候補リストにさらに追加し、前記さらに追加されたインター予測情報候補リストに含まれるインター予測情報候補の数が前記指定されたインター予測情報候補の数に満たない場合、前記さらに追加されたインター予測情報候補リストに含まれるインター予測情報候補の数が前記指定されたインター予測情報候補の数に達するまで、予測モード、参照インデックス、および動きベクトルのそれぞれに所定の値を有するインター予測情報候補を導出し、前記さらに追加されたインター予測情報候補リストにさらに追加する候補追加ステップと、
前記インター予測情報候補リストに含まれるインター予測情報候補から1つのインター予測情報の候補を選択して、その選択したインター予測情報の候補により前記符号化対象の予測ブロックのインター予測を行う動き補償予測ステップとをコンピュータに実行させることを特徴とする動画像符号化プログラム。 - 動画像の各ピクチャを分割したブロック単位で動き補償予測を用いて前記動画像を符号化する動画像符号化プログラムであって、
指定される候補数のインター予測情報の候補を格納するインター予測情報候補リストに前記指定される候補数分の所定の予測モード、参照インデックスおよび動きベクトルをもつインター予測情報の候補をあらかじめ格納して初期化した上で、符号化対象の予測ブロックに隣接する予測ブロック、または符号化対象の予測ブロックと時間的に異なる符号化済みのピクチャにおける前記符号化対象の予測ブロックと同一位置あるいはその付近に存在する予測ブロックのインター予測情報から、インター予測情報の候補を導出する予測情報導出ステップと、
前記インター予測情報候補リストに含まれるインター予測情報候補から1つのインター予測情報の候補を選択して、その選択したインター予測情報の候補により前記符号化対象の予測ブロックのインター予測を行う動き補償予測ステップとをコンピュータに実行させることを特徴とする動画像符号化プログラム。 - 動画像の各ピクチャを分割したブロック単位で動き補償予測を用いて前記動画像を符号化する動画像符号化プログラムであって、
指定されたインター予測情報候補の数を示す情報を符号化する予測情報数符号化ステップと、
符号化対象の予測ブロックに隣接する予測ブロック、または符号化対象の予測ブロックと時間的に異なる符号化済みのピクチャにおける前記符号化対象の予測ブロックと同一位置あるいはその付近に存在する予測ブロックのインター予測情報から、インター予測情報の候補数として指定される候補数に基づいて、インター予測情報の候補を導出する予測情報導出ステップと、
前記導出されたインター予測情報候補からインター予測情報候補リストを構築する候補リスト構築ステップと、
前記指定される候補数が1以上の際に、前記インター予測情報候補リストに含まれるインター予測情報候補から1つのインター予測情報の候補を選択して、その選択したインター予測情報の候補により前記符号化対象の予測ブロックのインター予測を行うと共に、前記指定される候補数が0の際に、規定値のインター予測情報により前記符号化対象の予測ブロックのインター予測を行う動き補償予測ステップとを備えることを特徴とする動画像符号化プログラム。 - 動画像の各ピクチャを分割したブロック単位で動き補償予測を用いて前記動画像を符号化する動画像符号化方法により符号化された符号化ビット列をパケット化して符号化ストリームを得るパケット処理部と、
パケット化された前記符号化ストリームを送信する送信部とを備え、
前記動画像符号化方法は、
指定されたインター予測情報候補の数を示す情報を符号化する予測情報数符号化ステップと、
符号化対象の予測ブロックに隣接する予測ブロック、または符号化対象の予測ブロックと時間的に異なる符号化済みのピクチャにおける前記符号化対象の予測ブロックと同一位置あるいはその付近に存在する予測ブロックのインター予測情報から、前記インター予測情報の候補を導出する予測情報導出ステップと、
前記導出されたインター予測情報候補からインター予測情報候補リストを構築する候補リスト構築ステップと、
前記構築されたインター予測情報候補リストに含まれるインター予測情報候補の数が前記指定されたインター予測情報候補の数未満であった場合、予測モード、参照インデックス、および動きベクトルのそれぞれに所定の値を有するインター予測情報候補を導出し、前記構築されたインター予測情報候補リストに追加し、前記追加されたインター予測情報候補リストに含まれるインター予測情報候補の数が前記指定されたインター予測情報候補の数未満であった場合、前記所定の値を有するインター予測情報候補に対して予測モード、参照インデックス、および動きベクトルの少なくも一つが変更された1つまたは複数のインター予測情報候補を導出し、前記追加されたインター予測情報候補リストにさらに追加する候補追加ステップと、
前記さらに追加されたインター予測情報候補リストに含まれるインター予測情報候補の数が前記指定されたインター予測情報候補の数に満たない場合、前記さらに追加されたインター予測情報候補リストに含まれるインター予測情報候補の数が前記指定されたインター予測情報候補の数に達するまで、予測モード、参照インデックス、および動きベクトルのそれぞれに所定の値を有するインター予測情報候補を導出し、前記さらに追加されたインター予測情報候補リストにさらに追加する候補補充ステップと、
前記インター予測情報候補リストに含まれるインター予測情報候補から1つのインター予測情報の候補を選択して、その選択したインター予測情報の候補により前記符号化対象の予測ブロックのインター予測を行う動き補償予測ステップとを備えることを特徴とする送信装置。 - 動画像の各ピクチャを分割したブロック単位で動き補償予測を用いて前記動画像を符号化する動画像符号化方法により符号化された符号化ビット列をパケット化して符号化ストリームを得るパケット処理ステップと、
パケット化された前記符号化ストリームを送信する送信ステップとを備え、
前記動画像符号化方法は、
指定されたインター予測情報候補の数を示す情報を符号化する予測情報数符号化ステップと、
符号化対象の予測ブロックに隣接する予測ブロック、または符号化対象の予測ブロックと時間的に異なる符号化済みのピクチャにおける前記符号化対象の予測ブロックと同一位置あるいはその付近に存在する予測ブロックのインター予測情報から、前記インター予測情報の候補を導出する予測情報導出ステップと、
前記導出されたインター予測情報候補からインター予測情報候補リストを構築する候補リスト構築ステップと、
前記構築されたインター予測情報候補リストに含まれるインター予測情報候補の数が前記指定されたインター予測情報候補の数未満であった場合、予測モード、参照インデックス、および動きベクトルのそれぞれに所定の値を有するインター予測情報候補を導出し、前記構築されたインター予測情報候補リストに追加し、前記追加されたインター予測情報候補リストに含まれるインター予測情報候補の数が前記指定されたインター予測情報候補の数未満であった場合、前記所定の値を有するインター予測情報候補に対して予測モード、参照インデックス、および動きベクトルの少なくも一つが変更された1つまたは複数のインター予測情報候補を導出し、前記追加されたインター予測情報候補リストにさらに追加する候補追加ステップと、
前記さらに追加されたインター予測情報候補リストに含まれるインター予測情報候補の数が前記指定されたインター予測情報候補の数に満たない場合、前記さらに追加されたインター予測情報候補リストに含まれるインター予測情報候補の数が前記指定されたインター予測情報候補の数に達するまで、予測モード、参照インデックス、および動きベクトルのそれぞれに所定の値を有するインター予測情報候補を導出し、前記さらに追加されたインター予測情報候補リストにさらに追加する候補補充ステップと、
前記インター予測情報候補リストに含まれるインター予測情報候補から1つのインター予測情報の候補を選択して、その選択したインター予測情報の候補により前記符号化対象の予測ブロックのインター予測を行う動き補償予測ステップとを備えることを特徴とする送信方法。 - 動画像の各ピクチャを分割したブロック単位で動き補償予測を用いて前記動画像を符号化する動画像符号化方法により符号化された符号化ビット列をパケット化して符号化ストリームを得るパケット処理ステップと、
パケット化された前記符号化ストリームを送信する送信ステップとコンピュータに実行させ、
前記動画像符号化方法は、
指定されたインター予測情報候補の数を示す情報を符号化する予測情報数符号化ステップと、
符号化対象の予測ブロックに隣接する予測ブロック、または符号化対象の予測ブロックと時間的に異なる符号化済みのピクチャにおける前記符号化対象の予測ブロックと同一位置あるいはその付近に存在する予測ブロックのインター予測情報から、前記インター予測情報の候補を導出する予測情報導出ステップと、
前記導出されたインター予測情報候補からインター予測情報候補リストを構築する候補リスト構築ステップと、
前記構築されたインター予測情報候補リストに含まれるインター予測情報候補の数が前記指定されたインター予測情報候補の数未満であった場合、予測モード、参照インデックス、および動きベクトルのそれぞれに所定の値を有するインター予測情報候補を導出し、前記構築されたインター予測情報候補リストに追加し、前記追加されたインター予測情報候補リストに含まれるインター予測情報候補の数が前記指定されたインター予測情報候補の数未満であった場合、前記所定の値を有するインター予測情報候補に対して予測モード、参照インデックス、および動きベクトルの少なくも一つが変更された1つまたは複数のインター予測情報候補を導出し、前記追加されたインター予測情報候補リストにさらに追加する候補追加ステップと、
前記さらに追加されたインター予測情報候補リストに含まれるインター予測情報候補の数が前記指定されたインター予測情報候補の数に満たない場合、前記さらに追加されたインター予測情報候補リストに含まれるインター予測情報候補の数が前記指定されたインター予測情報候補の数に達するまで、予測モード、参照インデックス、および動きベクトルのそれぞれに所定の値を有するインター予測情報候補を導出し、前記さらに追加されたインター予測情報候補リストにさらに追加する候補補充ステップと、
前記インター予測情報候補リストに含まれるインター予測情報候補から1つのインター予測情報の候補を選択して、その選択したインター予測情報の候補により前記符号化対象の予測ブロックのインター予測を行う動き補償予測ステップとを備えることを特徴とする送信プログラム。 - 動画像の各ピクチャを分割したブロック単位で動き補償予測を用いて前記動画像が符号化された符号化ビット列を復号する動画像復号装置であって、
予め指定されたインター予測情報候補の数を示す情報を復号する予測情報復号部と、
復号対象の予測ブロックに隣接する予測ブロック、または復号対象の予測ブロックと時間的に異なる復号済みのピクチャにおける前記復号対象の予測ブロックと同一位置又はその付近に存在する予測ブロックのインター予測情報から、前記インター予測情報候補を導出する予測情報導出部と、
前記導出されたインター予測情報候補からインター予測情報候補リストを構築する候補リスト構築部と、
前記構築されたインター予測情報候補リストに含まれるインター予測情報候補の数が前記予め指定されたインター予測情報候補の数未満であった場合、前記インター予測情報候補リストに含まれるインター予測情報候補の数が前記予め指定されたインター予測情報候補の数に達するまで、予測モード、参照インデックス、および動きベクトルのそれぞれに所定の値を有するインター予測情報候補を導出し、前記構築されたインター予測情報候補リストに追加する候補補充部と、
前記インター予測情報候補から1つのインター予測情報候補を選択して、その選択されたインター予測情報候補により前記復号対象の予測ブロックのインター予測を行う動き補償予測部とを備えることを特徴とする動画像復号装置。 - 動画像の各ピクチャを分割したブロック単位で動き補償予測を用いて前記動画像が符号化された符号化ビット列を復号する動画像復号装置であって、
予め指定されたインター予測情報候補の数を示す情報を復号する予測情報復号部と、
復号対象の予測ブロックに隣接する予測ブロック、または復号対象の予測ブロックと時間的に異なる復号済みのピクチャにおける前記復号対象の予測ブロックと同一位置又はその付近に存在する予測ブロックのインター予測情報から、前記インター予測情報候補を導出する予測情報導出部と、
前記導出されたインター予測情報候補からインター予測情報候補リストを構築する候補リスト構築部と、
前記構築されたインター予測情報候補リストに含まれるインター予測情報候補の数が前記予め指定されたインター予測情報候補の数未満であった場合、予測モード、参照インデックス、および動きベクトルのそれぞれに所定の値を有するインター予測情報候補を導出し、前記構築されたインター予測情報候補リストに追加し、前記追加されたインター予測情報候補リストに含まれるインター予測情報候補の数が前記予め指定されたインター予測情報候補の数未満であった場合、前記所定の値を有するインター予測情報候補に対して予測モード、参照インデックス、および動きベクトルの少なくも一つが変更された1つまたは複数のインター予測情報候補を導出し、前記追加されたインター予測情報候補リストにさらに追加する候補追加部と、
前記さらに追加されたインター予測情報候補リストに含まれるインター予測情報候補の数が前記予め指定されたインター予測情報候補の数に満たない場合、前記さらに追加されたインター予測情報候補リストに含まれるインター予測情報候補の数が前記予め指定されたインター予測情報候補の数に達するまで、予測モード、参照インデックス、および動きベクトルのそれぞれに所定の値を有するインター予測情報候補を導出し、前記さらに追加されたインター予測情報候補リストにさらに追加する候補補充部と、
前記インター予測情報の候補から1つのインター予測情報の候補を選択して、その選択したインター予測情報の候補により前記復号対象の予測ブロックのインター予測を行う動き補償予測部とを備えることを特徴とする動画像復号装置。 - 動画像の各ピクチャを分割したブロック単位で動き補償予測を用いて前記動画像が符号化された符号化ビット列を復号する動画像復号装置であって、
予め指定されたインター予測情報候補の数を示す情報を復号する予測情報復号部と、
復号対象の予測ブロックに隣接する予測ブロック、または復号対象の予測ブロックと時間的に異なる復号済みのピクチャにおける前記復号対象の予測ブロックと同一位置又はその付近に存在する予測ブロックのインター予測情報から、前記インター予測情報候補を導出する予測情報導出部と、
前記導出されたインター予測情報候補からインター予測情報候補リストを構築する候補リスト構築部と、
前記構築されたインター予測情報候補リストに含まれるインター予測情報候補の数が前記予め指定されたインター予測情報候補の数未満であった場合、予測モード、参照インデックス、および動きベクトルのそれぞれに所定の値を有するインター予測情報候補を導出し、前記構築されたインター予測情報候補リストに追加し、前記追加されたインター予測情報候補リストに含まれるインター予測情報候補の数が前記予め指定されたインター予測情報候補の数未満であった場合、前記所定の値を有するインター予測情報候補に対して予測モードと動きベクトルについては同じ値を有し、参照インデックスについては値が変更された1つまたは複数のインター予測情報候補を導出し、前記追加されたインター予測情報候補リストにさらに追加し、前記さらに追加されたインター予測情報候補リストに含まれるインター予測情報候補の数が前記予め指定されたインター予測情報候補の数に満たない場合、前記さらに追加されたインター予測情報候補リストに含まれるインター予測情報候補の数が前記予め指定されたインター予測情報候補の数に達するまで、予測モード、参照インデックス、および動きベクトルのそれぞれに所定の値を有するインター予測情報候補を導出し、前記さらに追加されたインター予測情報候補リストにさらに追加する候補補充部と、
前記インター予測情報の候補から1つのインター予測情報の候補を選択して、その選択したインター予測情報の候補により前記復号対象の予測ブロックのインター予測を行う動き補償予測部とを備えることを特徴とする動画像復号装置。 - 前記候補補充部は、スライスタイプに応じた予測モードのインター予測情報候補を追加することを特徴とする請求項24から26のいずれかに記載の動画像復号装置。
- 前記候補補充部は、前記ピクチャにおけるBスライスに対しては予測モードが双予測、Pスライスに対しては予測モードがL0予測のインター予測情報候補を追加することを特徴とする請求項24から26のいずれかに記載の動画像復号装置。
- 前記候補補充部は、参照インデックスの値が0のインター予測情報候補を追加することを特徴とする請求項24から26のいずれかに記載の動画像復号装置。
- 前記候補補充部は、動きベクトルの値が(0,0)のインター予測情報の候補を追加することを特徴とする請求項24から26のいずれかに記載の動画像復号装置。
- 動画像の各ピクチャを分割したブロック単位で動き補償予測を用いて前記動画像が符号化された符号化ビット列を復号する動画像復号装置であって、
インター予測情報の指定される候補数を復号する予測情報復号部と、
指定される候補数のインター予測情報の候補を格納するインター予測情報候補リストに前記指定される候補数分の所定の予測モード、参照インデックスおよび動きベクトルをもつインター予測情報の候補をあらかじめ格納して初期化した上で、復号対象の予測ブロックに隣接する予測ブロック、または復号対象の予測ブロックと時間的に異なる復号済みのピクチャにおける前記復号対象の予測ブロックと同一位置あるいはその付近に存在する予測ブロックのインター予測情報から、インター予測情報の候補を導出する予測情報導出部と、
前記インター予測情報候補リストに含まれるインター予測情報候補から1つのインター予測情報の候補を選択して、その選択したインター予測情報の候補により前記復号対象の予測ブロックのインター予測を行う動き補償予測部とを備えることを特徴とする動画像復号装置。 - 動画像の各ピクチャを分割したブロック単位で動き補償予測を用いて前記動画像を符号化する動画像符号化装置であって、
予め指定されたインター予測情報候補の数を示す情報を復号する予測情報復号部と、
復号対象の予測ブロックに隣接する予測ブロック、または符号化対象の予測ブロックと時間的に異なる符号化済みのピクチャにおける前記符号化対象の予測ブロックと同一位置あるいはその付近に存在する予測ブロックのインター予測情報から、インター予測情報の候補数として指定される候補数に基づいて、インター予測情報の候補を導出する予測情報導出部と、
前記導出されたインター予測情報候補からインター予測情報候補リストを構築する候補リスト構築部と、
前記予め指定された候補数が1以上の際に、前記インター予測情報候補リストに含まれるインター予測情報候補から1つのインター予測情報の候補を選択して、その選択したインター予測情報の候補により前記符号化対象の予測ブロックのインター予測を行うと共に、前記予め指定された候補数が0の際に、規定値のインター予測情報により前記符号化対象の予測ブロックのインター予測を行う動き補償予測部とを備えることを特徴とする動画像復号装置。 - 動画像の各ピクチャを分割したブロック単位で動き補償予測を用いて前記動画像が符号化された符号化ビット列を復号する動画像復号装置であって、
復号対象の予測ブロックに隣接する予測ブロック、または復号対象の予測ブロックと時間的に異なる復号済みのピクチャにおける前記復号対象の予測ブロックと同一位置あるいはその付近に存在する予測ブロックのインター予測情報から、インター予測情報の候補を導出する予測情報導出部と、
インター予測情報の候補数が指定される候補数に満たない場合、前記指定される候補数になるまで、予測モード、参照インデックスおよび動きベクトル値が同じインター予測情報の候補を補充する候補補充部と、
前記インター予測情報の候補から1つのインター予測情報の候補を選択して、その選択したインター予測情報の候補により前記復号対象の予測ブロックのインター予測を行う動き補償予測部とを備えることを特徴とする動画像復号装置。 - 動画像の各ピクチャを分割したブロック単位で動き補償予測を用いて前記動画像が符号化された符号化ビット列を復号する動画像復号方法であって、
予め指定されたインター予測情報候補の数を示す情報を復号する予測情報復号ステップと、
復号対象の予測ブロックに隣接する予測ブロック、または復号対象の予測ブロックと時間的に異なる復号済みのピクチャにおける前記復号対象の予測ブロックと同一位置又はその付近に存在する予測ブロックのインター予測情報から、前記インター予測情報候補を導出する予測情報導出ステップと、
前記導出されたインター予測情報候補からインター予測情報候補リストを構築する候補リスト構築ステップと、
前記構築されたインター予測情報候補リストに含まれるインター予測情報候補の数が前記予め指定されたインター予測情報候補の数未満であった場合、前記インター予測情報候補リストに含まれるインター予測情報候補の数が前記予め指定されたインター予測情報候補の数に達するまで、予測モード、参照インデックス、および動きベクトルのそれぞれに所定の値を有するインター予測情報候補を導出し、前記構築されたインター予測情報候補リストに追加する候補補充ステップと、
前記インター予測情報候補から1つのインター予測情報候補を選択して、その選択されたインター予測情報候補により前記復号対象の予測ブロックのインター予測を行う動き補償予測ステップとを備えることを特徴とする動画像復号方法。 - 動画像の各ピクチャを分割したブロック単位で動き補償予測を用いて前記動画像が符号化された符号化ビット列を復号する動画像復号方法であって、
予め指定されたインター予測情報候補の数を示す情報を復号する予測情報数復号ステップと、
復号対象の予測ブロックに隣接する予測ブロック、または復号対象の予測ブロックと時間的に異なる復号済みのピクチャにおける前記復号対象の予測ブロックと同一位置又はその付近に存在する予測ブロックのインター予測情報から、前記インター予測情報候補を導出する予測情報導出ステップと、
前記導出されたインター予測情報候補からインター予測情報候補リストを構築する候補リスト構築ステップと、
前記構築されたインター予測情報候補リストに含まれるインター予測情報候補の数が前記予め指定されたインター予測情報候補の数未満であった場合、予測モード、参照インデックス、および動きベクトルのそれぞれに所定の値を有するインター予測情報候補を導出し、前記構築されたインター予測情報候補リストに追加し、前記追加されたインター予測情報候補リストに含まれるインター予測情報候補の数が前記予め指定されたインター予測情報候補の数未満であった場合、前記所定の値を有するインター予測情報候補に対して予測モード、参照インデックス、および動きベクトルの少なくも一つが変更された1つまたは複数のインター予測情報候補を導出し、前記追加されたインター予測情報候補リストにさらに追加する候補追加ステップと、
前記さらに追加されたインター予測情報候補リストに含まれるインター予測情報候補の数が前記予め指定されたインター予測情報候補の数に満たない場合、前記さらに追加されたインター予測情報候補リストに含まれるインター予測情報候補の数が前記予め指定されたインター予測情報候補の数に達するまで、予測モード、参照インデックス、および動きベクトルのそれぞれに所定の値を有するインター予測情報候補を導出し、前記さらに追加されたインター予測情報候補リストにさらに追加する候補補充ステップと、
前記インター予測情報の候補から1つのインター予測情報の候補を選択して、その選択したインター予測情報の候補により前記復号対象の予測ブロックのインター予測を行う動き補償予測ステップとを備えることを特徴とする動画像復号方法。 - 動画像の各ピクチャを分割したブロック単位で動き補償予測を用いて前記動画像が符号化された符号化ビット列を復号する動画像復号方法であって、
予め指定されたインター予測情報候補の数を示す情報を復号する予測情報数復号ステップと、
復号対象の予測ブロックに隣接する予測ブロック、または復号対象の予測ブロックと時間的に異なる復号済みのピクチャにおける前記復号対象の予測ブロックと同一位置又はその付近に存在する予測ブロックのインター予測情報から、前記インター予測情報候補を導出する予測情報導出ステップと、
前記導出されたインター予測情報候補からインター予測情報候補リストを構築する候補リスト構築ステップと、
前記構築されたインター予測情報候補リストに含まれるインター予測情報候補の数が前記予め指定されたインター予測情報候補の数未満であった場合、予測モード、参照インデックス、および動きベクトルのそれぞれに所定の値を有するインター予測情報候補を導出し、前記構築されたインター予測情報候補リストに追加し、前記追加されたインター予測情報候補リストに含まれるインター予測情報候補の数が前記予め指定されたインター予測情報候補の数未満であった場合、前記所定の値を有するインター予測情報候補に対して予測モードと動きベクトルについては同じ値を有し、参照インデックスについては値が変更された1つまたは複数のインター予測情報候補を導出し、前記追加されたインター予測情報候補リストにさらに追加し、前記さらに追加されたインター予測情報候補リストに含まれるインター予測情報候補の数が前記予め指定されたインター予測情報候補の数に満たない場合、前記さらに追加されたインター予測情報候補リストに含まれるインター予測情報候補の数が前記予め指定されたインター予測情報候補の数に達するまで、予測モード、参照インデックス、および動きベクトルのそれぞれに所定の値を有するインター予測情報候補を導出し、前記さらに追加されたインター予測情報候補リストにさらに追加する候補追加ステップと、
前記インター予測情報の候補から1つのインター予測情報の候補を選択して、その選択したインター予測情報の候補により前記復号対象の予測ブロックのインター予測を行う動き補償予測ステップとを備えることを特徴とする動画像復号方法。 - 動画像の各ピクチャを分割したブロック単位で動き補償予測を用いて前記動画像が符号化された符号化ビット列を復号する動画像復号方法であって、
インター予測情報の指定される候補数を復号する予測情報数復号ステップと、
指定される候補数のインター予測情報の候補を格納するインター予測情報候補リストに前記指定される候補数分の所定の予測モード、参照インデックスおよび動きベクトルをもつインター予測情報の候補をあらかじめ格納して初期化した上で、復号対象の予測ブロックに隣接する予測ブロック、または復号対象の予測ブロックと時間的に異なる復号済みのピクチャにおける前記復号対象の予測ブロックと同一位置あるいはその付近に存在する予測ブロックのインター予測情報から、インター予測情報の候補を導出する予測情報導出ステップと、
前記インター予測情報候補リストに含まれるインター予測情報候補から1つのインター予測情報の候補を選択して、その選択したインター予測情報の候補により前記復号対象の予測ブロックのインター予測を行う動き補償予測ステップとを備えることを特徴とする動画像復号方法。 - 動画像の各ピクチャを分割したブロック単位で動き補償予測を用いて前記動画像を符号化する動画像符号化方法であって、
予め指定されたインター予測情報候補の数を示す情報を復号する予測情報数復号ステップと、
復号対象の予測ブロックに隣接する予測ブロック、または符号化対象の予測ブロックと時間的に異なる符号化済みのピクチャにおける前記符号化対象の予測ブロックと同一位置あるいはその付近に存在する予測ブロックのインター予測情報から、インター予測情報の候補数として指定される候補数に基づいて、インター予測情報の候補を導出する予測情報導出ステップと、
前記導出されたインター予測情報候補からインター予測情報候補リストを構築する候補リスト構築ステップと、
前記予め指定された候補数が1以上の際に、前記インター予測情報候補リストに含まれるインター予測情報候補から1つのインター予測情報の候補を選択して、その選択したインター予測情報の候補により前記符号化対象の予測ブロックのインター予測を行うと共に、前記予め指定された候補数が0の際に、規定値のインター予測情報により前記符号化対象の予測ブロックのインター予測を行う動き補償予測ステップとを備えることを特徴とする動画像復号方法。 - 動画像の各ピクチャを分割したブロック単位で動き補償予測を用いて前記動画像が符号化された符号化ビット列を復号する動画像復号プログラムであって、
予め指定されたインター予測情報候補の数を示す情報を復号する予測情報復号ステップと、
復号対象の予測ブロックに隣接する予測ブロック、または復号対象の予測ブロックと時間的に異なる復号済みのピクチャにおける前記復号対象の予測ブロックと同一位置又はその付近に存在する予測ブロックのインター予測情報から、前記インター予測情報候補を導出する予測情報導出ステップと、
前記導出されたインター予測情報候補からインター予測情報候補リストを構築する候補リスト構築ステップと、
前記構築されたインター予測情報候補リストに含まれるインター予測情報候補の数が前記予め指定されたインター予測情報候補の数未満であった場合、前記インター予測情報候補リストに含まれるインター予測情報候補の数が前記予め指定されたインター予測情報候補の数に達するまで、予測モード、参照インデックス、および動きベクトルのそれぞれに所定の値を有するインター予測情報候補を導出し、前記構築されたインター予測情報候補リストに追加する候補補充ステップと、
前記インター予測情報候補から1つのインター予測情報候補を選択して、その選択されたインター予測情報候補により前記復号対象の予測ブロックのインター予測を行う動き補償予測ステップとをコンピュータに実行させることを特徴とする動画像復号プログラム。 - 動画像の各ピクチャを分割したブロック単位で動き補償予測を用いて前記動画像が符号化された符号化ビット列を復号する動画像復号プログラムであって、
予め指定されたインター予測情報候補の数を示す情報を復号する予測情報数復号ステップと、
復号対象の予測ブロックに隣接する予測ブロック、または復号対象の予測ブロックと時間的に異なる復号済みのピクチャにおける前記復号対象の予測ブロックと同一位置又はその付近に存在する予測ブロックのインター予測情報から、前記インター予測情報候補を導出する予測情報導出ステップと、
前記導出されたインター予測情報候補からインター予測情報候補リストを構築する候補リスト構築ステップと、
前記構築されたインター予測情報候補リストに含まれるインター予測情報候補の数が前記予め指定されたインター予測情報候補の数未満であった場合、予測モード、参照インデックス、および動きベクトルのそれぞれに所定の値を有するインター予測情報候補を導出し、前記構築されたインター予測情報候補リストに追加し、前記追加されたインター予測情報候補リストに含まれるインター予測情報候補の数が前記予め指定されたインター予測情報候補の数未満であった場合、前記所定の値を有するインター予測情報候補に対して予測モード、参照インデックス、および動きベクトルの少なくも一つが変更された1つまたは複数のインター予測情報候補を導出し、前記追加されたインター予測情報候補リストにさらに追加する候補追加ステップと、
前記さらに追加されたインター予測情報候補リストに含まれるインター予測情報候補の数が前記予め指定されたインター予測情報候補の数に満たない場合、前記さらに追加されたインター予測情報候補リストに含まれるインター予測情報候補の数が前記予め指定されたインター予測情報候補の数に達するまで、予測モード、参照インデックス、および動きベクトルのそれぞれに所定の値を有するインター予測情報候補を導出し、前記さらに追加されたインター予測情報候補リストにさらに追加する候補補充ステップと、
前記インター予測情報の候補から1つのインター予測情報の候補を選択して、その選択したインター予測情報の候補により前記復号対象の予測ブロックのインター予測を行う動き補償予測ステップとをコンピュータに実行させることを特徴とする動画像復号プログラム。 - 動画像の各ピクチャを分割したブロック単位で動き補償予測を用いて前記動画像が符号化された符号化ビット列を復号する動画像復号プログラムであって、
予め指定されたインター予測情報候補の数を示す情報を復号する予測情報数復号ステップと、
復号対象の予測ブロックに隣接する予測ブロック、または復号対象の予測ブロックと時間的に異なる復号済みのピクチャにおける前記復号対象の予測ブロックと同一位置又はその付近に存在する予測ブロックのインター予測情報から、前記インター予測情報候補を導出する予測情報導出ステップと、
前記導出されたインター予測情報候補からインター予測情報候補リストを構築する候補リスト構築ステップと、
前記構築されたインター予測情報候補リストに含まれるインター予測情報候補の数が前記予め指定されたインター予測情報候補の数未満であった場合、予測モード、参照インデックス、および動きベクトルのそれぞれに所定の値を有するインター予測情報候補を導出し、前記構築されたインター予測情報候補リストに追加し、前記追加されたインター予測情報候補リストに含まれるインター予測情報候補の数が前記予め指定されたインター予測情報候補の数未満であった場合、前記所定の値を有するインター予測情報候補に対して予測モードと動きベクトルについては同じ値を有し、参照インデックスについては値が変更された1つまたは複数のインター予測情報候補を導出し、前記追加されたインター予測情報候補リストにさらに追加し、前記さらに追加されたインター予測情報候補リストに含まれるインター予測情報候補の数が前記予め指定されたインター予測情報候補の数に満たない場合、前記さらに追加されたインター予測情報候補リストに含まれるインター予測情報候補の数が前記予め指定されたインター予測情報候補の数に達するまで、予測モード、参照インデックス、および動きベクトルのそれぞれに所定の値を有するインター予測情報候補を導出し、前記さらに追加されたインター予測情報候補リストにさらに追加する候補追加ステップと、
前記インター予測情報の候補から1つのインター予測情報の候補を選択して、その選択したインター予測情報の候補により前記復号対象の予測ブロックのインター予測を行う動き補償予測ステップとをコンピュータに実行させることを特徴とする動画像復号プログラム。 - 動画像の各ピクチャを分割したブロック単位で動き補償予測を用いて前記動画像が符号化された符号化ビット列を復号する動画像復号プログラムであって、
インター予測情報の指定される候補数を復号する予測情報数復号ステップと、
指定される候補数のインター予測情報の候補を格納するインター予測情報候補リストに前記指定される候補数分の所定の予測モード、参照インデックスおよび動きベクトルをもつインター予測情報の候補をあらかじめ格納して初期化した上で、復号対象の予測ブロックに隣接する予測ブロック、または復号対象の予測ブロックと時間的に異なる復号済みのピクチャにおける前記復号対象の予測ブロックと同一位置あるいはその付近に存在する予測ブロックのインター予測情報から、インター予測情報の候補を導出する予測情報導出ステップと、
前記インター予測情報候補リストに含まれるインター予測情報候補から1つのインター予測情報の候補を選択して、その選択したインター予測情報の候補により前記復号対象の予測ブロックのインター予測を行う動き補償予測ステップとをコンピュータに実行させることを特徴とする動画像復号プログラム。 - 動画像の各ピクチャを分割したブロック単位で動き補償予測を用いて前記動画像を符号化する動画像符号化プログラムであって、
予め指定されたインター予測情報候補の数を示す情報を復号する予測情報数復号ステップと、
復号対象の予測ブロックに隣接する予測ブロック、または符号化対象の予測ブロックと時間的に異なる符号化済みのピクチャにおける前記符号化対象の予測ブロックと同一位置あるいはその付近に存在する予測ブロックのインター予測情報から、インター予測情報の候補数として指定される候補数に基づいて、インター予測情報の候補を導出する予測情報導出ステップと、
前記導出されたインター予測情報候補からインター予測情報候補リストを構築する候補リスト構築ステップと、
前記予め指定された候補数が1以上の際に、前記インター予測情報候補リストに含まれるインター予測情報候補から1つのインター予測情報の候補を選択して、その選択したインター予測情報の候補により前記符号化対象の予測ブロックのインター予測を行うと共に、前記予め指定された候補数が0の際に、規定値のインター予測情報により前記符号化対象の予測ブロックのインター予測を行う動き補償予測ステップとをコンピュータに実行させることを特徴とする動画像復号プログラム。 - 動画像が符号化された符号化ビット列を受信して復号する受信装置であって、
動画像の各ピクチャを分割したブロック単位で動き補償予測を用いて前記動画像が符号化された符号化ビット列がパケット化された符号化ストリームを受信する受信部と、
受信された前記符号化ストリームをパケット処理して元の符号化ビット列を復元する復元部と、
復元された前記符号化ビット列から、予め指定されたインター予測情報候補の数を示す情報を復号する予測情報復号部と、
復号対象の予測ブロックに隣接する予測ブロック、または復号対象の予測ブロックと時間的に異なる復号済みのピクチャにおける前記復号対象の予測ブロックと同一位置又はその付近に存在する予測ブロックのインター予測情報から、前記インター予測情報候補を導出する予測情報導出部と、
前記導出されたインター予測情報候補からインター予測情報候補リストを構築する候補リスト構築部と、
前記構築されたインター予測情報候補リストに含まれるインター予測情報候補の数が前記予め指定されたインター予測情報候補の数未満であった場合、予測モード、参照インデックス、および動きベクトルのそれぞれに所定の値を有するインター予測情報候補を導出し、前記構築されたインター予測情報候補リストに追加し、前記追加されたインター予測情報候補リストに含まれるインター予測情報候補の数が前記予め指定されたインター予測情報候補の数未満であった場合、前記所定の値を有するインター予測情報候補に対して予測モード、参照インデックス、および動きベクトルの少なくも一つが変更された1つまたは複数のインター予測情報候補を導出し、前記追加されたインター予測情報候補リストにさらに追加する候補追加部と、
前記さらに追加されたインター予測情報候補リストに含まれるインター予測情報候補の数が前記予め指定されたインター予測情報候補の数に満たない場合、前記さらに追加されたインター予測情報候補リストに含まれるインター予測情報候補の数が前記予め指定されたインター予測情報候補の数に達するまで、予測モード、参照インデックス、および動きベクトルのそれぞれに所定の値を有するインター予測情報候補を導出し、前記さらに追加されたインター予測情報候補リストにさらに追加する候補補充部と、
前記インター予測情報の候補から1つのインター予測情報の候補を選択して、その選択したインター予測情報の候補により前記復号対象の予測ブロックのインター予測を行う動き補償予測部とを備えることを特徴とする受信装置。 - 動画像が符号化された符号化ビット列を受信して復号する受信方法であって、
動画像の各ピクチャを分割したブロック単位で動き補償予測を用いて前記動画像が符号化された符号化ビット列がパケット化された符号化ストリームを受信する受信ステップと、
受信された前記符号化ストリームをパケット処理して元の符号化ビット列を復元する復元ステップと、
復元された前記符号化ビット列から、予め指定されたインター予測情報候補の数を示す情報を復号する予測情報数復号ステップと、
復号対象の予測ブロックに隣接する予測ブロック、または復号対象の予測ブロックと時間的に異なる復号済みのピクチャにおける前記復号対象の予測ブロックと同一位置又はその付近に存在する予測ブロックのインター予測情報から、前記インター予測情報候補を導出する予測情報導出ステップと、
前記導出されたインター予測情報候補からインター予測情報候補リストを構築する候補リスト構築ステップと、
前記構築されたインター予測情報候補リストに含まれるインター予測情報候補の数が前記予め指定されたインター予測情報候補の数未満であった場合、予測モード、参照インデックス、および動きベクトルのそれぞれに所定の値を有するインター予測情報候補を導出し、前記構築されたインター予測情報候補リストに追加し、前記追加されたインター予測情報候補リストに含まれるインター予測情報候補の数が前記予め指定されたインター予測情報候補の数未満であった場合、前記所定の値を有するインター予測情報候補に対して予測モード、参照インデックス、および動きベクトルの少なくも一つが変更された1つまたは複数のインター予測情報候補を導出し、前記追加されたインター予測情報候補リストにさらに追加する候補追加ステップと、
前記さらに追加されたインター予測情報候補リストに含まれるインター予測情報候補の数が前記予め指定されたインター予測情報候補の数に満たない場合、前記さらに追加されたインター予測情報候補リストに含まれるインター予測情報候補の数が前記予め指定されたインター予測情報候補の数に達するまで、予測モード、参照インデックス、および動きベクトルのそれぞれに所定の値を有するインター予測情報候補を導出し、前記さらに追加されたインター予測情報候補リストにさらに追加する候補補充ステップと、
前記インター予測情報の候補から1つのインター予測情報の候補を選択して、その選択したインター予測情報の候補により前記復号対象の予測ブロックのインター予測を行う動き補償予測ステップとを備えることを特徴とする受信方法。 - 動画像が符号化された符号化ビット列を受信して復号する受信プログラムであって、
動画像の各ピクチャを分割したブロック単位で動き補償予測を用いて前記動画像が符号化された符号化ビット列がパケット化された符号化ストリームを受信する受信ステップと、
受信された前記符号化ストリームをパケット処理して元の符号化ビット列を復元する復元ステップと、
復元された前記符号化ビット列から、予め指定されたインター予測情報候補の数を示す情報を復号する予測情報数復号ステップと、
復号対象の予測ブロックに隣接する予測ブロック、または復号対象の予測ブロックと時間的に異なる復号済みのピクチャにおける前記復号対象の予測ブロックと同一位置又はその付近に存在する予測ブロックのインター予測情報から、前記インター予測情報候補を導出する予測情報導出ステップと、
前記導出されたインター予測情報候補からインター予測情報候補リストを構築する候補リスト構築ステップと、
前記構築されたインター予測情報候補リストに含まれるインター予測情報候補の数が前記予め指定されたインター予測情報候補の数未満であった場合、予測モード、参照インデックス、および動きベクトルのそれぞれに所定の値を有するインター予測情報候補を導出し、前記構築されたインター予測情報候補リストに追加し、前記追加されたインター予測情報候補リストに含まれるインター予測情報候補の数が前記予め指定されたインター予測情報候補の数未満であった場合、前記所定の値を有するインター予測情報候補に対して予測モード、参照インデックス、および動きベクトルの少なくも一つが変更された1つまたは複数のインター予測情報候補を導出し、前記追加されたインター予測情報候補リストにさらに追加する候補追加ステップと、
前記さらに追加されたインター予測情報候補リストに含まれるインター予測情報候補の数が前記予め指定されたインター予測情報候補の数に満たない場合、前記さらに追加されたインター予測情報候補リストに含まれるインター予測情報候補の数が前記予め指定されたインター予測情報候補の数に達するまで、予測モード、参照インデックス、および動きベクトルのそれぞれに所定の値を有するインター予測情報候補を導出し、前記さらに追加されたインター予測情報候補リストにさらに追加する候補補充ステップと、
前記インター予測情報の候補から1つのインター予測情報の候補を選択して、その選択したインター予測情報の候補により前記復号対象の予測ブロックのインター予測を行う動き補償予測ステップとをコンピュータに実行させることを特徴とする受信プログラム。
Priority Applications (24)
Application Number | Priority Date | Filing Date | Title |
---|---|---|---|
EP20174155.0A EP3716621B1 (en) | 2012-04-12 | 2013-04-12 | Moving picture coding device, moving picture coding method, moving picture coding program, and moving picture decoding device, moving picture decoding method, moving picture decoding program |
PL21153576T PL3833023T3 (pl) | 2012-04-12 | 2013-04-12 | Konstrukcja listy kandydatów do scalania |
EP21153579.4A EP3833024B1 (en) | 2012-04-12 | 2013-04-12 | Merge candidate list construction |
DK13775584.9T DK2838261T3 (da) | 2012-04-12 | 2013-04-12 | Videodekodningsapparat, videodekodningsmetode, videodekodningsprogram |
EP21153576.0A EP3833023B1 (en) | 2012-04-12 | 2013-04-12 | Merge candidate list construction |
PL13775584T PL2838261T3 (pl) | 2012-04-12 | 2013-04-12 | Urządzenie do dekodowania wideo, sposób dekodowania wideo, program dekodujący wideo |
PL21153568T PL3833022T3 (pl) | 2012-04-12 | 2013-04-12 | Konstrukcja listy kandydatów do scalania |
EP13775584.9A EP2838261B1 (en) | 2012-04-12 | 2013-04-12 | Video decoding device, video decoding method, video decoding program |
EP21153568.7A EP3833022B1 (en) | 2012-04-12 | 2013-04-12 | Merge candidate list construction |
BR112014024294-1A BR112014024294B1 (pt) | 2012-04-12 | 2013-04-12 | Dispositivo de decodificação de imagem em movimento e método de decodificação de imagem em movimento |
RU2014145257A RU2617920C9 (ru) | 2012-04-12 | 2013-04-12 | Устройство кодирования движущегося изображения, способ кодирования движущегося изображения и программа кодирования движущегося изображения, а также устройство декодирования движущегося изображения, способ декодирования движущегося изображения и программа декодирования движущегося изображения |
SI201331744T SI2838261T1 (sl) | 2012-04-12 | 2013-04-12 | Naprava za dekodiranje videa, postopek za dekodiranje videa, program za dekodiranje videa |
PL20174155T PL3716621T3 (pl) | 2012-04-12 | 2013-04-12 | Urządzenie do kodowania ruchomych obrazów, sposób kodowania ruchomych obrazów, program kodujący ruchome obrazy oraz urządzenie do dekodowania ruchomych obrazów, sposób dekodowania ruchomych obrazów, program dekodujący ruchome obrazy |
US14/479,139 US9681134B2 (en) | 2012-04-12 | 2014-09-05 | Moving picture coding device, moving picture coding method, and moving picture coding program, and moving picture decoding device, moving picture decoding method, and moving picture decoding program |
US15/206,015 US9667966B2 (en) | 2012-04-12 | 2016-07-08 | Moving picture coding device, moving picture coding method, and moving picture coding program, and moving picture decoding device, moving picture decoding method, and moving picture decoding program |
US15/490,521 US9918103B2 (en) | 2012-04-12 | 2017-04-18 | Moving picture decoding device, moving picture decoding method, and moving picture decoding program |
US15/490,445 US9872038B2 (en) | 2012-04-12 | 2017-04-18 | Moving picture coding device, moving picture coding method, and moving picture coding program |
US15/490,337 US9872037B2 (en) | 2012-04-12 | 2017-04-18 | Moving picture decoding device, moving picture decoding method, and moving picture decoding program |
US15/861,763 US10230975B2 (en) | 2012-04-12 | 2018-01-04 | Moving picture decoding device, moving picture decoding method, and moving picture decoding program |
US16/276,661 US10523962B2 (en) | 2012-04-12 | 2019-02-15 | Moving picture coding device, moving picture coding method, moving picture coding program, moving picture decoding device, moving picture decoding method, and moving picture decoding program |
US16/729,774 US10791336B2 (en) | 2012-04-12 | 2019-12-30 | Moving picture coding device, moving picture coding method, moving picture coding program, moving picture decoding device, moving picture decoding method, and moving picture decoding program |
US17/032,531 US11206421B2 (en) | 2012-04-12 | 2020-09-25 | Moving picture coding device, moving picture coding method, moving picture coding program, moving picture decoding device, moving picture decoding method, and moving picture decoding program |
US17/528,700 US11831898B2 (en) | 2012-04-12 | 2021-11-17 | Moving picture coding device, moving picture coding method, moving picture coding program, moving picture decoding device, moving picture decoding method, and moving picture decoding program |
US18/490,340 US20240048754A1 (en) | 2012-04-12 | 2023-10-19 | Moving picture coding device, moving picture coding method, moving picture coding program, moving picture decoding device, moving picture decoding method, and moving picture decoding program |
Applications Claiming Priority (8)
Application Number | Priority Date | Filing Date | Title |
---|---|---|---|
JP2012091386 | 2012-04-12 | ||
JP2012-091386 | 2012-04-12 | ||
JP2012-091385 | 2012-04-12 | ||
JP2012091385 | 2012-04-12 | ||
JP2013083577A JP6020323B2 (ja) | 2012-04-12 | 2013-04-12 | 動画像符号化装置、動画像符号化方法、動画像符号化プログラム、送信装置、送信方法及び送信プログラム |
JP2013-083577 | 2013-04-12 | ||
JP2013-083578 | 2013-04-12 | ||
JP2013083578A JP5633597B2 (ja) | 2012-04-12 | 2013-04-12 | 動画像復号装置、動画像復号方法、動画像復号プログラム、受信装置、受信方法及び受信プログラム |
Related Child Applications (1)
Application Number | Title | Priority Date | Filing Date |
---|---|---|---|
US14/479,139 Continuation US9681134B2 (en) | 2012-04-12 | 2014-09-05 | Moving picture coding device, moving picture coding method, and moving picture coding program, and moving picture decoding device, moving picture decoding method, and moving picture decoding program |
Publications (1)
Publication Number | Publication Date |
---|---|
WO2013153823A1 true WO2013153823A1 (ja) | 2013-10-17 |
Family
ID=50157654
Family Applications (1)
Application Number | Title | Priority Date | Filing Date |
---|---|---|---|
PCT/JP2013/002513 WO2013153823A1 (ja) | 2012-04-12 | 2013-04-12 | 動画像符号化装置、動画像符号化方法、動画像符号化プログラム、送信装置、送信方法及び送信プログラム、並びに動画像復号装置、動画像復号方法、動画像復号プログラム、受信装置、受信方法及び受信プログラム |
Country Status (12)
Country | Link |
---|---|
US (11) | US9681134B2 (ja) |
EP (5) | EP3833024B1 (ja) |
BR (1) | BR112014024294B1 (ja) |
DK (4) | DK3716621T3 (ja) |
ES (4) | ES2951398T3 (ja) |
HR (4) | HRP20211884T1 (ja) |
HU (3) | HUE056708T2 (ja) |
PL (5) | PL3716621T3 (ja) |
RU (5) | RU2658146C9 (ja) |
SI (4) | SI3833022T1 (ja) |
TW (5) | TWI578760B (ja) |
WO (1) | WO2013153823A1 (ja) |
Families Citing this family (40)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
WO2011145819A2 (ko) * | 2010-05-19 | 2011-11-24 | 에스케이텔레콤 주식회사 | 영상 부호화/복호화 장치 및 방법 |
CN107071458B (zh) * | 2010-12-14 | 2020-01-03 | M&K控股株式会社 | 用于编码运动画面的设备 |
PL3716621T3 (pl) | 2012-04-12 | 2022-01-24 | Jvckenwood Corporation | Urządzenie do kodowania ruchomych obrazów, sposób kodowania ruchomych obrazów, program kodujący ruchome obrazy oraz urządzenie do dekodowania ruchomych obrazów, sposób dekodowania ruchomych obrazów, program dekodujący ruchome obrazy |
US10205950B2 (en) | 2014-02-21 | 2019-02-12 | Panasonic Corporation | Image decoding method, image encoding method, image decoding apparatus, and image encoding apparatus |
KR20220162877A (ko) * | 2014-10-31 | 2022-12-08 | 삼성전자주식회사 | 고정밀 스킵 부호화를 이용한 비디오 부호화 장치 및 비디오 복호화 장치 및 그 방법 |
JP6678357B2 (ja) * | 2015-03-31 | 2020-04-08 | リアルネットワークス,インコーポレーテッド | ビデオ符号化システムにおける動きベクトルの選択及び予測方法 |
KR102583501B1 (ko) * | 2015-04-27 | 2023-09-27 | 엘지전자 주식회사 | 비디오 신호의 처리 방법 및 이를 위한 장치 |
WO2017010073A1 (ja) | 2015-07-10 | 2017-01-19 | 日本電気株式会社 | 動画像符号化装置、動画像符号化方法および動画像符号化プログラムを記憶する記録媒体 |
US10652569B2 (en) | 2015-12-22 | 2020-05-12 | Realnetworks, Inc. | Motion vector selection and prediction in video coding systems and methods |
CN116708782A (zh) * | 2016-07-12 | 2023-09-05 | 韩国电子通信研究院 | 图像编码/解码方法以及用于该方法的记录介质 |
WO2018152749A1 (en) * | 2017-02-23 | 2018-08-30 | Realnetworks, Inc. | Coding block bitstream structure and syntax in video coding systems and methods |
CN117061738A (zh) | 2018-04-02 | 2023-11-14 | 深圳市大疆创新科技有限公司 | 用于图像处理的方法和图像处理装置 |
WO2019216325A1 (en) | 2018-05-09 | 2019-11-14 | Sharp Kabushiki Kaisha | Systems and methods for performing motion vector prediction using a derived set of motion vectors |
GB2593020B (en) | 2018-05-24 | 2022-12-28 | Kt Corp | Method and apparatus for processing video signal |
US11412246B2 (en) | 2018-06-08 | 2022-08-09 | Kt Corporation | Method and apparatus for processing video signal |
KR20240007299A (ko) | 2018-06-29 | 2024-01-16 | 베이징 바이트댄스 네트워크 테크놀로지 컴퍼니, 리미티드 | 룩업 테이블의 업데이트: fifo, 제약된 fifo |
BR112020024142A2 (pt) | 2018-06-29 | 2021-03-02 | Beijing Bytedance Network Technology Co., Ltd. | método para processamento de vídeo, aparelho para codificação de dados de vídeo, meio de armazenamento e meio de gravação legíveis por computador não transitório |
KR20240005239A (ko) | 2018-06-29 | 2024-01-11 | 베이징 바이트댄스 네트워크 테크놀로지 컴퍼니, 리미티드 | Lut와 amvp 사이의 상호작용 |
JP7137008B2 (ja) | 2018-06-29 | 2022-09-13 | 北京字節跳動網絡技術有限公司 | 1つまたは複数のルックアップテーブルを使用して、以前コーディングされた動き情報を順に記憶させてそれらを後続のブロックのコーディングに使用する概念 |
KR102646649B1 (ko) | 2018-06-29 | 2024-03-13 | 베이징 바이트댄스 네트워크 테크놀로지 컴퍼니, 리미티드 | Lut에서의 모션 후보들의 검사 순서 |
CN110662057B (zh) | 2018-06-29 | 2022-06-21 | 北京字节跳动网络技术有限公司 | 视频处理方法、装置、设备以及存储比特流的方法 |
JP7460617B2 (ja) | 2018-06-29 | 2024-04-02 | 北京字節跳動網絡技術有限公司 | Lut更新条件 |
KR102611261B1 (ko) * | 2018-06-29 | 2023-12-08 | 베이징 바이트댄스 네트워크 테크놀로지 컴퍼니, 리미티드 | 업데이트 대상 lut 또는 비업데이트 lut의 구별 |
WO2020008350A1 (en) | 2018-07-02 | 2020-01-09 | Beijing Bytedance Network Technology Co., Ltd. | Luts with intra prediction modes and intra mode prediction from non-adjacent blocks |
US10924731B2 (en) * | 2018-08-28 | 2021-02-16 | Tencent America LLC | Complexity constraints on merge candidates list construction |
CN110868601B (zh) * | 2018-08-28 | 2024-03-15 | 华为技术有限公司 | 帧间预测方法、装置以及视频编码器和视频解码器 |
TW202025760A (zh) | 2018-09-12 | 2020-07-01 | 大陸商北京字節跳動網絡技術有限公司 | 要檢查多少個hmvp候選 |
CN112740705A (zh) * | 2018-09-21 | 2021-04-30 | 夏普株式会社 | 用于在视频编码中发送信号通知参考图片的系统和方法 |
KR102658887B1 (ko) * | 2018-10-10 | 2024-04-18 | 삼성전자주식회사 | 움직임 벡터 차분값을 이용한 비디오 부호화 및 복호화 방법, 및 움직임 정보의 부호화 및 복호화 장치 |
CN116248870A (zh) | 2018-11-08 | 2023-06-09 | Oppo广东移动通信有限公司 | 视频信号编码/解码方法以及用于所述方法的设备 |
WO2020096425A1 (ko) | 2018-11-08 | 2020-05-14 | 주식회사 엑스리스 | 영상 신호 부호화/복호화 방법 및 이를 위한 장치 |
KR20200054113A (ko) | 2018-11-08 | 2020-05-19 | 광동 오포 모바일 텔레커뮤니케이션즈 코포레이션 리미티드 | 영상 신호 부호화/복호화 방법 및 이를 위한 장치 |
WO2020130714A1 (ko) * | 2018-12-21 | 2020-06-25 | 주식회사 엑스리스 | 영상 신호 부호화/복호화 방법 및 이를 위한 장치 |
WO2020143741A1 (en) | 2019-01-10 | 2020-07-16 | Beijing Bytedance Network Technology Co., Ltd. | Invoke of lut updating |
CN113383554B (zh) | 2019-01-13 | 2022-12-16 | 北京字节跳动网络技术有限公司 | LUT和共享Merge列表之间的交互 |
CN113330739A (zh) | 2019-01-16 | 2021-08-31 | 北京字节跳动网络技术有限公司 | Lut中的运动候选的插入顺序 |
CN113615193A (zh) | 2019-03-22 | 2021-11-05 | 北京字节跳动网络技术有限公司 | Merge列表构建和其他工具之间的交互 |
CN113853783B (zh) | 2019-05-25 | 2023-12-15 | 北京字节跳动网络技术有限公司 | 对帧内块复制编解码的块的块矢量的编解码 |
CN113924782A (zh) * | 2019-06-25 | 2022-01-11 | Jvc建伍株式会社 | 动图像编码装置、动图像编码方法及动图像编码程序、动图像解码装置、动图像解码方法及动图像解码程序 |
AU2020363130A1 (en) * | 2019-10-10 | 2022-05-05 | Huawei Technologies Co., Ltd. | An encoder, a decoder and corresponding methods for simplifying signalling picture header |
Citations (1)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
JP2004129191A (ja) | 2002-10-04 | 2004-04-22 | Lg Electronics Inc | Bピクチャのダイレクトモード動きベクトル演算方法 |
Family Cites Families (25)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
KR100399932B1 (ko) * | 2001-05-07 | 2003-09-29 | 주식회사 하이닉스반도체 | 메모리의 양을 감소시키기 위한 비디오 프레임의압축/역압축 하드웨어 시스템 |
US20040001546A1 (en) * | 2002-06-03 | 2004-01-01 | Alexandros Tourapis | Spatiotemporal prediction for bidirectionally predictive (B) pictures and motion vector prediction for multi-picture reference motion compensation |
US7400681B2 (en) * | 2003-11-28 | 2008-07-15 | Scientific-Atlanta, Inc. | Low-complexity motion vector prediction for video codec with two lists of reference pictures |
US7453938B2 (en) * | 2004-02-06 | 2008-11-18 | Apple Inc. | Target bitrate estimator, picture activity and buffer management in rate control for video coder |
US8879857B2 (en) * | 2005-09-27 | 2014-11-04 | Qualcomm Incorporated | Redundant data encoding methods and device |
JP4999859B2 (ja) * | 2006-10-30 | 2012-08-15 | 日本電信電話株式会社 | 予測参照情報生成方法、動画像符号化及び復号方法、それらの装置、及びそれらのプログラム並びにプログラムを記録した記憶媒体 |
JP4821723B2 (ja) | 2007-07-13 | 2011-11-24 | 富士通株式会社 | 動画像符号化装置及びプログラム |
JP2009164880A (ja) * | 2008-01-07 | 2009-07-23 | Mitsubishi Electric Corp | トランスコーダ及び受信機 |
JP4600574B2 (ja) | 2009-01-07 | 2010-12-15 | 日本電気株式会社 | 動画像復号装置、動画像復号方法、及びプログラム |
TWI387314B (zh) * | 2009-03-10 | 2013-02-21 | Univ Nat Central | Image processing apparatus and method thereof |
KR101490646B1 (ko) * | 2009-05-29 | 2015-02-06 | 미쓰비시덴키 가부시키가이샤 | 화상 복호 장치 및 화상 복호 방법 |
CN102783152A (zh) * | 2010-03-02 | 2012-11-14 | 三星电子株式会社 | 用于使用可分级视频编码方案进行自适应流传输的方法和设备 |
WO2012008506A1 (ja) | 2010-07-15 | 2012-01-19 | シャープ株式会社 | 画像フィルタ装置、復号装置、動画像復号装置、符号化装置、動画像符号化装置、および、データ構造 |
HRP20231669T1 (hr) * | 2010-09-02 | 2024-03-15 | Lg Electronics, Inc. | Inter predikcija |
US8861617B2 (en) * | 2010-10-05 | 2014-10-14 | Mediatek Inc | Method and apparatus of region-based adaptive loop filtering |
CN107105258B (zh) * | 2010-10-08 | 2020-09-01 | Ge视频压缩有限责任公司 | 编码器和编码方法与解码器和解码方法 |
EP2717575B1 (en) | 2011-05-27 | 2018-10-10 | Sun Patent Trust | Image decoding method and image decoding device |
US9485518B2 (en) | 2011-05-27 | 2016-11-01 | Sun Patent Trust | Decoding method and apparatus with candidate motion vectors |
ES2779226T3 (es) | 2011-05-31 | 2020-08-14 | Sun Patent Trust | Procedimiento de descodificación de vídeo y dispositivo de descodificación de vídeo |
US9866859B2 (en) * | 2011-06-14 | 2018-01-09 | Texas Instruments Incorporated | Inter-prediction candidate index coding independent of inter-prediction candidate list construction in video coding |
JP5488666B2 (ja) | 2011-09-28 | 2014-05-14 | 株式会社Jvcケンウッド | 動画像復号装置、動画像復号方法、動画像復号プログラム、受信装置、受信方法及び受信プログラム |
CN108881903B (zh) | 2011-10-19 | 2022-01-04 | 太阳专利托管公司 | 图像编码方法及装置、图像解码方法及装置、编解码装置 |
WO2013108613A1 (ja) | 2012-01-17 | 2013-07-25 | パナソニック株式会社 | 動画像符号化方法、動画像復号化方法、動画像符号化装置、動画像復号化装置および動画像符号化復号化装置 |
BR122022002972B1 (pt) * | 2012-01-19 | 2022-11-22 | Electronics And Telecommunications Research Institute | Aparelho de decodificação de imagem |
PL3716621T3 (pl) | 2012-04-12 | 2022-01-24 | Jvckenwood Corporation | Urządzenie do kodowania ruchomych obrazów, sposób kodowania ruchomych obrazów, program kodujący ruchome obrazy oraz urządzenie do dekodowania ruchomych obrazów, sposób dekodowania ruchomych obrazów, program dekodujący ruchome obrazy |
-
2013
- 2013-04-12 PL PL20174155T patent/PL3716621T3/pl unknown
- 2013-04-12 DK DK20174155.0T patent/DK3716621T3/da active
- 2013-04-12 EP EP21153579.4A patent/EP3833024B1/en active Active
- 2013-04-12 PL PL21153568T patent/PL3833022T3/pl unknown
- 2013-04-12 DK DK21153568.7T patent/DK3833022T3/da active
- 2013-04-12 DK DK21153576.0T patent/DK3833023T3/da active
- 2013-04-12 HR HRP20211884TT patent/HRP20211884T1/hr unknown
- 2013-04-12 PL PL21153576T patent/PL3833023T3/pl unknown
- 2013-04-12 PL PL21153579.4T patent/PL3833024T3/pl unknown
- 2013-04-12 BR BR112014024294-1A patent/BR112014024294B1/pt active IP Right Grant
- 2013-04-12 PL PL13775584T patent/PL2838261T3/pl unknown
- 2013-04-12 SI SI201331980T patent/SI3833022T1/sl unknown
- 2013-04-12 RU RU2017109867A patent/RU2658146C9/ru active
- 2013-04-12 TW TW105142270A patent/TWI578760B/zh active
- 2013-04-12 SI SI201331744T patent/SI2838261T1/sl unknown
- 2013-04-12 EP EP21153576.0A patent/EP3833023B1/en active Active
- 2013-04-12 SI SI201331952T patent/SI3716621T1/sl unknown
- 2013-04-12 DK DK13775584.9T patent/DK2838261T3/da active
- 2013-04-12 SI SI201331981T patent/SI3833023T1/sl unknown
- 2013-04-12 TW TW105142267A patent/TWI580248B/zh active
- 2013-04-12 RU RU2014145257A patent/RU2617920C9/ru active
- 2013-04-12 HR HRP20230687TT patent/HRP20230687T1/hr unknown
- 2013-04-12 TW TW102113227A patent/TWI575935B/zh active
- 2013-04-12 TW TW105142266A patent/TWI586145B/zh active
- 2013-04-12 HU HUE20174155A patent/HUE056708T2/hu unknown
- 2013-04-12 ES ES21153579T patent/ES2951398T3/es active Active
- 2013-04-12 HR HRP20220541TT patent/HRP20220541T1/hr unknown
- 2013-04-12 HR HRP20220542TT patent/HRP20220542T1/hr unknown
- 2013-04-12 EP EP21153568.7A patent/EP3833022B1/en active Active
- 2013-04-12 ES ES21153576T patent/ES2912134T3/es active Active
- 2013-04-12 ES ES21153568T patent/ES2912132T3/es active Active
- 2013-04-12 ES ES20174155T patent/ES2902536T3/es active Active
- 2013-04-12 EP EP13775584.9A patent/EP2838261B1/en active Active
- 2013-04-12 WO PCT/JP2013/002513 patent/WO2013153823A1/ja active Application Filing
- 2013-04-12 TW TW105142269A patent/TWI578759B/zh active
- 2013-04-12 HU HUE13775584A patent/HUE050755T2/hu unknown
- 2013-04-12 HU HUE21153579A patent/HUE062452T2/hu unknown
- 2013-04-12 EP EP20174155.0A patent/EP3716621B1/en active Active
-
2014
- 2014-09-05 US US14/479,139 patent/US9681134B2/en active Active
-
2016
- 2016-07-08 US US15/206,015 patent/US9667966B2/en active Active
-
2017
- 2017-04-18 US US15/490,445 patent/US9872038B2/en active Active
- 2017-04-18 US US15/490,337 patent/US9872037B2/en active Active
- 2017-04-18 US US15/490,521 patent/US9918103B2/en active Active
-
2018
- 2018-01-04 US US15/861,763 patent/US10230975B2/en active Active
- 2018-05-25 RU RU2018119332A patent/RU2682368C9/ru active
-
2019
- 2019-02-15 US US16/276,661 patent/US10523962B2/en active Active
- 2019-03-06 RU RU2019106385A patent/RU2708359C9/ru active
- 2019-03-06 RU RU2019106386A patent/RU2708360C9/ru active
- 2019-12-30 US US16/729,774 patent/US10791336B2/en active Active
-
2020
- 2020-09-25 US US17/032,531 patent/US11206421B2/en active Active
-
2021
- 2021-11-17 US US17/528,700 patent/US11831898B2/en active Active
-
2023
- 2023-10-19 US US18/490,340 patent/US20240048754A1/en active Pending
Patent Citations (1)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
JP2004129191A (ja) | 2002-10-04 | 2004-04-22 | Lg Electronics Inc | Bピクチャのダイレクトモード動きベクトル演算方法 |
Non-Patent Citations (4)
Title |
---|
BENJAMIN BROSS ET AL.: "WD5: Working Draft 5 of High-Efficiency Video Coding", JOINT COLLABORATIVE TEAM ON VIDEO CODING (JCT-VC) OF ITU-T SG16 WP3 AND ISO/IEC JTC1/SC29/WG11, JCTVC-G1103_D9, 7TH MEETING, February 2012 (2012-02-01), GENEVA, CH, pages 32 - 33, 70-73, 107-122, XP055054453 * |
JIANLE CHEN ET AL.: "MVP index parsing with fixed number of candidates", JOINT COLLABORATIVE TEAM ON VIDEO CODING (JCT-VC) OF ITU-T SG16 WP3 AND ISO/IEC JTC1/SC29/WG11, JCTVC-F402R1, 6TH MEETING, July 2011 (2011-07-01), TORINO, IT, XP030049391 * |
KEN MCCANN ET AL.: "HM5: High Efficiency Video Coding (HEVC) Test Model 5 Encoder Description", JOINT COLLABORATIVE TEAM ON VIDEO CODING (JCT- VC) OF ITU-T SG16 WP3 AND ISO/IEC JTC1/SC29/ WG11, JCTVC-G1102_DL, 7TH MEETING, January 2012 (2012-01-01), GENEVA, CH, pages 1 - 21, XP030018840 * |
TOSHIYASU SUGIO ET AL.: "Parsing Robustness for Merge/AMVP", JOINT COLLABORATIVE TEAM ON VIDEO CODING (JCT-VC) OF ITU-T SG16 WP3 AND ISO/IEC JTC1/SC29/WG11, JCTVC-F470_R4, 6TH MEETING, July 2011 (2011-07-01), TORINO, IT, pages 1 - 15, XP030049463 * |
Also Published As
Similar Documents
Publication | Publication Date | Title |
---|---|---|
WO2013153823A1 (ja) | 動画像符号化装置、動画像符号化方法、動画像符号化プログラム、送信装置、送信方法及び送信プログラム、並びに動画像復号装置、動画像復号方法、動画像復号プログラム、受信装置、受信方法及び受信プログラム | |
JP6206558B2 (ja) | 動画像復号装置、動画像復号方法、動画像復号プログラム、受信装置、受信方法及び受信プログラム | |
JP6090512B2 (ja) | 動画像符号化装置、動画像符号化方法、動画像符号化プログラム、送信装置、送信方法及び送信プログラム | |
JP6217820B2 (ja) | 動画像符号化装置、動画像符号化方法、動画像符号化プログラム、送信装置、送信方法及び送信プログラム | |
JP6229806B2 (ja) | 動画像復号装置、動画像復号方法、動画像復号プログラム、受信装置、受信方法及び受信プログラム | |
JP6644270B2 (ja) | 動画像符号化装置、動画像符号化方法、動画像符号化プログラム、送信装置、送信方法及び送信プログラム | |
WO2013065301A1 (ja) | 動画像符号化装置、動画像符号化方法、動画像符号化プログラム、送信装置、送信方法及び送信プログラム、並びに動画像復号装置、動画像復号方法、動画像復号プログラム、受信装置、受信方法及び受信プログラム | |
JP2013102260A (ja) | 動画像復号装置、動画像復号方法及び動画像復号プログラム | |
JP2013102259A (ja) | 動画像符号化装置、動画像符号化方法及び動画像符号化プログラム | |
JP2013098733A (ja) | 動画像復号装置、動画像復号方法及び動画像復号プログラム | |
JP2013098732A (ja) | 動画像符号化装置、動画像符号化方法及び動画像符号化プログラム |
Legal Events
Date | Code | Title | Description |
---|---|---|---|
121 | Ep: the epo has been informed by wipo that ep was designated in this application |
Ref document number: 13775584 Country of ref document: EP Kind code of ref document: A1 |
|
WWE | Wipo information: entry into national phase |
Ref document number: 2013775584 Country of ref document: EP |
|
NENP | Non-entry into the national phase |
Ref country code: DE |
|
REG | Reference to national code |
Ref country code: BR Ref legal event code: B01A Ref document number: 112014024294 Country of ref document: BR |
|
ENP | Entry into the national phase |
Ref document number: 2014145257 Country of ref document: RU Kind code of ref document: A |
|
ENP | Entry into the national phase |
Ref document number: 112014024294 Country of ref document: BR Kind code of ref document: A2 Effective date: 20140929 |