WO2015056735A1 - Image decoding device - Google Patents

Image decoding device Download PDF

Info

Publication number
WO2015056735A1
WO2015056735A1 PCT/JP2014/077530 JP2014077530W WO2015056735A1 WO 2015056735 A1 WO2015056735 A1 WO 2015056735A1 JP 2014077530 W JP2014077530 W JP 2014077530W WO 2015056735 A1 WO2015056735 A1 WO 2015056735A1
Authority
WO
WIPO (PCT)
Prior art keywords
picture
reference picture
prediction
unit
layer
Prior art date
Application number
PCT/JP2014/077530
Other languages
French (fr)
Japanese (ja)
Inventor
知宏 猪飼
健史 筑波
山本 智幸
Original Assignee
シャープ株式会社
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Application filed by シャープ株式会社 filed Critical シャープ株式会社
Priority to JP2015542651A priority Critical patent/JP6401707B2/en
Publication of WO2015056735A1 publication Critical patent/WO2015056735A1/en

Links

Images

Classifications

    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04NPICTORIAL COMMUNICATION, e.g. TELEVISION
    • H04N19/00Methods or arrangements for coding, decoding, compressing or decompressing digital video signals
    • H04N19/30Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using hierarchical techniques, e.g. scalability
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04NPICTORIAL COMMUNICATION, e.g. TELEVISION
    • H04N19/00Methods or arrangements for coding, decoding, compressing or decompressing digital video signals
    • H04N19/50Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using predictive coding
    • H04N19/597Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using predictive coding specially adapted for multi-view video sequence encoding
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04NPICTORIAL COMMUNICATION, e.g. TELEVISION
    • H04N19/00Methods or arrangements for coding, decoding, compressing or decompressing digital video signals
    • H04N19/70Methods or arrangements for coding, decoding, compressing or decompressing digital video signals characterised by syntax aspects related to video coding, e.g. related to compression standards

Definitions

  • the present invention relates to an image decoding apparatus that performs residual prediction between a target layer and a reference layer.
  • HEVC High-Efficiency Video Coding
  • Non-patent Document 1 There is also known a technique for generating encoded data from a plurality of moving images by encoding a plurality of mutually related moving images into layers (hierarchies). This is called an encoding technique (or scalable encoding technique).
  • MV-HEVC Multi-View HEVC
  • HEVC based on HEVC
  • Non-patent Document 2 MV-HEVC (Multi-View HEVC) based on HEVC
  • MV-HEVC supports view scalability.
  • view scalability a moving image corresponding to a plurality of different viewpoints (views) is divided into layers and encoded to generate hierarchical encoded data.
  • a moving image corresponding to a basic viewpoint (base view) is encoded as a lower layer.
  • a moving image corresponding to a different viewpoint is encoded as an upper layer after applying inter-layer prediction.
  • 3D-HEVC High Efficiency Efficiency Video Coding
  • ARP Advanced Residual Prediction
  • pictures that have been decoded at different viewpoints are used to estimate residuals in the target picture. More specifically, in the ARP, the residual of the target picture is estimated from the difference between the decoded image (refIvRefPic) of the reference layer other than the target time and the decoded image (currIvRefPic) of the reference layer at the target time.
  • Pictures P101, P102, and P103 are pictures in the target view (layer), and pictures P201, P202, and P203 are pictures of the reference view (RefViewIdx). Note that the number assigned to each picture indicates the decoding order.
  • a residual is derived from the picture corresponding to the current POC or the picture P203 specified by currIvRefPic and the picture P202 specified by the first picture or refIvRefPic in RefPicListX.
  • the picture P202 specified by refIvRefPic is not referenced from the picture P203 specified by currIvRefPic, if it is deleted from the DPB before decoding of the picture P103, The picture P202 specified by refIvRefPic cannot be used.
  • the target picture is decoded while referring to the pictures stored in the reference picture list. Since the reference picture list changes in units of slices, the position of the picture P102 (arpRefpic) in the reference picture list is in units of slices. There is also the problem that it may change. Furthermore, there is a possibility that the POC of the picture P102 to be referred to matches the POC of the picture P103 of the target picture to be decoded.
  • the present invention has been made in view of the above problems, and an object of the present invention is to determine whether there is no reference picture in the reference layer, or the location of the reference picture in the target layer changes between slices, or decoding during residual prediction.
  • An object of the present invention is to realize an image decoding apparatus or the like that can avoid a situation in which the picture order of a target picture to be matched matches the picture order of a picture in a target layer to be referred to.
  • An image decoding apparatus includes a reference picture determination unit that determines whether or not a residual prediction reference picture is usable, and a residual that performs residual prediction using the residual prediction reference picture.
  • a difference prediction application unit wherein the reference picture determination unit performs the determination according to whether or not a reference picture of the residual prediction reference layer is stored in a DPB (Decoded Picture) Buffer). .
  • DPB Decoded Picture
  • An image decoding apparatus includes a reference picture deriving unit that derives a reference picture for residual prediction, and a residual prediction applying unit that performs residual prediction using the residual prediction reference picture.
  • the reference picture deriving unit derives, as the residual prediction reference picture, a reference picture having a picture order different from the picture order of the target picture among the reference pictures included in the reference picture list.
  • An image decoding apparatus uses a reference picture selection unit that derives a reference picture for residual prediction, and the residual prediction reference picture when the residual prediction reference picture is available
  • a residual prediction application unit that performs residual prediction, scans the reference picture list in order from the top, and the absolute value of the difference between the POC of the reference picture RefPicListX [i] and the POC (PicOrderCntVal) of the target picture currPic is When the difference is smaller than the POC difference, the reference picture is set as a residual prediction reference picture.
  • An image decoding apparatus provides a reference picture derivation that derives, as a residual prediction reference picture, a reference picture having a picture order different from the picture order of the target picture among reference pictures included in a reference picture list And an inter prediction parameter decoding control unit for decoding the residual prediction flag iv_res_pred_weight_idx when the residual prediction reference picture is available.
  • the image decoding apparatus has an effect that it is possible to avoid a situation in which a decoded picture of the reference layer cannot be used during residual prediction for a target picture.
  • the image decoding apparatus according to another aspect of the present invention, it is possible to ensure that the arpRefpic POC is different from the POC of the current picture.
  • (A) is a conceptual diagram which shows an example of a reference picture list
  • (b) is a conceptual diagram which shows an example of a vector candidate.
  • A) is a figure which shows the syntax table utilized at the time of decoding of short-term reference picture set information by the entropy decoding part of the said image decoding apparatus.
  • (B) is a part of a syntax table used at the time of slice header decoding by the entropy decoding unit of the image decoding apparatus, and shows a part related to a reference picture set.
  • (A) is a part of syntax table referred to at the time of decoding of the VPS extension (vps_extension) included in the VPS, and shows a part corresponding to IL-RPS information.
  • (B) is a diagram showing a part corresponding to IL-RPS information, which is a part of a syntax table referred to at the time of slice decoding.
  • (A) is a figure which shows the relationship between the dependence type in the case where there exist inter-layer image prediction and inter-layer motion prediction in the type of inter-layer prediction, and the availability of each inter-layer prediction.
  • (B) is a diagram illustrating the relationship between sub RPS (inter-layer pixel RPS and inter-layer motion limited RPS) included in the inter-layer RPS generated in the image decoding apparatus and the dependency type. It is a flowchart showing the derivation
  • (A) is a part of the syntax table used at the time of slice header decoding by the entropy decoding unit of the image decoding apparatus, and shows a part related to the reference picture list.
  • (B) is a figure which shows the syntax table utilized at the time of decoding of reference picture list correction information by the entropy decoding part of the said image decoding apparatus.
  • FIG. 2 is a schematic diagram showing the configuration of the image transmission system 1 according to the present embodiment.
  • the image transmission system 1 is a system that transmits a code obtained by encoding a plurality of layer images and displays an image obtained by decoding the transmitted code.
  • the image transmission system 1 includes an image encoding device 11, a network 21, an image decoding device 31, and an image display device 41.
  • the signal T indicating a plurality of layer images (also referred to as texture images) is input to the image encoding device 11.
  • a layer image is an image that is viewed or photographed at a certain resolution and a certain viewpoint.
  • each of the plurality of layer images is referred to as a viewpoint image.
  • the viewpoint corresponds to the position or observation point of the photographing apparatus.
  • the plurality of viewpoint images are images taken by the left and right photographing devices toward the subject.
  • the image encoding device 11 encodes each of the signals to generate an encoded stream Te (encoded data). Details of the encoded stream Te will be described later.
  • a viewpoint image is a two-dimensional image (planar image) observed at a certain viewpoint.
  • the viewpoint image is indicated by, for example, a luminance value or a color signal value for each pixel arranged in a two-dimensional plane.
  • one viewpoint image or a signal indicating the viewpoint image is referred to as a picture.
  • the plurality of layer images include a base layer image having a low resolution and an enhancement layer image having a high resolution.
  • SNR scalable encoding is performed using a plurality of layer images
  • the plurality of layer images are composed of a base layer image with low image quality and an extended layer image with high image quality.
  • view scalable coding, spatial scalable coding, and SNR scalable coding may be arbitrarily combined.
  • encoding and decoding of an image including at least a base layer image and an image other than the base layer image is handled as the plurality of layer images.
  • the image on the reference side is referred to as a first layer image
  • the image on the reference side is referred to as a second layer image.
  • the base layer image is treated as a first layer image and the enhancement layer image is treated as a second layer image.
  • the enhancement layer image include an image of a viewpoint other than the base view and a depth image.
  • the network 21 transmits the encoded stream Te generated by the image encoding device 11 to the image decoding device 31.
  • the network 21 can be configured by, for example, the Internet, a wide area network (WAN: Wide Area Network), a small network (LAN: Local Area Network), or a combination thereof.
  • the network 21 is not necessarily limited to a bidirectional communication network, and may be a unidirectional or bidirectional communication network that transmits broadcast waves such as terrestrial digital broadcasting and satellite broadcasting.
  • the network 21 may be replaced by a storage medium that records an encoded stream Te such as a DVD (Digital Versatile Disc) or a BD (Blue-ray Disc).
  • the image decoding device 31 decodes each of the encoded streams Te transmitted by the network 21, and generates a plurality of decoded layer images Td (decoded viewpoint images Td).
  • the image display device 41 displays all or part of the plurality of decoded layer images Td generated by the image decoding device 31. For example, in view scalable coding, a 3D image (stereoscopic image) and a free viewpoint image are displayed in all cases, and a 2D image is displayed in some cases.
  • the image display device 41 includes, for example, a display device such as a liquid crystal display or an organic EL (Electro-luminescence) display.
  • a display device such as a liquid crystal display or an organic EL (Electro-luminescence) display.
  • the spatial scalable coding and SNR scalable coding when the image decoding device 31 and the image display device 41 have a high processing capability, a high-quality enhancement layer image is displayed and only a lower processing capability is provided. Displays a base layer image that does not require higher processing capability and display capability as an extension layer.
  • FIG. 3 is a diagram showing a hierarchical structure of data in the encoded stream Te.
  • the encoded stream Te illustratively includes a sequence and a plurality of pictures constituting the sequence.
  • (A) to (f) of FIG. 3 respectively show a sequence layer that defines a sequence SEQ, a picture layer that defines a picture PICT, a slice layer that defines a slice S, a slice data layer that defines slice data, and a slice data.
  • Coding Unit CU
  • sequence layer In the sequence layer, a set of data referred to by the image decoding device 31 for decoding a sequence SEQ to be processed (hereinafter also referred to as a target sequence) is defined.
  • the sequence SEQ includes a video parameter set VPS (Video Parameter Set) sequence parameter set SPS (Sequence Parameter Parameter Set), a picture parameter set PPS (Picture Parameter Parameter Set), a picture PICT, and an addition.
  • Extension information SEI SupplementallementEnhancement Information
  • the value indicated after # indicates the layer ID.
  • FIG. 3 shows an example in which encoded data of # 0 and # 1, that is, layer 0 and layer 1, exists, but the type of layer and the number of layers are not dependent on this.
  • the video parameter set VPS is a set of encoding parameters common to a plurality of moving images, a plurality of layers included in the moving image, and encoding parameters related to individual layers in a moving image composed of a plurality of layers.
  • a set is defined.
  • the sequence parameter set SPS defines a set of encoding parameters that the image decoding device 31 refers to in order to decode the target sequence. For example, the width and height of the picture are defined.
  • a set of encoding parameters referred to by the image decoding device 31 in order to decode each picture in the target sequence is defined.
  • a quantization width reference value (pic_init_qp_minus26) used for picture decoding and a flag (weighted_pred_flag) indicating application of weighted prediction are included.
  • a plurality of PPS may exist. In that case, one of a plurality of PPSs is selected from each picture in the target sequence.
  • Picture layer In the picture layer, a set of data referred to by the image decoding device 31 for decoding a picture PICT to be processed (hereinafter also referred to as a target picture) is defined. As shown in FIG. 3B, the picture PICT includes slices S0 to SNS-1 (NS is the total number of slices included in the picture PICT).
  • slice layer In the slice layer, a set of data referred to by the image decoding device 31 for decoding a slice S (also referred to as a target slice) to be processed is defined. As shown in FIG. 3C, the slice S includes a slice header SH and slice data SDATA.
  • the slice header SH includes a coding parameter group that the image decoding device 31 refers to in order to determine a decoding method of the target slice.
  • the slice type designation information (slice_type) that designates the slice type is an example of an encoding parameter included in the slice header SH.
  • I slice using only intra prediction at the time of encoding (2) P slice using unidirectional prediction or intra prediction at the time of encoding, (3) B-slice using unidirectional prediction, bidirectional prediction, or intra prediction at the time of encoding may be used.
  • the slice header SH may include a reference (pic_parameter_set_id) to the picture parameter set PPS included in the sequence layer.
  • the slice data SDATA includes a coding tree block (CTB: Coded ⁇ ⁇ ⁇ Tree Block) (coding tree unit CTU) as shown in FIG.
  • CTB is a block of a fixed size (for example, 64 ⁇ 64) constituting a slice, and may be called a maximum coding unit (LCU: LargegestLCording Unit).
  • the encoding tree layer defines a set of data that the image decoding device 31 refers to in order to decode the encoding tree block to be processed.
  • the coding tree unit is divided by recursive quadtree division.
  • a node having a tree structure obtained by recursive quadtree partitioning is called a coding tree.
  • An intermediate node of the quadtree is a coded quadtree tree (CQT: Coded Quad Tree), and the CTU is defined as including the highest CQT.
  • the CQT includes a split flag (split_flag). When the split_flag is 1, the CQT is divided into four CQTs (including four CQTs). When split_flag is 0, CQT includes a coding unit (CU: Coded Unit) that is a terminal node.
  • the encoding unit CU is a basic unit of the encoding process.
  • the encoding unit layer defines a set of data referred to by the image decoding device 31 in order to decode the encoding unit to be processed.
  • the encoding unit includes a CU header CUH, a prediction tree, a conversion tree, and a CU header CUH.
  • the CU header CUH it is defined whether the coding unit is a unit using intra prediction or a unit using inter prediction.
  • the encoding unit is the root of a prediction tree (prediction tree; PT) and a transformation tree (transform tree; TT).
  • the CU header CUH is included between the prediction tree and the conversion tree or after the conversion tree.
  • the coding unit is divided into one or a plurality of prediction blocks, and the position and size of each prediction block are defined.
  • the prediction block is one or a plurality of non-overlapping areas constituting the coding unit.
  • the prediction tree includes one or a plurality of prediction blocks obtained by the above division.
  • Prediction processing is performed for each prediction block.
  • a prediction block which is a unit of prediction is also referred to as a prediction unit (PU, prediction unit).
  • Intra prediction is prediction within the same picture
  • inter prediction refers to prediction processing performed between different pictures (for example, between display times and between layer images).
  • the division method is encoded by part_mode of encoded data, and 2N ⁇ 2N (the same size as the encoding unit), 2N ⁇ N, 2N ⁇ nU, 2N ⁇ nD, N ⁇ 2N, nL X2N, nRx2N, and NxN.
  • 2N ⁇ nU indicates that a 2N ⁇ 2N encoding unit is divided into two regions of 2N ⁇ 0.5N and 2N ⁇ 1.5N in order from the top.
  • 2N ⁇ nD indicates that a 2N ⁇ 2N encoding unit is divided into two regions of 2N ⁇ 1.5N and 2N ⁇ 0.5N in order from the top.
  • nL ⁇ 2N indicates that a 2N ⁇ 2N encoding unit is divided into two regions of 0.5N ⁇ 2N and 1.5N ⁇ 2N in order from the left.
  • nR ⁇ 2N indicates that a 2N ⁇ 2N encoding unit is divided into two regions of 1.5N ⁇ 2N and 0.5N ⁇ 1.5N in order from the left. Since the number of divisions is one of 1, 2, and 4, PUs included in the CU are 1 to 4. These PUs are expressed as PU0, PU1, PU2, and PU3 in order.
  • the encoding unit is divided into one or a plurality of transform blocks, and the position and size of each transform block are defined.
  • the transform block is one or a plurality of non-overlapping areas constituting the encoding unit.
  • the conversion tree includes one or a plurality of conversion blocks obtained by the above division.
  • the division in the transformation tree includes the one in which an area having the same size as that of the encoding unit is assigned as the transformation block, and the one in the recursive quadtree division like the above-described division in the tree block.
  • a transform block that is a unit of transformation is also referred to as a transform unit (TU).
  • the prediction image of the prediction unit is derived by the prediction parameter attached to the prediction unit.
  • the prediction parameters include a prediction parameter for intra prediction or a prediction parameter for inter prediction.
  • prediction parameters for inter prediction inter prediction (inter prediction parameters) will be described.
  • the inter prediction parameter includes prediction list use flags predFlagL0 and predFlagL1, reference picture indexes refIdxL0 and refIdxL1, and vectors mvL0 and mvL1.
  • the prediction list use flags predFlagL0 and predFlagL1 are flags indicating whether or not reference picture lists called L0 list and L1 list are used, respectively, and a reference picture list corresponding to a value of 1 is used.
  • prediction list use flag information can also be expressed by an inter prediction flag inter_pred_idc described later.
  • a prediction list use flag is used in a prediction image generation unit and a prediction parameter memory described later, and an inter prediction flag inter_pred_idc is used when decoding information on which reference picture list is used from encoded data. It is done.
  • Syntax elements for deriving inter prediction parameters included in the encoded data include, for example, a partition mode part_mode, a merge flag merge_flag, a merge index merge_idx, an inter prediction flag inter_pred_idc, a reference picture index refIdxLX, a prediction vector index mvp_LX_idx, and a difference There is a vector mvdLX.
  • FIG. 4A is a conceptual diagram illustrating an example of a reference picture list.
  • the reference picture list 601 five rectangles arranged in a line on the left and right indicate reference pictures, respectively.
  • the codes P1, P2, Q0, P3, and P4 shown in order from the left end to the right are codes indicating respective reference pictures.
  • P such as P1 indicates the viewpoint P
  • Q of Q0 indicates a viewpoint Q different from the viewpoint P.
  • the subscripts P and Q indicate the picture order number POC.
  • a downward arrow directly below refIdxLX indicates that the reference picture index refIdxLX is an index that refers to the reference picture Q0 stored in the DPB 3061 of the decoded picture management unit 306.
  • FIG. 5 is a conceptual diagram illustrating an example of a reference picture.
  • the horizontal axis indicates the display time
  • the vertical axis indicates the viewpoint.
  • the rectangles shown in FIG. 5 with 2 rows and 3 columns (6 in total) indicate pictures.
  • the rectangle in the second column from the left in the lower row indicates a picture to be decoded (target picture), and the remaining five rectangles indicate reference pictures.
  • a reference picture Q0 indicated by an upward arrow from the target picture is a picture that has the same display time as the target picture and a different viewpoint. In the displacement prediction based on the target picture, the reference picture Q0 is used.
  • a reference picture P1 indicated by a left-pointing arrow from the target picture is a past picture at the same viewpoint as the target picture.
  • a reference picture P2 indicated by a right-pointing arrow from the target picture is a future picture at the same viewpoint as the target picture. In motion prediction based on the target picture, the reference picture P1 or P2 is used.
  • Inter prediction flag and prediction list use flag The relationship between the inter prediction flag and the prediction list use flags predFlagL0 and predFlagL1 can be mutually converted as follows. Therefore, as an inter prediction parameter, a prediction list use flag may be used, or an inter prediction flag may be used. In addition, hereinafter, the determination using the prediction list use flag may be replaced with the inter prediction flag. Conversely, the determination using the inter prediction flag can be performed by replacing the prediction list use flag.
  • >> is a right shift
  • is a left shift.
  • the prediction parameter decoding (encoding) method includes merge prediction (merge) mode and AMVP (Adaptive ⁇ Motion Vector Prediction, adaptive motion vector prediction) mode.
  • the merge flag merge_flag identifies these It is a flag to do.
  • the prediction parameter of the target PU is derived using the prediction parameter of the already processed block.
  • the merge prediction mode is a mode that uses the prediction parameter already derived without including the prediction list use flag predFlagLX (inter prediction flag inter_pred_idcinter_pred_idc), the reference picture index refIdxLX, and the vector mvLX in the encoded data.
  • the prediction flag inter_pred_idcinter_pred_idc, the reference picture index refIdxLX, and the vector mvLX are included in the encoded data.
  • the vector mvLX is encoded as a prediction vector index mvp_LX_idx indicating a prediction vector and a difference vector (mvdLX).
  • the inter prediction flag inter_pred_idc is data indicating the type and number of reference pictures, and takes one of the values Pred_L0, Pred_L1, and Pred_Bi.
  • Pred_L0 and Pred_L1 indicate that reference pictures stored in reference picture lists called an L0 list and an L1 list are used, respectively, and that both use one reference picture (single prediction). Prediction using the L0 list and the L1 list are referred to as L0 prediction and L1 prediction, respectively.
  • Pred_Bi indicates that two reference pictures are used (bi-prediction), and indicates that two reference pictures stored in the L0 list and the L1 list are used.
  • the prediction vector index mvp_LX_idx is an index indicating a prediction vector
  • the reference picture index refIdxLX is an index indicating a reference picture stored in the reference picture list.
  • LX is a description method used when L0 prediction and L1 prediction are not distinguished.
  • refIdxL0 is a reference picture index used for L0 prediction
  • refIdxL1 is a reference picture index used for L1 prediction
  • refIdx (refIdxLX) is a notation used when refIdxL0 and refIdxL1 are not distinguished.
  • the merge index merge_idx is an index indicating which one of the prediction parameter candidates (merge candidates) derived from the processed block is used as the prediction parameter of the decoding target block.
  • the vector mvLX includes a motion vector and a displacement vector (disparity vector).
  • a motion vector is a positional shift between the position of a block in a picture at a certain display time of a layer and the position of the corresponding block in a picture of the same layer at a different display time (for example, an adjacent discrete time). It is a vector which shows.
  • the displacement vector is a vector indicating a positional shift between the position of a block in a picture at a certain display time of a certain layer and the position of a corresponding block in a picture of a different layer at the same display time.
  • the pictures in different layers may be pictures from different viewpoints or pictures with different resolutions.
  • a displacement vector corresponding to pictures of different viewpoints is called a disparity vector.
  • a vector mvLX A prediction vector and a difference vector related to the vector mvLX are referred to as a prediction vector mvpLX and a difference vector mvdLX, respectively.
  • Whether the vector mvLX and the difference vector mvdLX are motion vectors or displacement vectors is determined using a reference picture index refIdxLX attached to the vectors.
  • FIG. 6 is a schematic diagram illustrating a configuration of the image decoding device 31 according to the present embodiment.
  • the image decoding device 31 includes an entropy decoding unit 301, a prediction parameter decoding unit 302, a decoded picture management unit (reference image storage unit, frame memory) 306, a predicted image generation unit 308, an inverse quantization / inverse DCT unit 311, and an addition unit 312 and a residual storage unit 313 (residual recording unit).
  • the prediction parameter decoding unit 302 includes an inter prediction parameter decoding unit 303 and an intra prediction parameter decoding unit 304.
  • the predicted image generation unit 308 includes an inter predicted image generation unit 309 and an intra predicted image generation unit 310.
  • the entropy decoding unit 301 performs entropy decoding on the encoded stream Te input from the outside, and separates and decodes individual codes (syntax elements).
  • the separated codes include prediction information for generating a prediction image and residual information for generating a difference image.
  • the entropy decoding unit 301 outputs a part of the separated code to the prediction parameter decoding unit 302.
  • Some of the separated codes are, for example, prediction mode PredMode, split mode part_mode, merge flag merge_flag, merge index merge_idx, inter prediction flag inter_pred_idcinter_pred_idc, reference picture index refIdxLX, prediction vector index mvp_LX_idx, difference vector mvdLX, residual prediction flag iv_res_pred_weight_idx. Control of which code to decode is performed based on an instruction from the prediction parameter decoding unit 302.
  • the entropy decoding unit 301 outputs the quantization coefficient to the inverse quantization / inverse DCT unit 311.
  • the quantization coefficient is a coefficient obtained by performing quantization by performing DCT (Discrete Cosine Transform) on the residual signal in the encoding process.
  • the inter prediction parameter decoding unit 303 refers to the prediction parameter stored in the prediction parameter memory 3067 of the decoded picture management unit 306 based on the code input from the entropy decoding unit 301 and decodes the inter prediction parameter.
  • the inter prediction parameter decoding unit 303 outputs the decoded inter prediction parameters to the prediction image generation unit 308 and stores them in the prediction parameter memory 3067 of the decoded picture management unit 306. Details of the inter prediction parameter decoding unit 303 will be described later.
  • the intra prediction parameter decoding unit 304 refers to the prediction parameter stored in the prediction parameter memory 3067 of the decoded picture management unit 306 based on the code input from the entropy decoding unit 301 and decodes the intra prediction parameter.
  • the intra prediction parameter is a parameter used in a process of predicting a picture block within one picture, for example, an intra prediction mode IntraPredMode.
  • the intra prediction parameter decoding unit 304 outputs the decoded intra prediction parameter to the prediction image generation unit 308 and stores it in the prediction parameter memory 3067 of the decoded picture management unit 306.
  • the intra prediction parameter decoding unit 304 may derive different intra prediction modes depending on luminance and color difference.
  • the intra prediction parameter decoding unit 304 decodes the luminance prediction mode IntraPredModeY as the luminance prediction parameter and the color difference prediction mode IntraPredModeC as the color difference prediction parameter.
  • the luminance prediction mode IntraPredModeY is a 35 mode and corresponds to planar prediction (0), DC prediction (1), and direction prediction (2 to 34).
  • the color difference prediction mode IntraPredModeC uses one of planar prediction (0), DC prediction (1), direction prediction (2, 3, 4), and LM mode (5).
  • the decoded picture management unit 306 stores the reference picture block (reference picture block) generated by the addition unit 312 in the DPB 3061 at a predetermined position for each decoding target picture and block, and at a predetermined timing.
  • a decoded viewpoint image Td in which blocks are integrated for each picture is output to the outside.
  • the decoded picture management unit 306 stores the prediction parameter in the prediction parameter memory 3067 at a predetermined position for each picture and block to be decoded. Details of the decoded picture management unit 306 will be described later with reference to FIG.
  • the prediction image generation unit 308 receives the prediction mode predMode input from the entropy decoding unit 301 and the prediction parameter from the prediction parameter decoding unit 302. Further, the predicted image generation unit 308 reads the reference picture from the DPB 3061 of the decoded picture management unit 306. The predicted image generation unit 308 generates a predicted picture block P (predicted image) using the input prediction parameter and the read reference picture in the prediction mode indicated by the prediction mode predMode.
  • the inter prediction image generation unit 309 uses the inter prediction parameter input from the inter prediction parameter decoding unit 303 and the read reference picture to perform the prediction picture block P by inter prediction. Is generated.
  • the predicted picture block P corresponds to the prediction unit PU.
  • the PU corresponds to a part of a picture composed of a plurality of pixels as a unit for performing the prediction process as described above, that is, a decoding target block on which the prediction process is performed at a time.
  • the inter prediction image generation unit 309 decodes a reference picture list (RPS: ReferenceSPicture List) (L0 list or L1 list) whose prediction list use flag predFlagLX is 1 from a reference picture indicated by a reference picture index refIdxLX.
  • the reference picture block at the position indicated by the vector mvLX is read from the DPB 3061 of the decoded picture management unit 306 with reference to the target block.
  • the inter prediction image generation unit 309 performs prediction on the read reference picture block to generate a prediction picture block P.
  • the inter prediction image generation unit 309 outputs the generated prediction picture block P to the addition unit 312.
  • the intra predicted image generation unit 310 performs intra prediction using the intra prediction parameter input from the intra prediction parameter decoding unit 304 and the read reference picture. Specifically, the intra-predicted image generation unit 310 extracts a reference picture block that is a decoding target picture and is in a predetermined range from the decoding target block among the already decoded blocks from the DPB 3061 of the decoded picture management unit 306. read out.
  • the predetermined range is, for example, any of the left, upper left, upper, and upper right adjacent blocks when the decoding target block sequentially moves in a so-called raster scan order, and varies depending on the intra prediction mode.
  • the raster scan order is an order in which each row is sequentially moved from the left end to the right end in each picture from the upper end to the lower end.
  • the intra predicted image generation unit 310 performs prediction in the prediction mode indicated by the intra prediction mode IntraPredMode for the read reference picture block, and generates a predicted picture block.
  • the intra predicted image generation unit 310 outputs the generated predicted picture block P to the addition unit 312.
  • the intra prediction image generation unit 310 performs planar prediction (0), DC prediction (1), direction according to the luminance prediction mode IntraPredModeY.
  • a prediction picture block of luminance is generated according to any of prediction (2 to 34), and planar prediction (0), DC prediction (1), direction prediction (2, 3, 4), LM according to the color difference prediction mode IntraPredModeC
  • a color difference prediction picture block is generated in any one of modes (5).
  • the inverse quantization / inverse DCT unit 311 inversely quantizes the quantization coefficient input from the entropy decoding unit 301 to obtain a DCT coefficient.
  • the inverse quantization / inverse DCT unit 311 performs inverse DCT (Inverse Discrete Cosine Transform) on the obtained DCT coefficient to calculate a decoded residual signal.
  • the inverse quantization / inverse DCT unit 311 outputs the calculated decoded residual signal to the addition unit 312 and the residual storage unit 313.
  • the adder 312 outputs the prediction picture block P input from the inter prediction image generation unit 309 and the intra prediction image generation unit 310 and the signal value of the decoded residual signal input from the inverse quantization / inverse DCT unit 311 for each pixel. Addition to generate a reference picture block.
  • the adding unit 312 stores the generated reference picture block (ie, decoded picture) in the DPB 3061 of the decoded picture management unit 306.
  • FIG. 7 is a schematic diagram illustrating a configuration of the inter prediction parameter decoding unit 303 according to the present embodiment.
  • the inter prediction parameter decoding unit 303 includes an inter prediction parameter decoding control unit 3031, an AMVP prediction parameter derivation unit 3032, an addition unit 3035, and a merge prediction parameter derivation unit 3036.
  • the inter prediction parameter decoding control unit 3031 instructs the entropy decoding unit 301 to decode a code (syntax element) related to inter prediction, and the code (synth For example, a partition mode part_mode, a merge flag merge_flag, a merge index merge_idx, an inter prediction flag inter_pred_idcinter_pred_idc, a reference picture index refIdxLX, a prediction vector index mvp_LX_idx, a difference vector mvdLX, and a residual prediction flag iv_res_pred_weight_idx are extracted.
  • a code syntax element
  • the code for example, a partition mode part_mode, a merge flag merge_flag, a merge index merge_idx, an inter prediction flag inter_pred_idcinter_pred_idc, a reference picture index refIdxLX, a prediction vector index mvp_LX_idx,
  • the inter prediction parameter decoding control unit 3031 first extracts a merge flag from the encoded data.
  • the inter prediction parameter decoding control unit 3031 expresses that a certain syntax element is to be extracted, it means that the entropy decoding unit 301 is instructed to decode a certain syntax element, and the corresponding syntax element is read from the encoded data. To do.
  • the inter prediction parameter decoding control unit 3031 extracts the merge index merge_idx as a prediction parameter related to merge prediction.
  • the inter prediction parameter decoding control unit 3031 outputs the extracted merge index merge_idx to the merge prediction parameter derivation unit 3036.
  • the inter prediction parameter decoding control unit 3031 uses the entropy decoding unit 301 to extract the AMVP prediction parameter from the encoded data.
  • AMVP prediction parameters include an inter prediction flag inter_pred_idc, a reference picture index refIdxLX, a vector index mvp_LX_idx, and a difference vector mvdLX.
  • the inter prediction parameter decoding control unit 3031 outputs the prediction list use flag predFlagLX derived from the extracted inter prediction flag inter_pred_idcinter_pred_idc and the reference picture index refIdxLX to the AMVP prediction parameter derivation unit 3032 and the prediction image generation unit 308 (FIG. 6). Further, it is stored in the DPB 3061 (FIG. 9) of the decoded picture management unit 306.
  • the inter prediction parameter decoding control unit 3031 outputs the extracted vector index mvp_LX_idx to the AMVP prediction parameter derivation unit 3032.
  • the inter prediction parameter decoding control unit 3031 outputs the extracted difference vector mvdLX to the addition unit 3035.
  • FIG. 23 is a schematic diagram showing the configuration of the merge prediction parameter deriving unit 3036 according to this embodiment.
  • the merge prediction parameter derivation unit 3036 includes a merge candidate derivation unit 30361 and a merge candidate selection unit 30362.
  • the merge candidate derivation unit 30361 includes a merge candidate storage unit 303611, an extended merge candidate derivation unit 303612, and a basic merge candidate derivation unit 303613.
  • the merge candidate storage unit 303611 stores the merge candidates input from the extended merge candidate derivation unit 303612 and the basic merge candidate derivation unit 303613.
  • the merge candidate includes a prediction list use flag predFlagLX, a vector mvLX, and a reference picture index refIdxLX.
  • an index is assigned to the stored merge candidates according to a predetermined rule. For example, “0” is assigned as an index to the merge candidate input from the extended merge candidate derivation unit 303612.
  • the extended merge candidate derivation unit 303612 includes a displacement vector acquisition unit 3036122, an inter-layer merge candidate derivation unit 3036121, a displacement merge candidate derivation unit 3036123, and a BVSP merge candidate derivation unit 3036124 (not shown).
  • the displacement vector acquisition unit 3036122 first acquires displacement vectors in order from a plurality of candidate blocks adjacent to the decoding target block (for example, blocks adjacent to the left, upper, and upper right). Specifically, the displacement vector acquisition unit 3036122 selects one of the candidate blocks, and uses the reference picture index refIdxLX of the candidate block to determine whether the selected candidate block vector is a displacement vector or a motion vector. If there is a displacement vector determined using a reference layer determination function (described later), it is set as the displacement vector. If there is no displacement vector in the candidate block, the next candidate block is scanned in order.
  • the displacement vector acquisition unit 3036122 When there is no displacement vector in the adjacent block, the displacement vector acquisition unit 3036122 attempts to acquire the displacement vector of the block at the position corresponding to the target block of the block included in the reference picture in the temporally different display order. When the displacement vector cannot be acquired, the displacement vector acquisition unit 3036122 sets a zero vector as the displacement vector. The displacement vector acquisition unit 3036122 outputs the displacement vector to the inter-layer merge candidate derivation unit 3036121 and the displacement merge candidate derivation unit.
  • the inter-layer merge candidate derivation unit 3036121 receives the displacement vector from the displacement vector acquisition unit 3036122.
  • the inter-layer merge candidate derivation unit 3036121 selects a block indicated only by the displacement vector input from the displacement vector acquisition unit 3036122 from a picture having the same POC as the decoding target picture of another layer (eg, base layer, base view).
  • the prediction parameter which is a motion vector included in the block, is read from the prediction parameter memory 307. More specifically, the prediction parameter read by the inter-layer merge candidate derivation unit 3036121 is a prediction parameter of a block including coordinates obtained by adding a displacement vector to the coordinates of the starting point when the center point of the target block is the starting point. .
  • the coordinates (xRef, yRef) of the reference block are the coordinates of the target block (xP, yP), the displacement vector (mvDisp [0], mvDisp [1]), and the width and height of the target block are nPSW, nPSH. Is derived by the following equation.
  • xRef Clip3 (0, PicWidthInSamples L -1, xP + ((nPSW-1) >> 1) + ((mvDisp [0] + 2) >> 2))
  • yRef Clip3 (0, PicHeightInSamples L -1, yP + ((nPSH-1) >> 1) + ((mvDisp [1] + 2) >> 2))
  • the inter-layer merge candidate derivation unit 3036121 determines whether the prediction parameter is a motion vector as false (not a displacement vector) in a determination method using a reference layer determination function described later included in the inter prediction parameter decoding control unit 3031. Judgment is made by the method.
  • the inter-layer merge candidate derivation unit 3036121 outputs the read prediction parameter as a merge candidate to the merge candidate storage unit 303611. Further, when the prediction parameter cannot be derived, the inter-layer merge candidate deriving unit 3036121 outputs that fact to the displacement merge candidate deriving unit.
  • This merge candidate is a motion prediction inter-layer candidate (inter-view candidate) and is also described as an inter-layer merge candidate (motion prediction).
  • the displacement merge candidate derivation unit 3036123 receives the displacement vector from the displacement vector acquisition unit 3036122.
  • the displacement merge candidate derivation unit 3036123 merges the input displacement vector and the reference picture index refIdxLX of the previous layer image indicated by the displacement vector (for example, the index of the base layer image having the same POC as the decoding target picture) as a merge candidate.
  • the data is output to the candidate storage unit 303611.
  • This merge candidate is a displacement prediction inter-layer candidate (inter-view candidate) and is also described as an inter-layer merge candidate (displacement prediction).
  • the BVSP merge candidate derivation unit 3036124 derives block view synthesis prediction (Block View Synthesis Prediction) merge candidates.
  • a BVSP merge candidate is a type of displacement merge candidate that generates a predicted image from another viewpoint image, but is a merge candidate that divides the PU into smaller blocks and performs a predicted image generation process.
  • the basic merge candidate derivation unit 303613 includes a spatial merge candidate derivation unit 3036131, a temporal merge candidate derivation unit 3036132, a merge merge candidate derivation unit 3036133, and a zero merge candidate derivation unit 3036134.
  • the spatial merge candidate derivation unit 3036131 reads the prediction parameters (prediction list use flag predFlagLX, vector mvLX, reference picture index refIdxLX) stored in the prediction parameter memory 307 according to a predetermined rule, and uses the read prediction parameters as merge candidates.
  • the prediction parameter to be read is a prediction parameter relating to each of the blocks within a predetermined range from the decoding target block (for example, all or a part of the blocks in contact with the lower left end, upper left upper end, and upper right end of the decoding target block, respectively). is there.
  • the derived merge candidates are stored in the merge candidate storage unit 303611.
  • the temporal merge candidate derivation unit 3036132 reads the prediction parameter of the block in the reference image including the lower right coordinate of the decoding target block from the prediction parameter memory 307 and sets it as a merge candidate.
  • the reference picture designation method may be, for example, the reference picture index refIdxLX designated in the slice header, or may be designated using the smallest reference picture index refIdxLX of the block adjacent to the decoding target block. .
  • the derived merge candidates are stored in the merge candidate storage unit 303611.
  • the merge merge candidate derivation unit 3036133 derives merge merge candidates by combining two different derived merge candidate vectors and reference picture indexes already derived and stored in the merge candidate storage unit 303611 as L0 and L1 vectors, respectively. To do.
  • the derived merge candidates are stored in the merge candidate storage unit 303611.
  • the zero merge candidate derivation unit 3036134 derives a merge candidate in which the reference picture index refIdxLX is 0 and both the X component and the Y component of the vector mvLX are 0.
  • the derived merge candidates are stored in the merge candidate storage unit 303611.
  • FIG. 8 shows an example of merge candidates derived by the merge candidate deriving unit 30361. If two merge candidates have the same prediction parameter, excluding the processing of reducing the order, the merge index order, layer merge candidate (lower left), spatial merge candidate (upper right), spatial merge candidate (upper right), Displacement merge candidates, BVSP merge candidates, spatial merge candidates (lower left), spatial merge candidates (upper left), and temporal merge candidates. Further, there are a merge merge candidate and a zero merge candidate thereafter, which are omitted in FIG.
  • the merge candidate selection unit 30362 selects, from the merge candidates stored in the merge candidate storage unit 303611, a merge candidate to which an index corresponding to the merge index merge_idx input from the inter prediction parameter decoding control unit 3031 is assigned. As an inter prediction parameter.
  • the merge candidate selection unit 30362 stores the selected merge candidate in the prediction parameter memory 3067 (FIG. 9) of the decoded picture management unit 306 and outputs it to the prediction image generation unit 308 (FIG. 6).
  • the displacement vector acquisition unit 3036122 extracts the displacement vector from the prediction parameter memory 3067 of the decoded picture management unit 306, refers to the prediction parameter memory 3067, A prediction flag predFlagLX, a reference picture index refIdxLX, and a vector mvLX of a block adjacent to the PU are read out.
  • the displacement vector acquisition unit 3036122 has a reference layer determination function.
  • the displacement vector acquisition unit 3036122 sequentially reads prediction parameters of blocks adjacent to the target PU, and determines whether or not the adjacent block has a displacement vector from the reference picture index of the adjacent block by the reference layer determination function. When the adjacent block includes a displacement vector, the displacement vector acquisition unit 3036122 outputs the displacement vector. If there is no displacement vector in the prediction parameter of the adjacent block, the zero vector is output as the displacement vector.
  • Reference layer determination function The reference layer determination function of the displacement vector acquisition unit 3036122 is based on the input reference picture index refIdxLX, and reference layer information reference_layer_info indicating the relationship between the reference picture indicated by the reference picture index refIdxLX and the target picture Is to determine.
  • Reference layer information reference_layer_info is information indicating whether the vector mvLX to the reference picture is a displacement vector or a motion vector.
  • Prediction when the target picture layer and the reference picture layer are the same layer is called the same layer prediction, and the vector obtained in this case is a motion vector.
  • Prediction when the target picture layer and the reference picture layer are different layers is called inter-layer prediction, and the vector obtained in this case is a displacement vector.
  • the AMVP prediction parameter deriving unit 3032 is a vector (motion vector or displacement vector) stored in the prediction parameter memory 3067 (FIG. 9) of the decoded picture management unit 306 based on the reference picture index refIdx. ) As a vector candidate mvpLX.
  • the vector read out by the AMVP prediction parameter derivation unit 3032 is a block within a predetermined range from the decoding target block (for example, all or a part of the blocks in contact with the lower left end, upper left upper end, and upper right end of the decoding target block, respectively). It is a vector concerning.
  • the AMVP prediction parameter derivation unit 3032 selects a vector candidate indicated by the vector index mvp_LX_idx input from the inter prediction parameter decoding control unit 3031 among the read vector candidates as a prediction vector mvpLX.
  • the AMVP prediction parameter derivation unit 3032 outputs the selected prediction vector mvpLX to the addition unit 3035.
  • FIG. 4B is a conceptual diagram showing an example of vector candidates.
  • a prediction vector list 602 illustrated in FIG. 4B is a list including a plurality of vector candidates derived by the AMVP prediction parameter deriving unit 3032.
  • five rectangles arranged in a line on the left and right indicate areas indicating prediction vectors, respectively.
  • the downward arrow directly below the second mvp_LX_idx from the left end and mvpLX below the mvp_LX_idx indicate that the vector index mvp_LX_idx is an index referring to the vector mvpLX in the prediction parameter memory 3067 of the decoded picture management unit 306.
  • the candidate vector is a block for which the decoding process has been completed, and is generated based on a vector related to the referenced block with reference to a block (for example, an adjacent block) in a predetermined range from the decoding target block.
  • the adjacent block has a block that is spatially adjacent to the target block, for example, the left block and the upper block, and a block that is temporally adjacent to the target block, for example, the same position as the target block, and has a different display time. Contains blocks derived from blocks.
  • the addition unit 3035 adds the prediction vector mvpLX input from the prediction vector selection unit 3034 and the difference vector mvdLX input from the inter prediction parameter decoding control unit to calculate a vector mvLX.
  • the adding unit 3035 outputs the calculated vector mvLX to the predicted image generation unit 308 (FIG. 6).
  • FIG. 9 is a functional block diagram illustrating the configuration of the decoded picture management unit 306.
  • the decoded picture management unit 306 includes a DPB 3061, an RPS derivation unit 3062, a reference picture control unit 3063, a reference layer picture control unit 3064, an RPL derivation unit 3065, an output control unit 3066, and a prediction parameter memory 3067.
  • the DPB 3061 is also called a decoded picture buffer (DecodedDePicture Buffer), and records the decoded picture of each picture of the target layer output from the adding unit 312.
  • the decoded pictures corresponding to the pictures in the target layer are recorded in association with the output order (POC: “Picture” Order “Count”).
  • POC output order
  • a reference mark and an output mark can be set for each DPB picture.
  • the reference mark is information indicating whether or not the picture on the DPB can be used for predicted image generation processing (for example, inter prediction or inter-layer image prediction) in decoding processing after the target picture.
  • reference marks are "short-term reference use” ("used for-short-term reference”), “long-term reference use” ("used for-long-term reference”), “reference not used” ("not usedfor One of the values of “reference”).
  • a reference mark may be set to the value of “inter-layer reference use” (“used for inter-layer reference”). Further, without distinguishing between “short-term reference use” and “long-term reference use”, the union of both may be defined as “reference use” (“used for reference”).
  • the output mark is information indicating whether or not it is necessary to output the picture on the DPB to the outside. Specifically, the output mark takes one of the values “output required” (“needed for output”) and “output unnecessary” (“not needed for output”). Note that the reference mark and the output mark may not be explicitly set at a specific timing of the decoding process or the encoding process. In this case, the reference mark or the output mark is determined as “undefined”.
  • the RPS deriving unit 3062 derives an RPS (Reference Picture Set) to be used for decoding the target picture based on the input syntax value, and the base reference picture control unit 144 To the reference picture control unit 143 and the RPL deriving unit 3065.
  • RPS Reference Picture Set
  • RPS generally represents a set of reference pictures that can be used in decoding processing of a target picture or decoding processing of pictures following the target picture in decoding order.
  • RPS can be divided into multiple sub RPSs based on the nature of the reference picture.
  • the RPS is composed of the following five types of sub-RPSs.
  • Forward short-term RPS a reference picture specified by a relative position in display order with respect to the target picture, including a reference picture that belongs to the same layer as the target picture and whose display order is earlier than the target picture Sub RPS.
  • Rear short-term RPS a reference picture specified by a relative position in the display order with respect to the target picture, which belongs to the same layer as the target picture and includes a reference picture whose display order is after the target picture Sub RPS.
  • Long-term RPS A sub-RPS that is a reference picture specified by an absolute position in display order and includes a reference picture belonging to the same layer as the target picture.
  • Inter-layer pixel RPS a sub-RPS that includes a reference picture that belongs to a different layer from the target picture and whose pixel value is referred to in inter-layer prediction.
  • Inter-layer motion limited RPS a sub-RPS that includes a reference picture that belongs to a layer different from the target picture and that does not refer to a pixel value by referring to motion information in inter-layer prediction.
  • short-term RPS forward short-term RPS and backward short-term RPS. That is, the short-term RPS is a reference picture specified by a relative position in display order with respect to the target picture, and includes a picture belonging to the same layer as the target picture.
  • inter-layer RPS The union of inter-layer pixel RPS and inter-layer motion limited RPS is also referred to as inter-layer RPS. That is, the inter-layer RPS includes a reference picture (inter-layer reference picture) belonging to a layer different from the target picture.
  • Short-term RPS The syntax for short-term RPS (forward short-term RPS and backward short-term RPS) is SPS short-term RPS information that is short-term reference picture set information included in the SPS, and short-term reference picture set information included in the slice header. There is SH short-term RPS information.
  • the SPS short-term RPS information includes information on a plurality of short-term RPS candidates that can be selected as short-term SPS in each picture referring to the SPS.
  • the short-term RPS is a set of pictures that can be a reference picture (short-term reference picture) specified by a relative position with respect to the target picture (for example, a POC difference from the target picture).
  • FIG. 16 illustrates a part of the SPS syntax table used at the time of SPS decoding.
  • the part (A) in FIG. 16 corresponds to SPS short-term RPS information.
  • the SPS short-term RPS information includes the number of short-term RPS included in the SPS (num_short_term_ref_pic_sets) and the definition of each short-term RPS (short_term_ref_pic_set (i)).
  • FIG. 17A illustrates a short-term RPS syntax table used at the time of SPS decoding and slice header decoding.
  • the short-term RPS information includes the number of short-term reference pictures whose display order is earlier than the target picture (num_negative_pics) and the number of short-term reference pictures whose display order is later than the target picture (num_positive_pics).
  • a short-term reference picture whose display order is earlier than the target picture is referred to as a front short-term reference picture
  • a short-term reference picture whose display order is later than the target picture is referred to as a rear short-term reference picture.
  • the short-term RPS information includes, for each forward short-term reference picture, the absolute value of the POC difference with respect to the target picture (delta_poc_s0_minus1 [i]), and the presence / absence of use as a reference picture of the target picture (used_by_curr_pic_s0_flag [ i]).
  • the absolute value of the POC difference with respect to the target picture delta_poc_s1_minus1 [i]
  • the possibility of being used as the reference picture of the target picture used_by_curr_pic_s1_flag [i]
  • the SH short-term RPS information includes information on a single short-term RPS that can be used from a picture referring to a slice header.
  • FIG. 17B illustrates a part of the slice header syntax table used when decoding the slice header.
  • the part (A) in FIG. 17B corresponds to the SH short-term RPS information.
  • the SH short-term RPS information includes a flag (short_term_ref_pic_set_sps_flag) indicating whether the short-term RPS is selected from among the short-term RPS candidates decoded by the SPS or explicitly included in the slice header.
  • an identifier short_term_ref_pic_set_idx
  • Short-term RPS Short-term RPS
  • forward short-term RPS A picture that can be referred to the current picture specified by the SPS short-term RPS information or the SH short-term RPS information, and includes a picture whose display order is earlier than the target picture.
  • Backward short-term RPS A picture that can be referred to the current picture specified by SPS short-term RPS information or SH short-term RPS information, and includes a picture whose display order is later than the target picture.
  • Subsequent reference short-term RPS includes pictures that are not referenced in the current picture but can be referenced in pictures that follow the current picture in decoding order.
  • a forward short-term RPS (ListStCurrBefore), a backward short-term RPS (ListStCurrAfter), and a subsequent reference short-term RPS (ListStFoll) are derived by the following procedure. Note that the front short-term RPS, the rear short-term RPS, and the subsequent reference short-term RPS are set to be empty before the start of the following processing. (S101) Based on the SPS short-term RPS information and the SH short-term RPS information, short-term RPS information used for decoding the target picture is specified.
  • short_term_ref_pic_set_sps included in the SH short-term RPS information when the value of short_term_ref_pic_set_sps included in the SH short-term RPS information is 0, the short-term RPS explicitly transmitted in the slice header included in the SH short-term RPS information is selected. Other than that (when the value of short_term_ref_pic_set_sps is 1, the short-term RPS indicated by short_term_ref_pic_set_idx included in the SH short-term RPS information is selected from a plurality of short-term RPS candidates included in the SPS short-term RPS information. (S102) The POC of each reference picture included in the selected short-term RPS is derived.
  • the reference picture POC is derived by subtracting the value of “delta_poc_s0_minus1 [i] +1” from the POC of the target picture.
  • the reference picture belongs to the backward short-term RPS it is derived by adding the value of “delta_poc_s1_minus1 [i] +1” to the POC of the target picture.
  • the forward reference pictures included in the short-term RPS are confirmed in the order of transmission, and when the associated used_by_curr_pic_s0_flag [i] value is 1, the forward reference picture is added to ListStCurrBefore.
  • the syntax related to long-term RPS includes SPS long-term RPS information that is long-term reference picture information included in the SPS and SH long-term RPS information that is long-term reference picture information included in the slice header.
  • the SPS long-term RPS information includes information on a plurality of long-term reference pictures that can be used from each picture referring to the SPS.
  • the long-term reference picture is a reference picture specified by an absolute position (for example, POC) in the sequence.
  • the SPS long-term RPS information includes a flag (long_term_ref_pics_present_flag) indicating whether or not long-term reference pictures are transmitted in SPS, the number of long-term reference pictures (num_long_term_ref_pics_sps) transmitted in SPS, and information on each long-term reference picture.
  • a flag long_term_ref_pics_present_flag
  • the long-term reference picture information includes the POC of the reference picture (lt_ref_pic_poc_lsb_sps [i]) and the presence / absence of the possibility that the long-term reference picture is referenced in the target picture (used_by_curr_pic_lt_sps_flag [i]).
  • the POC of the reference picture may be the POC value associated with the reference picture, or the LSB of the POC (Least Significant Bit), that is, the POC divided by a predetermined power of two. The remainder value may be used.
  • the SH long-term RPS information includes information on a long-term reference picture that can be used from a picture that references a slice header.
  • SH long-term RPS information is included in the slice header when use of a long-term reference picture is indicated by a flag (long_term_ref_pic_present_flag).
  • SPS long-term RPS information includes one or more long-term reference pictures (num_long_term_ref_pics_sps> 0)
  • the number of reference pictures (num_long_term_sps) that can be referred to by the target picture among the long-term reference pictures decoded by SPS is the SH long-term RPS information. included.
  • the number of long-term reference pictures (num_long_term_pics) explicitly transmitted in the slice header is included in the SH long-term RPS information.
  • information (lt_idx_sps [i]) for selecting the num_long_term_sps number of long-term reference pictures from the long-term reference pictures included in the SPS long-term RPS information is included in the SH long-term RPS information.
  • the number of reference pictures POC poc_lsb_lt [i]
  • the presence / absence of use as a reference picture of the target picture used_by_curr_pic_lt_flag
  • a long-term RPS is derived from long-term RPS information.
  • a subsequent reference long-term RPS is also derived.
  • Long-term RPS Contains a picture that can be referred to the current picture specified by SPS long-term RPS information or SH long-term RPS information.
  • Subsequent reference long-term RPS includes a reference picture that is not referenced in the current picture but can be referenced in a picture that follows the current picture in decoding order.
  • the long-term RPS (ListLtCurr) and subsequent reference long-term RPS (ListLtFoll) are derived by the following procedure. Note that the long-term RPS and the subsequent reference long-term RPS are set to be empty before starting the following processing.
  • S201 Based on the SPS long-term RPS information and the SH long-term RP information, the long-term reference picture used for decoding the current picture is specified. Specifically, num_long_term_sps reference pictures are selected from the reference pictures included in the SPS long-term RPS information and added to the long-term RPS. The selected reference picture is the reference picture indicated by lt_idx_sps [i].
  • Inter-layer RPS The syntax related to inter-layer pixel RPS and inter-layer motion limited RPS includes IL-RPS information (inter-layer RPS information, inter-layer reference picture set information).
  • the IL-RPS information includes information on an inter-layer reference picture that can be referred to in inter-layer prediction from a picture including a slice header.
  • the IL-RPS information will be described with reference to FIGS. 18A and 18B.
  • (A) in FIG. 18 is a part of a syntax table that is referred to when the VPS extension (vps_extension) included in the VPS is decoded, and corresponds to the IL-RPS information.
  • the VPS includes max_one_active_ref_layer_flag, direct_dep_type_len_minus2, and direct_dependency_type [i] [j], which are syntaxes included in the IL-RPS information.
  • the syntax max_one_active_ref_layer_flag is a flag indicating whether the maximum value of the layer referred to when decoding an arbitrary picture of an arbitrary layer is 1 or less. When the maximum value is 1 or less, the flag value is set to 1. When the maximum value is other than (the maximum value is 2 or more), the flag value is set to 0.
  • the syntax direct_dep_type_len_minus2 is a value representing the number of bits of the syntax direct_dependency_type [i] [j].
  • the number of bits of direct_dependency_type [i] [j] is (direct_dep_type_len_minus2 + 2).
  • the syntax direct_dependency_type [i] [j] is a value indicating the type of inter-layer prediction that can be used when referring to the layer indicated by “j” from the layer indicated by “i”.
  • direct_dependency_type [i] [j] is also referred to as a dependency type when referring to a reference layer (layer j) from a target layer (layer i).
  • layer i and layer j are omitted, and are also referred to as dependency types (direct_dependency_type).
  • FIG. 19A shows the relationship between the dependency type and inter-layer prediction availability when there are inter-layer image prediction and inter-layer motion prediction as the types of inter-layer prediction.
  • dependency type When the dependency type is “0”, it is pixel-dependent and motion-dependent.
  • dependency type When the dependency type is “1”, it is pixel-dependent and not motion-dependent.
  • dependency type When the dependency type is “2”, it is motion-dependent and not pixel-dependent.
  • the target layer i can use the pixel of the reference layer j for prediction. For example, it can be used for inter-layer image prediction.
  • the target layer i can use the motion information (motion vector and reference picture index) of the reference layer j for prediction.
  • motion information motion vector and reference picture index
  • inter-layer motion prediction is a process of generating a predicted image of a target picture using pixel values of a decoded image of a reference layer.
  • inter-layer motion prediction is processing for generating a predicted image of a target picture by directly or indirectly using reference layer motion information (motion vector, reference picture index, inter prediction type).
  • the dependency type when the dependency type is “0”, it means that both the decoded pixel of the reference layer (the pixel value of the decoded image) and the motion information may be used in the decoding process of the target picture.
  • the dependency type when the dependency type is “1”, it means that the decoded pixel of the reference layer may be used, but the motion information may not be used.
  • the dependency type is “2”, it means that the motion information of the reference layer may be used, but the decoded pixel may not be used.
  • the dependency type indicates that the decoded pixel of the reference layer is referenced (in the above definition, the dependency type is “0” or “1”)
  • the dependency type indicates pixel dependency.
  • the dependency type indicates that the decoded pixel of the reference layer is not referred to (in the above definition, the dependency type is “2”)
  • the dependency type indicates pixel independence.
  • the dependency type indicates that the motion information of the reference layer is referred to (in the above definition, the dependency type is “0” or “2”)
  • the dependency type indicates motion dependency.
  • the dependency type indicates that the decoded pixel of the reference layer is not referred to (in the above definition, the dependency type is “2”)
  • the dependency type indicates motion independence.
  • the motion dependency flag MotionEnableFlag [i] [j] indicating motion dependency can be derived from the following equation.
  • the pixel dependency flag SampleEnableFlag [i] [j] and the motion dependency flag MotionEnableFlag [i] [j] are derived from the following equations.
  • a dependency type other than pixel dependency and motion dependency may be indicated depending on the dependency type.
  • the type of dependence can be block division information, transform coefficient information (such as the presence or absence of transform coefficients), loop filter information, and the like.
  • the pixel dependence flag and the motion dependence flag can be derived by the above-described formulas and the like, and flags indicating whether or not there is an additional dependence (for example, XXX dependence flag XXXEnableFlag, XXX is a block Division information, transform coefficient information, loop filter information, etc.) can be derived from the following equations.
  • XXXEnableFlag [i] [j] (direct_dependency_type [i] [j] + 1) & 4)
  • 2 (B) of FIG. 18 is a part of the syntax table referred to when the slice header is decoded, and corresponds to the IL-RPS information.
  • the IL-RPS information includes an inter-layer prediction enabled flag (inter_layer_pred_enabled_flag). Furthermore, when the inter-layer prediction valid flag is 1 (inter-layer prediction is valid) and the number of reference layers that can be referred from the target picture (NumDirectRefLayers [nuh_layer_id]) is greater than 1, the number of inter-layer reference pictures is set.
  • the expressing syntax (num_inter_layr_ref_pics_minus1) is included in the IL-RPS information.
  • the number of reference pictures between active layers (NumActiveRefLayerPics) is set to a value of “num_inter_layer_ref_pics_minus1 + 1”.
  • the number of reference pictures between active layers corresponds to the number of reference pictures between layers that can be referred to by inter-layer prediction in the target picture.
  • a layer identifier (inter_layer_pred_layer_idc [i]) indicating a layer to which each inter-layer reference picture belongs is included in the IL-RPS information.
  • each syntax included in the IL-RPS information may be omitted if it is obvious. For example, when the number of inter-layer reference pictures that can be referenced from one picture is limited to one, the syntax related to the number of inter-layer reference pictures is not necessary.
  • Inter-layer RPS derivation process Inter-layer RPS, that is, inter-layer pixel RPS and inter-layer motion limited RPS are derived from IL-RPS information.
  • FIG. 19B is a diagram illustrating the relationship between the sub RPS (inter-layer pixel RPS and inter-layer motion limited RPS) included in the inter-layer RPS and the dependency type.
  • the inter-layer RPS includes two sub-RPSs, an inter-layer pixel RPS and an inter-layer motion limited RPS.
  • the inter-layer pixel RPS includes a reference picture whose dependency type is “0” and an inter-layer reference picture whose dependency type is “1”.
  • the inter-layer motion limited RPS includes an inter-layer reference picture whose dependency type is “2”.
  • the inter-layer pixel RPS includes an inter-layer reference picture that can refer to a decoded pixel.
  • the inter-layer motion limited RPS does not include an inter-layer reference picture that may refer to a decoded pixel, but includes an inter-layer reference picture that may refer to motion information.
  • FIG. 20 is a flowchart showing a derivation process of sub RPS (inter-layer pixel RPS and inter-layer motion limited RPS) included in the inter-layer RPS.
  • the list IL-RPS0 representing the inter-layer pixel RPS and the list IL-RPS1 representing the inter-layer motion limited RPS are respectively set to be empty.
  • the inter-layer pixel RPS and inter-layer motion limited RPS having the properties described with reference to FIG. 19B can be derived based on the IL-RPS information.
  • Inter-layer pixel RPS and inter-layer motion limited RPS have specific characteristics, i.e., inter-layer pixel RPS includes an inter-layer reference picture to which a decoded pixel may be referenced, and inter-layer motion limited RPS includes a decoded pixel
  • inter-layer motion limited RPS includes a decoded pixel
  • Different methods may be used as long as the property that an inter-layer reference picture that may be referred to is not included and an inter-layer reference picture that motion information may be referred to is included is satisfied.
  • the sub-RPS of the inter-layer pixel RPS and the inter-layer motion limited RPS is derived.
  • the inter-layer pixel RPS and the inter-layer pixel-independent RPS sub-RPS may be derived. In this case, the determinations in S303 and S305 are replaced with the following determinations S303r2 and S305r2, respectively.
  • the reference picture control unit 3063 updates the DPB 3061 based on the input RPS.
  • the reference picture control unit 3063 performs “reference use” (“short-term reference use” or “short-term reference use” or a reference mark of a picture that is indicated by the inter prediction of the target picture (current picture) in the input RPS. Set to “Long term reference use”).
  • a decoded picture of the target layer recorded in the DPB which is not marked as “reference use” in the above process, is set to “reference nonuse”.
  • the reference mark of the inter-layer reference picture on the DPB is not changed by the reference picture control unit 3063.
  • the reference mark change of the picture on the DPB derived from the base decoded picture is not performed by the reference picture control unit 3063 but by the reference layer picture control unit 3064 described later.
  • the reference layer picture control unit 3064 updates the DPB 3061 based on the input decoded picture and RPS of the reference layer.
  • the reference layer picture control unit 3064 records, in the DPB, a decoded picture of a reference layer corresponding to a picture that can be referred to in inter-layer inter prediction of the target picture (current picture) in the input RPS.
  • the reference mark of the picture is set to “use reference” (“use of short-term reference” or “use of long-term reference” on the DPB.
  • the output mark of the picture is set to “output unnecessary” on the DPB.
  • the decoded picture of the reference layer when recording the decoded picture of the reference layer, it may be recorded in the picture buffer after applying scaling and filtering as necessary, in particular, the resolution of the output picture of the reference layer and the target layer is different. In this case (in the case of spatial scalability), it is necessary to scale the decoded picture of the reference layer according to the resolution of the output picture of the target layer.
  • the RPL deriving unit 3065 uses a reference picture list to be used for inter prediction or inter-layer prediction of the target slice of the target picture based on the input RPS and the RPL information included in the input syntax value. Derived and output.
  • the RPL information is a syntax value that is decoded from the SPS or slice header in order to construct the reference picture list RPL.
  • the RPL information includes SPS list correction information and SH list correction information.
  • the SPS list modification information is information included in the SPS, and is information related to restrictions on modification of the reference picture list.
  • the SPS list correction information will be described with reference to FIG. 16 again.
  • the part (C) in FIG. 16 corresponds to SPS list correction information.
  • a flag restricted_ref_pic_lists_flag
  • a flag whether or not information related to list rearrangement exists in the slice header ( lists_modification_present_flag).
  • the SH list correction information is information included in the slice header, and the update information of the length of the reference picture list (reference list length) applied to the target picture, and the reordering information of the reference picture list (reference list reordering information) ) Is included.
  • the SH list correction information will be described with reference to FIG. FIG. 21A illustrates a part of a slice header syntax table used when decoding a slice header.
  • the part (C) in FIG. 21A corresponds to the SH list correction information.
  • the reference list length update information includes a flag (num_ref_idx_active_override_flag) indicating whether or not the list length is updated.
  • a flag (num_ref_idx_active_override_flag) indicating whether or not the list length is updated.
  • information (num_ref_idx_l0_active_minus1) indicating the reference list length after the change of the L0 reference list and information (num_ref_idx_l1_active_minus1) indicating the reference list length after the change of the L1 reference list are included.
  • FIG. 21B illustrates a syntax table of reference list rearrangement information used at the time of decoding the slice header.
  • the reference list rearrangement information includes an L0 reference list rearrangement presence / absence flag (ref_pic_list_modification_flag_l0).
  • the L0 reference list rearrangement order (list_entry_l0 [i]) is included in the reference list rearrangement information.
  • NumPocTotalCurr is a variable representing the number of reference pictures that can be used in the current picture. Therefore, the L0 reference list rearrangement order is included in the slice header only when the L0 reference list is rearranged and the number of reference pictures available in the current picture is larger than two.
  • the L1 reference list rearrangement presence / absence flag (ref_pic_list_modification_flag_l1) is included in the reference list rearrangement information.
  • the L1 reference list rearrangement order (list_entry_l1 [i]) is included in the reference list rearrangement information.
  • the L1 reference list rearrangement order is included in the slice header only when the L1 reference list is rearranged and the number of reference pictures available in the current picture is larger than two.
  • the reference picture list deriving unit generates a reference picture list RPL used for decoding the target picture based on the reference picture set RPS and the RPL correction information.
  • FIG. 22 is a diagram showing an outline of the temporary L0 reference list and the temporary L1 reference list generated in the intermediate process of the L0 reference list and L1 reference list derivation in the RPL deriving unit 3065.
  • the provisional L0 reference list includes a forward short-term RPS (StBef in the figure), an inter-layer pixel RPS (ILSample in the figure), and a rear in order from the top of the list (in order of priority).
  • Sub RPSs are included in the order of short-term RPS (StAft in the figure), long-term RPS (Lt in the figure), and inter-layer motion limited RPS (ILMotion in the figure).
  • the provisional L1 reference list includes a backward short-term RPS (StAft in the figure), forward short-term RPS (StBef in the figure), and long-term RPS in order from the top of the list (in order of priority).
  • StAft backward short-term RPS
  • StBef forward short-term RPS
  • long-term RPS long-term RPS in order from the top of the list (in order of priority).
  • Lt in the drawing inter-layer pixel RPS
  • IMotion in the drawing inter-layer motion limited RPS
  • the provisional reference list (the provisional reference L0 reference list and the provisional L1 reference list) generated by the RPL deriving unit 3065 is a position closer to the top of the list of the inter-layer pixel RPS than the inter-layer motion limited RPS. Is included.
  • the provisional reference list includes the inter-layer pixel RPS at a position corresponding to a higher priority than the inter-layer motion limited RPS.
  • the L0 reference list is constructed according to the procedure shown in S401 to S409 below.
  • RefPicList0 [rIdx] of the L0 reference list corresponding to the reference picture index rIdx is derived by the following equation.
  • RefListTemp0 [i] represents the i-th element of the provisional L0 reference list.
  • RefPicList0 [rIdx] RefPicListTemp0 [list_entry_l0 [rIdx]]
  • the value recorded at the position indicated by the reference picture index rIdx is referred to, and the reference recorded at the position of the value in the provisional L0 reference list
  • the picture is stored as a reference picture at the position of rIdx in the L0 reference list.
  • the provisional L0 reference list is set as the L0 reference list.
  • L1 reference list construction procedure Next, the L1 reference list construction procedure will be described.
  • the construction of the L1 reference list is executed according to the procedure described in S501 to S509 below.
  • the elements of the provisional L1 reference list are rearranged based on the value of the reference list rearrangement order list_entry_l1 ⁇ ⁇ ⁇ [i] to obtain an L1 reference list.
  • the element RefPicList1 [rIdx] of the L1 reference list corresponding to the reference picture index rIdx is derived by the following equation.
  • RefListTemp1 [i] represents the i-th element of the provisional L1 reference list.
  • RefPicList1 [rIdx] RefPicListTemp1 [list_entry_l1 [rIdx]]
  • the value recorded at the position indicated by the reference picture index rIdx is referred to, and the reference recorded at the position of the value in the provisional L1 reference list
  • the picture is stored as a reference picture at the position of rIdx in the L1 reference list.
  • the provisional L1 reference list is set as the L1 reference list.
  • the reference picture list selects and selects the corresponding provisional reference list (provisional reference list L0 and provisional reference list L1) based on the RPL correction information. Generated by sorting.
  • the provisional reference list includes the inter-layer pixel RPS at a position corresponding to a higher priority than the inter-layer motion limited RPS.
  • the provisional reference list includes the inter-layer pixel RPS at a position closer to the top of the list than the inter-layer motion limited RPS.
  • the inter-layer reference picture included in the inter-layer pixel RPS is closer to the top of the reference picture list than the inter-layer reference picture included in the inter-layer motion limited RPS ( Associated with the smaller reference picture index). Therefore, the inter-layer reference picture included in the inter-layer pixel RPS can be designated with a reference picture index having a smaller value than the inter-layer reference picture included in the inter-layer motion limited RPS.
  • the inter-layer reference picture included in the inter-layer pixel RPS is an inter-layer reference picture that may be used for inter-layer pixel prediction.
  • the inter-layer reference picture included in the inter-layer motion limited RPS is an inter-layer reference picture that may not be used for inter-layer pixel prediction but may be used for inter-layer motion prediction.
  • a larger number of reference picture indexes for specifying reference pictures used for inter-layer image prediction are included than reference picture indexes for specifying reference pictures used for inter-layer motion prediction.
  • the former is included for each prediction unit, while the latter is included for each slice. Therefore, the inter-layer reference picture included in the inter-layer pixel RPS is assigned a smaller reference picture index than the inter-layer reference picture included in the inter-layer motion limited RPS so that it can be specified with a smaller code amount.
  • the code amount of the entire encoded data can be reduced.
  • the position in the provisional reference list is designated by the reference list reordering information (list_entry_l0, list_entry_l1), thereby reordering the provisional reference list to generate a reference list.
  • the code amount of the RPL correction information it is possible to reduce the code amount of the RPL correction information by making it possible to specify a reference picture that is likely to be rearranged at the head of the reference picture list with reference list rearrangement information having a smaller value.
  • the inter-layer reference picture included in the inter-layer pixel RPS is less likely to be rearranged at the top of the reference picture list, compared to the inter-layer reference picture included in the inter-layer motion limited RPS. Since it is near the top of the list, it can be specified with reference list sorting information with a smaller value. Therefore, the code amount of the RPL correction information can be reduced by using the provisional reference list derived by the above procedure.
  • the above provisional reference list includes a motion limited RPS at the end of the provisional reference list.
  • the provisional reference list includes the motion limited RPS at a position closer to the back of the list than the short-term RPS.
  • the inter-layer reference pictures included in the motion-limited reference list have a lower selection frequency than the reference pictures included in the short-term RPS and long-term RPS that are sub-RPSs related to inter prediction in the same layer. The information related to the selection of the reference picture can be decoded with a smaller code amount.
  • the reference picture list derivation process (RPL derivation) is performed.
  • the inter-layer motion limited RPS is read as an inter-layer pixel-independent RPS and processed. More specifically, the processes of S406 and S506 are replaced with the following processes.
  • the provisional reference list includes the inter-layer pixel RPS at a position corresponding to a higher priority than the inter-layer pixel-independent RPS.
  • the temporary reference list includes the inter-layer pixel RPS at a position closer to the top of the list than the inter-layer pixel-independent RPS.
  • the provisional reference list includes a pixel-independent limited RPS at the end of the provisional reference list.
  • the provisional reference list includes a pixel-independent RPS at a position closer to the back of the list than the short-term RPS.
  • an inter-layer reference picture included in an inter-layer pixel RPS can be designated with a smaller code amount by assigning a smaller reference picture index than an inter-layer reference picture included in an inter-layer pixel-independent RPS.
  • the code amount of the entire encoded data can be reduced.
  • the output control unit 3066 generally outputs the DPB 3061 picture to the outside and updates the output mark at a predetermined timing. Specifically, the picture output processing by the output control unit 3066 is executed in the following procedure.
  • a picture with a minimum POC is output from the pictures on the DPB whose output mark is “output required”.
  • the output mark of the output picture is set to “output unnecessary”.
  • a picture whose reference mark is “reference not used” and whose output mark is “output unnecessary” is selected from the pictures on the DPB, and the picture is deleted from the DPB.
  • the prediction parameter memory 3067 includes inter prediction parameters decoded by the inter prediction parameter decoding unit 303, intra prediction parameters decoded by the intra prediction parameter decoding unit 304, and prediction modes separated by the entropy decoding unit 301.
  • predMode is stored.
  • the inter prediction parameters stored in the prediction parameter memory 3067 include, for example, a prediction list use flag predFlagLX (inter prediction flag inter_pred_idcinter_pred_idc), a reference picture index refIdxLX, and a vector mvLX.
  • FIG. 10 is a schematic diagram illustrating a configuration of the inter predicted image generation unit 309 according to the present embodiment.
  • the inter predicted image generation unit 309 includes a motion displacement compensation unit 3091, a residual prediction unit 3092, an illuminance compensation unit 3093, a weight prediction unit 3094, and a reference image determination unit 3095.
  • the motion displacement compensation unit 3091 uses the reference picture index refIdxLX from the DPB 3061 of the decoded picture management unit 306.
  • a motion displacement compensation image is generated by reading out a block at a position shifted by the vector mvLX starting from the position of the target block of the designated reference picture.
  • a motion displacement compensation image is generated by applying a filter called a motion compensation filter (or displacement compensation filter) for generating a pixel at a decimal position.
  • the above processing is called motion compensation
  • the vector mvLX is a displacement vector
  • it is called displacement compensation
  • it is collectively referred to as motion displacement compensation.
  • the motion displacement compensation image for L0 prediction is referred to as predSamplesL0
  • the motion displacement compensation image for L1 prediction is referred to as predSamplesL1. If the two are not distinguished, they are called predSamplesLX.
  • predSamplesLX the motion displacement compensation image for L0 prediction
  • predSamplesL1 the motion displacement compensation image for L1 prediction
  • These output images are also referred to as motion displacement compensation images predSamplesLX.
  • the input image is expressed as predSamplesLX and the output image is expressed as predSamplesLX ′.
  • the reference image determination unit 3095 uses a reference picture (decoded image) (refIvRefPic) in a reference layer (reference view) (hereinafter referred to as “reference layer reference picture refIvRefPic” for convenience of explanation) used for residual prediction. It is determined whether or not (refer) is available. When the reference layer reference picture refIvRefPic is available, the reference image determination unit 3095 sets 1 to the reference layer reference picture use flag refIvRefPicAvailable (refIvRefPicAvailable2).
  • the reference image determination unit 3095 sets the reference layer reference picture use flag refIvRefPicAvailable (refIvRefPicAvailable2) to 0.
  • the reference image determination unit 3095 outputs the reference layer reference picture use flag refIvRefPicAvailable (refIvRefPicAvailable2) set in this way to the residual prediction unit 3092.
  • the reference image determination unit 3095 may determine whether or not the reference picture can be used as follows.
  • the reference picture in the target layer is referred to as an ARP reference picture arpRefPic.
  • an index indicating the view identifier is described as ViewIdx
  • the view identifier of the reference layer (view) in the coordinates of the target block (xP, yP) is described as refViewIdx [xP] [yP].
  • PicOrderCnt (X) indicates the picture order number POC of picture X.
  • the index arpRefIdxLX may be set to 0, or the method shown in option Y5 of the fourth embodiment described later may be used.
  • PicOrderCnt (Pic) is PicOrderCnt (arpRefPic), and ViewIdx of Pic is refViewIdx [xP] [yP].
  • the reference picture Pic that satisfies the condition (1-1) is present in the RPS of the picture on the reference layer whose POC is the same as the target picture and whose ViewIdx is refViewIdx [xP] [yP] To do.
  • RPS Since RPS is invariant among a plurality of slices included in a picture and exists even when the slice type is I slice, determination by RPS is appropriate.
  • (Option 2) Determine whether or not a reference picture exists in the RPL. If there is a reference picture Pic that satisfies the above (1-1) and the following (2-2), the reference picture Pic is used as the reference layer reference picture refIvRefPic. And the reference layer reference picture usage flag refIvRefPicAvailable is set to 1. Otherwise, the reference layer reference picture usage flag refIvRefPicAvailable is set to 0.
  • the reference picture Pic that satisfies the above condition (1-1) is the same POC as the target picture, and the slice type on the reference layer (ViewIdx is refViewIdx [xP] [yP]) slice_type) exists in the reference layer list RPL (RefPicListX) of a picture that is not an I slice.
  • the RefPicListX [] of the first slice can be used for the determination.
  • the present invention is not limited to this, and the RefPicListX [] of a predetermined slice among the slices included in the target picture may be used between the slices included in the target picture.
  • the last slice can be used.
  • the presence in the reference picture list does not exist in a specific order element of the reference picture list, but may exist in any order.
  • the L0 list RefPicList0 [] and the L1 list RefPicList1 [] as reference picture lists they may exist as elements of any list.
  • a reference picture Pic that satisfies the above condition (1-1) exists in the decoded picture buffer (DPB) having the same POC as that of the target picture and ViewIdx of refViewIdx [xP] [yP]. To do.
  • DPB 3061 Since DPB 3061 stores usable pictures, DPB may be used for determination as described above.
  • the picture storage state in the DPB may not be reliable. Specifically, the DPB is not marked as need-for-output and discards the picture that is marked for reference use, but the output timing may be decoder-dependent, so some decoders In this case, the picture is discarded, but it can be said that the picture is not discarded in other decoders. In this case, the determination of whether a reference picture (used for ARP) exists in the DPB may differ depending on the decoder. Since the mark for reference use is set according to the RPS that is explicitly decoded, there is no difference between the decoders. For this reason, as described above, it is preferable to determine whether or not the reference mark of the reference picture is “use of reference”.
  • (1-1b) may be used instead of the setting (1-1).
  • PicOrderCnt (Pic) is PicOrderCnt (arpRefPic)
  • Pic's ViewIdx is refViewIdx [xP] [yP]
  • Pic's DepthFlag is 0.
  • (1-1b) when the reference picture is a depth picture (when DepthFlag is 1), it is not set as the target reference picture Pic.
  • the residual prediction unit 3092 performs the residual prediction (ARP: Advanced Residual Prediction) on the input motion displacement compensation image predSamplesLX when the residual prediction execution flag resPredFlag is 1. When the residual prediction execution flag resPredFlag is 0, the input motion displacement compensation image predSamplesLX is output as it is.
  • ARP Advanced Residual Prediction
  • Residual prediction is an image (motion compensation) in which a residual of a reference layer (first layer image) different from a target layer (second layer image) that is a target of predicted image generation is predicted from a reference picture of the target layer. This is performed by adding to the motion displacement compensation image predSamplesLX. This is called the first residual prediction.
  • residual prediction refers to a residual at a different time (POC) from a target picture between a target layer (second layer image) and a reference layer (first layer image) that are targets of predicted image generation. It can also be performed by adding to the motion displacement compensation image predSamplesLX which is an image (displacement compensation image) predicted from the reference picture of the layer. This is called second residual prediction.
  • the picture refIvRefPic at a different time (POC) from the target picture of the already derived reference layer, and the reference layer target The residual between the pictures currIvRefPic at the same time (POC) as the picture is used as an estimate of the residual in motion compensation prediction from a picture (for example, arpRefPic) at a different time (POC) from the target picture of the target layer.
  • the residuals of the reference layer picture refIvRefPic and the inter-picture arpRefPic of the picture at the time (POC) different from the already derived target picture are set at the same time (POC) as the target picture. This is used as an estimate of the residual in displacement compensation prediction from the reference layer picture currIvRefPic.
  • both the first residual prediction and the second residual prediction are used. However, for simplification, only one of the first residual prediction and the second residual prediction is used. I do not care.
  • FIG. 1 is a block diagram showing the configuration of the residual prediction unit 3092.
  • the residual prediction unit 3092 includes a residual prediction execution flag deriving unit 30921, a reference image acquisition unit 30922, and a residual synthesis unit 30923.
  • the residual prediction execution flag deriving unit 30921 has a residual prediction execution flag resPredFlag when (1) the residual prediction flag iv_res_pred_weight_idx is not 0 and (2) the reference picture use flag is “1”. 1 is set to indicate that difference prediction is to be executed. On the other hand, when the residual prediction flag iv_res_pred_weight_idx is 0, or when the reference picture use flag is not “1” (in the case of disparity compensation), 0 is set to the residual prediction execution flag resPredFlag.
  • the residual prediction execution flag deriving unit 30921 may derive the residual prediction execution flag resPredFlag according to the following conditional expression (R-1).
  • PicOrderCntVal is the picture order number POC of the target picture.
  • the residual prediction execution flag deriving unit 30921 determines (3) whether the target block is motion compensation in addition to (1) and (2) above.
  • (PicOrderCnt (RefPicListX [refIdxLX])! PicOrderCntVal). That is, the residual prediction execution flag resPredFlag may be derived from the following conditional expression (R-2).
  • the reference image acquisition unit 30922 stores the motion vector mvLX and the residual prediction displacement vector mvDisp input from the inter prediction parameter decoding unit 303, and the decoded picture management unit 306.
  • the corresponding block currIvSamplesLX and the reference block refIvSamplesLX of the corresponding block are read.
  • FIG. 11A is a diagram for explaining the corresponding block currIvSamplesLX.
  • the corresponding block Cor1 corresponding to the target block Tar1 on the target layer has a positional relationship between the reference layer and the target layer, starting from the position of the target block Tar1 ′ of the image on the reference layer. Is located in a block at a position shifted by a displacement vector mvDisp, which is a vector indicating.
  • the reference image acquisition unit 30922 derives a pixel at a position where the coordinates (x, y) of the pixel of the target block are shifted by the displacement vector mvDisp of the target block.
  • the displacement vector mvDisp has a decimal precision of 1/4 pel
  • the residual generation unit 30922 uses the X of the pixel R0 with integer precision corresponding to the case where the coordinates of the pixel of the target block are (xP, yP).
  • XInt xPb + (mvLX [0] >> 2)
  • the coordinates xInt and Y coordinates yInt, and the fractional part xFrac of the displacement vector mvDisp and the fractional part yFrac of the Y component yInt yPb + (mvLX [1] >> 2)
  • xFrac mvLX [0] & 3
  • yFrac mvLX [1] & 3 It is derived by the following formula.
  • X & 3 is a mathematical expression for extracting only the lower 2 bits of X.
  • the reference image acquisition unit 30922 generates an interpolation pixel predPartLX [x] [y] in consideration of the fact that the displacement vector mvDisp has a pel resolution of 1/4 pel.
  • xA Clip3 (0, picWidthInSamples-1, xInt)
  • xB Clip3 (0, picWidthInSamples-1, xInt + 1)
  • xC Clip3 (0, picWidthInSamples-1, xInt)
  • xD Clip3 (0, picWidthInSamples-1, xInt + 1)
  • yA Clip3 (0, picHeightInSamples-1, yInt)
  • yB Clip3 (0, picHeightInSamples-1, yInt)
  • yC Clip3 (0, picHeightInSamples-1, yInt)
  • the integer pixel A is a pixel corresponding to the pixel R0
  • the integer pixels B, C, and D are integer precision pixels adjacent to the right, bottom, and bottom right of the integer pixel A, respectively
  • Clip3 (x, y , z) is a function that limits (clips) z to be greater than or equal to x and less than or equal to y.
  • the reference image acquisition unit 30922 includes reference pixels refPicLX [xA] [yA], refPicLX [xB] [yB], refPicLX [xC] [yC], and refPicLX [corresponding to the integer pixels A, B, C, and D, respectively.
  • xD] [yD] is read from the DPB 3061 of the decoded picture management unit 306.
  • the reference image acquisition unit 30922 includes the reference pixel refPicLX [xA] [yA], refPicLX [xB] [yB], refPicLX [xC] [yC], refPicLX [xD] [yD] and the X component of the displacement vector mvDisp.
  • An interpolated pixel predPartLX [x] [y] which is a pixel shifted by the decimal part of the displacement vector mvDisp from the pixel R0, is derived using the fractional part xFrac and the fractional part yFrac of the Y component.
  • predPartLX [x] [y] (refPicLX [xA] [yA] * (8-xFrac) * (8-yFrac) + refPicLX [xB] [yB] * (8-yFrac) * xFrac + refPicLX [xC] [yC] * (8-xFrac) * yFrac + refPicLX [xD] [yD] * xFrac * yFrac)
  • the reference image acquisition unit 30922 performs the above interpolation pixel derivation process on each pixel in the target block, and sets a set of interpolation pixels as an interpolation block predPartLX.
  • the reference image acquisition unit 30922 outputs the derived interpolation block predPartLX to the residual synthesis unit 30923 as the corresponding block currIvSamplesLX.
  • FIG. 11B is a diagram for explaining the reference block refIvSamplesLX. As shown in FIG. 11B, the reference block corresponding to the corresponding block on the reference layer is shifted to the position shifted by the motion vector mvLX of the target block from the position of the corresponding block of the reference image on the reference layer. Located in the block.
  • the reference image acquisition unit 30922 except for the process of deriving the corresponding block currIvSamplesLX and replacing the displacement vector mvDisp with a vector (mvDisp [0] + mvLX [0], mvDisp [1] + mvLX [1])
  • the corresponding block refIvSamplesLX is derived by performing the same processing.
  • the reference image acquisition unit 30922 outputs the corresponding block refIvSamplesLX to the residual synthesis unit 30923.
  • the residual synthesis unit 30923 derives a corrected motion displacement compensation image predSamplesLX ′ from the motion displacement compensation image predSamplesLX, the corresponding block currIvSamplesLX, the reference block refIvSamplesLX, and the residual prediction flag iv_res_pred_weight_idx. To do.
  • the corrected motion displacement compensation image predSamplesLX ⁇ predSamplesLX ⁇ predSamplesLX + ((currIvSamplesLX-refIvSamplesLX) >> (iv_res_pred_weight_idx-1)) It is calculated using the following formula.
  • the residual prediction execution flag resPredFlag is 0, the residual synthesis unit 30923 outputs the motion displacement compensation image predSamplesLX as it is.
  • the illumination compensation unit 3093 performs illumination compensation on the input motion displacement compensation image predSamplesLX.
  • the illumination compensation flag ic_enable_flag is 0, the input motion displacement compensation image predSamplesLX is output as it is.
  • the motion displacement compensation image predSamplesLX input to the illuminance compensation unit 3093 is an output image of the motion displacement compensation unit 3091 when the residual prediction is off, and the residual prediction unit when the residual prediction is on. 3092 is an output image.
  • the weight prediction unit 3094 generates a predicted picture block P (predicted image) by multiplying the input motion displacement image predSamplesLX by a weight coefficient.
  • the input motion displacement image predSamplesLX is an image on which residual prediction and illuminance compensation are performed.
  • the input motion displacement image predSamplesLX (LX is L0 or L1) is set to the number of pixel bits. The following formula is processed.
  • predSamples [x] [y] Clip3 (0, (1 ⁇ bitDepth)-1, (predSamplesLX [x] [y] + offset1) >> shift1)
  • shift1 14-bitDepth
  • offset1 1 ⁇ (shift1-1).
  • predFlagL0 or predFlagL1 are 1 (in the case of bi-prediction) and weight prediction is not used
  • the input motion displacement images predSamplesL0 and predSamplesL1 are averaged to obtain the number of pixel bits. The following formula is processed.
  • predSamples [x] [y] Clip3 (0, (1 ⁇ bitDepth)-1, (predSamplesL0 [x] [y] + predSamplesL1 [x] [y] + offset2) >> shift2)
  • shift2 15-bitDepth
  • offset2 1 ⁇ (shift2-1).
  • the weight prediction unit 3094 when performing weight prediction, derives the weight prediction coefficient w0 and the offset o0, and performs the following processing.
  • predSamples [x] [y] Clip3 (0, (1 ⁇ bitDepth)-1, ((predSamplesLX [x] [y] * w0 + 2log2WD-1) >> log2WD) + o0)
  • log2WD is a variable indicating a predetermined shift amount.
  • the weight prediction unit 3094 derives weight prediction coefficients w0, w1, o0, o1, and performs the processing of the following equation.
  • predSamples [x] [y] Clip3 (0, (1 ⁇ bitDepth)-1, (predSamplesL0 [x] [y] * w0 + predSamplesL1 [x] [y] * w1 + ((o0 + o1 + 1) ⁇ log2WD)) >> (log2WD + 1)) (Modification 1)
  • refIvRefPic is determined using the motion compensation level condition (resPredFlag).
  • the present invention is not limited to this, and conditions in the parsing stage may be used.
  • the reference image determination unit 3095 may determine whether or not the ARP reference picture arpRefPic can be used in the parsing stage, and supply the determination result arpRefPicAvailable to the inter prediction parameter decoding control unit 3031.
  • a flag indicating the availability of the ARP reference picture arpRefPic in the parsing stage is distinguished as arpRefPicAvailable
  • a flag indicating the availability of the ARP reference picture arpRefPic in the motion compensation stage is distinguished as refIvRefPicAvailable.
  • the reference image determination unit 3095 may determine whether or not the ARP reference picture arpRefPic can be used.
  • the inter prediction parameter decoding control unit 3031 may decode iv_res_pred_weight_idx.
  • FIG. 12 is a block diagram illustrating a configuration of the image encoding device 11 according to the present embodiment.
  • the image encoding device 11 includes a predicted image generation unit 101, a subtraction unit 102, a DCT / quantization unit 103, an entropy encoding unit 104, an inverse quantization / inverse DCT unit 105, an addition unit 106, a decoded picture management unit (reference image) (Storage unit, frame memory) 109, encoding parameter determination unit 110, prediction parameter encoding unit 111, and residual storage unit 313 (residual recording unit).
  • the prediction parameter encoding unit 111 includes an inter prediction parameter encoding unit 112 and an intra prediction parameter encoding unit 113.
  • the predicted image generation unit 101 generates a predicted picture block P for each block which is an area obtained by dividing the picture for each viewpoint of the layer image T input from the outside.
  • the predicted image generation unit 101 reads the reference picture block from the decoded picture management unit 109 based on the prediction parameter input from the prediction parameter encoding unit 111.
  • the prediction parameter input from the prediction parameter encoding unit 111 is, for example, a motion vector or a displacement vector.
  • the predicted image generation unit 101 reads the reference picture block of the block at the position indicated by the motion vector or the displacement vector predicted from the encoding target block.
  • the prediction image generation unit 101 generates a prediction picture block P using one prediction method among a plurality of prediction methods for the read reference picture block.
  • the predicted image generation unit 101 outputs the generated predicted picture block P to the subtraction unit 102. Note that since the predicted image generation unit 101 performs the same operation as the predicted image generation unit 308 already described, details of generation of the predicted picture block P are omitted.
  • the predicted image generation unit 101 calculates an error value based on a difference between a signal value for each pixel of a block included in the layer image and a signal value for each corresponding pixel of the predicted picture block P. Select the prediction method to minimize.
  • the method for selecting the prediction method is not limited to this.
  • the plurality of prediction methods are intra prediction, motion prediction, and merge prediction.
  • Motion prediction is prediction between display times among the above-mentioned inter predictions.
  • the merge prediction is a prediction that uses the same reference picture block and prediction parameter as a block that has already been encoded and is within a predetermined range from the encoding target block.
  • the plurality of prediction methods are intra prediction, motion prediction, merge prediction, and displacement prediction.
  • the displacement prediction is prediction between different layer images (different viewpoint images) in the above-described inter prediction. Furthermore, motion prediction, merge prediction, and displacement prediction. For displacement prediction (disparity prediction), there are predictions with and without additional prediction (residual prediction and illuminance compensation).
  • the prediction image generation unit 101 outputs a prediction mode predMode indicating the intra prediction mode used when generating the prediction picture block P to the prediction parameter encoding unit 111 when intra prediction is selected.
  • the prediction image generation unit 101 stores the motion vector mvLX used when generating the prediction picture block P in the decoded picture management unit 109 and outputs the motion vector mvLX to the inter prediction parameter encoding unit 112.
  • the motion vector mvLX indicates a vector from the position of the encoding target block to the position of the reference picture block when the predicted picture block P is generated.
  • the information indicating the motion vector mvLX may include information indicating a reference picture (for example, a reference picture index refIdxLX, a picture order number POC), and may represent a prediction parameter.
  • the predicted image generation unit 101 outputs a prediction mode predMode indicating the inter prediction mode to the prediction parameter encoding unit 111.
  • the prediction image generation unit 101 When the prediction image generation unit 101 selects displacement prediction, the prediction image generation unit 101 stores the displacement vector used in generating the prediction picture block P in the decoded picture management unit 109 and outputs it to the inter prediction parameter encoding unit 112.
  • the displacement vector dvLX indicates a vector from the position of the encoding target block to the position of the reference picture block when the predicted picture block P is generated.
  • the information indicating the displacement vector dvLX may include information indicating a reference picture (for example, reference picture index refIdxLX, view IDview_id) and may represent a prediction parameter.
  • the predicted image generation unit 101 outputs a prediction mode predMode indicating the inter prediction mode to the prediction parameter encoding unit 111.
  • the prediction image generation unit 101 selects merge prediction
  • the prediction image generation unit 101 outputs a merge index merge_idx indicating the selected reference picture block to the inter prediction parameter encoding unit 112. Further, the predicted image generation unit 101 outputs a prediction mode predMode indicating the merge prediction mode to the prediction parameter encoding unit 111.
  • the prediction image generation unit 101 when the residual prediction execution flag resPredFlag indicates that the residual prediction is performed, as described above, the prediction image generation unit 101 Residual prediction is performed in the included residual prediction unit 3092.
  • the subtraction unit 102 subtracts the signal value of the prediction picture block P input from the prediction image generation unit 101 for each pixel from the signal value of the corresponding block of the layer image T input from the outside, and generates a residual signal. Generate.
  • the subtraction unit 102 outputs the generated residual signal to the DCT / quantization unit 103 and the encoding parameter determination unit 110.
  • the DCT / quantization unit 103 performs DCT on the residual signal input from the subtraction unit 102 and calculates a DCT coefficient.
  • the DCT / quantization unit 103 quantizes the calculated DCT coefficient to obtain a quantization coefficient.
  • the DCT / quantization unit 103 outputs the obtained quantization coefficient to the entropy encoding unit 104 and the inverse quantization / inverse DCT unit 105.
  • the entropy coding unit 104 receives the quantization coefficient from the DCT / quantization unit 103 and the coding parameter from the coding parameter determination unit 110.
  • Input coding parameters include codes such as a reference picture index refIdxLX, a vector index mvp_LX_idx, a difference vector mvdLX, a prediction mode predMode, a residual prediction flag iv_res_pred_weight_idx, and a merge index merge_idx.
  • the entropy encoding unit 104 generates an encoded stream Te by entropy encoding the input quantization coefficient and encoding parameter, and outputs the generated encoded stream Te to the outside.
  • the inverse quantization / inverse DCT unit 105 inversely quantizes the quantization coefficient input from the DCT / quantization unit 103 to obtain a DCT coefficient.
  • the inverse quantization / inverse DCT unit 105 performs inverse DCT on the obtained DCT coefficient to calculate a decoded residual signal.
  • the inverse quantization / inverse DCT unit 105 outputs the calculated decoded residual signal to the addition unit 106.
  • the addition unit 106 adds the signal value of the predicted picture block P input from the predicted image generation unit 101 and the signal value of the decoded residual signal input from the inverse quantization / inverse DCT unit 105 for each pixel, and refers to them. Generate a picture block.
  • the adding unit 106 stores the generated reference picture block in the decoded picture management unit 109.
  • the decoded picture management unit 109 Similar to the decoded picture management unit 306 of the image decoding device 31, the decoded picture management unit 109 has a prediction parameter memory (not shown), and the prediction parameter generated by the prediction parameter encoding unit 111 in the prediction parameter memory. Are stored in a predetermined position for each picture and block to be encoded.
  • the decoded picture management unit 109 has a DPB (not shown), and encodes the reference picture block generated by the addition unit 106 in the DPB. Store in a predetermined position for each target picture and block.
  • decoded picture management unit 109 the details of the decoded picture management unit 109 are the same as those described for the decoded picture management unit 306 of the image decoding device 31, and thus the description thereof is omitted here.
  • the encoding parameter determination unit 110 selects one set from among a plurality of sets of encoding parameters.
  • the encoding parameter is a parameter to be encoded that is generated in association with the above-described prediction parameter or the prediction parameter.
  • the predicted image generation unit 101 generates a predicted picture block P using each of these sets of encoding parameters.
  • the encoding parameter determination unit 110 calculates a cost value indicating the amount of information and the encoding error for each of a plurality of sets.
  • the cost value is, for example, the sum of a code amount and a square error multiplied by a coefficient ⁇ .
  • the code amount is the information amount of the encoded stream Te obtained by entropy encoding the quantization error and the encoding parameter.
  • the square error is the sum between pixels regarding the square value of the residual value of the residual signal calculated by the subtracting unit 102.
  • the coefficient ⁇ is a real number larger than a preset zero.
  • the encoding parameter determination unit 110 selects a set of encoding parameters that minimizes the calculated cost value. As a result, the entropy encoding unit 104 outputs the selected set of encoding parameters to the outside as the encoded stream Te, and does not output the set of unselected encoding parameters.
  • the prediction parameter encoding unit 111 derives a prediction parameter used when generating a prediction picture based on the parameter input from the prediction image generation unit 101, and encodes the derived prediction parameter to generate a set of encoding parameters. To do.
  • the prediction parameter encoding unit 111 outputs the generated set of encoding parameters to the entropy encoding unit 104.
  • the prediction parameter encoding unit 111 stores, in the decoded picture management unit 109, a prediction parameter corresponding to the generated encoding parameter set selected by the encoding parameter determination unit 110.
  • the prediction parameter encoding unit 111 operates the inter prediction parameter encoding unit 112 when the prediction mode predMode input from the prediction image generation unit 101 indicates the inter prediction mode.
  • the prediction parameter encoding unit 111 operates the intra prediction parameter encoding unit 113 when the prediction mode predMode indicates the intra prediction mode.
  • the inter prediction parameter encoding unit 112 derives an inter prediction parameter based on the prediction parameter input from the encoding parameter determination unit 110.
  • the inter prediction parameter encoding unit 112 includes the same configuration as the configuration in which the inter prediction parameter decoding unit 303 (see FIG. 6 and the like) derives the inter prediction parameter as a configuration for deriving the inter prediction parameter.
  • the configuration of the inter prediction parameter encoding unit 112 will be described later.
  • the intra prediction parameter encoding unit 113 determines the intra prediction mode IntraPredMode indicated by the prediction mode predMode input from the encoding parameter determination unit 110 as a set of inter prediction parameters.
  • the inter prediction parameter encoding unit 112 is means corresponding to the inter prediction parameter decoding unit 303.
  • FIG. 13 is a schematic diagram illustrating a configuration of the inter prediction parameter encoding unit 112 according to the present embodiment.
  • the inter prediction parameter encoding unit 112 includes an inter prediction parameter encoding control unit 1031, a merge prediction parameter derivation unit 1121, an AMVP prediction parameter derivation unit 1122, a subtraction unit 1123, and a prediction parameter integration unit 1126.
  • the merge prediction parameter derivation unit 1121 has the same configuration as the merge prediction parameter derivation unit 3036 (see FIG. 7).
  • the inter prediction parameter encoding control unit 1031 instructs the entropy encoding unit 104 to decode a code related to the inter prediction (syntax element decoding), for example, a code (syntax element) included in the encoded data.
  • a code related to the inter prediction for example, a code (syntax element) included in the encoded data.
  • Merge flag merge_flag, merge index merge_idx, inter prediction flag inter_pred_idcinter_pred_idc, reference picture index refIdxLX, prediction vector index mvp_LX_idx, and difference vector mvdLX are encoded.
  • the inter prediction parameter encoding control unit 1031 includes an additional prediction flag encoding unit 10311, a merge index encoding unit 10312, a vector candidate index encoding unit 10313, and a split mode encoding unit, a merge flag encoding unit, an inter not shown.
  • a prediction flag encoding unit, a reference picture index encoding unit, and a vector difference encoding unit are configured.
  • the division mode encoding unit, the merge flag encoding unit, the merge index encoding unit, the inter prediction flag encoding unit, the reference picture index encoding unit, the vector candidate index encoding unit 10313, and the vector difference encoding unit are respectively divided modes.
  • merge flag merge_flag merge index merge_idx
  • inter prediction flag inter_pred_idcinter_pred_idc reference picture index refIdxLX
  • prediction vector index mvp_LX_idx reference picture index refIdxLX
  • difference vector mvdLX difference vector
  • the additional prediction flag encoding unit 1031 encodes the illumination compensation flag ic_enable_flag and the residual prediction flag iv_res_pred_weight_flsg to indicate whether or not additional prediction is performed.
  • the merge index merge_idx is input from the encoding parameter determination unit 110 to the merge prediction parameter derivation unit 1121 when the prediction mode predMode input from the prediction image generation unit 101 indicates the merge prediction mode.
  • the merge index merge_idx is output to the prediction parameter integration unit 1126.
  • the merge prediction parameter deriving unit 1121 reads the reference picture index refIdxLX and vector mvLX of the reference block indicated by the merge index merge_idx from the merge candidates from the decoded picture management unit 109.
  • the merge candidate is a reference block (for example, a reference block in contact with the lower left end, upper left end, and upper right end of the encoding target block) within a predetermined range from the encoding target block to be encoded, This is a reference block for which the encoding process has been completed.
  • the AMVP prediction parameter derivation unit 1122 has the same configuration as the AMVP prediction parameter derivation unit 3032 (see FIG. 7).
  • the AMVP prediction parameter derivation unit 1122 receives the vector mvLX from the encoding parameter determination unit 110 when the prediction mode predMode input from the prediction image generation unit 101 indicates the inter prediction mode.
  • the AMVP prediction parameter derivation unit 1122 derives a prediction vector mvpLX based on the input vector mvLX.
  • the AMVP prediction parameter derivation unit 1122 outputs the derived prediction vector mvpLX to the subtraction unit 1123. Note that the reference picture index refIdx and the vector index mvp_LX_idx are output to the prediction parameter integration unit 1126.
  • the subtraction unit 1123 subtracts the prediction vector mvpLX input from the AMVP prediction parameter derivation unit 1122 from the vector mvLX input from the coding parameter determination unit 110 to generate a difference vector mvdLX.
  • the difference vector mvdLX is output to the prediction parameter integration unit 1126.
  • the prediction parameter integration unit 1126 When the prediction mode predMode input from the predicted image generation unit 101 indicates the merge prediction mode, the prediction parameter integration unit 1126 outputs the merge index merge_idx input from the encoding parameter determination unit 110 to the entropy encoding unit 104. To do.
  • the prediction parameter integration unit 1126 performs the following process.
  • the prediction parameter integration unit 1126 integrates the reference picture index refIdxLX and the vector index mvp_LX_idx input from the encoding parameter determination unit 110 and the difference vector mvdLX input from the subtraction unit 1123.
  • the prediction parameter integration unit 1126 outputs the integrated code to the entropy encoding unit 104.
  • bitstream conformance The second embodiment of the present invention will be described below with reference to the drawings.
  • bitstream conformance the constraint on which the bitstream is to be imposed.
  • the prediction parameter encoding unit 111 of the image encoding device 11 may encode an ARP on / off flag at the slice level.
  • the image encoding device 11 satisfies the following conditions (B1-1) and (B1-2) as bitstream conformance.
  • PicOrderCnt (Pic) is PicOrderCnt (arpRefPic), and ViewIdx is refViewIdx [xP] [yP].
  • the image decoding device 31 decodes the encoded data having the above bitstream conformance. By imposing restrictions on the ARP reference picture as a bit stream to be decoded by the image decoding device 31, it is possible to prevent the decoding process from failing because the ARP reference picture cannot be referred to.
  • the image decoding device 31 may be configured as follows. That is, the image decoding device 31 receives the encoded stream generated according to the bitstream conformance, and sets the ARP on / off flag encoded at the slice level (slice header / segment header) to the inter prediction parameter decoding unit. In 303, decryption is performed. In addition, the residual prediction unit 3092 executes a residual prediction process according to the decoded ARP on / off flag. By encoding the ARP on / off in units of slice headers, the ARP can be turned off if the bitstream conformance cannot be observed in the reference picture list configuration of the reference layer.
  • the prediction parameter encoding unit 111 of the image encoding device 11 may encode the ARP on / off flag at the slice level.
  • the image encoding device 11 satisfies the following (B2-1) and (B2-2) as bitstream conformance.
  • PicOrderCnt (Pic) is PicOrderCnt (arpRefPic) and ViewIdx is refViewIdx [xP] [yP].
  • the image decoding device 31 decodes the encoded data having the above bitstream conformance. By imposing restrictions on the ARP reference picture as a bit stream to be decoded by the image decoding device 31, it is possible to prevent the decoding process from failing because the ARP reference picture cannot be referred to.
  • the image decoding device 31 may be configured as follows. That is, the image decoding device 31 receives the encoded stream generated according to the bit stream conformance, and the inter prediction parameter decoding unit 303 decodes the ARP on / off flag encoded at the slice level. In addition, the residual prediction unit 3092 executes a residual prediction process according to the decoded ARP on / off flag.
  • a RefPicListX [0] of a predetermined slice among slices that are not I slices of the target picture can be used.
  • RefPicListX [0] of the last slice can be used.
  • not only the 0th picture in the reference picture list but also a reference picture at a predetermined position may be targeted.
  • (Option B3) It is determined whether or not a reference picture exists in the DPB.
  • the prediction parameter encoding unit 111 of the image encoding device 11 encodes an ARP on / off flag at the slice level.
  • the image encoding device 11 satisfies the following (B3-1) and (B3-2) as bitstream conformance.
  • the image decoding device 31 decodes the encoded data having the above bitstream conformance. By imposing restrictions on the ARP reference picture as a bit stream to be decoded by the image decoding device 31, it is possible to prevent the decoding process from failing because the ARP reference picture cannot be referred to.
  • the image decoding device 31 may be configured as follows. That is, the image decoding device 31 receives the encoded stream generated according to the bit stream conformance, and the inter prediction parameter decoding unit 303 decodes the ARP on / off flag encoded at the slice level. In addition, the residual prediction unit 3092 executes a residual prediction process according to the decoded ARP on / off flag.
  • ARP reference picture arpRefPic is the reference picture RefPicListX [arpRefIdxLX] of the arpRefIdxLX of the reference picture list RefPicListX
  • the reference picture list differs depending on the picture.
  • the reference picture (arpRefPic and corresponding curIvRefPic) used for the ARP reference picture has the following bit stream conformance as (Option X1) or (Option X1 ′) so that it is the same between slices included in the target picture. It is preferable to introduce.
  • the reference picture referenced by arpRefPicL1 when performing ARP using the L1 reference picture is the same for all slices whose slice type (slice_type) is “P” To unify.
  • the image decoding device 31 decodes the encoded data having the above bitstream conformance.
  • an increase in the amount of data transfer due to the change of the ARP reference picture in the picture (such as no cache hit) Can be prevented.
  • the image decoding device 31 may be configured as follows. That is, the image decoding apparatus 31 receives the encoded stream generated according to the bit stream conformance, and the inter prediction parameter decoding unit 303 decodes the ARP flag encoded at the slice level. Further, the residual prediction unit 3092 executes a residual prediction process according to the decoded ARP flag.
  • RefPicList0 (arpRefIdxL0] is unified for all slices whose slice type (slice_type) is “P”.
  • the image decoding device 31 decodes the encoded data having the above bitstream conformance.
  • an increase in the amount of data transfer due to the change of the ARP reference picture in the picture (such as no cache hit) Can be prevented.
  • the image decoding device 31 may be configured as follows. That is, the image decoding apparatus 31 receives the encoded stream generated according to the bit stream conformance, and the inter prediction parameter decoding unit 303 decodes the ARP flag encoded at the slice level. Further, the residual prediction unit 3092 executes a residual prediction process according to the decoded ARP flag.
  • the ARP reference picture arpRefPic is limited to RefPicListX [0] of the first slice whose slice type (slice_type) is not “I” in the picture.
  • the reference image acquisition unit 30922 may acquire the ARP reference picture arpRefPic from RefPicListX [0] of the first slice whose slice type (slice_type) is not “I” in the picture. .
  • the present invention is not limited to the above, and a predetermined slice among slices other than the I slice of the target picture can be used.
  • a predetermined slice among slices other than the I slice of the target picture can be used.
  • RefPicListX [0] of the last slice between slices included in the current picture may be targeted.
  • a residual is predicted using a reference picture in a reference picture order that is a POC different from the target picture order that is a POC of the target picture.
  • this reference picture order is derived as the POC of a specific reference picture in the reference picture list (for example, the POC of the first reference picture RefPicListX [0] of the reference picture list RefPicListX [])
  • the reference picture order is the same as the target picture. May be equal. In this case, ARP cannot be used.
  • a reference picture to be used for the ARP in order to prevent an inter-layer picture having the same POC as the target picture from being referred to as a reference picture having a POC different from the target picture, a reference picture to be used for the ARP is used. The method to identify is demonstrated.
  • Option Y1 bitstream conformance regarding a reference picture used for ARP will be described.
  • the ARP reference picture arpRefPic is the reference picture RefPicListX [arpRefIdxLX] of the arpRefIdxLX of the reference picture list RefPicListX
  • the image encoding device 11 indicates that the following (Y1-1) and (Y1-2) are satisfied. Let's do it.
  • DiffPicOrderCnt (RefPicList0 [arpRefIdxL0], currPic) is not 0 and the slice type (slice_type) is not I.
  • DiffPicOrderCnt (RefPicList1 [arpRefIdxL1], currPic) is not 0 and the slice type (slice_type) is B.
  • the image decoding device 31 receives the encoded stream generated according to the bit stream conformance and executes a residual prediction process.
  • the image decoding device 31 decodes the encoded data having the above bitstream conformance. By imposing restrictions on the ARP reference picture as a bit stream to be decoded by the image decoding device 31, it is possible to prevent the decoding process from failing because the ARP reference picture cannot be referred to.
  • any one of the inter prediction parameter decoding unit 303 and the entropy decoding unit 301 may include a reference image determination unit 3095.
  • the reference image determination unit 3095 may derive the ARP reference picture use flag arpRefPicAvailable flag at the parsing stage (CU level). Further, as shown in SYN21 and SYN22 in FIG. 15, at the parsing stage, that is, in either the inter prediction parameter decoding unit 303 or the entropy decoding unit 301, the reference image determination unit 3095 determines whether or not the ARP reference picture can be used (ARP The residual prediction flag (ARP flag) iv_res_pred_weight_idx may be decoded according to the reference picture usage flag arpRefPicAvailable).
  • iv_res_pred_flag [nuh_layer_id] is not 0
  • arpRefPicAvailable is not 0
  • the reference image determination unit 3095 derives arpRefPicAvailable according to the following equation (Y2-1) in the parsing stage.
  • true (1) is set for arpRefPicAvailable when the following (Y2.1) and (Y2.2) are satisfied.
  • the inter prediction image generation unit 309 performs determination in the motion compensation stage.
  • the picture (RefPicListX [arpRefIdxLX]) of the index arpRefIdxLX of the reference picture list is referred to as an ARP reference picture arpRefPic.
  • the reference image determination unit 3095 refers to the reference picture arpRefPic depending on whether the picture order of the arpRefPic (POCPO (arpRefPic)) is equal to the picture order PicOrderCntVal of the target picture currPic.
  • Set the layer reference picture availability flag (refIvRefPicAvailable2) Specifically, if DiffPicOrderCnt (arpRefPic, currPic) is not 0, refIvRefPicAvailable2 is set to 1; otherwise, refIvRefPicAvailable2 is set to 0.
  • the reference image determination unit 3095 sets a reference layer reference picture use flag (refIvRefPicAvailable2) for the reference picture arpRefPic depending on whether the picture order of arpRefPic (PicOrderCnt (arpRefPic)) is equal to the picture order PicOrderCntVal of the target picture currPic To do. Specifically, if DiffPicOrderCnt (arpRefPic, currPic) is not 0, refIvRefPicAvailable2 is set to non-zero, otherwise refIvRefPicAvailable2 is set to 0.
  • arpRefPic is derived from the element RefPicListX [arpRefIdxLX] of the reference index arpRefIdxLX of the reference picture list RefPicListX [], it is derived by the following equation (Y3-2).
  • the residual prediction implementation flag deriving unit 30921 derives resPredFlag according to the following equation (C3-1) so that the implementation flag resPredFlag is non-zero only when the derived refIvRefPicAvailable2 is non-zero.
  • the residual prediction implementation flag deriving unit 30921 may derive the implementation flag resPredFlag using the refIvRefPicAvailable derived in the first embodiment together. Specifically, the following formula may be used by combining the formula (R-1) and the formula (Y3-1) of Embodiment 1.
  • the motion compensation image is determined according to whether the ARP reference picture arpRefPic is equal to the POC of the target picture currPic. Determine whether to perform residual prediction (resPredFlag).
  • residual prediction residual prediction
  • either the inter prediction parameter decoding unit 303 or the entropy decoding unit 301 is configured to include the reference image determination unit 3095.
  • option Y4 the determination at the parsing stage in the inter prediction parameter decoding unit 303 and the entropy decoding unit 301 and the determination at the motion compensation stage in the inter prediction image generation unit 309 are used together.
  • the reference image determination unit 3095 derives the arpRefPicAvailable shown in FIG. 15 according to the above equation (Y2-1) and its modification. May be.
  • the inter prediction parameter decoding unit 303 and the entropy decoding unit 301 may decode the residual prediction flag (ARP flag) iv_res_pred_weight_idx according to the ARP reference picture use flag arpRefPicAvailable flag. That is, as shown in FIG. 16, iv_res_pred_weight_idx may be decoded only when the arpRefPicAvailable flag is true.
  • ARP flag residual prediction flag
  • the reference image determination unit 3095 is modified to perform the determination as follows.
  • the reference image determination unit 3095 derives refIvRefPicAvailable2 in accordance with the above equation (Y3-1) and its modifications.
  • the residual prediction execution flag deriving unit 30921 derives resPredFlag according to the equation (Y3-1) described in the third embodiment and its modification.
  • the iv_res_pred_weight_idx becomes a useless flag when the ARP reference picture arpRefPic is equal to the POC of the target picture currPic. Therefore, the amount of code can be reduced.
  • the ARP reference picture arpRefPic is the target.
  • the reference image determination unit 3095 of this embodiment also derives a reference picture arpRefPic.
  • arpRefIdxL0 and arpRefIdxL1 are set to negative values (in this case, ⁇ 1) which are invalid values as reference picture indexes, and arpRefPicAvailable is 0 indicating false. Set to.
  • arpRefIdxLX In the initial state, as the reference picture index arpRefIdxLX, a negative value (here, -1) is given, and the above search is performed. When a reference picture that satisfies the condition is obtained, arpRefIdxLX is set to a value of 0 or more. The search ends when a reference picture that satisfies the condition is obtained, that is, when arpRefIdxLX becomes 0 or more.
  • Option Y5 may also perform determination at the parsing stage in the inter prediction parameter decoding unit 303 and the entropy decoding unit 301, similarly to the option Y2 and the option Y4.
  • either the inter prediction parameter decoding unit 303 or the entropy decoding unit 301 is configured to include the reference image determination unit 3095.
  • the reference image determination unit 3095 follows the above equation (Y2-1) and the modification example (Y2-1 ′) as described above. Although it is possible to derive arpRefPicAvailable shown in FIG. 15, it is also possible to derive it using the following equation (Y5-1).
  • arpRefIdxL1> 0
  • the reference image determination unit 3095 has a value (0 or more) in which the arpRefIdxL0 that is the reference picture index of L0 is valid or a value (0 or more) in which the arpRefIdxL1 that is the reference picture index of L1 is valid.
  • arpRefPicAvailable is set to true (a value other than 0). That is, when a picture other than the inter-layer picture is found in either the L0 reference picture list or the L1 reference picture list, arpRefPicAvailable is set to true (non-zero).
  • the option Y5 can also make a determination at the motion compensation stage in the inter-predicted image generation unit 309, like the options Y3 and Y4.
  • the reference image determination unit 3095 can derive refIvRefPicAvailable2 in accordance with the above-described equation (Y3-1) and its modification as described above. It is also possible to derive in the equation (Y5-2).
  • the residual prediction execution flag deriving unit 30921 derives resPredFlag according to the equation (C3-1) described in the option Y3 and its modification.
  • the arpRefIdxLX is selected so that the arpRefPic having a POC different from the PictureOrderCntVal that is the POC of the target picture is selected. Therefore, there is an effect of eliminating the problem that the ARP does not operate.
  • the ARP reference picture arpRefPic is the target picture currPic as in the option Y2. Since iv_res_pred_weight_idx, which is a useless flag when it is equal to the POC, is not decoded, the code amount can be reduced.
  • ARP is performed similarly to the option Y3. If this is not possible, that is, if the ARP reference picture arpRefPic is equal to the POC of the target picture currPic, the ARP operation is not performed, so that an invalid operation is avoided.
  • the reference picture is determined so as to set an arpRefPic having a POC different from the PictureOrderCntVal that is the POC of the target picture.
  • the reference image determination unit 3095 may derive the ARP reference picture arpRecPic from the reference pictures included in the reference picture list RefPicListX according to the following pseudo code.
  • miniDiffPOC indicating the provisional minimum value of the POC difference is set to a sufficiently large predetermined value (eg, 2 16 in this case ), and ARP reference is made.
  • ArpRefPicAvailable which is a flag indicating availability of a picture, is set to 0 (false).
  • DiffPicOrderCnt (RefPicListX [rIdx], currPic) (C1-1)
  • DiffPicOrderCnt (PicA, PicB) is a function for obtaining a value obtained by subtracting the POC of the picture PicB from the POC of the picture PicA. (The same applies to the following configurations).
  • DiffPicOrderCnt (picA, currPic) may be POC (picA) -PicOrderCntVal.
  • conditional expression C (1-1) there is a reference picture having the same POC as the reference picture stored in RefPicListX [rIdx], and the POC of the rIdx-th reference picture and the POC (PicOrderCntVal of the target picture currPic) ) And the absolute value of the difference is less than minDiffPOC, it is “true”.
  • minDiffPOC is updated by abs (DiffPicOrderCnt (RefPicListX [rIdx], currPic), and rIdx becomes a tentative ARP reference picture candidate (rIdxSel).
  • arpRefPicAvailable is set to 1.
  • a reference picture having a POC closest to the POC of the reference picture is derived as an index indicating the ARP reference picture (rIdxSel).
  • the reference image determination unit 3095 derives the ARP reference picture arpRefPi as the reference picture RefPicListX [rIdxSel] of the index rIdxSel of the reference picture list RefPicListX [] for the rIdxSel obtained in this way.
  • the arpRefIdxLX is selected so that the arpRefPic having a POC different from the PictureOrderCntVal that is the POC of the target picture is set. Therefore, the POC of the arpRefPic is the POC of the target picture Since it is equal to PictureOrderCntVal, the effect that ARP does not operate is solved.
  • the parsing stage may be determined in the same manner as option Y5.
  • the reference image determination unit 3095 and the entropy decoding unit 301 (inter prediction parameter decoding unit 303) of the option Y5 similarly to the option Y2, it is useless when the ARP reference picture arpRefPic is equal to the POC of the target picture currPic. Since iv_res_pred_weight_idx, which is a bad flag, is not decoded, the code amount can be reduced.
  • the motion compensation stage may be determined in the same manner as option Y5.
  • the reference image determination unit 3095 and the residual prediction execution flag deriving unit 30921 (inter prediction image generation unit 309) of the option Y5 as in the case of the option Y3, that is, the ARP
  • the reference picture arpRefPic is equal to the POC of the target picture currPic, by not performing the ARP operation, there is an effect of avoiding an invalid operation.
  • the program for realizing the control function may be recorded on a computer-readable recording medium, and the program recorded on the recording medium may be read by a computer system and executed.
  • the “computer system” here is a computer system built in the image encoding device 11 and the image decoding device 31, and includes an OS and hardware such as peripheral devices.
  • the “computer-readable recording medium” refers to a storage device such as a flexible medium, a magneto-optical disk, a portable medium such as a ROM or a CD-ROM, and a hard disk incorporated in a computer system.
  • the “computer-readable recording medium” is a medium that dynamically holds a program for a short time, such as a communication line when transmitting a program via a network such as the Internet or a communication line such as a telephone line,
  • a volatile memory inside a computer system serving as a server or a client may be included and a program that holds a program for a certain period of time.
  • the program may be a program for realizing a part of the functions described above, and may be a program capable of realizing the functions described above in combination with a program already recorded in a computer system.
  • part or all of the image encoding device 11 and the image decoding device 31 in the above-described embodiment may be realized as an integrated circuit such as an LSI (Large Scale Integration).
  • LSI Large Scale Integration
  • Each functional block of the image encoding device 11 and the image decoding device 31 may be individually made into a processor, or a part or all of them may be integrated into a processor.
  • the method of circuit integration is not limited to LSI, and may be realized by a dedicated circuit or a general-purpose processor. Further, in the case where an integrated circuit technology that replaces LSI appears due to progress in semiconductor technology, an integrated circuit based on the technology may be used.
  • An image decoding device is an image decoding device that generates a predicted image of a target picture by applying residual prediction using a reference layer different from the target picture to the motion compensated image.
  • a reference picture determination unit for determining whether or not a reference picture of the reference layer having a picture order different from the picture order of the target picture is available, and at least according to the determination result,
  • a residual prediction application unit (residual synthesis) that applies the residual prediction based on the reference picture of the reference layer and the decoded picture of the reference layer having the same picture order as the target picture to the motion compensated image Part 30923).
  • the target picture is a picture to be decoded.
  • the reference layer is a layer different from the target layer to which the target picture belongs. When performing residual prediction in the reference layer, reference pictures belonging to the reference layer are referred to.
  • the residual prediction is a technique for estimating the residual in the reference layer as the residual in the target layer.
  • the residual prediction is performed based on the reference picture of the reference layer having a picture order different from the picture order of the target picture and the decoded picture of the reference layer having the same picture order as the picture order of the target picture.
  • the residual prediction can be applied to the motion compensated image depending on whether or not a reference picture of the reference layer is available.
  • the reference picture determination unit refers to the reference layer in a reference picture set indicating a picture to be referred to when decoding the decoded picture of the reference layer. The above determination may be performed depending on whether or not a picture exists.
  • the reference picture set is derived for each picture. Therefore, it is commonly used in decoding of slices included in the target picture.
  • the reference picture determination unit refers to the reference layer at a predetermined position in a reference picture list of a slice that is not the predetermined I-th slice of the target picture. The above determination may be performed depending on whether or not a picture exists.
  • the determination process can be shared. .
  • the reference picture determination unit performs the determination according to whether or not the reference picture of the reference layer is stored in a DPB (Decoded Picture Buffer). May be performed.
  • DPB Decoded Picture Buffer
  • the reference picture determination unit determines whether or not the reference mark in the DPB (Decoded Picture Buffer) of the reference picture of the reference layer is “reference use”. Depending on whether or not, the above determination may be made.
  • DPB Decoded Picture Buffer
  • Some decoders do not support DPB processing as defined by HRD (Hypothetical Reference Decoder), so the DPB state may be inaccurate. Therefore, by checking the reference mark in the DPB (Decoded Picture Buffer) of the reference picture of the reference layer, it is possible to more accurately determine whether or not the reference picture of the reference layer is usable.
  • HRD Hydrophiltical Reference Decoder
  • the reference mark is “reference use”, it may be determined that the reference picture of the reference layer exists.
  • the image decoding device is an image decoding device that generates a predicted image of a target picture by applying residual prediction using a reference layer different from the target picture to the motion compensated image.
  • Reference for obtaining a reference picture belonging to a target layer to which the target picture belongs and having a picture order different from the picture order of the target picture, from a predetermined position in a reference picture list of a slice that is not the predetermined I-th slice of the target picture A picture acquisition unit (reference image acquisition unit 30922), a reference picture of the reference layer having the same picture order as the picture order of the acquired reference picture, and a reference layer having the same picture order as the picture order of the target picture.
  • the residual prediction based on the decoded picture of Comprising residual prediction application unit that applies to the image (the residual synthesis section 30923), the.
  • a reference picture belonging to the target layer to which the target picture belongs and having a picture order different from the picture order of the target picture is not a predetermined I-th slice of the target picture. Obtained from a predetermined position in the reference picture list of the slice.
  • the slice that is not the predetermined I-slice of the target picture is preferably a slice that is not the first I-slice of the target picture.
  • the predetermined position of the reference picture list may be the 0th position (first position) of the reference picture list. If the slices included in the target picture are common, the slice number and the position of the reference picture list are arbitrary.
  • the determination of the ARP reference picture can be made common among the slices included in the target picture.
  • the image decoding device is an image decoding device that generates a predicted image of a target picture by applying residual prediction using a reference layer different from the target picture to the motion compensated image.
  • a flag decoding unit (inter prediction parameter decoding unit 303) that decodes a residual prediction execution flag that instructs execution of residual prediction, and a reference picture of the reference layer that has a picture order different from the picture order of the target picture.
  • a receiving unit that receives a usable bitstream, and a residual prediction execution unit (residual prediction unit 3092) that executes the residual prediction according to the residual prediction execution flag.
  • An image encoding device is an image encoding device that generates a predicted image of a target picture by applying residual prediction using a reference layer different from the target picture to the motion compensated image.
  • a flag encoding unit (prediction parameter encoding unit 111) that encodes a residual prediction execution flag that instructs execution of the residual prediction, and the reference layer having a picture order different from the picture order of the target picture
  • a bit stream generation unit (prediction parameter encoding unit 111) that generates a bit stream in which the reference picture can be used, and a bit stream transmission unit (entropy code) that transmits the generated bit stream to the image decoding apparatus And a conversion unit 104).
  • the image decoding device according to aspect 7 or the image encoding device according to aspect 8 includes feature points corresponding to the image decoding device according to aspect 1. Therefore, the image decoding device according to aspect 7 or the image encoding device according to aspect 8 can achieve the same effects as those of the image decoding device according to aspect 1.
  • the image decoding device is an image decoding device that generates a prediction image of a target picture by applying residual prediction using a reference layer of a layer different from the target picture to the motion compensated image.
  • a reference picture selection unit that selects a reference picture that belongs to the target layer to which the target picture belongs and has a picture order different from the picture order of the target picture, and the selected reference
  • the residual prediction based on the reference picture of the reference layer having the same picture order as the picture order of the picture and the decoded picture of the reference layer having the same picture order as the picture order of the target picture, and the motion compensation
  • a residual prediction application unit to be applied to the image.
  • the arpRefpic POC is different from the POC of the current picture. That is, the picture order of the reference picture of the reference layer having the same picture order as the picture order of the selected reference picture, and the picture order of the decoded picture of the reference layer having the same picture order as the picture order of the target picture Can be guaranteed to be different.
  • the present invention can be suitably applied to an image decoding apparatus that decodes encoded data obtained by encoding image data and an image encoding apparatus that generates encoded data obtained by encoding image data. Further, the present invention can be suitably applied to the data structure of encoded data generated by an image encoding device and referenced by the image decoding device.
  • Displacement compensation Unit 3092 Residual prediction unit (residual prediction execution unit) 30921 ... Residual prediction execution flag deriving unit 30922 ... Reference image acquisition unit (reference picture acquisition unit) 30923 ... Residual synthesis unit (residual prediction application unit) 3093: Illuminance compensation unit 3094 ... Weight prediction unit 3095 ... Reference image determination unit (reference picture determination unit) 310 ... Intra predicted image generation unit 311 ... Inverse quantization / inverse DCT unit 312 ... Adder 313 ... Residual storage unit 41 ... Image display device

Landscapes

  • Engineering & Computer Science (AREA)
  • Multimedia (AREA)
  • Signal Processing (AREA)
  • Compression Or Coding Systems Of Tv Signals (AREA)

Abstract

The present invention prevents a situation in which there is no reference picture of a reference layer during advanced residual prediction (ARP) processing. This image decoding device, which performs residual prediction, comprises: a reference picture determination unit (3095) which determines whether a residual prediction reference picture can be used; and a residual prediction unit (3092) which applies residual prediction to a motion compensation image in a reference layer if the reference picture of the reference layer can be used. The reference picture determination unit performs the determination according to whether the reference picture of the residual prediction reference layer is stored in a decoded picture buffer (DPB).

Description

画像復号装置Image decoding device
 本発明は、対象レイヤおよび参照レイヤの間で残差予測を行う画像復号装置に関する。 The present invention relates to an image decoding apparatus that performs residual prediction between a target layer and a reference layer.
 近年、動画像を符号化して伝送・蓄積する動画像符号化方式の一つとしてHEVC(High-Efficiency Video Coding)が知られている(非特許文献1)。また、複数の相互に関連性のある動画像をレイヤ(階層)に分けて符号化することで、複数の動画像から符号化データを生成する技術も知られており、このような技術は階層符号化技術(あるいはスケーラブル符号化技術)と呼ばれる。 In recent years, HEVC (High-Efficiency Video Coding) is known as one of video encoding methods for encoding, transmitting, and storing video (Non-patent Document 1). There is also known a technique for generating encoded data from a plurality of moving images by encoding a plurality of mutually related moving images into layers (hierarchies). This is called an encoding technique (or scalable encoding technique).
 また、階層符号化技術の一つとしてHEVCを基礎とするMV-HEVC(Multi View HEVC)が提案されている(非特許文献2)。 Also, MV-HEVC (Multi-View HEVC) based on HEVC has been proposed as one of hierarchical coding technologies (Non-patent Document 2).
 MV-HEVCではビュースケーラビリティをサポートする。ビュースケーラビリティでは、複数の異なる視点(ビュー)に対応する動画像をレイヤに分けて符号化して階層符号化データを生成する。例えば、基本となる視点(ベースビュー)に対応する動画像を下位レイヤとして符号化する。次に、異なる視点に対応する動画像を、レイヤ間予測を適用した上で、上位レイヤとして符号化する。 MV-HEVC supports view scalability. In view scalability, a moving image corresponding to a plurality of different viewpoints (views) is divided into layers and encoded to generate hierarchical encoded data. For example, a moving image corresponding to a basic viewpoint (base view) is encoded as a lower layer. Next, a moving image corresponding to a different viewpoint is encoded as an upper layer after applying inter-layer prediction.
 さらには、HEVCおよびMV-HEVCを基礎とした3D-HEVC(3D High Efficiency Video Coding)という3D動画技術が検討されている。 Furthermore, 3D video technology called 3D-HEVC (3D High Efficiency Efficiency Video Coding) based on HEVC and MV-HEVC is being studied.
 また、3D-HEVCでは、ARP(Advanced Residual Prediction)という予測技術も提案されている。ARPでは、異なる視点において復号済みのピクチャを対象ピクチャにおける残差の推定に用いる。より具体的には、ARPでは、対象時刻以外の参照レイヤの復号画像(refIvRefPic)と、対象時刻の参照レイヤの復号画像(currIvRefPic)の差から、対象ピクチャの残差を推定する。 In 3D-HEVC, a prediction technique called ARP (Advanced Residual Prediction) has also been proposed. In ARP, pictures that have been decoded at different viewpoints are used to estimate residuals in the target picture. More specifically, in the ARP, the residual of the target picture is estimated from the difference between the decoded image (refIvRefPic) of the reference layer other than the target time and the decoded image (currIvRefPic) of the reference layer at the target time.
 しかしながら、上述のような従来技術では、refIvRefPicがDPB(Decoded Picture Buffer)から既に破棄されて、存在しない場合があるという課題がある。 However, the conventional technology as described above has a problem that refIvRefPic is already discarded from DPB (Decoded Picture Buffer) and may not exist.
 図24を用いて具体的に説明すると次のとおりである。図24において、2つのビューが符号化されている。ピクチャP101、P102、およびP103は、対象ビュー(レイヤ)におけるピクチャであり、ピクチャP201、P202、およびP203は、参照ビュー(RefViewIdx)のピクチャである。なお、各ピクチャに付した番号は、復号順を示している。 It will be specifically described with reference to FIG. In FIG. 24, two views are encoded. Pictures P101, P102, and P103 are pictures in the target view (layer), and pictures P201, P202, and P203 are pictures of the reference view (RefViewIdx). Note that the number assigned to each picture indicates the decoding order.
 ここで、ピクチャP103を復号する場合において、ピクチャP103が、ピクチャP101から予測されるとする。 Here, when decoding the picture P103, it is assumed that the picture P103 is predicted from the picture P101.
 この場合、ARPでは、現行POCの対応するピクチャまたはcurrIvRefPicにて指定されるピクチャP203と、RefPicListXにおいて最初に位置するピクチャまたはrefIvRefPicにて指定されるピクチャP202とから残差が導出される。 In this case, in ARP, a residual is derived from the picture corresponding to the current POC or the picture P203 specified by currIvRefPic and the picture P202 specified by the first picture or refIvRefPic in RefPicListX.
 しかしながら、図24に示す例では、refIvRefPicにて指定されるピクチャP202がcurrIvRefPicにて指定されるピクチャP203から参照されていないため、ピクチャP103の復号の前に、DPBから削除された場合には、refIvRefPicにて指定されるピクチャP202が利用できない。また、対象ピクチャの復号は、参照ピクチャリストに格納されたピクチャを参照しながら行われるが、参照ピクチャリストはスライス単位で変わるため、ピクチャP102(arpRefpic)の参照ピクチャリストにおける位置が、スライス単位で変わる可能性があるという問題もある。さらには、参照しようとするピクチャP102のPOCと、復号する対象ピクチャのピクチャP103のPOCとが一致していしまう可能性もある。 However, in the example shown in FIG. 24, since the picture P202 specified by refIvRefPic is not referenced from the picture P203 specified by currIvRefPic, if it is deleted from the DPB before decoding of the picture P103, The picture P202 specified by refIvRefPic cannot be used. The target picture is decoded while referring to the pictures stored in the reference picture list. Since the reference picture list changes in units of slices, the position of the picture P102 (arpRefpic) in the reference picture list is in units of slices. There is also the problem that it may change. Furthermore, there is a possibility that the POC of the picture P102 to be referred to matches the POC of the picture P103 of the target picture to be decoded.
 本発明は上記問題に鑑みてなされたものであり、その目的は、残差予測時に、参照レイヤの参照ピクチャが存在しない、または、対象レイヤにおける参照ピクチャの所在がスライス間で変わる、または、復号する対象ピクチャのピクチャ順序と参照しようとする対象レイヤにおけるピクチャのピクチャ順序とが一致する、という事態を避けることができる画像復号装置等を実現することにある。 The present invention has been made in view of the above problems, and an object of the present invention is to determine whether there is no reference picture in the reference layer, or the location of the reference picture in the target layer changes between slices, or decoding during residual prediction. An object of the present invention is to realize an image decoding apparatus or the like that can avoid a situation in which the picture order of a target picture to be matched matches the picture order of a picture in a target layer to be referred to.
 本発明の一態様に係る画像復号装置は、残差予測参照ピクチャが利用可能であるか否かの判定を行う参照ピクチャ判定部と、上記残差予測参照ピクチャを用いた残差予測を行う残差予測適用部とを備え上記参照ピクチャ判定部は、上記残差予測参照レイヤの参照ピクチャがDPB(Decoded Picture Buffer)において格納されているか否かに応じて、上記判定を行うことを特徴とする。 An image decoding apparatus according to an aspect of the present invention includes a reference picture determination unit that determines whether or not a residual prediction reference picture is usable, and a residual that performs residual prediction using the residual prediction reference picture. A difference prediction application unit, wherein the reference picture determination unit performs the determination according to whether or not a reference picture of the residual prediction reference layer is stored in a DPB (Decoded Picture) Buffer). .
 本発明の別の態様に係る画像復号装置は、残差予測用参照ピクチャを導出する参照ピクチャ導出部と、上記残差予測参照ピクチャを用いた残差予測を行う残差予測適用部とを備え、
 上記参照ピクチャ導出部は、参照ピクチャリストに含まれる参照ピクチャのうち、対象ピクチャのピクチャ順序と異なるピクチャ順序を有する参照ピクチャを上記残差予測用参照ピクチャとして導出することを特徴とする。
An image decoding apparatus according to another aspect of the present invention includes a reference picture deriving unit that derives a reference picture for residual prediction, and a residual prediction applying unit that performs residual prediction using the residual prediction reference picture. ,
The reference picture deriving unit derives, as the residual prediction reference picture, a reference picture having a picture order different from the picture order of the target picture among the reference pictures included in the reference picture list.
 本発明の別の態様に係る画像復号装置は、残差予測用参照ピクチャを導出する参照ピクチャ選択部と、上記残差予測参照ピクチャが利用可能な場合に、上記残差予測参照ピクチャを用いた残差予測を行う残差予測適用部とを備え、参照ピクチャリストを先頭から順に走査し、参照ピクチャRefPicListX[i]のPOCと、対象ピクチャcurrPicのPOC(PicOrderCntVal)の差分の絶対値が、上記POC差分よりも小さい場合に、その参照ピクチャを残差予測用参照ピクチャに設定することを特徴とする。 An image decoding apparatus according to another aspect of the present invention uses a reference picture selection unit that derives a reference picture for residual prediction, and the residual prediction reference picture when the residual prediction reference picture is available A residual prediction application unit that performs residual prediction, scans the reference picture list in order from the top, and the absolute value of the difference between the POC of the reference picture RefPicListX [i] and the POC (PicOrderCntVal) of the target picture currPic is When the difference is smaller than the POC difference, the reference picture is set as a residual prediction reference picture.
 本発明の別の態様に係る画像復号装置は、参照ピクチャリストに含まれる参照ピクチャのうち、上記対象ピクチャのピクチャ順序と異なるピクチャ順序を有する参照ピクチャを残差予測参照ピクチャとして導出する参照ピクチャ導出部と、上記残差予測参照ピクチャが利用可能な場合に、残差予測フラグiv_res_pred_weight_idxを復号するインター予測パラメータ復号制御部を備える。 An image decoding apparatus according to another aspect of the present invention provides a reference picture derivation that derives, as a residual prediction reference picture, a reference picture having a picture order different from the picture order of the target picture among reference pictures included in a reference picture list And an inter prediction parameter decoding control unit for decoding the residual prediction flag iv_res_pred_weight_idx when the residual prediction reference picture is available.
 本発明の一態様に係る画像復号装置よれば、対象ピクチャについての残差予測時に、上記参照レイヤの復号ピクチャが利用できないという事態を回避することができるという効果を奏する。 The image decoding apparatus according to an aspect of the present invention has an effect that it is possible to avoid a situation in which a decoded picture of the reference layer cannot be used during residual prediction for a target picture.
 また、本発明の他の態様に係る画像復号装置によれば、arpRefpicのPOCが現ピクチャのPOCと異なることを保障することができるという効果を奏する。 Also, according to the image decoding apparatus according to another aspect of the present invention, it is possible to ensure that the arpRefpic POC is different from the POC of the current picture.
本発明の一実施形態に係る画像復号装置に含まれる残差予測部の構成を示す機能ブロック図である。It is a functional block diagram which shows the structure of the residual prediction part contained in the image decoding apparatus which concerns on one Embodiment of this invention. 本実施形態に係る画像伝送システムの構成を示す概略図である。It is the schematic which shows the structure of the image transmission system which concerns on this embodiment. 本実施形態に係る符号化ストリームのデータの階層構造を示す図である。It is a figure which shows the hierarchical structure of the data of the encoding stream which concerns on this embodiment. (a)は、参照ピクチャリストの一例を示す概念図であり、(b)は、ベクトル候補の一例を示す概念図である。(A) is a conceptual diagram which shows an example of a reference picture list, (b) is a conceptual diagram which shows an example of a vector candidate. 参照ピクチャの例を示す概念図である。It is a conceptual diagram which shows the example of a reference picture. 本実施形態に係る画像復号装置の構成を示す概略図である。It is the schematic which shows the structure of the image decoding apparatus which concerns on this embodiment. 本実施形態に係るインター予測パラメータ復号部の構成を示す概略図である。It is the schematic which shows the structure of the inter prediction parameter decoding part which concerns on this embodiment. マージインデックスmerge_idxと、残差予測の重み数の関係を示す表である。It is a table | surface which shows the relationship between merge index merge_idx and the weight number of residual prediction. 上記画像復号装置に含まれる復号ピクチャ管理部の構成を例示した機能ブロック図である。It is the functional block diagram which illustrated the structure of the decoded picture management part contained in the said image decoding apparatus. 本実施形態に係るインター予測画像生成部の構成を示す概略図である。It is the schematic which shows the structure of the inter estimated image generation part which concerns on this embodiment. (a)は、本実施形態に係る残差予測の概念図(その1)であり、(b)は、本実施形態に係る残差予測の概念図(その2)である。(A) is a conceptual diagram (part 1) of residual prediction according to the present embodiment, and (b) is a conceptual diagram (part 2) of residual prediction according to the present embodiment. 本実施形態に係る画像符号化装置の構成を示すブロック図である。It is a block diagram which shows the structure of the image coding apparatus which concerns on this embodiment. 本実施形態に係るインター予測パラメータ符号化部の構成を示す概略図である。It is the schematic which shows the structure of the inter prediction parameter encoding part which concerns on this embodiment. CUレベルで、arpRefPicAvailableフラグを判定する例を示すシンタックスである。It is a syntax which shows the example which determines an arpRefPicAvailable flag in CU level. CUレベルで、arpRefPicAvailableフラグを判定する別の例を示すシンタックスである。It is a syntax which shows another example which determines an arpRefPicAvailable flag in CU level. 上記画像復号装置のエントロピー復号部によりSPS復号時に利用されるシンタックス表の一部であって、参照ピクチャセットおよび参照ピクチャリストに係る部分を示す図である。It is a figure which is a part of syntax table utilized at the time of SPS decoding by the entropy decoding part of the said image decoding apparatus, Comprising: It is a figure which shows the part which concerns on a reference picture set and a reference picture list. (a)は、上記画像復号装置のエントロピー復号部により短期参照ピクチャセット情報の復号時に利用されるシンタックス表を示す図である。(b)は、上記画像復号装置のエントロピー復号部によりスライスヘッダ復号時に利用されるシンタックス表の一部であって、参照ピクチャセットに係る部分を示す図である。(A) is a figure which shows the syntax table utilized at the time of decoding of short-term reference picture set information by the entropy decoding part of the said image decoding apparatus. (B) is a part of a syntax table used at the time of slice header decoding by the entropy decoding unit of the image decoding apparatus, and shows a part related to a reference picture set. (a)は、VPSに含まれるVPS拡張(vps_extension)の復号時に参照されるシンタックス表の一部であって、IL-RPS情報に相当する部分を示す図である。(b)は、スライス復号時に参照されるシンタックス表の一部であって、IL-RPS情報に相当する部分を示す図である。(A) is a part of syntax table referred to at the time of decoding of the VPS extension (vps_extension) included in the VPS, and shows a part corresponding to IL-RPS information. (B) is a diagram showing a part corresponding to IL-RPS information, which is a part of a syntax table referred to at the time of slice decoding. (a)は、レイヤ間予測の種類にレイヤ間画像予測とレイヤ間動き予測がある場合の依存タイプと各レイヤ間予測の使用可否の関係を示す図である。(b)は、上記画像復号装置において生成されるレイヤ間RPSに含まれるサブRPS(レイヤ間画素RPSとレイヤ間動き限定RPS)と依存タイプの関係を例示する図である。(A) is a figure which shows the relationship between the dependence type in the case where there exist inter-layer image prediction and inter-layer motion prediction in the type of inter-layer prediction, and the availability of each inter-layer prediction. (B) is a diagram illustrating the relationship between sub RPS (inter-layer pixel RPS and inter-layer motion limited RPS) included in the inter-layer RPS generated in the image decoding apparatus and the dependency type. レイヤ間RPSに含まれるサブRPS(レイヤ間画素RPSとレイヤ間動き限定RPS)の導出処理を表わすフロー図である。It is a flowchart showing the derivation | leading-out process of sub RPS (inter-layer pixel RPS and inter-layer motion limited RPS) included in inter-layer RPS. (a)は、上記画像復号装置のエントロピー復号部によりスライスヘッダ復号時に利用されるシンタックス表の一部であって、参照ピクチャリストに係る部分を示す図である。(b)は、上記画像復号装置のエントロピー復号部により参照ピクチャリスト修正情報の復号時に利用されるシンタックス表を示す図である。(A) is a part of the syntax table used at the time of slice header decoding by the entropy decoding unit of the image decoding apparatus, and shows a part related to the reference picture list. (B) is a figure which shows the syntax table utilized at the time of decoding of reference picture list correction information by the entropy decoding part of the said image decoding apparatus. 上記画像復号装置のRPL導出部におけるL0参照リストとL1参照リスト導出の中間過程で生成される暫定L0参照リストと暫定L1参照リストを示す概略図である。It is the schematic which shows the temporary L0 reference list and temporary L1 reference list which are produced | generated in the intermediate process of L0 reference list and L1 reference list derivation | leading-out in the RPL derivation part of the said image decoding apparatus. 上記画像復号装置に含まれるマージ予測パラメータ導出部の構成を示す概略図である。It is the schematic which shows the structure of the merge prediction parameter derivation | leading-out part contained in the said image decoding apparatus. ARPにおける対象レイヤのピクチャおよび参照レイヤの間の関連性を説明する図である。It is a figure explaining the relationship between the picture of the object layer in ARP, and a reference layer.
 〔第1の実施形態〕
 以下、図面を参照しながら本発明の実施形態について説明する。
[First Embodiment]
Hereinafter, embodiments of the present invention will be described with reference to the drawings.
 まず、図2を用いて、本実施形態に係る画像伝送システム1の概略的構成について説明する。図2は、本実施形態に係る画像伝送システム1の構成を示す概略図である。 First, a schematic configuration of the image transmission system 1 according to the present embodiment will be described with reference to FIG. FIG. 2 is a schematic diagram showing the configuration of the image transmission system 1 according to the present embodiment.
 画像伝送システム1は、複数のレイヤ画像を符号化した符号を伝送し、伝送された符号を復号した画像を表示するシステムである。画像伝送システム1は、画像符号化装置11、ネットワーク21、画像復号装置31及び画像表示装置41を含んで構成される。 The image transmission system 1 is a system that transmits a code obtained by encoding a plurality of layer images and displays an image obtained by decoding the transmitted code. The image transmission system 1 includes an image encoding device 11, a network 21, an image decoding device 31, and an image display device 41.
 画像符号化装置11には、複数のレイヤ画像(テクスチャ画像ともいう)を示す信号Tが入力される。レイヤ画像とは、ある解像度及びある視点で視認もしくは撮影される画像である。複数のレイヤ画像を用いて3次元画像を符号化するビュースケーラブル符号化を行う場合、複数のレイヤ画像のそれぞれは、視点画像と呼ばれる。ここで、視点は撮影装置の位置又は観測点に相当する。例えば、複数の視点画像は、被写体に向かって左右の撮影装置のそれぞれが撮影した画像である。画像符号化装置11は、この信号のそれぞれを符号化して符号化ストリームTe(符号化データ)を生成する。符号化ストリームTeの詳細については、後述する。視点画像とは、ある視点において観測される2次元画像(平面画像)である。視点画像は、例えば2次元平面内に配置された画素毎の輝度値、又は色信号値で示される。以下では、1枚の視点画像又は、その視点画像を示す信号をピクチャ(picture)と呼ぶ。また、複数のレイヤ画像を用いて空間スケーラブル符号化を行う場合、その複数のレイヤ画像は、解像度の低いベースレイヤ画像と、解像度の高い拡張レイヤ画像からなる。複数のレイヤ画像を用いてSNRスケーラブル符号化を行う場合、その複数のレイヤ画像は、画質の低いベースレイヤ画像と、画質の高い拡張レイヤ画像からなる。なお、ビュースケーラブル符号化、空間スケーラブル符号化、SNRスケーラブル符号化を任意に組み合わせて行っても良い。本実施形態では、複数のレイヤ画像として、少なくともベースレイヤ画像と、ベースレイヤ画像以外の画像(拡張レイヤ画像)を含む画像の符号化および復号を扱う。複数のレイヤのうち、画像もしくは符号化パラメータにおいて参照関係(依存関係)にある2つのレイヤについて、参照される側の画像を、第1レイヤ画像、参照する側の画像を第2レイヤ画像と呼ぶ。例えば、ベースレイヤを参照して符号化される(ベースレイヤ以外の)エンハンスレイヤ画像がある場合、ベースレイヤ画像を第1レイヤ画像、エンハンスレイヤ画像を第2レイヤ画像として扱う。なお、エンハンスレイヤ画像の例としては、ベースビュー以外の視点の画像、デプス(depth)の画像などがある。 The signal T indicating a plurality of layer images (also referred to as texture images) is input to the image encoding device 11. A layer image is an image that is viewed or photographed at a certain resolution and a certain viewpoint. When performing view scalable coding in which a three-dimensional image is coded using a plurality of layer images, each of the plurality of layer images is referred to as a viewpoint image. Here, the viewpoint corresponds to the position or observation point of the photographing apparatus. For example, the plurality of viewpoint images are images taken by the left and right photographing devices toward the subject. The image encoding device 11 encodes each of the signals to generate an encoded stream Te (encoded data). Details of the encoded stream Te will be described later. A viewpoint image is a two-dimensional image (planar image) observed at a certain viewpoint. The viewpoint image is indicated by, for example, a luminance value or a color signal value for each pixel arranged in a two-dimensional plane. Hereinafter, one viewpoint image or a signal indicating the viewpoint image is referred to as a picture. In addition, when performing spatial scalable coding using a plurality of layer images, the plurality of layer images include a base layer image having a low resolution and an enhancement layer image having a high resolution. When SNR scalable encoding is performed using a plurality of layer images, the plurality of layer images are composed of a base layer image with low image quality and an extended layer image with high image quality. Note that view scalable coding, spatial scalable coding, and SNR scalable coding may be arbitrarily combined. In the present embodiment, encoding and decoding of an image including at least a base layer image and an image other than the base layer image (enhancement layer image) is handled as the plurality of layer images. Of the multiple layers, for two layers that have a reference relationship (dependency relationship) in the image or encoding parameter, the image on the reference side is referred to as a first layer image, and the image on the reference side is referred to as a second layer image. . For example, when there is an enhancement layer image (other than the base layer) that is encoded with reference to the base layer, the base layer image is treated as a first layer image and the enhancement layer image is treated as a second layer image. Note that examples of the enhancement layer image include an image of a viewpoint other than the base view and a depth image.
 ネットワーク21は、画像符号化装置11が生成した符号化ストリームTeを画像復号装置31に伝送する。ネットワーク21は、例えば、インターネット(internet)、広域ネットワーク(WAN:Wide Area Network)、小規模ネットワーク(LAN:Local Area Network)又はこれらの組み合わせによって構成することができる。ネットワーク21は、必ずしも双方向の通信網に限らず、地上波ディジタル放送、衛星放送等の放送波を伝送する一方向又は双方向の通信網であっても良い。また、ネットワーク21は、DVD(Digital Versatile Disc)、BD(Blue-ray Disc)等の符号化ストリームTeを記録した記憶媒体で代替されても良い。 The network 21 transmits the encoded stream Te generated by the image encoding device 11 to the image decoding device 31. The network 21 can be configured by, for example, the Internet, a wide area network (WAN: Wide Area Network), a small network (LAN: Local Area Network), or a combination thereof. The network 21 is not necessarily limited to a bidirectional communication network, and may be a unidirectional or bidirectional communication network that transmits broadcast waves such as terrestrial digital broadcasting and satellite broadcasting. The network 21 may be replaced by a storage medium that records an encoded stream Te such as a DVD (Digital Versatile Disc) or a BD (Blue-ray Disc).
 画像復号装置31は、ネットワーク21が伝送した符号化ストリームTeのそれぞれを復号し、それぞれ復号した複数の復号レイヤ画像Td(復号視点画像Td)を生成する。 The image decoding device 31 decodes each of the encoded streams Te transmitted by the network 21, and generates a plurality of decoded layer images Td (decoded viewpoint images Td).
 画像表示装置41は、画像復号装置31が生成した複数の復号レイヤ画像Tdの全部又は一部を表示する。例えば、ビュースケーラブル符号化においては、全部の場合、3次元画像(立体画像)や自由視点画像が表示され、一部の場合、2次元画像が表示される。画像表示装置41は、例えば、液晶ディスプレイ、有機EL(Electro-luminescence)ディスプレイ等の表示デバイスを備える。また、空間スケーラブル符号化、SNRスケーラブル符号化では、画像復号装置31、画像表示装置41が高い処理能力を有する場合には、画質の高い拡張レイヤ画像を表示し、より低い処理能力しか有しない場合には、拡張レイヤほど高い処理能力、表示能力を必要としないベースレイヤ画像を表示する。 The image display device 41 displays all or part of the plurality of decoded layer images Td generated by the image decoding device 31. For example, in view scalable coding, a 3D image (stereoscopic image) and a free viewpoint image are displayed in all cases, and a 2D image is displayed in some cases. The image display device 41 includes, for example, a display device such as a liquid crystal display or an organic EL (Electro-luminescence) display. In addition, in the spatial scalable coding and SNR scalable coding, when the image decoding device 31 and the image display device 41 have a high processing capability, a high-quality enhancement layer image is displayed and only a lower processing capability is provided. Displays a base layer image that does not require higher processing capability and display capability as an extension layer.
 本実施形態に係る画像符号化装置11および画像復号装置31の詳細な説明に先立って、画像符号化装置11によって生成され、画像復号装置31によって復号される符号化ストリームTeのデータ構造について説明する。 Prior to detailed description of the image encoding device 11 and the image decoding device 31 according to the present embodiment, the data structure of the encoded stream Te generated by the image encoding device 11 and decoded by the image decoding device 31 will be described. .
 図3は、符号化ストリームTeにおけるデータの階層構造を示す図である。符号化ストリームTeは、例示的に、シーケンス、およびシーケンスを構成する複数のピクチャを含む。図3の(a)~(f)は、それぞれ、シーケンスSEQを既定するシーケンスレイヤ、ピクチャPICTを規定するピクチャレイヤ、スライスSを規定するスライスレイヤ、スライスデータを規定するスライスデータレイヤ、スライスデータに含まれる符号化ツリーユニットを規定する符号化ツリーレイヤ、符号化ツリーに含まれる符号化単位(Coding Unit;CU)を規定する符号化ユニットレイヤを示す図である。 FIG. 3 is a diagram showing a hierarchical structure of data in the encoded stream Te. The encoded stream Te illustratively includes a sequence and a plurality of pictures constituting the sequence. (A) to (f) of FIG. 3 respectively show a sequence layer that defines a sequence SEQ, a picture layer that defines a picture PICT, a slice layer that defines a slice S, a slice data layer that defines slice data, and a slice data. It is a figure which shows the encoding unit layer which prescribes | regulates the encoding tree layer which prescribes | regulates the encoding tree unit contained, and the coding unit (Coding Unit; CU) contained in a coding tree.
  (シーケンスレイヤ) シーケンスレイヤでは、処理対象のシーケンスSEQ(以下、対象シーケンスとも称する)を復号するために画像復号装置31が参照するデータの集合が規定されている。シーケンスSEQは、図3の(a)に示すように、ビデオパラメータセットVPS(Video Parameter Set)シーケンスパラメータセットSPS(Sequence Parameter Set)、ピクチャパラメータセットPPS(Picture Parameter Set)、ピクチャPICT、及び、付加拡張情報SEI(Supplemental Enhancement Information)を含んでいる。ここで#の後に示される値はレイヤIDを示す。図3では、#0と#1すなわちレイヤ0とレイヤ1の符号化データが存在する例を示すが、レイヤの種類およびレイヤの数はこれによらない。 (Sequence layer) In the sequence layer, a set of data referred to by the image decoding device 31 for decoding a sequence SEQ to be processed (hereinafter also referred to as a target sequence) is defined. As shown in FIG. 3A, the sequence SEQ includes a video parameter set VPS (Video Parameter Set) sequence parameter set SPS (Sequence Parameter Parameter Set), a picture parameter set PPS (Picture Parameter Parameter Set), a picture PICT, and an addition. Extension information SEI (SupplementallementEnhancement Information) is included. Here, the value indicated after # indicates the layer ID. FIG. 3 shows an example in which encoded data of # 0 and # 1, that is, layer 0 and layer 1, exists, but the type of layer and the number of layers are not dependent on this.
 ビデオパラメータセットVPSは、複数のレイヤから構成されている動画像において、複数の動画像に共通する符号化パラメータの集合および動画像に含まれる複数のレイヤおよび個々のレイヤに関連する符号化パラメータの集合が規定されている。 The video parameter set VPS is a set of encoding parameters common to a plurality of moving images, a plurality of layers included in the moving image, and encoding parameters related to individual layers in a moving image composed of a plurality of layers. A set is defined.
 シーケンスパラメータセットSPSでは、対象シーケンスを復号するために画像復号装置31が参照する符号化パラメータの集合が規定されている。例えば、ピクチャの幅や高さが規定される。 The sequence parameter set SPS defines a set of encoding parameters that the image decoding device 31 refers to in order to decode the target sequence. For example, the width and height of the picture are defined.
 ピクチャパラメータセットPPSでは、対象シーケンス内の各ピクチャを復号するために画像復号装置31が参照する符号化パラメータの集合が規定されている。例えば、ピクチャの復号に用いられる量子化幅の基準値(pic_init_qp_minus26)や重み付き予測の適用を示すフラグ(weighted_pred_flag)が含まれる。なお、PPSは複数存在してもよい。その場合、対象シーケンス内の各ピクチャから複数のPPSの何れかを選択する。 In the picture parameter set PPS, a set of encoding parameters referred to by the image decoding device 31 in order to decode each picture in the target sequence is defined. For example, a quantization width reference value (pic_init_qp_minus26) used for picture decoding and a flag (weighted_pred_flag) indicating application of weighted prediction are included. A plurality of PPS may exist. In that case, one of a plurality of PPSs is selected from each picture in the target sequence.
  (ピクチャレイヤ) ピクチャレイヤでは、処理対象のピクチャPICT(以下、対象ピクチャとも称する)を復号するために画像復号装置31が参照するデータの集合が規定されている。ピクチャPICTは、図3の(b)に示すように、スライスS0~SNS-1を含んでいる(NSはピクチャPICTに含まれるスライスの総数)。 (Picture layer) In the picture layer, a set of data referred to by the image decoding device 31 for decoding a picture PICT to be processed (hereinafter also referred to as a target picture) is defined. As shown in FIG. 3B, the picture PICT includes slices S0 to SNS-1 (NS is the total number of slices included in the picture PICT).
 なお、以下、スライスS0~SNS-1のそれぞれを区別する必要が無い場合、符号の添え字を省略して記述することがある。また、以下に説明する符号化ストリームTeに含まれるデータであって、添え字を付している他のデータについても同様である。 It should be noted that, hereinafter, when it is not necessary to distinguish each of the slices S0 to SNS-1, the subscripts may be omitted. The same applies to data included in an encoded stream Te described below and to which other subscripts are attached.
  (スライスレイヤ) スライスレイヤでは、処理対象のスライスS(対象スライスとも称する)を復号するために画像復号装置31が参照するデータの集合が規定されている。スライスSは、図3の(c)に示すように、スライスヘッダSH、および、スライスデータSDATAを含んでいる。 (Slice layer) In the slice layer, a set of data referred to by the image decoding device 31 for decoding a slice S (also referred to as a target slice) to be processed is defined. As shown in FIG. 3C, the slice S includes a slice header SH and slice data SDATA.
 スライスヘッダSHには、対象スライスの復号方法を決定するために画像復号装置31が参照する符号化パラメータ群が含まれる。スライスタイプを指定するスライスタイプ指定情報(slice_type)は、スライスヘッダSHに含まれる符号化パラメータの一例である。 The slice header SH includes a coding parameter group that the image decoding device 31 refers to in order to determine a decoding method of the target slice. The slice type designation information (slice_type) that designates the slice type is an example of an encoding parameter included in the slice header SH.
 スライスタイプ指定情報により指定可能なスライスタイプとしては、(1)符号化の際にイントラ予測のみを用いるIスライス、(2)符号化の際に単方向予測、または、イントラ予測を用いるPスライス、(3)符号化の際に単方向予測、双方向予測、または、イントラ予測を用いるBスライスなどが挙げられる。 As slice types that can be specified by the slice type specification information, (1) I slice using only intra prediction at the time of encoding, (2) P slice using unidirectional prediction or intra prediction at the time of encoding, (3) B-slice using unidirectional prediction, bidirectional prediction, or intra prediction at the time of encoding may be used.
 なお、スライスヘッダSHには、上記シーケンスレイヤに含まれる、ピクチャパラメータセットPPSへの参照(pic_parameter_set_id)を含んでいても良い。 In addition, the slice header SH may include a reference (pic_parameter_set_id) to the picture parameter set PPS included in the sequence layer.
  (スライスデータレイヤ) スライスデータレイヤでは、処理対象のスライスデータSDATAを復号するために画像復号装置31が参照するデータの集合が規定されている。スライスデータSDATAは、図3の(d)に示すように、符号化ツリーブロック(CTB:Coded Tree Block)(符号化ツリーユニットCTU)を含んでいる。CTBは、スライスを構成する固定サイズ(例えば64×64)のブロックであり、最大符号化単位(LCU:Largest Cording Unit)と呼ぶこともある。 (Slice data layer) In the slice data layer, a set of data referred to by the image decoding device 31 for decoding the slice data SDATA to be processed is defined. The slice data SDATA includes a coding tree block (CTB: Coded ツ リ ー Tree Block) (coding tree unit CTU) as shown in FIG. CTB is a block of a fixed size (for example, 64 × 64) constituting a slice, and may be called a maximum coding unit (LCU: LargegestLCording Unit).
  (符号化ツリーレイヤ) 符号化ツリーレイヤは、図2の(e)に示すように、処理対象の符号化ツリーブロックを復号するために画像復号装置31が参照するデータの集合が規定されている。符号化ツリーユニットは、再帰的な4分木分割により分割される。再帰的な4分木分割により得られる木構造のノードのことを符号化ツリー(coding tree)と称する。4分木の中間ノードは、符号化4分木ツリー(CQT:Coded Quad Tree)であり、CTUは、最上位のCQTを含むものとして規定される。CQTは、分割フラグ(split_flag)を含み、split_flagが1の場合には、4つのCQTに分割される(4つのCQTを含む)。split_flagが0の場合には、CQTは末端ノードである符号化ユニット(CU:Coded Unit)を含む。符号化ユニットCUは、符号化処理の基本的な単位となる。 (Encoding tree layer) As shown in (e) of FIG. 2, the encoding tree layer defines a set of data that the image decoding device 31 refers to in order to decode the encoding tree block to be processed. . The coding tree unit is divided by recursive quadtree division. A node having a tree structure obtained by recursive quadtree partitioning is called a coding tree. An intermediate node of the quadtree is a coded quadtree tree (CQT: Coded Quad Tree), and the CTU is defined as including the highest CQT. The CQT includes a split flag (split_flag). When the split_flag is 1, the CQT is divided into four CQTs (including four CQTs). When split_flag is 0, CQT includes a coding unit (CU: Coded Unit) that is a terminal node. The encoding unit CU is a basic unit of the encoding process.
  (符号化ユニットレイヤ) 符号化ユニットレイヤは、図3の(f)に示すように、処理対象の符号化ユニットを復号するために画像復号装置31が参照するデータの集合が規定されている。具体的には、符号化ユニットは、CUヘッダCUH、予測ツリー、変換ツリー、CUヘッダCUHから構成される。CUヘッダCUHでは、符号化ユニットが、イントラ予測を用いるユニットであるか、インター予測を用いるユニットであるかなどが規定される。符号化ユニットは、予測ツリー(prediction tree;PT)および変換ツリー(transform tree;TT)のルートとなる。CUヘッダCUHは、予測ツリーと変換ツリーの間、もしくは、変換ツリーの後に含まれる。 (Encoding unit layer) As shown in (f) of FIG. 3, the encoding unit layer defines a set of data referred to by the image decoding device 31 in order to decode the encoding unit to be processed. Specifically, the encoding unit includes a CU header CUH, a prediction tree, a conversion tree, and a CU header CUH. In the CU header CUH, it is defined whether the coding unit is a unit using intra prediction or a unit using inter prediction. The encoding unit is the root of a prediction tree (prediction tree; PT) and a transformation tree (transform tree; TT). The CU header CUH is included between the prediction tree and the conversion tree or after the conversion tree.
 予測ツリーは、符号化ユニットが1または複数の予測ブロックに分割され、各予測ブロックの位置とサイズとが規定される。別の表現でいえば、予測ブロックは、符号化ユニットを構成する1または複数の重複しない領域である。また、予測ツリーは、上述の分割により得られた1または複数の予測ブロックを含む。 In the prediction tree, the coding unit is divided into one or a plurality of prediction blocks, and the position and size of each prediction block are defined. In other words, the prediction block is one or a plurality of non-overlapping areas constituting the coding unit. The prediction tree includes one or a plurality of prediction blocks obtained by the above division.
 予測処理は、この予測ブロックごとに行われる。以下、予測の単位である予測ブロックのことを、予測単位(prediction unit;PU、予測ユニット)とも称する。 Prediction processing is performed for each prediction block. Hereinafter, a prediction block which is a unit of prediction is also referred to as a prediction unit (PU, prediction unit).
 予測ツリーにおける分割の種類は、大まかにいえば、イントラ予測の場合と、インター予測の場合との2つがある。イントラ予測とは、同一ピクチャ内の予測であり、インター予測とは、互いに異なるピクチャ間(例えば、表示時刻間、レイヤ画像間)で行われる予測処理を指す。 There are roughly two types of division in the prediction tree: intra prediction and inter prediction. Intra prediction is prediction within the same picture, and inter prediction refers to prediction processing performed between different pictures (for example, between display times and between layer images).
 イントラ予測の場合、分割方法は、2N×2N(符号化ユニットと同一サイズ)と、N×Nとがある。 In the case of intra prediction, there are 2N × 2N (the same size as the encoding unit) and N × N division methods.
 また、インター予測の場合、分割方法は、符号化データのpart_modeにより符号化され、2N×2N(符号化ユニットと同一サイズ)、2N×N、2N×nU、2N×nD、N×2N、nL×2N、nR×2N、および、N×Nなどがある。なお、2N×nUは、2N×2Nの符号化ユニットを上から順に2N×0.5Nと2N×1.5Nの2領域に分割することを示す。2N×nDは、2N×2Nの符号化ユニットを上から順に2N×1.5Nと2N×0.5Nの2領域に分割することを示す。nL×2Nは、2N×2Nの符号化ユニットを左から順に0.5N×2Nと1.5N×2Nの2領域に分割することを示す。nR×2Nは、2N×2Nの符号化ユニットを左から順に1.5N×2Nと0.5N×1.5Nの2領域に分割することを示す。分割数は1、2、4のいずれかであるため、CUに含まれるPUは1個から4個である。これらのPUを順にPU0、PU1、PU2、PU3と表現する。 Further, in the case of inter prediction, the division method is encoded by part_mode of encoded data, and 2N × 2N (the same size as the encoding unit), 2N × N, 2N × nU, 2N × nD, N × 2N, nL X2N, nRx2N, and NxN. Note that 2N × nU indicates that a 2N × 2N encoding unit is divided into two regions of 2N × 0.5N and 2N × 1.5N in order from the top. 2N × nD indicates that a 2N × 2N encoding unit is divided into two regions of 2N × 1.5N and 2N × 0.5N in order from the top. nL × 2N indicates that a 2N × 2N encoding unit is divided into two regions of 0.5N × 2N and 1.5N × 2N in order from the left. nR × 2N indicates that a 2N × 2N encoding unit is divided into two regions of 1.5N × 2N and 0.5N × 1.5N in order from the left. Since the number of divisions is one of 1, 2, and 4, PUs included in the CU are 1 to 4. These PUs are expressed as PU0, PU1, PU2, and PU3 in order.
 また、変換ツリーにおいては、符号化ユニットが1または複数の変換ブロックに分割され、各変換ブロックの位置とサイズとが規定される。別の表現でいえば、変換ブロックは、符号化ユニットを構成する1または複数の重複しない領域のことである。また、変換ツリーは、上述の分割より得られた1または複数の変換ブロックを含む。 Also, in the transform tree, the encoding unit is divided into one or a plurality of transform blocks, and the position and size of each transform block are defined. In other words, the transform block is one or a plurality of non-overlapping areas constituting the encoding unit. The conversion tree includes one or a plurality of conversion blocks obtained by the above division.
 変換ツリーにおける分割には、符号化ユニットと同一のサイズの領域を変換ブロックとして割り付けるものと、上述したツリーブロックの分割と同様、再帰的な4分木分割によるものがある。 The division in the transformation tree includes the one in which an area having the same size as that of the encoding unit is assigned as the transformation block, and the one in the recursive quadtree division like the above-described division in the tree block.
 変換処理は、この変換ブロックごとに行われる。以下、変換の単位である変換ブロックのことを、変換単位(transform unit;TU)とも称する。 Conversion processing is performed for each conversion block. Hereinafter, a transform block that is a unit of transformation is also referred to as a transform unit (TU).
  (予測パラメータ) 予測ユニットの予測画像は、予測ユニットに付随する予測パラメータによって導出される。予測パラメータには、イントラ予測の予測パラメータもしくはインター予測の予測パラメータがある。以下、インター予測の予測パラメータ(インター予測パラメータ)について説明する。インター予測パラメータは、予測リスト利用フラグpredFlagL0、predFlagL1と、参照ピクチャインデックスrefIdxL0、refIdxL1と、ベクトルmvL0、mvL1から構成される。予測リスト利用フラグpredFlagL0、predFlagL1は、各々L0リスト、L1リストと呼ばれる参照ピクチャリストが用いられるか否かを示すフラグであり、値が1の場合に対応する参照ピクチャリストが用いられる。なお、本明細書中「XXであるか否かを示すフラグ」と記す場合、1をXXである場合、0をXXではない場合とし、論理否定、論理積などでは1を真、0を偽と扱う(以下同様)。但し、実際の装置や方法では真値、偽値として他の値を用いることもできる。2つの参照ピクチャリストが用いられる場合、つまり、predFlagL0=1, predFlagL1=1の場合が、双予測に対応し、1つの参照ピクチャリストを用いる場合、すなわち(predFlagL0, predFlagL1) = (1, 0)もしくは(predFlagL0, predFlagL1) = (0, 1)の場合が単予測に対応する。なお、予測リスト利用フラグの情報は、後述のインター予測フラグinter_pred_idcで表現することもできる。通常、後述の予測画像生成部、予測パラメータメモリでは、予測リスト利用フラグが用いれ、符号化データから、どの参照ピクチャリストが用いられるか否かの情報を復号する場合にはインター予測フラグinter_pred_idcが用いられる。 (Prediction parameter) The prediction image of the prediction unit is derived by the prediction parameter attached to the prediction unit. The prediction parameters include a prediction parameter for intra prediction or a prediction parameter for inter prediction. Hereinafter, prediction parameters for inter prediction (inter prediction parameters) will be described. The inter prediction parameter includes prediction list use flags predFlagL0 and predFlagL1, reference picture indexes refIdxL0 and refIdxL1, and vectors mvL0 and mvL1. The prediction list use flags predFlagL0 and predFlagL1 are flags indicating whether or not reference picture lists called L0 list and L1 list are used, respectively, and a reference picture list corresponding to a value of 1 is used. In this specification, when “flag indicating whether or not XX” is described, 1 is XX, 0 is not XX, 1 is true and 0 is false in logical negation and logical product. (The same applies hereinafter). However, other values can be used as true values and false values in an actual apparatus or method. When two reference picture lists are used, that is, when predFlagL0 = 1 and predFlagL1 = 1 correspond to bi-prediction, when one reference picture list is used, that is, (predFlagL0, predFlagL1) = (1, 0) Alternatively, the case of (predFlagL0, predFlagL1) = (0, 1) corresponds to single prediction. Note that the prediction list use flag information can also be expressed by an inter prediction flag inter_pred_idc described later. Normally, a prediction list use flag is used in a prediction image generation unit and a prediction parameter memory described later, and an inter prediction flag inter_pred_idc is used when decoding information on which reference picture list is used from encoded data. It is done.
 符号化データに含まれるインター予測パラメータを導出するためのシンタックス要素には、例えば、分割モードpart_mode、マージフラグmerge_flag、マージインデックスmerge_idx、インター予測フラグinter_pred_idc、参照ピクチャインデックスrefIdxLX、予測ベクトルインデックスmvp_LX_idx、差分ベクトルmvdLXがある。 Syntax elements for deriving inter prediction parameters included in the encoded data include, for example, a partition mode part_mode, a merge flag merge_flag, a merge index merge_idx, an inter prediction flag inter_pred_idc, a reference picture index refIdxLX, a prediction vector index mvp_LX_idx, and a difference There is a vector mvdLX.
  (参照ピクチャリストの一例) 次に、参照ピクチャリストの一例について説明する。参照ピクチャリストとは、復号ピクチャ管理部306のDPB3061(図9:詳細は後述)に記憶された参照ピクチャからなる列である。図4の(a)は、参照ピクチャリストの一例を示す概念図である。参照ピクチャリスト601において、左右に一列に配列された5個の長方形は、それぞれ参照ピクチャを示す。左端から右へ順に示されている符号、P1、P2、Q0、P3、P4は、それぞれの参照ピクチャを示す符号である。P1等のPとは、視点Pを示し、そしてQ0のQとは、視点Pとは異なる視点Qを示す。P及びQの添字は、ピクチャ順序番号POCを示す。refIdxLXの真下の下向きの矢印は、参照ピクチャインデックスrefIdxLXが、復号ピクチャ管理部306のDPB3061に記憶された参照ピクチャQ0を参照するインデックスであることを示す。 (Example of reference picture list) Next, an example of the reference picture list will be described. The reference picture list is a sequence of reference pictures stored in the DPB 3061 (FIG. 9: details will be described later) of the decoded picture management unit 306. FIG. 4A is a conceptual diagram illustrating an example of a reference picture list. In the reference picture list 601, five rectangles arranged in a line on the left and right indicate reference pictures, respectively. The codes P1, P2, Q0, P3, and P4 shown in order from the left end to the right are codes indicating respective reference pictures. P such as P1 indicates the viewpoint P, and Q of Q0 indicates a viewpoint Q different from the viewpoint P. The subscripts P and Q indicate the picture order number POC. A downward arrow directly below refIdxLX indicates that the reference picture index refIdxLX is an index that refers to the reference picture Q0 stored in the DPB 3061 of the decoded picture management unit 306.
  (参照ピクチャの例) 次に、ベクトルを導出する際に用いる参照ピクチャの例について説明する。図5は、参照ピクチャの例を示す概念図である。図5において、横軸は表示時刻を示し、縦軸は視点を示す。図5に示されている、縦2行、横3列(計6個)の長方形は、それぞれピクチャを示す。6個の長方形のうち、下行の左から2列目の長方形は復号対象のピクチャ(対象ピクチャ)を示し、残りの5個の長方形がそれぞれ参照ピクチャを示す。対象ピクチャから上向きの矢印で示される参照ピクチャQ0は対象ピクチャと同表示時刻であって視点が異なるピクチャである。対象ピクチャを基準とする変位予測においては、参照ピクチャQ0が用いられる。対象ピクチャから左向きの矢印で示される参照ピクチャP1は、対象ピクチャと同じ視点であって、過去のピクチャである。対象ピクチャから右向きの矢印で示される参照ピクチャP2は、対象ピクチャと同じ視点であって、未来のピクチャである。対象ピクチャを基準とする動き予測においては、参照ピクチャP1又はP2が用いられる。 (Example of reference picture) Next, an example of a reference picture used to derive a vector will be described. FIG. 5 is a conceptual diagram illustrating an example of a reference picture. In FIG. 5, the horizontal axis indicates the display time, and the vertical axis indicates the viewpoint. The rectangles shown in FIG. 5 with 2 rows and 3 columns (6 in total) indicate pictures. Among the six rectangles, the rectangle in the second column from the left in the lower row indicates a picture to be decoded (target picture), and the remaining five rectangles indicate reference pictures. A reference picture Q0 indicated by an upward arrow from the target picture is a picture that has the same display time as the target picture and a different viewpoint. In the displacement prediction based on the target picture, the reference picture Q0 is used. A reference picture P1 indicated by a left-pointing arrow from the target picture is a past picture at the same viewpoint as the target picture. A reference picture P2 indicated by a right-pointing arrow from the target picture is a future picture at the same viewpoint as the target picture. In motion prediction based on the target picture, the reference picture P1 or P2 is used.
  (インター予測フラグと予測リスト利用フラグ) インター予測フラグと、予測リスト利用フラグpredFlagL0、predFlagL1の関係は以下のように相互に変換可能である。そのため、インター予測パラメータとしては、予測リスト利用フラグを用いても良いし、インター予測フラグを用いてもよい。また、以下、予測リスト利用フラグを用いた判定は、インター予測フラグに置き替えても可能である。逆に、インター予測フラグを用いた判定は、予測リスト利用フラグに置き替えても可能である。 (Inter prediction flag and prediction list use flag) The relationship between the inter prediction flag and the prediction list use flags predFlagL0 and predFlagL1 can be mutually converted as follows. Therefore, as an inter prediction parameter, a prediction list use flag may be used, or an inter prediction flag may be used. In addition, hereinafter, the determination using the prediction list use flag may be replaced with the inter prediction flag. Conversely, the determination using the inter prediction flag can be performed by replacing the prediction list use flag.
 インター予測フラグ = (predFlagL1<<1) + predFlagL0
 predFlagL0 =インター予測フラグ & 1
 predFlagL1 =インター予測フラグ >> 1
 ここで、>>は右シフト、<<は左シフトである。
Inter prediction flag = (predFlagL1 << 1) + predFlagL0
predFlagL0 = Inter prediction flag & 1
predFlagL1 = Inter prediction flag >> 1
Here, >> is a right shift, and << is a left shift.
  (マージ予測とAMVP予測) 予測パラメータの復号(符号化)方法には、マージ予測(merge)モードとAMVP(Adaptive Motion Vector Prediction、適応動きベクトル予測)モードがある、マージフラグmerge_flagは、これらを識別するためのフラグである。マージ予測モードでも、AMVPモードでも、既に処理済みのブロックの予測パラメータを用いて、対象PUの予測パラメータが導出される。マージ予測モードは、予測リスト利用フラグpredFlagLX(インター予測フラグinter_pred_idcinter_pred_idc)、参照ピクチャインデックスrefIdxLX、ベクトルmvLXを符号化データに含めずに、既に導出した予測パラメータをそのまま用いるモードであり、AMVPモードは、インター予測フラグinter_pred_idcinter_pred_idc、参照ピクチャインデックスrefIdxLX、ベクトルmvLXを符号化データに含めるモードである。なおベクトルmvLXは、予測ベクトルを示す予測ベクトルインデックスmvp_LX_idxと差分ベクトル(mvdLX)として符号化される。 (Merge prediction and AMVP prediction) The prediction parameter decoding (encoding) method includes merge prediction (merge) mode and AMVP (Adaptive 、 Motion Vector Prediction, adaptive motion vector prediction) mode. The merge flag merge_flag identifies these It is a flag to do. In both the merge prediction mode and the AMVP mode, the prediction parameter of the target PU is derived using the prediction parameter of the already processed block. The merge prediction mode is a mode that uses the prediction parameter already derived without including the prediction list use flag predFlagLX (inter prediction flag inter_pred_idcinter_pred_idc), the reference picture index refIdxLX, and the vector mvLX in the encoded data. In this mode, the prediction flag inter_pred_idcinter_pred_idc, the reference picture index refIdxLX, and the vector mvLX are included in the encoded data. The vector mvLX is encoded as a prediction vector index mvp_LX_idx indicating a prediction vector and a difference vector (mvdLX).
 インター予測フラグinter_pred_idcは、参照ピクチャの種類および数を示すデータであり、Pred_L0、Pred_L1、Pred_Biの何れかの値をとる。Pred_L0、Pred_L1は、各々L0リスト、L1リストと呼ばれる参照ピクチャリストに記憶された参照ピクチャが用いられることを示し、共に1枚の参照ピクチャを用いること(単予測)を示す。L0リスト、L1リストを用いた予測を各々L0予測、L1予測と呼ぶ。Pred_Biは2枚の参照ピクチャを用いること(双予測)を示し、L0リストとL1リストに記憶された参照ピクチャの2つを用いることを示す。予測ベクトルインデックスmvp_LX_idxは予測ベクトルを示すインデックスであり、参照ピクチャインデックスrefIdxLXは、参照ピクチャリストに記憶された参照ピクチャを示すインデックスである。なお、LXは、L0予測とL1予測を区別しない場合に用いられる記述方法であり、LXをL0、L1に置き換えることでL0リストに対するパラメータとL1リストに対するパラメータを区別するする。例えば、refIdxL0はL0予測に用いる参照ピクチャインデックス、refIdxL1はL1予測に用いる参照ピクチャインデックス、refIdx(refIdxLX)は、refIdxL0とrefIdxL1を区別しない場合に用いられる表記である。 The inter prediction flag inter_pred_idc is data indicating the type and number of reference pictures, and takes one of the values Pred_L0, Pred_L1, and Pred_Bi. Pred_L0 and Pred_L1 indicate that reference pictures stored in reference picture lists called an L0 list and an L1 list are used, respectively, and that both use one reference picture (single prediction). Prediction using the L0 list and the L1 list are referred to as L0 prediction and L1 prediction, respectively. Pred_Bi indicates that two reference pictures are used (bi-prediction), and indicates that two reference pictures stored in the L0 list and the L1 list are used. The prediction vector index mvp_LX_idx is an index indicating a prediction vector, and the reference picture index refIdxLX is an index indicating a reference picture stored in the reference picture list. Note that LX is a description method used when L0 prediction and L1 prediction are not distinguished. By replacing LX with L0 and L1, parameters for the L0 list and parameters for the L1 list are distinguished. For example, refIdxL0 is a reference picture index used for L0 prediction, refIdxL1 is a reference picture index used for L1 prediction, and refIdx (refIdxLX) is a notation used when refIdxL0 and refIdxL1 are not distinguished.
 マージインデックスmerge_idxは、処理が完了したブロックから導出される予測パラメータ候補(マージ候補)のうち、いずれかの予測パラメータを復号対象ブロックの予測パラメータとして用いるかを示すインデックスである。 The merge index merge_idx is an index indicating which one of the prediction parameter candidates (merge candidates) derived from the processed block is used as the prediction parameter of the decoding target block.
  (動きベクトルと変位ベクトル) ベクトルmvLXには、動きベクトルと変位ベクトル(disparity vector、視差ベクトル)がある。動きベクトルとは、あるレイヤのある表示時刻でのピクチャにおけるブロックの位置と、異なる表示時刻(例えば、隣接する離散時刻)における同一のレイヤのピクチャにおける対応するブロックの位置との間の位置のずれを示すベクトルである。変位ベクトルとは、あるレイヤのある表示時刻でのピクチャにおけるブロックの位置と、同一の表示時刻における異なるレイヤのピクチャにおける対応するブロックの位置との間の位置のずれを示すベクトルである。異なるレイヤのピクチャとしては、異なる視点のピクチャである場合、もしくは、異なる解像度のピクチャである場合などがある。特に、異なる視点のピクチャに対応する変位ベクトルを視差ベクトルと呼ぶ。以下の説明では、動きベクトルと変位ベクトルを区別しない場合には、単にベクトルmvLXと呼ぶ。ベクトルmvLXに関する予測ベクトル、差分ベクトルを、それぞれ予測ベクトルmvpLX、差分ベクトルmvdLXと呼ぶ。ベクトルmvLXおよび差分ベクトルmvdLXが、動きベクトルであるか、変位ベクトルであるかは、ベクトルに付随する参照ピクチャインデックスrefIdxLXを用いて判別される。 (Motion vector and displacement vector) The vector mvLX includes a motion vector and a displacement vector (disparity vector). A motion vector is a positional shift between the position of a block in a picture at a certain display time of a layer and the position of the corresponding block in a picture of the same layer at a different display time (for example, an adjacent discrete time). It is a vector which shows. The displacement vector is a vector indicating a positional shift between the position of a block in a picture at a certain display time of a certain layer and the position of a corresponding block in a picture of a different layer at the same display time. The pictures in different layers may be pictures from different viewpoints or pictures with different resolutions. In particular, a displacement vector corresponding to pictures of different viewpoints is called a disparity vector. In the following description, when a motion vector and a displacement vector are not distinguished, they are simply referred to as a vector mvLX. A prediction vector and a difference vector related to the vector mvLX are referred to as a prediction vector mvpLX and a difference vector mvdLX, respectively. Whether the vector mvLX and the difference vector mvdLX are motion vectors or displacement vectors is determined using a reference picture index refIdxLX attached to the vectors.
 次に、本実施形態に係る画像復号装置31の構成について説明する。図6は、本実施形態に係る画像復号装置31の構成を示す概略図である。画像復号装置31は、エントロピー復号部301、予測パラメータ復号部302、復号ピクチャ管理部(参照画像記憶部、フレームメモリ)306、予測画像生成部308、逆量子化・逆DCT部311、及び加算部312、残差格納部313(残差記録部)を含んで構成される。 Next, the configuration of the image decoding device 31 according to the present embodiment will be described. FIG. 6 is a schematic diagram illustrating a configuration of the image decoding device 31 according to the present embodiment. The image decoding device 31 includes an entropy decoding unit 301, a prediction parameter decoding unit 302, a decoded picture management unit (reference image storage unit, frame memory) 306, a predicted image generation unit 308, an inverse quantization / inverse DCT unit 311, and an addition unit 312 and a residual storage unit 313 (residual recording unit).
 また、予測パラメータ復号部302は、インター予測パラメータ復号部303及びイントラ予測パラメータ復号部304を含んで構成される。予測画像生成部308は、インター予測画像生成部309及びイントラ予測画像生成部310を含んで構成される。 The prediction parameter decoding unit 302 includes an inter prediction parameter decoding unit 303 and an intra prediction parameter decoding unit 304. The predicted image generation unit 308 includes an inter predicted image generation unit 309 and an intra predicted image generation unit 310.
 エントロピー復号部301は、外部から入力された符号化ストリームTeに対してエントロピー復号を行って、個々の符号(シンタックス要素)を分離し復号する。分離された符号には、予測画像を生成するための予測情報および、差分画像を生成するための残差情報などがある。 The entropy decoding unit 301 performs entropy decoding on the encoded stream Te input from the outside, and separates and decodes individual codes (syntax elements). The separated codes include prediction information for generating a prediction image and residual information for generating a difference image.
 エントロピー復号部301は、分離した符号の一部を予測パラメータ復号部302に出力する。分離した符号の一部とは、例えば、予測モードPredMode、分割モードpart_mode、マージフラグmerge_flag、マージインデックスmerge_idx、インター予測フラグinter_pred_idcinter_pred_idc、参照ピクチャインデックスrefIdxLX、予測ベクトルインデックmvp_LX_idx、差分ベクトルmvdLX、残差予測フラグiv_res_pred_weight_idxである。どの符号を復号するか否かの制御は、予測パラメータ復号部302の指示に基づいて行われる。エントロピー復号部301は、量子化係数を逆量子化・逆DCT部311に出力する。この量子化係数は、符号化処理において、残差信号に対してDCT(Discrete Cosine Transform、離散コサイン変換)を行い量子化して得られる係数である。 The entropy decoding unit 301 outputs a part of the separated code to the prediction parameter decoding unit 302. Some of the separated codes are, for example, prediction mode PredMode, split mode part_mode, merge flag merge_flag, merge index merge_idx, inter prediction flag inter_pred_idcinter_pred_idc, reference picture index refIdxLX, prediction vector index mvp_LX_idx, difference vector mvdLX, residual prediction flag iv_res_pred_weight_idx. Control of which code to decode is performed based on an instruction from the prediction parameter decoding unit 302. The entropy decoding unit 301 outputs the quantization coefficient to the inverse quantization / inverse DCT unit 311. The quantization coefficient is a coefficient obtained by performing quantization by performing DCT (Discrete Cosine Transform) on the residual signal in the encoding process.
 インター予測パラメータ復号部303は、エントロピー復号部301から入力された符号に基づいて、復号ピクチャ管理部306の予測パラメータメモリ3067に記憶された予測パラメータを参照してインター予測パラメータを復号する。 The inter prediction parameter decoding unit 303 refers to the prediction parameter stored in the prediction parameter memory 3067 of the decoded picture management unit 306 based on the code input from the entropy decoding unit 301 and decodes the inter prediction parameter.
 インター予測パラメータ復号部303は、復号したインター予測パラメータを予測画像生成部308に出力し、また復号ピクチャ管理部306の予測パラメータメモリ3067に記憶する。インター予測パラメータ復号部303の詳細については後述する。 The inter prediction parameter decoding unit 303 outputs the decoded inter prediction parameters to the prediction image generation unit 308 and stores them in the prediction parameter memory 3067 of the decoded picture management unit 306. Details of the inter prediction parameter decoding unit 303 will be described later.
 イントラ予測パラメータ復号部304は、エントロピー復号部301から入力された符号に基づいて、復号ピクチャ管理部306の予測パラメータメモリ3067に記憶された予測パラメータを参照してイントラ予測パラメータを復号する。イントラ予測パラメータとは、ピクチャブロックを1つのピクチャ内で予測する処理で用いるパラメータ、例えば、イントラ予測モードIntraPredModeである。イントラ予測パラメータ復号部304は、復号したイントラ予測パラメータを予測画像生成部308に出力し、また復号ピクチャ管理部306の予測パラメータメモリ3067に記憶する。 The intra prediction parameter decoding unit 304 refers to the prediction parameter stored in the prediction parameter memory 3067 of the decoded picture management unit 306 based on the code input from the entropy decoding unit 301 and decodes the intra prediction parameter. The intra prediction parameter is a parameter used in a process of predicting a picture block within one picture, for example, an intra prediction mode IntraPredMode. The intra prediction parameter decoding unit 304 outputs the decoded intra prediction parameter to the prediction image generation unit 308 and stores it in the prediction parameter memory 3067 of the decoded picture management unit 306.
 イントラ予測パラメータ復号部304は、輝度と色差で異なるイントラ予測モードを導出しても良い。この場合、イントラ予測パラメータ復号部304は、輝度の予測パラメータとして輝度予測モードIntraPredModeY、色差の予測パラメータとして、色差予測モードIntraPredModeCを復号する。輝度予測モードIntraPredModeYは、35モードであり、プレーナ予測(0)、DC予測(1)、方向予測(2~34)が対応する。色差予測モードIntraPredModeCは、プレーナ予測(0)、DC予測(1)、方向予測(2、3、4)、LMモード(5)の何れかを用いるもの。 The intra prediction parameter decoding unit 304 may derive different intra prediction modes depending on luminance and color difference. In this case, the intra prediction parameter decoding unit 304 decodes the luminance prediction mode IntraPredModeY as the luminance prediction parameter and the color difference prediction mode IntraPredModeC as the color difference prediction parameter. The luminance prediction mode IntraPredModeY is a 35 mode and corresponds to planar prediction (0), DC prediction (1), and direction prediction (2 to 34). The color difference prediction mode IntraPredModeC uses one of planar prediction (0), DC prediction (1), direction prediction (2, 3, 4), and LM mode (5).
 復号ピクチャ管理部306は、加算部312が生成した参照ピクチャのブロック(参照ピクチャブロック)を、DPB3061において、復号対象のピクチャ及びブロック毎に予め定めた位置に記憶するとともに、所定のタイミングで参照ピクチャブロックをピクチャ毎に統合した復号視点画像Tdを外部に出力する。 The decoded picture management unit 306 stores the reference picture block (reference picture block) generated by the addition unit 312 in the DPB 3061 at a predetermined position for each decoding target picture and block, and at a predetermined timing. A decoded viewpoint image Td in which blocks are integrated for each picture is output to the outside.
 また、復号ピクチャ管理部306は、予測パラメータを、予測パラメータメモリ3067において、復号対象のピクチャ及びブロック毎に予め定めた位置に記憶する。復号ピクチャ管理部306の詳細については、図9を用いて後に説明する。 Also, the decoded picture management unit 306 stores the prediction parameter in the prediction parameter memory 3067 at a predetermined position for each picture and block to be decoded. Details of the decoded picture management unit 306 will be described later with reference to FIG.
 予測画像生成部308には、エントロピー復号部301から入力された予測モードpredModeが入力され、また予測パラメータ復号部302から予測パラメータが入力される。また、予測画像生成部308は、復号ピクチャ管理部306のDPB3061から参照ピクチャを読み出す。予測画像生成部308は、予測モードpredModeが示す予測モードで、入力された予測パラメータと読み出した参照ピクチャを用いて予測ピクチャブロックP(予測画像)を生成する。 The prediction image generation unit 308 receives the prediction mode predMode input from the entropy decoding unit 301 and the prediction parameter from the prediction parameter decoding unit 302. Further, the predicted image generation unit 308 reads the reference picture from the DPB 3061 of the decoded picture management unit 306. The predicted image generation unit 308 generates a predicted picture block P (predicted image) using the input prediction parameter and the read reference picture in the prediction mode indicated by the prediction mode predMode.
 ここで、予測モードpredModeがインター予測モードを示す場合、インター予測画像生成部309は、インター予測パラメータ復号部303から入力されたインター予測パラメータと読み出した参照ピクチャを用いてインター予測により予測ピクチャブロックPを生成する。予測ピクチャブロックPは予測単位PUに対応する。PUは、上述したように予測処理を行う単位となる複数の画素からなるピクチャの一部分、つまり1度に予測処理が行われる復号対象ブロックに相当する。 Here, when the prediction mode predMode indicates the inter prediction mode, the inter prediction image generation unit 309 uses the inter prediction parameter input from the inter prediction parameter decoding unit 303 and the read reference picture to perform the prediction picture block P by inter prediction. Is generated. The predicted picture block P corresponds to the prediction unit PU. The PU corresponds to a part of a picture composed of a plurality of pixels as a unit for performing the prediction process as described above, that is, a decoding target block on which the prediction process is performed at a time.
 インター予測画像生成部309は、予測リスト利用フラグpredFlagLXが1である参照ピクチャリスト(RPS:Reference Picture List)(L0リスト、もしくはL1リスト)に対し、参照ピクチャインデックスrefIdxLXで示される参照ピクチャから、復号対象ブロックを基準としてベクトルmvLXが示す位置にある参照ピクチャブロックを復号ピクチャ管理部306のDPB3061から読み出す。インター予測画像生成部309は、読み出した参照ピクチャブロックについて予測を行って予測ピクチャブロックPを生成する。インター予測画像生成部309は、生成した予測ピクチャブロックPを加算部312に出力する。 The inter prediction image generation unit 309 decodes a reference picture list (RPS: ReferenceSPicture List) (L0 list or L1 list) whose prediction list use flag predFlagLX is 1 from a reference picture indicated by a reference picture index refIdxLX. The reference picture block at the position indicated by the vector mvLX is read from the DPB 3061 of the decoded picture management unit 306 with reference to the target block. The inter prediction image generation unit 309 performs prediction on the read reference picture block to generate a prediction picture block P. The inter prediction image generation unit 309 outputs the generated prediction picture block P to the addition unit 312.
 予測モードpredModeがイントラ予測モードを示す場合、イントラ予測画像生成部310は、イントラ予測パラメータ復号部304から入力されたイントラ予測パラメータと読み出した参照ピクチャを用いてイントラ予測を行う。具体的には、イントラ予測画像生成部310は、復号対象のピクチャであって、既に復号されたブロックのうち復号対象ブロックから予め定めた範囲にある参照ピクチャブロックを復号ピクチャ管理部306のDPB3061から読み出す。予め定めた範囲とは、復号対象ブロックがいわゆるラスタースキャンの順序で順次移動する場合、例えば、左、左上、上、右上の隣接ブロックのうちのいずれかであり、イントラ予測モードによって異なる。ラスタースキャンの順序とは、各ピクチャにおいて、上端から下端まで各行について、順次左端から右端まで移動させる順序である。 When the prediction mode predMode indicates the intra prediction mode, the intra predicted image generation unit 310 performs intra prediction using the intra prediction parameter input from the intra prediction parameter decoding unit 304 and the read reference picture. Specifically, the intra-predicted image generation unit 310 extracts a reference picture block that is a decoding target picture and is in a predetermined range from the decoding target block among the already decoded blocks from the DPB 3061 of the decoded picture management unit 306. read out. The predetermined range is, for example, any of the left, upper left, upper, and upper right adjacent blocks when the decoding target block sequentially moves in a so-called raster scan order, and varies depending on the intra prediction mode. The raster scan order is an order in which each row is sequentially moved from the left end to the right end in each picture from the upper end to the lower end.
 イントラ予測画像生成部310は、読み出した参照ピクチャブロックについてイントラ予測モードIntraPredModeが示す予測モードで予測を行って予測ピクチャブロックを生成する。イントラ予測画像生成部310は、生成した予測ピクチャブロックPを加算部312に出力する。 The intra predicted image generation unit 310 performs prediction in the prediction mode indicated by the intra prediction mode IntraPredMode for the read reference picture block, and generates a predicted picture block. The intra predicted image generation unit 310 outputs the generated predicted picture block P to the addition unit 312.
 イントラ予測パラメータ復号部304において、輝度と色差で異なるイントラ予測モードを導出する場合、イントラ予測画像生成部310は、輝度予測モードIntraPredModeYに応じて、プレーナ予測(0)、DC予測(1)、方向予測(2~34)の何れかによって輝度の予測ピクチャブロックを生成し、色差予測モードIntraPredModeCに応じて、プレーナ予測(0)、DC予測(1)、方向予測(2、3、4)、LMモード(5)の何れかによって色差の予測ピクチャブロックを生成する。 When the intra prediction parameter decoding unit 304 derives an intra prediction mode different in luminance and color difference, the intra prediction image generation unit 310 performs planar prediction (0), DC prediction (1), direction according to the luminance prediction mode IntraPredModeY. A prediction picture block of luminance is generated according to any of prediction (2 to 34), and planar prediction (0), DC prediction (1), direction prediction (2, 3, 4), LM according to the color difference prediction mode IntraPredModeC A color difference prediction picture block is generated in any one of modes (5).
 逆量子化・逆DCT部311は、エントロピー復号部301から入力された量子化係数を逆量子化してDCT係数を求める。逆量子化・逆DCT部311は、求めたDCT係数について逆DCT(Inverse Discrete Cosine Transform、逆離散コサイン変換)を行い、復号残差信号を算出する。逆量子化・逆DCT部311は、算出した復号残差信号を加算部312および残差格納部313に出力する。 The inverse quantization / inverse DCT unit 311 inversely quantizes the quantization coefficient input from the entropy decoding unit 301 to obtain a DCT coefficient. The inverse quantization / inverse DCT unit 311 performs inverse DCT (Inverse Discrete Cosine Transform) on the obtained DCT coefficient to calculate a decoded residual signal. The inverse quantization / inverse DCT unit 311 outputs the calculated decoded residual signal to the addition unit 312 and the residual storage unit 313.
 加算部312は、インター予測画像生成部309及びイントラ予測画像生成部310から入力された予測ピクチャブロックPと逆量子化・逆DCT部311から入力された復号残差信号の信号値を画素毎に加算して、参照ピクチャブロックを生成する。加算部312は、生成した参照ピクチャブロック(すなわち復号ピクチャ)を復号ピクチャ管理部306のDPB3061に記憶する。 The adder 312 outputs the prediction picture block P input from the inter prediction image generation unit 309 and the intra prediction image generation unit 310 and the signal value of the decoded residual signal input from the inverse quantization / inverse DCT unit 311 for each pixel. Addition to generate a reference picture block. The adding unit 312 stores the generated reference picture block (ie, decoded picture) in the DPB 3061 of the decoded picture management unit 306.
  [インター予測パラメータ復号部の構成] 次に、インター予測パラメータ復号部303の構成について説明する。図7は、本実施形態に係るインター予測パラメータ復号部303の構成を示す概略図である。インター予測パラメータ復号部303は、インター予測パラメータ復号制御部3031、AMVP予測パラメータ導出部3032、加算部3035及びマージ予測パラメータ導出部3036を含んで構成される。 [Configuration of Inter Prediction Parameter Decoding Unit] Next, the configuration of the inter prediction parameter decoding unit 303 will be described. FIG. 7 is a schematic diagram illustrating a configuration of the inter prediction parameter decoding unit 303 according to the present embodiment. The inter prediction parameter decoding unit 303 includes an inter prediction parameter decoding control unit 3031, an AMVP prediction parameter derivation unit 3032, an addition unit 3035, and a merge prediction parameter derivation unit 3036.
  [[インター予測パラメータ復号制御部]] インター予測パラメータ復号制御部3031は、インター予測に関連する符号(シンタックス要素)の復号をエントロピー復号部301に指示し、符号化データに含まれる符号(シンタックス要素)を例えば、分割モードpart_mode、マージフラグmerge_flag、マージインデックスmerge_idx、インター予測フラグinter_pred_idcinter_pred_idc、参照ピクチャインデックスrefIdxLX、予測ベクトルインデックスmvp_LX_idx、差分ベクトルmvdLX、残差予測フラグiv_res_pred_weight_idxを抽出する。 [[Inter Prediction Parameter Decoding Control Unit]] The inter prediction parameter decoding control unit 3031 instructs the entropy decoding unit 301 to decode a code (syntax element) related to inter prediction, and the code (synth For example, a partition mode part_mode, a merge flag merge_flag, a merge index merge_idx, an inter prediction flag inter_pred_idcinter_pred_idc, a reference picture index refIdxLX, a prediction vector index mvp_LX_idx, a difference vector mvdLX, and a residual prediction flag iv_res_pred_weight_idx are extracted.
 インター予測パラメータ復号制御部3031は、まず、符号化データからマージフラグを抽出する。インター予測パラメータ復号制御部3031が、あるシンタックス要素を抽出すると表現する場合は、あるシンタックス要素の復号をエントロピー復号部301に指示し、該当のシンタックス要素を符号化データから読み出すことを意味する。ここで、マージフラグが示す値が1、すなわち、マージ予測モードを示す場合、インター予測パラメータ復号制御部3031は、マージ予測に係る予測パラメータとして、マージインデックスmerge_idxを抽出する。インター予測パラメータ復号制御部3031は、抽出したマージインデックスmerge_idxをマージ予測パラメータ導出部3036に出力する。 The inter prediction parameter decoding control unit 3031 first extracts a merge flag from the encoded data. When the inter prediction parameter decoding control unit 3031 expresses that a certain syntax element is to be extracted, it means that the entropy decoding unit 301 is instructed to decode a certain syntax element, and the corresponding syntax element is read from the encoded data. To do. Here, when the value indicated by the merge flag is 1, that is, indicates the merge prediction mode, the inter prediction parameter decoding control unit 3031 extracts the merge index merge_idx as a prediction parameter related to merge prediction. The inter prediction parameter decoding control unit 3031 outputs the extracted merge index merge_idx to the merge prediction parameter derivation unit 3036.
 マージフラグmerge_flagが0、すなわち、AMVP予測モードを示す場合、インター予測パラメータ復号制御部3031は、エントロピー復号部301を用いて符号化データからAMVP予測パラメータを抽出する。AMVP予測パラメータとして、例えば、インター予測フラグinter_pred_idc、参照ピクチャインデックスrefIdxLX、ベクトルインデックスmvp_LX_idx、差分ベクトルmvdLXがある。インター予測パラメータ復号制御部3031は、抽出したインター予測フラグinter_pred_idcinter_pred_idcから導出した予測リスト利用フラグpredFlagLXと、参照ピクチャインデックスrefIdxLXをAMVP予測パラメータ導出部3032及び予測画像生成部308(図6)に出力し、また復号ピクチャ管理部306のDPB3061(図9)に記憶する。インター予測パラメータ復号制御部3031は、抽出したベクトルインデックスmvp_LX_idxをAMVP予測パラメータ導出部3032に出力する。インター予測パラメータ復号制御部3031は、抽出した差分ベクトルmvdLXを加算部3035に出力する。 When the merge flag merge_flag is 0, that is, indicates the AMVP prediction mode, the inter prediction parameter decoding control unit 3031 uses the entropy decoding unit 301 to extract the AMVP prediction parameter from the encoded data. Examples of AMVP prediction parameters include an inter prediction flag inter_pred_idc, a reference picture index refIdxLX, a vector index mvp_LX_idx, and a difference vector mvdLX. The inter prediction parameter decoding control unit 3031 outputs the prediction list use flag predFlagLX derived from the extracted inter prediction flag inter_pred_idcinter_pred_idc and the reference picture index refIdxLX to the AMVP prediction parameter derivation unit 3032 and the prediction image generation unit 308 (FIG. 6). Further, it is stored in the DPB 3061 (FIG. 9) of the decoded picture management unit 306. The inter prediction parameter decoding control unit 3031 outputs the extracted vector index mvp_LX_idx to the AMVP prediction parameter derivation unit 3032. The inter prediction parameter decoding control unit 3031 outputs the extracted difference vector mvdLX to the addition unit 3035.
  [[マージ予測パラメータ導出部]] 図23は、本実施形態に係るマージ予測パラメータ導出部3036の構成を示す概略図である。マージ予測パラメータ導出部3036は、マージ候補導出部30361とマージ候補選択部30362を備える。マージ候補導出部30361は、マージ候補格納部303611と、拡張マージ候補導出部303612と基本マージ候補導出部303613を含んで構成される。 [[Merge Prediction Parameter Deriving Unit]] FIG. 23 is a schematic diagram showing the configuration of the merge prediction parameter deriving unit 3036 according to this embodiment. The merge prediction parameter derivation unit 3036 includes a merge candidate derivation unit 30361 and a merge candidate selection unit 30362. The merge candidate derivation unit 30361 includes a merge candidate storage unit 303611, an extended merge candidate derivation unit 303612, and a basic merge candidate derivation unit 303613.
 マージ候補格納部303611は、拡張マージ候補導出部303612及び基本マージ候補導出部303613から入力されたマージ候補を格納する。なお、マージ候補は、予測リスト利用フラグpredFlagLX、ベクトルmvLX、参照ピクチャインデックスrefIdxLXを含んで構成されている。マージ候補格納部303611において、格納されたマージ候補には、所定の規則に従ってインデックスが割り当てられる。例えば、拡張マージ候補導出部303612から入力されたマージ候補には、インデックスとして「0」を割り当てる。 The merge candidate storage unit 303611 stores the merge candidates input from the extended merge candidate derivation unit 303612 and the basic merge candidate derivation unit 303613. The merge candidate includes a prediction list use flag predFlagLX, a vector mvLX, and a reference picture index refIdxLX. In the merge candidate storage unit 303611, an index is assigned to the stored merge candidates according to a predetermined rule. For example, “0” is assigned as an index to the merge candidate input from the extended merge candidate derivation unit 303612.
 拡張マージ候補導出部303612は、変位ベクトル取得部3036122と、レイヤ間マージ候補導出部3036121と変位マージ候補導出部3036123、図示しないBVSPマージ候補導出部3036124を含んで構成される。 The extended merge candidate derivation unit 303612 includes a displacement vector acquisition unit 3036122, an inter-layer merge candidate derivation unit 3036121, a displacement merge candidate derivation unit 3036123, and a BVSP merge candidate derivation unit 3036124 (not shown).
 変位ベクトル取得部3036122は、まず、復号対象ブロックに隣接する複数の候補ブロック(例えば、左、上、右上に隣接するブロック)から順に変位ベクトルを取得する。具体的には、変位ベクトル取得部3036122は、候補ブロックの一つを選択し、選択した候補ブロックのベクトルが変位ベクトルであるか動きベクトルであるかを、候補ブロックの参照ピクチャインデックスrefIdxLXを用いてリファレンスレイヤ判定機能(後述)を用いて判定し変位ベクトルが有る場合には、それを変位ベクトルとする。候補ブロックに変位ベクトルがない場合には、次の候補ブロックを順に走査する。隣接するブロックに変位ベクトルがない場合、変位ベクトル取得部3036122は、時間的に別の表示順の参照ピクチャに含まれるブロックの対象ブロックに対応する位置のブロックの変位ベクトルの取得を試みる。変位ベクトルが取得できなかった場合には、変位ベクトル取得部3036122は、変位ベクトルとしてゼロベクトルを設定する。変位ベクトル取得部3036122は、変位ベクトルをレイヤ間マージ候補導出部3036121及び変位マージ候補導出部に出力する。 The displacement vector acquisition unit 3036122 first acquires displacement vectors in order from a plurality of candidate blocks adjacent to the decoding target block (for example, blocks adjacent to the left, upper, and upper right). Specifically, the displacement vector acquisition unit 3036122 selects one of the candidate blocks, and uses the reference picture index refIdxLX of the candidate block to determine whether the selected candidate block vector is a displacement vector or a motion vector. If there is a displacement vector determined using a reference layer determination function (described later), it is set as the displacement vector. If there is no displacement vector in the candidate block, the next candidate block is scanned in order. When there is no displacement vector in the adjacent block, the displacement vector acquisition unit 3036122 attempts to acquire the displacement vector of the block at the position corresponding to the target block of the block included in the reference picture in the temporally different display order. When the displacement vector cannot be acquired, the displacement vector acquisition unit 3036122 sets a zero vector as the displacement vector. The displacement vector acquisition unit 3036122 outputs the displacement vector to the inter-layer merge candidate derivation unit 3036121 and the displacement merge candidate derivation unit.
 レイヤ間マージ候補導出部3036121は、変位ベクトル取得部3036122から変位ベクトルを入力される。レイヤ間マージ候補導出部3036121は、別レイヤ(例えばベースレイヤ、ベースビュー)の復号対象ピクチャと同一POCを持つピクチャ内から、変位ベクトル取得部3036122から入力された変位ベクトルだけが示すブロックを選択し、該ブロックが有する動きベクトルである予測パラメータを予測パラメータメモリ307から読み出す。より具体的には、レイヤ間マージ候補導出部3036121が読みだす予測パラメータは、対象ブロックの中心点を起点にしたときに、起点の座標に変位ベクトルを加算した座標を含むブロックの予測パラメータである。
参照ブロックの座標(xRef、yRef)は、対象ブロックの座標が(xP、yP)、変位ベクトルが(mvDisp[0]、mvDisp[1])、対象ブロックの幅と高さがnPSW、nPSHの場合に以下の式により導出する。
The inter-layer merge candidate derivation unit 3036121 receives the displacement vector from the displacement vector acquisition unit 3036122. The inter-layer merge candidate derivation unit 3036121 selects a block indicated only by the displacement vector input from the displacement vector acquisition unit 3036122 from a picture having the same POC as the decoding target picture of another layer (eg, base layer, base view). The prediction parameter, which is a motion vector included in the block, is read from the prediction parameter memory 307. More specifically, the prediction parameter read by the inter-layer merge candidate derivation unit 3036121 is a prediction parameter of a block including coordinates obtained by adding a displacement vector to the coordinates of the starting point when the center point of the target block is the starting point. .
The coordinates (xRef, yRef) of the reference block are the coordinates of the target block (xP, yP), the displacement vector (mvDisp [0], mvDisp [1]), and the width and height of the target block are nPSW, nPSH. Is derived by the following equation.
xRef = Clip3( 0, PicWidthInSamples- 1, xP + ( ( nPSW - 1 ) >> 1 ) + ( ( mvDisp[0] + 2 ) >> 2 ) )
yRef = Clip3( 0, PicHeightInSamples- 1, yP + ( ( nPSH - 1 ) >> 1 ) + ( ( mvDisp[1] + 2 ) >> 2 ))
 なお、レイヤ間マージ候補導出部3036121は、予測パラメータが動きベクトルか否かを、インター予測パラメータ復号制御部3031に含まれる後述するリファレンスレイヤ判定機能による判定方法において偽(変位ベクトルではない)と判定した方法により判定する。レイヤ間マージ候補導出部3036121は、読みだした予測パラメータをマージ候補としてマージ候補格納部303611に出力する。また、レイヤ間マージ候補導出部3036121は、予測パラメータを導出出来なかった際には、その旨を変位マージ候補導出部に出力する。本マージ候補は、動き予測のインターレイヤ候補(インタービュー候補)でありレイヤ間マージ候補(動き予測)とも記載する。
xRef = Clip3 (0, PicWidthInSamples L -1, xP + ((nPSW-1) >> 1) + ((mvDisp [0] + 2) >> 2))
yRef = Clip3 (0, PicHeightInSamples L -1, yP + ((nPSH-1) >> 1) + ((mvDisp [1] + 2) >> 2))
Note that the inter-layer merge candidate derivation unit 3036121 determines whether the prediction parameter is a motion vector as false (not a displacement vector) in a determination method using a reference layer determination function described later included in the inter prediction parameter decoding control unit 3031. Judgment is made by the method. The inter-layer merge candidate derivation unit 3036121 outputs the read prediction parameter as a merge candidate to the merge candidate storage unit 303611. Further, when the prediction parameter cannot be derived, the inter-layer merge candidate deriving unit 3036121 outputs that fact to the displacement merge candidate deriving unit. This merge candidate is a motion prediction inter-layer candidate (inter-view candidate) and is also described as an inter-layer merge candidate (motion prediction).
 変位マージ候補導出部3036123は、変位ベクトル取得部3036122から変位ベクトルを入力される。変位マージ候補導出部3036123は、入力された変位ベクトルと、変位ベクトルが指す先のレイヤ画像の参照ピクチャインデックスrefIdxLX(例えば、復号対象ピクチャと同一POCを持つベースレイヤ画像のインデックス)をマージ候補としてマージ候補格納部303611に出力する。本マージ候補は、変位予測のインターレイヤ候補(インタービュー候補)でありレイヤ間マージ候補(変位予測)とも記載する。 The displacement merge candidate derivation unit 3036123 receives the displacement vector from the displacement vector acquisition unit 3036122. The displacement merge candidate derivation unit 3036123 merges the input displacement vector and the reference picture index refIdxLX of the previous layer image indicated by the displacement vector (for example, the index of the base layer image having the same POC as the decoding target picture) as a merge candidate. The data is output to the candidate storage unit 303611. This merge candidate is a displacement prediction inter-layer candidate (inter-view candidate) and is also described as an inter-layer merge candidate (displacement prediction).
 BVSPマージ候補導出部3036124は、ブロック視点合成予測(Block View Synthesis Prediction)マージ候補を導出する。BVSPマージ候補は、別の視点画像から予測画像を生成する変位マージ候補の一種であるが、PU内をさらに小さいブロックに分割して、予測画像生成処理を行うマージ候補である。 The BVSP merge candidate derivation unit 3036124 derives block view synthesis prediction (Block View Synthesis Prediction) merge candidates. A BVSP merge candidate is a type of displacement merge candidate that generates a predicted image from another viewpoint image, but is a merge candidate that divides the PU into smaller blocks and performs a predicted image generation process.
 基本マージ候補導出部303613は、空間マージ候補導出部3036131と時間マージ候補導出部3036132と結合マージ候補導出部3036133とゼロマージ候補導出部3036134を含んで構成される。 The basic merge candidate derivation unit 303613 includes a spatial merge candidate derivation unit 3036131, a temporal merge candidate derivation unit 3036132, a merge merge candidate derivation unit 3036133, and a zero merge candidate derivation unit 3036134.
 空間マージ候補導出部3036131は、所定の規則に従って、予測パラメータメモリ307が記憶している予測パラメータ(予測リスト利用フラグpredFlagLX、ベクトルmvLX、参照ピクチャインデックスrefIdxLX)を読み出し、読み出した予測パラメータをマージ候補として導出する。読み出される予測パラメータは、復号対象ブロックから予め定めた範囲内にあるブロック(例えば、復号対象ブロックの左下端、左上端、右上端にそれぞれ接するブロックの全部又は一部)のそれぞれに係る予測パラメータである。導出されたマージ候補はマージ候補格納部303611に格納される。 The spatial merge candidate derivation unit 3036131 reads the prediction parameters (prediction list use flag predFlagLX, vector mvLX, reference picture index refIdxLX) stored in the prediction parameter memory 307 according to a predetermined rule, and uses the read prediction parameters as merge candidates. To derive. The prediction parameter to be read is a prediction parameter relating to each of the blocks within a predetermined range from the decoding target block (for example, all or a part of the blocks in contact with the lower left end, upper left upper end, and upper right end of the decoding target block, respectively). is there. The derived merge candidates are stored in the merge candidate storage unit 303611.
 時間マージ候補導出部3036132は、復号対象ブロックの右下の座標を含む参照画像中のブロックの予測パラメータを予測パラメータメモリ307から読みだしマージ候補とする。参照画像の指定方法は、例えば、スライスヘッダに置いて指定された参照ピクチャインデックスrefIdxLXでも良いし、復号対象ブロックに隣接するブロックの参照ピクチャインデックスrefIdxLXのうち最小のものを用いて指定しても良い。導出されたマージ候補はマージ候補格納部303611に格納される。 The temporal merge candidate derivation unit 3036132 reads the prediction parameter of the block in the reference image including the lower right coordinate of the decoding target block from the prediction parameter memory 307 and sets it as a merge candidate. The reference picture designation method may be, for example, the reference picture index refIdxLX designated in the slice header, or may be designated using the smallest reference picture index refIdxLX of the block adjacent to the decoding target block. . The derived merge candidates are stored in the merge candidate storage unit 303611.
 結合マージ候補導出部3036133は、既に導出されマージ候補格納部303611に格納された2つの異なる導出済マージ候補のベクトルと参照ピクチャインデックスを、それぞれL0、L1のベクトルとして組み合わせることで結合マージ候補を導出する。導出されたマージ候補はマージ候補格納部303611に格納される。 The merge merge candidate derivation unit 3036133 derives merge merge candidates by combining two different derived merge candidate vectors and reference picture indexes already derived and stored in the merge candidate storage unit 303611 as L0 and L1 vectors, respectively. To do. The derived merge candidates are stored in the merge candidate storage unit 303611.
 ゼロマージ候補導出部3036134は、参照ピクチャインデックスrefIdxLXが0であり、ベクトルmvLXのX成分、Y成分が共に0であるマージ候補を導出する。導出されたマージ候補はマージ候補格納部303611に格納される。 The zero merge candidate derivation unit 3036134 derives a merge candidate in which the reference picture index refIdxLX is 0 and both the X component and the Y component of the vector mvLX are 0. The derived merge candidates are stored in the merge candidate storage unit 303611.
 図8は、マージ候補導出部30361が導出するマージ候補の例を示すものである。2つのマージ候補が同じ予測パラメータである場合に順番を詰める処理を除くと、マージインデックス順に、レイヤ間マージ候補、空間マージ候補(左下)、空間マージ候補(右上)、空間マージ候補(右上)、変位マージ候補、BVSPマージ候補、空間マージ候補(左下)、空間マージ候補(左上)、時間マージ候補の順になる。また、それ以降に、結合マージ候補、ゼロマージ候補があるが、図8では省略している。 FIG. 8 shows an example of merge candidates derived by the merge candidate deriving unit 30361. If two merge candidates have the same prediction parameter, excluding the processing of reducing the order, the merge index order, layer merge candidate (lower left), spatial merge candidate (upper right), spatial merge candidate (upper right), Displacement merge candidates, BVSP merge candidates, spatial merge candidates (lower left), spatial merge candidates (upper left), and temporal merge candidates. Further, there are a merge merge candidate and a zero merge candidate thereafter, which are omitted in FIG.
 マージ候補選択部30362は、マージ候補格納部303611に格納されているマージ候補のうち、インター予測パラメータ復号制御部3031から入力されたマージインデックスmerge_idxに対応するインデックスが割り当てられたマージ候補を、対象PUのインター予測パラメータとして選択する。マージ候補選択部30362は選択したマージ候補を復号ピクチャ管理部306の予測パラメータメモリ3067(図9)に記憶するとともに、予測画像生成部308(図6)に出力する。 The merge candidate selection unit 30362 selects, from the merge candidates stored in the merge candidate storage unit 303611, a merge candidate to which an index corresponding to the merge index merge_idx input from the inter prediction parameter decoding control unit 3031 is assigned. As an inter prediction parameter. The merge candidate selection unit 30362 stores the selected merge candidate in the prediction parameter memory 3067 (FIG. 9) of the decoded picture management unit 306 and outputs it to the prediction image generation unit 308 (FIG. 6).
  [[[変位ベクトル取得部について]]] 以下において、変位ベクトル取得部3036122の詳細について説明する。変位ベクトル取得部3036122は、対象PUに隣接するブロックが変位ベクトルを持つ場合には、その変位ベクトルを復号ピクチャ管理部306の予測パラメータメモリ3067から抽出し、当該予測パラメータメモリ3067を参照し、対象PUに隣接するブロックの予測フラグpredFlagLX、参照ピクチャインデックスrefIdxLXとベクトルmvLXを読み出す。変位ベクトル取得部3036122は、参照レイヤ判定機能を有している。変位ベクトル取得部3036122は、対象PUに隣接するブロックの予測パラメータを順に読み出し、参照レイヤ判定機能によって、隣接ブロックの参照ピクチャインデックスから隣接ブロックが変位ベクトルを備えるか否かを判定する。隣接ブロックが変位ベクトルを備える場合、変位ベクトル取得部3036122は、その変位ベクトルを出力する。隣接ブロックの予測パラメータに変位ベクトルが無い場合にはゼロベクトルを変位ベクトルとして出力する。 [[[About Displacement Vector Acquisition Unit]]] Details of the displacement vector acquisition unit 3036122 will be described below. When the block adjacent to the target PU has a displacement vector, the displacement vector acquisition unit 3036122 extracts the displacement vector from the prediction parameter memory 3067 of the decoded picture management unit 306, refers to the prediction parameter memory 3067, A prediction flag predFlagLX, a reference picture index refIdxLX, and a vector mvLX of a block adjacent to the PU are read out. The displacement vector acquisition unit 3036122 has a reference layer determination function. The displacement vector acquisition unit 3036122 sequentially reads prediction parameters of blocks adjacent to the target PU, and determines whether or not the adjacent block has a displacement vector from the reference picture index of the adjacent block by the reference layer determination function. When the adjacent block includes a displacement vector, the displacement vector acquisition unit 3036122 outputs the displacement vector. If there is no displacement vector in the prediction parameter of the adjacent block, the zero vector is output as the displacement vector.
 (参照レイヤ判定機能) 変位ベクトル取得部3036122が有する参照レイヤ判定機能は、入力された参照ピクチャインデックスrefIdxLXに基づいて、参照ピクチャインデックスrefIdxLXが指す参照ピクチャと、対象ピクチャの関係を示すリファレンスレイヤ情報reference_layer_infoを定めるというものである。リファレンスレイヤ情報reference_layer_infoは、参照ピクチャへのベクトルmvLXが変位ベクトルであるか動きベクトルであるかを示す情報である。 (Reference layer determination function) The reference layer determination function of the displacement vector acquisition unit 3036122 is based on the input reference picture index refIdxLX, and reference layer information reference_layer_info indicating the relationship between the reference picture indicated by the reference picture index refIdxLX and the target picture Is to determine. Reference layer information reference_layer_info is information indicating whether the vector mvLX to the reference picture is a displacement vector or a motion vector.
 対象ピクチャのレイヤと参照ピクチャのレイヤが同じレイヤである場合の予測を、同一レイヤ予測と呼び、この場合に得られるベクトルは動きベクトルである。対象ピクチャのレイヤと、参照ピクチャのレイヤが異なるレイヤである場合の予測をレイヤ間予測と呼び、この場合に得られるベクトルは変位ベクトルである。 Prediction when the target picture layer and the reference picture layer are the same layer is called the same layer prediction, and the vector obtained in this case is a motion vector. Prediction when the target picture layer and the reference picture layer are different layers is called inter-layer prediction, and the vector obtained in this case is a displacement vector.
  [[AMVP予測パラメータ導出部]] AMVP予測パラメータ導出部3032は、参照ピクチャインデックスrefIdxに基づいて復号ピクチャ管理部306の予測パラメータメモリ3067(図9)に記憶されているベクトル(動きベクトル又は変位ベクトル)をベクトル候補mvpLXとして読み出す。AMVP予測パラメータ導出部3032が読み出すベクトルは、復号対象ブロックから予め定めた範囲内にあるブロック(例えば、復号対象ブロックの左下端、左上端、右上端にそれぞれ接するブロックの全部又は一部)のそれぞれに係るベクトルである。 [[AMVP Prediction Parameter Deriving Unit]] The AMVP prediction parameter deriving unit 3032 is a vector (motion vector or displacement vector) stored in the prediction parameter memory 3067 (FIG. 9) of the decoded picture management unit 306 based on the reference picture index refIdx. ) As a vector candidate mvpLX. The vector read out by the AMVP prediction parameter derivation unit 3032 is a block within a predetermined range from the decoding target block (for example, all or a part of the blocks in contact with the lower left end, upper left upper end, and upper right end of the decoding target block, respectively). It is a vector concerning.
 AMVP予測パラメータ導出部3032は、読み出したベクトル候補のうち、インター予測パラメータ復号制御部3031から入力されたベクトルインデックスmvp_LX_idxが示すベクトル候補を予測ベクトルmvpLXとして選択する。AMVP予測パラメータ導出部3032は、選択した予測ベクトルmvpLXを加算部3035に出力する。 The AMVP prediction parameter derivation unit 3032 selects a vector candidate indicated by the vector index mvp_LX_idx input from the inter prediction parameter decoding control unit 3031 among the read vector candidates as a prediction vector mvpLX. The AMVP prediction parameter derivation unit 3032 outputs the selected prediction vector mvpLX to the addition unit 3035.
 図4の(b)は、ベクトル候補の一例を示す概念図である。図4の(b)に示す予測ベクトルリスト602は、AMVP予測パラメータ導出部3032において導出される複数のベクトル候補からなるリストである。予測ベクトルリスト602において、左右に一列に配列された5個の長方形は、それぞれ予測ベクトルを示す領域を示す。左端から2番目のmvp_LX_idxの真下の下向きの矢印とその下のmvpLXは、ベクトルインデックスmvp_LX_idxが、復号ピクチャ管理部306の予測パラメータメモリ3067においてベクトルmvpLXを参照するインデックスであることを示す。 FIG. 4B is a conceptual diagram showing an example of vector candidates. A prediction vector list 602 illustrated in FIG. 4B is a list including a plurality of vector candidates derived by the AMVP prediction parameter deriving unit 3032. In the prediction vector list 602, five rectangles arranged in a line on the left and right indicate areas indicating prediction vectors, respectively. The downward arrow directly below the second mvp_LX_idx from the left end and mvpLX below the mvp_LX_idx indicate that the vector index mvp_LX_idx is an index referring to the vector mvpLX in the prediction parameter memory 3067 of the decoded picture management unit 306.
 候補ベクトルは、復号処理が完了したブロックであって、復号対象ブロックから予め定めた範囲のブロック(例えば、隣接ブロック)を参照し、参照したブロックに係るベクトルに基づいて生成される。なお、隣接ブロックには、対象ブロックに空間的に隣接するブロック、例えば、左ブロック、上ブロックの他、対象ブロックに時間的に隣接するブロック、例えば、対象ブロックと同じ位置で、表示時刻が異なるブロックから得られたブロックを含む。 The candidate vector is a block for which the decoding process has been completed, and is generated based on a vector related to the referenced block with reference to a block (for example, an adjacent block) in a predetermined range from the decoding target block. The adjacent block has a block that is spatially adjacent to the target block, for example, the left block and the upper block, and a block that is temporally adjacent to the target block, for example, the same position as the target block, and has a different display time. Contains blocks derived from blocks.
 加算部3035は、予測ベクトル選択部3034から入力された予測ベクトルmvpLXとインター予測パラメータ復号制御部から入力された差分ベクトルmvdLXを加算してベクトルmvLXを算出する。加算部3035は、算出したベクトルmvLXを予測画像生成部308(図6)に出力する。 The addition unit 3035 adds the prediction vector mvpLX input from the prediction vector selection unit 3034 and the difference vector mvdLX input from the inter prediction parameter decoding control unit to calculate a vector mvLX. The adding unit 3035 outputs the calculated vector mvLX to the predicted image generation unit 308 (FIG. 6).
  [復号ピクチャ管理部] 図9を用いて、復号ピクチャ管理部306の詳細構成を説明する。図9は復号ピクチャ管理部306の構成を例示した機能ブロック図である。 [Decoded picture management unit] The detailed configuration of the decoded picture management unit 306 will be described with reference to FIG. FIG. 9 is a functional block diagram illustrating the configuration of the decoded picture management unit 306.
 図9に示すように、復号ピクチャ管理部306は、DPB3061、RPS導出部3062、参照ピクチャ制御部3063、参照レイヤピクチャ制御部3064、RPL導出部3065、出力制御部3066、予測パラメータメモリ3067を備える。 As shown in FIG. 9, the decoded picture management unit 306 includes a DPB 3061, an RPS derivation unit 3062, a reference picture control unit 3063, a reference layer picture control unit 3064, an RPL derivation unit 3065, an output control unit 3066, and a prediction parameter memory 3067. .
  [[DPB3061]] DPB3061は、復号ピクチャバッファ(Decoded Picture Buffer)とも呼ばれ、加算部312から出力される対象レイヤの各ピクチャの復号ピクチャを記録する。DPBには、対象レイヤの各ピクチャに対応する復号ピクチャが出力順(POC: Picture Order Count)に関連付けて記録されている。加えて、DPBの各ピクチャに対しては、参照マークおよび出力マークを設定できる。 [[DPB 3061]] The DPB 3061 is also called a decoded picture buffer (DecodedDePicture Buffer), and records the decoded picture of each picture of the target layer output from the adding unit 312. In the DPB, the decoded pictures corresponding to the pictures in the target layer are recorded in association with the output order (POC: “Picture” Order “Count”). In addition, a reference mark and an output mark can be set for each DPB picture.
 参照マークは、DPB上のピクチャが対象ピクチャ以降の復号処理における予測画像生成処理(例えば、インター予測やレイヤ間画像予測)への利用可否を示す情報である。参照マークは、具体的には、「短期参照使用」(「used for short-term reference」)、「長期参照使用」(「used for long-term reference」)、「参照不使用」(「not usedfor reference」)のいずれかの値を取る。 The reference mark is information indicating whether or not the picture on the DPB can be used for predicted image generation processing (for example, inter prediction or inter-layer image prediction) in decoding processing after the target picture. Specifically, reference marks are "short-term reference use" ("used for-short-term reference"), "long-term reference use" ("used for-long-term reference"), "reference not used" ("not usedfor One of the values of “reference”).
 なお、参照マークの取り得る値を上記とするがそれに限らない。例えば、「レイヤ間参照使用」(「used for inter-layer reference」)の値に参照マークを設定できてもよい。また、「短期参照使用」と「長期参照使用」の区別をせず、両者の和集合を「参照使用」(「used for reference」)と定義してもよい。 In addition, although the value which a reference mark can take is set as the above, it is not restricted to it. For example, a reference mark may be set to the value of “inter-layer reference use” (“used for inter-layer reference”). Further, without distinguishing between “short-term reference use” and “long-term reference use”, the union of both may be defined as “reference use” (“used for reference”).
 出力マークは、DPB上のピクチャを外部へ出力する必要性の有無を示す情報である。出力マークは、具体的には、「出力要」(「needed for output」)、「出力不要」(「not needed for output」)のいずれかの値を取る。なお、参照マークや出力マークは復号処理や符号化処理の特定のタイミングで明示的に設定されていなくてもよく、その場合、参照マークまたは出力マークは「未定義」と判定する。 The output mark is information indicating whether or not it is necessary to output the picture on the DPB to the outside. Specifically, the output mark takes one of the values “output required” (“needed for output”) and “output unnecessary” (“not needed for output”). Note that the reference mark and the output mark may not be explicitly set at a specific timing of the decoding process or the encoding process. In this case, the reference mark or the output mark is determined as “undefined”.
  [[RPS導出部3062]] RPS導出部3062は、入力されるシンタックス値に基づいて対象ピクチャの復号処理に用いるRPS(Reference Picture Set;参照ピクチャセット)を導出してベース参照ピクチャ制御部144、参照ピクチャ制御部143、およびRPL導出部3065に出力する。 [[RPS Deriving Unit 3062]] The RPS deriving unit 3062 derives an RPS (Reference Picture Set) to be used for decoding the target picture based on the input syntax value, and the base reference picture control unit 144 To the reference picture control unit 143 and the RPL deriving unit 3065.
 RPSは、概略的には、対象ピクチャの復号処理、または、復号順で対象ピクチャに後続するピクチャの復号処理において利用されうる参照ピクチャの集合を表わす。 RPS generally represents a set of reference pictures that can be used in decoding processing of a target picture or decoding processing of pictures following the target picture in decoding order.
   (サブRPSの定義) RPSは、参照ピクチャの性質に基づいて複数のサブRPSに分割できる。本実施形態では、RPSは以下の5種類のサブRPSから構成される。 (Sub RPS definition) RPS can be divided into multiple sub RPSs based on the nature of the reference picture. In the present embodiment, the RPS is composed of the following five types of sub-RPSs.
 (1)前方短期RPS:対象ピクチャを基準とする表示順の相対位置により指定される参照ピクチャであって、対象ピクチャと同一レイヤに属し、かつ、表示順が対象ピクチャより前の参照ピクチャを含むサブRPS。 (1) Forward short-term RPS: a reference picture specified by a relative position in display order with respect to the target picture, including a reference picture that belongs to the same layer as the target picture and whose display order is earlier than the target picture Sub RPS.
 (2)後方短期RPS:対象ピクチャを基準とする表示順の相対位置により指定される参照ピクチャであって、対象ピクチャと同一レイヤに属し、かつ、表示順が対象ピクチャより後の参照ピクチャを含むサブRPS。 (2) Rear short-term RPS: a reference picture specified by a relative position in the display order with respect to the target picture, which belongs to the same layer as the target picture and includes a reference picture whose display order is after the target picture Sub RPS.
 (3)長期RPS:表示順の絶対位置により指定される参照ピクチャであって、対象ピクチャと同一レイヤに属する参照ピクチャを含むサブRPS。 (3) Long-term RPS: A sub-RPS that is a reference picture specified by an absolute position in display order and includes a reference picture belonging to the same layer as the target picture.
 (4)レイヤ間画素RPS:対象ピクチャと異なるレイヤに属し、かつ、レイヤ間予測で画素値が参照される参照ピクチャを含むサブRPS。 (4) Inter-layer pixel RPS: a sub-RPS that includes a reference picture that belongs to a different layer from the target picture and whose pixel value is referred to in inter-layer prediction.
 (5)レイヤ間動き限定RPS:対象ピクチャと異なるレイヤに属し、かつ、レイヤ間予測で動き情報が参照されて画素値が参照されない参照ピクチャを含むサブRPS。 (5) Inter-layer motion limited RPS: a sub-RPS that includes a reference picture that belongs to a layer different from the target picture and that does not refer to a pixel value by referring to motion information in inter-layer prediction.
 なお、以下では、前方短期RPSと後方短期RPSの和集合を短期RPSとも称する。つまり、短期RPSは、対象ピクチャを基準とする表示順の相対位置により指定される参照ピクチャであって、対象ピクチャと同一レイヤに属するピクチャを含む。また、レイヤ間画素RPSとレイヤ間動き限定RPSの和集合をレイヤ間RPSとも称する。つまり、レイヤ間RPSは、対象ピクチャと異なるレイヤに属する参照ピクチャ(レイヤ間参照ピクチャ)を含む。 In the following, the union of forward short-term RPS and backward short-term RPS is also referred to as short-term RPS. That is, the short-term RPS is a reference picture specified by a relative position in display order with respect to the target picture, and includes a picture belonging to the same layer as the target picture. The union of inter-layer pixel RPS and inter-layer motion limited RPS is also referred to as inter-layer RPS. That is, the inter-layer RPS includes a reference picture (inter-layer reference picture) belonging to a layer different from the target picture.
   (サブRPSの導出) RPS導出部3062におけるRPS導出はサブRPS毎に実行される。以下では、前述のサブRPS毎に関連シンタックスを説明するとともに、当該シンタックスからサブRPSを導出する処理を説明する。 (Sub-RPS derivation) RPS derivation in the RPS derivation unit 3062 is executed for each sub-RPS. Hereinafter, the related syntax will be described for each sub RPS described above, and the process for deriving the sub RPS from the syntax will be described.
   (短期RPS) 短期RPS(前方短期RPSと後方短期RPS)に係るシンタックスに、SPSに含まれる短期参照ピクチャセット情報であるSPS短期RPS情報と、スライスヘッダに含まれる短期参照ピクチャセット情報であるSH短期RPS情報がある。 (Short-term RPS) The syntax for short-term RPS (forward short-term RPS and backward short-term RPS) is SPS short-term RPS information that is short-term reference picture set information included in the SPS, and short-term reference picture set information included in the slice header. There is SH short-term RPS information.
 (SPS短期RPS情報) SPS短期RPS情報は、SPSを参照する各ピクチャにおいて短期SPSとして選択され得る複数の短期RPS候補の情報を含む。なお、短期RPSは、対象ピクチャに対する相対的な位置(例えば対象ピクチャとのPOC差分)により指定される参照ピクチャ(短期参照ピクチャ)となり得るピクチャの集合である。 (SPS short-term RPS information) The SPS short-term RPS information includes information on a plurality of short-term RPS candidates that can be selected as short-term SPS in each picture referring to the SPS. Note that the short-term RPS is a set of pictures that can be a reference picture (short-term reference picture) specified by a relative position with respect to the target picture (for example, a POC difference from the target picture).
 SPS短期RPS情報を、図16を参照して説明する。図16は、SPS復号時に利用されるSPSシンタックス表の一部を例示している。図16の(A)の部分がSPS短期RPS情報に相当する。SPS短期RPS情報には、SPSに含まれる短期RPS数(num_short_term_ref_pic_sets)、および、各短期RPSの定義(short_term_ref_pic_set(i))が含まれる。 SPS short-term RPS information will be described with reference to FIG. FIG. 16 illustrates a part of the SPS syntax table used at the time of SPS decoding. The part (A) in FIG. 16 corresponds to SPS short-term RPS information. The SPS short-term RPS information includes the number of short-term RPS included in the SPS (num_short_term_ref_pic_sets) and the definition of each short-term RPS (short_term_ref_pic_set (i)).
 短期RPS情報について、図17の(a)を参照して説明する。図17の(a)は、SPS復号時、および、スライスヘッダ復号時に利用される短期RPSのシンタックス表を例示している。 The short-term RPS information will be described with reference to FIG. FIG. 17A illustrates a short-term RPS syntax table used at the time of SPS decoding and slice header decoding.
 短期RPS情報には、対象ピクチャより表示順が早い短期参照ピクチャ数(num_negative_pics)、および、対象ピクチャより表示順が遅い短期参照ピクチャ数(num_positive_pics)が含まれる。なお、以下では、対象ピクチャより表示順が早い短期参照ピクチャを前方短期参照ピクチャ、対象ピクチャより表示順が遅い短期参照ピクチャを後方短期参照ピクチャと呼ぶ。 The short-term RPS information includes the number of short-term reference pictures whose display order is earlier than the target picture (num_negative_pics) and the number of short-term reference pictures whose display order is later than the target picture (num_positive_pics). In the following, a short-term reference picture whose display order is earlier than the target picture is referred to as a front short-term reference picture, and a short-term reference picture whose display order is later than the target picture is referred to as a rear short-term reference picture.
 また、短期RPS情報には、各前方短期参照ピクチャに対して、対象ピクチャに対するPOC差分の絶対値(delta_poc_s0_minus1[i])、および、対象ピクチャの参照ピクチャとして使用される可能性の有無(used_by_curr_pic_s0_flag[i])が含まれる。加えて、各後方短期参照ピクチャに対して、対象ピクチャに対するPOC差分の絶対値(delta_poc_s1_minus1[i])、および、対象ピクチャの参照ピクチャとして使用される可能性の有無(used_by_curr_pic_s1_flag[i])が含まれる。 The short-term RPS information includes, for each forward short-term reference picture, the absolute value of the POC difference with respect to the target picture (delta_poc_s0_minus1 [i]), and the presence / absence of use as a reference picture of the target picture (used_by_curr_pic_s0_flag [ i]). In addition, for each backward short-term reference picture, the absolute value of the POC difference with respect to the target picture (delta_poc_s1_minus1 [i]) and the possibility of being used as the reference picture of the target picture (used_by_curr_pic_s1_flag [i]) are included It is.
 (SH短期RPS情報) SH短期RPS情報は、スライスヘッダを参照するピクチャから利用され得る単一の短期RPSの情報を含む。 (SH short-term RPS information) The SH short-term RPS information includes information on a single short-term RPS that can be used from a picture referring to a slice header.
 SH短期RPS情報の復号について、図17の(b)を参照して説明する。図17の(b)は、スライスヘッダ復号時に利用されるスライスヘッダシンタックス表の一部を例示している。図17の(b)の(A)の部分がSH短期RPS情報に相当する。SH短期RPS情報は、短期RPSをSPSで復号済みの短期RPS候補の中から選択するか、スライスヘッダに明示的に含めるかを示すフラグ(short_term_ref_pic_set_sps_flag)を含む。短期RPS候補から選択する場合、何れか一つの短期RPS候補を選択する識別子(short_term_ref_pic_set_idx)が含まれる。スライスヘッダに明示的に含める場合は、前述の図17の(a)を参照して説明したシンタックス表(short_term_ref_pic_set(idx))に相当する情報が、SPS短期RPS情報に含まれる。 Decoding of SH short-term RPS information will be described with reference to FIG. FIG. 17B illustrates a part of the slice header syntax table used when decoding the slice header. The part (A) in FIG. 17B corresponds to the SH short-term RPS information. The SH short-term RPS information includes a flag (short_term_ref_pic_set_sps_flag) indicating whether the short-term RPS is selected from among the short-term RPS candidates decoded by the SPS or explicitly included in the slice header. When selecting from short-term RPS candidates, an identifier (short_term_ref_pic_set_idx) for selecting any one short-term RPS candidate is included. When explicitly included in the slice header, information corresponding to the syntax table (short_term_ref_pic_set (idx)) described with reference to FIG. 17A is included in the SPS short-term RPS information.
   (短期RPS導出処理) 短期RPS情報から短期RPS、すなわち、前方短期RPSと後方短期RPSが導出される。また、後続参照短期RPSも導出される。
・前方短期RPS:SPS短期RPS情報またはSH短期RPS情報により指定される現ピクチャ参照可能ピクチャであって、表示順が対象ピクチャより早いピクチャを含む。
・後方短期RPS:SPS短期RPS情報またはSH短期RPS情報により指定される現ピクチャ参照可能ピクチャであって、表示順が対象ピクチャより遅いピクチャを含む。
・後続参照短期RPS:現ピクチャでは参照されないが、復号順で現ピクチャに後続するピクチャにおいて参照されえるピクチャを含む。
(Short-term RPS derivation processing) Short-term RPS, that is, forward short-term RPS and backward short-term RPS are derived from the short-term RPS information. A subsequent reference short-term RPS is also derived.
Forward short RPS: A picture that can be referred to the current picture specified by the SPS short-term RPS information or the SH short-term RPS information, and includes a picture whose display order is earlier than the target picture.
Backward short-term RPS: A picture that can be referred to the current picture specified by SPS short-term RPS information or SH short-term RPS information, and includes a picture whose display order is later than the target picture.
Subsequent reference short-term RPS: includes pictures that are not referenced in the current picture but can be referenced in pictures that follow the current picture in decoding order.
 前方短期RPS(ListStCurrBefore)、後方短期RPS(ListStCurrAfter)、後続参照短期RPS(ListStFoll)を以下の手順で導出する。なお、前方短期RPS、後方短期RPS、および、後続参照短期RPSは、以下の処理の開始前に空に設定されている。
(S101)SPS短期RPS情報、および、SH短期RPS情報に基づいて、対象ピクチャの復号に用いる短期RPS情報を特定する。具体的には、SH短期RPS情報に含まれるshort_term_ref_pic_set_spsの値が0である場合、SH短期RPS情報に含まれるスライスヘッダで明示的に伝送された短期RPSを選択する。それ以外(short_term_ref_pic_set_spsの値が1の場合、SH短期RPS情報に含まれるshort_term_ref_pic_set_idxが示す短期RPSを、SPS短期RPS情報に含まれる複数の短期RPS候補から選択する。
(S102)選択された短期RPSに含まれる参照ピクチャ各々のPOCを導出する。参照ピクチャのPOCは、参照ピクチャが前方短期RPSに属する場合、対象ピクチャのPOCから「delta_poc_s0_minus1[i]+1」の値を減算して導出する。一方、参照ピクチャが後方短期RPSに属する場合、対象ピクチャのPOCに「delta_poc_s1_minus1[i]+1」の値を加算して導出する。
(S103)短期RPSに含まれる前方参照ピクチャを伝送された順に確認し、関連付けられているused_by_curr_pic_s0_flag[i]の値が1である場合、当該前方参照ピクチャをListStCurrBeforeに追加する。それ以外(used_by_curr_pic_s0_flag[i]の値が0)の場合、当該前方参照ピクチャをListStFollに追加する。
(S104)短期RPSに含まれる後方参照ピクチャを伝送された順に確認し、関連付けられているused_by_curr_pic_s1_flag[i]の値が1である場合、当該後方参照ピクチャをListStCurrAfterに追加する。それ以外(used_by_curr_pic_s1_flag[i]の値が0)の場合、当該後方参照ピクチャをListStFollに追加する。
A forward short-term RPS (ListStCurrBefore), a backward short-term RPS (ListStCurrAfter), and a subsequent reference short-term RPS (ListStFoll) are derived by the following procedure. Note that the front short-term RPS, the rear short-term RPS, and the subsequent reference short-term RPS are set to be empty before the start of the following processing.
(S101) Based on the SPS short-term RPS information and the SH short-term RPS information, short-term RPS information used for decoding the target picture is specified. Specifically, when the value of short_term_ref_pic_set_sps included in the SH short-term RPS information is 0, the short-term RPS explicitly transmitted in the slice header included in the SH short-term RPS information is selected. Other than that (when the value of short_term_ref_pic_set_sps is 1, the short-term RPS indicated by short_term_ref_pic_set_idx included in the SH short-term RPS information is selected from a plurality of short-term RPS candidates included in the SPS short-term RPS information.
(S102) The POC of each reference picture included in the selected short-term RPS is derived. When the reference picture belongs to the forward short-term RPS, the reference picture POC is derived by subtracting the value of “delta_poc_s0_minus1 [i] +1” from the POC of the target picture. On the other hand, when the reference picture belongs to the backward short-term RPS, it is derived by adding the value of “delta_poc_s1_minus1 [i] +1” to the POC of the target picture.
(S103) The forward reference pictures included in the short-term RPS are confirmed in the order of transmission, and when the associated used_by_curr_pic_s0_flag [i] value is 1, the forward reference picture is added to ListStCurrBefore. Otherwise (used_by_curr_pic_s0_flag [i] has a value of 0), the forward reference picture is added to ListStFoll.
(S104) The backward reference pictures included in the short-term RPS are confirmed in the order of transmission. When the value of used_by_curr_pic_s1_flag [i] associated with the backward reference picture is 1, the backward reference picture is added to ListStCurrAfter. In other cases (used_by_curr_pic_s1_flag [i] has a value of 0), the backward reference picture is added to ListStFoll.
   (長期RPS) 長期RPSに係るシンタックスに、SPSに含まれる長期参照ピクチャ情報であるSPS長期RPS情報と、スライスヘッダに含まれる長期参照ピクチャ情報であるSH長期RPS情報がある。 (Long-term RPS) The syntax related to long-term RPS includes SPS long-term RPS information that is long-term reference picture information included in the SPS and SH long-term RPS information that is long-term reference picture information included in the slice header.
 (SPS長期RPS情報) SPS長期RPS情報は、SPSを参照する各ピクチャから利用され得る複数の長期参照ピクチャの情報を含む。なお、長期参照ピクチャとは、シーケンス内の絶対的な位置(例えばPOC)により指定される参照ピクチャである。 (SPS long-term RPS information) The SPS long-term RPS information includes information on a plurality of long-term reference pictures that can be used from each picture referring to the SPS. Note that the long-term reference picture is a reference picture specified by an absolute position (for example, POC) in the sequence.
 SPS長期RPS情報の復号について、図16を再び参照して説明する。図16の(B)の部分がSPS長期RPS情報に相当する。SPS長期RPS情報には、SPSでの長期参照ピクチャ伝送有無を示すフラグ(long_term_ref_pics_present_flag)、SPSで伝送される長期参照ピクチャ数(num_long_term_ref_pics_sps)、および、各長期参照ピクチャの情報が含まれる。長期参照ピクチャの情報には、参照ピクチャのPOC(lt_ref_pic_poc_lsb_sps[i])、および、長期参照ピクチャが対象ピクチャにおいて参照される可能性の有無(used_by_curr_pic_lt_sps_flag[i])が含まれる。 Decoding of SPS long-term RPS information will be described with reference to FIG. 16 again. The part (B) in FIG. 16 corresponds to the SPS long-term RPS information. The SPS long-term RPS information includes a flag (long_term_ref_pics_present_flag) indicating whether or not long-term reference pictures are transmitted in SPS, the number of long-term reference pictures (num_long_term_ref_pics_sps) transmitted in SPS, and information on each long-term reference picture. The long-term reference picture information includes the POC of the reference picture (lt_ref_pic_poc_lsb_sps [i]) and the presence / absence of the possibility that the long-term reference picture is referenced in the target picture (used_by_curr_pic_lt_sps_flag [i]).
 なお、上記参照ピクチャのPOCは、参照ピクチャに関連付けられたPOCの値自体であってもよいし、POCのLSB(Least Significant Bit)、すなわち、POCを既定の2の冪乗の数で割った余りの値を用いてもよい。 The POC of the reference picture may be the POC value associated with the reference picture, or the LSB of the POC (Least Significant Bit), that is, the POC divided by a predetermined power of two. The remainder value may be used.
 (SH長期RPS情報) SH長期RPS情報は、スライスヘッダを参照するピクチャから利用され得る長期参照ピクチャの情報を含む。 (SH long-term RPS information) The SH long-term RPS information includes information on a long-term reference picture that can be used from a picture that references a slice header.
 SH長期RPS情報の復号について、図17の(b)を再び参照して説明する。図17の(b)の(B)の部分がSH長期RPS情報に相当する。SH長期RPS情報は、長期参照ピクチャを利用することがフラグ(long_term_ref_pic_present_flag)で示されている場合にスライスヘッダに含まれる。SPS長期RPS情報が1以上の長期参照ピクチャを含む場合(num_long_term_ref_pics_sps>0)、SPSで復号済の長期参照ピクチャの中で対象ピクチャで参照され得る参照ピクチャの数(num_long_term_sps)がSH長期RPS情報に含まれる。また、スライスヘッダで明示的に伝送される長期参照ピクチャ数(num_long_term_pics)がSH長期RPS情報に含まれる。加えて、上記num_long_term_spsの数の長期参照ピクチャをSPS長期RPS情報に含まれる長期参照ピクチャから選択する情報(lt_idx_sps[i])がSH長期RPS情報に含まれる。さらに、スライスヘッダに明示的に含める長期参照ピクチャの情報として、上記num_long_term_picsの数だけ、参照ピクチャのPOC(poc_lsb_lt [i])、および、対象ピクチャの参照ピクチャとして使用される可能性の有無(used_by_curr_pic_lt_flag[i])が含まれる。 Decoding of SH long-term RPS information will be described with reference again to FIG. The part (B) of FIG. 17B corresponds to the SH long-term RPS information. The SH long-term RPS information is included in the slice header when use of a long-term reference picture is indicated by a flag (long_term_ref_pic_present_flag). When the SPS long-term RPS information includes one or more long-term reference pictures (num_long_term_ref_pics_sps> 0), the number of reference pictures (num_long_term_sps) that can be referred to by the target picture among the long-term reference pictures decoded by SPS is the SH long-term RPS information. included. In addition, the number of long-term reference pictures (num_long_term_pics) explicitly transmitted in the slice header is included in the SH long-term RPS information. In addition, information (lt_idx_sps [i]) for selecting the num_long_term_sps number of long-term reference pictures from the long-term reference pictures included in the SPS long-term RPS information is included in the SH long-term RPS information. Furthermore, as information on long-term reference pictures to be explicitly included in the slice header, the number of reference pictures POC (poc_lsb_lt [i]) and the presence / absence of use as a reference picture of the target picture (used_by_curr_pic_lt_flag) [i]) is included.
   (長期RPS導出処理) 長期RPS情報から長期RPSが導出される。また、後続参照長期RPSも導出される。
・長期RPS:SPS長期RPS情報またはSH長期RPS情報により指定される現ピクチャ参照可能ピクチャを含む。
・後続参照長期RPS:現ピクチャでは参照されないが、復号順で現ピクチャに後続するピクチャにおいて参照されえる参照ピクチャを含む。
(Long-term RPS derivation process) A long-term RPS is derived from long-term RPS information. A subsequent reference long-term RPS is also derived.
Long-term RPS: Contains a picture that can be referred to the current picture specified by SPS long-term RPS information or SH long-term RPS information.
Subsequent reference long-term RPS: includes a reference picture that is not referenced in the current picture but can be referenced in a picture that follows the current picture in decoding order.
 長期RPS(ListLtCurr)、後続参照長期RPS(ListLtFoll)を以下の手順で導出する。なお、長期RPS、および、後続参照長期RPSは、以下の処理の開始前に空に設定されている。(S201) SPS長期RPS情報、および、SH長期RP情報に基づいて、対象ピクチャの復号に用いる長期参照ピクチャを特定する。具体的には、num_long_term_spsの数の参照ピクチャをSPS長期RPS情報に含まれる参照ピクチャの中から選択して、長期RPSに追加すする。選択される参照ピクチャは、lt_idx_sps[i]の示す参照ピクチャである。
(S202)続いて、num_long_term_picsの数のSH長期RPS情報に含まれる参照ピクチャを順に長期RPSに追加する。
(S203)長期RPSに含まれる参照ピクチャ各々のPOCを導出する。長期参照ピクチャのPOCは、関連付けて復号されたpoc_lst_lt[i]、または、lt_ref_pic_poc_lsb_sps[i]の値から直接導出される。
(S204)長期RPSに含まれる参照ピクチャを順に確認し、関連付けられているused_by_curr_pic_lt_flag[i]、または、used_by_curr_pic_lt_sps_flag[i]の値が1である場合、当該長期参照ピクチャをListLtCurrに追加する。それ以外(used_by_curr_pic_lt_flag[i]、または、used_by_curr_pic_lt_sps_flag[i]の値が0)の場合、当該長期参照ピクチャを後続参照長期RPS(ListLtFoll)に追加する。
The long-term RPS (ListLtCurr) and subsequent reference long-term RPS (ListLtFoll) are derived by the following procedure. Note that the long-term RPS and the subsequent reference long-term RPS are set to be empty before starting the following processing. (S201) Based on the SPS long-term RPS information and the SH long-term RP information, the long-term reference picture used for decoding the current picture is specified. Specifically, num_long_term_sps reference pictures are selected from the reference pictures included in the SPS long-term RPS information and added to the long-term RPS. The selected reference picture is the reference picture indicated by lt_idx_sps [i].
(S202) Subsequently, reference pictures included in the SH long-term RPS information of the number num_long_term_pics are sequentially added to the long-term RPS.
(S203) The POC of each reference picture included in the long-term RPS is derived. The POC of the long-term reference picture is directly derived from the value of poc_lst_lt [i] or lt_ref_pic_poc_lsb_sps [i] decoded in association with each other.
(S204) The reference pictures included in the long-term RPS are checked in order, and when the value of associated used_by_curr_pic_lt_flag [i] or used_by_curr_pic_lt_sps_flag [i] is 1, the long-term reference picture is added to ListLtCurr. In other cases (used_by_curr_pic_lt_flag [i] or used_by_curr_pic_lt_sps_flag [i] is 0), the long-term reference picture is added to the subsequent reference long-term RPS (ListLtFoll).
   (レイヤ間RPS) レイヤ間画素RPSとレイヤ間動き限定RPSに係るシンタックスに、IL-RPS情報(レイヤ間RPS情報、レイヤ間参照ピクチャセット情報)がある。 (Inter-layer RPS) The syntax related to inter-layer pixel RPS and inter-layer motion limited RPS includes IL-RPS information (inter-layer RPS information, inter-layer reference picture set information).
 (IL-RPS情報) IL-RPS情報は、スライスヘッダを含むピクチャからレイヤ間予測で参照され得るレイヤ間参照ピクチャの情報を含む。 (IL-RPS information) The IL-RPS information includes information on an inter-layer reference picture that can be referred to in inter-layer prediction from a picture including a slice header.
 IL-RPS情報について、図18の(a)と図18の(b)を参照して説明する。 The IL-RPS information will be described with reference to FIGS. 18A and 18B.
 図18の(a)はVPSに含まれるVPS拡張(vps_extension)の復号時に参照されるシンタックス表の一部であって、IL-RPS情報に相当する部分である。図18の(a)に示すように、VPSには、IL-RPS情報に含まれるシンタックスであるmax_one_active_ref_layer_flag、direct_dep_type_len_minus2、direct_dependency_type[i][j]が含まれる。 (A) in FIG. 18 is a part of a syntax table that is referred to when the VPS extension (vps_extension) included in the VPS is decoded, and corresponds to the IL-RPS information. As illustrated in FIG. 18A, the VPS includes max_one_active_ref_layer_flag, direct_dep_type_len_minus2, and direct_dependency_type [i] [j], which are syntaxes included in the IL-RPS information.
 シンタックスmax_one_active_ref_layer_flagは、任意レイヤの任意ピクチャの復号時に参照されるレイヤの最大値が1以下であるかを示すフラグである。最大値が1以下の場合に当該フラグの値に1、それ以外(最大値が2以上)の場合に当該フラグの値に0が設定される。 The syntax max_one_active_ref_layer_flag is a flag indicating whether the maximum value of the layer referred to when decoding an arbitrary picture of an arbitrary layer is 1 or less. When the maximum value is 1 or less, the flag value is set to 1. When the maximum value is other than (the maximum value is 2 or more), the flag value is set to 0.
 シンタックスdirect_dep_type_len_minus2は、シンタックスdirect_dependency_type[i][j]のビット数を表わす値である。direct_dependency_type[i][j]のビット数は(direct_dep_type_len_minus2+2)となる。 The syntax direct_dep_type_len_minus2 is a value representing the number of bits of the syntax direct_dependency_type [i] [j]. The number of bits of direct_dependency_type [i] [j] is (direct_dep_type_len_minus2 + 2).
 シンタックスdirect_dependency_type[i][j]は、“i”で示されるレイヤから“j”で示されるレイヤを参照する際に使用可能なレイヤ間予測の種類を示す値である。以降の説明では、direct_dependency_type[i][j]のことを、対象レイヤ(レイヤi)から参照レイヤ(レイヤj)参照時の依存タイプ、とも称する。また、レイヤiとレイヤjを省略して、依存タイプ(direct_dependency_type)とも称する。 The syntax direct_dependency_type [i] [j] is a value indicating the type of inter-layer prediction that can be used when referring to the layer indicated by “j” from the layer indicated by “i”. In the following description, direct_dependency_type [i] [j] is also referred to as a dependency type when referring to a reference layer (layer j) from a target layer (layer i). Also, layer i and layer j are omitted, and are also referred to as dependency types (direct_dependency_type).
 レイヤ間予測の種類にレイヤ間画像予測とレイヤ間動き予測がある場合の依存タイプと各レイヤ間予測の使用可否の関係を図19の(a)に示す。依存タイプが「0」の場合、画素依存であり、かつ、動き依存である。依存タイプが「1」の場合、画素依存であり、かつ、動き依存ではない。依存タイプが「2」の場合、動き依存であり、かつ、画素依存ではない。 FIG. 19A shows the relationship between the dependency type and inter-layer prediction availability when there are inter-layer image prediction and inter-layer motion prediction as the types of inter-layer prediction. When the dependency type is “0”, it is pixel-dependent and motion-dependent. When the dependency type is “1”, it is pixel-dependent and not motion-dependent. When the dependency type is “2”, it is motion-dependent and not pixel-dependent.
 ここで、依存タイプが画素依存を示す場合には、対象レイヤiは、参照レイヤjの画素を予測に用いることができる。例えば、レイヤ間画像予測に用いることができる。依存タイプが動き依存を示す場合には、対象レイヤiは、参照レイヤjの動き情報(動きベクトルおよび参照ピクチャインデックス)を予測に用いることができる。例えば、レイヤ間動き予測に用いることができる。レイヤ間画像予測は、参照レイヤの復号画像の画素値を用いて対象ピクチャの予測画像を生成する処理である。また、レイヤ間動き予測は、参照レイヤの動き情報(動きベクトル、参照ピクチャインデックス、インター予測タイプ)を直接または間接的に用いて対象ピクチャの予測画像を生成する処理である。 Here, when the dependency type indicates pixel dependency, the target layer i can use the pixel of the reference layer j for prediction. For example, it can be used for inter-layer image prediction. When the dependency type indicates motion dependency, the target layer i can use the motion information (motion vector and reference picture index) of the reference layer j for prediction. For example, it can be used for inter-layer motion prediction. Inter-layer image prediction is a process of generating a predicted image of a target picture using pixel values of a decoded image of a reference layer. In addition, inter-layer motion prediction is processing for generating a predicted image of a target picture by directly or indirectly using reference layer motion information (motion vector, reference picture index, inter prediction type).
 したがって、依存タイプが「0」の場合は、参照レイヤの復号画素(復号画像の画素値)と動き情報のいずれも対象ピクチャの復号処理で利用される可能性があることを意味する。依存タイプが「1」の場合は、参照レイヤの復号画素は利用される可能性があるが、動き情報は利用される可能性がないことを意味する。依存タイプが「2」の場合は、参照レイヤの動き情報は利用される可能性があるが、復号画素は利用される可能性がないことを意味する。 Therefore, when the dependency type is “0”, it means that both the decoded pixel of the reference layer (the pixel value of the decoded image) and the motion information may be used in the decoding process of the target picture. When the dependency type is “1”, it means that the decoded pixel of the reference layer may be used, but the motion information may not be used. When the dependency type is “2”, it means that the motion information of the reference layer may be used, but the decoded pixel may not be used.
 依存タイプが参照レイヤの復号画素を参照することを示す場合(上記の定義でいえば、依存タイプが「0」または「1」)の場合、依存タイプが画素依存を示すと表現できる。一方、依存タイプが参照レイヤの復号画素を参照しないことを示す場合(上記の定義でいえば、依存タイプが「2」)の場合、依存タイプが画素非依存を示すと表現できる。 When the dependency type indicates that the decoded pixel of the reference layer is referenced (in the above definition, the dependency type is “0” or “1”), it can be expressed that the dependency type indicates pixel dependency. On the other hand, when the dependency type indicates that the decoded pixel of the reference layer is not referred to (in the above definition, the dependency type is “2”), it can be expressed that the dependency type indicates pixel independence.
 同様に、依存タイプが参照レイヤの動き情報を参照することを示す場合(上記の定義でいえば、依存タイプが「0」または「2」)の場合、依存タイプが動き依存を示すと表現できる。一方、依存タイプが参照レイヤの復号画素を参照しないことを示す場合(上記の定義でいえば、依存タイプが「2」)の場合、依存タイプが動き非依存を示すと表現できる。 Similarly, when the dependency type indicates that the motion information of the reference layer is referred to (in the above definition, the dependency type is “0” or “2”), it can be expressed that the dependency type indicates motion dependency. . On the other hand, when the dependency type indicates that the decoded pixel of the reference layer is not referred to (in the above definition, the dependency type is “2”), it can be expressed that the dependency type indicates motion independence.
 例えば、対象レイヤiが参照レイヤjの復号画素を用いるか否か(画素依存)を示す画素依存フラグSampleEnableFlag[i][j]、対象レイヤiが参照レイヤjの動き情報を用いるか否か(動き依存)を示す動き依存フラグMotionEnableFlag[i][j]は、以下の式において導出することができる。 For example, a pixel dependency flag SampleEnableFlag [i] [j] indicating whether or not the target layer i uses a decoded pixel of the reference layer j (pixel dependency), and whether or not the target layer i uses motion information of the reference layer j ( The motion dependency flag MotionEnableFlag [i] [j] indicating motion dependency can be derived from the following equation.
 SampleEnableFlag[i][j] = (3 - direct_dependency_type[i][j]) & 2) >> 1
 MotionEnableFlag[i][j] = (3 - direct_dependency_type[i][j]) & 1)
 なお、依存タイプの値の意味は必ずしも上記に限らない。例えば、依存タイプが「0」の場合に画素依存かつ動き非依存を示し、依存タイプが「1」の場合に動き依存かつ画素非依存を示し、依存タイプが「2」の場合に画素依存かつ動き依存を示していてもよい。
SampleEnableFlag [i] [j] = (3-direct_dependency_type [i] [j]) & 2) >> 1
MotionEnableFlag [i] [j] = (3-direct_dependency_type [i] [j]) & 1)
The meaning of the dependency type value is not necessarily limited to the above. For example, when the dependency type is “0”, pixel dependency and motion independence are indicated, when the dependency type is “1”, motion dependency and pixel independence are indicated, and when the dependency type is “2”, pixel dependency and It may indicate movement dependence.
 この例では、上記、画素依存フラグSampleEnableFlag[i][j]、動き依存フラグMotionEnableFlag[i][j]は以下の式から導出される。 In this example, the pixel dependency flag SampleEnableFlag [i] [j] and the motion dependency flag MotionEnableFlag [i] [j] are derived from the following equations.
 SampleEnableFlag[i][j] = (direct_dependency_type[i][j] + 1) & 1)
 MotionEnableFlag[i][j] = (direct_dependency_type[i][j] + 1) & 2) >> 1
 また、対象レイヤが、参照レイヤの復号画素および動き情報以外を参照することを示すために、依存タイプによって、画素依存および動き依存以外の依存の種類を示しても良い。例えば、依存の種類には、ブロック分割情報、変換係数情報(変換係数の有無など)、ループフィルタ情報などが可能である。この場合も、画素依存フラグ、動き依存フラグは、上述の式などにより、導出することが可能であり、追加の依存があるか否かを示すフラグ(例えば、XXX依存フラグXXXEnableFlag、XXXは、ブロック分割情報、変換係数情報、ループフィルタ情報など)は以下の式により導出することができる。
SampleEnableFlag [i] [j] = (direct_dependency_type [i] [j] + 1) & 1)
MotionEnableFlag [i] [j] = (direct_dependency_type [i] [j] + 1) & 2) >> 1
In addition, in order to indicate that the target layer refers to other than the decoded pixel and motion information of the reference layer, a dependency type other than pixel dependency and motion dependency may be indicated depending on the dependency type. For example, the type of dependence can be block division information, transform coefficient information (such as the presence or absence of transform coefficients), loop filter information, and the like. Also in this case, the pixel dependence flag and the motion dependence flag can be derived by the above-described formulas and the like, and flags indicating whether or not there is an additional dependence (for example, XXX dependence flag XXXEnableFlag, XXX is a block Division information, transform coefficient information, loop filter information, etc.) can be derived from the following equations.
 XXXEnableFlag[i][j] = (direct_dependency_type[i][j] + 1) & 4) >> 2
 図18の(b)はスライスヘッダ復号時に参照されるシンタックス表の一部であって、IL-RPS情報に相当する部分である。
XXXEnableFlag [i] [j] = (direct_dependency_type [i] [j] + 1) & 4) >> 2
(B) of FIG. 18 is a part of the syntax table referred to when the slice header is decoded, and corresponds to the IL-RPS information.
 IL-RPS情報は、レイヤ間予測有効フラグ(inter_layer_pred_enabled_flag)を含む。さらに、レイヤ間予測有効フラグが1であり(レイヤ間予測が有効であり)、かつ、対象ピクチャから参照可能な参照レイヤ数(NumDirectRefLayers[nuh_layer_id])が1より大きい場合、レイヤ間参照ピクチャ数を表すシンタックス(num_inter_layr_ref_pics_minus1)がIL-RPS情報に含まれる。アクティブレイヤ間参照ピクチャ数(NumActiveRefLayerPics)は“num_inter_layer_ref_pics_minus1 + 1”の値に設定される。アクティブレイヤ間参照ピクチャ数は、対象ピクチャにおいてレイヤ間予測により参照可能なレイヤ間参照ピクチャ数に相当する。加えて、各レイヤ間参照ピクチャの属するレイヤを示すレイヤ識別子(inter_layer_pred_layer_idc[i])がIL-RPS情報に含まれる。 The IL-RPS information includes an inter-layer prediction enabled flag (inter_layer_pred_enabled_flag). Furthermore, when the inter-layer prediction valid flag is 1 (inter-layer prediction is valid) and the number of reference layers that can be referred from the target picture (NumDirectRefLayers [nuh_layer_id]) is greater than 1, the number of inter-layer reference pictures is set. The expressing syntax (num_inter_layr_ref_pics_minus1) is included in the IL-RPS information. The number of reference pictures between active layers (NumActiveRefLayerPics) is set to a value of “num_inter_layer_ref_pics_minus1 + 1”. The number of reference pictures between active layers corresponds to the number of reference pictures between layers that can be referred to by inter-layer prediction in the target picture. In addition, a layer identifier (inter_layer_pred_layer_idc [i]) indicating a layer to which each inter-layer reference picture belongs is included in the IL-RPS information.
 なお、IL-RPS情報に含まれる上記の各シンタックスは、自明である場合は省略されてもよい。例えば、1つのピクチャから参照可能なレイヤ間参照ピクチャ数が1枚に制限されている場合、レイヤ間参照ピクチャ数に係るシンタックスは不要である。 Note that each syntax included in the IL-RPS information may be omitted if it is obvious. For example, when the number of inter-layer reference pictures that can be referenced from one picture is limited to one, the syntax related to the number of inter-layer reference pictures is not necessary.
   (レイヤ間RPS導出処理) IL-RPS情報からレイヤ間RPS、すなわちレイヤ間画素RPSとレイヤ間動き限定RPSが導出される。 (Inter-layer RPS derivation process) Inter-layer RPS, that is, inter-layer pixel RPS and inter-layer motion limited RPS are derived from IL-RPS information.
 導出処理の説明に先立って、本実施形態のRPS導出部3062により導出されるレイヤ間画素RPSとレイヤ間動き限定RPSと依存タイプの関係について図19の(b)を用いて説明する。図19の(b)はレイヤ間RPSに含まれるサブRPS(レイヤ間画素RPSとレイヤ間動き限定RPS)と依存タイプの関係を例示する図である。 Prior to the description of the derivation process, the relationship between the inter-layer pixel RPS, the inter-layer motion limited RPS, and the dependency type, which are derived by the RPS deriving unit 3062 according to the present embodiment, will be described with reference to FIG. FIG. 19B is a diagram illustrating the relationship between the sub RPS (inter-layer pixel RPS and inter-layer motion limited RPS) included in the inter-layer RPS and the dependency type.
 図19の(b)に示すように、レイヤ間RPSはレイヤ間画素RPSとレイヤ間動き限定RPSの2つのサブRPSを含む。以下、レイヤ識別子がx、依存タイプがyであるレイヤ間参照ピクチャを「LID=x,DT=y」と表記する。図19の(b)の例では、レイヤ間RPSは3枚のレイヤ間参照ピクチャ(「LID=0,DT=0」、「LID=1,DT=1」、「LID=3,DT=0」)を含む。レイヤ間動き限定RPSは2枚のレイヤ間参照ピクチャ(「LID=2,DT=2」、「LID=4,DT=2」)を含む。 As shown in FIG. 19B, the inter-layer RPS includes two sub-RPSs, an inter-layer pixel RPS and an inter-layer motion limited RPS. Hereinafter, an inter-layer reference picture whose layer identifier is x and whose dependency type is y is referred to as “LID = x, DT = y”. In the example of FIG. 19B, the inter-layer RPS includes three inter-layer reference pictures (“LID = 0, DT = 0”, “LID = 1, DT = 1”, “LID = 3, DT = 0” ")including. The inter-layer motion limited RPS includes two inter-layer reference pictures (“LID = 2, DT = 2”, “LID = 4, DT = 2”).
 すなわち、レイヤ間画素RPSは依存タイプが「0」の参照ピクチャと依存タイプが「1」のレイヤ間参照ピクチャを含む。一方、レイヤ間動き限定RPSは依存タイプが「2」のレイヤ間参照ピクチャを含む。換言すると、レイヤ間画素RPSは、復号画素が参照される可能性のあるレイヤ間参照ピクチャを含む。一方、レイヤ間動き限定RPSは、復号画素が参照される可能性のあるレイヤ間参照ピクチャを含まず、動き情報が参照される可能性のあるレイヤ間参照ピクチャを含む。 That is, the inter-layer pixel RPS includes a reference picture whose dependency type is “0” and an inter-layer reference picture whose dependency type is “1”. On the other hand, the inter-layer motion limited RPS includes an inter-layer reference picture whose dependency type is “2”. In other words, the inter-layer pixel RPS includes an inter-layer reference picture that can refer to a decoded pixel. On the other hand, the inter-layer motion limited RPS does not include an inter-layer reference picture that may refer to a decoded pixel, but includes an inter-layer reference picture that may refer to motion information.
 レイヤ間画素RPS(IL-RPS0)、レイヤ間動き限定RPS(IL-RPS1)の導出手順を図20を用いて説明する。図20は、レイヤ間RPSに含まれるサブRPS(レイヤ間画素RPSとレイヤ間動き限定RPS)の導出処理を表わすフロー図である。 The procedure for deriving the inter-layer pixel RPS (IL-RPS0) and the inter-layer motion limited RPS (IL-RPS1) will be described with reference to FIG. FIG. 20 is a flowchart showing a derivation process of sub RPS (inter-layer pixel RPS and inter-layer motion limited RPS) included in the inter-layer RPS.
 (S301)レイヤ間画素RPSを表わすリストIL-RPS0、および、レイヤ間動き限定RPSを表わすリストIL-RPS1をそれぞれ空に設定する。 (S301) The list IL-RPS0 representing the inter-layer pixel RPS and the list IL-RPS1 representing the inter-layer motion limited RPS are respectively set to be empty.
 (S302)変数iを「0」に設定する。S303に進む。 (S302) The variable i is set to “0”. The process proceeds to S303.
 (S303)i番目のアクティブレイヤ間参照ピクチャの依存タイプが「0」または「1」の場合(YESの場合)、S304に進む。それ以外の場合(NOの場合)、S305に進む。 (S303) If the dependency type of the i-th active layer reference picture is “0” or “1” (YES), the process proceeds to S304. In other cases (NO), the process proceeds to S305.
 (S304)i番目のアクティブレイヤ間参照ピクチャをIL-RPS0(レイヤ間画素RPS)の末尾に追加して、S307に進む。 (S304) The i-th active layer reference picture is added to the end of IL-RPS0 (interlayer pixel RPS), and the process proceeds to S307.
 (S305)i番目のアクティブレイヤ間参照ピクチャの依存タイプが「2」の場合(YESの場合)、S306に進む。それ以外の場合(NOの場合)、S307に進む。 (S305) If the dependency type of the i-th active layer reference picture is “2” (YES), the process proceeds to S306. In other cases (NO), the process proceeds to S307.
 (S306)i番目のアクティブレイヤ間参照ピクチャをIL-RPS1(レイヤ間動き限定RPS)の末尾に追加して、S307に進む。 (S306) The i-th active layer reference picture is added to the end of IL-RPS1 (inter-layer motion limited RPS), and the process proceeds to S307.
 (S307)iの値がアクティブレイヤ間参照ピクチャ数(NumActiveRefLayerPics)より小さい場合(YESの場合)、S308に進む。それ以外の場合(NOの場合)、処理を終了する。 (S307) If the value of i is smaller than the number of reference pictures between active layers (NumActiveRefLayerPics) (YES), the process proceeds to S308. In other cases (in the case of NO), the process is terminated.
 (S308)変数iの値を1加算して、S303に進む。 (S308) Add 1 to the value of the variable i, and proceed to S303.
 以上の処理により、図19の(b)を用いて説明した性質を持つレイヤ間画素RPSとレイヤ間動き限定RPSをIL-RPS情報に基づいて導出できる。 Through the above processing, the inter-layer pixel RPS and inter-layer motion limited RPS having the properties described with reference to FIG. 19B can be derived based on the IL-RPS information.
 なお、S303とS305の判定は、より一般的には次のように表現できる。 Note that the determinations in S303 and S305 can be expressed more generally as follows.
 (S303r1)i番目のアクティブレイヤ間参照ピクチャの依存タイプが、レイヤ間参照ピクチャの復号画素が参照される可能性のあることを示す場合(i番目のアクティブレイヤ間参照ピクチャに対応する参照ピクチャのSampleEnableFlagが1の場合)、S304に進む。それ以外の場合、S305に進む。 (S303r1) When the dependency type of the i-th active layer reference picture indicates that the decoded pixel of the inter-layer reference picture may be referred to (the reference picture corresponding to the i-th active layer reference picture) When SampleEnableFlag is 1, the process proceeds to S304. Otherwise, the process proceeds to S305.
 (S305r1)i番目のアクティブレイヤ間参照ピクチャの依存タイプが、レイヤ間参照ピクチャの動き情報が参照される可能性のあることを示す場合(i番目のアクティブレイヤ間参照ピクチャに対応する参照ピクチャのSampleEnableFlagが0かつMotionEnableFlagが1の場合)、S306に進む。それ以外の場合、S307に進む。 (S305r1) When the dependency type of the i-th active layer reference picture indicates that the motion information of the inter-layer reference picture may be referred to (the reference picture corresponding to the i-th active layer reference picture) When SampleEnableFlag is 0 and MotionEnableFlag is 1, the process proceeds to S306. Otherwise, the process proceeds to S307.
 上記の図20を用いて説明したレイヤ間画素RPSとレイヤ間動き限定RPSの導出方法はあくまで一例である。レイヤ間画素RPSとレイヤ間動き限定RPSが特定の性質、すなわち、レイヤ間画素RPSは復号画素が参照される可能性のあるレイヤ間参照ピクチャを含み、また、レイヤ間動き限定RPSは復号画素が参照される可能性のあるレイヤ間参照ピクチャを含まず、動き情報が参照される可能性のあるレイヤ間参照ピクチャを含む、という性質を満たす限りにおいては、異なる方法で導出しても構わない。なお、上記では、レイヤ間画素RPSとレイヤ間動き限定RPSのサブRPSを導出したが、レイヤ間画素RPSとレイヤ間画素非依存RPSのサブRPSを導出しても良い。この場合、S303とS305の判定を、各々、以下の判定S303r2とS305r2に置き換える。 The method for deriving the inter-layer pixel RPS and the inter-layer motion limited RPS described with reference to FIG. 20 is merely an example. Inter-layer pixel RPS and inter-layer motion limited RPS have specific characteristics, i.e., inter-layer pixel RPS includes an inter-layer reference picture to which a decoded pixel may be referenced, and inter-layer motion limited RPS includes a decoded pixel Different methods may be used as long as the property that an inter-layer reference picture that may be referred to is not included and an inter-layer reference picture that motion information may be referred to is included is satisfied. In the above description, the sub-RPS of the inter-layer pixel RPS and the inter-layer motion limited RPS is derived. However, the inter-layer pixel RPS and the inter-layer pixel-independent RPS sub-RPS may be derived. In this case, the determinations in S303 and S305 are replaced with the following determinations S303r2 and S305r2, respectively.
 (S303r2)i番目のアクティブレイヤ間参照ピクチャの依存タイプが、レイヤ間参照ピクチャの復号画素が参照される可能性のあることを示す場合(i番目のアクティブレイヤ間参照ピクチャに対応する参照ピクチャのSampleEnableFlagが1の場合)、S304に進む。それ以外の場合、S305に進む。 (S303r2) When the dependency type of the i-th active layer reference picture indicates that the decoded pixel of the inter-layer reference picture may be referred to (the reference picture corresponding to the i-th active layer reference picture) When SampleEnableFlag is 1, the process proceeds to S304. Otherwise, the process proceeds to S305.
 (S305r2)i番目のアクティブレイヤ間参照ピクチャの依存タイプが、レイヤ間参照ピクチャの復号画素が参照されることを示さない場合(i番目のアクティブレイヤ間参照ピクチャに対応する参照ピクチャのSampleEnableFlagが0の場合)、S306に進む。それ以外の場合、S307に進む。なお、依存タイプの種類が、画素依存と動き依存の2つのみの場合は、レイヤ間動き限定RPSとレイヤ間画素非依存RPSは等しくなる。 (S305r2) When the dependency type of the i-th active layer reference picture does not indicate that the decoded pixel of the inter-layer reference picture is referred to (SampleEnableFlag of the reference picture corresponding to the i-th active layer reference picture is 0) ), The process proceeds to S306. Otherwise, the process proceeds to S307. Note that when there are only two types of dependency types, pixel-dependent and motion-dependent, the inter-layer motion limited RPS and the inter-layer pixel-independent RPS are equal.
  [[参照ピクチャ制御部3063]] 参照ピクチャ制御部3063は、入力されるRPSに基づいて、DPB3061を更新する。概略的には、参照ピクチャ制御部3063は、入力されたRPSにおいて 対象ピクチャ(現ピクチャ)のインター予測で参照可能と示されているピクチャの参照マークを「参照使用」(「短期参照使用」または「長期参照使用」)に設定する。加えて、DPBに記録されている対象レイヤの復号ピクチャであって、前記の処理で「参照使用」とマークされないピクチャを「参照不使用」に設定する。なお、DPB上のレイヤ間参照ピクチャの参照マークは参照ピクチャ制御部3063では変更しない。言い換えると、ベース復号ピクチャに由来するDPB上のピクチャの参照マークの変更は参照ピクチャ制御部3063では行わず、後述の参照レイヤピクチャ制御部3064で行う。 [[Reference picture control unit 3063]] The reference picture control unit 3063 updates the DPB 3061 based on the input RPS. Schematically, the reference picture control unit 3063 performs “reference use” (“short-term reference use” or “short-term reference use” or a reference mark of a picture that is indicated by the inter prediction of the target picture (current picture) in the input RPS. Set to “Long term reference use”). In addition, a decoded picture of the target layer recorded in the DPB, which is not marked as “reference use” in the above process, is set to “reference nonuse”. Note that the reference mark of the inter-layer reference picture on the DPB is not changed by the reference picture control unit 3063. In other words, the reference mark change of the picture on the DPB derived from the base decoded picture is not performed by the reference picture control unit 3063 but by the reference layer picture control unit 3064 described later.
  [[参照レイヤピクチャ制御部3064]] 参照レイヤピクチャ制御部3064は、入力される参照レイヤの復号ピクチャとRPSに基づいて、DPB3061を更新する。概略的には、参照レイヤピクチャ制御部3064は、入力されたRPSにおいて 対象ピクチャ(現ピクチャ)のレイヤ間インター予測で参照可能と示されているピクチャに対応する参照レイヤの復号ピクチャをDPBに記録する。加えて、DPB上で当該ピクチャの参照マークを「参照使用」(「短期参照使用」または「長期参照使用」に設定する。加えて、DPB上で当該ピクチャの出力マークを「出力不要」に設定する。なお、参照レイヤの復号ピクチャを記録する際に、必要に応じてスケーリングやフィルタリングを適用した後に前記ピクチャバッファに記録してもよい。特に、参照レイヤと対象レイヤの出力ピクチャの解像度が異なる場合(空間スケーラビリティの場合)、参照レイヤの復号ピクチャを、対象レイヤの出力ピクチャの解像度に合わせてスケーリングする必要がある。 [[Reference layer picture control unit 3064]] The reference layer picture control unit 3064 updates the DPB 3061 based on the input decoded picture and RPS of the reference layer. Schematically, the reference layer picture control unit 3064 records, in the DPB, a decoded picture of a reference layer corresponding to a picture that can be referred to in inter-layer inter prediction of the target picture (current picture) in the input RPS. To do. In addition, the reference mark of the picture is set to “use reference” (“use of short-term reference” or “use of long-term reference” on the DPB. In addition, the output mark of the picture is set to “output unnecessary” on the DPB. In addition, when recording the decoded picture of the reference layer, it may be recorded in the picture buffer after applying scaling and filtering as necessary, in particular, the resolution of the output picture of the reference layer and the target layer is different. In this case (in the case of spatial scalability), it is necessary to scale the decoded picture of the reference layer according to the resolution of the output picture of the target layer.
  [[RPL導出部3065]] RPL導出部3065は、入力RPS、および、入力シンタックス値に含まれるRPL情報に基づいて対象ピクチャの対象スライスのインター予測またはレイヤ間予測で利用する参照ピクチャリストを導出して出力する。 [[RPL Deriving Unit 3065]] The RPL deriving unit 3065 uses a reference picture list to be used for inter prediction or inter-layer prediction of the target slice of the target picture based on the input RPS and the RPL information included in the input syntax value. Derived and output.
 RPL情報は、参照ピクチャリストRPLを構築するためにSPSまたはスライスヘッダより復号されるシンタックス値である。RPL情報は、SPSリスト修正情報、および、SHリスト修正情報から構成される。 The RPL information is a syntax value that is decoded from the SPS or slice header in order to construct the reference picture list RPL. The RPL information includes SPS list correction information and SH list correction information.
 SPSリスト修正情報はSPSに含まれる情報であり、参照ピクチャリスト修正の制約に係る情報である。SPSリスト修正情報について、図16を再び参照して説明する。図16の(C)の部分がSPSリスト修正情報に相当する。SPSリスト修正情報には、ピクチャに含まれる前スライスで参照ピクチャリストが共通か否かを示すフラグ(restricted_ref_pic_lists_flag)、および、スライスヘッダ内にリスト並べ替えに関する情報が存在するか否かを示すフラグ(lists_modification_present_flag)が含まれる。 The SPS list modification information is information included in the SPS, and is information related to restrictions on modification of the reference picture list. The SPS list correction information will be described with reference to FIG. 16 again. The part (C) in FIG. 16 corresponds to SPS list correction information. In the SPS list correction information, a flag (restricted_ref_pic_lists_flag) indicating whether or not the reference picture list is common in the previous slice included in the picture, and a flag (whether or not information related to list rearrangement exists in the slice header ( lists_modification_present_flag).
 SHリスト修正情報はスライスヘッダに含まれる情報であり、対象ピクチャに適用される参照ピクチャリストの長さ(参照リスト長)の更新情報、および、参照ピクチャリストの並べ替え情報(参照リスト並べ替え情報)が含まれる。SHリスト修正情報について、図21の(a)を参照して説明する。図21の(a)はスライスヘッダ復号時に利用されるスライスヘッダシンタックス表の一部を例示している。図21の(a)の(C)の部分がSHリスト修正情報に相当する。 The SH list correction information is information included in the slice header, and the update information of the length of the reference picture list (reference list length) applied to the target picture, and the reordering information of the reference picture list (reference list reordering information) ) Is included. The SH list correction information will be described with reference to FIG. FIG. 21A illustrates a part of a slice header syntax table used when decoding a slice header. The part (C) in FIG. 21A corresponds to the SH list correction information.
 参照リスト長更新情報として、リスト長の更新有無を示すフラグ(num_ref_idx_active_override_flag)が含まれる。加えて、L0参照リストの変更後の参照リスト長を表す情報(num_ref_idx_l0_active_minus1)、および、L1参照リストの変更後の参照リスト長を表す情報(num_ref_idx_l1_active_minus1)が含まれる。 The reference list length update information includes a flag (num_ref_idx_active_override_flag) indicating whether or not the list length is updated. In addition, information (num_ref_idx_l0_active_minus1) indicating the reference list length after the change of the L0 reference list and information (num_ref_idx_l1_active_minus1) indicating the reference list length after the change of the L1 reference list are included.
 参照リスト並べ替え情報としてスライスヘッダに含まれる情報について、図21の(b)を参照して説明する。図21の(b)はスライスヘッダ復号時に利用される参照リスト並べ替え情報のシンタックス表を例示している。 Information included in the slice header as reference list rearrangement information will be described with reference to FIG. FIG. 21B illustrates a syntax table of reference list rearrangement information used at the time of decoding the slice header.
 参照リスト並べ替え情報には、L0参照リスト並べ替え有無フラグ(ref_pic_list_modification_flag_l0)が含まれる。前記フラグの値が1(L0参照リストの並べ替えが有る場合)、かつ、NumPocTotalCurrが2より大きい場合、L0参照リスト並べ替え順序(list_entry_l0[i])が参照リスト並べ替え情報に含まれる。ここで、NumPocTotalCurrは、現ピクチャにおいて利用可能な参照ピクチャの数を表す変数である。したがって、L0参照リストの並べ替えが有る場合であって、かつ、現ピクチャにおいて利用可能な参照ピクチャ数が2より大きい場合にのみ、L0参照リスト並べ替え順序がスライスヘッダに含まれる。 The reference list rearrangement information includes an L0 reference list rearrangement presence / absence flag (ref_pic_list_modification_flag_l0). When the value of the flag is 1 (when the L0 reference list is rearranged) and NumPocTotalCurr is larger than 2, the L0 reference list rearrangement order (list_entry_l0 [i]) is included in the reference list rearrangement information. Here, NumPocTotalCurr is a variable representing the number of reference pictures that can be used in the current picture. Therefore, the L0 reference list rearrangement order is included in the slice header only when the L0 reference list is rearranged and the number of reference pictures available in the current picture is larger than two.
 同様に、参照ピクチャがBスライスである場合、つまり、対象ピクチャにおいてL1参照リストが利用可能である場合、L1参照リスト並べ替え有無フラグ(ref_pic_list_modification_flag_l1)が参照リスト並べ替え情報に含まれる。前記フラグの値が1、かつ、NumPocTotalCurrが2より大きい場合、L1参照リスト並べ替え順序(list_entry_l1[i])が参照リスト並べ替え情報に含まれる。言い換えると、L1参照リストの並べ替えが有る場合であって、かつ、現ピクチャにおいて利用可能な参照ピクチャ数が2より大きい場合にのみ、L1参照リスト並べ替え順序がスライスヘッダに含まれる。 Similarly, when the reference picture is a B slice, that is, when the L1 reference list is available in the target picture, the L1 reference list rearrangement presence / absence flag (ref_pic_list_modification_flag_l1) is included in the reference list rearrangement information. When the value of the flag is 1 and NumPocTotalCurr is greater than 2, the L1 reference list rearrangement order (list_entry_l1 [i]) is included in the reference list rearrangement information. In other words, the L1 reference list rearrangement order is included in the slice header only when the L1 reference list is rearranged and the number of reference pictures available in the current picture is larger than two.
   (RPL導出処理) 参照ピクチャリスト導出処理(RPL導出処理)の詳細を説明する。参照ピクチャリスト導出部は、参照ピクチャセットRPSと、RPL修正情報に基づいて、対象ピクチャの復号に用いる参照ピクチャリストRPLを生成する。 (RPL derivation process) Details of the reference picture list derivation process (RPL derivation process) will be described. The reference picture list deriving unit generates a reference picture list RPL used for decoding the target picture based on the reference picture set RPS and the RPL correction information.
 参照ピクチャリストにはL0参照リストとL1参照リストの2つのリストがある。各参照ピクチャリスト導出の説明に先立って、導出処理の過程で生成される暫定L0参照リストと暫定L1参照リストの特徴を図22を用いて説明しておく。図22は、RPL導出部3065におけるL0参照リストとL1参照リスト導出の中間過程で生成される暫定L0参照リストと暫定L1参照リストの概略を示す図である。 There are two lists of reference picture lists: an L0 reference list and an L1 reference list. Prior to the description of deriving each reference picture list, the characteristics of the provisional L0 reference list and the provisional L1 reference list generated in the course of the derivation process will be described with reference to FIG. FIG. 22 is a diagram showing an outline of the temporary L0 reference list and the temporary L1 reference list generated in the intermediate process of the L0 reference list and L1 reference list derivation in the RPL deriving unit 3065.
 図22(a)に示すように、暫定L0参照リストは、リストの先頭から順に(優先度の高い順に)前方短期RPS(図中でStBef)、レイヤ間画素RPS(図中でILSample)、後方短期RPS(図中でStAft)、長期RPS(図中でLt)、レイヤ間動き限定RPS(図中でILMotion)の順にサブRPSを含む。 As shown in FIG. 22 (a), the provisional L0 reference list includes a forward short-term RPS (StBef in the figure), an inter-layer pixel RPS (ILSample in the figure), and a rear in order from the top of the list (in order of priority). Sub RPSs are included in the order of short-term RPS (StAft in the figure), long-term RPS (Lt in the figure), and inter-layer motion limited RPS (ILMotion in the figure).
 図22(b)に示すように、暫定L1参照リストは、リストの先頭から順に(優先度の高い順に)後方短期RPS(図中でStAft)、前方短期RPS(図中でStBef)、長期RPS(図中でLt)、レイヤ間画素RPS(図中でILSample)、レイヤ間動き限定RPS(図中でILMotion)の順にサブRPSを含む。 As shown in FIG. 22 (b), the provisional L1 reference list includes a backward short-term RPS (StAft in the figure), forward short-term RPS (StBef in the figure), and long-term RPS in order from the top of the list (in order of priority). (Lt in the drawing), inter-layer pixel RPS (ILSample in the drawing), and inter-layer motion limited RPS (ILMotion in the drawing) are included in this order.
 つまり、RPL導出部3065にて生成される暫定参照リスト(暫定参照L0参照リストおよび暫定L1参照リスト)は、レイヤ間画素RPSを、レイヤ間動き限定RPSに較べて、よりリストの先頭に近い位置に含んでいる。換言すると、暫定参照リストは、レイヤ間画素RPSを、レイヤ間動き限定RPSに較べてより高い優先度に対応する位置に含んでいる。 That is, the provisional reference list (the provisional reference L0 reference list and the provisional L1 reference list) generated by the RPL deriving unit 3065 is a position closer to the top of the list of the inter-layer pixel RPS than the inter-layer motion limited RPS. Is included. In other words, the provisional reference list includes the inter-layer pixel RPS at a position corresponding to a higher priority than the inter-layer motion limited RPS.
 続いて、L0参照リストとL1参照リストの導出手順を説明する。L0参照リストは、以下のS401~S409に示す手順で構築される。 Subsequently, the procedure for deriving the L0 reference list and the L1 reference list will be described. The L0 reference list is constructed according to the procedure shown in S401 to S409 below.
 (S401)暫定L0参照リストを生成して、空のリストに初期化する。 (S401) A temporary L0 reference list is generated and initialized to an empty list.
 (S402)暫定L0参照リストに対し、前方短期RPSに含まれる参照ピクチャを順に追加する。 (S402) Reference pictures included in the forward short-term RPS are sequentially added to the provisional L0 reference list.
 (S403)暫定L0参照リストに対し、レイヤ間画素RPSに含まれる参照ピクチャを順に追加する。 (S403) Reference pictures included in the inter-layer pixel RPS are sequentially added to the provisional L0 reference list.
 (S404)暫定L0参照リストに対し、後方短期RPSに含まれる参照ピクチャを順に追加する。 (S404) The reference pictures included in the backward short-term RPS are sequentially added to the provisional L0 reference list.
 (S405)暫定L0参照リストに対し、長期RPSに含まれる参照ピクチャを順に追加する。 (S405) Reference pictures included in the long-term RPS are sequentially added to the provisional L0 reference list.
 (S406)暫定L0参照リストに対し、レイヤ間動き限定RPSに含まれる参照ピクチャを順に追加する。 (S406) Reference pictures included in the inter-layer motion limited RPS are sequentially added to the provisional L0 reference list.
 (S407)参照ピクチャリストが修正される場合(RPL修正情報に含まれるlists_modification_present_flagの値が1の場合)、以下のS408に進む。それ以外の場合(lists_modification_present_flagの値が0の場合)、S409に進む。 (S407) When the reference picture list is modified (when the value of lists_modification_present_flag included in the RPL modification information is 1), the process proceeds to S408 below. In other cases (when the value of lists_modification_present_flag is 0), the process proceeds to S409.
 (S408)参照リスト並べ替え順序list_entry_l0[i]の値に基づいて、暫定L0参照リストの要素を並べ換えて、L0参照リストとする。参照ピクチャインデックスrIdxに対応するL0参照リストの要素RefPicList0[rIdx]は、次式により導出される。ここで、RefListTemp0[i]は、暫定L0参照リストのi番目の要素を表す。 (S408) Based on the value of the reference list rearrangement order list_entry_l0 [i], the elements of the provisional L0 reference list are rearranged to form the L0 reference list. The element RefPicList0 [rIdx] of the L0 reference list corresponding to the reference picture index rIdx is derived by the following equation. Here, RefListTemp0 [i] represents the i-th element of the provisional L0 reference list.
 RefPicList0[ rIdx ] = RefPicListTemp0[ list_entry_l0[ rIdx ] ]
上記の式によれば、参照リスト並べ替え順序list_entry_l0[i]において、参照ピクチャインデックスrIdxの示す位置に記録されている値を参照し、暫定L0参照リストにおいて前記値の位置に記録されている参照ピクチャを、L0参照リストのrIdxの位置の参照ピクチャとして格納する。
RefPicList0 [rIdx] = RefPicListTemp0 [list_entry_l0 [rIdx]]
According to the above formula, in the reference list rearrangement order list_entry_l0 [i], the value recorded at the position indicated by the reference picture index rIdx is referred to, and the reference recorded at the position of the value in the provisional L0 reference list The picture is stored as a reference picture at the position of rIdx in the L0 reference list.
 (S409)暫定L0参照リストをL0参照リストとする。 (S409) The provisional L0 reference list is set as the L0 reference list.
 次にL1参照リストの構築手順を説明する。L1参照リストの構築は、以下のS501~S509に記載の手順で実行される。 Next, the L1 reference list construction procedure will be described. The construction of the L1 reference list is executed according to the procedure described in S501 to S509 below.
 (S501)暫定L1参照リストを生成して、空のリストに初期化する。 (S501) A provisional L1 reference list is generated and initialized to an empty list.
 (S502)暫定L1参照リストに対し、後方短期RPSに含まれる参照ピクチャを順に追加する。 (S502) The reference pictures included in the backward short-term RPS are sequentially added to the provisional L1 reference list.
 (S503)暫定L1参照リストに対し、前方短期RPSに含まれる参照ピクチャを順に追加する。 (S503) The reference pictures included in the forward short-term RPS are sequentially added to the provisional L1 reference list.
 (S504)暫定L1参照リストに対し、長期RPSに含まれる参照ピクチャを順に追加する。 (S504) Reference pictures included in the long-term RPS are sequentially added to the provisional L1 reference list.
 (S505)暫定L1参照リストに対し、レイヤ間画素RPSに含まれる参照ピクチャを順に追加する。 (S505) Reference pictures included in the inter-layer pixel RPS are sequentially added to the provisional L1 reference list.
 (S506)暫定L1参照リストに対し、レイヤ間動き限定RPSに含まれる参照ピクチャを順に追加する。 (S506) Reference pictures included in the inter-layer motion limited RPS are sequentially added to the provisional L1 reference list.
 (S507)参照ピクチャリストが修正される場合(RPL修正情報に含まれるlists_modification_present_flagの値が1の場合)、以下のS508に進む。それ以外の場合(lists_modification_present_flagの値が0の場合)、S509に進む。 (S507) When the reference picture list is modified (when the value of lists_modification_present_flag included in the RPL modification information is 1), the process proceeds to S508 below. In other cases (when the value of lists_modification_present_flag is 0), the process proceeds to S509.
 (S508)参照リスト並べ替え順序list_entry_l1 [i]の値に基づいて、暫定L1参照リストの要素を並べ換えて、L1参照リストとする。参照ピクチャインデックスrIdxに対応するL1参照リストの要素RefPicList1[rIdx]は、次式により導出される。ここで、RefListTemp1[i]は、暫定L1参照リストのi番目の要素を表す。 (S508) The elements of the provisional L1 reference list are rearranged based on the value of the reference list rearrangement order list_entry_l1 リ ス ト [i] to obtain an L1 reference list. The element RefPicList1 [rIdx] of the L1 reference list corresponding to the reference picture index rIdx is derived by the following equation. Here, RefListTemp1 [i] represents the i-th element of the provisional L1 reference list.
 RefPicList1[ rIdx ] = RefPicListTemp1[ list_entry_l1[ rIdx ] ]
上記の式によれば、参照リスト並べ替え順序list_entry_l1[i]において、参照ピクチャインデックスrIdxの示す位置に記録されている値を参照し、暫定L1参照リストにおいて前記値の位置に記録されている参照ピクチャを、L1参照リストのrIdxの位置の参照ピクチャとして格納する。
RefPicList1 [rIdx] = RefPicListTemp1 [list_entry_l1 [rIdx]]
According to the above formula, in the reference list rearrangement order list_entry_l1 [i], the value recorded at the position indicated by the reference picture index rIdx is referred to, and the reference recorded at the position of the value in the provisional L1 reference list The picture is stored as a reference picture at the position of rIdx in the L1 reference list.
 (S509)暫定L1参照リストをL1参照リストとする。 (S509) The provisional L1 reference list is set as the L1 reference list.
 上記の参照ピクチャリスト導出手順によれば、参照ピクチャリスト(参照リストL0と参照リストL1)が、対応する暫定参照リスト(暫定参照リストL0と暫定参照リストL1)をRPL修正情報に基づいて選択および並べ替えることにより生成される。その際、暫定参照リストは、レイヤ間画素RPSを、レイヤ間動き限定RPSに較べてより高い優先度に対応する位置に含んでいる。言い換えると、暫定参照リストは、レイヤ間画素RPSを、レイヤ間動き限定RPSに較べてよりリストの先頭に近い位置に含んでいる。 According to the above reference picture list derivation procedure, the reference picture list (reference list L0 and reference list L1) selects and selects the corresponding provisional reference list (provisional reference list L0 and provisional reference list L1) based on the RPL correction information. Generated by sorting. At that time, the provisional reference list includes the inter-layer pixel RPS at a position corresponding to a higher priority than the inter-layer motion limited RPS. In other words, the provisional reference list includes the inter-layer pixel RPS at a position closer to the top of the list than the inter-layer motion limited RPS.
 RPL修正情報が並べ替えを行わないことを示す場合、レイヤ間画素RPSに含まれるレイヤ間参照ピクチャが、レイヤ間動き限定RPSに含まれるレイヤ間参照ピクチャ較べて参照ピクチャリストの先頭に近い位置(より小さい参照ピクチャインデックスに対応する位置)に関連付けられる。したがって、レイヤ間画素RPSに含まれるレイヤ間参照ピクチャを、レイヤ間動き限定RPSに含まれるレイヤ間参照ピクチャ較べてより小さい値の参照ピクチャインデックスで指定できる。レイヤ間画素RPSに含まれるレイヤ間参照ピクチャは、レイヤ間画素予測に使われる可能性のあるレイヤ間参照ピクチャである。一方、レイヤ間動き限定RPSに含まれるレイヤ間参照ピクチャは、レイヤ間画素予測に使われる可能性はなく、レイヤ間動き予測に使われる可能性のあるレイヤ間参照ピクチャである。一般に、レイヤ間画像予測に用いる参照ピクチャを指定するための参照ピクチャインデックスの方が、レイヤ間動き予測に用いる参照ピクチャを指定するための参照ピクチャインデックスに較べてより多数含まれている。例えば、前者は予測単位毎に含まれるのに対し、後者はスライス毎に含まれる。したがって、レイヤ間画素RPSに含まれるレイヤ間参照ピクチャを、レイヤ間動き限定RPSに含まれるレイヤ間参照ピクチャに較べて小さい参照ピクチャインデックスを割り当てて、より少ない符号量で指定できるようにすることで、符号化データ全体の符号量を低減できる。 When the RPL correction information indicates that no reordering is performed, the inter-layer reference picture included in the inter-layer pixel RPS is closer to the top of the reference picture list than the inter-layer reference picture included in the inter-layer motion limited RPS ( Associated with the smaller reference picture index). Therefore, the inter-layer reference picture included in the inter-layer pixel RPS can be designated with a reference picture index having a smaller value than the inter-layer reference picture included in the inter-layer motion limited RPS. The inter-layer reference picture included in the inter-layer pixel RPS is an inter-layer reference picture that may be used for inter-layer pixel prediction. On the other hand, the inter-layer reference picture included in the inter-layer motion limited RPS is an inter-layer reference picture that may not be used for inter-layer pixel prediction but may be used for inter-layer motion prediction. In general, a larger number of reference picture indexes for specifying reference pictures used for inter-layer image prediction are included than reference picture indexes for specifying reference pictures used for inter-layer motion prediction. For example, the former is included for each prediction unit, while the latter is included for each slice. Therefore, the inter-layer reference picture included in the inter-layer pixel RPS is assigned a smaller reference picture index than the inter-layer reference picture included in the inter-layer motion limited RPS so that it can be specified with a smaller code amount. Thus, the code amount of the entire encoded data can be reduced.
 RPL修正情報が並べ変えを行うことを示す場合、参照リスト並べ替え情報(list_entry_l0、list_entry_l1)により暫定参照リスト中の位置を指定することで、暫定参照リストを並べ替えて参照リストを生成する。この場合、参照ピクチャリストの先頭に並べ替えられる可能性の高い参照ピクチャを、より小さい値の参照リスト並べ替え情報で指定できるようにすることで、RPL修正情報の符号量を削減できる。暫定参照リストでは、レイヤ間画素RPSに含まれるレイヤ間参照ピクチャが、参照ピクチャリストの先頭に並べ替えられる可能性のより低いレイヤ間動き限定RPSに含まれるレイヤ間参照ピクチャに較べて、暫定参照リストの先頭に近い位置にあるため、より小さい値の参照リスト並べ替え情報で指定できる。したがって、上記手順で導出した暫定参照リストを用いることで、RPL修正情報の符号量を低減できる。 When the RPL correction information indicates that reordering is to be performed, the position in the provisional reference list is designated by the reference list reordering information (list_entry_l0, list_entry_l1), thereby reordering the provisional reference list to generate a reference list. In this case, it is possible to reduce the code amount of the RPL correction information by making it possible to specify a reference picture that is likely to be rearranged at the head of the reference picture list with reference list rearrangement information having a smaller value. In the provisional reference list, the inter-layer reference picture included in the inter-layer pixel RPS is less likely to be rearranged at the top of the reference picture list, compared to the inter-layer reference picture included in the inter-layer motion limited RPS. Since it is near the top of the list, it can be specified with reference list sorting information with a smaller value. Therefore, the code amount of the RPL correction information can be reduced by using the provisional reference list derived by the above procedure.
 加えて、上記の暫定参照リストは、動き限定RPSを暫定参照リストの末尾に含んでいる。また、上記の暫定参照リストは、動き限定RPSを短期RPSよりもリストの後方に近い位置に含んでいる。動き限定参照リストに含まれるレイヤ間参照ピクチャは、同一レイヤ内のインター予測に係るサブRPSである短期RPSや長期RPSに含まれる参照ピクチャよりも選択頻度が低いため、上記のように暫定参照リストを構成することで、参照ピクチャの選択に係る情報をより少ない符号量で復号できる。 In addition, the above provisional reference list includes a motion limited RPS at the end of the provisional reference list. The provisional reference list includes the motion limited RPS at a position closer to the back of the list than the short-term RPS. The inter-layer reference pictures included in the motion-limited reference list have a lower selection frequency than the reference pictures included in the short-term RPS and long-term RPS that are sub-RPSs related to inter prediction in the same layer. The information related to the selection of the reference picture can be decoded with a smaller code amount.
 なお、レイヤ間画素RPSとレイヤ間動き限定RPSのサブRPSの代わりに、レイヤ間画素RPSとレイヤ間画素非依存RPSのサブRPSを導出する場合には、上記、参照ピクチャリスト導出処理(RPL導出処理)において、レイヤ間動き限定RPSを、レイヤ間画素非依存RPSに読み替えて処理する。より具体的には、S406とS506の処理を以下の処理に置き換える。 When deriving the sub-RPS of the inter-layer pixel RPS and the inter-layer pixel-independent RPS instead of the sub-RPS of the inter-layer pixel RPS and the inter-layer motion limited RPS, the reference picture list derivation process (RPL derivation) is performed. In the processing), the inter-layer motion limited RPS is read as an inter-layer pixel-independent RPS and processed. More specifically, the processes of S406 and S506 are replaced with the following processes.
 (S406r2)暫定L0参照リストに対し、レイヤ間画素非依存RPSに含まれる参照ピクチャを順に追加する。 (S406r2) The reference pictures included in the inter-layer pixel-independent RPS are sequentially added to the provisional L0 reference list.
 (S506r2)暫定L1参照リストに対し、レイヤ間画素非依存RPSに含まれる参照ピクチャを順に追加する。 (S506r2) Reference pictures included in the inter-layer pixel-independent RPS are sequentially added to the provisional L1 reference list.
 これにより、暫定参照リストは、レイヤ間画素RPSを、レイヤ間画素非依存RPSに較べてより高い優先度に対応する位置に含んでいる。言い換えると、暫定参照リストは、レイヤ間画素RPSを、レイヤ間画素非依存RPSに較べてよりリストの先頭に近い位置に含んでいる。また、上記の暫定参照リストは、画素非依存限定RPSを暫定参照リストの末尾に含んでいる。また、上記の暫定参照リストは、画素非依存RPSを短期RPSよりもリストの後方に近い位置に含んでいる。 Thereby, the provisional reference list includes the inter-layer pixel RPS at a position corresponding to a higher priority than the inter-layer pixel-independent RPS. In other words, the temporary reference list includes the inter-layer pixel RPS at a position closer to the top of the list than the inter-layer pixel-independent RPS. In addition, the provisional reference list includes a pixel-independent limited RPS at the end of the provisional reference list. The provisional reference list includes a pixel-independent RPS at a position closer to the back of the list than the short-term RPS.
 上述のように、レイヤ間画像予測に用いる参照ピクチャを指定するための参照ピクチャインデックスは予測単位毎に含まれるため、一般の場合においても、レイヤ間画像予測以外の予測(例えば、動き情報依存の場合のレイヤ間動き予測)の場合に、参照ピクチャを指定するための参照ピクチャインデックスよりも、多数含まれると考えられる。従って、レイヤ間画素RPSに含まれるレイヤ間参照ピクチャを、レイヤ間画素非依存RPSに含まれるレイヤ間参照ピクチャに較べて小さい参照ピクチャインデックスを割り当てて、より少ない符号量で指定できるようにすることで、符号化データ全体の符号量を低減できる。 As described above, since a reference picture index for specifying a reference picture used for inter-layer image prediction is included in each prediction unit, prediction other than inter-layer image prediction (for example, motion information dependent) is also used in general cases. In the case of motion prediction between layers), it is considered that a larger number than the reference picture index for specifying the reference picture is included. Therefore, an inter-layer reference picture included in an inter-layer pixel RPS can be designated with a smaller code amount by assigning a smaller reference picture index than an inter-layer reference picture included in an inter-layer pixel-independent RPS. Thus, the code amount of the entire encoded data can be reduced.
  [[出力制御部3066]] 出力制御部3066は、概略的には、所定のタイミングでDPB3061のピクチャを外部に出力するとともに出力マークを更新する。具体的には、出力制御部3066によるピクチャ出力処理は次の手順で実行される。 [[Output Control Unit 3066]] The output control unit 3066 generally outputs the DPB 3061 picture to the outside and updates the output mark at a predetermined timing. Specifically, the picture output processing by the output control unit 3066 is executed in the following procedure.
 まず、DPB上のピクチャであって、出力マークが「出力要」であるピクチャのうちPOCが最小であるピクチャを出力する。次に、出力されたピクチャの出力マークを「出力不要」に設定する。最後に、DPB上のピクチャの中から、参照マークが「参照不使用」であり、かつ、出力マークが「出力不要」であるピクチャを選択して、当該ピクチャをDPBから削除する。 First, a picture with a minimum POC is output from the pictures on the DPB whose output mark is “output required”. Next, the output mark of the output picture is set to “output unnecessary”. Finally, a picture whose reference mark is “reference not used” and whose output mark is “output unnecessary” is selected from the pictures on the DPB, and the picture is deleted from the DPB.
  [[予測パラメータメモリ3067]] 予測パラメータメモリ3067には、インター予測パラメータ復号部303が復号したインター予測パラメータ、イントラ予測パラメータ復号部304が復号したイントラ予測パラメータ及びエントロピー復号部301が分離した予測モードpredModeが記憶される。予測パラメータメモリ3067に記憶されるインター予測パラメータには、例えば、予測リスト利用フラグpredFlagLX(インター予測フラグinter_pred_idcinter_pred_idc)、参照ピクチャインデックスrefIdxLX、ベクトルmvLXがある。 [[Prediction parameter memory 3067]] The prediction parameter memory 3067 includes inter prediction parameters decoded by the inter prediction parameter decoding unit 303, intra prediction parameters decoded by the intra prediction parameter decoding unit 304, and prediction modes separated by the entropy decoding unit 301. predMode is stored. The inter prediction parameters stored in the prediction parameter memory 3067 include, for example, a prediction list use flag predFlagLX (inter prediction flag inter_pred_idcinter_pred_idc), a reference picture index refIdxLX, and a vector mvLX.
  [インター予測画像生成部309] 図10は、本実施形態に係るインター予測画像生成部309の構成を示す概略図である。インター予測画像生成部309は、動き変位補償部3091、残差予測部3092、照度補償部3093、重み予測部3094、参照画像判定部3095を含んで構成される。 [Inter predicted image generation unit 309] FIG. 10 is a schematic diagram illustrating a configuration of the inter predicted image generation unit 309 according to the present embodiment. The inter predicted image generation unit 309 includes a motion displacement compensation unit 3091, a residual prediction unit 3092, an illuminance compensation unit 3093, a weight prediction unit 3094, and a reference image determination unit 3095.
 動き変位補償部3091は、インター予測パラメータ復号部303から入力された、予測リスト利用フラグpredFlagLX、参照ピクチャインデックスrefIdxLX、動きベクトルmvLXに基づいて、復号ピクチャ管理部306のDPB3061から、参照ピクチャインデックスrefIdxLXで指定された参照ピクチャの対象ブロックの位置を起点として、ベクトルmvLXだけずれた位置にあるブロックを読み出すことによって動き変位補償画像を生成する。ここで、ベクトルmvLXが整数ベクトルでない場合には、動き補償フィルタ(もしくは変位補償フィルタ)と呼ばれる小数位置の画素を生成するためのフィルタを施して、動き変位補償画像を生成する。一般に、ベクトルmvLXが動きベクトルの場合、上記処理を動き補償と呼び、変位ベクトルの場合は変位補償と呼ぶ。ここでは総称して動き変位補償と表現する。以下、L0予測の動き変位補償画像をpredSamplesL0、L1予測の動き変位補償画像をpredSamplesL1と呼ぶ。両者を区別しない場合predSamplesLXと呼ぶ。以下、動き変位補償部3091で得られた動き変位補償画像predSamplesLXに、さらに残差予測および照度補償が行われる例を説明するが、これらの出力画像もまた、動き変位補償画像predSamplesLXと呼ぶ。なお、以下の残差予測および照度補償において、入力画像と出力画像を区別する場合には、入力画像をpredSamplesLX、出力画像をpredSamplesLX´と表現する。 Based on the prediction list use flag predFlagLX, the reference picture index refIdxLX, and the motion vector mvLX input from the inter prediction parameter decoding unit 303, the motion displacement compensation unit 3091 uses the reference picture index refIdxLX from the DPB 3061 of the decoded picture management unit 306. A motion displacement compensation image is generated by reading out a block at a position shifted by the vector mvLX starting from the position of the target block of the designated reference picture. Here, when the vector mvLX is not an integer vector, a motion displacement compensation image is generated by applying a filter called a motion compensation filter (or displacement compensation filter) for generating a pixel at a decimal position. In general, when the vector mvLX is a motion vector, the above processing is called motion compensation, and when the vector mvLX is a displacement vector, it is called displacement compensation. Here, it is collectively referred to as motion displacement compensation. Hereinafter, the motion displacement compensation image for L0 prediction is referred to as predSamplesL0, and the motion displacement compensation image for L1 prediction is referred to as predSamplesL1. If the two are not distinguished, they are called predSamplesLX. Hereinafter, an example in which residual prediction and illuminance compensation are further performed on the motion displacement compensation image predSamplesLX obtained by the motion displacement compensation unit 3091 will be described. These output images are also referred to as motion displacement compensation images predSamplesLX. In the following residual prediction and illuminance compensation, when an input image and an output image are distinguished, the input image is expressed as predSamplesLX and the output image is expressed as predSamplesLX ′.
 参照画像判定部3095は、残差予測を行うために用いられる、参照レイヤ(参照ビュー)における参照ピクチャ(復号画像)(refIvRefPic)(以下、説明の便宜のため“参照レイヤ参照ピクチャrefIvRefPic”にて参照する)が利用可能か否かを判定する。参照画像判定部3095は、参照レイヤ参照ピクチャrefIvRefPicが利用可能な場合、参照レイヤ参照ピクチャ利用フラグrefIvRefPicAvailable(refIvRefPicAvailable2)に1を設定する。一方、参照画像判定部3095は、参照レイヤ参照ピクチャrefIvRefPicが利用不可能な場合、参照レイヤ参照ピクチャ利用フラグrefIvRefPicAvailable(refIvRefPicAvailable2)に0を設定する。参照画像判定部3095は、このようにして設定した参照レイヤ参照ピクチャ利用フラグrefIvRefPicAvailable(refIvRefPicAvailable2)を残差予測部3092に出力する。 The reference image determination unit 3095 uses a reference picture (decoded image) (refIvRefPic) in a reference layer (reference view) (hereinafter referred to as “reference layer reference picture refIvRefPic” for convenience of explanation) used for residual prediction. It is determined whether or not (refer) is available. When the reference layer reference picture refIvRefPic is available, the reference image determination unit 3095 sets 1 to the reference layer reference picture use flag refIvRefPicAvailable (refIvRefPicAvailable2). On the other hand, when the reference layer reference picture refIvRefPic cannot be used, the reference image determination unit 3095 sets the reference layer reference picture use flag refIvRefPicAvailable (refIvRefPicAvailable2) to 0. The reference image determination unit 3095 outputs the reference layer reference picture use flag refIvRefPicAvailable (refIvRefPicAvailable2) set in this way to the residual prediction unit 3092.
 参照画像判定部3095は、参照ピクチャの利用可否を以下のとおり判定してもよい。なお、以下では、対象レイヤにおける参照ピクチャをARP参照ピクチャarpRefPicと称する。ARP参照ピクチャarpRefPicは、上記対象ピクチャのピクチャ順序と異なるピクチャ順序を有する。また、参照ピクチャリストRefPicListX(X=0,1)のインデックスarpRefIdxLXで参照されるピクチャ(RefPicListX[arpRefIdxLX ])がARP参照ピクチャarpRefPicであるものとする。また、ビュー識別子を示すインデックスをViewIdx、対象ブロックの座標を(xP、yP)における参照レイヤ(ビュー)のビュー識別子をrefViewIdx[xP][yP]と記述する。また、PicOrderCnt(X)は、ピクチャXのピクチャ順序番号POCを示す。インデックスarpRefIdxLXは、0を設定しても良いし、後述する第4の実施形態のオプションY5に示す方法を用いても良い。 The reference image determination unit 3095 may determine whether or not the reference picture can be used as follows. Hereinafter, the reference picture in the target layer is referred to as an ARP reference picture arpRefPic. The ARP reference picture arpRefPic has a picture order different from the picture order of the target picture. Also, it is assumed that the picture (RefPicListX [arpRefIdxLX]) referenced by the index arpRefIdxLX of the reference picture list RefPicListX (X = 0, 1) is the ARP reference picture arpRefPic. Also, an index indicating the view identifier is described as ViewIdx, and the view identifier of the reference layer (view) in the coordinates of the target block (xP, yP) is described as refViewIdx [xP] [yP]. Also, PicOrderCnt (X) indicates the picture order number POC of picture X. The index arpRefIdxLX may be set to 0, or the method shown in option Y5 of the fourth embodiment described later may be used.
 (オプション1):参照ピクチャがRPSに存在するか否かを判定する
 以下の(1-1)および(1-2)を満たす参照ピクチャPicが存在する場合、参照レイヤ参照ピクチャrefIvRefPicに参照ピクチャPicを設定し、参照レイヤ参照ピクチャ利用フラグrefIvRefPicAvailableに1を設定する。そうでなければ、参照レイヤ参照ピクチャ利用フラグrefIvRefPicAvailableに0を設定する。
(Option 1): Determine whether or not the reference picture exists in the RPS. If there is a reference picture Pic that satisfies the following (1-1) and (1-2), the reference picture Pic is included in the reference layer reference picture refIvRefPic. And the reference layer reference picture usage flag refIvRefPicAvailable is set to 1. Otherwise, the reference layer reference picture usage flag refIvRefPicAvailable is set to 0.
 (1-1)PicOrderCnt(Pic)が、PicOrderCnt(arpRefPic)であり、かつ、PicのViewIdxが refViewIdx[xP][yP]となる。 (1-1) PicOrderCnt (Pic) is PicOrderCnt (arpRefPic), and ViewIdx of Pic is refViewIdx [xP] [yP].
 (1-2)上記(1-1)の条件を満たす参照ピクチャPicが、対象ピクチャと同じPOCであり、かつ、ViewIdxが refViewIdx[xP][yP]である参照レイヤ上のピクチャのRPSに存在する。 (1-2) The reference picture Pic that satisfies the condition (1-1) is present in the RPS of the picture on the reference layer whose POC is the same as the target picture and whose ViewIdx is refViewIdx [xP] [yP] To do.
 RPSは、あるピクチャに含まれる複数のスライス間で不変であり、スライスタイプがIスライスの場合にも存在することから、RPSによる判定は適当である。 Since RPS is invariant among a plurality of slices included in a picture and exists even when the slice type is I slice, determination by RPS is appropriate.
 (オプション2):参照ピクチャがRPLに存在するか否かを判定する
 上記(1-1)および下記(2-2)を満たす参照ピクチャPicが存在する場合、参照レイヤ参照ピクチャrefIvRefPicとして参照ピクチャPicを採用し、参照レイヤ参照ピクチャ利用フラグrefIvRefPicAvailableに1を設定する。そうでなければ、参照レイヤ参照ピクチャ利用フラグrefIvRefPicAvailableに0を設定する。
(Option 2): Determine whether or not a reference picture exists in the RPL. If there is a reference picture Pic that satisfies the above (1-1) and the following (2-2), the reference picture Pic is used as the reference layer reference picture refIvRefPic. And the reference layer reference picture usage flag refIvRefPicAvailable is set to 1. Otherwise, the reference layer reference picture usage flag refIvRefPicAvailable is set to 0.
 (2-2)上記(1-1)の条件を満たす参照ピクチャPicが、対象ピクチャと同じPOCであり、かつ、ViewIdxが refViewIdx[xP][yP]である、参照レイヤ上の、スライスタイプ(slice_type)がIスライスでないピクチャの参照レイヤリストRPL(RefPicListX)に存在する。 (2-2) The reference picture Pic that satisfies the above condition (1-1) is the same POC as the target picture, and the slice type on the reference layer (ViewIdx is refViewIdx [xP] [yP]) slice_type) exists in the reference layer list RPL (RefPicListX) of a picture that is not an I slice.
 RPLは多くの場合、簡単に導出することができる。なお、RPLの長さ、および/または、内容は、スライス毎に異なるため、以上に説明したように、判定には第1番目のスライスのRefPicListX[]を用いることができる。しかしながら、これに限られず、対象ピクチャに含まれるスライスのうち、所定番目のスライスのRefPicListX[]を、対象ピクチャに含まれるスライス間で用いるように構成することもできる。例えば、最後のスライスを用いることも可能である。上記判定において、参照ピクチャリストに存在するとは、参照ピクチャリストの特定の順番の要素に存在するのではなく、順序によらず存在すれば良い。また、参照ピクチャリストとしてL0リストRefPicList0[]とL1リストRefPicList1[]が有る場合には、何れかのリストの要素として存在すれば良い。 RPL can be easily derived in many cases. Note that since the length and / or content of the RPL is different for each slice, as described above, the RefPicListX [] of the first slice can be used for the determination. However, the present invention is not limited to this, and the RefPicListX [] of a predetermined slice among the slices included in the target picture may be used between the slices included in the target picture. For example, the last slice can be used. In the above determination, the presence in the reference picture list does not exist in a specific order element of the reference picture list, but may exist in any order. Further, when there are the L0 list RefPicList0 [] and the L1 list RefPicList1 [] as reference picture lists, they may exist as elements of any list.
 (オプション3a):参照ピクチャがDPBに存在するか否かを判定する
 上記(1-1)および下記(3-1)を満たす参照ピクチャPicが存在する場合、参照レイヤ参照ピクチャrefIvRefPicとして参照ピクチャPicを採用し、参照レイヤ参照ピクチャ利用フラグrefIvRefPicAvailableに1を設定する。そうでなければ、参照レイヤ参照ピクチャ利用フラグrefIvRefPicAvailableに0を設定する。
(Option 3a): Determine whether or not a reference picture exists in the DPB. If there is a reference picture Pic that satisfies the above (1-1) and the following (3-1), the reference picture Pic is used as the reference layer reference picture refIvRefPic. And the reference layer reference picture usage flag refIvRefPicAvailable is set to 1. Otherwise, the reference layer reference picture usage flag refIvRefPicAvailable is set to 0.
 (3-1)上記(1-1)の条件を満たす参照ピクチャPicが、対象ピクチャと同じPOCであり、かつ、ViewIdxが refViewIdx[xP][yP]である、デコードピクチャバッファ(DPB)に存在する。 (3-1) A reference picture Pic that satisfies the above condition (1-1) exists in the decoded picture buffer (DPB) having the same POC as that of the target picture and ViewIdx of refViewIdx [xP] [yP]. To do.
 DPB3061は、利用可能なピクチャを記憶しているため、以上のようにDPBを判定に用いてもよい。 Since DPB 3061 stores usable pictures, DPB may be used for determination as described above.
 (オプション3b):参照ピクチャがDPBに存在し、さらに、参照使用(used for reference)にマークされているか否かを判定する
 上記(1-1)および以下の(3b-2)を満たす参照ピクチャPicが存在する場合、参照レイヤ参照ピクチャrefIvRefPicとして参照ピクチャPicを採用し、参照レイヤ参照レイヤ参照ピクチャ利用フラグrefIvRefPicAvailableに1を設定する。そうでなければ、参照レイヤ参照レイヤ参照ピクチャ利用フラグrefIvRefPicAvailableに0を設定する。
(Option 3b): It is determined whether or not a reference picture exists in the DPB and is marked as used for reference. Reference picture satisfying the above (1-1) and the following (3b-2) When Pic exists, the reference picture Pic is adopted as the reference layer reference picture refIvRefPic, and 1 is set to the reference layer reference layer reference picture use flag refIvRefPicAvailable. Otherwise, the reference layer reference layer reference picture usage flag refIvRefPicAvailable is set to 0.
 (3b-2)上記(1-1)の条件を満たす参照ピクチャPicが、対象ピクチャと同じPOCであり、かつ、ViewIdxが refViewIdx[xP][yP]である、かつ、参照使用(used forreference)とマークされているピクチャが、デコードピクチャバッファ(DPB)に存在する。 (3b-2) The reference picture Pic that satisfies the above condition (1-1) is the same POC as the target picture, the ViewIdx is refViewIdx [xP] [yP], and the reference use (used forreference) The picture marked with is present in the decoded picture buffer (DPB).
 DPBの動作はデコーダ依存であるため、DPBにおけるピクチャの記憶状態が信頼できない場合がある。具体的には、DPBは出力必要(need for output)とマークされておらず、参照使用とマークされているピクチャを破棄するが、出力するタイミングは、デコーダ依存である場合があるため、あるデコーダではピクチャが破棄されているが、他のデコーダではピクチャが破棄されていないということがおきうる。この場合、(ARPに利用される)ある参照ピクチャがDPBに存在するか否かの判定がデコーダにより異なる可能性がある。参照使用のマークは、明示的に復号されるRPSに応じて設定されるため、デコーダ間で異なることがない。このため、以上に示したように、参照ピクチャの参照マークが「参照使用」であるか否かを判定することが好ましい。 Since the DPB operation is decoder-dependent, the picture storage state in the DPB may not be reliable. Specifically, the DPB is not marked as need-for-output and discards the picture that is marked for reference use, but the output timing may be decoder-dependent, so some decoders In this case, the picture is discarded, but it can be said that the picture is not discarded in other decoders. In this case, the determination of whether a reference picture (used for ARP) exists in the DPB may differ depending on the decoder. Since the mark for reference use is set according to the RPS that is explicitly decoded, there is no difference between the decoders. For this reason, as described above, it is preferable to determine whether or not the reference mark of the reference picture is “use of reference”.
 なお、上記オプション1からオプション3bにおいて、(1-1)の設定の代わりに、以下の(1-1b)を用いても良い。
(1-1b)PicOrderCnt(Pic)が、PicOrderCnt(arpRefPic)であり、かつ、PicのViewIdxが refViewIdx[xP][yP]、かつ、PicのDepthFlagが0である。
In the above options 1 to 3b, the following (1-1b) may be used instead of the setting (1-1).
(1-1b) PicOrderCnt (Pic) is PicOrderCnt (arpRefPic), Pic's ViewIdx is refViewIdx [xP] [yP], and Pic's DepthFlag is 0.
 (1-1b)によれば、参照ピクチャが、デプスピクチャである場合(DepthFlagが1である場合)に、目的とする参照ピクチャPicであるとして設定することがない。ARPは、テクスチャピクチャ(DepthFlag=0)を参照ピクチャとして使用されることで機能し、デプスピクチャ(DepthFlag=1)を参照ピクチャとして使用する場合には機能しない(残差予測により適当な予測画像が生成できない)ことから、上記、(1-1b)の判定は、残差予測の効果を生じさせる効果を奏する。つまり、RPS(オプション1の場合)、RPL(オプション2の場合)、DPB(オプション3a、オプション3b)の場合で、それらから参照ピクチャPicを設定する場合に、誤ってデプスピクチャを設定して残差予測の効果が得られない課題を防ぐ効果を奏する。 According to (1-1b), when the reference picture is a depth picture (when DepthFlag is 1), it is not set as the target reference picture Pic. ARP functions when a texture picture (DepthFlag = 0) is used as a reference picture, and does not function when a depth picture (DepthFlag = 1) is used as a reference picture. Therefore, the determination of (1-1b) has an effect of causing the effect of residual prediction. That is, in the case of RPS (option 1), RPL (option 2), and DPB (option 3a, option 3b), when setting the reference picture Pic from them, the depth picture is set by mistake and left. There is an effect of preventing a problem that the effect of the difference prediction cannot be obtained.
 残差予測部3092は、残差予測実施フラグresPredFlagが1の場合に、入力された動き変位補償画像predSamplesLXに対して、残差予測(ARP:Advanced Residual Prediction)を行う。残差予測実施フラグresPredFlagが0の場合には、入力された動き変位補償画像predSamplesLXをそのまま出力する。 The residual prediction unit 3092 performs the residual prediction (ARP: Advanced Residual Prediction) on the input motion displacement compensation image predSamplesLX when the residual prediction execution flag resPredFlag is 1. When the residual prediction execution flag resPredFlag is 0, the input motion displacement compensation image predSamplesLX is output as it is.
 残差予測は、予測画像生成の対象とする対象レイヤ(第2のレイヤ画像)とは異なる参照レイヤ(第1のレイヤ画像)の残差を、対象レイヤの参照ピクチャから予測した画像(動き補償画像)である動き変位補償画像predSamplesLXに加えることにより行われる。これを、第1の残差予測と呼ぶ。また、残差予測は、予測画像生成の対象とする対象レイヤ(第2のレイヤ画像)と参照レイヤ(第1のレイヤ画像)間の対象ピクチャとは異なる時刻(POC)における残差を、参照レイヤの参照ピクチャから予測した画像(変位補償画像)である動き変位補償画像predSamplesLXに加えることにより行うこともできる。これを第2の残差予測と呼ぶ。 Residual prediction is an image (motion compensation) in which a residual of a reference layer (first layer image) different from a target layer (second layer image) that is a target of predicted image generation is predicted from a reference picture of the target layer. This is performed by adding to the motion displacement compensation image predSamplesLX. This is called the first residual prediction. In addition, residual prediction refers to a residual at a different time (POC) from a target picture between a target layer (second layer image) and a reference layer (first layer image) that are targets of predicted image generation. It can also be performed by adding to the motion displacement compensation image predSamplesLX which is an image (displacement compensation image) predicted from the reference picture of the layer. This is called second residual prediction.
 第1の残差予測では、参照レイヤと同様の残差が対象レイヤにも生じると仮定して、既に導出された参照レイヤの対象ピクチャと異なる時刻(POC)のピクチャrefIvRefPicと、参照レイヤの対象ピクチャと同じ時刻(POC)のピクチャcurrIvRefPic間の残差を対象レイヤの対象ピクチャと異なる時刻(POC)のピクチャ(例えばarpRefPic)からの動き補償予測における残差の推定値として用いる。 In the first residual prediction, assuming that a residual similar to the reference layer also occurs in the target layer, the picture refIvRefPic at a different time (POC) from the target picture of the already derived reference layer, and the reference layer target The residual between the pictures currIvRefPic at the same time (POC) as the picture is used as an estimate of the residual in motion compensation prediction from a picture (for example, arpRefPic) at a different time (POC) from the target picture of the target layer.
 第2の残差予測では、既に導出された対象ピクチャと異なる時刻(POC)のピクチャにおける参照レイヤのピクチャrefIvRefPicと、対象ピクチャのピクチャ間arpRefPicの残差を、対象ピクチャと同じ時刻(POC)における、参照レイヤのピクチャcurrIvRefPicからの変位補償予測における残差の推定値として用いる。 In the second residual prediction, the residuals of the reference layer picture refIvRefPic and the inter-picture arpRefPic of the picture at the time (POC) different from the already derived target picture are set at the same time (POC) as the target picture. This is used as an estimate of the residual in displacement compensation prediction from the reference layer picture currIvRefPic.
 以下の実施形態では、第1の残差予測と第2の残差予測の両方を用いるが、簡略化のために第1の残差予測と第2の残差予測の一方のみを用いる構成でも構わない。 In the following embodiments, both the first residual prediction and the second residual prediction are used. However, for simplification, only one of the first residual prediction and the second residual prediction is used. I do not care.
 図1は残差予測部3092の構成を示すブロック図である。残差予測部3092は、残差予測実施フラグ導出部30921と、参照画像取得部30922と、残差合成部30923から構成される。 FIG. 1 is a block diagram showing the configuration of the residual prediction unit 3092. The residual prediction unit 3092 includes a residual prediction execution flag deriving unit 30921, a reference image acquisition unit 30922, and a residual synthesis unit 30923.
 残差予測実施フラグ導出部30921は、(1)残差予測フラグiv_res_pred_weight_idxが0ではなく、かつ、(2)参照ピクチャ利用フラグが、“1”である場合に、残差予測実施フラグresPredFlagに残差予測を実行することを示す1を設定する。一方、残差予測フラグiv_res_pred_weight_idxが0である、または、参照ピクチャ利用フラグが、“1”でない場合(視差補償である場合)に、残差予測実施フラグresPredFlagに0を設定する。 The residual prediction execution flag deriving unit 30921 has a residual prediction execution flag resPredFlag when (1) the residual prediction flag iv_res_pred_weight_idx is not 0 and (2) the reference picture use flag is “1”. 1 is set to indicate that difference prediction is to be executed. On the other hand, when the residual prediction flag iv_res_pred_weight_idx is 0, or when the reference picture use flag is not “1” (in the case of disparity compensation), 0 is set to the residual prediction execution flag resPredFlag.
 すなわち、残差予測実施フラグ導出部30921は、以下の条件式(R-1)により、残差予測実施フラグresPredFlagを導出してもよい。 That is, the residual prediction execution flag deriving unit 30921 may derive the residual prediction execution flag resPredFlag according to the following conditional expression (R-1).
 resPredFlag = ( iv_res_pred_weight_idx != 0 ) && refIvRefPicAvailable ・・・(R-1)
 ここで、PicOrderCntValは、対象ピクチャのピクチャ順序番号POCである。
resPredFlag = (iv_res_pred_weight_idx! = 0) && refIvRefPicAvailable (R-1)
Here, PicOrderCntVal is the picture order number POC of the target picture.
 なお、動き予測の場合に限定する場合には、残差予測実施フラグ導出部30921は、上記(1)および(2)に加えて、(3)対象ブロックが動き補償であるか否かを判定してもよい( PicOrderCnt( RefPicListX[ refIdxLX ] ) != PicOrderCntVal )。すなわち、以下の条件式(R-2)により、残差予測実施フラグresPredFlagを導出してもよい。 In the case of limiting to motion prediction, the residual prediction execution flag deriving unit 30921 determines (3) whether the target block is motion compensation in addition to (1) and (2) above. (PicOrderCnt (RefPicListX [refIdxLX])! = PicOrderCntVal). That is, the residual prediction execution flag resPredFlag may be derived from the following conditional expression (R-2).
 resPredFlag = ( iv_res_pred_weight_idx != 0 )
 ( PicOrderCnt( RefPicListX[ refIdxLX ] ) != PicOrderCntVal ) && refIvRefPicAvail ・・・ (R-2)
 参照画像取得部30922は、残差予測実施フラグresPredFlagが1の場合には、インター予測パラメータ復号部303から入力された動きベクトルmvLXと残差予測変位ベクトルmvDisp、及び復号ピクチャ管理部306に格納された対応ブロックcurrIvSamplesLXと対応ブロックの参照ブロックrefIvSamplesLXを読み出す。
resPredFlag = (iv_res_pred_weight_idx! = 0)
(PicOrderCnt (RefPicListX [refIdxLX])! = PicOrderCntVal) && refIvRefPicAvail ・ ・ ・ (R-2)
When the residual prediction execution flag resPredFlag is 1, the reference image acquisition unit 30922 stores the motion vector mvLX and the residual prediction displacement vector mvDisp input from the inter prediction parameter decoding unit 303, and the decoded picture management unit 306. The corresponding block currIvSamplesLX and the reference block refIvSamplesLX of the corresponding block are read.
 図11の(a)は、対応ブロックcurrIvSamplesLXを説明するための図である。図11の(a)に示すように、対象レイヤ上の対象ブロックTar1に対応する対応ブロックCor1は、参照レイヤ上の画像の対象ブロックTar1´の位置を起点として、参照レイヤと対象レイヤの位置関係を示すベクトルである変位ベクトルmvDispだけずれた位置になるブロックに位置する。 FIG. 11A is a diagram for explaining the corresponding block currIvSamplesLX. As shown in FIG. 11A, the corresponding block Cor1 corresponding to the target block Tar1 on the target layer has a positional relationship between the reference layer and the target layer, starting from the position of the target block Tar1 ′ of the image on the reference layer. Is located in a block at a position shifted by a displacement vector mvDisp, which is a vector indicating.
 具体的には、参照画像取得部30922は、対象ブロックの画素の座標(x,y)を、対象ブロックの変位ベクトルmvDispだけずらした位置の画素を導出する。変位ベクトルmvDispが1/4ペルの小数精度であることを考慮し、残差生成部30922は、対象ブロックの画素の座標が(xP、yP)である場合に対応する整数精度の画素R0のX座標xIntとY座標yInt、及び変位ベクトルmvDispのX成分の小数部分xFracとY成分の小数部分yFracを
 xInt = xPb + ( mvLX[ 0 ] >> 2 )
 yInt = yPb + ( mvLX[ 1 ] >> 2 )
 xFrac = mvLX[ 0 ] & 3 
 yFrac = mvLX[ 1 ] & 3 
の式により導出する。ここで、X & 3は、Xの下位2ビットのみを取り出す数式である。
Specifically, the reference image acquisition unit 30922 derives a pixel at a position where the coordinates (x, y) of the pixel of the target block are shifted by the displacement vector mvDisp of the target block. Considering that the displacement vector mvDisp has a decimal precision of 1/4 pel, the residual generation unit 30922 uses the X of the pixel R0 with integer precision corresponding to the case where the coordinates of the pixel of the target block are (xP, yP). XInt = xPb + (mvLX [0] >> 2) The coordinates xInt and Y coordinates yInt, and the fractional part xFrac of the displacement vector mvDisp and the fractional part yFrac of the Y component
yInt = yPb + (mvLX [1] >> 2)
xFrac = mvLX [0] & 3
yFrac = mvLX [1] & 3
It is derived by the following formula. Here, X & 3 is a mathematical expression for extracting only the lower 2 bits of X.
 次に、参照画像取得部30922は、変位ベクトルmvDispが1/4ペルの小数精度であることを考慮し、補間画素predPartLX[ x ][ y ]を生成する。まず、整数画素A(xA,yB)、B(xB,yB)、C(xC,yC)及びD(xD,yD)の座標を
 xA = Clip3( 0, picWidthInSamples - 1, xInt )
 xB = Clip3( 0, picWidthInSamples - 1, xInt + 1 )
 xC = Clip3( 0, picWidthInSamples - 1, xInt )
 xD = Clip3( 0, picWidthInSamples - 1, xInt + 1 ) 
 yA = Clip3( 0, picHeightInSamples - 1, yInt )
 yB = Clip3( 0, picHeightInSamples - 1, yInt )
 yC = Clip3( 0, picHeightInSamples - 1, yInt + 1 ) 
 yD = Clip3( 0, picHeightInSamples - 1, yInt + 1 )
の式により導出する。ここで、整数画素Aは画素R0に対応した画素であり、整数画素B,C,Dはそれぞれ整数画素Aの右、下、右下に隣接する整数精度の画素であり、Clip3(x, y, z)は、zをx以上、y以下に制限(クリップ)する関数である。参照画像取得部30922は、各整数画素A、B、C、及びDに対応する参照画素refPicLX[ xA ][ yA ]、refPicLX[ xB ][ yB ]、refPicLX[ xC ][ yC ]、及びrefPicLX[ xD ][ yD ]を復号ピクチャ管理部306のDPB3061から読み出す。
Next, the reference image acquisition unit 30922 generates an interpolation pixel predPartLX [x] [y] in consideration of the fact that the displacement vector mvDisp has a pel resolution of 1/4 pel. First, the coordinates of integer pixels A (xA, yB), B (xB, yB), C (xC, yC) and D (xD, yD) are set to xA = Clip3 (0, picWidthInSamples-1, xInt)
xB = Clip3 (0, picWidthInSamples-1, xInt + 1)
xC = Clip3 (0, picWidthInSamples-1, xInt)
xD = Clip3 (0, picWidthInSamples-1, xInt + 1)
yA = Clip3 (0, picHeightInSamples-1, yInt)
yB = Clip3 (0, picHeightInSamples-1, yInt)
yC = Clip3 (0, picHeightInSamples-1, yInt + 1)
yD = Clip3 (0, picHeightInSamples-1, yInt + 1)
It is derived by the following formula. Here, the integer pixel A is a pixel corresponding to the pixel R0, and the integer pixels B, C, and D are integer precision pixels adjacent to the right, bottom, and bottom right of the integer pixel A, respectively, and Clip3 (x, y , z) is a function that limits (clips) z to be greater than or equal to x and less than or equal to y. The reference image acquisition unit 30922 includes reference pixels refPicLX [xA] [yA], refPicLX [xB] [yB], refPicLX [xC] [yC], and refPicLX [corresponding to the integer pixels A, B, C, and D, respectively. xD] [yD] is read from the DPB 3061 of the decoded picture management unit 306.
 そして、参照画像取得部30922は、参照画素refPicLX[ xA ][ yA ]、refPicLX[ xB][ yB ]、refPicLX[ xC ][ yC ]、refPicLX[ xD ][ yD ]と変位ベクトルmvDispのX成分の小数部分xFracとY成分の小数部分yFracを用いて、画素R0から変位ベクトルmvDispの小数部分だけずらした位置の画素である補間画素predPartLX[ x ][ y ]を導出する。具体的には、
 predPartLX[ x ][ y ] = (refPicLX[ xA ][ yA ] * ( 8 - xFrac ) * ( 8 - yFrac ) 
           + refPicLX[ xB ][ yB ] * ( 8 - yFrac ) * xFrac
            + refPicLX[ xC ][ yC ] * ( 8 - xFrac ) * yFrac
            + refPicLX[ xD ][ yD ] * xFrac * yFrac ) >> 6
の式により導出する。
Then, the reference image acquisition unit 30922 includes the reference pixel refPicLX [xA] [yA], refPicLX [xB] [yB], refPicLX [xC] [yC], refPicLX [xD] [yD] and the X component of the displacement vector mvDisp. An interpolated pixel predPartLX [x] [y], which is a pixel shifted by the decimal part of the displacement vector mvDisp from the pixel R0, is derived using the fractional part xFrac and the fractional part yFrac of the Y component. In particular,
predPartLX [x] [y] = (refPicLX [xA] [yA] * (8-xFrac) * (8-yFrac)
+ refPicLX [xB] [yB] * (8-yFrac) * xFrac
+ refPicLX [xC] [yC] * (8-xFrac) * yFrac
+ refPicLX [xD] [yD] * xFrac * yFrac) >> 6
It is derived by the following formula.
 参照画像取得部30922は、上記の補間画素導出処理を、対象ブロック内の各画素に対して行い、補間画素の集合を補間ブロックpredPartLXとする。参照画像取得部30922は、導出した補間ブロックpredPartLXを、対応ブロックcurrIvSamplesLXとして、残差合成部30923に出力する。 The reference image acquisition unit 30922 performs the above interpolation pixel derivation process on each pixel in the target block, and sets a set of interpolation pixels as an interpolation block predPartLX. The reference image acquisition unit 30922 outputs the derived interpolation block predPartLX to the residual synthesis unit 30923 as the corresponding block currIvSamplesLX.
 図11の(b)は、参照ブロックrefIvSamplesLXを説明するための図である。図11の(b)に示すように、参照レイヤ上の対応ブロックに対応する参照ブロックは、参照レイヤ上の参照画像の対応ブロックの位置を起点として、対象ブロックの動きベクトルmvLXだけずれた位置になるブロックに位置する。 FIG. 11B is a diagram for explaining the reference block refIvSamplesLX. As shown in FIG. 11B, the reference block corresponding to the corresponding block on the reference layer is shifted to the position shifted by the motion vector mvLX of the target block from the position of the corresponding block of the reference image on the reference layer. Located in the block.
 参照画像取得部30922は、対応ブロックcurrIvSamplesLXを導出した処理と、変位ベクトルmvDispをベクトル(mvDisp[ 0 ] + mvLX [ 0 ]、mvDisp[ 1 ] + mvLX [ 1 ])に置き換えている点を除いて、同様の処理を行うことで、対応ブロックrefIvSamplesLXを導出する。参照画像取得部30922は、対応ブロックrefIvSamplesLXを残差合成部30923に出力する。 The reference image acquisition unit 30922 except for the process of deriving the corresponding block currIvSamplesLX and replacing the displacement vector mvDisp with a vector (mvDisp [0] + mvLX [0], mvDisp [1] + mvLX [1]) The corresponding block refIvSamplesLX is derived by performing the same processing. The reference image acquisition unit 30922 outputs the corresponding block refIvSamplesLX to the residual synthesis unit 30923.
 残差合成部30923は、残差予測実施フラグresPredFlagが1の場合には、動き変位補償画像predSamplesLX、対応ブロックcurrIvSamplesLX、参照ブロックrefIvSamplesLX及び残差予測フラグiv_res_pred_weight_idxから、補正動き変位補償画像predSamplesLX´を導出する。補正動き変位補償画像predSamplesLX´は、
 predSamplesLX´ = predSamplesLX + ((currIvSamplesLX - refIvSamplesLX) >> (iv_res_pred_weight_idx - 1))
の式を用いて求める。残差合成部30923は、残差予測実施フラグresPredFlagが0の場合には、動き変位補償画像predSamplesLXをそのまま出力する。
When the residual prediction execution flag resPredFlag is 1, the residual synthesis unit 30923 derives a corrected motion displacement compensation image predSamplesLX ′ from the motion displacement compensation image predSamplesLX, the corresponding block currIvSamplesLX, the reference block refIvSamplesLX, and the residual prediction flag iv_res_pred_weight_idx. To do. The corrected motion displacement compensation image predSamplesLX´
predSamplesLX´ = predSamplesLX + ((currIvSamplesLX-refIvSamplesLX) >> (iv_res_pred_weight_idx-1))
It is calculated using the following formula. When the residual prediction execution flag resPredFlag is 0, the residual synthesis unit 30923 outputs the motion displacement compensation image predSamplesLX as it is.
 照度補償部3093は、照度補償フラグic_enable_flagが1の場合に、入力された動き変位補償画像predSamplesLXに対して、照度補償を行う。照度補償フラグic_enable_flagが0の場合には、入力された動き変位補償画像predSamplesLXをそのまま出力する。照度補償部3093に入力される動き変位補償画像predSamplesLXは、残差予測がオフの場合には、動き変位補償部3091の出力画像であり、残差予測がオンの場合には、残差予測部3092の出力画像である。 When the illumination compensation flag ic_enable_flag is 1, the illumination compensation unit 3093 performs illumination compensation on the input motion displacement compensation image predSamplesLX. When the illumination compensation flag ic_enable_flag is 0, the input motion displacement compensation image predSamplesLX is output as it is. The motion displacement compensation image predSamplesLX input to the illuminance compensation unit 3093 is an output image of the motion displacement compensation unit 3091 when the residual prediction is off, and the residual prediction unit when the residual prediction is on. 3092 is an output image.
 重み予測部3094は、入力される動き変位画像predSamplesLXに重み係数を乗算することにより予測ピクチャブロックP(予測画像)を生成する。入力される動き変位画像predSamplesLXは、残差予測、照度補償が行われる場合には、それらが施された画像である。参照リスト利用フラグの一方(predFlagL0もしくはpredFlagL1)が1の場合(単予測の場合)で、重み予測を用いない場合には入力された動き変位画像predSamplesLX(LXはL0もしくはL1)を画素ビット数に合わせる以下の式の処理を行う。 The weight prediction unit 3094 generates a predicted picture block P (predicted image) by multiplying the input motion displacement image predSamplesLX by a weight coefficient. The input motion displacement image predSamplesLX is an image on which residual prediction and illuminance compensation are performed. When one of the reference list usage flags (predFlagL0 or predFlagL1) is 1 (in the case of single prediction) and weight prediction is not used, the input motion displacement image predSamplesLX (LX is L0 or L1) is set to the number of pixel bits. The following formula is processed.
predSamples[ x ][ y ] = Clip3( 0, ( 1 << bitDepth ) - 1, ( predSamplesLX[ x ][ y ] + offset1 ) >> shift1 )
ここで、shift1=14-bitDepth、offset1=1<<(shift1-1)である。
predSamples [x] [y] = Clip3 (0, (1 << bitDepth)-1, (predSamplesLX [x] [y] + offset1) >> shift1)
Here, shift1 = 14-bitDepth, offset1 = 1 << (shift1-1).
 また、参照リスト利用フラグの両者(predFlagL0もしくはpredFlagL1)が1の場合(双予測の場合)で、重み予測を用いない場合には、入力された動き変位画像predSamplesL0、predSamplesL1を平均し画素ビット数に合わせる以下の式の処理を行う。 When both of the reference list usage flags (predFlagL0 or predFlagL1) are 1 (in the case of bi-prediction) and weight prediction is not used, the input motion displacement images predSamplesL0 and predSamplesL1 are averaged to obtain the number of pixel bits. The following formula is processed.
predSamples[ x ][ y ] = Clip3( 0, ( 1 << bitDepth ) - 1, ( predSamplesL0[ x ][ y ] + predSamplesL1[ x ][ y ] + offset2 ) >> shift2 )
ここで、shift2=15-bitDepth、offset2=1<<(shift2-1)である。
predSamples [x] [y] = Clip3 (0, (1 << bitDepth)-1, (predSamplesL0 [x] [y] + predSamplesL1 [x] [y] + offset2) >> shift2)
Here, shift2 = 15-bitDepth, offset2 = 1 << (shift2-1).
 さらに、単予測の場合で、重み予測を行う場合には、重み予測部3094は、重み予測係数w0とオフセットo0を導出し、以下の式の処理を行う。 Furthermore, in the case of single prediction, when performing weight prediction, the weight prediction unit 3094 derives the weight prediction coefficient w0 and the offset o0, and performs the following processing.
predSamples[ x ][ y ] = Clip3( 0, ( 1 << bitDepth ) - 1, ( (predSamplesLX[ x ][ y ] * w0 + 2log2WD - 1) >> log2WD ) + o0 )
 ここで、log2WDは所定のシフト量を示す変数である。
predSamples [x] [y] = Clip3 (0, (1 << bitDepth)-1, ((predSamplesLX [x] [y] * w0 + 2log2WD-1) >> log2WD) + o0)
Here, log2WD is a variable indicating a predetermined shift amount.
 さらに、双予測の場合で、重み予測を行う場合には、重み予測部3094は、重み予測係数w0、w1、o0、o1を導出し、以下の式の処理を行う。 Furthermore, in the case of bi-prediction, when weight prediction is performed, the weight prediction unit 3094 derives weight prediction coefficients w0, w1, o0, o1, and performs the processing of the following equation.
predSamples[ x ][ y ] = Clip3( 0, ( 1 << bitDepth ) - 1, ( predSamplesL0 [ x ][ y ] * w0 + predSamplesL1[ x ][ y ] * w1 +  ((o0 + o1 + 1) << log2WD) ) >> (log2WD + 1) )
  (変形例1) 上記(参照画像判定)では、動き補償レベルの条件(resPredFlag)を用いて、参照レイヤ参照ピクチャrefIvRefPicの利用可否を判定することについて説明した。しかしながら、これに限られず、パース段階の条件を用いてもよい。
predSamples [x] [y] = Clip3 (0, (1 << bitDepth)-1, (predSamplesL0 [x] [y] * w0 + predSamplesL1 [x] [y] * w1 + ((o0 + o1 + 1) << log2WD)) >> (log2WD + 1))
(Modification 1) In the above (reference image determination), it has been described that the use of the reference layer reference picture refIvRefPic is determined using the motion compensation level condition (resPredFlag). However, the present invention is not limited to this, and conditions in the parsing stage may be used.
 すなわち、参照画像判定部3095は、パース段階においてARP参照ピクチャarpRefPicの利用可否を判定し、当該判定結果arpRefPicAvailableをインター予測パラメータ復号制御部3031に供給してもよい。以下、パース段階におけるARP参照ピクチャarpRefPicの利用可否を示すフラグをarpRefPicAvailable、動き補償段階におけるARP参照ピクチャarpRefPicの利用可否を示すフラグをrefIvRefPicAvailableとして区別する。 That is, the reference image determination unit 3095 may determine whether or not the ARP reference picture arpRefPic can be used in the parsing stage, and supply the determination result arpRefPicAvailable to the inter prediction parameter decoding control unit 3031. Hereinafter, a flag indicating the availability of the ARP reference picture arpRefPic in the parsing stage is distinguished as arpRefPicAvailable, and a flag indicating the availability of the ARP reference picture arpRefPic in the motion compensation stage is distinguished as refIvRefPicAvailable.
 図14を用いて、より具体的に説明すると次のとおりである。図14のSYN11およびSYN12に示すように、CUレベルの残差予測フラグ(ARPフラグ)iv_res_pred_weight_idxの導出の際、参照画像判定部3095は、ARP参照ピクチャarpRefPicの利用可否を判定してもよい。 It will be described in more detail with reference to FIG. As shown in SYN11 and SYN12 in FIG. 14, when deriving a CU level residual prediction flag (ARP flag) iv_res_pred_weight_idx, the reference image determination unit 3095 may determine whether or not the ARP reference picture arpRefPic can be used.
 そして、CUレベルのパラメータ復号において、インタービュー残差予測が利用可能なレイヤであって(iv_res_pred_flag[nuh_layer_id]が0でない)、かつARP参照ピクチャarpRefPicが利用可能である(arpRefPicAvailableが0でない)場合、インター予測パラメータ復号制御部3031は、iv_res_pred_weight_idxを復号してもよい。 In the CU level parameter decoding, when the inter-view residual prediction is a usable layer (iv_res_pred_flag [nuh_layer_id] is not 0) and the ARP reference picture arpRefPic is available (arpRefPicAvailable is not 0), The inter prediction parameter decoding control unit 3031 may decode iv_res_pred_weight_idx.
 これにより、残差予測処理において、参照ピクチャが利用可能でない事態が発生することを抑制することができる。 Thereby, it is possible to suppress the occurrence of a situation in which the reference picture cannot be used in the residual prediction process.
 次に、本実施形態に係る画像符号化装置11の構成について説明する。図12は、本実施形態に係る画像符号化装置11の構成を示すブロック図である。画像符号化装置11は、予測画像生成部101、減算部102、DCT・量子化部103、エントロピー符号化部104、逆量子化・逆DCT部105、加算部106、復号ピクチャ管理部(参照画像記憶部、フレームメモリ)109、符号化パラメータ決定部110、予測パラメータ符号化部111、残差格納部313(残差記録部)を含んで構成される。予測パラメータ符号化部111は、インター予測パラメータ符号化部112及びイントラ予測パラメータ符号化部113を含んで構成される。 Next, the configuration of the image encoding device 11 according to the present embodiment will be described. FIG. 12 is a block diagram illustrating a configuration of the image encoding device 11 according to the present embodiment. The image encoding device 11 includes a predicted image generation unit 101, a subtraction unit 102, a DCT / quantization unit 103, an entropy encoding unit 104, an inverse quantization / inverse DCT unit 105, an addition unit 106, a decoded picture management unit (reference image) (Storage unit, frame memory) 109, encoding parameter determination unit 110, prediction parameter encoding unit 111, and residual storage unit 313 (residual recording unit). The prediction parameter encoding unit 111 includes an inter prediction parameter encoding unit 112 and an intra prediction parameter encoding unit 113.
 予測画像生成部101は、外部から入力されたレイヤ画像Tの視点毎の各ピクチャについて、そのピクチャを分割した領域であるブロック毎に予測ピクチャブロックPを生成する。ここで、予測画像生成部101は、予測パラメータ符号化部111から入力された予測パラメータに基づいて復号ピクチャ管理部109から参照ピクチャブロックを読み出す。予測パラメータ符号化部111から入力された予測パラメータとは、例えば、動きベクトル又は変位ベクトルである。予測画像生成部101は、符号化対象ブロックを起点として予測された動きベクトル又は変位ベクトルが示す位置にあるブロックの参照ピクチャブロックを読み出す。予測画像生成部101は、読み出した参照ピクチャブロックについて複数の予測方式のうちの1つの予測方式を用いて予測ピクチャブロックPを生成する。予測画像生成部101は、生成した予測ピクチャブロックPを減算部102に出力する。なお、予測画像生成部101は、既に説明した予測画像生成部308と同じ動作であるため予測ピクチャブロックPの生成の詳細は省略する。 The predicted image generation unit 101 generates a predicted picture block P for each block which is an area obtained by dividing the picture for each viewpoint of the layer image T input from the outside. Here, the predicted image generation unit 101 reads the reference picture block from the decoded picture management unit 109 based on the prediction parameter input from the prediction parameter encoding unit 111. The prediction parameter input from the prediction parameter encoding unit 111 is, for example, a motion vector or a displacement vector. The predicted image generation unit 101 reads the reference picture block of the block at the position indicated by the motion vector or the displacement vector predicted from the encoding target block. The prediction image generation unit 101 generates a prediction picture block P using one prediction method among a plurality of prediction methods for the read reference picture block. The predicted image generation unit 101 outputs the generated predicted picture block P to the subtraction unit 102. Note that since the predicted image generation unit 101 performs the same operation as the predicted image generation unit 308 already described, details of generation of the predicted picture block P are omitted.
 予測画像生成部101は、予測方式を選択するために、例えば、レイヤ画像に含まれるブロックの画素毎の信号値と予測ピクチャブロックPの対応する画素毎の信号値との差分に基づく誤差値を最小にする予測方式を選択する。予測方式を選択する方法は、これには限られない。 In order to select a prediction method, the predicted image generation unit 101, for example, calculates an error value based on a difference between a signal value for each pixel of a block included in the layer image and a signal value for each corresponding pixel of the predicted picture block P. Select the prediction method to minimize. The method for selecting the prediction method is not limited to this.
 符号化対象のピクチャがベースビューピクチャである場合には、複数の予測方式とは、イントラ予測、動き予測及びマージ予測である。動き予測とは、上述のインター予測のうち、表示時刻間の予測である。マージ予測とは、既に符号化されたブロックであって、符号化対象ブロックから予め定めた範囲内にあるブロックと同一の参照ピクチャブロック及び予測パラメータを用いる予測である。符号化対象のピクチャがノンベースビューピクチャである場合には、複数の予測方式とは、イントラ予測、動き予測、マージ予測、及び変位予測である。変位予測(視差予測)とは、上述のインター予測のうち、別レイヤ画像(別視点画像)間の予測である。さらに、動き予測、マージ予測、及び変位予測である。変位予測(視差予測)に対して、追加予測(残差予測および照度補償)を行う場合と行わない場合の予測がある。 When the picture to be encoded is a base view picture, the plurality of prediction methods are intra prediction, motion prediction, and merge prediction. Motion prediction is prediction between display times among the above-mentioned inter predictions. The merge prediction is a prediction that uses the same reference picture block and prediction parameter as a block that has already been encoded and is within a predetermined range from the encoding target block. When the picture to be encoded is a non-base view picture, the plurality of prediction methods are intra prediction, motion prediction, merge prediction, and displacement prediction. The displacement prediction (disparity prediction) is prediction between different layer images (different viewpoint images) in the above-described inter prediction. Furthermore, motion prediction, merge prediction, and displacement prediction. For displacement prediction (disparity prediction), there are predictions with and without additional prediction (residual prediction and illuminance compensation).
 予測画像生成部101は、イントラ予測を選択した場合、予測ピクチャブロックPを生成する際に用いたイントラ予測モードを示す予測モードpredModeを予測パラメータ符号化部111に出力する。 The prediction image generation unit 101 outputs a prediction mode predMode indicating the intra prediction mode used when generating the prediction picture block P to the prediction parameter encoding unit 111 when intra prediction is selected.
 予測画像生成部101は、動き予測を選択した場合、予測ピクチャブロックPを生成する際に用いた動きベクトルmvLXを復号ピクチャ管理部109に記憶し、インター予測パラメータ符号化部112に出力する。動きベクトルmvLXは、符号化対象ブロックの位置から予測ピクチャブロックPを生成する際の参照ピクチャブロックの位置までのベクトルを示す。動きベクトルmvLXを示す情報には、参照ピクチャを示す情報(例えば、参照ピクチャインデックスrefIdxLX、ピクチャ順序番号POC)を含み、予測パラメータを表すものであっても良い。また、予測画像生成部101は、インター予測モードを示す予測モードpredModeを予測パラメータ符号化部111に出力する。 When the motion picture prediction is selected, the prediction image generation unit 101 stores the motion vector mvLX used when generating the prediction picture block P in the decoded picture management unit 109 and outputs the motion vector mvLX to the inter prediction parameter encoding unit 112. The motion vector mvLX indicates a vector from the position of the encoding target block to the position of the reference picture block when the predicted picture block P is generated. The information indicating the motion vector mvLX may include information indicating a reference picture (for example, a reference picture index refIdxLX, a picture order number POC), and may represent a prediction parameter. Further, the predicted image generation unit 101 outputs a prediction mode predMode indicating the inter prediction mode to the prediction parameter encoding unit 111.
 予測画像生成部101は、変位予測を選択した場合、予測ピクチャブロックPを生成する際に用いた変位ベクトルを復号ピクチャ管理部109に記憶し、インター予測パラメータ符号化部112に出力する。変位ベクトルdvLXは、符号化対象ブロックの位置から予測ピクチャブロックPを生成する際の参照ピクチャブロックの位置までのベクトルを示す。変位ベクトルdvLXを示す情報には、参照ピクチャを示す情報(例えば、参照ピクチャインデックスrefIdxLX、ビューIDview_id)を含み、予測パラメータを表すものであっても良い。また、予測画像生成部101は、インター予測モードを示す予測モードpredModeを予測パラメータ符号化部111に出力する。 When the prediction image generation unit 101 selects displacement prediction, the prediction image generation unit 101 stores the displacement vector used in generating the prediction picture block P in the decoded picture management unit 109 and outputs it to the inter prediction parameter encoding unit 112. The displacement vector dvLX indicates a vector from the position of the encoding target block to the position of the reference picture block when the predicted picture block P is generated. The information indicating the displacement vector dvLX may include information indicating a reference picture (for example, reference picture index refIdxLX, view IDview_id) and may represent a prediction parameter. Further, the predicted image generation unit 101 outputs a prediction mode predMode indicating the inter prediction mode to the prediction parameter encoding unit 111.
 予測画像生成部101は、マージ予測を選択した場合、選択した参照ピクチャブロックを示すマージインデックスmerge_idxをインター予測パラメータ符号化部112に出力する。また、予測画像生成部101は、マージ予測モードを示す予測モードpredModeを予測パラメータ符号化部111に出力する。 When the prediction image generation unit 101 selects merge prediction, the prediction image generation unit 101 outputs a merge index merge_idx indicating the selected reference picture block to the inter prediction parameter encoding unit 112. Further, the predicted image generation unit 101 outputs a prediction mode predMode indicating the merge prediction mode to the prediction parameter encoding unit 111.
 上記の、動き予測、変位予測、マージ予測において、予測画像生成部101は、残差予測実施フラグresPredFlagが残差予測を行うことを示す場合には、既に説明したように予測画像生成部101に含まれる残差予測部3092において残差予測を行う。 In the above-described motion prediction, displacement prediction, and merge prediction, the prediction image generation unit 101, when the residual prediction execution flag resPredFlag indicates that the residual prediction is performed, as described above, the prediction image generation unit 101 Residual prediction is performed in the included residual prediction unit 3092.
 減算部102は、予測画像生成部101から入力された予測ピクチャブロックPの信号値を、外部から入力されたレイヤ画像Tの対応するブロックの信号値から画素毎に減算して、残差信号を生成する。減算部102は、生成した残差信号をDCT・量子化部103と符号化パラメータ決定部110に出力する。 The subtraction unit 102 subtracts the signal value of the prediction picture block P input from the prediction image generation unit 101 for each pixel from the signal value of the corresponding block of the layer image T input from the outside, and generates a residual signal. Generate. The subtraction unit 102 outputs the generated residual signal to the DCT / quantization unit 103 and the encoding parameter determination unit 110.
 DCT・量子化部103は、減算部102から入力された残差信号についてDCTを行い、DCT係数を算出する。DCT・量子化部103は、算出したDCT係数を量子化して量子化係数を求める。DCT・量子化部103は、求めた量子化係数をエントロピー符号化部104及び逆量子化・逆DCT部105に出力する。 The DCT / quantization unit 103 performs DCT on the residual signal input from the subtraction unit 102 and calculates a DCT coefficient. The DCT / quantization unit 103 quantizes the calculated DCT coefficient to obtain a quantization coefficient. The DCT / quantization unit 103 outputs the obtained quantization coefficient to the entropy encoding unit 104 and the inverse quantization / inverse DCT unit 105.
 エントロピー符号化部104には、DCT・量子化部103から量子化係数が入力され、符号化パラメータ決定部110から符号化パラメータが入力される。入力される符号化パラメータには、例えば、参照ピクチャインデックスrefIdxLX、ベクトルインデックスmvp_LX_idx、差分ベクトルmvdLX、予測モードpredMode、残差予測フラグiv_res_pred_weight_idx、及びマージインデックスmerge_idx等の符号がある。 The entropy coding unit 104 receives the quantization coefficient from the DCT / quantization unit 103 and the coding parameter from the coding parameter determination unit 110. Input coding parameters include codes such as a reference picture index refIdxLX, a vector index mvp_LX_idx, a difference vector mvdLX, a prediction mode predMode, a residual prediction flag iv_res_pred_weight_idx, and a merge index merge_idx.
 エントロピー符号化部104は、入力された量子化係数と符号化パラメータをエントロピー符号化して符号化ストリームTeを生成し、生成した符号化ストリームTeを外部に出力する。 The entropy encoding unit 104 generates an encoded stream Te by entropy encoding the input quantization coefficient and encoding parameter, and outputs the generated encoded stream Te to the outside.
 逆量子化・逆DCT部105は、DCT・量子化部103から入力された量子化係数を逆量子化してDCT係数を求める。逆量子化・逆DCT部105は、求めたDCT係数について逆DCTを行い、復号残差信号を算出する。逆量子化・逆DCT部105は、算出した復号残差信号を加算部106に出力する。 The inverse quantization / inverse DCT unit 105 inversely quantizes the quantization coefficient input from the DCT / quantization unit 103 to obtain a DCT coefficient. The inverse quantization / inverse DCT unit 105 performs inverse DCT on the obtained DCT coefficient to calculate a decoded residual signal. The inverse quantization / inverse DCT unit 105 outputs the calculated decoded residual signal to the addition unit 106.
 加算部106は、予測画像生成部101から入力された予測ピクチャブロックPの信号値と逆量子化・逆DCT部105から入力された復号残差信号の信号値を画素毎に加算して、参照ピクチャブロックを生成する。加算部106は、生成した参照ピクチャブロックを復号ピクチャ管理部109に記憶する。 The addition unit 106 adds the signal value of the predicted picture block P input from the predicted image generation unit 101 and the signal value of the decoded residual signal input from the inverse quantization / inverse DCT unit 105 for each pixel, and refers to them. Generate a picture block. The adding unit 106 stores the generated reference picture block in the decoded picture management unit 109.
 復号ピクチャ管理部109は、画像復号装置31の復号ピクチャ管理部306と同様、予測パラメータメモリ(不図示)を有しており、当該予測パラメータメモリにおいて、予測パラメータ符号化部111が生成した予測パラメータを、符号化対象のピクチャ及びブロック毎に予め定めた位置に記憶する。 Similar to the decoded picture management unit 306 of the image decoding device 31, the decoded picture management unit 109 has a prediction parameter memory (not shown), and the prediction parameter generated by the prediction parameter encoding unit 111 in the prediction parameter memory. Are stored in a predetermined position for each picture and block to be encoded.
 また、復号ピクチャ管理部109は、画像復号装置31の復号ピクチャ管理部306と同様、DPB(不図示)を有しており、当該DPBにおいて、加算部106が生成した参照ピクチャブロックを、符号化対象のピクチャ及びブロック毎に予め定めた位置に記憶する。 Similarly to the decoded picture management unit 306 of the image decoding device 31, the decoded picture management unit 109 has a DPB (not shown), and encodes the reference picture block generated by the addition unit 106 in the DPB. Store in a predetermined position for each target picture and block.
 なお、復号ピクチャ管理部109の詳細については、画像復号装置31の復号ピクチャ管理部306について説明したものと同様であるので、ここではその説明は省略する。 Note that the details of the decoded picture management unit 109 are the same as those described for the decoded picture management unit 306 of the image decoding device 31, and thus the description thereof is omitted here.
 符号化パラメータ決定部110は、符号化パラメータの複数のセットのうち、1つのセットを選択する。符号化パラメータとは、上述した予測パラメータやこの予測パラメータに関連して生成される符号化の対象となるパラメータである。予測画像生成部101は、これらの符号化パラメータのセットの各々を用いて予測ピクチャブロックPを生成する。 The encoding parameter determination unit 110 selects one set from among a plurality of sets of encoding parameters. The encoding parameter is a parameter to be encoded that is generated in association with the above-described prediction parameter or the prediction parameter. The predicted image generation unit 101 generates a predicted picture block P using each of these sets of encoding parameters.
 符号化パラメータ決定部110は、複数のセットの各々について情報量の大きさと符号化誤差を示すコスト値を算出する。コスト値は、例えば、符号量と二乗誤差に係数λを乗じた値との和である。符号量は、量子化誤差と符号化パラメータをエントロピー符号化して得られる符号化ストリームTeの情報量である。二乗誤差は、減算部102において算出された残差信号の残差値の二乗値についての画素間の総和である。係数λは、予め設定されたゼロよりも大きい実数である。符号化パラメータ決定部110は、算出したコスト値が最小となる符号化パラメータのセットを選択する。これにより、エントロピー符号化部104は、選択した符号化パラメータのセットを符号化ストリームTeとして外部に出力し、選択されなかった符号化パラメータのセットを出力しない。 The encoding parameter determination unit 110 calculates a cost value indicating the amount of information and the encoding error for each of a plurality of sets. The cost value is, for example, the sum of a code amount and a square error multiplied by a coefficient λ. The code amount is the information amount of the encoded stream Te obtained by entropy encoding the quantization error and the encoding parameter. The square error is the sum between pixels regarding the square value of the residual value of the residual signal calculated by the subtracting unit 102. The coefficient λ is a real number larger than a preset zero. The encoding parameter determination unit 110 selects a set of encoding parameters that minimizes the calculated cost value. As a result, the entropy encoding unit 104 outputs the selected set of encoding parameters to the outside as the encoded stream Te, and does not output the set of unselected encoding parameters.
 予測パラメータ符号化部111は、予測画像生成部101から入力されたパラメータに基づいて予測ピクチャを生成する際に用いる予測パラメータを導出し、導出した予測パラメータを符号化して符号化パラメータのセットを生成する。予測パラメータ符号化部111は、生成した符号化パラメータのセットをエントロピー符号化部104に出力する。 The prediction parameter encoding unit 111 derives a prediction parameter used when generating a prediction picture based on the parameter input from the prediction image generation unit 101, and encodes the derived prediction parameter to generate a set of encoding parameters. To do. The prediction parameter encoding unit 111 outputs the generated set of encoding parameters to the entropy encoding unit 104.
 予測パラメータ符号化部111は、生成した符号化パラメータのセットのうち符号化パラメータ決定部110が選択したものに対応する予測パラメータを復号ピクチャ管理部109に記憶する。 The prediction parameter encoding unit 111 stores, in the decoded picture management unit 109, a prediction parameter corresponding to the generated encoding parameter set selected by the encoding parameter determination unit 110.
 予測パラメータ符号化部111は、予測画像生成部101から入力された予測モードpredModeがインター予測モードを示す場合、インター予測パラメータ符号化部112を動作させる。予測パラメータ符号化部111は、予測モードpredModeがイントラ予測モードを示す場合、イントラ予測パラメータ符号化部113を動作させる。 The prediction parameter encoding unit 111 operates the inter prediction parameter encoding unit 112 when the prediction mode predMode input from the prediction image generation unit 101 indicates the inter prediction mode. The prediction parameter encoding unit 111 operates the intra prediction parameter encoding unit 113 when the prediction mode predMode indicates the intra prediction mode.
 インター予測パラメータ符号化部112は、符号化パラメータ決定部110から入力された予測パラメータに基づいてインター予測パラメータを導出する。インター予測パラメータ符号化部112は、インター予測パラメータを導出する構成として、インター予測パラメータ復号部303(図6等、参照)がインター予測パラメータを導出する構成と同一の構成を含む。インター予測パラメータ符号化部112の構成については、後述する。 The inter prediction parameter encoding unit 112 derives an inter prediction parameter based on the prediction parameter input from the encoding parameter determination unit 110. The inter prediction parameter encoding unit 112 includes the same configuration as the configuration in which the inter prediction parameter decoding unit 303 (see FIG. 6 and the like) derives the inter prediction parameter as a configuration for deriving the inter prediction parameter. The configuration of the inter prediction parameter encoding unit 112 will be described later.
 イントラ予測パラメータ符号化部113は、符号化パラメータ決定部110から入力された予測モードpredModeが示すイントラ予測モードIntraPredModeをインター予測パラメータのセットとして定める。 The intra prediction parameter encoding unit 113 determines the intra prediction mode IntraPredMode indicated by the prediction mode predMode input from the encoding parameter determination unit 110 as a set of inter prediction parameters.
 次に、インター予測パラメータ符号化部112の構成について説明する。インター予測パラメータ符号化部112は、インター予測パラメータ復号部303に対応する手段である。 Next, the configuration of the inter prediction parameter encoding unit 112 will be described. The inter prediction parameter encoding unit 112 is means corresponding to the inter prediction parameter decoding unit 303.
 図13は、本実施形態に係るインター予測パラメータ符号化部112の構成を示す概略図である。 FIG. 13 is a schematic diagram illustrating a configuration of the inter prediction parameter encoding unit 112 according to the present embodiment.
 インター予測パラメータ符号化部112は、インター予測パラメータ符号化制御部1031、マージ予測パラメータ導出部1121、AMVP予測パラメータ導出部1122、減算部1123、及び予測パラメータ統合部1126を含んで構成される。 The inter prediction parameter encoding unit 112 includes an inter prediction parameter encoding control unit 1031, a merge prediction parameter derivation unit 1121, an AMVP prediction parameter derivation unit 1122, a subtraction unit 1123, and a prediction parameter integration unit 1126.
 マージ予測パラメータ導出部1121は、上述のマージ予測パラメータ導出部3036(図7参照)と同様な構成を有する。 The merge prediction parameter derivation unit 1121 has the same configuration as the merge prediction parameter derivation unit 3036 (see FIG. 7).
 インター予測パラメータ符号化制御部1031は、インター予測に関連する符号(シンタックス要素の復号をエントロピー符号化部104に指示し、符号化データに含まれる符号(シンタックス要素)を例えば、分割モードpart_mode、マージフラグmerge_flag、マージインデックスmerge_idx、インター予測フラグinter_pred_idcinter_pred_idc、参照ピクチャインデックスrefIdxLX、予測ベクトルインデックスmvp_LX_idx、差分ベクトルmvdLXを符号化する。 The inter prediction parameter encoding control unit 1031 instructs the entropy encoding unit 104 to decode a code related to the inter prediction (syntax element decoding), for example, a code (syntax element) included in the encoded data. , Merge flag merge_flag, merge index merge_idx, inter prediction flag inter_pred_idcinter_pred_idc, reference picture index refIdxLX, prediction vector index mvp_LX_idx, and difference vector mvdLX are encoded.
 インター予測パラメータ符号化制御部1031は、追加予測フラグ符号化部10311、マージインデックス符号化部10312、ベクトル候補インデックス符号化部10313、及び図示しない、分割モード符号化部、マージフラグ符号化部、インター予測フラグ符号化部、参照ピクチャインデックス符号化部、ベクトル差分符号化部を含んで構成される。分割モード符号化部、マージフラグ符号化部、マージインデックス符号化部、インター予測フラグ符号化部、参照ピクチャインデックス符号化部、ベクトル候補インデックス符号化部10313、ベクトル差分符号化部は各々、分割モードpart_mode、マージフラグmerge_flag、マージインデックスmerge_idx、インター予測フラグinter_pred_idcinter_pred_idc、参照ピクチャインデックスrefIdxLX、予測ベクトルインデックスmvp_LX_idx、差分ベクトルmvdLXを符号化する。 The inter prediction parameter encoding control unit 1031 includes an additional prediction flag encoding unit 10311, a merge index encoding unit 10312, a vector candidate index encoding unit 10313, and a split mode encoding unit, a merge flag encoding unit, an inter not shown. A prediction flag encoding unit, a reference picture index encoding unit, and a vector difference encoding unit are configured. The division mode encoding unit, the merge flag encoding unit, the merge index encoding unit, the inter prediction flag encoding unit, the reference picture index encoding unit, the vector candidate index encoding unit 10313, and the vector difference encoding unit are respectively divided modes. Part_mode, merge flag merge_flag, merge index merge_idx, inter prediction flag inter_pred_idcinter_pred_idc, reference picture index refIdxLX, prediction vector index mvp_LX_idx, and difference vector mvdLX are encoded.
 追加予測フラグ符号化部1031は、追加予測が行われるか否かを示すために、照度補償フラグic_enable_flag、残差予測フラグiv_res_pred_weight_flsgを符号化する。 The additional prediction flag encoding unit 1031 encodes the illumination compensation flag ic_enable_flag and the residual prediction flag iv_res_pred_weight_flsg to indicate whether or not additional prediction is performed.
 マージ予測パラメータ導出部1121には、予測画像生成部101から入力された予測モードpredModeがマージ予測モードを示す場合、符号化パラメータ決定部110からマージインデックスmerge_idxが入力される。マージインデックスmerge_idxは、予測パラメータ統合部1126に出力される。マージ予測パラメータ導出部1121は、マージ候補のうちマージインデックスmerge_idxが示す参照ブロックの参照ピクチャインデックスrefIdxLX、ベクトルmvLXを復号ピクチャ管理部109から読み出す。マージ候補とは、符号化対象となる符号化対象ブロックから予め定めた範囲にある参照ブロック(例えば、符号化対象ブロックの左下端、左上端、右上端に接する参照ブロックのうち)であって、符号化処理が完了した参照ブロックである。 The merge index merge_idx is input from the encoding parameter determination unit 110 to the merge prediction parameter derivation unit 1121 when the prediction mode predMode input from the prediction image generation unit 101 indicates the merge prediction mode. The merge index merge_idx is output to the prediction parameter integration unit 1126. The merge prediction parameter deriving unit 1121 reads the reference picture index refIdxLX and vector mvLX of the reference block indicated by the merge index merge_idx from the merge candidates from the decoded picture management unit 109. The merge candidate is a reference block (for example, a reference block in contact with the lower left end, upper left end, and upper right end of the encoding target block) within a predetermined range from the encoding target block to be encoded, This is a reference block for which the encoding process has been completed.
 AMVP予測パラメータ導出部1122は、上述のAMVP予測パラメータ導出部3032(図7参照)と同様な構成を有する。 The AMVP prediction parameter derivation unit 1122 has the same configuration as the AMVP prediction parameter derivation unit 3032 (see FIG. 7).
 AMVP予測パラメータ導出部1122には、予測画像生成部101から入力された予測モードpredModeがインター予測モードを示す場合、符号化パラメータ決定部110からベクトルmvLXが入力される。AMVP予測パラメータ導出部1122は、入力されたベクトルmvLXに基づいて予測ベクトルmvpLXを導出する。AMVP予測パラメータ導出部1122は、導出した予測ベクトルmvpLXを減算部1123に出力する。なお、参照ピクチャインデックスrefIdx及びベクトルインデックスmvp_LX_idxは、予測パラメータ統合部1126に出力される。 The AMVP prediction parameter derivation unit 1122 receives the vector mvLX from the encoding parameter determination unit 110 when the prediction mode predMode input from the prediction image generation unit 101 indicates the inter prediction mode. The AMVP prediction parameter derivation unit 1122 derives a prediction vector mvpLX based on the input vector mvLX. The AMVP prediction parameter derivation unit 1122 outputs the derived prediction vector mvpLX to the subtraction unit 1123. Note that the reference picture index refIdx and the vector index mvp_LX_idx are output to the prediction parameter integration unit 1126.
 減算部1123は、符号化パラメータ決定部110から入力されたベクトルmvLXからAMVP予測パラメータ導出部1122から入力された予測ベクトルmvpLXを減算して差分ベクトルmvdLXを生成する。差分ベクトルmvdLXは予測パラメータ統合部1126に出力する。 The subtraction unit 1123 subtracts the prediction vector mvpLX input from the AMVP prediction parameter derivation unit 1122 from the vector mvLX input from the coding parameter determination unit 110 to generate a difference vector mvdLX. The difference vector mvdLX is output to the prediction parameter integration unit 1126.
 予測画像生成部101から入力された予測モードpredModeがマージ予測モードを示す場合には、予測パラメータ統合部1126は、符号化パラメータ決定部110から入力されたマージインデックスmerge_idxをエントロピー符号化部104に出力する。 When the prediction mode predMode input from the predicted image generation unit 101 indicates the merge prediction mode, the prediction parameter integration unit 1126 outputs the merge index merge_idx input from the encoding parameter determination unit 110 to the entropy encoding unit 104. To do.
 予測画像生成部101から入力された予測モードpredModeがインター予測モードを示す場合には、予測パラメータ統合部1126は、次の処理を行う。 When the prediction mode predMode input from the predicted image generation unit 101 indicates the inter prediction mode, the prediction parameter integration unit 1126 performs the following process.
 予測パラメータ統合部1126は、符号化パラメータ決定部110から入力された参照ピクチャインデックスrefIdxLX及びベクトルインデックスmvp_LX_idx、減算部1123から入力された差分ベクトルmvdLXを統合する。予測パラメータ統合部1126は、統合した符号をエントロピー符号化部104に出力する。 The prediction parameter integration unit 1126 integrates the reference picture index refIdxLX and the vector index mvp_LX_idx input from the encoding parameter determination unit 110 and the difference vector mvdLX input from the subtraction unit 1123. The prediction parameter integration unit 1126 outputs the integrated code to the entropy encoding unit 104.
 〔第2の実施形態〕
 以下、図面を参照しながら本発明の第2の実施形態について説明する。第2の実施形態では、参照レイヤ参照ピクチャrefIvRefPicの利用可否が判定できるようビットストリームに制約を課す例について説明する。以下では、ビットストリームが課されるべき制約のことをビットストリームコンフォーマンスと称する。
[Second Embodiment]
The second embodiment of the present invention will be described below with reference to the drawings. In the second embodiment, an example will be described in which a restriction is imposed on the bitstream so that the availability of the reference layer reference picture refIvRefPic can be determined. Hereinafter, the constraint on which the bitstream is to be imposed is referred to as bitstream conformance.
 (オプションB1):RPS制約によるビットストリームコンフォーマンス
 画像符号化装置11の予測パラメータ符号化部111は、スライスレベルでARPのオン/オフフラグを符号化してもよい。また、画像符号化装置11は、インター予測処理が実行されるとき、以下の条件(B1-1)および(B1-2)を満たすことをビットストリームコンフォーマンスとする。
(Option B1): Bitstream conformance due to RPS restriction The prediction parameter encoding unit 111 of the image encoding device 11 may encode an ARP on / off flag at the slice level. In addition, when the inter prediction process is performed, the image encoding device 11 satisfies the following conditions (B1-1) and (B1-2) as bitstream conformance.
 (B1-1)対象ピクチャのRPS において、ViewIdx = refViewIdx[xP][yP]となる参照ピクチャPicが存在する。 (B1-1) In the RPS of the target picture, there is a reference picture Pic with ViewIdx = refViewIdx [xP] [yP].
 (B1-2)PicOrderCnt(Pic)が、PicOrderCnt(arpRefPic)であり、かつ、ViewIdxが refViewIdx[xP][yP]となる。 (B1-2) PicOrderCnt (Pic) is PicOrderCnt (arpRefPic), and ViewIdx is refViewIdx [xP] [yP].
 画像復号装置31は、上記をビットストリームコンフォーマンスとする符号化データを復号する。画像復号装置31の復号対象とするビットストリームとして、ARP参照ピクチャに関する制約を課すことによって、ARP参照ピクチャが参照できないことにより復号処理が失敗することを防ぐことができる。 The image decoding device 31 decodes the encoded data having the above bitstream conformance. By imposing restrictions on the ARP reference picture as a bit stream to be decoded by the image decoding device 31, it is possible to prevent the decoding process from failing because the ARP reference picture cannot be referred to.
 また、画像復号装置31は、次のように構成しても良い。すなわち、画像復号装置31は、上記ビットストリームコンフォーマンスに従って生成された、符号化ストリームを受信し、スライスレベル(スライスヘッダ/セグメントヘッダ)で符号化されたARPのオン/オフフラグをインター予測パラメータ復号部303において復号する。また、復号したARPのオン/オフフラグに応じて、残差予測部3092において残差予測処理を実行する。スライスヘッダ単位でARPのオンオフを符号化することで、参照レイヤの参照ピクチャリストの構成において、上記ビットストリームコンフォーマンスを遵守できない場合には、ARPをオフとすることができる。 Further, the image decoding device 31 may be configured as follows. That is, the image decoding device 31 receives the encoded stream generated according to the bitstream conformance, and sets the ARP on / off flag encoded at the slice level (slice header / segment header) to the inter prediction parameter decoding unit. In 303, decryption is performed. In addition, the residual prediction unit 3092 executes a residual prediction process according to the decoded ARP on / off flag. By encoding the ARP on / off in units of slice headers, the ARP can be turned off if the bitstream conformance cannot be observed in the reference picture list configuration of the reference layer.
 (オプションB2):RPL制約によるビットストリームコンフォーマンス
 画像符号化装置11の予測パラメータ符号化部111は、スライスレベルでARPのオン/オフフラグを符号化してもよい。また、画像符号化装置11は、インター予測処理が実行されるとき、以下の(B2-1)および(B2-2)を満たすことをビットストリームコンフォーマンスとする。
(Option B2): Bitstream Conformance Based on RPL Constraint The prediction parameter encoding unit 111 of the image encoding device 11 may encode the ARP on / off flag at the slice level. In addition, when the inter prediction process is executed, the image encoding device 11 satisfies the following (B2-1) and (B2-2) as bitstream conformance.
 (B2-1)参照ピクチャPicが、ViewIdx = refViewIdx[xP][yP]であり、かつスライスタイプ(slice_type)が対象ピクチャのIスライスでない第1番目のスライスのRefPicListX[0]に存在する。 (B2-1) The reference picture Pic is ViewIdx = refViewIdx [xP] [yP], and the slice type (slice_type) exists in the RefPicListX [0] of the first slice that is not the I slice of the target picture.
 (B2-2)PicOrderCnt(Pic)が、PicOrderCnt(arpRefPic)であり、かつ、ViewIdxが refViewIdx[xP][yP]となる。 (B2-2) PicOrderCnt (Pic) is PicOrderCnt (arpRefPic) and ViewIdx is refViewIdx [xP] [yP].
 画像復号装置31は、上記をビットストリームコンフォーマンスとする符号化データを復号する。画像復号装置31の復号対象とするビットストリームとして、ARP参照ピクチャに関する制約を課すことによって、ARP参照ピクチャが参照できないことにより復号処理が失敗することを防ぐことができる。 The image decoding device 31 decodes the encoded data having the above bitstream conformance. By imposing restrictions on the ARP reference picture as a bit stream to be decoded by the image decoding device 31, it is possible to prevent the decoding process from failing because the ARP reference picture cannot be referred to.
 また、画像復号装置31は、次のように構成しても良い。すなわち、画像復号装置31は、上記ビットストリームコンフォーマンスに従って生成された、符号化ストリームを受信し、スライスレベルで符号化されたARPのオン/オフフラグをインター予測パラメータ復号部303において復号する。また、復号したARPのオン/オフフラグに応じて、残差予測部3092において残差予測処理を実行する。 Further, the image decoding device 31 may be configured as follows. That is, the image decoding device 31 receives the encoded stream generated according to the bit stream conformance, and the inter prediction parameter decoding unit 303 decodes the ARP on / off flag encoded at the slice level. In addition, the residual prediction unit 3092 executes a residual prediction process according to the decoded ARP on / off flag.
 なお、上記に限られず、(B2-1)では、対象ピクチャのIスライスでないスライスのうち、所定番目のスライスのRefPicListX[0]を用いるように構成することもできる。例えば、最後のスライスのRefPicListX[0]を用いることも可能である。また、上記判定では、参照ピクチャリストの0番目だけでなく、所定番目の位置の参照ピクチャを対象としてもよい。 Note that the present invention is not limited to the above, and in (B2-1), a RefPicListX [0] of a predetermined slice among slices that are not I slices of the target picture can be used. For example, RefPicListX [0] of the last slice can be used. Further, in the above determination, not only the 0th picture in the reference picture list but also a reference picture at a predetermined position may be targeted.
 (オプションB3):参照ピクチャがDPBに存在するか否かを判定する
 画像符号化装置11の予測パラメータ符号化部111は、スライスレベルでARPのオン/オフフラグを符号化する。また、画像符号化装置11は、インター予測処理が実行されるとき、以下の(B3-1)および(B3-2)を満たすことをビットストリームコンフォーマンスとする。
(Option B3): It is determined whether or not a reference picture exists in the DPB. The prediction parameter encoding unit 111 of the image encoding device 11 encodes an ARP on / off flag at the slice level. In addition, when the inter prediction process is executed, the image encoding device 11 satisfies the following (B3-1) and (B3-2) as bitstream conformance.
 (B3-1)参照ピクチャPicが、ViewIdx = refViewIdx[xP][yP]であり、PicOrderCnt(Pic)が、PicOrderCnt(arpRefPic)であり、かつ、DPBに存在する。 (B3-1) The reference picture Pic is ViewIdx = refViewIdx [xP] [yP], PicOrderCnt (Pic) is PicOrderCnt (arpRefPic), and exists in the DPB.
 (B3-2)参照ピクチャPicの参照マークは「参照使用」である。 (B3-2) The reference mark of the reference picture Pic is “use reference”.
 画像復号装置31は、上記をビットストリームコンフォーマンスとする符号化データを復号する。画像復号装置31の復号対象とするビットストリームとして、ARP参照ピクチャに関する制約を課すことによって、ARP参照ピクチャが参照できないことにより復号処理が失敗することを防ぐことができる。 The image decoding device 31 decodes the encoded data having the above bitstream conformance. By imposing restrictions on the ARP reference picture as a bit stream to be decoded by the image decoding device 31, it is possible to prevent the decoding process from failing because the ARP reference picture cannot be referred to.
 また、画像復号装置31は、次のように構成しても良い。すなわち、画像復号装置31は、上記ビットストリームコンフォーマンスに従って生成された、符号化ストリームを受信し、スライスレベルで符号化されたARPのオン/オフフラグをインター予測パラメータ復号部303において復号する。また、復号したARPのオン/オフフラグに応じて、残差予測部3092において残差予測処理を実行する。 Further, the image decoding device 31 may be configured as follows. That is, the image decoding device 31 receives the encoded stream generated according to the bit stream conformance, and the inter prediction parameter decoding unit 303 decodes the ARP on / off flag encoded at the slice level. In addition, the residual prediction unit 3092 executes a residual prediction process according to the decoded ARP on / off flag.
 〔第3の実施形態〕
 以下、第3の実施形態について説明する。ARP参照ピクチャarpRefPicが参照ピクチャリストRefPicListXのarpRefIdxLXの参照ピクチャRefPicListX[arpRefIdxLX]である場合において、参照ピクチャリストは、ピクチャによって異なる。第3の実施形態では、対象レイヤにおける参照ピクチャarpRefPicの参照ピクチャリストにおける位置が、スライス単位で変わるのを防ぐための構成について説明する。なお、arpRefIdxLX(X=0 or 1)は、ARP参照ピクチャarpRefPicを特定するためのインデックスである。
[Third Embodiment]
Hereinafter, a third embodiment will be described. When the ARP reference picture arpRefPic is the reference picture RefPicListX [arpRefIdxLX] of the arpRefIdxLX of the reference picture list RefPicListX, the reference picture list differs depending on the picture. In the third embodiment, a configuration for preventing the position in the reference picture list of the reference picture arpRefPic in the target layer from changing in units of slices will be described. Note that arpRefIdxLX (X = 0 or 1) is an index for specifying the ARP reference picture arpRefPic.
 ARP参照ピクチャに用いられる参照ピクチャ(arpRefPicおよび対応curIvRefPic)は、対象ピクチャに含まれるスライス間において同じになるように、以下の(オプションX1)または(オプションX1´)のようなビットストリームコンフォーマンスを導入することが好ましい。 The reference picture (arpRefPic and corresponding curIvRefPic) used for the ARP reference picture has the following bit stream conformance as (Option X1) or (Option X1 ′) so that it is the same between slices included in the target picture. It is preferable to introduce.
 (オプションX1)
 画像符号化装置11は、符号化されるピクチャのARPフラグ(残差予測フラグ)が、“ARP残差予測が可能である”ことを示す場合、以下の(X1-1)および(X1-2)を満たすことをビットストリームコンフォーマンスとする。
(Option X1)
When the ARP flag (residual prediction flag) of the picture to be encoded indicates that “ARP residual prediction is possible”, the image encoding device 11 performs the following (X1-1) and (X1-2) ) Is bitstream conformance.
 (X1-1) 符号化されるピクチャにおいて、L0の参照画像を用いてARPを行う場合にarpRefPicL0によって参照される参照ピクチャを、スライスタイプ(slice_type)が“P”である全スライスについて、同じものに統一する。 (X1-1) In the picture to be encoded, the reference picture referenced by arpRefPicL0 when performing ARP using the reference picture of L0 is the same for all slices whose slice type (slice_type) is “P” To unify.
 (X1-2) 符号化されるピクチャにおいて、L1の参照画像を用いてARPを行う場合にarpRefPicL1によって参照される参照ピクチャを、スライスタイプ(slice_type)が“P”である全スライスについて、同じものに統一する。 (X1-2) In the picture to be encoded, the reference picture referenced by arpRefPicL1 when performing ARP using the L1 reference picture is the same for all slices whose slice type (slice_type) is “P” To unify.
 画像復号装置31は、上記をビットストリームコンフォーマンスとする符号化データを復号する。画像復号装置31の復号対象とするビットストリームとして、ARP参照ピクチャに関する制約を課すことによって、ARP参照ピクチャがピクチャ内で変化することに伴うデータ転送上の処理量の増加(キャッシュが当たらない等)を防ぐことができる。 The image decoding device 31 decodes the encoded data having the above bitstream conformance. By imposing restrictions on the ARP reference picture as a bit stream to be decoded by the image decoding apparatus 31, an increase in the amount of data transfer due to the change of the ARP reference picture in the picture (such as no cache hit) Can be prevented.
 また、画像復号装置31は、次のように構成してもよい。すなわち、画像復号装置31は、上記ビットストリームコンフォーマンスに従って生成された、符号化ストリームを受信し、スライスレベルで符号化されたARPフラグをインター予測パラメータ復号部303において復号する。また、復号したARPフラグに応じて、残差予測部3092において残差予測処理を実行する。 Further, the image decoding device 31 may be configured as follows. That is, the image decoding apparatus 31 receives the encoded stream generated according to the bit stream conformance, and the inter prediction parameter decoding unit 303 decodes the ARP flag encoded at the slice level. Further, the residual prediction unit 3092 executes a residual prediction process according to the decoded ARP flag.
 (オプションX1´)
 画像符号化装置11は、符号化されるピクチャのARPフラグ(残差予測フラグ)が、“ARP残差予測が可能である”ことを示す場合、以下の(X1-1´)および(X1-2´)を満たすことをビットストリームコンフォーマンスとする。
(Option X1 ')
When the ARP flag (residual prediction flag) of the picture to be encoded indicates that “ARP residual prediction is possible”, the image encoding device 11 performs the following (X1-1 ′) and (X1- Satisfying 2 ′) is defined as bit stream conformance.
 (X1-1´) 符号化されるピクチャにおいて、RefPicList0[arpRefIdxL0]によって参照される参照ピクチャを、スライスタイプ(slice_type)が“P”である全スライスについて、同じものに統一する。 (X1-1 ′) In a picture to be encoded, the reference picture referred to by RefPicList0 [arpRefIdxL0] is unified for all slices whose slice type (slice_type) is “P”.
 (X1-2´) 符号化されるピクチャにおいて、RefPicList1[arpRefIdxL1]によって参照される参照ピクチャを、スライスタイプ(slice_type)が“B”である全スライスについて、同じものに統一する。 (X1-2 ′) In the picture to be encoded, the reference picture referenced by RefPicList1 [arpRefIdxL1] is unified to the same for all slices whose slice type (slice_type) is “B”.
 画像復号装置31は、上記をビットストリームコンフォーマンスとする符号化データを復号する。画像復号装置31の復号対象とするビットストリームとして、ARP参照ピクチャに関する制約を課すことによって、ARP参照ピクチャがピクチャ内で変化することに伴うデータ転送上の処理量の増加(キャッシュが当たらない等)を防ぐことができる。 The image decoding device 31 decodes the encoded data having the above bitstream conformance. By imposing restrictions on the ARP reference picture as a bit stream to be decoded by the image decoding apparatus 31, an increase in the amount of data transfer due to the change of the ARP reference picture in the picture (such as no cache hit) Can be prevented.
 また、画像復号装置31は、次のように構成しても良い。すなわち、画像復号装置31は、上記ビットストリームコンフォーマンスに従って生成された、符号化ストリームを受信し、スライスレベルで符号化されたARPフラグをインター予測パラメータ復号部303において復号する。また、復号したARPフラグに応じて、残差予測部3092において残差予測処理を実行する。 Further, the image decoding device 31 may be configured as follows. That is, the image decoding apparatus 31 receives the encoded stream generated according to the bit stream conformance, and the inter prediction parameter decoding unit 303 decodes the ARP flag encoded at the slice level. Further, the residual prediction unit 3092 executes a residual prediction process according to the decoded ARP flag.
 なお、ARP参照ピクチャarpRefPicが参照ピクチャリストRefPicListXの先頭の要素として導出される場合には、上記のビットストリームコンフォーマンスにおいてarpRefIdxL0=0、arpRefIdxL1=0とすれば良い。 When the ARP reference picture arpRefPic is derived as the head element of the reference picture list RefPicListX, arpRefIdxL0 = 0 and arpRefIdxL1 = 0 may be set in the above bitstream conformance.
 (オプションX2)
 ARP参照ピクチャarpRefPicを、ピクチャにおける、スライスタイプ(slice_type)が“I”でない第1番目のスライスのRefPicListX[0]に限定する。
(Option X2)
The ARP reference picture arpRefPic is limited to RefPicListX [0] of the first slice whose slice type (slice_type) is not “I” in the picture.
 例えば、残差予測時に、参照画像取得部30922において、ピクチャにおける、スライスタイプ(slice_type)が“I”でない第1番目のスライスのRefPicListX[0]からARP参照ピクチャarpRefPicを取得するようにすればよい。 For example, at the time of residual prediction, the reference image acquisition unit 30922 may acquire the ARP reference picture arpRefPic from RefPicListX [0] of the first slice whose slice type (slice_type) is not “I” in the picture. .
 なお、上記に限られず、対象ピクチャのIスライスでないスライスのうち、所定番目のスライスを用いるように構成することもできる。例えば、対象ピクチャに含まれるスライスの間で、最後のスライスのRefPicListX[0]を用いることも可能である。また、上記判定では、参照ピクチャリストの0番目だけでなく、所定番目の位置(arpRefIdxLX)の参照ピクチャ(すなわち、RefPicListX[arpRefIdxLX])を対象としてもよい。 It should be noted that the present invention is not limited to the above, and a predetermined slice among slices other than the I slice of the target picture can be used. For example, it is possible to use RefPicListX [0] of the last slice between slices included in the current picture. In the above determination, not only the 0th reference picture list but also the reference picture (that is, RefPicListX [arpRefIdxLX]) at the predetermined position (arpRefIdxLX) may be targeted.
 〔第4の実施形態〕
 以下、第4の実施形態について説明する。ARPでは対象ピクチャのPOCである対象ピクチャ順序とは異なるPOCである参照ピクチャ順序の参照ピクチャを利用して残差を予測する。この参照ピクチャ順序を、参照ピクチャリスト中の特定の参照ピクチャのPOC(例えば参照ピクチャリストRefPicListX[]の先頭の参照ピクチャRefPicListX[0]のPOC)として導出する場合、参照ピクチャ順序が、対象ピクチャと等しくなることがある。この場合には、ARPを利用することができない。第4の実施形態では、ARPにおいて、対象ピクチャと同じPOCを有するレイヤ間ピクチャが、対象ピクチャと異なるPOCを有する参照ピクチャとして、参照されることを防ぐために、ARPに使用するための参照ピクチャを特定する手法について説明する。
[Fourth Embodiment]
Hereinafter, a fourth embodiment will be described. In ARP, a residual is predicted using a reference picture in a reference picture order that is a POC different from the target picture order that is a POC of the target picture. When this reference picture order is derived as the POC of a specific reference picture in the reference picture list (for example, the POC of the first reference picture RefPicListX [0] of the reference picture list RefPicListX []), the reference picture order is the same as the target picture. May be equal. In this case, ARP cannot be used. In the fourth embodiment, in ARP, in order to prevent an inter-layer picture having the same POC as the target picture from being referred to as a reference picture having a POC different from the target picture, a reference picture to be used for the ARP is used. The method to identify is demonstrated.
 (オプションY1):
 オプションY1では、ARPに用いられる参照ピクチャに関するビットストリームコンフォーマンスについて説明する。ARP参照ピクチャarpRefPicが参照ピクチャリストRefPicListXのarpRefIdxLXの参照ピクチャRefPicListX[arpRefIdxLX]である場合において、画像符号化装置11は、以下の(Y1-1)および(Y1-2)を満たすことをビットストリームコンフォーマンスとする。
(Option Y1):
In option Y1, bitstream conformance regarding a reference picture used for ARP will be described. When the ARP reference picture arpRefPic is the reference picture RefPicListX [arpRefIdxLX] of the arpRefIdxLX of the reference picture list RefPicListX, the image encoding device 11 indicates that the following (Y1-1) and (Y1-2) are satisfied. Let's do it.
 (Y1-1) ARPが適用されるピクチャでは、DiffPicOrderCnt(RefPicList0[arpRefIdxL0], currPic)が0でなく、スライスタイプ(slice_type)がIでない。 (Y1-1) In a picture to which ARP is applied, DiffPicOrderCnt (RefPicList0 [arpRefIdxL0], currPic) is not 0 and the slice type (slice_type) is not I.
 (Y1-2) ARPが適用されるピクチャでは、DiffPicOrderCnt(RefPicList1[arpRefIdxL1], currPic)が0でなく、スライスタイプ(slice_type)がBである。 (Y1-2) In a picture to which ARP is applied, DiffPicOrderCnt (RefPicList1 [arpRefIdxL1], currPic) is not 0 and the slice type (slice_type) is B.
 なお、上記において「DiffPicOrderCnt(PicA、currPic)が0ではない」の式は、「PicOrderCnt(PicA) != PicOrderCntVal」に置き替え可能である(以下同様)。 In the above, the expression “DiffPicOrderCnt (PicA, currPic) is not 0” can be replaced with “PicOrderCnt (PicA)! = PicOrderCntVal” (the same applies hereinafter).
 画像復号装置31は、上記ビットストリームコンフォーマンスに従って生成された、符号化ストリームを受信し、残差予測処理を実行する。 The image decoding device 31 receives the encoded stream generated according to the bit stream conformance and executes a residual prediction process.
 画像復号装置31は、上記をビットストリームコンフォーマンスとする符号化データを復号する。画像復号装置31の復号対象とするビットストリームとして、ARP参照ピクチャに関する制約を課すことによって、ARP参照ピクチャが参照できないことにより復号処理が失敗することを防ぐことができる。 The image decoding device 31 decodes the encoded data having the above bitstream conformance. By imposing restrictions on the ARP reference picture as a bit stream to be decoded by the image decoding device 31, it is possible to prevent the decoding process from failing because the ARP reference picture cannot be referred to.
 なお、ARP参照ピクチャarpRefPicが参照ピクチャリストRefPicListXの先頭の要素として導出される場合には、上記のビットストリームコンフォーマンスにおいてarpRefIdxL0=0、arpRefIdxL1=0とすれば良い。 When the ARP reference picture arpRefPic is derived as the head element of the reference picture list RefPicListX, arpRefIdxL0 = 0 and arpRefIdxL1 = 0 may be set in the above bitstream conformance.
 (オプションY2):
 インター予測パラメータ復号部303およびエントロピー復号部301のいずれかが、参照画像判定部3095を備える構成であってもよい。
(Option Y2):
Any one of the inter prediction parameter decoding unit 303 and the entropy decoding unit 301 may include a reference image determination unit 3095.
 そして、図15に示すように、パース段階(CUレベル)で、参照画像判定部3095がARP参照ピクチャ利用フラグarpRefPicAvailableフラグを導出してもよい。また、図15のSYN21およびSYN22に示すように、パース段階で、つまり、インター予測パラメータ復号部303およびエントロピー復号部301のいずれかにおいて、参照画像判定部3095が、ARP参照ピクチャの利用可否(ARP参照ピクチャ利用フラグarpRefPicAvailable)に応じて残差予測フラグ(ARPフラグ)iv_res_pred_weight_idxを復号してもよい。すなわち、パース段階のパラメータ復号において、インタービュー残差予測が利用可能なレイヤであって(iv_res_pred_flag[nuh_layer_id]が0でない)、かつARP参照ピクチャarpRefPicが利用可能である(arpRefPicAvailableが0でない)場合、iv_res_pred_weight_idxを復号してもよい。 Then, as shown in FIG. 15, the reference image determination unit 3095 may derive the ARP reference picture use flag arpRefPicAvailable flag at the parsing stage (CU level). Further, as shown in SYN21 and SYN22 in FIG. 15, at the parsing stage, that is, in either the inter prediction parameter decoding unit 303 or the entropy decoding unit 301, the reference image determination unit 3095 determines whether or not the ARP reference picture can be used (ARP The residual prediction flag (ARP flag) iv_res_pred_weight_idx may be decoded according to the reference picture usage flag arpRefPicAvailable). That is, in the parsing stage parameter decoding, when the inter-view residual prediction is available (iv_res_pred_flag [nuh_layer_id] is not 0) and the ARP reference picture arpRefPic is available (arpRefPicAvailable is not 0), iv_res_pred_weight_idx may be decoded.
 なお、参照画像判定部3095は、パース段階では、arpRefPicAvailableを、以下の式(Y2-1)に従って導出する。 The reference image determination unit 3095 derives arpRefPicAvailable according to the following equation (Y2-1) in the parsing stage.
 arpRefPicAvailable = !(DiffPicOrderCnt(RefPicList0[arpRefIdxL0], currPic) == 0)
&& !(DiffPicOrderCnt(RefPicList1[arpRefIdxL1] , currPic) == 0) && slice_type == B)・・・ (Y2-1)
 式(Y2-1)では、以下の(Y2.1)および(Y2.2)を満たす場合、arpRefPicAvailableに対して真(1)が設定される。
(Y2.1) L0参照ピクチャリストのインデックスarpRefIdxL0の参照ピクチャのPOCと、対象ピクチャcurrPicのPOC(PicOrderCntVal)との差分が0でない。
(Y2.2) L1参照ピクチャリストのインデックスarpRefIdxL1の参照ピクチャのPOCと、対象ピクチャcurrPicのPOC(PicOrderCntVal)との差分が0でなく、かつ、スライスタイプ(slice_type)がBである。
arpRefPicAvailable =! (DiffPicOrderCnt (RefPicList0 [arpRefIdxL0], currPic) == 0)
&&! (DiffPicOrderCnt (RefPicList1 [arpRefIdxL1], currPic) == 0) && slice_type == B) ... (Y2-1)
In the equation (Y2-1), true (1) is set for arpRefPicAvailable when the following (Y2.1) and (Y2.2) are satisfied.
(Y2.1) The difference between the POC of the reference picture with the index arpRefIdxL0 in the L0 reference picture list and the POC (PicOrderCntVal) of the target picture currPic is not zero.
(Y2.2) The difference between the POC of the reference picture of the index arpRefIdxL1 of the L1 reference picture list and the POC (PicOrderCntVal) of the target picture currPic is not 0, and the slice type (slice_type) is B.
 すなわち、L0リストの参照ピクチャarpRefPicL0(上記式ではRefPicList0[arpRefIdxL0])のPOCと、L1リストの参照ピクチャarpRefPicL1(上記式ではRefPicList1[arpRefIdxL1])のPOCの何れかが、対象ピクチャcurrPicのPOCと等しくなければ上記式(Y4-1)は真となる。スライスタイプslice_typeが双方向予測可能を示すBの場合のみ、L1リストが用いられるため、上記式では、slice_type == Bの場合のみ、L1リストに関する条件が可否に利用されるようにしている。(Y2-1)は、変形すれば、(Y2-1´)とも記載できる。 That is, either the POC of the reference picture arpRefPicL0 (RefPicList0 [arpRefIdxL0] in the above expression) of the L0 list or the POC of the reference picture arpRefPicL1 (RefPicList1 [arpRefIdxL1] in the above expression) of the L0 list is equal to the POC of the target picture currPic. Otherwise, the above formula (Y4-1) is true. Since the L1 list is used only when the slice type slice_type is B indicating that bi-directional prediction is possible, in the above formula, the condition relating to the L1 list is used only when slice_type == B. (Y2-1) can also be described as (Y2-1 ′) if it is modified.
 arpRefPicAvailable = !( PicOrderCnt (RefPicListL0[arpRefPicL0]) == PicOrderCntVal) || !( PicOrderCnt (RefPicListL1[arpRefPicL1]) == PicOrderCntVal) && slice_type == B) ・・・ (Y2-1´)
 ここで、PicOrderCnt(Pic)はPicのPOCを導出する関数である。
arpRefPicAvailable =! (PicOrderCnt (RefPicListL0 [arpRefPicL0]) == PicOrderCntVal) ||! (PicOrderCnt (RefPicListL1 [arpRefPicL1]) == PicOrderCntVal) && slice_type == B) ... (Y2-1 ')
Here, PicOrderCnt (Pic) is a function for deriving the POC of Pic.
 なお、ARP参照ピクチャarpRefPicが参照ピクチャリストRefPicListXの先頭の要素として導出される場合には、上記の式(Y2-1)においてarpRefIdxL0=0、arpRefIdxL1=0とした以下の式(Y2-1´´)を用いてarpRefPicAvailableを導出すれば良い。 When the ARP reference picture arpRefPic is derived as the first element of the reference picture list RefPicListX, the following expression (Y2-1 ″) with arpRefIdxL0 = 0 and arpRefIdxL1 = 0 in the above expression (Y2-1) ) To derive arpRefPicAvailable.
 arpRefPicAvailable = !( PicOrderCnt (RefPicListL0]) == PicOrderCntVal) || !( PicOrderCnt (RefPicListL1[0]) == PicOrderCntVal) && slice_type == B) ・・・ (Y2-1´´)
 オプションY2の参照画像判定部3095およびエントロピー復号部301(インター予測パラメータ復号部303)によれば、ARP参照ピクチャarpRefPicが対象ピクチャcurrPicのPOCと等しいかに応じて、残差予測フラグ(ARPフラグ)iv_res_pred_weight_idxを復号する。ARPを行うことができない場合、つまり、ARP参照ピクチャarpRefPicが対象ピクチャcurrPicのPOCと等しい場合に無駄なフラグとなるiv_res_pred_weight_idxを復号しないため、符号量低減の効果を奏する。
arpRefPicAvailable =! (PicOrderCnt (RefPicListL0]) == PicOrderCntVal) ||! (PicOrderCnt (RefPicListL1 [0]) == PicOrderCntVal) && slice_type == B) ・ ・ ・ (Y2-1´´)
According to the reference image determination unit 3095 and the entropy decoding unit 301 (inter prediction parameter decoding unit 303) of the option Y2, depending on whether the ARP reference picture arpRefPic is equal to the POC of the target picture currPic, a residual prediction flag (ARP flag) Decode iv_res_pred_weight_idx. When ARP cannot be performed, that is, when the ARP reference picture arpRefPic is equal to the POC of the target picture currPic, iv_res_pred_weight_idx, which is a useless flag, is not decoded.
 (オプションY3)
 また、オプション3では、インター予測画像生成部309において、動き補償段階における判定を行う。以下では、参照ピクチャリストのインデックスarpRefIdxLXのピクチャ(RefPicListX[arpRefIdxLX])をARP参照ピクチャarpRefPicと称する。
(Option Y3)
In option 3, the inter prediction image generation unit 309 performs determination in the motion compensation stage. Hereinafter, the picture (RefPicListX [arpRefIdxLX]) of the index arpRefIdxLX of the reference picture list is referred to as an ARP reference picture arpRefPic.
 まず、インター予測画像生成部309において、参照画像判定部3095は、参照ピクチャarpRefPicについて、arpRefPicのピクチャ順序(POC (arpRefPic))が、対象ピクチャcurrPicのピクチャ順序PicOrderCntValと等しいか否かに応じて参照レイヤ参照ピクチャ利用フラグ(refIvRefPicAvailable2)を設定する。具体的には、DiffPicOrderCnt(arpRefPic, currPic)が0でなければ、refIvRefPicAvailable2を1に設定する一方で、それ以外の場合は、refIvRefPicAvailable2を0に設定する。 First, in the inter predicted image generation unit 309, the reference image determination unit 3095 refers to the reference picture arpRefPic depending on whether the picture order of the arpRefPic (POCPO (arpRefPic)) is equal to the picture order PicOrderCntVal of the target picture currPic. Set the layer reference picture availability flag (refIvRefPicAvailable2). Specifically, if DiffPicOrderCnt (arpRefPic, currPic) is not 0, refIvRefPicAvailable2 is set to 1; otherwise, refIvRefPicAvailable2 is set to 0.
 参照画像判定部3095は、参照ピクチャarpRefPicについて、arpRefPicのピクチャ順序(PicOrderCnt (arpRefPic))が、対象ピクチャcurrPicのピクチャ順序PicOrderCntValと等しいか否かに応じて参照レイヤ参照ピクチャ利用フラグ(refIvRefPicAvailable2)を設定する。具体的には、DiffPicOrderCnt(arpRefPic, currPic)が0でなければ、refIvRefPicAvailable2を非0に設定する一方で、それ以外の場合は、refIvRefPicAvailable2を0に設定する。 The reference image determination unit 3095 sets a reference layer reference picture use flag (refIvRefPicAvailable2) for the reference picture arpRefPic depending on whether the picture order of arpRefPic (PicOrderCnt (arpRefPic)) is equal to the picture order PicOrderCntVal of the target picture currPic To do. Specifically, if DiffPicOrderCnt (arpRefPic, currPic) is not 0, refIvRefPicAvailable2 is set to non-zero, otherwise refIvRefPicAvailable2 is set to 0.
 refIvRefPicAvailable2=!DiffPicOrderCnt(arpRefPic, currPic) (Y3-1)
 ここで、式(Y3-1)は、以下の式(Y3-1´)と表現することも可能である。
refIvRefPicAvailable2 =! DiffPicOrderCnt (arpRefPic, currPic) (Y3-1)
Here, the expression (Y3-1) can also be expressed as the following expression (Y3-1 ′).
 refIvRefPicAvailable2=PicOrderCnt(arpRefPic) != PicOrderCntVal (Y3-1´)
 なお、参照ピクチャarpRefPicが、参照ピクチャリストRefPicListX[]の参照インデックスarpRefIdxLXの要素RefPicListX[arpRefIdxLX]から導出される場合には、以下の式(Y3-2)で導出する。
refIvRefPicAvailable2 = PicOrderCnt (arpRefPic)! = PicOrderCntVal (Y3-1 ')
When the reference picture arpRefPic is derived from the element RefPicListX [arpRefIdxLX] of the reference index arpRefIdxLX of the reference picture list RefPicListX [], it is derived by the following equation (Y3-2).
 refIvRefPicAvailable2=!DiffPicOrderCnt(RefPicListX[arpRefIdxLX], currPic) (Y3―2)
 ここでは、LX(X=0, 1)リストから導出される動き補償画像に残差予測を実施する場合に、そのリストLXの参照ピクチャインデックスであるarpRefIdxLXが0以上であるか(LXリストに有効な参照ピクチャarpRefPicが存在するかを判定している。式(Y3-2)は、以下の式(Y3-2´)と表現することも可能である。
refIvRefPicAvailable2 =! DiffPicOrderCnt (RefPicListX [arpRefIdxLX], currPic) (Y3-2)
Here, when residual prediction is performed on a motion compensated image derived from the LX (X = 0, 1) list, whether the reference picture index of the list LX is arpRefIdxLX is 0 or more (valid for the LX list) The equation (Y3-2) can also be expressed as the following equation (Y3-2 ′).
 refIvRefPicAvailable2=PicOrderCnt(RefPicListX[arpRefIdxLX]) != PicOrderCntVal
 (Y3-2´)
 残差予測実施フラグ導出部30921は、導出したrefIvRefPicAvailable2が非0である場合のみ、実施フラグresPredFlagが非0となるように、以下の式(C3-1)に従ってresPredFlagを導出する。
refIvRefPicAvailable2 = PicOrderCnt (RefPicListX [arpRefIdxLX])! = PicOrderCntVal
(Y3-2 ')
The residual prediction implementation flag deriving unit 30921 derives resPredFlag according to the following equation (C3-1) so that the implementation flag resPredFlag is non-zero only when the derived refIvRefPicAvailable2 is non-zero.
 resPredFlag = (iv_res_pred_weight_idx != 0) && arpRefPicAvailable2 ・・・ 式(C3-1)
 なお、残差予測を、動き予測の場合に限定する場合には、式(Y3-1)の変形例である以下の式(C3-2)を用いてresPredFlagを導出しても良い(以降の実施形態も同様)
 resPredFlag = (iv_res_pred_weight_idx != 0) &&
 (PicOrderCnt(RefPicListX[refIdxLX]) != PicOrderCntVal) && arpRefPicAvailable2
 ・・・(C3-2)
 残差予測実施フラグ導出部30921は、実施形態1で導出したrefIvRefPicAvailableを併用して、実施フラグresPredFlagを導出しても良い。具体的には、実施形態1の式(R-1)と、式(Y3-1)をあわせて以下の式を用いても良い。
resPredFlag = (iv_res_pred_weight_idx! = 0) && arpRefPicAvailable2 Expression (C3-1)
When the residual prediction is limited to motion prediction, resPredFlag may be derived using the following equation (C3-2), which is a modification of equation (Y3-1) The same applies to the embodiment)
resPredFlag = (iv_res_pred_weight_idx! = 0) &&
(PicOrderCnt (RefPicListX [refIdxLX])! = PicOrderCntVal) && arpRefPicAvailable2
... (C3-2)
The residual prediction implementation flag deriving unit 30921 may derive the implementation flag resPredFlag using the refIvRefPicAvailable derived in the first embodiment together. Specifically, the following formula may be used by combining the formula (R-1) and the formula (Y3-1) of Embodiment 1.
 resPredFlag = ( iv_res_pred_weight_idx != 0 ) && refIvRefPicAvailable && refIvRefPicAvailable2 ・・・(C3-3)
 なお、残差予測を、動き予測の場合に限定する場合には、式(Y3-3)の変形例である以下の式(C3-4)を用いてresPredFlagを導出しても良い(以降の実施形態も同様)
 resPredFlag = (iv_res_pred_weight_idx != 0) &&
 (PicOrderCnt(RefPicListX[refIdxLX]) != PicOrderCntVal) && arpRefPicAvailable1 && arpRefPicAvailable2 ・・・(C3-4)
 オプションY3の参照画像判定部3095および残差予測実施フラグ導出部30921(インター予測画像生成部309)によれば、ARP参照ピクチャarpRefPicが対象ピクチャcurrPicのPOCと等しいかに応じて、動き補償画像に残差予測を行うかどうか(resPredFlag)を決定する。ARPを行うことができない場合、つまり、ARP参照ピクチャarpRefPicが対象ピクチャcurrPicのPOCと等しい場合にARPの動作を行わないことによって、無効な動作を避ける効果を奏する。
resPredFlag = (iv_res_pred_weight_idx! = 0) && refIvRefPicAvailable && refIvRefPicAvailable2 (C3-3)
When the residual prediction is limited to motion prediction, resPredFlag may be derived using the following equation (C3-4), which is a modification of equation (Y3-3) The same applies to the embodiment)
resPredFlag = (iv_res_pred_weight_idx! = 0) &&
(PicOrderCnt (RefPicListX [refIdxLX])! = PicOrderCntVal) && arpRefPicAvailable1 && arpRefPicAvailable2 (C3-4)
According to the reference image determination unit 3095 and the residual prediction execution flag derivation unit 30921 (inter prediction image generation unit 309) of the option Y3, the motion compensation image is determined according to whether the ARP reference picture arpRefPic is equal to the POC of the target picture currPic. Determine whether to perform residual prediction (resPredFlag). When the ARP cannot be performed, that is, when the ARP reference picture arpRefPic is equal to the POC of the target picture currPic, the ARP operation is not performed, so that an invalid operation is avoided.
 (オプションY4)
 オプションY4では、インター予測パラメータ復号部303およびエントロピー復号部301のいずれかが、参照画像判定部3095を備える構成とする。
(Option Y4)
In the option Y4, either the inter prediction parameter decoding unit 303 or the entropy decoding unit 301 is configured to include the reference image determination unit 3095.
 また、オプションY4では、インター予測パラメータ復号部303およびエントロピー復号部301におけるパース段階での判定とインター予測画像生成部309における動き補償段階における判定を併用する。 Also, in option Y4, the determination at the parsing stage in the inter prediction parameter decoding unit 303 and the entropy decoding unit 301 and the determination at the motion compensation stage in the inter prediction image generation unit 309 are used together.
 パース段階では、インター予測パラメータ復号部303およびエントロピー復号部301のいずれかにおいて、参照画像判定部3095は、上述の式(Y2-1)およびその変形例に従って、図15に示したarpRefPicAvailableを導出してもよい。 In the parsing stage, in either the inter prediction parameter decoding unit 303 or the entropy decoding unit 301, the reference image determination unit 3095 derives the arpRefPicAvailable shown in FIG. 15 according to the above equation (Y2-1) and its modification. May be.
 また、インター予測パラメータ復号部303およびエントロピー復号部301では、ARP参照ピクチャ利用フラグarpRefPicAvailableフラグに応じて、残差予測フラグ(ARPフラグ)iv_res_pred_weight_idxを復号してもよい。すなわち図16に示すように、arpRefPicAvailableフラグが真の場合のみiv_res_pred_weight_idxを復号してもよい。 Also, the inter prediction parameter decoding unit 303 and the entropy decoding unit 301 may decode the residual prediction flag (ARP flag) iv_res_pred_weight_idx according to the ARP reference picture use flag arpRefPicAvailable flag. That is, as shown in FIG. 16, iv_res_pred_weight_idx may be decoded only when the arpRefPicAvailable flag is true.
 一方、動き補償段階では、以下のとおり判定を行うよう参照画像判定部3095を変形する。 On the other hand, in the motion compensation stage, the reference image determination unit 3095 is modified to perform the determination as follows.
 参照画像判定部3095は、上述の式(Y3-1)およびその変形例に従って、refIvRefPicAvailable2を導出する。 The reference image determination unit 3095 derives refIvRefPicAvailable2 in accordance with the above equation (Y3-1) and its modifications.
 また、残差予測実施フラグ導出部30921は、実施形態3において説明した式(Y3-1)およびその変形例に従ってresPredFlagを導出する。 Also, the residual prediction execution flag deriving unit 30921 derives resPredFlag according to the equation (Y3-1) described in the third embodiment and its modification.
 オプションY4の参照画像判定部3095およびエントロピー復号部301(インター予測パラメータ復号部303)によれば、オプションY2と同様、ARP参照ピクチャarpRefPicが対象ピクチャcurrPicのPOCと等しい場合に無駄なフラグとなるiv_res_pred_weight_idxを復号しないため、符号量低減の効果を奏する。 According to the reference image determination unit 3095 and the entropy decoding unit 301 (inter prediction parameter decoding unit 303) of the option Y4, similarly to the option Y2, the iv_res_pred_weight_idx becomes a useless flag when the ARP reference picture arpRefPic is equal to the POC of the target picture currPic. Therefore, the amount of code can be reduced.
 オプションY4の参照画像判定部3095および残差予測実施フラグ導出部30921(インター予測画像生成部309)によれば、オプションY3と同様、ARPを行うことができない場合、つまり、ARP参照ピクチャarpRefPicが対象ピクチャcurrPicのPOCと等しい場合にARPの動作を行わないことによって、無効な動作を避ける効果を奏する。 According to the reference image determination unit 3095 and the residual prediction execution flag deriving unit 30921 (inter prediction image generation unit 309) of the option Y4, as in the case of the option Y3, that is, the ARP reference picture arpRefPic is the target. By not performing the ARP operation when it is equal to the POC of the picture currPic, there is an effect of avoiding an invalid operation.
 (オプションY5)
 オプション5では、ARP参照ピクチャarpRefPicが参照ピクチャリストRefPicListXのarpRefIdxLXの参照ピクチャRefPicListX[arpRefIdxLX]である場合において、arpRefPicがレイヤ間ピクチャとはならないよう、つまり、可能であれば、対象ピクチャのPOCであるPictureOrderCntValと異なるPOCを有するarpRefPicを設定するように、arpRefIdxLXを選択する。
(Option Y5)
In option 5, when the ARP reference picture arpRefPic is the reference picture RefPicListX [arpRefIdxLX] of the arpRefIdxLX of the reference picture list RefPicListX, the arpRefPic is not an inter-layer picture, that is, if possible, it is the POC of the target picture Select arpRefIdxLX to set arpRefPic having a POC different from PictureOrderCntVal.
 本実施例の参照画像判定部3095は、参照ピクチャarpRefPicの導出も行う。 The reference image determination unit 3095 of this embodiment also derives a reference picture arpRefPic.
 参照画像判定部3095は、以下の疑似コードに従い、X = 0 or 1について、インターレイヤピクチャではないという条件を満たす参照ピクチャリストRefPicListX[]の要素を検索し、ARP参照ピクチャのインデックスarpRefIdxLXと、arpRefPicAvailableとを導出してもよい。 The reference image determination unit 3095 searches for an element of the reference picture list RefPicListX [] that satisfies the condition that it is not an inter-layer picture with respect to X = 0 or に 従 い 1 according to the following pseudo code, an ARP reference picture index arpRefIdxLX, and arpRefPicAvailable And may be derived.
 なお、以下の疑似コードにおいて、X = 0 or 1である。また、以下の疑似コードの実行前の初期化処理として、arpRefIdxL0およびarpRefIdxL1は、参照ピクチャインデックスとしては、無効な値である負値(ここでは-1)に設定するとともに、arpRefPicAvailableは偽を示す0に設定する。
================================================================================for(i = 0; i <= num_ref_idx_lX_active_minus1; i++) {
 if(DiffPicOrderCnt(RefPicListX[i], currPic)) {
  arpRefIdxLX = i
  arpRefPicAvailable = 1
  }
 }
================================================================================
 上記疑似コードでは、参照ピクチャリストRefPicListX[]中の0からnum_ref_idx_lX_active_minus1までのインデックスiの参照ピクチャRefPicListX[i]を順に走査し、参照ピクチャRefPicListX[i]のPOCと、対象ピクチャcurrPicのPOC(PicOrderCntVal)との差が0ではない条件を満たす参照ピクチャarpRefPicLXを検索している。上記疑似コードによって得られたarpRefIdxLXに関し、参照画像判定部3095は、RefPIcListX[arpRefIdxLX]を、ARP参照ピクチャarpRefPicとして採用してもよい。
In the following pseudo code, X = 0 or 1. As initialization processing before execution of the following pseudo code, arpRefIdxL0 and arpRefIdxL1 are set to negative values (in this case, −1) which are invalid values as reference picture indexes, and arpRefPicAvailable is 0 indicating false. Set to.
================================================== ============================== for (i = 0; i <= num_ref_idx_lX_active_minus1; i ++) {
if (DiffPicOrderCnt (RefPicListX [i], currPic)) {
arpRefIdxLX = i
arpRefPicAvailable = 1
}
}
================================================== ==============================
In the above pseudo code, the reference picture RefPicListX [i] of index i from 0 to num_ref_idx_lX_active_minus1 in the reference picture list RefPicListX [] is sequentially scanned, and the POC of the reference picture RefPicListX [i] and the POC of the target picture currPic (PicOrderCntVal) The reference picture arpRefPicLX that satisfies the condition that the difference between and is not 0 is searched. Regarding the arpRefIdxLX obtained by the pseudo code, the reference image determination unit 3095 may adopt RefPIcListX [arpRefIdxLX] as the ARP reference picture arpRefPic.
 なお、上記処理は、以下の擬似コードを用いても処理できる。 The above processing can also be performed using the following pseudo code.
 arpRefIdxL0 = -1
 for(i = 0; i <= num_ref_idx_l0_active_minus1 && arpRefIdxL0 < 0; i++)
  if (PicOrderCnt(RefPicList0[i]) != PicOrderCntVal)
   arpRefIdxL0 = i
  arpRefIdxL1 = -1
 if (slice_type != B)
 for(i = 0; i <= num_ref_idx_l1_active_minus1 && arpRefIdxL1 < 0; i++)
  if (PicOrderCnt(RefPicList1[i]) != PicOrderCntVal)
   arpRefIdxL1 = i
 上記、擬似コードにおいては、X=0、X=1の各々の参照ピクチャリストLXについて、参照ピクチャインデックスarpRefIdxLXとして0未満の値(ここでは-1)で初期化した上で、参照ピクチャリストRefPicListX[]中の0からnum_ref_idx_lX_active_minus1までのインデックスiの参照ピクチャRefPicListX[i]を順に走査し、参照ピクチャRefPicListX[i]のPOCと、対象ピクチャのPOC(PicOrderCntVal)が異なる条件を満たす参照ピクチャ(レイヤ間ピクチャ以外の参照ピクチャ)を検索し、条件を満たす参照ピクチャの参照ピクチャインデックスiをarpRefIdxLXとして保持する。初期状態では、参照ピクチャインデックスarpRefIdxLXとして、負の値(ここでは-1)を与え、上記検索を行う。条件を満たす参照ピクチャが得られた場合でarpRefIdxLXは0以上の値が設定される。なお、上記検索は、条件を満たす参照ピクチャが得られた時点、つまり、arpRefIdxLXが0以上になった時点で終了する。
arpRefIdxL0 = -1
for (i = 0; i <= num_ref_idx_l0_active_minus1 && arpRefIdxL0 <0; i ++)
if (PicOrderCnt (RefPicList0 [i])! = PicOrderCntVal)
arpRefIdxL0 = i
arpRefIdxL1 = -1
if (slice_type! = B)
for (i = 0; i <= num_ref_idx_l1_active_minus1 && arpRefIdxL1 <0; i ++)
if (PicOrderCnt (RefPicList1 [i])! = PicOrderCntVal)
arpRefIdxL1 = i
In the above pseudo code, each reference picture list LX with X = 0 and X = 1 is initialized with a reference picture index arpRefIdxLX with a value less than 0 (here, −1), and then the reference picture list RefPicListX [ ] Is sequentially scanned from 0 to num_ref_idx_lX_active_minus1, and the reference picture (inter-layer picture) that satisfies the different POC of the reference picture RefPicListX [i] and the POC (PicOrderCntVal) of the target picture (Reference picture other than) and the reference picture index i of the reference picture satisfying the condition is held as arpRefIdxLX. In the initial state, as the reference picture index arpRefIdxLX, a negative value (here, -1) is given, and the above search is performed. When a reference picture that satisfies the condition is obtained, arpRefIdxLX is set to a value of 0 or more. The search ends when a reference picture that satisfies the condition is obtained, that is, when arpRefIdxLX becomes 0 or more.
 オプションY5も、オプションY2、オプションY4と同様、インター予測パラメータ復号部303およびエントロピー復号部301におけるパース段階での判定を行っても良い。この場合、インター予測パラメータ復号部303およびエントロピー復号部301のいずれかが、参照画像判定部3095を備える構成とする。 Option Y5 may also perform determination at the parsing stage in the inter prediction parameter decoding unit 303 and the entropy decoding unit 301, similarly to the option Y2 and the option Y4. In this case, either the inter prediction parameter decoding unit 303 or the entropy decoding unit 301 is configured to include the reference image determination unit 3095.
 パース段階では、インター予測パラメータ復号部303およびエントロピー復号部301のいずれかにおいて、参照画像判定部3095は、既に説明したように上述の式(Y2-1)および変形例(Y2-1´)に従って、図15に示したarpRefPicAvailableを導出することも可能だが、以下の式(Y5-1)において導出することも可能である。 In the parsing stage, in either the inter prediction parameter decoding unit 303 or the entropy decoding unit 301, the reference image determination unit 3095 follows the above equation (Y2-1) and the modification example (Y2-1 ′) as described above. Although it is possible to derive arpRefPicAvailable shown in FIG. 15, it is also possible to derive it using the following equation (Y5-1).
 arpRefPicAvailable = arpRefIdxL0 >= 0 || arpRefIdxL1 >= 0
 ここでは、参照画像判定部3095は、パース段階においては、L0の参照ピクチャインデックスであるarpRefIdxL0が有効である値(0以上)もしくはL1の参照ピクチャインデックスであるarpRefIdxL1が有効である値(0以上)の場合に、arpRefPicAvailableに真(0以外の値)を設定する。すなわち、L0の参照ピクチャリストもしくはL1の参照ピクチャリストのいずれかでレイヤ間ピクチャ以外のピクチャが見つかった場合には、arpRefPicAvailableに真(非0)を設定する。
arpRefPicAvailable = arpRefIdxL0> = 0 || arpRefIdxL1> = 0
Here, in the parsing stage, the reference image determination unit 3095 has a value (0 or more) in which the arpRefIdxL0 that is the reference picture index of L0 is valid or a value (0 or more) in which the arpRefIdxL1 that is the reference picture index of L1 is valid. In this case, arpRefPicAvailable is set to true (a value other than 0). That is, when a picture other than the inter-layer picture is found in either the L0 reference picture list or the L1 reference picture list, arpRefPicAvailable is set to true (non-zero).
 また、オプションY5も、オプションY3、オプションY4と同様、インター予測画像生成部309における動き補償段階における判定を行うことも可能である。 Also, the option Y5 can also make a determination at the motion compensation stage in the inter-predicted image generation unit 309, like the options Y3 and Y4.
 動き補償段階では、インター予測画像生成部309において、参照画像判定部3095は、既に説明したように上述の式(Y3-1)およびその変形例に従って、refIvRefPicAvailable2を導出することも可能だが、以下の式(Y5-2)において導出することも可能である。 In the motion compensation stage, in the inter prediction image generation unit 309, the reference image determination unit 3095 can derive refIvRefPicAvailable2 in accordance with the above-described equation (Y3-1) and its modification as described above. It is also possible to derive in the equation (Y5-2).
 refIvRefPicAvailable2 = arpRefIdxLX >= 0 式(Y5-2)
 ここでは、動き補償段階において、LX(X=0, 1)リストから導出される動き補償画像に残差予測を実施する場合に、そのリストLXの参照ピクチャインデックスであるarpRefIdxLXが0以上であるか(LXリストに有効な参照ピクチャarpRefPicが存在するか)により判定を行っている。
refIvRefPicAvailable2 = arpRefIdxLX> = 0 Formula (Y5-2)
Here, in the motion compensation stage, when residual prediction is performed on a motion compensated image derived from the LX (X = 0, 1) list, whether the reference picture index arpRefIdxLX of the list LX is 0 or more The determination is made based on (whether there is a valid reference picture arpRefPic in the LX list).
 また、残差予測実施フラグ導出部30921は、オプションY3において説明した式(C3-1)およびその変形例に従ってresPredFlagを導出する。 Also, the residual prediction execution flag deriving unit 30921 derives resPredFlag according to the equation (C3-1) described in the option Y3 and its modification.
 オプションY5の参照画像判定部3095によれば、対象ピクチャのPOCであるPictureOrderCntValと異なるPOCを有するarpRefPicを設定するように、arpRefIdxLXを選択するため、arpRefPicのPOCが対象ピクチャのPOCであるPictureOrderCntValと等しいためにARPが動作しないという課題を解消する効果を奏する。 According to the reference image determination unit 3095 of option Y5, the arpRefIdxLX is selected so that the arpRefPic having a POC different from the PictureOrderCntVal that is the POC of the target picture is selected. Therefore, there is an effect of eliminating the problem that the ARP does not operate.
 さらに、パース段階の判定を行う場合には、オプションY5の参照画像判定部3095およびエントロピー復号部301(インター予測パラメータ復号部303)によれば、オプションY2と同様、ARP参照ピクチャarpRefPicが対象ピクチャcurrPicのPOCと等しい場合に無駄なフラグとなるiv_res_pred_weight_idxを復号しないため、符号量低減の効果を奏する。 Further, when the parsing stage is determined, according to the reference image determination unit 3095 and the entropy decoding unit 301 (inter prediction parameter decoding unit 303) of the option Y5, the ARP reference picture arpRefPic is the target picture currPic as in the option Y2. Since iv_res_pred_weight_idx, which is a useless flag when it is equal to the POC, is not decoded, the code amount can be reduced.
  さらに、動き補償段階の判定を行う場合には、オプションY5の参照画像判定部3095および残差予測実施フラグ導出部30921(インター予測画像生成部309)によれば、オプションY3と同様、ARPを行うことができない場合、つまり、ARP参照ピクチャarpRefPicが対象ピクチャcurrPicのPOCと等しい場合にARPの動作を行わないことによって、無効な動作を避ける効果を奏する。 Further, when the determination of the motion compensation stage is performed, according to the reference image determination unit 3095 and the residual prediction execution flag deriving unit 30921 (inter prediction image generation unit 309) of the option Y5, ARP is performed similarly to the option Y3. If this is not possible, that is, if the ARP reference picture arpRefPic is equal to the POC of the target picture currPic, the ARP operation is not performed, so that an invalid operation is avoided.
 (オプションY5の変形例)
 以下、オプションY5の変形例を説明する。オプションY5の変形例も、対象ピクチャのPOCであるPictureOrderCntValと異なるPOCを有するarpRefPicを設定するように参照ピクチャを定める。
(Modification of option Y5)
Hereinafter, a modified example of the option Y5 will be described. In the modified example of the option Y5, the reference picture is determined so as to set an arpRefPic having a POC different from the PictureOrderCntVal that is the POC of the target picture.
 参照画像判定部3095は、ARP参照ピクチャarpRecPicを、参照ピクチャリストRefPicListXに含まれる参照ピクチャから、以下に示す疑似コードに従って導出してもよい。なお、以下の疑似コードの実行前に初期化処理として、POCの差分の暫定的な最小値を示すminiDiffPOCを、十分に大きい所定の値(ここでは、例えば216)に設定するとともに、ARP参照ピクチャの利用可否を示すフラグであるarpRefPicAvailableを、0(偽)に設定する。
================================================================================for (rIdx = 0; rIdx <= num_ref_idx_lX_active_minus1; rIdx++) {
 if (isNearerPOC(rIdx, minDiffPOC)) {
  minDiffPOC = abs(DiffPicOrderCnt(RefPicListX[rIdx], currPic)
  rIdxSel = rIdx
  arpRefPicAvailable = 1 }
 }
================================================================================
 上記疑似コードでは、forループの処理が参照ピクチャリストに含まれる各参照ピクチャに対して実行される。
The reference image determination unit 3095 may derive the ARP reference picture arpRecPic from the reference pictures included in the reference picture list RefPicListX according to the following pseudo code. In addition, as an initialization process before the execution of the following pseudo code, miniDiffPOC indicating the provisional minimum value of the POC difference is set to a sufficiently large predetermined value (eg, 2 16 in this case ), and ARP reference is made. ArpRefPicAvailable, which is a flag indicating availability of a picture, is set to 0 (false).
================================================== ======================= for (rIdx = 0; rIdx <= num_ref_idx_lX_active_minus1; rIdx ++) {
if (isNearerPOC (rIdx, minDiffPOC)) {
minDiffPOC = abs (DiffPicOrderCnt (RefPicListX [rIdx], currPic)
rIdxSel = rIdx
arpRefPicAvailable = 1}
}
================================================== ==============================
In the above pseudo code, the for loop processing is executed for each reference picture included in the reference picture list.
 ここで、isNearerPOC(rIdx, minDiffPOC)は、参照ピクチャリストのrIdx番目の参照ピクチャが、対象ピクチャのRPSに含まれ、ViewIdx = RefViewIdx[xP][yP]であり、POCが以下の条件式C(1-1)を満たすような参照ピクチャが存在するか否かを判定する関数である。 Here, isNearerPOC (rIdx, minDiffPOC) includes the rIdx-th reference picture of the reference picture list included in the RPS of the target picture, and ViewIdxId = RefViewIdx [xP] [yP], and POC is the following conditional expression C ( This function determines whether or not there is a reference picture that satisfies 1-1).
 (参照ピクチャがRefPicListX[rIdx]に格納されている参照ピクチャと同じPOCを有する) && minDiffPOC > abs(DiffPicOrderCnt(RefPicListX[rIdx], currPic))・・・(C1-1)
 条件式(C1-1)において、DiffPicOrderCnt(PicA, PicB)は、ピクチャPicAのPOCからピクチャPicBのPOCを差し引いた値を求める関数である。(以下の構成においても同様)。対象ピクチャcurrPicのPOCはPicOrderCntValであるから、DiffPicOrderCnt(picA, currPic)をPOC(picA) - PicOrderCntValとしても良い。条件式C(1-1)では、RefPicListX[rIdx]に格納されている参照ピクチャと同じPOCを有する参照ピクチャが存在し、かつ、rIdx番目の参照ピクチャのPOCと、対象ピクチャcurrPicのPOC(PicOrderCntVal)との差分の絶対値が、minDiffPOC未満である場合、“真”となる。
(The reference picture has the same POC as the reference picture stored in RefPicListX [rIdx]) &&minDiffPOC> abs (DiffPicOrderCnt (RefPicListX [rIdx], currPic)) (C1-1)
In the conditional expression (C1-1), DiffPicOrderCnt (PicA, PicB) is a function for obtaining a value obtained by subtracting the POC of the picture PicB from the POC of the picture PicA. (The same applies to the following configurations). Since the POC of the target picture currPic is PicOrderCntVal, DiffPicOrderCnt (picA, currPic) may be POC (picA) -PicOrderCntVal. In the conditional expression C (1-1), there is a reference picture having the same POC as the reference picture stored in RefPicListX [rIdx], and the POC of the rIdx-th reference picture and the POC (PicOrderCntVal of the target picture currPic) ) And the absolute value of the difference is less than minDiffPOC, it is “true”.
 条件式(C1-1)が“真”である場合、minDiffPOCが、abs(DiffPicOrderCnt(RefPicListX[rIdx], currPic)によって更新され、rIdxが、暫定的なARP参照ピクチャの候補となる(rIdxSel)。また、このとき、arpRefPicAvailableは、1に設定される。 When the conditional expression (C1-1) is “true”, minDiffPOC is updated by abs (DiffPicOrderCnt (RefPicListX [rIdx], currPic), and rIdx becomes a tentative ARP reference picture candidate (rIdxSel). At this time, arpRefPicAvailable is set to 1.
 このようにして、各参照ピクチャについて、forループが実行されることで、最も参照ピクチャのPOCと近いPOCを有する参照ピクチャが、ARP参照ピクチャを示すインデックスとして(rIdxSel)導出される。 In this way, by executing the for loop for each reference picture, a reference picture having a POC closest to the POC of the reference picture is derived as an index indicating the ARP reference picture (rIdxSel).
 すなわち、参照画像判定部3095は、このようにして得られたrIdxSelに関し、ARP参照ピクチャarpRefPiを、が参照ピクチャリストRefPicListX[]のインデックスrIdxSelの参照ピクチャRefPicListX[rIdxSel]として導出する。 That is, the reference image determination unit 3095 derives the ARP reference picture arpRefPi as the reference picture RefPicListX [rIdxSel] of the index rIdxSel of the reference picture list RefPicListX [] for the rIdxSel obtained in this way.
 オプションY5の変形例の参照画像判定部3095によれば、対象ピクチャのPOCであるPictureOrderCntValと異なるPOCを有するarpRefPicを設定するように、arpRefIdxLXを選択するため、arpRefPicのPOCが対象ピクチャのPOCであるPictureOrderCntValと等しいためにARPが動作しないという課題を解消する効果を奏する。 According to the reference image determination unit 3095 of the modified example of the option Y5, the arpRefIdxLX is selected so that the arpRefPic having a POC different from the PictureOrderCntVal that is the POC of the target picture is set. Therefore, the POC of the arpRefPic is the POC of the target picture Since it is equal to PictureOrderCntVal, the effect that ARP does not operate is solved.
 さらに、オプションY5同様にパース段階の判定を行っても良い。この場合には、オプションY5の参照画像判定部3095およびエントロピー復号部301(インター予測パラメータ復号部303)によれば、オプションY2と同様、ARP参照ピクチャarpRefPicが対象ピクチャcurrPicのPOCと等しい場合に無駄なフラグとなるiv_res_pred_weight_idxを復号しないため、符号量低減の効果を奏する。 Furthermore, the parsing stage may be determined in the same manner as option Y5. In this case, according to the reference image determination unit 3095 and the entropy decoding unit 301 (inter prediction parameter decoding unit 303) of the option Y5, similarly to the option Y2, it is useless when the ARP reference picture arpRefPic is equal to the POC of the target picture currPic. Since iv_res_pred_weight_idx, which is a bad flag, is not decoded, the code amount can be reduced.
  さらに、オプションY5同様に動き補償段階の判定を行っても良い。この場合には、オプションY5の参照画像判定部3095および残差予測実施フラグ導出部30921(インター予測画像生成部309)によれば、オプションY3と同様、ARPを行うことができない場合、つまり、ARP参照ピクチャarpRefPicが対象ピクチャcurrPicのPOCと等しい場合にARPの動作を行わないことによって、無効な動作を避ける効果を奏する。 Furthermore, the motion compensation stage may be determined in the same manner as option Y5. In this case, according to the reference image determination unit 3095 and the residual prediction execution flag deriving unit 30921 (inter prediction image generation unit 309) of the option Y5, as in the case of the option Y3, that is, the ARP When the reference picture arpRefPic is equal to the POC of the target picture currPic, by not performing the ARP operation, there is an effect of avoiding an invalid operation.
 なお、上述した実施形態における画像符号化装置11、画像復号装置31の各部をコンピュータで実現するようにしても良い。その場合、この制御機能を実現するためのプログラムをコンピュータ読み取り可能な記録媒体に記録して、この記録媒体に記録されたプログラムをコンピュータシステムに読み込ませ、実行することによって実現しても良い。なお、ここでいう「コンピュータシステム」とは、画像符号化装置11、画像復号装置31に内蔵されたコンピュータシステムであって、OSや周辺機器等のハードウェアを含むものとする。また、「コンピュータ読み取り可能な記録媒体」とは、フレキシブルディスク、光磁気ディスク、ROM、CD-ROM等の可搬媒体、コンピュータシステムに内蔵されるハードディスク等の記憶装置のことをいう。さらに「コンピュータ読み取り可能な記録媒体」とは、インターネット等のネットワークや電話回線等の通信回線を介してプログラムを送信する場合の通信線のように、短時間、動的にプログラムを保持するもの、その場合のサーバやクライアントとなるコンピュータシステム内部の揮発性メモリのように、一定時間プログラムを保持しているものも含んでも良い。また上記プログラムは、前述した機能の一部を実現するためのものであっても良く、さらに前述した機能をコンピュータシステムにすでに記録されているプログラムとの組み合わせで実現できるものであっても良い。 In addition, you may make it implement | achieve each part of the image coding apparatus 11 in the embodiment mentioned above, and the image decoding apparatus 31 with a computer. In that case, the program for realizing the control function may be recorded on a computer-readable recording medium, and the program recorded on the recording medium may be read by a computer system and executed. The “computer system” here is a computer system built in the image encoding device 11 and the image decoding device 31, and includes an OS and hardware such as peripheral devices. The “computer-readable recording medium” refers to a storage device such as a flexible medium, a magneto-optical disk, a portable medium such as a ROM or a CD-ROM, and a hard disk incorporated in a computer system. Furthermore, the “computer-readable recording medium” is a medium that dynamically holds a program for a short time, such as a communication line when transmitting a program via a network such as the Internet or a communication line such as a telephone line, In such a case, a volatile memory inside a computer system serving as a server or a client may be included and a program that holds a program for a certain period of time. The program may be a program for realizing a part of the functions described above, and may be a program capable of realizing the functions described above in combination with a program already recorded in a computer system.
 また、上述した実施形態における画像符号化装置11、画像復号装置31の一部、または全部を、LSI(Large Scale Integration)等の集積回路として実現しても良い。画像符号化装置11、画像復号装置31の各機能ブロックは個別にプロセッサ化しても良いし、一部、または全部を集積してプロセッサ化しても良い。また、集積回路化の手法はLSIに限らず専用回路、または汎用プロセッサで実現しても良い。また、半導体技術の進歩によりLSIに代替する集積回路化の技術が出現した場合、当該技術による集積回路を用いても良い。 Further, part or all of the image encoding device 11 and the image decoding device 31 in the above-described embodiment may be realized as an integrated circuit such as an LSI (Large Scale Integration). Each functional block of the image encoding device 11 and the image decoding device 31 may be individually made into a processor, or a part or all of them may be integrated into a processor. Further, the method of circuit integration is not limited to LSI, and may be realized by a dedicated circuit or a general-purpose processor. Further, in the case where an integrated circuit technology that replaces LSI appears due to progress in semiconductor technology, an integrated circuit based on the technology may be used.
 以上、図面を参照してこの発明の一実施形態について詳しく説明してきたが、具体的な構成は上述のものに限られることはなく、この発明の要旨を逸脱しない範囲内において様々な設計変更等をすることが可能である。 As described above, the embodiment of the present invention has been described in detail with reference to the drawings. However, the specific configuration is not limited to the above, and various design changes and the like can be made without departing from the scope of the present invention. It is possible to
 〔まとめ〕
 本発明の態様1に係る画像復号装置は、動き補償画像に対象ピクチャとは異なるレイヤの参照レイヤを用いた残差予測を適用することで対象ピクチャの予測画像を生成する画像復号装置において、上記対象ピクチャのピクチャ順序と異なるピクチャ順序を有する上記参照レイヤの参照ピクチャが利用可能であるか否かを判定する参照ピクチャ判定部(参照画像判定部3095)と、少なくとも上記判定結果に応じて、上記参照レイヤの参照ピクチャと、対象ピクチャのピクチャ順序と同一のピクチャ順序を有する参照レイヤの復号済ピクチャとに基づく上記残差予測を、上記動き補償画像に適用する残差予測適用部(残差合成部30923)と、を備える。
[Summary]
An image decoding device according to aspect 1 of the present invention is an image decoding device that generates a predicted image of a target picture by applying residual prediction using a reference layer different from the target picture to the motion compensated image. A reference picture determination unit (reference image determination unit 3095) for determining whether or not a reference picture of the reference layer having a picture order different from the picture order of the target picture is available, and at least according to the determination result, A residual prediction application unit (residual synthesis) that applies the residual prediction based on the reference picture of the reference layer and the decoded picture of the reference layer having the same picture order as the target picture to the motion compensated image Part 30923).
 対象ピクチャとは復号の対象となるピクチャである。参照レイヤとは、対象ピクチャが属するレイヤである対象レイヤと異なるレイヤのことである。参照レイヤにおける残差予測を行う場合、参照レイヤに属する参照ピクチャを参照する。 The target picture is a picture to be decoded. The reference layer is a layer different from the target layer to which the target picture belongs. When performing residual prediction in the reference layer, reference pictures belonging to the reference layer are referred to.
 上記残差予測は、参照レイヤにおける残差を、対象レイヤにおける残差として推定するという技術である。上記残差予測は、対象ピクチャのピクチャ順序と異なるピクチャ順序を有する参照レイヤの参照ピクチャと、対象ピクチャのピクチャ順序と同一のピクチャ順序を有する参照レイヤの復号済ピクチャとに基づいて行われる。 The residual prediction is a technique for estimating the residual in the reference layer as the residual in the target layer. The residual prediction is performed based on the reference picture of the reference layer having a picture order different from the picture order of the target picture and the decoded picture of the reference layer having the same picture order as the picture order of the target picture.
 上記構成によれば、上記参照レイヤの参照ピクチャが利用可能であるか否かに応じて、上記動き補償画像に上記残差予測を適用することができる。 According to the above configuration, the residual prediction can be applied to the motion compensated image depending on whether or not a reference picture of the reference layer is available.
 よって、対象ピクチャについての残差予測時に、上記参照レイヤの復号ピクチャが利用できないという事態を回避することができるという効果を奏する。 Therefore, it is possible to avoid a situation in which the decoded picture of the reference layer cannot be used at the time of residual prediction for the target picture.
 本発明の態様2に係る画像復号装置では、上記態様1において、上記参照ピクチャ判定部は、上記参照レイヤの復号済ピクチャの復号時に参照されるピクチャを示す参照ピクチャセットにおいて、上記参照レイヤの参照ピクチャが存在するか否かに応じて、上記判定を行ってもよい。 In the image decoding device according to aspect 2 of the present invention, in the aspect 1, the reference picture determination unit refers to the reference layer in a reference picture set indicating a picture to be referred to when decoding the decoded picture of the reference layer. The above determination may be performed depending on whether or not a picture exists.
 参照ピクチャセットは、ピクチャ単位に導出される。よって、対象ピクチャに含まれるスライスの復号において共通して用いられる。 The reference picture set is derived for each picture. Therefore, it is commonly used in decoding of slices included in the target picture.
 上記構成のように、参照ピクチャセットに基づいて上記参照レイヤの上記参照ピクチャの存否を判定することで、スライス間で共通の基準に基づいて上記参照レイヤの上記参照ピクチャの存否を判定することができる。 As in the above configuration, by determining the presence or absence of the reference picture of the reference layer based on a reference picture set, it is possible to determine the presence or absence of the reference picture of the reference layer based on a common criterion between slices. it can.
 本発明の態様3に係る画像復号装置では、上記態様1において、上記参照ピクチャ判定部は、上記対象ピクチャの所定番目のIスライスでないスライスの参照ピクチャリストの所定の位置において、上記参照レイヤの参照ピクチャが存在するか否かに応じて、上記判定を行ってもよい。 In the image decoding device according to aspect 3 of the present invention, in the aspect 1, the reference picture determination unit refers to the reference layer at a predetermined position in a reference picture list of a slice that is not the predetermined I-th slice of the target picture. The above determination may be performed depending on whether or not a picture exists.
 上記構成によれば、対象ピクチャに含まれるスライスの間において、上記対象ピクチャの所定番目のIスライスでないスライスの参照ピクチャリストの所定の位置を判定に用いるため、判定処理を共通化することができる。 According to the above configuration, since a predetermined position in the reference picture list of a slice that is not the predetermined I-th slice of the target picture is used for determination between slices included in the target picture, the determination process can be shared. .
 本発明の態様4に係る画像復号装置では、上記態様1において、上記参照ピクチャ判定部は、上記参照レイヤの参照ピクチャがDPB(Decoded Picture Buffer)において格納されているか否かに応じて、上記判定を行ってもよい。 In the image decoding device according to aspect 4 of the present invention, in the aspect 1, the reference picture determination unit performs the determination according to whether or not the reference picture of the reference layer is stored in a DPB (Decoded Picture Buffer). May be performed.
 このように、復号済みのピクチャが保持されるDPBに上記参照レイヤの参照ピクチャが定義されているか否かに応じて判定を行うことも可能である。 Thus, it is also possible to make a determination according to whether or not the reference picture of the reference layer is defined in the DPB that holds the decoded picture.
 本発明の態様5に係る画像復号装置では、上記態様4において、上記参照ピクチャ判定部は、上記参照レイヤの参照ピクチャのDPB(Decoded Picture Buffer)における参照マークが、「参照使用」であるか否かに応じて、上記判定を行ってもよい。 In the image decoding device according to aspect 5 of the present invention, in the aspect 4, the reference picture determination unit determines whether or not the reference mark in the DPB (Decoded Picture Buffer) of the reference picture of the reference layer is “reference use”. Depending on whether or not, the above determination may be made.
 デコーダによっては、HRD(Hypothetical Reference Decoder)で規定されるようなDPB処理をサポートしていないものもあるため、DPBの状態は不正確な場合がある。よって、上記参照レイヤの参照ピクチャのDPB(Decoded Picture Buffer)における参照マークを確認することにより、上記参照レイヤの参照ピクチャが利用可能か否かの上記判定をより正確に行うことができる。 Some decoders do not support DPB processing as defined by HRD (Hypothetical Reference Decoder), so the DPB state may be inaccurate. Therefore, by checking the reference mark in the DPB (Decoded Picture Buffer) of the reference picture of the reference layer, it is possible to more accurately determine whether or not the reference picture of the reference layer is usable.
 すなわち、参照マークが、「参照使用」である場合、上記参照レイヤの参照ピクチャが存在すると判定すればよい。 That is, when the reference mark is “reference use”, it may be determined that the reference picture of the reference layer exists.
 本発明の態様6に係る画像復号装置は、動き補償画像に対象ピクチャとは異なるレイヤの参照レイヤを用いた残差予測を適用することで対象ピクチャの予測画像を生成する画像復号装置において、上記対象ピクチャの所定番目のIスライスでないスライスの参照ピクチャリストの所定の位置から、上記対象ピクチャが属する対象レイヤに属しており、上記対象ピクチャのピクチャ順序と異なるピクチャ順序を有する参照ピクチャを取得する参照ピクチャ取得部(参照画像取得部30922)と、上記取得した参照ピクチャのピクチャ順序と同一のピクチャ順序を有する上記参照レイヤの参照ピクチャと、上記対象ピクチャのピクチャ順序と同一のピクチャ順序を有する参照レイヤの復号済ピクチャとに基づく上記残差予測を、上記動き補償画像に適用する残差予測適用部(残差合成部30923)と、を備える。 The image decoding device according to aspect 6 of the present invention is an image decoding device that generates a predicted image of a target picture by applying residual prediction using a reference layer different from the target picture to the motion compensated image. Reference for obtaining a reference picture belonging to a target layer to which the target picture belongs and having a picture order different from the picture order of the target picture, from a predetermined position in a reference picture list of a slice that is not the predetermined I-th slice of the target picture A picture acquisition unit (reference image acquisition unit 30922), a reference picture of the reference layer having the same picture order as the picture order of the acquired reference picture, and a reference layer having the same picture order as the picture order of the target picture The residual prediction based on the decoded picture of Comprising residual prediction application unit that applies to the image (the residual synthesis section 30923), the.
 上記構成によれば、上記対象ピクチャが属する対象レイヤに属しており、上記対象ピクチャのピクチャ順序と異なるピクチャ順序を有する参照ピクチャ(いわゆるARP参照ピクチャ)を、上記対象ピクチャの所定番目のIスライスでないスライスの参照ピクチャリストの所定の位置から取得する。 According to the above configuration, a reference picture (a so-called ARP reference picture) belonging to the target layer to which the target picture belongs and having a picture order different from the picture order of the target picture is not a predetermined I-th slice of the target picture. Obtained from a predetermined position in the reference picture list of the slice.
 上記対象ピクチャの所定番目のIスライスでないスライスは、対象ピクチャの第1番目のIスライスでないスライスであることが好ましい。また、参照ピクチャリストの所定の位置は、参照ピクチャリストの第0番目の位置(先頭)であってもよい。対象ピクチャに含まれるスライス間で共通であれば、スライスの番号、参照ピクチャリストの位置は、任意である。 The slice that is not the predetermined I-slice of the target picture is preferably a slice that is not the first I-slice of the target picture. Further, the predetermined position of the reference picture list may be the 0th position (first position) of the reference picture list. If the slices included in the target picture are common, the slice number and the position of the reference picture list are arbitrary.
 上記構成によれば、対象ピクチャに含まれるスライス間で、ARP参照ピクチャの判定を共通化することができる。 According to the above configuration, the determination of the ARP reference picture can be made common among the slices included in the target picture.
 本発明の態様7に係る画像復号装置は、動き補償画像に対象ピクチャとは異なるレイヤの参照レイヤを用いた残差予測を適用することで対象ピクチャの予測画像を生成する画像復号装置において、上記残差予測を実行することを指示する残差予測実行フラグを復号するフラグ復号部(インター予測パラメータ復号部303)と、上記対象ピクチャのピクチャ順序と異なるピクチャ順序を有する上記参照レイヤの参照ピクチャが利用可能となっているビットストリームを受信する受信部と、上記残差予測実行フラグに従って、上記残差予測を実行する残差予測実行部(残差予測部3092)と、を備える。 The image decoding device according to aspect 7 of the present invention is an image decoding device that generates a predicted image of a target picture by applying residual prediction using a reference layer different from the target picture to the motion compensated image. A flag decoding unit (inter prediction parameter decoding unit 303) that decodes a residual prediction execution flag that instructs execution of residual prediction, and a reference picture of the reference layer that has a picture order different from the picture order of the target picture. A receiving unit that receives a usable bitstream, and a residual prediction execution unit (residual prediction unit 3092) that executes the residual prediction according to the residual prediction execution flag.
 本発明の態様8に係る画像符号化装置は、動き補償画像に対象ピクチャとは異なるレイヤの参照レイヤを用いた残差予測を適用することで対象ピクチャの予測画像を生成する画像符号化装置において、上記残差予測を実行することを指示する残差予測実行フラグを符号化するフラグ符号化部(予測パラメータ符号化部111)と、上記対象ピクチャのピクチャ順序と異なるピクチャ順序を有する上記参照レイヤの参照ピクチャが利用可能となっているビットストリームを生成するビットストリーム生成部(予測パラメータ符号化部111)と、上記生成したビットストリームを画像復号装置に対して送信するビットストリーム送信部(エントロピー符号化部104)と、を備える。 An image encoding device according to aspect 8 of the present invention is an image encoding device that generates a predicted image of a target picture by applying residual prediction using a reference layer different from the target picture to the motion compensated image. A flag encoding unit (prediction parameter encoding unit 111) that encodes a residual prediction execution flag that instructs execution of the residual prediction, and the reference layer having a picture order different from the picture order of the target picture A bit stream generation unit (prediction parameter encoding unit 111) that generates a bit stream in which the reference picture can be used, and a bit stream transmission unit (entropy code) that transmits the generated bit stream to the image decoding apparatus And a conversion unit 104).
 上記態様7に係る画像復号装置または上記態様8に係る画像符号化装置は、上記態様1に係る画像復号装置に対応する特徴点を含む。よって、上記態様7に係る画像復号装置または上記態様8に係る画像符号化装置によっても、上記態様1に係る画像復号装置と同様の効果を得ることができる。 The image decoding device according to aspect 7 or the image encoding device according to aspect 8 includes feature points corresponding to the image decoding device according to aspect 1. Therefore, the image decoding device according to aspect 7 or the image encoding device according to aspect 8 can achieve the same effects as those of the image decoding device according to aspect 1.
 本発明の態様9に係る画像復号装置は、動き補償画像に対象ピクチャとは異なるレイヤの参照レイヤを用いた残差予測を適用することで対象ピクチャの予測画像を生成する画像復号装置において、参照ピクチャリストに含まれる参照ピクチャのうち、上記対象ピクチャが属する対象レイヤに属しており、上記対象ピクチャのピクチャ順序と異なるピクチャ順序を有する参照ピクチャを選択する参照ピクチャ選択部と、上記選択された参照ピクチャのピクチャ順序と同一のピクチャ順序を有する上記参照レイヤの参照ピクチャと、上記対象ピクチャのピクチャ順序と同一のピクチャ順序を有する参照レイヤの復号済ピクチャとに基づく上記残差予測を、上記動き補償画像に適用する残差予測適用部と、を備えることを特徴とする。 The image decoding device according to aspect 9 of the present invention is an image decoding device that generates a prediction image of a target picture by applying residual prediction using a reference layer of a layer different from the target picture to the motion compensated image. Among the reference pictures included in the picture list, a reference picture selection unit that selects a reference picture that belongs to the target layer to which the target picture belongs and has a picture order different from the picture order of the target picture, and the selected reference The residual prediction based on the reference picture of the reference layer having the same picture order as the picture order of the picture and the decoded picture of the reference layer having the same picture order as the picture order of the target picture, and the motion compensation A residual prediction application unit to be applied to the image.
 上記構成によれば、arpRefpicのPOCが現ピクチャのPOCと異なることを保障することができる。すなわち、選択された参照ピクチャのピクチャ順序と同一のピクチャ順序を有する上記参照レイヤの参照ピクチャのピクチャ順序と、上記対象ピクチャのピクチャ順序と同一のピクチャ順序を有する参照レイヤの復号済ピクチャのピクチャ順序とが異なることを保障することができる。 According to the above configuration, it is possible to ensure that the arpRefpic POC is different from the POC of the current picture. That is, the picture order of the reference picture of the reference layer having the same picture order as the picture order of the selected reference picture, and the picture order of the decoded picture of the reference layer having the same picture order as the picture order of the target picture Can be guaranteed to be different.
 本発明は上述した各実施形態に限定されるものではなく、請求項に示した範囲で種々の変更が可能であり、異なる実施形態にそれぞれ開示された技術的手段を適宜組み合わせて得られる実施形態についても本発明の技術的範囲に含まれる。さらに、各実施形態にそれぞれ開示された技術的手段を組み合わせることにより、新しい技術的特徴を形成することができる。 The present invention is not limited to the above-described embodiments, and various modifications are possible within the scope shown in the claims, and embodiments obtained by appropriately combining technical means disclosed in different embodiments. Is also included in the technical scope of the present invention. Furthermore, a new technical feature can be formed by combining the technical means disclosed in each embodiment.
 本発明は、画像データが符号化された符号化データを復号する画像復号装置、および、画像データが符号化された符号化データを生成する画像符号化装置に好適に適用することができる。また、画像符号化装置によって生成され、画像復号装置によって参照される符号化データのデータ構造に好適に適用することができる。 The present invention can be suitably applied to an image decoding apparatus that decodes encoded data obtained by encoding image data and an image encoding apparatus that generates encoded data obtained by encoding image data. Further, the present invention can be suitably applied to the data structure of encoded data generated by an image encoding device and referenced by the image decoding device.
1…画像伝送システム
11…画像符号化装置
101…予測画像生成部
102…減算部
103…DCT・量子化部
104…エントロピー符号化部
105…逆量子化・逆DCT部
106…加算部
109…復号ピクチャ管理部
110…符号化パラメータ決定部
111…予測パラメータ符号化部(フラグ符号化部、ビットストリーム生成部)
112…インター予測パラメータ符号化部
1121…マージ予測パラメータ導出部
1122…AMVP予測パラメータ導出部
1123…減算部
1126…予測パラメータ統合部
113…イントラ予測パラメータ符号化部
21…ネットワーク
31…画像復号装置
301…エントロピー復号部
302…予測パラメータ復号部
303…インター予測パラメータ復号部(フラグ復号部)
3031…インター予測パラメータ復号制御部
3032…AMVP予測パラメータ導出部
3035…加算部
304…イントラ予測パラメータ復号部
306…復号ピクチャ管理部
3061…DPB
3062…RPS導出部
3063…参照ピクチャ制御部
3064…参照レイヤピクチャ制御部
3065…RPL導出部
3066…出力制御部
3067…予測パラメータメモリ
308…予測画像生成部
309…インター予測画像生成部
3091…変位補償部
3092…残差予測部(残差予測実行部)
30921…残差予測実施フラグ導出部
30922…参照画像取得部(参照ピクチャ取得部)
30923…残差合成部(残差予測適用部)
3093…照度補償部
3094…重み予測部
3095…参照画像判定部(参照ピクチャ判定部)
310…イントラ予測画像生成部
311…逆量子化・逆DCT部
312…加算部
313…残差格納部
41…画像表示装置
DESCRIPTION OF SYMBOLS 1 ... Image transmission system 11 ... Image encoding apparatus 101 ... Prediction image generation part 102 ... Subtraction part 103 ... DCT / quantization part 104 ... Entropy encoding part 105 ... Inverse quantization / inverse DCT part 106 ... Addition part 109 ... Decoding Picture management unit 110 ... coding parameter determination unit 111 ... prediction parameter coding unit (flag coding unit, bit stream generation unit)
DESCRIPTION OF SYMBOLS 112 ... Inter prediction parameter encoding part 1121 ... Merge prediction parameter derivation part 1122 ... AMVP prediction parameter derivation part 1123 ... Subtraction part 1126 ... Prediction parameter integration part 113 ... Intra prediction parameter encoding part 21 ... Network 31 ... Image decoding apparatus 301 ... Entropy decoding unit 302 ... Prediction parameter decoding unit 303 ... Inter prediction parameter decoding unit (flag decoding unit)
3031 ... Inter prediction parameter decoding control unit 3032 ... AMVP prediction parameter derivation unit 3035 ... Adder 304 ... Intra prediction parameter decoding unit 306 ... Decoded picture management unit 3061 ... DPB
3062 ... RPS derivation unit 3063 ... Reference picture control unit 3064 ... Reference layer picture control unit 3065 ... RPL derivation unit 3066 ... Output control unit 3067 ... Prediction parameter memory 308 ... Prediction image generation unit 309 ... Inter prediction image generation unit 3091 ... Displacement compensation Unit 3092 ... Residual prediction unit (residual prediction execution unit)
30921 ... Residual prediction execution flag deriving unit 30922 ... Reference image acquisition unit (reference picture acquisition unit)
30923 ... Residual synthesis unit (residual prediction application unit)
3093: Illuminance compensation unit 3094 ... Weight prediction unit 3095 ... Reference image determination unit (reference picture determination unit)
310 ... Intra predicted image generation unit 311 ... Inverse quantization / inverse DCT unit 312 ... Adder 313 ... Residual storage unit 41 ... Image display device

Claims (6)

  1.  残差予測参照ピクチャが利用可能であるか否かの判定を行う参照ピクチャ判定部と、
    上記残差予測参照ピクチャを用いた残差予測を行う残差予測適用部とを備え
     上記参照ピクチャ判定部は、上記残差予測参照レイヤの参照ピクチャがDPB(Decoded Picture Buffer)において格納されているか否かに応じて、上記判定を行うことを特徴とする画像復号装置。
    A reference picture determination unit that determines whether a residual prediction reference picture is available;
    A residual prediction applying unit that performs residual prediction using the residual prediction reference picture, wherein the reference picture determination unit is configured to determine whether a reference picture of the residual prediction reference layer is stored in a DPB (Decoded Picture Buffer) An image decoding apparatus that performs the above determination according to whether or not.
  2.  上記参照ピクチャ判定部は、上記参照レイヤの参照ピクチャがDPB(Decoded Picture Buffer)において、参照可能か否かを示す情報にもとづいて、上記判定を行うことを特徴とする請求項1に記載の画像復号装置。 2. The image according to claim 1, wherein the reference picture determination unit performs the determination based on information indicating whether or not the reference picture of the reference layer can be referred to in a DPB (Decoded Picture Buffer). Decoding device.
  3.  残差予測用参照ピクチャを導出する参照ピクチャ導出部と、上記残差予測参照ピクチャを用いた残差予測を行う残差予測適用部とを備え、
     上記参照ピクチャ導出部は、参照ピクチャリストに含まれる参照ピクチャのうち、対象ピクチャのピクチャ順序と異なるピクチャ順序を有する参照ピクチャを上記残差予測用参照ピクチャとして導出することを特徴とする画像復号装置。
    A reference picture deriving unit for deriving a reference picture for residual prediction, and a residual prediction applying unit for performing residual prediction using the residual prediction reference picture,
    The reference picture deriving unit derives, as the residual prediction reference picture, a reference picture having a picture order different from the picture order of the target picture among the reference pictures included in the reference picture list .
  4.  上記参照ピクチャ導出部は、参照ピクチャリストを先頭から順に走査し、参照ピクチャRefPicListX[i]のPOCと対象ピクチャcurrPicのPOC(PicOrderCntVal)が異なる場合の参照ピクチャを上記残差予測用参照ピクチャとして導出することを特徴とする請求項3に記載の画像復号装置。 The reference picture deriving unit sequentially scans the reference picture list from the top, and derives a reference picture when the POC of the reference picture RefPicListX [i] and the POC (PicOrderCntVal) of the target picture currPic are different as the residual prediction reference picture The image decoding apparatus according to claim 3, wherein:
  5.  残差予測用参照ピクチャを導出する参照ピクチャ選択部と、上記残差予測参照ピクチャが利用可能な場合に、上記残差予測参照ピクチャを用いた残差予測を行う残差予測適用部とを備え、
    参照ピクチャリストを先頭から順に走査し、参照ピクチャRefPicListX[i]のPOCと、対象ピクチャcurrPicのPOC(PicOrderCntVal)の差分の絶対値が、上記POC差分よりも小さい場合に、その参照ピクチャを残差予測用参照ピクチャに設定することを特徴とする画像復号装置。
    A reference picture selection unit that derives a reference picture for residual prediction, and a residual prediction application unit that performs residual prediction using the residual prediction reference picture when the residual prediction reference picture is available ,
    The reference picture list is scanned in order from the top, and if the absolute value of the difference between the POC of the reference picture RefPicListX [i] and the POC (PicOrderCntVal) of the target picture currPic is smaller than the POC difference, the reference picture is determined as a residual. An image decoding apparatus characterized by being set to a reference picture for prediction.
  6.  参照ピクチャリストに含まれる参照ピクチャのうち、上記対象ピクチャのピクチャ順序と異なるピクチャ順序を有する参照ピクチャを残差予測参照ピクチャとして導出する参照ピクチャ導出部と、
     上記残差予測参照ピクチャが利用可能な場合に、
    残差予測フラグiv_res_pred_weight_idxを復号するインター予測パラメータ復号制御部を備える画像復号装置。
    A reference picture deriving unit for deriving a reference picture having a picture order different from the picture order of the target picture among the reference pictures included in the reference picture list as a residual prediction reference picture;
    When the residual prediction reference picture is available,
    An image decoding apparatus including an inter prediction parameter decoding control unit that decodes a residual prediction flag iv_res_pred_weight_idx.
PCT/JP2014/077530 2013-10-16 2014-10-16 Image decoding device WO2015056735A1 (en)

Priority Applications (1)

Application Number Priority Date Filing Date Title
JP2015542651A JP6401707B2 (en) 2013-10-16 2014-10-16 Image decoding apparatus, image decoding method, and recording medium

Applications Claiming Priority (2)

Application Number Priority Date Filing Date Title
JP2013215881 2013-10-16
JP2013-215881 2013-10-16

Publications (1)

Publication Number Publication Date
WO2015056735A1 true WO2015056735A1 (en) 2015-04-23

Family

ID=52828176

Family Applications (1)

Application Number Title Priority Date Filing Date
PCT/JP2014/077530 WO2015056735A1 (en) 2013-10-16 2014-10-16 Image decoding device

Country Status (2)

Country Link
JP (1) JP6401707B2 (en)
WO (1) WO2015056735A1 (en)

Family Cites Families (4)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US10034018B2 (en) * 2011-09-23 2018-07-24 Velos Media, Llc Decoded picture buffer management
US9615090B2 (en) * 2012-12-28 2017-04-04 Qualcomm Incorporated Parsing syntax elements in three-dimensional video coding
WO2014104242A1 (en) * 2012-12-28 2014-07-03 シャープ株式会社 Image decoding device and image encoding device
US9380305B2 (en) * 2013-04-05 2016-06-28 Qualcomm Incorporated Generalized residual prediction in high-level syntax only SHVC and signaling and management thereof

Non-Patent Citations (4)

* Cited by examiner, † Cited by third party
Title
GERHARD TECH ET AL.: "MV-HEVC Working Draft 1", JOINT COLLABORATIVE TEAM ON 3D VIDEO CODING EXTENSION DEVELOPMENT OF ITU-T SG 16 WP 3 AND ISO/IEC JTC 1/SC 29/WG 11, JCT3V-A1004_D0, 1ST MEETING, August 2012 (2012-08-01), STOCKHOLM, SE, pages I - III, 1-16 *
LI ZHANG ET AL.: "CE4: Advanced residual prediction for multiview coding", JOINT COLLABORATIVE TEAM ON 3D VIDEO CODING EXTENSIONS OF ITU-T SG 16 WP 3 AND ISO/IEC JTC 1/SC 29/WG 11, JCT3V-D0177, 4TH MEETING, April 2013 (2013-04-01), INCHEON, KR, pages 1 - 9 *
LI ZHANG ET AL.: "Further improvements on advanced residual prediction", JOINT COLLABORATIVE TEAM ON 3D VIDEO CODING EXTENSIONS OF ITU-T SG 16 WP 3 AND ISO/IEC JTC 1/SC 29/WG 11, JCT3V-E0124, 5TH MEETING, July 2013 (2013-07-01), VIENNA, AU, pages 1 - 6 *
TOMOHIRO IKAI: "CE4-related: ARP reference picture selection and its availability check", JOINT COLLABORATIVE TEAM ON 3D VIDEO CODING EXTENSIONS OF ITU-T SG 16 WP 3 AND ISO/IEC JTC 1/SC 29/WG 11, JCT3V-FOI05_R1, 6TH MEETING, October 2013 (2013-10-01), GENEVA, CH, pages 1 - 6 *

Also Published As

Publication number Publication date
JP6401707B2 (en) 2018-10-10
JPWO2015056735A1 (en) 2017-03-09

Similar Documents

Publication Publication Date Title
US10200712B2 (en) Merge candidate derivation device, image decoding device, and image coding device
US10306235B2 (en) Image decoding apparatus, image coding apparatus, and prediction-vector deriving device
US9967592B2 (en) Block-based advanced residual prediction for 3D video coding
JP6469588B2 (en) Residual prediction device, image decoding device, image coding device, residual prediction method, image decoding method, and image coding method
KR102060857B1 (en) Simplified advanced motion prediction for 3d-hevc
JP6397421B2 (en) Image decoding apparatus and image encoding apparatus
US20150326866A1 (en) Image decoding device and data structure
KR20140147102A (en) Motion vector coding and bi-prediction in hevc and its extensions
US20160191933A1 (en) Image decoding device and image coding device
WO2015056719A1 (en) Image decoding device and image coding device
JP6473078B2 (en) Image decoding device
WO2015056620A1 (en) Image decoding device and image coding device
JP6118199B2 (en) Image decoding apparatus, image encoding apparatus, image decoding method, image encoding method, and computer-readable recording medium
WO2014103600A1 (en) Encoded data structure and image decoding device
JP6401707B2 (en) Image decoding apparatus, image decoding method, and recording medium
JPWO2015141696A1 (en) Image decoding apparatus, image encoding apparatus, and prediction apparatus
JP2016066864A (en) Image decoding device, image encoding device, and merge mode parameter derivation device
JP2017212477A (en) Image decoding apparatus, image encoding apparatus, displacement array deriving apparatus, displacement vector deriving apparatus, default reference view index deriving apparatus, and depth lookup table deriving apparatus
JP2015015626A (en) Image decoder and image encoder
JP2014204327A (en) Image decoding apparatus and image encoding apparatus
JP2015080053A (en) Image decoder and image encoder

Legal Events

Date Code Title Description
121 Ep: the epo has been informed by wipo that ep was designated in this application

Ref document number: 14854489

Country of ref document: EP

Kind code of ref document: A1

ENP Entry into the national phase

Ref document number: 2015542651

Country of ref document: JP

Kind code of ref document: A

NENP Non-entry into the national phase

Ref country code: DE

122 Ep: pct application non-entry in european phase

Ref document number: 14854489

Country of ref document: EP

Kind code of ref document: A1