WO2012043678A1 - Image decoding device, image encoding device, and data structure - Google Patents

Image decoding device, image encoding device, and data structure Download PDF

Info

Publication number
WO2012043678A1
WO2012043678A1 PCT/JP2011/072283 JP2011072283W WO2012043678A1 WO 2012043678 A1 WO2012043678 A1 WO 2012043678A1 JP 2011072283 W JP2011072283 W JP 2011072283W WO 2012043678 A1 WO2012043678 A1 WO 2012043678A1
Authority
WO
WIPO (PCT)
Prior art keywords
prediction
unit
prediction mode
target
partition
Prior art date
Application number
PCT/JP2011/072283
Other languages
French (fr)
Japanese (ja)
Inventor
山本 智幸
将伸 八杉
知宏 猪飼
Original Assignee
シャープ株式会社
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Application filed by シャープ株式会社 filed Critical シャープ株式会社
Publication of WO2012043678A1 publication Critical patent/WO2012043678A1/en

Links

Images

Classifications

    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04NPICTORIAL COMMUNICATION, e.g. TELEVISION
    • H04N19/00Methods or arrangements for coding, decoding, compressing or decompressing digital video signals
    • H04N19/10Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using adaptive coding
    • H04N19/134Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using adaptive coding characterised by the element, parameter or criterion affecting or controlling the adaptive coding
    • H04N19/146Data rate or code amount at the encoder output
    • H04N19/152Data rate or code amount at the encoder output by measuring the fullness of the transmission buffer
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04NPICTORIAL COMMUNICATION, e.g. TELEVISION
    • H04N19/00Methods or arrangements for coding, decoding, compressing or decompressing digital video signals
    • H04N19/50Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using predictive coding

Definitions

  • the present invention relates to an image decoding device that decodes encoded data, and an image encoding device that generates encoded data.
  • the present invention also relates to a data structure of encoded data generated by the image encoding device and referenced by the image decoding device.
  • a moving image encoding device that generates encoded data by encoding the moving image, and decoding the encoded data
  • a video decoding device image decoding device that generates a decoded image
  • As a specific moving picture encoding method for example, H.264 is used. H.264 / MPEG-4.
  • Adopted in the KTA software which is a codec for joint development in AVC (Non-Patent Document 1) and VCEG (Video Coding Expert Group), and TMuC (Test Model Under Consideration) software, the successor codec The method that has been used.
  • an image (picture) constituting a moving image includes a slice obtained by dividing the image, a coding unit obtained by dividing the slice (macroblock or maximum coding unit (Largest Coding) And a hierarchical structure composed of blocks and partitions obtained by dividing the coding unit.
  • a predicted image is usually generated based on a locally decoded image obtained by encoding / decoding an input image, and the predicted image is subtracted from the input image (original image).
  • the prediction residual (which may be referred to as “difference image” or “residual image”) is encoded.
  • examples of the method for generating a predicted image include inter-screen prediction (inter prediction) and intra-screen prediction (intra prediction).
  • a predicted image in a frame being decoded is generated for each prediction unit by applying motion compensation using a motion vector with the decoded frame as a reference frame.
  • intra prediction a predicted image in a frame being decoded is generated for each prediction unit based on a decoded area of the frame being decoded.
  • a predetermined prediction mode group also referred to as “basic prediction mode group”
  • a method for generating a pixel value on the prediction unit by selecting a prediction mode and (2) extrapolating a pixel value of a decoded area in an extrapolation direction (prediction direction) corresponding to the selected prediction mode (“ Sometimes referred to as “basic prediction”).
  • Non-Patent Document 2 for each prediction unit, the edge direction in the prediction unit is estimated based on the pixel values of pixels located around the prediction unit, and the pixel value of the decoded area is extrapolated to the estimated edge direction.
  • a method for generating a pixel value on the prediction unit a method called Differential Coding of Mode Intra Modes (DCIM), sometimes called “edge prediction” or “edge based prediction” is disclosed.
  • DCIM Differential Coding of Mode Intra Modes
  • FIG. 24A is a diagram schematically illustrating a prediction unit (CurrentBlock) to be processed and pixels located around the prediction unit.
  • an edge vector a i 1 to N, N is the total number of surrounding pixels to be referred to).
  • a Sobel operator also referred to as a Sobel filter
  • an argument ⁇ * argmaxS ( ⁇ ) that maximizes the function S ( ⁇ ) is calculated, and the direction represented by ⁇ * is set as the prediction direction.
  • the pixel value on the prediction unit is generated by extrapolating the pixel value of the decoded area in the prediction direction.
  • the argument ⁇ * that maximizes the function S ( ⁇ ) represents an estimated value of the edge direction in the prediction unit to be processed, and is also referred to as “neighbors' suggested prediction direction”. Note that the calculation of the argument ⁇ * is performed in both the encoding device and the decoding device, and thus the argument ⁇ * itself is not encoded.
  • the direction represented by ⁇ * + ⁇ can be used as the prediction direction.
  • indicates direction adjustment, and ⁇ used in the encoding device needs to be encoded and transmitted to the decoding device.
  • the decoding apparatus sets ⁇ * + k ⁇ ⁇ as the prediction direction.
  • FIG. 24C shows a parameter set including the syntax adjust_neighb_dir for specifying the quantization index k. In the example shown in FIG.
  • Non-Patent Document 2 H. H.264 / MPEG-4.
  • the number of selectable prediction modes increases, so that the prediction accuracy can be improved.
  • Non-Patent Document 2 requires a syntax for designating one of the one or more prediction modes added to the basic prediction mode group, the encoded data There is also an aspect that the amount of codes increases. For this reason, when the additional prediction mode is frequently used, there has been a problem that the encoding efficiency may not be improved, or the encoding efficiency may not be improved as expected.
  • the present invention has been made in view of the above-described problems, and the object of the present invention is to provide a code amount of encoded data while maintaining high prediction accuracy even when a prediction mode is added to the basic prediction mode group. It is to realize an image encoding device and an image decoding device capable of improving the encoding efficiency by suppressing the increase in the image quality.
  • an image decoding apparatus is a prediction mode belonging to a prediction mode group in a prediction residual decoded from encoded data together with prediction mode designation information and other side information,
  • the image decoding apparatus that generates a decoded image by adding the prediction images generated according to the prediction mode specified by the prediction mode specification information, the number N of prediction modes to be added to the prediction mode group related to the target prediction unit is used as the side information.
  • a setting unit that is an encoding parameter that is included and is set with reference to an encoding unit that includes the target prediction unit or an encoding parameter that is decoded prior to the encoding unit; Prediction parameters corresponding to each of the N prediction modes to be added to the prediction mode group related to the target prediction unit are decoded. It is characterized in that it comprises a derivation means for deriving from the decoded region of the image, a.
  • the prediction unit is a type that increases the prediction accuracy by increasing the number of prediction modes belonging to the prediction mode group, and the coding efficiency is improved by reducing the number of prediction modes belonging to the prediction mode group and improving the coding efficiency. Some types improve efficiency. According to the knowledge obtained by the inventors, whether a certain prediction unit belongs to the former type or the latter type is usually encoded that is decoded before the prediction unit or the encoding unit. It can be determined from the coding parameters for the unit.
  • the coding parameter is a parameter that is referred to in order to generate a decoded image, a locally decoded image, or a predicted image, and is referred to in a motion vector or intra prediction that is referred to in inter-screen prediction.
  • the prediction parameters such as the prediction mode, the size and shape of the partition, the size and shape of the block, the size and shape of the encoding unit, and residual data between the original image and the predicted image are included.
  • the side information refers to a set of all information except for the residual data among the information included in the encoding parameter.
  • the encoding parameter related to the encoding unit including the target prediction unit is related to the encoding parameter associated with the encoding unit and one or a plurality of prediction units included in the encoding unit. Encoding parameters are included.
  • the prediction unit may be a PU (Prediction Unit) described in the embodiment, or may be a partition obtained by dividing the PU.
  • the coding unit may be a maximum coding unit (LCU: Largegest Coding Unit) described in the embodiment, or a coding unit (CU: Coding Unit) obtained by dividing an LCU. Also good.
  • the image coding apparatus designates the selected prediction mode based on the prediction residual obtained by subtracting the predicted image generated according to the prediction mode selected from the prediction mode group for each prediction unit from the original image.
  • the number N of prediction modes to be added to the prediction mode group related to the target prediction unit is encoded in the side information.
  • a parameter setting unit that refers to a coding unit that includes the target prediction unit or a coding parameter that is encoded prior to the coding unit; and the target prediction Deriving prediction parameters corresponding to each of the N prediction modes to be added to the prediction mode group related to the unit from the decoded region of the locally decoded image It is characterized in that comprises a derivation means that, the.
  • the prediction unit is a type that increases the prediction accuracy by increasing the number of prediction modes belonging to the prediction mode group, and the coding efficiency is improved by reducing the number of prediction modes belonging to the prediction mode group and improving the coding efficiency. Some types improve efficiency. According to the knowledge obtained by the inventors, whether a certain prediction unit belongs to the former type or the latter type is usually a code encoded before the prediction unit or the encoding unit. It can be determined from the encoding parameters for the encoding unit.
  • the encoding efficiency is increased by referring to the encoding parameter including the target prediction unit or the encoding parameter related to the encoding unit encoded before the encoding unit.
  • the number of prediction modes to be added to the prediction mode group related to the target prediction unit can be set. Therefore, according to said structure, encoding efficiency can be improved by suppressing the increase in the code amount of coding data, maintaining high prediction accuracy.
  • the coding parameter is a parameter that is referred to in order to generate a decoded image, a locally decoded image, or a predicted image, and is referred to in a motion vector or intra prediction that is referred to in inter-screen prediction.
  • the prediction parameters such as the prediction mode, the size and shape of the partition, the size and shape of the block, the size and shape of the encoding unit, and residual data between the original image and the predicted image are included.
  • the side information refers to a set of all information except for the residual data among the information included in the encoding parameter.
  • the encoding parameter related to the encoding unit including the target prediction unit is related to the encoding parameter associated with the encoding unit and one or a plurality of prediction units included in the encoding unit. Encoding parameters are included.
  • the prediction unit may be a PU (Prediction Unit) described in the embodiment, or may be a partition obtained by dividing the PU.
  • the coding unit may be a maximum coding unit (LCU: Largegest Coding Unit) described in the embodiment, or a coding unit (CU: Coding Unit) obtained by dividing an LCU. Also good.
  • the data structure of the encoded data according to the present invention includes a prediction residual obtained by subtracting, from the original image, a prediction image generated according to a prediction mode selected from a prediction mode group for each prediction unit.
  • the side information includes the number of prediction modes to be added to the prediction mode group related to the target prediction unit, which is a data structure of encoded data generated by encoding together with the prediction mode specifying information to be specified and other side information. Coding parameters that are expressed implicitly by a coding unit that includes the target prediction unit or a coding parameter that is coded before the coding unit. It is characterized by.
  • the image decoding apparatus includes the prediction mode belonging to the prediction mode group in the prediction residual decoded from the encoded data together with the prediction mode specifying information and other side information, and the prediction mode specifying information.
  • the image decoding device that generates a decoded image by adding the prediction images generated according to the prediction mode specified by the encoding
  • the number N of prediction modes to be added to the prediction mode group related to the target prediction unit is encoded in the side information
  • a parameter setting unit that refers to a coding unit that includes the target prediction unit or a coding parameter that is decoded prior to the coding unit; and the target prediction unit
  • Has a a derivation means for deriving from.
  • the above image decoding apparatus it is possible to improve the encoding efficiency by suppressing an increase in the code amount of the encoded data while maintaining high prediction accuracy.
  • FIG. 4 is a diagram illustrating a configuration of a picture layer of encoded data
  • (b) is a diagram illustrating a configuration of a slice layer included in the picture layer
  • (c) is a configuration of an LCU layer included in the slice layer.
  • (D) is a figure which shows the structure of the leaf CU contained in an LCU layer
  • (e) is a figure which shows the structure of the inter prediction information about leaf CU
  • (f) is It is a figure which shows the structure of the intra prediction information about leaf CU.
  • FIG. 1 It is a flowchart which shows the flow of the prediction image generation process by the moving image decoding apparatus which concerns on the 1st Embodiment of this invention. It is a figure for demonstrating the prediction image production
  • FIG. 1 It is a block diagram which shows the structure of the leaf CU encoding part with which the LCU encoding part which concerns on the 1st Embodiment of this invention is provided. It is a flowchart which shows the flow of the encoding process by the moving image encoder which concerns on the 1st Embodiment of this invention. It is a figure for demonstrating the prediction image generation process by the moving image decoding apparatus which concerns on the 1st Embodiment of this invention, and a moving image encoding apparatus, Comprising: A mode that an edge exists ranging over a target partition and an adjacent partition.
  • the conventional moving image decoding apparatus it is a figure for demonstrating the production
  • FIG. 2B is a block diagram showing a configuration of a receiving device equipped with a video decoding device.
  • BRIEF DESCRIPTION OF THE DRAWINGS It is a figure for demonstrating that the moving image decoding apparatus and moving image encoding apparatus which concern on embodiment of this invention can be utilized for recording and reproduction
  • Embodiments of an image decoding apparatus and an image encoding apparatus according to the present invention will be described below with reference to the drawings.
  • the image decoding apparatus according to the present embodiment decodes a moving image from encoded data. Therefore, hereinafter, this is referred to as “moving image decoding apparatus”.
  • the image encoding device according to the present embodiment generates encoded data by encoding a moving image. Therefore, in the following, this is referred to as a “video encoding device”.
  • the scope of application of the present invention is not limited to this. That is, as will be apparent from the following description, the feature of the present invention lies in intra prediction, and is established without assuming a plurality of frames. That is, the present invention can be applied to general image decoding apparatuses and general image encoding apparatuses regardless of whether the target is a moving image or a still image.
  • the encoded data # 1 has a hierarchical structure including a sequence layer, a GOP (Group Of Pictures) layer, a picture layer, a slice layer, and a maximum coding unit (LCU: Large Coding Unit) layer.
  • GOP Group Of Pictures
  • LCU Large Coding Unit
  • FIG. 2 shows the hierarchical structure below the picture layer in the encoded data # 1.
  • (A) to (f) in FIG. 2 are a picture layer P, a slice layer S, an LCU layer LCU, a leaf CU included in the LCU (denoted as CUL in FIG. 2 (d)), and inter prediction (inter-screen)
  • PI_Inter which is the prediction information PI about (prediction) partition
  • PI_Intra intra prediction information
  • the picture layer P is a set of data that is referenced by the video decoding device 1 in order to decode a target picture that is a processing target picture. As shown in FIG. 2A, the picture layer P includes a picture header PH and slice layers S1 to SNs (Ns is the total number of slice layers included in the picture layer P).
  • the picture header PH includes a coding parameter group referred to by the video decoding device 1 in order to determine a decoding method of the target picture.
  • the encoding mode information (entoropy_coding_mode_flag) indicating the variable length encoding mode used in encoding by the moving image encoding device 2 is an example of an encoding parameter included in the picture header PH.
  • entorpy_coding_mode_flag When entorpy_coding_mode_flag is 0, the picture is encoded by CAVLC (Context-based Adaptive Variable Length Coding). It has become.
  • Each slice layer S included in the picture layer P is a set of data referred to by the video decoding device 1 in order to decode a target slice that is a slice to be processed.
  • the slice layer S includes a slice header SH and LCU layers LCU1 to LCUn (Nc is the total number of LCUs included in the slice S).
  • the slice header SH includes a coding parameter group that the moving image decoding apparatus 1 refers to in order to determine a decoding method of the target slice.
  • Slice type designation information (slice_type) for designating a slice type is an example of an encoding parameter included in the slice header SH.
  • the slice header SH includes a filter parameter FP that is referred to by a loop filter included in the video decoding device 1.
  • slice types that can be specified by the slice type specification information, (1) I slice using only intra prediction at the time of encoding, and (2) P using unidirectional prediction or intra prediction at the time of encoding. Slice, (3) B-slice using unidirectional prediction, bidirectional prediction, or intra prediction at the time of encoding.
  • Each LCU layer LCU included in the slice layer S is a set of data that the video decoding device 1 refers to in order to decode the target LCU that is the processing target LCU.
  • the LCU layer LCU includes an LCU header LCUH and a plurality of coding units (CU: Coding Units) obtained by dividing the LCU by quadtree division.
  • the size that each CU can take depends on the LCU size and hierarchical depth included in the sequence parameter set SPS of the encoded data # 1. For example, when the size of the LCU is 128 ⁇ 128 pixels and the maximum hierarchical depth is 5, the CU included in the LCU has five types of sizes, that is, 128 ⁇ 128 pixels, 64 ⁇ 64 pixels, Any of 32 ⁇ 32 pixels, 16 ⁇ 16 pixels, and 8 ⁇ 8 pixels can be taken. A CU that is not further divided is called a leaf CU.
  • the LCU header LCUH includes an encoding parameter referred to by the video decoding device 1 in order to determine a decoding method of the target LCU. Specifically, as shown in FIG. 2 (c), CU partition information SP_CU that specifies a partition pattern of each target LCU into each leaf CU, and a quantization parameter difference ⁇ qp that specifies the size of the quantization step. (Mb_qp_delta) is included.
  • CU division information SP_CU is information that specifies the shape and size of each CU (and leaf CU) included in the target LCU, and the position in the target LCU.
  • the CU partition information SP_CU does not necessarily need to explicitly include the shape and size of the leaf CU.
  • the CU partition information SP_CU may be a set of flags (split_coding_unit_flag) indicating whether or not the entire LCU or a partial region of the LCU is divided into four. In that case, the shape and size of each leaf CU can be specified by using the shape and size of the LCU together.
  • the quantization parameter difference ⁇ qp is a difference qp ⁇ qp ′ between the quantization parameter qp in the target LCU and the quantization parameter qp ′ in the LCU encoded immediately before the LCU.
  • CU leaf A CU (CU leaf) that cannot be further divided is handled as a prediction unit (PU: Prediction Unit) and a transform unit (TU: Transform Unit).
  • PU Prediction Unit
  • TU Transform Unit
  • the leaf CU (denoted as CUL in (d) of FIG. 2) is (1) PU information PUI that is referred to when the moving image decoding apparatus 1 generates a predicted image. And (2) the TU information TUI that is referred to when the moving image decoding apparatus 1 decodes the residual data.
  • the PU information PUI may include a skip flag SKIP. When the value of the skip flag SKIP is 1, the TU information is omitted.
  • the PU information PUI includes prediction type information PT and prediction information PI, as shown in FIG.
  • the prediction type information PT is information that specifies whether intra prediction or inter prediction is used as a predicted image generation method for the target leaf CU (target PU).
  • the prediction information PI includes intra prediction information PI_Intra or inter prediction information PI_Inter depending on which prediction method is specified by the prediction type information PT.
  • a PU to which intra prediction is applied is also referred to as an intra PU
  • a PU to which inter prediction is applied is also referred to as an inter PU.
  • the PU information PUI includes information specifying the shape and size of each partition included in the target PU and the position in the target PU.
  • the partition is one or a plurality of non-overlapping areas constituting the target leaf CU, and the generation of the predicted image is performed in units of partitions.
  • the TU information TUI includes TU partition information SP_TU that specifies a partition pattern for each block of the target leaf CU (target TU), and quantized prediction residuals QD1 to QDNT (NT Includes the total number of blocks included in the target TU).
  • TU partition information SP_TU is information that specifies the shape and size of each block included in the target TU and the position in the target TU.
  • Each TU can be, for example, a size from 64 ⁇ 64 pixels to 2 ⁇ 2 pixels.
  • the block is one or a plurality of non-overlapping areas constituting the target leaf CU, and prediction residual encoding / decoding is performed in units of TUs or blocks obtained by dividing TUs.
  • Each quantized prediction residual QD is encoded data generated by the moving image encoding apparatus 2 performing the following processes 1 to 3 on a target block that is a processing target block.
  • Process 1 DCT transform (Discrete Cosine Transform) is performed on the prediction residual obtained by subtracting the prediction image from the encoding target image.
  • Process 2 The DCT coefficient obtained in Process 1 is quantized.
  • Process 3 The DCT coefficient quantized in Process 2 is variable length encoded.
  • the inter prediction information PI_Inter includes a coding parameter that is referred to when the video decoding device 1 generates an inter prediction image by inter prediction. As shown in FIG. 2 (e), the inter prediction information PI_Inter includes inter PU partition information SP_Inter that specifies a partition pattern of the target PU into each partition, and inter prediction parameters PP_Inter1 to PP_InterNe (Ne for each partition). , The total number of inter prediction partitions included in the target PU).
  • the inter-PU partition information SP_Inter is information for designating the shape and size of each inter prediction partition included in the target PU (inter PU) and the position in the target PU.
  • the inter PU is composed of four symmetric splittings of 2N ⁇ 2N pixels, 2N ⁇ N pixels, N ⁇ 2N pixels, and N ⁇ N pixels, and 2N ⁇ nU pixels, 2N ⁇ nD pixels, and nL ⁇ 2N. It is possible to divide into 8 types of partitions in total by four asymmetric splits of pixels and nR ⁇ 2N pixels.
  • the specific value of N is defined by the size of the CU to which the PU belongs, and the specific values of nU, nD, nL, and nR are determined according to the value of N.
  • an inter PU of 128 ⁇ 128 pixels is 128 ⁇ 128 pixels, 128 ⁇ 64 pixels, 64 ⁇ 128 pixels, 64 ⁇ 64 pixels, 128 ⁇ 32 pixels, 128 ⁇ 96 pixels, 32 ⁇ 128 pixels, and 96 ⁇ It is possible to divide into 128-pixel inter prediction partitions.
  • the inter prediction parameter PP_Inter includes a reference image index RI, an estimated motion vector index PMVI, and a motion vector residual MVD.
  • the intra prediction information PI_Intra includes an encoding parameter that is referred to when the video decoding device 1 generates an intra predicted image by intra prediction.
  • the intra prediction information PI_Intra includes intra PU partition information SP_Intra that specifies a partition pattern of the target PU (intra PU) into each partition, and intra prediction parameters PP_Intra1 to PP_IntraNa (Na is the total number of intra prediction partitions included in the target PU).
  • the intra-PU partition information SP_Intra is information that specifies the shape and size of each intra-predicted partition included in the target PU, and the position in the target PU.
  • the intra PU split information SP_Intra includes an intra split flag (intra_split_flag) that specifies whether or not the target PU is split into partitions. If the intra partition flag is 1, the target PU is divided symmetrically into four partitions. If the intra partition flag is 0, the target PU is not divided and the target PU itself is one partition.
  • N 2 n , n is an arbitrary integer of 1 or more.
  • a 128 ⁇ 128 pixel intra PU can be divided into 128 ⁇ 128 pixel and 64 ⁇ 64 pixel intra prediction partitions.
  • the intra prediction parameter PP_Intra includes an edge-based prediction flag EF, a sub-direction index AI, an estimation flag MPM, and a residual prediction mode index RIPM as shown in (f) of FIG.
  • the intra prediction parameter PP_Intra is a parameter for designating an intra prediction method (prediction mode) for each partition.
  • the edge-based prediction flag EF is a flag that specifies whether or not the edge-based prediction mode is applied to the target partition that is the processing target partition. In the following description, it is assumed that the edge-based prediction mode is applied to the target partition when the edge-based prediction flag is 1.
  • the correction direction index (also referred to as an additional index) AI is an index included in the prediction parameter PP_Intra when the edge-based prediction mode is applied to the target partition, and is the main direction derived in the edge-based prediction mode. This is an index for designating whether or not to add the correction direction, and for specifying the correction direction to be added when adding the correction direction.
  • the total number of correction directions that can be selected by the additional index AI differs depending on the size of the target partition. For example, when the size of the target partition is 4 ⁇ 4 pixels or 8 ⁇ 8 pixels, the total number of correction directions selectable by the additional index AI is 8, and when the size of the target partition is 16 ⁇ 16 pixels, the additional index When the total number of correction directions that can be selected by AI is 2, and the size of the target partition is 32 ⁇ 32 pixels or 64 ⁇ 64 pixels, the total number of correction directions that can be selected by the additional index AI is 0 (that is, the correction direction is not used). The main direction is always used).
  • the estimation flag MPM is a flag that is included in the prediction parameter PP_Intra when the edge-based prediction mode is not applied to the target partition, and is an estimated prediction mode that is estimated based on the prediction mode assigned to the surrounding partitions of the target partition. It is a flag indicating whether or not the prediction mode for the target partition is the same.
  • the residual prediction mode index RIPM specifies the prediction mode assigned to the target partition when the edge-based prediction mode is not applied to the target partition and the estimated prediction mode and the prediction mode for the target partition are different. It is an index to do.
  • the moving picture decoding apparatus 1 includes H.264 as a part thereof.
  • KTA software which is a codec for joint development in the Video Coding Expert Group (VCEG)
  • TMuC Traffic Model under Consideration
  • the video decoding device 1 generates a prediction image for each prediction unit, generates a decoded image # 2 by adding the generated prediction image and a prediction residual decoded from the encoded data # 1, The generated decoded image # 2 is output to the outside.
  • the generation of the predicted image is performed with reference to the encoding parameter obtained by decoding the encoded data # 1.
  • the encoding parameter is a parameter referred to in order to generate a prediction image, and in addition to a prediction parameter such as a motion vector referred to in inter-screen prediction and a prediction mode referred to in intra-screen prediction. Partition size and shape, block size and shape, and residual data between the original image and the predicted image.
  • side information a set of all information excluding the residual data among the information included in the encoding parameter is referred to as side information.
  • the prediction unit is a partition constituting the LCU
  • the present embodiment is not limited to this, and the prediction unit is a unit larger than the partition.
  • the present invention can also be applied to the case where the prediction unit is a unit smaller than the partition.
  • a frame, a slice, an LCU, a block, and a partition to be decoded are referred to as a target frame, a target slice, a target LCU, a target block, and a target partition, respectively.
  • the LCU size is, for example, 64 ⁇ 64 pixels
  • the partition size is, for example, 64 ⁇ 64 pixels, 32 ⁇ 32 pixels, 16 ⁇ 16 pixels, 8 ⁇ 8 pixels, 4 ⁇ 4 pixels, or the like. These sizes do not limit the present embodiment, and the size and partition of the LCU may be other sizes.
  • FIG. 3 is a block diagram showing a configuration of the moving picture decoding apparatus 1.
  • the video decoding device 1 includes a variable length code demultiplexing unit 11, a header information decoding unit 12, an LCU setting unit 13, an LCU decoding unit 14, and a frame memory 15.
  • the encoded data # 1 input to the video decoding device 1 is input to the variable length code demultiplexing unit 11.
  • the variable-length code demultiplexing unit 11 demultiplexes the input encoded data # 1, thereby converting the encoded data # 1 into header encoded data # 11a that is encoded data related to header information, and a slice.
  • the encoded data # 11a is output to the header information decoding unit 12 and the encoded data # 11b is output to the LCU setting unit 13, respectively.
  • the header information decoding unit 12 decodes the header information # 12 from the encoded header data # 11a.
  • the header information # 12 is (1) information about the size, shape and position of the LCU belonging to the target slice, and (2) the size, shape and shape of the leaf CU belonging to each LCU. And (3) information about the size and shape of the partition belonging to each leaf CU and the position within the target leaf CU.
  • the LCU setting unit 13 separates the encoded data # 11b into encoded data # 13 corresponding to each LCU based on the input header information # 12 and sequentially outputs the encoded data # 11b to the LCU decoding unit 14.
  • the LCU decoding unit 14 generates and outputs a decoded image # 2 corresponding to each LCU by sequentially decoding the encoded data # 13 corresponding to each input LCU.
  • the decoded image # 2 is stored in the frame memory 15. The configuration of the LCU decoding unit 14 will be described later and will not be described here.
  • the decoded image # 2 is recorded in the frame memory 15.
  • decoded images corresponding to all the LCUs decoded before the target LCU are recorded. .
  • the decoded image # corresponding to the encoded data for one frame input to the video decoding device 1 2 generation processing is completed.
  • LCU decoding unit 14 LCU decoding unit 14
  • the LCU decoding unit 14 will be described more specifically with reference to different drawings.
  • FIG. 4 is a block diagram showing a configuration of the LCU decoding unit 14.
  • the LCU decoding unit 14 includes an LCU division structure setting unit 141, a leaf CU scanning unit 142, and a leaf CU decoding unit 143.
  • the LCU division structure setting unit 141 sets the division structure of the target LCU into leaf CUs with reference to the header information # 12.
  • the division structure information # 141 that specifies the division structure of the target LCU into leaf CUs is supplied to the leaf CU scanning unit 142 together with the encoded data # 13 about the target LCU.
  • FIG. 5 is a diagram illustrating an example of a division structure into which the target LCU is divided into leaf CUs and specified by the division structure information # 141.
  • the target LCU is divided into CUL1 to CUL7 that are leaf CUs.
  • the leaf CU scan unit 142 scans the leaf CUs included in the target LCU in a predetermined order (for example, raster scan order), and encodes data # 142a and CU information for the target leaf CU that is the leaf CU to be processed.
  • # 142b is supplied to the leaf CU decoding unit 143.
  • CU information # 142b includes (1) size, shape, and position in the target LCU for the target leaf CU, (2) PU information about the target leaf CU, and (3) TU information about the target leaf CU. , Including.
  • FIG. 5 is a diagram illustrating a case where CUL4 is set as the target leaf CU by the leaf CU scanning unit 142 among the CUL1 to CUL7 which are the leaf CUs included in the target LCU.
  • the leaf CU scanning unit 142 supplies the encoded data for CUL4, which is the leaf CU, and the CU information for CUL4 to the leaf CU decoding unit 143.
  • the leaf CU decoding unit 143 refers to the encoded data and CU information about the target leaf CU, and the decoded pixel values stored in the frame memory 15, and generates a predicted image for the target leaf CU.
  • the prediction residual for the target leaf CU is decoded.
  • the leaf CU decoding unit 143 generates a decoded image for the target leaf CU by adding the generated predicted image and the decoded prediction residual.
  • the leaf CU decoding unit 143 integrates the decoded images for each leaf CU belonging to the target LCU, and generates a decoded image # 2 for the target LCU.
  • FIG. 6 is a block diagram showing a configuration of the leaf CU decoding unit 143.
  • the leaf CU decoding unit 143 includes a PU decoding unit 431, a predicted image generation unit 432, a TU decoding unit 433, a prediction residual restoration unit 434, and a decoded image generation unit 435.
  • the PU decoding unit 431 determines the division pattern for each partition of the target leaf CU by decoding the PU information about the target leaf CU, determines the prediction mode for each partition, and determines each determined prediction mode # 431 is assigned to each partition.
  • the prediction mode # 431 assigned to each partition is supplied to the predicted image generation unit 432.
  • the prediction mode designates a generation method for generating a prediction image by intra-screen prediction (intra prediction).
  • intra prediction intra-screen prediction
  • a basic prediction mode set or edge-based prediction is used for each partition. Any prediction mode is selected from the mode set.
  • the basic prediction mode set is (1) a direction prediction mode for generating a prediction image for the target partition by extrapolating a decoded image around the target partition along a predetermined prediction direction, and (2) A DC prediction mode for generating a prediction image for the target partition by taking an average value of decoded pixel values around the target partition.
  • FIG. 7 is a diagram illustrating prediction modes included in the basic prediction mode set according to the present embodiment, together with a prediction mode index. Moreover, in FIG. 7, the prediction direction of each direction prediction mode is shown. As shown in FIG. 7, the basic prediction mode set in this embodiment includes a direction prediction mode specified by indexes 0, 1, 3 to 8 and a DC prediction mode specified by index 2, respectively.
  • the basic prediction mode set is exemplified as a case where a prediction mode that specifies any of eight different direction predictions is included.
  • the present embodiment is not limited to this. Absent.
  • a set including a prediction mode that specifies any of nine or more different directions may be used as the basic prediction mode set.
  • a set including a prediction mode for designating any of 16 different directions and a prediction mode for designating any of 32 different directions can be given.
  • the edge-based prediction mode set is (1) deriving the main direction for the target partition based on the decoded pixel values around the target partition, and surrounding the target partition along the main direction.
  • An edge-based prediction mode (hereinafter also referred to as a main direction edge-based prediction mode) that generates a predicted image in the target partition by extrapolating the decoded pixel values of (2), and (2)
  • Edge-based prediction mode for deriving a sub-direction by adding direction correction and generating a predicted image in the target partition by extrapolating decoded pixel values around the target partition along the sub-direction (Hereinafter also referred to as a sub-direction edge-based prediction mode).
  • FIG. 1 is a flowchart showing a flow of PU information decoding processing by the PU decoding unit 431.
  • FIG. 8A is a table showing the value of the edge-based prediction flag EF, the value of the additional index AI, the value of the estimation flag MPM, and the value of the residual prediction mode index RIPM that can be taken for the target partition.
  • FIG. 8B is a diagram illustrating the prediction direction corresponding to each value of the decoded additional index AI.
  • the PU decoding unit 431 refers to an intra partition flag (intra_split_flag) that specifies whether or not to divide the target leaf CU into partitions among the PU information about the target leaf CU, and converts the target leaf CU into a plurality of partitions. It is determined whether to divide or handle the target leaf CU as one partition. Specifically, the PU decoding unit 431 divides the target leaf CU symmetrically into four partitions if the intra split flag is 1, and splits the target leaf CU if the intra split flag is 0. It handles as one partition without doing.
  • intra_split_flag an intra partition flag
  • Step S102 the PU decoding unit 431 initializes the value of the loop variable part_id to 0, and starts a loop process that sets the increment value of the loop variable part_id for each loop to 1 for part_id that satisfies part_id ⁇ Npart-1.
  • the loop variable part_id is an index for mutually identifying partitions included in the target leaf CU
  • Npart is the total number of partitions included in the target leaf CU.
  • Step S103 Subsequently, the PU decoding unit 431 decodes the edge-based prediction flag EF for the target partition specified by the loop variable part_id.
  • Step S104 the PU decoding unit 431 determines whether or not the edge-based prediction flag EF decoded in step S103 indicates that the edge-based prediction mode is applied to the target partition.
  • Step S105 When the edge-based prediction mode is not applied to the target partition (No in step S104), the PU decoding unit 431 decodes the estimation flag MPM for the target partition. If the estimation flag MPM is 0, the PU decoding unit 431 continues to decode the residual prediction mode index RIPM for the target partition.
  • the residual prediction mode index RIPM is the prediction mode specified by the estimation flag MPM from the prediction mode included in the basic prediction mode set. Any of a total of eight types of prediction modes excluding the above is designated. Therefore, in the present embodiment, the residual prediction mode index RIPM is expressed by a 3-bit bit string.
  • Step S106 the PU decoding unit 431 selects a prediction mode for the target partition from the basic prediction mode set, and assigns the prediction mode to the target partition. Specifically, if the estimated MPM flag decoded in step S105 is 1, the estimated prediction mode obtained by referring to the prediction modes assigned to the decoded partitions around the target partition is set for the target partition. Set to the prediction mode. If the estimation flag MPM decoded in step S105 is 0, the PU decoding unit 431 sets the prediction mode specified by the residual prediction mode index RIPM as the prediction mode for the target partition.
  • Step S107 On the other hand, when the edge-based prediction mode is applied to the target partition (Yes in step S104), the PU decoding unit 431 determines the total number of correction directions that can be selected for the target partition according to the size of the target partition. . More specifically, the PU decoding unit 431 determines the total number of correction directions that can be selected for the target partition so as to have a negative correlation with the size of the target partition.
  • the total number of correction directions that can be selected for the target partition is set to 8
  • the target when the total number of correction directions that can be selected for a partition is set to 2 and the size of the target partition is 32 ⁇ 32 pixels or 64 ⁇ 64 pixels, the total number of correction directions that can be selected for the target partition is set to zero.
  • the PU decoding unit 431 sets the total number of correction directions that can be selected for the target partition to Nad1
  • the total number of pixels belonging to the target partition is Npix2.
  • the total number of correction directions that can be selected for the target partition is set to Nad2 (Nad1 ⁇ Nad2).
  • the PU decoding unit 431 is the division information included in the encoded data # 1, and includes CU division information SP_CU that specifies a division pattern for each leaf CU of the target LCU including the target partition, and the target PU (target The size of the target partition can be identified by referring to the intra PU partition information SP_Intra that specifies the partition pattern of each leaf CU).
  • Step S108 the PU decoding unit 431 decodes the additional index AI for the target partition.
  • the most significant bit of the additional index AI functions as a flag indicating whether the main direction is used as it is or as shown in FIG. 8A.
  • the bit string other than the most significant bit of the additional index AI functions as information for designating any correction direction from selectable correction directions.
  • Step S110 When the most significant bit of the additional index AI indicates that the main direction is used as it is (YES in step S109), the PU decoding unit 431 sets the prediction mode for the target partition to the main direction edge-based prediction mode. To do.
  • Step S111 When the most significant bit of the additional index AI indicates that the sub-direction is used (NO in step S109), the PU decoding unit 431 decodes a bit string (binary representation) other than the most significant bit of the additional index AI. Then, the sub-direction index k (decimal expression) designated by the bit string is determined. Also, the PU decoding unit 431 sets the prediction mode for the target partition to the sub-direction edge-based prediction mode specified by the sub-direction index k.
  • the possible values of the sub-direction index k are, for example, any one of ⁇ 1 and +1 It is.
  • the total number of correction directions that can be selected is 8, and thus the possible values of the sub-direction index k are, for example, ⁇ 4, -3, -2, -1, +1, +2, +3, and +4.
  • the total number of correction directions that can be selected is 0, and the additional index AI is not decoded.
  • the bit string other than the most significant bit of the additional index AI corresponding to the sub-direction index k is composed of a bit string corresponding to abs (k) and a bit string corresponding to sign (k). it can.
  • the bit string corresponding to sign (k) is 0 when k is a negative value and 1 when k is a positive value.
  • FIG. 8B is a diagram illustrating an example of the main direction and the sub-direction set by the PU decoding unit 431 when the size of the target partition is 4 ⁇ 4 pixels.
  • the angle ⁇ sk that designates the sub direction (sub direction vector) includes the sub direction index k and the correction.
  • ⁇ sk ⁇ m + k ⁇ ⁇ (1-1) It is expressed.
  • the parameter ⁇ is derived by the predicted image generation unit 432 according to a method described later.
  • Step S112 This step is the end of the loop.
  • the above is the flow of PU information decoding processing by the PU decoding unit 431.
  • the prediction mode set for each partition included in the target leaf CU in step S106, step S110, and step S111 is supplied to the predicted image generation unit 432 as a prediction mode # 431.
  • the prediction image generation unit 432 refers to the prediction image for each partition included in the target leaf CU with reference to the prediction mode # 431 supplied from the PU decoding unit 431 and the decoded pixel values around the partition. Generate.
  • FIG. 9 is a flowchart showing the flow of predicted image generation processing for the target leaf CU by the predicted image generation unit 432.
  • the predicted image generation unit 432 initializes the value of the loop variable part_id to 0, and starts loop processing that sets the increment value of the loop variable part_id for each loop to 1 for part_id that satisfies part_id ⁇ Npart-1.
  • the loop variable part_id is the same loop variable as described above. That is, the loop variable part_id is an index for mutually identifying partitions included in the target leaf CU, and Npart is the total number of partitions included in the target leaf CU.
  • the predicted image generation unit 432 refers to the prediction mode # 431 supplied from the PU decoding unit 431, refers to the prediction mode for the target partition specified by the loop variable part_id, and performs edge-based prediction for the target partition. Determine whether the mode is applied.
  • Step S203 When the edge-based prediction mode is not applied to the target partition (No in step S202), the predicted image generation unit 432 uses the direction prediction mode as the prediction mode for the target partition, that is, a prediction mode that uses a predetermined prediction direction. It is determined whether or not.
  • Step S204 When the prediction mode for the target partition is not the direction prediction mode (No in step S203), the predicted image generation unit 432 averages the decoded pixel values around the target partition for the predicted image for the target partition. Generate by.
  • FIG. 10 is a diagram illustrating each pixel (prediction target pixel) of the target partition, which is 4 ⁇ 4 pixels, and pixels (reference pixels) around the target partition.
  • the prediction target pixels are denoted by symbols a to p
  • the reference pixels are denoted by symbols A to M
  • the pixel value of the pixel X (X is any of a to p or A to M) It will be expressed as Further, it is assumed that the reference pixels A to M have all been decoded.
  • ave Indicates that an element included in parentheses is averaged.
  • Step S205 When the prediction mode for the target partition is the direction prediction mode (Yes in step S203), the predicted image generation unit 432 sets the prediction direction corresponding to the direction prediction mode as the prediction direction for the target partition.
  • Step S206 On the other hand, when the prediction mode for the target partition is the edge-based prediction mode (Yes in step S202), the predicted image generation unit 432 is based on decoded pixel values (pixel values of reference pixels) around the target partition. Thus, the main direction assigned to the target partition is derived.
  • the predicted image generation unit 432 performs, for example, the following sub-step S206-1 to sub-step S206-4.
  • FIG. 11 is a diagram showing the target partition OP together with adjacent partitions NP2 and NP3 adjacent to the target partition OP, and a partition NP1 sharing the top left vertex of the target partition.
  • the pixel values of the pixels included in the adjacent partitions NP1 to NP3 shown in FIG. 11 have been decoded.
  • pixels included in partitions NP1 to NP3 around the target partition OP can be used as reference pixels.
  • FIG. 11 shows the case where the target partition OP and the adjacent partitions NP1 to NP3 are all 4 ⁇ 4 pixels, but the size of these partitions is 8 ⁇ 8 pixels or other sizes. The same applies to the case of.
  • the predicted image generation unit 432 initializes the value of the loop variable i to 1, and starts a loop process that sets the increment value of the loop variable i for each loop to 1 for j that satisfies i ⁇ M.
  • M is the number of reference pixels that are referred to in order to derive the main direction for the target partition.
  • the prediction image generation unit 432 calculates the edge vectors b i.
  • the calculation of the edge vectors b i is Sobel operator (Sobel operators, also referred to as a Sobel filter) Gx, and may be used to Gy.
  • the Sobel filters Gx and Gy are filter matrices used for calculating the image gradient along the x direction and the image gradient along the y direction, respectively. For example, as a 3 ⁇ 3 matrix, ,
  • Prediction image generating unit 432 Given by. Prediction image generating unit 432, image gradient for the calculated x-direction, and calculates an edge vector b i that is perpendicular to the image gradients represented by the image gradient along the y-direction.
  • the predicted image generation unit 432 has a function T ( ⁇ ) shown below.
  • T ( ⁇ ) ⁇ ⁇ e, b i > 2
  • e represents a unit vector whose angle between its own direction and the horizontal direction (x direction) is ⁇
  • represents the inner product of both vectors.
  • the symbol ⁇ indicates that the subscript i is to be summed from 1 to M.
  • the predicted image generation unit 432 calculates the main direction
  • the partition adjacent to the upper side of the target partition, the partition adjacent to the left side of the target partition, and the upper left vertex of the target partition are shared.
  • the case where the pixel values of the pixels belonging to the partition are referred to has been taken as an example, but the present embodiment is not limited to this, and the predicted image generation unit 432 is more generally a reference set around the target partition. With reference to the decoded pixel values belonging to the region, the main direction for the target partition can be calculated.
  • the method of deriving the main direction by the predicted image generation unit 432 is not limited to the above example.
  • the predicted image generation unit 432 may be configured to calculate, as the main direction, the edge vector having the largest magnitude among the edge vectors b i calculated for the reference pixel, that is, the direction indicated by the edge vector having the largest norm. Good.
  • Step S207 the predicted image generation unit 432 determines whether or not the prediction mode for the target partition is the main direction prediction mode.
  • Step S208 When the prediction mode for the target partition is the main direction prediction mode (Yes in step S207), the predicted image generation unit 432 sets the main direction calculated in step S206 as the prediction direction for the target partition.
  • Step S209 On the other hand, when the prediction mode for the target partition is not the main direction prediction mode (No in step S207), the predicted image generation unit 432 performs a quantization step for specifying the roughness of quantization when the correction angle is quantized. A parameter ⁇ representing the size is calculated. In the present embodiment, the predicted image generation unit 432 has confidence measures s shown below.
  • the parameter ⁇ is calculated as a function of Note that the parameter ⁇ derived by the predicted image generation unit 432 has a property of being a decreasing function of the certainty factor s.
  • the predicted image generation unit 432 may be configured to use a predetermined parameter ⁇ , for example, instead of the configuration that calculates the parameter ⁇ individually for each partition.
  • the parameter ⁇ is encoded for each slice or LCU, the parameter ⁇ is included in the encoded data # 1, and the predicted image generation unit 432 A configuration using the parameter ⁇ decoded from the encoded data # 1 may be adopted.
  • the predicted image generation unit 432 calculates an angle ⁇ sk indicating the sub-direction for the target partition using Equation (1-1) using the sub-direction index k obtained by decoding the additional index AI.
  • Step S210 Subsequently, the predicted image generation unit 432 sets the direction indicated by the angle ⁇ sk calculated in step S209 as the predicted direction for the target partition.
  • Step S211 the predicted image generation unit 432 extrapolates the decoded pixel values around the target partition along the prediction direction for the target partition set in step S205, step S208, and step S210. Then, a predicted image for the target partition is generated.
  • the predicted image generation unit 432 uses the pixel position of the prediction target pixel in the target partition as a starting point, and among the decoded pixels located on the virtual line segment facing the reverse direction of the prediction direction, By setting the pixel value of the pixel closest to the pixel (hereinafter also referred to as the closest pixel) to the pixel value of the prediction target pixel, a predicted image in the target partition is generated. Further, the pixel value of the prediction target pixel may be a value calculated using the pixel value of the nearest pixel and the pixel values of the pixels around the nearest pixel.
  • the predicted image generation unit 432 includes the pixel value a in the example illustrated in FIG. 10.
  • “>>” represents a right shift operation, and for any positive integer x, s, the value of x >> s is equal to the value obtained by rounding down the decimal part of x
  • the predicted image generation unit 432 can calculate the pixel values a to p by the same method for the basic prediction modes other than the above prediction modes. Also, the predicted image generation unit 432 can calculate the pixel values a to p by the same method for the edge-based prediction mode.
  • Step S212 This step is the end of the loop whose loop variable is id_part.
  • the predicted image generation unit 432 generates a predicted image for the target leaf CU by performing the above processing. Further, the generated predicted image for the target leaf CU is supplied to the decoded image generation unit 435 as a predicted image # 432.
  • the TU decoding unit 433 decodes TU information for the target leaf CU. Also, the TU decoding unit 433 divides the target leaf CU into one or a plurality of blocks with reference to the TU partition information SP_TU included in the TU information for the target leaf CU. Also, the quantization prediction residual for each block is decoded from the encoded data # 142a for the target leaf CU, and the decoded quantization prediction residual is supplied to the prediction residual restoration unit 434.
  • the prediction residual restoration unit 434 performs inverse quantization and inverse DCT transform (Inverse Discrete Cosine Transform) on the quantized prediction residual supplied from the TU decoding unit 433, so that the prediction residual for each partition is obtained. Restore the difference.
  • the restored prediction residual # 434 is supplied to the decoded image generation unit 435.
  • the decoded image generation unit 435 adds the prediction image # 432 supplied from the prediction image generation unit 432 and the prediction residual supplied from the prediction residual restoration unit 434, thereby obtaining a decoded image for the target leaf CU. Generate.
  • the decoded image generation unit 435 generates a decoded image # 2 for the target LCU by integrating the decoded images for each leaf CU included in the target LCU.
  • FIG. 12A is a flowchart showing an overall flow of decoding processing by the leaf CU decoding unit 143.
  • Step S301 the leaf CU decoding unit 143 uses the TU decoding unit 433 to decode the TU information for the target leaf CU. Further, the target leaf CU is divided into one or a plurality of blocks with reference to the TU partition information SP_TU included in the TU information regarding the target leaf CU. Also, the quantization prediction residual for each block is decoded from the encoded data # 142a for the target leaf CU, and the decoded quantization prediction residual is supplied to the prediction residual restoration unit 434.
  • Step S302 the leaf CU decoding unit 143 performs inverse quantization and inverse DCT transform on the quantized prediction residual decoded in step S301 by the prediction residual restoration unit 434, so that each pixel for the target leaf CU is processed. Restore the prediction residual of.
  • Step S303 the leaf CU decoding unit 143 determines the division pattern for each partition of the target leaf CU by decoding the PU information about the target leaf CU in the PU decoding unit 431, and also predicts each partition. A mode is determined, and each determined prediction mode is assigned to each partition. Since specific PU information decoding processing has already been described, description thereof will be omitted.
  • Step S304 the leaf CU decoding unit 143 causes the predicted image generation unit 432 to initialize the value of the loop variable part_id to 0, and for the part_id satisfying part_id ⁇ Npart ⁇ 1, the increment value of the loop variable part_id for each loop is set to 1.
  • the loop processing is started.
  • the loop variable part_id is the same loop variable as described above. That is, the loop variable part_id is an index for mutually identifying partitions included in the target leaf CU, and Npart is the total number of partitions included in the target leaf CU.
  • Step S305 the leaf CU decoding unit 143 generates a predicted image for the partition specified by part_id in the predicted image generation unit 432. Since specific prediction image generation processing has already been described, description thereof will be omitted.
  • Step S306 the leaf CU decoding unit 143 adds the prediction image # 432 supplied from the prediction image generation unit 432 and the prediction residual supplied from the prediction residual restoration unit 434, thereby adding a prediction for the target leaf CU.
  • a decoded image generation unit 435 generates a decoded image.
  • Step S307 This step is the end of the loop.
  • the video decoding device 1 includes the prediction mode belonging to the prediction mode group in the prediction residual decoded from the encoded data together with the prediction mode designation information and other side information, and the prediction In an image decoding apparatus that generates a decoded image by adding predicted images generated according to a prediction mode specified by mode specifying information, the number of prediction modes N (N is an arbitrary natural number) added to the prediction mode group related to the target prediction unit Referring to a coding parameter included in the side information and including a coding unit including the target prediction unit or a coding unit decoded prior to the coding unit.
  • PU decoding unit 431 to set and each of the N prediction modes to be added to the prediction mode group related to the target prediction unit
  • the prediction unit is a type that increases the prediction accuracy by increasing the number of prediction modes belonging to the prediction mode group, and the coding efficiency is improved by reducing the number of prediction modes belonging to the prediction mode group and improving the coding efficiency. Some types improve efficiency. According to the knowledge obtained by the inventors, whether a certain prediction unit belongs to the former type or the latter type is usually encoded that is decoded before the prediction unit or the encoding unit. It can be determined from the coding parameters for the unit.
  • the amount of code of side information necessary for encoding an encoding target image for a certain encoding unit has a positive correlation with the total number of prediction units included in the encoding unit. That is, if the total number of prediction units is large, the code amount of the side information increases.
  • the total number of prediction units included in a certain coding unit has a negative correlation with the size of each prediction unit.
  • the PU decoding unit 431 determines the total number of correction directions that can be selected for the target partition so as to have a negative correlation with the size of the target partition. Specifically, when the size of the target partition is 4 ⁇ 4 pixels or 8 ⁇ 8 pixels, the total number of correction directions that can be selected for the target partition is set to 8, and the size of the target partition is 16 ⁇ 16 pixels. In this case, when the total number of correction directions that can be selected for the target partition is set to 2 and the size of the target partition is 32 ⁇ 32 pixels or 64 ⁇ 64 pixels, the total number of correction directions that can be selected for the target partition is set to 0. .
  • FIG. 12B shows the result of measuring the relationship between the number of correction directions and the effect of introducing edge-based prediction for each partition size.
  • the horizontal axis indicates the number of correction directions
  • sd0, sd1, sd2, sd3, and sd4 indicate the number of correction directions of 0, 2, 4, 6, and 8, respectively.
  • the vertical axis indicates the effect of introducing edge-based prediction. More precisely, the rate distortion cost reduction achieved by making it possible to select the additional prediction mode in addition to the basic prediction mode is reduced based on the rate distortion cost when only the basic prediction mode can be selected. Expressed as a percentage.
  • FIG. 12B shows that when the partition size is 4 ⁇ 4 pixels or 8 ⁇ 8 pixels, the rate distortion cost can be reduced most by setting sd4, that is, eight correction directions.
  • sd4 that is, eight correction directions.
  • 16 ⁇ 16 pixels it can be seen that if sd1, that is, two correction directions are used, the rate distortion cost can be sufficiently reduced, and the effect is not recognized even if more correction directions are added.
  • sd0 that is, when the number of correction directions is 0 (only the main direction is used)
  • the reduction rate of the rate distortion cost is the maximum. It turns out that it is.
  • the addition of the correction direction is related to both the increase of the code amount of the side information and the improvement of the prediction accuracy.
  • the influence of the side information increase per unit area is larger than in the case of the large size.
  • the improvement range of the prediction accuracy is large as compared with the case of the large size. This is because there is a high possibility that a straight line exists in a block at a small size. From the experimental results shown in FIG. 12 (b), it can be said that the influence of the latter prediction accuracy improvement exceeds the influence of the former side information increase.
  • Moving picture encoding apparatus 2 (Moving picture encoding device 2)
  • the moving picture encoding apparatus (encoding apparatus) 2 according to the present embodiment will be described with reference to FIGS.
  • the same part as the part already demonstrated is attached
  • FIG. 13 is a block diagram showing a configuration of the moving picture encoding apparatus 2.
  • the moving image encoding device 2 includes an LCU header information determination unit 21, a header information encoding unit 22, an LCU setting unit 23, an LCU encoding unit 24, a variable length code multiplexing unit 25, and an LCU decoding.
  • a unit 26 and a frame memory 27 are provided.
  • the video encoding device 2 is a device that generates and outputs encoded data # 1 by encoding the input image # 100, in brief.
  • the moving image encoding apparatus 2 includes, as part thereof, H.264. H.264 / MPEG-4 AVC standard technology, VCEG (Video Coding Expert Group) technology used in KTA software, which is a joint development codec, and successor codec TMuC (Test Model Under Consideration) )
  • KTA software which is a joint development codec, and successor codec TMuC (Test Model Under Consideration)
  • the LCU header information determination unit 21 determines LCU header information based on the input image # 100.
  • the determined LCU header information is output as LCU header information # 21.
  • the LCU header information # 21 includes (1) information about the size and shape of the LCU belonging to the target slice and the position in the target slice, and (2) the size, shape and target of the leaf CU belonging to each LCU. Contains information about the location within the LCU.
  • the LCU header information # 21 is input to the LCU setting unit 23 and also supplied to the header information encoding unit 22.
  • the header information encoding unit 22 encodes the LCU header information # 21 supplied from the LCU header information determination unit 21 and the leaf CU header information # 53 supplied from the LCU encoding unit 24, and has encoded header information. # 22 is output. The encoded header information # 22 is supplied to the variable length code multiplexer 25.
  • the LCU setting unit 23 divides the input image (encoding target image) # 100 into a plurality of LCUs based on the LCU header information # 21, and outputs an LCU image # 23 related to each LCU.
  • the LCU image # 23 is sequentially supplied to the LCU encoding unit 24.
  • the LCU encoding unit 24 encodes sequentially input LCU image # 23 to generate LCU encoded data # 24.
  • the generated LCU encoded data # 24 is supplied to the variable length code multiplexing unit 25 and the LCU decoding unit 26. Since the configuration of the LCU encoding unit 24 will be described later, description thereof is omitted here.
  • variable length code multiplexing unit 25 generates encoded data # 1 by multiplexing encoded header information # 22 and LCU encoded data # 24, and outputs the encoded data # 1.
  • the LCU decoding unit 26 sequentially decodes the LCU encoded data # 24 corresponding to each input LCU, thereby generating and outputting a decoded image # 26 corresponding to each LCU.
  • the decoded image # 26 is supplied to the frame memory 27. Since the configuration of the LCU decoding unit 26 is the same as the configuration of the LCU decoding unit 14 included in the video decoding device 1, detailed description thereof is omitted here.
  • the input decoded image # 26 is recorded in the frame memory 27.
  • decoded images corresponding to all the LCUs preceding the target LCU in the raster scan order are recorded in the frame memory 27.
  • the decoded image recorded in the frame memory is called decoded image # 27. That is, the decoded image # 27 corresponds to the decoded image # 26 recorded in the frame memory 27.
  • LCU encoder 24 LCU encoder 24
  • the LCU encoding unit 24 will be described more specifically with reference to different drawings.
  • FIG. 14 is a block diagram showing a configuration of the LCU encoding unit 24.
  • the LCU encoding unit 24 includes an LCU division structure determining unit 241, a leaf CU scanning unit 242, and a leaf CU encoding unit 243.
  • the LCU partition structure determining unit 241 refers to the LCU header information # 21 to determine the partition structure of the target LCU into leaf CUs.
  • Division structure information # 241 for designating the division structure of the target LCU into leaf CUs is supplied to the leaf CU scanning unit 242 together with the LCU image # 23 for the target LCU.
  • An example of the divided structure determined by the LCU divided structure determining unit 241 is the one shown in FIG.
  • the leaf CU scanning unit 242 scans the leaf CUs included in the target LCU in a predetermined order (for example, raster scan order), and the leaf CU image # 242a and CU information about the target leaf CU that is the leaf CU to be processed.
  • # 242b is supplied to the leaf CU encoding unit 243.
  • the CU information # 242b includes the size and shape of the target leaf CU and the position in the target LCU.
  • the leaf CU encoding unit 243 refers to the leaf CU image and CU information for the target leaf CU, and the decoded pixel value stored in the frame memory 15, and generates a predicted image for the target leaf CU. At the same time, the prediction residual between the leaf CU image and the prediction image is encoded. Also, the leaf CU encoding unit 243 integrates the encoded prediction residuals for each leaf CU belonging to the target LCU, and generates LCU encoded data # 24 for the target LCU.
  • FIG. 15 is a block diagram showing the configuration of the leaf CU encoding unit 243.
  • the leaf CU encoding unit 243 includes a PU determination unit 531, a predicted image generation unit 532, a decoded image generation unit 533, a prediction residual calculation unit 534, a TU determination unit 535, and a prediction residual restoration unit 536.
  • the PU determination unit 531 refers to the leaf CU image # 242a and the CU information # 242b, and determines (1) a division pattern for each partition of the target leaf CU and (2) a prediction mode to be allocated to each partition. To do.
  • a division pattern into the partitions of the target leaf CU for example, a division pattern that is handled as one partition without dividing the target leaf CU, or the target leaf CU is divided into four partitions symmetrically. One of the division patterns can be selected.
  • the PU determination unit 531 determines the total number of correction directions that can be selected for the target partition according to the size of the target partition. More specifically, the PU determination unit 531 determines the total number of correction directions that can be selected for the target partition so as to have a positive correlation with the size of the target partition.
  • the total number of correction directions that can be selected for the target partition is set to 8
  • the target when the total number of correction directions that can be selected for the partition is set to 2 and the size of the target partition is 32 ⁇ 32 or 64 ⁇ 64, the total number of correction directions that can be selected for the target partition is set to 0.
  • the PU determination unit 531 determines the total number of correction directions that can be selected for the target partition as Nad1, and the total number of pixels belonging to the target partition is Npix2 ( When Npix1 ⁇ Npix2), the total number of correction directions that can be selected for the target partition is determined to be Nad2 (Nad1 ⁇ Nad2).
  • PU information including (1) an intra split flag (intra_split_flag) that specifies whether or not the target leaf CU is split into four partitions, and (2) prediction mode information that specifies a prediction mode assigned to each partition.
  • # 531 is supplied to the predicted image generation unit 532 and the LCU encoded data generation unit 538.
  • the PU information # 531 includes an additional index AI including a 1-bit flag that specifies whether to use the main direction or the sub-direction for the target partition.
  • the additional index AI includes a plurality of sub-directions and an index that specifies which of the plurality of sub-directions whose total number is determined according to the size of the target partition is used. It is.
  • the PU determination unit 531 applies to each of all combinations of (1) a possible division pattern for each partition of the target leaf CU and (2) a prediction mode that can be assigned to each partition. Corresponding PU information is generated. In addition, generation of a prediction image, a prediction residual, and a decoded image, which will be described later, is performed for each piece of PU information.
  • the PU determination unit 531 supplies PU information that optimizes rate distortion supplied from the rate distortion evaluation unit 537 described later to the header information encoding unit 22 as part of the leaf CU header information # 53. Then, the PU information for optimizing the rate distortion is supplied to the predicted image generation unit 532 and the LCU encoded data generation unit 538 as the final PU information for the target leaf CU.
  • the predicted image generation unit 532 generates a predicted image # 532 for the target leaf CU with reference to the decoded image # 27 stored in the frame memory 27 and the PU information # 531 supplied from the PU determination unit 531. To do. Since the predicted image generation process by the predicted image generation unit 532 is the same as the predicted image generation unit 432 included in the video decoding device 1, description thereof is omitted here.
  • the prediction residual calculation unit 534 generates a prediction residual # 534 by subtracting the prediction image # 532 from the leaf CU image # 242a.
  • the generated prediction residual # 534 is supplied to the TU determination unit 535.
  • the TU determination unit 535 refers to the leaf CU image # 242a and the CU information # 242b, and determines a division pattern of the target leaf CU into one or a plurality of blocks. Also, according to the determined division pattern, the prediction residual # 534 for the target leaf CU supplied from the prediction residual calculation unit 534 is divided into prediction residuals for each block.
  • the TU determination unit 535 generates a prediction residual in the frequency domain by performing DCT transform (DiscretecreCosine Transform) on the prediction residual for each block, and then quantizes the prediction residual in the frequency domain. Generate a quantized prediction residual for each block.
  • the generated quantization prediction residual for each block is transmitted to the prediction residual restoration unit 536 and the LCU encoded data generation unit 538 as TU information # 535 together with TU partition information SP_TU that specifies the partition pattern of the target leaf CU. Supplied.
  • the TU determination unit 535 generates TU information # 535 corresponding to all possible division patterns for each block of the target leaf CU. Further, the generation of the quantized prediction residual is performed for all of the divided patterns.
  • the TU determination unit 535 supplies the TU information for optimizing the rate distortion supplied from the rate distortion evaluation unit 537 described later to the header information encoding unit 22 as leaf CU header information # 53, and the rate.
  • TU information for optimizing distortion is supplied to the prediction residual restoration unit 536 and the LCU encoded data generation unit 538 as final TU information for the target leaf CU.
  • the prediction residual restoration unit 536 performs prediction for each block by performing inverse quantization and inverse DCT transform (Inverse Discrete Cosine Transform) on the quantization prediction residual for each block supplied from the TU determination unit 535. Restore the residual. Further, the prediction residuals for each block are integrated according to the division pattern determined by the TU determination unit 535, and the prediction residual # 536 for the target leaf CU is generated. The prediction residual # 536 for the generated target leaf CU is supplied to the decoded image generation unit 533.
  • inverse quantization and inverse DCT transform Inverse Discrete Cosine Transform
  • the decoded image generation unit 533 adds the prediction image # 532 for the target leaf CU supplied from the prediction image generation unit 532 and the prediction residual # 536 for the target leaf CU supplied from the prediction residual restoration unit 536. As a result, a decoded image # 533 for the target leaf CU is generated. The generated decoded image # 533 for the target leaf CU is supplied to the rate distortion evaluation unit 537.
  • the rate distortion evaluation unit 537 refers to the decoded image # 533 for the target leaf CU and the leaf CU image # 242a for the target leaf CU supplied from the decoded image generation unit 533, and calculates the rate distortion for the target leaf CU. evaluate.
  • the rate distortion evaluation unit 537 includes (1) a possible division pattern for each partition of the target leaf CU, (2) a prediction mode that can be assigned to each partition, and (3) each block of the target leaf CU. Among possible combinations of divided patterns, a combination that optimizes the rate distortion is determined.
  • the PU information and TU information specified by the determined combination are supplied to the PU determination unit 531 and the TU determination unit 535, respectively.
  • the LCU encoded data generation unit 538 performs variable length encoding on the final PU information supplied from the PU determination unit 531 and the final TU information supplied from the TU determination unit 535, thereby performing the target leaf CU.
  • the encoded data for is generated.
  • the LCU encoded data generation unit 538 generates encoded data # 24 for the target LCU by multiplexing the encoded data for each leaf CU belonging to the target LCU.
  • FIG. 16 is a flowchart showing the flow of the encoding process performed by the leaf CU encoder 243.
  • Step S401 the leaf CU encoding unit 243 initializes the value of the loop variable pu_param_id to 0, and for the pu_param_id that satisfies pu_param_id ⁇ Npu ⁇ 1, performs a first loop process in which the increment value of the loop variable pu_param_id for each loop is 1.
  • the loop variable pu_param_id is an index for mutually identifying combinations of possible division patterns for each partition of the target leaf CU and possible prediction modes to be assigned to each partition.
  • Npu represents the total number of combinations of possible division patterns for each partition of the target leaf CU and prediction modes that can be assigned to each partition.
  • Npu 1 ⁇ Nmode + 4 ⁇ Nmode
  • pu_param_id functions as an index for identifying these combinations of 1 ⁇ Nmode + 4 ⁇ Nmode from each other.
  • Step S402 the leaf CU encoding unit 243 initializes the value of the loop variable tu_param_id to 0, and for the tu_param_id that satisfies tu_param_id ⁇ Ntu ⁇ 1, the second loop processing sets the increment value of the loop variable tu_param_id to 1 for each loop.
  • the loop variable tu_param_id is an index for mutually identifying possible division patterns for each block of the target leaf CU.
  • Ntu represents the total number of possible division patterns for each block of the target leaf CU.
  • the size of the target leaf CU is 8 ⁇ 8 pixels
  • the division pattern in which the target leaf CU is handled as it is without being divided and the target leaf CU is symmetrical to four 4 ⁇ 4 pixel blocks. If the division pattern to be divided is possible, the total number Ntu of possible division patterns for each block of the target leaf CU is 2, and the loop variable tu_param_id identifies these two division patterns from each other. Function as an index for
  • Step S403 the leaf CU encoding unit 243 initializes the value of the loop variable part_id to 0, and performs a third loop process in which the increment value of the loop variable part_id for each loop is set to 1 for part_id that satisfies part_id ⁇ Npart ⁇ 1.
  • the loop variable part_id is an index for identifying each partition
  • Npart is the total number of the partitions when the target leaf CU is divided into partitions according to the division pattern corresponding to the loop variable pu_param_id.
  • Step S404 the leaf CU encoding unit 243 generates a predicted image for the target partition specified by the set of loop variables (pu_param_id, tu_param_id, part_id) in the predicted image generation unit 532.
  • Step S405 the leaf CU encoding unit 243 calculates the prediction residual for the partition specified by the set of loop variables (pu_param_id, part_id) by the prediction residual calculation unit 534.
  • Step S406 the leaf CU encoding unit 243 generates a quantized prediction residual by DCT transforming and quantizing the prediction residual for the target block belonging to the target leaf CU in the TU determination unit 535.
  • Step S407 the leaf CU encoding unit 243 restores the prediction residual by performing inverse quantization and inverse DCT transform on the quantized prediction residual generated in step S406.
  • Step S408 the leaf CU encoding unit 243 generates a decoded image for the target partition by adding the prediction image generated in step S404 and the prediction residual restored in step S407.
  • Step S409 This step is the end of the third loop. Generation of the decoded image for the target leaf CU is completed by the third loop processing.
  • the rate distortion evaluation unit 537 refers to the code amount and the distortion for the target leaf CU and evaluates the rate distortion for the target leaf CU.
  • Step S411 This step is the end of the second loop.
  • Step S412 This step is the end of the first loop. Note that the first and second loop processes complete the rate distortion evaluation for all possible combinations of the loop variable pu_param_id and the loop variable tu_param_id.
  • Step S413 Subsequently, the leaf CU encoding unit 243, in the rate distortion evaluation unit 537, out of the rate distortions for all possible combinations of the loop variable pu_param_id and the loop variable tu_param_id, the loop variable pu_param_id and the loop corresponding to the optimum rate distortion The combination of the variable tu_param_id is determined.
  • Step S414 The leaf CU encoding unit 243 generates encoded data for the target leaf CU using a combination of the loop variable pu_param_id and the loop variable tu_param_id corresponding to the optimum rate distortion determined in step S413. That is, the leaf CU encoding unit 243 generates encoded data for the target leaf CU using a division pattern for dividing the target leaf CU into partitions and blocks and optimizing rate distortion.
  • the video encoding device 2 has selected the prediction residual obtained by subtracting the prediction image generated according to the prediction mode selected from the prediction mode group for each prediction unit from the original image.
  • An image encoding apparatus that generates encoded data by encoding together with prediction mode specifying information for specifying a prediction mode and other side information, and the number N of prediction modes to be added to the prediction mode group related to the target prediction unit,
  • the encoding parameter included in the side information is set with reference to an encoding unit including the target prediction unit or an encoding parameter related to an encoding unit encoded before the encoding unit.
  • Prediction parameters corresponding to each of the N prediction modes to be added to the setting unit (PU determination unit 531) and the prediction mode group related to the target prediction unit Includes a deriving means for deriving from the decoded region of the local decoded image (predicted image generation unit 532), the.
  • the prediction unit is a type that increases the prediction accuracy by increasing the number of prediction modes belonging to the prediction mode group, and the coding efficiency is improved by reducing the number of prediction modes belonging to the prediction mode group and improving the coding efficiency. Some types improve efficiency. According to the knowledge obtained by the inventors, whether a certain prediction unit belongs to the former type or the latter type is usually encoded that is decoded before the prediction unit or the encoding unit. It can be determined from the coding parameters for the unit.
  • the total number of basic prediction modes that can be selected for the target partition is a specified value
  • the PU decoding unit 431 included in the video decoding device 1 and the PU determination unit 531 included in the video encoding device 2 are Although the case where the total number of sub-directions that can be selected for a partition is changed according to the size of the target partition has been described, the present embodiment is not limited to this.
  • the total number of basic prediction modes that can be selected for each partition may be changed according to the size of the partition. Specifically, the total number of basic prediction modes that can be selected for a partition of 32 ⁇ 32 pixels is set to 9, and the total number of basic prediction modes that can be selected for a partition of 8 ⁇ 8 pixels is set to 33. .
  • a flag indicating the total number of basic prediction modes may be included in the encoded data, and the video decoding device 1 may be configured to decode the basic prediction mode with reference to the flags.
  • the PU decoding unit 431 included in the video decoding device 1 and the PU determination unit 531 included in the video encoding device 2 determine the total number of sub-directions that can be selected for the target partition.
  • the configuration may be changed in accordance with the total number of basic prediction modes that can be selected for the target partition. For example, when the number of basic prediction modes that can be selected for the target partition is 9, the total number of sub-directions that can be selected for the target partition is 0, and the number of basic prediction modes that can be selected for the target partition is 33 A configuration is also possible in which the total number of sub-directions that can be selected for the target partition is 8.
  • the total number of basic prediction modes that can be selected for the target partition is Nba1
  • the total number of sub-directions that can be selected for the target partition is determined as Nad1
  • the target partition can be selected.
  • the total number of basic prediction modes is Nba2 (Nba1 ⁇ Nba2)
  • the total number of correction directions that can be selected for the target partition is determined as Nad2 (Nad1 ⁇ Nad2).
  • the PU decoding unit 431 included in the video decoding device 1 and the PU determination unit 531 included in the video encoding device 2 are the partitions adjacent to the target partition, and the edge-based prediction mode for the decoded partition Depending on whether or not is applied, the total number of sub-directions that can be selected for the target partition may be changed.
  • the PU decoding unit 431 included in the video decoding device 1 and the PU determination unit 531 included in the video encoding device 2 have an edge-based prediction mode for at least one of the decoded partitions adjacent to the target partition. If applied, the total number of sub-directions that can be selected for the target partition is 2. If the edge-based prediction mode is not applied to any of the decoded partitions adjacent to the target partition, the target partition can be selected. The total number of sub-directions may be four.
  • FIG. 17 is a diagram illustrating a case where the edge-based prediction mode is applied to an adjacent partition adjacent to the left side of the target partition. As shown in FIG. 17, when the edge-based prediction mode is applied to the adjacent partition, the prediction direction derived using the edge-based prediction mode for the adjacent partition is also an appropriate prediction direction for the target partition. There is a tendency.
  • the PU decoding unit 431 included in the video decoding device 1 and the PU determination unit 531 included in the video encoding device 2 determine the number of sub-directions that can be selected for the target partition.
  • the configuration can be reduced compared to the case where the edge-based prediction mode is not applied to any partition adjacent to the partition.
  • the prediction direction derived using the edge-based prediction mode for the adjacent partition adjacent to the target partition tends to be an appropriate prediction direction for the target partition.
  • the main direction derived for the target partition tends to be close to the prediction direction derived using the edge-based prediction mode for the adjacent partition.
  • the number of sub-directions for the sub-direction edge-based prediction mode that is unlikely to be selected can be reduced, so that the code amount of the encoded data # 1 can be reduced while maintaining high prediction accuracy. Since it can be reduced, encoding efficiency can be improved.
  • the PU decoding unit 431 included in the video decoding device 1 and the PU determination unit 531 included in the video encoding device 2 are applied with the edge-based prediction mode for the decoded adjacent partitions adjacent to the target partition.
  • the value of the parameter ⁇ may be set smaller than the value of the parameter ⁇ when the edge-based prediction mode is not applied to the adjacent partition.
  • the prediction direction derived using the edge-based prediction mode for the adjacent partition adjacent to the target partition tends to be close to an appropriate prediction direction for the target partition.
  • the main direction derived for the target partition tends to be close to the optimum prediction direction for the target partition.
  • the above configuration increases the possibility that the sub-direction derived for the target partition matches the optimal prediction direction for the target partition, so that the prediction accuracy is improved and the encoding efficiency is improved.
  • the PU decoding unit 431 included in the video decoding device 1 and the PU determination unit 531 included in the video encoding device 2 determine the number of correction directions that can be selected in the target partition based on the partition size of the target partition. Although determined, the number of correction directions may be determined based on the size of the leaf CU including the target partition. In that case, the number of executions of the determination process of the number of correction directions can be reduced, and the processing amount of the decoding process or the encoding process is reduced.
  • the PU decoding unit 431 included in the video decoding device 1 and the PU determination unit 531 included in the video encoding device 2 use the partition size of the target partition and the size of the leaf CU including the target partition in combination.
  • the number of correction directions may be determined. For example, even when the partition size is the same, if the size of the leaf CU to which the partition belongs is different, a different number of correction directions may be selected. In this case, even if the regions have the same partition size, a different number of correction directions can be added, so that an appropriate number of sub-directions can be set for each region, and the coding efficiency is improved.
  • ⁇ Modification 5 of Embodiment 1> The total number of selectable basic prediction modes for a 4 ⁇ 4 pixel partition is 18 (direction prediction mode: 16, DC prediction mode: 1, Planar prediction mode: 1), and 8 ⁇ 8 pixels, 16 The total number of selectable basic prediction modes for a partition of ⁇ 16 pixels and 32 ⁇ 32 pixels is 35 (direction prediction mode: 33, DC prediction mode: 1, Planar prediction mode: 1). Good.
  • the number of sub-directions for edge-based prediction is set to 0 for 4 ⁇ 4 pixel and 8 ⁇ 8 pixel partitions, In the partition of 16 ⁇ 16 pixels and 32 ⁇ 32 pixels, it is preferable to set the number of sub-directions to four.
  • decoding of the additional index AI can be omitted in the PU information decoding process by the PU decoding unit 431. That is, the decoding process of the additional index in step S108 in FIG. 1 is omitted, and the main direction prediction mode is set as the prediction mode for the target partition.
  • the Planar prediction mode (also referred to as plane prediction mode or plane prediction mode) refers to a mode in which a prediction image in the target partition is predicted two-dimensionally from pixel values of surrounding partitions. Indicates a mode in which the pixel value predSamples [x, y] (where [x, y] represents the coordinates of the pixel) of the predicted image in the target partition is calculated by, for example, the following equation.
  • the number of main directions derived in the edge-based prediction mode is one for each partition. However, depending on the image characteristics of the encoding target image, a plurality of main directions are derived for each partition. It is good also as a possible structure. Also, the total number of main directions that can be derived is changed according to the size of the target partition, and when a plurality of main directions can be derived, one of the main directions is selected and used. It is good.
  • the total number of main directions that can be derived is changed according to the size of the target partition, and when a plurality of main directions can be derived, the plurality of main directions are used.
  • a configuration for selecting and using any of the main directions will be described. In the following, the same parts as those already described are denoted by the same reference numerals, and the description thereof is omitted.
  • the encoded data generated by the moving image encoding device 4 according to the present embodiment and decoded by the moving image decoding device 3 according to the present embodiment has substantially the same configuration as the encoded data # 1 in the first embodiment. However, the following points are different.
  • the intra prediction parameter PP_Intra for a partition of a predetermined size or larger includes a main direction designation index MDI that designates one of a plurality of main directions.
  • the encoded data according to the present embodiment may include the additional index AI described in the first embodiment, or may not include the additional index AI.
  • FIG. 18 illustrates a configuration of a leaf CU decoding unit included in the video decoding device according to the present embodiment.
  • Other configurations of the video decoding device according to the present embodiment are the same as the configurations of the video decoding device 1 according to the first embodiment.
  • the PU decoding unit 431 ′ decodes the PU information about the target leaf CU, thereby determining a division pattern for each partition of the target leaf CU and determining a prediction mode for each partition.
  • Mode # 431 is assigned to each partition.
  • the prediction mode # 431 assigned to each partition is supplied to the predicted image generation unit 432 ′.
  • the prediction mode designates a generation method for generating a prediction image by intra prediction (intra prediction).
  • intra prediction intra prediction
  • Any prediction mode is selected from the basic prediction mode set or the edge-based prediction mode set.
  • the PU decoding unit 431 ' refers to the size of the target partition and determines the number of main directions that can be derived for the target partition. In addition, when a plurality of main directions can be derived for the target partition, the PU decoding unit 431 ′ decodes a main direction designation index that designates any of the plurality of main directions. Also, the PU decoding unit 431 'sets the main direction specified by the main direction index as the prediction direction for the target partition.
  • FIG. 19 is a flowchart showing a flow of PU information decoding processing by the PU decoding unit 431 '.
  • FIG. 20 is a table showing the value of the edge-based prediction flag EF, the value of the main direction designation index MDI, the value of the estimation flag MPM, and the value of the residual prediction mode index RIPM that can be taken for the target partition.
  • Step S501 to (Step S506) and (Step S510) Steps S501 to S506 and S510 shown in FIG. 19 are the same as steps S101 to S106 and S112 of the decoding process performed by the PU decoding unit 431 described in the first embodiment, and thus description thereof is omitted.
  • Step S507 When the edge-based prediction mode is applied to the target partition, the PU decoding unit 431 ′ determines the total number of main directions derived for the target partition according to the size of the target partition.
  • the total number of main directions derived for the target partition preferably has a negative correlation with the size of the target partition.
  • the PU decoding unit 431 ′ sets the total number of main directions derived for the target partition to 2, and when the size of the target partition is 8 ⁇ 8 pixels, The total number of main directions derived for the target partition is set to 1. More generally, when the total number of pixels belonging to the target partition is Npix1, the PU decoding unit 431 ′ sets the total number of main directions that can be selected for the target partition to Nmain1, and the total number of pixels belonging to the target partition is When Npix2 (Npix1 ⁇ Npix2), the total number of main directions that can be selected for the target partition is set to Nmain2 (Nmain1 ⁇ Nmain2).
  • the PU decoding unit 431 ′ is the division information included in the encoded data, and includes CU division information SP_CU that specifies a division pattern for each leaf CU of the target LCU including the target partition, and the target PU (target leaf).
  • the size of the target partition can be identified by referring to the intra PU partition information SP_Intra that specifies the partition pattern of each partition of (CU).
  • the PU decoding unit 431 ′ decodes the main direction designation index MDI for the target partition.
  • Step S509 the PU decoding unit 431 ′ sets the prediction mode for the target partition to the edge-based prediction mode in which the main direction specified by the main direction specifying index decoded in step S508 is the prediction direction.
  • the above is the flow of PU information decoding processing by the PU decoding unit 431 '.
  • the prediction mode set for each partition included in the target leaf CU in step S506 and step S509 is supplied to the predicted image generation unit 432 'as a prediction mode # 431'.
  • the PU decoding unit 431 ′ decodes the additional index AI and sets the prediction mode for the target partition in the main direction specified in step S508. Is set to the edge-based prediction mode in which the sub-direction obtained by adding the correction angle specified by the additional index AI is the prediction direction.
  • the prediction image generation unit 432 ′ refers to the prediction image for each partition included in the target leaf CU with reference to the prediction mode # 431 ′ supplied from the PU decoding unit 431 ′ and the decoded pixel values around the partition And generate.
  • FIG. 21 is a flowchart showing the flow of predicted image generation processing for the target leaf CU by the predicted image generation unit 432 '.
  • Step S601 to (Step S605), and (Step S609) Steps S601 to S605 and S609 shown in FIG. 21 are the same as steps S201 to S205 and S212 of the predicted image generation processing by the predicted image generation unit 432 described in the first embodiment, and thus description thereof is omitted. To do.
  • Step S607 the predicted image generation unit 432 ′ sets the derived main direction as the predicted direction for the target partition.
  • Step S608 The predicted image generation unit 432 ′ generates a predicted image for the target partition by extrapolating the decoded pixel values around the target partition along the prediction direction set for the target partition.
  • the processing in this step is the same as that in step S211 described in the first embodiment.
  • the predicted image generation unit 432 ' generates a predicted image for the target leaf CU by performing the above processing. Also, the generated predicted image for the target leaf CU is supplied to the decoded image generation unit 435 as a predicted image # 432 '.
  • the predicted image generation unit 432 ′ adds the correction angle specified by the additional index AI to the main direction specified by the PU decoding unit 431 ′.
  • the predicted image for the target partition is generated by extrapolating the decoded pixel values around the target partition along the prediction direction obtained as described above.
  • the predicted image generation unit 432 ′ sets, as the main direction 1, the direction indicated by the edge vector having the maximum norm among the edge vectors calculated for the pixels belonging to the reference area set in the decoded area around the target partition.
  • the edge vector for each pixel belonging to the reference region can be calculated, for example, by applying a Sobel filter.
  • the predicted image generation unit 432 ′ derives the main direction 2 with reference to the predicted direction assigned to the decoded partitions around the target partition.
  • the predicted image generation unit 432 ′ determines the main direction (or main direction) derived for the plurality of adjacent partitions.
  • the main direction 2 for the target partition is derived by taking the average of the sub-directions obtained by adding the correction directions.
  • the predicted image generation unit 432 ′ determines the main direction derived by the edge-based prediction mode.
  • the main direction 2 for the target partition is set.
  • the prediction mode having a smaller prediction mode index among the basic prediction modes applied to the plurality of adjacent partitions is indicated.
  • the direction is set to the main direction 2 for the target partition.
  • the predicted image generation unit 432 ′ derives the main direction 2 for the target partition by taking the average of the edge vectors calculated for the pixels belonging to the reference region set in the decoded region around the target partition.
  • the predicted image generation unit 432 ′ sets the direction indicated by the edge vector having the maximum norm among the edge vectors calculated for each pixel belonging to the adjacent partition adjacent to the left side of the target partition as the main direction 1 for the target partition. To do.
  • the predicted image generation unit 432 ′ sets the direction indicated by the edge vector having the maximum norm among the edge vectors calculated for each pixel belonging to the adjacent partition adjacent to the upper side of the target partition as the main direction 2 for the target partition. To do.
  • the predicted image generation unit 432 ′ sets, as the main direction 1, the direction indicated by the edge vector having the maximum norm among the edge vectors calculated for the pixels belonging to the reference area set in the decoded area around the target partition. To derive.
  • the predicted image generation unit 432 ′ determines the direction indicated by the edge vector having the second largest norm among the edge vectors calculated for the pixels belonging to the reference region set in the decoded region around the target partition as the main direction 2.
  • the predicted image generation unit 432 ′ classifies each reference pixel belonging to the reference region set around the target partition into two clusters (cluster A and cluster B), and each of the reference pixels belonging to the cluster A Of the edge vectors calculated for the reference pixel, the direction indicated by the edge vector having the maximum norm is set as the main direction 1, and the norm is the maximum among the edge vectors derived for each reference pixel belonging to the cluster B. The direction indicated by the edge vector is set as the main direction 2.
  • the predicted image generation unit 432 ′ calculates an edge vector for each reference pixel belonging to the reference region, and calculates an edge strength of the edge vector.
  • Step S702 Clustering processing is applied to the three-dimensional vector (dx, dy, es) defined for each reference pixel (or a partial vector of the three-dimensional vector), and each reference pixel is classified into either cluster A or cluster B To do.
  • a k-average method, SVM (Support Vector Machine), or the like can be used as a specific example of the clustering process.
  • Step S703 the predicted image generation unit 432 ′ sets the direction indicated by the edge vector having the maximum norm among the edge vectors for each reference pixel classified into cluster A as the main direction 1 for the target partition. Also, the predicted image generation unit 432 ′ sets the direction indicated by the edge vector having the maximum norm among the edge vectors for each reference pixel classified into the cluster B as the main direction 2 for the target partition.
  • condition for determining which of the two clusters is the cluster A is shared by both the video encoding device according to the present embodiment and the video decoding device according to the present embodiment. Preferably it is.
  • FIG. 22 shows a reference area composed of an adjacent partition adjacent to the target partition OP and a partition sharing the upper left vertex of the target partition, and a cluster A and a cluster B into which each reference pixel belonging to the reference area is classified.
  • weak edges in the substantially vertical direction exist in the target partition OP and all areas of the partitions NP1 to NP3. Further, a strong edge in the substantially horizontal direction exists across the target partition OP and the adjacent partition NP3 adjacent to the left side of the target partition.
  • the predicted image generation unit 432 ′ overlaps the reference area composed of the partitions NP1 to NP3 with the cluster A that includes the strong edge and the cluster A as described above.
  • the main direction is set for each cluster with reference to each cluster, so that the main direction in the substantially horizontal direction which is the optimum prediction direction for the target partition OP is divided.
  • a direction can be derived.
  • the reference area can be divided into Nreg cluster areas, and the main direction can be derived for each cluster area.
  • FIG. 23 illustrates a configuration of a leaf CU encoding unit included in the moving image encoding device 4 according to the present embodiment.
  • Other configurations of the video encoding device 4 according to the present embodiment are the same as the configurations of the video encoding device 2 according to the first embodiment.
  • the PU determination unit 531 ′ refers to the leaf CU image # 242a and the CU information # 242b, and (1) the division pattern of the target leaf CU into each partition, and (2) the prediction mode assigned to each partition. decide.
  • a division pattern into the partitions of the target leaf CU for example, a division pattern that is handled as one partition without dividing the target leaf CU, or the target leaf CU is divided into four partitions symmetrically. One of the division patterns can be selected.
  • the PU determination unit 531 determines the total number of main directions that can be selected for the target partition according to the size of the target partition. More specifically, the PU determination unit 531 determines the total number of main directions that can be selected for the target partition so as to have a positive correlation with the size of the target partition.
  • the total number of main directions that can be selected for the target partition is set to 2, and when the size of the target partition is 8 ⁇ 8 pixels, the target partition can be selected.
  • Set the total number of main directions to 1. More generally, when the total number of pixels belonging to the target partition is Npix1, the PU determining unit 531 ′ determines the total number of main directions that can be selected for the target partition as Nmain1, and the total number of pixels belonging to the target partition is When Npix2 (Npix1 ⁇ Npix2), the total number of main directions that can be selected for the target partition is determined as Nmain2 (Nmain1 ⁇ Nmain2).
  • PU information including (1) an intra split flag (intra_split_flag) that specifies whether or not the target leaf CU is split into four partitions, and (2) prediction mode information that specifies the prediction mode assigned to each partition.
  • # 531 ′ is supplied to the predicted image generation unit 532 ′ and the LCU encoded data generation unit 538.
  • the PU information # 531 ′ includes a main direction designation index MDI that designates which of a plurality of main directions is used, the total number of which is determined according to the size of the target partition. It is.
  • the PU information # 531 ' may include a 1-bit flag that specifies whether to use the main direction or the sub direction for the target partition.
  • the additional index AI may include an index that specifies which of the plurality of sub directions is used.
  • the PU determination unit 531 ′ performs each of all combinations of (1) a possible division pattern for each partition of the target leaf CU and (2) a prediction mode that can be assigned to each partition. PU information corresponding to is generated. In addition, generation of a prediction image, a prediction residual, and a decoded image, which will be described later, is performed for each piece of PU information.
  • the PU determination unit 531 ′ supplies the PU information for optimizing the rate distortion supplied from the rate distortion evaluation unit 537 described later to the header information encoding unit 22 as a part of the leaf CU header information # 53.
  • the PU information for optimizing the rate distortion is supplied to the predicted image generation unit 532 ′ and the LCU encoded data generation unit 538 as final PU information for the target leaf CU.
  • the predicted image generation unit 532 ′ refers to the decoded image # 27 stored in the frame memory 27 and the PU information # 531 ′ supplied from the PU determination unit 531 ′, thereby predicting the predicted image # for the target leaf CU. 532 ′ is generated.
  • the prediction image generation process by the prediction image generation unit 532 ′ is the same as that of the prediction image generation unit 432 ′ included in the video decoding device according to the present embodiment, and thus description thereof is omitted here.
  • the above-described moving image encoding device 2 and moving image decoding device 1 can be used by being mounted on various devices that perform transmission, reception, recording, and reproduction of moving images. The same applies to the moving image encoding device 4 and the moving image decoding device 3, but the following description will be made by taking the moving image encoding device 2 and the moving image decoding device 1 as examples.
  • the moving picture encoding apparatus 2 and the moving picture decoding apparatus 1 described above can be used for transmission and reception of moving pictures.
  • FIG. 25 is a block diagram illustrating a configuration of a transmission device PROD_A in which the moving image encoding device 2 is mounted.
  • the transmission device PROD_A modulates a carrier wave with an encoding unit PROD_A1 that obtains encoded data by encoding a moving image, and the encoded data obtained by the encoding unit PROD_A1.
  • a modulation unit PROD_A2 that obtains a modulation signal and a transmission unit PROD_A3 that transmits the modulation signal obtained by the modulation unit PROD_A2 are provided.
  • the moving image encoding apparatus 2 described above is used as the encoding unit PROD_A1.
  • the transmission device PROD_A is a camera PROD_A4 that captures a moving image, a recording medium PROD_A5 that records the moving image, and an input terminal PROD_A6 for inputting the moving image from the outside as a supply source of the moving image input to the encoding unit PROD_A1. May be further provided.
  • FIG. 25A illustrates a configuration in which the transmission apparatus PROD_A includes all of these, but a part of the configuration may be omitted.
  • the recording medium PROD_A5 may be a recording of a non-encoded moving image, or a recording of a moving image encoded by a recording encoding scheme different from the transmission encoding scheme. It may be a thing. In the latter case, a decoding unit (not shown) for decoding the encoded data read from the recording medium PROD_A5 according to the recording encoding method may be interposed between the recording medium PROD_A5 and the encoding unit PROD_A1.
  • FIG. 25 is a block diagram illustrating a configuration of the receiving device PROD_B in which the moving image decoding device 1 is mounted.
  • the reception device PROD_B includes a reception unit PROD_B1 that receives a modulation signal, a demodulation unit PROD_B2 that obtains encoded data by demodulating the modulation signal received by the reception unit PROD_B1, and a demodulation A decoding unit PROD_B3 that obtains a moving image by decoding the encoded data obtained by the unit PROD_B2.
  • the moving picture decoding apparatus 1 described above is used as the decoding unit PROD_B3.
  • the receiving device PROD_B has a display PROD_B4 for displaying a moving image, a recording medium PROD_B5 for recording the moving image, and an output terminal for outputting the moving image to the outside as a supply destination of the moving image output by the decoding unit PROD_B3.
  • PROD_B6 may be further provided.
  • FIG. 25B illustrates a configuration in which the reception apparatus PROD_B includes all of these, but a part of the configuration may be omitted.
  • the recording medium PROD_B5 may be used for recording a non-encoded moving image, or may be encoded using a recording encoding method different from the transmission encoding method. May be. In the latter case, an encoding unit (not shown) for encoding the moving image acquired from the decoding unit PROD_B3 according to the recording encoding method may be interposed between the decoding unit PROD_B3 and the recording medium PROD_B5.
  • the transmission medium for transmitting the modulation signal may be wireless or wired.
  • the transmission mode for transmitting the modulated signal may be broadcasting (here, a transmission mode in which the transmission destination is not specified in advance) or communication (here, transmission in which the transmission destination is specified in advance). Refers to the embodiment). That is, the transmission of the modulation signal may be realized by any of wireless broadcasting, wired broadcasting, wireless communication, and wired communication.
  • a terrestrial digital broadcast broadcasting station (broadcasting equipment or the like) / receiving station (such as a television receiver) is an example of a transmitting device PROD_A / receiving device PROD_B that transmits and receives a modulated signal by wireless broadcasting.
  • a broadcasting station (such as broadcasting equipment) / receiving station (such as a television receiver) of cable television broadcasting is an example of a transmitting device PROD_A / receiving device PROD_B that transmits and receives a modulated signal by cable broadcasting.
  • a server workstation etc.
  • Client television receiver, personal computer, smart phone etc.
  • VOD Video On Demand
  • video sharing service using the Internet is a transmitting device for transmitting and receiving modulated signals by communication.
  • PROD_A / reception device PROD_B usually, either a wireless or wired transmission medium is used in a LAN, and a wired transmission medium is used in a WAN.
  • the personal computer includes a desktop PC, a laptop PC, and a tablet PC.
  • the smartphone also includes a multi-function mobile phone terminal.
  • the video sharing service client has a function of encoding a moving image captured by the camera and uploading it to the server. That is, the client of the video sharing service functions as both the transmission device PROD_A and the reception device PROD_B.
  • moving image encoding device 2 and moving image decoding device 1 can be used for recording and reproduction of moving images.
  • FIG. 26 (a) is a block diagram showing a configuration of a recording apparatus PROD_C in which the above-described moving picture encoding apparatus 2 is mounted.
  • the recording device PROD_C includes an encoding unit PROD_C1 that obtains encoded data by encoding a moving image, and the encoded data obtained by the encoding unit PROD_C1 on the recording medium PROD_M.
  • the moving image encoding apparatus 2 described above is used as the encoding unit PROD_C1.
  • the recording medium PROD_M may be of a type built in the recording device PROD_C, such as (1) HDD (Hard Disk Drive) or SSD (Solid State Drive), or (2) SD memory. It may be of the type connected to the recording device PROD_C, such as a card or USB (Universal Serial Bus) flash memory, or (3) DVD (Digital Versatile Disk) or BD (Blu-ray Disk: registration) Or a drive device (not shown) built in the recording device PROD_C.
  • HDD Hard Disk Drive
  • SSD Solid State Drive
  • SD memory such as a card or USB (Universal Serial Bus) flash memory, or (3) DVD (Digital Versatile Disk) or BD (Blu-ray Disk: registration) Or a drive device (not shown) built in the recording device PROD_C.
  • the recording device PROD_C receives a moving image as a supply source of a moving image to be input to the encoding unit PROD_C1, a camera PROD_C3 that captures a moving image, an input terminal PROD_C4 for inputting a moving image from the outside, and a moving image.
  • the receiving unit PROD_C5 may be further provided. In FIG. 26A, a configuration in which all of these are provided in the recording apparatus PROD_C is illustrated, but a part may be omitted.
  • the receiving unit PROD_C5 may receive a non-encoded moving image, or may receive encoded data encoded by a transmission encoding scheme different from the recording encoding scheme. You may do. In the latter case, a transmission decoding unit (not shown) that decodes encoded data encoded by the transmission encoding method may be interposed between the reception unit PROD_C5 and the encoding unit PROD_C1.
  • Examples of such a recording device PROD_C include a DVD recorder, a BD recorder, and an HD (Hard Disk) recorder (in this case, the input terminal PROD_C4 or the receiving unit PROD_C5 is a main source of moving images).
  • a camcorder in this case, the camera PROD_C3 is a main source of moving images
  • a personal computer in this case, the receiving unit PROD_C5 is a main source of moving images
  • a smartphone in this case, the camera PROD_C3 or The receiving unit PROD_C5 is a main source of moving images
  • the recording apparatus PROD_C is an example of such a recording apparatus PROD_C.
  • FIG. 26 is a block showing a configuration of a playback device PROD_D equipped with the above-described video decoding device 1.
  • the playback device PROD_D reads a moving image by decoding a read unit PROD_D1 that reads encoded data written to the recording medium PROD_M and a coded data read by the read unit PROD_D1. And a decoding unit PROD_D2 to be obtained.
  • the moving picture decoding apparatus 1 described above is used as the decoding unit PROD_D2.
  • the recording medium PROD_M may be of the type built into the playback device PROD_D, such as (1) HDD or SSD, or (2) such as an SD memory card or USB flash memory, It may be of a type connected to the playback device PROD_D, or (3) may be loaded into a drive device (not shown) built in the playback device PROD_D, such as DVD or BD. Good.
  • the playback device PROD_D has a display PROD_D3 that displays a moving image, an output terminal PROD_D4 that outputs the moving image to the outside, and a transmission unit that transmits the moving image as a supply destination of the moving image output by the decoding unit PROD_D2.
  • PROD_D5 may be further provided.
  • FIG. 26B illustrates a configuration in which the playback apparatus PROD_D includes all of these, but a part may be omitted.
  • the transmission unit PROD_D5 may transmit an unencoded moving image, or transmits encoded data encoded by a transmission encoding method different from the recording encoding method. You may do. In the latter case, it is preferable to interpose an encoding unit (not shown) that encodes a moving image with an encoding method for transmission between the decoding unit PROD_D2 and the transmission unit PROD_D5.
  • Examples of such a playback device PROD_D include a DVD player, a BD player, and an HDD player (in this case, an output terminal PROD_D4 to which a television receiver or the like is connected is a main supply destination of moving images). .
  • a television receiver in this case, the display PROD_D3 is a main destination of moving images
  • a desktop PC in this case, the output terminal PROD_D4 or the transmission unit PROD_D5 is a main destination of moving images
  • Laptop type or tablet type PC in this case, display PROD_D3 or transmission unit PROD_D5 is the main supply destination of moving images
  • smartphone in this case, display PROD_D3 or transmission unit PROD_D5 is the main supply destination of moving images)
  • each block of the above-described moving picture decoding apparatuses 1 and 3 and moving picture encoding apparatuses 2 and 4 may be realized in hardware by a logic circuit formed on an integrated circuit (IC chip). Alternatively, it may be realized by software using a CPU (Central Processing Unit).
  • IC chip integrated circuit
  • CPU Central Processing Unit
  • each device includes a CPU that executes instructions of a program that realizes each function, a ROM (Read (Memory) that stores the program, a RAM (Random Memory) that expands the program, the program, and various types
  • a storage device such as a memory for storing data is provided.
  • An object of the present invention is to provide a recording medium in which a program code (execution format program, intermediate code program, source program) of a control program of each of the above devices, which is software that realizes the above-described functions, is recorded so as to be readable by a computer. This can also be achieved by supplying to each of the above devices and reading and executing the program code recorded on the recording medium by the computer (or CPU or MPU).
  • Examples of the recording medium include tapes such as magnetic tapes and cassette tapes, magnetic disks such as floppy (registered trademark) disks / hard disks, and disks including optical disks such as CD-ROM / MO / MD / DVD / CD-R.
  • IC cards including memory cards
  • semiconductor memories such as mask ROM / EPROM / EEPROM / flash ROM, or PLD (Programmable logic device) or FPGA (Field Programmable Gate Array) Logic circuits can be used.
  • each of the above devices may be configured to be connectable to a communication network, and the program code may be supplied via the communication network.
  • the communication network is not particularly limited as long as it can transmit the program code.
  • the Internet intranet, extranet, LAN, ISDN, VAN, CATV communication network, virtual private network (Virtual Private Network), telephone line network, mobile communication network, satellite communication network, etc. can be used.
  • the transmission medium constituting the communication network may be any medium that can transmit the program code, and is not limited to a specific configuration or type.
  • wired lines such as IEEE 1394, USB, power line carrier, cable TV line, telephone line, ADSL (Asymmetric Digital Subscriber Line) line, infrared rays such as IrDA and remote control, Bluetooth (registered trademark), IEEE 802.11 wireless, HDR ( It can also be used by wireless such as High Data Rate, NFC (Near Field Communication), DLNA (Digital Living Network Alliance), mobile phone network, satellite line, and terrestrial digital network.
  • wired lines such as IEEE 1394, USB, power line carrier, cable TV line, telephone line, ADSL (Asymmetric Digital Subscriber Line) line, infrared rays such as IrDA and remote control, Bluetooth (registered trademark), IEEE 802.11 wireless, HDR ( It can also be used by wireless such as High Data Rate, NFC (Near Field Communication), DLNA (Digital Living Network Alliance), mobile phone network, satellite line, and terrestrial digital network.
  • wired lines such as IEEE 1394, USB, power line carrier, cable TV line, telephone line, ADSL (Asymmetric Digital Subscriber Line) line,
  • the image decoding apparatus includes a prediction mode belonging to a prediction mode group in a prediction residual decoded from encoded data together with prediction mode specification information and other side information, and the prediction mode specification information
  • the number N of prediction modes to be added to the prediction mode group related to the target prediction unit is encoded in the side information
  • a parameter setting unit that refers to a coding unit that includes the target prediction unit or a coding parameter that is decoded prior to the coding unit; and the target prediction unit A prediction parameter corresponding to each of the N prediction modes to be added to the prediction mode group relating to the decoded region of the decoded image. It is characterized in that it comprises a, and deriving means for deriving from.
  • the prediction unit is a type that increases the prediction accuracy by increasing the number of prediction modes belonging to the prediction mode group, and the coding efficiency is improved by reducing the number of prediction modes belonging to the prediction mode group and improving the coding efficiency. Some types improve efficiency. According to the knowledge obtained by the inventors, whether a certain prediction unit belongs to the former type or the latter type is usually encoded that is decoded before the prediction unit or the encoding unit. It can be determined from the coding parameters for the unit.
  • the coding parameter is a parameter that is referred to in order to generate a decoded image, a locally decoded image, or a predicted image, and is referred to in a motion vector or intra prediction that is referred to in inter-screen prediction.
  • the prediction parameters such as the prediction mode, the size and shape of the partition, the size and shape of the block, the size and shape of the encoding unit, and residual data between the original image and the predicted image are included.
  • the side information refers to a set of all information except for the residual data among the information included in the encoding parameter.
  • the encoding parameter related to the encoding unit including the target prediction unit is related to the encoding parameter associated with the encoding unit and one or a plurality of prediction units included in the encoding unit. Encoding parameters are included.
  • the prediction unit may be a PU (Prediction Unit) described in the embodiment, or may be a partition obtained by dividing the PU.
  • the coding unit may be a maximum coding unit (LCU: Largegest Coding Unit) described in the embodiment, or a coding unit (CU: Coding Unit) obtained by dividing an LCU. Also good.
  • the deriving unit derives a main prediction parameter corresponding to a main prediction mode included in the prediction mode group from the decoded region and adds the main prediction parameter to the prediction mode group. It is preferable that a sub prediction parameter corresponding to each of the N sub prediction modes and having a specific relationship with the main prediction parameter is derived from the main prediction parameter.
  • the prediction accuracy is improved by using the prediction mode group including the main prediction mode. Can be achieved.
  • the sub prediction parameters corresponding to each of the N sub prediction modes to be added to the prediction mode group, the sub prediction parameters having a specific relationship with the main prediction parameter are converted into the main prediction parameters. Since it derives from the prediction parameters, the prediction accuracy can be further improved.
  • the total number of sub-prediction modes can be determined with reference to a coding unit that includes the target prediction unit or a coding parameter related to a coding unit that is decoded prior to the coding unit. It is possible to prevent a decrease in coding efficiency due to an increase in the amount of code that can be caused by excessively adding the sub prediction mode.
  • the main prediction parameter represents the edge direction ⁇ m in the target prediction unit estimated from the pixel value of the decoded region
  • the main prediction mode is the value of the decoded region.
  • This is a prediction mode for generating a prediction image on the target prediction unit by extrapolating the pixel value from the extrapolation direction represented by the main prediction parameter, and the N sub-prediction parameters have ⁇ m as the edge direction.
  • the N sub-prediction modes exclude the pixel values of the decoded region from the extrapolation direction represented by each of the N sub-prediction parameters. It is preferable that it is a prediction mode which produces
  • the target prediction unit when the edge exists in the encoding target image can be generated with high prediction accuracy.
  • the sub prediction mode can be used, so high prediction accuracy can be achieved. Can keep.
  • the total number of sub-prediction modes can be determined with reference to a coding unit that includes the target prediction unit or a coding parameter related to a coding unit that is decoded prior to the coding unit. It is possible to prevent a decrease in coding efficiency due to an increase in the amount of code that can be caused by excessively adding the sub prediction mode.
  • the prediction mode group includes a basic prediction mode in which a prediction parameter is predetermined in addition to the main prediction mode and the sub prediction mode, and the derivation is performed.
  • the means is configured such that the prediction mode designated by the prediction mode designation information relating to the adjacent prediction unit adjacent to the target prediction unit, the interval ⁇ in the extrapolation direction represented by each of the main prediction parameter and the sub prediction parameter is the main prediction mode. Or it is preferable to adjust to the value according to whether it is the said sub prediction mode or the said basic prediction mode.
  • the prediction unit in which the main prediction mode is selected tends to have a relatively strong edge.
  • such strong edges often exist continuously across a plurality of prediction units. Therefore, the main prediction parameter derived using the main prediction mode for the adjacent prediction unit tends to be close to the optimal prediction parameter for the target prediction unit.
  • the extrapolation direction interval ⁇ represented by each of the main prediction parameter and the sub prediction parameter for the target prediction unit is the main prediction mode, the sub prediction mode, or the basic prediction mode.
  • the side information includes an encoding parameter that defines a size of the target prediction unit, and the setting unit adds the number N of prediction modes to be added to the prediction mode group. Is preferably set to a value according to the size of the target prediction unit defined by the encoding parameter.
  • the code amount of side information necessary for encoding an encoding target image for a certain encoding unit has a positive correlation with the total number of prediction units included in the encoding unit. That is, if the total number of prediction units is large, the code amount of the side information increases.
  • the total number of prediction units included in a certain coding unit has a negative correlation with the size of each prediction unit.
  • the said setting means is high by setting the prediction mode number N added to the said prediction mode group to the value according to the size of the said object prediction unit prescribed
  • the side information includes an encoding parameter that defines the number M of basic prediction modes
  • the prediction mode group includes M basic prediction modes.
  • the setting means preferably sets the number N of prediction modes to be added to the prediction mode group to a value corresponding to the number M of basic prediction modes defined by the encoding parameter.
  • the prediction unit around the prediction unit in which the total number of selectable basic prediction modes is set to be large is that the coding efficiency is improved by increasing the number of prediction modes belonging to the prediction mode group and increasing the prediction accuracy. Tend.
  • the prediction mode group includes a basic prediction mode in which a prediction parameter is predetermined in addition to the main prediction mode and the sub prediction mode, and the setting The means is that the prediction mode designated by the prediction mode designation information related to the adjacent prediction unit adjacent to the target prediction unit for the number N of prediction modes to be added to the prediction mode group is the main prediction mode or the sub prediction mode. It is preferable to set the value according to whether the basic prediction mode is selected.
  • the prediction unit in which the main prediction mode is selected tends to have a relatively strong edge.
  • such strong edges often exist continuously across a plurality of prediction units. Therefore, the main prediction parameter derived using the main prediction mode for the adjacent prediction unit tends to be close to the optimal prediction parameter for the target prediction unit.
  • designated information regarding the adjacent prediction unit adjacent to the said prediction unit for the prediction mode number N added to the said prediction mode group is the said main prediction mode or the said sub prediction mode.
  • the deriving unit derives a main prediction parameter corresponding to each of N main prediction modes to be added to the prediction mode group from the decoded region. .
  • the deriving unit divides the decoded region into N regions and derives the main prediction mode from each of the obtained N regions.
  • an encoding target image is often composed of a plurality of regions having different image directivities.
  • the derivation unit divides the decoded region into N regions, and derives the main prediction mode from each of the obtained N regions, so that the decoded images are mutually imaged. Even if it is comprised from several area
  • the image coding apparatus designates the selected prediction mode based on the prediction residual obtained by subtracting the predicted image generated according to the prediction mode selected from the prediction mode group for each prediction unit from the original image.
  • the number N of prediction modes to be added to the prediction mode group related to the target prediction unit is encoded in the side information.
  • a parameter setting unit that refers to a coding unit that includes the target prediction unit or a coding parameter that is encoded prior to the coding unit; and the target prediction Deriving prediction parameters corresponding to each of the N prediction modes to be added to the prediction mode group related to the unit from the decoded region of the locally decoded image It is characterized in that comprises a derivation means that, the.
  • the prediction unit is a type that increases the prediction accuracy by increasing the number of prediction modes belonging to the prediction mode group, and the coding efficiency is improved by reducing the number of prediction modes belonging to the prediction mode group and improving the coding efficiency. Some types improve efficiency. According to the knowledge obtained by the inventors, whether a certain prediction unit belongs to the former type or the latter type is usually a code encoded before the prediction unit or the encoding unit. It can be determined from the encoding parameters for the encoding unit.
  • the encoding efficiency is increased by referring to the encoding parameter including the target prediction unit or the encoding parameter related to the encoding unit encoded before the encoding unit.
  • the number of prediction modes to be added to the prediction mode group related to the target prediction unit can be set. Therefore, according to said structure, encoding efficiency can be improved by suppressing the increase in the code amount of coding data, maintaining high prediction accuracy.
  • the coding parameter is a parameter that is referred to in order to generate a decoded image, a locally decoded image, or a predicted image, and is referred to in a motion vector or intra prediction that is referred to in inter-screen prediction.
  • the prediction parameters such as the prediction mode, the size and shape of the partition, the size and shape of the block, the size and shape of the encoding unit, and residual data between the original image and the predicted image are included.
  • the side information refers to a set of all information except for the residual data among the information included in the encoding parameter.
  • the encoding parameter related to the encoding unit including the target prediction unit is related to the encoding parameter associated with the encoding unit and one or a plurality of prediction units included in the encoding unit. Encoding parameters are included.
  • the prediction unit may be a PU (Prediction Unit) described in the embodiment, or may be a partition obtained by dividing the PU.
  • the coding unit may be a maximum coding unit (LCU: Largegest Coding Unit) described in the embodiment, or a coding unit (CU: Coding Unit) obtained by dividing an LCU. Also good.
  • the data structure of the encoded data according to the present invention includes a prediction residual obtained by subtracting, from the original image, a prediction image generated according to a prediction mode selected from a prediction mode group for each prediction unit.
  • the side information includes the number of prediction modes to be added to the prediction mode group related to the target prediction unit, which is a data structure of encoded data generated by encoding together with the prediction mode specifying information to be specified and other side information. Coding parameters that are expressed implicitly by a coding unit that includes the target prediction unit or a coding parameter that is coded before the coding unit. It is characterized by.
  • the present invention can be suitably applied to a decoding device that decodes encoded data and an encoding device that generates encoded data. Further, the present invention can be suitably applied to the data structure of encoded data generated by the encoding device and referenced by the decoding device.
  • Video decoding device 11 Variable length code demultiplexing unit 12 Header information decoding unit 13 LCU setting unit 14 LCU decoding unit (setting means, deriving means) 15 Frame memory 2, 4 Video encoding device 21 LCU header information determination unit 22 Header information encoding unit 23 LCU setting unit 24 LCU encoding unit (setting unit, deriving unit) 25 Variable length code multiplexing unit 26 LCU decoding unit 27 Frame memory

Landscapes

  • Engineering & Computer Science (AREA)
  • Multimedia (AREA)
  • Signal Processing (AREA)
  • Compression Or Coding Systems Of Tv Signals (AREA)

Abstract

An LCU decoding unit (14) with which a moving image decoding device (1) is equipped sets the number of prediction modes N that are added to the prediction mode group pertaining to a target prediction unit by referencing an encoding unit that contains the aforementioned target prediction unit or by referencing an encoding parameter pertaining to an encoding unit that has been decoded prior to said encoding unit and, based on the regions of the decoded image that previously have been decoded, derives the prediction parameters corresponding to each of the N prediction modes that have been added to the prediction mode group pertaining to the target prediction group.

Description

画像復号装置、画像符号化装置、および、データ構造Image decoding apparatus, image encoding apparatus, and data structure
 本発明は、符号化データを復号する画像復号装置、および、符号化データを生成する画像符号化装置に関する。また、画像符号化装置によって生成され、画像復号装置によって参照される符号化データのデータ構造に関する。 The present invention relates to an image decoding device that decodes encoded data, and an image encoding device that generates encoded data. The present invention also relates to a data structure of encoded data generated by the image encoding device and referenced by the image decoding device.
 動画像を効率的に伝送または記録するために、動画像を符号化することによって符号化データを生成する動画像符号化装置(画像符号化装置)、および、当該符号化データを復号することによって復号画像を生成する動画像復号装置(画像復号装置)が用いられている。具体的な動画像符号化方式としては、例えば、H.264/MPEG-4.AVC(非特許文献1)、および、VCEG(Video Coding Expert Group)における共同開発用コーデックであるKTAソフトウェアに採用されている方式、および、その後継コーデックであるTMuC(Test Model under Consideration)ソフトウェアに採用されている方式などが挙げられる。 In order to efficiently transmit or record a moving image, a moving image encoding device (image encoding device) that generates encoded data by encoding the moving image, and decoding the encoded data A video decoding device (image decoding device) that generates a decoded image is used. As a specific moving picture encoding method, for example, H.264 is used. H.264 / MPEG-4. Adopted in the KTA software, which is a codec for joint development in AVC (Non-Patent Document 1) and VCEG (Video Coding Expert Group), and TMuC (Test Model Under Consideration) software, the successor codec The method that has been used.
 このような符号化方式において、動画像を構成する画像(ピクチャ)は、画像を分割することにより得られるスライス、スライスを分割することにより得られる符号化単位(マクロブロックまたは最大コーディングユニット(Largest Coding Unit)と呼ばれることもある)、及び、符号化単位を分割することより得られるブロックおよびパーティションからなる階層構造により管理される。 In such an encoding method, an image (picture) constituting a moving image includes a slice obtained by dividing the image, a coding unit obtained by dividing the slice (macroblock or maximum coding unit (Largest Coding) And a hierarchical structure composed of blocks and partitions obtained by dividing the coding unit.
 また、このような符号化方式においては、通常、入力画像を符号化/復号することによって得られる局所復号画像に基づいて予測画像が生成され、当該予測画像を入力画像(原画像)から減算して得られる予測残差(「差分画像」または「残差画像」と呼ぶこともある)が符号化される。また、予測画像の生成方法としては、画面間予測(インター予測)、および、画面内予測(イントラ予測)が挙げられる。 In such an encoding method, a predicted image is usually generated based on a locally decoded image obtained by encoding / decoding an input image, and the predicted image is subtracted from the input image (original image). The prediction residual (which may be referred to as “difference image” or “residual image”) is encoded. In addition, examples of the method for generating a predicted image include inter-screen prediction (inter prediction) and intra-screen prediction (intra prediction).
 インター予測においては、復号済みのフレームを参照フレームとして、動きベクトルを用いた動き補償を適用することによって、復号中のフレーム内の予測画像が予測単位毎に生成される。 In inter prediction, a predicted image in a frame being decoded is generated for each prediction unit by applying motion compensation using a motion vector with the decoded frame as a reference frame.
 一方、イントラ予測においては、復号中のフレームの復号済領域に基づいて、復号中のフレーム内の予測画像が予測単位毎に生成される。H.264/MPEG-4.AVCに用いられているイントラ予測の一例としては、予測単位(例えば、パーティション)毎に、(1)予め定められた予測モード群(「基本予測モード群」とも呼ぶことがある)から何れかの予測モードを選択し、(2)復号済領域の画素値を選択した予測モードに対応する外挿方向(予測方向)に外挿することによって、該予測単位上の画素値を生成する方法(「基本予測」とも呼ぶことがある)が挙げられる。 On the other hand, in intra prediction, a predicted image in a frame being decoded is generated for each prediction unit based on a decoded area of the frame being decoded. H. H.264 / MPEG-4. As an example of intra prediction used in AVC, any one of (1) a predetermined prediction mode group (also referred to as “basic prediction mode group”) for each prediction unit (for example, partition) A method for generating a pixel value on the prediction unit by selecting a prediction mode and (2) extrapolating a pixel value of a decoded area in an extrapolation direction (prediction direction) corresponding to the selected prediction mode (“ Sometimes referred to as “basic prediction”).
 非特許文献2には、予測単位毎に、該予測単位の周辺に位置する画素の画素値に基づき該予測単位におけるエッジ方向を推定し、復号済領域の画素値を推定したエッジ方向に外挿することによって、該予測単位上の画素値を生成する方法(Differential Coding of Intra Modes(DCIM)とも呼ばれる方法、「エッジ予測」または「エッジベース予測」とも呼ぶことがある)が開示されている。 In Non-Patent Document 2, for each prediction unit, the edge direction in the prediction unit is estimated based on the pixel values of pixels located around the prediction unit, and the pixel value of the decoded area is extrapolated to the estimated edge direction. Thus, a method for generating a pixel value on the prediction unit (a method called Differential Coding of Mode Intra Modes (DCIM), sometimes called “edge prediction” or “edge based prediction”) is disclosed.
 以下では、図24の(a)~(c)を参照して、非特許文献2に開示されている予測画像生成方法についてより具体的に説明する。図24の(a)は、処理対象の予測単位(CurrentBlock)と、該予測単位の周辺に位置する画素とを模式的に示す図である。 Hereinafter, the predicted image generation method disclosed in Non-Patent Document 2 will be described more specifically with reference to FIGS. FIG. 24A is a diagram schematically illustrating a prediction unit (CurrentBlock) to be processed and pixels located around the prediction unit.
 非特許文献2に開示されている方法によれば、まず、該予測単位の周辺に位置する画素の各々について、エッジベクトルai(i=1~N、Nは参照する周辺画素の総数)が算出される。ここで、エッジベクトルaiの算出には、ソーベル演算子(Sobel operators、ソーベルフィルタとも呼ぶ)が用いられる。 According to the method disclosed in Non-Patent Document 2, first, for each pixel located in the vicinity of the prediction unit, an edge vector a i (i = 1 to N, N is the total number of surrounding pixels to be referred to). Calculated. Here, a Sobel operator (also referred to as a Sobel filter) is used to calculate the edge vector a i .
 続いて、関数S(θ)=Σ<e,ai2が導出される。ここで、eは、自身の方向と水平方向とのなす角がθである単位ベクトルを表しており、記号<,>は、両者のベクトルの内積を表している。また、記号Σは、添え字iについて1からNまでの和をとることを示している。 Subsequently, the function S (θ) = Σ <e, a i > 2 is derived. Here, e represents a unit vector whose angle between the direction of itself and the horizontal direction is θ, and the symbol <,> represents the inner product of both vectors. The symbol Σ indicates that a sum from 1 to N is taken for the subscript i.
 続いて、関数S(θ)を最大化する引数θ*=argmaxS(θ)が算出され、θ*によって表される方向が予測方向に設定される。最後に、復号済領域の画素値を、当該予測方向に外挿することによって、該予測単位上の画素値が生成される。ここで、関数S(θ)を最大化する引数θ*は、処理対象の予測単位におけるエッジ方向の推定値を表し、「近傍により示された予測方向」(neighbors' suggested prediction direction)とも呼ばれる。なお、引数θ*の算出は、符号化装置、および、復号装置の双方において行われるため、引数θ*自体は符号化されない。 Subsequently, an argument θ * = argmaxS (θ) that maximizes the function S (θ) is calculated, and the direction represented by θ * is set as the prediction direction. Finally, the pixel value on the prediction unit is generated by extrapolating the pixel value of the decoded area in the prediction direction. Here, the argument θ * that maximizes the function S (θ) represents an estimated value of the edge direction in the prediction unit to be processed, and is also referred to as “neighbors' suggested prediction direction”. Note that the calculation of the argument θ * is performed in both the encoding device and the decoding device, and thus the argument θ * itself is not encoded.
 また、非特許文献2に開示されている方法によれば、予測方向として、θ*+Δθによって表される方向を用いることもできる。ここで、Δθは、方向補正(direction adjustment)を示しており、符号化装置にて用いられたΔθは、符号化され復号装置に伝送される必要がある。具体的には、Δθは、量子化ステップサイズδθを用いて量子化され、以下の量子化インデックスk=Δθ/δθが符号化される。復号装置は、θ*+k×δθを予測方向に設定する。図24の(b)には、量子化インデックスk(k=-2、-1、0、1、2)によって指定される予測方向の例が示されている。また、図24の(c)には、量子化インデックスkを指定するためのシンタックスadjust_neighb_dirを含むパラメータセットが示されている。図24の(b)に示した例では、基本予測モード群に対して、量子化インデックスk=-2、-1、0、1、2にそれぞれ対応する合計5つの追加予測モードが追加され、それら5つの追加予測モードの何れを用いるかは、シンタックスadjust_neighb_dirによって指定される。 Further, according to the method disclosed in Non-Patent Document 2, the direction represented by θ * + Δθ can be used as the prediction direction. Here, Δθ indicates direction adjustment, and Δθ used in the encoding device needs to be encoded and transmitted to the decoding device. Specifically, Δθ is quantized using a quantization step size δθ, and the following quantization index k = Δθ / δθ is encoded. The decoding apparatus sets θ * + k × δθ as the prediction direction. FIG. 24B shows an example of the prediction direction specified by the quantization index k (k = −2, −1, 0, 1, 2). FIG. 24C shows a parameter set including the syntax adjust_neighb_dir for specifying the quantization index k. In the example shown in FIG. 24B, a total of five additional prediction modes respectively corresponding to the quantization indexes k = −2, −1, 0, 1 and 2 are added to the basic prediction mode group. Which of these five additional prediction modes is used is specified by the syntax adjust_neighb_dir.
 このように、非特許文献2に開示されている方法によれば、H.264/MPEG-4.AVCにおいて用いられている基本予測モード群に対して、1または複数の予測モードを追加することにより、選択可能な予測モード数が増えるので、予測精度の向上を図ることができる。 Thus, according to the method disclosed in Non-Patent Document 2, H. H.264 / MPEG-4. By adding one or a plurality of prediction modes to the basic prediction mode group used in AVC, the number of selectable prediction modes increases, so that the prediction accuracy can be improved.
 しかしながら、非特許文献2に開示されている方法では、基本予測モード群に追加した1又は複数の予測モードのうち何れかの予測モードを指定するためのシンタックスが必要となるため、符号化データの符号量が増大という側面もある。このため、追加予測モードが頻繁に用いられる場合には、符号化効率が向上しない、あるいは、符号化効率が期待される程向上しない場合があるという問題が生じていた。 However, since the method disclosed in Non-Patent Document 2 requires a syntax for designating one of the one or more prediction modes added to the basic prediction mode group, the encoded data There is also an aspect that the amount of codes increases. For this reason, when the additional prediction mode is frequently used, there has been a problem that the encoding efficiency may not be improved, or the encoding efficiency may not be improved as expected.
 本発明は上記の問題に鑑みてなされたものであり、その目的は、基本予測モード群に、予測モードを追加する場合であっても、高い予測精度を維持しつつ、符号化データの符号量の増大を抑制することによって、符号化効率を向上させることのできる画像符号化装置および画像復号装置を実現することにある。 The present invention has been made in view of the above-described problems, and the object of the present invention is to provide a code amount of encoded data while maintaining high prediction accuracy even when a prediction mode is added to the basic prediction mode group. It is to realize an image encoding device and an image decoding device capable of improving the encoding efficiency by suppressing the increase in the image quality.
 上記の問題を解決するために、本発明に係る画像復号装置は、予測モード指定情報その他のサイド情報と共に符号化データから復号した予測残差に、予測モード群に属する予測モードであって、上記予測モード指定情報により指定される予測モードに従って生成した予測画像を加算することによって復号画像を生成する画像復号装置において、対象予測単位に関する予測モード群に追加する予測モード数Nを、上記サイド情報に含まれる符号化パラメータであって、上記対象予測単位を包含する符号化単位、または、該符号化単位よりも先に復号された符号化単位に関する符号化パラメータを参照して設定する設定手段と、上記対象予測単位に関する予測モード群に追加するN個の予測モードの各々に対応する予測パラメータを、上記復号画像の復号済領域から導出する導出手段と、を備えていることを特徴としている。 In order to solve the above problem, an image decoding apparatus according to the present invention is a prediction mode belonging to a prediction mode group in a prediction residual decoded from encoded data together with prediction mode designation information and other side information, In the image decoding apparatus that generates a decoded image by adding the prediction images generated according to the prediction mode specified by the prediction mode specification information, the number N of prediction modes to be added to the prediction mode group related to the target prediction unit is used as the side information. A setting unit that is an encoding parameter that is included and is set with reference to an encoding unit that includes the target prediction unit or an encoding parameter that is decoded prior to the encoding unit; Prediction parameters corresponding to each of the N prediction modes to be added to the prediction mode group related to the target prediction unit are decoded. It is characterized in that it comprises a derivation means for deriving from the decoded region of the image, a.
 予測単位には、予測モード群に属する予測モード数を増やして予測精度を上げる方が符号化効率が向上するタイプと、予測モード群に属する予測モード数を減らして符号量を下げる方が符号化効率が向上するタイプとがある。発明者らが得た知見によれば、ある予測単位が前者のタイプに属するのか後者のタイプに属するのかは、普通、その予測単位、又は、その符号化単位よりも先に復号される符号化単位に関する符号化パラメータから判定することができる。 The prediction unit is a type that increases the prediction accuracy by increasing the number of prediction modes belonging to the prediction mode group, and the coding efficiency is improved by reducing the number of prediction modes belonging to the prediction mode group and improving the coding efficiency. Some types improve efficiency. According to the knowledge obtained by the inventors, whether a certain prediction unit belongs to the former type or the latter type is usually encoded that is decoded before the prediction unit or the encoding unit. It can be determined from the coding parameters for the unit.
 上記の構成によれば、対象予測単位を包含する符号化単位、又は、その符号化単位よりも先に復号される符号化単位に関する符号化パラメータを参照することによって、符号化効率を大きくするように(望ましくは符号化効率を最大化するように)、対象予測単位に関する予測モード群に追加する予測モード数を設定することができる。したがって、上記の構成によれば、高い予測精度を維持しつつ、符号化データの符号量の増大を抑制することによって、符号化効率を向上させることができる。 According to the above configuration, it is possible to increase the coding efficiency by referring to the coding parameter relating to the coding unit including the target prediction unit or the coding unit decoded prior to the coding unit. In addition, it is possible to set the number of prediction modes to be added to the prediction mode group related to the target prediction unit (desirably to maximize the coding efficiency). Therefore, according to said structure, encoding efficiency can be improved by suppressing the increase in the code amount of coding data, maintaining high prediction accuracy.
 なお、上記符号化パラメータとは、復号画像、局所復号画像、または、予測画像を生成するために参照されるパラメータのことであり、画面間予測において参照される動きベクトルや画面内予測において参照される予測モードなどの予測パラメータに加えて、パーティションのサイズや形状、ブロックのサイズや形状、符号化単位のサイズや形状、および、原画像と予測画像との残差データなどが含まれる。また、上記サイド情報とは、符号化パラメータに含まれる情報のうち、上記残差データを除く全ての情報の集合のことを指す。 Note that the coding parameter is a parameter that is referred to in order to generate a decoded image, a locally decoded image, or a predicted image, and is referred to in a motion vector or intra prediction that is referred to in inter-screen prediction. In addition to the prediction parameters such as the prediction mode, the size and shape of the partition, the size and shape of the block, the size and shape of the encoding unit, and residual data between the original image and the predicted image are included. The side information refers to a set of all information except for the residual data among the information included in the encoding parameter.
 また、上記対象予測単位を包含する符号化単位に関する符号化パラメータには、当該符号化単位に関連付けられた符号化パラメータ、および、当該符号化単位に含まれる1または複数の予測単位に関連付けられた符号化パラメータが含まれる。 In addition, the encoding parameter related to the encoding unit including the target prediction unit is related to the encoding parameter associated with the encoding unit and one or a plurality of prediction units included in the encoding unit. Encoding parameters are included.
 また、上記予測単位は、実施形態において説明するPU(Prediction Unit)であってもよいし、PUを分割して得られるパーティションであってもよい。また、上記符号化単位は、実施形態において説明する最大符号化単位(LCU:Largest Coding Unit)であってもよいし、LCUを分割して得られる符号化単位(CU:Coding Unit)であってもよい。 Also, the prediction unit may be a PU (Prediction Unit) described in the embodiment, or may be a partition obtained by dividing the PU. In addition, the coding unit may be a maximum coding unit (LCU: Largegest Coding Unit) described in the embodiment, or a coding unit (CU: Coding Unit) obtained by dividing an LCU. Also good.
 また、本発明に係る画像符号化装置は、予測単位毎に予測モード群から選択した予測モードに従って生成した予測画像を原画像から減算して得られる予測残差を、選択した予測モードを指定する予測モード指定情報その他のサイド情報と共に符号化することによって符号化データを生成する画像符号化装置において、対象予測単位に関する予測モード群に追加する予測モード数Nを、上記サイド情報に含まれる符号化パラメータであって、上記対象予測単位を包含する符号化単位、または、該符号化単位よりも先に符号化された符号化単位に関する符号化パラメータを参照して設定する設定手段と、上記対象予測単位に関する予測モード群に追加するN個の予測モードの各々に対応する予測パラメータを局所復号画像の復号済領域から導出する導出手段と、を備えていることを特徴としている。 In addition, the image coding apparatus according to the present invention designates the selected prediction mode based on the prediction residual obtained by subtracting the predicted image generated according to the prediction mode selected from the prediction mode group for each prediction unit from the original image. In an image encoding device that generates encoded data by encoding together with prediction mode designation information and other side information, the number N of prediction modes to be added to the prediction mode group related to the target prediction unit is encoded in the side information. A parameter setting unit that refers to a coding unit that includes the target prediction unit or a coding parameter that is encoded prior to the coding unit; and the target prediction Deriving prediction parameters corresponding to each of the N prediction modes to be added to the prediction mode group related to the unit from the decoded region of the locally decoded image It is characterized in that comprises a derivation means that, the.
 予測単位には、予測モード群に属する予測モード数を増やして予測精度を上げる方が符号化効率が向上するタイプと、予測モード群に属する予測モード数を減らして符号量を下げる方が符号化効率が向上するタイプとがある。発明者らが得た知見によれば、ある予測単位が前者のタイプに属するのか後者のタイプに属するのかは、普通、その予測単位、又は、その符号化単位よりも先に符号化される符号化単位に関する符号化パラメータから判定することができる。 The prediction unit is a type that increases the prediction accuracy by increasing the number of prediction modes belonging to the prediction mode group, and the coding efficiency is improved by reducing the number of prediction modes belonging to the prediction mode group and improving the coding efficiency. Some types improve efficiency. According to the knowledge obtained by the inventors, whether a certain prediction unit belongs to the former type or the latter type is usually a code encoded before the prediction unit or the encoding unit. It can be determined from the encoding parameters for the encoding unit.
 上記の構成によれば、対象予測単位を包含する符号化単位、又は、その符号化単位よりも先に符号化される符号化単位に関する符号化パラメータを参照することによって、符号化効率を大きくするように(望ましくは符号化効率を最大化するように)、対象予測単位に関する予測モード群に追加する予測モード数を設定することができる。したがって、上記の構成によれば、高い予測精度を維持しつつ、符号化データの符号量の増大を抑制することによって、符号化効率を向上させることができる。 According to the above configuration, the encoding efficiency is increased by referring to the encoding parameter including the target prediction unit or the encoding parameter related to the encoding unit encoded before the encoding unit. As such (desirably to maximize encoding efficiency), the number of prediction modes to be added to the prediction mode group related to the target prediction unit can be set. Therefore, according to said structure, encoding efficiency can be improved by suppressing the increase in the code amount of coding data, maintaining high prediction accuracy.
 なお、上記符号化パラメータとは、復号画像、局所復号画像、または、予測画像を生成するために参照されるパラメータのことであり、画面間予測において参照される動きベクトルや画面内予測において参照される予測モードなどの予測パラメータに加えて、パーティションのサイズや形状、ブロックのサイズや形状、符号化単位のサイズや形状、および、原画像と予測画像との残差データなどが含まれる。また、上記サイド情報とは、符号化パラメータに含まれる情報のうち、上記残差データを除く全ての情報の集合のことを指す。 Note that the coding parameter is a parameter that is referred to in order to generate a decoded image, a locally decoded image, or a predicted image, and is referred to in a motion vector or intra prediction that is referred to in inter-screen prediction. In addition to the prediction parameters such as the prediction mode, the size and shape of the partition, the size and shape of the block, the size and shape of the encoding unit, and residual data between the original image and the predicted image are included. The side information refers to a set of all information except for the residual data among the information included in the encoding parameter.
 また、上記対象予測単位を包含する符号化単位に関する符号化パラメータには、当該符号化単位に関連付けられた符号化パラメータ、および、当該符号化単位に含まれる1または複数の予測単位に関連付けられた符号化パラメータが含まれる。 In addition, the encoding parameter related to the encoding unit including the target prediction unit is related to the encoding parameter associated with the encoding unit and one or a plurality of prediction units included in the encoding unit. Encoding parameters are included.
 また、上記予測単位は、実施形態において説明するPU(Prediction Unit)であってもよいし、PUを分割して得られるパーティションであってもよい。また、上記符号化単位は、実施形態において説明する最大符号化単位(LCU:Largest Coding Unit)であってもよいし、LCUを分割して得られる符号化単位(CU:Coding Unit)であってもよい。 Also, the prediction unit may be a PU (Prediction Unit) described in the embodiment, or may be a partition obtained by dividing the PU. In addition, the coding unit may be a maximum coding unit (LCU: Largegest Coding Unit) described in the embodiment, or a coding unit (CU: Coding Unit) obtained by dividing an LCU. Also good.
 また、本発明に係る符号化データのデータ構造は、予測単位毎に予測モード群から選択した予測モードに従って生成した予測画像を原画像から減算して得られる予測残差を、選択した予測モードを指定する予測モード指定情報その他のサイド情報と共に符号化することによって生成された符号化データのデータ構造であって、対象予測単位に関する予測モード群に追加する予測モードの個数が、上記サイド情報に含まれる符号化パラメータであって、上記対象予測単位を包含する符号化単位、または、該符号化単位よりも先に符号化された符号化単位に関する符号化パラメータによって陰的に表現されている、ことを特徴としている。 In addition, the data structure of the encoded data according to the present invention includes a prediction residual obtained by subtracting, from the original image, a prediction image generated according to a prediction mode selected from a prediction mode group for each prediction unit. The side information includes the number of prediction modes to be added to the prediction mode group related to the target prediction unit, which is a data structure of encoded data generated by encoding together with the prediction mode specifying information to be specified and other side information. Coding parameters that are expressed implicitly by a coding unit that includes the target prediction unit or a coding parameter that is coded before the coding unit. It is characterized by.
 上記のように構成された符号化データのデータ構造によれば、本発明に係る画像復号装置および画像符号化装置と同様の効果を奏する。 According to the data structure of the encoded data configured as described above, the same effects as those of the image decoding apparatus and the image encoding apparatus according to the present invention are achieved.
 以上のように、本発明に係る画像復号装置は、予測モード指定情報その他のサイド情報と共に符号化データから復号した予測残差に、予測モード群に属する予測モードであって、上記予測モード指定情報により指定される予測モードに従って生成した予測画像を加算することによって復号画像を生成する画像復号装置において、対象予測単位に関する予測モード群に追加する予測モード数Nを、上記サイド情報に含まれる符号化パラメータであって、上記対象予測単位を包含する符号化単位、または、該符号化単位よりも先に復号された符号化単位に関する符号化パラメータを参照して設定する設定手段と、上記対象予測単位に関する予測モード群に追加するN個の予測モードの各々に対応する予測パラメータを、上記復号画像の復号済領域から導出する導出手段と、を備えている。 As described above, the image decoding apparatus according to the present invention includes the prediction mode belonging to the prediction mode group in the prediction residual decoded from the encoded data together with the prediction mode specifying information and other side information, and the prediction mode specifying information. In the image decoding device that generates a decoded image by adding the prediction images generated according to the prediction mode specified by the encoding, the number N of prediction modes to be added to the prediction mode group related to the target prediction unit is encoded in the side information A parameter setting unit that refers to a coding unit that includes the target prediction unit or a coding parameter that is decoded prior to the coding unit; and the target prediction unit A prediction parameter corresponding to each of the N prediction modes to be added to the prediction mode group relating to the decoded region of the decoded image. Has a, a derivation means for deriving from.
 上記画像復号装置によれば、高い予測精度を維持しつつ、符号化データの符号量の増大を抑制することによって、符号化効率を向上させることができる。 According to the above image decoding apparatus, it is possible to improve the encoding efficiency by suppressing an increase in the code amount of the encoded data while maintaining high prediction accuracy.
本発明の第1の実施形態に係る動画像復号装置によるPU情報復号処理の流れを示すフローチャートである。It is a flowchart which shows the flow of PU information decoding processing by the moving image decoding apparatus which concerns on the 1st Embodiment of this invention. 本発明の第1の実施形態に係る動画像符号化装置によって生成され、本発明の実施形態に係る動画像復号装置によって参照される符号化データのデータ構造を示すものであって、(a)は、符号化データのピクチャレイヤの構成を示す図であり、(b)は、ピクチャレイヤに含まれるスライスレイヤの構成を示す図であり、(c)は、スライスレイヤに含まれるLCUレイヤの構成を示す図であり、(d)は、LCUレイヤに含まれるリーフCUの構成を示す図であり、(e)は、リーフCUについてのインター予測情報の構成を示す図であり、(f)は、リーフCUについてのイントラ予測情報の構成を示す図である。The data structure of the encoding data produced | generated by the moving image encoder which concerns on the 1st Embodiment of this invention, and referred by the moving image decoder which concerns on embodiment of this invention, Comprising: (a) FIG. 4 is a diagram illustrating a configuration of a picture layer of encoded data, (b) is a diagram illustrating a configuration of a slice layer included in the picture layer, and (c) is a configuration of an LCU layer included in the slice layer. (D) is a figure which shows the structure of the leaf CU contained in an LCU layer, (e) is a figure which shows the structure of the inter prediction information about leaf CU, (f) is It is a figure which shows the structure of the intra prediction information about leaf CU. 本発明の第1の実施形態に係る動画像復号装置の構成を示すブロック図である。It is a block diagram which shows the structure of the moving image decoding apparatus which concerns on the 1st Embodiment of this invention. 本発明の第1の実施形態に係る動画像復号装置の備えるLCU復号部の構成を示すブロック図である。It is a block diagram which shows the structure of the LCU decoding part with which the moving image decoding apparatus which concerns on the 1st Embodiment of this invention is provided. 本発明の第1の実施形態に係るLCU分割構造設定部により参照される分割構造情報を説明するための図であって、(a)は、分割構造情報によって指定されるLCU分割構造を例示する図であり、(b)は、分割構造情報に従って分割されたLCUにおける対象リーフCUを例示する図である。It is a figure for demonstrating the division | segmentation structure information referred by the LCU division | segmentation structure setting part which concerns on the 1st Embodiment of this invention, Comprising: (a) illustrates the LCU division | segmentation structure designated with division | segmentation structure information. (B) is a figure which illustrates object leaf CU in LCU divided | segmented according to division | segmentation structure information. 本発明の第1の実施形態に係るLCU復号部の備えるリーフCU復号部の構成を示すブロック図である。It is a block diagram which shows the structure of the leaf CU decoding part with which the LCU decoding part which concerns on the 1st Embodiment of this invention is provided. 本発明の第1の実施形態に係る動画像復号装置が参照する予測モードであって、複数の基本予測モードよりなる基本予測モードセットに含まれる予測モードを、予測モードインデックスと共に示す図である。It is a prediction mode which the moving image decoding apparatus which concerns on the 1st Embodiment of this invention refers, Comprising: It is a figure which shows the prediction mode contained in the basic prediction mode set which consists of several basic prediction modes with a prediction mode index. 本発明の第1の実施形態に係る動画像復号装置が参照するイントラ予測情報を説明するためのものであって、(a)は、対象パーティションについて取り得るエッジベース予測フラグの値、追加インデックスの値、推定フラグの値、および、残余予測モードインデックスの各値を示す表であり、(b)は、復号された追加インデックスの各値に対応する予測方向を示す図である。It is for demonstrating the intra prediction information which the moving image decoding apparatus which concerns on the 1st Embodiment of this invention refers, Comprising: (a) is the value of the edge base prediction flag which can be taken about an object partition, an additional index It is a table | surface which shows each value of a value, the value of an estimation flag, and a residual prediction mode index, (b) is a figure which shows the prediction direction corresponding to each value of the decoded additional index. 本発明の第1の実施形態に係る動画像復号装置による予測画像生成処理の流れを示すフローチャートである。It is a flowchart which shows the flow of the prediction image generation process by the moving image decoding apparatus which concerns on the 1st Embodiment of this invention. 本発明の第1の実施形態に係る動画像復号装置による予測画像生成処理を説明するための図であって、対象パーティションに属する画素、および、その周辺の復号済み画素を示す図である。It is a figure for demonstrating the prediction image production | generation process by the moving image decoding apparatus which concerns on the 1st Embodiment of this invention, Comprising: It is a figure which shows the pixel which belongs to an object partition, and its surrounding decoded pixel. 本発明の第1の実施形態に係る動画像復号装置において、エッジベース予測モードが選択された場合の、予測画像生成処理を説明するための図であって、対象パーティションと対象パーティションの周辺のパーティションとを示す図である。In the moving picture decoding apparatus which concerns on the 1st Embodiment of this invention, it is a figure for demonstrating a prediction image generation process when edge-based prediction mode is selected, Comprising: The partition around a target partition and a target partition FIG. 本発明の第1の実施形態に係る動画像復号装置による復号処理を説明するためのものであって、(a)は、本発明の第1の実施形態に係る動画像復号装置による復号処理の流れを示すフローチャートであり、(b)は、本発明の第1の実施形態に係る動画像復号装置によるレート歪コストの減少幅を示すグラフである。It is for demonstrating the decoding process by the moving image decoding apparatus which concerns on the 1st Embodiment of this invention, Comprising: (a) is the decoding process by the moving image decoding apparatus which concerns on the 1st Embodiment of this invention. It is a flowchart which shows a flow, (b) is a graph which shows the reduction width of the rate distortion cost by the moving image decoding apparatus which concerns on the 1st Embodiment of this invention. 本発明の第1の実施形態に係る動画像符号化装置の構成を示すブロック図である。It is a block diagram which shows the structure of the moving image encoder which concerns on the 1st Embodiment of this invention. 本発明の第1の実施形態に係る動画像符号化装置の備えるLCU符号化部の構成を示すブロック図である。It is a block diagram which shows the structure of the LCU encoding part with which the moving image encoder which concerns on the 1st Embodiment of this invention is provided. 本発明の第1の実施形態に係るLCU符号化部の備えるリーフCU符号化部の構成を示すブロック図である。It is a block diagram which shows the structure of the leaf CU encoding part with which the LCU encoding part which concerns on the 1st Embodiment of this invention is provided. 本発明の第1の実施形態に係る動画像符号化装置による符号化処理の流れを示すフローチャートである。It is a flowchart which shows the flow of the encoding process by the moving image encoder which concerns on the 1st Embodiment of this invention. 本発明の第1の実施形態に係る動画像復号装置および動画像符号化装置による予測画像生成処理を説明するための図であって、対象パーティションと隣接パーティションとに跨ってエッジが存在する様子を示す図である。It is a figure for demonstrating the prediction image generation process by the moving image decoding apparatus which concerns on the 1st Embodiment of this invention, and a moving image encoding apparatus, Comprising: A mode that an edge exists ranging over a target partition and an adjacent partition. FIG. 本発明の第2の実施形態に係る動画像復号装置の備えるリーフCU復号部の構成を示すブロック図である。It is a block diagram which shows the structure of the leaf CU decoding part with which the moving image decoding apparatus which concerns on the 2nd Embodiment of this invention is provided. 本発明の第2の実施形態に係る動画像復号装置によるPU情報復号処理の流れを示すフローチャートである。It is a flowchart which shows the flow of PU information decoding process by the moving image decoding apparatus which concerns on the 2nd Embodiment of this invention. 本発明の第2の実施形態に係る動画像復号装置が参照するイントラ予測情報を説明するためのものであって、対象パーティションについて取り得るエッジベース予測フラグの値、主方向指定インデックスの値、推定フラグの値、および、残余予測モードインデックスの各値を示す表である。It is for demonstrating the intra prediction information which the moving image decoding apparatus which concerns on the 2nd Embodiment of this invention refers, Comprising: The value of the edge base prediction flag which can be taken about a object partition, the value of a main direction designation | designated index, estimation It is a table | surface which shows each value of the value of a flag, and a residual prediction mode index. 本発明の第2の実施形態に係る動画像復号装置による予測画像生成処理の流れを示すフローチャートである。It is a flowchart which shows the flow of the estimated image generation process by the moving image decoding apparatus which concerns on the 2nd Embodiment of this invention. 本発明の第2の実施形態に係る動画像復号装置による予測画像生成処理を説明するための図であって、対象パーティションと当該対象パーティションの周辺のパーティションに、強い横方向のエッジと、弱い縦方向のエッジが存在する様子を示す図である。It is a figure for demonstrating the prediction image production | generation process by the moving image decoding apparatus which concerns on the 2nd Embodiment of this invention, Comprising: A strong horizontal direction edge and a weak vertical direction are carried out to the partition around the object partition and the said object partition. It is a figure which shows a mode that the edge of a direction exists. 本発明の第2の実施形態に係る動画像符号化装置の備えるリーフCU符号化部の構成を示すブロック図である。It is a block diagram which shows the structure of the leaf CU encoding part with which the moving image encoder which concerns on the 2nd Embodiment of this invention is provided. 従来の動画像復号装置において、エッジベース予測モードが選択された場合の、イントラ予測画像の生成処理を説明するための図であって、(a)は、対象パーティションを対象パーティションの周辺のパーティションと共に示す図であり、(b)は、補正角度を指定するパラメータを、補正後の予測方向と共に示す図であり、(c)は、従来の動画像復号装置が参照するイントラ予測パラメータの構成を示す表である。In the conventional moving image decoding apparatus, it is a figure for demonstrating the production | generation process of an intra prediction image when edge-based prediction mode is selected, Comprising: (a) is a target partition with the periphery partition of a target partition. (B) is a figure which shows the parameter which designates a correction angle with the prediction direction after correction | amendment, (c) shows the structure of the intra prediction parameter which the conventional moving image decoding apparatus refers to. It is a table. 本発明の実施形態に係る動画像復号装置および動画像符号化装置が、動画像の送受信に利用できることを説明するための図であり、(a)は、動画像符号化装置を搭載した送信装置の構成を示したブロック図であり、(b)は、動画像復号装置を搭載した受信装置の構成を示したブロック図である。BRIEF DESCRIPTION OF THE DRAWINGS It is a figure for demonstrating that the moving image decoding apparatus and moving image encoding apparatus which concern on embodiment of this invention can be utilized for transmission / reception of a moving image, (a) is the transmitter which mounts a moving image encoding apparatus FIG. 2B is a block diagram showing a configuration of a receiving device equipped with a video decoding device. 本発明の実施形態に係る動画像復号装置および動画像符号化装置が、動画像の記録および再生に利用できることを説明するための図であり、(a)は、動画像符号化装置を搭載した記録装置の構成を示したブロック図であり、(b)は、動画像復号装置を搭載した再生装置の構成を示したブロックである。BRIEF DESCRIPTION OF THE DRAWINGS It is a figure for demonstrating that the moving image decoding apparatus and moving image encoding apparatus which concern on embodiment of this invention can be utilized for recording and reproduction | regeneration of a moving image, (a) is mounted with the moving image encoding apparatus. It is the block diagram which showed the structure of the recording device, (b) is the block which showed the structure of the reproducing | regenerating apparatus carrying a moving image decoding apparatus.
 本発明に係る画像復号装置および画像符号化装置の実施形態について図面に基づいて説明すれば以下のとおりである。なお、本実施形態に係る画像復号装置は、符号化データから動画像を復号するものである。したがって、以下では、これを「動画像復号装置」と呼称する。また、本実施形態に係る画像符号化装置は、動画像を符号化することによって符号化データを生成するものである。したがって、以下では、これを「動画像符号化装置」と呼ぶ。 Embodiments of an image decoding apparatus and an image encoding apparatus according to the present invention will be described below with reference to the drawings. Note that the image decoding apparatus according to the present embodiment decodes a moving image from encoded data. Therefore, hereinafter, this is referred to as “moving image decoding apparatus”. In addition, the image encoding device according to the present embodiment generates encoded data by encoding a moving image. Therefore, in the following, this is referred to as a “video encoding device”.
 ただし、本発明の適用範囲はこれに限定されるものではない。すなわち、以下の説明からも明らかなように、本発明の特徴はイントラ予測にあり、複数のフレームを前提としなくとも成立するものである。すなわち、動画像を対象とするか静止画像を対象とするかを問わず、画像復号装置一般および画像符号化装置一般に適用できるものである。 However, the scope of application of the present invention is not limited to this. That is, as will be apparent from the following description, the feature of the present invention lies in intra prediction, and is established without assuming a plurality of frames. That is, the present invention can be applied to general image decoding apparatuses and general image encoding apparatuses regardless of whether the target is a moving image or a still image.
 〔実施形態1〕
 以下では、本発明の第1の実施形態について説明する。
[Embodiment 1]
Hereinafter, a first embodiment of the present invention will be described.
 (符号化データ#1の構成)
 本実施形態に係る動画像復号装置1の説明に先立ち、本実施形態に係る動画像符号化装置2によって生成され、動画像復号装置1によって復号される符号化データ#1の構成について、図2を参照して説明する。符号化データ#1は、シーケンスレイヤ、GOP(Group Of Pictures)レイヤ、ピクチャレイヤ、スライスレイヤ、及び、最大符号化単位(LCU:Largest Coding Unit)レイヤからなる階層構造を有している。
(Configuration of encoded data # 1)
Prior to the description of the moving picture decoding apparatus 1 according to the present embodiment, the configuration of the encoded data # 1 generated by the moving picture encoding apparatus 2 according to the present embodiment and decoded by the moving picture decoding apparatus 1 will be described with reference to FIG. Will be described with reference to FIG. The encoded data # 1 has a hierarchical structure including a sequence layer, a GOP (Group Of Pictures) layer, a picture layer, a slice layer, and a maximum coding unit (LCU: Large Coding Unit) layer.
 符号化データ#1におけるピクチャレイヤ以下の階層の構造を図2に示す。図2の(a)~(f)は、それぞれ、ピクチャレイヤP、スライスレイヤS、LCUレイヤLCU、LCUに含まれるリーフCU(図2の(d)ではCULと表記)、インター予測(画面間予測)パーティションについての予測情報PIであるインター予測情報PI_Inter、イントラ予測(画面内予測)パーティションについての予測情報PIであるイントラ予測情報PI_Intraの構造を示す図である。 FIG. 2 shows the hierarchical structure below the picture layer in the encoded data # 1. (A) to (f) in FIG. 2 are a picture layer P, a slice layer S, an LCU layer LCU, a leaf CU included in the LCU (denoted as CUL in FIG. 2 (d)), and inter prediction (inter-screen) It is a figure which shows the structure of inter prediction information PI_Inter which is the prediction information PI about (prediction) partition, and intra prediction information PI_Intra which is the prediction information PI about intra prediction (prediction in a screen) partition.
 (ピクチャレイヤ)
 ピクチャレイヤPは、処理対象のピクチャである対象ピクチャを復号するために動画像復号装置1が参照するデータの集合である。ピクチャレイヤPは、図2の(a)に示すように、ピクチャヘッダPH、及び、スライスレイヤS1~SNsを含んでいる(NsはピクチャレイヤPに含まれるスライスレイヤの総数)。
(Picture layer)
The picture layer P is a set of data that is referenced by the video decoding device 1 in order to decode a target picture that is a processing target picture. As shown in FIG. 2A, the picture layer P includes a picture header PH and slice layers S1 to SNs (Ns is the total number of slice layers included in the picture layer P).
 ピクチャヘッダPHには、対象ピクチャの復号方法を決定するために動画像復号装置1が参照する符号化パラメータ群が含まれている。例えば、動画像符号化装置2が符号化の際に用いた可変長符号化のモードを示す符号化モード情報(entoropy_coding_mode_flag)は、ピクチャヘッダPHに含まれる符号化パラメータの一例である。entorpy_coding_mode_flagが0の場合、当該ピクチャは、CAVLC(Context-based Adaptive Variable Length Coding)によって符号化されており、entorpy_coding_mode_flagが1である場合、当該ピクチャは、CABAC(Context-based Adaptive Binary Arithmetic Coding)によって符号化されている。 The picture header PH includes a coding parameter group referred to by the video decoding device 1 in order to determine a decoding method of the target picture. For example, the encoding mode information (entoropy_coding_mode_flag) indicating the variable length encoding mode used in encoding by the moving image encoding device 2 is an example of an encoding parameter included in the picture header PH. When entorpy_coding_mode_flag is 0, the picture is encoded by CAVLC (Context-based Adaptive Variable Length Coding). It has become.
 (スライスレイヤ)
 ピクチャレイヤPに含まれる各スライスレイヤSは、処理対象のスライスである対象スライスを復号するために動画像復号装置1が参照するデータの集合である。スライスレイヤSは、図2の(b)に示すように、スライスヘッダSH、及び、LCUレイヤLCU1~LCUNc(NcはスライスSに含まれるLCUの総数)を含んでいる。
(Slice layer)
Each slice layer S included in the picture layer P is a set of data referred to by the video decoding device 1 in order to decode a target slice that is a slice to be processed. As shown in FIG. 2B, the slice layer S includes a slice header SH and LCU layers LCU1 to LCUn (Nc is the total number of LCUs included in the slice S).
 スライスヘッダSHには、対象スライスの復号方法を決定するために動画像復号装置1が参照する符号化パラメータ群が含まれる。スライスタイプを指定するスライスタイプ指定情報(slice_type)は、スライスヘッダSHに含まれる符号化パラメータの一例である。また、スライスヘッダSHには、動画像復号装置1の備えるループフィルタによって参照されるフィルタパラメータFPが含まれる。 The slice header SH includes a coding parameter group that the moving image decoding apparatus 1 refers to in order to determine a decoding method of the target slice. Slice type designation information (slice_type) for designating a slice type is an example of an encoding parameter included in the slice header SH. Further, the slice header SH includes a filter parameter FP that is referred to by a loop filter included in the video decoding device 1.
 なお、スライスタイプ指定情報により指定可能なスライスタイプとしては、(1)符号化の際にイントラ予測のみを用いるIスライス、(2)符号化の際に単方向予測、又は、イントラ予測を用いるPスライス、(3)符号化の際に単方向予測、双方向予測、又は、イントラ予測を用いるBスライスなどが挙げられる。 As slice types that can be specified by the slice type specification information, (1) I slice using only intra prediction at the time of encoding, and (2) P using unidirectional prediction or intra prediction at the time of encoding. Slice, (3) B-slice using unidirectional prediction, bidirectional prediction, or intra prediction at the time of encoding.
 (LCUレイヤ)
 スライスレイヤSに含まれる各LCUレイヤLCUは、処理対象のLCUである対象LCUを復号するために動画像復号装置1が参照するデータの集合である。LCUレイヤLCUは、図2の(c)に示すように、LCUヘッダLCUH、および、LCUを4分木分割することにより得られる複数の符号化単位(CU:Coding Unit)を含んでいる。
(LCU layer)
Each LCU layer LCU included in the slice layer S is a set of data that the video decoding device 1 refers to in order to decode the target LCU that is the processing target LCU. As shown in FIG. 2C, the LCU layer LCU includes an LCU header LCUH and a plurality of coding units (CU: Coding Units) obtained by dividing the LCU by quadtree division.
 各CUのとり得るサイズは、符号化データ#1のシーケンスパラメータセットSPSに含まれる、LCUのサイズおよび階層深度(hierarchial depth)に依存する。例えば、LCUのサイズが128×128画素であって、最大階層深度が5である場合には、当該LCUに含まれるCUは、5種類のサイズ、すなわち、128×128画素、64×64画素、32×32画素、16×16画素、および、8×8画素の何れかをとり得る。また、それ以上分割されないCUをリーフCUと呼ぶことにする。 The size that each CU can take depends on the LCU size and hierarchical depth included in the sequence parameter set SPS of the encoded data # 1. For example, when the size of the LCU is 128 × 128 pixels and the maximum hierarchical depth is 5, the CU included in the LCU has five types of sizes, that is, 128 × 128 pixels, 64 × 64 pixels, Any of 32 × 32 pixels, 16 × 16 pixels, and 8 × 8 pixels can be taken. A CU that is not further divided is called a leaf CU.
 (LCUヘッダ)
 LCUヘッダLCUHには、対象LCUの復号方法を決定するために動画像復号装置1が参照する符号化パラメータが含まれる。具体的には、図2の(c)に示すように、対象LCUの各リーフCUへの分割パターンを指定するCU分割情報SP_CU、および、量子化ステップの大きさを指定する量子化パラメータ差分Δqp(mb_qp_delta)が含まれる。
(LCU header)
The LCU header LCUH includes an encoding parameter referred to by the video decoding device 1 in order to determine a decoding method of the target LCU. Specifically, as shown in FIG. 2 (c), CU partition information SP_CU that specifies a partition pattern of each target LCU into each leaf CU, and a quantization parameter difference Δqp that specifies the size of the quantization step. (Mb_qp_delta) is included.
 CU分割情報SP_CUは、具体的には、対象LCUに含まれる各CU(およびリーフCU)の形状、サイズ、および、対象LCU内での位置を指定する情報である。なお、CU分割情報SP_CUは、リーフCUの形状やサイズを明示的に含んでいる必要は必ずしもない。例えばCU分割情報SP_CUは、LCU全体またはLCUの部分領域を四分割するか否かを示すフラグ(split_coding_unit_flag)の集合であってもよい。その場合、LCUの形状やサイズを併用することで各リーフCUの形状やサイズを特定できる。 CU division information SP_CU is information that specifies the shape and size of each CU (and leaf CU) included in the target LCU, and the position in the target LCU. Note that the CU partition information SP_CU does not necessarily need to explicitly include the shape and size of the leaf CU. For example, the CU partition information SP_CU may be a set of flags (split_coding_unit_flag) indicating whether or not the entire LCU or a partial region of the LCU is divided into four. In that case, the shape and size of each leaf CU can be specified by using the shape and size of the LCU together.
 また、量子化パラメータ差分Δqpは、対象LCUにおける量子化パラメータqpと、そのLCUの直前に符号化されたLCUにおける量子化パラメータqp’との差分qp-qp’である。 Further, the quantization parameter difference Δqp is a difference qp−qp ′ between the quantization parameter qp in the target LCU and the quantization parameter qp ′ in the LCU encoded immediately before the LCU.
 (リーフCU)
 それ以上分割することができないCU(CUのリーフ)は予測単位(PU:Prediction Unit)、および、変換単位(TU:Transform Unit)として取り扱われる。
(Leaf CU)
A CU (CU leaf) that cannot be further divided is handled as a prediction unit (PU: Prediction Unit) and a transform unit (TU: Transform Unit).
 図2の(d)に示すように、リーフCU(図2の(d)においてはCULと表記)は、(1)動画像復号装置1により予測画像を生成する際に参照されるPU情報PUI、および、(2)動画像復号装置1により残差データを復号する際に参照されるTU情報TUIを含んでいる。なお、PU情報PUIにはスキップフラグSKIPが含まれている場合があり、スキップフラグSKIPの値が1の場合、TU情報は省略される。 As shown in (d) of FIG. 2, the leaf CU (denoted as CUL in (d) of FIG. 2) is (1) PU information PUI that is referred to when the moving image decoding apparatus 1 generates a predicted image. And (2) the TU information TUI that is referred to when the moving image decoding apparatus 1 decodes the residual data. The PU information PUI may include a skip flag SKIP. When the value of the skip flag SKIP is 1, the TU information is omitted.
 PU情報PUIは、図2の(d)に示すように、予測タイプ情報PT、および、予測情報PIを含んでいる。予測タイプ情報PTは、対象リーフCU(対象PU)についての予測画像生成方法として、イントラ予測を用いるのか、または、インター予測を用いるのかを指定する情報である。予測情報PIは、予測タイプ情報PTが何れの予測方法を指定するのかに応じて、イントラ予測情報PI_Intra、または、インター予測情報PI_Interより構成される。以下では、イントラ予測が適用されるPUをイントラPUとも呼称し、インター予測が適用されるPUをインターPUとも呼称する。 The PU information PUI includes prediction type information PT and prediction information PI, as shown in FIG. The prediction type information PT is information that specifies whether intra prediction or inter prediction is used as a predicted image generation method for the target leaf CU (target PU). The prediction information PI includes intra prediction information PI_Intra or inter prediction information PI_Inter depending on which prediction method is specified by the prediction type information PT. Hereinafter, a PU to which intra prediction is applied is also referred to as an intra PU, and a PU to which inter prediction is applied is also referred to as an inter PU.
 PU情報PUIには、対象PUに含まれる各パーティションの形状、サイズ、および、対象PU内での位置を指定する情報が含まれる。ここで、パーティションとは、対象リーフCUを構成する1又は複数の重複しない領域のことであり、予測画像の生成は、パーティションを単位として行われる。 The PU information PUI includes information specifying the shape and size of each partition included in the target PU and the position in the target PU. Here, the partition is one or a plurality of non-overlapping areas constituting the target leaf CU, and the generation of the predicted image is performed in units of partitions.
 TU情報TUIは、図2の(d)に示すように、対象リーフCU(対象TU)の各ブロックへの分割パターンを指定するTU分割情報SP_TU、および、量子化予測残差QD1~QDNT(NTは、対象TUに含まれるブロックの総数)を含んでいる。 As shown in FIG. 2D, the TU information TUI includes TU partition information SP_TU that specifies a partition pattern for each block of the target leaf CU (target TU), and quantized prediction residuals QD1 to QDNT (NT Includes the total number of blocks included in the target TU).
 TU分割情報SP_TUは、具体的には、対象TUに含まれる各ブロックの形状、サイズ、および、対象TU内での位置を指定する情報である。各TUは、例えば、64×64画素から2×2画素までのサイズをとり得る。ここで、ブロックとは、対象リーフCUを構成する1又は複数の重複しない領域のことであり、予測残差の符号化・復号は、TUあるいはTUを分割したブロックを単位として行われる。 TU partition information SP_TU is information that specifies the shape and size of each block included in the target TU and the position in the target TU. Each TU can be, for example, a size from 64 × 64 pixels to 2 × 2 pixels. Here, the block is one or a plurality of non-overlapping areas constituting the target leaf CU, and prediction residual encoding / decoding is performed in units of TUs or blocks obtained by dividing TUs.
 各量子化予測残差QDは、動画像符号化装置2が以下の処理1~3を、処理対象のブロックである対象ブロックに施すことによって生成した符号化データである。処理1:符号化対象画像から予測画像を減算した予測残差をDCT変換(Discrete Cosine Transform)する。処理2:処理1にて得られたDCT係数を量子化する。処理3:処理2にて量子化されたDCT係数を可変長符号化する。上述した量子化パラメータqpは、動画像符号化装置2がDCT係数を量子化する際に用いた量子化ステップQPの大きさを表す(QP=2qp/6)。 Each quantized prediction residual QD is encoded data generated by the moving image encoding apparatus 2 performing the following processes 1 to 3 on a target block that is a processing target block. Process 1: DCT transform (Discrete Cosine Transform) is performed on the prediction residual obtained by subtracting the prediction image from the encoding target image. Process 2: The DCT coefficient obtained in Process 1 is quantized. Process 3: The DCT coefficient quantized in Process 2 is variable length encoded. The quantization parameter qp described above represents the magnitude of the quantization step QP used when the moving picture coding apparatus 2 quantizes the DCT coefficient (QP = 2 qp / 6 ).
 (インター予測情報PI_Inter)
 インター予測情報PI_Interは、動画像復号装置1が、インター予測によってインター予測画像を生成する際に参照される符号化パラメータを含んでいる。図2の(e)に示すように、インター予測情報PI_Interは、対象PUの各パーティションへの分割パターンを指定するインターPU分割情報SP_Inter、および、各パーティションについてのインター予測パラメータPP_Inter1~PP_InterNe(Neは、対象PUに含まれるインター予測パーティションの総数)を含んでいる。
(Inter prediction information PI_Inter)
The inter prediction information PI_Inter includes a coding parameter that is referred to when the video decoding device 1 generates an inter prediction image by inter prediction. As shown in FIG. 2 (e), the inter prediction information PI_Inter includes inter PU partition information SP_Inter that specifies a partition pattern of the target PU into each partition, and inter prediction parameters PP_Inter1 to PP_InterNe (Ne for each partition). , The total number of inter prediction partitions included in the target PU).
 インターPU分割情報SP_Interは、具体的には、対象PU(インターPU)に含まれる各インター予測パーティションの形状、サイズ、および、対象PU内での位置を指定する情報である。 Specifically, the inter-PU partition information SP_Inter is information for designating the shape and size of each inter prediction partition included in the target PU (inter PU) and the position in the target PU.
 インターPUは、2N×2N画素、2N×N画素、N×2N画素、およびN×N画素の4つの対称的分割(symmetric splittings)、並びに、2N×nU画素、2N×nD画素、nL×2N画素、およびnR×2N画素の4つの非対称的分割(asymmetric splittings)により、合計8種類のパーティションに分割することが可能である。ここで、Nの具体的な値は、当該PUが属するCUのサイズによって規定され、nU、nD、nL、および、nRの具体的な値は、Nの値に応じて定められる。例えば、128×128画素のインターPUは、128×128画素、128×64画素、64×128画素、64×64画素、128×32画素、128×96画素、32×128画素、および、96×128画素のインター予測パーティションへ分割することが可能である。 The inter PU is composed of four symmetric splittings of 2N × 2N pixels, 2N × N pixels, N × 2N pixels, and N × N pixels, and 2N × nU pixels, 2N × nD pixels, and nL × 2N. It is possible to divide into 8 types of partitions in total by four asymmetric splits of pixels and nR × 2N pixels. Here, the specific value of N is defined by the size of the CU to which the PU belongs, and the specific values of nU, nD, nL, and nR are determined according to the value of N. For example, an inter PU of 128 × 128 pixels is 128 × 128 pixels, 128 × 64 pixels, 64 × 128 pixels, 64 × 64 pixels, 128 × 32 pixels, 128 × 96 pixels, 32 × 128 pixels, and 96 × It is possible to divide into 128-pixel inter prediction partitions.
 (インター予測パラメータ)
 インター予測パラメータPP_Interは、図2の(e)に示すように、参照画像インデックスRIと、推定動きベクトルインデックスPMVIと、動きベクトル残差MVDとを含んでいる。
(Inter prediction parameter)
As shown in FIG. 2E, the inter prediction parameter PP_Inter includes a reference image index RI, an estimated motion vector index PMVI, and a motion vector residual MVD.
 (イントラ予測情報PI_Intra)
 イントラ予測情報PI_Intraは、動画像復号装置1が、イントラ予測によってイントラ予測画像を生成する際に参照される符号化パラメータを含んでいる。図2の(f)に示すように、イントラ予測情報PI_Intraは、対象PU(イントラPU)の各パーティションへの分割パターンを指定するイントラPU分割情報SP_Intra、および、各パーティションについてのイントラ予測パラメータPP_Intra1~PP_IntraNa(Naは、対象PUに含まれるイントラ予測パーティションの総数)を含んでいる。
(Intra prediction information PI_Intra)
The intra prediction information PI_Intra includes an encoding parameter that is referred to when the video decoding device 1 generates an intra predicted image by intra prediction. As shown in FIG. 2 (f), the intra prediction information PI_Intra includes intra PU partition information SP_Intra that specifies a partition pattern of the target PU (intra PU) into each partition, and intra prediction parameters PP_Intra1 to PP_IntraNa (Na is the total number of intra prediction partitions included in the target PU).
 イントラPU分割情報SP_Intraは、具体的には、対象PUに含まれる各イントラ予測パーティションの形状、サイズ、および、対象PU内での位置を指定する情報である。イントラPU分割情報SP_Intraには、対象PUをパーティションに分割するか否かを指定するイントラ分割フラグ(intra_split_flag)を含んでいる。イントラ分割フラグが1であれば、対象PUは、4つのパーティションへと対称的に分割され、イントラ分割フラグが0であれば、対象PUは、分割されることなく、対象PU自身が1つのパーティションとして取り扱われる。したがって、対象PUのサイズを2N×2N画素とすると、イントラ予測パーティションは、2N×2N画素(分割なし)、および、N×N画素(4分割)の何れかのサイズを取り得る(ここで、N=2n、nは1以上の任意の整数)。例えば、128×128画素のイントラPUは、128×128画素、および、64×64画素のイントラ予測パーティションへ分割することが可能である。 Specifically, the intra-PU partition information SP_Intra is information that specifies the shape and size of each intra-predicted partition included in the target PU, and the position in the target PU. The intra PU split information SP_Intra includes an intra split flag (intra_split_flag) that specifies whether or not the target PU is split into partitions. If the intra partition flag is 1, the target PU is divided symmetrically into four partitions. If the intra partition flag is 0, the target PU is not divided and the target PU itself is one partition. Are treated as Therefore, if the size of the target PU is 2N × 2N pixels, the intra prediction partition can take any of 2N × 2N pixels (no division) and N × N pixels (four divisions) (where, N = 2 n , n is an arbitrary integer of 1 or more). For example, a 128 × 128 pixel intra PU can be divided into 128 × 128 pixel and 64 × 64 pixel intra prediction partitions.
 (イントラ予測パラメータPP_Intra)
 イントラ予測パラメータPP_Intraは、図2の(f)に示すように、エッジベース予測フラグEF、副方向インデックスAI、推定フラグMPM、および、残余予測モードインデックスRIPMを含んでいる。イントラ予測パラメータPP_Intraは、各パーティションについてのイントラ予測方法(予測モード)を指定するためのパラメータである。
(Intra prediction parameter PP_Intra)
The intra prediction parameter PP_Intra includes an edge-based prediction flag EF, a sub-direction index AI, an estimation flag MPM, and a residual prediction mode index RIPM as shown in (f) of FIG. The intra prediction parameter PP_Intra is a parameter for designating an intra prediction method (prediction mode) for each partition.
 エッジベース予測フラグEFは、処理対象のパーティションである対象パーティションについて、エッジベース予測モードを適用するのか否かを指定するフラグである。以下の説明では、エッジベース予測フラグが1であるときに、対象パーティションについてエッジベース予測モードが適用されるものとする。 The edge-based prediction flag EF is a flag that specifies whether or not the edge-based prediction mode is applied to the target partition that is the processing target partition. In the following description, it is assumed that the edge-based prediction mode is applied to the target partition when the edge-based prediction flag is 1.
 補正方向インデックス(追加インデックスとも呼ぶ)AIは、対象パーティションについてエッジベース予測モードが適用される場合に、予測パラメータPP_Intraに含められるインデックスであり、エッジベース予測モードにて導出される主方向に対して補正方向を加算するか否かを指定すると共に、補正方向を加算する場合には、加算する補正方向を指定するためのインデックスである。 The correction direction index (also referred to as an additional index) AI is an index included in the prediction parameter PP_Intra when the edge-based prediction mode is applied to the target partition, and is the main direction derived in the edge-based prediction mode. This is an index for designating whether or not to add the correction direction, and for specifying the correction direction to be added when adding the correction direction.
 ここで、追加インデックスAIによって選択可能な補正方向の総数は、対象パーティションのサイズによって異なる。例えば、対象パーティションのサイズが4×4画素または8×8画素である場合、追加インデックスAIによって選択可能な補正方向の総数を8とし、対象パーティションのサイズが16×16画素である場合、追加インデックスAIによって選択可能な補正方向の総数を2とし、対象パーティションのサイズが32×32画素または64×64画素の場合、追加インデックスAIによって選択可能な補正方向の総数を0(すなわち補正方向は用いず常に主方向を用いる)とすることができる。 Here, the total number of correction directions that can be selected by the additional index AI differs depending on the size of the target partition. For example, when the size of the target partition is 4 × 4 pixels or 8 × 8 pixels, the total number of correction directions selectable by the additional index AI is 8, and when the size of the target partition is 16 × 16 pixels, the additional index When the total number of correction directions that can be selected by AI is 2, and the size of the target partition is 32 × 32 pixels or 64 × 64 pixels, the total number of correction directions that can be selected by the additional index AI is 0 (that is, the correction direction is not used). The main direction is always used).
 推定フラグMPMは、対象パーティションについてエッジベース予測モードが適用されない場合に、予測パラメータPP_Intraに含められるフラグであり、対象パーティションの周辺のパーティションに割り付けられた予測モードに基づいて推定された推定予測モードと当該対象パーティションについての予測モードとが同じであるか否かを示すフラグである。 The estimation flag MPM is a flag that is included in the prediction parameter PP_Intra when the edge-based prediction mode is not applied to the target partition, and is an estimated prediction mode that is estimated based on the prediction mode assigned to the surrounding partitions of the target partition. It is a flag indicating whether or not the prediction mode for the target partition is the same.
 残余予測モードインデックスRIPMは、対象パーティションについてエッジベース予測モードが適用されない場合であって、推定された予測モードと対象パーティションについての予測モードとが異なる場合に、当該対象パーティションに割り付けられる予測モードを指定するためのインデックスである。 The residual prediction mode index RIPM specifies the prediction mode assigned to the target partition when the edge-based prediction mode is not applied to the target partition and the estimated prediction mode and the prediction mode for the target partition are different. It is an index to do.
 (動画像復号装置1)
 以下では、本実施形態に係る動画像復号装置(復号装置)1の構成について、図1~図11を参照して説明する。動画像復号装置1は、その一部に、H.264/MPEG-4 AVC規格に採用されている技術、VCEG(Video Coding Expert Group)における共同開発用コーデックであるKTAソフトウェアに採用されている技術、および、その後継コーデックであるTMuC(Test Model under Consideration)ソフトウェアに採用されている技術を用いている動画像復号装置である。
(Moving picture decoding apparatus 1)
Hereinafter, the configuration of the moving picture decoding apparatus (decoding apparatus) 1 according to the present embodiment will be described with reference to FIGS. The moving picture decoding apparatus 1 includes H.264 as a part thereof. Technology adopted in the H.264 / MPEG-4 AVC standard, technology adopted in KTA software which is a codec for joint development in the Video Coding Expert Group (VCEG), and TMuC (Test Model under Consideration) which is the successor codec ) A moving picture decoding apparatus using a technique adopted in software.
 動画像復号装置1は、予測単位毎に予測画像を生成し、生成された予測画像と、符号化データ#1から復号された予測残差とを加算することによって復号画像#2を生成し、生成された復号画像#2を外部に出力する。 The video decoding device 1 generates a prediction image for each prediction unit, generates a decoded image # 2 by adding the generated prediction image and a prediction residual decoded from the encoded data # 1, The generated decoded image # 2 is output to the outside.
 また、予測画像の生成は、符号化データ#1を復号することによって得られる符号化パラメータを参照して行われる。ここで、符号化パラメータとは、予測画像を生成するために参照されるパラメータのことであり、画面間予測において参照される動きベクトルや画面内予測において参照される予測モードなどの予測パラメータに加えて、パーティションのサイズや形状、ブロックのサイズや形状、および、原画像と予測画像との残差データなどが含まれる。以下では、符号化パラメータに含まれる情報のうち、上記残差データを除く全ての情報の集合を、サイド情報と呼ぶ。 Also, the generation of the predicted image is performed with reference to the encoding parameter obtained by decoding the encoded data # 1. Here, the encoding parameter is a parameter referred to in order to generate a prediction image, and in addition to a prediction parameter such as a motion vector referred to in inter-screen prediction and a prediction mode referred to in intra-screen prediction. Partition size and shape, block size and shape, and residual data between the original image and the predicted image. Hereinafter, a set of all information excluding the residual data among the information included in the encoding parameter is referred to as side information.
 また、以下では、上記予測単位が、LCUを構成するパーティションである場合を例に挙げ説明を行うが、本実施形態はこれに限定されるものではなく、予測単位がパーティションよりも大きい単位である場合、および、予測単位がパーティションよりも小さい単位である場合に対しても適用することができる。 Further, in the following description, the case where the prediction unit is a partition constituting the LCU will be described as an example. However, the present embodiment is not limited to this, and the prediction unit is a unit larger than the partition. The present invention can also be applied to the case where the prediction unit is a unit smaller than the partition.
 また、以下では、復号の対象となるフレーム、スライス、LCU、ブロック、および、パーティションをそれぞれ、対象フレーム、対象スライス、対象LCU、対象ブロック、および、対象パーティションと呼ぶことにする。 In the following description, a frame, a slice, an LCU, a block, and a partition to be decoded are referred to as a target frame, a target slice, a target LCU, a target block, and a target partition, respectively.
 なお、LCUのサイズは、例えば64×64画素であり、パーティションのサイズは、例えば、64×64画素、32×32画素、16×16画素、8×8画素や4×4画素などであるが、これらのサイズは本実施形態を限定するものではなく、LCUのサイズおよびパーティションはそれらのサイズ以外のサイズであってもよい。 The LCU size is, for example, 64 × 64 pixels, and the partition size is, for example, 64 × 64 pixels, 32 × 32 pixels, 16 × 16 pixels, 8 × 8 pixels, 4 × 4 pixels, or the like. These sizes do not limit the present embodiment, and the size and partition of the LCU may be other sizes.
 図3は、動画像復号装置1の構成を示すブロック図である。図3に示すように、動画像復号装置1は、可変長符号逆多重化部11、ヘッダ情報復号部12、LCU設定部13、LCU復号部14、および、フレームメモリ15を備えている。 FIG. 3 is a block diagram showing a configuration of the moving picture decoding apparatus 1. As illustrated in FIG. 3, the video decoding device 1 includes a variable length code demultiplexing unit 11, a header information decoding unit 12, an LCU setting unit 13, an LCU decoding unit 14, and a frame memory 15.
 動画像復号装置1に入力された符号化データ#1は可変長符号逆多重化部11へ入力される。可変長符号逆多重化部11は、入力された符号化データ#1を逆多重化することによって、符号化データ#1を、ヘッダ情報に関する符号化データであるヘッダ符号化データ#11aと、スライスに関する符号化データである符号化データ#11bとに分離し、ヘッダ符号化データ#11aをヘッダ情報復号部12に、符号化データ#11bをLCU設定部13にそれぞれ出力する。 The encoded data # 1 input to the video decoding device 1 is input to the variable length code demultiplexing unit 11. The variable-length code demultiplexing unit 11 demultiplexes the input encoded data # 1, thereby converting the encoded data # 1 into header encoded data # 11a that is encoded data related to header information, and a slice. And the encoded data # 11a is output to the header information decoding unit 12 and the encoded data # 11b is output to the LCU setting unit 13, respectively.
 ヘッダ情報復号部12では、ヘッダ符号化データ#11aからヘッダ情報#12を復号する。ここで、ヘッダ情報#12は、(1)対象スライスに属するLCUのサイズ、形状および対象スライス内での位置についての情報、(2)各LCUに属するリーフCUのサイズ、形状および対象LCU内での位置についての情報、並びに、(3)各リーフCUに属するパーティションのサイズ、形状および対象リーフCU内での位置についての情報、を含んでいる。 The header information decoding unit 12 decodes the header information # 12 from the encoded header data # 11a. Here, the header information # 12 is (1) information about the size, shape and position of the LCU belonging to the target slice, and (2) the size, shape and shape of the leaf CU belonging to each LCU. And (3) information about the size and shape of the partition belonging to each leaf CU and the position within the target leaf CU.
 LCU設定部13では、入力されたヘッダ情報#12に基づいて、符号化データ#11bを個々のLCUに対応する符号化データ#13に分離し、LCU復号部14に対して順次出力する。 The LCU setting unit 13 separates the encoded data # 11b into encoded data # 13 corresponding to each LCU based on the input header information # 12 and sequentially outputs the encoded data # 11b to the LCU decoding unit 14.
 LCU復号部14は、入力された個々のLCUに対応する符号化データ#13を順次復号することにより、個々のLCUに対応する復号画像#2を生成し、出力する。また、復号画像#2は、フレームメモリ15に格納される。LCU復号部14の構成については、後述するためここでは説明を省略する。 The LCU decoding unit 14 generates and outputs a decoded image # 2 corresponding to each LCU by sequentially decoding the encoded data # 13 corresponding to each input LCU. The decoded image # 2 is stored in the frame memory 15. The configuration of the LCU decoding unit 14 will be described later and will not be described here.
 フレームメモリ15には、復号画像#2が記録される。フレームメモリ15には、対象LCUを復号する時点において、当該対象LCUよりも先に復号された全てのLCU(例えば、ラスタスキャン順で先行する全てのLCU)に対応する復号画像が記録されている。 The decoded image # 2 is recorded in the frame memory 15. In the frame memory 15, at the time of decoding the target LCU, decoded images corresponding to all the LCUs decoded before the target LCU (for example, all the LCUs preceding in the raster scan order) are recorded. .
 画像内の全てのLCUに対して、LCU復号部14によるLCU単位の復号画像生成処理が終わった時点で、動画像復号装置1に入力された1フレーム分の符号化データに対応する復号画像#2の生成処理が完了する。 When all the LCUs in the image have been processed by the LCU decoding unit 14 to generate the decoded image in units of LCUs, the decoded image # corresponding to the encoded data for one frame input to the video decoding device 1 2 generation processing is completed.
 (LCU復号部14)
 以下では、LCU復号部14について、参照する図面を替えてより具体的に説明する。
(LCU decoding unit 14)
Hereinafter, the LCU decoding unit 14 will be described more specifically with reference to different drawings.
 図4は、LCU復号部14の構成を示すブロック図である。図4に示すように、LCU復号部14は、LCU分割構造設定部141、リーフCUスキャン部142、および、リーフCU復号部143を備えている。 FIG. 4 is a block diagram showing a configuration of the LCU decoding unit 14. As illustrated in FIG. 4, the LCU decoding unit 14 includes an LCU division structure setting unit 141, a leaf CU scanning unit 142, and a leaf CU decoding unit 143.
 LCU分割構造設定部141は、ヘッダ情報#12を参照して、対象LCUのリーフCUへの分割構造を設定する。対象LCUのリーフCUへの分割構造を指定する分割構造情報#141は、対象LCUについての符号化データ#13と共にリーフCUスキャン部142に供給される。 The LCU division structure setting unit 141 sets the division structure of the target LCU into leaf CUs with reference to the header information # 12. The division structure information # 141 that specifies the division structure of the target LCU into leaf CUs is supplied to the leaf CU scanning unit 142 together with the encoded data # 13 about the target LCU.
 図5の(a)は、対象LCUのリーフCUへの分割構造であって、分割構造情報#141によって指定される分割構造の例を示す図である。図5の(a)に示す例においては、対象LCUは、それぞれリーフCUであるCUL1~CUL7へと分割されている。 (A) of FIG. 5 is a diagram illustrating an example of a division structure into which the target LCU is divided into leaf CUs and specified by the division structure information # 141. In the example shown in FIG. 5A, the target LCU is divided into CUL1 to CUL7 that are leaf CUs.
 リーフCUスキャン部142は、対象LCUに含まれるリーフCUを所定の順序(例えば、ラスタスキャン順)でスキャンし、処理対象となるリーフCUである対象リーフCUについての符号化データ#142aおよびCU情報#142bを、リーフCU復号部143に供給する。ここで、CU情報#142bは、対象リーフCUについての(1)サイズ、形状および対象LCUでの位置、(2)対象リーフCUについてのPU情報、並びに、(3)対象リーフCUについてのTU情報、を含んでいる。 The leaf CU scan unit 142 scans the leaf CUs included in the target LCU in a predetermined order (for example, raster scan order), and encodes data # 142a and CU information for the target leaf CU that is the leaf CU to be processed. # 142b is supplied to the leaf CU decoding unit 143. Here, CU information # 142b includes (1) size, shape, and position in the target LCU for the target leaf CU, (2) PU information about the target leaf CU, and (3) TU information about the target leaf CU. , Including.
 図5の(b)は、対象LCUに含まれるリーフCUであるCUL1~CUL7のうち、リーフCUスキャン部142によって、CUL4が対象リーフCUに設定される場合を例示する図である。図5の(b)に示す例においては、リーフCUスキャン部142は、リーフCUであるCUL4についての符号化データ、および、CUL4についてのCU情報を、リーフCU復号部143に供給する。 (B) of FIG. 5 is a diagram illustrating a case where CUL4 is set as the target leaf CU by the leaf CU scanning unit 142 among the CUL1 to CUL7 which are the leaf CUs included in the target LCU. In the example illustrated in FIG. 5B, the leaf CU scanning unit 142 supplies the encoded data for CUL4, which is the leaf CU, and the CU information for CUL4 to the leaf CU decoding unit 143.
 (リーフCU復号部143)
 リーフCU復号部143は、対象リーフCUについての符号化データおよびCU情報、並びに、フレームメモリ15に格納されている復号済みの画素値を参照して、対象リーフCUについての予測画像を生成すると共に、対象リーフCUについての予測残差を復号する。また、リーフCU復号部143は、生成した予測画像と、復号した予測残差とを加算することによって、対象リーフCUについての復号画像を生成する。また、リーフCU復号部143は、対象LCUに属する各リーフCUについての復号画像を統合し、当該対象LCUについての復号画像#2を生成する。
(Leaf CU decoding unit 143)
The leaf CU decoding unit 143 refers to the encoded data and CU information about the target leaf CU, and the decoded pixel values stored in the frame memory 15, and generates a predicted image for the target leaf CU. The prediction residual for the target leaf CU is decoded. In addition, the leaf CU decoding unit 143 generates a decoded image for the target leaf CU by adding the generated predicted image and the decoded prediction residual. In addition, the leaf CU decoding unit 143 integrates the decoded images for each leaf CU belonging to the target LCU, and generates a decoded image # 2 for the target LCU.
 (リーフCU復号部143の構成および処理の流れ)
 以下では、リーフCU復号部143の構成について、図1、および、図6~図11を参照して説明する。図6はリーフCU復号部143の構成を示すブロック図である。図6に示すように、リーフCU復号部143は、PU復号部431、予測画像生成部432、TU復号部433、予測残差復元部434、および、復号画像生成部435を備えている。
(Configuration and processing flow of leaf CU decoding unit 143)
In the following, the configuration of leaf CU decoding section 143 will be described with reference to FIG. 1 and FIGS. FIG. 6 is a block diagram showing a configuration of the leaf CU decoding unit 143. As illustrated in FIG. 6, the leaf CU decoding unit 143 includes a PU decoding unit 431, a predicted image generation unit 432, a TU decoding unit 433, a prediction residual restoration unit 434, and a decoded image generation unit 435.
 (PU復号部431)
 PU復号部431は、対象リーフCUについてのPU情報を復号することによって、対象リーフCUの各パーティションへの分割パターンを決定すると共に、各パーティションについての予測モードを決定し、決定された各予測モード#431を、各パーティションに割り付ける。各パーティションに割り付けられた予測モード#431は、予測画像生成部432に対して供給される。ここで、予測モードとは、画面内予測(イントラ予測)によって予測画像を生成する生成方法を指定するものであり、本実施形態においては、各パーティションについて、基本予測モードセット、または、エッジベース予測モードセットから何れかの予測モードが選択される。
(PU decoding unit 431)
The PU decoding unit 431 determines the division pattern for each partition of the target leaf CU by decoding the PU information about the target leaf CU, determines the prediction mode for each partition, and determines each determined prediction mode # 431 is assigned to each partition. The prediction mode # 431 assigned to each partition is supplied to the predicted image generation unit 432. Here, the prediction mode designates a generation method for generating a prediction image by intra-screen prediction (intra prediction). In this embodiment, for each partition, a basic prediction mode set or edge-based prediction is used. Any prediction mode is selected from the mode set.
 基本予測モードセットとは、(1)対象パーティションの周辺の復号済み画像を、予め定められた予測方向に沿って外挿することによって当該対象パーティションについての予測画像を生成する方向予測モード、および、(2)対象パーティションの周辺の復号済み画素値の平均値をとることによって、当該対象パーティションについての予測画像を生成するDC予測モードより構成される。 The basic prediction mode set is (1) a direction prediction mode for generating a prediction image for the target partition by extrapolating a decoded image around the target partition along a predetermined prediction direction, and (2) A DC prediction mode for generating a prediction image for the target partition by taking an average value of decoded pixel values around the target partition.
 図7は、本実施形態における基本予測モードセットに含まれる予測モードを、予測モードインデックスと共に例示する図である。また、図7においては、各方向予測モードの予測方向が示されている。図7に示すように、本実施形態における基本予測モードセットには、インデックス0、1、3~8によってそれぞれ指定される方向予測モード、および、インデックス2によって指定されるDC予測モードが含まれる。 FIG. 7 is a diagram illustrating prediction modes included in the basic prediction mode set according to the present embodiment, together with a prediction mode index. Moreover, in FIG. 7, the prediction direction of each direction prediction mode is shown. As shown in FIG. 7, the basic prediction mode set in this embodiment includes a direction prediction mode specified by indexes 0, 1, 3 to 8 and a DC prediction mode specified by index 2, respectively.
 なお、図7においては、基本予測モードセットが、互いに異なる8つの方向予測の何れかを指定する予測モードを含んでいる場合を例に挙げたが、本実施形態はこれに限定されるものではない。例えば、基本予測モードセットとして、互いに異なる9以上の方向の何れかを指定する予測モードを含むセットを用いてもよい。そのような例としては、例えば、互いに異なる16方向の何れかを指定する予測モードや、互いに異なる32方向の何れかを指定する予測モードを含むセットが挙げられる。 In FIG. 7, the basic prediction mode set is exemplified as a case where a prediction mode that specifies any of eight different direction predictions is included. However, the present embodiment is not limited to this. Absent. For example, as the basic prediction mode set, a set including a prediction mode that specifies any of nine or more different directions may be used. As such an example, for example, a set including a prediction mode for designating any of 16 different directions and a prediction mode for designating any of 32 different directions can be given.
 一方で、エッジベース予測モードセットとは、(1)対象パーティションの周辺の復号済みの画素値に基づいて、対象パーティションについての主方向を導出し、当該主方向に沿って、当該対象パーティションの周辺の復号済みの画素値を外挿することにより当該対象パーティションにおける予測画像を生成するエッジベース予測モード(以下、主方向エッジベース予測モードとも呼ぶ)、および、(2)上記主方向に対して、方向補正を加算することによって副方向を導出し、当該副方向に沿って、当該対象パーティションの周辺の復号済みの画素値を外挿することにより当該対象パーティションにおける予測画像を生成するエッジベース予測モード(以下、副方向エッジベース予測モードとも呼ぶ)より構成される。 On the other hand, the edge-based prediction mode set is (1) deriving the main direction for the target partition based on the decoded pixel values around the target partition, and surrounding the target partition along the main direction. An edge-based prediction mode (hereinafter also referred to as a main direction edge-based prediction mode) that generates a predicted image in the target partition by extrapolating the decoded pixel values of (2), and (2) Edge-based prediction mode for deriving a sub-direction by adding direction correction and generating a predicted image in the target partition by extrapolating decoded pixel values around the target partition along the sub-direction (Hereinafter also referred to as a sub-direction edge-based prediction mode).
 以下では、図1、および、図8の(a)~(b)を参照して、PU復号部431によるPU情報復号処理の流れについて説明する。図1は、PU復号部431によるPU情報復号処理の流れを示すフローチャートである。また、図8の(a)は、対象パーティションについて取り得るエッジベース予測フラグEFの値、追加インデックスAIの値、推定フラグMPMの値、および、残余予測モードインデックスRIPMの値を示す表である。また、図8の(b)は、復号された追加インデックスAIの各値に対応する予測方向を示す図である。 Hereinafter, the flow of the PU information decoding process performed by the PU decoding unit 431 will be described with reference to FIGS. 1 and 8A to 8B. FIG. 1 is a flowchart showing a flow of PU information decoding processing by the PU decoding unit 431. FIG. 8A is a table showing the value of the edge-based prediction flag EF, the value of the additional index AI, the value of the estimation flag MPM, and the value of the residual prediction mode index RIPM that can be taken for the target partition. FIG. 8B is a diagram illustrating the prediction direction corresponding to each value of the decoded additional index AI.
 (ステップS101)
 まず、PU復号部431は、対象リーフCUについてのPU情報のうち、対象リーフCUをパーティションに分割するか否かを指定するイントラ分割フラグ(intra_split_flag)を参照し、対象リーフCUを複数のパーティションに分割するのか、または、対象リーフCUを1つのパーティションとして取り扱うのかを決定する。具体的には、PU復号部431は、上記イントラ分割フラグが1であれば、対象リーフCUを4つのパーティションへと対称的に分割し、イントラ分割フラグが0であれば、対象リーフCUを分割することなく、1つのパーティションとして取り扱う。
(Step S101)
First, the PU decoding unit 431 refers to an intra partition flag (intra_split_flag) that specifies whether or not to divide the target leaf CU into partitions among the PU information about the target leaf CU, and converts the target leaf CU into a plurality of partitions. It is determined whether to divide or handle the target leaf CU as one partition. Specifically, the PU decoding unit 431 divides the target leaf CU symmetrically into four partitions if the intra split flag is 1, and splits the target leaf CU if the intra split flag is 0. It handles as one partition without doing.
 (ステップS102)
 続いて、PU復号部431は、ループ変数part_idの値を0に初期化し、part_id≦Npart-1を満たすpart_idについて、ループ毎のループ変数part_idの増分値を1とするループ処理を開始する。ここで、ループ変数part_idは、対象リーフCUに含まれるパーティションを互いに識別するためのインデックスであり、Npartは、対象リーフCUに含まれるパーティションの総数である。
(Step S102)
Subsequently, the PU decoding unit 431 initializes the value of the loop variable part_id to 0, and starts a loop process that sets the increment value of the loop variable part_id for each loop to 1 for part_id that satisfies part_id ≦ Npart-1. Here, the loop variable part_id is an index for mutually identifying partitions included in the target leaf CU, and Npart is the total number of partitions included in the target leaf CU.
 (ステップS103)
 続いて、PU復号部431は、ループ変数part_idによって指定される対象パーティションについてのエッジベース予測フラグEFを復号する。
(Step S103)
Subsequently, the PU decoding unit 431 decodes the edge-based prediction flag EF for the target partition specified by the loop variable part_id.
 (ステップS104)
 続いて、PU復号部431は、ステップS103にて復号したエッジベース予測フラグEFが、対象パーティションについてエッジベース予測モードを適用することを示しているか否かを判別する。
(Step S104)
Subsequently, the PU decoding unit 431 determines whether or not the edge-based prediction flag EF decoded in step S103 indicates that the edge-based prediction mode is applied to the target partition.
 (ステップS105)
 対象パーティションについてエッジベース予測モードが適用されない場合(ステップS104でNo)、PU復号部431は、対象パーティションについての推定フラグMPMを復号する。また、推定フラグMPMが0であれば、PU復号部431は、引き続き対象パーティションについての残余予測モードインデックスRIPMを復号する。本実施形態においては、基本予測モードセットに含まれる予測モードの数は9種類であるので、残余予測モードインデックスRIPMは、基本予測モードセットに含まれる予測モードから推定フラグMPMによって指定される予測モードを除いた合計8種類の予測モードの何れかを指定することになる。したがって、本実施形態においては、残余予測モードインデックスRIPMは3ビットのビット列によって表現される。
(Step S105)
When the edge-based prediction mode is not applied to the target partition (No in step S104), the PU decoding unit 431 decodes the estimation flag MPM for the target partition. If the estimation flag MPM is 0, the PU decoding unit 431 continues to decode the residual prediction mode index RIPM for the target partition. In the present embodiment, since the number of prediction modes included in the basic prediction mode set is nine, the residual prediction mode index RIPM is the prediction mode specified by the estimation flag MPM from the prediction mode included in the basic prediction mode set. Any of a total of eight types of prediction modes excluding the above is designated. Therefore, in the present embodiment, the residual prediction mode index RIPM is expressed by a 3-bit bit string.
 (ステップS106)
 続いて、PU復号部431は、対象パーティションについての予測モードを基本予測モードセットから選択し、当該対象パーティションに割り付ける。具体的には、ステップS105にて復号した推定MPMフラグが1であれば、当該対象パーティションの周辺の復号済みパーティションに割り付けられた予測モードを参照して得られる推定予測モードを、当該対象パーティションについての予測モードに設定する。また、ステップS105にて復号した推定フラグMPMが0であれば、PU復号部431は、残余予測モードインデックスRIPMによって指定される予測モードを、当該対象パーティションについての予測モードに設定する。
(Step S106)
Subsequently, the PU decoding unit 431 selects a prediction mode for the target partition from the basic prediction mode set, and assigns the prediction mode to the target partition. Specifically, if the estimated MPM flag decoded in step S105 is 1, the estimated prediction mode obtained by referring to the prediction modes assigned to the decoded partitions around the target partition is set for the target partition. Set to the prediction mode. If the estimation flag MPM decoded in step S105 is 0, the PU decoding unit 431 sets the prediction mode specified by the residual prediction mode index RIPM as the prediction mode for the target partition.
 (ステップS107)
 一方で、対象パーティションについてエッジベース予測モードが適用される場合(ステップS104でYes)、PU復号部431は、対象パーティションのサイズに応じて、当該対象パーティションについて選択可能な補正方向の総数を決定する。より具体的には、PU復号部431は、対象パーティションのサイズと負の相関を有するように、当該対象パーティションについて選択可能な補正方向の総数を決定する。
(Step S107)
On the other hand, when the edge-based prediction mode is applied to the target partition (Yes in step S104), the PU decoding unit 431 determines the total number of correction directions that can be selected for the target partition according to the size of the target partition. . More specifically, the PU decoding unit 431 determines the total number of correction directions that can be selected for the target partition so as to have a negative correlation with the size of the target partition.
 例えば、対象パーティションのサイズが4×4画素や8×8画素である場合、対象パーティションについて選択可能な補正方向の総数を8に設定し、対象パーティションのサイズが16×16画素である場合、対象パーティションについて選択可能な補正方向の総数を2に設定し、対象パーティションのサイズが32×32画素や64×64画素である場合、対象パーティションについて選択可能な補正方向の総数を0とする。より一般には、PU復号部431は、対象パーティションに属する画素の総数がNpix1であるときに、当該対象パーティションについて選択可能な補正方向の総数をNad1に設定し、対象パーティションに属する画素の総数がNpix2(Npix1<Npix2)であるときに、当該対象パーティションについて選択可能な補正方向の総数をNad2(Nad1≧Nad2)に設定する。 For example, when the size of the target partition is 4 × 4 pixels or 8 × 8 pixels, the total number of correction directions that can be selected for the target partition is set to 8, and when the size of the target partition is 16 × 16 pixels, the target When the total number of correction directions that can be selected for a partition is set to 2 and the size of the target partition is 32 × 32 pixels or 64 × 64 pixels, the total number of correction directions that can be selected for the target partition is set to zero. More generally, when the total number of pixels belonging to the target partition is Npix1, the PU decoding unit 431 sets the total number of correction directions that can be selected for the target partition to Nad1, and the total number of pixels belonging to the target partition is Npix2. When (Npix1 <Npix2), the total number of correction directions that can be selected for the target partition is set to Nad2 (Nad1 ≧ Nad2).
 なお、PU復号部431は、符号化データ#1に含まれる分割情報であって、対象パーティションを含む対象LCUの各リーフCUへの分割パターンを指定するCU分割情報SP_CU、および、対象PU(対象リーフCU)の各パーティションへの分割パターンを指定するイントラPU分割情報SP_Intraを参照することによって、対象パーティションのサイズを識別することができる。 The PU decoding unit 431 is the division information included in the encoded data # 1, and includes CU division information SP_CU that specifies a division pattern for each leaf CU of the target LCU including the target partition, and the target PU (target The size of the target partition can be identified by referring to the intra PU partition information SP_Intra that specifies the partition pattern of each leaf CU).
 (ステップS108)
 続いて、PU復号部431は、対象パーティションについての追加インデックスAIを復号する。ここで、追加インデックスAIの最上位ビットは、図8の(a)に示すように主方向をそのまま用いるのか、副方向を用いるのかを示すフラグとして機能する。また、追加インデックスAIの最上位以外のビット列は、選択可能な補正方向から何れかの補正方向を指定するための情報として機能する。
(Step S108)
Subsequently, the PU decoding unit 431 decodes the additional index AI for the target partition. Here, the most significant bit of the additional index AI functions as a flag indicating whether the main direction is used as it is or as shown in FIG. 8A. Further, the bit string other than the most significant bit of the additional index AI functions as information for designating any correction direction from selectable correction directions.
 (ステップS109)
 続いて、PU復号部431は、ステップS108にて復号された追加インデックスAIの最上位ビットを参照し、主方向をそのまま用いる(最上位ビット=1)のか、副方向を用いる(最上位ビット=0)のかを判別する。
(Step S109)
Subsequently, the PU decoding unit 431 refers to the most significant bit of the additional index AI decoded in step S108 and uses the main direction as it is (most significant bit = 1) or uses the sub direction (most significant bit = 0).
 (ステップS110)
 追加インデックスAIの最上位ビットが主方向をそのまま用いることを示している場合(ステップS109でYES)には、PU復号部431は、対象パーティションについての予測モードを、主方向エッジベース予測モードに設定する。
(Step S110)
When the most significant bit of the additional index AI indicates that the main direction is used as it is (YES in step S109), the PU decoding unit 431 sets the prediction mode for the target partition to the main direction edge-based prediction mode. To do.
 (ステップS111)
 追加インデックスAIの最上位ビットが副方向を用いることを示している場合(ステップS109でNO)には、PU復号部431は、追加インデックスAIの最上位ビット以外のビット列(2進表現)を復号し、当該ビット列によって指定される副方向インデックスk(10進表現)を決定する。また、PU復号部431は、対象パーティションについての予測モードを、副方向インデックスkによって指定される副方向エッジベース予測モードに設定する。
(Step S111)
When the most significant bit of the additional index AI indicates that the sub-direction is used (NO in step S109), the PU decoding unit 431 decodes a bit string (binary representation) other than the most significant bit of the additional index AI. Then, the sub-direction index k (decimal expression) designated by the bit string is determined. Also, the PU decoding unit 431 sets the prediction mode for the target partition to the sub-direction edge-based prediction mode specified by the sub-direction index k.
 ここで、対象パーティションのサイズが16×16画素である場合には、選択可能な補正方向の総数は2であるので、副方向インデックスkのとり得る値は、例えば、-1、および+1の何れかである。ここで、k=-1は、追加インデックスAIの最上位ビット以外のビット列が0である場合に対応し、k=+1は、追加インデックスAIの最上位ビット以外のビット列が1である場合に対応する。 Here, when the size of the target partition is 16 × 16 pixels, the total number of correction directions that can be selected is 2. Therefore, the possible values of the sub-direction index k are, for example, any one of −1 and +1 It is. Here, k = −1 corresponds to the case where the bit string other than the most significant bit of the additional index AI is 0, and k = + 1 corresponds to the case where the bit string other than the most significant bit of the additional index AI is 1. To do.
 また、対象パーティションのサイズが4×4画素または8×8画素である場合には、選択可能な補正方向の総数は8であるので、副方向インデックスkのとり得る値は、例えば、-4、-3、-2、-1、+1、+2、+3、および+4の何れかである。ここで、k=-4、k=-3、k=-2、k=-1は、追加インデックスAIの最上位ビット以外のビット列が1110、1100、100、00である場合にそれぞれ対応し、k=+4、k=+3、k=+2、k=+1は、追加インデックスAIの最上位ビット以外のビット列が1111、1101、101、01である場合にそれぞれ対応する。 In addition, when the size of the target partition is 4 × 4 pixels or 8 × 8 pixels, the total number of correction directions that can be selected is 8, and thus the possible values of the sub-direction index k are, for example, −4, -3, -2, -1, +1, +2, +3, and +4. Here, k = −4, k = −3, k = −2, and k = −1 correspond to cases where the bit string other than the most significant bit of the additional index AI is 1110, 1100, 100, 00, respectively. k = + 4, k = + 3, k = + 2, and k = + 1 correspond to cases where the bit strings other than the most significant bit of the additional index AI are 1111, 1101, 101, and 01, respectively.
 また、対象パーティションのサイズが32×32画素または64×64画素である場合には、選択可能な補正方向の総数は0であるので、追加インデックスAIの復号は行わない。 In addition, when the size of the target partition is 32 × 32 pixels or 64 × 64 pixels, the total number of correction directions that can be selected is 0, and the additional index AI is not decoded.
 一般には、補正方向の総数が2Kである場合、副方向インデックスkに対応する追加インデックスAIの最上位以外のビット列は、abs(k)に対応するビット列とsign(k)に対応するビット列により構成できる。上記の場合、abs(k)に対応するビット列は、k<Kの場合はk-1個の連続する1の後に0を付加したビット列であり、k=Kの場合はk-1個の連続する1によるビット列である。また、sign(k)に対応するビット列は、kが負値の場合に0であり、kが正値の場合に1である。 In general, when the total number of correction directions is 2K, the bit string other than the most significant bit of the additional index AI corresponding to the sub-direction index k is composed of a bit string corresponding to abs (k) and a bit string corresponding to sign (k). it can. In the above case, the bit string corresponding to abs (k) is a bit string in which 0 is added after k−1 consecutive 1s when k <K, and k−1 consecutives when k = K. Is a bit string according to 1. The bit string corresponding to sign (k) is 0 when k is a negative value and 1 when k is a positive value.
 図8の(b)は、対象パーティションのサイズが4×4画素である場合に、PU復号部431によって設定される主方向、および、副方向の例を示す図である。図8の(b)に示すように、主方向(主方向ベクトル)が角度θmによって表されるとすると、副方向(副方向ベクトル)を指定する角度θskは、副方向インデックスk、および、補正角度を量子化する際の量子化の粗さを指定する量子化ステップサイズを表すパラメータδを用いて、
 θsk=θm+k×δ   …(1-1)
と表される。なお、主方向および副方向は、何れも、水平右向きを0度とし、時計回りを正として表されるものとする(以下の角度の表現についても同様)。また、パラメータδは、予測画像生成部432によって、後述する方法に従って導出される。
FIG. 8B is a diagram illustrating an example of the main direction and the sub-direction set by the PU decoding unit 431 when the size of the target partition is 4 × 4 pixels. As shown in FIG. 8B, when the main direction (main direction vector) is represented by an angle θm, the angle θsk that designates the sub direction (sub direction vector) includes the sub direction index k and the correction. Using the parameter δ representing the quantization step size that specifies the roughness of the quantization when the angle is quantized,
θsk = θm + k × δ (1-1)
It is expressed. Note that the main direction and the sub-direction are both expressed with the horizontal right direction being 0 degrees and the clockwise direction being positive (the same applies to the expression of the following angles). The parameter δ is derived by the predicted image generation unit 432 according to a method described later.
 (ステップS112)
 本ステップは、ループの終端である。
(Step S112)
This step is the end of the loop.
 以上が、PU復号部431によるPU情報復号処理の流れである。ステップS106、ステップS110、および、ステップS111にて対象リーフCUに含まれる各パーティションについて設定された予測モードは、予測モード#431として予測画像生成部432に供給される。 The above is the flow of PU information decoding processing by the PU decoding unit 431. The prediction mode set for each partition included in the target leaf CU in step S106, step S110, and step S111 is supplied to the predicted image generation unit 432 as a prediction mode # 431.
 (予測画像生成部432)
 予測画像生成部432は、対象リーフCUに含まれる各パーティションについての予測画像を、PU復号部431より供給される予測モード#431、および、該パーティションの周辺の復号済み画素値を参照して、生成する。
(Predicted image generation unit 432)
The prediction image generation unit 432 refers to the prediction image for each partition included in the target leaf CU with reference to the prediction mode # 431 supplied from the PU decoding unit 431 and the decoded pixel values around the partition. Generate.
 以下では、図9~図11を参照して、予測画像生成部432による、対象リーフCUについての予測画像生成処理について説明する。 Hereinafter, a predicted image generation process for the target leaf CU by the predicted image generation unit 432 will be described with reference to FIGS.
 図9は、予測画像生成部432による、対象リーフCUについての予測画像生成処理の流れを示すフローチャートである。 FIG. 9 is a flowchart showing the flow of predicted image generation processing for the target leaf CU by the predicted image generation unit 432.
 (ステップS201)
 まず、予測画像生成部432は、ループ変数part_idの値を0に初期化し、part_id≦Npart-1を満たすpart_idについて、ループ毎のループ変数part_idの増分値を1とするループ処理を開始する。ここで、ループ変数part_idは、上述したものと同様のループ変数である。すなわち、ループ変数part_idは、対象リーフCUに含まれるパーティションを互いに識別するためのインデックスであり、Npartは、対象リーフCUに含まれるパーティションの総数である。
(Step S201)
First, the predicted image generation unit 432 initializes the value of the loop variable part_id to 0, and starts loop processing that sets the increment value of the loop variable part_id for each loop to 1 for part_id that satisfies part_id ≦ Npart-1. Here, the loop variable part_id is the same loop variable as described above. That is, the loop variable part_id is an index for mutually identifying partitions included in the target leaf CU, and Npart is the total number of partitions included in the target leaf CU.
 (ステップS202)
 続いて、予測画像生成部432は、PU復号部431より供給される予測モード#431を参照して、ループ変数part_idによって指定される対象パーティションについての予測モードを参照し、対象パーティションについてエッジベース予測モードが適用されているか否かを判別する。
(Step S202)
Subsequently, the predicted image generation unit 432 refers to the prediction mode # 431 supplied from the PU decoding unit 431, refers to the prediction mode for the target partition specified by the loop variable part_id, and performs edge-based prediction for the target partition. Determine whether the mode is applied.
 (ステップS203)
 対象パーティションについてエッジベース予測モードが適用されていない場合(ステップS202でNo)、予測画像生成部432は、対象パーティションについての予測モードが方向予測モード、すなわち、予め定められた予測方向を用いる予測モードであるか否かを判別する。
(Step S203)
When the edge-based prediction mode is not applied to the target partition (No in step S202), the predicted image generation unit 432 uses the direction prediction mode as the prediction mode for the target partition, that is, a prediction mode that uses a predetermined prediction direction. It is determined whether or not.
 (ステップS204)
 対象パーティションについての予測モードが方向予測モードではない場合(ステップS203でNo)、予測画像生成部432は、対象パーティションについての予測画像を、当該対象パーティションの周辺の復号済み画素値の平均をとることによって生成する。
(Step S204)
When the prediction mode for the target partition is not the direction prediction mode (No in step S203), the predicted image generation unit 432 averages the decoded pixel values around the target partition for the predicted image for the target partition. Generate by.
 図10は、4×4画素である対象パーティションの各画素(予測対象画素)と、当該対象パーティションの周辺の画素(参照画素)とを示す図である。図10に示すように、予測対象画素には符号a~p、参照画素には符号A~Mを付し、画素X(Xはa~p、A~Mの何れか)の画素値をXと表すことにする。また、参照画素A~Mは、何れも復号済みであるとする。図10に示す例においては、予測画像生成部432は、画素値a~pを、以下の式
 a~p=ave(A,B,C,D,I,J,K,L)
によって生成する。ここで、ave(…)は、括弧内に含まれる要素の平均をとることを示している。
FIG. 10 is a diagram illustrating each pixel (prediction target pixel) of the target partition, which is 4 × 4 pixels, and pixels (reference pixels) around the target partition. As shown in FIG. 10, the prediction target pixels are denoted by symbols a to p, the reference pixels are denoted by symbols A to M, and the pixel value of the pixel X (X is any of a to p or A to M) It will be expressed as Further, it is assumed that the reference pixels A to M have all been decoded. In the example illustrated in FIG. 10, the predicted image generation unit 432 uses pixel values a to p as the following formulas a to p = ave (A, B, C, D, I, J, K, L).
Generate by. Here, ave (...) Indicates that an element included in parentheses is averaged.
 (ステップS205)
 対象パーティションについての予測モードが方向予測モードである場合(ステップS203でYes)、予測画像生成部432は、当該方向予測モードに対応する予測方向を、当該対象パーティションについての予測方向に設定する。
(Step S205)
When the prediction mode for the target partition is the direction prediction mode (Yes in step S203), the predicted image generation unit 432 sets the prediction direction corresponding to the direction prediction mode as the prediction direction for the target partition.
 (ステップS206)
 一方で、対象パーティションについての予測モードがエッジベース予測モードである場合(ステップS202でYes)、予測画像生成部432は、対象パーティションの周辺の復号済みの画素値(参照画素の画素値)に基づいて、対象パーティションに割り付ける主方向を導出する。
(Step S206)
On the other hand, when the prediction mode for the target partition is the edge-based prediction mode (Yes in step S202), the predicted image generation unit 432 is based on decoded pixel values (pixel values of reference pixels) around the target partition. Thus, the main direction assigned to the target partition is derived.
 本ステップにおける主方向の導出処理について、図11を参照して説明すれば以下の通りである。本ステップにおいては、予測画像生成部432は、例えば、以下のサブステップS206-1~サブステップS206-4の処理を行う。 The main direction derivation process in this step will be described with reference to FIG. In this step, the predicted image generation unit 432 performs, for example, the following sub-step S206-1 to sub-step S206-4.
 図11は、対象パーティションOPを、対象パーティションOPに隣接する隣接パーティションNP2およびNP3、並びに、対象パーティションの左上の頂点を共有するパーティションNP1と共に示す図である。ここで、図11に示す隣接パーティションNP1~3に含まれる画素の画素値は、何れも復号済みであるものとする。図11に示すように、対象パーティションOPの周辺のパーティションNP1~NP3に含まれる画素を参照画素とすることができる。なお、図11においては、対象パーティションOP、および、隣接パーティションNP1~3が何れも4×4画素である場合を示すが、それらのパーティションのサイズが8×8画素である場合や、その他のサイズである場合についても同様である。 FIG. 11 is a diagram showing the target partition OP together with adjacent partitions NP2 and NP3 adjacent to the target partition OP, and a partition NP1 sharing the top left vertex of the target partition. Here, it is assumed that the pixel values of the pixels included in the adjacent partitions NP1 to NP3 shown in FIG. 11 have been decoded. As shown in FIG. 11, pixels included in partitions NP1 to NP3 around the target partition OP can be used as reference pixels. FIG. 11 shows the case where the target partition OP and the adjacent partitions NP1 to NP3 are all 4 × 4 pixels, but the size of these partitions is 8 × 8 pixels or other sizes. The same applies to the case of.
 (サブステップS206-1)
 予測画像生成部432は、ループ変数iの値を1に初期化し、i≦Mを満たすjについて、ループ毎のループ変数iの増分値を1とするループ処理を開始する。ここで、Mは、対象パーティションについての主方向を導出するために参照される参照画素の数である。
(Substep S206-1)
The predicted image generation unit 432 initializes the value of the loop variable i to 1, and starts a loop process that sets the increment value of the loop variable i for each loop to 1 for j that satisfies i ≦ M. Here, M is the number of reference pixels that are referred to in order to derive the main direction for the target partition.
 (サブステップS206-2)
 続いて、予測画像生成部432は、i番目の参照画素について、エッジベクトルbiを算出する。ここで、エッジベクトルbiの算出には、ソーベル演算子(Sobel operators、ソーベルフィルタとも呼ぶ)Gx、および、Gyを用いればよい。ここでソーベルフィルタGx、および、Gyは、それぞれ、x方向に沿った画像勾配、および、y方向に沿った画像勾配を算出するために用いられるフィルタ行列であり、例えば、3×3行列として、
(Substep S206-2)
Subsequently, the prediction image generation unit 432, the i-th reference pixel, and calculates the edge vectors b i. Here, the calculation of the edge vectors b i is Sobel operator (Sobel operators, also referred to as a Sobel filter) Gx, and may be used to Gy. Here, the Sobel filters Gx and Gy are filter matrices used for calculating the image gradient along the x direction and the image gradient along the y direction, respectively. For example, as a 3 × 3 matrix, ,
Figure JPOXMLDOC01-appb-M000001
Figure JPOXMLDOC01-appb-M000001
によって与えられる。予測画像生成部432は、算出されたx方向についての画像勾配、および、y方向に沿った画像勾配によって表される画像勾配に直交するエッジベクトルbiを算出する。 Given by. Prediction image generating unit 432, image gradient for the calculated x-direction, and calculates an edge vector b i that is perpendicular to the image gradients represented by the image gradient along the y-direction.
 (サブステップS206-3)
 本サブステップは、ループ変数をiとするループの終端である。
(Substep S206-3)
This substep is the end of the loop with the loop variable i.
 (サブステップS206-4)
 続いて、予測画像生成部432は、以下に示す関数T(α)
 T(α)=Σ<e,bi2
を定義する。ここで、eは、自身の方向と水平方向(x方向)とのなす角がαである単位ベクトルを表しており、記号<,>は、両者のベクトルの内積を表している。また、記号Σは、添え字iについて1からMまでの和をとることを示している。
(Substep S206-4)
Subsequently, the predicted image generation unit 432 has a function T (α) shown below.
T (α) = Σ <e, b i > 2
Define Here, e represents a unit vector whose angle between its own direction and the horizontal direction (x direction) is α, and the symbol <,> represents the inner product of both vectors. The symbol Σ indicates that the subscript i is to be summed from 1 to M.
 また、予測画像生成部432は、関数T(α)を最大化する引数θm
 θm=argmaxS(α)
を算出し、θmによって表される方向を、対象パーティションについての主方向ベクトルに設定する。なお、角度θmは、水平右向きを0度とし、時計回りを正として表されるものとする。
The predicted image generation unit 432 also uses the argument θm to maximize the function T (α)
θm = argmaxS (α)
And the direction represented by θm is set as the main direction vector for the target partition. Note that the angle θm is represented with the horizontal right direction being 0 degrees and the clockwise direction being positive.
 なお、上記の説明では、予測画像生成部432が主方向を算出する際に、対象パーティションの上辺に隣接するパーティション、対象パーティションの左辺に隣接するパーティション、および、対象パーティションの左上の頂点を共有するパーティションに属する画素の画素値を参照する場合を例に挙げたが、本実施形態はこれに限定されるものではなく、予測画像生成部432は、より一般に、対象パーティションの周辺に設定される参照領域に属する復号済みの画素値を参照して、当該対象パーティションについての主方向を算出する構成とすることができる。 In the above description, when the predicted image generation unit 432 calculates the main direction, the partition adjacent to the upper side of the target partition, the partition adjacent to the left side of the target partition, and the upper left vertex of the target partition are shared. The case where the pixel values of the pixels belonging to the partition are referred to has been taken as an example, but the present embodiment is not limited to this, and the predicted image generation unit 432 is more generally a reference set around the target partition. With reference to the decoded pixel values belonging to the region, the main direction for the target partition can be calculated.
 また、予測画像生成部432による主方向の導出方法は、上記の例に限定されるものではない。例えば、予測画像生成部432は、参照画素について算出したエッジベクトルbiのうち、大きさが最大のエッジベクトル、すなわち、ノルムが最大のエッジベクトルの示す方向を、主方向として算出する構成としてもよい。 In addition, the method of deriving the main direction by the predicted image generation unit 432 is not limited to the above example. For example, the predicted image generation unit 432 may be configured to calculate, as the main direction, the edge vector having the largest magnitude among the edge vectors b i calculated for the reference pixel, that is, the direction indicated by the edge vector having the largest norm. Good.
 (ステップS207)
 続いて、予測画像生成部432は、対象パーティションについての予測モードが、主方向予測モードであるか否かを判別する。
(Step S207)
Subsequently, the predicted image generation unit 432 determines whether or not the prediction mode for the target partition is the main direction prediction mode.
 (ステップS208)
 対象パーティションについての予測モードが、主方向予測モードである場合(ステップS207でYes)、予測画像生成部432は、ステップS206にて算出された主方向を、対象パーティションについての予測方向に設定する。
(Step S208)
When the prediction mode for the target partition is the main direction prediction mode (Yes in step S207), the predicted image generation unit 432 sets the main direction calculated in step S206 as the prediction direction for the target partition.
 (ステップS209)
 一方で、対象パーティションについての予測モードが、主方向予測モードでない場合(ステップS207でNo)、予測画像生成部432は、補正角度を量子化する際の量子化の粗さを指定する量子化ステップサイズを表すパラメータδを算出する。本実施形態においては、予測画像生成部432は、以下に示す確信度(confidence measure)s
(Step S209)
On the other hand, when the prediction mode for the target partition is not the main direction prediction mode (No in step S207), the predicted image generation unit 432 performs a quantization step for specifying the roughness of quantization when the correction angle is quantized. A parameter δ representing the size is calculated. In the present embodiment, the predicted image generation unit 432 has confidence measures s shown below.
Figure JPOXMLDOC01-appb-M000002
Figure JPOXMLDOC01-appb-M000002
の関数としてパラメータδを算出する。なお、予測画像生成部432により導出されたパラメータδをは、確信度sの減少関数であるという性質がある。また、予測画像生成部432は、パラメータδを、パーティション毎に個別に算出する構成に代えて、例えば、予め定められたパラメータδを用いる構成としてもよい。また、符号化データ#1を生成する動画像符号化装置において、スライス毎やLCU毎にパラメータδを符号化し、当該パラメータδを符号化データ#1に含めておき、予測画像生成部432が、符号化データ#1から復号されたパラメータδを用いる構成としてもよい。 The parameter δ is calculated as a function of Note that the parameter δ derived by the predicted image generation unit 432 has a property of being a decreasing function of the certainty factor s. In addition, the predicted image generation unit 432 may be configured to use a predetermined parameter δ, for example, instead of the configuration that calculates the parameter δ individually for each partition. Further, in the video encoding apparatus that generates the encoded data # 1, the parameter δ is encoded for each slice or LCU, the parameter δ is included in the encoded data # 1, and the predicted image generation unit 432 A configuration using the parameter δ decoded from the encoded data # 1 may be adopted.
 続いて、予測画像生成部432は、追加インデックスAIを復号することによって得られた副方向インデックスkを用いて、数式(1-1)により対象パーティションについての副方向を示す角度θskを算出する。 Subsequently, the predicted image generation unit 432 calculates an angle θsk indicating the sub-direction for the target partition using Equation (1-1) using the sub-direction index k obtained by decoding the additional index AI.
 (ステップS210)
 続いて、予測画像生成部432は、ステップS209にて算出した角度θskの示す方向を、対象パーティションについての予測方向に設定する。
(Step S210)
Subsequently, the predicted image generation unit 432 sets the direction indicated by the angle θsk calculated in step S209 as the predicted direction for the target partition.
 (ステップS211)
 最後に、予測画像生成部432は、ステップS205、ステップS208、および、ステップS210にて設定された対象パーティションについての予測方向に沿って、対象パーティションの周辺の復号済み画素値を外挿することによって、当該対象パーティションについての予測画像を生成する。
(Step S211)
Finally, the predicted image generation unit 432 extrapolates the decoded pixel values around the target partition along the prediction direction for the target partition set in step S205, step S208, and step S210. Then, a predicted image for the target partition is generated.
 具体的には、予測画像生成部432は、当該対象パーティションにおける予測対象画素の画素位置を始点とする、上記予測方向の逆方向を向いた仮想線分上に位置する復号済み画素のうち、当該画素に最も近い画素(以下、最近接画素とも呼ぶ)の画素値を、当該予測対象画素の画素値に設定することにより、当該対象パーティションにおける予測画像を生成する。また、最近接画素の画素値、および、最近接画素の周辺の画素の画素値を用いて算出される値を、当該予測対象画素の画素値としてもよい。 Specifically, the predicted image generation unit 432 uses the pixel position of the prediction target pixel in the target partition as a starting point, and among the decoded pixels located on the virtual line segment facing the reverse direction of the prediction direction, By setting the pixel value of the pixel closest to the pixel (hereinafter also referred to as the closest pixel) to the pixel value of the prediction target pixel, a predicted image in the target partition is generated. Further, the pixel value of the prediction target pixel may be a value calculated using the pixel value of the nearest pixel and the pixel values of the pixels around the nearest pixel.
 例えば、対象パーティションについての方向予測モードが、図7に示す予測モードインデックス0によって指定されるものである場合には、予測画像生成部432は、図10に示す例においては、画素値a~pを、以下の式
 a,e,i,m=A,
 b,f,j,n=B,
 c,g,k,o=C,
 d,h,l,p=D
によって生成する。
For example, when the direction prediction mode for the target partition is designated by the prediction mode index 0 shown in FIG. 7, the prediction image generation unit 432 uses the pixel values a to p in the example shown in FIG. With the following equations a, e, i, m = A,
b, f, j, n = B,
c, g, k, o = C,
d, h, l, p = D
Generate by.
 また、例えば、対象パーティションについての方向予測モードが、図7に示す予測モードインデックス4によって指定されるものである場合には、予測画像生成部432は、図10に示す例においては、画素値a~pを以下の式
 d=(B+(C×2)+D+2)>>2,
 c,h=(A+(B×2)+C+2)>>2,
 b,g,l=(M+(A×2)+B+2)>>2,
 a,f,k,p=(I+(M×2)+A+2)>>2,
 e,j,o=(J+(I×2)+M+2)>>2,
 i,n=(K+(J×2)+I+2)>>2,
 m=(L+(K×2)+J+2)>>2
によって生成する。ここで、「>>」は右シフト演算を表し、任意の正の整数x、sに対し、x>>sの値は、x÷(2^s)の少数部分を切り捨てた値と等しい。
For example, when the direction prediction mode for the target partition is specified by the prediction mode index 4 illustrated in FIG. 7, the predicted image generation unit 432 includes the pixel value a in the example illustrated in FIG. 10. ˜p is expressed by the following formula d = (B + (C × 2) + D + 2) >> 2,
c, h = (A + (B × 2) + C + 2) >> 2,
b, g, l = (M + (A × 2) + B + 2) >> 2,
a, f, k, p = (I + (M × 2) + A + 2) >> 2,
e, j, o = (J + (I × 2) + M + 2) >> 2,
i, n = (K + (J × 2) + I + 2) >> 2,
m = (L + (K × 2) + J + 2) >> 2
Generate by. Here, “>>” represents a right shift operation, and for any positive integer x, s, the value of x >> s is equal to the value obtained by rounding down the decimal part of x ÷ (2 ^ s).
 また、予測画像生成部432は、上記の予測モード以外の基本予測モードに対しても、同様の方法によって画素値a~pを算出することができる。また、予測画像生成部432は、エッジベース予測モードに対しても、同様の方法によって画素値a~pを算出することができる。 Also, the predicted image generation unit 432 can calculate the pixel values a to p by the same method for the basic prediction modes other than the above prediction modes. Also, the predicted image generation unit 432 can calculate the pixel values a to p by the same method for the edge-based prediction mode.
 (ステップS212)
 本ステップは、ループ変数をid_partとするループの終端である。
(Step S212)
This step is the end of the loop whose loop variable is id_part.
 予測画像生成部432は、以上の処理を行うことによって対象リーフCUについての予測画像を生成する。また、生成された対象リーフCUについての予測画像は、予測画像#432として復号画像生成部435に供給される。 The predicted image generation unit 432 generates a predicted image for the target leaf CU by performing the above processing. Further, the generated predicted image for the target leaf CU is supplied to the decoded image generation unit 435 as a predicted image # 432.
 (TU復号部433)
 続いて、リーフCU復号部143の備えるTU復号部433について説明する。TU復号部433は、対象リーフCUについてのTU情報を復号する。また、TU復号部433は、対象リーフCUについてのTU情報に含まれるTU分割情報SP_TUを参照して、対象リーフCUを1または複数のブロックへと分割する。また、各ブロックについての量子化予測残差を、対象リーフCUについての符号化データ#142aより復号し、復号された量子化予測残差を、予測残差復元部434に供給する。
(TU decoding unit 433)
Next, the TU decoding unit 433 included in the leaf CU decoding unit 143 will be described. The TU decoding unit 433 decodes TU information for the target leaf CU. Also, the TU decoding unit 433 divides the target leaf CU into one or a plurality of blocks with reference to the TU partition information SP_TU included in the TU information for the target leaf CU. Also, the quantization prediction residual for each block is decoded from the encoded data # 142a for the target leaf CU, and the decoded quantization prediction residual is supplied to the prediction residual restoration unit 434.
 (予測残差復元部434)
 予測残差復元部434は、TU復号部433より供給される量子化予測残差を、逆量子化および逆DCT変換(Inverse Discrete Cosine Transform)することによって、各パーティションについての、画素毎の予測残差を復元する。復元された予測残差#434は、復号画像生成部435に供給される。
(Prediction residual restoration unit 434)
The prediction residual restoration unit 434 performs inverse quantization and inverse DCT transform (Inverse Discrete Cosine Transform) on the quantized prediction residual supplied from the TU decoding unit 433, so that the prediction residual for each partition is obtained. Restore the difference. The restored prediction residual # 434 is supplied to the decoded image generation unit 435.
 (復号画像生成部435)
 復号画像生成部435は、予測画像生成部432より供給される予測画像#432と、予測残差復元部434より供給される予測残差とを加算することによって、対象リーフCUについての復号画像を生成する。
(Decoded image generation unit 435)
The decoded image generation unit 435 adds the prediction image # 432 supplied from the prediction image generation unit 432 and the prediction residual supplied from the prediction residual restoration unit 434, thereby obtaining a decoded image for the target leaf CU. Generate.
 また、復号画像生成部435は、対象LCUに含まれる各リーフCUについての復号画像を統合することによって、対象LCUについての復号画像#2を生成する。 Also, the decoded image generation unit 435 generates a decoded image # 2 for the target LCU by integrating the decoded images for each leaf CU included in the target LCU.
 (リーフCU復号部143による復号処理の全体の流れ)
 以下では、リーフCU復号部143による復号処理の全体の流れについて、図12の(a)を参照して説明する。図12の(a)は、リーフCU復号部143による復号処理の全体の流れを示すフローチャートである。
(Overall Flow of Decoding Process by Leaf CU Decoding Unit 143)
Hereinafter, an overall flow of the decoding process performed by the leaf CU decoding unit 143 will be described with reference to FIG. FIG. 12A is a flowchart showing an overall flow of decoding processing by the leaf CU decoding unit 143.
 (ステップS301)
 まず、リーフCU復号部143は、TU復号部433にて、対象リーフCUについてのTU情報を復号する。また、対象リーフCUについてのTU情報に含まれるTU分割情報SP_TUを参照して、対象リーフCUを1または複数のブロックへと分割する。また、各ブロックについての量子化予測残差を、対象リーフCUについての符号化データ#142aより復号し、復号された量子化予測残差を、予測残差復元部434に供給する。
(Step S301)
First, the leaf CU decoding unit 143 uses the TU decoding unit 433 to decode the TU information for the target leaf CU. Further, the target leaf CU is divided into one or a plurality of blocks with reference to the TU partition information SP_TU included in the TU information regarding the target leaf CU. Also, the quantization prediction residual for each block is decoded from the encoded data # 142a for the target leaf CU, and the decoded quantization prediction residual is supplied to the prediction residual restoration unit 434.
 (ステップS302)
 続いて、リーフCU復号部143は、ステップS301にて復号された量子化予測残差を、予測残差復元部434にて逆量子化および逆DCT変換することによって、対象リーフCUについての画素毎の予測残差を復元する。
(Step S302)
Subsequently, the leaf CU decoding unit 143 performs inverse quantization and inverse DCT transform on the quantized prediction residual decoded in step S301 by the prediction residual restoration unit 434, so that each pixel for the target leaf CU is processed. Restore the prediction residual of.
 (ステップS303)
 続いて、リーフCU復号部143は、PU復号部431にて、対象リーフCUについてのPU情報を復号することによって、対象リーフCUの各パーティションへの分割パターンを決定すると共に、各パーティションについての予測モードを決定し、決定された各予測モードを、各パーティションに割り付ける。具体的なPU情報の復号処理については既に説明したため、説明を省略する。
(Step S303)
Subsequently, the leaf CU decoding unit 143 determines the division pattern for each partition of the target leaf CU by decoding the PU information about the target leaf CU in the PU decoding unit 431, and also predicts each partition. A mode is determined, and each determined prediction mode is assigned to each partition. Since specific PU information decoding processing has already been described, description thereof will be omitted.
 (ステップS304)
 続いて、リーフCU復号部143は、予測画像生成部432にて、ループ変数part_idの値を0に初期化し、part_id≦Npart-1を満たすpart_idについて、ループ毎のループ変数part_idの増分値を1とするループ処理を開始する。ここで、ループ変数part_idは、上述したものと同様のループ変数である。すなわち、ループ変数part_idは、対象リーフCUに含まれるパーティションを互いに識別するためのインデックスであり、Npartは、対象リーフCUに含まれるパーティションの総数である。
(Step S304)
Subsequently, the leaf CU decoding unit 143 causes the predicted image generation unit 432 to initialize the value of the loop variable part_id to 0, and for the part_id satisfying part_id ≦ Npart−1, the increment value of the loop variable part_id for each loop is set to 1. The loop processing is started. Here, the loop variable part_id is the same loop variable as described above. That is, the loop variable part_id is an index for mutually identifying partitions included in the target leaf CU, and Npart is the total number of partitions included in the target leaf CU.
 (ステップS305)
 続いて、リーフCU復号部143は、part_idによって指定されるパーティションについての予測画像を、予測画像生成部432にて生成する。具体的な予測画像の生成処理については、既に説明したため、説明を省略する。
(Step S305)
Subsequently, the leaf CU decoding unit 143 generates a predicted image for the partition specified by part_id in the predicted image generation unit 432. Since specific prediction image generation processing has already been described, description thereof will be omitted.
 (ステップS306)
 続いて、リーフCU復号部143は、予測画像生成部432より供給される予測画像#432と、予測残差復元部434より供給される予測残差とを加算することによって、対象リーフCUについての復号画像を、復号画像生成部435にて生成する。
(Step S306)
Subsequently, the leaf CU decoding unit 143 adds the prediction image # 432 supplied from the prediction image generation unit 432 and the prediction residual supplied from the prediction residual restoration unit 434, thereby adding a prediction for the target leaf CU. A decoded image generation unit 435 generates a decoded image.
 (ステップS307)
 本ステップは、ループの終端である。
(Step S307)
This step is the end of the loop.
 以上のように、本実施形態に係る動画像復号装置1は、予測モード指定情報その他のサイド情報と共に符号化データから復号した予測残差に、予測モード群に属する予測モードであって、上記予測モード指定情報により指定される予測モードに従って生成した予測画像を加算することによって復号画像を生成する画像復号装置において、対象予測単位に関する予測モード群に追加する予測モード数N(Nは任意の自然数)を、上記サイド情報に含まれる符号化パラメータであって、上記対象予測単位を包含する符号化単位、または、該符号化単位よりも先に復号された符号化単位に関する符号化パラメータを参照して設定する設定手段(PU復号部431)と、上記対象予測単位に関する予測モード群に追加するN個の予測モードの各々に対応する予測パラメータを、上記復号画像の復号済領域から導出する導出手段(予測画像生成部432)と、を備えている。 As described above, the video decoding device 1 according to the present embodiment includes the prediction mode belonging to the prediction mode group in the prediction residual decoded from the encoded data together with the prediction mode designation information and other side information, and the prediction In an image decoding apparatus that generates a decoded image by adding predicted images generated according to a prediction mode specified by mode specifying information, the number of prediction modes N (N is an arbitrary natural number) added to the prediction mode group related to the target prediction unit Referring to a coding parameter included in the side information and including a coding unit including the target prediction unit or a coding unit decoded prior to the coding unit. Setting means (PU decoding unit 431) to set and each of the N prediction modes to be added to the prediction mode group related to the target prediction unit The prediction parameters for response, and a, and derivation means (predicted image generation unit 432) for deriving from the decoded region of the decoded image.
 予測単位には、予測モード群に属する予測モード数を増やして予測精度を上げる方が符号化効率が向上するタイプと、予測モード群に属する予測モード数を減らして符号量を下げる方が符号化効率が向上するタイプとがある。発明者らが得た知見によれば、ある予測単位が前者のタイプに属するのか後者のタイプに属するのかは、普通、その予測単位、又は、その符号化単位よりも先に復号される符号化単位に関する符号化パラメータから判定することができる。 The prediction unit is a type that increases the prediction accuracy by increasing the number of prediction modes belonging to the prediction mode group, and the coding efficiency is improved by reducing the number of prediction modes belonging to the prediction mode group and improving the coding efficiency. Some types improve efficiency. According to the knowledge obtained by the inventors, whether a certain prediction unit belongs to the former type or the latter type is usually encoded that is decoded before the prediction unit or the encoding unit. It can be determined from the coding parameters for the unit.
 上記の構成によれば、対象予測単位を包含する符号化単位、又は、その符号化単位よりも先に復号される符号化単位に関する符号化パラメータを参照することによって、符号化効率を大きくするように(望ましくは符号化効率を最大化するように)、対象予測単位に関する予測モード群に追加する予測モード数を設定することができる。したがって、上記の構成によれば、高い予測精度を維持しつつ、符号化データの符号量の増大を抑制することによって、符号化効率を向上させることができる。 According to the above configuration, it is possible to increase the coding efficiency by referring to the coding parameter relating to the coding unit including the target prediction unit or the coding unit decoded prior to the coding unit. In addition, it is possible to set the number of prediction modes to be added to the prediction mode group related to the target prediction unit (desirably to maximize the coding efficiency). Therefore, according to said structure, encoding efficiency can be improved by suppressing the increase in the code amount of coding data, maintaining high prediction accuracy.
 また、一般に、ある符号化単位についての符号化対象画像を符号化するために必要なサイド情報の符号量は、当該符号化単位に含まれる予測単位の総数と正の相関を有する。すなわち、予測単位の総数が大きければサイド情報の符号量が増大する。また、ある符号化単位に含まれる予測単位の総数は、各予測単位のサイズと負の相関を有する。 Also, in general, the amount of code of side information necessary for encoding an encoding target image for a certain encoding unit has a positive correlation with the total number of prediction units included in the encoding unit. That is, if the total number of prediction units is large, the code amount of the side information increases. The total number of prediction units included in a certain coding unit has a negative correlation with the size of each prediction unit.
 本実施形態のように、上記予測モード群に追加する予測モード数Nを、上記符号化パラメータにより規定される上記対象予測単位のサイズに応じた値に設定することにより、高い予測精度を保ちつつ、符号化対象画像を符号化するために必要なサイド情報の符号量の増加を抑制することができるので、符号化効率が向上する。 As in this embodiment, by setting the number of prediction modes N to be added to the prediction mode group to a value according to the size of the target prediction unit defined by the encoding parameter, high prediction accuracy is maintained. Since an increase in the code amount of side information necessary for encoding the encoding target image can be suppressed, encoding efficiency is improved.
 なお、本実施形態において、PU復号部431は、対象パーティションのサイズと負の相関を有するように、当該対象パーティションについて選択可能な補正方向の総数を決定している。具体的には、対象パーティションのサイズが4×4画素や8×8画素である場合、対象パーティションについて選択可能な補正方向の総数を8に設定し、対象パーティションのサイズが16×16画素である場合、対象パーティションについて選択可能な補正方向の総数を2に設定し、対象パーティションのサイズが32×32画素や64×64画素である場合、対象パーティションについて選択可能な補正方向の総数を0としている。 In this embodiment, the PU decoding unit 431 determines the total number of correction directions that can be selected for the target partition so as to have a negative correlation with the size of the target partition. Specifically, when the size of the target partition is 4 × 4 pixels or 8 × 8 pixels, the total number of correction directions that can be selected for the target partition is set to 8, and the size of the target partition is 16 × 16 pixels. In this case, when the total number of correction directions that can be selected for the target partition is set to 2 and the size of the target partition is 32 × 32 pixels or 64 × 64 pixels, the total number of correction directions that can be selected for the target partition is set to 0. .
 このような設定が有効である理由を図12の(b)に示す実験結果に基づいて説明する。図12の(b)は補正方向数とエッジベース予測導入の効果の関係をパーティションサイズ毎に測定した結果を示す。図12の(b)において横軸は補正方向数を示し、sd0、sd1、sd2、sd3、sd4はそれぞれ0個、2個、4個、6個、8個の補正方向数を表す。また、図12の(b)において縦軸はエッジベース予測導入の効果を示す。より厳密には、基本予測モードに加えて追加予測モードを選択可能とすることで実現されるレート歪コストの減少幅を、基本予測モードのみを選択可能とする場合のレート歪コストを基準として削減率で表わしている。 The reason why such a setting is effective will be described based on the experimental results shown in FIG. FIG. 12B shows the result of measuring the relationship between the number of correction directions and the effect of introducing edge-based prediction for each partition size. In FIG. 12B, the horizontal axis indicates the number of correction directions, and sd0, sd1, sd2, sd3, and sd4 indicate the number of correction directions of 0, 2, 4, 6, and 8, respectively. In FIG. 12B, the vertical axis indicates the effect of introducing edge-based prediction. More precisely, the rate distortion cost reduction achieved by making it possible to select the additional prediction mode in addition to the basic prediction mode is reduced based on the rate distortion cost when only the basic prediction mode can be selected. Expressed as a percentage.
 図12の(b)から、パーティションサイズが4×4画素や8×8画素の場合はsd4すなわち補正方向を8個とすることでレート歪コストを最も削減できることが分かる。一方、16×16画素の場合はsd1すなわち補正方向を2個とすれば十分レート歪コストを削減でき、それ以上補正方向を追加しても効果が認められないことが分かる。さらに、32×32画素や64×64画素の場合はsd0すなわち補正方向を0個(主方向のみを用いる)場合にレート歪コストの削減幅が最大であり、それ以上補正方向を追加すると逆効果であることが分かる。 FIG. 12B shows that when the partition size is 4 × 4 pixels or 8 × 8 pixels, the rate distortion cost can be reduced most by setting sd4, that is, eight correction directions. On the other hand, in the case of 16 × 16 pixels, it can be seen that if sd1, that is, two correction directions are used, the rate distortion cost can be sufficiently reduced, and the effect is not recognized even if more correction directions are added. Further, in the case of 32 × 32 pixels or 64 × 64 pixels, sd0, that is, when the number of correction directions is 0 (only the main direction is used), the reduction rate of the rate distortion cost is the maximum. It turns out that it is.
 一般に補正方向の追加はサイド情報の符号量増加と予測精度の向上の両方に関係する。小さいサイズのパーティションでは、大きいサイズの場合に較べて、単位面積当たりのサイド情報増加の影響が大きい。一方で、小さいサイズのパーティションでは、大きいサイズの場合に較べて、予測精度の向上幅が大きい。これは、小さいサイズではブロック内に直線が存在する可能性が高いことに起因する。図12の(b)に示す実験結果から、後者の予測精度向上幅の影響が、前者のサイド情報増加の影響を上回っているといえる。 Generally, the addition of the correction direction is related to both the increase of the code amount of the side information and the improvement of the prediction accuracy. In the small size partition, the influence of the side information increase per unit area is larger than in the case of the large size. On the other hand, in the small size partition, the improvement range of the prediction accuracy is large as compared with the case of the large size. This is because there is a high possibility that a straight line exists in a block at a small size. From the experimental results shown in FIG. 12 (b), it can be said that the influence of the latter prediction accuracy improvement exceeds the influence of the former side information increase.
 (動画像符号化装置2)
 以下では、本実施形態に係る動画像符号化装置(符号化装置)2について、図13~図17を参照して説明する。なお、すでに説明した部分と同じ部分については、同じ符号を付し、その説明を省略する。
(Moving picture encoding device 2)
Hereinafter, the moving picture encoding apparatus (encoding apparatus) 2 according to the present embodiment will be described with reference to FIGS. In addition, the same part as the part already demonstrated is attached | subjected the same code | symbol, and the description is abbreviate | omitted.
 図13は、動画像符号化装置2の構成を示すブロック図である。図13に示すように、動画像符号化装置2は、LCUヘッダ情報決定部21、ヘッダ情報符号化部22、LCU設定部23、LCU符号化部24、可変長符号多重化部25、LCU復号部26、および、フレームメモリ27を備えている。 FIG. 13 is a block diagram showing a configuration of the moving picture encoding apparatus 2. As illustrated in FIG. 13, the moving image encoding device 2 includes an LCU header information determination unit 21, a header information encoding unit 22, an LCU setting unit 23, an LCU encoding unit 24, a variable length code multiplexing unit 25, and an LCU decoding. A unit 26 and a frame memory 27 are provided.
 動画像符号化装置2は、概略的に言えば、入力画像#100を符号化することによって符号化データ#1を生成し、出力する装置である。また、動画像符号化装置2は、その一部に、H.264/MPEG-4 AVC規格に採用されている技術、VCEG(Video Coding Expert Group)における共同開発用コーデックであるKTAソフトウェアに採用されている技術、および、その後継コーデックであるTMuC(Test Model under Consideration)ソフトウェアに採用されている技術を用いている動画像符号化装置である。 The video encoding device 2 is a device that generates and outputs encoded data # 1 by encoding the input image # 100, in brief. In addition, the moving image encoding apparatus 2 includes, as part thereof, H.264. H.264 / MPEG-4 AVC standard technology, VCEG (Video Coding Expert Group) technology used in KTA software, which is a joint development codec, and successor codec TMuC (Test Model Under Consideration) ) A moving picture coding apparatus using a technique adopted in software.
 LCUヘッダ情報決定部21は、入力画像#100に基づいて、LCUヘッダ情報を決定する。決定されたLCUヘッダ情報はLCUヘッダ情報#21として出力される。ここで、LCUヘッダ情報#21は、(1)対象スライスに属するLCUのサイズ、形状および対象スライス内での位置についての情報、並びに、(2)各LCUに属するリーフCUのサイズ、形状および対象LCU内での位置についての情報を含んでいる。LCUヘッダ情報#21は、LCU設定部23に入力されると共に、ヘッダ情報符号化部22に供給される。 The LCU header information determination unit 21 determines LCU header information based on the input image # 100. The determined LCU header information is output as LCU header information # 21. Here, the LCU header information # 21 includes (1) information about the size and shape of the LCU belonging to the target slice and the position in the target slice, and (2) the size, shape and target of the leaf CU belonging to each LCU. Contains information about the location within the LCU. The LCU header information # 21 is input to the LCU setting unit 23 and also supplied to the header information encoding unit 22.
 ヘッダ情報符号化部22は、LCUヘッダ情報決定部21より供給されるLCUヘッダ情報#21、および、LCU符号化部24より供給されるリーフCUヘッダ情報#53を符号化し、符号化済ヘッダ情報#22を出力する。符号化済ヘッダ情報#22は、可変長符号多重化部25に供給される。 The header information encoding unit 22 encodes the LCU header information # 21 supplied from the LCU header information determination unit 21 and the leaf CU header information # 53 supplied from the LCU encoding unit 24, and has encoded header information. # 22 is output. The encoded header information # 22 is supplied to the variable length code multiplexer 25.
 LCU設定部23は、LCUヘッダ情報#21に基づいて、入力画像(符号化対象画像)#100を複数のLCUに分割し、各LCUに関するLCU画像#23を出力する。LCU画像#23は、LCU符号化部24に順次供給される。 The LCU setting unit 23 divides the input image (encoding target image) # 100 into a plurality of LCUs based on the LCU header information # 21, and outputs an LCU image # 23 related to each LCU. The LCU image # 23 is sequentially supplied to the LCU encoding unit 24.
 LCU符号化部24は、順次入力されるLCU画像#23を符号化し、LCU符号化データ#24を生成する。生成されたLCU符号化データ#24は、可変長符号多重化部25、およびLCU復号部26に供給される。LCU符号化部24の構成については、後述するため、ここでは説明を省略する。 The LCU encoding unit 24 encodes sequentially input LCU image # 23 to generate LCU encoded data # 24. The generated LCU encoded data # 24 is supplied to the variable length code multiplexing unit 25 and the LCU decoding unit 26. Since the configuration of the LCU encoding unit 24 will be described later, description thereof is omitted here.
 可変長符号多重化部25は、符号化済ヘッダ情報#22と、LCU符号化データ#24とを多重化することによって符号化データ#1を生成し、出力する。 The variable length code multiplexing unit 25 generates encoded data # 1 by multiplexing encoded header information # 22 and LCU encoded data # 24, and outputs the encoded data # 1.
 LCU復号部26は、入力された個々のLCUに対応するLCU符号化データ#24を順次復号することにより、個々のLCUに対応する復号画像#26を生成し、出力する。復号画像#26は、フレームメモリ27に供給される。LCU復号部26の構成は、動画像復号装置1の備えるLCU復号部14の構成と同様であるため、ここではその詳細な説明を省略する。 The LCU decoding unit 26 sequentially decodes the LCU encoded data # 24 corresponding to each input LCU, thereby generating and outputting a decoded image # 26 corresponding to each LCU. The decoded image # 26 is supplied to the frame memory 27. Since the configuration of the LCU decoding unit 26 is the same as the configuration of the LCU decoding unit 14 included in the video decoding device 1, detailed description thereof is omitted here.
 フレームメモリ27には、入力された復号画像#26が記録される。対象LCUを符号化する時点では、当該対象LCUよりもラスタスキャン順で先行する全てのLCUに対応する復号画像がフレームメモリ27に記録されている。なお、フレームメモリに記録された復号画像を復号画像#27と呼ぶ。すなわち、復号画像#27は、フレームメモリ27に記録された復号画像#26に対応している。 The input decoded image # 26 is recorded in the frame memory 27. At the time of encoding the target LCU, decoded images corresponding to all the LCUs preceding the target LCU in the raster scan order are recorded in the frame memory 27. The decoded image recorded in the frame memory is called decoded image # 27. That is, the decoded image # 27 corresponds to the decoded image # 26 recorded in the frame memory 27.
 (LCU符号化部24)
 以下では、LCU符号化部24について、参照する図面を替えてより具体的に説明する。
(LCU encoder 24)
Hereinafter, the LCU encoding unit 24 will be described more specifically with reference to different drawings.
 図14は、LCU符号化部24の構成を示すブロック図である。図14に示すように、LCU符号化部24は、LCU分割構造決定部241、リーフCUスキャン部242、および、リーフCU符号化部243を備えている。 FIG. 14 is a block diagram showing a configuration of the LCU encoding unit 24. As illustrated in FIG. 14, the LCU encoding unit 24 includes an LCU division structure determining unit 241, a leaf CU scanning unit 242, and a leaf CU encoding unit 243.
 LCU分割構造決定部241は、LCUヘッダ情報#21を参照して、対象LCUのリーフCUへの分割構造を決定する。対象LCUのリーフCUへの分割構造を指定する分割構造情報#241は、対象LCUについてのLCU画像#23と共にリーフCUスキャン部242に供給される。なお、LCU分割構造決定部241により決定される分割構造の例としては、図5の(a)に示したものが挙げられる。 The LCU partition structure determining unit 241 refers to the LCU header information # 21 to determine the partition structure of the target LCU into leaf CUs. Division structure information # 241 for designating the division structure of the target LCU into leaf CUs is supplied to the leaf CU scanning unit 242 together with the LCU image # 23 for the target LCU. An example of the divided structure determined by the LCU divided structure determining unit 241 is the one shown in FIG.
 リーフCUスキャン部242は、対象LCUに含まれるリーフCUを所定の順序(例えば、ラスタスキャン順)でスキャンし、処理対象となるリーフCUである対象リーフCUについてのリーフCU画像#242aおよびCU情報#242bを、リーフCU符号化部243に供給する。ここで、CU情報#242bは、対象リーフCUについてのサイズ、形状および対象LCUでの位置、を含んでいる。 The leaf CU scanning unit 242 scans the leaf CUs included in the target LCU in a predetermined order (for example, raster scan order), and the leaf CU image # 242a and CU information about the target leaf CU that is the leaf CU to be processed. # 242b is supplied to the leaf CU encoding unit 243. Here, the CU information # 242b includes the size and shape of the target leaf CU and the position in the target LCU.
 (リーフCU符号化部243)
 リーフCU符号化部243は、対象リーフCUについてのリーフCU画像およびCU情報、並びに、フレームメモリ15に格納されている復号済みの画素値を参照して、対象リーフCUについての予測画像を生成すると共に、リーフCU画像と予測画像との予測残差を符号化する。また、リーフCU符号化部243は、対象LCUに属する各リーフCUについての符号化された予測残差を統合し、当該対象LCUについてのLCU符号化データ#24を生成する。
(Leaf CU encoding unit 243)
The leaf CU encoding unit 243 refers to the leaf CU image and CU information for the target leaf CU, and the decoded pixel value stored in the frame memory 15, and generates a predicted image for the target leaf CU. At the same time, the prediction residual between the leaf CU image and the prediction image is encoded. Also, the leaf CU encoding unit 243 integrates the encoded prediction residuals for each leaf CU belonging to the target LCU, and generates LCU encoded data # 24 for the target LCU.
 図15は、リーフCU符号化部243の構成を示すブロック図である。図15に示すように、リーフCU符号化部243は、PU決定部531、予測画像生成部532、復号画像生成部533、予測残差算出部534、TU決定部535、予測残差復元部536、レート歪評価部537、および、LCU符号化データ生成部538を備えている。 FIG. 15 is a block diagram showing the configuration of the leaf CU encoding unit 243. As illustrated in FIG. 15, the leaf CU encoding unit 243 includes a PU determination unit 531, a predicted image generation unit 532, a decoded image generation unit 533, a prediction residual calculation unit 534, a TU determination unit 535, and a prediction residual restoration unit 536. A rate distortion evaluation unit 537 and an LCU encoded data generation unit 538.
 (PU決定部531)
 PU決定部531は、リーフCU画像#242a、および、CU情報#242bを参照し、(1)対象リーフCUの各パーティションへの分割パターン、および、(2)各パーティションに割り付ける予測モード、を決定する。ここで、対象リーフCUの各パーティションへの分割パターンとしては、例えば、対象リーフCUを分割せずにそのまま1つのパーティションとして取り扱う分割パターン、または、対象リーフCUを対称的に4つのパーティションに分割する分割パターン、の何れかが選択可能である。
(PU determination unit 531)
The PU determination unit 531 refers to the leaf CU image # 242a and the CU information # 242b, and determines (1) a division pattern for each partition of the target leaf CU and (2) a prediction mode to be allocated to each partition. To do. Here, as a division pattern into the partitions of the target leaf CU, for example, a division pattern that is handled as one partition without dividing the target leaf CU, or the target leaf CU is divided into four partitions symmetrically. One of the division patterns can be selected.
 また、PU決定部531は、対象パーティションのサイズに応じて、当該対象パーティションについて選択可能な補正方向の総数を決定する。より具体的には、PU決定部531は、対象パーティションのサイズと正の相関を有するように、当該対象パーティションについて選択可能な補正方向の総数を決定する。 Also, the PU determination unit 531 determines the total number of correction directions that can be selected for the target partition according to the size of the target partition. More specifically, the PU determination unit 531 determines the total number of correction directions that can be selected for the target partition so as to have a positive correlation with the size of the target partition.
 例えば、対象パーティションのサイズが4×4画素や8×8画素である場合、対象パーティションについて選択可能な補正方向の総数を8に設定し、対象パーティションのサイズが16×16画素である場合、対象パーティションについて選択可能な補正方向の総数を2に設定し、対象パーティションのサイズが32×32や64×64である場合、対象パーティションについて選択可能な補正方向の総数を0に設定する。より一般には、PU決定部531、対象パーティションに属する画素の総数がNpix1であるときに、当該対象パーティションについて選択可能な補正方向の総数をNad1に決定し、対象パーティションに属する画素の総数がNpix2(Npix1≧Npix2)であるときに、当該対象パーティションについて選択可能な補正方向の総数をNad2(Nad1<Nad2)に決定する。 For example, when the size of the target partition is 4 × 4 pixels or 8 × 8 pixels, the total number of correction directions that can be selected for the target partition is set to 8, and when the size of the target partition is 16 × 16 pixels, the target When the total number of correction directions that can be selected for the partition is set to 2 and the size of the target partition is 32 × 32 or 64 × 64, the total number of correction directions that can be selected for the target partition is set to 0. More generally, when the total number of pixels belonging to the target partition is Npix1, the PU determination unit 531 determines the total number of correction directions that can be selected for the target partition as Nad1, and the total number of pixels belonging to the target partition is Npix2 ( When Npix1 ≧ Npix2), the total number of correction directions that can be selected for the target partition is determined to be Nad2 (Nad1 <Nad2).
 (1)対象リーフCUを4つのパーティションに分割するか否かを指定するイントラ分割フラグ(intra_split_flag)、および、(2)各パーティションに割り付けられた予測モードを指定する予測モード情報、を含むPU情報#531は、予測画像生成部532、および、LCU符号化データ生成部538に供給される。 PU information including (1) an intra split flag (intra_split_flag) that specifies whether or not the target leaf CU is split into four partitions, and (2) prediction mode information that specifies a prediction mode assigned to each partition. # 531 is supplied to the predicted image generation unit 532 and the LCU encoded data generation unit 538.
 なお、PU情報#531には、対象パーティションに対して主方向を用いるのか、副方向を用いるのかを指定する1ビットのフラグを含む追加インデックスAIが含まれる。また、副方向を用いる場合、追加インデックスAIには、複数の副方向であって、その総数が対象パーティションのサイズに応じて決定された複数の副方向の何れを用いるのかを指定するインデックスが含まれる。 Note that the PU information # 531 includes an additional index AI including a 1-bit flag that specifies whether to use the main direction or the sub-direction for the target partition. In addition, when the sub-direction is used, the additional index AI includes a plurality of sub-directions and an index that specifies which of the plurality of sub-directions whose total number is determined according to the size of the target partition is used. It is.
 なお、後述するように、PU決定部531は、(1)対象リーフCUの各パーティションへの可能な分割パターン、および、(2)各パーティションに割り付ける可能な予測モード、の全ての組み合わせの各々に対応するPU情報を生成する。また、後述する予測画像、予測残差、および、復号画像の生成は、それら各々のPU情報について逐一行われる。 As will be described later, the PU determination unit 531 applies to each of all combinations of (1) a possible division pattern for each partition of the target leaf CU and (2) a prediction mode that can be assigned to each partition. Corresponding PU information is generated. In addition, generation of a prediction image, a prediction residual, and a decoded image, which will be described later, is performed for each piece of PU information.
 また、PU決定部531は、後述するレート歪評価部537より供給される、レート歪を最適化するPU情報を、リーフCUヘッダ情報#53の一部としてヘッダ情報符号化部22に供給すると共に、当該レート歪を最適化するPU情報を、対象リーフCUについての最終的なPU情報として、予測画像生成部532、および、LCU符号化データ生成部538に供給する。 In addition, the PU determination unit 531 supplies PU information that optimizes rate distortion supplied from the rate distortion evaluation unit 537 described later to the header information encoding unit 22 as part of the leaf CU header information # 53. Then, the PU information for optimizing the rate distortion is supplied to the predicted image generation unit 532 and the LCU encoded data generation unit 538 as the final PU information for the target leaf CU.
 (予測画像生成部532)
 予測画像生成部532は、フレームメモリ27に格納されている復号画像#27、および、PU決定部531より供給されるPU情報#531を参照して、対象リーフCUについての予測画像#532を生成する。予測画像生成部532による予測画像生成処理は、動画像復号装置1の備える予測画像生成部432と同様であるので、ここでは説明を省略する。
(Predicted image generation unit 532)
The predicted image generation unit 532 generates a predicted image # 532 for the target leaf CU with reference to the decoded image # 27 stored in the frame memory 27 and the PU information # 531 supplied from the PU determination unit 531. To do. Since the predicted image generation process by the predicted image generation unit 532 is the same as the predicted image generation unit 432 included in the video decoding device 1, description thereof is omitted here.
 (予測残差算出部534)
 予測残差算出部534は、リーフCU画像#242aから予測画像#532を減算することによって、予測残差#534を生成する。生成された予測残差#534は、TU決定部535に供給される。
(Prediction residual calculation unit 534)
The prediction residual calculation unit 534 generates a prediction residual # 534 by subtracting the prediction image # 532 from the leaf CU image # 242a. The generated prediction residual # 534 is supplied to the TU determination unit 535.
 (TU決定部535)
 TU決定部535は、リーフCU画像#242a、および、CU情報#242bを参照し、対象リーフCUの1または複数のブロックへの分割パターンを決定する。また、決定された分割パターンに従って、予測残差算出部534より供給される対象リーフCUについての予測残差#534を、各ブロックについての予測残差に分割する。
(TU determination unit 535)
The TU determination unit 535 refers to the leaf CU image # 242a and the CU information # 242b, and determines a division pattern of the target leaf CU into one or a plurality of blocks. Also, according to the determined division pattern, the prediction residual # 534 for the target leaf CU supplied from the prediction residual calculation unit 534 is divided into prediction residuals for each block.
 また、TU決定部535は、各ブロックについての予測残差をDCT変換(Discrete Cosine Transform)することによって周波数領域における予測残差を生成した後、当該周波数領域における予測残差を量子化することによってブロック毎の量子化予測残差を生成する。生成されたブロック毎の量子化予測残差は、対象リーフCUの分割パターンを指定するTU分割情報SP_TUと共に、TU情報#535として予測残差復元部536、および、LCU符号化データ生成部538に供給される。 The TU determination unit 535 generates a prediction residual in the frequency domain by performing DCT transform (DiscretecreCosine Transform) on the prediction residual for each block, and then quantizes the prediction residual in the frequency domain. Generate a quantized prediction residual for each block. The generated quantization prediction residual for each block is transmitted to the prediction residual restoration unit 536 and the LCU encoded data generation unit 538 as TU information # 535 together with TU partition information SP_TU that specifies the partition pattern of the target leaf CU. Supplied.
 なお、後述するように、TU決定部535は、対象リーフCUの各ブロックへの可能な分割パターンの全てに対応するTU情報#535を生成する。また、量子化予測残差の生成は、それら全ての分割パターンについて行われる。 As will be described later, the TU determination unit 535 generates TU information # 535 corresponding to all possible division patterns for each block of the target leaf CU. Further, the generation of the quantized prediction residual is performed for all of the divided patterns.
 また、TU決定部535は、後述するレート歪評価部537より供給される、レート歪を最適化するTU情報を、リーフCUヘッダ情報#53としてヘッダ情報符号化部22に供給すると共に、当該レート歪を最適化するTU情報を、対象リーフCUについての最終的なTU情報として、予測残差復元部536、および、LCU符号化データ生成部538に供給する。 Further, the TU determination unit 535 supplies the TU information for optimizing the rate distortion supplied from the rate distortion evaluation unit 537 described later to the header information encoding unit 22 as leaf CU header information # 53, and the rate. TU information for optimizing distortion is supplied to the prediction residual restoration unit 536 and the LCU encoded data generation unit 538 as final TU information for the target leaf CU.
 (予測残差復元部536)
 予測残差復元部536は、TU決定部535より供給される、ブロック毎の量子化予測残差を、逆量子化、および、逆DCT変換(Inverse Discrete Cosine Transform)することによって、ブロック毎の予測残差を復元する。また、ブロック毎の予測残差を、TU決定部535によって決定された分割パターンに従って統合し、対象リーフCUについての予測残差#536を生成する。生成された対象リーフCUについての予測残差#536は、復号画像生成部533に供給される。
(Prediction residual restoration unit 536)
The prediction residual restoration unit 536 performs prediction for each block by performing inverse quantization and inverse DCT transform (Inverse Discrete Cosine Transform) on the quantization prediction residual for each block supplied from the TU determination unit 535. Restore the residual. Further, the prediction residuals for each block are integrated according to the division pattern determined by the TU determination unit 535, and the prediction residual # 536 for the target leaf CU is generated. The prediction residual # 536 for the generated target leaf CU is supplied to the decoded image generation unit 533.
 (復号画像生成部533)
 復号画像生成部533は、予測画像生成部532より供給される対象リーフCUについての予測画像#532と、予測残差復元部536より供給される対象リーフCUについての予測残差#536とを加算することによって、対象リーフCUについての復号画像#533を生成する。生成された対象リーフCUについての復号画像#533は、レート歪評価部537に供給される。
(Decoded image generation unit 533)
The decoded image generation unit 533 adds the prediction image # 532 for the target leaf CU supplied from the prediction image generation unit 532 and the prediction residual # 536 for the target leaf CU supplied from the prediction residual restoration unit 536. As a result, a decoded image # 533 for the target leaf CU is generated. The generated decoded image # 533 for the target leaf CU is supplied to the rate distortion evaluation unit 537.
 (レート歪評価部537)
 レート歪評価部537は、復号画像生成部533より供給される対象リーフCUについての復号画像#533、および、対象リーフCUについてのリーフCU画像#242aを参照し、対象リーフCUについてのレート歪を評価する。
(Rate distortion evaluation unit 537)
The rate distortion evaluation unit 537 refers to the decoded image # 533 for the target leaf CU and the leaf CU image # 242a for the target leaf CU supplied from the decoded image generation unit 533, and calculates the rate distortion for the target leaf CU. evaluate.
 また、レート歪評価部537は、(1)対象リーフCUの各パーティションへの可能な分割パターン、(2)各パーティションに割り付ける可能な予測モード、および、(3)対象リーフCUの各ブロックへの可能な分割パターンの組み合わせのうち、レート歪を最適化する組み合わせを決定する。 Further, the rate distortion evaluation unit 537 includes (1) a possible division pattern for each partition of the target leaf CU, (2) a prediction mode that can be assigned to each partition, and (3) each block of the target leaf CU. Among possible combinations of divided patterns, a combination that optimizes the rate distortion is determined.
 決定された組み合わせによって指定されるPU情報、および、TU情報は、それぞれ、PU決定部531、および、TU決定部535に供給される。 The PU information and TU information specified by the determined combination are supplied to the PU determination unit 531 and the TU determination unit 535, respectively.
 (LCU符号化データ生成部538)
 LCU符号化データ生成部538は、PU決定部531より供給される最終的なPU情報、および、TU決定部535より供給される最終的なTU情報を可変長符号化することによって、対象リーフCUについての符号化データを生成する。また、LCU符号化データ生成部538は、対象LCUに属する各リーフCUについての符号化データを多重化することによって、対象LCUについての符号化データ#24を生成する。
(LCU encoded data generation unit 538)
The LCU encoded data generation unit 538 performs variable length encoding on the final PU information supplied from the PU determination unit 531 and the final TU information supplied from the TU determination unit 535, thereby performing the target leaf CU. The encoded data for is generated. Also, the LCU encoded data generation unit 538 generates encoded data # 24 for the target LCU by multiplexing the encoded data for each leaf CU belonging to the target LCU.
 (リーフCU符号化部243による符号化処理の流れ)
 以下では、リーフCU符号化部243による符号化処理の流れについて、図16を参照して説明する。図16は、リーフCU符号化部243による符号化処理の流れを示すフローチャートである。
(Flow of encoding process by leaf CU encoding unit 243)
Hereinafter, the flow of the encoding process performed by the leaf CU encoder 243 will be described with reference to FIG. FIG. 16 is a flowchart showing the flow of the encoding process performed by the leaf CU encoder 243.
 (ステップS401)
 まず、リーフCU符号化部243は、ループ変数pu_param_idの値を0に初期化し、pu_param_id≦Npu-1を満たすpu_param_idについて、ループ毎のループ変数pu_param_idの増分値を1とする第1のループ処理を開始する。ここで、ループ変数pu_param_idは、対象リーフCUの各パーティションへの可能な分割パターン、および、各パーティションへ割り付ける可能な予測モードの組み合わせを互いに識別するためのインデックスである。また、Npuは、対象リーフCUの各パーティションへの可能な分割パターン、および、各パーティションへ割り付ける可能な予測モードの組み合わせの総数を表している。
(Step S401)
First, the leaf CU encoding unit 243 initializes the value of the loop variable pu_param_id to 0, and for the pu_param_id that satisfies pu_param_id ≦ Npu−1, performs a first loop process in which the increment value of the loop variable pu_param_id for each loop is 1. Start. Here, the loop variable pu_param_id is an index for mutually identifying combinations of possible division patterns for each partition of the target leaf CU and possible prediction modes to be assigned to each partition. Npu represents the total number of combinations of possible division patterns for each partition of the target leaf CU and prediction modes that can be assigned to each partition.
 例えば、対象リーフCUが、1または4つのパーティションに分割が可能であり、各パーティションについて選択可能な予測モードの数がNmodeであるとすると、
 Npu=1×Nmode+4×Nmode
であり、ループ変数pu_param_idは、それら1×Nmode+4×Nmodeの組み合わせを互いに識別するためのインデックスとして機能する。
For example, if the target leaf CU can be divided into one or four partitions and the number of prediction modes that can be selected for each partition is Nmode,
Npu = 1 × Nmode + 4 × Nmode
The loop variable pu_param_id functions as an index for identifying these combinations of 1 × Nmode + 4 × Nmode from each other.
 (ステップS402)
 続いて、リーフCU符号化部243は、ループ変数tu_param_idの値を0に初期化し、tu_param_id≦Ntu-1を満たすtu_param_idについて、ループ毎のループ変数tu_param_idの増分値を1とする第2のループ処理を開始する。ここで、ループ変数tu_param_idは、対象リーフCUの各ブロックへの可能な分割パターンを互いに識別するためのインデックスである。また、Ntuは、対象リーフCUの各ブロックへの可能な分割パターンの総数を表している。
(Step S402)
Subsequently, the leaf CU encoding unit 243 initializes the value of the loop variable tu_param_id to 0, and for the tu_param_id that satisfies tu_param_id ≦ Ntu−1, the second loop processing sets the increment value of the loop variable tu_param_id to 1 for each loop. To start. Here, the loop variable tu_param_id is an index for mutually identifying possible division patterns for each block of the target leaf CU. Ntu represents the total number of possible division patterns for each block of the target leaf CU.
 例えば、対象リーフCUのサイズが8×8画素であり、当該対象リーフCUが分割されずにそのままブロックとして扱われる分割パターン、および、当該対象リーフCUが、4つの4×4画素のブロックへ対称的に分割される分割パターン、が可能であるとすると、当該対象リーフCUの各ブロックへの可能な分割パターンの総数Ntuは、2であり、ループ変数tu_param_idは、それら2つの分割パターンを互いに識別するためのインデックスとして機能する。 For example, the size of the target leaf CU is 8 × 8 pixels, the division pattern in which the target leaf CU is handled as it is without being divided, and the target leaf CU is symmetrical to four 4 × 4 pixel blocks. If the division pattern to be divided is possible, the total number Ntu of possible division patterns for each block of the target leaf CU is 2, and the loop variable tu_param_id identifies these two division patterns from each other. Function as an index for
 (ステップS403)
 まず、リーフCU符号化部243は、ループ変数part_idの値を0に初期化し、part_id≦Npart-1を満たすpart_idについて、ループ毎のループ変数part_idの増分値を1とする第3のループ処理を開始する。ここで、ループ変数part_idは、各パーティションを互いに識別するためのインデックスであり、Npartは、対象リーフCUを、ループ変数pu_param_idに対応する分割パターンに従ってパーティションに分割した場合の当該パーティションの総数である。
(Step S403)
First, the leaf CU encoding unit 243 initializes the value of the loop variable part_id to 0, and performs a third loop process in which the increment value of the loop variable part_id for each loop is set to 1 for part_id that satisfies part_id ≦ Npart−1. Start. Here, the loop variable part_id is an index for identifying each partition, and Npart is the total number of the partitions when the target leaf CU is divided into partitions according to the division pattern corresponding to the loop variable pu_param_id.
 (ステップS404)
 続いて、リーフCU符号化部243は、ループ変数のセット(pu_param_id,tu_param_id,part_id)によって指定される対象パーティションについての予測画像を、予測画像生成部532にて生成する。
(Step S404)
Subsequently, the leaf CU encoding unit 243 generates a predicted image for the target partition specified by the set of loop variables (pu_param_id, tu_param_id, part_id) in the predicted image generation unit 532.
 (ステップS405)
 続いて、リーフCU符号化部243は、ループ変数のセット(pu_param_id,part_id)によって指定されるパーティションについての予測残差を、予測残差算出部534にて算出する。
(Step S405)
Subsequently, the leaf CU encoding unit 243 calculates the prediction residual for the partition specified by the set of loop variables (pu_param_id, part_id) by the prediction residual calculation unit 534.
 (ステップS406)
 続いて、リーフCU符号化部243は、対象リーフCUに属する対象ブロックについての予測残差を、TU決定部535にて、DCT変換および量子化することによって、量子化予測残差を生成する。
(Step S406)
Subsequently, the leaf CU encoding unit 243 generates a quantized prediction residual by DCT transforming and quantizing the prediction residual for the target block belonging to the target leaf CU in the TU determination unit 535.
 (ステップS407)
 続いて、リーフCU符号化部243は、ステップS406にて生成された量子化予測残差を逆量子化および逆DCT変換することによって、予測残差を復元する。
(Step S407)
Subsequently, the leaf CU encoding unit 243 restores the prediction residual by performing inverse quantization and inverse DCT transform on the quantized prediction residual generated in step S406.
 (ステップS408)
 続いて、リーフCU符号化部243は、ステップS404にて生成された予測画像と、ステップS407にて復元された予測残差とを加算することによって対象パーティションについての復号画像を生成する。
(Step S408)
Subsequently, the leaf CU encoding unit 243 generates a decoded image for the target partition by adding the prediction image generated in step S404 and the prediction residual restored in step S407.
 (ステップS409)
 本ステップは、第3のループの終端である。第3のループ処理によって、対象リーフCUについての復号画像の生成が完了する。
(Step S409)
This step is the end of the third loop. Generation of the decoded image for the target leaf CU is completed by the third loop processing.
 (ステップS410)
 続いて、リーフCU符号化部243は、レート歪評価部537にて、対象リーフCUについての符号量および歪を参照して、当該対象リーフCUについてのレート歪を評価する。
(Step S410)
Subsequently, in the leaf CU encoding unit 243, the rate distortion evaluation unit 537 refers to the code amount and the distortion for the target leaf CU and evaluates the rate distortion for the target leaf CU.
 (ステップS411)
 本ステップは、第2のループの終端である。
(Step S411)
This step is the end of the second loop.
 (ステップS412)
 本ステップは、第1のループの終端である。なお、第1のループ処理および第2のループ処理によって、ループ変数pu_param_idおよびループ変数tu_param_idの全ての可能な組み合わせについてのレート歪の評価が完了する。
(Step S412)
This step is the end of the first loop. Note that the first and second loop processes complete the rate distortion evaluation for all possible combinations of the loop variable pu_param_id and the loop variable tu_param_id.
 (ステップS413)
 続いて、リーフCU符号化部243は、レート歪評価部537にてループ変数pu_param_idおよびループ変数tu_param_idの全ての可能な組み合わせについてのレート歪のうち、最適なレート歪に対応するループ変数pu_param_idおよびループ変数tu_param_idの組み合わせを決定する。
(Step S413)
Subsequently, the leaf CU encoding unit 243, in the rate distortion evaluation unit 537, out of the rate distortions for all possible combinations of the loop variable pu_param_id and the loop variable tu_param_id, the loop variable pu_param_id and the loop corresponding to the optimum rate distortion The combination of the variable tu_param_id is determined.
 (ステップS414)
 リーフCU符号化部243は、ステップS413にて決定された最適なレート歪に対応するループ変数pu_param_idおよびループ変数tu_param_idの組み合わせを用いて、対象リーフCUについての符号化データを生成する。すなわち、リーフCU符号化部243は、対象リーフCUのパーティションおよびブロックへの分割パターンであって、レート歪を最適化する分割パターンを用いて、当該対象リーフCUについての符号化データを生成する。
(Step S414)
The leaf CU encoding unit 243 generates encoded data for the target leaf CU using a combination of the loop variable pu_param_id and the loop variable tu_param_id corresponding to the optimum rate distortion determined in step S413. That is, the leaf CU encoding unit 243 generates encoded data for the target leaf CU using a division pattern for dividing the target leaf CU into partitions and blocks and optimizing rate distortion.
 このように、本実施形態に係る動画像符号化装置2は、予測単位毎に予測モード群から選択した予測モードに従って生成した予測画像を原画像から減算して得られる予測残差を、選択した予測モードを指定する予測モード指定情報その他のサイド情報と共に符号化することによって符号化データを生成する画像符号化装置であって、対象予測単位に関する予測モード群に追加する予測モード数Nを、上記サイド情報に含まれる符号化パラメータであって、上記対象予測単位を包含する符号化単位、または、該符号化単位よりも先に符号化された符号化単位に関する符号化パラメータを参照して設定する設定手段(PU決定部531)と、上記対象予測単位に関する予測モード群に追加するN個の予測モードの各々に対応する予測パラメータを局所復号画像の復号済領域から導出する導出手段(予測画像生成部532)と、を備えている。 As described above, the video encoding device 2 according to the present embodiment has selected the prediction residual obtained by subtracting the prediction image generated according to the prediction mode selected from the prediction mode group for each prediction unit from the original image. An image encoding apparatus that generates encoded data by encoding together with prediction mode specifying information for specifying a prediction mode and other side information, and the number N of prediction modes to be added to the prediction mode group related to the target prediction unit, The encoding parameter included in the side information is set with reference to an encoding unit including the target prediction unit or an encoding parameter related to an encoding unit encoded before the encoding unit. Prediction parameters corresponding to each of the N prediction modes to be added to the setting unit (PU determination unit 531) and the prediction mode group related to the target prediction unit Includes a deriving means for deriving from the decoded region of the local decoded image (predicted image generation unit 532), the.
 予測単位には、予測モード群に属する予測モード数を増やして予測精度を上げる方が符号化効率が向上するタイプと、予測モード群に属する予測モード数を減らして符号量を下げる方が符号化効率が向上するタイプとがある。発明者らが得た知見によれば、ある予測単位が前者のタイプに属するのか後者のタイプに属するのかは、普通、その予測単位、又は、その符号化単位よりも先に復号される符号化単位に関する符号化パラメータから判定することができる。 The prediction unit is a type that increases the prediction accuracy by increasing the number of prediction modes belonging to the prediction mode group, and the coding efficiency is improved by reducing the number of prediction modes belonging to the prediction mode group and improving the coding efficiency. Some types improve efficiency. According to the knowledge obtained by the inventors, whether a certain prediction unit belongs to the former type or the latter type is usually encoded that is decoded before the prediction unit or the encoding unit. It can be determined from the coding parameters for the unit.
 上記の構成によれば、対象予測単位を包含する符号化単位、又は、その符号化単位よりも先に復号される符号化単位に関する符号化パラメータを参照することによって、符号化効率を大きくするように(望ましくは符号化効率を最大化するように)、対象予測単位に関する予測モード群に追加する予測モード数を設定することができる。したがって、上記の構成によれば、高い予測精度を維持しつつ、符号化データの符号量の増大を抑制することによって、符号化効率を向上させることができる。 According to the above configuration, it is possible to increase the coding efficiency by referring to the coding parameter relating to the coding unit including the target prediction unit or the coding unit decoded prior to the coding unit. In addition, it is possible to set the number of prediction modes to be added to the prediction mode group related to the target prediction unit (desirably to maximize the coding efficiency). Therefore, according to said structure, encoding efficiency can be improved by suppressing the increase in the code amount of coding data, maintaining high prediction accuracy.
 <実施形態1の変形例>
 上記の例では、対象パーティションについて選択可能な基本予測モードの総数を規定値とし、動画像復号装置1の備えるPU復号部431、および、動画像符号化装置2の備えるPU決定部531が、対象パーティションについて選択可能な副方向の総数を、当該対象パーティションのサイズに応じて変更する場合について説明したが、本実施形態はこれに限定されるものではない。
<Modification of Embodiment 1>
In the above example, the total number of basic prediction modes that can be selected for the target partition is a specified value, and the PU decoding unit 431 included in the video decoding device 1 and the PU determination unit 531 included in the video encoding device 2 are Although the case where the total number of sub-directions that can be selected for a partition is changed according to the size of the target partition has been described, the present embodiment is not limited to this.
 例えば、各パーティションについて選択可能な基本予測モードの総数を、該パーティションのサイズに応じて変更する構成とすることもできる。具体的には、32×32画素のパーティションについて選択可能な基本予測モードの総数を9とし、8×8画素のパーティションについて選択可能な基本予測モードの総数を33とするような構成も可能である。このような構成の場合、基本予測モードの総数を示すフラグを符号化データに含めておき、動画像復号装置1は、該フラグを参照して基本予測モードを復号する構成とすることができる。 For example, the total number of basic prediction modes that can be selected for each partition may be changed according to the size of the partition. Specifically, the total number of basic prediction modes that can be selected for a partition of 32 × 32 pixels is set to 9, and the total number of basic prediction modes that can be selected for a partition of 8 × 8 pixels is set to 33. . In such a configuration, a flag indicating the total number of basic prediction modes may be included in the encoded data, and the video decoding device 1 may be configured to decode the basic prediction mode with reference to the flags.
 また、上記のような構成とする場合、動画像復号装置1の備えるPU復号部431、および、動画像符号化装置2の備えるPU決定部531は、対象パーティションについて選択可能な副方向の総数を、当該対象パーティションについて選択可能な基本予測モードの総数に応じて変更する構成とすることもできる。例えば、対象パーティションについて選択可能な基本予測モードの数が9である場合、当該対象パーティションについて選択可能な副方向の総数を0とし、対象パーティションについて選択可能な基本予測モードの数が33である場合、当該対象パーティションについて選択可能な副方向の総数を8とするような構成も可能である。 In the case of the above configuration, the PU decoding unit 431 included in the video decoding device 1 and the PU determination unit 531 included in the video encoding device 2 determine the total number of sub-directions that can be selected for the target partition. The configuration may be changed in accordance with the total number of basic prediction modes that can be selected for the target partition. For example, when the number of basic prediction modes that can be selected for the target partition is 9, the total number of sub-directions that can be selected for the target partition is 0, and the number of basic prediction modes that can be selected for the target partition is 33 A configuration is also possible in which the total number of sub-directions that can be selected for the target partition is 8.
 より一般には、PU決定部531、対象パーティションについて選択可能な基本予測モードの総数がNba1であるときに、当該対象パーティションについて選択可能な副方向の総数をNad1に決定し、対象パーティションについて選択可能な基本予測モードの総数がNba2(Nba1<Nba2)であるときに、当該対象パーティションについて選択可能な補正方向の総数をNad2(Nad1<Nad2)に決定する。 More generally, when the total number of basic prediction modes that can be selected for the target partition is Nba1, the total number of sub-directions that can be selected for the target partition is determined as Nad1, and the target partition can be selected. When the total number of basic prediction modes is Nba2 (Nba1 <Nba2), the total number of correction directions that can be selected for the target partition is determined as Nad2 (Nad1 <Nad2).
 上記の構成をとることによって、基本予測モードの総数が小さいパーティションについてのサイド情報の増加を防ぐことができるので、符号化データ#1の符号量が削減され、符号化効率が向上する。 By adopting the above configuration, it is possible to prevent an increase in side information for a partition having a small total number of basic prediction modes, so that the amount of encoded data # 1 is reduced and the encoding efficiency is improved.
 <実施形態1の変形例2>
 また、動画像復号装置1の備えるPU復号部431、および、動画像符号化装置2の備えるPU決定部531は、対象パーティションに隣接するパーティションであって、復号済みのパーティションについて、エッジベース予測モードが適用されているか否かに応じて、当該対象パーティションについて選択可能な副方向の総数を変更する構成としてもよい。
<Modification 2 of Embodiment 1>
The PU decoding unit 431 included in the video decoding device 1 and the PU determination unit 531 included in the video encoding device 2 are the partitions adjacent to the target partition, and the edge-based prediction mode for the decoded partition Depending on whether or not is applied, the total number of sub-directions that can be selected for the target partition may be changed.
 例えば、動画像復号装置1の備えるPU復号部431、および、動画像符号化装置2の備えるPU決定部531は、対象パーティションに隣接する復号済みのパーティションの少なくとも何れかについて、エッジベース予測モードが適用されている場合、対象パーティションについて選択可能な副方向の総数を2とし、対象パーティションに隣接する復号済みのパーティションの何れにも、エッジベース予測モードが適用されていない場合、対象パーティションについて選択可能な副方向の総数を4とする構成としてもよい。 For example, the PU decoding unit 431 included in the video decoding device 1 and the PU determination unit 531 included in the video encoding device 2 have an edge-based prediction mode for at least one of the decoded partitions adjacent to the target partition. If applied, the total number of sub-directions that can be selected for the target partition is 2. If the edge-based prediction mode is not applied to any of the decoded partitions adjacent to the target partition, the target partition can be selected. The total number of sub-directions may be four.
 図17は、対象パーティションの左辺に隣接する隣接パーティションについてエッジベース予測モードが適用されている場合を示す図である。図17に示すように、隣接パーティションについてエッジベース予測モードが適用されている場合には、隣接パーティションについてエッジベース予測モードを用いて導出された予測方向が、対象パーティションについても適切な予測方向であるという傾向がある。 FIG. 17 is a diagram illustrating a case where the edge-based prediction mode is applied to an adjacent partition adjacent to the left side of the target partition. As shown in FIG. 17, when the edge-based prediction mode is applied to the adjacent partition, the prediction direction derived using the edge-based prediction mode for the adjacent partition is also an appropriate prediction direction for the target partition. There is a tendency.
 図17に示す例においては、動画像復号装置1の備えるPU復号部431、および、動画像符号化装置2の備えるPU決定部531は、対象パーティションついて選択可能な副方向数の数を、対象パーティションに隣接するパーティションの何れについてもエッジベース予測モードが適用されていない場合に比べて減らす構成とすることができる。 In the example illustrated in FIG. 17, the PU decoding unit 431 included in the video decoding device 1 and the PU determination unit 531 included in the video encoding device 2 determine the number of sub-directions that can be selected for the target partition. The configuration can be reduced compared to the case where the edge-based prediction mode is not applied to any partition adjacent to the partition.
 一般に、エッジベース予測モードが選択されるパーティションには、比較的強いエッジが存在する場合が多い。また、そのような強いエッジは、複数のパーティションに跨って連続的に存在している場合が多い。したがって、対象パーティションに隣接する隣接パーティションについてエッジベース予測モードを用いて導出された予測方向は、対象パーティションについても適切な予測方向であるという傾向がある。また、このような場合、対象パーティションについて導出される主方向は、隣接パーティションについてエッジベース予測モードを用いて導出された予測方向に近いという傾向がある。 In general, there are often relatively strong edges in a partition for which the edge-based prediction mode is selected. Such strong edges often exist continuously across a plurality of partitions. Therefore, the prediction direction derived using the edge-based prediction mode for the adjacent partition adjacent to the target partition tends to be an appropriate prediction direction for the target partition. In such a case, the main direction derived for the target partition tends to be close to the prediction direction derived using the edge-based prediction mode for the adjacent partition.
 したがって、隣接パーティションについてエッジベース予測モードが選択されている場合には、対象パーティションについて副方向エッジベース予測モードが選択される可能性が低いという傾向がある。 Therefore, when the edge-based prediction mode is selected for the neighboring partition, there is a low possibility that the sub-direction edge-based prediction mode is selected for the target partition.
 上記の構成をとることによって、選択される可能性の低い副方向エッジベース予測モードについての副方向数を減らすことができるので、高い予測精度を維持しつつ、符号化データ#1の符号量を削減することができるので、符号化効率を向上させることができる。 By adopting the above configuration, the number of sub-directions for the sub-direction edge-based prediction mode that is unlikely to be selected can be reduced, so that the code amount of the encoded data # 1 can be reduced while maintaining high prediction accuracy. Since it can be reduced, encoding efficiency can be improved.
 <実施形態1の変形例3>
 また、動画像復号装置1の備えるPU復号部431、および、動画像符号化装置2の備えるPU決定部531は、対象パーティションに隣接する復号済みのパーティションについて、エッジベース予測モードが適用されているか否かに応じて、副方向を規定するパラメータδの値を変更する構成としてもよい。
<Modification 3 of Embodiment 1>
Whether the PU decoding unit 431 included in the video decoding device 1 and the PU determination unit 531 included in the video encoding device 2 are applied with the edge-based prediction mode for the decoded partition adjacent to the target partition. Depending on whether or not, the value of the parameter δ that defines the sub-direction may be changed.
 例えば、動画像復号装置1の備えるPU復号部431、および、動画像符号化装置2の備えるPU決定部531は、対象パーティションに隣接する復号済みの隣接パーティションについてエッジベース予測モードが適用されている場合のパラメータδの値を、隣接パーティションについてエッジベース予測モードが適用されていない場合のパラメータδの値よりも小さく設定する構成としてもよい。 For example, the PU decoding unit 431 included in the video decoding device 1 and the PU determination unit 531 included in the video encoding device 2 are applied with the edge-based prediction mode for the decoded adjacent partitions adjacent to the target partition. In this case, the value of the parameter δ may be set smaller than the value of the parameter δ when the edge-based prediction mode is not applied to the adjacent partition.
 一般に、エッジベース予測モードが選択されるパーティションには、比較的強いエッジが存在する場合が多い。また、そのような強いエッジは、複数のパーティションに跨って連続的に存在している場合が多い。したがって、対象パーティションに隣接する隣接パーティションについてエッジベース予測モードを用いて導出された予測方向は、対象パーティションについても適切な予測方向に近いという傾向がある。また、このような場合、対象パーティションについて導出される主方向は、対象パーティションについての最適な予測方向に近いという傾向がある。 In general, there are often relatively strong edges in a partition for which the edge-based prediction mode is selected. Such strong edges often exist continuously across a plurality of partitions. Therefore, the prediction direction derived using the edge-based prediction mode for the adjacent partition adjacent to the target partition tends to be close to an appropriate prediction direction for the target partition. In such a case, the main direction derived for the target partition tends to be close to the optimum prediction direction for the target partition.
 上記の構成をとることによって、対象パーティションについて導出される副方向が、当該対象パーティションについての最適な予測方向に一致する可能性が高まるので、予測精度が向上し、符号化効率が向上する。 The above configuration increases the possibility that the sub-direction derived for the target partition matches the optimal prediction direction for the target partition, so that the prediction accuracy is improved and the encoding efficiency is improved.
 <実施形態1の変形例4>
 また、動画像復号装置1の備えるPU復号部431、および、動画像符号化装置2の備えるPU決定部531は、対象パーティションにて選択可能な補正方向数を当該対象パーティションのパーティションサイズに基づいて決定しているが、対象パーティションを包含するリーフCUのサイズに基づいて補正方向数を決定してもよい。その場合、補正方向数の決定処理の実行回数が削減でき、復号処理または符号化処理の処理量が軽減される。
<Modification 4 of Embodiment 1>
The PU decoding unit 431 included in the video decoding device 1 and the PU determination unit 531 included in the video encoding device 2 determine the number of correction directions that can be selected in the target partition based on the partition size of the target partition. Although determined, the number of correction directions may be determined based on the size of the leaf CU including the target partition. In that case, the number of executions of the determination process of the number of correction directions can be reduced, and the processing amount of the decoding process or the encoding process is reduced.
 <実施形態1の変形例5>
 また、動画像復号装置1の備えるPU復号部431、および、動画像符号化装置2の備えるPU決定部531は、対象パーティションのパーティションサイズと、対象パーティションを包含するリーフCUのサイズを併用して補正方向数を決定してもよい。例えば、パーティションサイズが同一の場合でも、当該パーティションが属するリーフCUのサイズが異なる場合には、異なる個数の補正方向数を選択可能としてもよい。その場合、パーティションサイズが同一の領域であっても、異なる補正方向数を追加できるようになるため、領域毎により適切な副方向数が設定できるようになり、符号化効率が向上する。
<Modification 5 of Embodiment 1>
Further, the PU decoding unit 431 included in the video decoding device 1 and the PU determination unit 531 included in the video encoding device 2 use the partition size of the target partition and the size of the leaf CU including the target partition in combination. The number of correction directions may be determined. For example, even when the partition size is the same, if the size of the leaf CU to which the partition belongs is different, a different number of correction directions may be selected. In this case, even if the regions have the same partition size, a different number of correction directions can be added, so that an appropriate number of sub-directions can be set for each region, and the coding efficiency is improved.
 <実施形態1の変形例5>
 また、4×4画素のパーティションについて選択可能な基本予測モードの総数を18個(方向予測モード:16個、DC予測モード:1個、Planar予測モード:1個)とし、8×8画素、16×16画素、及び32×32画素のパーティションについて選択可能な基本予測モードの総数を35個(方向予測モード:33個、DC予測モード:1個、Planar予測モード:1個)とする構成としてもよい。
<Modification 5 of Embodiment 1>
The total number of selectable basic prediction modes for a 4 × 4 pixel partition is 18 (direction prediction mode: 16, DC prediction mode: 1, Planar prediction mode: 1), and 8 × 8 pixels, 16 The total number of selectable basic prediction modes for a partition of × 16 pixels and 32 × 32 pixels is 35 (direction prediction mode: 33, DC prediction mode: 1, Planar prediction mode: 1). Good.
 また、基本予測モードの総数を本変形例のように設定する場合には、4×4画素及び8×8画素のパーティションに対しては、エッジベース予測の副方向数を0個に設定し、16×16画素及び32×32画素のパーティションでは副方向数を4個に設定することが好ましい。 When the total number of basic prediction modes is set as in this modification, the number of sub-directions for edge-based prediction is set to 0 for 4 × 4 pixel and 8 × 8 pixel partitions, In the partition of 16 × 16 pixels and 32 × 32 pixels, it is preferable to set the number of sub-directions to four.
 なお、副方向数が0個である場合には、PU復号部431によるPU情報復号処理において、追加インデックスAIの復号を省略することができる。すなわち、図1のステップS108における追加インデックスの復号処理が省略され、主方向予測モードが対象パーティションについての予測モードとして設定される。 If the number of sub-directions is 0, decoding of the additional index AI can be omitted in the PU information decoding process by the PU decoding unit 431. That is, the decoding process of the additional index in step S108 in FIG. 1 is omitted, and the main direction prediction mode is set as the prediction mode for the target partition.
 以上の構成とすることによって、基本予測モードの総数が小さいパーティションについてのサイド情報の増加を防ぐことができるので、符号化データ#1の符号量が削減され、符号化効率が向上する。 With the above configuration, it is possible to prevent an increase in side information for a partition having a small total number of basic prediction modes, so that the amount of encoded data # 1 is reduced and the encoding efficiency is improved.
 なお、上記の説明において、Planar予測モード(プレーン予測モード、平面予測モードとも呼ばれる)とは、対象パーティションにおける予測画像を周辺のパーティションの画素値から平面的に予測するモードのことを指し、具体的には、対象パーティションにおける予測画像の画素値predSamples[x,y](ここで、[x,y]は画素の座標を表す)を、例えば以下の数式によって算出するモードのことを指す。 In the above description, the Planar prediction mode (also referred to as plane prediction mode or plane prediction mode) refers to a mode in which a prediction image in the target partition is predicted two-dimensionally from pixel values of surrounding partitions. Indicates a mode in which the pixel value predSamples [x, y] (where [x, y] represents the coordinates of the pixel) of the predicted image in the target partition is calculated by, for example, the following equation.
 predSamples[x,y] = ((nS-1-x) * p[-1,y] + (x+1) * p[nS-1,-1] 
+ (nS-1-y) * p[x,-1] + (y+1) * p[-1,nS-1] + nS) >> (k+1)
ここで、x, y = 0...nS-1 であり、k = log2(nS)である。また、nS は、対象パーティションの一辺の画素数を表している。
predSamples [x, y] = ((nS-1-x) * p [-1, y] + (x + 1) * p [nS-1, -1]
+ (nS-1-y) * p [x, -1] + (y + 1) * p [-1, nS-1] + nS) >> (k + 1)
Here, x, y = 0 ... nS-1 and k = log 2 (nS). NS represents the number of pixels on one side of the target partition.
 〔実施形態2〕
 実施形態1においては、エッジベース予測モードにて導出される主方向の数は各パーティションについて1つであったが、符号化対象画像の画像特性如何によっては、各パーティションについて複数の主方向を導出可能な構成としてもよい。また、対象パーティションのサイズに応じて、導出可能な主方向の総数を変更し、複数の主方向が導出可能な場合には、当該複数の主方向から何れかの主方向を選択して用いる構成としてもよい。
[Embodiment 2]
In the first embodiment, the number of main directions derived in the edge-based prediction mode is one for each partition. However, depending on the image characteristics of the encoding target image, a plurality of main directions are derived for each partition. It is good also as a possible structure. Also, the total number of main directions that can be derived is changed according to the size of the target partition, and when a plurality of main directions can be derived, one of the main directions is selected and used. It is good.
 以下では、本発明の第2の実施形態として、対象パーティションのサイズに応じて、導出可能な主方向の総数を変更し、複数の主方向が導出可能な場合には、当該複数の主方向から何れかの主方向を選択して用いる構成について説明する。なお、以下では、すでに説明した部分と同じ部分については同じ符号を付し、その説明を省略する。 Hereinafter, as the second embodiment of the present invention, the total number of main directions that can be derived is changed according to the size of the target partition, and when a plurality of main directions can be derived, the plurality of main directions are used. A configuration for selecting and using any of the main directions will be described. In the following, the same parts as those already described are denoted by the same reference numerals, and the description thereof is omitted.
 (符号化データの構成)
 本実施形態に係る動画像符号化装置4によって生成され、本実施形態に係る動画像復号装置3によって復号される符号化データは、実施形態1における符号化データ#1とほぼ同様の構成であるが、以下の点において異なる。
(Configuration of encoded data)
The encoded data generated by the moving image encoding device 4 according to the present embodiment and decoded by the moving image decoding device 3 according to the present embodiment has substantially the same configuration as the encoded data # 1 in the first embodiment. However, the following points are different.
 すなわち、本実施形態に係る符号化データにおいては、所定のサイズ以上のパーティションについてのイントラ予測パラメータPP_Intraが、複数の主方向の何れかを指定する主方向指定インデックスMDIを含んでいる。 That is, in the encoded data according to the present embodiment, the intra prediction parameter PP_Intra for a partition of a predetermined size or larger includes a main direction designation index MDI that designates one of a plurality of main directions.
 また、本実施形態に係る符号化データは、実施形態1において説明した追加インデックスAIを含む構成としてもよいし、追加インデックスAIを含まない構成としてもよい。 Also, the encoded data according to the present embodiment may include the additional index AI described in the first embodiment, or may not include the additional index AI.
 (動画像復号装置)
 以下では、本実施形態に係る動画像復号装置3について説明する。本実施形態に係る動画像復号装置3は、実施形態1に係る動画像復号装置1の備えるPU復号部431、および、予測画像生成部432に代えて、それぞれ、PU復号部431’、および、予測画像生成部432’を備えている。図18に、本実施形態に係る動画像復号装置の備えるリーフCU復号部の構成を示す。本実施形態に係る動画像復号装置のその他の構成は、実施形態1に係る動画像復号装置1の構成と同様である。
(Video decoding device)
Below, the moving image decoding apparatus 3 which concerns on this embodiment is demonstrated. The video decoding device 3 according to the present embodiment replaces the PU decoding unit 431 and the predicted image generation unit 432 included in the video decoding device 1 according to Embodiment 1, respectively, with the PU decoding unit 431 ′ and A predicted image generation unit 432 ′ is provided. FIG. 18 illustrates a configuration of a leaf CU decoding unit included in the video decoding device according to the present embodiment. Other configurations of the video decoding device according to the present embodiment are the same as the configurations of the video decoding device 1 according to the first embodiment.
 (PU復号部431’)
 PU復号部431’は、対象リーフCUについてのPU情報を復号することによって、対象リーフCUの各パーティションへの分割パターンを決定すると共に、各パーティションについての予測モードを決定し、決定された各予測モード#431を、各パーティションに割り付ける。各パーティションに割り付けられた予測モード#431は、予測画像生成部432’に対して供給される。ここで、予測モードとは、実施形態1にて説明したように、画面内予測(イントラ予測)によって予測画像を生成する生成方法を指定するものであり、本実施形態においては、各パーティションについて、基本予測モードセット、または、エッジベース予測モードセットから何れかの予測モードが選択される。
(PU decoding unit 431 ′)
The PU decoding unit 431 ′ decodes the PU information about the target leaf CU, thereby determining a division pattern for each partition of the target leaf CU and determining a prediction mode for each partition. Mode # 431 is assigned to each partition. The prediction mode # 431 assigned to each partition is supplied to the predicted image generation unit 432 ′. Here, as described in the first embodiment, the prediction mode designates a generation method for generating a prediction image by intra prediction (intra prediction). In the present embodiment, for each partition, Any prediction mode is selected from the basic prediction mode set or the edge-based prediction mode set.
 また、PU復号部431’は、対象パーティションのサイズを参照して、当該対象パーティションについて導出可能な主方向の数を決定する。また、PU復号部431’は、対象パーティションについて複数の主方向が導出可能な場合には、当該複数の主方向の何れを指定する主方向指定インデックスを復号する。また、PU復号部431’は、当該主方向インデックスによって指定される主方向を、対象パーティションについての予測方向に設定する。 Also, the PU decoding unit 431 'refers to the size of the target partition and determines the number of main directions that can be derived for the target partition. In addition, when a plurality of main directions can be derived for the target partition, the PU decoding unit 431 ′ decodes a main direction designation index that designates any of the plurality of main directions. Also, the PU decoding unit 431 'sets the main direction specified by the main direction index as the prediction direction for the target partition.
 以下では、図19~20を参照して、PU復号部431’によるPU情報復号処理について説明する。図19は、PU復号部431’によるPU情報復号処理の流れを示すフローチャートである。図20は、対象パーティションについて取り得るエッジベース予測フラグEFの値、主方向指定インデックスMDIの値、推定フラグMPMの値、および、残余予測モードインデックスRIPMの値を示す表である。 Hereinafter, the PU information decoding process by the PU decoding unit 431 ′ will be described with reference to FIGS. 19 to 20. FIG. 19 is a flowchart showing a flow of PU information decoding processing by the PU decoding unit 431 '. FIG. 20 is a table showing the value of the edge-based prediction flag EF, the value of the main direction designation index MDI, the value of the estimation flag MPM, and the value of the residual prediction mode index RIPM that can be taken for the target partition.
 (ステップS501)~(ステップS506)、および、(ステップS510)
 図19に示すステップS501~S506、および、S510は、それぞれ、実施形態1において説明したPU復号部431による復号処理のステップS101~S106、および、S112と同様であるので、説明を省略する。
(Step S501) to (Step S506) and (Step S510)
Steps S501 to S506 and S510 shown in FIG. 19 are the same as steps S101 to S106 and S112 of the decoding process performed by the PU decoding unit 431 described in the first embodiment, and thus description thereof is omitted.
 (ステップS507)
 PU復号部431’は、対象パーティションについてエッジベース予測モードが適用される場合、PU復号部431’は、対象パーティションのサイズに応じて、当該対象パーティションについて導出する主方向の総数を決定する。ここで、当該対象パーティションについて導出する主方向の総数は、対象パーティションのサイズと負の相関を有することが好ましい。
(Step S507)
When the edge-based prediction mode is applied to the target partition, the PU decoding unit 431 ′ determines the total number of main directions derived for the target partition according to the size of the target partition. Here, the total number of main directions derived for the target partition preferably has a negative correlation with the size of the target partition.
 例えば、PU復号部431’は、対象パーティションのサイズが4×4画素である場合、対象パーティションについて導出する主方向の総数を2に設定し、対象パーティションのサイズが8×8画素である場合、対象パーティションについて導出する主方向の総数を1に設定する。より一般には、PU復号部431’は、対象パーティションに属する画素の総数がNpix1であるときに、当該対象パーティションについて選択可能な主方向の総数をNmain1に設定し、対象パーティションに属する画素の総数がNpix2(Npix1<Npix2)であるときに、当該対象パーティションについて選択可能な主方向の総数をNmain2(Nmain1≧Nmain2)に設定する。 For example, when the size of the target partition is 4 × 4 pixels, the PU decoding unit 431 ′ sets the total number of main directions derived for the target partition to 2, and when the size of the target partition is 8 × 8 pixels, The total number of main directions derived for the target partition is set to 1. More generally, when the total number of pixels belonging to the target partition is Npix1, the PU decoding unit 431 ′ sets the total number of main directions that can be selected for the target partition to Nmain1, and the total number of pixels belonging to the target partition is When Npix2 (Npix1 <Npix2), the total number of main directions that can be selected for the target partition is set to Nmain2 (Nmain1 ≧ Nmain2).
 なお、PU復号部431’は、符号化データに含まれる分割情報であって、対象パーティションを含む対象LCUの各リーフCUへの分割パターンを指定するCU分割情報SP_CU、および、対象PU(対象リーフCU)の各パーティションへの分割パターンを指定するイントラPU分割情報SP_Intraを参照することによって、対象パーティションのサイズを識別することができる。 Note that the PU decoding unit 431 ′ is the division information included in the encoded data, and includes CU division information SP_CU that specifies a division pattern for each leaf CU of the target LCU including the target partition, and the target PU (target leaf). The size of the target partition can be identified by referring to the intra PU partition information SP_Intra that specifies the partition pattern of each partition of (CU).
 (ステップS508)
 続いて、PU復号部431’は、対象パーティションについての主方向指定インデックスMDIを復号する。ここで、主方向指定インデックスMDIは、対象パーティションについて導出される複数の主方向の何れかを指定するインデックスである。例えば、図20に示すように、導出される主方向の総数が2つ(主方向1、主方向2)である場合には、主方向指定インデックスMDIは、当該2つの主方向の何れかを指定する。より具体的には、主方向1は、主方向指定インデックス=0によって指定され、主方向2は、主方向指定インデックス=1によって指定される。なお、主方向1、および、主方向2の具体例については、後述するためここでは説明を省略する。
(Step S508)
Subsequently, the PU decoding unit 431 ′ decodes the main direction designation index MDI for the target partition. Here, the main direction designation index MDI is an index for designating any of a plurality of main directions derived for the target partition. For example, as shown in FIG. 20, when the total number of derived main directions is two (main direction 1, main direction 2), the main direction designation index MDI indicates one of the two main directions. specify. More specifically, the main direction 1 is designated by a main direction designation index = 0, and the main direction 2 is designated by a main direction designation index = 1. In addition, since the specific example of the main direction 1 and the main direction 2 is mentioned later, description is abbreviate | omitted here.
 (ステップS509)
 続いて、PU復号部431’は、対象パーティションについての予測モードを、ステップS508にて復号された主方向指定インデックスの指定する主方向を予測方向とするエッジベース予測モードに設定する。
(Step S509)
Subsequently, the PU decoding unit 431 ′ sets the prediction mode for the target partition to the edge-based prediction mode in which the main direction specified by the main direction specifying index decoded in step S508 is the prediction direction.
 以上が、PU復号部431’によるPU情報復号処理の流れである。ステップS506、および、ステップS509にて対象リーフCUに含まれる各パーティションについて設定された予測モードは、予測モード#431’として予測画像生成部432’に供給される。 The above is the flow of PU information decoding processing by the PU decoding unit 431 '. The prediction mode set for each partition included in the target leaf CU in step S506 and step S509 is supplied to the predicted image generation unit 432 'as a prediction mode # 431'.
 なお、符号化データに追加インデックスAIが含まれている場合には、PU復号部431’は、当該追加インデックスAIを復号し、対象パーティションについての予測モードを、ステップS508にて指定される主方向に追加インデックスAIによって指定される補正角度を加算して得られる副方向を予測方向とするエッジベース予測モードに設定する。 If the encoded data includes the additional index AI, the PU decoding unit 431 ′ decodes the additional index AI and sets the prediction mode for the target partition in the main direction specified in step S508. Is set to the edge-based prediction mode in which the sub-direction obtained by adding the correction angle specified by the additional index AI is the prediction direction.
 (予測画像生成部432’)
 予測画像生成部432’は、対象リーフCUに含まれる各パーティションについての予測画像を、PU復号部431’より供給される予測モード#431’、および、該パーティションの周辺の復号済み画素値を参照して、生成する。
(Predicted image generation unit 432 ′)
The prediction image generation unit 432 ′ refers to the prediction image for each partition included in the target leaf CU with reference to the prediction mode # 431 ′ supplied from the PU decoding unit 431 ′ and the decoded pixel values around the partition And generate.
 以下では、図21を参照して、予測画像生成部432’による、対象リーフCUについての予測画像生成処理について説明する。 Hereinafter, a predicted image generation process for the target leaf CU by the predicted image generation unit 432 ′ will be described with reference to FIG. 21.
 図21は、予測画像生成部432’による、対象リーフCUについての予測画像生成処理の流れを示すフローチャートである。 FIG. 21 is a flowchart showing the flow of predicted image generation processing for the target leaf CU by the predicted image generation unit 432 '.
 (ステップS601)~(ステップS605)、および、(ステップS609)
 図21に示すステップS601~S605、および、S609は、それぞれ、実施形態1において説明した予測画像生成部432による予測画像生成処理のステップS201~S205、および、S212と同様であるので、説明を省略する。
(Step S601) to (Step S605), and (Step S609)
Steps S601 to S605 and S609 shown in FIG. 21 are the same as steps S201 to S205 and S212 of the predicted image generation processing by the predicted image generation unit 432 described in the first embodiment, and thus description thereof is omitted. To do.
 (ステップS606)
 対象パーティションについての予測モードがエッジベース予測モードである場合、予測画像生成部432’は、主方向指定インデックスによって指定される主方向(例えば、主方向インデックス=0によって指定される主方向1、または、主方向インデックス=1によって指定される主方向2)を、対象パーティションの周辺に設定された参照領域に属する復号済み画素値(参照画素の画素値)に基づいて導出する。
(Step S606)
When the prediction mode for the target partition is the edge-based prediction mode, the predicted image generation unit 432 ′ uses the main direction specified by the main direction specifying index (for example, the main direction 1 specified by the main direction index = 0, or The main direction 2 specified by the main direction index = 1 is derived based on the decoded pixel values (pixel values of the reference pixels) belonging to the reference area set around the target partition.
 (ステップS607)
 続いて、予測画像生成部432’は、導出した主方向を対象パーティションについての予測方向に設定する。
(Step S607)
Subsequently, the predicted image generation unit 432 ′ sets the derived main direction as the predicted direction for the target partition.
 (ステップS608)
 予測画像生成部432’は、対象パーティションについての予測画像を、当該対象パーティションについて設定された予測方向に沿って、当該対象パーティションの周辺の復号済み画素値を外挿することによって生成する。本ステップの処理は、実施形態1において説明したステップS211と同様である。
(Step S608)
The predicted image generation unit 432 ′ generates a predicted image for the target partition by extrapolating the decoded pixel values around the target partition along the prediction direction set for the target partition. The processing in this step is the same as that in step S211 described in the first embodiment.
 予測画像生成部432’は、以上の処理を行うことによって対象リーフCUについての予測画像を生成する。また、生成された対象リーフCUについての予測画像は、予測画像#432’として復号画像生成部435に供給される。 The predicted image generation unit 432 'generates a predicted image for the target leaf CU by performing the above processing. Also, the generated predicted image for the target leaf CU is supplied to the decoded image generation unit 435 as a predicted image # 432 '.
 なお、符号化データに追加インデックスAIが含まれている場合には、予測画像生成部432’は、PU復号部431’にて指定される主方向に追加インデックスAIによって指定される補正角度を加算して得られる予測方向に沿って、対象パーティションの周辺の復号済み画素値を外挿することによって、当該対象パーティションについての予測画像を生成する。 If the encoded data includes the additional index AI, the predicted image generation unit 432 ′ adds the correction angle specified by the additional index AI to the main direction specified by the PU decoding unit 431 ′. The predicted image for the target partition is generated by extrapolating the decoded pixel values around the target partition along the prediction direction obtained as described above.
 (主方向1および主方向2の具体例)
 以下では、上記ステップS606にて予測画像生成部432’により導出される主方向1および主方向2の具体例について説明する。
(Specific examples of main direction 1 and main direction 2)
Hereinafter, specific examples of the main direction 1 and the main direction 2 derived by the predicted image generation unit 432 ′ in step S606 will be described.
 (具体例1)
 (主方向1)
 予測画像生成部432’は、対象パーティションの周辺の復号済み領域に設定された参照領域に属する画素について算出されるエッジベクトルのうち、ノルムが最大であるエッジベクトルの指す方向を、主方向1として導出する。ここで、参照領域に属する各画素についてのエッジベクトルは、実施形態1において説明したように、例えば、ソーベルフィルタを作用させることによって算出することができる。
(Specific example 1)
(Main direction 1)
The predicted image generation unit 432 ′ sets, as the main direction 1, the direction indicated by the edge vector having the maximum norm among the edge vectors calculated for the pixels belonging to the reference area set in the decoded area around the target partition. To derive. Here, as described in the first embodiment, the edge vector for each pixel belonging to the reference region can be calculated, for example, by applying a Sobel filter.
 (主方向2)
 予測画像生成部432’は、対象パーティションの周辺の復号済みパーティションに割り付けられた予測方向を参照して、主方向2を導出する。
(Main direction 2)
The predicted image generation unit 432 ′ derives the main direction 2 with reference to the predicted direction assigned to the decoded partitions around the target partition.
 例えば、対象パーティションに隣接する複数の隣接パーティションの何れについてもエッジベース予測モードが適用されていた場合、予測画像生成部432’は、当該複数の隣接パーティションについて導出された主方向(または主方向に補正方向を加算して得られる副方向)の平均をとることによって当該対象パーティションについての主方向2を導出する。 For example, when the edge-based prediction mode is applied to any of a plurality of adjacent partitions adjacent to the target partition, the predicted image generation unit 432 ′ determines the main direction (or main direction) derived for the plurality of adjacent partitions. The main direction 2 for the target partition is derived by taking the average of the sub-directions obtained by adding the correction directions.
 また、対象パーティションに隣接する複数の隣接パーティションのうち何れかの隣接パーティションについてエッジベース予測モードが適用されていた場合、予測画像生成部432’は、当該エッジベース予測モードによって導出された主方向を、対象パーティションについての主方向2に設定する。 Further, when the edge-based prediction mode is applied to any one of a plurality of adjacent partitions adjacent to the target partition, the predicted image generation unit 432 ′ determines the main direction derived by the edge-based prediction mode. The main direction 2 for the target partition is set.
 また、対象パーティションに隣接する複数の隣接パーティションの何れについても基本予測モードが適用されていた場合、当該複数の隣接パーティションについて適用された基本予測モードのうち、予測モードインデックスのより小さい予測モードの示す方向を、対象パーティションについての主方向2に設定する。 In addition, when the basic prediction mode is applied to any of the plurality of adjacent partitions adjacent to the target partition, the prediction mode having a smaller prediction mode index among the basic prediction modes applied to the plurality of adjacent partitions is indicated. The direction is set to the main direction 2 for the target partition.
 (具体例2)
 (主方向1)
 具体例1における主方向1と同様である。
(Specific example 2)
(Main direction 1)
The same as the main direction 1 in the specific example 1.
 (主方向2)
 予測画像生成部432’は、対象パーティションの周辺の復号済み領域に設定された参照領域に属する画素について算出されるエッジベクトルの平均をとることによって、対象パーティションについての主方向2を導出する。
(Main direction 2)
The predicted image generation unit 432 ′ derives the main direction 2 for the target partition by taking the average of the edge vectors calculated for the pixels belonging to the reference region set in the decoded region around the target partition.
 (具体例3)
 (主方向1)
 予測画像生成部432’は、対象パーティションの左辺に隣接する隣接パーティションに属する各画素について算出したエッジベクトルのうち、ノルムが最大であるエッジベクトルの指す方向を、対象パーティションについての主方向1に設定する。
(Specific example 3)
(Main direction 1)
The predicted image generation unit 432 ′ sets the direction indicated by the edge vector having the maximum norm among the edge vectors calculated for each pixel belonging to the adjacent partition adjacent to the left side of the target partition as the main direction 1 for the target partition. To do.
 (主方向2)
 予測画像生成部432’は、対象パーティションの上辺に隣接する隣接パーティションに属する各画素について算出したエッジベクトルのうち、ノルムが最大であるエッジベクトルの指す方向を、対象パーティションについての主方向2に設定する。
(Main direction 2)
The predicted image generation unit 432 ′ sets the direction indicated by the edge vector having the maximum norm among the edge vectors calculated for each pixel belonging to the adjacent partition adjacent to the upper side of the target partition as the main direction 2 for the target partition. To do.
 (具体例4)
 (主方向1)
 予測画像生成部432’は、対象パーティションの周辺の復号済み領域に設定された参照領域に属する画素について算出されるエッジベクトルのうち、ノルムが最大であるエッジベクトルの指す方向を、主方向1として導出する。
(Specific example 4)
(Main direction 1)
The predicted image generation unit 432 ′ sets, as the main direction 1, the direction indicated by the edge vector having the maximum norm among the edge vectors calculated for the pixels belonging to the reference area set in the decoded area around the target partition. To derive.
 (主方向2)
 予測画像生成部432’は、対象パーティションの周辺の復号済み領域に設定された参照領域に属する画素について算出されるエッジベクトルのうち、ノルムが2番目に大きいエッジベクトルの指す方向を、主方向2として導出する。
(Main direction 2)
The predicted image generation unit 432 ′ determines the direction indicated by the edge vector having the second largest norm among the edge vectors calculated for the pixels belonging to the reference region set in the decoded region around the target partition as the main direction 2. Derived as
 (具体例5)
 本具体例においては、予測画像生成部432’は、対象パーティションの周辺に設定された参照領域に属する各参照画素を、2つのクラスタ(クラスタA、クラスタB)に分類し、クラスタAに属する各参照画素について算出されるエッジベクトルのうち、ノルムが最大であるエッジベクトルの指す方向を主方向1に設定し、クラスタBに属する各参照画素について導出されるエッジベクトルのうち、ノルムが最大であるエッジベクトルの指す方向を主方向2に設定する。
(Specific example 5)
In this specific example, the predicted image generation unit 432 ′ classifies each reference pixel belonging to the reference region set around the target partition into two clusters (cluster A and cluster B), and each of the reference pixels belonging to the cluster A Of the edge vectors calculated for the reference pixel, the direction indicated by the edge vector having the maximum norm is set as the main direction 1, and the norm is the maximum among the edge vectors derived for each reference pixel belonging to the cluster B. The direction indicated by the edge vector is set as the main direction 2.
 例えば、予測画像生成部432’は、以下の処理(ステップS701~ステップS703)によって、参照領域に属する各参照画素を各クラスタへ分類することができる。 For example, the predicted image generation unit 432 'can classify each reference pixel belonging to the reference region into each cluster by the following processing (steps S701 to S703).
 (ステップS701)
 まず、予測画像生成部432’は、参照領域に属する各参照画素についてエッジベクトルを算出すると共に、当該エッジベクトルのエッジ強度を算出する。ここで、エッジベクトルの強度は、当該エッジベクトルのノルムによって評価することができる。すなわち、ある参照画素について算出されたエッジベクトルを(dx、dy)と表すとすると、当該エッジベクトルの強度esは、es=((dx)2+(dy)21/2と表すことができる。
(Step S701)
First, the predicted image generation unit 432 ′ calculates an edge vector for each reference pixel belonging to the reference region, and calculates an edge strength of the edge vector. Here, the strength of the edge vector can be evaluated by the norm of the edge vector. That is, if an edge vector calculated for a certain reference pixel is expressed as (dx, dy), the intensity es of the edge vector can be expressed as es = ((dx) 2 + (dy) 2 ) 1/2. it can.
 (ステップS702)
 各参照画素について定義される3次元ベクトル(dx、dy、es)(または当該3次元ベクトルの部分ベクトル)に対してクラスタリング処理を適用し、各参照画素をクラスタAおよびクラスタBの何れかに分類する。上記クラスタリング処理の具体例としては、k-平均法やSVM(Support Vector Machine)などを用いることができる。
(Step S702)
Clustering processing is applied to the three-dimensional vector (dx, dy, es) defined for each reference pixel (or a partial vector of the three-dimensional vector), and each reference pixel is classified into either cluster A or cluster B To do. As a specific example of the clustering process, a k-average method, SVM (Support Vector Machine), or the like can be used.
 (ステップS703)
 続いて、予測画像生成部432’は、クラスタAに分類された各参照画素についてのエッジベクトルのうち、ノルムが最大であるエッジベクトルの指す方向を、対象パーティションについての主方向1に設定する。また、予測画像生成部432’は、クラスタBに分類された各参照画素についてのエッジベクトルのうち、ノルムが最大であるエッジベクトルの指す方向を、対象パーティションについての主方向2に設定する。
(Step S703)
Subsequently, the predicted image generation unit 432 ′ sets the direction indicated by the edge vector having the maximum norm among the edge vectors for each reference pixel classified into cluster A as the main direction 1 for the target partition. Also, the predicted image generation unit 432 ′ sets the direction indicated by the edge vector having the maximum norm among the edge vectors for each reference pixel classified into the cluster B as the main direction 2 for the target partition.
 なお、2つのクラスタのうち、何れのクラスタをクラスタAとするかの判定条件は、本実施形態に係る動画像符号化装置、および、本実施形態に係る動画像復号装置の双方において共有されていることが好ましい。 Note that the condition for determining which of the two clusters is the cluster A is shared by both the video encoding device according to the present embodiment and the video decoding device according to the present embodiment. Preferably it is.
 例えば、本実施形態に係る動画像符号化装置、および、本実施形態に係る動画像復号装置は、より多くの参照画素から構成されているクラスタをクラスタAに設定し、参照領域において、クラスタAに属する参照画素以外の参照画素から構成されているクラスタをクラスタBに設定することができる。 For example, the moving picture encoding apparatus according to the present embodiment and the moving picture decoding apparatus according to the present embodiment set a cluster composed of more reference pixels as cluster A, and in the reference area, cluster A A cluster composed of reference pixels other than the reference pixels belonging to can be set as cluster B.
 図22は、対象パーティションOPに隣接する隣接パーティション、および、対象パーティションの左上の頂点を共有するパーティションよりなる参照領域と、当該参照領域に属する各参照画素が分類されるクラスタA、および、クラスタBを示す図である。 FIG. 22 shows a reference area composed of an adjacent partition adjacent to the target partition OP and a partition sharing the upper left vertex of the target partition, and a cluster A and a cluster B into which each reference pixel belonging to the reference area is classified. FIG.
 図22に示す例では、対象パーティションOP、および、パーティションNP1~3の全ての領域において略垂直方向の弱いエッジが存在する。また、対象パーティションOPと、対象パーティションの左辺に隣接する隣接パーティションNP3とに跨って、略水平方向の強いエッジが存在する。 In the example shown in FIG. 22, weak edges in the substantially vertical direction exist in the target partition OP and all areas of the partitions NP1 to NP3. Further, a strong edge in the substantially horizontal direction exists across the target partition OP and the adjacent partition NP3 adjacent to the left side of the target partition.
 図22に示す例においては、予測画像生成部432’は、上述のように、パーティションNP1~3より構成される参照領域を、上記強いエッジを含む領域であるクラスタA、および、クラスタAと重複しない領域であって上記弱いエッジを含むクラスタBとに分割し、それぞれのクラスタを参照して、クラスタ毎に主方向を設定するので、対象パーティションOPについて最適な予測方向である略横方向の主方向を導出することができる。 In the example shown in FIG. 22, the predicted image generation unit 432 ′ overlaps the reference area composed of the partitions NP1 to NP3 with the cluster A that includes the strong edge and the cluster A as described above. The main direction is set for each cluster with reference to each cluster, so that the main direction in the substantially horizontal direction which is the optimum prediction direction for the target partition OP is divided. A direction can be derived.
 なお、本具体例においては、より一般に、参照領域をNregのクラスタ領域に分割し、各クラスタ領域について主方向を導出する構成とすることができる。 In this specific example, more generally, the reference area can be divided into Nreg cluster areas, and the main direction can be derived for each cluster area.
 (動画像符号化装置)
 以下では、本実施形態に係る動画像符号化装置4について説明する。本実施形態に係る動画像符号化装置4は、実施形態1に係る動画像符号化装置2の備えるPU決定部531、および、予測画像生成部532に代えて、それぞれ、PU決定部531’、および、予測画像生成部532’を備えている。図23に、本実施形態に係る動画像符号化装置4の備えるリーフCU符号化部の構成を示す。本実施形態に係る動画像符号化装置4のその他の構成は、実施形態1に係る動画像符号化装置2の構成と同様である。
(Moving picture encoding device)
Below, the moving image encoder 4 which concerns on this embodiment is demonstrated. The video encoding device 4 according to the present embodiment replaces the PU determination unit 531 and the predicted image generation unit 532 included in the video encoding device 2 according to Embodiment 1, respectively, with a PU determination unit 531 ′, And the prediction image generation part 532 'is provided. FIG. 23 illustrates a configuration of a leaf CU encoding unit included in the moving image encoding device 4 according to the present embodiment. Other configurations of the video encoding device 4 according to the present embodiment are the same as the configurations of the video encoding device 2 according to the first embodiment.
 (PU決定部531’)
 PU決定部531’は、リーフCU画像#242a、および、CU情報#242bを参照し、(1)対象リーフCUの各パーティションへの分割パターン、および、(2)各パーティションに割り付ける予測モード、を決定する。ここで、対象リーフCUの各パーティションへの分割パターンとしては、例えば、対象リーフCUを分割せずにそのまま1つのパーティションとして取り扱う分割パターン、または、対象リーフCUを対称的に4つのパーティションに分割する分割パターン、の何れかが選択可能である。
(PU determination unit 531 ′)
The PU determination unit 531 ′ refers to the leaf CU image # 242a and the CU information # 242b, and (1) the division pattern of the target leaf CU into each partition, and (2) the prediction mode assigned to each partition. decide. Here, as a division pattern into the partitions of the target leaf CU, for example, a division pattern that is handled as one partition without dividing the target leaf CU, or the target leaf CU is divided into four partitions symmetrically. One of the division patterns can be selected.
 また、PU決定部531’は、対象パーティションのサイズに応じて、当該対象パーティションについて選択可能な主方向の総数を決定する。より具体的には、PU決定部531は、対象パーティションのサイズと正の相関を有するように、当該対象パーティションについて選択可能な主方向の総数を決定する。 Also, the PU determination unit 531 'determines the total number of main directions that can be selected for the target partition according to the size of the target partition. More specifically, the PU determination unit 531 determines the total number of main directions that can be selected for the target partition so as to have a positive correlation with the size of the target partition.
 例えば、対象パーティションのサイズが4×4画素である場合、対象パーティションについて選択可能な主方向の総数を2に設定し、対象パーティションのサイズが8×8画素である場合、対象パーティションについて選択可能な主方向の総数を1に設定する。より一般には、PU決定部531’は、対象パーティションに属する画素の総数がNpix1であるときに、当該対象パーティションについて選択可能な主方向の総数をNmain1に決定し、対象パーティションに属する画素の総数がNpix2(Npix1<Npix2)であるときに、当該対象パーティションについて選択可能な主方向の総数をNmain2(Nmain1≧Nmain2)に決定する。 For example, when the size of the target partition is 4 × 4 pixels, the total number of main directions that can be selected for the target partition is set to 2, and when the size of the target partition is 8 × 8 pixels, the target partition can be selected. Set the total number of main directions to 1. More generally, when the total number of pixels belonging to the target partition is Npix1, the PU determining unit 531 ′ determines the total number of main directions that can be selected for the target partition as Nmain1, and the total number of pixels belonging to the target partition is When Npix2 (Npix1 <Npix2), the total number of main directions that can be selected for the target partition is determined as Nmain2 (Nmain1 ≧ Nmain2).
 (1)対象リーフCUを4つのパーティションに分割するか否かを指定するイントラ分割フラグ(intra_split_flag)、および、(2)各パーティションに割り付けられた予測モードを指定する予測モード情報、を含むPU情報#531’は、予測画像生成部532’、および、LCU符号化データ生成部538に供給される。 PU information including (1) an intra split flag (intra_split_flag) that specifies whether or not the target leaf CU is split into four partitions, and (2) prediction mode information that specifies the prediction mode assigned to each partition. # 531 ′ is supplied to the predicted image generation unit 532 ′ and the LCU encoded data generation unit 538.
 なお、PU情報#531’には、複数の主方向であって、その総数が対象パーティションのサイズに応じて決定された複数の主方向の何れを用いるのかを指定する主方向指定インデックスMDIが含まれる。 The PU information # 531 ′ includes a main direction designation index MDI that designates which of a plurality of main directions is used, the total number of which is determined according to the size of the target partition. It is.
 また、PU情報#531’は、対象パーティションに対して主方向を用いるのか、副方向を用いるのかを指定する1ビットのフラグを含む構成としてもよい。また、副方向を用いる場合、追加インデックスAIには、複数の副方向の何れを用いるのかを指定するインデックスを含む構成としてもよい。 Also, the PU information # 531 'may include a 1-bit flag that specifies whether to use the main direction or the sub direction for the target partition. When the sub direction is used, the additional index AI may include an index that specifies which of the plurality of sub directions is used.
 また、後述するように、PU決定部531’は、(1)対象リーフCUの各パーティションへの可能な分割パターン、および、(2)各パーティションに割り付ける可能な予測モード、の全ての組み合わせの各々に対応するPU情報を生成する。また、後述する予測画像、予測残差、および、復号画像の生成は、それら各々のPU情報について逐一行われる。 Further, as will be described later, the PU determination unit 531 ′ performs each of all combinations of (1) a possible division pattern for each partition of the target leaf CU and (2) a prediction mode that can be assigned to each partition. PU information corresponding to is generated. In addition, generation of a prediction image, a prediction residual, and a decoded image, which will be described later, is performed for each piece of PU information.
 また、PU決定部531’は、後述するレート歪評価部537より供給される、レート歪を最適化するPU情報を、リーフCUヘッダ情報#53の一部としてヘッダ情報符号化部22に供給すると共に、当該レート歪を最適化するPU情報を、対象リーフCUについての最終的なPU情報として、予測画像生成部532’、および、LCU符号化データ生成部538に供給する。 Also, the PU determination unit 531 ′ supplies the PU information for optimizing the rate distortion supplied from the rate distortion evaluation unit 537 described later to the header information encoding unit 22 as a part of the leaf CU header information # 53. At the same time, the PU information for optimizing the rate distortion is supplied to the predicted image generation unit 532 ′ and the LCU encoded data generation unit 538 as final PU information for the target leaf CU.
 (予測画像生成部532’)
 予測画像生成部532’は、フレームメモリ27に格納されている復号画像#27、および、PU決定部531’より供給されるPU情報#531’を参照して、対象リーフCUについての予測画像#532’を生成する。予測画像生成部532’による予測画像生成処理は、本実施形態に係る動画像復号装置の備える予測画像生成部432’と同様であるので、ここでは説明を省略する。
(Predicted image generation unit 532 ′)
The predicted image generation unit 532 ′ refers to the decoded image # 27 stored in the frame memory 27 and the PU information # 531 ′ supplied from the PU determination unit 531 ′, thereby predicting the predicted image # for the target leaf CU. 532 ′ is generated. The prediction image generation process by the prediction image generation unit 532 ′ is the same as that of the prediction image generation unit 432 ′ included in the video decoding device according to the present embodiment, and thus description thereof is omitted here.
 〔応用例〕
 上述した動画像符号化装置2及び動画像復号装置1は、動画像の送信、受信、記録、再生を行う各種装置に搭載して利用することができる。動画像符号化装置4及び動画像復号装置3についても同様であるが、以下では、動画像符号化装置2及び動画像復号装置1を例にとり説明を行う。
[Application example]
The above-described moving image encoding device 2 and moving image decoding device 1 can be used by being mounted on various devices that perform transmission, reception, recording, and reproduction of moving images. The same applies to the moving image encoding device 4 and the moving image decoding device 3, but the following description will be made by taking the moving image encoding device 2 and the moving image decoding device 1 as examples.
 まず、上述した動画像符号化装置2及び動画像復号装置1を、動画像の送信及び受信に利用できることを、図25を参照して説明する。 First, it will be described with reference to FIG. 25 that the moving picture encoding apparatus 2 and the moving picture decoding apparatus 1 described above can be used for transmission and reception of moving pictures.
 図25の(a)は、動画像符号化装置2を搭載した送信装置PROD_Aの構成を示したブロック図である。図25の(a)に示すように、送信装置PROD_Aは、動画像を符号化することによって符号化データを得る符号化部PROD_A1と、符号化部PROD_A1が得た符号化データで搬送波を変調することによって変調信号を得る変調部PROD_A2と、変調部PROD_A2が得た変調信号を送信する送信部PROD_A3と、を備えている。上述した動画像符号化装置2は、この符号化部PROD_A1として利用される。 (A) of FIG. 25 is a block diagram illustrating a configuration of a transmission device PROD_A in which the moving image encoding device 2 is mounted. As illustrated in (a) of FIG. 25, the transmission device PROD_A modulates a carrier wave with an encoding unit PROD_A1 that obtains encoded data by encoding a moving image, and the encoded data obtained by the encoding unit PROD_A1. Thus, a modulation unit PROD_A2 that obtains a modulation signal and a transmission unit PROD_A3 that transmits the modulation signal obtained by the modulation unit PROD_A2 are provided. The moving image encoding apparatus 2 described above is used as the encoding unit PROD_A1.
 送信装置PROD_Aは、符号化部PROD_A1に入力する動画像の供給源として、動画像を撮像するカメラPROD_A4、動画像を記録した記録媒体PROD_A5、及び、動画像を外部から入力するための入力端子PROD_A6を更に備えていてもよい。図25の(a)においては、これら全てを送信装置PROD_Aが備えた構成を例示しているが、一部を省略しても構わない。 The transmission device PROD_A is a camera PROD_A4 that captures a moving image, a recording medium PROD_A5 that records the moving image, and an input terminal PROD_A6 for inputting the moving image from the outside as a supply source of the moving image input to the encoding unit PROD_A1. May be further provided. FIG. 25A illustrates a configuration in which the transmission apparatus PROD_A includes all of these, but a part of the configuration may be omitted.
 なお、記録媒体PROD_A5は、符号化されていない動画像を記録したものであってもよいし、伝送用の符号化方式とは異なる記録用の符号化方式で符号化された動画像を記録したものであってもよい。後者の場合、記録媒体PROD_A5と符号化部PROD_A1との間に、記録媒体PROD_A5から読み出した符号化データを記録用の符号化方式に従って復号する復号部(不図示)を介在させるとよい。 The recording medium PROD_A5 may be a recording of a non-encoded moving image, or a recording of a moving image encoded by a recording encoding scheme different from the transmission encoding scheme. It may be a thing. In the latter case, a decoding unit (not shown) for decoding the encoded data read from the recording medium PROD_A5 according to the recording encoding method may be interposed between the recording medium PROD_A5 and the encoding unit PROD_A1.
 図25の(b)は、動画像復号装置1を搭載した受信装置PROD_Bの構成を示したブロック図である。図25の(b)に示すように、受信装置PROD_Bは、変調信号を受信する受信部PROD_B1と、受信部PROD_B1が受信した変調信号を復調することによって符号化データを得る復調部PROD_B2と、復調部PROD_B2が得た符号化データを復号することによって動画像を得る復号部PROD_B3と、を備えている。上述した動画像復号装置1は、この復号部PROD_B3として利用される。 (B) of FIG. 25 is a block diagram illustrating a configuration of the receiving device PROD_B in which the moving image decoding device 1 is mounted. As illustrated in (b) of FIG. 25, the reception device PROD_B includes a reception unit PROD_B1 that receives a modulation signal, a demodulation unit PROD_B2 that obtains encoded data by demodulating the modulation signal received by the reception unit PROD_B1, and a demodulation A decoding unit PROD_B3 that obtains a moving image by decoding the encoded data obtained by the unit PROD_B2. The moving picture decoding apparatus 1 described above is used as the decoding unit PROD_B3.
 受信装置PROD_Bは、復号部PROD_B3が出力する動画像の供給先として、動画像を表示するディスプレイPROD_B4、動画像を記録するための記録媒体PROD_B5、及び、動画像を外部に出力するための出力端子PROD_B6を更に備えていてもよい。図25の(b)においては、これら全てを受信装置PROD_Bが備えた構成を例示しているが、一部を省略しても構わない。 The receiving device PROD_B has a display PROD_B4 for displaying a moving image, a recording medium PROD_B5 for recording the moving image, and an output terminal for outputting the moving image to the outside as a supply destination of the moving image output by the decoding unit PROD_B3. PROD_B6 may be further provided. FIG. 25B illustrates a configuration in which the reception apparatus PROD_B includes all of these, but a part of the configuration may be omitted.
 なお、記録媒体PROD_B5は、符号化されていない動画像を記録するためのものであってもよいし、伝送用の符号化方式とは異なる記録用の符号化方式で符号化されたものであってもよい。後者の場合、復号部PROD_B3と記録媒体PROD_B5との間に、復号部PROD_B3から取得した動画像を記録用の符号化方式に従って符号化する符号化部(不図示)を介在させるとよい。 The recording medium PROD_B5 may be used for recording a non-encoded moving image, or may be encoded using a recording encoding method different from the transmission encoding method. May be. In the latter case, an encoding unit (not shown) for encoding the moving image acquired from the decoding unit PROD_B3 according to the recording encoding method may be interposed between the decoding unit PROD_B3 and the recording medium PROD_B5.
 なお、変調信号を伝送する伝送媒体は、無線であってもよいし、有線であってもよい。また、変調信号を伝送する伝送態様は、放送(ここでは、送信先が予め特定されていない送信態様を指す)であってもよいし、通信(ここでは、送信先が予め特定されている送信態様を指す)であってもよい。すなわち、変調信号の伝送は、無線放送、有線放送、無線通信、及び有線通信の何れによって実現してもよい。 Note that the transmission medium for transmitting the modulation signal may be wireless or wired. Further, the transmission mode for transmitting the modulated signal may be broadcasting (here, a transmission mode in which the transmission destination is not specified in advance) or communication (here, transmission in which the transmission destination is specified in advance). Refers to the embodiment). That is, the transmission of the modulation signal may be realized by any of wireless broadcasting, wired broadcasting, wireless communication, and wired communication.
 例えば、地上デジタル放送の放送局(放送設備など)/受信局(テレビジョン受像機など)は、変調信号を無線放送で送受信する送信装置PROD_A/受信装置PROD_Bの一例である。また、ケーブルテレビ放送の放送局(放送設備など)/受信局(テレビジョン受像機など)は、変調信号を有線放送で送受信する送信装置PROD_A/受信装置PROD_Bの一例である。 For example, a terrestrial digital broadcast broadcasting station (broadcasting equipment or the like) / receiving station (such as a television receiver) is an example of a transmitting device PROD_A / receiving device PROD_B that transmits and receives a modulated signal by wireless broadcasting. Further, a broadcasting station (such as broadcasting equipment) / receiving station (such as a television receiver) of cable television broadcasting is an example of a transmitting device PROD_A / receiving device PROD_B that transmits and receives a modulated signal by cable broadcasting.
 また、インターネットを用いたVOD(Video On Demand)サービスや動画共有サービスなどのサーバ(ワークステーションなど)/クライアント(テレビジョン受像機、パーソナルコンピュータ、スマートフォンなど)は、変調信号を通信で送受信する送信装置PROD_A/受信装置PROD_Bの一例である(通常、LANにおいては伝送媒体として無線又は有線の何れかが用いられ、WANにおいては伝送媒体として有線が用いられる)。ここで、パーソナルコンピュータには、デスクトップ型PC、ラップトップ型PC、及びタブレット型PCが含まれる。また、スマートフォンには、多機能携帯電話端末も含まれる。 Also, a server (workstation etc.) / Client (television receiver, personal computer, smart phone etc.) such as VOD (Video On Demand) service and video sharing service using the Internet is a transmitting device for transmitting and receiving modulated signals by communication. This is an example of PROD_A / reception device PROD_B (usually, either a wireless or wired transmission medium is used in a LAN, and a wired transmission medium is used in a WAN). Here, the personal computer includes a desktop PC, a laptop PC, and a tablet PC. The smartphone also includes a multi-function mobile phone terminal.
 なお、動画共有サービスのクライアントは、サーバからダウンロードした符号化データを復号してディスプレイに表示する機能に加え、カメラで撮像した動画像を符号化してサーバにアップロードする機能を有している。すなわち、動画共有サービスのクライアントは、送信装置PROD_A及び受信装置PROD_Bの双方として機能する。 In addition to the function of decoding the encoded data downloaded from the server and displaying it on the display, the video sharing service client has a function of encoding a moving image captured by the camera and uploading it to the server. That is, the client of the video sharing service functions as both the transmission device PROD_A and the reception device PROD_B.
 次に、上述した動画像符号化装置2及び動画像復号装置1を、動画像の記録及び再生に利用できることを、図26を参照して説明する。 Next, it will be described with reference to FIG. 26 that the above-described moving image encoding device 2 and moving image decoding device 1 can be used for recording and reproduction of moving images.
 図26の(a)は、上述した動画像符号化装置2を搭載した記録装置PROD_Cの構成を示したブロック図である。図26の(a)に示すように、記録装置PROD_Cは、動画像を符号化することによって符号化データを得る符号化部PROD_C1と、符号化部PROD_C1が得た符号化データを記録媒体PROD_Mに書き込む書込部PROD_C2と、を備えている。上述した動画像符号化装置2は、この符号化部PROD_C1として利用される。 FIG. 26 (a) is a block diagram showing a configuration of a recording apparatus PROD_C in which the above-described moving picture encoding apparatus 2 is mounted. As shown in (a) of FIG. 26, the recording device PROD_C includes an encoding unit PROD_C1 that obtains encoded data by encoding a moving image, and the encoded data obtained by the encoding unit PROD_C1 on the recording medium PROD_M. A writing unit PROD_C2 for writing. The moving image encoding apparatus 2 described above is used as the encoding unit PROD_C1.
 なお、記録媒体PROD_Mは、(1)HDD(Hard Disk Drive)やSSD(Solid State Drive)などのように、記録装置PROD_Cに内蔵されるタイプのものであってもよいし、(2)SDメモリカードやUSB(Universal Serial Bus)フラッシュメモリなどのように、記録装置PROD_Cに接続されるタイプのものであってもよいし、(3)DVD(Digital Versatile Disc)やBD(Blu-ray Disk:登録商標)などのように、記録装置PROD_Cに内蔵されたドライブ装置(不図示)に装填されるものであってもよい。 The recording medium PROD_M may be of a type built in the recording device PROD_C, such as (1) HDD (Hard Disk Drive) or SSD (Solid State Drive), or (2) SD memory. It may be of the type connected to the recording device PROD_C, such as a card or USB (Universal Serial Bus) flash memory, or (3) DVD (Digital Versatile Disk) or BD (Blu-ray Disk: registration) Or a drive device (not shown) built in the recording device PROD_C.
 また、記録装置PROD_Cは、符号化部PROD_C1に入力する動画像の供給源として、動画像を撮像するカメラPROD_C3、動画像を外部から入力するための入力端子PROD_C4、及び、動画像を受信するための受信部PROD_C5を更に備えていてもよい。図26の(a)においては、これら全てを記録装置PROD_Cが備えた構成を例示しているが、一部を省略しても構わない。 The recording device PROD_C receives a moving image as a supply source of a moving image to be input to the encoding unit PROD_C1, a camera PROD_C3 that captures a moving image, an input terminal PROD_C4 for inputting a moving image from the outside, and a moving image. The receiving unit PROD_C5 may be further provided. In FIG. 26A, a configuration in which all of these are provided in the recording apparatus PROD_C is illustrated, but a part may be omitted.
 なお、受信部PROD_C5は、符号化されていない動画像を受信するものであってもよいし、記録用の符号化方式とは異なる伝送用の符号化方式で符号化された符号化データを受信するものであってもよい。後者の場合、受信部PROD_C5と符号化部PROD_C1との間に、伝送用の符号化方式で符号化された符号化データを復号する伝送用復号部(不図示)を介在させるとよい。 The receiving unit PROD_C5 may receive a non-encoded moving image, or may receive encoded data encoded by a transmission encoding scheme different from the recording encoding scheme. You may do. In the latter case, a transmission decoding unit (not shown) that decodes encoded data encoded by the transmission encoding method may be interposed between the reception unit PROD_C5 and the encoding unit PROD_C1.
 このような記録装置PROD_Cとしては、例えば、DVDレコーダ、BDレコーダ、HD(Hard Disk)レコーダなどが挙げられる(この場合、入力端子PROD_C4又は受信部PROD_C5が動画像の主な供給源となる)。また、カムコーダ(この場合、カメラPROD_C3が動画像の主な供給源となる)、パーソナルコンピュータ(この場合、受信部PROD_C5が動画像の主な供給源となる)、スマートフォン(この場合、カメラPROD_C3又は受信部PROD_C5が動画像の主な供給源となる)なども、このような記録装置PROD_Cの一例である。 Examples of such a recording device PROD_C include a DVD recorder, a BD recorder, and an HD (Hard Disk) recorder (in this case, the input terminal PROD_C4 or the receiving unit PROD_C5 is a main source of moving images). In addition, a camcorder (in this case, the camera PROD_C3 is a main source of moving images), a personal computer (in this case, the receiving unit PROD_C5 is a main source of moving images), a smartphone (in this case, the camera PROD_C3 or The receiving unit PROD_C5 is a main source of moving images) is an example of such a recording apparatus PROD_C.
 図26の(b)は、上述した動画像復号装置1を搭載した再生装置PROD_Dの構成を示したブロックである。図26の(b)に示すように、再生装置PROD_Dは、記録媒体PROD_Mに書き込まれた符号化データを読み出す読出部PROD_D1と、読出部PROD_D1が読み出した符号化データを復号することによって動画像を得る復号部PROD_D2と、を備えている。上述した動画像復号装置1は、この復号部PROD_D2として利用される。 (B) of FIG. 26 is a block showing a configuration of a playback device PROD_D equipped with the above-described video decoding device 1. As shown in (b) of FIG. 26, the playback device PROD_D reads a moving image by decoding a read unit PROD_D1 that reads encoded data written to the recording medium PROD_M and a coded data read by the read unit PROD_D1. And a decoding unit PROD_D2 to be obtained. The moving picture decoding apparatus 1 described above is used as the decoding unit PROD_D2.
 なお、記録媒体PROD_Mは、(1)HDDやSSDなどのように、再生装置PROD_Dに内蔵されるタイプのものであってもよいし、(2)SDメモリカードやUSBフラッシュメモリなどのように、再生装置PROD_Dに接続されるタイプのものであってもよいし、(3)DVDやBDなどのように、再生装置PROD_Dに内蔵されたドライブ装置(不図示)に装填されるものであってもよい。 Note that the recording medium PROD_M may be of the type built into the playback device PROD_D, such as (1) HDD or SSD, or (2) such as an SD memory card or USB flash memory, It may be of a type connected to the playback device PROD_D, or (3) may be loaded into a drive device (not shown) built in the playback device PROD_D, such as DVD or BD. Good.
 また、再生装置PROD_Dは、復号部PROD_D2が出力する動画像の供給先として、動画像を表示するディスプレイPROD_D3、動画像を外部に出力するための出力端子PROD_D4、及び、動画像を送信する送信部PROD_D5を更に備えていてもよい。図26の(b)においては、これら全てを再生装置PROD_Dが備えた構成を例示しているが、一部を省略しても構わない。 In addition, the playback device PROD_D has a display PROD_D3 that displays a moving image, an output terminal PROD_D4 that outputs the moving image to the outside, and a transmission unit that transmits the moving image as a supply destination of the moving image output by the decoding unit PROD_D2. PROD_D5 may be further provided. FIG. 26B illustrates a configuration in which the playback apparatus PROD_D includes all of these, but a part may be omitted.
  なお、送信部PROD_D5は、符号化されていない動画像を送信するものであってもよいし、記録用の符号化方式とは異なる伝送用の符号化方式で符号化された符号化データを送信するものであってもよい。後者の場合、復号部PROD_D2と送信部PROD_D5との間に、動画像を伝送用の符号化方式で符号化する符号化部(不図示)を介在させるとよい。 The transmission unit PROD_D5 may transmit an unencoded moving image, or transmits encoded data encoded by a transmission encoding method different from the recording encoding method. You may do. In the latter case, it is preferable to interpose an encoding unit (not shown) that encodes a moving image with an encoding method for transmission between the decoding unit PROD_D2 and the transmission unit PROD_D5.
 このような再生装置PROD_Dとしては、例えば、DVDプレイヤ、BDプレイヤ、HDDプレイヤなどが挙げられる(この場合、テレビジョン受像機等が接続される出力端子PROD_D4が動画像の主な供給先となる)。また、テレビジョン受像機(この場合、ディスプレイPROD_D3が動画像の主な供給先となる)、デスクトップ型PC(この場合、出力端子PROD_D4又は送信部PROD_D5が動画像の主な供給先となる)、ラップトップ型又はタブレット型PC(この場合、ディスプレイPROD_D3又は送信部PROD_D5が動画像の主な供給先となる)、スマートフォン(この場合、ディスプレイPROD_D3又は送信部PROD_D5が動画像の主な供給先となる)なども、このような再生装置PROD_Dの一例である。 Examples of such a playback device PROD_D include a DVD player, a BD player, and an HDD player (in this case, an output terminal PROD_D4 to which a television receiver or the like is connected is a main supply destination of moving images). . In addition, a television receiver (in this case, the display PROD_D3 is a main destination of moving images), a desktop PC (in this case, the output terminal PROD_D4 or the transmission unit PROD_D5 is a main destination of moving images), Laptop type or tablet type PC (in this case, display PROD_D3 or transmission unit PROD_D5 is the main supply destination of moving images), smartphone (in this case, display PROD_D3 or transmission unit PROD_D5 is the main supply destination of moving images) ) Is an example of such a playback device PROD_D.
 (付記事項1)
 最後に、上述した動画像復号装置1,3および動画像符号化装置2,4の各ブロックは、集積回路(ICチップ)上に形成された論理回路によってハードウェア的に実現してもよいし、CPU(Central Processing Unit)を用いてソフトウェア的に実現してもよい。
(Appendix 1)
Finally, each block of the above-described moving picture decoding apparatuses 1 and 3 and moving picture encoding apparatuses 2 and 4 may be realized in hardware by a logic circuit formed on an integrated circuit (IC chip). Alternatively, it may be realized by software using a CPU (Central Processing Unit).
 後者の場合、上記各装置は、各機能を実現するプログラムの命令を実行するCPU、上記プログラムを格納したROM(Read Only Memory)、上記プログラムを展開するRAM(Random Access Memory)、上記プログラムおよび各種データを格納するメモリ等の記憶装置(記録媒体)などを備えている。そして、本発明の目的は、上述した機能を実現するソフトウェアである上記各装置の制御プログラムのプログラムコード(実行形式プログラム、中間コードプログラム、ソースプログラム)をコンピュータで読み取り可能に記録した記録媒体を、上記各装置に供給し、そのコンピュータ(またはCPUやMPU)が記録媒体に記録されているプログラムコードを読み出し実行することによっても、達成可能である。 In the latter case, each device includes a CPU that executes instructions of a program that realizes each function, a ROM (Read (Memory) that stores the program, a RAM (Random Memory) that expands the program, the program, and various types A storage device (recording medium) such as a memory for storing data is provided. An object of the present invention is to provide a recording medium in which a program code (execution format program, intermediate code program, source program) of a control program of each of the above devices, which is software that realizes the above-described functions, is recorded so as to be readable by a computer. This can also be achieved by supplying to each of the above devices and reading and executing the program code recorded on the recording medium by the computer (or CPU or MPU).
 上記記録媒体としては、例えば、磁気テープやカセットテープ等のテープ類、フロッピー(登録商標)ディスク/ハードディスク等の磁気ディスクやCD-ROM/MO/MD/DVD/CD-R等の光ディスクを含むディスク類、ICカード(メモリカードを含む)/光カード等のカード類、マスクROM/EPROM/EEPROM/フラッシュROM等の半導体メモリ類、あるいはPLD(Programmable logic device)やFPGA(Field Programmable Gate Array)等の論理回路類などを用いることができる。 Examples of the recording medium include tapes such as magnetic tapes and cassette tapes, magnetic disks such as floppy (registered trademark) disks / hard disks, and disks including optical disks such as CD-ROM / MO / MD / DVD / CD-R. IC cards (including memory cards) / optical cards, semiconductor memories such as mask ROM / EPROM / EEPROM / flash ROM, or PLD (Programmable logic device) or FPGA (Field Programmable Gate Array) Logic circuits can be used.
 また、上記各装置を通信ネットワークと接続可能に構成し、上記プログラムコードを通信ネットワークを介して供給してもよい。この通信ネットワークは、プログラムコードを伝送可能であればよく、特に限定されない。例えば、インターネット、イントラネット、エキストラネット、LAN、ISDN、VAN、CATV通信網、仮想専用網(Virtual Private Network)、電話回線網、移動体通信網、衛星通信網等が利用可能である。また、この通信ネットワークを構成する伝送媒体も、プログラムコードを伝送可能な媒体であればよく、特定の構成または種類のものに限定されない。例えば、IEEE1394、USB、電力線搬送、ケーブルTV回線、電話線、ADSL(Asymmetric Digital Subscriber Line)回線等の有線でも、IrDAやリモコンのような赤外線、Bluetooth(登録商標)、IEEE802.11無線、HDR(High Data Rate)、NFC(Near Field Communication)、DLNA(Digital Living Network Alliance)、携帯電話網、衛星回線、地上波デジタル網等の無線でも利用可能である。 Also, each of the above devices may be configured to be connectable to a communication network, and the program code may be supplied via the communication network. The communication network is not particularly limited as long as it can transmit the program code. For example, the Internet, intranet, extranet, LAN, ISDN, VAN, CATV communication network, virtual private network (Virtual Private Network), telephone line network, mobile communication network, satellite communication network, etc. can be used. The transmission medium constituting the communication network may be any medium that can transmit the program code, and is not limited to a specific configuration or type. For example, even in the case of wired lines such as IEEE 1394, USB, power line carrier, cable TV line, telephone line, ADSL (Asymmetric Digital Subscriber Line) line, infrared rays such as IrDA and remote control, Bluetooth (registered trademark), IEEE 802.11 wireless, HDR ( It can also be used by wireless such as High Data Rate, NFC (Near Field Communication), DLNA (Digital Living Network Alliance), mobile phone network, satellite line, and terrestrial digital network.
 (付記事項2)
 上述したように、本発明に係る画像復号装置は、予測モード指定情報その他のサイド情報と共に符号化データから復号した予測残差に、予測モード群に属する予測モードであって、上記予測モード指定情報により指定される予測モードに従って生成した予測画像を加算することによって復号画像を生成する画像復号装置において、対象予測単位に関する予測モード群に追加する予測モード数Nを、上記サイド情報に含まれる符号化パラメータであって、上記対象予測単位を包含する符号化単位、または、該符号化単位よりも先に復号された符号化単位に関する符号化パラメータを参照して設定する設定手段と、上記対象予測単位に関する予測モード群に追加するN個の予測モードの各々に対応する予測パラメータを、上記復号画像の復号済領域から導出する導出手段と、を備えていることを特徴としている。
(Appendix 2)
As described above, the image decoding apparatus according to the present invention includes a prediction mode belonging to a prediction mode group in a prediction residual decoded from encoded data together with prediction mode specification information and other side information, and the prediction mode specification information In the image decoding device that generates a decoded image by adding the prediction images generated according to the prediction mode specified by the encoding, the number N of prediction modes to be added to the prediction mode group related to the target prediction unit is encoded in the side information A parameter setting unit that refers to a coding unit that includes the target prediction unit or a coding parameter that is decoded prior to the coding unit; and the target prediction unit A prediction parameter corresponding to each of the N prediction modes to be added to the prediction mode group relating to the decoded region of the decoded image. It is characterized in that it comprises a, and deriving means for deriving from.
 予測単位には、予測モード群に属する予測モード数を増やして予測精度を上げる方が符号化効率が向上するタイプと、予測モード群に属する予測モード数を減らして符号量を下げる方が符号化効率が向上するタイプとがある。発明者らが得た知見によれば、ある予測単位が前者のタイプに属するのか後者のタイプに属するのかは、普通、その予測単位、又は、その符号化単位よりも先に復号される符号化単位に関する符号化パラメータから判定することができる。 The prediction unit is a type that increases the prediction accuracy by increasing the number of prediction modes belonging to the prediction mode group, and the coding efficiency is improved by reducing the number of prediction modes belonging to the prediction mode group and improving the coding efficiency. Some types improve efficiency. According to the knowledge obtained by the inventors, whether a certain prediction unit belongs to the former type or the latter type is usually encoded that is decoded before the prediction unit or the encoding unit. It can be determined from the coding parameters for the unit.
 上記の構成によれば、対象予測単位を包含する符号化単位、又は、その符号化単位よりも先に復号される符号化単位に関する符号化パラメータを参照することによって、符号化効率を大きくするように(望ましくは符号化効率を最大化するように)、対象予測単位に関する予測モード群に追加する予測モード数を設定することができる。したがって、上記の構成によれば、高い予測精度を維持しつつ、符号化データの符号量の増大を抑制することによって、符号化効率を向上させることができる。 According to the above configuration, it is possible to increase the coding efficiency by referring to the coding parameter relating to the coding unit including the target prediction unit or the coding unit decoded prior to the coding unit. In addition, it is possible to set the number of prediction modes to be added to the prediction mode group related to the target prediction unit (desirably to maximize the coding efficiency). Therefore, according to said structure, encoding efficiency can be improved by suppressing the increase in the code amount of coding data, maintaining high prediction accuracy.
 なお、上記符号化パラメータとは、復号画像、局所復号画像、または、予測画像を生成するために参照されるパラメータのことであり、画面間予測において参照される動きベクトルや画面内予測において参照される予測モードなどの予測パラメータに加えて、パーティションのサイズや形状、ブロックのサイズや形状、符号化単位のサイズや形状、および、原画像と予測画像との残差データなどが含まれる。また、上記サイド情報とは、符号化パラメータに含まれる情報のうち、上記残差データを除く全ての情報の集合のことを指す。 Note that the coding parameter is a parameter that is referred to in order to generate a decoded image, a locally decoded image, or a predicted image, and is referred to in a motion vector or intra prediction that is referred to in inter-screen prediction. In addition to the prediction parameters such as the prediction mode, the size and shape of the partition, the size and shape of the block, the size and shape of the encoding unit, and residual data between the original image and the predicted image are included. The side information refers to a set of all information except for the residual data among the information included in the encoding parameter.
 また、上記対象予測単位を包含する符号化単位に関する符号化パラメータには、当該符号化単位に関連付けられた符号化パラメータ、および、当該符号化単位に含まれる1または複数の予測単位に関連付けられた符号化パラメータが含まれる。 In addition, the encoding parameter related to the encoding unit including the target prediction unit is related to the encoding parameter associated with the encoding unit and one or a plurality of prediction units included in the encoding unit. Encoding parameters are included.
 また、上記予測単位は、実施形態において説明するPU(Prediction Unit)であってもよいし、PUを分割して得られるパーティションであってもよい。また、上記符号化単位は、実施形態において説明する最大符号化単位(LCU:Largest Coding Unit)であってもよいし、LCUを分割して得られる符号化単位(CU:Coding Unit)であってもよい。 Also, the prediction unit may be a PU (Prediction Unit) described in the embodiment, or may be a partition obtained by dividing the PU. In addition, the coding unit may be a maximum coding unit (LCU: Largegest Coding Unit) described in the embodiment, or a coding unit (CU: Coding Unit) obtained by dividing an LCU. Also good.
 また、本発明に係る画像復号装置においては、上記導出手段は、上記予測モード群に含まれる主予測モードに対応する主予測パラメータを上記復号済領域から導出すると共に、上記予測モード群に追加するN個の副予測モードの各々に対応する副予測パラメータであって、上記主予測パラメータと特定の関係にある副予測パラメータを上記主予測パラメータから導出する、ことが好ましい。 In the image decoding device according to the present invention, the deriving unit derives a main prediction parameter corresponding to a main prediction mode included in the prediction mode group from the decoded region and adds the main prediction parameter to the prediction mode group. It is preferable that a sub prediction parameter corresponding to each of the N sub prediction modes and having a specific relationship with the main prediction parameter is derived from the main prediction parameter.
 上記の構成によれば、上記導出手段によって導出される主予測パラメータは、上記復号済領域を参照して導出されるので、上記主予測モードを含む予測モード群を用いることによって、予測精度の向上を図ることができる。 According to the above configuration, since the main prediction parameter derived by the deriving unit is derived with reference to the decoded region, the prediction accuracy is improved by using the prediction mode group including the main prediction mode. Can be achieved.
 さらに、上記の構成によれば、上記予測モード群に追加するN個の副予測モードの各々に対応する副予測パラメータであって、上記主予測パラメータと特定の関係にある副予測パラメータを上記主予測パラメータから導出するので、予測精度の更なる向上を図ることができる。 Further, according to the above configuration, the sub prediction parameters corresponding to each of the N sub prediction modes to be added to the prediction mode group, the sub prediction parameters having a specific relationship with the main prediction parameter are converted into the main prediction parameters. Since it derives from the prediction parameters, the prediction accuracy can be further improved.
 また、上記副予測モードの総数は、対象予測単位を包含する符号化単位、又は、その符号化単位よりも先に復号される符号化単位に関する符号化パラメータを参照して定めることができるので、副予測モードを過剰に追加することによって生じ得る符号量の増加に起因した符号化効率の低下を防ぐことができる。 The total number of sub-prediction modes can be determined with reference to a coding unit that includes the target prediction unit or a coding parameter related to a coding unit that is decoded prior to the coding unit. It is possible to prevent a decrease in coding efficiency due to an increase in the amount of code that can be caused by excessively adding the sub prediction mode.
 また、本発明に係る画像復号装置においては、上記主予測パラメータは、上記復号済領域の画素値から推定した上記対象予測単位におけるエッジ方向θmを表し、上記主予測モードは、上記復号済領域の画素値を上記主予測パラメータが表す外挿方向から外挿することによって、上記対象予測単位上の予測画像を生成する予測モードであり、上記N個の副予測パラメータは、θmを上記エッジ方向として、方向θm±k×δ(1≦k≦N)を表し、上記N個の副予測モードは、上記復号済領域の画素値を上記N個の副予測パラメータの各々が表す外挿方向から外挿することによって、上記対象予測単位上の予測画像を生成する予測モードである、ことが好ましい。 In the image decoding device according to the present invention, the main prediction parameter represents the edge direction θm in the target prediction unit estimated from the pixel value of the decoded region, and the main prediction mode is the value of the decoded region. This is a prediction mode for generating a prediction image on the target prediction unit by extrapolating the pixel value from the extrapolation direction represented by the main prediction parameter, and the N sub-prediction parameters have θm as the edge direction. , Direction θm ± k × δ (1 ≦ k ≦ N), and the N sub-prediction modes exclude the pixel values of the decoded region from the extrapolation direction represented by each of the N sub-prediction parameters. It is preferable that it is a prediction mode which produces | generates the estimated image on the said object prediction unit by inserting.
 上記の構成によれば、主予測パラメータは、上記復号済領域の画素値から推定した上記対象予測単位におけるエッジ方向θmを表すので、符号化対象画像にエッジが存在する場合に、上記対象予測単位についての予測画像を高い予測精度で生成することができる。また、符号化対象画像におけるエッジ以外の成分により、上記エッジ方向θmが、対象予測単位についての最適な予測方向に一致しない場合には、上記副予測モードを用いることができるので、高い予測精度を保つことができる。 According to the above configuration, since the main prediction parameter represents the edge direction θm in the target prediction unit estimated from the pixel value of the decoded region, the target prediction unit when the edge exists in the encoding target image Can be generated with high prediction accuracy. In addition, when the edge direction θm does not match the optimal prediction direction for the target prediction unit due to components other than the edges in the encoding target image, the sub prediction mode can be used, so high prediction accuracy can be achieved. Can keep.
 また、上記副予測モードの総数は、対象予測単位を包含する符号化単位、又は、その符号化単位よりも先に復号される符号化単位に関する符号化パラメータを参照して定めることができるので、副予測モードを過剰に追加することによって生じ得る符号量の増加に起因した符号化効率の低下を防ぐことができる。 The total number of sub-prediction modes can be determined with reference to a coding unit that includes the target prediction unit or a coding parameter related to a coding unit that is decoded prior to the coding unit. It is possible to prevent a decrease in coding efficiency due to an increase in the amount of code that can be caused by excessively adding the sub prediction mode.
 また、本発明に係る画像復号装置においては、上記予測モード群には、上記主予測モードおよび上記副予測モードに加えて、予測パラメータが予め定められた基本予測モードが含まれており、上記導出手段は、上記主予測パラメータおよび副予測パラメータの各々が表す外挿方向の間隔δを、上記対象予測単位に隣接する隣接予測単位に関する予測モード指定情報により指定された予測モードが、上記主予測モードまたは上記副予測モードであるか、上記基本予測モードであるかに応じた値に調整する、ことが好ましい。 In the image decoding device according to the present invention, the prediction mode group includes a basic prediction mode in which a prediction parameter is predetermined in addition to the main prediction mode and the sub prediction mode, and the derivation is performed. The means is configured such that the prediction mode designated by the prediction mode designation information relating to the adjacent prediction unit adjacent to the target prediction unit, the interval δ in the extrapolation direction represented by each of the main prediction parameter and the sub prediction parameter is the main prediction mode. Or it is preferable to adjust to the value according to whether it is the said sub prediction mode or the said basic prediction mode.
 上記主予測モードが選択される予測単位には、比較的強いエッジが存在するという傾向がある。また、そのような強いエッジは、複数の予測単位に跨って連続的に存在している場合が多い。したがって、上記隣接予測単位について主予測モードを用いて導出された主予測パラメータは、上記対象予測単位についての最適な予測パラメータに近いという傾向がある。 The prediction unit in which the main prediction mode is selected tends to have a relatively strong edge. In addition, such strong edges often exist continuously across a plurality of prediction units. Therefore, the main prediction parameter derived using the main prediction mode for the adjacent prediction unit tends to be close to the optimal prediction parameter for the target prediction unit.
 上記の構成によれば、対象予測単位についての主予測パラメータと副予測パラメータの各々が表す外挿方向の間隔δを、上記主予測モードまたは上記副予測モードであるか、上記基本予測モードであるかに応じた値に調整することによって、予測精度が向上し、符号化効率が向上する。 According to the above configuration, the extrapolation direction interval δ represented by each of the main prediction parameter and the sub prediction parameter for the target prediction unit is the main prediction mode, the sub prediction mode, or the basic prediction mode. By adjusting to a value corresponding to this, the prediction accuracy is improved and the coding efficiency is improved.
 また、本発明に係る画像復号装置においては、上記サイド情報には、上記対象予測単位のサイズを規定する符号化パラメータが含まれ、上記設定手段は、上記予測モード群に追加する予測モード数Nを、上記符号化パラメータにより規定される上記対象予測単位のサイズに応じた値に設定する、ことが好ましい。 In the image decoding device according to the present invention, the side information includes an encoding parameter that defines a size of the target prediction unit, and the setting unit adds the number N of prediction modes to be added to the prediction mode group. Is preferably set to a value according to the size of the target prediction unit defined by the encoding parameter.
 一般に、ある符号化単位についての符号化対象画像を符号化するために必要なサイド情報の符号量は、当該符号化単位に含まれる予測単位の総数と正の相関を有する。すなわち、予測単位の総数が大きければサイド情報の符号量が増大する。また、ある符号化単位に含まれる予測単位の総数は、各予測単位のサイズと負の相関を有する。 Generally, the code amount of side information necessary for encoding an encoding target image for a certain encoding unit has a positive correlation with the total number of prediction units included in the encoding unit. That is, if the total number of prediction units is large, the code amount of the side information increases. The total number of prediction units included in a certain coding unit has a negative correlation with the size of each prediction unit.
 上記の構成によれは、上記設定手段は、上記予測モード群に追加する予測モード数Nを、上記符号化パラメータにより規定される上記対象予測単位のサイズに応じた値に設定することにより、高い予測精度を保ちつつ、符号化対象画像を符号化するために必要なサイド情報の符号量の増加を抑制することができるので、符号化効率が向上する。 According to said structure, the said setting means is high by setting the prediction mode number N added to the said prediction mode group to the value according to the size of the said object prediction unit prescribed | regulated by the said encoding parameter. Since it is possible to suppress an increase in the code amount of the side information necessary for encoding the encoding target image while maintaining the prediction accuracy, the encoding efficiency is improved.
 また、本発明に係る画像復号装置においては、上記サイド情報には、基本予測モード数Mを規定する符号化パラメータが含まれ、上記予測モード群には、M個の基本予測モードが含まれ、上記設定手段は、上記予測モード群に追加する予測モード数Nを、上記符号化パラメータにより規定される基本予測モード数Mに応じた値に設定する、ことが好ましい。 In the image decoding apparatus according to the present invention, the side information includes an encoding parameter that defines the number M of basic prediction modes, and the prediction mode group includes M basic prediction modes. The setting means preferably sets the number N of prediction modes to be added to the prediction mode group to a value corresponding to the number M of basic prediction modes defined by the encoding parameter.
 選択可能な基本予測モードの総数が大きく設定されている予測単位の周辺の予測単位は、予測モード群に属する予測モード数を増やして予測精度を上げる方が符号化効率が向上するタイプであるという傾向がある。 The prediction unit around the prediction unit in which the total number of selectable basic prediction modes is set to be large is that the coding efficiency is improved by increasing the number of prediction modes belonging to the prediction mode group and increasing the prediction accuracy. Tend.
 上記の構成によれば、上記予測モード群に追加する予測モード数Nを、上記符号化パラメータにより規定される基本予測モード数Mに応じた値に設定することにより、予測精度、および、符号化効率の双方が向上する。 According to said structure, by setting the prediction mode number N added to the said prediction mode group to the value according to the basic prediction mode number M prescribed | regulated by the said encoding parameter, prediction accuracy and encoding Both efficiencies are improved.
 また、本発明に係る画像復号装置においては、上記予測モード群には、上記主予測モードおよび上記副予測モードに加えて、予測パラメータが予め定められた基本予測モードが含まれており、上記設定手段は、上記予測モード群に追加する予測モード数Nを、上記対象予測単位に隣接する隣接予測単位に関する予測モード指定情報により指定された予測モードが、上記主予測モードまたは上記副予測モードであるか、上記基本予測モードであるかに応じた値に設定する、ことが好ましい。 In the image decoding apparatus according to the present invention, the prediction mode group includes a basic prediction mode in which a prediction parameter is predetermined in addition to the main prediction mode and the sub prediction mode, and the setting The means is that the prediction mode designated by the prediction mode designation information related to the adjacent prediction unit adjacent to the target prediction unit for the number N of prediction modes to be added to the prediction mode group is the main prediction mode or the sub prediction mode. It is preferable to set the value according to whether the basic prediction mode is selected.
 上記主予測モードが選択される予測単位には、比較的強いエッジが存在するという傾向がある。また、そのような強いエッジは、複数の予測単位に跨って連続的に存在している場合が多い。したがって、上記隣接予測単位について主予測モードを用いて導出された主予測パラメータは、上記対象予測単位についての最適な予測パラメータに近いという傾向がある。 The prediction unit in which the main prediction mode is selected tends to have a relatively strong edge. In addition, such strong edges often exist continuously across a plurality of prediction units. Therefore, the main prediction parameter derived using the main prediction mode for the adjacent prediction unit tends to be close to the optimal prediction parameter for the target prediction unit.
 上記の構成によれば、上記予測モード群に追加する予測モード数Nを、上記対象予測単位に隣接する隣接予測単位に関する予測モード指定情報により指定された予測モードが、上記主予測モードまたは上記副予測モードであるか、上記基本予測モードであるかに応じた値に設定することによって、符号量が削減され、符号化効率が向上する。 According to said structure, the prediction mode designated with the prediction mode designation | designated information regarding the adjacent prediction unit adjacent to the said prediction unit for the prediction mode number N added to the said prediction mode group is the said main prediction mode or the said sub prediction mode. By setting the value according to whether it is the prediction mode or the basic prediction mode, the code amount is reduced and the coding efficiency is improved.
 また、本発明に係る画像復号装置においては、上記導出手段は、上記予測モード群に追加するN個の主予測モードの各々に対応する主予測パラメータを上記復号済領域から導出する、ことが好ましい。 In the image decoding device according to the present invention, it is preferable that the deriving unit derives a main prediction parameter corresponding to each of N main prediction modes to be added to the prediction mode group from the decoded region. .
 上記の構成によれば、上記N個の主予測モードの各々に対応する主予測パラメータは、上記復号済領域を参照して導出されるので、主予測モードを1つのみ追加する場合に比べて、更なる予測精度の向上を図ることができる。 According to the above configuration, since the main prediction parameters corresponding to each of the N main prediction modes are derived with reference to the decoded region, compared with a case where only one main prediction mode is added. Therefore, the prediction accuracy can be further improved.
 また、上記導出手段は、上記復号済み領域をN個の領域に分割し、得られたN個の領域の各々から上記主予測モードを導出する、ことが好ましい。 Further, it is preferable that the deriving unit divides the decoded region into N regions and derives the main prediction mode from each of the obtained N regions.
 一般に、符号化対象画像は、互いに画像の方向性の異なる複数の領域から構成されていることが多い。上記の構成によれば、上記導出手段が、上記復号済み領域をN個の領域に分割し、得られたN個の領域の各々から上記主予測モードを導出するので、復号済画像が互いに画像の方向性の異なる複数の領域から構成されている場合であっても、対象予測単位についての適切な主予測パラメータを導出することができる。したがって、上記の構成によれば、符号化対象画像が、互いに画像の方向性の異なる複数の領域から構成されている場合であっても、高い予測精度を保つことができる。 Generally, an encoding target image is often composed of a plurality of regions having different image directivities. According to the above configuration, the derivation unit divides the decoded region into N regions, and derives the main prediction mode from each of the obtained N regions, so that the decoded images are mutually imaged. Even if it is comprised from several area | regions from which directionality differs, the appropriate main prediction parameter about an object prediction unit can be derived | led-out. Therefore, according to the above configuration, high prediction accuracy can be maintained even when the encoding target image is composed of a plurality of regions having different image directivities.
 また、本発明に係る画像符号化装置は、予測単位毎に予測モード群から選択した予測モードに従って生成した予測画像を原画像から減算して得られる予測残差を、選択した予測モードを指定する予測モード指定情報その他のサイド情報と共に符号化することによって符号化データを生成する画像符号化装置において、対象予測単位に関する予測モード群に追加する予測モード数Nを、上記サイド情報に含まれる符号化パラメータであって、上記対象予測単位を包含する符号化単位、または、該符号化単位よりも先に符号化された符号化単位に関する符号化パラメータを参照して設定する設定手段と、上記対象予測単位に関する予測モード群に追加するN個の予測モードの各々に対応する予測パラメータを局所復号画像の復号済領域から導出する導出手段と、を備えていることを特徴としている。 In addition, the image coding apparatus according to the present invention designates the selected prediction mode based on the prediction residual obtained by subtracting the predicted image generated according to the prediction mode selected from the prediction mode group for each prediction unit from the original image. In an image encoding device that generates encoded data by encoding together with prediction mode designation information and other side information, the number N of prediction modes to be added to the prediction mode group related to the target prediction unit is encoded in the side information. A parameter setting unit that refers to a coding unit that includes the target prediction unit or a coding parameter that is encoded prior to the coding unit; and the target prediction Deriving prediction parameters corresponding to each of the N prediction modes to be added to the prediction mode group related to the unit from the decoded region of the locally decoded image It is characterized in that comprises a derivation means that, the.
 予測単位には、予測モード群に属する予測モード数を増やして予測精度を上げる方が符号化効率が向上するタイプと、予測モード群に属する予測モード数を減らして符号量を下げる方が符号化効率が向上するタイプとがある。発明者らが得た知見によれば、ある予測単位が前者のタイプに属するのか後者のタイプに属するのかは、普通、その予測単位、又は、その符号化単位よりも先に符号化される符号化単位に関する符号化パラメータから判定することができる。 The prediction unit is a type that increases the prediction accuracy by increasing the number of prediction modes belonging to the prediction mode group, and the coding efficiency is improved by reducing the number of prediction modes belonging to the prediction mode group and improving the coding efficiency. Some types improve efficiency. According to the knowledge obtained by the inventors, whether a certain prediction unit belongs to the former type or the latter type is usually a code encoded before the prediction unit or the encoding unit. It can be determined from the encoding parameters for the encoding unit.
 上記の構成によれば、対象予測単位を包含する符号化単位、又は、その符号化単位よりも先に符号化される符号化単位に関する符号化パラメータを参照することによって、符号化効率を大きくするように(望ましくは符号化効率を最大化するように)、対象予測単位に関する予測モード群に追加する予測モード数を設定することができる。したがって、上記の構成によれば、高い予測精度を維持しつつ、符号化データの符号量の増大を抑制することによって、符号化効率を向上させることができる。 According to the above configuration, the encoding efficiency is increased by referring to the encoding parameter including the target prediction unit or the encoding parameter related to the encoding unit encoded before the encoding unit. As such (desirably to maximize encoding efficiency), the number of prediction modes to be added to the prediction mode group related to the target prediction unit can be set. Therefore, according to said structure, encoding efficiency can be improved by suppressing the increase in the code amount of coding data, maintaining high prediction accuracy.
 なお、上記符号化パラメータとは、復号画像、局所復号画像、または、予測画像を生成するために参照されるパラメータのことであり、画面間予測において参照される動きベクトルや画面内予測において参照される予測モードなどの予測パラメータに加えて、パーティションのサイズや形状、ブロックのサイズや形状、符号化単位のサイズや形状、および、原画像と予測画像との残差データなどが含まれる。また、上記サイド情報とは、符号化パラメータに含まれる情報のうち、上記残差データを除く全ての情報の集合のことを指す。 Note that the coding parameter is a parameter that is referred to in order to generate a decoded image, a locally decoded image, or a predicted image, and is referred to in a motion vector or intra prediction that is referred to in inter-screen prediction. In addition to the prediction parameters such as the prediction mode, the size and shape of the partition, the size and shape of the block, the size and shape of the encoding unit, and residual data between the original image and the predicted image are included. The side information refers to a set of all information except for the residual data among the information included in the encoding parameter.
 また、上記対象予測単位を包含する符号化単位に関する符号化パラメータには、当該符号化単位に関連付けられた符号化パラメータ、および、当該符号化単位に含まれる1または複数の予測単位に関連付けられた符号化パラメータが含まれる。 In addition, the encoding parameter related to the encoding unit including the target prediction unit is related to the encoding parameter associated with the encoding unit and one or a plurality of prediction units included in the encoding unit. Encoding parameters are included.
 また、上記予測単位は、実施形態において説明するPU(Prediction Unit)であってもよいし、PUを分割して得られるパーティションであってもよい。また、上記符号化単位は、実施形態において説明する最大符号化単位(LCU:Largest Coding Unit)であってもよいし、LCUを分割して得られる符号化単位(CU:Coding Unit)であってもよい。 Also, the prediction unit may be a PU (Prediction Unit) described in the embodiment, or may be a partition obtained by dividing the PU. In addition, the coding unit may be a maximum coding unit (LCU: Largegest Coding Unit) described in the embodiment, or a coding unit (CU: Coding Unit) obtained by dividing an LCU. Also good.
 また、本発明に係る符号化データのデータ構造は、予測単位毎に予測モード群から選択した予測モードに従って生成した予測画像を原画像から減算して得られる予測残差を、選択した予測モードを指定する予測モード指定情報その他のサイド情報と共に符号化することによって生成された符号化データのデータ構造であって、対象予測単位に関する予測モード群に追加する予測モードの個数が、上記サイド情報に含まれる符号化パラメータであって、上記対象予測単位を包含する符号化単位、または、該符号化単位よりも先に符号化された符号化単位に関する符号化パラメータによって陰的に表現されている、ことを特徴としている。 In addition, the data structure of the encoded data according to the present invention includes a prediction residual obtained by subtracting, from the original image, a prediction image generated according to a prediction mode selected from a prediction mode group for each prediction unit. The side information includes the number of prediction modes to be added to the prediction mode group related to the target prediction unit, which is a data structure of encoded data generated by encoding together with the prediction mode specifying information to be specified and other side information. Coding parameters that are expressed implicitly by a coding unit that includes the target prediction unit or a coding parameter that is coded before the coding unit. It is characterized by.
 上記のように構成された符号化データのデータ構造によれば、本発明に係る画像復号装置および画像符号化装置と同様の効果を奏する。 According to the data structure of the encoded data configured as described above, the same effects as those of the image decoding apparatus and the image encoding apparatus according to the present invention are achieved.
 本発明は上述した各実施形態に限定されるものではなく、請求項に示した範囲で種々の変更が可能であり、異なる実施形態にそれぞれ開示された技術的手段を適宜組み合わせて得られる実施形態についても本発明の技術的範囲に含まれる。 The present invention is not limited to the above-described embodiments, and various modifications are possible within the scope shown in the claims, and embodiments obtained by appropriately combining technical means disclosed in different embodiments. Is also included in the technical scope of the present invention.
 本発明は、符号化データを復号する復号装置、および、符号化データを生成する符号化装置に好適に適用することができる。また、符号化装置によって生成され、復号装置によって参照される符号化データのデータ構造に好適に適用することができる。 The present invention can be suitably applied to a decoding device that decodes encoded data and an encoding device that generates encoded data. Further, the present invention can be suitably applied to the data structure of encoded data generated by the encoding device and referenced by the decoding device.
1,3    動画像復号装置
11     可変長符号逆多重化部
12     ヘッダ情報復号部
13     LCU設定部
14     LCU復号部(設定手段、導出手段)
15     フレームメモリ
2,4    動画像符号化装置
21     LCUヘッダ情報決定部
22     ヘッダ情報符号化部
23     LCU設定部
24     LCU符号化部(設定手段、導出手段)
25     可変長符号多重化部
26     LCU復号部
27     フレームメモリ
1, 3 Video decoding device 11 Variable length code demultiplexing unit 12 Header information decoding unit 13 LCU setting unit 14 LCU decoding unit (setting means, deriving means)
15 Frame memory 2, 4 Video encoding device 21 LCU header information determination unit 22 Header information encoding unit 23 LCU setting unit 24 LCU encoding unit (setting unit, deriving unit)
25 Variable length code multiplexing unit 26 LCU decoding unit 27 Frame memory

Claims (11)

  1.  予測モード指定情報その他のサイド情報と共に符号化データから復号した予測残差に、予測モード群に属する予測モードであって、上記予測モード指定情報により指定される予測モードに従って生成した予測画像を加算することによって復号画像を生成する画像復号装置において、
     対象予測単位に関する予測モード群に追加する予測モード数Nを、上記サイド情報に含まれる符号化パラメータであって、上記対象予測単位を包含する符号化単位、または、該符号化単位よりも先に復号された符号化単位に関する符号化パラメータを参照して設定する設定手段と、
     上記対象予測単位に関する予測モード群に追加するN個の予測モードの各々に対応する予測パラメータを、上記復号画像の復号済領域から導出する導出手段と、を備えていることを特徴とする画像復号装置。
    Prediction images, which are prediction modes belonging to the prediction mode group and generated according to the prediction mode specified by the prediction mode specification information, are added to the prediction residual decoded from the encoded data together with the prediction mode specification information and other side information. In an image decoding device that generates a decoded image by
    The number of prediction modes N to be added to the prediction mode group related to the target prediction unit is a coding parameter included in the side information, and is a coding unit including the target prediction unit, or prior to the coding unit. Setting means for setting with reference to a coding parameter relating to a decoded coding unit;
    Decoding means for deriving prediction parameters corresponding to each of the N prediction modes to be added to the prediction mode group related to the target prediction unit from the decoded region of the decoded image apparatus.
  2.  上記導出手段は、上記予測モード群に含まれる主予測モードに対応する主予測パラメータを上記復号済領域から導出すると共に、上記予測モード群に追加するN個の副予測モードの各々に対応する副予測パラメータであって、上記主予測パラメータと特定の関係にある副予測パラメータを上記主予測パラメータから導出する、
    ことを特徴とする請求項1に記載の画像復号装置。
    The derivation means derives a main prediction parameter corresponding to a main prediction mode included in the prediction mode group from the decoded region, and outputs a sub prediction mode corresponding to each of the N sub prediction modes to be added to the prediction mode group. A sub-prediction parameter that is a prediction parameter and has a specific relationship with the main prediction parameter is derived from the main prediction parameter;
    The image decoding apparatus according to claim 1.
  3.  上記主予測パラメータは、上記復号済領域の画素値から推定した上記対象予測単位におけるエッジ方向θmを表し、上記主予測モードは、上記復号済領域の画素値を上記主予測パラメータが表す外挿方向から外挿することによって、上記対象予測単位上の予測画像を生成する予測モードであり、
     上記N個の副予測パラメータは、θmを上記エッジ方向として、方向θm±k×δ(1≦k≦N)を表し、上記N個の副予測モードは、上記復号済領域の画素値を上記N個の副予測パラメータの各々が表す外挿方向から外挿することによって、上記対象予測単位上の予測画像を生成する予測モードである、
    ことを特徴とする請求項2に記載の画像復号装置。
    The main prediction parameter represents the edge direction θm in the target prediction unit estimated from the pixel value of the decoded area, and the main prediction mode represents an extrapolation direction in which the main prediction parameter represents the pixel value of the decoded area Is a prediction mode for generating a prediction image on the target prediction unit by extrapolating from
    The N sub-prediction parameters represent a direction θm ± k × δ (1 ≦ k ≦ N) with θm as the edge direction, and the N sub-prediction modes represent the pixel values of the decoded region as the values It is a prediction mode for generating a prediction image on the target prediction unit by extrapolating from the extrapolation direction represented by each of the N sub-prediction parameters.
    The image decoding apparatus according to claim 2.
  4.  上記予測モード群には、上記主予測モードおよび上記副予測モードに加えて、予測パラメータが予め定められた基本予測モードが含まれており、
     上記導出手段は、上記主予測パラメータおよび副予測パラメータの各々が表す外挿方向の間隔δを、上記対象予測単位に隣接する隣接予測単位に関する予測モード指定情報により指定された予測モードが、上記主予測モードまたは上記副予測モードであるか、上記基本予測モードであるかに応じた値に調整する、
    ことを特徴とする請求項3に記載の画像復号装置。
    In addition to the main prediction mode and the sub prediction mode, the prediction mode group includes a basic prediction mode in which a prediction parameter is predetermined,
    The derivation means uses the prediction mode designated by the prediction mode designation information related to the adjacent prediction unit adjacent to the target prediction unit for the interval δ in the extrapolation direction represented by each of the main prediction parameter and the sub prediction parameter. Adjust to a value depending on whether the prediction mode or the sub-prediction mode or the basic prediction mode,
    The image decoding apparatus according to claim 3.
  5.  上記サイド情報には、上記対象予測単位のサイズを規定する符号化パラメータが含まれ、
     上記設定手段は、上記予測モード群に追加する予測モード数Nを、上記符号化パラメータにより規定される上記対象予測単位のサイズに応じた値に設定する、
    ことを特徴とする請求項1から4までの何れか1項に記載の画像復号装置。
    The side information includes an encoding parameter that defines the size of the target prediction unit,
    The setting means sets the prediction mode number N to be added to the prediction mode group to a value according to the size of the target prediction unit defined by the encoding parameter.
    The image decoding apparatus according to any one of claims 1 to 4, wherein the image decoding apparatus according to any one of claims 1 to 4 is provided.
  6.  上記サイド情報には、基本予測モード数Mを規定する符号化パラメータが含まれ、
     上記予測モード群には、M個の基本予測モードが含まれ、
     上記設定手段は、上記予測モード群に追加する予測モード数Nを、上記符号化パラメータにより規定される基本予測モード数Mに応じた値に設定する、
    ことを特徴とする請求項1から4までの何れか1項に記載の画像復号装置。
    The side information includes an encoding parameter that defines the number M of basic prediction modes,
    The prediction mode group includes M basic prediction modes,
    The setting means sets the prediction mode number N to be added to the prediction mode group to a value according to the basic prediction mode number M defined by the encoding parameter.
    The image decoding apparatus according to any one of claims 1 to 4, wherein the image decoding apparatus according to any one of claims 1 to 4 is provided.
  7.  上記予測モード群には、上記主予測モードおよび上記副予測モードに加えて、予測パラメータが予め定められた基本予測モードが含まれており、
     上記設定手段は、上記予測モード群に追加する予測モード数Nを、上記対象予測単位に隣接する隣接予測単位に関する予測モード指定情報により指定された予測モードが、上記主予測モードまたは上記副予測モードであるか、上記基本予測モードであるかに応じた値に設定する、
    ことを特徴とする請求項3または4に記載の画像復号装置。
    In addition to the main prediction mode and the sub prediction mode, the prediction mode group includes a basic prediction mode in which a prediction parameter is predetermined,
    The setting means is configured such that the prediction mode designated by the prediction mode designation information related to the adjacent prediction unit adjacent to the target prediction unit is the prediction mode number N to be added to the prediction mode group. Or a value corresponding to whether the basic prediction mode is set,
    The image decoding apparatus according to claim 3 or 4, wherein
  8.  上記導出手段は、上記予測モード群に追加するN個の主予測モードの各々に対応する主予測パラメータを上記復号済領域から導出する、
    ことを特徴とする請求項1に記載の画像復号装置。
    The derivation means derives a main prediction parameter corresponding to each of the N main prediction modes to be added to the prediction mode group from the decoded region.
    The image decoding apparatus according to claim 1.
  9.  上記導出手段は、上記復号済領域をN個の領域に分割し、得られたN個の領域の各々から上記主予測モードを導出する、
    ことを特徴とする請求項8に記載の画像復号装置。
    The derivation means divides the decoded region into N regions, and derives the main prediction mode from each of the obtained N regions.
    The image decoding apparatus according to claim 8.
  10.  予測単位毎に予測モード群から選択した予測モードに従って生成した予測画像を原画像から減算して得られる予測残差を、選択した予測モードを指定する予測モード指定情報その他のサイド情報と共に符号化することによって符号化データを生成する画像符号化装置において、
     対象予測単位に関する予測モード群に追加する予測モード数Nを、上記サイド情報に含まれる符号化パラメータであって、上記対象予測単位を包含する符号化単位、または、該符号化単位よりも先に符号化された符号化単位に関する符号化パラメータを参照して設定する設定手段と、
     上記対象予測単位に関する予測モード群に追加するN個の予測モードの各々に対応する予測パラメータを局所復号画像の復号済領域から導出する導出手段と、を備えていることを特徴とする画像符号化装置。
    The prediction residual obtained by subtracting the prediction image generated according to the prediction mode selected from the prediction mode group for each prediction unit from the original image is encoded together with the prediction mode specifying information specifying the selected prediction mode and other side information. In an image encoding device that generates encoded data by
    The number of prediction modes N to be added to the prediction mode group related to the target prediction unit is a coding parameter included in the side information, and is a coding unit including the target prediction unit, or prior to the coding unit. Setting means for setting with reference to a coding parameter relating to a coded coding unit;
    Deriving means for deriving prediction parameters corresponding to each of the N prediction modes to be added to the prediction mode group related to the target prediction unit from the decoded region of the local decoded image, apparatus.
  11.  予測単位毎に予測モード群から選択した予測モードに従って生成した予測画像を原画像から減算して得られる予測残差を、選択した予測モードを指定する予測モード指定情報その他のサイド情報と共に符号化することによって生成された符号化データのデータ構造であって、
     対象予測単位に関する予測モード群に追加する予測モードの個数が、上記サイド情報に含まれる符号化パラメータであって、上記対象予測単位を包含する符号化単位、または、該符号化単位よりも先に符号化された符号化単位に関する符号化パラメータによって陰的に表現されている、ことを特徴とする符号化データのデータ構造。
    The prediction residual obtained by subtracting the prediction image generated according to the prediction mode selected from the prediction mode group for each prediction unit from the original image is encoded together with the prediction mode specifying information specifying the selected prediction mode and other side information. A data structure of encoded data generated by
    The number of prediction modes to be added to the prediction mode group related to the target prediction unit is a coding parameter included in the side information, and the coding unit including the target prediction unit or the coding unit before the coding unit. A data structure of encoded data, characterized in that it is implicitly represented by an encoding parameter relating to an encoded encoding unit.
PCT/JP2011/072283 2010-09-30 2011-09-28 Image decoding device, image encoding device, and data structure WO2012043678A1 (en)

Applications Claiming Priority (2)

Application Number Priority Date Filing Date Title
JP2010-222234 2010-09-30
JP2010222234 2010-09-30

Publications (1)

Publication Number Publication Date
WO2012043678A1 true WO2012043678A1 (en) 2012-04-05

Family

ID=45893117

Family Applications (1)

Application Number Title Priority Date Filing Date
PCT/JP2011/072283 WO2012043678A1 (en) 2010-09-30 2011-09-28 Image decoding device, image encoding device, and data structure

Country Status (1)

Country Link
WO (1) WO2012043678A1 (en)

Citations (3)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
JP2009111691A (en) * 2007-10-30 2009-05-21 Hitachi Ltd Image-encoding device and encoding method, and image-decoding device and decoding method
WO2009090884A1 (en) * 2008-01-18 2009-07-23 Panasonic Corporation Image encoding method and image decoding method
JP2010056701A (en) * 2008-08-27 2010-03-11 Nippon Telegr & Teleph Corp <Ntt> Intra-image predictive encoding method, intra-image predictive decoding method, devices thereof, programs therefor, and recording medium recording the programs

Patent Citations (3)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
JP2009111691A (en) * 2007-10-30 2009-05-21 Hitachi Ltd Image-encoding device and encoding method, and image-decoding device and decoding method
WO2009090884A1 (en) * 2008-01-18 2009-07-23 Panasonic Corporation Image encoding method and image decoding method
JP2010056701A (en) * 2008-08-27 2010-03-11 Nippon Telegr & Teleph Corp <Ntt> Intra-image predictive encoding method, intra-image predictive decoding method, devices thereof, programs therefor, and recording medium recording the programs

Non-Patent Citations (3)

* Cited by examiner, † Cited by third party
Title
EHSAN MAANI ET AL.: "Differential Coding of Intra Modes (DCIM)", JOINT COLLABORATIVE TEAM ON VIDEO CODING (JCT-VC) OF ITU-T SG16 WP3 AND ISO/IEC JTC1/SC29/WG11, JCTVC-B109, 2ND MEETING, July 2010 (2010-07-01), GENEVA, CH, pages 1 - 7 *
TEST MODEL UNDER CONSIDERATION: "Output Document (draft000)", JOINT COLLABORATIVE TEAM ON VIDEO CODING (JCT-VC) OF ITU-T SG16 WP3 AND ISO/IEC JTC1/SC29/WG11, JCTVC-B205, 2ND MEETING, July 2010 (2010-07-01), GENEVA, CH, pages 1 - 14,53-67,132-136 *
TOMOYUKI YAMAMOTO ET AL.: "Flexible Representation of Intra Prediction Modes", JOINT COLLABORATIVE TEAM ON VIDEO CODING (JCT-VC) OF ITU-T SG16 WP3 AND ISO/IEC JTC1/SC29/WG11, JCTVC-B063, 2ND MEETING, July 2010 (2010-07-01), GENEVA, CH, pages 1 - 7 *

Similar Documents

Publication Publication Date Title
US11627337B2 (en) Image decoding device
JP7200320B2 (en) Image filter device, filter method and moving image decoding device
US10547861B2 (en) Image decoding device
WO2017068856A1 (en) Predictive image generation device, image decoding device, and image encoding device
JP5551274B2 (en) Image filter device
WO2017195532A1 (en) Image decoding device and image encoding device
JP6114787B2 (en) Decoding device and encoding device
WO2015199040A1 (en) Dmm prediction unit, image decoding device, and image encoding device
JP5995448B2 (en) Image decoding apparatus and image encoding apparatus
WO2012077719A1 (en) Image decoding device and image coding device
WO2018110281A1 (en) Code prediction device, image decoding device, and image encoding device
JP2013192118A (en) Arithmetic decoder, image decoder, arithmetic encoder, and image encoder
WO2013046990A1 (en) Offset decoding apparatus, offset encoding apparatus, image filter apparatus, and data structure
US20210006836A1 (en) Image encoding apparatus, image encoding method, image decoding apparatus, and image decoding method
WO2012077795A1 (en) Image coding device, image decoding device, and data structure
WO2012090962A1 (en) Image decoding device, image encoding device, data structure of encoded data, arithmetic decoding device, and arithmetic encoding device
WO2012043678A1 (en) Image decoding device, image encoding device, and data structure
JP6162289B2 (en) Image decoding apparatus and image decoding method
WO2012043766A1 (en) Image decoding device, image encoding device, and data structure for encoded data
WO2012043676A1 (en) Decoding device, encoding device, and data structure
JP2016143962A (en) Region splitting image generating apparatus, image decoder, and encoder
JP2016072941A (en) Dmm prediction device, image decoding device and image encoding device
WO2012081706A1 (en) Image filter device, filter device, decoder, encoder, and data structure
JP2013251827A (en) Image filter device, image decoding device, image encoding device, and data structure
JP2016076904A (en) Dc offset information decoder, image decoder and dc offset information encoder

Legal Events

Date Code Title Description
121 Ep: the epo has been informed by wipo that ep was designated in this application

Ref document number: 11829231

Country of ref document: EP

Kind code of ref document: A1

NENP Non-entry into the national phase

Ref country code: DE

122 Ep: pct application non-entry in european phase

Ref document number: 11829231

Country of ref document: EP

Kind code of ref document: A1

NENP Non-entry into the national phase

Ref country code: JP