WO2012081636A1 - Image decoding device, image coding device, and data structure of coded data - Google Patents

Image decoding device, image coding device, and data structure of coded data Download PDF

Info

Publication number
WO2012081636A1
WO2012081636A1 PCT/JP2011/078953 JP2011078953W WO2012081636A1 WO 2012081636 A1 WO2012081636 A1 WO 2012081636A1 JP 2011078953 W JP2011078953 W JP 2011078953W WO 2012081636 A1 WO2012081636 A1 WO 2012081636A1
Authority
WO
WIPO (PCT)
Prior art keywords
image
channel
prediction
decoded
channels
Prior art date
Application number
PCT/JP2011/078953
Other languages
French (fr)
Japanese (ja)
Inventor
将伸 八杉
Original Assignee
シャープ株式会社
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Application filed by シャープ株式会社 filed Critical シャープ株式会社
Publication of WO2012081636A1 publication Critical patent/WO2012081636A1/en

Links

Images

Classifications

    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04NPICTORIAL COMMUNICATION, e.g. TELEVISION
    • H04N19/00Methods or arrangements for coding, decoding, compressing or decompressing digital video signals
    • H04N19/10Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using adaptive coding
    • H04N19/102Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using adaptive coding characterised by the element, parameter or selection affected or controlled by the adaptive coding
    • H04N19/103Selection of coding mode or of prediction mode
    • H04N19/105Selection of the reference unit for prediction within a chosen coding or prediction mode, e.g. adaptive choice of position and number of pixels used for prediction
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04NPICTORIAL COMMUNICATION, e.g. TELEVISION
    • H04N19/00Methods or arrangements for coding, decoding, compressing or decompressing digital video signals
    • H04N19/10Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using adaptive coding
    • H04N19/169Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using adaptive coding characterised by the coding unit, i.e. the structural portion or semantic portion of the video signal being the object or the subject of the adaptive coding
    • H04N19/186Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using adaptive coding characterised by the coding unit, i.e. the structural portion or semantic portion of the video signal being the object or the subject of the adaptive coding the unit being a colour or a chrominance component
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04NPICTORIAL COMMUNICATION, e.g. TELEVISION
    • H04N19/00Methods or arrangements for coding, decoding, compressing or decompressing digital video signals
    • H04N19/44Decoders specially adapted therefor, e.g. video decoders which are asymmetric with respect to the encoder
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04NPICTORIAL COMMUNICATION, e.g. TELEVISION
    • H04N19/00Methods or arrangements for coding, decoding, compressing or decompressing digital video signals
    • H04N19/46Embedding additional information in the video signal during the compression process
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04NPICTORIAL COMMUNICATION, e.g. TELEVISION
    • H04N19/00Methods or arrangements for coding, decoding, compressing or decompressing digital video signals
    • H04N19/50Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using predictive coding
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04NPICTORIAL COMMUNICATION, e.g. TELEVISION
    • H04N19/00Methods or arrangements for coding, decoding, compressing or decompressing digital video signals
    • H04N19/70Methods or arrangements for coding, decoding, compressing or decompressing digital video signals characterised by syntax aspects related to video coding, e.g. related to compression standards

Definitions

  • the present invention relates to an image decoding device that performs intra prediction on image color differences, an image encoding device, and a data structure of encoded data.
  • a moving image encoding device that generates encoded data by encoding the moving image, and a moving image that generates a decoded image by decoding the encoded data
  • An image decoding device is used.
  • Non-Patent Document 2 Specific examples of the moving image encoding method include H.264. H.264 / MPEG-4. Adopted in KTA software, which is a codec for joint development in AVC (Non-patent Document 1) and VCEG (Video Coding Expert Group), and TMuC (Test Model Under Consideration) software, which is the successor codec (Non-Patent Document 2).
  • a predicted image is usually generated based on a locally decoded image obtained by encoding / decoding an input image, and the predicted image is subtracted from the input image (original image).
  • the prediction residual (which may be referred to as “difference image” or “residual image”) is encoded.
  • examples of the method for generating a predicted image include inter-screen prediction (inter prediction) and intra-screen prediction (intra prediction).
  • TMuC Non-Patent Document 2 proposes to divide a moving image and manage it by the following hierarchical structure.
  • an image (picture) constituting a moving image is divided into slices.
  • the slice is also divided into maximum coding units (sometimes called macroblocks or maximum coding units (Largest (Coding Units)).
  • the maximum coding unit can be divided into smaller coding units (Coding Unit) by quadtree division.
  • a coding unit (leaf CU) that cannot be further divided is treated as a conversion unit and a prediction unit. These are sometimes called blocks.
  • a unit called a partition that uses a prediction unit as it is or is further divided is defined.
  • intra prediction is performed in units of partitions.
  • Non-Patent Document 1 an expression format in which a pixel is represented by a combination of a luminance component Y and color difference components C b and C r is employed.
  • Luminance and color difference are inherently independent components. Therefore, H.I. H.264 / MPEG-4.
  • AVC Non-Patent Document 1
  • horizontal prediction, vertical prediction, DC prediction, and plane prediction are used for intra prediction of color difference signals.
  • the human eye has a visual characteristic that it is sensitive to pixel luminance changes but insensitive to color changes.
  • the resolution of the chrominance pixel is lowered, there is less visual influence than when the resolution of the luminance pixel is lowered. Therefore, in the encoding of moving images, the resolution of color difference pixels is made lower than the resolution of luminance pixels to reduce the data amount.
  • FIG. 27 shows an example of a correspondence relationship between an original image, a luminance pixel, and a color difference pixel in the prior art.
  • 27A shows an image (YUV) to be encoded
  • FIG. 27B shows a luminance pixel
  • FIG. 27C shows a color difference pixel (exemplified for U). Show.
  • the resolution of the luminance pixel is 8 ⁇ 8
  • the color difference pixel has a resolution of 4 ⁇ 4.
  • Non-Patent Document 3 discloses predicting a color difference image (UV) from a luminance image (Y) by linear conversion. Specifically, performing linear transformation according to the following formula (A1) is disclosed.
  • PredC predicted image (color difference) [X C , y C ], [x Y , y Y ]: coordinates indicating the position of the same sample
  • RecY decoded image (luminance) ⁇ C , ⁇ C : coefficients derived by the least squares method from the pixel values of the surrounding decoded images (hereinafter referred to as local decoded images)
  • the coordinates of [x C , y C ] and [x Y , y Y ] need to be appropriately converted.
  • Non-Patent Document 3 also mentions how to take sample points when the resolution is different between the luminance image (Y) and the color difference image (U, V), as illustrated in FIG. .
  • FIG. 28A shows the case where sample points are taken from a 2N ⁇ 2N size luminance image
  • FIG. 28B shows the case where samples are taken from an N ⁇ N color difference image.
  • Non-Patent Document 3 describes that the sample point smpl100 shown in (a) of FIG. 28 is associated with the sample point smpl200 shown in (b) of FIG.
  • the conventional technology as described above performs linear conversion by the least square method in inter-channel prediction. For this reason, when the local decoded image used as a sample is not suitable for linear transformation, the accuracy of inter-channel prediction may not be sufficient.
  • the sample distribution is distributed in two regions of the group Gr101 and the group Gr102, and the sample distribution varies in each region.
  • the error may become large in the linear conversion.
  • the sample points Smpl300 are distributed at positions far away from the values approximated by the equation (A1).
  • the present invention has been made in view of the above-mentioned problems, and its purpose is to perform a prediction between channels in accordance with a linear correlation because each component of each pixel included in a locally decoded image varies.
  • an object of the present invention is to realize an image decoding apparatus capable of improving the possibility that higher prediction accuracy can be obtained even when the accuracy of prediction is lowered.
  • an image decoding apparatus generates a prediction image for each of a plurality of channels indicating each component constituting an image, and adds a prediction residual to the generated prediction image
  • channel decoding means for decoding one or a plurality of channels among the plurality of channels for the processing target block, and each of the plurality of channels has been decoded
  • a nonlinear correlation between the one or more channels that have been decoded by the channel decoding means and the other channels to be decoded is referred to with reference to a locally decoded image located around the block to be processed.
  • the correlation derivation means to be derived and the processing target block are restored according to the above derived correlation. From already the one or more channels of the decoded image, characterized in that it comprises, the predicted image generating means for generating the prediction image of the other channels.
  • a channel is a generalized component that constitutes an image.
  • the luminance component and the color difference component correspond to channels. That is, in this example, the channel includes a luminance channel and a color difference channel.
  • the color difference channel includes a U channel indicating the U component of the color difference and a V channel indicating the V component of the color difference.
  • the channel may relate to the RGB color space.
  • the image decoding process is performed for each channel.
  • the local decoded image is a decoded image that has been decoded for the plurality of channels and is located around the block to be processed.
  • Processing target block refers to various processing units in the decoding process.
  • a coding unit, a conversion unit, a prediction unit, and the like can be given.
  • the processing unit includes a unit obtained by further subdividing the encoding unit, the conversion unit, and the prediction unit.
  • the periphery of the processing target block includes, for example, a pixel adjacent to the target block, a block adjacent to the left side of the target block, a block adjacent to the upper side of the target block, and the like.
  • the non-linear correlation can be derived, for example, by examining the correspondence of each point composed of the luminance value and the color difference value.
  • the non-linear correlation for example, when describing the YUV color space, it can be derived from the correspondence between the luminance value of each pixel included in the locally decoded image and the color difference value.
  • the correlation may be realized as an LUT in which a decoded channel and a decoding target channel are associated with each other.
  • the correlation may be expressed by a function including a relational expression established between the decoded channel and the decoding target channel.
  • the channel to be decoded in the processing target block is predicted from the channel that has been decoded in the processing target block according to the nonlinear correlation derived in this way.
  • such prediction is also referred to as inter-channel prediction.
  • the pixel value of the decoded image of the decoded channel is converted according to the nonlinear correlation, and the pixel value of the predicted image of the channel to be decoded is obtained.
  • the pixel value is a generalized value of a component constituting the image.
  • an image encoding device encodes a prediction residual obtained by subtracting, from an original image, a prediction image generated for each of a plurality of channels indicating each component constituting the image.
  • a channel decoding unit that decodes one or a plurality of channels among the plurality of channels and a plurality of channels that have been decoded for the processing target block
  • a non-linear correlation between the one or more channels decoded by the channel decoding means and other channels to be decoded is derived with reference to a local decoded image located around the processing target block.
  • the correlation derivation means and the processing target block are decoded according to the derived correlation.
  • Mino from the decoded image of the one or more channels characterized in that it comprises a prediction image generating means for generating the prediction image of the other channels.
  • the data structure of the encoded data according to the present invention includes a prediction residual obtained by subtracting, from an original image, a prediction image generated for each of a plurality of channels indicating each component constituting the image.
  • a prediction residual obtained by subtracting, from an original image, a prediction image generated for each of a plurality of channels indicating each component constituting the image.
  • channel decoding processing order information indicating in which order the plurality of channels are decoded for the processing target block, and which of the decoded 1 above for the processing target block
  • the processing target that has been decoded for each of the plurality of channels from the decoded images of the plurality of channels
  • the predicted image of the other channel is determined according to a non-linear correlation between the one or more decoded channels and the other channel to be decoded.
  • prediction source channel designation information for designating whether to generate.
  • the same effects as those of the image decoding device according to the present invention can be obtained.
  • the image encoding device specifies channel decoding processing order information indicating in which order the plurality of channels are to be decoded, and from which of the decoded channels the channel to be decoded is to be predicted
  • the prediction source channel information may be included in the data structure of the encoded data.
  • the image encoding device may encode the information in, for example, side information.
  • the image decoding apparatus includes a channel decoding unit that decodes one or a plurality of channels among a plurality of channels, and a periphery of the processing target block that has been decoded for each of the plurality of channels.
  • a correlation deriving means for deriving a non-linear correlation between the one or more channels already decoded by the channel decoding means and the other channels to be decoded with reference to the local decoded image located at
  • the block includes a predicted image generating unit that generates the predicted image of the other channel from the decoded image of the one or more channels that has been decoded in accordance with the derived correlation.
  • the image encoding device includes a channel decoding unit that decodes one or more channels among a plurality of channels for the processing target block, and the processing target block that has been decoded for each of the plurality of channels.
  • Correlation deriving means for deriving a non-linear correlation between the one or more channels already decoded by the channel decoding means and other channels to be decoded with reference to locally decoded images located in the periphery, and processing
  • the target block includes a predicted image generating unit that generates the predicted image of the other channel from the decoded images of the one or more channels that have been decoded in accordance with the derived correlation.
  • the data structure of the encoded data according to the present invention is an image that decodes encoded image data by generating a prediction image for each of the plurality of channels and adding a prediction residual to the generated prediction image.
  • channel decoding processing order information indicating in which order the plurality of channels are decoded for the processing target block, and the decoded image of the one or more channels already decoded for the processing target block
  • Prediction specifying whether to generate the predicted image of the other channel according to the nonlinear correlation of And channel information, a data structure containing.
  • FIG. 4 is a diagram illustrating an image format of a YUV format, where (a) to (d) are a 4: 2: 0 format, a 4: 4: 4 format, a 4: 2: 2 format, and a 4: 1: 1 format, respectively.
  • FIG. 6 is a diagram illustrating sample positions of color difference pixels, and (a) to (c) show three sample positions. It is a figure shown about the pattern of the processing order of a color difference channel. It is the flowchart which illustrated about the schematic flow of the color difference estimated image generation process in the said estimated image generation part.
  • FIG. 4 is a diagram illustrating an example of an image in which luminance-color difference distributions do not overlap and extend over two regions.
  • 11 is a graph plotting luminance (Y) -color difference (U) of pixels included in the image shown in FIG.
  • FIG. 12 is a flowchart showing a modification of the flow of LUT derivation processing by the LUT derivation unit. It is a figure which shows an example of the derived
  • FIG. 5 is a graph plotting luminance (Y) -color difference (U) of an image, and two types of images are shown in (a) and (b).
  • (A) shows a transmitting apparatus equipped with a moving picture coding apparatus, and (b) shows a receiving apparatus equipped with a moving picture decoding apparatus. It is the figure shown about the structure of the recording device which mounts the said moving image encoder, and the reproducing
  • (A) shows a recording apparatus equipped with a moving picture coding apparatus, and (b) shows a reproduction apparatus equipped with a moving picture decoding apparatus.
  • FIG. 2 is a functional block diagram showing a schematic configuration of the moving picture decoding apparatus 1.
  • VCEG Video Coding Expert Group
  • KTA software which is a joint development codec
  • TMuC successor Codec
  • the moving image decoding apparatus 1 receives encoded data (data structure of encoded data) # 1 obtained by encoding a moving image by the moving image encoding apparatus 2.
  • the video decoding device 1 decodes the input encoded data # 1 and outputs the video # 2 to the outside.
  • the configuration of the encoded data # 1 will be described below.
  • the configuration of encoded data # 1 that is generated by the video encoding device 2 and decoded by the video decoding device 1 will be described with reference to FIG.
  • the encoded data # 1 has a hierarchical structure including a sequence layer, a GOP (Group Of Pictures) layer, a picture layer, a slice layer, and a maximum coding unit (LCU: Large Coding Unit) layer.
  • GOP Group Of Pictures
  • LCU Large Coding Unit
  • FIG. 3 shows the hierarchical structure below the picture layer in the encoded data # 1.
  • 3A to 3F show a picture layer PICT, a slice layer S, an LCU layer LCU, a leaf CU included in the LCU (denoted as CUL in FIG. 3D), and inter prediction (between screens).
  • PI_Inter which is the prediction information PI about (prediction) partition
  • PI_Intra which is the prediction information PI about intra prediction (prediction in a screen) partition.
  • the picture layer PICT is a set of data referred to by the video decoding device 1 in order to decode a target picture that is a processing target picture.
  • the picture layer PICT includes a picture header PH and slice layers S 1 to S NS (NS is the total number of slice layers included in the picture layer PICT).
  • NS is the total number of slice layers included in the picture layer PICT.
  • the reference numerals may be omitted. The same applies to other configurations included in the encoded data # 1.
  • the picture header PH includes a coding parameter group referred to by the video decoding device 1 in order to determine a decoding method of the target picture.
  • the encoding mode information (entoropy_coding_mode_flag) indicating the variable length encoding mode used in encoding by the moving image encoding device 2 is an example of an encoding parameter included in the picture header PH.
  • entorpy_coding_mode_flag When entorpy_coding_mode_flag is 0, the picture is encoded by CAVLC (Context-based Adaptive Variable Length Coding). It has become.
  • Each slice layer S included in the picture layer PICT is a set of data referred to by the video decoding device 1 in order to decode a target slice that is a processing target slice.
  • the slice layer S includes a slice header SH and LCU layers LCU 1 to LCU NC (NC is the total number of LCUs included in the slice S).
  • the slice header SH includes a coding parameter group that the moving image decoding apparatus 1 refers to in order to determine a decoding method of the target slice.
  • Slice type designation information (slice_type) for designating a slice type is an example of an encoding parameter included in the slice header SH.
  • the slice header SH includes a filter parameter FP that is referred to by a loop filter included in the video decoding device 1.
  • slice types that can be specified by the slice type specification information, (1) I slice using only intra prediction at the time of encoding, and (2) P using unidirectional prediction or intra prediction at the time of encoding. Slice, (3) B-slice using unidirectional prediction, bidirectional prediction, or intra prediction at the time of encoding.
  • Each LCU layer LCU included in the slice layer S is a set of data that the video decoding device 1 refers to in order to decode the target LCU that is the processing target LCU.
  • LCU layer LCU as shown in (c) of FIG. 3, LCU header LCUH, and a plurality of coding units obtained by the quadtree dividing the LCU: the (CU Coding Unit) CU 1 ⁇ CU NL Contains.
  • the size that each CU can take depends on the LCU size and the hierarchical depth included in the sequence parameter set SPS of the encoded data # 1. For example, when the size of the LCU is 128 ⁇ 128 pixels and the maximum hierarchical depth is 5, the CU included in the LCU has five types of sizes, that is, 128 ⁇ 128 pixels, 64 ⁇ 64 pixels, Any of 32 ⁇ 32 pixels, 16 ⁇ 16 pixels, and 8 ⁇ 8 pixels can be taken. A CU that is not further divided is called a leaf CU.
  • the LCU header LCUH includes an encoding parameter referred to by the video decoding device 1 in order to determine a decoding method of the target LCU. Specifically, as shown in FIG. 3C, CU partition information SP_CU that specifies a partition pattern for each leaf CU of the target LCU, and a quantization parameter difference ⁇ qp that specifies the size of the quantization step. (Mb_qp_delta) is included.
  • CU division information SP_CU is information that specifies the shape and size of each CU (and leaf CU) included in the target LCU, and the position in the target LCU.
  • the CU partition information SP_CU does not necessarily need to explicitly include the shape and size of the leaf CU.
  • the CU partition information SP_CU may be a set of flags (split_coding_unit_flag) indicating whether or not the entire LCU or a partial region of the LCU is divided into four. In that case, the shape and size of each leaf CU can be specified by using the shape and size of the LCU together.
  • the quantization parameter difference ⁇ qp is a difference qp ⁇ qp ′ between the quantization parameter qp in the target LCU and the quantization parameter qp ′ in the LCU encoded immediately before the LCU.
  • Leaf CU A CU (leaf CU) that cannot be further divided is treated as a prediction unit (PU: Prediction Unit) and a transform unit (TU: Transform Unit).
  • PU Prediction Unit
  • TU Transform Unit
  • the leaf CU (denoted as CUL in (d) of FIG. 3) is (1) PU information PUI that is referred to when the moving image decoding apparatus 1 generates a predicted image. And (2) the TU information TUI that is referred to when the moving image decoding apparatus 1 decodes the residual data.
  • the PU information PUI may include a skip flag SKIP. When the value of the skip flag SKIP is 1, the TU information is omitted.
  • the PU information PUI includes prediction type information PT and prediction information PI, as shown in FIG.
  • the prediction type information PT is information that specifies whether intra prediction or inter prediction is used as a predicted image generation method for the target leaf CU (target PU).
  • the prediction information PI includes intra prediction information PI_Intra or inter prediction information PI_Inter depending on which prediction method is specified by the prediction type information PT.
  • a PU to which intra prediction is applied is also referred to as an intra PU
  • a PU to which inter prediction is applied is also referred to as an inter PU.
  • the PU information PUI includes information specifying the shape and size of each partition included in the target PU and the position in the target PU.
  • the partition is one or a plurality of non-overlapping areas constituting the target leaf CU, and the generation of the predicted image is performed in units of partitions.
  • the TU information TUI includes TU partition information SP_TU that specifies a partition pattern for each block of the target leaf CU (target TU), and quantized prediction residuals QD 1 to QD NT.
  • target TU the target leaf CU
  • QD 1 to QD NT the total number of blocks included in the target TU
  • TU partition information SP_TU is information that specifies the shape and size of each block included in the target TU and the position in the target TU.
  • Each TU can be, for example, a size from 64 ⁇ 64 pixels to 2 ⁇ 2 pixels.
  • the block is one or a plurality of non-overlapping areas constituting the target leaf CU, and prediction residual encoding / decoding is performed in units of TUs or blocks obtained by dividing TUs.
  • Each quantized prediction residual QD is encoded data generated by the moving image encoding apparatus 2 performing the following processes 1 to 3 on a target block that is a processing target block.
  • Process 1 DCT transform (Discrete Cosine Transform) is performed on the prediction residual obtained by subtracting the prediction image from the encoding target image.
  • Process 2 The DCT coefficient obtained in Process 1 is quantized.
  • Process 3 The DCT coefficient quantized in Process 2 is variable length encoded.
  • the inter prediction information PI_Inter includes a coding parameter that is referred to when the video decoding device 1 generates an inter prediction image by inter prediction. As shown in FIG. 3E, the inter prediction information PI_Inter includes inter PU partition information SP_Inter that specifies a partition pattern of the target PU into each partition, and inter prediction parameters PP_Inter1 to PP_InterNe (Ne for each partition). , The total number of inter prediction partitions included in the target PU).
  • the inter-PU partition information SP_Inter is information for designating the shape and size of each inter prediction partition included in the target PU (inter PU) and the position in the target PU.
  • the inter PU is composed of four symmetric splittings of 2N ⁇ 2N pixels, 2N ⁇ N pixels, N ⁇ 2N pixels, and N ⁇ N pixels, and 2N ⁇ nU pixels, 2N ⁇ nD pixels, and nL ⁇ 2N. It is possible to divide into 8 types of partitions in total by four asymmetric splits of pixels and nR ⁇ 2N pixels.
  • the specific value of N is defined by the size of the CU to which the PU belongs, and the specific values of nU, nD, nL, and nR are determined according to the value of N.
  • an inter PU of 128 ⁇ 128 pixels is 128 ⁇ 128 pixels, 128 ⁇ 64 pixels, 64 ⁇ 128 pixels, 64 ⁇ 64 pixels, 128 ⁇ 32 pixels, 128 ⁇ 96 pixels, 32 ⁇ 128 pixels, and 96 ⁇ It is possible to divide into 128-pixel inter prediction partitions.
  • the inter prediction parameter PP_Inter includes a reference image index RI, an estimated motion vector index PMVI, and a motion vector residual MVD.
  • the intra prediction information PI_Intra includes an encoding parameter that is referred to when the video decoding device 1 generates an intra predicted image by intra prediction.
  • the intra prediction information PI_Intra includes intra PU partition information SP_Intra that specifies a partition pattern of the target PU (intra PU) into each partition, and intra prediction parameters PP_Intra 1 for each partition.
  • -PP_Intra NA NA is the total number of intra prediction partitions included in the target PU).
  • the intra-PU partition information SP_Intra is information that specifies the shape and size of each intra-predicted partition included in the target PU, and the position in the target PU.
  • the intra PU split information SP_Intra includes an intra split flag (intra_split_flag) that specifies whether or not the target PU is split into partitions. If the intra partition flag is 1, the target PU is divided symmetrically into four partitions. If the intra partition flag is 0, the target PU is not divided and the target PU itself is one partition.
  • N 2 n , n is an arbitrary integer of 1 or more.
  • a 128 ⁇ 128 pixel intra PU can be divided into 128 ⁇ 128 pixel and 64 ⁇ 64 pixel intra prediction partitions.
  • the video decoding device 1 generates a predicted image for each partition, generates a decoded image # 2 by adding the generated predicted image and the prediction residual decoded from the encoded data # 1, and generates The decoded image # 2 is output to the outside.
  • the generation of the predicted image is performed with reference to the encoding parameter obtained by decoding the encoded data # 1.
  • the encoding parameter is a parameter referred to in order to generate a prediction image, and in addition to a prediction parameter such as a motion vector referred to in inter-screen prediction and a prediction mode referred to in intra-screen prediction. Partition size and shape, block size and shape, and residual data between the original image and the predicted image.
  • side information a set of all information excluding the residual data among the information included in the encoding parameter is referred to as side information.
  • the prediction unit is a partition constituting the LCU
  • the present embodiment is not limited to this, and the prediction unit is a unit larger than the partition.
  • the present invention can also be applied to the case where the prediction unit is a unit smaller than the partition.
  • a frame (picture), a slice, an LCU, a block, and a partition to be decoded are referred to as a target frame, a target slice, a target LCU, a target block, and a target partition, respectively.
  • the LCU size is, for example, 64 ⁇ 64 pixels
  • the partition size is, for example, 64 ⁇ 64 pixels, 32 ⁇ 32 pixels, 16 ⁇ 16 pixels, 8 ⁇ 8 pixels, 4 ⁇ 4 pixels, or the like. These sizes do not limit the present embodiment, and the size and partition of the LCU may be other sizes.
  • FIG. 2 is a functional block diagram showing a schematic configuration of the moving picture decoding apparatus 1.
  • the moving picture decoding apparatus 1 includes a variable length code demultiplexing unit 11, an inverse quantization / inverse conversion unit 12, a predicted image generation unit (channel decoding unit) 13, and an adder (channel decoding unit) 14. And a frame memory 15.
  • variable-length code demultiplexing unit 11 demultiplexes the encoded data # 1 for one frame input to the video decoding device 1 to obtain various kinds of information included in the hierarchical structure shown in FIG. To separate.
  • the variable length code demultiplexing unit 11 refers to information included in various headers, and sequentially separates the encoded data # 1 into slices and LCUs.
  • the various headers include (1) information on the method of dividing the target frame into slices, and (2) information on the size, shape, and position of the LCU belonging to the target slice.
  • variable length code demultiplexer 11 refers to the CU partition information SP_CU included in the encoded LCU header LCUH and divides the target LCU into leaf CUs. In addition, the variable-length code demultiplexing unit 11 acquires the TU information TUI and the PU information PUI for the target leaf CU: CUL.
  • variable length code demultiplexing unit 11 supplies the TU information TUI obtained for the target leaf CU to the dequantization / inverse transform unit 12. Further, the variable length code demultiplexing unit 11 supplies the PU information PUI obtained for the target leaf CU to the predicted image generation unit 13.
  • the inverse quantization / inverse transform unit 12 performs inverse quantization / inverse transform of the quantization prediction residual for each block for the target leaf CU.
  • the inverse quantization / inverse transform unit 12 first decodes the TU partition information SP_TU from the TU information TUI about the target leaf CU supplied from the variable length code demultiplexer 11.
  • the inverse quantization / inverse transform unit 12 divides the target leaf CU into one or a plurality of blocks according to the decoded TU partition information SP_TU.
  • the inverse quantization / inverse transform unit 12 decodes the TU partition information SP_TU and the quantized prediction residual QD from the TU information TUI for each block.
  • the inverse quantization / inverse transform unit 12 restores the prediction residual D for each pixel for each target partition by performing inverse quantization and inverse DCT transform (Inverse DiscretecreCosine Transform).
  • the inverse quantization / inverse transform unit 12 supplies the restored prediction residual D to the adder 14.
  • the predicted image generation unit 13 For each partition included in the target leaf CU, the predicted image generation unit 13 refers to a local decoded image P ′ that is a decoded image around the partition, and generates a predicted image Pred by intra prediction or inter prediction.
  • intra prediction and inter prediction include luminance prediction and color difference prediction, respectively.
  • the predicted image generation unit 13 refers to a luminance decoded image.
  • the video decoding device 1 may generate the predicted image Pred by inter prediction.
  • Intra prediction is sometimes referred to as intra prediction or spatial prediction, but in the following, it is unified with the expression intra prediction.
  • the local decoded image P ′ includes a luminance local decoded image P ′ Y related to luminance and a color difference local decoded image P ′ C related to color difference.
  • the predicted image generation unit 13 operates as follows. First, the predicted image generation unit 13 decodes the PU information PUI for the target leaf CU supplied from the variable length code demultiplexing unit 11. Subsequently, the predicted image generation unit 13 determines a division pattern for each partition of the target leaf CU according to the PU information PUI. Further, the predicted image generation unit 13 selects a prediction mode of each partition according to the PU information PUI, and assigns each selected prediction mode to each partition.
  • the predicted image generation unit 13 generates a predicted image Pred for each partition included in the target leaf CU with reference to the selected prediction mode and the pixel values of the local decoded image P ′ around the partition.
  • the predicted image generation unit 13 supplies the predicted image Pred generated for the target leaf CU to the adder 14.
  • the predicted image Pred specifically includes a luminance predicted image PredY related to luminance and a color difference predicted image PredC related to color difference.
  • the color difference prediction image PredC includes a color difference prediction image PredU for the U channel and a color difference prediction image PredV for the V channel. Further, a more specific configuration of the predicted image generation unit 13 will be described later.
  • the adder 14 adds the predicted image Pred supplied from the predicted image generation unit 13 and the prediction residual D supplied from the inverse quantization / inverse transform unit 12, thereby decoding the decoded image P for the target leaf CU. Is generated.
  • Decoded image P includes, luminance decoded image (hereinafter, referred to by the luminance decoded picture P Y) with a color difference decoded image.
  • the decoded image P that has been decoded is sequentially recorded in the frame memory 15.
  • decoded images corresponding to all the LCUs decoded before the target LCU are recorded. .
  • the moving image decoding apparatus 1 one frame of encoded data # 1 input to the moving image decoding apparatus 1 at the time when the decoded image generation processing for each LCU is completed for all the LCUs in the image.
  • the decoded image # 2 corresponding to is output to the outside.
  • intra prediction parameter PP data structure of encoded data
  • the intra prediction parameter PP (data structure of encoded data) illustratively includes an inter-channel prediction flag, a color difference channel processing order flag, and a second channel prediction source channel specifier.
  • the inter-channel prediction flag is a flag indicating whether or not the color difference is predicted by inter-channel prediction. For example, if the inter-channel prediction flag is “1”, it indicates that the color difference is predicted by the inter-channel prediction, and if “0”, 1 bit indicates that the color difference is predicted without using the inter-channel prediction. Information.
  • the color difference channel processing order flag is a flag for designating whether the prediction process is performed from the U channel or the V channel.
  • the color difference channel processing order flag is, for example, 1-bit information indicating that processing is performed in the order of U and V if “0” and in the order of V and U if “1”.
  • the second channel prediction source channel specifier is information for designating which channel the second predicted channel is predicted from. That is, the second predicted channel can be predicted from either the Y channel, the first predicted first channel, or both. For example, if the second channel prediction source channel specifier is “0”, it indicates that prediction is performed from the Y channel, if “10”, it indicates that prediction is performed from the first channel, and “11”. If it is, it is 1 or 2-bit information indicating that prediction is performed from both the Y channel and the first channel.
  • the information indicating the combination of the processing order and the prediction source channel is encoded as separate information of the color difference channel processing order flag and the second channel prediction source channel specifier, respectively.
  • FIG. 4 is a diagram for explaining the YUV color space, and shows the relationship between the luminance Y and the U component and V component, which are the components of the color difference.
  • FIG. 5 is a diagram showing an image format of YUV format. (A) to (d) are respectively 4: 2: 0 format, 4: 4: 4 format, 4: 2: 2 format, and 4: 1: 1 format is shown.
  • the YUV color space will be described with reference to FIG.
  • an image is expressed by luminance Y and U and V components that are color differences.
  • the luminance Y is defined as an independent coordinate system orthogonal to the U-V plane with respect to the U component and the V component.
  • the luminance Y, U component, and V component each take a value from 0 to 255.
  • the image When the color difference value of the U component is close to 0, the image is generally green, and when it is close to 255, the image is generally red. In addition, when the color difference value of the V component is close to 0, the image is generally yellow, and when it is close to 255, the image is generally blue.
  • the types of pixel values used are limited.
  • the luminance Y has a correlation with each of the U component and the V component. Therefore, locally, it is possible to derive the U component and the V component from the luminance Y using this correlation.
  • the luminance Y is called the Y channel
  • the color difference consisting of the U component and the V component is called the color difference channel.
  • a channel is a generalized concept of luminance Y, U component, and V component.
  • U component and the V component when it is necessary to distinguish the U component and the V component in the color difference channel, they are referred to as a U channel and a V channel, respectively.
  • Prediction of the U channel and the V channel from the Y channel using the correlation between the luminance Y and the U component and the V component is referred to as inter-channel prediction.
  • the image format of luminance and color difference will be described as follows. Even if the resolution of the color difference is lowered, the visual effect is less than when the luminance resolution is lowered. Therefore, the data amount can be reduced by reducing the resolution of the color difference.
  • the data amount is reduced by the following data structure. Note that the left block of FIG. 5A shows the resolution of the luminance Y, and the right block shows the resolution of the U component and the V component. The same applies to (b) to (d) below.
  • the resolution of the color difference is 1 ⁇ 2 of the luminance resolution in both the horizontal and vertical directions. That is, as a whole, the resolution of the color difference is 1/4 of the resolution of the luminance.
  • the 4: 2: 0 format is used in television broadcasting and consumer video equipment.
  • the luminance resolution and the color difference resolution are the same.
  • the 4: 4: 4 format is used, for example, in specialized equipment for image processing when high image quality is required rather than reducing the amount of data.
  • FIG. 1 is a functional block diagram illustrating an example of the configuration of the predicted image generation unit 13.
  • the predicted image generation unit 13 includes a local image input unit 131, a luminance predicted image generation unit 132, an inter-channel prediction determination unit 133, an LUT derivation unit (correlation derivation unit) 134, and a color difference prediction image generation unit ( A prediction image generation unit, a processing information acquisition unit, a prediction control unit) 135, and a prediction image output unit 136.
  • the local image input unit 131 acquires the luminance local decoded image P ′ Y and the color difference local decoded image P ′ C from the local decoded image P ′.
  • the local image input unit 131 transfers the luminance local decoded image P ′ Y to the luminance predicted image generation unit 132 and transfers the color difference local decoded image P ′ C to the inter-channel prediction determination unit 133.
  • Brightness prediction image generation unit 132 refers to the luminance local decoded image P 'Y, when predictions are based on PU information PUI, it generates a brightness prediction image PredY.
  • the predicted brightness image generation unit 132 transmits the generated predicted brightness image PredY to the predicted image output unit 136.
  • the inter-channel prediction determination unit 133 refers to an inter-channel prediction flag included in the intra-prediction parameter PP, and a channel in which intra prediction of color difference (hereinafter simply referred to as color difference prediction) generates a color difference prediction image by inter-channel prediction. It is determined whether or not it is an inter prediction mode.
  • the inter-channel prediction determination unit 133 informs the LUT derivation unit 134 and the inter-channel prediction unit 351 (described later) of the color difference prediction image generation unit 135 that it is the inter-channel prediction mode. Notice. As a result of the determination, if it is not between channels prediction mode, the prediction determining section 133 among the channels, and transfers the color difference local decoded image P 'C channel prediction unit 352 of the color difference prediction image generation unit 135 (described later).
  • LUT deriving unit 134 for each target partition, it derived based an LUT (Look Up Table) for performing inter-channel prediction chrominance local decoded image P 'Y.
  • the LUT derived by the LUT deriving unit is illustratively structured as follows. That is, the LUT, the pixel position of the color difference local decoded image P 'Y [x Y, y Y] in association with the luminance value in the above pixel position [x Y, y Y] corresponding to the color difference local decoded image P 'pixel positions C [x C, y C] chrominance values in is stored.
  • the LUT deriving unit 134 transmits the derived LUT to the inter-channel prediction unit 351 (predicted image generation means; described later) of the color difference predicted image generation unit 135. Details of the operation of the LUT deriving unit 134 will be described later.
  • the color difference predicted image generation unit 135 predicts a color difference image and generates a color difference predicted image PredC. More specifically, the color difference predicted image generation unit 135 includes an inter-channel prediction unit 351 and an intra-channel prediction unit 352.
  • Inter-channel prediction unit 351 at the time the color difference prediction is a prediction mode among channels, with reference to the luminance decoded image P Y, to generate the color difference prediction image PredC performs prediction of the color difference images by inter-channel prediction.
  • the channel prediction unit 352 at the time the color difference prediction not prediction mode among channels, with reference to the chrominance local decoded image P 'C, to produce a color difference prediction image PredC performs prediction of the color difference images.
  • the prediction of the color difference image by the intra-channel prediction unit 352 is performed by, for example, direction prediction or DC prediction.
  • inter-channel prediction unit 351 Details of the operation of the inter-channel prediction unit 351 will be described later.
  • the predicted image output unit 136 outputs the luminance predicted image PredY generated by the luminance predicted image generating unit 132 and the color difference predicted image PredC generated by the color difference predicted image generating unit 135 as the predicted image Pred.
  • FIG. 6 is a diagram illustrating a correspondence relationship between luminance and color difference pixel positions in the 4: 2: 0 format.
  • A) of FIG. 6 shows the pixel position of the luminance decoded image P Y
  • B) shows the pixel position of the color difference prediction image PredU to be predicted.
  • Pixel positions of the luminance decoded image P Y and the color difference prediction image PredU are both represented by relative coordinates with the origin at the upper left of the block.
  • FIGS. 7A to 7B are diagrams illustrating sample positions of three color difference pixels.
  • FIG. 7A shows the case already described with reference to FIG. That is, as shown in FIG. 7A, the sample position of the color difference pixel may be set at the upper left of the block.
  • the sample position of the color difference pixel may be set to the left from the center of the block.
  • the sample position of the color difference pixel may be set at the center of the block.
  • the luminance value corresponding to the value of a certain color difference pixel is derived by filtering the luminance value in the vicinity of the pixel position obtained in accordance with the correspondence relationship shown in (a) to (c) of FIG.
  • the vicinity of the pixel position is a coordinate obtained when the value of each coordinate of the pixel position obtained from the correspondence relation is rounded up and down. That is, in the example shown in FIG. 7B, [x Y , y Y ] and [x Y , y Y +1]. Further, in the example shown in FIG. 7 (c), in [x Y, y Y], [x Y + 1, y Y], [x Y, y Y +1] and [x Y + 1, y Y +1] is there.
  • the luminance value at the pixel position [x Y , y Y ] may be used as a luminance value as a sample as it is. Good.
  • a smoothing filter is mentioned as an example of this filtering.
  • FIG. 8 is a diagram showing a pattern of the color difference channel processing order when there is one prediction source channel.
  • the U channel and the V channel can be predicted using the luminance (Y) channel as a base point, as shown in FIG.
  • Y luminance
  • FIG. 8 there are the following three prediction patterns.
  • the first is a pattern indicated by a solid line in FIG. That is, while predicting the U channel from the Y channel, the V channel is predicted from the Y channel. Which of the U channel and the V channel is predicted first can be arbitrarily selected. Further, the U channel and V channel prediction processing may be performed in parallel.
  • the second is a pattern indicated by a dotted line in FIG. That is, the U channel is first predicted from the Y channel, and then the V channel is predicted from the predicted U channel.
  • the third is the reverse of the second, and is a pattern indicated by a broken line in FIG. That is, the V channel is first predicted from the Y channel, and then the U channel is predicted from the predicted V channel.
  • the chrominance channel processing order flag indicating which processing order is used, and the second channel prediction source channel specifier that specifies the prediction source channel of the second prediction channel are the codes in the moving picture encoding device 2. Encoded in the encoding process and transmitted to the video decoding device 1. Then, the inter-channel prediction unit 351 of the video decoding device 1 performs inter-channel prediction according to the color difference channel processing order flag and the second channel prediction source channel specifier.
  • inter-channel prediction unit 35 in accordance with the following equation (1), by performing the inter-channel prediction of the U channel, and generates the color difference prediction image PredU from the luminance decoded image P Y.
  • PredV [x V, y V] LUT V [RecY [x Y, y Y]] ... (2)
  • the meaning of each symbol in the formula (2) is the same as that in the formula (1), and the description thereof is omitted.
  • FIG. 9 is a flowchart illustrating a schematic flow of color difference predicted image generation processing in the predicted image generation unit 13.
  • the inter-channel prediction determination unit 133 refers to the inter-channel prediction flag and determines whether or not the inter-channel prediction mode is set (S10).
  • the intra-channel prediction unit 352 generates the color difference prediction image PredC without depending on the inter-channel prediction (S11), and the process ends.
  • LUT deriving unit 134 derives the LUT with reference to the luminance decoded image P Y (S12). Then, the inter-channel prediction unit 351 generates a color difference prediction image PredC by inter-channel prediction with reference to the LUT derived by the LUT deriving unit 134 (S13). This is the end of the process.
  • FIG. 10 is a diagram showing an example of an image B in which the luminance-color difference distributions do not overlap and extend over two regions.
  • the image B shown in FIG. 10 is composed of six pixel areas.
  • the pixel regions R1 to R3 are regions where the luminance value (Y) is low and the color difference value (U) is high.
  • the pixel regions R4 to R6 are regions having a high luminance value (Y) and a low color difference value (U).
  • the color difference value (U) gradually increases from the pixel region R1 to R3.
  • the color difference value (U) decreases from the pixel region R4 to R6.
  • a plot of the luminance (Y) -color difference (U) of the pixels included in such an image B is shown in the graph of FIG.
  • the pixels included in the pixel regions R1 to R3 are plotted in the group Gr1 in the graph shown in FIG.
  • the luminance (Y) of the sample is low and the color difference value (U) is high.
  • the pixels included in the pixel regions R4 to R6 are plotted in the group Gr2 in the graph shown in FIG.
  • the luminance (Y) of the sample is high and the color difference value (U) is low.
  • Image B is not suitable for linear approximation because there are variations in brightness and color difference in the image. Such variations tend to be often seen in images that include boundaries between multiple objects or textures of multiple colors.
  • FIG. 12 is a flowchart showing an example of the flow of LUT derivation processing by the LUT derivation unit 134.
  • the LUT derivation unit 134 initializes the LUT (S100). To initialize the LUT is to make the LUT unregistered.
  • the LUT deriving unit 134 enters a loop LP11 of registration processing for each luminance pixel adjacent to the target partition (S101).
  • the LUT deriving unit 134 determines whether or not a color difference value is registered in the LUT [n] (S103).
  • the LUT deriving unit 134 registers the color difference value m acquired in Step S102 as it is in the LUT [n]. The process returns to the beginning of the loop LP11 (S106).
  • the LUT derivation unit 134 calculates (m + LUT [n] +1) / 2, thereby obtaining the acquired color difference value m and the registered color difference value. And the average value is calculated. The LUT deriving unit 134 substitutes the average value thus calculated for the color difference value m (S104). Subsequently, the LUT deriving unit 134 registers the color difference value m into which the average value is substituted in step S104 in the LUT [n] (S105), and returns to the top of the loop LP11 (S106).
  • the LUT deriving unit 134 determines whether or not LUT [n] is unregistered (S108).
  • the LUT deriving unit 134 searches for the latest registered entries before and after n as shown in FIG. 13 (S109). Specifically, the LUT deriving unit 134 searches for the latest registered entry before and after n as follows.
  • the LUT deriving unit 134 searches for a registered entry for n forward, that is, nL smaller than n, with n as a base point. That is, with reference to FIG. 13, a search is made for a sample point Smpl1 in the n forward direction.
  • the LUT deriving unit 134 searches for an entry registered for n backward, that is, nR larger than n, with n as a base point. That is, with reference to FIG. 13, a search is made for a sample point Smpl2 in the backward direction of n.
  • the LUT derivation unit 134 registers a value obtained by linear interpolation between the LUT [nL] and the LUT [nR] in the LUT [n] (S110).
  • step S110 an interpolation process is performed to connect the nearest sample point Smpl1 ahead of n and the sample point Smpl2 behind n by a straight line L1. Yes.
  • the value at n of the straight line L1 is registered in LUT [n].
  • step S104 if an entry has already been registered in the registration process, the average value of the acquired color difference value m and the registered color difference value is registered as a new entry. This is due to the following reason.
  • step S104 if the acquired color difference value m is overwritten with the registered entry value, if the acquired color difference value m is statistical noise, the noise is directly registered in the entry. It becomes. On the other hand, such noise can be reduced by acquiring an average value of the acquired color difference value m and the registered color difference value.
  • the average value acquired here is a weighted average in the order of registration. Further, the present invention is not limited to this, and in step S104, a weighted average in which weights are arbitrarily set may be acquired.
  • the moving average process may be performed on the entire table. By performing the moving average process on the entire table, a rapid change in color difference can be suppressed.
  • FIG. 14 shows a graph of the LUT values derived from the image B, that is, entries.
  • a graph L11 shown in FIG. 14 shows LUT entry values derived from the image B.
  • the graph L11 is a line passing through each sample point.
  • the entry between each sample point is created by linear interpolation. For example, a straight line is connected between the leftmost sample of the group Gr1 and the leftmost sample of the group Gr2.
  • an average value is taken for sample points where n values overlap, and this value is used as an actual sample point for inter-channel prediction.
  • an unregistered entry that is, an entry between each sample point
  • the present invention is not limited to this.
  • the entry may be created by cubic interpolation. Thereby, the prediction accuracy of the table can be improved.
  • FIG. 15 is a flowchart illustrating an example of a flow of color difference prediction image generation processing by inter-channel prediction in the inter-channel prediction unit 351.
  • the inter-channel prediction unit 351 sets the color difference A and color difference A prediction source channels and the color difference B and color difference B prediction source channels according to the intra prediction parameter PP (S120). .
  • the contents set by the inter-channel prediction unit 351 are specifically as follows. That is, the inter-channel prediction unit 351 sets the color difference A and the color difference B according to the color difference channel processing order flag. In the following, the color difference B is processed after the color difference A. For example, the inter-channel prediction unit 351 may set the color difference A as “U channel” and the color difference B as “V channel”.
  • the inter-channel prediction unit 351 sets “Y channel” as the color difference A prediction source channel.
  • the inter-channel prediction unit 351 sets a color difference B prediction source channel according to the second channel prediction source channel specifier. For example, the inter-channel prediction unit 351 may set the prediction source channel of the U channel with the color difference A as “Y channel” and set the prediction source channel of the V channel with the color difference B as “Y channel and U channel”. Good.
  • the inter-channel prediction unit 351 selects a color difference A prediction mode (S121), and generates a color difference prediction image by inter-channel prediction for the color difference A according to the setting in step S120 (S122).
  • the inter-channel prediction unit 351 selects a color difference B prediction mode (S123), and generates a color difference prediction image by inter-channel prediction for the color difference B according to the setting in step S120 (S124). The process ends as described above.
  • the video decoding device 1 generates the prediction image Pred for each of the luminance (Y) channel and the color difference (U, V) channel, and adds the prediction residual D to the generated prediction image Pred.
  • the prediction image generation unit 13 and the adder 14 that decode the luminance (Y) channel and generate the luminance decoded image PY for the target partition, and the target partition
  • An LUT deriving unit 134 for deriving a non-linear correlation between a decoded luminance (Y) channel and a color difference (U, V) channel to be decoded as an LUT with reference to a local decoded image P ′ located in the vicinity. If, for the target partition, according to the LUT, inter-channel pre from luminance decoded image P Y, to generate the color difference prediction image PredC It includes a section 351, a.
  • the luminance component of each pixel included in the locally decoded image P ′ varies, and when prediction between channels is performed according to a linear correlation, the prediction accuracy is reduced. Even if it exists, there exists an effect that possibility that a higher prediction precision can be improved can be improved.
  • FIG. 16 and FIG. 17 are diagrams showing the pattern of the color difference channel processing order when there are two prediction source channels for the second channel to be predicted.
  • the U channel may be predicted using the Y channel as a base point, and then (2) the V channel may be predicted based on a combination of the Y channel and the U channel. Also, as shown in FIG. 17, (1) the V channel may be predicted using the Y channel as a base point, and then (2) the U channel may be predicted based on a combination of the Y channel and the V channel.
  • the LUT may be extended to two dimensions. Even when the LUT is extended to two dimensions, the same technique as the one-dimensional LUT derivation can be adopted, and a known technique used when creating a two-dimensional LUT can be adopted. Is possible.
  • the LUT deriving unit 134 derives the LUT as follows. That is, first, a primary table is derived for the luminance value Y and the color difference value U of the sample points as described above. Subsequently, the color difference value V of the sample point is registered in the table in association with the luminance Y and the color difference value U of the sample point. The same applies to the processing order shown in FIG.
  • the LUT can be configured to extend to more than two dimensions.
  • Another example of a configuration that expands the LUT to two or more dimensions includes a configuration that uses luminance values Y in a plurality of pixels. More specifically, in the LUT entry, the color difference value U or the color difference value V may be looked up by a combination of adjacent luminance values.
  • the color difference can be predicted from a plurality of channels or from a plurality of pixel values of the same channel by using a two-dimensional or more LUT as described above, the accuracy of the predicted image can be improved.
  • the intra prediction parameter PP is illustratively configured to include the inter-channel prediction flag, the color difference channel processing order flag, and the second channel prediction source channel specifier.
  • the intra prediction parameter PP may be configured to include an inter-channel prediction index indicating a combination of processing order and prediction source channels as follows.
  • the above-mentioned inter-channel prediction index is encoded and transmitted to the video decoding device 1 in the encoding process in the video encoding device 2. Then, the inter-channel prediction unit 351 of the video decoding device 1 performs inter-channel prediction according to the inter-channel prediction index transmitted from the video encoding device 2.
  • the intra prediction parameter PP indicating the combination of the processing order and the prediction source channel is included in the intra prediction parameter PP, but the present invention is not limited to this.
  • information indicating the combination of the processing order and the prediction source channel may be stored in the header of the processing unit other than the slice header, and the processing order or the like may be changed according to the processing unit.
  • information indicating a combination of processing order and prediction source channel may be stored in a sequence header (SPS: Sequence Parameter Set) or a picture header (PPS: Picture Parameter Set).
  • processing order or the like may be changed in units of processing smaller than slices, for example, in units of LCUs.
  • the LUT derivation unit can be changed within a range in which there is a correlation between luminance and color difference.
  • the information indicating the combination of the processing order and the prediction source channel may be encoded in different processing units.
  • the information indicating the processing order may be encoded in LCU units
  • the information indicating the combination of prediction source channels may be encoded in PU units.
  • PredU [x U , y U ] is calculated as follows.
  • PredU [x U , y U ] LUT [RecY [x Y , y Y ] / 2]
  • PredU [x U , y U ] (LUT [RecY [x Y, y Y] / 2] + LUT [RecY [x Y, y Y] / 2 + 1]) / 2
  • FIG. 18 is a flowchart showing a modified example of the flow of the LUT derivation process by the LUT derivation unit 134.
  • the LUT derivation unit 134 initializes the LUT (S130).
  • the LUT deriving unit 134 enters a registration processing loop LP11A for each luminance pixel adjacent to the target partition (S131).
  • the LUT deriving unit 134 substitutes sum [n] + m for sum [n] and substitutes count [n] +1 for count [n] (S133).
  • the LUT deriving unit 134 determines whether count [n] is greater than 0 (S136).
  • count [n] 0
  • count [n] 0
  • the LUT deriving unit 134 returns to the top of the loop LP12A (S138), and continues processing for the next entry.
  • the LUT deriving unit 134 substitutes sum [n] / count [n] into LUT [n] (S137). That is, in step S137, the arithmetic average of the color difference values at the luminance value n is acquired and substituted into LUT [n].
  • the LUT deriving unit 134 performs interpolation processing for unregistered entries (S139). Since the process in step S139 is the same as the process in steps S107 to S111 (loop LP13) shown in FIG. 12, the description thereof is omitted here. Thereafter, the LUT derivation process ends.
  • FIG. 19 is a diagram illustrating an example of the derived LUT.
  • the LUT may hold only a set of samples until it is referred to.
  • the LUT deriving unit 134 does not interpolate unregistered entries other than the entries for 16 sets of samples during the LUT deriving process.
  • the LUT deriving unit 134 derives the referenced unregistered entry.
  • the LUT deriving unit 134 derives the referenced unregistered entry n (nL ⁇ n ⁇ nR) by linear interpolation according to the following equation (3) as an example.
  • LUT [n] (LUT [nL] ⁇ (n ⁇ nL) + LUT [nR] ⁇ (nR ⁇ n) / (nR ⁇ nL)) (3)
  • nL and nR are the numbers of the most recently registered entries before and after n, respectively, as described with reference to FIG.
  • the memory area may be prepared for the number of registered entries. That is, when such a configuration is adopted, only a memory area proportional to the number of samples is consumed in creating the LUT table.
  • the inter-channel prediction based on the conventional linear transformation and the inter-channel prediction based on the LUT may be switched.
  • the higher accuracy may be used for inter-channel prediction.
  • a configuration using linear conversion can be used.
  • the moving image encoding device 2 is a device that generates and outputs encoded data # 1 by encoding the input image # 10.
  • FIG. 21 is a functional block diagram showing the configuration of the moving image encoding device 2.
  • the moving image encoding device 2 includes an encoding setting unit 21, an inverse quantization / inverse conversion unit 22, a predicted image generation unit 23, an adder 24, a frame memory 25, a subtractor 26, a conversion / A quantization unit 27 and a variable length coding unit 28 are provided.
  • the encoding setting unit 21 generates image data related to encoding and various setting information based on the input image # 10.
  • the encoding setting unit 21 generates the next image data and setting information.
  • the encoding setting unit 21 generates the leaf CU image # 100 for the target leaf CU by sequentially dividing the input image # 10 into slice units and LCU units.
  • the encoding setting unit 21 generates header setting information H ′ based on the result of the division process.
  • the header information H ′ includes (1) information about the size, shape and position of the LCU belonging to the target slice, and (2) the size, shape and shape of the leaf CU belonging to each LCU. It includes CU information CU ′ about the position.
  • the encoding setting unit 21 refers to the leaf CU image # 100 and the CU information CU 'to generate PU setting information PUI'.
  • the PU setting information PUI ' includes information on all combinations of (1) possible division patterns for each partition of the target leaf CU and (2) prediction modes that can be assigned to each partition.
  • the encoding setting unit 21 supplies the leaf CU image # 100 to the subtractor 26.
  • the encoding setting unit 21 supplies the header information H ′ to the variable length encoding unit 28. Also, the encoding setting unit 21 supplies the PU setting information PUI ′ to the predicted image generation unit 23.
  • the inverse quantization / inverse transform unit 22 performs inverse quantization and inverse DCT transform (Inverse Discrete Cosine Transform) on the quantization prediction residual for each block supplied from the transform / quantization unit 27, Restore the prediction residual for each block. Further, the inverse quantization / inverse transform unit 22 integrates the prediction residual for each block according to the division pattern specified by the TU partition information, and generates a prediction residual D for the target leaf CU. The inverse quantization / inverse transform unit 22 supplies the prediction residual D for the generated target leaf CU to the adder 24.
  • inverse quantization / inverse transform unit 22 supplies the prediction residual D for the generated target leaf CU to the adder 24.
  • the predicted image generation unit 23 refers to the locally decoded image P ′ recorded in the frame memory 25 and the PU setting information PUI ′, and generates a predicted image Pred for the target leaf CU.
  • the prediction image generation unit 23 refers to the luminance decoded image P Y.
  • the predicted image generation unit 23 sets the prediction parameter obtained by the predicted image generation process in the PU setting information PUI ′, and transfers the set PU setting information PUI ′ to the variable length encoding unit 28. Note that the predicted image generation process performed by the predicted image generation unit 23 is the same as that performed by the predicted image generation unit 13 included in the video decoding device 1, and thus description thereof is omitted here.
  • the adder 24 adds the predicted image Pred supplied from the predicted image generation unit 23 and the prediction residual D supplied from the inverse quantization / inverse transform unit 22 to add a decoded image P for the target leaf CU. Is generated.
  • Decoded decoded image P is sequentially recorded in the frame memory 25.
  • decoded images corresponding to all the LCUs decoded before the target LCU for example, all the LCUs preceding in the raster scan order) at the time of decoding the target LCU are recorded. .
  • the subtracter 26 generates a prediction residual D for the target leaf CU by subtracting the prediction image Pred from the leaf CU image # 100.
  • the subtractor 26 supplies the generated prediction residual D to the transform / quantization unit 27.
  • the transform / quantization unit 27 performs a DCT transform (Discrete Cosine Transform) and quantization on the prediction residual D to generate a quantized prediction residual.
  • DCT transform Discrete Cosine Transform
  • the transform / quantization unit 27 refers to the leaf CU image # 100 and the CU information CU ', and determines the division pattern of the target leaf CU into one or a plurality of blocks. Further, according to the determined division pattern, the prediction residual D is divided into prediction residuals for each block.
  • the transform / quantization unit 27 generates a prediction residual in the frequency domain by performing DCT transform (DiscretecreCosine Transform) on the prediction residual for each block, and then quantizes the prediction residual in the frequency domain. Thus, a quantized prediction residual for each block is generated.
  • DCT transform DiscretecreCosine Transform
  • the transform / quantization unit 27 relates to the generated quantization prediction residual for each block, TU partition information that specifies the partition pattern of the target leaf CU, and all possible partition patterns for each block of the target leaf CU.
  • TU setting information TUI ′ including the information is generated.
  • the transform / quantization unit 27 supplies the generated TU setting information TUI 'to the inverse quantization / inverse transform unit 22 and the variable length coding unit 28.
  • variable length encoding unit 28 generates and outputs encoded data # 1 based on the TU setting information TUI ′, the PU setting information PUI ′, and the header information H ′.
  • FIG. 22 is a flowchart illustrating an example of the flow of the color difference channel processing order and the prediction source channel encoding process in the moving image encoding apparatus 2.
  • the predicted image generation unit 23 selects a luminance prediction mode in the luminance predicted image generation process for the target partition (S200).
  • the predicted image generation unit 23 generates a brightness predicted image based on the selected prediction mode (S201).
  • the predicted image generation unit 23 enters the color difference predicted image creation processing loop LP21 for each pattern in the color difference channel processing order.
  • the predicted image generation unit 23 performs the color difference A and the prediction source channel of the color difference A according to the pattern of each processing order described using FIG. 8, FIG. 16, and FIG.
  • the color difference B and the prediction source channel of the color difference B are set (S203).
  • step S203 the inter-channel prediction unit 351 sets one of the U channel and the V channel to the color difference A for which channel estimation is performed first, and the other is set to the color difference B for which estimation is performed after the color difference A. Further, the inter-channel prediction unit 351 sets one or a plurality of prediction source channels for the color difference A and the color difference B, respectively.
  • the predicted image generation unit 23 selects a prediction mode for the color difference A (S204), and generates a color difference predicted image for the color difference A by inter-channel prediction according to the setting in step S120 (S205).
  • the predicted image generation unit 23 selects a prediction mode for the color difference B (S206), and generates a color difference predicted image by inter-channel prediction for the color difference B according to the setting in step S120 (S207). Furthermore, the color difference predicted image creation process by inter-channel prediction continues, and the process returns to the top of the loop LP21 (S208).
  • the predicted image generation unit 23 selects a color difference channel processing order and a combination of prediction source channels that are most suitable for encoding (S209).
  • the prediction image generation unit 23 performs encoding by including the inter-channel prediction flag, the color difference channel processing order flag, and the second channel prediction source channel specifier in the intra prediction parameter PP (S210).
  • the predicted image generation unit 23 may encode the inter-channel prediction index indicating the combination of the processing order and the prediction source channel in step S210.
  • Embodiment 2 Another embodiment of the present invention will be described below with reference to FIGS. For convenience of explanation, members having the same functions as those in the drawings described in the first embodiment are denoted by the same reference numerals and description thereof is omitted.
  • FIG. 23 is a functional block diagram illustrating another example of the configuration of the predicted image generation unit 13.
  • the predicted image generation unit 13A initializes the LUT in units of LCU, and updates the LUT in each block (target partition) in the same LCU.
  • the inter-channel prediction unit 351 is changed to an inter-channel prediction unit 351A.
  • the predicted image generation unit 13A is provided with an LUT derivation unit (correlation derivation means) 16 separately from the predicted image generation unit 13A.
  • the LUT derivation unit 16 newly derives an LUT for each target LCU.
  • the LUT deriving unit 16 updates the LUT in units of partitions.
  • the difference between the LUT deriving unit 16 and the LUT deriving unit 134 is a unit for deriving this LUT.
  • the inter-channel prediction unit 351A is changed to refer to the LUT deriving unit 16.
  • FIG. 24 is a flowchart illustrating a schematic flow of color difference predicted image generation processing in the predicted image generation unit 13A.
  • the LUT deriving unit 16 determines whether or not the target partition is a block that is first processed by the LCU (S30). When the target partition is a block processed first by the LCU (YES in S30), the LUT is initialized (S31).
  • LUT_F [n] 1 indicates that there is a sample
  • LUT_F [n] 0 indicates that there is no sample.
  • the LUT deriving unit 16 updates the LUT of each channel while referring to the local decoded image P ′ (S32). Details of the LUT update processing will be described later.
  • the LUT deriving unit 16 executes the LUT update process without initializing the LUT (S32).
  • the inter-channel prediction determination unit 133 refers to the inter-channel prediction flag and determines whether or not the inter-channel prediction mode is set (S33).
  • the intra-channel prediction unit 352 generates the color difference prediction image PredC without using the inter-channel prediction (S36), and the process ends.
  • the inter-channel prediction unit 351A refers to the LUT updated by the LUT deriving unit 16, and the color difference prediction image PredC by the inter-channel prediction Is generated (S35). This is the end of the process.
  • FIG. 25 is a flowchart showing an example of the flow of LUT update processing by the LUT deriving unit 16.
  • the LUT deriving unit 16 enters a registration process loop LP41 for each luminance pixel adjacent to the target partition (S400).
  • the LUT deriving unit 16 determines whether or not a color difference value is registered in the LUT [n] (S402). In other words, the LUT deriving unit 16 determines that the color difference value is registered if LUT_F [n] is 1 (with a sample), whereas if LUT_F [n] is 0 (no sample), the color difference is determined. It is determined that the value is not registered.
  • the LUT derivation unit 134 registers the color difference value m acquired in Step S401 as it is in LUT [n]. At the same time, 1 (with sample) is substituted into LUT_F [n] (S404), and the process returns to the top of the loop LP11 (S405).
  • the LUT derivation unit 16 calculates (m + LUT [n] +1) / 2, thereby obtaining the acquired color difference value m and the registered color difference value. And the average value is calculated. The LUT deriving unit 16 substitutes the average value thus calculated for the color difference value m (S403). Subsequently, the LUT deriving unit 16 registers the color difference value m into which the average value is substituted in step S403 in LUT [n], and substitutes 1 (with sample) into LUT_F [n] (S404). Return to the top of LP41 (S405).
  • the LUT deriving unit 134 determines whether or not LUT_F [n] is 0 (no sample) (S407).
  • the LUT itself is retained while processing the same LCU.
  • all entries are registered by interpolation. Therefore, here, instead of confirming whether or not LUT [n] has been registered, it is confirmed whether or not a sample exists. An entry for which no sample exists is again subject to interpolation.
  • LUT_F [n] is not 0 (no sample), that is, if LUT [n] has already been registered (NO in S407), the interpolation process continues and returns to the top of the loop LP42 (S410). .
  • the sample points Smpl1 and Smpl2 are registered in the immediately preceding target partition and the interpolation process is performed, but the sample point Smpl3 is newly registered in the target partition.
  • the LUT deriving unit 16 searches for entries with the latest sample before and after n.
  • the LUT deriving unit 16 searches for registered entries for n forwards, that is, nL smaller than n, with n as a base point. That is, here, the sample point Smpl1 shown in FIG. 26 is searched.
  • the LUT deriving unit 16 searches for an entry registered for n backward, that is, nR larger than n, with n as a base point.
  • nR larger than n
  • the sample point Smpl2 shown in FIG. 26 has been searched for in the target partition up to immediately before, but the sample point Smpl3 is searched for when the sample point Smpl3 is registered in the target partition.
  • the LUT deriving unit 16 registers a value obtained by linear interpolation between the LUT [nL] and the LUT [nR] in the LUT [n] (S409).
  • step S408 If only one of nL and nR is detected as a result of the search in step S408, the value of the registered entry of the detected one is registered in LUT [n].
  • step S403 the LUT deriving unit 16 calculates the average value of the acquired color difference value m and the registered color difference value by calculating (m + LUT [n] +1) / 2. Not limited to. For example, a weighted average of 1: 3 may be calculated for the acquired color difference value m and the registered color difference value.
  • the predicted image generation unit 13A is configured to initialize the LUT in units of LCUs and update the LUT in each block (target partition) in the same LCU.
  • the LUT tends to increase in accuracy as the number of samples used for derivation increases. However, if the sample is acquired too widely, the correlation may be lost or it may become extremely small.
  • the LCU is a range that is wider than the target partition and is assumed to have a correlation.
  • the above-described moving image encoding device 2 and moving image decoding device 1 can be used by being mounted on various devices that perform transmission, reception, recording, and reproduction of moving images.
  • the moving image may be a natural moving image captured by a camera or the like, or may be an artificial moving image (including CG and GUI) generated by a computer or the like.
  • moving image encoding device 2 and moving image decoding device 1 can be used for transmission and reception of moving images.
  • FIG. 30 (a) is a block diagram illustrating a configuration of a transmission device PROD_A in which the moving image encoding device 2 is mounted.
  • the transmission device PROD_A modulates a carrier wave with an encoding unit PROD_A1 that obtains encoded data by encoding a moving image, and the encoded data obtained by the encoding unit PROD_A1.
  • a modulation unit PROD_A2 that obtains a modulation signal and a transmission unit PROD_A3 that transmits the modulation signal obtained by the modulation unit PROD_A2 are provided.
  • the moving image encoding apparatus 2 described above is used as the encoding unit PROD_A1.
  • the transmission device PROD_A is a camera PROD_A4 that captures a moving image, a recording medium PROD_A5 that records the moving image, an input terminal PROD_A6 that inputs the moving image from the outside, as a supply source of the moving image input to the encoding unit PROD_A1.
  • An image processing unit A7 that generates or processes an image may be further provided.
  • FIG. 30A illustrates a configuration in which the transmission apparatus PROD_A includes all of these, but a part of the configuration may be omitted.
  • the recording medium PROD_A5 may be a recording of a non-encoded moving image, or a recording of a moving image encoded by a recording encoding scheme different from the transmission encoding scheme. It may be a thing. In the latter case, a decoding unit (not shown) for decoding the encoded data read from the recording medium PROD_A5 according to the recording encoding method may be interposed between the recording medium PROD_A5 and the encoding unit PROD_A1.
  • FIG. 30 is a block diagram illustrating a configuration of the receiving device PROD_B in which the moving image decoding device 1 is mounted.
  • the receiving device PROD_B includes a receiving unit PROD_B1 that receives a modulated signal, a demodulating unit PROD_B2 that obtains encoded data by demodulating the modulated signal received by the receiving unit PROD_B1, and a demodulator.
  • a decoding unit PROD_B3 that obtains a moving image by decoding the encoded data obtained by the unit PROD_B2.
  • the moving picture decoding apparatus 1 described above is used as the decoding unit PROD_B3.
  • the receiving device PROD_B has a display PROD_B4 for displaying a moving image, a recording medium PROD_B5 for recording the moving image, and an output terminal for outputting the moving image to the outside as a supply destination of the moving image output by the decoding unit PROD_B3.
  • PROD_B6 may be further provided.
  • FIG. 30B illustrates a configuration in which the reception device PROD_B includes all of these, but a part of the configuration may be omitted.
  • the recording medium PROD_B5 may be used for recording a non-encoded moving image, or may be encoded using a recording encoding method different from the transmission encoding method. May be. In the latter case, an encoding unit (not shown) for encoding the moving image acquired from the decoding unit PROD_B3 according to the recording encoding method may be interposed between the decoding unit PROD_B3 and the recording medium PROD_B5.
  • the transmission medium for transmitting the modulation signal may be wireless or wired.
  • the transmission mode for transmitting the modulated signal may be broadcasting (here, a transmission mode in which the transmission destination is not specified in advance) or communication (here, transmission in which the transmission destination is specified in advance). Refers to the embodiment). That is, the transmission of the modulation signal may be realized by any of wireless broadcasting, wired broadcasting, wireless communication, and wired communication.
  • a terrestrial digital broadcast broadcasting station (broadcasting equipment or the like) / receiving station (such as a television receiver) is an example of a transmitting device PROD_A / receiving device PROD_B that transmits and receives a modulated signal by wireless broadcasting.
  • a broadcasting station (such as broadcasting equipment) / receiving station (such as a television receiver) of cable television broadcasting is an example of a transmitting device PROD_A / receiving device PROD_B that transmits and receives a modulated signal by cable broadcasting.
  • a server workstation etc.
  • Client television receiver, personal computer, smart phone etc.
  • VOD Video On Demand
  • video sharing service using the Internet is a transmitting device for transmitting and receiving modulated signals by communication.
  • PROD_A / reception device PROD_B usually, either a wireless or wired transmission medium is used in a LAN, and a wired transmission medium is used in a WAN.
  • the personal computer includes a desktop PC, a laptop PC, and a tablet PC.
  • the smartphone also includes a multi-function mobile phone terminal.
  • the video sharing service client has a function of encoding a moving image captured by the camera and uploading it to the server. That is, the client of the video sharing service functions as both the transmission device PROD_A and the reception device PROD_B.
  • moving picture encoding apparatus 2 and moving picture decoding apparatus 1 can be used for recording and reproduction of moving pictures.
  • FIG. 31 (a) is a block diagram showing a configuration of a recording apparatus PROD_C in which the above-described moving picture encoding apparatus 2 is mounted.
  • the recording device PROD_C has an encoding unit PROD_C1 that obtains encoded data by encoding a moving image, and the encoded data obtained by the encoding unit PROD_C1 on the recording medium PROD_M.
  • a writing unit PROD_C2 for writing.
  • the moving image encoding apparatus 2 described above is used as the encoding unit PROD_C1.
  • the recording medium PROD_M may be of a type built in the recording device PROD_C, such as (1) HDD (Hard Disk Drive) or SSD (Solid State Drive), or (2) SD memory. It may be of the type connected to the recording device PROD_C, such as a card or USB (Universal Serial Bus) flash memory, or (3) DVD (Digital Versatile Disc) or BD (Blu-ray Disc: registration) Or a drive device (not shown) built in the recording device PROD_C.
  • HDD Hard Disk Drive
  • SSD Solid State Drive
  • SD memory such as a card or USB (Universal Serial Bus) flash memory, or (3) DVD (Digital Versatile Disc) or BD (Blu-ray Disc: registration) Or a drive device (not shown) built in the recording device PROD_C.
  • the recording device PROD_C is a camera PROD_C3 that captures moving images as a supply source of moving images to be input to the encoding unit PROD_C1, an input terminal PROD_C4 for inputting moving images from the outside, and reception for receiving moving images.
  • the unit PROD_C5 and an image processing unit C6 that generates or processes an image may be further provided.
  • FIG. 31A illustrates a configuration in which the recording apparatus PROD_C includes all of these, but a part of the configuration may be omitted.
  • the receiving unit PROD_C5 may receive a non-encoded moving image, or may receive encoded data encoded by a transmission encoding scheme different from the recording encoding scheme. You may do. In the latter case, a transmission decoding unit (not shown) that decodes encoded data encoded by the transmission encoding method may be interposed between the reception unit PROD_C5 and the encoding unit PROD_C1.
  • Examples of such a recording device PROD_C include a DVD recorder, a BD recorder, and an HDD (Hard Disk Drive) recorder (in this case, the input terminal PROD_C4 or the receiving unit PROD_C5 is a main supply source of moving images).
  • a camcorder in this case, the camera PROD_C3 is a main source of moving images
  • a personal computer in this case, the receiving unit PROD_C5 or the image processing unit C6 is a main source of moving images
  • a smartphone in this case In this case, the camera PROD_C3 or the receiving unit PROD_C5 is a main supply source of moving images
  • the camera PROD_C3 or the receiving unit PROD_C5 is a main supply source of moving images
  • FIG. 31 is a block showing a configuration of a playback device PROD_D equipped with the above-described video decoding device 1.
  • the playback device PROD_D reads a moving image by decoding a read unit PROD_D1 that reads encoded data written to the recording medium PROD_M and a read unit PROD_D1 that reads the encoded data. And a decoding unit PROD_D2 to be obtained.
  • the moving picture decoding apparatus 1 described above is used as the decoding unit PROD_D2.
  • the recording medium PROD_M may be of the type built into the playback device PROD_D, such as (1) HDD or SSD, or (2) such as an SD memory card or USB flash memory, It may be of a type connected to the playback device PROD_D, or (3) may be loaded into a drive device (not shown) built in the playback device PROD_D, such as DVD or BD. Good.
  • the playback device PROD_D has a display PROD_D3 that displays a moving image, an output terminal PROD_D4 that outputs the moving image to the outside, and a transmission unit that transmits the moving image as a supply destination of the moving image output by the decoding unit PROD_D2.
  • PROD_D5 may be further provided.
  • FIG. 31B illustrates a configuration in which the playback apparatus PROD_D includes all of these, but some of them may be omitted.
  • the transmission unit PROD_D5 may transmit an unencoded moving image, or transmits encoded data encoded by a transmission encoding method different from the recording encoding method. You may do. In the latter case, it is preferable to interpose an encoding unit (not shown) that encodes a moving image with an encoding method for transmission between the decoding unit PROD_D2 and the transmission unit PROD_D5.
  • Examples of such a playback device PROD_D include a DVD player, a BD player, and an HDD player (in this case, an output terminal PROD_D4 to which a television receiver or the like is connected is a main supply destination of moving images).
  • a television receiver in this case, the display PROD_D3 is a main supply destination of moving images
  • a digital signage also referred to as an electronic signboard or an electronic bulletin board
  • the display PROD_D3 or the transmission unit PROD_D5 is the main supply of moving images.
  • each block of the moving picture decoding apparatus 1 and the moving picture encoding apparatus 2 described above may be realized in hardware by a logic circuit formed on an integrated circuit (IC chip), or may be a CPU (Central It may be realized by software using a Processing Unit).
  • each device includes a CPU that executes instructions of a program that realizes each function, a ROM (Read (Memory) that stores the program, a RAM (Random Memory) that expands the program, the program, and various types
  • a storage device such as a memory for storing data is provided.
  • An object of the present invention is to provide a recording medium in which a program code (execution format program, intermediate code program, source program) of a control program of each of the above devices, which is software that realizes the above-described functions, is recorded so as to be readable by a computer. This can also be achieved by supplying to each of the above devices and reading and executing the program code recorded on the recording medium by the computer (or CPU or MPU).
  • Examples of the recording medium include tapes such as magnetic tape and cassette tape, magnetic disks such as floppy (registered trademark) disks / hard disks, and CD-ROM / MO / MD / DVD / CD-R / Blu-ray disks (registered trademarks). ) And other optical disks, IC cards (including memory cards) / optical cards, semiconductor memories such as mask ROM / EPROM / EEPROM / flash ROM, PLD (Programmable logic device) and FPGA ( Logic circuits such as Field Programmable Gate Array can be used.
  • tapes such as magnetic tape and cassette tape
  • magnetic disks such as floppy (registered trademark) disks / hard disks
  • CD-ROM / MO / MD / DVD / CD-R / Blu-ray disks registered trademarks
  • IC cards including memory cards
  • semiconductor memories such as mask ROM / EPROM / EEPROM / flash ROM, PLD (Programmable logic device) and FPGA ( Logic circuits
  • each of the above devices may be configured to be connectable to a communication network, and the program code may be supplied via the communication network.
  • the communication network is not particularly limited as long as it can transmit the program code.
  • the Internet intranet, extranet, LAN, ISDN, VAN, CATV communication network, virtual private network (Virtual Private Network), telephone line network, mobile communication network, satellite communication network, etc. can be used.
  • the transmission medium constituting the communication network may be any medium that can transmit the program code, and is not limited to a specific configuration or type.
  • wired lines such as IEEE 1394, USB, power line carrier, cable TV line, telephone line, ADSL (Asymmetric Digital Subscriber Line) line, infrared rays such as IrDA and remote control, Bluetooth (registered trademark), IEEE 802.11 wireless, HDR ( It can also be used by wireless such as High Data Rate, NFC (Near Field Communication), DLNA (Digital Living Network Alliance), mobile phone network, satellite line, and terrestrial digital network.
  • the present invention can also be realized in the form of a computer data signal embedded in a carrier wave in which the program code is embodied by electronic transmission.
  • the image decoding device decodes a moving image from encoded data.
  • the image decoding device regardless of whether the image is a moving image or a still image, the image decoding device generally Applicable. The same applies to the image encoding device.
  • LCU Large Coding Unit
  • HEVC High Efficiency Video Coding
  • a leaf CU is a CU (Coding Unit, coding) It is also called the leaf of the tree).
  • PU and TU in the said embodiment are respectively equivalent to the prediction tree (Prediction Tree) and transformation tree (transform tree) in HEVC.
  • the partition of PU in the said embodiment is corresponded to PU (Prediction Unit) in HEVC.
  • a block obtained by dividing a TU corresponds to a TU (Transformation Unit) in HEVC.
  • each component of each pixel included in the locally decoded image has a variation and is linear. This is a configuration that improves the possibility that higher prediction accuracy can be obtained even if prediction between the channels is performed according to the correlation, even if the prediction accuracy is reduced.
  • the prediction image generation unit 13 refers to the local decoded image P ′ positioned around the target partition, and performs nonlinearity between the decoded luminance (Y) channel and the color difference (U, V) channel to be decoded.
  • a LUT derivation unit 134 for deriving a correlation as LUT, the target partition, according to the LUT LUT deriving unit 134 derives, from the luminance decoded image P Y, and inter-channel prediction unit 351 for generating a color difference prediction image PredC, the Prepare.
  • the image decoding apparatus generates a prediction image for each of a plurality of channels indicating each component constituting an image, and adds a prediction residual to the generated prediction image.
  • the image decoding device that decodes the image data encoded according to the above, for each block to be processed, channel decoding means for decoding one or more channels among the plurality of channels, and each of the plurality of channels has been decoded. Then, with reference to the local decoded image located around the block to be processed, a non-linear correlation between the one or more channels already decoded by the channel decoding means and the other channels to be decoded is derived. For the correlation deriving means and the processing target block, the decoded block is added according to the derived correlation. From one or more channels of the decoded image, which is configured to include, the predicted image generating means for generating the prediction image of the other channels.
  • a channel is a generalized component that constitutes an image.
  • the luminance component and the color difference component correspond to channels. That is, in this example, the channel includes a luminance channel and a color difference channel.
  • the color difference channel includes a U channel indicating the U component of the color difference and a V channel indicating the V component of the color difference.
  • the channel may relate to the RGB color space.
  • the image decoding process is performed for each channel.
  • the local decoded image is a decoded image that has been decoded for the plurality of channels and is located around the block to be processed.
  • Processing target block refers to various processing units in the decoding process.
  • a coding unit, a conversion unit, a prediction unit, and the like can be given.
  • the processing unit includes a unit obtained by further subdividing the encoding unit, the conversion unit, and the prediction unit.
  • the periphery of the processing target block includes, for example, a pixel adjacent to the target block, a block adjacent to the left side of the target block, a block adjacent to the upper side of the target block, and the like.
  • the non-linear correlation can be derived, for example, by examining the correspondence of each point composed of the luminance value and the color difference value.
  • the non-linear correlation for example, when describing the YUV color space, it can be derived from the correspondence between the luminance value of each pixel included in the locally decoded image and the color difference value.
  • the correlation may be realized as an LUT in which a decoded channel and a decoding target channel are associated with each other.
  • the correlation may be expressed by a function including a relational expression established between the decoded channel and the decoding target channel.
  • the channel to be decoded in the processing target block is predicted from the channel that has been decoded in the processing target block according to the nonlinear correlation derived in this way.
  • such prediction is also referred to as inter-channel prediction.
  • the pixel value of the decoded image of the decoded channel is converted according to the nonlinear correlation, and the pixel value of the predicted image of the channel to be decoded is obtained.
  • the pixel value is a generalized value of a component constituting the image.
  • the correlation deriving unit is configured to detect a decoded channel when a pixel value of the decoded image of the decoded channel does not exist as a corresponding pixel value of a pixel included in the local decoded image. It is preferable to derive the nonlinear correlation by performing interpolation using pixel values of pixels included in the local decoded image having pixel values within a predetermined range of pixel values of the decoded image.
  • the pixel value of the image is the value of any component that forms the image.
  • the prediction source when the pixel value of the decoded image of the decoded channel does not exist as the corresponding pixel value of the pixel included in the local decoded image, in the example of the luminance value, the prediction source The pixel value of the decoded image of the decoded luminance channel that does not appear as the luminance value of the pixel included in the local decoded image.
  • the nonlinear correlation may be derived in advance, or the pixel value of the decoded image of the decoded channel does not exist as the corresponding pixel value of the pixel included in the local decoded image. You may derive
  • a correlation can be obtained by linear interpolation from previous and subsequent samples near the value. For example, for each sample point consisting of a luminance value and a color difference value, the correlation can be derived by linearly interpolating adjacent points. As another example of nonlinear correlation, each point can be derived by approximating each point by cubic interpolation.
  • the value for the decoded channel is the value that does not appear as the pixel value of the pixel included in the locally decoded image.
  • the value for the decoding target channel can be predicted with high accuracy.
  • the correlation deriving unit derives a relationship between a plurality of decoded channels and a decoding target channel as a correlation.
  • the V channel to be decoded can be predicted from the decoded luminance channel and U channel by the above configuration.
  • the prediction between channels is performed using the relationship between a plurality of decoded channels and the channel to be decoded as a correlation, the prediction accuracy can be improved.
  • the correlation deriving unit derives a correlation between a plurality of pixel values included in a locally decoded image of a decoded channel and a channel to be decoded.
  • the above configuration derives a correlation between a plurality of luminance values included in the locally decoded image of the decoded luminance channel and the color difference to be decoded.
  • the plurality of luminance values are luminance values within a predetermined range.
  • channel decoding processing order information indicating in which order the plurality of channels are to be decoded, and from which of the decoded channels the channel to be decoded should be predicted
  • the processing information acquisition means for acquiring the specified prediction source channel information, and the plurality of channels as decoding targets in the order indicated by the channel decoding processing order information, and the decoding target channels as the prediction source It is preferable to include a prediction control unit that performs control so as to perform prediction from the decoded channel specified in the channel information.
  • inter-channel prediction can be controlled based on designation of channel decoding process order information and prediction source channel information.
  • the channel decoding processing order information and the prediction source channel information are included in encoded data including encoded image data, for example. Therefore, for example, it is possible to cope with an image encoding device that encodes channel decoding process order information and prediction source channel information into encoded data and transmits the encoded data.
  • the processing information acquisition means acquires the channel decoding processing order information and the prediction source channel information that are encoded in a predetermined processing unit in decoding processing
  • the prediction control means Preferably, the control is performed in accordance with the channel decoding processing order information and the prediction source channel information acquired by the processing information acquisition unit.
  • control can be changed according to the processing unit.
  • For each processing unit it is possible to set the actual decoding target order and prediction source channel.
  • the correlation deriving means uses, for the processing target block included in the block group including a plurality of blocks, the local decoded image decoded in the processed block included in the block group. It is preferable to derive the correlation.
  • the local decoded image “located around the processing target block” is the local decoded image decoded in “the processed block included in the processing target block group”.
  • the more sample points the higher the accuracy of prediction based on the above correlation. Therefore, if as many sample points as possible can be acquired between blocks having a spatial correlation, the prediction accuracy of inter-channel prediction can be improved.
  • the image encoding device encodes the prediction residual obtained by subtracting the prediction image generated for each of a plurality of channels indicating each component constituting the image from the original image.
  • channel decoding means for decoding one or more channels among the plurality of channels, and each of the plurality of channels has been decoded.
  • the derivation means and the processing target block are processed in accordance with the above derived correlation. From one or more channels of the decoded image, which is configured to include, the predicted image generating means for generating the prediction image of the other channels.
  • the data structure of the encoded data encodes the prediction residual obtained by subtracting the prediction image generated for each of a plurality of channels indicating each component constituting the image from the original image.
  • a data structure of encoded data generated by converting to an image encoded by generating a predicted image for each of the plurality of channels and adding a prediction residual to the generated predicted image
  • channel decoding processing order information that indicates in which order the plurality of channels are decoded for the processing target block, and one or more decoded ones for the processing target block From the decoded image of the channel, the block to be processed has been decoded for each of the plurality of channels.
  • the predicted image of the other channel is generated according to a non-linear correlation between the one or more channels that have been decoded and the other channel to be decoded with reference to a locally decoded image located in the vicinity of
  • the data structure includes prediction source channel designation information for designating whether or not to perform.
  • the same effects as those of the image decoding device according to the present invention can be obtained.
  • the image encoding device specifies channel decoding processing order information indicating in which order the plurality of channels are to be decoded, and from which of the decoded channels the channel to be decoded is to be predicted
  • the prediction source channel information may be included in the data structure of the encoded data.
  • the image encoding device may encode the information in, for example, side information.
  • the present invention can be suitably applied to a decoding device that decodes encoded data and an encoding device that generates encoded data. Further, the present invention can be suitably applied to the data structure of encoded data generated by the encoding device and referenced by the decoding device.

Landscapes

  • Engineering & Computer Science (AREA)
  • Multimedia (AREA)
  • Signal Processing (AREA)
  • Compression Or Coding Systems Of Tv Signals (AREA)

Abstract

Provided is an image decoding device comprising: an LUT derivation unit (134) which queries a local decoded image (P') which is located on the periphery of a subject partition, and derives, in the form of an LUT, a nonlinear correlation between a decoded brightness (Y) channel and a color differential (U, V) channel to be decoded; and an inter-channel prediction unit (351) which generates a color differential prediction image (PredC) from a brightness decoded image (PY) for the subject partition in accordance with the LUT.

Description

画像復号装置、画像符号化装置、および符号化データのデータ構造Image decoding apparatus, image encoding apparatus, and data structure of encoded data
 本発明は画像の色差についてイントラ予測を行う画像復号装置、画像符号化装置、および符号化データのデータ構造に関する。 The present invention relates to an image decoding device that performs intra prediction on image color differences, an image encoding device, and a data structure of encoded data.
 動画像を効率的に伝送または記録するために、動画像を符号化することによって符号化データを生成する動画像符号化装置、および、当該符号化データを復号することによって復号画像を生成する動画像復号装置が用いられている。 In order to efficiently transmit or record a moving image, a moving image encoding device that generates encoded data by encoding the moving image, and a moving image that generates a decoded image by decoding the encoded data An image decoding device is used.
 具体的な動画像符号化方式としては、例えば、H.264/MPEG-4.AVC(非特許文献1)、および、VCEG(Video Coding Expert Group)における共同開発用コーデックであるKTAソフトウェアに採用されている方式、および、その後継コーデックであるTMuC(Test Model under Consideration)ソフトウェアに採用されている方式などが挙げられる(非特許文献2)。 Specific examples of the moving image encoding method include H.264. H.264 / MPEG-4. Adopted in KTA software, which is a codec for joint development in AVC (Non-patent Document 1) and VCEG (Video Coding Expert Group), and TMuC (Test Model Under Consideration) software, which is the successor codec (Non-Patent Document 2).
 また、このような符号化方式においては、通常、入力画像を符号化/復号することによって得られる局所復号画像に基づいて予測画像が生成され、当該予測画像を入力画像(原画像)から減算して得られる予測残差(「差分画像」または「残差画像」と呼ぶこともある)が符号化される。また、予測画像の生成方法としては、画面間予測(インター予測)、および、画面内予測(イントラ予測)が挙げられる。 In such an encoding method, a predicted image is usually generated based on a locally decoded image obtained by encoding / decoding an input image, and the predicted image is subtracted from the input image (original image). The prediction residual (which may be referred to as “difference image” or “residual image”) is encoded. In addition, examples of the method for generating a predicted image include inter-screen prediction (inter prediction) and intra-screen prediction (intra prediction).
 なお、TMuC(非特許文献2)では、動画像を分割して次のような階層構造により管理することを提案している。 Note that TMuC (Non-Patent Document 2) proposes to divide a moving image and manage it by the following hierarchical structure.
 まず、動画像を構成する画像(ピクチャ)は、スライスに分割される。また、スライスは、最大符号化単位(マクロブロックまたは最大コーディングユニット(Largest Coding Unit)と呼ばれることもある)に分割される。最大符号化単位は、4分木分割により、さらに細かい符号化単位(Coding Unit)に分割可能である。 First, an image (picture) constituting a moving image is divided into slices. The slice is also divided into maximum coding units (sometimes called macroblocks or maximum coding units (Largest (Coding Units)). The maximum coding unit can be divided into smaller coding units (Coding Unit) by quadtree division.
 そして、それ以上分割することができない符号化単位(リーフCU)は、変換単位および予測単位として取り扱われる。これらは、ブロックと呼ばれることもある。また、予測単位を、そのまま援用した、または、さらに分割したパーティションと呼ばれる単位が定義されている。さらに付言しておくと、このパーティション単位でイントラ予測が行われる。 Further, a coding unit (leaf CU) that cannot be further divided is treated as a conversion unit and a prediction unit. These are sometimes called blocks. In addition, a unit called a partition that uses a prediction unit as it is or is further divided is defined. In addition, intra prediction is performed in units of partitions.
 H.264/MPEG-4.AVC(非特許文献1)等では、画素を、輝度成分Yと、色差成分C,Cの組み合わせで表す表現形式が採用されている。 H. H.264 / MPEG-4. In AVC (Non-Patent Document 1) and the like, an expression format in which a pixel is represented by a combination of a luminance component Y and color difference components C b and C r is employed.
 また、色差についても、イントラ予測が行われる。輝度と色差とは本来互いに独立した成分である。よって、H.264/MPEG-4.AVC(非特許文献1)等では、色差信号のイントラ予測に、水平予測、垂直予測、DC予測、プレーン予測が用いられていた。 Also, intra prediction is performed for color differences. Luminance and color difference are inherently independent components. Therefore, H.I. H.264 / MPEG-4. In AVC (Non-Patent Document 1) and the like, horizontal prediction, vertical prediction, DC prediction, and plane prediction are used for intra prediction of color difference signals.
 また、人間の眼は画素の輝度変化には敏感であるが、色変化には鈍感であるという視覚特性がある。 Also, the human eye has a visual characteristic that it is sensitive to pixel luminance changes but insensitive to color changes.
 よって、色差画素の解像度を下げても、輝度画素の解像度を下げる場合に比べて、視覚上の影響が少ない。そこで、動画像の符号化においては、色差画素の解像度を、輝度画素の解像度よりも下げてデータ量の削減を図っている。 Therefore, even if the resolution of the chrominance pixel is lowered, there is less visual influence than when the resolution of the luminance pixel is lowered. Therefore, in the encoding of moving images, the resolution of color difference pixels is made lower than the resolution of luminance pixels to reduce the data amount.
 図27に、従来技術における、原画像と、輝度画素および色差画素との対応関係の一例について示す。図27の(a)は、符号化の対象となる画像(YUV)について示しており、(b)は、輝度画素について示しており、また、(c)は、色差画素(Uについて例示)を示している。 FIG. 27 shows an example of a correspondence relationship between an original image, a luminance pixel, and a color difference pixel in the prior art. 27A shows an image (YUV) to be encoded, FIG. 27B shows a luminance pixel, and FIG. 27C shows a color difference pixel (exemplified for U). Show.
 図27の(a)~(c)に示すように、例えば、輝度画素の解像度は、8×8であるのに対して、色差画素は、4×4の解像度である。 As shown in FIGS. 27A to 27C, for example, the resolution of the luminance pixel is 8 × 8, whereas the color difference pixel has a resolution of 4 × 4.
 ところで、局所的な空間においては、画素値の種類は限られている。このため、特定の状況下、例えば、このように局所的な空間においては、輝度(Y)と色差(UV)との間で相関があることが知られている。そして、近年、この相関を色差のイントラ予測に用いる技術が提案されている。いわゆるチャンネル間予測の技術である。 By the way, in the local space, the types of pixel values are limited. For this reason, it is known that there is a correlation between the luminance (Y) and the color difference (UV) in a specific situation, for example, in such a local space. In recent years, a technique for using this correlation for intra prediction of color difference has been proposed. This is a so-called inter-channel prediction technique.
 例えば、非特許文献3では、輝度画像(Y)から、線形変換により色差画像(UV)を予測することが開示されている。具体的には、次の式(A1)に従って線形変換を行うことが開示されている。 For example, Non-Patent Document 3 discloses predicting a color difference image (UV) from a luminance image (Y) by linear conversion. Specifically, performing linear transformation according to the following formula (A1) is disclosed.
 PredC[x,y]=αRecY[x,y]+β … (A1)
 式(A1)における各記号の意味は次のとおりである。
PredC [x C , y C ] = α C RecY [x Y , y Y ] + β C (A1)
The meaning of each symbol in the formula (A1) is as follows.
 PredC:予測画像(色差)
[x,y]、[x,y]:同じサンプルの位置を示す座標
 RecY :復号画像(輝度)
 α,β :周辺の復号済み画像(以下、局所復号画像と称する)の画素値から最小二乗法により導出した係数
 なお、上述のとおり、輝度画像と、色差画像との解像度の違いから、[x,y]および[x,y]の座標は適宜変換する必要がある。
PredC: predicted image (color difference)
[X C , y C ], [x Y , y Y ]: coordinates indicating the position of the same sample RecY: decoded image (luminance)
α C , β C : coefficients derived by the least squares method from the pixel values of the surrounding decoded images (hereinafter referred to as local decoded images) As described above, from the difference in resolution between the luminance image and the color difference image, The coordinates of [x C , y C ] and [x Y , y Y ] need to be appropriately converted.
 また、非特許文献3では、図28に例示するように、輝度画像(Y)と、色差画像(U,V)との間で解像度が違う場合のサンプル点の採り方についても言及している。図28の(a)は、2N×2Nサイズの輝度画像からサンプル点を採る場合について示しており、図28の(b)は、N×Nサイズの色差画像からサンプルをとる場合について示している。非特許文献3では、図28の(a)に示すサンプル点smpl100を、図28の(b)に示すサンプル点smpl200に対応付けることが記載されている。 Further, Non-Patent Document 3 also mentions how to take sample points when the resolution is different between the luminance image (Y) and the color difference image (U, V), as illustrated in FIG. . FIG. 28A shows the case where sample points are taken from a 2N × 2N size luminance image, and FIG. 28B shows the case where samples are taken from an N × N color difference image. . Non-Patent Document 3 describes that the sample point smpl100 shown in (a) of FIG. 28 is associated with the sample point smpl200 shown in (b) of FIG.
 上述のような従来技術は、チャンネル間予測において、最小二乗法により線形的な変換を行っていた。このため、サンプルとした局所復号画像が、線形変換に適していない場合には、チャンネル間予測の精度が十分でない場合があった。 The conventional technology as described above performs linear conversion by the least square method in inter-channel prediction. For this reason, when the local decoded image used as a sample is not suitable for linear transformation, the accuracy of inter-channel prediction may not be sufficient.
 図29を参照しながら、この問題について説明すると次のとおりである。図29の(a)に示すように、輝度-色差平面におけるサンプルの分布が、グループGr100に1つの領域に収まる場合には、上述の式(A1)の予測の精度は、ある程度期待できる。 This problem will be described with reference to FIG. 29 as follows. As shown in FIG. 29A, when the sample distribution on the luminance-chrominance plane falls within one region in the group Gr100, the accuracy of prediction of the above-described equation (A1) can be expected to some extent.
 しかしながら、図29の(b)のように、サンプルの分布が、グループGr101と、グループGr102との2つの領域に分散している場合であって、なおかつ各領域においても、サンプルの分布がばらついているような場合、線形変換では、誤差が大きくなる場合がある。例えば、サンプル点Smpl300は、式(A1)で近似される値から大きく離れた位置に分布している。 However, as shown in FIG. 29B, the sample distribution is distributed in two regions of the group Gr101 and the group Gr102, and the sample distribution varies in each region. In such a case, the error may become large in the linear conversion. For example, the sample points Smpl300 are distributed at positions far away from the values approximated by the equation (A1).
 このように、線形的な相関に従って、各チャンネル間の予測を行うと、予測の精度が低下してしまうという問題があった。 As described above, when the prediction between the channels is performed according to the linear correlation, there is a problem that the accuracy of the prediction is lowered.
 本発明は、上記の問題点に鑑みてなされたものであり、その目的は、局所復号画像に含まれる各画素の各成分にばらつきがあり、線形的な相関に従って、各チャンネル間の予測を行うと、予測の精度が低下してしまうような場合であっても、より高い予測精度が得られる可能性を向上させることができる画像復号装置を実現することにある。 The present invention has been made in view of the above-mentioned problems, and its purpose is to perform a prediction between channels in accordance with a linear correlation because each component of each pixel included in a locally decoded image varies. Thus, an object of the present invention is to realize an image decoding apparatus capable of improving the possibility that higher prediction accuracy can be obtained even when the accuracy of prediction is lowered.
 上記の課題を解決するために、本発明に係る画像復号装置は、画像を構成する各成分を示す複数のチャンネルそれぞれについての予測画像を生成して、生成した上記予測画像に予測残差を加算することにより符号化された画像データを復号する画像復号装置において、処理対象ブロックについて、上記複数のチャンネルのうち、1または複数のチャンネルについて復号するチャンネル復号手段と、上記複数のチャンネルそれぞれについて復号済みである、上記処理対象ブロックの周辺に位置する局所復号画像を参照して、上記チャンネル復号手段により復号済みの上記1または複数のチャンネルと復号対象の他のチャンネルとの間の非線形的な相関を導出する相関導出手段と、処理対象ブロックについて、上記導出された相関に従って、復号済みの上記1または複数のチャンネルの復号画像から、上記他のチャンネルの上記予測画像を生成する予測画像生成手段と、を備えることを特徴とする。 In order to solve the above problem, an image decoding apparatus according to the present invention generates a prediction image for each of a plurality of channels indicating each component constituting an image, and adds a prediction residual to the generated prediction image In the image decoding apparatus that decodes the encoded image data, channel decoding means for decoding one or a plurality of channels among the plurality of channels for the processing target block, and each of the plurality of channels has been decoded A nonlinear correlation between the one or more channels that have been decoded by the channel decoding means and the other channels to be decoded is referred to with reference to a locally decoded image located around the block to be processed. The correlation derivation means to be derived and the processing target block are restored according to the above derived correlation. From already the one or more channels of the decoded image, characterized in that it comprises, the predicted image generating means for generating the prediction image of the other channels.
 上記構成において、チャンネルとは、画像を構成する成分を一般化したものである。例えば、YUV色空間においては、輝度成分および色差成分がチャンネルに該当する。すなわち、この例の場合、チャンネルには、輝度チャンネル、色差チャンネルが含まれる。また、色差チャンネルには、色差のU成分を示すUチャンネル、および、色差のV成分を示すVチャンネルが含まれる。なお、チャンネルはRGB色空間に関するものであってもよい。画像の復号処理は、この各チャンネルに対して行われる。 In the above configuration, a channel is a generalized component that constitutes an image. For example, in the YUV color space, the luminance component and the color difference component correspond to channels. That is, in this example, the channel includes a luminance channel and a color difference channel. The color difference channel includes a U channel indicating the U component of the color difference and a V channel indicating the V component of the color difference. The channel may relate to the RGB color space. The image decoding process is performed for each channel.
 また、局所復号画像とは、上記複数のチャンネルについて復号がなされたものであり、なおかつ処理対象ブロックの周辺に位置する復号画像のことである。 Also, the local decoded image is a decoded image that has been decoded for the plurality of channels and is located around the block to be processed.
 処理対象ブロックとは、復号処理における各種の処理単位のことである。例えば、符号化単位、変換単位および予測単位などが挙げられる。なお、処理単位には、符号化単位、変換単位および予測単位をさらに細かく細分化した単位も含まれる。 Processing target block refers to various processing units in the decoding process. For example, a coding unit, a conversion unit, a prediction unit, and the like can be given. Note that the processing unit includes a unit obtained by further subdividing the encoding unit, the conversion unit, and the prediction unit.
 処理対象ブロックの周辺とは、例えば、対象ブロックに隣接する画素や、対象ブロックの左側に隣接するブロックや、対象ブロックの上側に隣接するブロックなどが挙げられる。 The periphery of the processing target block includes, for example, a pixel adjacent to the target block, a block adjacent to the left side of the target block, a block adjacent to the upper side of the target block, and the like.
 また、局所的な空間においては、輝度チャンネルと色差チャンネルとの間で相関があることが知られている。そこで、上記構成では、上記処理対象ブロックの周辺に位置する局所復号画像をサンプルとして参照することで、上記チャンネル復号手段により復号済みの上記1または複数のチャンネルと復号対象の他のチャンネルとの間の非線形的な相関を導出する。 Also, in a local space, it is known that there is a correlation between the luminance channel and the color difference channel. Therefore, in the above configuration, by referring to the locally decoded image located around the processing target block as a sample, between the one or more channels decoded by the channel decoding means and the other channels to be decoded. The nonlinear correlation of is derived.
 ここで非線形的な相関は、例えば、輝度値および色差値からなる各点の対応を調べることにより導出することができる。 Here, the non-linear correlation can be derived, for example, by examining the correspondence of each point composed of the luminance value and the color difference value.
 非線形的な相関としては、例えば、YUV色空間について説明すると、局所復号画像に含まれる各画素の輝度値と、色差値とを対応付けたものから導出することができる。また、上記相関は、復号済みのチャンネルと復号対象のチャンネルとが対応付けられたLUTとして実現されていてもよい。また、上記相関は、復号済みのチャンネルと復号対象のチャンネルと間に成立する関係式からなる関数によって表現されていてもよい。 As the non-linear correlation, for example, when describing the YUV color space, it can be derived from the correspondence between the luminance value of each pixel included in the locally decoded image and the color difference value. The correlation may be realized as an LUT in which a decoded channel and a decoding target channel are associated with each other. The correlation may be expressed by a function including a relational expression established between the decoded channel and the decoding target channel.
 上記構成によれば、このようにして導出した非線形的な相関に従って、処理対象ブロックにて復号済みのチャンネルから、処理対象ブロックにて復号対象のチャンネルを予測する。以下、このような予測をチャンネル間予測ともいう。 According to the above configuration, the channel to be decoded in the processing target block is predicted from the channel that has been decoded in the processing target block according to the nonlinear correlation derived in this way. Hereinafter, such prediction is also referred to as inter-channel prediction.
 チャンネル間予測では、例えば、上記非線形的な相関に従って、復号済みのチャンネルの復号画像の画素値を変換して、復号対象のチャンネルの予測画像の画素値を得る。ここで、画素値とは、画像を構成する成分の値を一般化したものである。 In the inter-channel prediction, for example, the pixel value of the decoded image of the decoded channel is converted according to the nonlinear correlation, and the pixel value of the predicted image of the channel to be decoded is obtained. Here, the pixel value is a generalized value of a component constituting the image.
 このため、局所復号画像に含まれる各画素の各成分にばらつきがあり、線形的な相関に従って、各チャンネル間の予測を行うと、予測の精度が低下してしまうような場合であっても、より高い予測精度が得られる可能性を向上させることができるという効果を奏する。 For this reason, there is variation in each component of each pixel included in the locally decoded image, and even when prediction between each channel is performed according to a linear correlation, the accuracy of the prediction is reduced. There is an effect that the possibility that higher prediction accuracy can be obtained can be improved.
 上記の課題を解決するために、本発明に係る画像符号化装置は、画像を構成する各成分を示す複数のチャンネルそれぞれについて生成した予測画像を原画像から減算して得られる予測残差を符号化することによって符号化データを生成する画像符号化装置において、処理対象ブロックについて、上記複数のチャンネルのうち、1または複数のチャンネルについて復号するチャンネル復号手段と、上記複数のチャンネルそれぞれについて復号済みである、上記処理対象ブロックの周辺に位置する局所復号画像を参照して、上記チャンネル復号手段により復号済みの上記1または複数のチャンネルと復号対象の他のチャンネルとの間の非線形的な相関を導出する相関導出手段と、処理対象ブロックについて、上記導出された相関に従って、復号済みの上記1または複数のチャンネルの復号画像から、上記他のチャンネルの上記予測画像を生成する予測画像生成手段と、を備えることを特徴とする。 In order to solve the above problems, an image encoding device according to the present invention encodes a prediction residual obtained by subtracting, from an original image, a prediction image generated for each of a plurality of channels indicating each component constituting the image. In the image encoding device that generates encoded data by converting, a channel decoding unit that decodes one or a plurality of channels among the plurality of channels and a plurality of channels that have been decoded for the processing target block A non-linear correlation between the one or more channels decoded by the channel decoding means and other channels to be decoded is derived with reference to a local decoded image located around the processing target block. The correlation derivation means and the processing target block are decoded according to the derived correlation. Mino from the decoded image of the one or more channels, characterized in that it comprises a prediction image generating means for generating the prediction image of the other channels.
 上記の課題を解決するために、本発明に係る符号化データのデータ構造は、画像を構成する各成分を示す複数のチャンネルそれぞれについて生成した予測画像を原画像から減算して得られる予測残差を符号化することによって生成された符号化データのデータ構造であって、上記複数のチャンネルそれぞれについての予測画像を生成して、生成した上記予測画像に予測残差を加算することにより符号化された画像データを復号する画像復号装置に、処理対象ブロックについて、上記複数のチャンネルを、いずれの順序で復号するかを示すチャンネル復号処理順序情報と、処理対象ブロックについて、いずれの復号済みの上記1または複数のチャンネルの復号画像から、上記複数のチャンネルそれぞれについて復号済みである、上記処理対象ブロックの周辺に位置する局所復号画像を参照して、上記復号済みの上記1または複数のチャンネルと復号対象の他のチャンネルとの間の非線形的な相関に従って、上記他のチャンネルの上記予測画像を生成するかを指定する予測元チャンネル指定情報と、を含むことを特徴とする。 In order to solve the above-described problem, the data structure of the encoded data according to the present invention includes a prediction residual obtained by subtracting, from an original image, a prediction image generated for each of a plurality of channels indicating each component constituting the image. Is a data structure of encoded data generated by encoding the image, and is generated by generating a prediction image for each of the plurality of channels and adding a prediction residual to the generated prediction image. In the image decoding apparatus that decodes the image data, channel decoding processing order information indicating in which order the plurality of channels are decoded for the processing target block, and which of the decoded 1 above for the processing target block Alternatively, the processing target that has been decoded for each of the plurality of channels from the decoded images of the plurality of channels With reference to a locally decoded image located around the lock, the predicted image of the other channel is determined according to a non-linear correlation between the one or more decoded channels and the other channel to be decoded. And prediction source channel designation information for designating whether to generate.
 上記のように構成された画像符号化装置または符号化データのデータ構造によれば、本発明に係る画像復号装置と同様の効果を奏する。 According to the image encoding device or the data structure of the encoded data configured as described above, the same effects as those of the image decoding device according to the present invention can be obtained.
 なお、画像符号化装置は、上記複数のチャンネルをいずれの順序で復号対象とするかを示すチャンネル復号処理順序情報と、復号対象とするチャンネルを復号済みのチャンネルのいずれから予測するべきかを指定する予測元チャンネル情報とを符号化データのデータ構造に含めてもよい。画像符号化装置は、上記情報を、例えば、サイド情報において符号化してもよい。 Note that the image encoding device specifies channel decoding processing order information indicating in which order the plurality of channels are to be decoded, and from which of the decoded channels the channel to be decoded is to be predicted The prediction source channel information may be included in the data structure of the encoded data. The image encoding device may encode the information in, for example, side information.
 本発明に係る画像復号装置は、処理対象ブロックについて、複数のチャンネルのうち、1または複数のチャンネルについて復号するチャンネル復号手段と、上記複数のチャンネルそれぞれについて復号済みである、上記処理対象ブロックの周辺に位置する局所復号画像を参照して、上記チャンネル復号手段により復号済みの上記1または複数のチャンネルと復号対象の他のチャンネルとの間の非線形的な相関を導出する相関導出手段と、処理対象ブロックについて、上記導出された相関に従って、復号済みの上記1または複数のチャンネルの復号画像から、上記他のチャンネルの上記予測画像を生成する予測画像生成手段と、を備える構成である。 The image decoding apparatus according to the present invention includes a channel decoding unit that decodes one or a plurality of channels among a plurality of channels, and a periphery of the processing target block that has been decoded for each of the plurality of channels. A correlation deriving means for deriving a non-linear correlation between the one or more channels already decoded by the channel decoding means and the other channels to be decoded with reference to the local decoded image located at The block includes a predicted image generating unit that generates the predicted image of the other channel from the decoded image of the one or more channels that has been decoded in accordance with the derived correlation.
 本発明に係る画像符号化装置は、処理対象ブロックについて、複数のチャンネルのうち、1または複数のチャンネルについて復号するチャンネル復号手段と、上記複数のチャンネルそれぞれについて復号済みである、上記処理対象ブロックの周辺に位置する局所復号画像を参照して、上記チャンネル復号手段により復号済みの上記1または複数のチャンネルと復号対象の他のチャンネルとの間の非線形的な相関を導出する相関導出手段と、処理対象ブロックについて、上記導出された相関に従って、復号済みの上記1または複数のチャンネルの復号画像から、上記他のチャンネルの上記予測画像を生成する予測画像生成手段と、を備える構成である。 The image encoding device according to the present invention includes a channel decoding unit that decodes one or more channels among a plurality of channels for the processing target block, and the processing target block that has been decoded for each of the plurality of channels. Correlation deriving means for deriving a non-linear correlation between the one or more channels already decoded by the channel decoding means and other channels to be decoded with reference to locally decoded images located in the periphery, and processing The target block includes a predicted image generating unit that generates the predicted image of the other channel from the decoded images of the one or more channels that have been decoded in accordance with the derived correlation.
 本発明に係る符号化データのデータ構造は、上記複数のチャンネルそれぞれについての予測画像を生成して、生成した上記予測画像に予測残差を加算することにより符号化された画像データを復号する画像復号装置に、処理対象ブロックについて、上記複数のチャンネルを、いずれの順番で復号するかを示すチャンネル復号処理順序情報と、処理対象ブロックについて、いずれの復号済みの上記1または複数のチャンネルの復号画像から、上記複数のチャンネルそれぞれについて復号済みである、上記処理対象ブロックの周辺に位置する局所復号画像を参照して、上記復号済みの上記1または複数のチャンネルと復号対象の他のチャンネルとの間の非線形的な相関に従って、上記他のチャンネルの上記予測画像を生成するかを指定する予測元チャンネル情報と、を含むデータ構造である。 The data structure of the encoded data according to the present invention is an image that decodes encoded image data by generating a prediction image for each of the plurality of channels and adding a prediction residual to the generated prediction image. In the decoding apparatus, channel decoding processing order information indicating in which order the plurality of channels are decoded for the processing target block, and the decoded image of the one or more channels already decoded for the processing target block To the local decoded image located around the block to be processed that has been decoded for each of the plurality of channels, and between the one or more channels that have been decoded and the other channels to be decoded. Prediction specifying whether to generate the predicted image of the other channel according to the nonlinear correlation of And channel information, a data structure containing.
 よって、局所復号画像に含まれる各画素の各成分にばらつきがあり、線形的な相関に従って、各チャンネル間の予測を行うと、予測の精度が低下してしまうような場合であっても、より好適な予測精度が得られる可能性を向上させることができるという効果を奏する。 Therefore, even if there is variation in each component of each pixel included in the locally decoded image and prediction between the channels is performed according to a linear correlation, the accuracy of the prediction may be reduced. There is an effect that the possibility of obtaining suitable prediction accuracy can be improved.
本発明の一実施形態に係る動画像復号装置が備える予測画像生成部の構成を示すブロック図である。It is a block diagram which shows the structure of the estimated image generation part with which the moving image decoding apparatus which concerns on one Embodiment of this invention is provided. 上記動画像復号装置の概略的構成を示す機能ブロック図である。It is a functional block diagram which shows the schematic structure of the said moving image decoding apparatus. 本発明の一実施形態に係る動画像符号化装置によって生成され、上記動画像復号装置によって復号される符号化データの構成を示す図である。It is a figure which shows the structure of the encoding data produced | generated by the moving image encoder which concerns on one Embodiment of this invention, and decoded by the said moving image decoder. YUV色空間について説明する図である。It is a figure explaining YUV color space. YUV形式の画像フォーマットについて示す図であり、(a)~(d)は、それぞれ、4:2:0形式、4:4:4形式、4:2:2形式、および4:1:1形式を示す。FIG. 4 is a diagram illustrating an image format of a YUV format, where (a) to (d) are a 4: 2: 0 format, a 4: 4: 4 format, a 4: 2: 2 format, and a 4: 1: 1 format, respectively. Indicates. 4:2:0形式における輝度および色差の画素位置の対応関係について示す図である。It is a figure shown about the correspondence of the pixel position of the brightness | luminance and color difference in 4: 2: 0 format. 色差画素のサンプル位置について例示する図であり、(a)~(c)は、3つのサンプル位置について示している。FIG. 6 is a diagram illustrating sample positions of color difference pixels, and (a) to (c) show three sample positions. 色差チャンネルの処理順序のパターンについて示す図である。It is a figure shown about the pattern of the processing order of a color difference channel. 上記予測画像生成部における色差予測画像生成処理の概略的な流れについて例示したフローチャートである。It is the flowchart which illustrated about the schematic flow of the color difference estimated image generation process in the said estimated image generation part. 輝度-色差の分布が重ならならず2領域にまたがるような画像の例を示す図である。FIG. 4 is a diagram illustrating an example of an image in which luminance-color difference distributions do not overlap and extend over two regions. 図10に示す画像に含まれる画素の輝度(Y)-色差(U)をプロットしたグラフである。11 is a graph plotting luminance (Y) -color difference (U) of pixels included in the image shown in FIG. LUT導出部によるLUT導出処理の流れの一例について示したフローチャートである。It is the flowchart shown about an example of the flow of the LUT derivation | leading-out process by a LUT derivation | leading-out part. エントリ間の線形補間について説明する図である。It is a figure explaining the linear interpolation between entries. 図10に示す画像から導出したLUTの値を示すグラフである。It is a graph which shows the value of LUT derived | led-out from the image shown in FIG. チャンネル間予測部におけるチャンネル間予測による色差予測画像生成処理の流れの一例について示したフローチャートである。It is the flowchart shown about an example of the flow of the color difference estimated image generation process by the prediction between channels in the prediction part between channels. 2つめに予測を行うチャンネルの予測元チャンネルが2つである場合における色差チャンネルの処理順序のパターンについて示す図である。It is a figure shown about the pattern of the processing order of a color difference channel in case the prediction origin channel of the channel which performs a 2nd prediction is two. 2つめに予測を行うチャンネルの予測元チャンネルが2つである場合における色差チャンネルの処理順序の他のパターンについて示す図である。It is a figure shown about the other pattern of the processing order of a color difference channel in case the prediction origin channel of the channel which performs a 2nd prediction is two. LUT導出部によるLUT導出処理の流れの一変形例について示したフローチャートである。12 is a flowchart showing a modification of the flow of LUT derivation processing by the LUT derivation unit. 導出されたLUTの一例を示す図である。It is a figure which shows an example of the derived | led-out LUT. LUTの代わりに、関数を用いる例について説明する図である。It is a figure explaining the example which uses a function instead of LUT. 本発明の一実施形態に係る動画像符号化装置の構成について示す機能ブロック図である。It is a functional block diagram shown about the structure of the moving image encoder which concerns on one Embodiment of this invention. 上記動画像符号化装置における色差チャンネル処理順序および予測元チャンネルの符号化処理の流れの一例について示したフローチャートである。It is the flowchart shown about the example of the flow of the color difference channel process order in the said moving image encoder, and the encoding process of a prediction source channel. 本発明の他の実施形態に係る動画像復号装置が備える予測画像生成部の構成の他の例について示す機能ブロック図である。It is a functional block diagram shown about the other example of a structure of the estimated image generation part with which the moving image decoding apparatus which concerns on other embodiment of this invention is provided. 上記予測画像生成部における色差予測画像生成処理の概略的な流れについて例示したフローチャートである。It is the flowchart which illustrated about the schematic flow of the color difference estimated image generation process in the said estimated image generation part. 上記動画像復号装置が備えるLUT導出部によるLUT更新処理の流れの一例について示したフローチャートである。It is the flowchart shown about an example of the flow of the LUT update process by the LUT derivation | leading-out part with which the said moving image decoding apparatus is provided. 追加登録されたエントリを用いた再補間について説明する図である。It is a figure explaining re-interpolation using the additionally registered entry. 原画像と、輝度画素および色差画素との対応関係の一例について示す図である。It is a figure shown about an example of the correspondence of an original image, a brightness | luminance pixel, and a color difference pixel. Yと、U,Vとの解像度が違う場合のサンプル点の採り方について示す図である。It is a figure shown about the sampling method when the resolutions of Y, U, and V are different. 画像の輝度(Y)-色差(U)をプロットしたグラフであり、(a)および(b)において2種類の画像について示している。FIG. 5 is a graph plotting luminance (Y) -color difference (U) of an image, and two types of images are shown in (a) and (b). 上記動画像符号化装置を搭載した送信装置、および、上記動画像復号装置を搭載した受信装置の構成について示した図である。(a)は、動画像符号化装置を搭載した送信装置を示しており、(b)は、動画像復号装置を搭載した受信装置を示している。It is the figure shown about the structure of the transmitter which mounts the said moving image encoder, and the receiver which mounts the said moving image decoder. (A) shows a transmitting apparatus equipped with a moving picture coding apparatus, and (b) shows a receiving apparatus equipped with a moving picture decoding apparatus. 上記動画像符号化装置を搭載した記録装置、および、上記動画像復号装置を搭載した再生装置の構成について示した図である。(a)は、動画像符号化装置を搭載した記録装置を示しており、(b)は、動画像復号装置を搭載した再生装置を示している。It is the figure shown about the structure of the recording device which mounts the said moving image encoder, and the reproducing | regenerating apparatus which mounts the said moving image decoder. (A) shows a recording apparatus equipped with a moving picture coding apparatus, and (b) shows a reproduction apparatus equipped with a moving picture decoding apparatus.
〔1〕実施形態1
 本発明の一実施形態について図1~図22を参照して説明する。まず、図2を参照しながら、動画像復号装置(画像復号装置)1および動画像符号化装置(画像符号化装置)2の概要について説明する。図2は、動画像復号装置1の概略的構成を示す機能ブロック図である。
[1] Embodiment 1
An embodiment of the present invention will be described with reference to FIGS. First, an overview of the moving picture decoding apparatus (image decoding apparatus) 1 and the moving picture encoding apparatus (image encoding apparatus) 2 will be described with reference to FIG. FIG. 2 is a functional block diagram showing a schematic configuration of the moving picture decoding apparatus 1.
 図2に示す動画像復号装置1および動画像符号化装置2は、H.264/MPEG-4 AVC規格に採用されている技術、VCEG(Video Coding Expert Group)における共同開発用コーデックであるKTAソフトウェアに採用されている技術、および、その後継コーデックであるTMuC(Test Model under Consideration)ソフトウェアに採用されている技術を実装している。 The video decoding device 1 and the video encoding device 2 shown in FIG. H.264 / MPEG-4 AVC standard technology, VCEG (Video Coding Expert Group) technology used in KTA software, which is a joint development codec, and successor codec TMuC (Test Model Under Consideration ) The technology used in the software is implemented.
 動画像復号装置1には、動画像符号化装置2が動画像を符号化した符号化データ(符号化データのデータ構造)#1が入力される。動画像復号装置1は、入力された符号化データ#1を復号して動画像#2を外部に出力する。動画像復号装置1の詳細な説明に先立ち、符号化データ#1の構成を以下に説明する。 The moving image decoding apparatus 1 receives encoded data (data structure of encoded data) # 1 obtained by encoding a moving image by the moving image encoding apparatus 2. The video decoding device 1 decodes the input encoded data # 1 and outputs the video # 2 to the outside. Prior to detailed description of the moving picture decoding apparatus 1, the configuration of the encoded data # 1 will be described below.
 〔符号化データの構成〕
 図3を用いて、動画像符号化装置2によって生成され、動画像復号装置1によって復号される符号化データ#1の構成について説明する。符号化データ#1は、シーケンスレイヤ、GOP(Group Of Pictures)レイヤ、ピクチャレイヤ、スライスレイヤ、及び、最大符号化単位(LCU:Largest Coding Unit)レイヤからなる階層構造を有している。
[Configuration of encoded data]
The configuration of encoded data # 1 that is generated by the video encoding device 2 and decoded by the video decoding device 1 will be described with reference to FIG. The encoded data # 1 has a hierarchical structure including a sequence layer, a GOP (Group Of Pictures) layer, a picture layer, a slice layer, and a maximum coding unit (LCU: Large Coding Unit) layer.
 符号化データ#1におけるピクチャレイヤ以下の階層の構造を図3に示す。図3の(a)~(f)は、それぞれ、ピクチャレイヤPICT、スライスレイヤS、LCUレイヤLCU、LCUに含まれるリーフCU(図3の(d)ではCULと表記)、インター予測(画面間予測)パーティションについての予測情報PIであるインター予測情報PI_Inter、イントラ予測(画面内予測)パーティションについての予測情報PIであるイントラ予測情報PI_Intraの構造を示す図である。 FIG. 3 shows the hierarchical structure below the picture layer in the encoded data # 1. 3A to 3F show a picture layer PICT, a slice layer S, an LCU layer LCU, a leaf CU included in the LCU (denoted as CUL in FIG. 3D), and inter prediction (between screens). It is a figure which shows the structure of inter prediction information PI_Inter which is the prediction information PI about (prediction) partition, and intra prediction information PI_Intra which is the prediction information PI about intra prediction (prediction in a screen) partition.
  (ピクチャレイヤ)
 ピクチャレイヤPICTは、処理対象のピクチャである対象ピクチャを復号するために動画像復号装置1が参照するデータの集合である。ピクチャレイヤPICTは、図3の(a)に示すように、ピクチャヘッダPH、及び、スライスレイヤS1~SNSを含んでいる(NSはピクチャレイヤPICTに含まれるスライスレイヤの総数)。なお、以下、スライスレイヤS1~SNSのそれぞれを区別する必要が無い場合、符号の添え字を省略して記述することがある。また、符号化データ#1に含まれる他の構成についても同様である。
(Picture layer)
The picture layer PICT is a set of data referred to by the video decoding device 1 in order to decode a target picture that is a processing target picture. As shown in FIG. 3A, the picture layer PICT includes a picture header PH and slice layers S 1 to S NS (NS is the total number of slice layers included in the picture layer PICT). Hereinafter, when it is not necessary to distinguish each of the slice layers S 1 to S NS , the reference numerals may be omitted. The same applies to other configurations included in the encoded data # 1.
 ピクチャヘッダPHには、対象ピクチャの復号方法を決定するために動画像復号装置1が参照する符号化パラメータ群が含まれている。例えば、動画像符号化装置2が符号化の際に用いた可変長符号化のモードを示す符号化モード情報(entoropy_coding_mode_flag)は、ピクチャヘッダPHに含まれる符号化パラメータの一例である。entorpy_coding_mode_flagが0の場合、当該ピクチャは、CAVLC(Context-based Adaptive Variable Length Coding)によって符号化されており、entorpy_coding_mode_flagが1である場合、当該ピクチャは、CABAC(Context-based Adaptive Binary Arithmetic Coding)によって符号化されている。 The picture header PH includes a coding parameter group referred to by the video decoding device 1 in order to determine a decoding method of the target picture. For example, the encoding mode information (entoropy_coding_mode_flag) indicating the variable length encoding mode used in encoding by the moving image encoding device 2 is an example of an encoding parameter included in the picture header PH. When entorpy_coding_mode_flag is 0, the picture is encoded by CAVLC (Context-based Adaptive Variable Length Coding). It has become.
  (スライスレイヤ)
 ピクチャレイヤPICTに含まれる各スライスレイヤSは、処理対象のスライスである対象スライスを復号するために動画像復号装置1が参照するデータの集合である。スライスレイヤSは、図3の(b)に示すように、スライスヘッダSH、及び、LCUレイヤLCU1~LCUNC(NCはスライスSに含まれるLCUの総数)を含んでいる。
(Slice layer)
Each slice layer S included in the picture layer PICT is a set of data referred to by the video decoding device 1 in order to decode a target slice that is a processing target slice. As shown in FIG. 3B, the slice layer S includes a slice header SH and LCU layers LCU 1 to LCU NC (NC is the total number of LCUs included in the slice S).
 スライスヘッダSHには、対象スライスの復号方法を決定するために動画像復号装置1が参照する符号化パラメータ群が含まれる。スライスタイプを指定するスライスタイプ指定情報(slice_type)は、スライスヘッダSHに含まれる符号化パラメータの一例である。また、スライスヘッダSHには、動画像復号装置1の備えるループフィルタによって参照されるフィルタパラメータFPが含まれる。 The slice header SH includes a coding parameter group that the moving image decoding apparatus 1 refers to in order to determine a decoding method of the target slice. Slice type designation information (slice_type) for designating a slice type is an example of an encoding parameter included in the slice header SH. Further, the slice header SH includes a filter parameter FP that is referred to by a loop filter included in the video decoding device 1.
 なお、スライスタイプ指定情報により指定可能なスライスタイプとしては、(1)符号化の際にイントラ予測のみを用いるIスライス、(2)符号化の際に単方向予測、又は、イントラ予測を用いるPスライス、(3)符号化の際に単方向予測、双方向予測、又は、イントラ予測を用いるBスライスなどが挙げられる。 As slice types that can be specified by the slice type specification information, (1) I slice using only intra prediction at the time of encoding, and (2) P using unidirectional prediction or intra prediction at the time of encoding. Slice, (3) B-slice using unidirectional prediction, bidirectional prediction, or intra prediction at the time of encoding.
  (LCUレイヤ)
 スライスレイヤSに含まれる各LCUレイヤLCUは、処理対象のLCUである対象LCUを復号するために動画像復号装置1が参照するデータの集合である。LCUレイヤLCUは、図3の(c)に示すように、LCUヘッダLCUH、および、LCUを4分木分割することにより得られる複数の符号化単位(CU:Coding Unit)CU1~CUNLを含んでいる。
(LCU layer)
Each LCU layer LCU included in the slice layer S is a set of data that the video decoding device 1 refers to in order to decode the target LCU that is the processing target LCU. LCU layer LCU, as shown in (c) of FIG. 3, LCU header LCUH, and a plurality of coding units obtained by the quadtree dividing the LCU: the (CU Coding Unit) CU 1 ~ CU NL Contains.
 各CUのとり得るサイズは、符号化データ#1のシーケンスパラメータセットSPSに含まれる、LCUのサイズおよび階層深度(hierarchical depth)に依存する。例えば、LCUのサイズが128×128画素であって、最大階層深度が5である場合には、当該LCUに含まれるCUは、5種類のサイズ、すなわち、128×128画素、64×64画素、32×32画素、16×16画素、および、8×8画素の何れかをとり得る。また、それ以上分割されないCUをリーフCUと呼ぶことにする。 The size that each CU can take depends on the LCU size and the hierarchical depth included in the sequence parameter set SPS of the encoded data # 1. For example, when the size of the LCU is 128 × 128 pixels and the maximum hierarchical depth is 5, the CU included in the LCU has five types of sizes, that is, 128 × 128 pixels, 64 × 64 pixels, Any of 32 × 32 pixels, 16 × 16 pixels, and 8 × 8 pixels can be taken. A CU that is not further divided is called a leaf CU.
  (LCUヘッダ)
 LCUヘッダLCUHには、対象LCUの復号方法を決定するために動画像復号装置1が参照する符号化パラメータが含まれる。具体的には、図3の(c)に示すように、対象LCUの各リーフCUへの分割パターンを指定するCU分割情報SP_CU、および、量子化ステップの大きさを指定する量子化パラメータ差分Δqp(mb_qp_delta)が含まれる。
(LCU header)
The LCU header LCUH includes an encoding parameter referred to by the video decoding device 1 in order to determine a decoding method of the target LCU. Specifically, as shown in FIG. 3C, CU partition information SP_CU that specifies a partition pattern for each leaf CU of the target LCU, and a quantization parameter difference Δqp that specifies the size of the quantization step. (Mb_qp_delta) is included.
 CU分割情報SP_CUは、具体的には、対象LCUに含まれる各CU(およびリーフCU)の形状、サイズ、および、対象LCU内での位置を指定する情報である。なお、CU分割情報SP_CUは、リーフCUの形状やサイズを明示的に含んでいる必要は必ずしもない。例えばCU分割情報SP_CUは、LCU全体またはLCUの部分領域を四分割するか否かを示すフラグ(split_coding_unit_flag)の集合であってもよい。その場合、LCUの形状やサイズを併用することで各リーフCUの形状やサイズを特定できる。 CU division information SP_CU is information that specifies the shape and size of each CU (and leaf CU) included in the target LCU, and the position in the target LCU. Note that the CU partition information SP_CU does not necessarily need to explicitly include the shape and size of the leaf CU. For example, the CU partition information SP_CU may be a set of flags (split_coding_unit_flag) indicating whether or not the entire LCU or a partial region of the LCU is divided into four. In that case, the shape and size of each leaf CU can be specified by using the shape and size of the LCU together.
 また、量子化パラメータ差分Δqpは、対象LCUにおける量子化パラメータqpと、そのLCUの直前に符号化されたLCUにおける量子化パラメータqp’との差分qp-qp’である。 Further, the quantization parameter difference Δqp is a difference qp−qp ′ between the quantization parameter qp in the target LCU and the quantization parameter qp ′ in the LCU encoded immediately before the LCU.
  (リーフCU)
 それ以上分割することができないCU(リーフCU)は予測単位(PU:Prediction Unit)、および、変換単位(TU:Transform Unit)として取り扱われる。
(Leaf CU)
A CU (leaf CU) that cannot be further divided is treated as a prediction unit (PU: Prediction Unit) and a transform unit (TU: Transform Unit).
 図3の(d)に示すように、リーフCU(図3の(d)においてはCULと表記)は、(1)動画像復号装置1により予測画像を生成する際に参照されるPU情報PUI、および、(2)動画像復号装置1により残差データを復号する際に参照されるTU情報TUIを含んでいる。なお、PU情報PUIにはスキップフラグSKIPが含まれている場合があり、スキップフラグSKIPの値が1の場合、TU情報は省略される。 As shown in (d) of FIG. 3, the leaf CU (denoted as CUL in (d) of FIG. 3) is (1) PU information PUI that is referred to when the moving image decoding apparatus 1 generates a predicted image. And (2) the TU information TUI that is referred to when the moving image decoding apparatus 1 decodes the residual data. The PU information PUI may include a skip flag SKIP. When the value of the skip flag SKIP is 1, the TU information is omitted.
 PU情報PUIは、図3の(d)に示すように、予測タイプ情報PT、および、予測情報PIを含んでいる。予測タイプ情報PTは、対象リーフCU(対象PU)についての予測画像生成方法として、イントラ予測を用いるのか、または、インター予測を用いるのかを指定する情報である。予測情報PIは、予測タイプ情報PTが何れの予測方法を指定するのかに応じて、イントラ予測情報PI_Intra、または、インター予測情報PI_Interより構成される。以下では、イントラ予測が適用されるPUをイントラPUとも呼称し、インター予測が適用されるPUをインターPUとも呼称する。 The PU information PUI includes prediction type information PT and prediction information PI, as shown in FIG. The prediction type information PT is information that specifies whether intra prediction or inter prediction is used as a predicted image generation method for the target leaf CU (target PU). The prediction information PI includes intra prediction information PI_Intra or inter prediction information PI_Inter depending on which prediction method is specified by the prediction type information PT. Hereinafter, a PU to which intra prediction is applied is also referred to as an intra PU, and a PU to which inter prediction is applied is also referred to as an inter PU.
 PU情報PUIには、対象PUに含まれる各パーティションの形状、サイズ、および、対象PU内での位置を指定する情報が含まれる。ここで、パーティションとは、対象リーフCUを構成する1又は複数の重複しない領域のことであり、予測画像の生成は、パーティションを単位として行われる。 The PU information PUI includes information specifying the shape and size of each partition included in the target PU and the position in the target PU. Here, the partition is one or a plurality of non-overlapping areas constituting the target leaf CU, and the generation of the predicted image is performed in units of partitions.
 TU情報TUIは、図3の(d)に示すように、対象リーフCU(対象TU)の各ブロックへの分割パターンを指定するTU分割情報SP_TU、および、量子化予測残差QD1~QDNT(NTは、対象TUに含まれるブロックの総数)を含んでいる。 As shown in FIG. 3 (d), the TU information TUI includes TU partition information SP_TU that specifies a partition pattern for each block of the target leaf CU (target TU), and quantized prediction residuals QD 1 to QD NT. (NT is the total number of blocks included in the target TU).
 TU分割情報SP_TUは、具体的には、対象TUに含まれる各ブロックの形状、サイズ、および、対象TU内での位置を指定する情報である。各TUは、例えば、64×64画素から2×2画素までのサイズをとり得る。ここで、ブロックとは、対象リーフCUを構成する1又は複数の重複しない領域のことであり、予測残差の符号化・復号は、TUあるいはTUを分割したブロックを単位として行われる。 TU partition information SP_TU is information that specifies the shape and size of each block included in the target TU and the position in the target TU. Each TU can be, for example, a size from 64 × 64 pixels to 2 × 2 pixels. Here, the block is one or a plurality of non-overlapping areas constituting the target leaf CU, and prediction residual encoding / decoding is performed in units of TUs or blocks obtained by dividing TUs.
 各量子化予測残差QDは、動画像符号化装置2が以下の処理1~3を、処理対象のブロックである対象ブロックに施すことによって生成した符号化データである。処理1:符号化対象画像から予測画像を減算した予測残差をDCT変換(Discrete Cosine Transform)する。処理2:処理1にて得られたDCT係数を量子化する。処理3:処理2にて量子化されたDCT係数を可変長符号化する。上述した量子化パラメータqpは、動画像符号化装置2がDCT係数を量子化する際に用いた量子化ステップQPの大きさを表す(QP=2qp/6)。 Each quantized prediction residual QD is encoded data generated by the moving image encoding apparatus 2 performing the following processes 1 to 3 on a target block that is a processing target block. Process 1: DCT transform (Discrete Cosine Transform) is performed on the prediction residual obtained by subtracting the prediction image from the encoding target image. Process 2: The DCT coefficient obtained in Process 1 is quantized. Process 3: The DCT coefficient quantized in Process 2 is variable length encoded. The quantization parameter qp described above represents the magnitude of the quantization step QP used when the moving picture coding apparatus 2 quantizes the DCT coefficient (QP = 2 qp / 6 ).
  (インター予測情報PI_Inter)
 インター予測情報PI_Interは、動画像復号装置1が、インター予測によってインター予測画像を生成する際に参照される符号化パラメータが含まれる。図3の(e)に示すように、インター予測情報PI_Interは、対象PUの各パーティションへの分割パターンを指定するインターPU分割情報SP_Inter、および、各パーティションについてのインター予測パラメータPP_Inter1~PP_InterNe(Neは、対象PUに含まれるインター予測パーティションの総数)を含んでいる。
(Inter prediction information PI_Inter)
The inter prediction information PI_Inter includes a coding parameter that is referred to when the video decoding device 1 generates an inter prediction image by inter prediction. As shown in FIG. 3E, the inter prediction information PI_Inter includes inter PU partition information SP_Inter that specifies a partition pattern of the target PU into each partition, and inter prediction parameters PP_Inter1 to PP_InterNe (Ne for each partition). , The total number of inter prediction partitions included in the target PU).
 インターPU分割情報SP_Interは、具体的には、対象PU(インターPU)に含まれる各インター予測パーティションの形状、サイズ、および、対象PU内での位置を指定する情報である。 Specifically, the inter-PU partition information SP_Inter is information for designating the shape and size of each inter prediction partition included in the target PU (inter PU) and the position in the target PU.
 インターPUは、2N×2N画素、2N×N画素、N×2N画素、およびN×N画素の4つの対称的分割(symmetric splittings)、並びに、2N×nU画素、2N×nD画素、nL×2N画素、およびnR×2N画素の4つの非対称的分割(asymmetric splittings)により、合計8種類のパーティションに分割することが可能である。ここで、Nの具体的な値は、当該PUが属するCUのサイズによって規定され、nU、nD、nL、および、nRの具体的な値は、Nの値に応じて定められる。例えば、128×128画素のインターPUは、128×128画素、128×64画素、64×128画素、64×64画素、128×32画素、128×96画素、32×128画素、および、96×128画素のインター予測パーティションへ分割することが可能である。 The inter PU is composed of four symmetric splittings of 2N × 2N pixels, 2N × N pixels, N × 2N pixels, and N × N pixels, and 2N × nU pixels, 2N × nD pixels, and nL × 2N. It is possible to divide into 8 types of partitions in total by four asymmetric splits of pixels and nR × 2N pixels. Here, the specific value of N is defined by the size of the CU to which the PU belongs, and the specific values of nU, nD, nL, and nR are determined according to the value of N. For example, an inter PU of 128 × 128 pixels is 128 × 128 pixels, 128 × 64 pixels, 64 × 128 pixels, 64 × 64 pixels, 128 × 32 pixels, 128 × 96 pixels, 32 × 128 pixels, and 96 × It is possible to divide into 128-pixel inter prediction partitions.
  (インター予測パラメータ)
 インター予測パラメータPP_Interは、図3の(e)に示すように、参照画像インデックスRIと、推定動きベクトルインデックスPMVIと、動きベクトル残差MVDとを含んでいる。
(Inter prediction parameter)
As shown in FIG. 3E, the inter prediction parameter PP_Inter includes a reference image index RI, an estimated motion vector index PMVI, and a motion vector residual MVD.
  (イントラ予測情報PI_Intra)
 イントラ予測情報PI_Intraは、動画像復号装置1が、イントラ予測によってイントラ予測画像を生成する際に参照される符号化パラメータが含まれる。図3の(f)に示すように、イントラ予測情報PI_Intraは、対象PU(イントラPU)の各パーティションへの分割パターンを指定するイントラPU分割情報SP_Intra、および、各パーティションについてのイントラ予測パラメータPP_Intra1~PP_IntraNA(NAは、対象PUに含まれるイントラ予測パーティションの総数)を含んでいる。
(Intra prediction information PI_Intra)
The intra prediction information PI_Intra includes an encoding parameter that is referred to when the video decoding device 1 generates an intra predicted image by intra prediction. As shown in (f) of FIG. 3, the intra prediction information PI_Intra includes intra PU partition information SP_Intra that specifies a partition pattern of the target PU (intra PU) into each partition, and intra prediction parameters PP_Intra 1 for each partition. -PP_Intra NA (NA is the total number of intra prediction partitions included in the target PU).
 イントラPU分割情報SP_Intraは、具体的には、対象PUに含まれる各イントラ予測パーティションの形状、サイズ、および、対象PU内での位置を指定する情報である。イントラPU分割情報SP_Intraには、対象PUをパーティションに分割するか否かを指定するイントラ分割フラグ(intra_split_flag)が含まれている。イントラ分割フラグが1であれば、対象PUは、4つのパーティションへと対称的に分割され、イントラ分割フラグが0であれば、対象PUは、分割されることなく、対象PU自身が1つのパーティションとして取り扱われる。したがって、対象PUのサイズを2N×2N画素とすると、イントラ予測パーティションは、2N×2N画素(分割なし)、および、N×N画素(4分割)の何れかのサイズをとり得る(ここで、N=2n、nは1以上の任意の整数)。例えば、128×128画素のイントラPUは、128×128画素、および、64×64画素のイントラ予測パーティションへ分割することが可能である。 Specifically, the intra-PU partition information SP_Intra is information that specifies the shape and size of each intra-predicted partition included in the target PU, and the position in the target PU. The intra PU split information SP_Intra includes an intra split flag (intra_split_flag) that specifies whether or not the target PU is split into partitions. If the intra partition flag is 1, the target PU is divided symmetrically into four partitions. If the intra partition flag is 0, the target PU is not divided and the target PU itself is one partition. Are treated as Therefore, if the size of the target PU is 2N × 2N pixels, the intra prediction partition can take any of 2N × 2N pixels (no division) and N × N pixels (four divisions) (where, N = 2 n , n is an arbitrary integer of 1 or more). For example, a 128 × 128 pixel intra PU can be divided into 128 × 128 pixel and 64 × 64 pixel intra prediction partitions.
 なお、イントラ予測パラメータPP_Intraの詳細について後述する。 Details of the intra prediction parameter PP_Intra will be described later.
 〔動画像復号装置〕
 以下では、本実施形態に係る動画像復号装置1の構成について、図1~図22を参照して説明する。
[Video decoding device]
Hereinafter, the configuration of the video decoding device 1 according to the present embodiment will be described with reference to FIGS.
  (動画像復号装置の概要)
 動画像復号装置1は、パーティション毎に予測画像を生成し、生成された予測画像と、符号化データ#1から復号された予測残差とを加算することによって復号画像#2を生成し、生成された復号画像#2を外部に出力する。
(Outline of video decoding device)
The video decoding device 1 generates a predicted image for each partition, generates a decoded image # 2 by adding the generated predicted image and the prediction residual decoded from the encoded data # 1, and generates The decoded image # 2 is output to the outside.
 ここで、予測画像の生成は、符号化データ#1を復号することによって得られる符号化パラメータを参照して行われる。ここで、符号化パラメータとは、予測画像を生成するために参照されるパラメータのことであり、画面間予測において参照される動きベクトルや画面内予測において参照される予測モードなどの予測パラメータに加えて、パーティションのサイズや形状、ブロックのサイズや形状、および、原画像と予測画像との残差データなどが含まれる。以下では、符号化パラメータに含まれる情報のうち、上記残差データを除く全ての情報の集合を、サイド情報と呼ぶ。 Here, the generation of the predicted image is performed with reference to the encoding parameter obtained by decoding the encoded data # 1. Here, the encoding parameter is a parameter referred to in order to generate a prediction image, and in addition to a prediction parameter such as a motion vector referred to in inter-screen prediction and a prediction mode referred to in intra-screen prediction. Partition size and shape, block size and shape, and residual data between the original image and the predicted image. Hereinafter, a set of all information excluding the residual data among the information included in the encoding parameter is referred to as side information.
 また、以下では、上記予測単位が、LCUを構成するパーティションである場合を例に挙げ説明を行うが、本実施形態はこれに限定されるものではなく、予測単位がパーティションよりも大きい単位である場合、および、予測単位がパーティションよりも小さい単位である場合に対しても適用することができる。 Further, in the following description, the case where the prediction unit is a partition constituting the LCU will be described as an example. However, the present embodiment is not limited to this, and the prediction unit is a unit larger than the partition. The present invention can also be applied to the case where the prediction unit is a unit smaller than the partition.
 また、以下では、復号の対象となるフレーム(ピクチャ)、スライス、LCU、ブロック、および、パーティションをそれぞれ、対象フレーム、対象スライス、対象LCU、対象ブロック、および、対象パーティションと呼ぶことにする。 In the following, a frame (picture), a slice, an LCU, a block, and a partition to be decoded are referred to as a target frame, a target slice, a target LCU, a target block, and a target partition, respectively.
 なお、LCUのサイズは、例えば64×64画素であり、パーティションのサイズは、例えば、64×64画素、32×32画素、16×16画素、8×8画素や4×4画素などであるが、これらのサイズは本実施形態を限定するものではなく、LCUのサイズおよびパーティションはそれらのサイズ以外のサイズであってもよい。 The LCU size is, for example, 64 × 64 pixels, and the partition size is, for example, 64 × 64 pixels, 32 × 32 pixels, 16 × 16 pixels, 8 × 8 pixels, 4 × 4 pixels, or the like. These sizes do not limit the present embodiment, and the size and partition of the LCU may be other sizes.
  (動画像復号装置の構成)
 再び、図2を参照して、動画像復号装置1の概略的構成について説明すると次のとおりである。図2は、動画像復号装置1の概略的構成について示した機能ブロック図である。
(Configuration of video decoding device)
Referring to FIG. 2 again, the schematic configuration of the moving picture decoding apparatus 1 will be described as follows. FIG. 2 is a functional block diagram showing a schematic configuration of the moving picture decoding apparatus 1.
 図2に示すように動画像復号装置1は、可変長符号逆多重化部11、逆量子化・逆変換部12、予測画像生成部(チャンネル復号手段)13、加算器(チャンネル復号手段)14およびフレームメモリ15を備えている。 As shown in FIG. 2, the moving picture decoding apparatus 1 includes a variable length code demultiplexing unit 11, an inverse quantization / inverse conversion unit 12, a predicted image generation unit (channel decoding unit) 13, and an adder (channel decoding unit) 14. And a frame memory 15.
   [可変長符号逆多重化部]
 可変長符号逆多重化部11は、動画像復号装置1に入力された1フレーム分の符号化データ#1を、逆多重化することで、図3に示した階層構造に含まれる各種情報に分離する。例えば、可変長符号逆多重化部11は、各種ヘッダに含まれる情報を参照して、符号化データ#1を、スライス、LCUに順次分離する。
[Variable length code demultiplexer]
The variable-length code demultiplexing unit 11 demultiplexes the encoded data # 1 for one frame input to the video decoding device 1 to obtain various kinds of information included in the hierarchical structure shown in FIG. To separate. For example, the variable length code demultiplexing unit 11 refers to information included in various headers, and sequentially separates the encoded data # 1 into slices and LCUs.
 ここで、各種ヘッダには、(1)対象フレームのスライスへの分割方法についての情報、および(2)対象スライスに属するLCUのサイズ、形状および対象スライス内での位置についての情報が含まれる。 Here, the various headers include (1) information on the method of dividing the target frame into slices, and (2) information on the size, shape, and position of the LCU belonging to the target slice.
 そして、可変長符号逆多重化部11は、符号化LCUヘッダLCUHに含まれるCU分割情報SP_CUを参照して、対象LCUを、リーフCUに分割する。また、可変長符号逆多重化部11は、対象リーフCU:CULについてTU情報TUI、およびPU情報PUIを取得する。 Then, the variable length code demultiplexer 11 refers to the CU partition information SP_CU included in the encoded LCU header LCUH and divides the target LCU into leaf CUs. In addition, the variable-length code demultiplexing unit 11 acquires the TU information TUI and the PU information PUI for the target leaf CU: CUL.
 可変長符号逆多重化部11は、対象リーフCUについて得られたTU情報TUIを逆量子化・逆変換部12に供給する。また、可変長符号逆多重化部11は、対象リーフCUについて得られたPU情報PUIを予測画像生成部13に供給する。 The variable length code demultiplexing unit 11 supplies the TU information TUI obtained for the target leaf CU to the dequantization / inverse transform unit 12. Further, the variable length code demultiplexing unit 11 supplies the PU information PUI obtained for the target leaf CU to the predicted image generation unit 13.
   [逆量子化・逆変換部]
 逆量子化・逆変換部12は、対象リーフCUについて、ブロックごとに、量子化予測残差の逆量子化・逆変換を行う。
[Inverse quantization / inverse transform unit]
The inverse quantization / inverse transform unit 12 performs inverse quantization / inverse transform of the quantization prediction residual for each block for the target leaf CU.
 具体的には、逆量子化・逆変換部12は、まず、可変長符号逆多重化部11から供給される、対象リーフCUについてのTU情報TUIからTU分割情報SP_TUを復号する。 Specifically, the inverse quantization / inverse transform unit 12 first decodes the TU partition information SP_TU from the TU information TUI about the target leaf CU supplied from the variable length code demultiplexer 11.
 また、逆量子化・逆変換部12は、復号したTU分割情報SP_TUに従って、対象リーフCUを1または複数のブロックへと分割する。 Further, the inverse quantization / inverse transform unit 12 divides the target leaf CU into one or a plurality of blocks according to the decoded TU partition information SP_TU.
 また、逆量子化・逆変換部12は、各ブロックについて、TU情報TUIからTU分割情報SP_TUおよび量子化予測残差QDを復号する。 Also, the inverse quantization / inverse transform unit 12 decodes the TU partition information SP_TU and the quantized prediction residual QD from the TU information TUI for each block.
 そして、逆量子化・逆変換部12は、逆量子化および逆DCT変換(Inverse Discrete Cosine Transform)することによって、各対象パーティションについて、画素毎の予測残差Dを復元する。逆量子化・逆変換部12は、復元した予測残差Dを加算器14に供給する。 The inverse quantization / inverse transform unit 12 restores the prediction residual D for each pixel for each target partition by performing inverse quantization and inverse DCT transform (Inverse DiscretecreCosine Transform). The inverse quantization / inverse transform unit 12 supplies the restored prediction residual D to the adder 14.
   [予測画像生成部]
 予測画像生成部13は、対象リーフCUに含まれる各パーティションについて、パーティションの周辺の復号済み画像である局所復号画像P’を参照して、イントラ予測またはインター予測により予測画像Predを生成する。
[Predicted image generator]
For each partition included in the target leaf CU, the predicted image generation unit 13 refers to a local decoded image P ′ that is a decoded image around the partition, and generates a predicted image Pred by intra prediction or inter prediction.
 なお、イントラ予測およびインター予測には、それぞれ輝度の予測および色差の予測がある。また、色差のイントラ予測においてチャンネル間予測を行う場合、予測画像生成部13は、輝度復号画像を参照する。 Note that intra prediction and inter prediction include luminance prediction and color difference prediction, respectively. In addition, when performing inter-channel prediction in color difference intra prediction, the predicted image generation unit 13 refers to a luminance decoded image.
 また、以下では、イントラ予測による予測画像Predの生成処理について説明するが、これに限られず、動画像復号装置1は、インター予測により予測画像Predを生成してもよい。 In the following, the generation process of the predicted image Pred by intra prediction will be described, but the present invention is not limited to this, and the video decoding device 1 may generate the predicted image Pred by inter prediction.
 また、イントラ予測は、画面内予測または空間予測と称されることがあるが、以下では、イントラ予測という表現に統一している。 Intra prediction is sometimes referred to as intra prediction or spatial prediction, but in the following, it is unified with the expression intra prediction.
 また、局所復号画像P’は、より具体的には、輝度に関する輝度局所復号画像P’と、色差に関する色差局所復号画像P’を含んでいる。また、色差局所復号画像P’は、Uチャンネルに関する色差局所復号画像P’と、および、Vチャンネルに関する色差局所復号画像P’とを含んでいる。 More specifically, the local decoded image P ′ includes a luminance local decoded image P ′ Y related to luminance and a color difference local decoded image P ′ C related to color difference. The color difference local decoded image P 'C, the color difference local decoded image P about U-channel' and includes a U, and the color difference relates V channel and the local decoded image P 'V.
 予測画像生成部13は、具体的には、次のように動作する。まず、予測画像生成部13は、可変長符号逆多重化部11から供給される、対象リーフCUについてのPU情報PUIを復号する。続いて、予測画像生成部13は、PU情報PUIに従って、対象リーフCUの各パーティションへの分割パターンを決定する。さらに、予測画像生成部13は、PU情報PUIに従って、各パーティションの予測モードを選択して、選択した各予測モードを各パーティションに割り付ける。 Specifically, the predicted image generation unit 13 operates as follows. First, the predicted image generation unit 13 decodes the PU information PUI for the target leaf CU supplied from the variable length code demultiplexing unit 11. Subsequently, the predicted image generation unit 13 determines a division pattern for each partition of the target leaf CU according to the PU information PUI. Further, the predicted image generation unit 13 selects a prediction mode of each partition according to the PU information PUI, and assigns each selected prediction mode to each partition.
 そして、予測画像生成部13は、対象リーフCUに含まれる各パーティションについての予測画像Predを、選択した予測モードと該パーティションの周辺の局所復号画像P’の画素値とを参照して生成する。予測画像生成部13は、対象リーフCUについて生成した予測画像Predを加算器14に供給する。 The predicted image generation unit 13 generates a predicted image Pred for each partition included in the target leaf CU with reference to the selected prediction mode and the pixel values of the local decoded image P ′ around the partition. The predicted image generation unit 13 supplies the predicted image Pred generated for the target leaf CU to the adder 14.
 なお、予測画像Predは、具体的には、輝度に関する輝度予測画像PredYと、色差に関する色差予測画像PredCを含む。また、色差予測画像PredCは、Uチャンネルに関する色差予測画像PredU、および、Vチャンネルに関する色差予測画像PredVを含む。また、予測画像生成部13のより具体的な構成については、後述する。 Note that the predicted image Pred specifically includes a luminance predicted image PredY related to luminance and a color difference predicted image PredC related to color difference. Further, the color difference prediction image PredC includes a color difference prediction image PredU for the U channel and a color difference prediction image PredV for the V channel. Further, a more specific configuration of the predicted image generation unit 13 will be described later.
   [加算器]
 加算器14は、予測画像生成部13より供給される予測画像Predと、逆量子化・逆変換部12より供給される予測残差Dとを加算することによって、対象リーフCUについての復号画像Pを生成する。復号画像Pは、輝度復号画像(以下、輝度復号画像Pにて参照する)と色差復号画像とを含む。
[Adder]
The adder 14 adds the predicted image Pred supplied from the predicted image generation unit 13 and the prediction residual D supplied from the inverse quantization / inverse transform unit 12, thereby decoding the decoded image P for the target leaf CU. Is generated. Decoded image P includes, luminance decoded image (hereinafter, referred to by the luminance decoded picture P Y) with a color difference decoded image.
   [フレームメモリ]
 フレームメモリ15には、復号された復号画像Pが順次記録される。フレームメモリ15には、対象LCUを復号する時点において、当該対象LCUよりも先に復号された全てのLCU(例えば、ラスタスキャン順で先行する全てのLCU)に対応する復号画像が記録されている。
[Frame memory]
The decoded image P that has been decoded is sequentially recorded in the frame memory 15. In the frame memory 15, at the time of decoding the target LCU, decoded images corresponding to all the LCUs decoded before the target LCU (for example, all the LCUs preceding in the raster scan order) are recorded. .
 なお、動画像復号装置1において、画像内の全てのLCUに対して、LCU単位の復号画像生成処理が終わった時点で、動画像復号装置1に入力された1フレーム分の符号化データ#1に対応する復号画像#2が外部に出力される。 Note that, in the moving image decoding apparatus 1, one frame of encoded data # 1 input to the moving image decoding apparatus 1 at the time when the decoded image generation processing for each LCU is completed for all the LCUs in the image. The decoded image # 2 corresponding to is output to the outside.
  (予測画像生成部について)
 以下では、予測画像生成部13についてより詳しく説明する。
(About predicted image generator)
Hereinafter, the predicted image generation unit 13 will be described in more detail.
   [イントラ予測パラメータのデータ構成]
 次に、イントラ予測パラメータPP_Intra(以下、“イントラ予測パラメータPP”にて参照する)のデータ構成について説明すると次のとおりである。イントラ予測パラメータPP(符号化データのデータ構造)は、例示的には、チャンネル間予測フラグ、色差チャンネル処理順序フラグおよび第2チャンネル予測元チャンネル指定子を含む。
[Data structure of intra prediction parameters]
Next, the data structure of the intra prediction parameter PP_Intra (hereinafter referred to as “intra prediction parameter PP”) will be described as follows. The intra prediction parameter PP (data structure of encoded data) illustratively includes an inter-channel prediction flag, a color difference channel processing order flag, and a second channel prediction source channel specifier.
 チャンネル間予測フラグは、色差をチャンネル間予測により予測するか否かを示すフラグである。チャンネル間予測フラグは、例えば、“1”であれば、色差をチャンネル間予測により予測することを示し、“0”であれば、色差をチャンネル間予測によらないで予測することを示す1ビットの情報である。 The inter-channel prediction flag is a flag indicating whether or not the color difference is predicted by inter-channel prediction. For example, if the inter-channel prediction flag is “1”, it indicates that the color difference is predicted by the inter-channel prediction, and if “0”, 1 bit indicates that the color difference is predicted without using the inter-channel prediction. Information.
 色差チャンネル処理順序フラグは、UチャンネルおよびVチャンネルのいずれから予測処理を行うかを指定するフラグである。色差チャンネル処理順序フラグは、例えば、“0”であればU、Vの順、“1”であればV、Uの順に処理することを示す1ビットの情報である。 The color difference channel processing order flag is a flag for designating whether the prediction process is performed from the U channel or the V channel. The color difference channel processing order flag is, for example, 1-bit information indicating that processing is performed in the order of U and V if “0” and in the order of V and U if “1”.
 第2チャンネル予測元チャンネル指定子は、2番目に予測するチャンネルをいずれのチャンネルから予測するかを指定する情報である。すなわち、2番目に予測するチャンネルは、Yチャンネル、1番目に予測した第1チャンネル、または、これらの両方のいずれかから予測することができる。第2チャンネル予測元チャンネル指定子は、例えば、“0”であれば、Yチャンネルから予測することを示し、“10”であれば、第1チャンネルから予測することを示し、また、“11”であれば、Yチャンネルおよび第1チャンネルの両方から予測することを示す1または2ビットの情報である。 The second channel prediction source channel specifier is information for designating which channel the second predicted channel is predicted from. That is, the second predicted channel can be predicted from either the Y channel, the first predicted first channel, or both. For example, if the second channel prediction source channel specifier is “0”, it indicates that prediction is performed from the Y channel, if “10”, it indicates that prediction is performed from the first channel, and “11”. If it is, it is 1 or 2-bit information indicating that prediction is performed from both the Y channel and the first channel.
 このように、例示的には、処理順序および予測元チャンネルの組み合わせを示す情報は、それぞれ色差チャンネル処理順序フラグおよび第2チャンネル予測元チャンネル指定子という別々の情報として符号化される。 Thus, illustratively, the information indicating the combination of the processing order and the prediction source channel is encoded as separate information of the color difference channel processing order flag and the second channel prediction source channel specifier, respectively.
 なお、以下では、予測元チャンネルを1つとする場合の例について主に説明する。予測元チャンネルが複数である場合については、後述する変形例において詳しく説明する。 In the following, an example in which there is one prediction source channel will be mainly described. A case where there are a plurality of prediction source channels will be described in detail in a later-described modification.
   [色差について]
 次に、図4および図5を用いて、色差について補足的に説明しておくと次のとおりである。図4は、YUV色空間について説明する図であり、輝度Yと、色差の各成分であるU成分、V成分との関係について示している。また、図5は、YUV形式の画像フォーマットについて示す図であり、(a)~(d)は、それぞれ、4:2:0形式、4:4:4形式、4:2:2形式、および4:1:1形式を示す。
[About color difference]
Next, the color difference will be supplementarily described with reference to FIGS. 4 and 5 as follows. FIG. 4 is a diagram for explaining the YUV color space, and shows the relationship between the luminance Y and the U component and V component, which are the components of the color difference. FIG. 5 is a diagram showing an image format of YUV format. (A) to (d) are respectively 4: 2: 0 format, 4: 4: 4 format, 4: 2: 2 format, and 4: 1: 1 format is shown.
 まず、図4を参照しながら、YUV色空間について説明すると次のとおりである。画像は、YUV形式によれば、輝度Yと、色差であるU成分およびV成分とにより表現される。そして、図4に示すように、本来は、YUV色空間において、輝度Yは、U成分およびV成分とはU-V平面に直交する独立した座標系として定義される。 First, the YUV color space will be described with reference to FIG. According to the YUV format, an image is expressed by luminance Y and U and V components that are color differences. As shown in FIG. 4, originally, in the YUV color space, the luminance Y is defined as an independent coordinate system orthogonal to the U-V plane with respect to the U component and the V component.
 図4において、U成分およびV成分を示す座標(U,V)が、原点付近である場合、無彩色となる。一方、U成分およびV成分を示す座標(U,V)が、原点からより遠くになると、より濃い色となる。 In FIG. 4, when the coordinates (U, V) indicating the U component and the V component are near the origin, the color is achromatic. On the other hand, when the coordinates (U, V) indicating the U component and the V component are further away from the origin, the color becomes darker.
 一実装例としては、輝度Y、U成分、V成分は、それぞれ0~255の値をとる。よって、輝度値が0に近くなればなるほど、画像は暗くなり、反対に255に近くなればなるほど、画像は明るくなる。また、U成分およびV成分が、色差値=128のとき、画像は無彩色となり、これを基準として、色差値が、0または255に近くなればなるほど色が濃くなる。 As an implementation example, the luminance Y, U component, and V component each take a value from 0 to 255. Thus, the closer the luminance value is to 0, the darker the image, and vice versa. Further, when the U component and the V component have a color difference value = 128, the image becomes an achromatic color, and with this as a reference, the color becomes darker as the color difference value approaches 0 or 255.
 画像は、U成分の色差値が、0に近いとき、概ね緑色を呈し、255に近いときは、概ね赤色を呈する。また、画像は、V成分の色差値が、0に近いとき、概ね黄色を呈し、255に近いときは、概ね青色を呈する。 When the color difference value of the U component is close to 0, the image is generally green, and when it is close to 255, the image is generally red. In addition, when the color difference value of the V component is close to 0, the image is generally yellow, and when it is close to 255, the image is generally blue.
 ところで、画像を局所的に見れば、使われている画素値の種類は限られている。そして、局所的に見れば、輝度Yは、U成分およびV成分それぞれとの相関を有する。よって、局所的には、この相関を利用して、輝度Yから、U成分およびV成分を導出することが可能となる。 By the way, if the image is viewed locally, the types of pixel values used are limited. When viewed locally, the luminance Y has a correlation with each of the U component and the V component. Therefore, locally, it is possible to derive the U component and the V component from the luminance Y using this correlation.
 以下、輝度YからU成分およびV成分を導出する過程に関する説明においては、次の用語を用いる。まず、輝度YのことをYチャンネルと呼び、U成分およびV成分からなる色差を色差チャンネルと呼ぶ。チャンネルとは、輝度Y、U成分およびV成分を一般化した概念である。 Hereinafter, the following terms will be used in the description of the process of deriving the U component and the V component from the luminance Y. First, the luminance Y is called the Y channel, and the color difference consisting of the U component and the V component is called the color difference channel. A channel is a generalized concept of luminance Y, U component, and V component.
 また、色差チャンネルにおいて、U成分およびV成分を区別する必要があるときは、それぞれUチャンネルおよびVチャンネルと称する。また、輝度Yと、U成分およびV成分との間にある相関を用いて、Yチャンネルから、UチャンネルおよびVチャンネルを予測することを、チャンネル間予測と称する。 Further, when it is necessary to distinguish the U component and the V component in the color difference channel, they are referred to as a U channel and a V channel, respectively. Prediction of the U channel and the V channel from the Y channel using the correlation between the luminance Y and the U component and the V component is referred to as inter-channel prediction.
 次に、輝度および色差の画像フォーマットについて説明すると次のとおりである。色差の解像度を下げても、輝度の解像度を下げたときに比べれば視覚上の影響が少ない。よって、色差の解像度を下げることによりデータ量の削減が図られる。例えば、図5の(a)に示す4:2:0形式では、次のようなデータ構造によりデータ量の削減を図っている。なお、図5の(a)の左側のブロックは、輝度Yの解像度を示しており、右側のブロックは、U成分およびV成分の解像度を示している。以下、(b)~(d)について同様である。 Next, the image format of luminance and color difference will be described as follows. Even if the resolution of the color difference is lowered, the visual effect is less than when the luminance resolution is lowered. Therefore, the data amount can be reduced by reducing the resolution of the color difference. For example, in the 4: 2: 0 format shown in FIG. 5A, the data amount is reduced by the following data structure. Note that the left block of FIG. 5A shows the resolution of the luminance Y, and the right block shows the resolution of the U component and the V component. The same applies to (b) to (d) below.
 図5の(a)に示すように、4:2:0形式では、色差の解像度は、水平方向および垂直方向ともに輝度の解像度の1/2となっている。つまり、全体としては、色差の解像度は、輝度の解像度の1/4である。一般的に、4:2:0形式はテレビジョン放送や民生用ビデオ機器で用いられている。 As shown in FIG. 5A, in the 4: 2: 0 format, the resolution of the color difference is ½ of the luminance resolution in both the horizontal and vertical directions. That is, as a whole, the resolution of the color difference is 1/4 of the resolution of the luminance. Generally, the 4: 2: 0 format is used in television broadcasting and consumer video equipment.
 図5をさらに参照して、YUV形式の画像フォーマットの他の例について説明すると次のとおりである。 Referring further to FIG. 5, another example of the YUV image format will be described as follows.
 まず、図5の(b)に示す4:4:4形式がある。4:4:4形式では、輝度の解像度と、色差の解像度とが同一である。4:4:4形式は、データ量の削減より、高画質が要求されるような場合、例えば、画像処理の専門機器などに用いられている。 First, there is a 4: 4: 4 format shown in FIG. In the 4: 4: 4 format, the luminance resolution and the color difference resolution are the same. The 4: 4: 4 format is used, for example, in specialized equipment for image processing when high image quality is required rather than reducing the amount of data.
 また、図5の(c)に示す4:2:2形式がある。これは、4:4:4形式において水平方向の解像度を1/2としたものである。 Also, there is a 4: 2: 2 format shown in FIG. In the 4: 4: 4 format, the horizontal resolution is halved.
 また、4:2:2形式において、さらに水平方向の解像度を1/2とした4:1:1形式がある(図5の(d))。 Also, in the 4: 2: 2 format, there is a 4: 1: 1 format in which the horizontal resolution is halved ((d) in FIG. 5).
   [予測画像生成部の構成]
 次に、図1を用いて、予測画像生成部13の構成についてより詳しく説明する。図1は、予測画像生成部13の構成の一例について示す機能ブロック図である。
[Configuration of predicted image generation unit]
Next, the configuration of the predicted image generation unit 13 will be described in more detail with reference to FIG. FIG. 1 is a functional block diagram illustrating an example of the configuration of the predicted image generation unit 13.
 図1に示すように、予測画像生成部13は、局所画像入力部131、輝度予測画像生成部132、チャンネル間予測判定部133、LUT導出部(相関導出手段)134、色差予測画像生成部(予測画像生成手段、処理情報取得手段、予測制御手段)135、および予測画像出力部136を備える。 As shown in FIG. 1, the predicted image generation unit 13 includes a local image input unit 131, a luminance predicted image generation unit 132, an inter-channel prediction determination unit 133, an LUT derivation unit (correlation derivation unit) 134, and a color difference prediction image generation unit ( A prediction image generation unit, a processing information acquisition unit, a prediction control unit) 135, and a prediction image output unit 136.
 局所画像入力部131は、局所復号画像P’から輝度局所復号画像P’と、色差局所復号画像P’とを取得する。局所画像入力部131は、輝度局所復号画像P’を輝度予測画像生成部132に転送するとともに、色差局所復号画像P’をチャンネル間予測判定部133に転送する。 The local image input unit 131 acquires the luminance local decoded image P ′ Y and the color difference local decoded image P ′ C from the local decoded image P ′. The local image input unit 131 transfers the luminance local decoded image P ′ Y to the luminance predicted image generation unit 132 and transfers the color difference local decoded image P ′ C to the inter-channel prediction determination unit 133.
 輝度予測画像生成部132は、輝度局所復号画像P’を参照し、PU情報PUIに基づいて予測を行って、輝度予測画像PredYを生成する。輝度予測画像生成部132は、生成した輝度予測画像PredYを予測画像出力部136に送信する。 Brightness prediction image generation unit 132 refers to the luminance local decoded image P 'Y, when predictions are based on PU information PUI, it generates a brightness prediction image PredY. The predicted brightness image generation unit 132 transmits the generated predicted brightness image PredY to the predicted image output unit 136.
 チャンネル間予測判定部133は、イントラ予測パラメータPPに含まれるチャンネル間予測フラグを参照して、色差のイントラ予測(以下、単に色差予測と称する)が、チャンネル間予測により色差予測画像を生成するチャンネル間予測モードであるか否かを判定する。 The inter-channel prediction determination unit 133 refers to an inter-channel prediction flag included in the intra-prediction parameter PP, and a channel in which intra prediction of color difference (hereinafter simply referred to as color difference prediction) generates a color difference prediction image by inter-channel prediction. It is determined whether or not it is an inter prediction mode.
 判定の結果、チャンネル間予測モードである場合、チャンネル間予測判定部133は、チャンネル間予測モードであることを、LUT導出部134および色差予測画像生成部135のチャンネル間予測部351(後述)に通知する。また、判定の結果、チャンネル間予測モードでない場合、チャンネル間予測判定部133は、色差予測画像生成部135のチャンネル内予測部352(後述)に色差局所復号画像P’を転送する。 As a result of the determination, if it is the inter-channel prediction mode, the inter-channel prediction determination unit 133 informs the LUT derivation unit 134 and the inter-channel prediction unit 351 (described later) of the color difference prediction image generation unit 135 that it is the inter-channel prediction mode. Notice. As a result of the determination, if it is not between channels prediction mode, the prediction determining section 133 among the channels, and transfers the color difference local decoded image P 'C channel prediction unit 352 of the color difference prediction image generation unit 135 (described later).
 LUT導出部134は、対象パーティションごとに、チャンネル間予測を行うためのLUT(Look Up Table)を色差局所復号画像P’に基づいて導出する。LUT導出部が導出するLUTは、例示的には、次のような構造である。すなわち、LUTには、色差局所復号画像P’の画素位置[x,y]における輝度値と対応付けて、上記画素位置[x,y]に対応する、色差局所復号画像P’の画素位置[x,y]における色差値が格納される。 LUT deriving unit 134, for each target partition, it derived based an LUT (Look Up Table) for performing inter-channel prediction chrominance local decoded image P 'Y. The LUT derived by the LUT deriving unit is illustratively structured as follows. That is, the LUT, the pixel position of the color difference local decoded image P 'Y [x Y, y Y] in association with the luminance value in the above pixel position [x Y, y Y] corresponding to the color difference local decoded image P 'pixel positions C [x C, y C] chrominance values in is stored.
 なお、色差局所復号画像P’における画素位置[x,y]と、色差予測画像Predにおける画素位置[x,y]との対応関係については、非特許文献3に記載の対応関係を採用することができる。 The correspondence relationship between the pixel position [x y , y y ] in the color difference local decoded image P ′ Y and the pixel position [x C , y C ] in the color difference prediction image Pred is described in Non-Patent Document 3. Relationships can be adopted.
 また、LUT導出部134は、導出したLUTを色差予測画像生成部135のチャンネル間予測部351(予測画像生成手段;後述)に送信する。LUT導出部134の動作詳細については後述する。 Further, the LUT deriving unit 134 transmits the derived LUT to the inter-channel prediction unit 351 (predicted image generation means; described later) of the color difference predicted image generation unit 135. Details of the operation of the LUT deriving unit 134 will be described later.
 色差予測画像生成部135は、色差画像の予測を行って、色差予測画像PredCを生成する。より詳細には、色差予測画像生成部135は、チャンネル間予測部351およびチャンネル内予測部352を備える構成である。 The color difference predicted image generation unit 135 predicts a color difference image and generates a color difference predicted image PredC. More specifically, the color difference predicted image generation unit 135 includes an inter-channel prediction unit 351 and an intra-channel prediction unit 352.
 チャンネル間予測部351は、色差予測がチャンネル間予測モードであるときにおいて、輝度復号画像Pを参照し、チャンネル間予測により色差画像の予測を行って色差予測画像PredCを生成する。 Inter-channel prediction unit 351, at the time the color difference prediction is a prediction mode among channels, with reference to the luminance decoded image P Y, to generate the color difference prediction image PredC performs prediction of the color difference images by inter-channel prediction.
 また、チャンネル内予測部352は、色差予測がチャンネル間予測モードでないときにおいて、色差局所復号画像P’を参照し、色差画像の予測を行って色差予測画像PredCを生成する。チャンネル内予測部352による色差画像の予測は、例えば、方向予測やDC予測によって行われる。 The channel prediction unit 352, at the time the color difference prediction not prediction mode among channels, with reference to the chrominance local decoded image P 'C, to produce a color difference prediction image PredC performs prediction of the color difference images. The prediction of the color difference image by the intra-channel prediction unit 352 is performed by, for example, direction prediction or DC prediction.
 なお、チャンネル間予測部351の動作詳細については後述する。 Details of the operation of the inter-channel prediction unit 351 will be described later.
 予測画像出力部136は、輝度予測画像生成部132が生成した輝度予測画像PredYおよび色差予測画像生成部135が生成した色差予測画像PredCを予測画像Predとして出力する。 The predicted image output unit 136 outputs the luminance predicted image PredY generated by the luminance predicted image generating unit 132 and the color difference predicted image PredC generated by the color difference predicted image generating unit 135 as the predicted image Pred.
 なお、以下では、色差の処理について説明するときは、色差のUチャンネルについて主に説明する。Uチャンネルの処理と、Vチャンネルの処理とは、概ね同様である。よって、特に断りがない限り、Uチャンネルの処理の説明において、Uの添え字を、Vに読み替えれば、Vチャンネルの処理の説明とすることができる。 In the following, when the color difference processing is described, the color difference U channel will be mainly described. U channel processing and V channel processing are generally the same. Therefore, unless otherwise noted, in the description of U channel processing, if the U suffix is read as V, the V channel processing can be described.
   [輝度復号画像における画素位置と色差予測画像における画素位置との対応関係]
 次に、図6および図7を用いて、輝度復号画像Pにおける画素位置[x,y]と、色差予測画像PredUにおける画素位置[x,y]との対応関係について説明する。図6は、4:2:0形式における輝度および色差の画素位置の対応関係について示す図である。図6の(a)は、輝度復号画像Pの画素位置について示しており、同図の(b)は、予測の対象となる色差予測画像PredUの画素位置について示している。
[Correspondence between pixel position in luminance decoded image and pixel position in color difference prediction image]
Next, the correspondence relationship between the pixel position [x Y , y Y ] in the luminance decoded image P Y and the pixel position [x U , y U ] in the color difference predicted image PredU will be described using FIG. 6 and FIG. 7. . FIG. 6 is a diagram illustrating a correspondence relationship between luminance and color difference pixel positions in the 4: 2: 0 format. (A) of FIG. 6 shows the pixel position of the luminance decoded image P Y, in FIG. (B) shows the pixel position of the color difference prediction image PredU to be predicted.
 輝度復号画像Pおよび色差予測画像PredUの画素位置は、いずれもブロックの左上を原点とする相対座標により表現される。 Pixel positions of the luminance decoded image P Y and the color difference prediction image PredU are both represented by relative coordinates with the origin at the upper left of the block.
 よって、図6の(a)に示す輝度復号画像Pにおいて、画素SP1の画素位置は、[x,y]=[6,2]となる。また、図6の(b)に示す色差局所復号画像PredUにおいて、画素SP1の画素位置は、[x,y]=[3,1]となる。 Therefore, in the luminance decoded image P Y shown in FIG. 6A, the pixel position of the pixel SP Y 1 is [x Y , y Y ] = [6, 2]. In the color difference local decoded image PredU shown in FIG. 6B, the pixel position of the pixel SP U 1 is [x U , y U ] = [3, 1].
 輝度復号画像Pにおける画素位置[x,y]と、色差予測画像PredUにおける画素位置[x,y]との対応関係は、例えば、x=2x,y=2yである。 The correspondence relationship between the pixel position [x Y , y Y ] in the luminance decoded image P Y and the pixel position [x U , y U ] in the color difference predicted image PredU is, for example, x Y = 2x U , y Y = 2y U. It is.
 つまり、この対応関係によると、上述の輝度復号画像Pにおける画素位置[x,y]=[6,2]は、色差局所復号画像PredUの画素位置[x,y]=[3,1]と対応している。 That is, according to this correspondence, the pixel position [x Y , y Y ] = [6, 2] in the luminance decoded image P Y described above is the pixel position [x U , y U ] = [ 3, 1].
 図7を用いて、さらに詳しく、色差画素のサンプル位置と、輝度復号画像Pにおける画素位置との対応関係について説明すると次のとおりである。図7の(a)~(b)は、3つの色差画素のサンプル位置について例示する図である。 With reference to FIG. 7, in more detail, the sample position of the color difference pixel, explaining correspondence between the pixel positions in the luminance decoded picture P Y are as follows. FIGS. 7A to 7B are diagrams illustrating sample positions of three color difference pixels.
 図7の(a)は、既に図6を用いて説明した場合について示している。すなわち、図7の(a)に示すように、色差画素のサンプル位置を、ブロックの左上に設定してもよい。この場合、画素位置の対応関係は、x=2x,y=2yである。 FIG. 7A shows the case already described with reference to FIG. That is, as shown in FIG. 7A, the sample position of the color difference pixel may be set at the upper left of the block. In this case, the correspondence relationship between the pixel positions is x Y = 2x U and y Y = 2y U.
 また、図7の(b)に示すように、色差画素のサンプル位置を、ブロックの中央から左寄りに設定してもよい。この場合、画素位置の対応関係は、x=2x,y=2y+0.5である。 Further, as shown in FIG. 7B, the sample position of the color difference pixel may be set to the left from the center of the block. In this case, the correspondence between the pixel positions, x Y = 2x U, a y Y = 2y U +0.5.
 また、図7の(c)に示すように、色差画素のサンプル位置を、ブロックの中央に設定してもよい。この場合、画素位置の対応関係は、x=2x+0.5,y=2y+0.5である。 Further, as shown in FIG. 7C, the sample position of the color difference pixel may be set at the center of the block. In this case, the correspondence between pixel locations, x Y = 2x U + 0.5 , a y Y = 2y U +0.5.
 そして、ある色差画素の値に対応する輝度値は、図7の(a)~(c)に示した対応関係に従って求めた画素位置の近傍の輝度値をフィルタリングすることにより導出される。 Then, the luminance value corresponding to the value of a certain color difference pixel is derived by filtering the luminance value in the vicinity of the pixel position obtained in accordance with the correspondence relationship shown in (a) to (c) of FIG.
 ここで、画素位置の近傍とは、上記対応関係から求めた画素位置の各座標の値を、切り上げ、および、切り下げしたときに得られる座標のことである。つまり、図7の(b)に示した例では、[x,y]および[x,y+1]である。また、図7の(c)に示した例では、[x,y]、[x+1,y]、[x,y+1]および[x+1,y+1]である。 Here, the vicinity of the pixel position is a coordinate obtained when the value of each coordinate of the pixel position obtained from the correspondence relation is rounded up and down. That is, in the example shown in FIG. 7B, [x Y , y Y ] and [x Y , y Y +1]. Further, in the example shown in FIG. 7 (c), in [x Y, y Y], [x Y + 1, y Y], [x Y, y Y +1] and [x Y + 1, y Y +1] is there.
 なお、図7の(a)に示した例では、xおよびyが整数値となるため、その画素位置[x,y]における輝度値をそのままサンプルとする輝度値として用いてもよい。また、このフィルタリングの例としては、平滑化フィルタが挙げられる。 In the example shown in FIG. 7A, since x Y and y Y are integer values, the luminance value at the pixel position [x Y , y Y ] may be used as a luminance value as a sample as it is. Good. Moreover, a smoothing filter is mentioned as an example of this filtering.
 より高精度に「ある色差画素の値に対応する輝度値を求める」には、このようなサンプル位置のうち、最も精度よく予測ができるものを選べばよい。 In order to obtain “a luminance value corresponding to a certain color difference pixel value” with higher accuracy, it is only necessary to select a sample position that can be predicted with the highest accuracy.
 なお、上述の関係は、Vチャンネルの場合についても同様であるので、ここではその説明を省略する。 Note that the above relationship is the same for the V channel, so the description thereof is omitted here.
   [色差チャンネルの処理順序のパターン]
 次に、図8を用いて、予測元チャンネルが1つである場合における色差チャンネルの処理順序のパターンについて説明する。図8は、予測元チャンネルが1つである場合における色差チャンネルの処理順序のパターンについて示す図である。予測元チャンネルが1つである場合、図8に示すように、輝度(Y)チャンネルを基点として、Uチャンネル、およびVチャンネルを予測することができる。図8に示すように、予測のパターンとしては、次の3つがある。
[Color difference channel processing order pattern]
Next, a color difference channel processing order pattern when there is one prediction source channel will be described with reference to FIG. FIG. 8 is a diagram showing a pattern of the color difference channel processing order when there is one prediction source channel. When there is one prediction source channel, the U channel and the V channel can be predicted using the luminance (Y) channel as a base point, as shown in FIG. As shown in FIG. 8, there are the following three prediction patterns.
 1つ目は、図8において実線で示すパターンである。すなわち、YチャンネルからUチャンネルを予測する一方で、YチャンネルからVチャンネルを予測する。UチャンネルおよびVチャンネルのいずれを先に予測するかは任意に選択することができる。また、UチャンネルおよびVチャンネルの予測処理を並列処理で行ってもよい。 The first is a pattern indicated by a solid line in FIG. That is, while predicting the U channel from the Y channel, the V channel is predicted from the Y channel. Which of the U channel and the V channel is predicted first can be arbitrarily selected. Further, the U channel and V channel prediction processing may be performed in parallel.
 2つ目は、図8において点線で示すパターンである。すなわち、Yチャンネルから、まず、Uチャンネルを予測し、その後、予測したUチャンネルからVチャンネルを予測する。 The second is a pattern indicated by a dotted line in FIG. That is, the U channel is first predicted from the Y channel, and then the V channel is predicted from the predicted U channel.
 3つ目は、2つめの逆で、図8において破線で示すパターンである。すなわち、Yチャンネルから、まず、Vチャンネルを予測し、その後、予測したVチャンネルからUチャンネルを予測する。 The third is the reverse of the second, and is a pattern indicated by a broken line in FIG. That is, the V channel is first predicted from the Y channel, and then the U channel is predicted from the predicted V channel.
 以上のいずれの処理順序を用いるかを示す色差チャンネル処理順序フラグ、および、2番目に予測するチャンネルの予測元チャンネルを指定する第2チャンネル予測元チャンネル指定子が、動画像符号化装置2における符号化処理において符号化されて動画像復号装置1に送信される。そして、動画像復号装置1のチャンネル間予測部351では、色差チャンネル処理順序フラグおよび第2チャンネル予測元チャンネル指定子に従ってチャンネル間予測を行う。 The chrominance channel processing order flag indicating which processing order is used, and the second channel prediction source channel specifier that specifies the prediction source channel of the second prediction channel are the codes in the moving picture encoding device 2. Encoded in the encoding process and transmitted to the video decoding device 1. Then, the inter-channel prediction unit 351 of the video decoding device 1 performs inter-channel prediction according to the color difference channel processing order flag and the second channel prediction source channel specifier.
 なお、予測元チャンネルが2つである場合については、後述する変形例にて説明する。 Note that the case where there are two prediction source channels will be described in a later-described modification.
 また、チャンネル間予測部351は、次の式(1)に従い、Uチャンネルのチャンネル間予測を行って、輝度復号画像Pから色差予測画像PredUを生成する。 Moreover, inter-channel prediction unit 351, in accordance with the following equation (1), by performing the inter-channel prediction of the U channel, and generates the color difference prediction image PredU from the luminance decoded image P Y.
 PredU[x,y]=LUT[RecY[x,y]] … (1)
 ここで、式(1)における各記号の意味について説明すると以下のとおりである。
PredU [x U, y U] = LUT U [RecY [x Y, y Y]] ... (1)
Here, the meaning of each symbol in the formula (1) will be described as follows.
 [x,y]:輝度復号画像Pにおける画素位置(以下、輝度画素位置と称する)
 [x,y]:輝度画素位置に対応する、色差予測画像PredUの画素位置(以下、色差画素位置と称する)
 LUT:UチャンネルのLUTのエントリを返す関数(以下、単にLUTと表記する。)
 また、念のため、Vチャンネルのチャンネル間予測についても説明しておくと次のとおりである。Vチャンネルについても、Uチャンネルと同様、次のような式(2)に従い、チャンネル間予測が行われる。
[X Y , y Y ]: Pixel position in the luminance decoded image P Y (hereinafter referred to as luminance pixel position)
[X U , y U ]: Pixel position of the color difference predicted image PredU corresponding to the luminance pixel position (hereinafter referred to as color difference pixel position)
LUT U : A function that returns an entry of a U channel LUT (hereinafter simply referred to as LUT)
As a precaution, the inter-channel prediction of the V channel is also described as follows. For the V channel, similarly to the U channel, inter-channel prediction is performed according to the following equation (2).
 PredV[x,y]=LUT[RecY[x,y]] … (2)
 式(2)の各記号の意味は、式(1)と同様であるので、その説明を省略する。
PredV [x V, y V] = LUT V [RecY [x Y, y Y]] ... (2)
The meaning of each symbol in the formula (2) is the same as that in the formula (1), and the description thereof is omitted.
  [色差予測画像生成処理の概要]
 次に、図9を用いて、予測画像生成部13における色差予測画像生成処理の概略的な流れについて説明する。図9は、予測画像生成部13における色差予測画像生成処理の概略的な流れについて例示したフローチャートである。
[Outline of color difference prediction image generation processing]
Next, a schematic flow of color difference predicted image generation processing in the predicted image generation unit 13 will be described with reference to FIG. FIG. 9 is a flowchart illustrating a schematic flow of color difference predicted image generation processing in the predicted image generation unit 13.
 色差予測画像生成処理が開始されると、チャンネル間予測判定部133が、チャンネル間予測フラグを参照して、チャンネル間予測モードであるか否かを判定する(S10)。 When the color difference prediction image generation process is started, the inter-channel prediction determination unit 133 refers to the inter-channel prediction flag and determines whether or not the inter-channel prediction mode is set (S10).
 ここで、判定の結果、チャンネル間予測モードでない場合(S10においてNO)、チャンネル内予測部352がチャンネル間予測によらないで色差予測画像PredCを生成し(S11)、処理は終了する。 Here, if the determination result shows that the mode is not the inter-channel prediction mode (NO in S10), the intra-channel prediction unit 352 generates the color difference prediction image PredC without depending on the inter-channel prediction (S11), and the process ends.
 これに対して、判定の結果、チャンネル間予測モードである場合(S10においてYES)、LUT導出部134が、輝度復号画像Pを参照してLUTを導出する(S12)。そして、チャンネル間予測部351が、LUT導出部134が導出したLUTを参照して、チャンネル間予測による色差予測画像PredCを生成する(S13)。以上で、処理は終了する。 In contrast, the result of the determination, if it is inter-channel prediction mode (YES at S10), LUT deriving unit 134 derives the LUT with reference to the luminance decoded image P Y (S12). Then, the inter-channel prediction unit 351 generates a color difference prediction image PredC by inter-channel prediction with reference to the LUT derived by the LUT deriving unit 134 (S13). This is the end of the process.
 なお、上記色差予測画像生成処理では、一例として、当該処理に先立って、輝度復号画像Pが生成されていること想定している。しかしながら、これに限られず、輝度復号画像Pは、ステップ10の後、ステップS13の開始までに生成されていればよい。 In the color difference prediction image generation process, as an example, it is assumed that a luminance decoded image PY is generated prior to the process. However, not limited to this, the luminance decoded image P Y after step 10, only to be generated by the start of step S13.
   [輝度-色差の分布と、チャンネル間予測との関係について]
 ここで、図10、図11、図12を用いて、輝度-色差の分布と、チャンネル間予測との関係について説明する。
[Relationship between luminance-color difference distribution and inter-channel prediction]
Here, the relationship between the luminance-color difference distribution and the inter-channel prediction will be described with reference to FIG. 10, FIG. 11, and FIG.
 図10は、輝度-色差の分布が重ならならず2領域にまたがるような画像Bの例を示す図である。 FIG. 10 is a diagram showing an example of an image B in which the luminance-color difference distributions do not overlap and extend over two regions.
 図10に示す画像Bは、6つの画素領域から構成される。画素領域R1~R3は、輝度値(Y)が低く色差値(U)が高い領域である。また、画素領域R4~R6は、輝度値(Y)が高く色差値(U)が低い領域である。 The image B shown in FIG. 10 is composed of six pixel areas. The pixel regions R1 to R3 are regions where the luminance value (Y) is low and the color difference value (U) is high. The pixel regions R4 to R6 are regions having a high luminance value (Y) and a low color difference value (U).
 また、画素領域R1~R3においては、画素領域R1からR3にかけて、徐々に色差値(U)が高くなっている。一方、画素領域R4~R6においては、画素領域R4からR6にかけて色差値(U)が低くなっている。 In the pixel regions R1 to R3, the color difference value (U) gradually increases from the pixel region R1 to R3. On the other hand, in the pixel regions R4 to R6, the color difference value (U) decreases from the pixel region R4 to R6.
 このような画像Bに含まれる画素の輝度(Y)-色差(U)をプロットしたものを、図11のグラフに示している。 A plot of the luminance (Y) -color difference (U) of the pixels included in such an image B is shown in the graph of FIG.
 画素領域R1~R3に含まれる画素は、図11に示すグラフにおいて、グループGr1にプロットされる。このグループGr1では、サンプルの輝度(Y)が低く、色差値(U)が高い。また、グループGr1では、さらに、輝度(Y)が高いほど、色差値(U)も高くなる傾向が見られる。 The pixels included in the pixel regions R1 to R3 are plotted in the group Gr1 in the graph shown in FIG. In this group Gr1, the luminance (Y) of the sample is low and the color difference value (U) is high. Further, in the group Gr1, there is a tendency that the higher the luminance (Y), the higher the color difference value (U).
 また、画素領域R4~R6に含まれる画素は、図11に示すグラフにおいて、グループGr2にプロットされる。グループGr2では、サンプルの輝度(Y)が高く、色差値(U)が低い。また、グループGr2では、さらに、輝度(Y)が高いほど、色差値(U)も高くなる傾向が見られる。 Further, the pixels included in the pixel regions R4 to R6 are plotted in the group Gr2 in the graph shown in FIG. In the group Gr2, the luminance (Y) of the sample is high and the color difference value (U) is low. Further, in the group Gr2, there is a tendency that the higher the luminance (Y), the higher the color difference value (U).
 このため、グループGr1およびグループGr2のそれぞれにおいて分布の広がる方向と、グループGr1およびグループGr2の位置関係とが、線形近似するのに適していない。 For this reason, the direction in which the distribution spreads in each of the groups Gr1 and Gr2 and the positional relationship between the groups Gr1 and Gr2 are not suitable for linear approximation.
 画像Bは、画像内において輝度や、色差にばらつきがあるので線形近似するのに適していない。このようなばらつきは、複数の物体の境界や、複数色のテクスチャが含まれていたりする画像に多く見られる傾向がある。 Image B is not suitable for linear approximation because there are variations in brightness and color difference in the image. Such variations tend to be often seen in images that include boundaries between multiple objects or textures of multiple colors.
 このような画像の場合において、線形近似によりLUTを作成してしまうと、誤差が大きくなる傾向がある。以下では、上記のサンプルを、非線形的に近似してLUTを導出し、このような誤差を抑制する手法について説明する。 In the case of such an image, if an LUT is created by linear approximation, the error tends to increase. In the following, a method for suppressing such an error by deriving a LUT by nonlinearly approximating the above sample will be described.
  [LUT導出処理の流れ]
 次に、図12を用いて、LUT導出部134によるLUT導出処理の流れについて説明する。図12は、LUT導出部134によるLUT導出処理の流れの一例について示したフローチャートである。
[LUT derivation process flow]
Next, the flow of LUT derivation processing by the LUT derivation unit 134 will be described using FIG. FIG. 12 is a flowchart showing an example of the flow of LUT derivation processing by the LUT derivation unit 134.
 図12に示すように、LUT導出処理が開始されると、まず、LUT導出部134は、LUTを初期化する(S100)。LUTを初期化するとは、LUTを未登録状態にすることである。 As shown in FIG. 12, when the LUT derivation process is started, first, the LUT derivation unit 134 initializes the LUT (S100). To initialize the LUT is to make the LUT unregistered.
 続いて、LUT導出部134は、対象パーティションに隣接する各輝度画素についての登録処理のループLP11に入る(S101)。 Subsequently, the LUT deriving unit 134 enters a loop LP11 of registration processing for each luminance pixel adjacent to the target partition (S101).
 このループLP11では、処理対象となる輝度画素について以下の登録処理が行われる。まず、LUT導出部134は、処理対象となる輝度画素における輝度値=nおよび、輝度画素位置に対応する色差画素位置における色差値=mを取得する(S102)。 In this loop LP11, the following registration process is performed for the luminance pixel to be processed. First, the LUT deriving unit 134 obtains the luminance value = n at the luminance pixel to be processed and the color difference value = m at the color difference pixel position corresponding to the luminance pixel position (S102).
 次に、LUT導出部134は、LUT[n]に色差値が登録されているか否かを判定する(S103)。 Next, the LUT deriving unit 134 determines whether or not a color difference value is registered in the LUT [n] (S103).
 ここで、判定の結果、LUT[n]に色差値が登録されていなかった場合(S103においてNO)、LUT導出部134は、ステップS102において取得した色差値mをそのまま、LUT[n]に登録して(S105)、ループLP11の先頭に戻る(S106)。 If the color difference value is not registered in the LUT [n] as a result of the determination (NO in S103), the LUT deriving unit 134 registers the color difference value m acquired in Step S102 as it is in the LUT [n]. The process returns to the beginning of the loop LP11 (S106).
 一方、判定の結果、LUT[n]が既に登録されていた場合、LUT導出部134は、(m+LUT[n]+1)/2を計算することより、取得した色差値mと登録済みの色差値との平均値を算出する。LUT導出部134は、このようにして算出した平均値を色差値mに代入する(S104)。続いて、LUT導出部134は、ステップS104において平均値を代入した色差値mを、LUT[n]に登録して(S105)、ループLP11の先頭に戻る(S106)。 On the other hand, if the result of determination is that LUT [n] has already been registered, the LUT derivation unit 134 calculates (m + LUT [n] +1) / 2, thereby obtaining the acquired color difference value m and the registered color difference value. And the average value is calculated. The LUT deriving unit 134 substitutes the average value thus calculated for the color difference value m (S104). Subsequently, the LUT deriving unit 134 registers the color difference value m into which the average value is substituted in step S104 in the LUT [n] (S105), and returns to the top of the loop LP11 (S106).
 各輝度画素についての登録処理が終了すると、ループLP11は終了する。 When the registration process for each luminance pixel ends, the loop LP11 ends.
 ループLP11が終了した後、未登録LUTエントリの補間処理のループLP12に入る(S107)。 After the loop LP11 is completed, the loop LP12 for interpolation processing of unregistered LUT entries is entered (S107).
 このループLP12では、n=0からn=255までの未登録エントリについて以下の補間処理が行われる。すなわち、ループLP12では、サンプルのなかったLUTのエントリを補間により埋め合わせる。まず、LUT導出部134は、LUT[n]が未登録であるか否かを判定する(S108)。 In this loop LP12, the following interpolation processing is performed for unregistered entries from n = 0 to n = 255. That is, in the loop LP12, the entry of the LUT that has no sample is made up by interpolation. First, the LUT deriving unit 134 determines whether or not LUT [n] is unregistered (S108).
 ここで、LUT[n]が未登録でない、すなわちLUT[n]が登録済みである場合(S108においてNO)、さらに補間処理は継続し、ループLP12の先頭に戻る(S111)。 Here, when LUT [n] is not registered, that is, when LUT [n] is registered (NO in S108), the interpolation process continues and returns to the top of the loop LP12 (S111).
 一方、LUT[n]が未登録である場合(S108においてYES)LUT導出部134は、図13に示すように、nの前後にある直近の登録済みエントリを探索する(S109)。具体的には、次のようにして、LUT導出部134は、nの前後にある直近の登録済みエントリを探索する。 On the other hand, when LUT [n] is not registered (YES in S108), the LUT deriving unit 134 searches for the latest registered entries before and after n as shown in FIG. 13 (S109). Specifically, the LUT deriving unit 134 searches for the latest registered entry before and after n as follows.
 まず、LUT導出部134は、nを基点として、nの前方向、すなわち、nよりも小さいnLについて登録済みエントリを探索する。つまり、図13を参照して説明すると、nの前方向にあるサンプル点Smpl1を探索している。 First, the LUT deriving unit 134 searches for a registered entry for n forward, that is, nL smaller than n, with n as a base point. That is, with reference to FIG. 13, a search is made for a sample point Smpl1 in the n forward direction.
 また、LUT導出部134は、nを基点として、nの後方向、すなわち、nよりも大きいnRについて登録済みであるエントリを探索する。つまり、図13を参照して説明すると、nの後方向にあるサンプル点Smpl2を探索している。 Also, the LUT deriving unit 134 searches for an entry registered for n backward, that is, nR larger than n, with n as a base point. That is, with reference to FIG. 13, a search is made for a sample point Smpl2 in the backward direction of n.
 次に、LUT導出部134は、LUT[nL]と、LUT[nR]とを直線補間した値を、LUT[n]に登録する(S110)。 Next, the LUT derivation unit 134 registers a value obtained by linear interpolation between the LUT [nL] and the LUT [nR] in the LUT [n] (S110).
 ここで、さらに図13を参照して説明すると、ステップS110では、nよりも前方にある直近のサンプル点Smpl1と、nよりも後方にあるサンプル点Smpl2とを直線L1で結ぶ補間処理を行っている。直線L1のnにおける値が、LUT[n]に登録される。 Now, with reference to FIG. 13, in step S110, an interpolation process is performed to connect the nearest sample point Smpl1 ahead of n and the sample point Smpl2 behind n by a straight line L1. Yes. The value at n of the straight line L1 is registered in LUT [n].
 なお、ステップS109の探索の結果、nLおよびnRのいずれか一方しか検知されなかった場合、検知された方の登録済みエントリの値が、LUT[n]に登録される。 If only one of nL and nR is detected as a result of the search in step S109, the value of the registered entry of the detected one is registered in LUT [n].
 さらに補間処理は継続し、ループLP12の先頭に戻る(S111)。 Further, the interpolation process is continued and the process returns to the top of the loop LP12 (S111).
 ループLP12において、n=0からn=255までのすべてのエントリについて補間処理が完了することにより、ループLP12は終了する。そしてLUT導出処理は終了する。 In the loop LP12, when the interpolation processing is completed for all entries from n = 0 to n = 255, the loop LP12 ends. Then, the LUT derivation process ends.
 なお、ステップS104では、登録処理においてエントリが既に登録済みであった場合、取得した色差値mと登録済みの色差値との平均値を、新たなエントリとして登録していた。これは、次の理由による。 In step S104, if an entry has already been registered in the registration process, the average value of the acquired color difference value m and the registered color difference value is registered as a new entry. This is due to the following reason.
 すなわち、ステップS104において、取得した色差値mで登録済みエントリの値を上書きしてしまうと、取得した色差値mが統計的なノイズである場合に、そのノイズをそのままエントリに登録してしまう結果となる。これに対して、取得した色差値mと登録済みの色差値との平均値を取得することで、このようなノイズを緩和することができる。なお、ここで取得している平均値は、登録順の重み付き平均である。また、これに限られず、ステップS104では、任意に重み付けを設定した加重平均を取得してもよい。 That is, in step S104, if the acquired color difference value m is overwritten with the registered entry value, if the acquired color difference value m is statistical noise, the noise is directly registered in the entry. It becomes. On the other hand, such noise can be reduced by acquiring an average value of the acquired color difference value m and the registered color difference value. The average value acquired here is a weighted average in the order of registration. Further, the present invention is not limited to this, and in step S104, a weighted average in which weights are arbitrarily set may be acquired.
 また、LUT導出後、テーブル全体に移動平均処理を行ってもよい。テーブル全体に移動平均処理を行うことで、急激な色差の変化を抑制できる。 Also, after the LUT is derived, the moving average process may be performed on the entire table. By performing the moving average process on the entire table, a rapid change in color difference can be suppressed.
 また、ここで、図10に示した画像Bに対して以上の処理を施すことにより導出されたLUTの値は、図14に示すとおりである。すなわち、画像Bから導出したLUTの値、すなわちエントリをグラフ化したものを、図14に示す。 Here, the LUT value derived by performing the above processing on the image B shown in FIG. 10 is as shown in FIG. That is, FIG. 14 shows a graph of the LUT values derived from the image B, that is, entries.
 図14に示すグラフL11は、画像Bから導出したLUTのエントリの値を示す。グラフL11は、各サンプル点を通る線である。グラフL11において、各サンプル点の間のエントリは、線形補間により作成されている。例えば、グループGr1の最も右側にあるサンプルとグループGr2の最も左側にあるサンプルの間は、直線で接続されている。 A graph L11 shown in FIG. 14 shows LUT entry values derived from the image B. The graph L11 is a line passing through each sample point. In the graph L11, the entry between each sample point is created by linear interpolation. For example, a straight line is connected between the leftmost sample of the group Gr1 and the leftmost sample of the group Gr2.
 なお、複数のサンプル点が存在するエントリについては、nの値が重複するサンプル点について平均値を取り、この値をチャンネル間予測のための実際のサンプル点としている。 Note that, for an entry having a plurality of sample points, an average value is taken for sample points where n values overlap, and this value is used as an actual sample point for inter-channel prediction.
 なお、以上において、線形補間により、未登録エントリ、つまり各サンプル点の間のエントリを作成したが、これに限られない。例えば、3次補間により当該エントリを作成してもよい。これにより、テーブルの予測精度の向上を図ることができる。 In the above, an unregistered entry, that is, an entry between each sample point, is created by linear interpolation. However, the present invention is not limited to this. For example, the entry may be created by cubic interpolation. Thereby, the prediction accuracy of the table can be improved.
   [チャンネル間予測による色差予測画像生成処理の流れについて]
 次に、図15を用いて、チャンネル間予測部351におけるチャンネル間予測による色差予測画像生成処理の流れについて説明する。図15は、チャンネル間予測部351におけるチャンネル間予測による色差予測画像生成処理の流れの一例について示したフローチャートである。
[Flow of color difference prediction image generation processing by inter-channel prediction]
Next, a flow of color difference prediction image generation processing by inter-channel prediction in the inter-channel prediction unit 351 will be described with reference to FIG. FIG. 15 is a flowchart illustrating an example of a flow of color difference prediction image generation processing by inter-channel prediction in the inter-channel prediction unit 351.
 チャンネル間予測処理が開始されると、チャンネル間予測部351が、イントラ予測パラメータPPに従い、色差Aおよび色差Aの予測元チャンネルと、色差Bおよび色差Bの予測元チャンネルとを設定する(S120)。 When the inter-channel prediction process is started, the inter-channel prediction unit 351 sets the color difference A and color difference A prediction source channels and the color difference B and color difference B prediction source channels according to the intra prediction parameter PP (S120). .
 チャンネル間予測部351が設定する内容は、具体的には、次のとおりである。すなわち、チャンネル間予測部351は、色差チャンネル処理順序フラグに従って、色差Aおよび色差Bを設定する。以下では、色差Bは、色差Aの後に処理される。チャンネル間予測部351は、例えば、色差Aを“Uチャンネル”と設定し、色差Bを“Vチャンネル”と設定してもよい。 The contents set by the inter-channel prediction unit 351 are specifically as follows. That is, the inter-channel prediction unit 351 sets the color difference A and the color difference B according to the color difference channel processing order flag. In the following, the color difference B is processed after the color difference A. For example, the inter-channel prediction unit 351 may set the color difference A as “U channel” and the color difference B as “V channel”.
 また、チャンネル間予測部351は、色差Aの予測元チャンネルに、“Yチャンネル”を設定する。 Also, the inter-channel prediction unit 351 sets “Y channel” as the color difference A prediction source channel.
 そして、チャンネル間予測部351は、第2チャンネル予測元チャンネル指定子に従って、色差Bの予測元チャンネルを設定する。チャンネル間予測部351は、例えば、色差AのUチャンネルの予測元チャンネルを、“Yチャンネル”と設定し、色差BのVチャンネルの予測元チャンネルを“YチャンネルおよびUチャンネル”と設定してもよい。 Then, the inter-channel prediction unit 351 sets a color difference B prediction source channel according to the second channel prediction source channel specifier. For example, the inter-channel prediction unit 351 may set the prediction source channel of the U channel with the color difference A as “Y channel” and set the prediction source channel of the V channel with the color difference B as “Y channel and U channel”. Good.
 次に、チャンネル間予測部351は、色差Aの予測モードを選択し(S121)、ステップS120における設定に従って、色差Aについてチャンネル間予測により色差予測画像を生成する(S122)。 Next, the inter-channel prediction unit 351 selects a color difference A prediction mode (S121), and generates a color difference prediction image by inter-channel prediction for the color difference A according to the setting in step S120 (S122).
 次に、チャンネル間予測部351は、色差Bの予測モードを選択し(S123)、ステップS120における設定に従って、色差Bについてチャンネル間予測により色差予測画像を生成する(S124)。以上により処理は終了する。 Next, the inter-channel prediction unit 351 selects a color difference B prediction mode (S123), and generates a color difference prediction image by inter-channel prediction for the color difference B according to the setting in step S120 (S124). The process ends as described above.
  (作用・効果)
 以上のように、動画像復号装置1は、輝度(Y)チャンネルおよび色差(U,V)チャンネルそれぞれについての予測画像Predを生成して、生成した上記予測画像Predに予測残差Dを加算することにより符号化された画像データを復号するものであり、対象パーティションについて、輝度(Y)チャンネルを復号して輝度復号画像Pを生成する予測画像生成部13および加算器14と、対象パーティションの周辺に位置する局所復号画像P’を参照して、復号済みの輝度(Y)チャンネルと復号対象の色差(U,V)チャンネルとの間の非線形的な相関をLUTとして導出するLUT導出部134と、対象パーティションについて、上記LUTに従って、輝度復号画像Pから、色差予測画像PredCを生成するチャンネル間予測部351と、を備える。
(Action / Effect)
As described above, the video decoding device 1 generates the prediction image Pred for each of the luminance (Y) channel and the color difference (U, V) channel, and adds the prediction residual D to the generated prediction image Pred. The prediction image generation unit 13 and the adder 14 that decode the luminance (Y) channel and generate the luminance decoded image PY for the target partition, and the target partition An LUT deriving unit 134 for deriving a non-linear correlation between a decoded luminance (Y) channel and a color difference (U, V) channel to be decoded as an LUT with reference to a local decoded image P ′ located in the vicinity. If, for the target partition, according to the LUT, inter-channel pre from luminance decoded image P Y, to generate the color difference prediction image PredC It includes a section 351, a.
 上記構成によれば、局所復号画像P’に含まれる各画素の輝度成分にばらつきがあり、線形的な相関に従って、チャンネル間の予測を行うと、予測の精度が低下してしまうような場合であっても、より高い予測精度が得られる可能性を向上させることができるという効果を奏する。 According to the above configuration, the luminance component of each pixel included in the locally decoded image P ′ varies, and when prediction between channels is performed according to a linear correlation, the prediction accuracy is reduced. Even if it exists, there exists an effect that possibility that a higher prediction precision can be improved can be improved.
  (変形例)
 以下において、動画像復号装置1の好ましい変形例について説明する。
(Modification)
Below, the preferable modification of the moving image decoding apparatus 1 is demonstrated.
   [色差チャンネルの処理順序のパターンの変形例]
 図16および図17は、2つめに予測を行うチャンネルの予測元チャンネルが2つである場合における色差チャンネルの処理順序のパターンについて示す図である。
[Modification of pattern of color difference channel processing order]
FIG. 16 and FIG. 17 are diagrams showing the pattern of the color difference channel processing order when there are two prediction source channels for the second channel to be predicted.
 図16に示すように、(1)Yチャンネルを基点として、Uチャンネルを予測し、その後、(2)YチャンネルおよびUチャンネルの組み合わせに基づいて、Vチャンネルを予測してもよい。また、図17に示すように、(1)Yチャンネルを基点として、Vチャンネルを予測し、その後、(2)YチャンネルおよびVチャンネルの組み合わせに基づいて、Uチャンネルを予測してもよい。 16, (1) the U channel may be predicted using the Y channel as a base point, and then (2) the V channel may be predicted based on a combination of the Y channel and the U channel. Also, as shown in FIG. 17, (1) the V channel may be predicted using the Y channel as a base point, and then (2) the U channel may be predicted based on a combination of the Y channel and the V channel.
 このように、2つめに予測を行うチャンネルの予測元チャンネルを2つにするには、例えば、LUTを2次元に拡張すればよい。LUTを2次元に拡張する場合においても、1次元のLUTの導出と同様の手法を採用することができるし、また、2次元のLUTを作成するときに用いられる公知の技術を採用することも可能である。 In this way, in order to make the prediction source channel of the second channel to be predicted two, for example, the LUT may be extended to two dimensions. Even when the LUT is extended to two dimensions, the same technique as the one-dimensional LUT derivation can be adopted, and a known technique used when creating a two-dimensional LUT can be adopted. Is possible.
 この場合、LUT導出部134は、例えば、図16に示す処理順序の場合には、次のようにしてLUTを導出する。すなわち、まず、サンプル点の輝度値Yおよび色差値Uについて1次のテーブルを、上述のとおり導出する。続いて、当該サンプル点の輝度Yおよび色差値Uに対応付けて、サンプル点の色差値Vを対応付けてテーブルに登録する。図17に示す処理順序の場合も同様である。 In this case, for example, in the case of the processing order shown in FIG. 16, the LUT deriving unit 134 derives the LUT as follows. That is, first, a primary table is derived for the luminance value Y and the color difference value U of the sample points as described above. Subsequently, the color difference value V of the sample point is registered in the table in association with the luminance Y and the color difference value U of the sample point. The same applies to the processing order shown in FIG.
 このように、LUTを2次元以上に拡張する構成とすることも可能である。LUTを2次元以上に拡張する構成の別の例としては、複数の画素における輝度値Yを用いる構成などが挙げられる。より具体的には、LUTのエントリにおいて、隣接する輝度値の組み合わせで色差値Uまたは色差値Vをルックアップできるようになっていてもよい。 In this way, the LUT can be configured to extend to more than two dimensions. Another example of a configuration that expands the LUT to two or more dimensions includes a configuration that uses luminance values Y in a plurality of pixels. More specifically, in the LUT entry, the color difference value U or the color difference value V may be looked up by a combination of adjacent luminance values.
 上述のような2次元以上のLUTを用いることで、複数のチャンネルから、あるいは同一チャンネルの複数の画素値から色差を予測することができるので、予測画像の精度の向上を図ることができる。 Since the color difference can be predicted from a plurality of channels or from a plurality of pixel values of the same channel by using a two-dimensional or more LUT as described above, the accuracy of the predicted image can be improved.
   [イントラ予測パラメータのデータ構成の変形例]
 以上では、イントラ予測パラメータPPは、例示的には、チャンネル間予測フラグ、色差チャンネル処理順序フラグおよび第2チャンネル予測元チャンネル指定子を含む構成とした。しかしながら、イントラ予測パラメータPPは、次のように、処理順序および予測元チャンネルの組み合わせを示すチャンネル間予測インデックスを含む構成としてもよい。
[Modified example of data structure of intra prediction parameter]
In the above, the intra prediction parameter PP is illustratively configured to include the inter-channel prediction flag, the color difference channel processing order flag, and the second channel prediction source channel specifier. However, the intra prediction parameter PP may be configured to include an inter-channel prediction index indicating a combination of processing order and prediction source channels as follows.
 すなわち、処理順序および予測元チャンネルの組み合わせに次のようにインデックスindex=0~5を割り当てる。 That is, the index index = 0 to 5 is assigned to the combination of the processing order and the prediction source channel as follows.
 index=0: 順序 U→V、Vの予測元チャンネル=Y
 index=1: 順序 U→V、Vの予測元チャンネル=U
 index=2: 順序 U→V、Vの予測元チャンネル=Y,U
 index=3: 順序 V→U、Vの予測元チャンネル=Y
 index=4: 順序 V→U、Vの予測元チャンネル=V
 index=5: 順序 V→U、Vの予測元チャンネル=Y,V
 上述のチャンネル間予測インデックスが、動画像符号化装置2における符号化処理において、符号化されて動画像復号装置1に送信される。そして、動画像復号装置1のチャンネル間予測部351では、動画像符号化装置2から送信されるチャンネル間予測インデックスに従ってチャンネル間予測を行う。
index = 0: Order U → V, V prediction source channel = Y
index = 1: Order U → V, V prediction source channel = U
index = 2: Order U → V, V prediction source channel = Y, U
index = 3: Order V → U, V prediction source channel = Y
index = 4: Order V → U, V prediction source channel = V
index = 5: Order V → U, V prediction source channel = Y, V
The above-mentioned inter-channel prediction index is encoded and transmitted to the video decoding device 1 in the encoding process in the video encoding device 2. Then, the inter-channel prediction unit 351 of the video decoding device 1 performs inter-channel prediction according to the inter-channel prediction index transmitted from the video encoding device 2.
 なお、以上では、処理順序および予測元チャンネルの組み合わせを示すイントラ予測パラメータPPにチャンネル間予測インデックス等を含める構成としたが、これに限られない。LUTの導出の単位に応じて種々の変更が可能である。例えば、処理順序および予測元チャンネルの組み合わせを示す情報を、スライスヘッダ以外の処理単位のヘッダに格納して、当該処理単位に応じて処理順序等の変更を行ってもよい。例えば、シーケンスヘッダ(SPS:Sequence Parameter Set)や、ピクチャヘッダ(PPS:Picture Parameter Set)に処理順序および予測元チャンネルの組み合わせを示す情報を格納してもよい。 In the above, the intra prediction parameter PP indicating the combination of the processing order and the prediction source channel is included in the intra prediction parameter PP, but the present invention is not limited to this. Various modifications are possible depending on the unit of LUT derivation. For example, information indicating the combination of the processing order and the prediction source channel may be stored in the header of the processing unit other than the slice header, and the processing order or the like may be changed according to the processing unit. For example, information indicating a combination of processing order and prediction source channel may be stored in a sequence header (SPS: Sequence Parameter Set) or a picture header (PPS: Picture Parameter Set).
 また、スライスよりも小さな処理単位、例えば、LCU単位で処理順序等の変更を行ってもよい。このように、輝度と色差との相関が存在する範囲で、LUTの導出の単位を変更することができる。 Also, the processing order or the like may be changed in units of processing smaller than slices, for example, in units of LCUs. In this way, the LUT derivation unit can be changed within a range in which there is a correlation between luminance and color difference.
 また、処理順序および予測元チャンネルの組み合わせを示す情報は、それぞれ別の処理単位で符号化されていてもよい。例えば、処理順序を示す情報は、LCU単位で符号化されており、予測元チャンネルの組み合わせを示す情報は、PU単位で符号化されていてもよい。 Also, the information indicating the combination of the processing order and the prediction source channel may be encoded in different processing units. For example, the information indicating the processing order may be encoded in LCU units, and the information indicating the combination of prediction source channels may be encoded in PU units.
   [LUT構造の変形例]
 以上では、LUTのエントリの数をn=0~255の256個としていた。これは、画素値を8bitで表現する場合の例である。しかしながら、これに限られず、エントリ数を2個より小さい数としてもよい。例えば、LUTの精度がそれほど要求されない場合おいて、次のようにしてエントリ数をn=0~127の128個に削減することができる。
[Modification of LUT structure]
In the above, the number of entries in the LUT is 256 (n = 0 to 255). This is an example when the pixel value is expressed by 8 bits. However, not limited to this, the number of entries may be two to eight smaller number. For example, when the accuracy of the LUT is not so required, the number of entries can be reduced to 128 (n = 0 to 127) as follows.
 すなわち、次のようにしてPredU[x,y]を算出する。 That is, PredU [x U , y U ] is calculated as follows.
 RecY[x,y]の値が偶数の場合:
  PredU[x,y]=LUT[RecY[x,y]/2]
 RecY[x,y]の値が奇数の場合:
  PredU[x,y]=
   (LUT[RecY[x,y]/2]+LUT[RecY[x,y]/2+1])/2
 これにより、LUTの精度をそれほど低下させることなく、LUT作成において使用する記憶領域を削減することができる。
When the value of RecY [x Y , y Y ] is an even number:
PredU [x U , y U ] = LUT [RecY [x Y , y Y ] / 2]
When the value of RecY [x Y , y Y ] is an odd number:
PredU [x U , y U ] =
(LUT [RecY [x Y, y Y] / 2] + LUT [RecY [x Y, y Y] / 2 + 1]) / 2
As a result, the storage area used in LUT creation can be reduced without significantly reducing the accuracy of the LUT.
 なお、PredV[x,y]についても同様である。 The same applies to PredV [x V , y V ].
  [LUT導出処理の流れの変形例]
 次に、図18を用いて、LUT導出部134によるLUT導出処理の流れの変形例について説明する。図12に示したフローチャートでは、登録順に重みつき平均をする例について示したが、以下では、図18を用いて、重み付けをすることなく全ての画素を均一に扱い平均値を取得する変形例について説明する。
[Modified example of flow of LUT derivation process]
Next, a modified example of the flow of the LUT derivation process by the LUT derivation unit 134 will be described with reference to FIG. In the flowchart shown in FIG. 12, an example in which weighted average is performed in the order of registration is shown. However, in the following, a modification example in which all pixels are treated uniformly without weighting and an average value is obtained will be described. explain.
 図18は、LUT導出部134によるLUT導出処理の流れの一変形例について示したフローチャートである。 FIG. 18 is a flowchart showing a modified example of the flow of the LUT derivation process by the LUT derivation unit 134.
 図18示すように、LUT導出処理が開始されると、まず、LUT導出部134は、LUTを初期化する(S130)。 As shown in FIG. 18, when the LUT derivation process is started, first, the LUT derivation unit 134 initializes the LUT (S130).
 ステップS130でのLUTの初期化においては、LUT[n]を未登録状態にするとともに、輝度値nにおける色差値の累積加算値sum[n]=0、輝度値nにおける色差値のサンプル数count[n]=0(LUT,sum、countについてn=0…255)を実行する。この初期化は、UチャンネルおよびVチャンネルの2チャンネル分行われる。 In the initialization of the LUT in step S130, the LUT [n] is set to an unregistered state, the cumulative addition value sum [n] = 0 of the color difference value at the luminance value n, and the number of samples of the color difference value at the luminance value n are counted. [N] = 0 (n = 0 to 255 for LUT, sum, count) is executed. This initialization is performed for two channels of the U channel and the V channel.
 続いて、LUT導出部134は、対象パーティションに隣接する各輝度画素についての登録処理のループLP11Aに入る(S131)。 Subsequently, the LUT deriving unit 134 enters a registration processing loop LP11A for each luminance pixel adjacent to the target partition (S131).
 このループLP11Aでは、処理対象となる輝度画素について以下の登録処理が行われる。まず、LUT導出部134は、処理対象となる輝度画素における輝度値=nおよび、当該輝度画素の画素位置に対応するサンプル位置における色差値=mを取得する(S132)。 In this loop LP11A, the following registration process is performed for the luminance pixel to be processed. First, the LUT deriving unit 134 obtains the luminance value = n at the luminance pixel to be processed and the color difference value = m at the sample position corresponding to the pixel position of the luminance pixel (S132).
 次に、LUT導出部134は、sum[n]に、sum[n]+mを代入するとともに、count[n]に、count[n]+1を代入する(S133)。 Next, the LUT deriving unit 134 substitutes sum [n] + m for sum [n] and substitutes count [n] +1 for count [n] (S133).
 そして、ループLP11の先頭に戻る(S134)。各輝度画素についての登録処理が終了すると、ループLP11Aは終了する。 Then, the process returns to the top of the loop LP11 (S134). When the registration process for each luminance pixel ends, the loop LP11A ends.
 ループLP11Aが終了した後、未登録LUTエントリの補間処理のループLP12Aに入る(S135)。 After the loop LP11A ends, the loop LP12A for interpolation processing of unregistered LUT entries is entered (S135).
 このループLP12Aでは、n=0からn=255までのエントリを処理対象となる輝度画素について以下の補間処理が行われる。まず、LUT導出部134は、count[n]が0より大きいか否か判定する(S136)。 In this loop LP12A, the following interpolation processing is performed for the luminance pixels to be processed with entries from n = 0 to n = 255. First, the LUT deriving unit 134 determines whether count [n] is greater than 0 (S136).
 ここで、count[n]が0以下(つまり、count[n]=0)の場合、LUT導出部134は、ループLP12Aの先頭に戻って(S138)、さらに次のエントリについて処理を継続する。 Here, if count [n] is 0 or less (that is, count [n] = 0), the LUT deriving unit 134 returns to the top of the loop LP12A (S138), and continues processing for the next entry.
 一方、count[n]が0より大きい場合、LUT導出部134は、LUT[n]に、sum[n]/count[n]を代入する(S137)。すなわち、ステップS137では、輝度値nにおける色差値の相加平均を取得し、LUT[n]に代入している。 On the other hand, when count [n] is larger than 0, the LUT deriving unit 134 substitutes sum [n] / count [n] into LUT [n] (S137). That is, in step S137, the arithmetic average of the color difference values at the luminance value n is acquired and substituted into LUT [n].
 その後、ループLP12Aの先頭に戻って(S138)、さらに次のエントリについて処理を継続する。 Thereafter, the process returns to the top of the loop LP12A (S138), and the process is continued for the next entry.
 ループLP12Aにおいて、n=0からn=255までのすべてのエントリについて補間処理が完了することにより、ループLP12Aは終了する。そしてLUT導出処理は終了する。 In the loop LP12A, when the interpolation processing is completed for all entries from n = 0 to n = 255, the loop LP12A ends. Then, the LUT derivation process ends.
 次に、LUT導出部134は、未登録エントリの補間処理を実行する(S139)。ステップS139の処理は、図12に示した、ステップS107~S111(ループLP13)の処理と同様であるので、ここでは、その説明を省略する。その後、LUT導出処理は終了する。 Next, the LUT deriving unit 134 performs interpolation processing for unregistered entries (S139). Since the process in step S139 is the same as the process in steps S107 to S111 (loop LP13) shown in FIG. 12, the description thereof is omitted here. Thereafter, the LUT derivation process ends.
   [LUT導出のために参照する画素の変形例]
 以上では、対象パーティションに隣接する画素を参照する構成について説明したが、これに限られない。例えば、隣接する2列の画素を参照する構成としてもよい。また、例えば、対象パーティションの上側に隣接するパーティションや、対象パーティションの左側に隣接するパーティションを参照する構成としてもよい。このように、輝度と色差との間で相関があるような領域の画素を適宜参照することができる。
[Modifications of Pixels Referenced for LUT Derivation]
In the above, the configuration in which the pixels adjacent to the target partition are referred to, but the configuration is not limited thereto. For example, a configuration in which two adjacent columns of pixels are referred to may be used. Further, for example, a configuration may be adopted in which a partition adjacent to the upper side of the target partition or a partition adjacent to the left side of the target partition is referred to. In this way, it is possible to appropriately refer to pixels in a region where there is a correlation between luminance and color difference.
   [LUT未登録エントリの補間処理の変形例]
 次に、図19を用いて、LUT未登録エントリの補間処理の変形例について説明する。図19は、導出されたLUTの一例を示す図である。
[Modified example of interpolation processing of LUT unregistered entry]
Next, a modified example of the interpolation process for the LUT unregistered entry will be described with reference to FIG. FIG. 19 is a diagram illustrating an example of the derived LUT.
 以下では、参照時にのみ補間処理を行う変形例について説明する。すなわち、図19に示すように、LUTには、参照されるまでは、サンプルの組しか保持しないようにしてもかまわない。 Hereinafter, a modified example in which interpolation processing is performed only at the time of reference will be described. That is, as shown in FIG. 19, the LUT may hold only a set of samples until it is referred to.
 図19に示す例では、輝度Y:40と、色差値(U):160との組み合わせをはじめとして、16組のサンプルの値のみがエントリとして登録されている(同図では、はじめの4組のみを示している)。 In the example shown in FIG. 19, only 16 sets of sample values including the combination of luminance Y: 40 and color difference value (U): 160 are registered as entries (in the figure, the first 4 sets). Only shows).
 つまり、本変形例では、LUT導出部134は、LUT導出処理時には、16組のサンプルについてのエントリ以外の未登録エントリを補間しない。 That is, in this modification, the LUT deriving unit 134 does not interpolate unregistered entries other than the entries for 16 sets of samples during the LUT deriving process.
 一方、色差予測画像生成処理において、LUT導出部134は、未登録エントリが参照された場合、参照された未登録エントリを導出する。 On the other hand, in the color difference prediction image generation process, when the unregistered entry is referred to, the LUT deriving unit 134 derives the referenced unregistered entry.
 未登録エントリが参照された場合、LUT導出部134は、一例として、次の式(3)に従って、参照された未登録エントリn(nL<n<nR)を線形補間により導出する。 When an unregistered entry is referred to, the LUT deriving unit 134 derives the referenced unregistered entry n (nL <n <nR) by linear interpolation according to the following equation (3) as an example.
 LUT[n] = (LUT[nL]×(n-nL)+LUT[nR]×(nR-n)/(nR-nL)) … 式(3)
 ここで、nLおよびnRは、図13を用いて説明したように、それぞれnの前後にある直近の登録済みエントリの番号である。
LUT [n] = (LUT [nL] × (n−nL) + LUT [nR] × (nR−n) / (nR−nL)) (3)
Here, nL and nR are the numbers of the most recently registered entries before and after n, respectively, as described with reference to FIG.
 以下に、式(3)の適用について、図19に示すLUTについて具体的に説明する。LUT[40]=160、LUT[44]=180が登録されているところ、LUT導出部134は、次のようにして参照された未登録エントリn=43を導出する。 Hereinafter, the application of Expression (3) will be described in detail for the LUT shown in FIG. When LUT [40] = 160 and LUT [44] = 180 are registered, the LUT deriving unit 134 derives an unregistered entry n = 43 referred to as follows.
 すなわち、式(3)より、
  LUT[n] = (LUT[40]×3+LUT[44]×1)/4
である。
That is, from equation (3),
LUT [n] = (LUT [40] × 3 + LUT [44] × 1) / 4
It is.
 本変形例によれば、未登録エントリが参照された場合には、補間処理のための計算が発生するものの、その一方で未登録エントリが参照されないままであった場合、LUT導出の補間処理を行わなくても済む。また、メモリ領域についても、登録済みエントリの分だけ用意しておく構成であってもよい。つまり、このような構成を採用した場合、LUTのテーブル作成において、サンプル数に比例するメモリ領域しか消費しない。 According to this modified example, when an unregistered entry is referred to, calculation for interpolation processing occurs. On the other hand, when an unregistered entry remains unreferenced, interpolation processing for LUT derivation is performed. You do n’t have to. Further, the memory area may be prepared for the number of registered entries. That is, when such a configuration is adopted, only a memory area proportional to the number of samples is consumed in creating the LUT table.
 なお、本変形例では、線形補間による補間処理について説明したが、3次補間による補間処理を採用してもかまわない。 In this modification, the interpolation processing by linear interpolation has been described, but interpolation processing by cubic interpolation may be adopted.
   [LUTを用いない変形例]
 次に、図20を用いて、LUTの代わりに、関数を用いる例について説明する。すなわち、符号化対象ブロックごとに、サンプルの組を近似する関数U=f(Y)を導出し、fを特定するパラメータだけをメモリに保持する。
[Variation without LUT]
Next, an example in which a function is used instead of the LUT will be described with reference to FIG. That is, for each encoding target block, a function U = f U (Y) that approximates a set of samples is derived, and only a parameter that specifies f U is stored in the memory.
 色差予測画像PredUは、f(Y)を用いて、すなわち、下記式(4)に従い、
 PredU = f[RecY[x,y]] … (4)
と導出される。図20にfの例について示している。図20に示すように、fは、各サンプル点を通過する曲線として導出されていてもよい。なお、fは、隣り合うサンプルの間において、個別に導出されてもよい。なお、別の観点からいえば、サンプルの組全体を、「fを特定するためのパラメータ」と言うこともできる。
The color difference prediction image PredU is obtained using f U (Y), that is, according to the following equation (4):
PredU = f U [RecY [x Y, y Y]] ... (4)
Is derived. It shows an example of f U in FIG. 20. As shown in FIG. 20, f U may be derived as a curve that passes through each sample point. Incidentally, f U, between adjacent samples may be derived independently. Incidentally, speaking from another perspective, the entire set of samples, can also be referred to as "parameters for specifying the f U".
   [線形変換との併用]
 最小二乗法エラー値の判定に基づいて、従来の線形変換によるチャンネル間予測と、上記LUTによるチャンネル間予測とを切り替えてもよい。これにより、より精度が高いほうをチャンネル間予測に用いてもかまわない。例えば、画像が単色で構成される場合や、画像のほとんどが単調なグラデーション領域により構成される場合などにおいて、線形変換を用いる構成とすることができる。
[Combination with linear transformation]
Based on the determination of the least square method error value, the inter-channel prediction based on the conventional linear transformation and the inter-channel prediction based on the LUT may be switched. As a result, the higher accuracy may be used for inter-channel prediction. For example, when the image is configured with a single color, or when most of the image is configured with a monotone gradation region, a configuration using linear conversion can be used.
 〔動画像符号化装置〕
 以下では、本実施形態に係る動画像符号化装置2の構成について、図21を参照して説明する。なお、すでに説明した部材と同じ部材については、同じ符号を付し、その説明を省略する。
[Moving picture encoding device]
Below, the structure of the moving image encoder 2 which concerns on this embodiment is demonstrated with reference to FIG. In addition, the same code | symbol is attached | subjected about the same member as the already demonstrated member, and the description is abbreviate | omitted.
  (動画像符号化装置の概要)
 動画像符号化装置2は、概略的に言えば、入力画像#10を符号化することによって符号化データ#1を生成し、出力する装置である。
(Outline of video encoding device)
Generally speaking, the moving image encoding device 2 is a device that generates and outputs encoded data # 1 by encoding the input image # 10.
  (動画像符号化装置の構成)
 図21は、動画像符号化装置2の構成について示す機能ブロック図である。図21に示すように、動画像符号化装置2は、符号化設定部21、逆量子化・逆変換部22、予測画像生成部23、加算器24、フレームメモリ25、減算器26、変換・量子化部27、および可変長符号化部28を備えている。
(Configuration of video encoding device)
FIG. 21 is a functional block diagram showing the configuration of the moving image encoding device 2. As illustrated in FIG. 21, the moving image encoding device 2 includes an encoding setting unit 21, an inverse quantization / inverse conversion unit 22, a predicted image generation unit 23, an adder 24, a frame memory 25, a subtractor 26, a conversion / A quantization unit 27 and a variable length coding unit 28 are provided.
 符号化設定部21は、入力画像#10に基づいて、符号化に関する画像データおよび各種の設定情報を生成する。 The encoding setting unit 21 generates image data related to encoding and various setting information based on the input image # 10.
 具体的には、符号化設定部21は、次の画像データおよび設定情報を生成する。 Specifically, the encoding setting unit 21 generates the next image data and setting information.
 まず、符号化設定部21は、入力画像#10を、スライス単位、LCU単位に順次分割することにより、対象リーフCUについてのリーフCU画像#100を生成する。 First, the encoding setting unit 21 generates the leaf CU image # 100 for the target leaf CU by sequentially dividing the input image # 10 into slice units and LCU units.
 また、符号化設定部21は、分割処理の結果に基づいて、ヘッダ設定情報H’を生成する。ヘッダ情報H’は、(1)対象スライスに属するLCUのサイズ、形状および対象スライス内での位置についての情報、並びに、(2)各LCUに属するリーフCUのサイズ、形状および対象LCU内での位置についてのCU情報CU’を含んでいる。 Also, the encoding setting unit 21 generates header setting information H ′ based on the result of the division process. The header information H ′ includes (1) information about the size, shape and position of the LCU belonging to the target slice, and (2) the size, shape and shape of the leaf CU belonging to each LCU. It includes CU information CU ′ about the position.
 さらに、符号化設定部21は、リーフCU画像#100、および、CU情報CU’を参照して、PU設定情報PUI’を生成する。PU設定情報PUI’には、(1)対象リーフCUの各パーティションへの可能な分割パターン、および、(2)各パーティションに割り付ける可能な予測モード、の全ての組み合わせに関する情報が含まれる。 Furthermore, the encoding setting unit 21 refers to the leaf CU image # 100 and the CU information CU 'to generate PU setting information PUI'. The PU setting information PUI 'includes information on all combinations of (1) possible division patterns for each partition of the target leaf CU and (2) prediction modes that can be assigned to each partition.
 符号化設定部21は、リーフCU画像#100を減算器26に供給する。 The encoding setting unit 21 supplies the leaf CU image # 100 to the subtractor 26.
 また、符号化設定部21は、ヘッダ情報H’を可変長符号化部28に供給する。また、符号化設定部21は、PU設定情報PUI’を予測画像生成部23に供給する。 Also, the encoding setting unit 21 supplies the header information H ′ to the variable length encoding unit 28. Also, the encoding setting unit 21 supplies the PU setting information PUI ′ to the predicted image generation unit 23.
 逆量子化・逆変換部22は、変換・量子化部27より供給される、ブロック毎の量子化予測残差を、逆量子化、および、逆DCT変換(Inverse Discrete Cosine Transform)することによって、ブロック毎の予測残差を復元する。また、逆量子化・逆変換部22は、ブロック毎の予測残差を、TU分割情報により指定される分割パターンに従って統合し、対象リーフCUについての予測残差Dを生成する。逆量子化・逆変換部22は、生成した対象リーフCUについての予測残差Dを、加算器24に供給する。 The inverse quantization / inverse transform unit 22 performs inverse quantization and inverse DCT transform (Inverse Discrete Cosine Transform) on the quantization prediction residual for each block supplied from the transform / quantization unit 27, Restore the prediction residual for each block. Further, the inverse quantization / inverse transform unit 22 integrates the prediction residual for each block according to the division pattern specified by the TU partition information, and generates a prediction residual D for the target leaf CU. The inverse quantization / inverse transform unit 22 supplies the prediction residual D for the generated target leaf CU to the adder 24.
 予測画像生成部23は、フレームメモリ25に記録されている局所復号画像P’、および、PU設定情報PUI’を参照して、対象リーフCUについての予測画像Predを生成する。色差の予測においてチャンネル間予測を行う場合、予測画像生成部23は、輝度復号画像Pを参照する。予測画像生成部23は、予測画像生成処理により得られた予測パラメータを、PU設定情報PUI’に設定し、設定後のPU設定情報PUI’を可変長符号化部28に転送する。なお、予測画像生成部23による予測画像生成処理は、動画像復号装置1の備える予測画像生成部13と同様であるので、ここでは説明を省略する。 The predicted image generation unit 23 refers to the locally decoded image P ′ recorded in the frame memory 25 and the PU setting information PUI ′, and generates a predicted image Pred for the target leaf CU. When performing inter-channel prediction in the prediction of chrominance, the prediction image generation unit 23 refers to the luminance decoded image P Y. The predicted image generation unit 23 sets the prediction parameter obtained by the predicted image generation process in the PU setting information PUI ′, and transfers the set PU setting information PUI ′ to the variable length encoding unit 28. Note that the predicted image generation process performed by the predicted image generation unit 23 is the same as that performed by the predicted image generation unit 13 included in the video decoding device 1, and thus description thereof is omitted here.
 加算器24は、予測画像生成部23より供給される予測画像Predと、逆量子化・逆変換部22より供給される予測残差Dとを加算することによって、対象リーフCUについての復号画像Pを生成する。 The adder 24 adds the predicted image Pred supplied from the predicted image generation unit 23 and the prediction residual D supplied from the inverse quantization / inverse transform unit 22 to add a decoded image P for the target leaf CU. Is generated.
 フレームメモリ25には、復号された復号画像Pが順次記録される。フレームメモリ25には、対象LCUを復号する時点において、当該対象LCUよりも先に復号された全てのLCU(例えば、ラスタスキャン順で先行する全てのLCU)に対応する復号画像が記録されている。 Decoded decoded image P is sequentially recorded in the frame memory 25. In the frame memory 25, decoded images corresponding to all the LCUs decoded before the target LCU (for example, all the LCUs preceding in the raster scan order) at the time of decoding the target LCU are recorded. .
 減算器26は、リーフCU画像#100から予測画像Predを減算することによって、対象リーフCUについての予測残差Dを生成する。減算器26は、生成した予測残差Dを、変換・量子化部27に供給する。 The subtracter 26 generates a prediction residual D for the target leaf CU by subtracting the prediction image Pred from the leaf CU image # 100. The subtractor 26 supplies the generated prediction residual D to the transform / quantization unit 27.
 変換・量子化部27は、予測残差Dに対して、DCT変換(Discrete Cosine Transform)および量子化を行うことで量子化予測残差を生成する。 The transform / quantization unit 27 performs a DCT transform (Discrete Cosine Transform) and quantization on the prediction residual D to generate a quantized prediction residual.
 具体的には、変換・量子化部27は、リーフCU画像#100、および、CU情報CU’を参照し、対象リーフCUの1または複数のブロックへの分割パターンを決定する。また、決定された分割パターンに従って、予測残差Dを、各ブロックについての予測残差に分割する。 Specifically, the transform / quantization unit 27 refers to the leaf CU image # 100 and the CU information CU ', and determines the division pattern of the target leaf CU into one or a plurality of blocks. Further, according to the determined division pattern, the prediction residual D is divided into prediction residuals for each block.
 また、変換・量子化部27は、各ブロックについての予測残差をDCT変換(Discrete Cosine Transform)することによって周波数領域における予測残差を生成した後、当該周波数領域における予測残差を量子化することによってブロック毎の量子化予測残差を生成する。 The transform / quantization unit 27 generates a prediction residual in the frequency domain by performing DCT transform (DiscretecreCosine Transform) on the prediction residual for each block, and then quantizes the prediction residual in the frequency domain. Thus, a quantized prediction residual for each block is generated.
 また、変換・量子化部27は、生成したブロック毎の量子化予測残差と、対象リーフCUの分割パターンを指定するTU分割情報と、対象リーフCUの各ブロックへの可能な全分割パターンに関する情報とを含むTU設定情報TUI’を生成する。 Also, the transform / quantization unit 27 relates to the generated quantization prediction residual for each block, TU partition information that specifies the partition pattern of the target leaf CU, and all possible partition patterns for each block of the target leaf CU. TU setting information TUI ′ including the information is generated.
 変換・量子化部27は、生成したTU設定情報TUI’を、逆量子化・逆変換部22、および、可変長符号化部28に供給する。 The transform / quantization unit 27 supplies the generated TU setting information TUI 'to the inverse quantization / inverse transform unit 22 and the variable length coding unit 28.
 可変長符号化部28は、TU設定情報TUI’、PU設定情報PUI’、ヘッダ情報H’に基づいて符号化データ#1を生成し、出力する。 The variable length encoding unit 28 generates and outputs encoded data # 1 based on the TU setting information TUI ′, the PU setting information PUI ′, and the header information H ′.
  (色差チャンネル処理順序および予測元チャンネルの符号化処理の流れ)
 次に、図22を用いて、動画像符号化装置2における色差チャンネル処理順序および予測元チャンネルの符号化処理の流れについて説明する。図22は、動画像符号化装置2における色差チャンネル処理順序および予測元チャンネルの符号化処理の流れの一例について示したフローチャートである。
(Color difference channel processing order and prediction source channel encoding process flow)
Next, the color difference channel processing order and the flow of the encoding process of the prediction source channel in the moving image encoding device 2 will be described with reference to FIG. FIG. 22 is a flowchart illustrating an example of the flow of the color difference channel processing order and the prediction source channel encoding process in the moving image encoding apparatus 2.
 符号化処理が開始されると、予測画像生成部23が、対象パーティションについての輝度予測画像生成処理における輝度予測モードを選択する(S200)。 When the encoding process is started, the predicted image generation unit 23 selects a luminance prediction mode in the luminance predicted image generation process for the target partition (S200).
 次に、予測画像生成部23は、選択した予測モードに基づいて輝度予測画像を生成する(S201)。 Next, the predicted image generation unit 23 generates a brightness predicted image based on the selected prediction mode (S201).
 次に、予測画像生成部23は、色差チャンネルの処理順序の各パターンについての色差予測画像作成処理ループLP21に入る。 Next, the predicted image generation unit 23 enters the color difference predicted image creation processing loop LP21 for each pattern in the color difference channel processing order.
 このループLP21では、対象パーティションについて、以下の色差予測画像作成処理が行われる。 In this loop LP21, the following color difference prediction image creation processing is performed for the target partition.
 まず、チャンネル間予測処理が開始されると、予測画像生成部23は、図8、図16および図17を用いて説明した各処理順序のパターンに従って、色差Aおよび色差Aの予測元チャンネルと、色差Bおよび色差Bの予測元チャンネルとを設定する(S203)。 First, when the inter-channel prediction process is started, the predicted image generation unit 23 performs the color difference A and the prediction source channel of the color difference A according to the pattern of each processing order described using FIG. 8, FIG. 16, and FIG. The color difference B and the prediction source channel of the color difference B are set (S203).
 すなわち、ステップS203では、チャンネル間予測部351が、UチャンネルおよびVチャンネルの一方を、先にチャンネル推定を行う色差Aに設定し、他方を、色差Aの後に推定を行う色差Bに設定する。また、チャンネル間予測部351は、色差Aおよび色差Bについて、それぞれ1または複数の予測元チャンネルを設定する。 That is, in step S203, the inter-channel prediction unit 351 sets one of the U channel and the V channel to the color difference A for which channel estimation is performed first, and the other is set to the color difference B for which estimation is performed after the color difference A. Further, the inter-channel prediction unit 351 sets one or a plurality of prediction source channels for the color difference A and the color difference B, respectively.
 次に、予測画像生成部23は、色差Aの予測モードを選択し(S204)、ステップS120における設定に従って、色差Aについてチャンネル間予測により色差予測画像を生成する(S205)。 Next, the predicted image generation unit 23 selects a prediction mode for the color difference A (S204), and generates a color difference predicted image for the color difference A by inter-channel prediction according to the setting in step S120 (S205).
 次に、予測画像生成部23は、色差Bの予測モードを選択し(S206)、ステップS120における設定に従って、色差Bについてチャンネル間予測により色差予測画像を生成する(S207)。さらにチャンネル間予測による色差予測画像作成処理は継続し、ループLP21の先頭に戻る(S208)。 Next, the predicted image generation unit 23 selects a prediction mode for the color difference B (S206), and generates a color difference predicted image by inter-channel prediction for the color difference B according to the setting in step S120 (S207). Furthermore, the color difference predicted image creation process by inter-channel prediction continues, and the process returns to the top of the loop LP21 (S208).
 色差チャンネルの処理順序の各パターンについての色差予測画像作成処理が完了することにより、ループLP21は終了する。 When the color difference prediction image creation processing for each pattern in the processing order of the color difference channel is completed, the loop LP21 ends.
 次に、予測画像生成部23は、ループLP21においてチャンネル間予測処理を行った結果、最も符号化に適した色差チャンネルの処理順序および予測元チャンネルの組み合わせを選択する(S209)。 Next, as a result of performing the inter-channel prediction processing in the loop LP21, the predicted image generation unit 23 selects a color difference channel processing order and a combination of prediction source channels that are most suitable for encoding (S209).
 予測画像生成部23は、チャンネル間予測フラグ、色差チャンネル処理順序フラグおよび第2チャンネル予測元チャンネル指定子をイントラ予測パラメータPPに含めて符号化を行う(S210)。 The prediction image generation unit 23 performs encoding by including the inter-channel prediction flag, the color difference channel processing order flag, and the second channel prediction source channel specifier in the intra prediction parameter PP (S210).
 なお、予測画像生成部23は、ステップS210において、処理順序および予測元チャンネルの組み合わせを示すチャンネル間予測インデックスを符号化してもよい。
〔2〕実施形態2
 本発明の他の実施形態について図23~図26に基づいて説明すると、以下の通りである。なお、説明の便宜上、前記実施形態1にて説明した図面と同じ機能を有する部材については、同じ符号を付記し、その説明を省略する。
Note that the predicted image generation unit 23 may encode the inter-channel prediction index indicating the combination of the processing order and the prediction source channel in step S210.
[2] Embodiment 2
Another embodiment of the present invention will be described below with reference to FIGS. For convenience of explanation, members having the same functions as those in the drawings described in the first embodiment are denoted by the same reference numerals and description thereof is omitted.
 〔画像復号装置〕
  (構成について)
 まず、図23を用いて、本実施形態に係る予測画像生成部13Aの構成について説明する。図23は、予測画像生成部13の構成の他の例について示す機能ブロック図である。
[Image decoding device]
(About configuration)
First, the configuration of the predicted image generation unit 13A according to the present embodiment will be described with reference to FIG. FIG. 23 is a functional block diagram illustrating another example of the configuration of the predicted image generation unit 13.
 予測画像生成部13Aでは、LUTをLCU単位で初期化し、同一LCU内における各ブロック(対象パーティション)では、LUTを更新する。 The predicted image generation unit 13A initializes the LUT in units of LCU, and updates the LUT in each block (target partition) in the same LCU.
 図21に示すように、予測画像生成部13Aでは、チャンネル間予測部351を、チャンネル間予測部351Aに変更している。 21, in the predicted image generation unit 13A, the inter-channel prediction unit 351 is changed to an inter-channel prediction unit 351A.
 また、予測画像生成部13Aでは、図1に示す予測画像生成部13とは異なり、予測画像生成部13Aとは別に、LUT導出部(相関導出手段)16を設けている。 Further, unlike the predicted image generation unit 13 shown in FIG. 1, the predicted image generation unit 13A is provided with an LUT derivation unit (correlation derivation means) 16 separately from the predicted image generation unit 13A.
 以下、この相違点について説明すると次のとおりである。 Hereafter, this difference will be explained as follows.
 まず、LUT導出部16は、対象パーティションごとにLUTを新たに導出するLUT導出部134と異なり、対象LCUごとにLUTを新たに導出する。そして、LUT導出部16は、パーティション単位で、LUTを更新する。LUT導出部16と、LUT導出部134との違いはこのLUTを導出する単位である。 First, unlike the LUT derivation unit 134 that newly derives an LUT for each target partition, the LUT derivation unit 16 newly derives an LUT for each target LCU. The LUT deriving unit 16 updates the LUT in units of partitions. The difference between the LUT deriving unit 16 and the LUT deriving unit 134 is a unit for deriving this LUT.
 それ以外については、同様の構成とすることができるので、ここではその詳細な説明は省略する。 Other than that, since the same configuration can be adopted, detailed description thereof is omitted here.
 チャンネル間予測部351Aは、LUT導出部16を参照するように変更している。 The inter-channel prediction unit 351A is changed to refer to the LUT deriving unit 16.
  (色差予測画像生成処理の概要)
 次に、図24を用いて、予測画像生成部13Aにおける色差予測画像生成処理の概略的な流れについて説明する。図24は、予測画像生成部13Aにおける色差予測画像生成処理の概略的な流れについて例示したフローチャートである。
(Outline of color difference prediction image generation processing)
Next, a schematic flow of color difference predicted image generation processing in the predicted image generation unit 13A will be described with reference to FIG. FIG. 24 is a flowchart illustrating a schematic flow of color difference predicted image generation processing in the predicted image generation unit 13A.
 色差予測画像生成処理が開始されると、LUT導出部16が、対象パーティションが、LCUで最初に処理されるブロックであるか否かを判定する(S30)。対象パーティションが、LCUで最初に処理されるブロックである場合(S30においてYES)、LUTを初期化する(S31)。 When the color difference prediction image generation process is started, the LUT deriving unit 16 determines whether or not the target partition is a block that is first processed by the LCU (S30). When the target partition is a block processed first by the LCU (YES in S30), the LUT is initialized (S31).
 ステップS31でのLUTの初期化においては、LUT[n]=128と、サンプル有りフラグLUT_F[n]=0との代入処理を行う(LUT,LUT_Fについてn=0…255)。この初期化は、UチャンネルおよびVチャンネルの2チャンネル分行われる。なお、サンプル有りフラグLUT_Fは、LUT_F[n]=1が、サンプル有りを意味し、LUT_F[n]=0が、サンプル無しを意味する。 In the initialization of the LUT in step S31, the LUT [n] = 128 and the sample presence flag LUT_F [n] = 0 are substituted (n = 0 to 255 for LUT and LUT_F). This initialization is performed for two channels of the U channel and the V channel. In the sample presence flag LUT_F, LUT_F [n] = 1 indicates that there is a sample, and LUT_F [n] = 0 indicates that there is no sample.
 次に、LUT導出部16は、局所復号画像P’を参照しながら、各チャンネルのLUTを更新する(S32)。このLUT更新処理の詳細については後述する。 Next, the LUT deriving unit 16 updates the LUT of each channel while referring to the local decoded image P ′ (S32). Details of the LUT update processing will be described later.
 なお、対象パーティションが、LCUで最初に処理されるブロックで無い場合(S30においてNO)、LUT導出部16は、LUTを初期化せずに、LUT更新処理を実行する(S32)。 If the target partition is not the first block processed by the LCU (NO in S30), the LUT deriving unit 16 executes the LUT update process without initializing the LUT (S32).
 次に、チャンネル間予測判定部133が、チャンネル間予測フラグを参照して、チャンネル間予測モードであるか否かを判定する(S33)。 Next, the inter-channel prediction determination unit 133 refers to the inter-channel prediction flag and determines whether or not the inter-channel prediction mode is set (S33).
 ここで、判定の結果、チャンネル間予測モードでない場合(S33においてNO)、チャンネル内予測部352がチャンネル間予測によらないで色差予測画像PredCを生成し(S36)、処理は終了する。 Here, if it is determined that the mode is not the inter-channel prediction mode (NO in S33), the intra-channel prediction unit 352 generates the color difference prediction image PredC without using the inter-channel prediction (S36), and the process ends.
 これに対して、判定の結果、チャンネル間予測モードである場合(S33においてYES)、チャンネル間予測部351Aが、LUT導出部16が更新したLUTを参照して、チャンネル間予測による色差予測画像PredCを生成する(S35)。以上で、処理は終了する。 On the other hand, if it is determined that the mode is the inter-channel prediction mode (YES in S33), the inter-channel prediction unit 351A refers to the LUT updated by the LUT deriving unit 16, and the color difference prediction image PredC by the inter-channel prediction Is generated (S35). This is the end of the process.
  (LUT導出処理の流れ)
 次に、図25を用いて、LUT導出部16によるLUT更新処理の流れについて説明する。図25は、LUT導出部16によるLUT更新処理の流れの一例について示したフローチャートである。
(LUT derivation process flow)
Next, the flow of LUT update processing by the LUT deriving unit 16 will be described with reference to FIG. FIG. 25 is a flowchart showing an example of the flow of LUT update processing by the LUT deriving unit 16.
 図25に示すように、LUT更新処理が開始されると、LUT導出部16は、対象パーティションに隣接する各輝度画素についての登録処理のループLP41に入る(S400)。 As shown in FIG. 25, when the LUT update process is started, the LUT deriving unit 16 enters a registration process loop LP41 for each luminance pixel adjacent to the target partition (S400).
 このループLP41では、処理対象となる輝度画素について以下の更新処理が行われる。まず、LUT導出部16は、処理対象となる輝度画素における輝度値=nおよび、輝度画素位置に対応する色差画素位置における色差値=mを取得する(S401)。 In this loop LP41, the following update process is performed for the luminance pixel to be processed. First, the LUT deriving unit 16 obtains the luminance value = n at the luminance pixel to be processed and the color difference value = m at the color difference pixel position corresponding to the luminance pixel position (S401).
 次に、LUT導出部16は、LUT[n]に色差値が登録されているか否かを判定する(S402)。すなわち、LUT導出部16は、LUT_F[n]が1(サンプル有り)であれば、色差値が登録されていると判定する一方で、LUT_F[n]が0(サンプル無し)であれば、色差値が登録されていないと判定する。 Next, the LUT deriving unit 16 determines whether or not a color difference value is registered in the LUT [n] (S402). In other words, the LUT deriving unit 16 determines that the color difference value is registered if LUT_F [n] is 1 (with a sample), whereas if LUT_F [n] is 0 (no sample), the color difference is determined. It is determined that the value is not registered.
 ここで、判定の結果、LUT[n]に色差値が登録されていなかった場合(S402においてNO)、LUT導出部134は、ステップS401において取得した色差値mをそのまま、LUT[n]に登録するとともに、LUT_F[n]に1(サンプル有り)を代入して(S404)、ループLP11の先頭に戻る(S405)。 If the color difference value is not registered in LUT [n] as a result of the determination (NO in S402), the LUT derivation unit 134 registers the color difference value m acquired in Step S401 as it is in LUT [n]. At the same time, 1 (with sample) is substituted into LUT_F [n] (S404), and the process returns to the top of the loop LP11 (S405).
 一方、判定の結果、LUT[n]が既に登録されていた場合、LUT導出部16は、(m+LUT[n]+1)/2を計算することより、取得した色差値mと登録済みの色差値との平均値を算出する。LUT導出部16は、このようにして算出した平均値を色差値mに代入する(S403)。続いて、LUT導出部16は、ステップS403において平均値を代入した色差値mを、LUT[n]に登録するとともに、LUT_F[n]に1(サンプル有り)を代入して(S404)、ループLP41の先頭に戻る(S405)。 On the other hand, if the result of determination is that LUT [n] has already been registered, the LUT derivation unit 16 calculates (m + LUT [n] +1) / 2, thereby obtaining the acquired color difference value m and the registered color difference value. And the average value is calculated. The LUT deriving unit 16 substitutes the average value thus calculated for the color difference value m (S403). Subsequently, the LUT deriving unit 16 registers the color difference value m into which the average value is substituted in step S403 in LUT [n], and substitutes 1 (with sample) into LUT_F [n] (S404). Return to the top of LP41 (S405).
 各輝度画素についての更新処理が終了すると、ループLP41は終了する。 When the update process for each luminance pixel ends, the loop LP41 ends.
 ループLP41が終了した後、未登録LUTエントリの補間処理のループLP42に入る(S406)。 After the loop LP41 is completed, the process enters the loop LP42 for interpolation processing of unregistered LUT entries (S406).
 このループLP42では、n=0からn=255までのエントリについて以下の補間処理が行われる。まず、LUT導出部134は、LUT_F[n]が0(サンプル無し)であるか否かを判定する(S407)。 In this loop LP42, the following interpolation processing is performed for entries from n = 0 to n = 255. First, the LUT deriving unit 134 determines whether or not LUT_F [n] is 0 (no sample) (S407).
 LUT自体は、同一LCUを処理している間は保持される。また、LCUで最初に処理されるブロックにおいて、補間により全てのエントリが登録済みとなる。よって、ここでは、LUT[n]が登録済みであるかどうかを確認する代わりに、サンプルが存在するかどうかを確認している。そして、サンプルが存在しないエントリについては、再度補間の対象とする。 The LUT itself is retained while processing the same LCU. In addition, in the block processed first by the LCU, all entries are registered by interpolation. Therefore, here, instead of confirming whether or not LUT [n] has been registered, it is confirmed whether or not a sample exists. An entry for which no sample exists is again subject to interpolation.
 ここで、LUT_F[n]が0(サンプル無し)で無い場合、すなわちLUT[n]が登録済みである場合(S407においてNO)、さらに補間処理は継続し、ループLP42の先頭に戻る(S410)。 If LUT_F [n] is not 0 (no sample), that is, if LUT [n] has already been registered (NO in S407), the interpolation process continues and returns to the top of the loop LP42 (S410). .
 一方、LUT_F[n]が0(サンプル無し)である場合(S407においてYES)、LUT導出部16は、nの前後にある直近のサンプル有りのエントリを探索する(S408)。 On the other hand, when LUT_F [n] is 0 (no sample) (YES in S407), the LUT deriving unit 16 searches for the most recent sample entry before and after n (S408).
 ここで、対象パーティションにおいて、直前の対象パーティションにおける補間のサンプル点よりも、近い値を有するサンプル点がエントリとしてLUTに登録されなかった場合は、図13に示したような補間が行われる。 Here, in the target partition, when a sample point having a value closer to the sample point of interpolation in the immediately preceding target partition is not registered as an entry in the LUT, interpolation as shown in FIG. 13 is performed.
 一方、対象パーティションにおいて、直前の対象パーティションにおける補間のサンプル点よりも、近い値を有するサンプル点がエントリとしてLUTに登録された場合は、図26に示すような再補間が行われる。 On the other hand, when a sample point having a value closer to the interpolation sample point in the immediately preceding target partition is registered in the LUT as an entry in the target partition, re-interpolation as shown in FIG. 26 is performed.
 図26に示す例では、直前の対象パーティションにおいて、サンプル点Smpl1およびSmpl2が登録され、補間処理が行われていたところ、対象パーティションにおいて、サンプル点Smpl3が新たに登録された場合について示している。 In the example shown in FIG. 26, the sample points Smpl1 and Smpl2 are registered in the immediately preceding target partition and the interpolation process is performed, but the sample point Smpl3 is newly registered in the target partition.
 まず、LUT導出部16は、nの前後にある直近のサンプル有りのエントリを探索する。 First, the LUT deriving unit 16 searches for entries with the latest sample before and after n.
 まず、LUT導出部16は、nを基点として、nの前方向、すなわち、nよりも小さいnLについて登録済みエントリを探索する。つまり、ここでは、図26に示すサンプル点Smpl1を探索している。 First, the LUT deriving unit 16 searches for registered entries for n forwards, that is, nL smaller than n, with n as a base point. That is, here, the sample point Smpl1 shown in FIG. 26 is searched.
 また、LUT導出部16は、nを基点として、nの後方向、すなわち、nよりも大きいnRについて登録済みであるエントリを探索する。ここでは、直前までの対象パーティションでは、図26に示すサンプル点Smpl2が探索されていたところ、対象パーティションにおいて、サンプル点Smpl3が登録されたことにより、サンプル点Smpl3を探索する。 Also, the LUT deriving unit 16 searches for an entry registered for n backward, that is, nR larger than n, with n as a base point. Here, the sample point Smpl2 shown in FIG. 26 has been searched for in the target partition up to immediately before, but the sample point Smpl3 is searched for when the sample point Smpl3 is registered in the target partition.
 次に、LUT導出部16は、LUT[nL]と、LUT[nR]とを直線補間した値を、LUT[n]に登録する(S409)。 Next, the LUT deriving unit 16 registers a value obtained by linear interpolation between the LUT [nL] and the LUT [nR] in the LUT [n] (S409).
 なお、ステップS408の探索の結果、nLおよびnRのいずれか一方しか検知されなかった場合、検知された方の登録済みエントリの値が、LUT[n]に登録される。 If only one of nL and nR is detected as a result of the search in step S408, the value of the registered entry of the detected one is registered in LUT [n].
 さらに補間処理は継続し、ループLP42の先頭に戻る(S410)。 Further, the interpolation process is continued and the process returns to the top of the loop LP42 (S410).
 ループLP42において、n=0からn=255までのすべてのエントリについて補間処理が完了することにより、ループLP42は終了する。そしてLUT導出処理は終了する。 In the loop LP42, when the interpolation processing is completed for all entries from n = 0 to n = 255, the loop LP42 ends. Then, the LUT derivation process ends.
 なお、ステップS403では、LUT導出部16が、(m+LUT[n]+1)/2を計算することより、取得した色差値mと登録済みの色差値との平均値を算出していたが、これに限られない。取得した色差値mと登録済みの色差値とについて例えば1:3の加重平均を算出してもよい。 In step S403, the LUT deriving unit 16 calculates the average value of the acquired color difference value m and the registered color difference value by calculating (m + LUT [n] +1) / 2. Not limited to. For example, a weighted average of 1: 3 may be calculated for the acquired color difference value m and the registered color difference value.
  (効果)
 以上のように、予測画像生成部13Aは、LUTをLCU単位で初期化し、同一LCU内における各ブロック(対象パーティション)では、LUTを更新する構成である。
(effect)
As described above, the predicted image generation unit 13A is configured to initialize the LUT in units of LCUs and update the LUT in each block (target partition) in the same LCU.
 また、LUTは、導出に用いるサンプル数が多いほど精度が高くなる傾向がある。しかし、あまり広範にサンプルを取得してしまうと相関が失われたり、著しく小さくなってしまったりするおそれがある。ここで、LCUは、対象パーティションよりも広い範囲であって、なおかつ相関が存在すると想定される範囲である。 Also, the LUT tends to increase in accuracy as the number of samples used for derivation increases. However, if the sample is acquired too widely, the correlation may be lost or it may become extremely small. Here, the LCU is a range that is wider than the target partition and is assumed to have a correlation.
 このようなLCUにおけるサンプル点の相関からLUTを導出することにより、対象パーティションよりも広い範囲でより多くのサンプル点を取得でき、チャンネル間予測の予測精度を向上させることができる。
〔応用例〕
 上述した動画像符号化装置2及び動画像復号装置1は、動画像の送信、受信、記録、再生を行う各種装置に搭載して利用することができる。なお、動画像は、カメラ等により撮像された自然動画像であってもよいし、コンピュータ等により生成された人工動画像(CGおよびGUIを含む)であってもよい。
By deriving the LUT from the correlation of sample points in such an LCU, more sample points can be acquired in a wider range than the target partition, and the prediction accuracy of inter-channel prediction can be improved.
[Application example]
The above-described moving image encoding device 2 and moving image decoding device 1 can be used by being mounted on various devices that perform transmission, reception, recording, and reproduction of moving images. The moving image may be a natural moving image captured by a camera or the like, or may be an artificial moving image (including CG and GUI) generated by a computer or the like.
 まず、上述した動画像符号化装置2及び動画像復号装置1を、動画像の送信及び受信に利用できることを、図30を参照して説明する。 First, it will be described with reference to FIG. 30 that the above-described moving image encoding device 2 and moving image decoding device 1 can be used for transmission and reception of moving images.
 図30の(a)は、動画像符号化装置2を搭載した送信装置PROD_Aの構成を示したブロック図である。図30の(a)に示すように、送信装置PROD_Aは、動画像を符号化することによって符号化データを得る符号化部PROD_A1と、符号化部PROD_A1が得た符号化データで搬送波を変調することによって変調信号を得る変調部PROD_A2と、変調部PROD_A2が得た変調信号を送信する送信部PROD_A3と、を備えている。上述した動画像符号化装置2は、この符号化部PROD_A1として利用される。 30 (a) is a block diagram illustrating a configuration of a transmission device PROD_A in which the moving image encoding device 2 is mounted. As illustrated in (a) of FIG. 30, the transmission device PROD_A modulates a carrier wave with an encoding unit PROD_A1 that obtains encoded data by encoding a moving image, and the encoded data obtained by the encoding unit PROD_A1. Thus, a modulation unit PROD_A2 that obtains a modulation signal and a transmission unit PROD_A3 that transmits the modulation signal obtained by the modulation unit PROD_A2 are provided. The moving image encoding apparatus 2 described above is used as the encoding unit PROD_A1.
 送信装置PROD_Aは、符号化部PROD_A1に入力する動画像の供給源として、動画像を撮像するカメラPROD_A4、動画像を記録した記録媒体PROD_A5、動画像を外部から入力するための入力端子PROD_A6、及び、画像を生成または加工する画像処理部A7を更に備えていてもよい。図30の(a)においては、これら全てを送信装置PROD_Aが備えた構成を例示しているが、一部を省略しても構わない。 The transmission device PROD_A is a camera PROD_A4 that captures a moving image, a recording medium PROD_A5 that records the moving image, an input terminal PROD_A6 that inputs the moving image from the outside, as a supply source of the moving image input to the encoding unit PROD_A1. An image processing unit A7 that generates or processes an image may be further provided. FIG. 30A illustrates a configuration in which the transmission apparatus PROD_A includes all of these, but a part of the configuration may be omitted.
 なお、記録媒体PROD_A5は、符号化されていない動画像を記録したものであってもよいし、伝送用の符号化方式とは異なる記録用の符号化方式で符号化された動画像を記録したものであってもよい。後者の場合、記録媒体PROD_A5と符号化部PROD_A1との間に、記録媒体PROD_A5から読み出した符号化データを記録用の符号化方式に従って復号する復号部(不図示)を介在させるとよい。 The recording medium PROD_A5 may be a recording of a non-encoded moving image, or a recording of a moving image encoded by a recording encoding scheme different from the transmission encoding scheme. It may be a thing. In the latter case, a decoding unit (not shown) for decoding the encoded data read from the recording medium PROD_A5 according to the recording encoding method may be interposed between the recording medium PROD_A5 and the encoding unit PROD_A1.
 図30の(b)は、動画像復号装置1を搭載した受信装置PROD_Bの構成を示したブロック図である。図30の(b)に示すように、受信装置PROD_Bは、変調信号を受信する受信部PROD_B1と、受信部PROD_B1が受信した変調信号を復調することによって符号化データを得る復調部PROD_B2と、復調部PROD_B2が得た符号化データを復号することによって動画像を得る復号部PROD_B3と、を備えている。上述した動画像復号装置1は、この復号部PROD_B3として利用される。 (B) of FIG. 30 is a block diagram illustrating a configuration of the receiving device PROD_B in which the moving image decoding device 1 is mounted. As illustrated in (b) of FIG. 30, the receiving device PROD_B includes a receiving unit PROD_B1 that receives a modulated signal, a demodulating unit PROD_B2 that obtains encoded data by demodulating the modulated signal received by the receiving unit PROD_B1, and a demodulator. A decoding unit PROD_B3 that obtains a moving image by decoding the encoded data obtained by the unit PROD_B2. The moving picture decoding apparatus 1 described above is used as the decoding unit PROD_B3.
 受信装置PROD_Bは、復号部PROD_B3が出力する動画像の供給先として、動画像を表示するディスプレイPROD_B4、動画像を記録するための記録媒体PROD_B5、及び、動画像を外部に出力するための出力端子PROD_B6を更に備えていてもよい。図30の(b)においては、これら全てを受信装置PROD_Bが備えた構成を例示しているが、一部を省略しても構わない。 The receiving device PROD_B has a display PROD_B4 for displaying a moving image, a recording medium PROD_B5 for recording the moving image, and an output terminal for outputting the moving image to the outside as a supply destination of the moving image output by the decoding unit PROD_B3. PROD_B6 may be further provided. FIG. 30B illustrates a configuration in which the reception device PROD_B includes all of these, but a part of the configuration may be omitted.
 なお、記録媒体PROD_B5は、符号化されていない動画像を記録するためのものであってもよいし、伝送用の符号化方式とは異なる記録用の符号化方式で符号化されたものであってもよい。後者の場合、復号部PROD_B3と記録媒体PROD_B5との間に、復号部PROD_B3から取得した動画像を記録用の符号化方式に従って符号化する符号化部(不図示)を介在させるとよい。 The recording medium PROD_B5 may be used for recording a non-encoded moving image, or may be encoded using a recording encoding method different from the transmission encoding method. May be. In the latter case, an encoding unit (not shown) for encoding the moving image acquired from the decoding unit PROD_B3 according to the recording encoding method may be interposed between the decoding unit PROD_B3 and the recording medium PROD_B5.
 なお、変調信号を伝送する伝送媒体は、無線であってもよいし、有線であってもよい。また、変調信号を伝送する伝送態様は、放送(ここでは、送信先が予め特定されていない送信態様を指す)であってもよいし、通信(ここでは、送信先が予め特定されている送信態様を指す)であってもよい。すなわち、変調信号の伝送は、無線放送、有線放送、無線通信、及び有線通信の何れによって実現してもよい。 Note that the transmission medium for transmitting the modulation signal may be wireless or wired. Further, the transmission mode for transmitting the modulated signal may be broadcasting (here, a transmission mode in which the transmission destination is not specified in advance) or communication (here, transmission in which the transmission destination is specified in advance). Refers to the embodiment). That is, the transmission of the modulation signal may be realized by any of wireless broadcasting, wired broadcasting, wireless communication, and wired communication.
 例えば、地上デジタル放送の放送局(放送設備など)/受信局(テレビジョン受像機など)は、変調信号を無線放送で送受信する送信装置PROD_A/受信装置PROD_Bの一例である。また、ケーブルテレビ放送の放送局(放送設備など)/受信局(テレビジョン受像機など)は、変調信号を有線放送で送受信する送信装置PROD_A/受信装置PROD_Bの一例である。 For example, a terrestrial digital broadcast broadcasting station (broadcasting equipment or the like) / receiving station (such as a television receiver) is an example of a transmitting device PROD_A / receiving device PROD_B that transmits and receives a modulated signal by wireless broadcasting. Further, a broadcasting station (such as broadcasting equipment) / receiving station (such as a television receiver) of cable television broadcasting is an example of a transmitting device PROD_A / receiving device PROD_B that transmits and receives a modulated signal by cable broadcasting.
 また、インターネットを用いたVOD(Video On Demand)サービスや動画共有サービスなどのサーバ(ワークステーションなど)/クライアント(テレビジョン受像機、パーソナルコンピュータ、スマートフォンなど)は、変調信号を通信で送受信する送信装置PROD_A/受信装置PROD_Bの一例である(通常、LANにおいては伝送媒体として無線又は有線の何れかが用いられ、WANにおいては伝送媒体として有線が用いられる)。ここで、パーソナルコンピュータには、デスクトップ型PC、ラップトップ型PC、及びタブレット型PCが含まれる。また、スマートフォンには、多機能携帯電話端末も含まれる。 Also, a server (workstation etc.) / Client (television receiver, personal computer, smart phone etc.) such as VOD (Video On Demand) service and video sharing service using the Internet is a transmitting device for transmitting and receiving modulated signals by communication. This is an example of PROD_A / reception device PROD_B (usually, either a wireless or wired transmission medium is used in a LAN, and a wired transmission medium is used in a WAN). Here, the personal computer includes a desktop PC, a laptop PC, and a tablet PC. The smartphone also includes a multi-function mobile phone terminal.
 なお、動画共有サービスのクライアントは、サーバからダウンロードした符号化データを復号してディスプレイに表示する機能に加え、カメラで撮像した動画像を符号化してサーバにアップロードする機能を有している。すなわち、動画共有サービスのクライアントは、送信装置PROD_A及び受信装置PROD_Bの双方として機能する。 In addition to the function of decoding the encoded data downloaded from the server and displaying it on the display, the video sharing service client has a function of encoding a moving image captured by the camera and uploading it to the server. That is, the client of the video sharing service functions as both the transmission device PROD_A and the reception device PROD_B.
 次に、上述した動画像符号化装置2及び動画像復号装置1を、動画像の記録及び再生に利用できることを、図31を参照して説明する。 Next, it will be described with reference to FIG. 31 that the above-described moving picture encoding apparatus 2 and moving picture decoding apparatus 1 can be used for recording and reproduction of moving pictures.
 図31の(a)は、上述した動画像符号化装置2を搭載した記録装置PROD_Cの構成を示したブロック図である。図31の(a)に示すように、記録装置PROD_Cは、動画像を符号化することによって符号化データを得る符号化部PROD_C1と、符号化部PROD_C1が得た符号化データを記録媒体PROD_Mに書き込む書込部PROD_C2と、を備えている。上述した動画像符号化装置2は、この符号化部PROD_C1として利用される。 FIG. 31 (a) is a block diagram showing a configuration of a recording apparatus PROD_C in which the above-described moving picture encoding apparatus 2 is mounted. As shown in FIG. 31 (a), the recording device PROD_C has an encoding unit PROD_C1 that obtains encoded data by encoding a moving image, and the encoded data obtained by the encoding unit PROD_C1 on the recording medium PROD_M. A writing unit PROD_C2 for writing. The moving image encoding apparatus 2 described above is used as the encoding unit PROD_C1.
 なお、記録媒体PROD_Mは、(1)HDD(Hard Disk Drive)やSSD(Solid State Drive)などのように、記録装置PROD_Cに内蔵されるタイプのものであってもよいし、(2)SDメモリカードやUSB(Universal Serial Bus)フラッシュメモリなどのように、記録装置PROD_Cに接続されるタイプのものであってもよいし、(3)DVD(Digital Versatile Disc)やBD(Blu-ray Disc:登録商標)などのように、記録装置PROD_Cに内蔵されたドライブ装置(不図示)に装填されるものであってもよい。 The recording medium PROD_M may be of a type built in the recording device PROD_C, such as (1) HDD (Hard Disk Drive) or SSD (Solid State Drive), or (2) SD memory. It may be of the type connected to the recording device PROD_C, such as a card or USB (Universal Serial Bus) flash memory, or (3) DVD (Digital Versatile Disc) or BD (Blu-ray Disc: registration) Or a drive device (not shown) built in the recording device PROD_C.
 また、記録装置PROD_Cは、符号化部PROD_C1に入力する動画像の供給源として、動画像を撮像するカメラPROD_C3、動画像を外部から入力するための入力端子PROD_C4、動画像を受信するための受信部PROD_C5、及び、画像を生成または加工する画像処理部C6を更に備えていてもよい。図31の(a)においては、これら全てを記録装置PROD_Cが備えた構成を例示しているが、一部を省略しても構わない。 The recording device PROD_C is a camera PROD_C3 that captures moving images as a supply source of moving images to be input to the encoding unit PROD_C1, an input terminal PROD_C4 for inputting moving images from the outside, and reception for receiving moving images. The unit PROD_C5 and an image processing unit C6 that generates or processes an image may be further provided. FIG. 31A illustrates a configuration in which the recording apparatus PROD_C includes all of these, but a part of the configuration may be omitted.
 なお、受信部PROD_C5は、符号化されていない動画像を受信するものであってもよいし、記録用の符号化方式とは異なる伝送用の符号化方式で符号化された符号化データを受信するものであってもよい。後者の場合、受信部PROD_C5と符号化部PROD_C1との間に、伝送用の符号化方式で符号化された符号化データを復号する伝送用復号部(不図示)を介在させるとよい。 The receiving unit PROD_C5 may receive a non-encoded moving image, or may receive encoded data encoded by a transmission encoding scheme different from the recording encoding scheme. You may do. In the latter case, a transmission decoding unit (not shown) that decodes encoded data encoded by the transmission encoding method may be interposed between the reception unit PROD_C5 and the encoding unit PROD_C1.
 このような記録装置PROD_Cとしては、例えば、DVDレコーダ、BDレコーダ、HDD(Hard Disk Drive)レコーダなどが挙げられる(この場合、入力端子PROD_C4又は受信部PROD_C5が動画像の主な供給源となる)。また、カムコーダ(この場合、カメラPROD_C3が動画像の主な供給源となる)、パーソナルコンピュータ(この場合、受信部PROD_C5又は画像処理部C6が動画像の主な供給源となる)、スマートフォン(この場合、カメラPROD_C3又は受信部PROD_C5が動画像の主な供給源となる)なども、このような記録装置PROD_Cの一例である。 Examples of such a recording device PROD_C include a DVD recorder, a BD recorder, and an HDD (Hard Disk Drive) recorder (in this case, the input terminal PROD_C4 or the receiving unit PROD_C5 is a main supply source of moving images). . In addition, a camcorder (in this case, the camera PROD_C3 is a main source of moving images), a personal computer (in this case, the receiving unit PROD_C5 or the image processing unit C6 is a main source of moving images), a smartphone (in this case In this case, the camera PROD_C3 or the receiving unit PROD_C5 is a main supply source of moving images) is also an example of such a recording device PROD_C.
 図31の(b)は、上述した動画像復号装置1を搭載した再生装置PROD_Dの構成を示したブロックである。図31の(b)に示すように、再生装置PROD_Dは、記録媒体PROD_Mに書き込まれた符号化データを読み出す読出部PROD_D1と、読出部PROD_D1が読み出した符号化データを復号することによって動画像を得る復号部PROD_D2と、を備えている。上述した動画像復号装置1は、この復号部PROD_D2として利用される。 (B) of FIG. 31 is a block showing a configuration of a playback device PROD_D equipped with the above-described video decoding device 1. As shown in (b) of FIG. 31, the playback device PROD_D reads a moving image by decoding a read unit PROD_D1 that reads encoded data written to the recording medium PROD_M and a read unit PROD_D1 that reads the encoded data. And a decoding unit PROD_D2 to be obtained. The moving picture decoding apparatus 1 described above is used as the decoding unit PROD_D2.
 なお、記録媒体PROD_Mは、(1)HDDやSSDなどのように、再生装置PROD_Dに内蔵されるタイプのものであってもよいし、(2)SDメモリカードやUSBフラッシュメモリなどのように、再生装置PROD_Dに接続されるタイプのものであってもよいし、(3)DVDやBDなどのように、再生装置PROD_Dに内蔵されたドライブ装置(不図示)に装填されるものであってもよい。 Note that the recording medium PROD_M may be of the type built into the playback device PROD_D, such as (1) HDD or SSD, or (2) such as an SD memory card or USB flash memory, It may be of a type connected to the playback device PROD_D, or (3) may be loaded into a drive device (not shown) built in the playback device PROD_D, such as DVD or BD. Good.
 また、再生装置PROD_Dは、復号部PROD_D2が出力する動画像の供給先として、動画像を表示するディスプレイPROD_D3、動画像を外部に出力するための出力端子PROD_D4、及び、動画像を送信する送信部PROD_D5を更に備えていてもよい。図31の(b)においては、これら全てを再生装置PROD_Dが備えた構成を例示しているが、一部を省略しても構わない。 In addition, the playback device PROD_D has a display PROD_D3 that displays a moving image, an output terminal PROD_D4 that outputs the moving image to the outside, and a transmission unit that transmits the moving image as a supply destination of the moving image output by the decoding unit PROD_D2. PROD_D5 may be further provided. FIG. 31B illustrates a configuration in which the playback apparatus PROD_D includes all of these, but some of them may be omitted.
  なお、送信部PROD_D5は、符号化されていない動画像を送信するものであってもよいし、記録用の符号化方式とは異なる伝送用の符号化方式で符号化された符号化データを送信するものであってもよい。後者の場合、復号部PROD_D2と送信部PROD_D5との間に、動画像を伝送用の符号化方式で符号化する符号化部(不図示)を介在させるとよい。 The transmission unit PROD_D5 may transmit an unencoded moving image, or transmits encoded data encoded by a transmission encoding method different from the recording encoding method. You may do. In the latter case, it is preferable to interpose an encoding unit (not shown) that encodes a moving image with an encoding method for transmission between the decoding unit PROD_D2 and the transmission unit PROD_D5.
 このような再生装置PROD_Dとしては、例えば、DVDプレイヤ、BDプレイヤ、HDDプレイヤなどが挙げられる(この場合、テレビジョン受像機等が接続される出力端子PROD_D4が動画像の主な供給先となる)。また、テレビジョン受像機(この場合、ディスプレイPROD_D3が動画像の主な供給先となる)、デジタルサイネージ(電子看板や電子掲示板等とも称され、ディスプレイPROD_D3又は送信部PROD_D5が動画像の主な供給先となる)、デスクトップ型PC(この場合、出力端子PROD_D4又は送信部PROD_D5が動画像の主な供給先となる)、ラップトップ型又はタブレット型PC(この場合、ディスプレイPROD_D3又は送信部PROD_D5が動画像の主な供給先となる)、スマートフォン(この場合、ディスプレイPROD_D3又は送信部PROD_D5が動画像の主な供給先となる)なども、このような再生装置PROD_Dの一例である。
〔結び〕
 最後に、上述した動画像復号装置1および動画像符号化装置2の各ブロックは、集積回路(ICチップ)上に形成された論理回路によってハードウェア的に実現してもよいし、CPU(Central Processing Unit)を用いてソフトウェア的に実現してもよい。
Examples of such a playback device PROD_D include a DVD player, a BD player, and an HDD player (in this case, an output terminal PROD_D4 to which a television receiver or the like is connected is a main supply destination of moving images). . In addition, a television receiver (in this case, the display PROD_D3 is a main supply destination of moving images), a digital signage (also referred to as an electronic signboard or an electronic bulletin board), and the display PROD_D3 or the transmission unit PROD_D5 is the main supply of moving images. Desktop PC (in this case, the output terminal PROD_D4 or the transmission unit PROD_D5 is the main video image supply destination), laptop or tablet PC (in this case, the display PROD_D3 or the transmission unit PROD_D5 is a moving image) A smartphone (which is a main image supply destination), a smartphone (in this case, the display PROD_D3 or the transmission unit PROD_D5 is a main moving image supply destination), and the like are also examples of such a playback device PROD_D.
[Conclusion]
Finally, each block of the moving picture decoding apparatus 1 and the moving picture encoding apparatus 2 described above may be realized in hardware by a logic circuit formed on an integrated circuit (IC chip), or may be a CPU (Central It may be realized by software using a Processing Unit).
 後者の場合、上記各装置は、各機能を実現するプログラムの命令を実行するCPU、上記プログラムを格納したROM(Read Only Memory)、上記プログラムを展開するRAM(Random Access Memory)、上記プログラムおよび各種データを格納するメモリ等の記憶装置(記録媒体)などを備えている。そして、本発明の目的は、上述した機能を実現するソフトウェアである上記各装置の制御プログラムのプログラムコード(実行形式プログラム、中間コードプログラム、ソースプログラム)をコンピュータで読み取り可能に記録した記録媒体を、上記各装置に供給し、そのコンピュータ(またはCPUやMPU)が記録媒体に記録されているプログラムコードを読み出し実行することによっても、達成可能である。 In the latter case, each device includes a CPU that executes instructions of a program that realizes each function, a ROM (Read (Memory) that stores the program, a RAM (Random Memory) that expands the program, the program, and various types A storage device (recording medium) such as a memory for storing data is provided. An object of the present invention is to provide a recording medium in which a program code (execution format program, intermediate code program, source program) of a control program of each of the above devices, which is software that realizes the above-described functions, is recorded so as to be readable by a computer. This can also be achieved by supplying to each of the above devices and reading and executing the program code recorded on the recording medium by the computer (or CPU or MPU).
 上記記録媒体としては、例えば、磁気テープやカセットテープ等のテープ類、フロッピー(登録商標)ディスク/ハードディスク等の磁気ディスクやCD-ROM/MO/MD/DVD/CD-R/ブルーレイディスク(登録商標)等の光ディスクを含むディスク類、ICカード(メモリカードを含む)/光カード等のカード類、マスクROM/EPROM/EEPROM/フラッシュROM等の半導体メモリ類、あるいはPLD(Programmable logic device)やFPGA(Field Programmable Gate Array)等の論理回路類などを用いることができる。 Examples of the recording medium include tapes such as magnetic tape and cassette tape, magnetic disks such as floppy (registered trademark) disks / hard disks, and CD-ROM / MO / MD / DVD / CD-R / Blu-ray disks (registered trademarks). ) And other optical disks, IC cards (including memory cards) / optical cards, semiconductor memories such as mask ROM / EPROM / EEPROM / flash ROM, PLD (Programmable logic device) and FPGA ( Logic circuits such as Field Programmable Gate Array can be used.
 また、上記各装置を通信ネットワークと接続可能に構成し、上記プログラムコードを通信ネットワークを介して供給してもよい。この通信ネットワークは、プログラムコードを伝送可能であればよく、特に限定されない。例えば、インターネット、イントラネット、エキストラネット、LAN、ISDN、VAN、CATV通信網、仮想専用網(Virtual Private Network)、電話回線網、移動体通信網、衛星通信網等が利用可能である。また、この通信ネットワークを構成する伝送媒体も、プログラムコードを伝送可能な媒体であればよく、特定の構成または種類のものに限定されない。例えば、IEEE1394、USB、電力線搬送、ケーブルTV回線、電話線、ADSL(Asymmetric Digital Subscriber Line)回線等の有線でも、IrDAやリモコンのような赤外線、Bluetooth(登録商標)、IEEE802.11無線、HDR(High Data Rate)、NFC(Near Field Communication)、DLNA(Digital Living Network Alliance)、携帯電話網、衛星回線、地上波デジタル網等の無線でも利用可能である。なお、本発明は、上記プログラムコードが電子的な伝送で具現化された、搬送波に埋め込まれたコンピュータデータ信号の形態でも実現され得る。 Also, each of the above devices may be configured to be connectable to a communication network, and the program code may be supplied via the communication network. The communication network is not particularly limited as long as it can transmit the program code. For example, the Internet, intranet, extranet, LAN, ISDN, VAN, CATV communication network, virtual private network (Virtual Private Network), telephone line network, mobile communication network, satellite communication network, etc. can be used. The transmission medium constituting the communication network may be any medium that can transmit the program code, and is not limited to a specific configuration or type. For example, even in the case of wired lines such as IEEE 1394, USB, power line carrier, cable TV line, telephone line, ADSL (Asymmetric Digital Subscriber Line) line, infrared rays such as IrDA and remote control, Bluetooth (registered trademark), IEEE 802.11 wireless, HDR ( It can also be used by wireless such as High Data Rate, NFC (Near Field Communication), DLNA (Digital Living Network Alliance), mobile phone network, satellite line, and terrestrial digital network. The present invention can also be realized in the form of a computer data signal embedded in a carrier wave in which the program code is embodied by electronic transmission.
 本発明は上述した実施形態に限定されるものではなく、請求項に示した範囲で種々の変更が可能である。すなわち、請求項に示した範囲で適宜変更した技術的手段を組み合わせて得られる実施形態についても本発明の技術的範囲に含まれる。 The present invention is not limited to the above-described embodiment, and various modifications can be made within the scope indicated in the claims. That is, embodiments obtained by combining technical means appropriately modified within the scope of the claims are also included in the technical scope of the present invention.
 例えば、上述の本実施形態に係る画像復号装置は、符号化データから動画像を復号するものとしたが、動画像を対象とするか静止画像を対象とするかを問わず、画像復号装置一般に適用できるものである。画像符号化装置についても同様である。 For example, the image decoding device according to the present embodiment described above decodes a moving image from encoded data. However, regardless of whether the image is a moving image or a still image, the image decoding device generally Applicable. The same applies to the image encoding device.
  (付記事項1)
 なお、上記実施形態における、LCU(最大コーディングユニット:Largest Coding Unit)は、H.264/MPEG-4 AVCの後継規格として提案されているHEVC(High Efficiency Video Coding)の符号化ツリー(Coding Tree)のrootに相当し、リーフCUは、HEVCのCU(コーディングユニット:Coding Unit、コーディングツリーのleafと呼ばれることもある)に相当する。また、上記実施形態における、PUおよびTUは、それぞれ、HEVCにおける、予測ツリー(Prediction Tree)および変換ツリー(transform tree)に相当する。また、上記実施形態における、PUのパーティションは、HEVCにおけるPU(Prediction Unit)に相当する。また、上記実施形態において、TUを分割して得られるブロックは、HEVCにおけるTU(Transformation Unit)に相当する。
(Appendix 1)
In the above embodiment, LCU (Largest Coding Unit) is H.264. It corresponds to the root of a coding tree (Coding Tree) of HEVC (High Efficiency Video Coding) proposed as a successor to H.264 / MPEG-4 AVC, and a leaf CU is a CU (Coding Unit, coding) It is also called the leaf of the tree). Moreover, PU and TU in the said embodiment are respectively equivalent to the prediction tree (Prediction Tree) and transformation tree (transform tree) in HEVC. Moreover, the partition of PU in the said embodiment is corresponded to PU (Prediction Unit) in HEVC. In the above embodiment, a block obtained by dividing a TU corresponds to a TU (Transformation Unit) in HEVC.
  (付記事項2)
 また、以上に示したように、本発明の一側面に係る動画像符号化装置2及び動画像復号装置1によれば、局所復号画像に含まれる各画素の各成分にばらつきがあり、線形的な相関に従って、各チャンネル間の予測を行うと、予測の精度が低下してしまうような場合であっても、より高い予測精度が得られる可能性を向上させる構成である。
(Appendix 2)
Further, as described above, according to the moving image encoding device 2 and the moving image decoding device 1 according to one aspect of the present invention, each component of each pixel included in the locally decoded image has a variation and is linear. This is a configuration that improves the possibility that higher prediction accuracy can be obtained even if prediction between the channels is performed according to the correlation, even if the prediction accuracy is reduced.
 すなわち、予測画像生成部13において、対象パーティションの周辺に位置する局所復号画像P’を参照して、復号済みの輝度(Y)チャンネルと復号対象の色差(U,V)チャンネルとの間の非線形的な相関をLUTとして導出するLUT導出部134と、対象パーティションについて、LUT導出部134が導出したLUTに従って、輝度復号画像Pから、色差予測画像PredCを生成するチャンネル間予測部351と、を備える。 That is, the prediction image generation unit 13 refers to the local decoded image P ′ positioned around the target partition, and performs nonlinearity between the decoded luminance (Y) channel and the color difference (U, V) channel to be decoded. a LUT derivation unit 134 for deriving a correlation as LUT, the target partition, according to the LUT LUT deriving unit 134 derives, from the luminance decoded image P Y, and inter-channel prediction unit 351 for generating a color difference prediction image PredC, the Prepare.
  (付記事項3)
 以上に示したように、本発明に係る画像復号装置は、画像を構成する各成分を示す複数のチャンネルそれぞれについての予測画像を生成して、生成した上記予測画像に予測残差を加算することにより符号化された画像データを復号する画像復号装置において、処理対象ブロックについて、上記複数のチャンネルのうち、1または複数のチャンネルについて復号するチャンネル復号手段と、上記複数のチャンネルそれぞれについて復号済みである、上記処理対象ブロックの周辺に位置する局所復号画像を参照して、上記チャンネル復号手段により復号済みの上記1または複数のチャンネルと復号対象の他のチャンネルとの間の非線形的な相関を導出する相関導出手段と、処理対象ブロックについて、上記導出された相関に従って、復号済みの上記1または複数のチャンネルの復号画像から、上記他のチャンネルの上記予測画像を生成する予測画像生成手段と、を備える構成である。
(Appendix 3)
As described above, the image decoding apparatus according to the present invention generates a prediction image for each of a plurality of channels indicating each component constituting an image, and adds a prediction residual to the generated prediction image. In the image decoding device that decodes the image data encoded according to the above, for each block to be processed, channel decoding means for decoding one or more channels among the plurality of channels, and each of the plurality of channels has been decoded. Then, with reference to the local decoded image located around the block to be processed, a non-linear correlation between the one or more channels already decoded by the channel decoding means and the other channels to be decoded is derived. For the correlation deriving means and the processing target block, the decoded block is added according to the derived correlation. From one or more channels of the decoded image, which is configured to include, the predicted image generating means for generating the prediction image of the other channels.
 上記構成において、チャンネルとは、画像を構成する成分を一般化したものである。例えば、YUV色空間においては、輝度成分および色差成分がチャンネルに該当する。すなわち、この例の場合、チャンネルには、輝度チャンネル、色差チャンネルが含まれる。また、色差チャンネルには、色差のU成分を示すUチャンネル、および、色差のV成分を示すVチャンネルが含まれる。なお、チャンネルはRGB色空間に関するものであってもよい。画像の復号処理は、この各チャンネルに対して行われる。 In the above configuration, a channel is a generalized component that constitutes an image. For example, in the YUV color space, the luminance component and the color difference component correspond to channels. That is, in this example, the channel includes a luminance channel and a color difference channel. The color difference channel includes a U channel indicating the U component of the color difference and a V channel indicating the V component of the color difference. The channel may relate to the RGB color space. The image decoding process is performed for each channel.
 また、局所復号画像とは、上記複数のチャンネルについて復号がなされたものであり、なおかつ処理対象ブロックの周辺に位置する復号画像のことである。 Also, the local decoded image is a decoded image that has been decoded for the plurality of channels and is located around the block to be processed.
 処理対象ブロックとは、復号処理における各種の処理単位のことである。例えば、符号化単位、変換単位および予測単位などが挙げられる。なお、処理単位には、符号化単位、変換単位および予測単位をさらに細かく細分化した単位も含まれる。 Processing target block refers to various processing units in the decoding process. For example, a coding unit, a conversion unit, a prediction unit, and the like can be given. Note that the processing unit includes a unit obtained by further subdividing the encoding unit, the conversion unit, and the prediction unit.
 処理対象ブロックの周辺とは、例えば、対象ブロックに隣接する画素や、対象ブロックの左側に隣接するブロックや、対象ブロックの上側に隣接するブロックなどが挙げられる。 The periphery of the processing target block includes, for example, a pixel adjacent to the target block, a block adjacent to the left side of the target block, a block adjacent to the upper side of the target block, and the like.
 また、局所的な空間においては、輝度チャンネルと色差チャンネルとの間で相関があることが知られている。そこで、上記構成では、上記処理対象ブロックの周辺に位置する局所復号画像をサンプルとして参照することで、上記チャンネル復号手段により復号済みの上記1または複数のチャンネルと復号対象の他のチャンネルとの間の非線形的な相関を導出する。 Also, in a local space, it is known that there is a correlation between the luminance channel and the color difference channel. Therefore, in the above configuration, by referring to the locally decoded image located around the processing target block as a sample, between the one or more channels decoded by the channel decoding means and the other channels to be decoded. The nonlinear correlation of is derived.
 ここで非線形的な相関は、例えば、輝度値および色差値からなる各点の対応を調べることにより導出することができる。 Here, the non-linear correlation can be derived, for example, by examining the correspondence of each point composed of the luminance value and the color difference value.
 非線形的な相関としては、例えば、YUV色空間について説明すると、局所復号画像に含まれる各画素の輝度値と、色差値とを対応付けたものから導出することができる。また、上記相関は、復号済みのチャンネルと復号対象のチャンネルとが対応付けられたLUTとして実現されていてもよい。また、上記相関は、復号済みのチャンネルと復号対象のチャンネルと間に成立する関係式からなる関数によって表現されていてもよい。 As the non-linear correlation, for example, when describing the YUV color space, it can be derived from the correspondence between the luminance value of each pixel included in the locally decoded image and the color difference value. The correlation may be realized as an LUT in which a decoded channel and a decoding target channel are associated with each other. The correlation may be expressed by a function including a relational expression established between the decoded channel and the decoding target channel.
 上記構成によれば、このようにして導出した非線形的な相関に従って、処理対象ブロックにて復号済みのチャンネルから、処理対象ブロックにて復号対象のチャンネルを予測する。以下、このような予測をチャンネル間予測ともいう。 According to the above configuration, the channel to be decoded in the processing target block is predicted from the channel that has been decoded in the processing target block according to the nonlinear correlation derived in this way. Hereinafter, such prediction is also referred to as inter-channel prediction.
 チャンネル間予測では、例えば、上記非線形的な相関に従って、復号済みのチャンネルの復号画像の画素値を変換して、復号対象のチャンネルの予測画像の画素値を得る。ここで、画素値とは、画像を構成する成分の値を一般化したものである。 In the inter-channel prediction, for example, the pixel value of the decoded image of the decoded channel is converted according to the nonlinear correlation, and the pixel value of the predicted image of the channel to be decoded is obtained. Here, the pixel value is a generalized value of a component constituting the image.
 このため、局所復号画像に含まれる各画素の各成分にばらつきがあり、線形的な相関に従って、各チャンネル間の予測を行うと、予測の精度が低下してしまうような場合であっても、より高い予測精度が得られる可能性を向上させることができるという効果を奏する。 For this reason, there is variation in each component of each pixel included in the locally decoded image, and even when prediction between each channel is performed according to a linear correlation, the accuracy of the prediction is reduced. There is an effect that the possibility that higher prediction accuracy can be obtained can be improved.
 本発明に係る画像復号装置では、上記相関導出手段は、上記復号済みのチャンネルの復号画像の画素値が、上記局所復号画像に含まれる画素の対応する画素値として存在しないとき、復号済みのチャンネルの復号画像の画素値の所定範囲内にある画素値を有する上記局所復号画像に含まれる画素の画素値を用いて補間することにより上記非線形的な相関を導出することが好ましい。 In the image decoding device according to the present invention, the correlation deriving unit is configured to detect a decoded channel when a pixel value of the decoded image of the decoded channel does not exist as a corresponding pixel value of a pixel included in the local decoded image. It is preferable to derive the nonlinear correlation by performing interpolation using pixel values of pixels included in the local decoded image having pixel values within a predetermined range of pixel values of the decoded image.
 上記構成において、画像の画素値とは、画像を構成するいずれかの成分の値のことである。 In the above configuration, the pixel value of the image is the value of any component that forms the image.
 また、上記復号済みのチャンネルの復号画像の画素値が、上記局所復号画像に含まれる画素の対応する画素値として存在しないときとは、上記の輝度値の例でいえば、上記予測の元となる復号済みの輝度チャンネルの復号画像の画素値が、局所復号画像に含まれる画素の輝度値として現れないときのことをいう。 In addition, when the pixel value of the decoded image of the decoded channel does not exist as the corresponding pixel value of the pixel included in the local decoded image, in the example of the luminance value, the prediction source The pixel value of the decoded image of the decoded luminance channel that does not appear as the luminance value of the pixel included in the local decoded image.
 なお、上記非線形的な相関は、予め導出しておいてもよいし、あるいは、上記復号済みのチャンネルの復号画像の画素値が、上記局所復号画像に含まれる画素の対応する画素値として存在しないと判明した段階において導出してもよい。 The nonlinear correlation may be derived in advance, or the pixel value of the decoded image of the decoded channel does not exist as the corresponding pixel value of the pixel included in the local decoded image. You may derive | lead-out in the stage which became clear.
 上記構成によれば、局所復号画像に含まれる画素の画素値として現れないような値について、当該値の近くにある前後のサンプルから線形補間によって相関を得ることができる。例えば、輝度値および色差値からなる各サンプル点について、隣り合う点どうしを線形補間することにより相関を導出することができる。また、非線形的な相関の別の例としては、各点を、3次補間により近似することにより導出することができる。 According to the above configuration, for a value that does not appear as a pixel value of a pixel included in the locally decoded image, a correlation can be obtained by linear interpolation from previous and subsequent samples near the value. For example, for each sample point consisting of a luminance value and a color difference value, the correlation can be derived by linearly interpolating adjacent points. As another example of nonlinear correlation, each point can be derived by approximating each point by cubic interpolation.
 このように局所復号画像のサンプルから得られた画素値を用いた補間により相関を導出するので、局所復号画像に含まれる画素の画素値として現れないような値について、復号済みのチャンネルについての値から、復号対象のチャンネルについての値を、精度よく予測することができる。 Since the correlation is derived by interpolation using the pixel value obtained from the sample of the locally decoded image in this way, the value for the decoded channel is the value that does not appear as the pixel value of the pixel included in the locally decoded image. Thus, the value for the decoding target channel can be predicted with high accuracy.
 本発明に係る画像復号装置では、上記相関導出手段は、複数の復号済みのチャンネルと、復号対象のチャンネルとの関係を、相関として導出することが好ましい。 In the image decoding apparatus according to the present invention, it is preferable that the correlation deriving unit derives a relationship between a plurality of decoded channels and a decoding target channel as a correlation.
 例えば、具体的に、YUV色空間についていえば、上記構成により、復号済みの輝度チャンネルおよびUチャンネルから、復号対象のVチャンネルを予測することができる。 For example, specifically, regarding the YUV color space, the V channel to be decoded can be predicted from the decoded luminance channel and U channel by the above configuration.
 上記構成によれば、複数の復号済みのチャンネルと復号対象のチャンネルとの関係を相関として用いて、チャンネル間の予測を行うので、予測の精度の向上を図ることができる。 According to the above configuration, since the prediction between channels is performed using the relationship between a plurality of decoded channels and the channel to be decoded as a correlation, the prediction accuracy can be improved.
 本発明に係る画像復号装置では、上記相関導出手段は、復号済みのチャンネルの局所復号画像に含まれる複数の画素値と、復号対象のチャンネルとの間の相関を導出することが好ましい。 In the image decoding apparatus according to the present invention, it is preferable that the correlation deriving unit derives a correlation between a plurality of pixel values included in a locally decoded image of a decoded channel and a channel to be decoded.
 例えば、具体的に、YUV色空間についていえば、上記構成により、復号済みの輝度チャンネルの局所復号画像に含まれる複数の輝度値と、復号対象の色差との間の相関を導出する。この場合、この複数の輝度値は、所定範囲内にある輝度値であることが望ましい。また、例えば、隣接する輝度値についての非線形的な相関を導出することが望ましい。これは、複数の輝度値から、予測を行うことでノイズの影響が平均化されるからである。 For example, specifically for the YUV color space, the above configuration derives a correlation between a plurality of luminance values included in the locally decoded image of the decoded luminance channel and the color difference to be decoded. In this case, it is desirable that the plurality of luminance values are luminance values within a predetermined range. For example, it is desirable to derive a non-linear correlation between adjacent luminance values. This is because the influence of noise is averaged by performing prediction from a plurality of luminance values.
 上記構成によれば、復号済みのチャンネルの複数の画素値と、復号対象のチャンネルとの間の相関を用いて、チャンネル間の予測を行うので、予測の精度の向上を図ることができる。 According to the above configuration, since prediction between channels is performed using correlation between a plurality of pixel values of a decoded channel and a channel to be decoded, prediction accuracy can be improved.
 本発明に係る画像復号装置では、上記複数のチャンネルをいずれの順序で復号対象とするかを示すチャンネル復号処理順序情報と、復号対象とするチャンネルを復号済みのチャンネルのいずれから予測するべきかを指定する予測元チャンネル情報とを取得する処理情報取得手段と、上記チャンネル復号処理順序情報が示す順序にて、上記複数のチャンネルを復号対象とするとともに、該復号対象としたチャンネルを、上記予測元チャンネル情報に指定された復号済みのチャンネルから予測するよう制御する予測制御手段と、を備えることが好ましい。 In the image decoding apparatus according to the present invention, channel decoding processing order information indicating in which order the plurality of channels are to be decoded, and from which of the decoded channels the channel to be decoded should be predicted The processing information acquisition means for acquiring the specified prediction source channel information, and the plurality of channels as decoding targets in the order indicated by the channel decoding processing order information, and the decoding target channels as the prediction source It is preferable to include a prediction control unit that performs control so as to perform prediction from the decoded channel specified in the channel information.
 上記構成によれば、チャンネル復号処理順序情報と、予測元チャンネル情報との指定に基づいてチャンネル間予測の制御を行うことができる。チャンネル復号処理順序情報と、予測元チャンネル情報とは、例えば、符号化された画像データが含まれる符号化データに含まれる。よって、例えば、チャンネル復号処理順序情報と、予測元チャンネル情報とを符号化データに符号化して送信してくる画像符号化装置に対応することができる。 According to the above configuration, inter-channel prediction can be controlled based on designation of channel decoding process order information and prediction source channel information. The channel decoding processing order information and the prediction source channel information are included in encoded data including encoded image data, for example. Therefore, for example, it is possible to cope with an image encoding device that encodes channel decoding process order information and prediction source channel information into encoded data and transmits the encoded data.
 本発明に係る画像復号装置では、上記処理情報取得手段は、復号処理における所定の処理単位で符号化されている上記チャンネル復号処理順序情報と上記予測元チャンネル情報とを取得し、上記予測制御手段は、上記処理情報取得手段が取得した上記チャンネル復号処理順序情報と上記予測元チャンネル情報とにそれぞれ従って、上記制御を行うことが好ましい。 In the image decoding apparatus according to the present invention, the processing information acquisition means acquires the channel decoding processing order information and the prediction source channel information that are encoded in a predetermined processing unit in decoding processing, and the prediction control means Preferably, the control is performed in accordance with the channel decoding processing order information and the prediction source channel information acquired by the processing information acquisition unit.
 上記構成によれば、処理単位に応じて、上記制御を変更することができる。処理単位ごとに、実際に復号対象とする順序や、予測元チャンネルを設定可能となる。 According to the above configuration, the control can be changed according to the processing unit. For each processing unit, it is possible to set the actual decoding target order and prediction source channel.
 本発明に係る画像復号装置では、上記相関導出手段は、ブロックを複数含むブロック群に含まれる上記処理対象ブロックについて、ブロック群に含まれる、処理済のブロックにおいて復号された上記局所復号画像を用いて相関を導出することが好ましい。 In the image decoding device according to the present invention, the correlation deriving means uses, for the processing target block included in the block group including a plurality of blocks, the local decoded image decoded in the processed block included in the block group. It is preferable to derive the correlation.
 上記構成では、ブロック群に含まれる各ブロックの間には、空間的な相関が存在する傾向がある。 In the above configuration, there is a tendency that a spatial correlation exists between each block included in the block group.
 つまり、ここでは、上記の“処理対象ブロックの周辺に位置する”局所復号画像を、“処理対象ブロック群に含まれる、処理済のブロック”において復号された局所復号画像としている。 That is, here, the local decoded image “located around the processing target block” is the local decoded image decoded in “the processed block included in the processing target block group”.
 また、サンプル点が多いほど、上記の相関による予測の精度は高くなる傾向がある。よって、空間的な相関が存在するブロックの間において、なるべく多くのサンプル点を取得できれば、チャンネル間予測の予測精度を向上させることができる。 Also, the more sample points, the higher the accuracy of prediction based on the above correlation. Therefore, if as many sample points as possible can be acquired between blocks having a spatial correlation, the prediction accuracy of inter-channel prediction can be improved.
 上記構成では、処理対象ブロックのチャンネル間予測において、ブロック群に含まれる処理済のブロック、すなわち空間的な相関が存在する可能性の高いブロックにおいて復号された局所復号画像から導出した相関を用いる。これにより、チャンネル間予測の予測精度の向上を図ることができる。 In the above configuration, in the inter-channel prediction of the processing target block, a correlation derived from a locally decoded image decoded in a processed block included in the block group, that is, a block having a high possibility of having a spatial correlation is used. Thereby, the prediction accuracy of inter-channel prediction can be improved.
 以上に示したように、本発明に係る画像符号化装置は、画像を構成する各成分を示す複数のチャンネルそれぞれについて生成した予測画像を原画像から減算して得られる予測残差を符号化することによって符号化データを生成する画像符号化装置において、処理対象ブロックについて、上記複数のチャンネルのうち、1または複数のチャンネルについて復号するチャンネル復号手段と、上記複数のチャンネルそれぞれについて復号済みである、上記処理対象ブロックの周辺に位置する局所復号画像を参照して、上記チャンネル復号手段により復号済みの上記1または複数のチャンネルと復号対象の他のチャンネルとの間の非線形的な相関を導出する相関導出手段と、処理対象ブロックについて、上記導出された相関に従って、復号済みの上記1または複数のチャンネルの復号画像から、上記他のチャンネルの上記予測画像を生成する予測画像生成手段と、を備える構成である。 As described above, the image encoding device according to the present invention encodes the prediction residual obtained by subtracting the prediction image generated for each of a plurality of channels indicating each component constituting the image from the original image. In the image encoding device that generates encoded data by this, for the processing target block, channel decoding means for decoding one or more channels among the plurality of channels, and each of the plurality of channels has been decoded. A correlation for deriving a non-linear correlation between the one or more channels decoded by the channel decoding means and the other channels to be decoded with reference to the locally decoded image located around the block to be processed The derivation means and the processing target block are processed in accordance with the above derived correlation. From one or more channels of the decoded image, which is configured to include, the predicted image generating means for generating the prediction image of the other channels.
 以上に示したように、本発明に係る符号化データのデータ構造は、画像を構成する各成分を示す複数のチャンネルそれぞれについて生成した予測画像を原画像から減算して得られる予測残差を符号化することによって生成された符号化データのデータ構造であって、上記複数のチャンネルそれぞれについての予測画像を生成して、生成した上記予測画像に予測残差を加算することにより符号化された画像データを復号する画像復号装置に、処理対象ブロックについて、上記複数のチャンネルを、いずれの順序で復号するかを示すチャンネル復号処理順序情報と、処理対象ブロックについて、いずれかの復号済みの1または複数の上記チャンネルの復号画像から、上記複数のチャンネルそれぞれについて復号済みである、上記処理対象ブロックの周辺に位置する局所復号画像を参照して、上記復号済みの上記1または複数のチャンネルと復号対象の他のチャンネルとの間の非線形的な相関に従って、上記他のチャンネルの上記予測画像を生成するかを指定する予測元チャンネル指定情報と、を含むデータ構造である。 As described above, the data structure of the encoded data according to the present invention encodes the prediction residual obtained by subtracting the prediction image generated for each of a plurality of channels indicating each component constituting the image from the original image. A data structure of encoded data generated by converting to an image encoded by generating a predicted image for each of the plurality of channels and adding a prediction residual to the generated predicted image In an image decoding apparatus that decodes data, channel decoding processing order information that indicates in which order the plurality of channels are decoded for the processing target block, and one or more decoded ones for the processing target block From the decoded image of the channel, the block to be processed has been decoded for each of the plurality of channels. The predicted image of the other channel is generated according to a non-linear correlation between the one or more channels that have been decoded and the other channel to be decoded with reference to a locally decoded image located in the vicinity of The data structure includes prediction source channel designation information for designating whether or not to perform.
 上記のように構成された画像符号化装置または符号化データのデータ構造によれば、本発明に係る画像復号装置と同様の効果を奏する。 According to the image encoding device or the data structure of the encoded data configured as described above, the same effects as those of the image decoding device according to the present invention can be obtained.
 なお、画像符号化装置は、上記複数のチャンネルをいずれの順序で復号対象とするかを示すチャンネル復号処理順序情報と、復号対象とするチャンネルを復号済みのチャンネルのいずれから予測するべきかを指定する予測元チャンネル情報とを符号化データのデータ構造に含めてもよい。画像符号化装置は、上記情報を、例えば、サイド情報において符号化してもよい。 Note that the image encoding device specifies channel decoding processing order information indicating in which order the plurality of channels are to be decoded, and from which of the decoded channels the channel to be decoded is to be predicted The prediction source channel information may be included in the data structure of the encoded data. The image encoding device may encode the information in, for example, side information.
 (付記事項4)
 以上、この発明の各実施形態について図面を参照して詳述してきたが、具体的な構成はこれらの実施形態に限られるものではなく、この発明の要旨を逸脱しない範囲の設計等も請求の範囲に含まれる。
(Appendix 4)
The embodiments of the present invention have been described in detail with reference to the drawings. However, the specific configuration is not limited to these embodiments, and the design and the like within the scope not departing from the gist of the present invention are also claimed. Included in the range.
 本発明は、符号化データを復号する復号装置、および、符号化データを生成する符号化装置に好適に適用することができる。また、符号化装置によって生成され、復号装置によって参照される符号化データのデータ構造に好適に適用することができる。 The present invention can be suitably applied to a decoding device that decodes encoded data and an encoding device that generates encoded data. Further, the present invention can be suitably applied to the data structure of encoded data generated by the encoding device and referenced by the decoding device.
  1 動画像復号装置(画像復号装置)
  2 動画像符号化装置(画像符号化装置)
 13 予測画像生成部(チャンネル復号手段)
 14 加算器(チャンネル復号手段)
 16 LUT導出部(相関導出手段)
134 LUT導出部(相関導出手段)
135 色差予測画像生成部(予測画像生成手段、処理情報取得手段、予測制御手段)
351 チャンネル間予測部(予測画像生成手段)
 PP イントラ予測パラメータ(符号化データのデータ構造)
1 video decoding device (image decoding device)
2 Video encoding device (image encoding device)
13 Predictive image generation unit (channel decoding means)
14 Adder (Channel decoding means)
16 LUT derivation unit (correlation derivation means)
134 LUT derivation unit (correlation derivation means)
135 Color difference prediction image generation unit (prediction image generation means, processing information acquisition means, prediction control means)
351 Inter-channel prediction unit (predicted image generation means)
PP intra prediction parameter (data structure of encoded data)

Claims (9)

  1.  画像を構成する各成分を示す複数のチャンネルそれぞれについての予測画像を生成して、生成した上記予測画像に予測残差を加算することにより符号化された画像データを復号する画像復号装置において、
     処理対象ブロックについて、上記複数のチャンネルのうち、1または複数のチャンネルについて復号するチャンネル復号手段と、
     上記複数のチャンネルそれぞれについて復号済みである、上記処理対象ブロックの周辺に位置する局所復号画像を参照して、上記チャンネル復号手段により復号済みの上記1または複数のチャンネルと復号対象の他のチャンネルとの間の非線形的な相関を導出する相関導出手段と、
     処理対象ブロックについて、上記導出された相関に従って、復号済みの上記1または複数のチャンネルの復号画像から、上記他のチャンネルの上記予測画像を生成する予測画像生成手段と、を備えることを特徴とする画像復号装置。
    In an image decoding apparatus that decodes image data encoded by generating a prediction image for each of a plurality of channels indicating each component constituting an image and adding a prediction residual to the generated prediction image,
    Channel decoding means for decoding one or a plurality of channels among the plurality of channels for the processing target block;
    With reference to the locally decoded image located around the processing target block that has been decoded for each of the plurality of channels, the one or more channels that have been decoded by the channel decoding means and the other channels to be decoded Correlation derivation means for deriving a nonlinear correlation between
    And a predicted image generating unit configured to generate the predicted image of the other channel from the decoded image of the one or more channels that has been decoded in accordance with the derived correlation. Image decoding device.
  2.  上記相関導出手段は、上記復号済みのチャンネルの復号画像の画素値が、上記局所復号画像に含まれる画素の対応する画素値として存在しないとき、復号済みのチャンネルの復号画像の画素値の所定範囲内にある画素値を有する上記局所復号画像に含まれる画素の画素値を用いて補間することにより上記非線形的な相関を導出することを特徴とする請求項1に記載の画像復号装置。 The correlation deriving means, when a pixel value of the decoded image of the decoded channel does not exist as a pixel value corresponding to a pixel included in the local decoded image, a predetermined range of pixel values of the decoded image of the decoded channel The image decoding apparatus according to claim 1, wherein the nonlinear correlation is derived by interpolation using a pixel value of a pixel included in the local decoded image having a pixel value within.
  3.  上記相関導出手段は、複数の復号済みのチャンネルと、復号対象のチャンネルとの関係を、相関として導出することを特徴とする請求項1または2に記載の画像復号装置。 The image decoding apparatus according to claim 1 or 2, wherein the correlation deriving means derives a relationship between a plurality of decoded channels and a channel to be decoded as a correlation.
  4.  上記相関導出手段は、復号済みのチャンネルの局所復号画像に含まれる複数の画素値と、復号対象のチャンネルとの間の相関を導出することを特徴とする請求項1または2に記載の画像復号装置。 3. The image decoding according to claim 1, wherein the correlation deriving unit derives a correlation between a plurality of pixel values included in a locally decoded image of a decoded channel and a channel to be decoded. apparatus.
  5.  上記複数のチャンネルをいずれの順序で復号対象とするかを示すチャンネル復号処理順序情報と、復号対象とするチャンネルを復号済みのチャンネルのいずれから予測するべきかを指定する予測元チャンネル情報とを取得する処理情報取得手段と、
     上記チャンネル復号処理順序情報が示す順序にて、上記複数のチャンネルを復号対象とするとともに、該復号対象としたチャンネルを、上記予測元チャンネル情報に指定された復号済みのチャンネルから予測するよう制御する予測制御手段と、を備えることを特徴とする請求項1から4のいずれか1項に記載の画像復号装置。
    Obtains channel decoding processing order information indicating in which order the plurality of channels are to be decoded, and prediction source channel information specifying which of the decoded channels is to be predicted as a decoding target channel Processing information acquisition means,
    The plurality of channels are set as decoding targets in the order indicated by the channel decoding processing order information, and the channels set as decoding targets are controlled to be predicted from the decoded channels specified in the prediction source channel information. The image decoding apparatus according to claim 1, further comprising: a prediction control unit.
  6.  上記処理情報取得手段は、復号処理における所定の処理単位で符号化されている上記チャンネル復号処理順序情報と上記予測元チャンネル情報とを取得し、
     上記予測制御手段は、上記処理情報取得手段が取得した上記チャンネル復号処理順序情報と上記予測元チャンネル情報とにそれぞれ従って、上記制御を行うことを特徴とする請求項5に記載の画像復号装置。
    The processing information acquisition means acquires the channel decoding processing order information and the prediction source channel information that are encoded in a predetermined processing unit in decoding processing,
    6. The image decoding apparatus according to claim 5, wherein the prediction control unit performs the control according to the channel decoding processing order information and the prediction source channel information acquired by the processing information acquisition unit.
  7.  上記相関導出手段は、ブロックを複数含むブロック群に含まれる上記処理対象ブロックについて、ブロック群に含まれる、処理済のブロックにおいて復号された上記局所復号画像を用いて相関を導出することを特徴とする請求項1から6のいずれか1項に記載の画像復号装置。 The correlation deriving means derives a correlation for the processing target block included in the block group including a plurality of blocks by using the local decoded image decoded in the processed block included in the block group. The image decoding device according to any one of claims 1 to 6.
  8.  画像を構成する各成分を示す複数のチャンネルそれぞれについて生成した予測画像を原画像から減算して得られる予測残差を符号化することによって符号化データを生成する画像符号化装置において、
     処理対象ブロックについて、上記複数のチャンネルのうち、1または複数のチャンネルについて復号するチャンネル復号手段と、
     上記複数のチャンネルそれぞれについて復号済みである、上記処理対象ブロックの周辺に位置する局所復号画像を参照して、上記チャンネル復号手段により復号済みの上記1または複数のチャンネルと復号対象の他のチャンネルとの間の非線形的な相関を導出する相関導出手段と、
     処理対象ブロックについて、上記導出された相関に従って、復号済みの上記1または複数のチャンネルの復号画像から、上記他のチャンネルの上記予測画像を生成する予測画像生成手段と、を備えることを特徴とする画像符号化装置。
    In an image encoding device that generates encoded data by encoding a prediction residual obtained by subtracting a prediction image generated for each of a plurality of channels indicating each component constituting an image from an original image,
    Channel decoding means for decoding one or a plurality of channels among the plurality of channels for the processing target block;
    With reference to the locally decoded image located around the processing target block that has been decoded for each of the plurality of channels, the one or more channels that have been decoded by the channel decoding means and the other channels to be decoded Correlation derivation means for deriving a nonlinear correlation between
    And a predicted image generating unit configured to generate the predicted image of the other channel from the decoded image of the one or more channels that has been decoded in accordance with the derived correlation. Image encoding device.
  9.  画像を構成する各成分を示す複数のチャンネルそれぞれについて生成した予測画像を原画像から減算して得られる予測残差を符号化することによって生成された符号化データのデータ構造であって、
     上記複数のチャンネルそれぞれについての予測画像を生成して、生成した上記予測画像に予測残差を加算することにより符号化された画像データを復号する画像復号装置に、
     処理対象ブロックについて、上記複数のチャンネルを、いずれの順序で復号するかを示すチャンネル復号処理順序情報と、
     処理対象ブロックについて、いずれかの復号済みの1または複数の上記チャンネルの復号画像から、上記複数のチャンネルそれぞれについて復号済みである、上記処理対象ブロックの周辺に位置する局所復号画像を参照して、上記復号済みの上記1または複数のチャンネルと復号対象の他のチャンネルとの間の非線形的な相関に従って、上記他のチャンネルの上記予測画像を生成するかを指定する予測元チャンネル情報と、を含むことを特徴とする符号化データのデータ構造。
    A data structure of encoded data generated by encoding a prediction residual obtained by subtracting a predicted image generated for each of a plurality of channels indicating each component constituting an image from an original image,
    An image decoding device that generates a prediction image for each of the plurality of channels and decodes encoded image data by adding a prediction residual to the generated prediction image.
    Channel decoding processing order information indicating in which order the plurality of channels are decoded for the processing target block;
    With respect to the processing target block, from one of the decoded images of one or more of the channels described above, with reference to a local decoded image located around the processing target block that has been decoded for each of the plurality of channels, Prediction source channel information that specifies whether to generate the prediction image of the other channel according to a non-linear correlation between the one or more channels that have been decoded and another channel to be decoded. A data structure of encoded data characterized by the above.
PCT/JP2011/078953 2010-12-15 2011-12-14 Image decoding device, image coding device, and data structure of coded data WO2012081636A1 (en)

Applications Claiming Priority (2)

Application Number Priority Date Filing Date Title
JP2010279454 2010-12-15
JP2010-279454 2010-12-15

Publications (1)

Publication Number Publication Date
WO2012081636A1 true WO2012081636A1 (en) 2012-06-21

Family

ID=46244732

Family Applications (1)

Application Number Title Priority Date Filing Date
PCT/JP2011/078953 WO2012081636A1 (en) 2010-12-15 2011-12-14 Image decoding device, image coding device, and data structure of coded data

Country Status (1)

Country Link
WO (1) WO2012081636A1 (en)

Citations (2)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
JPH03182171A (en) * 1989-12-11 1991-08-08 Dainippon Printing Co Ltd Coding and reproducing method or color picture information
JP2009534876A (en) * 2006-03-23 2009-09-24 サムスン エレクトロニクス カンパニー リミテッド Image encoding method and apparatus, decoding method and apparatus

Patent Citations (2)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
JPH03182171A (en) * 1989-12-11 1991-08-08 Dainippon Printing Co Ltd Coding and reproducing method or color picture information
JP2009534876A (en) * 2006-03-23 2009-09-24 サムスン エレクトロニクス カンパニー リミテッド Image encoding method and apparatus, decoding method and apparatus

Non-Patent Citations (1)

* Cited by examiner, † Cited by third party
Title
HARUHISA KATO ET AL.: "Adaptive Inter-channel Prediction for Intra Prediction Error in H.264", THE JOURNAL OF THE INSTITUTE OF IMAGE INFORMATION AND TELEVISION ENGINEERS, THE INSTITUTE OF IMAGE INFORMATION AND TELEVISION ENGINEERS, vol. 64, no. 11, 1 November 2010 (2010-11-01), pages 1711 - 1717 *

Similar Documents

Publication Publication Date Title
JP7200320B2 (en) Image filter device, filter method and moving image decoding device
JP7001768B2 (en) Arithmetic decoding device
US10547861B2 (en) Image decoding device
JP6441418B2 (en) Image decoding apparatus, image decoding method, image coding apparatus, and image coding method
WO2014115283A1 (en) Image decoding device and image encoding device
WO2016203981A1 (en) Image decoding device and image encoding device
US10136161B2 (en) DMM prediction section, image decoding device, and image coding device
WO2017068856A1 (en) Predictive image generation device, image decoding device, and image encoding device
JP5995448B2 (en) Image decoding apparatus and image encoding apparatus
WO2017195532A1 (en) Image decoding device and image encoding device
JP2013236358A (en) Image filter device, image decoding device, image encoding device, and data structure
TW202133613A (en) (無)
WO2013046990A1 (en) Offset decoding apparatus, offset encoding apparatus, image filter apparatus, and data structure
CN114902669A (en) Method and apparatus for encoding/decoding image using color space conversion and method of transmitting bitstream
JP2021535653A (en) Deblocking filter for video coding and processing
JP2013141094A (en) Image decoding device, image encoding device, image filter device, and data structure of encoded data
CN114902665A (en) Image encoding/decoding method and apparatus using adaptive transform and method of transmitting bitstream
WO2012121352A1 (en) Video decoding device, video coding device, and data structure
WO2012090962A1 (en) Image decoding device, image encoding device, data structure of encoded data, arithmetic decoding device, and arithmetic encoding device
WO2012077795A1 (en) Image coding device, image decoding device, and data structure
JP6162289B2 (en) Image decoding apparatus and image decoding method
WO2012081636A1 (en) Image decoding device, image coding device, and data structure of coded data
US11627341B2 (en) Method and device for signaling information relating to slice type in picture header in image/video coding system
WO2014050554A1 (en) Image decoding device and image coding device
JP2013251827A (en) Image filter device, image decoding device, image encoding device, and data structure

Legal Events

Date Code Title Description
121 Ep: the epo has been informed by wipo that ep was designated in this application

Ref document number: 11848672

Country of ref document: EP

Kind code of ref document: A1

NENP Non-entry into the national phase

Ref country code: DE

122 Ep: pct application non-entry in european phase

Ref document number: 11848672

Country of ref document: EP

Kind code of ref document: A1

NENP Non-entry into the national phase

Ref country code: JP