WO2012077719A1 - Image decoding device and image coding device - Google Patents

Image decoding device and image coding device Download PDF

Info

Publication number
WO2012077719A1
WO2012077719A1 PCT/JP2011/078332 JP2011078332W WO2012077719A1 WO 2012077719 A1 WO2012077719 A1 WO 2012077719A1 JP 2011078332 W JP2011078332 W JP 2011078332W WO 2012077719 A1 WO2012077719 A1 WO 2012077719A1
Authority
WO
WIPO (PCT)
Prior art keywords
prediction
unit
smoothing
image
pixel
Prior art date
Application number
PCT/JP2011/078332
Other languages
French (fr)
Japanese (ja)
Inventor
将伸 八杉
Original Assignee
シャープ株式会社
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Application filed by シャープ株式会社 filed Critical シャープ株式会社
Publication of WO2012077719A1 publication Critical patent/WO2012077719A1/en

Links

Images

Classifications

    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04NPICTORIAL COMMUNICATION, e.g. TELEVISION
    • H04N19/00Methods or arrangements for coding, decoding, compressing or decompressing digital video signals
    • H04N19/80Details of filtering operations specially adapted for video compression, e.g. for pixel interpolation
    • H04N19/82Details of filtering operations specially adapted for video compression, e.g. for pixel interpolation involving filtering within a prediction loop
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04NPICTORIAL COMMUNICATION, e.g. TELEVISION
    • H04N19/00Methods or arrangements for coding, decoding, compressing or decompressing digital video signals
    • H04N19/10Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using adaptive coding
    • H04N19/102Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using adaptive coding characterised by the element, parameter or selection affected or controlled by the adaptive coding
    • H04N19/103Selection of coding mode or of prediction mode
    • H04N19/11Selection of coding mode or of prediction mode among a plurality of spatial predictive coding modes
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04NPICTORIAL COMMUNICATION, e.g. TELEVISION
    • H04N19/00Methods or arrangements for coding, decoding, compressing or decompressing digital video signals
    • H04N19/10Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using adaptive coding
    • H04N19/102Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using adaptive coding characterised by the element, parameter or selection affected or controlled by the adaptive coding
    • H04N19/117Filters, e.g. for pre-processing or post-processing
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04NPICTORIAL COMMUNICATION, e.g. TELEVISION
    • H04N19/00Methods or arrangements for coding, decoding, compressing or decompressing digital video signals
    • H04N19/10Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using adaptive coding
    • H04N19/169Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using adaptive coding characterised by the coding unit, i.e. the structural portion or semantic portion of the video signal being the object or the subject of the adaptive coding
    • H04N19/182Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using adaptive coding characterised by the coding unit, i.e. the structural portion or semantic portion of the video signal being the object or the subject of the adaptive coding the unit being a pixel
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04NPICTORIAL COMMUNICATION, e.g. TELEVISION
    • H04N19/00Methods or arrangements for coding, decoding, compressing or decompressing digital video signals
    • H04N19/50Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using predictive coding
    • H04N19/593Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using predictive coding involving spatial prediction techniques
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04NPICTORIAL COMMUNICATION, e.g. TELEVISION
    • H04N19/00Methods or arrangements for coding, decoding, compressing or decompressing digital video signals
    • H04N19/60Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using transform coding
    • H04N19/61Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using transform coding in combination with predictive coding

Definitions

  • the present invention relates to an image decoding apparatus that decodes encoded data, and an image encoding apparatus that generates encoded data.
  • a moving image encoding device that generates encoded data by encoding the moving image, and decoding the encoded data
  • a video decoding device image decoding device that generates a decoded image
  • As a specific moving picture encoding method for example, H.264 is used. H.264 / MPEG-4.
  • AVC Non-Patent Document 1
  • VCEG Video Coding Expert Group
  • KTA Key Integrity Expert Group
  • TMuC Test Model Under Consideration
  • an image (picture) constituting a moving image is a slice obtained by dividing an image, and a coding unit (Coding ⁇ ⁇ ⁇ ⁇ Unit) obtained by dividing the slice.
  • a hierarchical structure consisting of blocks and partitions obtained by dividing an encoding unit, and is usually encoded block by block.
  • a predicted image is usually generated based on a local decoded image obtained by encoding / decoding an input image, and a difference image (“residual” between the predicted image and the input image is generated.
  • difference image or “prediction residual”.
  • inter prediction inter-screen prediction
  • intra-screen prediction intra-screen prediction
  • a prediction image in a prediction target frame is generated for each prediction unit by applying motion compensation using a motion vector to a reference image in a reference frame (decoded image) obtained by decoding the entire frame. Is done.
  • intra prediction predicted images in the frame are sequentially generated based on locally decoded images in the same frame.
  • H. H.264 / MPEG-4 As an example of intra prediction used in AVC, any prediction direction is selected from prediction directions included in a predetermined prediction direction group for each prediction unit (for example, partition), and a locally decoded image is used. And a method of generating a pixel value on the prediction unit by extrapolating the pixel value of the reference pixel in the selected prediction direction (sometimes referred to as “basic prediction”).
  • Non-Patent Document 2 for each prediction unit, the edge direction is calculated based on the pixel values of pixels around the prediction unit, and the pixel value of the reference pixel in the local decoded image is out of the calculated edge direction.
  • a method for generating a pixel value on the prediction unit by inserting (a method called Differential Coding of Intra Modes (DCIM), sometimes called “edge prediction” or “edge-based prediction”) is disclosed. .
  • DCIM Differential Coding of Intra Modes
  • Non-Patent Document 3 H. H.264 / MPEG-4.
  • AVC 8 ⁇ 8 intra-frame (intra) predictive coding before generating pixel values on a prediction unit, pixel values of pixels that are in contact with the upper and left sides of the prediction unit, which are reference pixels, are smoothed ( smoothing). This eliminates block distortion that may occur at the boundary between 4 ⁇ 4 pixel prediction units when two 4 ⁇ 4 pixel prediction units are adjacent to the upper or left side of the 8 ⁇ 8 pixel target prediction unit. can do. Therefore, according to the technique described in Non-Patent Document 3, the pixel value including the block distortion is used as it is for the prediction, thereby preventing the pixel value of the pixel on the prediction unit from containing a large error. Can do.
  • a 3-tap filter whose filter coefficients are (1/4, 1/2, 1/4) in order to remove high-frequency components from neighboring pixels in contact with the upper and left sides of the prediction unit 800.
  • Is applied to perform smoothing For example, when the pixel 800b in FIG. 10 is smoothed, if the pixel values after the pixels 800a, b, and c are encoded and decoded are Q1, Q2, and Q3, the pixel value after the smoothing of the pixel 800b is (Q1 / 4 + Q2 / 2 + Q3 / 4). This is performed for all peripheral pixels in contact with the upper and left sides of the prediction unit. Note that the same smoothing process is performed for the upper left, upper right, and lower right pixels.
  • Non-Patent Document 4 describes a technique for switching ON / OFF of smoothing by adaptively specifying whether smoothing is performed for each prediction unit in TMuC. For example, as shown in FIG. 11, whether to perform smoothing for each of the prediction units 401 to 407 is designated by a flag. In the case shown in FIG. 11, the prediction units 401, 402, and 407 are smoothed, and the prediction units 403 to 406 are not smoothed. In the encoder, whether to perform smoothing is compared with the case where smoothing is performed for each prediction unit and the case where smoothing is not performed, and the one with lower RD (Rate-Distortion) cost is selected. Then, the presence or absence of smoothing is designated by a flag. The decoder switches the smoothing ON / OFF based on the flag.
  • Non-Patent Documents 5 and 6 describe a technique for switching ON / OFF of smoothing according to the size of a prediction unit and a prediction direction (prediction mode) without using a flag in TMuC. .
  • the prediction direction for smoothing is determined in advance for each size of the prediction unit, and smoothing is performed when this is true, and smoothing is not performed when this is not the case.
  • TMuC can take various sizes as a prediction unit, even if the technique disclosed in Non-Patent Document 3 is used, block distortion that may occur at the boundary between prediction units adjacent to the target prediction unit. There is a problem that it cannot be removed properly. Even if the techniques disclosed in Non-Patent Documents 4 to 6 are used, there is a possibility that the edges inherent in the encoding target image may be smoothed, so that the encoding efficiency is not improved. Alternatively, there is a problem that the encoding efficiency is not improved as expected.
  • the present invention has been made in view of the above-described problems, and an object of the present invention is to realize an image decoding apparatus or the like that performs smoothing more appropriately than the conventional technique on reference pixels used for intra prediction. It is in.
  • the image decoding apparatus for each prediction unit, calculates a pixel value of a pixel in the prediction image from a pixel value of a reference pixel at another position in the same screen.
  • An image decoding apparatus that generates a decoded image by adding the generated predicted image to a prediction residual decoded from encoded data, and for each reference pixel adjacent to the reference pixel.
  • a reference pixel smoothing determination unit that determines whether or not to perform smoothing with another pixel, and a reference pixel that the reference pixel smoothing determination unit determines to perform smoothing as the other pixel.
  • Smoothing means that performs smoothing between the reference pixel and the reference pixel that is determined to be smoothed by the reference pixel smoothing determination means among the reference pixels, after the smoothing means smoothes the reference pixel To generate the predicted image Is characterized in that it comprises a, the predicted image generating means for.
  • the prediction unit may be a PU described in the embodiment or a partition obtained by dividing the PU.
  • the image encoding device for each prediction unit, calculates a pixel of a pixel in the prediction image from a pixel value of a reference pixel at another position in the same screen.
  • An image encoding apparatus that generates encoded data by encoding a prediction residual between a generated predicted image and an original image, by predicting a value, and for each of the reference pixels
  • a reference pixel smoothing determining unit that determines whether or not to perform smoothing between the pixel and another adjacent pixel, and the reference pixel determined by the reference pixel smoothing determining unit to perform the smoothing Smoothing means that performs smoothing between the reference pixel and reference pixels that are determined to be smoothed by the reference pixel smoothing determination means among the reference pixels after the smoothing means performs smoothing.
  • Prediction using the reference pixels It is characterized in that it comprises the predicted image generating means for generating an image, a.
  • the prediction unit may be a PU described in the embodiment or a partition obtained by dividing the PU.
  • the image decoding apparatus is a reference pixel smoothing determination unit that determines whether or not to perform smoothing between each reference pixel and another pixel adjacent to the reference pixel.
  • the reference pixel determined to be smoothed is configured to include a predicted image generation unit that generates the predicted image using the reference pixel after smoothing by the smoothing unit.
  • FIG. 1 shows an embodiment of the present invention, is a block diagram illustrating a main configuration of a moving picture decoding apparatus.
  • FIG. FIG. 2A shows a data structure of encoded data referred to by the moving picture decoding apparatus
  • FIG. 2A shows a configuration of a picture layer of the encoded data
  • FIG. FIG. 2C is a diagram illustrating a configuration of a slice layer included in the picture layer
  • FIG. 2C is a diagram illustrating a configuration of each CU configuring the LCU layer included in the slice layer
  • FIG. FIG. 2E is a diagram illustrating a configuration of a leaf CU included in a layer
  • FIG. 2E is a diagram illustrating a configuration of inter prediction information for the leaf CU
  • FIG. 2F is intra prediction information for a leaf CU.
  • FIG. It is a figure for demonstrating the reference image which performs smoothing. It is a flowchart which shows the flow which determines whether smoothing is performed about a reference pixel. It is a flowchart which shows the flow which determines whether smoothing is performed about a reference pixel. It is a flowchart which shows the flow which determines whether smoothing is performed about a reference pixel. It is a flowchart which shows the flow which determines whether smoothing is performed about a reference pixel.
  • FIGS. 7A and 7B are diagrams for explaining the operation of the video decoding device, in which FIG. 7A illustrates a prediction mode referred to by the video decoding device together with a prediction mode index, and FIG. FIG. 4 is a diagram illustrating pixels belonging to a target partition and decoded pixels in the vicinity thereof.
  • FIG. 7A illustrates a prediction mode referred to by the video decoding device together with a prediction mode index
  • FIG. 8A is a diagram for describing intra-prediction image generation processing when the edge-based prediction mode is selected in the video decoding device, and FIG. 8A illustrates a target partition as a partition around the target partition.
  • FIG. 8B is a diagram illustrating parameters for specifying the correction angle together with the predicted direction after correction. It is a block diagram which shows the principal part structure of the moving image encoder which concerns on this invention. It is a figure which shows a prior art and is an explanatory view of the method of smoothing a reference pixel. It is a figure which shows a prior art and shows whether smoothing is performed for every block.
  • FIG. 8B is a block diagram illustrating a configuration of a playback device equipped with a video decoding device.
  • a moving image decoding apparatus (image decoding apparatus) 1 decodes a moving image from encoded data.
  • the moving image encoding apparatus according to the present embodiment generates encoded data by encoding a moving image.
  • the scope of application of the present invention is not limited to this. That is, as will be apparent from the following description, the feature of the present invention lies in intra prediction, and is established without assuming a plurality of frames. That is, the present invention can be applied to a general decoding apparatus and a general encoding apparatus regardless of whether the target is a moving image or a still image.
  • the encoded data # 1 has a hierarchical structure including a sequence layer, a GOP (Group Of Pictures) layer, a picture layer, a slice layer, and a maximum coding unit (LCU) layer.
  • a GOP Group Of Pictures
  • LCU maximum coding unit
  • FIG. 2 shows the hierarchical structure below the picture layer in the encoded data # 1.
  • FIGS. 2A to 2F are a picture layer P, a slice layer S, an LCU layer LCU, a leaf CU included in the LCU (denoted as CUL in FIG. 2D), and inter prediction (inter-screen prediction), respectively. It is a figure which shows the structure of inter prediction information PI_Inter which is the prediction information PI about a partition, and intra prediction information PI_Intra which is the prediction information PI about an intra prediction (prediction in a screen) partition.
  • PI_Inter is the prediction information PI about a partition
  • intra prediction information PI_Intra which is the prediction information PI about an intra prediction (prediction in a screen) partition.
  • the picture layer P is a set of data that is referenced by the video decoding device 1 in order to decode a target picture that is a processing target picture. As shown in FIG. 2A, the picture layer P includes a picture header PH and slice layers S1 to SNs (Ns is the total number of slice layers included in the picture layer P).
  • the picture header PH includes a coding parameter group referred to by the video decoding device 1 in order to determine a decoding method of the target picture.
  • the encoding mode information (entropy_coding_mode_flag) indicating the variable length encoding mode used in encoding by the moving image encoding device 2 is an example of an encoding parameter included in the picture header PH.
  • entropy_coding_mode_flag When entropy_coding_mode_flag is 0, the picture is encoded by CAVLC (Context-based Adaptive Variable Length Coding).
  • entropy_coding_mode_flag When entropy_coding_mode_flag is 1, the picture is encoded by CABAC (Context-based Adaptive Binary Arithmetic Coding). It has become.
  • Each slice layer S included in the picture layer P is a set of data referred to by the video decoding device 1 in order to decode a target slice that is a slice to be processed.
  • the slice layer S includes a slice header SH and LCU layers LCU1 to LCUn (Nc is the total number of LCUs included in the slice S).
  • the slice header SH includes a coding parameter group that the moving image decoding apparatus 1 refers to in order to determine a decoding method of the target slice.
  • Slice type designation information (slice_type) for designating a slice type is an example of an encoding parameter included in the slice header SH.
  • the slice header SH includes a filter parameter FP that is referred to by a loop filter included in the video decoding device 1.
  • slice types that can be specified by the slice type specification information, (1) I slice using only intra prediction at the time of encoding, and (2) P using unidirectional prediction or intra prediction at the time of encoding. Slice, (3) B-slice using unidirectional prediction, bidirectional prediction, or intra prediction at the time of encoding.
  • Each LCU layer LCU included in the slice layer S is a set of data that the video decoding device 1 refers to in order to decode the target LCU that is the processing target LCU.
  • the LCU layer LCU is composed of a plurality of coding units (CU: Coding Units) obtained by hierarchically dividing the LCU into a quadtree.
  • the LCU layer LCU is a coding unit corresponding to the highest level in a hierarchical structure that recursively includes a plurality of CUs.
  • each CU included in the LCU layer LCU has a hierarchical structure that recursively includes a CU header CUH and a plurality of CUs obtained by dividing the CU into quadtrees. is doing.
  • each CU excluding the LCU is half the size of the CU to which the CU directly belongs (that is, the CU one layer higher than the CU), and the size that each CU can take is encoded data # 1.
  • a CU that is not further divided is called a leaf CU.
  • the CU header CUH includes a coding parameter referred to by the video decoding device 1 in order to determine a decoding method of the target CU. Specifically, as shown in FIG. 2C, a CU division flag SP_CU that specifies whether or not the target CU is further divided into four subordinate CUs is included. When the CU division flag SP_CU is 0, that is, when the CU is not further divided, the CU is a leaf CU.
  • a CU (CU leaf) that is not further divided is handled as a prediction unit PU (Prediction Unit) and a transform unit TU (Transform Unit).
  • PU Prediction Unit
  • TU Transform Unit
  • the leaf CU (denoted as CUL in FIG. 2 (d)) includes (1) PU information PUI that is referred to when the moving image decoding apparatus 1 generates a predicted image, and (2) The TU information TUI that is referred to when the residual data is decoded by the moving picture decoding apparatus 1 is included.
  • the skip flag SKIP is a flag indicating whether or not the skip mode is applied to the target PU.
  • the value of the skip flag SKIP is 1, that is, when the skip mode is applied to the target leaf, PU information PUI and TU information TUI in the leaf CU are omitted. Note that the skip flag SKIP is omitted for the I slice.
  • the PU information PUI includes a skip flag SKIP, prediction type information PT, and prediction information PI as shown in FIG.
  • the prediction type information PT is information that specifies whether intra prediction or inter prediction is used as a predicted image generation method for the target leaf CU (target PU).
  • the prediction information PI includes intra prediction information PI_Intra or inter prediction information PI_Inter depending on which prediction method is specified by the prediction type information PT.
  • a PU to which intra prediction is applied is also referred to as an intra PU
  • a PU to which inter prediction is applied is also referred to as an inter PU.
  • the PU information PUI includes information specifying the shape and size of each partition included in the target PU and the position in the target PU.
  • the partition is one or a plurality of non-overlapping areas constituting the target leaf CU, and the generation of the predicted image is performed in units of partitions.
  • the TU information TUI specifies a quantization parameter difference ⁇ qp (tu_qp_delta) that specifies the magnitude of the quantization step, and a division pattern for each block of the target leaf CU (target TU).
  • TU partition information SP_TU and quantized prediction residuals QD1 to QDNT are included.
  • the quantization parameter difference ⁇ qp is a difference qp ⁇ qp ′ between the quantization parameter qp in the target TU and the quantization parameter qp ′ in the TU encoded immediately before the TU.
  • TU partition information SP_TU is information that specifies the shape and size of each block included in the target TU and the position in the target TU.
  • Each TU can be, for example, a size from 64 ⁇ 64 pixels to 2 ⁇ 2 pixels.
  • the block is one or a plurality of non-overlapping areas constituting the target leaf CU, and encoding / decoding of the prediction residual is performed in units of blocks.
  • Each quantized prediction residual QD is encoded data generated by the moving image encoding apparatus 2 performing the following processes 1 to 3 on a target block that is a processing target block.
  • Process 1 DCT transform (Discrete Cosine Transform) is performed on the prediction residual obtained by subtracting the prediction image from the encoding target image.
  • Process 2 The DCT coefficient obtained in Process 1 is quantized.
  • Process 3 The DCT coefficient quantized in Process 2 is variable length encoded.
  • the inter prediction information PI_Inter includes a coding parameter that is referred to when the video decoding device 1 generates an inter prediction image by inter prediction. As shown in FIG. 2 (e), the inter prediction information PI_Inter includes inter PU partition information SP_Inter that specifies a partition pattern for each partition of the target PU, and inter prediction parameters PP_Inter1 to PP_InterNe (Ne for each partition). The total number of inter prediction partitions included in the target PU).
  • the inter-PU partition information SP_Inter is information for designating the shape and size of each inter prediction partition included in the target PU (inter PU) and the position in the target PU.
  • the inter PU is composed of four symmetric splittings of 2N ⁇ 2N pixels, 2N ⁇ N pixels, N ⁇ 2N pixels, and N ⁇ N pixels, and 2N ⁇ nU pixels, 2N ⁇ nD pixels, and nL ⁇ 2N. It is possible to divide into 8 types of partitions in total by four asymmetric splits of pixels and nR ⁇ 2N pixels.
  • the specific value of N is defined by the size of the CU to which the PU belongs, and the specific values of nU, nD, nL, and nR are determined according to the value of N.
  • an inter PU of 128 ⁇ 128 pixels is 128 ⁇ 128 pixels, 128 ⁇ 64 pixels, 64 ⁇ 128 pixels, 64 ⁇ 64 pixels, 128 ⁇ 32 pixels, 128 ⁇ 96 pixels, 32 ⁇ 128 pixels, and 96 ⁇ It is possible to divide into 128-pixel inter prediction partitions.
  • the inter prediction parameter PP_Inter includes a reference image index RI, an estimated motion vector index PMVI, and a motion vector residual MVD.
  • the intra prediction information PI_Intra includes an encoding parameter that is referred to when the video decoding device 1 generates an intra predicted image by intra prediction.
  • the intra prediction information PI_Intra includes intra PU partition information SP_Intra that specifies a partition pattern of the target PU (intra PU) into each partition, and intra prediction parameters PP_Intra1 to PP_IntraNa for each partition. (Na is the total number of intra prediction partitions included in the target PU).
  • the intra-PU partition information SP_Intra is information that specifies the shape and size of each intra-predicted partition included in the target PU, and the position in the target PU.
  • the intra PU split information SP_Intra includes an intra split flag (intra_split_flag) that specifies whether or not the target PU is split into partitions. If the intra partition flag is 1, the target PU is divided symmetrically into four partitions. If the intra partition flag is 0, the target PU is not divided and the target PU itself is one partition.
  • N 2 n , n is an arbitrary integer of 1 or more.
  • a 128 ⁇ 128 pixel intra PU can be divided into 128 ⁇ 128 pixel and 64 ⁇ 64 pixel intra prediction partitions.
  • the intra prediction parameter PP_Intra includes an estimation flag MPM, a residual prediction mode index RIPM, and an additional index AI.
  • the intra prediction parameter PP_Intra is a parameter for designating an intra prediction method (prediction mode) for each partition.
  • the estimation flag MPM is a flag indicating whether or not the prediction mode estimated based on the prediction mode allocated to the peripheral partition of the target partition that is the processing target is the same as the prediction mode for the target partition.
  • examples of partitions around the target partition include a partition adjacent to the upper side of the target partition and a partition adjacent to the left side of the target partition.
  • the residual prediction mode index RIPM is an index included in the intra prediction parameter PP_Intra when the estimated prediction mode and the prediction mode for the target partition are different, and is an index for designating a prediction mode assigned to the target partition. It is.
  • the additional index AI is an index for specifying the intra prediction method for the target partition in more detail when the prediction mode assigned to the target partition is a predetermined prediction mode.
  • the moving picture decoding apparatus 1 includes H.264 as a part thereof. H.264 / MPEG-4.
  • Decoding device including technology adopted in KTA software which is a codec for joint development in AVC and VCEG (Video Coding Expert Group), and technology adopted in TMuC (Test Model under Consideration) software which is a successor codec It is.
  • FIG. 1 is a block diagram showing a configuration of the moving picture decoding apparatus 1.
  • the moving image decoding apparatus 1 includes a variable length code decoding unit 11, a predicted image generation unit 12, an inverse quantization / inverse conversion unit 13, an adder 14, a frame memory 15, and a loop filter 16.
  • the predicted image generation unit 12 includes a motion vector restoration unit 12a, an inter predicted image generation unit 12b, an intra predicted image generation unit (reference pixel smoothing determination unit, smoothing unit, predicted image generation unit). ) 12c and a prediction method determination unit 12d.
  • the moving picture decoding apparatus 1 is an apparatus for generating moving picture # 2 by decoding encoded data # 1.
  • variable-length code decoding unit 11 decodes the prediction parameter PP related to each partition from the encoded data # 1, and supplies the decoded prediction parameter PP to the predicted image generation unit 12. Specifically, for the inter prediction partition, the variable-length code decoding unit 11 receives the inter prediction parameter PP_Inter including the reference image index RI, the estimated motion vector index PMVI, and the motion vector residual MVD from the encoded data # 1. These are decoded and supplied to the motion vector restoration unit 12a.
  • the intra prediction parameter PP_Intra including the estimation flag MPM, the residual index RIPM, and the additional index AI is decoded from the encoded data # 1, and these are supplied to the intra prediction image generation unit 12c.
  • the variable length code decoding unit 11 supplies size designation information for designating the size of the partition to the intra predicted image generation unit 12c (not shown).
  • variable length code decoding unit 11 decodes the prediction type information PT for each partition from the encoded data # 1, and supplies this to the prediction method determination unit 12d. Furthermore, the variable-length code decoding unit 11 decodes the quantization prediction residual QD for each block and the quantization parameter difference ⁇ qp for the TU including the block from the encoded data # 1, and dequantizes and reverses them. This is supplied to the conversion unit 13. Further, the variable length code decoding unit 11 decodes the filter parameter FP from the encoded data # 1 and supplies this to the loop filter 16.
  • CABAC Context-based Adaptive Binary Arithmetic Coding
  • CABAC Context-based Adaptive Binary Arithmetic Coding
  • CALVC Context-based Adaptive Variable Variable Length Coding
  • CABAC is an encoding / decoding scheme that performs adaptive binary arithmetic coding based on context
  • CALVC is an encoding / decoding scheme that uses a set of variable length codes that adaptively switch contexts. It is.
  • CABAC has a larger code amount reduction effect than CAVLC, but also has an aspect of increasing the processing amount.
  • variable length code decoding unit 11 refers to the encoding mode information (entropy_coding_mode_flag) included in the picture header PH of the encoded data # 1 to determine whether the target picture has been encoded by CABAC or by CAVLC. Can be identified. In addition, the variable length code decoding unit 11 decodes the target picture using a decoding method corresponding to the identified encoding method.
  • encoding mode information entropy_coding_mode_flag
  • the predicted image generation unit 12 identifies whether each partition is an inter prediction partition for performing inter prediction or an intra prediction partition for performing intra prediction based on the prediction type information PT for each partition. In the former case, the inter prediction image Pred_Inter is generated, and the generated inter prediction image Pred_Inter is supplied to the adder 14 as the prediction image Pred. In the latter case, the intra prediction image Pred_Intra is generated, The generated intra predicted image Pred_Intra is supplied to the adder 14. Note that, when the skip mode is applied to the processing target PU, the predicted image generation unit 12 omits decoding of other parameters belonging to the PU.
  • the motion vector restoration unit 12a restores the motion vector mv related to each inter prediction partition from the motion vector residual MVD related to that partition and the restored motion vector mv ′ related to another partition. Specifically, (1) the estimated motion vector pmv is derived from the restored motion vector mv ′ according to the estimation method specified by the estimated motion vector index PMVI, and (2) the derived estimated motion vector pmv and the motion vector remaining are derived. The motion vector mv is obtained by adding the difference MVD. It should be noted that the restored motion vector mv ′ relating to other partitions can be read from the frame memory 15. The motion vector restoration unit 12a supplies the restored motion vector mv to the inter predicted image generation unit 12b together with the corresponding reference image index RI.
  • the inter prediction image generation unit 12b generates a motion compensated image mc related to each inter prediction partition by inter-screen prediction. Specifically, using the motion vector mv supplied from the motion vector restoration unit 12a, the motion compensation image mc from the filtered decoded image P_ALF ′ designated by the reference image index RI supplied from the motion vector restoration unit 12a. Is generated.
  • the filtered decoded image P_ALF ′ is an image obtained by performing the filtering process by the loop filter 16 on the decoded image that has already been decoded for the entire frame, and the inter predicted image generation unit 12b. Can read out the pixel value of each pixel constituting the filtered decoded image P_ALF ′ from the frame memory 15.
  • the motion compensated image mc generated by the inter predicted image generation unit 12b is supplied to the prediction method determination unit 12d as an inter predicted image Pred_Inter.
  • the intra predicted image generation unit 12c generates a predicted image Pred_Intra related to each intra prediction partition. Specifically, first, a prediction mode is specified based on the intra prediction parameter PP_Intra supplied from the variable length code decoding unit 11, and the specified prediction mode is assigned to the target partition in, for example, raster scan order. Subsequently, a predicted image Pred_Intra is generated from the (local) decoded image P by intra prediction according to the prediction method indicated by the prediction mode. When performing intra prediction, whether or not smoothing is necessary is determined for the neighboring pixels of the target partition. And about the pixel determined to perform smoothing, after performing smoothing, intra prediction is performed.
  • the reference pixels 301a to 301q used for generating a prediction image of the prediction unit (target partition) 301 it is determined whether or not to smooth the reference pixels 301a to 301q used for generating a prediction image of the prediction unit (target partition) 301.
  • This determination is made, for example, depending on whether the reference pixel is a pixel near the boundary of a block such as PU or CU. Then, smoothing is performed only for pixels near the boundary.
  • the reference pixels 301c, d, e, and f which are pixels within two pixels from the boundary between the adjacent blocks 302 and 303 are smoothed. Since high-frequency distortion tends to occur at the block boundary, if only pixels near the boundary are smoothed, distortion can be reduced and distortion can be prevented from propagating to the predicted image. Therefore, the image quality can be improved.
  • it is not necessary to encode a flag indicating whether smoothing is performed in units of blocks it is possible to improve encoding efficiency.
  • the reference pixels 301l, m, n, and o which are pixels within two pixels from the boundary of the adjacent block are smoothed. .
  • the intra predicted image Pred_Intra generated by the intra predicted image generating unit 12c is supplied to the prediction method determining unit 12d.
  • the intra predicted image generation unit 12c may be configured to generate the predicted image Pred_Intra from the filtered decoded image P_ALF by intra prediction. The smoothing necessity determination process will be described later.
  • the prediction method determination unit 12d determines whether each partition is an inter prediction partition that should perform inter prediction or an intra prediction partition that should perform intra prediction based on the prediction type information PT about the PU to which each partition belongs. To do. In the former case, the inter prediction image Pred_Inter generated by the inter prediction image generation unit 12b is supplied to the adder 14 as the prediction image Pred. In the latter case, the inter prediction image generation unit 12c generates the inter prediction image Pred_Inter. The intra predicted image Pred_Intra that has been processed is supplied to the adder 14 as the predicted image Pred.
  • the inverse quantization / inverse transform unit 13 (1) inversely quantizes the quantized prediction residual QD, (2) performs inverse DCT (Discrete Cosine Transform) transform on the DCT coefficient obtained by the inverse quantization, and (3) The prediction residual D obtained by the inverse DCT transform is supplied to the adder 14.
  • the inverse quantization / inverse transform unit 13 derives the quantization step QP from the quantization parameter difference ⁇ qp supplied from the variable length code decoding unit 11.
  • the generation of the prediction residual D by the inverse quantization / inverse transform unit 13 is performed in units of blocks obtained by dividing TUs or TUs.
  • the adder 14 generates the decoded image P by adding the prediction image Pred supplied from the prediction image generation unit 12 and the prediction residual D supplied from the inverse quantization / inverse conversion unit 13.
  • the generated decoded image P is stored in the frame memory 15.
  • loop filter 16 The loop filter 16 reads the decoded image P from the frame memory 15 and performs block noise reduction processing (deblocking processing) at one or both of the partition boundary and the block boundary of the decoded image P. In addition, the loop filter 16 performs adaptive filter processing using the filter parameter FP decoded from the encoded data # 1 on the decoded image subjected to the block noise reduction processing, and the adaptive filter processing Is output to the frame memory 15 as a filtered decoded image P_ALF.
  • smoothing necessity determination processing by the intra predicted image generation unit 12c smoothing necessity determination processing in the intra predicted image generation unit 12c will be described with reference to FIG.
  • the smoothing necessity determination process is performed on all reference pixels used for generating a predicted image, that is, a pixel in contact with either the upper side or the left side of the target partition, and an upper left pixel of the target partition (the upper left vertex of the target partition This is performed for the pixels to be shared (S31).
  • the pixel exists in the vicinity of the boundary between adjacent blocks in contact with the target partition (S32). Whether or not the boundary is near is determined by whether or not it is within a predetermined distance d th pixels from the boundary.
  • the adjacent block (unit area) may be any of PU (prediction unit), CU (processing unit), TU (transform unit), and the partition obtained by dividing PU and TU Any of the obtained blocks may be used. Further, these may be used in combination. Further, only the boundary where the boundary between the PU and the TU overlaps may be a boundary used for the determination process.
  • the intra predicted image generation unit 12c performs a predicted image generation process.
  • the smoothing necessity determination process is performed for all reference pixels used for generating a predicted image, that is, a pixel in contact with either the upper side or the left side of the target partition, and the upper left pixel ( This is performed for pixels sharing the top left vertex of the target partition (S41).
  • the intra estimated image generation part 12c determines whether the pixel concerned exists in the boundary vicinity of the adjacent blocks which are in contact with the object partition about the pixel used as a decision object (S42). This determination process is performed in the same manner as in step S32 described above.
  • the intra predicted image generation unit 12c determines whether or not the block boundary strength (Bs value: Boundary Strength) of the deblocking process to be performed later is greater than or equal to the threshold for the pixel to be determined (S43).
  • Bs value Boundary Strength
  • the Bs value is set in five stages from 0 to 4. This is set as follows.
  • the intra CU indicates a leaf CU encoded using intra prediction.
  • the intra predicted image generation unit 12c smoothes the reference pixel (S44).
  • the smoothing process is performed by the same method as in step S33 described above.
  • the intra predicted image generation unit 12c performs a predicted image generation process.
  • the intra prediction information PI_Intra includes a flag (smoothing information) indicating whether or not smoothing is performed on the pixels near the boundary, thereby determining whether or not smoothing is performed. You may decide.
  • smoothing necessity determination processing is performed for all reference pixels used for generating a predicted image, that is, a pixel in contact with either the upper side or the left side of the target partition, and the upper left pixel ( This is performed for pixels that share the top left vertex of the target partition (S51). Then, the intra predicted image generation unit 12c determines whether the pixel to be determined is an edge pixel (S52).
  • the determination as to whether or not the pixel is an edge pixel is as follows: 3 ⁇ 3 pixels centered on the pixel (x, y ⁇ 1) It is determined by whether or not the value to which the Sobel filter as shown is applied is greater than or equal to the threshold value.
  • the intra predicted image generation unit 12c determines that the pixel to be determined is an edge pixel (YES in S52)
  • the intra predicted image generation unit 12c smoothes the pixel (S53).
  • the smoothing process is performed by the same method as in step S33 described above.
  • the intra predicted image generation unit 12c performs a predicted image generation process.
  • smoothing may be performed when the value to which the Sobel filter is applied is equal to or greater than the first threshold and equal to or less than the second threshold.
  • the second threshold value can be determined in advance according to the strength of the edge of the encoding target image.
  • the right and left (when touching the upper side of the prediction unit) or the upper and lower (when touching the left side of the prediction unit) of the pixel to be determined You may determine by the difference of a pixel value being more than a threshold value. For example, when determining the pixel (x, y) in contact with the upper side of the prediction unit, the pixel (x, y) that satisfies
  • FIG. 7 is a diagram for explaining the operation of the video decoding device 1, and FIG. 7A is a prediction mode referred to by the video decoding device 1, and includes a plurality of basic prediction modes, and FIG. 7B is a diagram illustrating a prediction mode included in an extended set including one edge-based prediction mode together with a prediction mode index, and FIG. 7B is a diagram illustrating pixels belonging to the target partition and decoded pixels in the vicinity thereof. It is.
  • the intra-predicted image generation unit 12c (1) a basic prediction mode that designates one or more predetermined prediction directions and DC prediction, and (2) calculation using pixel values around the target partition Prediction mode that determines the prediction direction by, for example, an edge whose prediction direction is the edge direction calculated from pixel values around the target partition (or the direction represented by the sum of the angle indicated by the edge direction and the correction angle) Based on the prediction mode specified by the intra prediction parameter PP_Intra among the base prediction modes, the intra prediction image Pred_Intra in the target partition is generated.
  • the intra-predicted image generation unit 12c is designated by an intra-prediction parameter PP_Intra from a set of prediction modes (hereinafter also referred to as “extended set”) including one or a plurality of basic prediction modes and an edge-based prediction mode.
  • extended set including one or a plurality of basic prediction modes and an edge-based prediction mode.
  • a prediction mode to be selected is selected, and an intra prediction image Pred_Intra in the target partition is generated based on the selected prediction mode.
  • a set of basic prediction modes that specify one or more predetermined prediction directions and DC prediction is also referred to as a basic prediction mode set. That is, the extended set includes a prediction mode included in the basic prediction mode set and an edge-based prediction mode.
  • FIG. 7A is a diagram showing each prediction mode included in the extended set together with a prediction mode index assigned to each prediction mode.
  • FIG. 7A shows each direction prediction mode belonging to the basic prediction mode set and the prediction direction indicated by each direction prediction mode.
  • the edge-based prediction mode is specified by index 1
  • the DC prediction mode included in the basic prediction mode is specified by index 0
  • each directional prediction mode included in the basic prediction mode set is specified by indices 2-9.
  • the information indicating the correspondence between each index and each prediction mode, and the information indicating the correspondence between each direction prediction mode belonging to the basic prediction mode set and each prediction direction are moving images that generate encoded data # 1.
  • a common configuration can be used for both the image encoding device and the moving image decoding device 1 that decodes the encoded data # 1.
  • the moving picture decoding apparatus 1 stores such information in its own memory, and whether the prediction mode specified by the decoded index is the edge-based prediction mode, the DC prediction mode, or It is possible to identify the direction prediction mode, and when the prediction mode specified by the decoded index is the direction prediction mode, which prediction direction the direction prediction mode specifies. Can be identified.
  • the information indicating the correspondence between each index and each prediction mode, and the information indicating the correspondence between each direction prediction mode belonging to the basic prediction mode set and each prediction direction are, for example, for each sequence, for each picture, or Further, a configuration may be adopted in which, for each slice, transmission is performed from the moving image encoding device to the moving image decoding device 1.
  • the edge-based prediction mode is assigned to the index 1, but the present embodiment is not limited to this, and the characteristics of the decoding target image and the edge-based prediction mode are selected. It is possible to adopt a configuration in which an optimal index is assigned in accordance with the frequency of execution. For example, in the configuration in which the prediction mode specified by a smaller index among the prediction modes assigned to the partitions around the target partition is set as the estimated prediction mode for the target partition, the prediction mode with the smaller index is selected. Will be more frequent. In the case of such a configuration, when the decoding target image includes many edges, it is preferable to assign a smaller index to the edge-based prediction mode. On the other hand, when the prediction mode having a smaller index is selected at a higher frequency and the image to be decoded includes many edges, the edge-based prediction mode is selected. It is preferable to assign a larger index to this.
  • the basic prediction mode set is exemplified as the case where the basic prediction mode set includes a prediction mode that specifies any one of eight different direction predictions.
  • the present embodiment is limited to this. It is not something.
  • a set including a prediction mode that specifies any of nine or more different directions may be used as the basic prediction mode set.
  • a set including a prediction mode for designating any of 16 different directions and a prediction mode for designating any of 32 different directions can be given.
  • a prediction mode included in the basic prediction mode set one of one or more predetermined directions or one or more non-directional prediction modes (for example, DC prediction) is designated.
  • the present embodiment is not limited by the number of prediction modes included in the basic prediction mode set.
  • FIG. 8 is a diagram for explaining the intra-prediction image generation process when the edge-based prediction mode is selected.
  • FIG. 8A illustrates the target partition OP as a partition adjacent to the target partition OP. It is a figure shown with partition NP1 which shares NP2 and NP3, and the upper left vertex of a target partition, and
  • FIG.8 (b) is a figure which shows the parameter which designates a correction angle with the prediction direction after correction
  • FIG. 8A shows a case where the target partition OP and the partitions NP1 to NP3 are all 4 ⁇ 4 pixels, but the present embodiment is not limited to this, and the target partition OP is four.
  • the present invention can also be applied when the size is other than ⁇ 4 pixels or when the partitions NP1 to NP3 are other than 4 ⁇ 4 pixels. It is assumed that the pixel values of the pixels included in the partitions NP1 to NP3 shown in FIG. 8A have been decoded.
  • the calculation of the edge vectors b i may be used Sobel filter Gx shown below, and Gy.
  • the Sobel filters Gx and Gy are filter matrices used for calculating an image gradient along the x direction and an image gradient along the y direction, respectively.
  • the intra predicted image generation unit 12c calculates, as the edge direction, a direction orthogonal to the calculated image gradient in the x direction and the image gradient represented by the image gradient along the y direction.
  • the intra predicted image generation unit 12c has a function T ( ⁇ ) shown below.
  • T ( ⁇ ) ⁇ ⁇ e, b i > 2
  • e represents a unit vector whose angle between its own direction and the horizontal direction (x direction) is ⁇
  • represents the inner product of both vectors.
  • the symbol ⁇ indicates that the subscript i is to be summed from 1 to M.
  • the intra predicted image generation unit 12c sets an argument ⁇ * that maximizes the function T ( ⁇ ) .
  • ⁇ * argmaxS ( ⁇ )
  • the direction represented by ⁇ * is set to the edge direction for the target partition.
  • the intra-predicted image generation unit 12c extrapolates the decoded pixel values for pixels around the target partition in the prediction direction determined as described above, so that An intra prediction image Pred_Intra is generated. Note that if there are decoded pixels on both sides along the prediction direction, the intra predicted image Pred_Intra may be generated by interpolating the pixel values of these pixels.
  • the pixel of the decoded image here is a pixel that has been smoothed if necessary according to the result of the necessity determination process for smoothing.
  • the intra-predicted image generation unit 12c sets the pixel among the decoded pixels that are located on the virtual line segment that faces the reverse direction of the prediction direction, starting from the pixel position of the prediction target pixel in the target partition.
  • the intra predicted image Pred_Intra in the target partition is generated by setting the pixel value of the closest pixel (hereinafter also referred to as the closest pixel) to the pixel value of the prediction target pixel.
  • the pixel value of the prediction target pixel may be a value calculated using the pixel value of the nearest pixel and the pixel values of the pixels around the nearest pixel.
  • the intra predicted image generation unit 12c calculates the edge direction
  • the partition adjacent to the upper side of the target partition, the partition adjacent to the left side of the target partition, and the upper left vertex of the target partition are shared.
  • the pixel values of the pixels belonging to the partition to be referred to is taken as an example, the present embodiment is not limited to this, and the intra-predicted image generation unit 12c is more generally set around the target partition.
  • the edge direction can be calculated with reference to the decoded pixel values belonging to the reference region.
  • the intra predicted image generation unit 12c When the DC prediction mode is selected for the target partition, the intra predicted image generation unit 12c generates an intra predicted image Pred_Intra for the target partition by taking an average value of decoded pixel values around the target partition. .
  • the intra predicted image generation unit 12c decodes the periphery of the target partition along the prediction direction indicated by the selected direction prediction mode.
  • the intra predicted image Pred_Intra for the target partition is generated by extrapolating the completed pixel values. Note that if there are decoded pixels on both sides along the prediction direction, the intra predicted image Pred_Intra may be generated by interpolating the pixel values of these pixels.
  • FIG. 7B is a diagram illustrating each pixel (prediction target pixel) of the target partition, which is 4 ⁇ 4 pixels, and pixels (reference pixels) around the target partition.
  • the prediction target pixels are denoted by reference signs a to p
  • the reference pixels are denoted by reference signs A to M
  • a pixel X (X is any one of a to p and A to M). Let the value be represented as X. Further, it is assumed that the reference pixels A to M have all been decoded.
  • ave Indicates that an element included in parentheses is averaged.
  • “>>” represents a right shift operation, and for any positive integer x, s, the value of x >> s is equal to the value obtained by rounding down the fractional part of x ⁇ (22s).
  • the intra predicted image generation unit 12c can calculate the pixel values a to p by the same method for the basic prediction modes other than the above prediction modes.
  • the intra predicted image generation unit 12c generates an intra predicted image Pred_Intra in the edge-based prediction mode by performing substantially the same process as described above using the prediction direction calculated in the edge-based prediction mode. Can do.
  • the smoothing may be performed when either the target pixel or a reference pixel adjacent to the target pixel is included in the intra CU and the boundary between the two pixels is a processing unit that is not a CU. Thereby, the same effect as the case where it switches whether it smoothes according to whether Bs value is 3 or more can be show
  • the number of reference images that are referenced when inter-predicting a partition that includes the target pixel, and a reference that is referenced when inter-predicting a partition that includes the reference pixel on the opposite side across the boundary with the target pixel Smoothing may be performed when the number of images is different. Further, a reference image that is referred to when inter-predicting a partition including the target pixel, and a reference image that is referred to when inter-predicting a partition including the reference pixel on the opposite side across the boundary with the target pixel If they are different, smoothing may be performed.
  • each pixel that is referred to when generating an intra-predicted image is smoothed by applying a 1: 2: 1 3-tap smoothing filter.
  • a filter having the above tap number may be applied.
  • the block distortion can be reduced with high accuracy.
  • the filter used for the conventional deblocking process is mentioned, for example.
  • the smoothing described in the present embodiment is effective not only in the above-described intra prediction modes but also in the case where an intra predicted image is generated using pixel values of reference pixels.
  • the intra prediction mode is DC prediction
  • the reference pixels are used only as an average value for prediction image generation, and therefore smoothing may be omitted.
  • the moving image encoding apparatus 2 includes H.264 as a part thereof.
  • FIG. 9 is a block diagram showing a configuration of the moving picture encoding apparatus 2.
  • the moving image encoding apparatus 2 includes a predicted image generation unit 21, a transform / quantization unit 22, an inverse quantization / inverse transform unit 23, an adder 24, a frame memory 25, a loop filter 26, a variable A long code encoding unit 27 and a subtracter 28 are provided.
  • the predicted image generation unit 21 includes an intra predicted image generation unit (reference pixel smoothing determination unit, smoothing unit, predicted image generation unit) 21a, a motion vector detection unit 21b, and an inter predicted image generation.
  • the moving image encoding device 2 is a device that generates encoded data # 1 by encoding moving image # 10 (encoding target image).
  • the predicted image generation unit 21 recursively divides the processing target LCU into one or a plurality of lower-order CUs, further divides each leaf CU into one or a plurality of partitions, and uses an inter-screen prediction for each partition.
  • a predicted image Pred_Inter or an intra predicted image Pred_Intra using intra prediction is generated.
  • the generated inter prediction image Pred_Inter and intra prediction image Pred_Intra are supplied to the adder 24 and the subtracter 28 as the prediction image Pred.
  • the prediction image generation unit 21 omits encoding of other parameters belonging to the PU for the PU to which the skip mode is applied. Also, (1) the mode of division into lower CUs and partitions in the target LCU, (2) whether to apply the skip mode, and (3) which of the inter predicted image Pred_Inter and the intra predicted image Pred_Intra for each partition Whether to generate is determined so as to optimize the encoding efficiency.
  • the intra predicted image generation unit 21a generates a predicted image Pred_Intra for each partition by intra prediction. Specifically, (1) a prediction mode used for intra prediction is selected for each partition, and (2) a prediction image Pred_Intra is generated from the decoded image P using the selected prediction mode. The intra predicted image generation unit 21a supplies the generated intra predicted image Pred_Intra to the prediction method control unit 21d.
  • the intra-prediction image generation unit 21a selects any one of the prediction modes included in the above-described basic prediction mode set and the prediction mode included in the extended set including the edge-based prediction mode.
  • the intra prediction image Pred_Intra is generated according to the method indicated by the selected prediction mode. Then, before generating the intra-predicted image Pred_Intra, whether or not smoothing is necessary for the neighboring pixels of the target partition is determined. Then, after smoothing the pixels determined to be smoothed, an intra predicted image Pred_Intra is generated.
  • the smoothing necessity determination process is the same as the process in the intra predicted image generation unit 12c included in the video decoding device 1.
  • the motion vector detection unit 21b detects a motion vector mv regarding each partition. Specifically, (1) the filtered decoded image P_ALF ′ used as the reference image is selected, and (2) the target partition is searched by searching for the region that best approximates the target partition in the selected filtered decoded image P_ALF ′. Detects a motion vector mv.
  • the filtered decoded image P_ALF ′ is an image obtained by performing adaptive filter processing by the loop filter 26 on the decoded image that has already been decoded for the entire frame, and is a motion vector detection unit. 21b can read out the pixel value of each pixel constituting the filtered decoded image P_ALF ′ from the frame memory 25.
  • the motion vector detection unit 21b supplies the detected motion vector mv to the inter prediction image generation unit 21c and the motion vector redundancy deletion unit 21e together with the reference image index RI that specifies the filtered decoded image P_ALF ′ used as the reference image. To do.
  • the inter prediction image generation unit 21c generates a motion compensated image mc related to each inter prediction partition by inter-screen prediction. Specifically, the motion compensation image mc is obtained from the filtered decoded image P_ALF ′ designated by the reference image index RI supplied from the motion vector detection unit 21b, using the motion vector mv supplied from the motion vector detection unit 21b. Generate. Similar to the motion vector detection unit 21b, the inter prediction image generation unit 21c can read out the pixel value of each pixel constituting the filtered decoded image P_ALF ′ from the frame memory 25. The inter prediction image generation unit 21c supplies the generated motion compensated image mc (inter prediction image Pred_Inter) together with the reference image index RI supplied from the motion vector detection unit 21b to the prediction method control unit 21d.
  • inter prediction image Pred_Inter inter prediction image Pred_Inter
  • the prediction scheme control unit 21d compares the intra predicted image Pred_Intra and the inter predicted image Pred_Inter with the encoding target image and selects whether to perform intra prediction or inter prediction.
  • the prediction scheme control unit 21d supplies the intra prediction image Pred_Intra as the prediction image Pred to the adder 24 and the subtracter 28, and also uses the intra prediction parameter PP_Intra supplied from the intra prediction image generation unit 21a. This is supplied to the variable length code encoding unit 27.
  • the prediction scheme control unit 21d supplies the inter prediction image Pred_Inter as the prediction image Pred to the adder 24 and the subtractor 28, and the reference image index RI and motion vector redundancy described later.
  • the estimated motion vector index PMVI and the motion vector residual MVD supplied from the deletion unit 21e are supplied to the variable length code encoding unit 27 as an inter prediction parameter PP_Inter.
  • the motion vector redundancy deletion unit 21e deletes redundancy in the motion vector mv detected by the motion vector detection unit 21b. Specifically, (1) an estimation method used for estimating the motion vector mv is selected, (2) an estimated motion vector pmv is derived according to the selected estimation method, and (3) the estimated motion vector pmv is subtracted from the motion vector mv. As a result, a motion vector residual MVD is generated. The motion vector redundancy deleting unit 21e supplies the generated motion vector residual MVD to the prediction method control unit 21d together with the estimated motion vector index PMVI indicating the selected estimation method.
  • the transform / quantization unit 22 (1) performs DCT transform (Discrete Cosine Transform) on the prediction residual D obtained by subtracting the predicted image Pred from the encoding target image (original image) for each block (transform unit), and (2) The DCT coefficient obtained by the DCT transform is quantized, and (3) the quantized prediction residual QD obtained by the quantization is supplied to the variable length code encoder 27 and the inverse quantization / inverse transform unit 23.
  • the transform / quantization unit 22 (1) selects a quantization step QP to be used for quantization for each TU, and (2) sets a quantization parameter difference ⁇ qp indicating the size of the selected quantization step QP.
  • the difference value obtained by subtracting the value of '.
  • the inverse quantization / inverse transform unit 23 (1) inversely quantizes the quantized prediction residual QD, (2) performs inverse DCT (Discrete Cosine Transform) transformation on the DCT coefficient obtained by the inverse quantization, and (3) The prediction residual D obtained by the inverse DCT transform is supplied to the adder 24.
  • the quantization prediction residual QD is inversely quantized, the quantization step QP supplied from the transform / quantization unit 22 is used.
  • the prediction residual D output from the inverse quantization / inverse transform unit 23 is obtained by adding a quantization error to the prediction residual D input to the transform / quantization unit 22. Common names are used for this purpose.
  • the adder 24 adds the predicted image Pred selected by the prediction scheme control unit 21d to the prediction residual D generated by the inverse quantization / inverse transform unit 23, thereby obtaining the (local) decoded image P. Generate.
  • the (local) decoded image P generated by the adder 24 is supplied to the loop filter 26 and stored in the frame memory 25, and is used as a reference image in intra prediction.
  • variable-length code encoding unit 27 (Variable-length code encoding unit 27) The variable length code encoding unit 27 (1) the quantization prediction residual QD and ⁇ qp supplied from the transform / quantization unit 22, and (2) the quantization parameter PP (inter prediction) supplied from the prediction scheme control unit 21d. The parameter PP_Inter and the intra prediction parameter PP_Intra) and (3) the filter parameter FP supplied from the loop filter 26 are variable-length-encoded to generate encoded data # 1.
  • CABAC Context-based Adaptive Binary Arithmetic Coding
  • CAVLC Context-based Adaptive VLC
  • the variable length code encoding unit 27 determines which encoding method of CABAC or CAVLC is to be used for each picture, performs encoding using the determined encoding method, and specifies the determined encoding method
  • the mode information (entropy_coding_mode_flag) is included in the picture header PH of the encoded data # 1.
  • the subtracter 28 generates the prediction residual D by subtracting the prediction image Pred selected by the prediction method control unit 21d from the encoding target image.
  • the prediction residual D generated by the subtracter 28 is DCT transformed / quantized by the transform / quantization unit 22.
  • the loop filter 26 reads the decoded image P from the frame memory 25 and performs block noise reduction processing (deblocking processing) at one or both of the partition boundary and the block boundary of the decoded image P.
  • the loop filter 26 performs adaptive filter processing using the adaptively calculated filter parameter FP on the decoded image subjected to block noise reduction processing, and the adaptive filter processing is performed.
  • the decoded image P is output to the frame memory 25 as a filtered decoded image P_ALF.
  • the filtered decoded image P_ALF is mainly used as a reference image in the inter predicted image generation unit 21c.
  • the moving picture decoding apparatus 1 and the moving picture encoding apparatus 2 described above can be used by being mounted on various apparatuses that perform moving picture transmission, reception, recording, and reproduction.
  • the moving image may be a natural moving image captured by a camera or the like, or may be an artificial moving image (including CG and GUI) generated by a computer or the like.
  • the moving picture decoding apparatus 1 and the moving picture encoding apparatus 2 described above can be used for transmission and reception of moving pictures.
  • FIG. 12A is a block diagram illustrating a configuration of the transmission apparatus A in which the moving picture encoding apparatus 2 is mounted.
  • the transmitting apparatus A encodes a moving image, obtains encoded data, and modulates a carrier wave with the encoded data obtained by the encoding unit A1.
  • a modulation unit A2 that obtains a modulation signal by the transmission unit A2 and a transmission unit A3 that transmits the modulation signal obtained by the modulation unit A2.
  • the moving image encoding device 2 described above is used as the encoding unit A1.
  • the transmission apparatus A has a camera A4 that captures a moving image, a recording medium A5 that records the moving image, an input terminal A6 for inputting the moving image from the outside, as a supply source of the moving image that is input to the encoding unit A1. You may further provide image processing part A7 which produces
  • FIG. 12A illustrates a configuration in which the transmission apparatus A includes all of these, but some of them may be omitted.
  • the recording medium A5 may be a recording of a non-encoded moving image, or a recording of a moving image encoded using a recording encoding scheme different from the transmission encoding scheme. It may be a thing. In the latter case, a decoding unit (not shown) for decoding the encoded data read from the recording medium A5 according to the recording encoding method may be interposed between the recording medium A5 and the encoding unit A1.
  • FIG. 12B is a block diagram illustrating a configuration of the receiving device B on which the moving image decoding device 1 is mounted.
  • the receiving device B includes a receiving unit B1 that receives a modulated signal, a demodulating unit B2 that obtains encoded data by demodulating the modulated signal received by the receiving unit B1, and a demodulating unit.
  • a decoding unit B3 that obtains a moving image by decoding the encoded data obtained by B2.
  • the moving picture decoding apparatus 1 described above is used as the decoding unit B3.
  • the receiving apparatus B has a display B4 for displaying a moving image, a recording medium B5 for recording the moving image, and an output terminal for outputting the moving image as a supply destination of the moving image output from the decoding unit B3.
  • B6 may be further provided.
  • FIG. 12B illustrates a configuration in which the receiving apparatus B includes all of these, but a part of the configuration may be omitted.
  • the recording medium B5 may be for recording an unencoded moving image, or is encoded by a recording encoding method different from the transmission encoding method. May be.
  • an encoding unit (not shown) that encodes the moving image acquired from the decoding unit B3 in accordance with the recording encoding method may be interposed between the decoding unit B3 and the recording medium B5.
  • the transmission medium for transmitting the modulation signal may be wireless or wired.
  • the transmission mode for transmitting the modulated signal may be broadcasting (here, a transmission mode in which the transmission destination is not specified in advance) or communication (here, transmission in which the transmission destination is specified in advance). Refers to the embodiment). That is, the transmission of the modulation signal may be realized by any of wireless broadcasting, wired broadcasting, wireless communication, and wired communication.
  • a terrestrial digital broadcast broadcasting station (such as broadcasting equipment) / receiving station (such as a television receiver) is an example of a transmitting apparatus A / receiving apparatus B that transmits and receives modulated signals by wireless broadcasting.
  • a broadcasting station (such as broadcasting equipment) / receiving station (such as a television receiver) for cable television broadcasting is an example of a transmitting device A / receiving device B that transmits and receives a modulated signal by cable broadcasting.
  • a server workstation etc.
  • Client television receiver, personal computer, smart phone etc.
  • VOD Video On Demand
  • video sharing service using the Internet is a transmitting device for transmitting and receiving modulated signals by communication.
  • a / reception device B usually, either wireless or wired is used as a transmission medium in a LAN, and wired is used as a transmission medium in a WAN.
  • the personal computer includes a desktop PC, a laptop PC, and a tablet PC.
  • the smartphone also includes a multi-function mobile phone terminal.
  • the video sharing service client has a function of encoding a moving image captured by the camera and uploading it to the server. That is, the client of the video sharing service functions as both the transmission device A and the reception device B.
  • FIG. 13A is a block diagram showing a configuration of a recording apparatus C on which the above-described moving picture decoding apparatus 1 is mounted.
  • the recording device C encodes a moving image to obtain encoded data, and writes the encoded data obtained by the encoding unit C1 to the recording medium M.
  • the moving image encoding device 2 described above is used as the encoding unit C1.
  • the recording medium M may be of a type built in the recording device C, such as (1) HDD (Hard Disk Drive) or SSD (Solid State Drive), or (2) SD memory. It may be of the type connected to the recording device C, such as a card or USB (Universal Serial Bus) flash memory, or (3) DVD (Digital Versatile Disc) or BD (Blu-ray Disc: registration) (Trademark) or the like may be mounted on a drive device (not shown) built in the recording apparatus C.
  • the recording apparatus C also serves as a moving image supply source to be input to the encoding unit C1, a camera C3 that captures moving images, an input terminal C4 for inputting moving images from the outside, and reception for receiving moving images.
  • a unit C5 and an image processing unit C6 that generates or processes an image may be further provided.
  • FIG. 13A illustrates a configuration in which the recording apparatus C includes all of these, but a part of the configuration may be omitted.
  • the receiving unit C5 may receive an unencoded moving image, or receives encoded data encoded by a transmission encoding method different from the recording encoding method. You may do. In the latter case, a transmission decoding unit (not shown) that decodes encoded data encoded by the transmission encoding method may be interposed between the reception unit C5 and the encoding unit C1.
  • Examples of such a recording device C include a DVD recorder, a BD recorder, and an HD (Hard Disk) recorder (in this case, the input terminal C4 or the receiving unit C5 is a main source of moving images).
  • a camcorder in this case, the camera C3 is a main source of moving images
  • a personal computer in this case, the receiving unit C5 or the image processing unit C6 is a main source of moving images
  • a smartphone this In this case, the camera C3 or the receiving unit C5 is a main source of moving images).
  • FIG. 13B is a block diagram showing the configuration of the playback device D equipped with the above-described video decoding device 1.
  • the playback device D obtains a moving image by decoding the read data D1 that reads the encoded data written on the recording medium M and the read data read by the read unit D1.
  • the moving picture decoding apparatus 1 described above is used as the decoding unit D2.
  • the recording medium M may be of a type built in the playback device D such as (1) HDD or SSD, or (2) such as an SD memory card or USB flash memory. It may be of a type connected to the playback device D, or (3) may be loaded into a drive device (not shown) built in the playback device D, such as DVD or BD. Good.
  • the playback device D has a display D3 for displaying a moving image, an output terminal D4 for outputting the moving image to the outside, and a transmitting unit for transmitting the moving image as a supply destination of the moving image output by the decoding unit D2.
  • D5 may be further provided.
  • FIG. 13B illustrates a configuration in which the playback apparatus D includes all of these, but some of them may be omitted.
  • the transmission unit D5 may transmit a non-encoded moving image, or transmits encoded data encoded by a transmission encoding method different from the recording encoding method. You may do. In the latter case, an encoding unit (not shown) that encodes a moving image with a transmission encoding method may be interposed between the decoding unit D2 and the transmission unit D5.
  • Examples of such a playback device D include a DVD player, a BD player, and an HDD player (in this case, an output terminal D4 to which a television receiver or the like is connected is a main moving image supply destination).
  • a television receiver in this case, the display D3 is a main destination of moving images
  • a desktop PC in this case, the output terminal D4 or the transmission unit D5 is a main destination of moving images
  • Laptop type or tablet type PC in this case, display D3 or transmission unit D5 is the main video image supply destination
  • smartphone in this case, display D3 or transmission unit D5 is the main video image supply destination
  • Digital signage also referred to as an electronic signboard or an electronic bulletin board, and the display D3 or the transmission unit D5 is the main supply destination of moving images
  • Digital signage also referred to as an electronic signboard or an electronic bulletin board, and the display D3 or the transmission unit D5 is the main supply destination of moving images
  • each block of the moving picture decoding apparatus 1 and the moving picture encoding apparatus 2 in particular, a variable length code decoding unit 11, a prediction image generation unit (motion vector restoration unit 12 a, inter prediction image generation unit 12 b, intra prediction image generation Unit 12c, prediction method determination unit 12d) 12, inverse quantization / inverse conversion unit 13, adder 14, frame memory 15, loop filter 16, prediction image generation unit (intra prediction image generation unit 21a, motion vector detection unit 21b, Inter prediction image generation unit 21c, prediction method control unit 21d, motion vector redundancy deletion unit 21e) 21, transform / quantization unit 22, inverse quantization / inverse transform unit 23, adder 24, frame memory 25, loop filter 26
  • the variable-length code encoding unit 27 and the subtracter 28 are implemented in hardware by a logic circuit formed on an integrated circuit (IC chip). It may be realized, or may be implemented in software using a CPU (central processing unit).
  • the video decoding device 1 and the video encoding device 2 develop a CPU that executes instructions of a control program that realizes each function, a ROM (read (only memory) that stores the program, and the program.
  • a RAM random access memory
  • a storage device such as a memory for storing the program and various data, and the like are provided.
  • the object of the present invention is to provide a computer program code (execution format program, intermediate code program, source program) of a control program for the video decoding device 1 and the video encoding device 2 that is software that implements the functions described above. Is supplied to the above-described moving picture decoding apparatus 1 and moving picture encoding apparatus 2 and the computer (or CPU or MPU (microprocessor unit)) is recorded on the recording medium. It can also be achieved by reading and executing the code.
  • the recording medium examples include tapes such as a magnetic tape and a cassette tape, a magnetic disk such as a floppy (registered trademark) disk / hard disk, a CD-ROM (compact disk-read-only memory) / MO (magneto-optical) / Discs including optical discs such as MD (Mini Disc) / DVD (digital versatile disc) / CD-R (CD Recordable), IC cards (including memory cards) / optical cards, mask ROM / EPROM (erasable) Programmable read-only memory) / EEPROM (electrically erasable and programmable programmable read-only memory) / semiconductor memory such as flash ROM, or logic circuits such as PLD (Programmable logic device) and FPGA (Field Programmable Gate Array) be able to.
  • a magnetic disk such as a floppy (registered trademark) disk / hard disk
  • the moving picture decoding apparatus 1 and the moving picture encoding apparatus 2 may be configured to be connectable to a communication network, and the program code may be supplied via the communication network.
  • the communication network is not particularly limited as long as it can transmit the program code.
  • Internet intranet, extranet, LAN (local area network), ISDN (integrated area services digital area), VAN (value-added area network), CATV (community area antenna television) communication network, virtual area private network (virtual area private network), A telephone line network, a mobile communication network, a satellite communication network, etc. can be used.
  • the transmission medium constituting the communication network may be any medium that can transmit the program code, and is not limited to a specific configuration or type.
  • IEEE institute of electrical and electronic engineers 1394, USB, power line carrier, cable TV line, telephone line, ADSL (asynchronous digital subscriber loop) line, etc. wired such as IrDA (infrared data association) or remote control , Bluetooth (registered trademark), IEEE802.11 wireless, HDR (high data rate), NFC (Near field communication), DLNA (Digital Living Network Alliance), mobile phone network, satellite line, terrestrial digital network, etc.
  • IrDA infrared data association
  • Bluetooth registered trademark
  • IEEE802.11 wireless wireless
  • HDR high data rate
  • NFC Near field communication
  • DLNA Digital Living Network Alliance
  • mobile phone network satellite line, terrestrial digital network, etc.
  • the present invention can also be realized in the form of a computer data signal embedded in a carrier wave in which the program code is embodied by electronic transmission.
  • the image decoding apparatus generates a prediction image by predicting a pixel value of a pixel in the prediction image from a pixel value of a reference pixel at another position in the same screen for each prediction unit,
  • An image decoding apparatus that generates a decoded image by adding a generated predicted image to a prediction residual decoded from encoded data, and for each reference pixel, between the reference pixel and another adjacent pixel
  • the reference pixel smoothing determining means for determining whether or not to perform smoothing, and the smoothing for smoothing the reference pixel determined to be smoothed by the reference pixel smoothing determining means with the other pixels.
  • the predicted image is generated using the reference pixels smoothed by the smoothing unit.
  • Predicted image generation means for It is characterized in that it comprises.
  • the prediction unit may be a PU described in the embodiment or a partition obtained by dividing the PU.
  • the reference pixel smoothing determining means is configured such that, when the reference pixel to be determined exists in the vicinity of a boundary between a plurality of unit regions in contact with the prediction unit to be processed, the reference pixel It is preferable to determine that smoothing is performed.
  • the unit area may be any of a partition obtained by dividing a CU, a TU, a PU, and a PU described in the embodiment, and a block obtained by dividing a TU.
  • the vicinity of the boundary means a range of a predetermined distance (for example, two pixels) from the boundary.
  • the boundary between unit areas is likely to cause high-frequency distortion. Therefore, according to the above configuration, since it is determined that the reference pixels near the boundary are to be smoothed, this high-frequency distortion can be appropriately reduced.
  • the reference pixel smoothing determining unit when the block boundary strength when a deblocking filter is applied in a later process for a reference pixel to be determined exceeds a threshold, It may be determined that the pixel is smoothed.
  • a pixel to which a deblocking filter having a high block boundary strength is applied is a pixel having a large block distortion. Therefore, it is possible to perform smoothing for a reference pixel having a large block distortion.
  • the reference pixel smoothing determining means includes at least one of a unit region including a reference pixel to be determined and an adjacent unit region in contact with the unit region at the boundary. May be determined to perform the smoothing of the reference pixel when the predicted region is generated by intra prediction.
  • At least one of the unit region including the reference pixel to be determined and the adjacent unit region that is in contact with the unit region at the boundary is near the boundary of the unit region where the prediction image is generated by intra prediction.
  • the reference pixel smoothing determining unit performs inter prediction on both the unit region including the reference pixel to be determined and the adjacent unit region in contact with the unit region at the boundary. If the angle formed by the motion vector assigned to each unit area exceeds a threshold when the predicted image is generated by the unit area, the reference pixel may be determined to be smoothed. .
  • the pixel at the boundary between the prediction units is likely to be distorted. Therefore, according to said structure, the reference pixel considered that distortion is large can be smoothed.
  • the reference pixel smoothing determining unit performs inter prediction on both the unit region including the reference pixel to be determined and the adjacent unit region in contact with the unit region at the boundary. If the difference between the motion vector components assigned to each unit area exceeds a threshold when the predicted image is generated by the unit area, the reference pixel is determined to be smoothed. Good.
  • the pixel at the boundary between the unit areas is likely to be distorted. Therefore, according to said structure, the reference pixel considered that distortion is large can be smoothed.
  • the reference pixel smoothing determining unit performs inter prediction on both the unit region including the reference pixel to be determined and the adjacent unit region in contact with the unit region at the boundary.
  • the unit region in which the predicted image is generated by the above when the number of reference images used for inter prediction in each unit region is different, it may be determined that the reference pixel is smoothed.
  • Pixels at the boundary between unit areas with different numbers of reference pixels used for inter prediction are highly likely to be distorted. Therefore, according to said structure, the reference pixel considered that distortion is large can be smoothed.
  • the reference pixel smoothing determining unit performs inter prediction on both the unit region including the reference pixel to be determined and the adjacent unit region in contact with the unit region at the boundary.
  • the prediction unit is the prediction unit in which the prediction image is generated
  • the reference pixel may be determined to be smoothed when the reference image used for the inter prediction in each unit region is different.
  • Pixels at the boundary between unit areas with different reference images used for inter prediction are likely to be distorted. Therefore, according to said structure, the reference pixel considered that distortion is large can be smoothed.
  • the reference pixel smoothing determining means includes a unit area including the reference pixel to be determined and quantization values of adjacent unit areas that are in contact with the unit area at the boundary. When the difference exceeds the threshold value, it may be determined that the reference pixel is smoothed.
  • the pixel at the boundary between the unit areas is likely to be distorted. Therefore, according to said structure, the reference pixel considered that distortion is large can be smoothed.
  • the reference pixel smoothing determining means may determine that the reference pixel is smoothed when the reference pixel to be determined is an edge pixel.
  • the reference pixel to be determined is an edge pixel
  • the reference pixel is smoothed. Edge pixels are very likely to have block distortion. Therefore, it is possible to smooth the reference pixels that are likely to have block distortion.
  • the encoded data includes smoothing information indicating that a reference pixel used for intra prediction is smoothed, and the smoothing unit detects the smoothing information.
  • the reference pixel existing in the vicinity of the boundary between the plurality of unit regions that are in contact with the prediction unit to be processed may be smoothed.
  • the reference pixel is smoothed by the smoothing information. Therefore, it is not necessary to determine whether or not a specific condition is satisfied, and the processing efficiency can be improved.
  • the image encoding device generates a prediction image by predicting a pixel value of a pixel in the prediction image from a pixel value of a reference pixel at another position in the same screen for each prediction unit.
  • An image encoding device that generates encoded data by encoding a prediction residual between a generated predicted image and an original image, and for each of the reference pixels, another pixel adjacent to the reference pixel.
  • the reference pixel smoothing determining means for determining whether to perform smoothing between the reference pixel and the reference pixel determined to be smoothed by the reference pixel smoothing determining means between the other pixels.
  • the reference pixel determined to be smoothed by the reference pixel smoothing determining means is used for the prediction using the reference pixel after the smoothing means smoothes.
  • Predictive image generation to generate images It is characterized in that it comprises a stage, a.
  • the prediction unit may be a PU described in the embodiment or a partition obtained by dividing the PU.
  • the present invention can be suitably applied to a decoding device that decodes encoded data and an encoding device that generates encoded data. Further, the present invention can be suitably applied to the data structure of encoded data generated by the encoding device and referenced by the decoding device.
  • Video decoding device (image decoding device) 2
  • Video encoding device (image encoding device) 12 prediction image generation unit 12b inter prediction image generation unit 12c intra prediction image generation unit (reference pixel smoothing determination unit, smoothing unit, prediction image generation unit) 21 prediction image generation unit 21a intra prediction image generation unit (reference pixel smoothing determination unit, smoothing unit, prediction image generation unit) 21c Inter prediction image generation unit

Landscapes

  • Engineering & Computer Science (AREA)
  • Multimedia (AREA)
  • Signal Processing (AREA)
  • Compression Or Coding Systems Of Tv Signals (AREA)

Abstract

This moving image decoding device (1) generates a prediction image by a prediction from the pixel value of a reference pixel present in a different position within the same screen, and is provided with an intra prediction image generation unit (12c) which determines, with respect to each of the reference pixels, whether or not smoothing is performed between the reference pixel and another pixel adjacent thereto, performs the smoothing between the determined reference pixel and the other pixel, and regarding the smoothed reference pixel, generates a prediction image using the smoothed reference pixel.

Description

画像復号装置、および画像符号化装置Image decoding apparatus and image encoding apparatus
 本発明は、符号化データを復号する画像復号装置、および符号化データを生成する画像符号化装置に関するものである。 The present invention relates to an image decoding apparatus that decodes encoded data, and an image encoding apparatus that generates encoded data.
 動画像を効率的に伝送または記録するために、動画像を符号化することによって符号化データを生成する動画像符号化装置(画像符号化装置)、および、当該符号化データを復号することによって復号画像を生成する動画像復号装置(画像復号装置)が用いられている。具体的な動画像符号化方式としては、例えば、H.264/MPEG-4.AVC(非特許文献1)、VCEG(Video Coding Expert Group)における共同開発用コーデックであるKTAソフトウェアに採用されている方式、および、その後継コーデックであるTMuC(Test Model under Consideration)ソフトウェアに採用されている方式などが挙げられる。 In order to efficiently transmit or record a moving image, a moving image encoding device (image encoding device) that generates encoded data by encoding the moving image, and decoding the encoded data A video decoding device (image decoding device) that generates a decoded image is used. As a specific moving picture encoding method, for example, H.264 is used. H.264 / MPEG-4. AVC (Non-Patent Document 1), VCEG (Video Coding Expert Group) adopted by KTA software, which is a codec for joint development, and TMuC (Test Model Under Consideration) software, a successor codec. There are some methods.
 このような符号化方式において、動画像を構成する画像(ピクチャ)は、画像を分割することにより得られるスライス、スライスを分割することにより得られる符号化単位(コーディングユニット(Coding Unit)と呼ばれることもある)、および、符号化単位を分割することより得られるブロックおよびパーティションからなる階層構造により管理され、普通、ブロックごとに符号化される。 In such an encoding method, an image (picture) constituting a moving image is a slice obtained by dividing an image, and a coding unit (Coding ス ラ イ ス Unit) obtained by dividing the slice. And a hierarchical structure consisting of blocks and partitions obtained by dividing an encoding unit, and is usually encoded block by block.
 また、このような符号化方式においては、通常、入力画像を符号化/復号することによって得られる局所復号画像に基づいて予測画像が生成され、当該予測画像と入力画像との差分画像(「残差画像」または「予測残差」と呼ぶこともある)が符号化される。また、予測画像の生成方法としては、画面間予測(インター予測)、および、画面内予測(イントラ予測)と呼ばれる方法が知られている。 In such an encoding method, a predicted image is usually generated based on a local decoded image obtained by encoding / decoding an input image, and a difference image (“residual” between the predicted image and the input image is generated. (Sometimes referred to as “difference image” or “prediction residual”). As methods for generating a predicted image, methods called inter-screen prediction (inter prediction) and intra-screen prediction (intra prediction) are known.
 インター予測においては、フレーム全体が復号された参照フレーム(復号画像)内の参照画像に対し、動きベクトルを用いた動き補償を適用することによって、予測対象フレーム内の予測画像が予測単位毎に生成される。 In inter prediction, a prediction image in a prediction target frame is generated for each prediction unit by applying motion compensation using a motion vector to a reference image in a reference frame (decoded image) obtained by decoding the entire frame. Is done.
 一方で、イントラ予測においては、同一フレーム内の局所復号画像に基づいて、当該フレームにおける予測画像が順次生成される。H.264/MPEG-4.AVCに用いられているイントラ予測の一例としては、予測単位(例えば、パーティション)毎に、予め定められた予測方向群に含まれる予測方向から何れかの予測方向が選択されると共に、局所復号画像における参照画素の画素値を、選択された予測方向に外挿することによって、該予測単位上の画素値を生成する方法(「基本予測」とも呼ぶことがある)が挙げられる。 On the other hand, in intra prediction, predicted images in the frame are sequentially generated based on locally decoded images in the same frame. H. H.264 / MPEG-4. As an example of intra prediction used in AVC, any prediction direction is selected from prediction directions included in a predetermined prediction direction group for each prediction unit (for example, partition), and a locally decoded image is used. And a method of generating a pixel value on the prediction unit by extrapolating the pixel value of the reference pixel in the selected prediction direction (sometimes referred to as “basic prediction”).
 また、非特許文献2には、予測単位毎に、該予測単位の周辺の画素の画素値に基づきエッジ方向を算出し、局所復号画像における参照画素の画素値を、算出されたエッジ方向に外挿することによって、該予測単位上の画素値を生成する方法(Differential Coding of Intra Modes(DCIM)とも呼ばれる方法、「エッジ予測」または「エッジベース予測」とも呼ぶことがある)が開示されている。 Further, in Non-Patent Document 2, for each prediction unit, the edge direction is calculated based on the pixel values of pixels around the prediction unit, and the pixel value of the reference pixel in the local decoded image is out of the calculated edge direction. A method for generating a pixel value on the prediction unit by inserting (a method called Differential Coding of Intra Modes (DCIM), sometimes called “edge prediction” or “edge-based prediction”) is disclosed. .
 また、非特許文献3に記載されているように、H.264/MPEG-4.AVCにおける8×8画面内(イントラ)予測符号化では、予測単位上の画素値を生成する前に、参照画素となる、予測単位の上辺および左辺に接している画素の画素値を平滑化(smoothing)している。これにより、8×8画素の対象予測単位の上辺または左辺に2つの4×4画素の予測単位が隣接している場合に、4×4画素の予測単位同士の境界に生じ得るブロック歪を除去することができる。したがって、非特許文献3に記載された技術によれば、ブロック歪を含んだ画素値をそのまま予測に用いることにより、予測単位上の画素の画素値が誤差を大きく含んでしまうことを防止することができる。 Also, as described in Non-Patent Document 3, H. H.264 / MPEG-4. In AVC 8 × 8 intra-frame (intra) predictive coding, before generating pixel values on a prediction unit, pixel values of pixels that are in contact with the upper and left sides of the prediction unit, which are reference pixels, are smoothed ( smoothing). This eliminates block distortion that may occur at the boundary between 4 × 4 pixel prediction units when two 4 × 4 pixel prediction units are adjacent to the upper or left side of the 8 × 8 pixel target prediction unit. can do. Therefore, according to the technique described in Non-Patent Document 3, the pixel value including the block distortion is used as it is for the prediction, thereby preventing the pixel value of the pixel on the prediction unit from containing a large error. Can do.
 具体的に図10を用いて説明する。図10に示すように、予測単位800の上辺および左辺に接する周辺画素に対し、高周波成分を除去するため、フィルタ係数が(1/4、1/2、1/4)である3タップのフィルタを適用して平滑化を行う。例えば、図10の画素800bを平滑化する場合、画素800a、b、cそれぞれが符号化・復号された後の画素値をQ1、Q2、Q3とすると、画素800bの平滑化後の画素値は(Q1/4+Q2/2+Q3/4)となる。これを、予測単位の上辺および左辺に接する周辺画素すべてについて行う。なお、左上、右上、右下の画素についてもほぼ同様の平滑化処理を行う。 Specific description will be given with reference to FIG. As shown in FIG. 10, a 3-tap filter whose filter coefficients are (1/4, 1/2, 1/4) in order to remove high-frequency components from neighboring pixels in contact with the upper and left sides of the prediction unit 800. Is applied to perform smoothing. For example, when the pixel 800b in FIG. 10 is smoothed, if the pixel values after the pixels 800a, b, and c are encoded and decoded are Q1, Q2, and Q3, the pixel value after the smoothing of the pixel 800b is (Q1 / 4 + Q2 / 2 + Q3 / 4). This is performed for all peripheral pixels in contact with the upper and left sides of the prediction unit. Note that the same smoothing process is performed for the upper left, upper right, and lower right pixels.
 また、非特許文献4には、TMuCにおいて、予測単位ごとに平滑化を行うか否かを適応的に指定して平滑化のON/OFFを切り替える技術が記載されている。例えば、図11に示すように、予測単位401~407の各々について平滑化を行うか否かがフラグで指定されている。図11に示す場合は、予測単位401、402、407で平滑化を行い、予測単位403~406は平滑化を行わない。エンコーダにおいて、平滑化を行うか否かは、予測単位ごとに平滑化を行った場合と行わなかった場合を比較し、RD(Rate-Distortion)コストの低い方を選択する。そして、平滑化の有無をフラグで指定する。デコーダでは、フラグに基づいて平滑化のON/OFFを切り替える。 Also, Non-Patent Document 4 describes a technique for switching ON / OFF of smoothing by adaptively specifying whether smoothing is performed for each prediction unit in TMuC. For example, as shown in FIG. 11, whether to perform smoothing for each of the prediction units 401 to 407 is designated by a flag. In the case shown in FIG. 11, the prediction units 401, 402, and 407 are smoothed, and the prediction units 403 to 406 are not smoothed. In the encoder, whether to perform smoothing is compared with the case where smoothing is performed for each prediction unit and the case where smoothing is not performed, and the one with lower RD (Rate-Distortion) cost is selected. Then, the presence or absence of smoothing is designated by a flag. The decoder switches the smoothing ON / OFF based on the flag.
 また、非特許文献5、6には、TMuCにおいて、フラグを用いずに、予測単位のサイズと予測方向(予測モード)とに応じて、平滑化のON/OFFを切り替える技術が記載されている。具体的には、予め予測単位のサイズごとに、平滑化を行う予測方向が決められており、これに該当する場合は平滑化を行い、該当しない場合は平滑化を行わない。 Non-Patent Documents 5 and 6 describe a technique for switching ON / OFF of smoothing according to the size of a prediction unit and a prediction direction (prediction mode) without using a flag in TMuC. . Specifically, the prediction direction for smoothing is determined in advance for each size of the prediction unit, and smoothing is performed when this is true, and smoothing is not performed when this is not the case.
 このように、イントラ予測にて参照される参照画素の平滑化を行う技術が複数開示されているが、何れも問題を有している。TMuCにおいては、予測単位として様々なサイズをとることが可能であるため、非特許文献3に開示された技術を用いたとしても、対象予測単位に隣接する予測単位同士の境界に生じ得るブロック歪を適切に除去することはできないという問題がある。また、非特許文献4~6に開示された技術を用いたとしても、符号化対象画像が本来有しているエッジについても平滑化を行ってしまう可能性があるため、符号化効率が向上しない、もしくは符号化効率が期待される程向上しないという問題がある。 As described above, a plurality of techniques for smoothing reference pixels referred to in intra prediction have been disclosed, but all have problems. Since TMuC can take various sizes as a prediction unit, even if the technique disclosed in Non-Patent Document 3 is used, block distortion that may occur at the boundary between prediction units adjacent to the target prediction unit. There is a problem that it cannot be removed properly. Even if the techniques disclosed in Non-Patent Documents 4 to 6 are used, there is a possibility that the edges inherent in the encoding target image may be smoothed, so that the encoding efficiency is not improved. Alternatively, there is a problem that the encoding efficiency is not improved as expected.
 本発明は、上記の問題点に鑑みてなされたものであり、その目的は、イントラ予測に用いられる参照画素に対して、従来よりも適切な平滑化を実行する画像復号装置等を実現することにある。 The present invention has been made in view of the above-described problems, and an object of the present invention is to realize an image decoding apparatus or the like that performs smoothing more appropriately than the conventional technique on reference pixels used for intra prediction. It is in.
 上記課題を解決するために、本発明に係る画像復号装置は、予測画像を、予測単位ごとに、同じ画面内の別の位置にある参照画素の画素値から該予測画像内の画素の画素値を予測することにより生成し、生成した予測画像を符号化データから復号した予測残差に加算して復号画像を生成する画像復号装置であって、上記参照画素それぞれについて、該参照画素と隣接している他の画素との間で平滑化を行うか否かを決定する参照画素平滑化決定手段と、上記参照画素平滑化決定手段が平滑化を行うと決定した参照画素を上記他の画素との間で平滑化を行う平滑化手段と、上記参照画素のうち、上記参照画素平滑化決定手段が平滑化を行うと決定した参照画素については、上記平滑化手段が平滑化した後の参照画素を用いて上記予測画像を生成する予測画像生成手段と、を備えていることを特徴としている。 In order to solve the above-described problem, the image decoding apparatus according to the present invention, for each prediction unit, calculates a pixel value of a pixel in the prediction image from a pixel value of a reference pixel at another position in the same screen. An image decoding apparatus that generates a decoded image by adding the generated predicted image to a prediction residual decoded from encoded data, and for each reference pixel adjacent to the reference pixel. A reference pixel smoothing determination unit that determines whether or not to perform smoothing with another pixel, and a reference pixel that the reference pixel smoothing determination unit determines to perform smoothing as the other pixel. Smoothing means that performs smoothing between the reference pixel and the reference pixel that is determined to be smoothed by the reference pixel smoothing determination means among the reference pixels, after the smoothing means smoothes the reference pixel To generate the predicted image Is characterized in that it comprises a, the predicted image generating means for.
 上記の構成によれば、予測単位ごとに同じ画面内の別の位置にある参照画素の画素値から該予測画像内の画素の画素値を予測するイントラ予測に用いる参照画素を平滑化するか否かを決定し、平滑化を行うと決定した参照画素のみ平滑化を行って予測画像を生成する。 According to said structure, whether the reference pixel used for the intra prediction which predicts the pixel value of the pixel in the said prediction image from the pixel value of the reference pixel in another position in the same screen for every prediction unit is smoothed Is determined, and only the reference pixels determined to be smoothed are smoothed to generate a predicted image.
 よって、参照画素ごとに平滑化を行うか、行わないかを決定することができる。したがって、従来のように、予測単位ごとに平滑化の有無が決定されることにより、当該予測単位の参照画素全てについて平滑化が行われてしまうということ、または参照画素全てについて平滑化が行われないということを防止することができる。 Therefore, it is possible to determine whether to perform smoothing for each reference pixel. Therefore, as in the prior art, whether or not smoothing is determined for each prediction unit results in smoothing for all reference pixels of the prediction unit, or smoothing for all reference pixels. It is possible to prevent the absence.
 また、参照画素ごとに平滑化の有無を決定することができるので、平滑化が不要な参照画素まで平滑化をしてしまうということがなく、より高精度な予測画像を生成することができる。 In addition, since it is possible to determine whether or not smoothing is performed for each reference pixel, it is possible to generate a predicted image with higher accuracy without performing smoothing to reference pixels that do not require smoothing.
 したがって、イントラ予測に用いられる参照画素に対して、従来よりも適切に平滑化を実行することができる。 Therefore, it is possible to perform smoothing more appropriately for the reference pixels used for intra prediction than before.
 なお、予測単位は、実施形態において説明するPUであってもよいし、PUを分割したパーティションであってもよい。 Note that the prediction unit may be a PU described in the embodiment or a partition obtained by dividing the PU.
 上記課題を解決するために、本発明に係る画像符号化装置は、予測画像を、予測単位ごとに、同じ画面内の別の位置にある参照画素の画素値から該予測画像内の画素の画素値を予測することにより生成し、生成した予測画像と原画像との予測残差を符号化することによって符号化データを生成する画像符号化装置であって、上記参照画素それぞれについて、該参照画素と隣接している他の画素との間で平滑化を行うか否かを決定する参照画素平滑化決定手段と、上記参照画素平滑化決定手段が平滑化を行うと決定した参照画素を上記他の画素との間で平滑化を行う平滑化手段と、上記参照画素のうち、上記参照画素平滑化決定手段が平滑化を行うと決定した参照画素については、上記平滑化手段が平滑化した後の参照画素を用いて上記予測画像を生成する予測画像生成手段と、を備えていることを特徴としている。 In order to solve the above-described problem, the image encoding device according to the present invention, for each prediction unit, calculates a pixel of a pixel in the prediction image from a pixel value of a reference pixel at another position in the same screen. An image encoding apparatus that generates encoded data by encoding a prediction residual between a generated predicted image and an original image, by predicting a value, and for each of the reference pixels A reference pixel smoothing determining unit that determines whether or not to perform smoothing between the pixel and another adjacent pixel, and the reference pixel determined by the reference pixel smoothing determining unit to perform the smoothing Smoothing means that performs smoothing between the reference pixel and reference pixels that are determined to be smoothed by the reference pixel smoothing determination means among the reference pixels after the smoothing means performs smoothing. Prediction using the reference pixels It is characterized in that it comprises the predicted image generating means for generating an image, a.
 上記の構成によれば、予測単位ごとに同じ画面内の別の位置にある参照画素の画素値から該予測画像内の画素の画素値を予測するイントラ予測に用いる参照画素を平滑化するか否かを決定し、平滑化を行うと決定した参照画素のみ平滑化を行って予測画像を生成する。 According to said structure, whether the reference pixel used for the intra prediction which predicts the pixel value of the pixel in the said prediction image from the pixel value of the reference pixel in another position in the same screen for every prediction unit is smoothed Is determined, and only the reference pixels determined to be smoothed are smoothed to generate a predicted image.
 よって、参照画素ごとに平滑化を行うか、行わないかを決定することができる。したがって、従来のように、予測単位ごとに平滑化の有無が決定されることにより、当該予測単位の参照画素全てについて平滑化が行われてしまうということ、または参照画素全てについて平滑化が行われないということを防止することができる。 Therefore, it is possible to determine whether to perform smoothing for each reference pixel. Therefore, as in the prior art, whether or not smoothing is determined for each prediction unit results in smoothing for all reference pixels of the prediction unit, or smoothing for all reference pixels. It is possible to prevent the absence.
 また、参照画素ごとに平滑化の有無を決定することができるので、平滑化が不要な参照画素まで平滑化をしてしまうということがなく、より高精度な予測画像を生成することができる。 In addition, since it is possible to determine whether or not smoothing is performed for each reference pixel, it is possible to generate a predicted image with higher accuracy without performing smoothing to reference pixels that do not require smoothing.
 したがって、イントラ予測に用いられる参照画素に対して、従来よりも適切に平滑化を実行することができる。 Therefore, it is possible to perform smoothing more appropriately for the reference pixels used for intra prediction than before.
 なお、予測単位は、実施形態において説明するPUであってもよいし、PUを分割したパーティションであってもよい。 Note that the prediction unit may be a PU described in the embodiment or a partition obtained by dividing the PU.
 以上のように、本発明に係る画像復号装置は、参照画素それぞれについて、該参照画素と隣接している他の画素との間で平滑化を行うか否かを決定する参照画素平滑化決定手段と、上記参照画素平滑化決定手段が平滑化を行うと決定した参照画素を上記他の画素との間で平滑化を行う平滑化手段と、上記参照画素のうち、上記参照画素平滑化決定手段が平滑化を行うと決定した参照画素については、上記平滑化手段が平滑化した後の参照画素を用いて上記予測画像を生成する予測画像生成手段と、を備えている構成である。 As described above, the image decoding apparatus according to the present invention is a reference pixel smoothing determination unit that determines whether or not to perform smoothing between each reference pixel and another pixel adjacent to the reference pixel. A smoothing means for smoothing the reference pixel determined to be smoothed by the reference pixel smoothing determining means with the other pixels, and among the reference pixels, the reference pixel smoothing determining means The reference pixel determined to be smoothed is configured to include a predicted image generation unit that generates the predicted image using the reference pixel after smoothing by the smoothing unit.
 よって、参照画素ごとに平滑化を行うか、行わないかを決定することができる。したがって、従来のように、予測単位ごとに平滑化の有無が決定されることにより、当該予測単位の参照画素全てについて平滑化が行われてしまうということ、または参照画素全てについて平滑化が行われないということを防止することができるという効果を奏する。 Therefore, it is possible to determine whether to perform smoothing for each reference pixel. Therefore, as in the prior art, whether or not smoothing is determined for each prediction unit results in smoothing for all reference pixels of the prediction unit, or smoothing for all reference pixels. There is an effect that it can be prevented.
 また、参照画素ごとに平滑化の有無を決定することができるので、平滑化が不要な参照画素まで平滑化をしてしまうということがなく、より高精度な予測画像を生成することができるという効果を奏する。 Further, since the presence or absence of smoothing can be determined for each reference pixel, it is possible to generate a predicted image with higher accuracy without smoothing to reference pixels that do not require smoothing. There is an effect.
 したがって、イントラ予測に用いられる参照画素に対して、従来よりも適切に平滑化を実行することができるという効果を奏する。 Therefore, there is an effect that smoothing can be performed more appropriately than in the past with respect to reference pixels used for intra prediction.
 また、従来のように予測単位ごとに平滑化を行うか否かを示すフラグを符号化する必要がないので、符号化効率も向上させることができるという効果を奏する。 In addition, since it is not necessary to encode a flag indicating whether smoothing is performed for each prediction unit as in the prior art, there is an effect that the encoding efficiency can be improved.
本発明の実施の形態を示すものであり、動画像復号装置の要部構成を示すブロック図である。1, showing an embodiment of the present invention, is a block diagram illustrating a main configuration of a moving picture decoding apparatus. FIG. 上記動画像復号装置によって参照される符号化データのデータ構造を示すものであって、図2(a)は、符号化データのピクチャレイヤの構成を示す図であり、図2(b)は、ピクチャレイヤに含まれるスライスレイヤの構成を示す図であり、図2(c)は、スライスレイヤに含まれるLCUレイヤを構成する各CUの構成を示す図であり、図2(d)は、CUレイヤに含まれるリーフCUの構成を示す図であり、図2(e)は、リーフCUについてのインター予測情報の構成を示す図であり、図2(f)は、リーフCUについてのイントラ予測情報の構成を示す図である。FIG. 2A shows a data structure of encoded data referred to by the moving picture decoding apparatus, FIG. 2A shows a configuration of a picture layer of the encoded data, and FIG. FIG. 2C is a diagram illustrating a configuration of a slice layer included in the picture layer, FIG. 2C is a diagram illustrating a configuration of each CU configuring the LCU layer included in the slice layer, and FIG. FIG. 2E is a diagram illustrating a configuration of a leaf CU included in a layer, FIG. 2E is a diagram illustrating a configuration of inter prediction information for the leaf CU, and FIG. 2F is intra prediction information for a leaf CU. FIG. 平滑化を行う参照画像を説明するための図である。It is a figure for demonstrating the reference image which performs smoothing. 参照画素について、平滑化を行うか否かを決定する流れを示すフローチャートである。It is a flowchart which shows the flow which determines whether smoothing is performed about a reference pixel. 参照画素について、平滑化を行うか否かを決定する流れを示すフローチャートである。It is a flowchart which shows the flow which determines whether smoothing is performed about a reference pixel. 参照画素について、平滑化を行うか否かを決定する流れを示すフローチャートである。It is a flowchart which shows the flow which determines whether smoothing is performed about a reference pixel. 上記動画像復号装置の動作を説明するための図であって、図7(a)は、上記動画像復号装置が参照する予測モードを予測モードインデックスと共に示す図であり、図7(b)は、対象パーティションに属する画素、および、その周辺の復号済み画素を示す図である。FIGS. 7A and 7B are diagrams for explaining the operation of the video decoding device, in which FIG. 7A illustrates a prediction mode referred to by the video decoding device together with a prediction mode index, and FIG. FIG. 4 is a diagram illustrating pixels belonging to a target partition and decoded pixels in the vicinity thereof. 上記動画像復号装置において、エッジベース予測モードが選択された場合の、イントラ予測画像の生成処理を説明するための図であって、図8(a)は、対象パーティションを対象パーティションの周辺のパーティションと共に示す図であり、図8(b)は、補正角度を指定するパラメータを、補正後の予測方向と共に示す図である。FIG. 8A is a diagram for describing intra-prediction image generation processing when the edge-based prediction mode is selected in the video decoding device, and FIG. 8A illustrates a target partition as a partition around the target partition. FIG. 8B is a diagram illustrating parameters for specifying the correction angle together with the predicted direction after correction. 本発明に係る動画像符号化装置の要部構成を示すブロック図である。It is a block diagram which shows the principal part structure of the moving image encoder which concerns on this invention. 従来技術を示すものであり、参照画素を平滑化する方法の説明図である。It is a figure which shows a prior art and is an explanatory view of the method of smoothing a reference pixel. 従来技術を示すものであり、ブロックごとに平滑化を行うか否かを示す図である。It is a figure which shows a prior art and shows whether smoothing is performed for every block. 動画像復号装置および動画像符号化装置が、動画像の送受信に利用できることを説明するための図であり、(a)は、動画像符号化装置を搭載した送信装置の構成を示したブロック図であり、(b)は、動画像復号装置を搭載した受信装置の構成を示したブロック図である。It is a figure for demonstrating that a moving image decoding apparatus and a moving image coding apparatus can be utilized for transmission / reception of a moving image, (a) is the block diagram which showed the structure of the transmission device carrying a moving image coding apparatus (B) is a block diagram showing a configuration of a receiving apparatus equipped with a moving picture decoding apparatus. 動画像復号装置および動画像符号化装置が、動画像の記録および再生に利用できることを説明するための図であり、(a)は、動画像符号化装置2を搭載した記録装置の構成を示したブロック図であり、(b)は、動画像復号装置を搭載した再生装置の構成を示したブロックである。It is a figure for demonstrating that a moving image decoding apparatus and a moving image encoding apparatus can be utilized for recording and reproduction | regeneration of a moving image, (a) shows the structure of the recording device carrying the moving image encoding apparatus 2. FIG. 8B is a block diagram illustrating a configuration of a playback device equipped with a video decoding device.
 本発明の一実施の形態について図1~9に基づいて説明すれば、以下のとおりである。本実施形態に係る動画像復号装置(画像復号装置)1は、符号化データから動画像を復号するものである。また、本実施形態に係る動画像符号化装置は、動画像を符号化することによって符号化データを生成するものである。ただし、本発明の適用範囲はこれに限定されるものではない。すなわち、以下の説明からも明らかなように、本発明の特徴はイントラ予測にあり、複数のフレームを前提としなくとも成立するものである。すなわち、動画像を対象とするか静止画像を対象とするかを問わず、復号装置一般および符号化装置一般に適用できるものである。 Referring to FIGS. 1 to 9, an embodiment of the present invention will be described as follows. A moving image decoding apparatus (image decoding apparatus) 1 according to the present embodiment decodes a moving image from encoded data. In addition, the moving image encoding apparatus according to the present embodiment generates encoded data by encoding a moving image. However, the scope of application of the present invention is not limited to this. That is, as will be apparent from the following description, the feature of the present invention lies in intra prediction, and is established without assuming a plurality of frames. That is, the present invention can be applied to a general decoding apparatus and a general encoding apparatus regardless of whether the target is a moving image or a still image.
 (符号化データ#1の構成)
 本実施形態に係る動画像復号装置1の説明に先立ち、本実施形態に係る動画像符号化装置2によって生成され、動画像復号装置1によって復号される符号化データ#1の構成について、図2を参照して説明する。符号化データ#1は、シーケンスレイヤ、GOP(Group Of Pictures)レイヤ、ピクチャレイヤ、スライスレイヤ、および、最大符号化単位(LCU:Largest Coding Unit)レイヤからなる階層構造を有している。
(Configuration of encoded data # 1)
Prior to the description of the moving picture decoding apparatus 1 according to the present embodiment, the configuration of the encoded data # 1 generated by the moving picture encoding apparatus 2 according to the present embodiment and decoded by the moving picture decoding apparatus 1 will be described with reference to FIG. Will be described with reference to FIG. The encoded data # 1 has a hierarchical structure including a sequence layer, a GOP (Group Of Pictures) layer, a picture layer, a slice layer, and a maximum coding unit (LCU) layer.
 符号化データ#1におけるピクチャレイヤ以下の階層の構造を図2に示す。図2(a)~(f)は、それぞれ、ピクチャレイヤP、スライスレイヤS、LCUレイヤLCU、LCUに含まれるリーフCU(図2(d)ではCULと表記)、インター予測(画面間予測)パーティションについての予測情報PIであるインター予測情報PI_Inter、イントラ予測(画面内予測)パーティションについての予測情報PIであるイントラ予測情報PI_Intraの構造を示す図である。 FIG. 2 shows the hierarchical structure below the picture layer in the encoded data # 1. FIGS. 2A to 2F are a picture layer P, a slice layer S, an LCU layer LCU, a leaf CU included in the LCU (denoted as CUL in FIG. 2D), and inter prediction (inter-screen prediction), respectively. It is a figure which shows the structure of inter prediction information PI_Inter which is the prediction information PI about a partition, and intra prediction information PI_Intra which is the prediction information PI about an intra prediction (prediction in a screen) partition.
 (ピクチャレイヤ)
 ピクチャレイヤPは、処理対象のピクチャである対象ピクチャを復号するために動画像復号装置1が参照するデータの集合である。ピクチャレイヤPは、図2(a)に示すように、ピクチャヘッダPH、および、スライスレイヤS1~SNsを含んでいる(NsはピクチャレイヤPに含まれるスライスレイヤの総数)。
(Picture layer)
The picture layer P is a set of data that is referenced by the video decoding device 1 in order to decode a target picture that is a processing target picture. As shown in FIG. 2A, the picture layer P includes a picture header PH and slice layers S1 to SNs (Ns is the total number of slice layers included in the picture layer P).
 ピクチャヘッダPHには、対象ピクチャの復号方法を決定するために動画像復号装置1が参照する符号化パラメータ群が含まれている。例えば、動画像符号化装置2が符号化の際に用いた可変長符号化のモードを示す符号化モード情報(entropy_coding_mode_flag)は、ピクチャヘッダPHに含まれる符号化パラメータの一例である。entropy_coding_mode_flagが0の場合、当該ピクチャは、CAVLC(Context-based Adaptive Variable Length Coding)によって符号化されており、entropy_coding_mode_flagが1である場合、当該ピクチャは、CABAC(Context-based Adaptive Binary Arithmetic Coding)によって符号化されている。 The picture header PH includes a coding parameter group referred to by the video decoding device 1 in order to determine a decoding method of the target picture. For example, the encoding mode information (entropy_coding_mode_flag) indicating the variable length encoding mode used in encoding by the moving image encoding device 2 is an example of an encoding parameter included in the picture header PH. When entropy_coding_mode_flag is 0, the picture is encoded by CAVLC (Context-based Adaptive Variable Length Coding). When entropy_coding_mode_flag is 1, the picture is encoded by CABAC (Context-based Adaptive Binary Arithmetic Coding). It has become.
 (スライスレイヤ)
 ピクチャレイヤPに含まれる各スライスレイヤSは、処理対象のスライスである対象スライスを復号するために動画像復号装置1が参照するデータの集合である。スライスレイヤSは、図2(b)に示すように、スライスヘッダSH、および、LCUレイヤLCU1~LCUNc(NcはスライスSに含まれるLCUの総数)を含んでいる。
(Slice layer)
Each slice layer S included in the picture layer P is a set of data referred to by the video decoding device 1 in order to decode a target slice that is a slice to be processed. As shown in FIG. 2B, the slice layer S includes a slice header SH and LCU layers LCU1 to LCUn (Nc is the total number of LCUs included in the slice S).
 スライスヘッダSHには、対象スライスの復号方法を決定するために動画像復号装置1が参照する符号化パラメータ群が含まれる。スライスタイプを指定するスライスタイプ指定情報(slice_type)は、スライスヘッダSHに含まれる符号化パラメータの一例である。また、スライスヘッダSHには、動画像復号装置1の備えるループフィルタによって参照されるフィルタパラメータFPが含まれる。 The slice header SH includes a coding parameter group that the moving image decoding apparatus 1 refers to in order to determine a decoding method of the target slice. Slice type designation information (slice_type) for designating a slice type is an example of an encoding parameter included in the slice header SH. Further, the slice header SH includes a filter parameter FP that is referred to by a loop filter included in the video decoding device 1.
 なお、スライスタイプ指定情報により指定可能なスライスタイプとしては、(1)符号化の際にイントラ予測のみを用いるIスライス、(2)符号化の際に単方向予測、又は、イントラ予測を用いるPスライス、(3)符号化の際に単方向予測、双方向予測、又は、イントラ予測を用いるBスライスなどが挙げられる。 As slice types that can be specified by the slice type specification information, (1) I slice using only intra prediction at the time of encoding, and (2) P using unidirectional prediction or intra prediction at the time of encoding. Slice, (3) B-slice using unidirectional prediction, bidirectional prediction, or intra prediction at the time of encoding.
 (LCUレイヤ)
 スライスレイヤSに含まれる各LCUレイヤLCUは、処理対象のLCUである対象LCUを復号するために動画像復号装置1が参照するデータの集合である。
(LCU layer)
Each LCU layer LCU included in the slice layer S is a set of data that the video decoding device 1 refers to in order to decode the target LCU that is the processing target LCU.
 LCUレイヤLCUは、当該LCUを階層的に4分木分割することにより得られる複数の符号化単位(CU:Coding Unit)より構成される。換言すれば、LCUレイヤLCUは、複数のCUを再帰的に含む階層構造のうち、最上位にあたる符号化単位である。LCUレイヤLCUに含まれる各CUは、図2(c)に示すように、CUヘッダCUH、および、当該CUを4分木分割することにより得られる複数のCUを再帰的に含む階層構造を有している。 The LCU layer LCU is composed of a plurality of coding units (CU: Coding Units) obtained by hierarchically dividing the LCU into a quadtree. In other words, the LCU layer LCU is a coding unit corresponding to the highest level in a hierarchical structure that recursively includes a plurality of CUs. As shown in FIG. 2C, each CU included in the LCU layer LCU has a hierarchical structure that recursively includes a CU header CUH and a plurality of CUs obtained by dividing the CU into quadtrees. is doing.
 LCUを除く各CUのサイズは、当該CUが直接に属するCU(すなわち、当該CUの1階層上位のCU)のサイズの縦横とも半分であり、各CUのとり得るサイズは、符号化データ#1のシーケンスパラメータセットSPSに含まれる、LCUのサイズおよび階層深度(hierarchical depth)に依存する。例えば、LCUのサイズが128×128画素であって、最大階層深度が5である場合には、当該LCU以下の階層におけるCUは、5種類のサイズ、すなわち、128×128画素、64×64画素、32×32画素、16×16画素、および、8×8画素の何れかをとり得る。また、それ以上分割されないCUをリーフCUと呼ぶことにする。 The size of each CU excluding the LCU is half the size of the CU to which the CU directly belongs (that is, the CU one layer higher than the CU), and the size that each CU can take is encoded data # 1. Dependent on the size and hierarchical depth of the LCU included in the sequence parameter set SPS. For example, when the size of the LCU is 128 × 128 pixels and the maximum hierarchical depth is 5, the CUs in the hierarchical level below the LCU have five sizes, that is, 128 × 128 pixels and 64 × 64 pixels. , 32 × 32 pixels, 16 × 16 pixels, and 8 × 8 pixels. A CU that is not further divided is called a leaf CU.
 (CUヘッダ)
 CUヘッダCUHには、対象CUの復号方法を決定するために動画像復号装置1が参照する符号化パラメータが含まれる。具体的には、図2(c)に示すように、対象CUをさらに下位のCUへ4分割するか否かを指定するCU分割フラグSP_CUが含まれる。CU分割フラグSP_CUが0である場合、すなわち、それ以上CUが分割されない場合、当該CUはリーフCUである。
(CU header)
The CU header CUH includes a coding parameter referred to by the video decoding device 1 in order to determine a decoding method of the target CU. Specifically, as shown in FIG. 2C, a CU division flag SP_CU that specifies whether or not the target CU is further divided into four subordinate CUs is included. When the CU division flag SP_CU is 0, that is, when the CU is not further divided, the CU is a leaf CU.
 (リーフCU)
 それ以上分割されないCU(CUのリーフ)は予測単位PU(Prediction Unit)、および、変換単位TU(Transform Unit)として取り扱われる。
(Leaf CU)
A CU (CU leaf) that is not further divided is handled as a prediction unit PU (Prediction Unit) and a transform unit TU (Transform Unit).
 図2(d)に示すように、リーフCU(図2(d)においてはCULと表記)は、(1)動画像復号装置1により予測画像を生成する際に参照されるPU情報PUI、および、(2)動画像復号装置1により残差データを復号する際に参照されるTU情報TUIを含んでいる。 As shown in FIG. 2 (d), the leaf CU (denoted as CUL in FIG. 2 (d)) includes (1) PU information PUI that is referred to when the moving image decoding apparatus 1 generates a predicted image, and (2) The TU information TUI that is referred to when the residual data is decoded by the moving picture decoding apparatus 1 is included.
 スキップフラグSKIPは、対象のPUについて、スキップモードが適用されているか否かを示すフラグであり、スキップフラグSKIPの値が1の場合、すなわち、対象のリーフにスキップモードが適用されている場合、そのリーフCUにおけるPU情報PUI、および、TU情報TUIは省略される。なお、スキップフラグSKIPは、Iスライスでは省略される。 The skip flag SKIP is a flag indicating whether or not the skip mode is applied to the target PU. When the value of the skip flag SKIP is 1, that is, when the skip mode is applied to the target leaf, PU information PUI and TU information TUI in the leaf CU are omitted. Note that the skip flag SKIP is omitted for the I slice.
 PU情報PUIは、図2(d)に示すように、スキップフラグSKIP、予測タイプ情報PT、および、予測情報PIを含んでいる。予測タイプ情報PTは、対象リーフCU(対象PU)についての予測画像生成方法として、イントラ予測を用いるのか、または、インター予測を用いるのかを指定する情報である。予測情報PIは、予測タイプ情報PTが何れの予測方法を指定するのかに応じて、イントラ予測情報PI_Intra、または、インター予測情報PI_Interより構成される。以下では、イントラ予測が適用されるPUをイントラPUとも呼称し、インター予測が適用されるPUをインターPUとも呼称する。 The PU information PUI includes a skip flag SKIP, prediction type information PT, and prediction information PI as shown in FIG. The prediction type information PT is information that specifies whether intra prediction or inter prediction is used as a predicted image generation method for the target leaf CU (target PU). The prediction information PI includes intra prediction information PI_Intra or inter prediction information PI_Inter depending on which prediction method is specified by the prediction type information PT. Hereinafter, a PU to which intra prediction is applied is also referred to as an intra PU, and a PU to which inter prediction is applied is also referred to as an inter PU.
 PU情報PUIには、対象PUに含まれる各パーティションの形状、サイズ、および、対象PU内での位置を指定する情報が含まれる。ここで、パーティションとは、対象リーフCUを構成する1又は複数の重複しない領域のことであり、予測画像の生成は、パーティションを単位として行われる。 The PU information PUI includes information specifying the shape and size of each partition included in the target PU and the position in the target PU. Here, the partition is one or a plurality of non-overlapping areas constituting the target leaf CU, and the generation of the predicted image is performed in units of partitions.
 TU情報TUIは、図2(d)に示すように、量子化ステップの大きさを指定する量子化パラメータ差分Δqp(tu_qp_delta)、対象リーフCU(対象TU)の各ブロックへの分割パターンを指定するTU分割情報SP_TU、および、量子化予測残差QD1~QDNT(NTは、対象TUに含まれるブロックの総数)を含んでいる。 As shown in FIG. 2D, the TU information TUI specifies a quantization parameter difference Δqp (tu_qp_delta) that specifies the magnitude of the quantization step, and a division pattern for each block of the target leaf CU (target TU). TU partition information SP_TU and quantized prediction residuals QD1 to QDNT (NT is the total number of blocks included in the target TU) are included.
 量子化パラメータ差分Δqpは、対象TUにおける量子化パラメータqpと、そのTUの直前に符号化されたTUにおける量子化パラメータqp’との差分qp-qp’である。 The quantization parameter difference Δqp is a difference qp−qp ′ between the quantization parameter qp in the target TU and the quantization parameter qp ′ in the TU encoded immediately before the TU.
 TU分割情報SP_TUは、具体的には、対象TUに含まれる各ブロックの形状、サイズ、および、対象TU内での位置を指定する情報である。各TUは、例えば、64×64画素から2×2画素までのサイズをとり得る。ここで、ブロックとは、対象リーフCUを構成する1又は複数の重複しない領域のことであり、予測残差の符号化・復号は、ブロックを単位として行われる。 TU partition information SP_TU is information that specifies the shape and size of each block included in the target TU and the position in the target TU. Each TU can be, for example, a size from 64 × 64 pixels to 2 × 2 pixels. Here, the block is one or a plurality of non-overlapping areas constituting the target leaf CU, and encoding / decoding of the prediction residual is performed in units of blocks.
 各量子化予測残差QDは、動画像符号化装置2が以下の処理1~3を、処理対象のブロックである対象ブロックに施すことによって生成した符号化データである。処理1:符号化対象画像から予測画像を減算した予測残差をDCT変換(Discrete Cosine Transform)する。処理2:処理1にて得られたDCT係数を量子化する。処理3:処理2にて量子化されたDCT係数を可変長符号化する。上述した量子化パラメータqpは、動画像符号化装置2がDCT係数を量子化する際に用いた量子化ステップQPの大きさを表す(QP=2qp/6)。 Each quantized prediction residual QD is encoded data generated by the moving image encoding apparatus 2 performing the following processes 1 to 3 on a target block that is a processing target block. Process 1: DCT transform (Discrete Cosine Transform) is performed on the prediction residual obtained by subtracting the prediction image from the encoding target image. Process 2: The DCT coefficient obtained in Process 1 is quantized. Process 3: The DCT coefficient quantized in Process 2 is variable length encoded. The quantization parameter qp described above represents the magnitude of the quantization step QP used when the moving picture coding apparatus 2 quantizes the DCT coefficient (QP = 2 qp / 6 ).
 (インター予測情報PI_Inter)
 インター予測情報PI_Interは、動画像復号装置1が、インター予測によってインター予測画像を生成する際に参照される符号化パラメータが含まれる。図2(e)に示すように、インター予測情報PI_Interは、対象PUの各パーティションへの分割パターンを指定するインターPU分割情報SP_Inter、および、各パーティションについてのインター予測パラメータPP_Inter1~PP_InterNe(Neは、対象PUに含まれるインター予測パーティションの総数)を含んでいる。
(Inter prediction information PI_Inter)
The inter prediction information PI_Inter includes a coding parameter that is referred to when the video decoding device 1 generates an inter prediction image by inter prediction. As shown in FIG. 2 (e), the inter prediction information PI_Inter includes inter PU partition information SP_Inter that specifies a partition pattern for each partition of the target PU, and inter prediction parameters PP_Inter1 to PP_InterNe (Ne for each partition). The total number of inter prediction partitions included in the target PU).
 インターPU分割情報SP_Interは、具体的には、対象PU(インターPU)に含まれる各インター予測パーティションの形状、サイズ、および、対象PU内での位置を指定する情報である。 Specifically, the inter-PU partition information SP_Inter is information for designating the shape and size of each inter prediction partition included in the target PU (inter PU) and the position in the target PU.
 インターPUは、2N×2N画素、2N×N画素、N×2N画素、およびN×N画素の4つの対称的分割(symmetric splittings)、並びに、2N×nU画素、2N×nD画素、nL×2N画素、およびnR×2N画素の4つの非対称的分割(asymmetric splittings)により、合計8種類のパーティションに分割することが可能である。ここで、Nの具体的な値は、当該PUが属するCUのサイズによって規定され、nU、nD、nL、および、nRの具体的な値は、Nの値に応じて定められる。例えば、128×128画素のインターPUは、128×128画素、128×64画素、64×128画素、64×64画素、128×32画素、128×96画素、32×128画素、および、96×128画素のインター予測パーティションへ分割することが可能である。 The inter PU is composed of four symmetric splittings of 2N × 2N pixels, 2N × N pixels, N × 2N pixels, and N × N pixels, and 2N × nU pixels, 2N × nD pixels, and nL × 2N. It is possible to divide into 8 types of partitions in total by four asymmetric splits of pixels and nR × 2N pixels. Here, the specific value of N is defined by the size of the CU to which the PU belongs, and the specific values of nU, nD, nL, and nR are determined according to the value of N. For example, an inter PU of 128 × 128 pixels is 128 × 128 pixels, 128 × 64 pixels, 64 × 128 pixels, 64 × 64 pixels, 128 × 32 pixels, 128 × 96 pixels, 32 × 128 pixels, and 96 × It is possible to divide into 128-pixel inter prediction partitions.
 (インター予測パラメータ)
 インター予測パラメータPP_Interは、図2(e)に示すように、参照画像インデックスRIと、推定動きベクトルインデックスPMVIと、動きベクトル残差MVDとを含んでいる。
(Inter prediction parameter)
As illustrated in FIG. 2E, the inter prediction parameter PP_Inter includes a reference image index RI, an estimated motion vector index PMVI, and a motion vector residual MVD.
 (イントラ予測情報PI_Intra)
 イントラ予測情報PI_Intraは、動画像復号装置1が、イントラ予測によってイントラ予測画像を生成する際に参照される符号化パラメータが含まれる。図2(f)に示すように、イントラ予測情報PI_Intraは、対象PU(イントラPU)の各パーティションへの分割パターンを指定するイントラPU分割情報SP_Intra、および、各パーティションについてのイントラ予測パラメータPP_Intra1~PP_IntraNa(Naは、対象PUに含まれるイントラ予測パーティションの総数)を含んでいる。
(Intra prediction information PI_Intra)
The intra prediction information PI_Intra includes an encoding parameter that is referred to when the video decoding device 1 generates an intra predicted image by intra prediction. As shown in FIG. 2 (f), the intra prediction information PI_Intra includes intra PU partition information SP_Intra that specifies a partition pattern of the target PU (intra PU) into each partition, and intra prediction parameters PP_Intra1 to PP_IntraNa for each partition. (Na is the total number of intra prediction partitions included in the target PU).
 イントラPU分割情報SP_Intraは、具体的には、対象PUに含まれる各イントラ予測パーティションの形状、サイズ、および、対象PU内での位置を指定する情報である。イントラPU分割情報SP_Intraには、対象PUをパーティションに分割するか否かを指定するイントラ分割フラグ(intra_split_flag)が含まれている。イントラ分割フラグが1であれば、対象PUは、4つのパーティションへと対称的に分割され、イントラ分割フラグが0であれば、対象PUは、分割されることなく、対象PU自身が1つのパーティションとして取り扱われる。したがって、対象PUのサイズを2N×2N画素とすると、イントラ予測パーティションは、2N×2N画素(分割なし)、および、N×N画素(4分割)の何れかのサイズを取り得る(ここで、N=2n、nは1以上の任意の整数)。例えば、128×128画素のイントラPUは、128×128画素、および、64×64画素のイントラ予測パーティションへ分割することが可能である。 Specifically, the intra-PU partition information SP_Intra is information that specifies the shape and size of each intra-predicted partition included in the target PU, and the position in the target PU. The intra PU split information SP_Intra includes an intra split flag (intra_split_flag) that specifies whether or not the target PU is split into partitions. If the intra partition flag is 1, the target PU is divided symmetrically into four partitions. If the intra partition flag is 0, the target PU is not divided and the target PU itself is one partition. Are treated as Therefore, if the size of the target PU is 2N × 2N pixels, the intra prediction partition can take any of 2N × 2N pixels (no division) and N × N pixels (four divisions) (where, N = 2 n , n is an arbitrary integer of 1 or more). For example, a 128 × 128 pixel intra PU can be divided into 128 × 128 pixel and 64 × 64 pixel intra prediction partitions.
 (イントラ予測パラメータPP_Intra)
 イントラ予測パラメータPP_Intraは、図2(f)に示すように、推定フラグMPM、残余予測モードインデックスRIPM、および、追加インデックスAIを含んでいる。イントラ予測パラメータPP_Intraは、各パーティションについてのイントラ予測方法(予測モード)を指定するためのパラメータである。
(Intra prediction parameter PP_Intra)
As shown in FIG. 2F, the intra prediction parameter PP_Intra includes an estimation flag MPM, a residual prediction mode index RIPM, and an additional index AI. The intra prediction parameter PP_Intra is a parameter for designating an intra prediction method (prediction mode) for each partition.
 推定フラグMPMは、処理対象である対象パーティションの周辺のパーティションに割り付けられた予測モードに基づいて推定された予測モードと当該対象パーティションについての予測モードとが同じであるか否かを示すフラグである。ここで、対象パーティションの周辺のパーティションの例としては、対象パーティションの上辺に隣接するパーティション、および、対象パーティションの左辺に隣接するパーティション等が挙げられる。 The estimation flag MPM is a flag indicating whether or not the prediction mode estimated based on the prediction mode allocated to the peripheral partition of the target partition that is the processing target is the same as the prediction mode for the target partition. . Here, examples of partitions around the target partition include a partition adjacent to the upper side of the target partition and a partition adjacent to the left side of the target partition.
 残余予測モードインデックスRIPMは、推定された予測モードと対象パーティションについての予測モードとが異なる場合に、イントラ予測パラメータPP_Intraに含められるインデックスであり、当該対象パーティションに割り付けられる予測モードを指定するためのインデックスである。 The residual prediction mode index RIPM is an index included in the intra prediction parameter PP_Intra when the estimated prediction mode and the prediction mode for the target partition are different, and is an index for designating a prediction mode assigned to the target partition. It is.
 追加インデックスAIは、対象パーティションに割り付けられた予測モードが、所定の予測モードである場合に、当該対象パーティションについてのイントラ予測方法をより詳細に指定するためのインデックスである。 The additional index AI is an index for specifying the intra prediction method for the target partition in more detail when the prediction mode assigned to the target partition is a predetermined prediction mode.
 (動画像復号装置1)
 以下では、本実施形態に係る動画像復号装置1について図1~7を参照して説明する。動画像復号装置1は、その一部に、H.264/MPEG-4.AVC、VCEG(Video Coding Expert Group)における共同開発用コーデックであるKTAソフトウェアに採用されている技術、および、その後継コーデックであるTMuC(Test Model under Consideration)ソフトウェアに採用されている技術を含む復号装置である。
(Moving picture decoding apparatus 1)
Hereinafter, the moving picture decoding apparatus 1 according to the present embodiment will be described with reference to FIGS. The moving picture decoding apparatus 1 includes H.264 as a part thereof. H.264 / MPEG-4. Decoding device including technology adopted in KTA software which is a codec for joint development in AVC and VCEG (Video Coding Expert Group), and technology adopted in TMuC (Test Model under Consideration) software which is a successor codec It is.
 図1は、動画像復号装置1の構成を示すブロック図である。図1に示すように、動画像復号装置1は、可変長符号復号部11、予測画像生成部12、逆量子化・逆変換部13、加算器14、フレームメモリ15、および、ループフィルタ16を備えている。また、図1に示すように、予測画像生成部12は、動きベクトル復元部12a、インター予測画像生成部12b、イントラ予測画像生成部(参照画素平滑化決定手段、平滑化手段、予測画像生成手段)12c、および、予測方式決定部12dを備えている。動画像復号装置1は、符号化データ#1を復号することによって動画像#2を生成するための装置である。 FIG. 1 is a block diagram showing a configuration of the moving picture decoding apparatus 1. As shown in FIG. 1, the moving image decoding apparatus 1 includes a variable length code decoding unit 11, a predicted image generation unit 12, an inverse quantization / inverse conversion unit 13, an adder 14, a frame memory 15, and a loop filter 16. I have. As shown in FIG. 1, the predicted image generation unit 12 includes a motion vector restoration unit 12a, an inter predicted image generation unit 12b, an intra predicted image generation unit (reference pixel smoothing determination unit, smoothing unit, predicted image generation unit). ) 12c and a prediction method determination unit 12d. The moving picture decoding apparatus 1 is an apparatus for generating moving picture # 2 by decoding encoded data # 1.
 (可変長符号復号部11)
 可変長符号復号部11は、各パーティションに関する予測パラメータPPを、符号化データ#1から復号し、予測画像生成部12に供給する。具体的には、可変長符号復号部11は、インター予測パーティションに関しては、参照画像インデックスRI、推定動きベクトルインデックスPMVI、および、動きベクトル残差MVDを含むインター予測パラメータPP_Interを符号化データ#1から復号し、これらを動きベクトル復元部12aに供給する。一方、イントラ予測パーティションに関しては、推定フラグMPM、残余インデックスRIPM、および、追加インデックスAIを含むイントラ予測パラメータPP_Intraを符号化データ#1から復号し、これらをイントラ予測画像生成部12cに供給する。また、可変長符号復号部11は、パーティションのサイズを指定するサイズ指定情報をイントラ予測画像生成部12cに供給する(図示省略)。
(Variable-length code decoding unit 11)
The variable length code decoding unit 11 decodes the prediction parameter PP related to each partition from the encoded data # 1, and supplies the decoded prediction parameter PP to the predicted image generation unit 12. Specifically, for the inter prediction partition, the variable-length code decoding unit 11 receives the inter prediction parameter PP_Inter including the reference image index RI, the estimated motion vector index PMVI, and the motion vector residual MVD from the encoded data # 1. These are decoded and supplied to the motion vector restoration unit 12a. On the other hand, for the intra prediction partition, the intra prediction parameter PP_Intra including the estimation flag MPM, the residual index RIPM, and the additional index AI is decoded from the encoded data # 1, and these are supplied to the intra prediction image generation unit 12c. In addition, the variable length code decoding unit 11 supplies size designation information for designating the size of the partition to the intra predicted image generation unit 12c (not shown).
 また、可変長符号復号部11は、各パーティションについての予測タイプ情報PTを符号化データ#1から復号し、これを予測方式決定部12dに供給する。更に、可変長符号復号部11は、各ブロックに関する量子化予測残差QD、および、そのブロックを含むTUに関する量子化パラメータ差分Δqpを符号化データ#1から復号し、これらを逆量子化・逆変換部13に供給する。また、可変長符号復号部11は、符号化データ#1からフィルタパラメータFPを復号し、これをループフィルタ16に供給する。 Also, the variable length code decoding unit 11 decodes the prediction type information PT for each partition from the encoded data # 1, and supplies this to the prediction method determination unit 12d. Furthermore, the variable-length code decoding unit 11 decodes the quantization prediction residual QD for each block and the quantization parameter difference Δqp for the TU including the block from the encoded data # 1, and dequantizes and reverses them. This is supplied to the conversion unit 13. Further, the variable length code decoding unit 11 decodes the filter parameter FP from the encoded data # 1 and supplies this to the loop filter 16.
 なお、可変長符号復号部11による具体的な復号方式としては、算術符号化/復号の一方式であるCABAC(Context-based Adaptive Binary Arithmetic Coding)、または、非算術符号化/復号の一方式であるCAVLC(Context-based Adaptive Variable Length Coding)が用いられる。ここで、CABACとは、コンテキストに基づく適応的なバイナリー算術符号化を行う符号化/復号方式であり、CALVCとは、コンテキストを適応的に切り替える可変長符号のセットを利用する符号化・復号方式である。CABACは、CAVLCに比べて、符号量削減効果が大きい半面、処理量が増大するという側面もある。 As a specific decoding method by the variable-length code decoding unit 11, CABAC (Context-based Adaptive Binary Arithmetic Coding) which is one arithmetic coding / decoding method, or one non-arithmetic encoding / decoding method is used. A certain CAVLC (Context-based Adaptive Variable Variable Length Coding) is used. Here, CABAC is an encoding / decoding scheme that performs adaptive binary arithmetic coding based on context, and CALVC is an encoding / decoding scheme that uses a set of variable length codes that adaptively switch contexts. It is. CABAC has a larger code amount reduction effect than CAVLC, but also has an aspect of increasing the processing amount.
 可変長符号復号部11は、符号化データ#1のピクチャヘッダPHに含まれる符号化モード情報(entropy_coding_mode_flag)を参照することにより、対象ピクチャがCABACにより符号化されたものであるのか、CAVLCにより符号化されたものであるのかを識別することができる。また、可変長符号復号部11は、識別した符号化方式に対応する復号方式により、対象ピクチャを復号する。 The variable length code decoding unit 11 refers to the encoding mode information (entropy_coding_mode_flag) included in the picture header PH of the encoded data # 1 to determine whether the target picture has been encoded by CABAC or by CAVLC. Can be identified. In addition, the variable length code decoding unit 11 decodes the target picture using a decoding method corresponding to the identified encoding method.
 (予測画像生成部12)
 予測画像生成部12は、各パーティションについての予測タイプ情報PTに基づいて、各パーティションがインター予測を行うべきインター予測パーティションであるのか、イントラ予測を行うべきイントラ予測パーティションであるのかを識別する。そして、前者の場合には、インター予測画像Pred_Interを生成すると共に、生成したインター予測画像Pred_Interを予測画像Predとして加算器14に供給し、後者の場合には、イントラ予測画像Pred_Intraを生成すると共に、生成したイントラ予測画像Pred_Intraを加算器14に供給する。なお、予測画像生成部12は、処理対象PUに対してスキップモードが適用されている場合には、当該PUに属する他のパラメータの復号を省略する。
(Predicted image generation unit 12)
The predicted image generation unit 12 identifies whether each partition is an inter prediction partition for performing inter prediction or an intra prediction partition for performing intra prediction based on the prediction type information PT for each partition. In the former case, the inter prediction image Pred_Inter is generated, and the generated inter prediction image Pred_Inter is supplied to the adder 14 as the prediction image Pred. In the latter case, the intra prediction image Pred_Intra is generated, The generated intra predicted image Pred_Intra is supplied to the adder 14. Note that, when the skip mode is applied to the processing target PU, the predicted image generation unit 12 omits decoding of other parameters belonging to the PU.
 (動きベクトル復元部12a)
 動きベクトル復元部12aは、各インター予測パーティションに関する動きベクトルmvを、そのパーティションに関する動きベクトル残差MVDと、他のパーティションに関する復元済みの動きベクトルmv’とから復元する。具体的には、(1)推定動きベクトルインデックスPMVIにより指定される推定方法に従って、復元済みの動きベクトルmv’から推定動きベクトルpmvを導出し、(2)導出した推定動きベクトルpmvと動きベクトル残差MVDとを加算することによって動きベクトルmvを得る。なお、他のパーティションに関する復元済みの動きベクトルmv’は、フレームメモリ15から読み出すことができる。動きベクトル復元部12aは、復元した動きベクトルmvを、対応する参照画像インデックスRIと共に、インター予測画像生成部12bに供給する。
(Motion vector restoration unit 12a)
The motion vector restoration unit 12a restores the motion vector mv related to each inter prediction partition from the motion vector residual MVD related to that partition and the restored motion vector mv ′ related to another partition. Specifically, (1) the estimated motion vector pmv is derived from the restored motion vector mv ′ according to the estimation method specified by the estimated motion vector index PMVI, and (2) the derived estimated motion vector pmv and the motion vector remaining are derived. The motion vector mv is obtained by adding the difference MVD. It should be noted that the restored motion vector mv ′ relating to other partitions can be read from the frame memory 15. The motion vector restoration unit 12a supplies the restored motion vector mv to the inter predicted image generation unit 12b together with the corresponding reference image index RI.
 (インター予測画像生成部12b)
 インター予測画像生成部12bは、画面間予測によって、各インター予測パーティションに関する動き補償画像mcを生成する。具体的には、動きベクトル復元部12aから供給された動きベクトルmvを用いて、同じく動きベクトル復元部12aから供給された参照画像インデックスRIによって指定されるフィルタ済復号画像P_ALF’から動き補償画像mcを生成する。ここで、フィルタ済復号画像P_ALF’は、既にフレーム全体の復号が完了した復号済みの復号画像に対して、ループフィルタ16によるフィルタ処理を施すことによって得られる画像であり、インター予測画像生成部12bは、フィルタ済復号画像P_ALF’を構成する各画素の画素値をフレームメモリ15から読み出すことができる。インター予測画像生成部12bによって生成された動き補償画像mcは、インター予測画像Pred_Interとして予測方式決定部12dに供給される。
(Inter prediction image generation unit 12b)
The inter prediction image generation unit 12b generates a motion compensated image mc related to each inter prediction partition by inter-screen prediction. Specifically, using the motion vector mv supplied from the motion vector restoration unit 12a, the motion compensation image mc from the filtered decoded image P_ALF ′ designated by the reference image index RI supplied from the motion vector restoration unit 12a. Is generated. Here, the filtered decoded image P_ALF ′ is an image obtained by performing the filtering process by the loop filter 16 on the decoded image that has already been decoded for the entire frame, and the inter predicted image generation unit 12b. Can read out the pixel value of each pixel constituting the filtered decoded image P_ALF ′ from the frame memory 15. The motion compensated image mc generated by the inter predicted image generation unit 12b is supplied to the prediction method determination unit 12d as an inter predicted image Pred_Inter.
 (イントラ予測画像生成部12c)
 イントラ予測画像生成部12cは、各イントラ予測パーティションに関する予測画像Pred_Intraを生成する。具体的には、まず、可変長符号復号部11から供給されたイントラ予測パラメータPP_Intraに基づいて予測モードを特定し、特定された予測モードを対象パーティションに対して、例えば、ラスタスキャン順に割り付ける。続いて、当該予測モードの示す予測方法に従って、画面内予測によって、(局所)復号画像Pから予測画像Pred_Intraを生成する。画面内予測を行う際には、対象パーティションの近傍画素について、平滑化の要否を決定する。そして、平滑化を行うと決定した画素については平滑化を行ったのち、画面内予測を行う。
(Intra predicted image generation unit 12c)
The intra predicted image generation unit 12c generates a predicted image Pred_Intra related to each intra prediction partition. Specifically, first, a prediction mode is specified based on the intra prediction parameter PP_Intra supplied from the variable length code decoding unit 11, and the specified prediction mode is assigned to the target partition in, for example, raster scan order. Subsequently, a predicted image Pred_Intra is generated from the (local) decoded image P by intra prediction according to the prediction method indicated by the prediction mode. When performing intra prediction, whether or not smoothing is necessary is determined for the neighboring pixels of the target partition. And about the pixel determined to perform smoothing, after performing smoothing, intra prediction is performed.
 具体的に、図3を用いて説明する。図3に示すように、予測単位(対象パーティション)301の予測画像を生成するために用いる参照画素301a~301qについて、平滑化を行うか否かの決定を行う。この決定は、例えば、当該参照画素がPUやCU等のブロックの境界近傍の画素であるかによって決定する。そして、境界近傍の画素についてのみ平滑化を行う。ここでは、隣接ブロック302、303の境界から2画素以内の画素である参照画素301c、d、e、fを平滑化する。ブロック境界は高周波の歪が出やすいため、境界近傍の画素についてのみ平滑化を行えば、歪を低減して予測画像に歪が伝播することを防止できる。よって、画質を向上することができる。また、ブロック単位で平滑化を行うか否かを示すフラグを符号化する必要がないので符号化効率も向上させることができる。 Specifically, this will be described with reference to FIG. As shown in FIG. 3, it is determined whether or not to smooth the reference pixels 301a to 301q used for generating a prediction image of the prediction unit (target partition) 301. This determination is made, for example, depending on whether the reference pixel is a pixel near the boundary of a block such as PU or CU. Then, smoothing is performed only for pixels near the boundary. Here, the reference pixels 301c, d, e, and f which are pixels within two pixels from the boundary between the adjacent blocks 302 and 303 are smoothed. Since high-frequency distortion tends to occur at the block boundary, if only pixels near the boundary are smoothed, distortion can be reduced and distortion can be prevented from propagating to the predicted image. Therefore, the image quality can be improved. In addition, since it is not necessary to encode a flag indicating whether smoothing is performed in units of blocks, it is possible to improve encoding efficiency.
 また、図3において、例えば、隣接ブロックの境界が参照画素301mと301nとの間にある場合、隣接ブロックの境界から2画素以内の画素である参照画素301l、m、n、oを平滑化する。 In FIG. 3, for example, when the boundary of the adjacent block is between the reference pixels 301m and 301n, the reference pixels 301l, m, n, and o which are pixels within two pixels from the boundary of the adjacent block are smoothed. .
 そして、イントラ予測画像生成部12cによって生成されたイントラ予測画像Pred_Intraは、予測方式決定部12dに供給される。なお、イントラ予測画像生成部12cは、画面内予測によって、フィルタ済復号画像P_ALFから予測画像Pred_Intraを生成する構成とすることも可能である。なお、平滑化の要否決定処理については、後述する。 Then, the intra predicted image Pred_Intra generated by the intra predicted image generating unit 12c is supplied to the prediction method determining unit 12d. Note that the intra predicted image generation unit 12c may be configured to generate the predicted image Pred_Intra from the filtered decoded image P_ALF by intra prediction. The smoothing necessity determination process will be described later.
 (予測方式決定部12d)
 予測方式決定部12dは、各パーティションが属するPUについての予測タイプ情報PTに基づいて、各パーティションがインター予測を行うべきインター予測パーティションであるのか、イントラ予測を行うべきイントラ予測パーティションであるのかを決定する。そして、前者の場合には、インター予測画像生成部12bにて生成されたインター予測画像Pred_Interを予測画像Predとして加算器14に供給し、後者の場合には、イントラ予測画像生成部12cにて生成されたイントラ予測画像Pred_Intraを予測画像Predとして加算器14に供給する。
(Prediction method determination unit 12d)
The prediction method determination unit 12d determines whether each partition is an inter prediction partition that should perform inter prediction or an intra prediction partition that should perform intra prediction based on the prediction type information PT about the PU to which each partition belongs. To do. In the former case, the inter prediction image Pred_Inter generated by the inter prediction image generation unit 12b is supplied to the adder 14 as the prediction image Pred. In the latter case, the inter prediction image generation unit 12c generates the inter prediction image Pred_Inter. The intra predicted image Pred_Intra that has been processed is supplied to the adder 14 as the predicted image Pred.
 (逆量子化・逆変換部13)
 逆量子化・逆変換部13は、(1)量子化予測残差QDを逆量子化し、(2)逆量子化によって得られたDCT係数を逆DCT(Discrete Cosine Transform)変換し、(3)逆DCT変換によって得られた予測残差Dを加算器14に供給する。なお、量子化予測残差QDを逆量子化する際に、逆量子化・逆変換部13は、可変長符号復号部11から供給された量子化パラメータ差分Δqpから量子化ステップQPを導出する。量子化パラメータqpは、直前に逆量子化/逆DCT変換したTUに関する量子化パラメータqp’に量子化パラメータ差分Δqpを加算することによって導出でき、量子化ステップQPは、量子化ステップqpから例えばQP=2pq/6によって導出できる。また、逆量子化・逆変換部13による予測残差Dの生成は、TUあるいはTUを分割したブロックを単位として行われる。
(Inverse quantization / inverse transform unit 13)
The inverse quantization / inverse transform unit 13 (1) inversely quantizes the quantized prediction residual QD, (2) performs inverse DCT (Discrete Cosine Transform) transform on the DCT coefficient obtained by the inverse quantization, and (3) The prediction residual D obtained by the inverse DCT transform is supplied to the adder 14. When the quantization prediction residual QD is inversely quantized, the inverse quantization / inverse transform unit 13 derives the quantization step QP from the quantization parameter difference Δqp supplied from the variable length code decoding unit 11. The quantization parameter qp can be derived by adding the quantization parameter difference Δqp to the quantization parameter qp ′ relating to the TU that has been inversely quantized / inversely DCT transformed immediately before, and the quantization step QP is derived from the quantization step qp, for example, QP = 2 pq / 6 . The generation of the prediction residual D by the inverse quantization / inverse transform unit 13 is performed in units of blocks obtained by dividing TUs or TUs.
 (加算器14)
 加算器14は、予測画像生成部12から供給された予測画像Predと、逆量子化・逆変換部13から供給された予測残差Dとを加算することによって復号画像Pを生成する。生成された復号画像Pは、フレームメモリ15に格納される。
(Adder 14)
The adder 14 generates the decoded image P by adding the prediction image Pred supplied from the prediction image generation unit 12 and the prediction residual D supplied from the inverse quantization / inverse conversion unit 13. The generated decoded image P is stored in the frame memory 15.
 (ループフィルタ16)
 ループフィルタ16は、フレームメモリ15から復号画像Pを読み出し、復号画像Pのパーティション境界およびブロック境界の何れか一方若しくは双方におけるブロックノイズ低減処理(デブロッキング処理)を施す。また、ループフィルタ16は、ブロックノイズ低減処理が施された復号画像に対して、符号化データ#1から復号されるフィルタパラメータFPを用いた適応的なフィルタ処理を施し、当該適応的なフィルタ処理が施された復号画像Pをフィルタ済復号画像P_ALFとしてフレームメモリ15に出力する。
(Loop filter 16)
The loop filter 16 reads the decoded image P from the frame memory 15 and performs block noise reduction processing (deblocking processing) at one or both of the partition boundary and the block boundary of the decoded image P. In addition, the loop filter 16 performs adaptive filter processing using the filter parameter FP decoded from the encoded data # 1 on the decoded image subjected to the block noise reduction processing, and the adaptive filter processing Is output to the frame memory 15 as a filtered decoded image P_ALF.
 (イントラ予測画像生成部12cによる平滑化の要否決定処理)
 次に、図4を用いて、イントラ予測画像生成部12cにおける平滑化の要否決定処理について説明する。平滑化の要否決定処理は、予測画像の生成に用いる参照画素全てについて、すなわち対象パーティションの上辺および左辺の何れかと接している画素および、対象パーティションの左上の画素(対象パーティションの左上の頂点を共有する画素)について行う(S31)。
(Smoothing necessity determination processing by the intra predicted image generation unit 12c)
Next, smoothing necessity determination processing in the intra predicted image generation unit 12c will be described with reference to FIG. The smoothing necessity determination process is performed on all reference pixels used for generating a predicted image, that is, a pixel in contact with either the upper side or the left side of the target partition, and an upper left pixel of the target partition (the upper left vertex of the target partition This is performed for the pixels to be shared (S31).
 まず、決定対象となる画素について、当該画素が、対象パーティションと接している隣接ブロック同士の境界近傍に存在するか否かを判定する(S32)。境界近傍であるか否かの判定は、境界から所定の距離dth画素以内であるか否かによって判定する。なお、隣接ブロック(単位領域)は、PU(予測単位)、CU(処理単位)、TU(変換単位)等のいずれであってもよいし、PUを分割して得られるパーティション、およびTUを分割して得られるブロックの何れであってもよい。さらにこられを併用するものであってもよい。また、PUとTUとの境界が重なる境界のみ、判定処理に用いる境界としてもよい。 First, with respect to a pixel to be determined, it is determined whether or not the pixel exists in the vicinity of the boundary between adjacent blocks in contact with the target partition (S32). Whether or not the boundary is near is determined by whether or not it is within a predetermined distance d th pixels from the boundary. The adjacent block (unit area) may be any of PU (prediction unit), CU (processing unit), TU (transform unit), and the partition obtained by dividing PU and TU Any of the obtained blocks may be used. Further, these may be used in combination. Further, only the boundary where the boundary between the PU and the TU overlaps may be a boundary used for the determination process.
 そして、境界近傍の画素であると判定すると(S32でYES)、イントラ予測画像生成部12cは、当該参照画素について平滑化を行う(S33)。平滑化は、例えば、1:2:1の平滑化フィルタを適用することによって行う。具体的には、(x,y)に存在する参照画素の画素値p(x,y)について、平滑化後の画素値は、p(x,y)=(p(x-1,y)+p(x,y)×2+p(x+1,y))/4となる。 If it is determined that the pixel is near the boundary (YES in S32), the intra predicted image generation unit 12c smoothes the reference pixel (S33). Smoothing is performed, for example, by applying a smoothing filter of 1: 2: 1. Specifically, for the pixel value p (x, y) of the reference pixel existing at (x, y), the pixel value after smoothing is p (x, y) = (p (x−1, y)) + P (x, y) × 2 + p (x + 1, y)) / 4.
 その後、参照画素全てについて上記処理が終了すると(S34)、イントラ予測画像生成部12cは、予測画像の生成処理を行う。 Thereafter, when the above process is completed for all the reference pixels (S34), the intra predicted image generation unit 12c performs a predicted image generation process.
  (平滑化の要否決定処理2)
 次に、図5を用いて、上記とは異なる方法による、イントラ予測画像生成部12cの平滑化の要否決定処理について説明する。
(Smoothing necessity determination process 2)
Next, smoothing necessity determination processing of the intra predicted image generation unit 12c by a method different from the above will be described with reference to FIG.
 図5に示すように、平滑化の要否決定処理は、予測画像の生成に用いる参照画素全てについて、すなわち対象パーティションの上辺および左辺の何れかと接している画素および、対象パーティションの左上の画素(対象パーティションの左上の頂点を共有する画素)について行う(S41)。そして、イントラ予測画像生成部12cは、決定対象となる画素について、当該画素が対象パーティションと接している隣接ブロック同士の境界近傍に存在するか否かを判定する(S42)。この判定処理は、上述したステップS32と同様の方法で行う。 As shown in FIG. 5, the smoothing necessity determination process is performed for all reference pixels used for generating a predicted image, that is, a pixel in contact with either the upper side or the left side of the target partition, and the upper left pixel ( This is performed for pixels sharing the top left vertex of the target partition (S41). And the intra estimated image generation part 12c determines whether the pixel concerned exists in the boundary vicinity of the adjacent blocks which are in contact with the object partition about the pixel used as a decision object (S42). This determination process is performed in the same manner as in step S32 described above.
 次に、イントラ予測画像生成部12cは、決定対象となる画素について、後に行うデブロッキング処理のブロック境界強度(Bs値:Boundary Strength)が、閾値以上か否かを判定する(S43)。閾値としては、例えば、Bs値=4とすることができる。 Next, the intra predicted image generation unit 12c determines whether or not the block boundary strength (Bs value: Boundary Strength) of the deblocking process to be performed later is greater than or equal to the threshold for the pixel to be determined (S43). As the threshold value, for example, Bs value = 4 can be set.
 デブロッキング処理は周知技術であるので、詳細な説明については省略するが、Bs値は、0~4の5段階で設定される。これは、以下のように設定される。ブロックの境界を挟む画素をそれぞれ、画素p、画素qとした場合に、画素pまたは画素qのうち少なくとも一方が、イントラ(画面内)CUに属し、なおかつ、CUの境界に位置する場合には、Bs=4となる。ここで、イントラCUとは、画面内予測を用いて符号化されたリーフCUのことを示す。 Since the deblocking process is a well-known technique, the detailed description is omitted, but the Bs value is set in five stages from 0 to 4. This is set as follows. When the pixels sandwiching the block boundary are the pixel p and the pixel q, respectively, when at least one of the pixel p or the pixel q belongs to the intra (in-screen) CU and is located at the boundary of the CU , Bs = 4. Here, the intra CU indicates a leaf CU encoded using intra prediction.
 また、画素pまたは画素qのうちいずれか一方が、イントラCUに属するが、CUの境界には位置しない場合には、Bs=3となる。 Further, when either one of the pixel p and the pixel q belongs to the intra CU but is not located at the boundary of the CU, Bs = 3.
 また、画素pおよび画素qの何れもイントラCUに属するものではなく、なおかつ、画素pまたは画素qの属する変換ブロックのうちどちらか一方が直交変換係数を持つ場合には、Bs=2となる。 Further, when neither the pixel p nor the pixel q belongs to the intra CU, and one of the transform blocks to which the pixel p or the pixel q belongs has an orthogonal transform coefficient, Bs = 2.
 また、画素pおよび画素qの何れもイントラCUに属するものではなく、なおかつ、画素pおよび画素qの属する変換ブロックの何れも直交変換係数を持たないが、参照ピクチャが異なるか、参照ピクチャの枚数が異なるか、または、動きベクトルの値が所定の閾値以上異なる場合には、Bs=1となる。 Also, neither the pixel p nor the pixel q belongs to the intra CU, and neither of the transform blocks to which the pixel p and the pixel q belong has an orthogonal transform coefficient, but the reference picture is different or the number of reference pictures Or when the motion vector values differ by a predetermined threshold value or more, Bs = 1.
 また、画素pおよび画素qの何れもCUに属するものではなく、なおかつ、画素pおよび画素qの何れも直交変換係数を持たず、参照ピクチャが同一であり、動きベクトルの値が所定の閾値未満である場合には、Bs=0となる。 Further, neither the pixel p nor the pixel q belongs to the CU, and neither the pixel p nor the pixel q has an orthogonal transformation coefficient, the reference picture is the same, and the value of the motion vector is less than a predetermined threshold value. In this case, Bs = 0.
 そして、Bs値が閾値以上であれば(S43でYES)、イントラ予測画像生成部12cは、当該参照画素について平滑化を行う(S44)。平滑化処理は、上述したステップS33と同様の方法で行う。 If the Bs value is equal to or greater than the threshold (YES in S43), the intra predicted image generation unit 12c smoothes the reference pixel (S44). The smoothing process is performed by the same method as in step S33 described above.
 その後、参照画素全てについて上記処理が終了すると(S45)、イントラ予測画像生成部12cは、予測画像の生成処理を行う。 After that, when the above process is completed for all the reference pixels (S45), the intra predicted image generation unit 12c performs a predicted image generation process.
 なお、これ以外にも、例えば、イントラ予測情報PI_Intraに、境界近傍画素に対し平滑化を行うか否かを示すフラグ(平滑化情報)を含ませ、これによって、平滑化を行うか否かを決定してもよい。 In addition to this, for example, the intra prediction information PI_Intra includes a flag (smoothing information) indicating whether or not smoothing is performed on the pixels near the boundary, thereby determining whether or not smoothing is performed. You may decide.
  (平滑化の要否決定処理3)
 次に、図6を用いて、上記とは異なる方法による、イントラ予測画像生成部12cの平滑化の要否決定処理について説明する。
(Smoothing necessity determination process 3)
Next, smoothing necessity determination processing of the intra predicted image generation unit 12c by a method different from the above will be described with reference to FIG.
 図6に示すように、平滑化の要否決定処理は、予測画像の生成に用いる参照画素全てについて、すなわち対象パーティションの上辺および左辺の何れかと接している画素および、対象パーティションの左上の画素(対象パーティションの左上の頂点を共有する画素)について行う(S51)。そして、イントラ予測画像生成部12cは、決定対象となる画素が、エッジ画素かどうかを判定する(S52)。 As shown in FIG. 6, smoothing necessity determination processing is performed for all reference pixels used for generating a predicted image, that is, a pixel in contact with either the upper side or the left side of the target partition, and the upper left pixel ( This is performed for pixels that share the top left vertex of the target partition (S51). Then, the intra predicted image generation unit 12c determines whether the pixel to be determined is an edge pixel (S52).
 エッジ画素か否かの判定は、例えば、判定対象となる画素(x,y)が対象パーティションの上辺に存在する場合、画素(x,y-1)を中心とした3×3画素に以下に示すようなソーベルフィルタを適用した値が、閾値以上であるか否かによって判定する。 For example, when the pixel (x, y) to be determined exists on the upper side of the target partition, the determination as to whether or not the pixel is an edge pixel is as follows: 3 × 3 pixels centered on the pixel (x, y−1) It is determined by whether or not the value to which the Sobel filter as shown is applied is greater than or equal to the threshold value.
Figure JPOXMLDOC01-appb-M000001
Figure JPOXMLDOC01-appb-M000001
 そして、イントラ予測画像生成部12cは、決定対象となる画素をエッジ画素と判定すると(S52でYES)、当該画素を平滑化する(S53)。平滑化処理は、上述したステップS33と同様の方法で行う。 When the intra predicted image generation unit 12c determines that the pixel to be determined is an edge pixel (YES in S52), the intra predicted image generation unit 12c smoothes the pixel (S53). The smoothing process is performed by the same method as in step S33 described above.
 その後、参照画素全てについて上記処理が終了すると(S54)、イントラ予測画像生成部12cは、予測画像の生成処理を行う。 After that, when the above process is completed for all the reference pixels (S54), the intra predicted image generation unit 12c performs a predicted image generation process.
 なお、決定対象となる画素が、ブロック歪に起因するエッジではなく、符号化対象画像が本来有しているエッジに相当する画素だった場合は、平滑化を行う必要はない。そこで、ソーベルフィルタを適用した値が、第1の閾値以上で、かつ、第2の閾値以下の場合に、平滑化を行う処理であってもよい。ここで第2の閾値は、符号化対象画像の有しているエッジの強度に応じて予め定めておくことができる。 Note that if the pixel to be determined is not an edge caused by block distortion but a pixel corresponding to an edge inherent in the encoding target image, it is not necessary to perform smoothing. Therefore, smoothing may be performed when the value to which the Sobel filter is applied is equal to or greater than the first threshold and equal to or less than the second threshold. Here, the second threshold value can be determined in advance according to the strength of the edge of the encoding target image.
 また、エッジ画素の判定にソーベルフィルタを用いるのではなく、決定対象画素の左右(予測単位の上辺に接している場合)、または、上下(予測単位の左辺に接している場合)の画素の画素値の差が閾値以上か否かで判定してもよい。例えば、予測単位の上辺に接している画素(x,y)について判定する場合、|p(x-1,y)-p(x+1,y)|≧閾値Thを満たす画素(x,y)をエッジ画素と判定してもよい。 Also, instead of using a Sobel filter for edge pixel determination, the right and left (when touching the upper side of the prediction unit) or the upper and lower (when touching the left side of the prediction unit) of the pixel to be determined You may determine by the difference of a pixel value being more than a threshold value. For example, when determining the pixel (x, y) in contact with the upper side of the prediction unit, the pixel (x, y) that satisfies | p (x−1, y) −p (x + 1, y) | ≧ threshold Th is determined. You may determine with an edge pixel.
 (イントラ予測画像生成部12cによるイントラ予測画像Pred_Intraの生成処理)
 次に、イントラ予測画像生成部12cによるイントラ予測画像Pred_Intraの生成処理について、図7を用いて説明する。図7は、動画像復号装置1の動作を説明するための図であって、図7(a)は、動画像復号装置1が参照する予測モードであって、複数の基本予測モード、および、1つのエッジベース予測モードよりなる拡張セットに含まれる予測モードを、予測モードインデックスと共に示す図であり、図7(b)は、対象パーティションに属する画素、および、その周辺の復号済み画素を示す図である。
(Generation process of intra prediction image Pred_Intra by intra prediction image generation unit 12c)
Next, generation processing of the intra predicted image Pred_Intra by the intra predicted image generation unit 12c will be described with reference to FIG. FIG. 7 is a diagram for explaining the operation of the video decoding device 1, and FIG. 7A is a prediction mode referred to by the video decoding device 1, and includes a plurality of basic prediction modes, and FIG. 7B is a diagram illustrating a prediction mode included in an extended set including one edge-based prediction mode together with a prediction mode index, and FIG. 7B is a diagram illustrating pixels belonging to the target partition and decoded pixels in the vicinity thereof. It is.
 イントラ予測画像生成部12cは、(1)予め定められた1または複数の予測方向およびDC予測の何れかを指定する基本予測モード、および、(2)対象パーティションの周辺の画素値を用いた計算により予測方向を定める予測モード、例えば、対象パーティションの周辺の画素値より算出されるエッジ方向(または、当該エッジ方向の示す角度と補正角度との和によって表される方向)を予測方向とするエッジベース予測モード、のうち、イントラ予測パラメータPP_Intraによって指定される予測モードに基づいて、対象パーティションにおけるイントラ予測画像Pred_Intraを生成する。 The intra-predicted image generation unit 12c (1) a basic prediction mode that designates one or more predetermined prediction directions and DC prediction, and (2) calculation using pixel values around the target partition Prediction mode that determines the prediction direction by, for example, an edge whose prediction direction is the edge direction calculated from pixel values around the target partition (or the direction represented by the sum of the angle indicated by the edge direction and the correction angle) Based on the prediction mode specified by the intra prediction parameter PP_Intra among the base prediction modes, the intra prediction image Pred_Intra in the target partition is generated.
 換言すれば、イントラ予測画像生成部12cは、1または複数の基本予測モード、および、エッジベース予測モードよりなる予測モードのセット(以下、「拡張セット」とも呼ぶ)から、イントラ予測パラメータPP_Intraによって指定される予測モードを選択し、選択された予測モードに基づいて、対象パーティションにおけるイントラ予測画像Pred_Intraを生成する。 In other words, the intra-predicted image generation unit 12c is designated by an intra-prediction parameter PP_Intra from a set of prediction modes (hereinafter also referred to as “extended set”) including one or a plurality of basic prediction modes and an edge-based prediction mode. A prediction mode to be selected is selected, and an intra prediction image Pred_Intra in the target partition is generated based on the selected prediction mode.
 以下では、予め定められた1または複数の予測方向およびDC予測の何れかを指定する基本予測モードよりなるセットを基本予測モードセットとも呼ぶことにする。すなわち、上記拡張セットには、基本予測モードセットに含まれる予測モードと、エッジベース予測モードとが含まれる。 Hereinafter, a set of basic prediction modes that specify one or more predetermined prediction directions and DC prediction is also referred to as a basic prediction mode set. That is, the extended set includes a prediction mode included in the basic prediction mode set and an edge-based prediction mode.
 図7(a)は、拡張セットに含まれる各予測モードを、当該各予測モードに付される予測モードインデックスと共に示す図である。また、図7(a)においては、基本予測モードセットに属する各方向予測モードと、当該各方向予測モードの示す予測方向とが示されている。図7(a)に示すように、エッジベース予測モードは、インデックス1によって指定され、基本予測モードに含まれるDC予測モードは、インデックス0によって指定され、基本予測モードセットに含まれる各方向予測モードは、インデックス2~9によって指定される。 FIG. 7A is a diagram showing each prediction mode included in the extended set together with a prediction mode index assigned to each prediction mode. FIG. 7A shows each direction prediction mode belonging to the basic prediction mode set and the prediction direction indicated by each direction prediction mode. As shown in FIG. 7A, the edge-based prediction mode is specified by index 1, and the DC prediction mode included in the basic prediction mode is specified by index 0, and each directional prediction mode included in the basic prediction mode set. Are specified by indices 2-9.
 なお、各インデックスと各予測モードとの対応関係を示す情報、および、基本予測モードセットに属する各方向予測モードと各予測方向との対応関係を示す情報は、符号化データ#1を生成する動画像符号化装置、および、符号化データ#1を復号する動画像復号装置1の双方において共通のものを用いる構成とすることができる。動画像復号装置1は、それらの情報を自身の備えるメモリに格納しておき、復号されたインデックスによって指定される予測モードが、エッジベース予測モードであるのか、DC予測モードであるのか、または、方向予測モードであるのかを識別することができ、また、復号されたインデックスが指定する予測モードが方向予測モードである場合には、当該方向予測モードが何れの予測方向を指定するものであるのかを識別することができる。 Note that the information indicating the correspondence between each index and each prediction mode, and the information indicating the correspondence between each direction prediction mode belonging to the basic prediction mode set and each prediction direction are moving images that generate encoded data # 1. A common configuration can be used for both the image encoding device and the moving image decoding device 1 that decodes the encoded data # 1. The moving picture decoding apparatus 1 stores such information in its own memory, and whether the prediction mode specified by the decoded index is the edge-based prediction mode, the DC prediction mode, or It is possible to identify the direction prediction mode, and when the prediction mode specified by the decoded index is the direction prediction mode, which prediction direction the direction prediction mode specifies. Can be identified.
 また、各インデックスと各予測モードとの対応関係を示す情報、および、基本予測モードセットに属する各方向予測モードと各予測方向との対応関係を示す情報は、例えば、シーケンス毎、ピクチャ毎、または、スライス毎に、動画像符号化装置から動画像復号装置1に伝送する構成としてもよい。 The information indicating the correspondence between each index and each prediction mode, and the information indicating the correspondence between each direction prediction mode belonging to the basic prediction mode set and each prediction direction are, for example, for each sequence, for each picture, or Further, a configuration may be adopted in which, for each slice, transmission is performed from the moving image encoding device to the moving image decoding device 1.
 なお、図7(a)においては、エッジベース予測モードをインデックス1に割り当てたが、本実施形態はこれに限定されるものではなく、復号対象の画像の特性、および、エッジベース予測モードが選択される頻度等に応じて、最適なインデックスを割り当てる構成とすることができる。例えば、対象パーティションの周辺のパーティションに割り付けられた予測モードのうち、より小さいインデックスによって指定される予測モードを当該対象パーティションについての推定予測モードに設定する構成においては、インデックスのより小さな予測モードが選択される頻度が高くなる。このような構成の場合であって、復号対象の画像にエッジが多く含まれているような場合には、エッジベース予測モードに対してより小さなインデックスを割り当てることが好ましい。一方で、インデックスのより小さな予測モードが、より高い頻度で選択されるような構成の場合であって、復号対象の画像にエッジが多く含まれているような場合には、エッジベース予測モードに対してより大きなインデックスを割り当てることが好ましい。 In FIG. 7A, the edge-based prediction mode is assigned to the index 1, but the present embodiment is not limited to this, and the characteristics of the decoding target image and the edge-based prediction mode are selected. It is possible to adopt a configuration in which an optimal index is assigned in accordance with the frequency of execution. For example, in the configuration in which the prediction mode specified by a smaller index among the prediction modes assigned to the partitions around the target partition is set as the estimated prediction mode for the target partition, the prediction mode with the smaller index is selected. Will be more frequent. In the case of such a configuration, when the decoding target image includes many edges, it is preferable to assign a smaller index to the edge-based prediction mode. On the other hand, when the prediction mode having a smaller index is selected at a higher frequency and the image to be decoded includes many edges, the edge-based prediction mode is selected. It is preferable to assign a larger index to this.
 また、図7(a)においては、基本予測モードセットが、互いに異なる8つの方向予測の何れかを指定する予測モードを含んでいる場合を例に挙げたが、本実施形態はこれに限定されるものではない。例えば、基本予測モードセットとして、互いに異なる9以上の方向の何れかを指定する予測モードを含むセットを用いてもよい。そのような例としては、例えば、互いに異なる16方向の何れかを指定する予測モードや、互いに異なる32方向の何れかを指定する予測モードを含むセットが挙げられる。 Further, in FIG. 7A, the basic prediction mode set is exemplified as the case where the basic prediction mode set includes a prediction mode that specifies any one of eight different direction predictions. However, the present embodiment is limited to this. It is not something. For example, as the basic prediction mode set, a set including a prediction mode that specifies any of nine or more different directions may be used. As such an example, for example, a set including a prediction mode for designating any of 16 different directions and a prediction mode for designating any of 32 different directions can be given.
 また、基本予測モードセットに含まれる予測モードとしては、予め定められている1または複数の方向の何れか、または、1または複数の無方向の予測モードの何れか(例えば、DC予測)を指定するものであればよく、本実施形態は、基本予測モードセットに含まれる予測モードの数によって限定されるものではない。 In addition, as a prediction mode included in the basic prediction mode set, one of one or more predetermined directions or one or more non-directional prediction modes (for example, DC prediction) is designated. The present embodiment is not limited by the number of prediction modes included in the basic prediction mode set.
 (エッジベース予測モードにおける予測画像の算出処理)
 次に、エッジベース予測モードにおける、イントラ予測画像生成部12cによる予測画像の算出処理について図8の(a)~(b)を参照して、具体的に説明する。図8は、エッジベース予測モードが選択された場合の、イントラ予測画像の生成処理を説明するための図であって、図8(a)は、対象パーティションOPを、対象パーティションOPに隣接するパーティションNP2およびNP3、並びに、対象パーティションの左上の頂点を共有するパーティションNP1と共に示す図であり、図8(b)は、補正角度を指定するパラメータを、補正後の予測方向と共に示す図である。
(Prediction image calculation processing in edge-based prediction mode)
Next, prediction image calculation processing by the intra-prediction image generation unit 12c in the edge-based prediction mode will be specifically described with reference to (a) to (b) of FIG. FIG. 8 is a diagram for explaining the intra-prediction image generation process when the edge-based prediction mode is selected. FIG. 8A illustrates the target partition OP as a partition adjacent to the target partition OP. It is a figure shown with partition NP1 which shares NP2 and NP3, and the upper left vertex of a target partition, and FIG.8 (b) is a figure which shows the parameter which designates a correction angle with the prediction direction after correction | amendment.
 図8(a)においては、対象パーティションOP、および、パーティションNP1~3が何れも4×4画素である場合を示すが、本実施形態はこれに限定されるものではなく、対象パーティションOPが4×4画素以外のサイズである場合や、パーティションNP1~3が4×4画素以外のサイズである場合にも適用することができる。また、図8(a)に示すパーティションNP1~3に含まれる画素の画素値は、何れも復号済みであるものとする。 FIG. 8A shows a case where the target partition OP and the partitions NP1 to NP3 are all 4 × 4 pixels, but the present embodiment is not limited to this, and the target partition OP is four. The present invention can also be applied when the size is other than × 4 pixels or when the partitions NP1 to NP3 are other than 4 × 4 pixels. It is assumed that the pixel values of the pixels included in the partitions NP1 to NP3 shown in FIG. 8A have been decoded.
 イントラ予測画像生成部12cは、まず、パーティションNP1~3に含まれる各画素について、エッジベクトルbi(i=1~M、MはパーティションNP1~3に含まれる画素の総数)を算出する。ここで、エッジベクトルbiの算出には、下記に示すソーベルフィルタGx、およびGyを用いればよい。 First, the intra predicted image generation unit 12c calculates edge vectors b i (i = 1 to M, M is the total number of pixels included in the partitions NP1 to 3) for each pixel included in the partitions NP1 to NP3. Here, the calculation of the edge vectors b i may be used Sobel filter Gx shown below, and Gy.
Figure JPOXMLDOC01-appb-M000002
Figure JPOXMLDOC01-appb-M000002
 なお、ソーベルフィルタGx、およびGyは、それぞれ、x方向に沿った画像勾配、および、y方向に沿った画像勾配を算出するために用いられるフィルタ行列である。 The Sobel filters Gx and Gy are filter matrices used for calculating an image gradient along the x direction and an image gradient along the y direction, respectively.
 イントラ予測画像生成部12cは、算出されたx方向についての画像勾配、および、y方向に沿った画像勾配によって表される画像勾配に直交する方向をエッジ方向として算出する。 The intra predicted image generation unit 12c calculates, as the edge direction, a direction orthogonal to the calculated image gradient in the x direction and the image gradient represented by the image gradient along the y direction.
 続いて、イントラ予測画像生成部12cは、以下に示す関数T(α)
 T(α)=Σ<e,bi2
を定義する。ここで、eは、自身の方向と水平方向(x方向)とのなす角がαである単位ベクトルを表しており、記号<,>は、両者のベクトルの内積を表している。また、記号Σは、添え字iについて1からMまでの和をとることを示している。
Subsequently, the intra predicted image generation unit 12c has a function T (α) shown below.
T (α) = Σ <e, b i > 2
Define Here, e represents a unit vector whose angle between its own direction and the horizontal direction (x direction) is α, and the symbol <,> represents the inner product of both vectors. The symbol Σ indicates that the subscript i is to be summed from 1 to M.
 続いて、イントラ予測画像生成部12cは、関数T(α)を最大化する引数α*
 α*=argmaxS(α)
を算出し、α*によって表される方向を、対象パーティションについてのエッジ方向に設定する。なお、上記の説明において、角度α、および、角度α*は、水平右向きを0度とし、時計回りを正として表されるものとする(以下の角度の表現についても同様)。
Subsequently, the intra predicted image generation unit 12c sets an argument α * that maximizes the function T (α) .
α * = argmaxS (α)
And the direction represented by α * is set to the edge direction for the target partition. In the above description, it is assumed that the angle α and the angle α * are represented with the horizontal right direction being 0 degrees and the clockwise direction being positive (the same applies to the expression of the following angles).
 図8(b)は、追加インデックスAIの示す値t(t=-2、-1、0、1、2)によって指定される予測方向の例を示している。 FIG. 8B shows an example of the prediction direction specified by the value t (t = −2, −1, 0, 1, 2) indicated by the additional index AI.
 イントラ予測画像生成部12cは、エッジベース予測モードにおいて、以上のようにして決定された予測方向に、対象パーティションの周辺の画素についての復号済みの画素値を外挿することによって、当該対象パーティションにおけるイントラ予測画像Pred_Intraを生成する。なお、予測方向に沿った両側に復号済み画素があれば、それらの画素の画素値を内挿することによってイントラ予測画像Pred_Intraを生成してもよい。ここでの復号済み画像の画素は、平滑化の要否決定処理の結果に応じて、必要であれば平滑化が実施された画素である。 In the edge-based prediction mode, the intra-predicted image generation unit 12c extrapolates the decoded pixel values for pixels around the target partition in the prediction direction determined as described above, so that An intra prediction image Pred_Intra is generated. Note that if there are decoded pixels on both sides along the prediction direction, the intra predicted image Pred_Intra may be generated by interpolating the pixel values of these pixels. The pixel of the decoded image here is a pixel that has been smoothed if necessary according to the result of the necessity determination process for smoothing.
 例えば、イントラ予測画像生成部12cは、当該対象パーティションにおける予測対象画素の画素位置を始点とする、上記予測方向の逆方向を向いた仮想線分上に位置する復号済み画素のうち、当該画素に最も近い画素(以下、最近接画素とも呼ぶ)の画素値を、当該予測対象画素の画素値に設定することにより、当該対象パーティションにおけるイントラ予測画像Pred_Intraを生成する。また、最近接画素の画素値、および、最近接画素の周辺の画素の画素値を用いて算出される値を、当該予測対象画素の画素値としてもよい。 For example, the intra-predicted image generation unit 12c sets the pixel among the decoded pixels that are located on the virtual line segment that faces the reverse direction of the prediction direction, starting from the pixel position of the prediction target pixel in the target partition. The intra predicted image Pred_Intra in the target partition is generated by setting the pixel value of the closest pixel (hereinafter also referred to as the closest pixel) to the pixel value of the prediction target pixel. Further, the pixel value of the prediction target pixel may be a value calculated using the pixel value of the nearest pixel and the pixel values of the pixels around the nearest pixel.
 なお、上記の説明では、イントラ予測画像生成部12cがエッジ方向を算出する際に、対象パーティションの上辺に隣接するパーティション、対象パーティションの左辺に隣接するパーティション、および、対象パーティションの左上の頂点を共有するパーティションに属する画素の画素値を参照する場合を例に挙げたが、本実施形態はこれに限定されるものではなく、イントラ予測画像生成部12cは、より一般に、対象パーティションの周辺に設定される参照領域に属する復号済みの画素値を参照してエッジ方向を算出する構成とすることができる。 In the above description, when the intra predicted image generation unit 12c calculates the edge direction, the partition adjacent to the upper side of the target partition, the partition adjacent to the left side of the target partition, and the upper left vertex of the target partition are shared. Although the case where the pixel values of the pixels belonging to the partition to be referred to is taken as an example, the present embodiment is not limited to this, and the intra-predicted image generation unit 12c is more generally set around the target partition. The edge direction can be calculated with reference to the decoded pixel values belonging to the reference region.
 (基本予測モードにおける予測画像の算出処理)
 続いて、基本予測モードにおける、イントラ予測画像生成部12cによる予測画像の生成処理について、具体的に説明する。
(Prediction image calculation processing in basic prediction mode)
Next, the prediction image generation processing by the intra prediction image generation unit 12c in the basic prediction mode will be specifically described.
 イントラ予測画像生成部12cは、対象パーティションについて、DC予測モードが選択された場合、対象パーティションの周辺の復号済み画素値の平均値をとることによって、当該対象パーティションについてのイントラ予測画像Pred_Intraを生成する。 When the DC prediction mode is selected for the target partition, the intra predicted image generation unit 12c generates an intra predicted image Pred_Intra for the target partition by taking an average value of decoded pixel values around the target partition. .
 また、イントラ予測画像生成部12cは、インデックス2~9の何れかによって指定される方向予測モードが選択された場合、選択された方向予測モードの示す予測方向に沿って、対象パーティションの周辺の復号済み画素値を外挿することによって、当該対象パーティションについてのイントラ予測画像Pred_Intraを生成する。なお、予測方向に沿った両側に復号済み画素があれば、それらの画素の画素値を内挿することによってイントラ予測画像Pred_Intraを生成してもよい。 Further, when the direction prediction mode designated by any of the indexes 2 to 9 is selected, the intra predicted image generation unit 12c decodes the periphery of the target partition along the prediction direction indicated by the selected direction prediction mode. The intra predicted image Pred_Intra for the target partition is generated by extrapolating the completed pixel values. Note that if there are decoded pixels on both sides along the prediction direction, the intra predicted image Pred_Intra may be generated by interpolating the pixel values of these pixels.
 以下では、図7(b)を参照して、イントラ予測画像生成部12cによる予測画像の算出処理例について説明する。なお、以下の例では、対象パーティションのサイズが4×4画素であるとして説明を行うが、これは本実施形態を限定するものではない。 Hereinafter, with reference to FIG. 7B, an example of a predicted image calculation process performed by the intra predicted image generation unit 12c will be described. In the following example, the description will be made on the assumption that the size of the target partition is 4 × 4 pixels, but this does not limit the present embodiment.
 図7(b)は、4×4画素である対象パーティションの各画素(予測対象画素)と、当該対象パーティションの周辺の画素(参照画素)とを示す図である。図7(b)に示すように、予測対象画素には符号a~p、参照画素には符号A~Mを付し、画素X(Xはa~p、A~Mの何れか)の画素値をXと表すことにする。また、参照画素A~Mは、何れも復号済みであるとする。 FIG. 7B is a diagram illustrating each pixel (prediction target pixel) of the target partition, which is 4 × 4 pixels, and pixels (reference pixels) around the target partition. As shown in FIG. 7B, the prediction target pixels are denoted by reference signs a to p, the reference pixels are denoted by reference signs A to M, and a pixel X (X is any one of a to p and A to M). Let the value be represented as X. Further, it is assumed that the reference pixels A to M have all been decoded.
 (予測モード0)
 割り付けられた予測モードのインデックスが0(DC予測)である場合、イントラ予測画像生成部12cは、画素値a~pを以下の式
 a~p=ave(A,B,C,D,I,J,K,L)
によって生成する。ここで、ave(…)は、括弧内に含まれる要素の平均をとることを示している。
(Prediction mode 0)
When the index of the allocated prediction mode is 0 (DC prediction), the intra predicted image generation unit 12c converts the pixel values a to p into the following formulas a to p = ave (A, B, C, D, I, J, K, L)
Generate by. Here, ave (...) Indicates that an element included in parentheses is averaged.
 (予測モード2)
 割り付けられた予測モードのインデックスが2である場合、イントラ予測画像生成部12cは、画素値a~pを、以下の式
 a,e,i,m=A,
 b,f,j,n=B,
 c,g,k,o=C,
 d,h,l,p=D
によって生成する。
(Prediction mode 2)
When the allocated prediction mode index is 2, the intra-predicted image generation unit 12c converts the pixel values a to p into the following formulas a, e, i, m = A,
b, f, j, n = B,
c, g, k, o = C,
d, h, l, p = D
Generate by.
 (予測モード5)
 割り付けられた予測モードのインデックスが5である場合、イントラ予測画像生成部12cは、画素値a~pを以下の式
 d=(B+(C×2)+D+2)>>2,
 c,h=(A+(B×2)+C+2)>>2,
 b,g,l=(M+(A×2)+B+2)>>2,
 a,f,k,p=(I+(M×2)+A+2)>>2,
 e,j,o=(J+(I×2)+M+2)>>2,
 i,n=(K+(J×2)+I+2)>>2,
 m=(L+(K×2)+J+2)>>2
によって生成する。ここで、「>>」は右シフト演算を表し、任意の正の整数x、sに対し、x>>sの値は、x÷(2^s)の小数部分を切り捨てた値と等しい。
(Prediction mode 5)
When the allocated prediction mode index is 5, the intra-predicted image generation unit 12c converts the pixel values a to p into the following expression d = (B + (C × 2) + D + 2) >> 2,
c, h = (A + (B × 2) + C + 2) >> 2,
b, g, l = (M + (A × 2) + B + 2) >> 2,
a, f, k, p = (I + (M × 2) + A + 2) >> 2,
e, j, o = (J + (I × 2) + M + 2) >> 2,
i, n = (K + (J × 2) + I + 2) >> 2,
m = (L + (K × 2) + J + 2) >> 2
Generate by. Here, “>>” represents a right shift operation, and for any positive integer x, s, the value of x >> s is equal to the value obtained by rounding down the fractional part of x ÷ (22s).
 また、イントラ予測画像生成部12cは、上記の予測モード以外の基本予測モードに対しても、同様の方法によって画素値a~pを算出することができる。 Also, the intra predicted image generation unit 12c can calculate the pixel values a to p by the same method for the basic prediction modes other than the above prediction modes.
 また、イントラ予測画像生成部12cは、エッジベース予測モードにおいて算出された予測方向を用いて、上記の処理と略同様の処理を行うことによって、エッジベース予測モードにおけるイントラ予測画像Pred_Intraを生成することができる。 In addition, the intra predicted image generation unit 12c generates an intra predicted image Pred_Intra in the edge-based prediction mode by performing substantially the same process as described above using the prediction direction calculated in the edge-based prediction mode. Can do.
 (付記事項1)
 上述した実施の形態では、Bs値を用いて、対象画素の平滑化を行うか否かを決定したが、Bs値を用いることなく、Bs値を用いた場合と同様の効果を奏する構成とすることもできる。
(Appendix 1)
In the above-described embodiment, it is determined whether or not to smooth the target pixel using the Bs value. However, the Bs value is used without using the Bs value. You can also.
 すなわち、対象画素および該対象画素と隣接する参照画素の両方がイントラCUに含まれ、2つの画素の間がCU境界である場合は、平滑化を行う。これにより、Bs値が4以上であるか否かに応じて平滑化するか否かを切り替える場合と同様の効果を、Bs値を算出することなく奏することができる。 That is, when both the target pixel and the reference pixel adjacent to the target pixel are included in the intra CU and the CU boundary is between the two pixels, smoothing is performed. Thereby, it is possible to achieve the same effect as when switching whether or not smoothing is performed according to whether or not the Bs value is 4 or more without calculating the Bs value.
 また、対象画素および該対象画素と隣接する参照画素のいずれかがイントラCUに含まれ、2つの画素の間がCUでない処理単位の境界である場合に、平滑化を行う構成としてもよい。これにより、Bs値が3以上であるか否かに応じて平滑化するか否かを切り替える場合と同様の効果を奏することができる。 Further, the smoothing may be performed when either the target pixel or a reference pixel adjacent to the target pixel is included in the intra CU and the boundary between the two pixels is a processing unit that is not a CU. Thereby, the same effect as the case where it switches whether it smoothes according to whether Bs value is 3 or more can be show | played.
 (付記事項2)
 また、対象画素が含まれるパーティションに割り付けられた動きベクトルmvと、該対象画素と境界を挟んで反対側の参照画素が含まれるパーティションに割り付けられた動きベクトルmvとのなす角度が閾値以上の場合や、動きベクトルmvの成分の差が閾値以上の場合に、平滑化を行うこととしてもよい。
(Appendix 2)
In addition, when the angle formed by the motion vector mv assigned to the partition including the target pixel and the motion vector mv allocated to the partition including the reference pixel opposite to the target pixel is greater than or equal to the threshold value Alternatively, smoothing may be performed when the difference between the components of the motion vector mv is equal to or greater than a threshold value.
 (付記事項3)
 また、対象画素が含まれるパーティションをインター予測する際に参照される参照画像の枚数と、該対象画素と境界を挟んで反対側の参照画素が含まれるパーティションをインター予測する際に参照される参照画像の枚数とが異なる場合に、平滑化を行うこととしてもよい。また、対象画素が含まれるパーティションをインター予測する際に参照される参照画像と、該対象画素と境界を挟んで反対側の参照画素が含まれるパーティションをインター予測する際に参照される参照画像とが異なる場合に、平滑化を行うこととしてもよい。
(Appendix 3)
Also, the number of reference images that are referenced when inter-predicting a partition that includes the target pixel, and a reference that is referenced when inter-predicting a partition that includes the reference pixel on the opposite side across the boundary with the target pixel Smoothing may be performed when the number of images is different. Further, a reference image that is referred to when inter-predicting a partition including the target pixel, and a reference image that is referred to when inter-predicting a partition including the reference pixel on the opposite side across the boundary with the target pixel If they are different, smoothing may be performed.
 (付記事項4)
 また、対象画素が含まれるTUについての量子化パラメータQP(QuantizationParameter)と、該対象画素と境界を挟んで反対側の参照画素が含まれるTUについての量子化パラメータQP(QuantizationParameter)との値の差が閾値以上の場合、平滑化を行うこととしてもよい。
(Appendix 4)
Also, the difference in value between the quantization parameter QP (QuantizationParameter) for the TU including the target pixel and the quantization parameter QP (QuantizationParameter) for the TU including the reference pixel opposite to the target pixel across the boundary. When is equal to or greater than the threshold, smoothing may be performed.
 (付記事項5)
 また、対象画素が含まれるTUと、該対象画素との境界を挟んで反対側の参照画素が含まれるTUとの少なくともいずれか一方で変換係数が符号化されていない場合、平滑化を行うこととしてもよい。
(Appendix 5)
Further, smoothing is performed when the transform coefficient is not encoded in at least one of the TU including the target pixel and the TU including the reference pixel on the opposite side across the boundary with the target pixel. It is good.
 (付記事項6)
 また、上述した実施形態では、イントラ予測画像を生成する際に参照される各画素を、1:2:1の3タップ平滑化フィルタを適用することによって平滑化したが、これの代わりに5タップ以上のタップ数を有するフィルタを適用してもよい。これにより、ブロック歪の低減を高精度に行うことができる。なお、このようなフィルタの例としては、例えば、従来のデブロッキング処理に用いられているフィルタが挙げられる。
(Appendix 6)
In the above-described embodiment, each pixel that is referred to when generating an intra-predicted image is smoothed by applying a 1: 2: 1 3-tap smoothing filter. A filter having the above tap number may be applied. Thereby, the block distortion can be reduced with high accuracy. In addition, as an example of such a filter, the filter used for the conventional deblocking process is mentioned, for example.
 (付記事項7)
 上述したイントラ予測の各モードに限らず、参照画素の画素値を用いてイントラ予測画像を生成する場合には、本実施形態で説明した平滑化は有効である。また、イントラ予測モードがDC予測である場合には、参照画素は平均値としてのみ予測画像生成に用いられるため、平滑化を省略するとよい。
(Appendix 7)
The smoothing described in the present embodiment is effective not only in the above-described intra prediction modes but also in the case where an intra predicted image is generated using pixel values of reference pixels. In addition, when the intra prediction mode is DC prediction, the reference pixels are used only as an average value for prediction image generation, and therefore smoothing may be omitted.
 (動画像符号化装置2)
 本実施形態に係る動画像符号化装置(画像符号化装置)2の構成について図9を参照して説明する。動画像符号化装置2は、その一部に、H.264/MPEG-4.AVC、VCEG(Video Coding Expert Group)における共同開発用コーデックであるKTAソフトウェアに採用されている技術、および、その後継コーデックであるTMuC(Test Model under Consideration)ソフトウェアに採用されている技術を含む符号化装置である。
(Moving picture encoding device 2)
The configuration of the moving image encoding apparatus (image encoding apparatus) 2 according to the present embodiment will be described with reference to FIG. The moving image encoding apparatus 2 includes H.264 as a part thereof. H.264 / MPEG-4. Coding including technology adopted in KTA software, which is a codec for joint development in AVC and VCEG (Video Coding Expert Group), and technology adopted in TMuC (Test Model under Consideration) software, which is the successor codec Device.
 図9は、動画像符号化装置2の構成を示すブロック図である。図9に示すように、動画像符号化装置2は、予測画像生成部21、変換・量子化部22、逆量子化・逆変換部23、加算器24、フレームメモリ25、ループフィルタ26、可変長符号符号化部27、および、減算器28を備えている。また、図9に示すように、予測画像生成部21は、イントラ予測画像生成部(参照画素平滑化決定手段、平滑化手段、予測画像生成手段)21a、動きベクトル検出部21b、インター予測画像生成部21c、予測方式制御部21d、および、動きベクトル冗長性削除部21eを備えている。動画像符号化装置2は、動画像#10(符号化対象画像)を符号化することによって、符号化データ#1を生成する装置である。 FIG. 9 is a block diagram showing a configuration of the moving picture encoding apparatus 2. As illustrated in FIG. 9, the moving image encoding apparatus 2 includes a predicted image generation unit 21, a transform / quantization unit 22, an inverse quantization / inverse transform unit 23, an adder 24, a frame memory 25, a loop filter 26, a variable A long code encoding unit 27 and a subtracter 28 are provided. As shown in FIG. 9, the predicted image generation unit 21 includes an intra predicted image generation unit (reference pixel smoothing determination unit, smoothing unit, predicted image generation unit) 21a, a motion vector detection unit 21b, and an inter predicted image generation. A unit 21c, a prediction scheme control unit 21d, and a motion vector redundancy deletion unit 21e. The moving image encoding device 2 is a device that generates encoded data # 1 by encoding moving image # 10 (encoding target image).
 (予測画像生成部21)
 予測画像生成部21は、処理対象LCUを、1または複数の下位CUに再帰的に分割し、各リーフCUをさらに1または複数のパーティションに分割し、パーティション毎に、画面間予測を用いたインター予測画像Pred_Inter、または、画面内予測を用いたイントラ予測画像Pred_Intraを生成する。生成されたインター予測画像Pred_Interおよびイントラ予測画像Pred_Intraは、予測画像Predとして、加算器24および減算器28に供給される。
(Predicted image generation unit 21)
The predicted image generation unit 21 recursively divides the processing target LCU into one or a plurality of lower-order CUs, further divides each leaf CU into one or a plurality of partitions, and uses an inter-screen prediction for each partition. A predicted image Pred_Inter or an intra predicted image Pred_Intra using intra prediction is generated. The generated inter prediction image Pred_Inter and intra prediction image Pred_Intra are supplied to the adder 24 and the subtracter 28 as the prediction image Pred.
 なお、予測画像生成部21は、スキップモードの適応されたPUについては、当該PUに属する他のパラメータの符号化を省略する。また、(1)対象LCUにおける下位CUおよびパーティションへの分割の態様、(2)スキップモードを適用するか否か、および、(3)パーティション毎にインター予測画像Pred_Interおよびイントラ予測画像Pred_Intraの何れを生成するか、は、符号化効率を最適化するように決定される。 Note that the prediction image generation unit 21 omits encoding of other parameters belonging to the PU for the PU to which the skip mode is applied. Also, (1) the mode of division into lower CUs and partitions in the target LCU, (2) whether to apply the skip mode, and (3) which of the inter predicted image Pred_Inter and the intra predicted image Pred_Intra for each partition Whether to generate is determined so as to optimize the encoding efficiency.
 (イントラ予測画像生成部21a)
 イントラ予測画像生成部21aは、画面内予測によって、各パーティションに関する予測画像Pred_Intraを生成する。具体的には、(1)各パーティションついてイントラ予測に用いる予測モードを選択し、(2)選択した予測モードを用いて、復号画像Pから予測画像Pred_Intraを生成する。イントラ予測画像生成部21aは、生成したイントラ予測画像Pred_Intraを、予測方式制御部21dに供給する。
(Intra predicted image generation unit 21a)
The intra predicted image generation unit 21a generates a predicted image Pred_Intra for each partition by intra prediction. Specifically, (1) a prediction mode used for intra prediction is selected for each partition, and (2) a prediction image Pred_Intra is generated from the decoded image P using the selected prediction mode. The intra predicted image generation unit 21a supplies the generated intra predicted image Pred_Intra to the prediction method control unit 21d.
 より具体的には、イントラ予測画像生成部21aは、上述した基本予測モードセットに含まれる予測モード、および、エッジベース予測モードよりなる拡張セットに含まれる予測モードのうち、何れかの予測モードを選択し、選択した予測モードの示す方法に従って、イントラ予測画像Pred_Intraを生成する。そして、イントラ予測画像Pred_Intraを生成する前に、対象パーティションの近傍画素について、平滑化の要否を決定する。そして、平滑化を行うと決定した画素については平滑化を行ったのち、イントラ予測画像Pred_Intraを生成する。平滑化の要否決定処理は、動画像復号装置1が備えるイントラ予測画像生成部12cにおける処理と同様である。 More specifically, the intra-prediction image generation unit 21a selects any one of the prediction modes included in the above-described basic prediction mode set and the prediction mode included in the extended set including the edge-based prediction mode. The intra prediction image Pred_Intra is generated according to the method indicated by the selected prediction mode. Then, before generating the intra-predicted image Pred_Intra, whether or not smoothing is necessary for the neighboring pixels of the target partition is determined. Then, after smoothing the pixels determined to be smoothed, an intra predicted image Pred_Intra is generated. The smoothing necessity determination process is the same as the process in the intra predicted image generation unit 12c included in the video decoding device 1.
 (動きベクトル検出部21b)
 動きベクトル検出部21bは、各パーティションに関する動きベクトルmvを検出する。具体的には、(1)参照画像として利用するフィルタ済復号画像P_ALF’を選択し、(2)選択したフィルタ済復号画像P_ALF’において対象パーティションを最良近似する領域を探索することによって、対象パーティションに関する動きベクトルmvを検出する。ここで、フィルタ済復号画像P_ALF’は、既にフレーム全体の復号が完了した復号済みの復号画像に対して、ループフィルタ26による適応的フィルタ処理を施すことによって得られる画像であり、動きベクトル検出部21bは、フィルタ済復号画像P_ALF’を構成する各画素の画素値をフレームメモリ25から読み出すことができる。動きベクトル検出部21bは、検出した動きベクトルmvを、参照画像として利用したフィルタ済復号画像P_ALF’を指定する参照画像インデックスRIと共に、インター予測画像生成部21cおよび動きベクトル冗長性削除部21eに供給する。
(Motion vector detection unit 21b)
The motion vector detection unit 21b detects a motion vector mv regarding each partition. Specifically, (1) the filtered decoded image P_ALF ′ used as the reference image is selected, and (2) the target partition is searched by searching for the region that best approximates the target partition in the selected filtered decoded image P_ALF ′. Detects a motion vector mv. Here, the filtered decoded image P_ALF ′ is an image obtained by performing adaptive filter processing by the loop filter 26 on the decoded image that has already been decoded for the entire frame, and is a motion vector detection unit. 21b can read out the pixel value of each pixel constituting the filtered decoded image P_ALF ′ from the frame memory 25. The motion vector detection unit 21b supplies the detected motion vector mv to the inter prediction image generation unit 21c and the motion vector redundancy deletion unit 21e together with the reference image index RI that specifies the filtered decoded image P_ALF ′ used as the reference image. To do.
 (インター予測画像生成部21c)
 インター予測画像生成部21cは、画面間予測によって、各インター予測パーティションに関する動き補償画像mcを生成する。具体的には、動きベクトル検出部21bから供給された動きベクトルmvを用いて、動きベクトル検出部21bから供給された参照画像インデックスRIによって指定されるフィルタ済復号画像P_ALF’から動き補償画像mcを生成する。動きベクトル検出部21bと同様に、インター予測画像生成部21cは、フィルタ済復号画像P_ALF’を構成する各画素の画素値をフレームメモリ25から読み出すことができる。インター予測画像生成部21cは、生成した動き補償画像mc(インター予測画像Pred_Inter)を、動きベクトル検出部21bから供給された参照画像インデックスRIと共に、予測方式制御部21dに供給する。
(Inter prediction image generation unit 21c)
The inter prediction image generation unit 21c generates a motion compensated image mc related to each inter prediction partition by inter-screen prediction. Specifically, the motion compensation image mc is obtained from the filtered decoded image P_ALF ′ designated by the reference image index RI supplied from the motion vector detection unit 21b, using the motion vector mv supplied from the motion vector detection unit 21b. Generate. Similar to the motion vector detection unit 21b, the inter prediction image generation unit 21c can read out the pixel value of each pixel constituting the filtered decoded image P_ALF ′ from the frame memory 25. The inter prediction image generation unit 21c supplies the generated motion compensated image mc (inter prediction image Pred_Inter) together with the reference image index RI supplied from the motion vector detection unit 21b to the prediction method control unit 21d.
 (予測方式制御部21d)
 予測方式制御部21dは、イントラ予測画像Pred_Intraおよびインター予測画像Pred_Interを符号化対象画像と比較し、イントラ予測を行うかインター予測を行うかを選択する。イントラ予測を選択した場合、予測方式制御部21dは、イントラ予測画像Pred_Intraを予測画像Predとして加算器24および減算器28に供給すると共に、イントラ予測画像生成部21aから供給されるイントラ予測パラメータPP_Intraを可変長符号符号化部27に供給する。一方、インター予測を選択した場合、予測方式制御部21dは、インター予測画像Pred_Interを予測画像Predとして加算器24および減算器28に供給すると共に、参照画像インデックスRI、並びに、後述する動きベクトル冗長性削除部21eから供給された推定動きベクトルインデックスPMVIおよび動きベクトル残差MVDをインター予測パラメータPP_Interとして可変長符号符号化部27に供給する。
(Prediction method controller 21d)
The prediction scheme control unit 21d compares the intra predicted image Pred_Intra and the inter predicted image Pred_Inter with the encoding target image and selects whether to perform intra prediction or inter prediction. When intra prediction is selected, the prediction scheme control unit 21d supplies the intra prediction image Pred_Intra as the prediction image Pred to the adder 24 and the subtracter 28, and also uses the intra prediction parameter PP_Intra supplied from the intra prediction image generation unit 21a. This is supplied to the variable length code encoding unit 27. On the other hand, when the inter prediction is selected, the prediction scheme control unit 21d supplies the inter prediction image Pred_Inter as the prediction image Pred to the adder 24 and the subtractor 28, and the reference image index RI and motion vector redundancy described later. The estimated motion vector index PMVI and the motion vector residual MVD supplied from the deletion unit 21e are supplied to the variable length code encoding unit 27 as an inter prediction parameter PP_Inter.
 (動きベクトル冗長性削除部21e)
 動きベクトル冗長性削除部21eは、動きベクトル検出部21bによって検出された動きベクトルmvにおける冗長性を削除する。具体的には、(1)動きベクトルmvの推定に用いる推定方法を選択し、(2)選択した推定方法に従って推定動きベクトルpmvを導出し、(3)動きベクトルmvから推定動きベクトルpmvを減算することにより動きベクトル残差MVDを生成する。動きベクトル冗長性削除部21eは、生成した動きベクトル残差MVDを、選択した推定方法を示す推定動きベクトルインデックスPMVIと共に、予測方式制御部21dに供給する。
(Motion vector redundancy deleting unit 21e)
The motion vector redundancy deletion unit 21e deletes redundancy in the motion vector mv detected by the motion vector detection unit 21b. Specifically, (1) an estimation method used for estimating the motion vector mv is selected, (2) an estimated motion vector pmv is derived according to the selected estimation method, and (3) the estimated motion vector pmv is subtracted from the motion vector mv. As a result, a motion vector residual MVD is generated. The motion vector redundancy deleting unit 21e supplies the generated motion vector residual MVD to the prediction method control unit 21d together with the estimated motion vector index PMVI indicating the selected estimation method.
 (変換・量子化部22)
 変換・量子化部22は、(1)符号化対象画像(原画像)から予測画像Predを減算した予測残差Dをブロック(変換単位)毎にDCT変換(Discrete Cosine Transform)し、(2)DCT変換により得られたDCT係数を量子化し、(3)量子化により得られた量子化予測残差QDを可変長符号符号化部27および逆量子化・逆変換部23に供給する。なお、変換・量子化部22は、(1)量子化の際に用いる量子化ステップQPをTU毎に選択し、(2)選択した量子化ステップQPの大きさを示す量子化パラメータ差分Δqpを可変長符号符号化部27に供給し、(3)選択した量子化ステップQPを逆量子化・逆変換部23に供給する。ここで、量子化パラメータ差分Δqpとは、DCT変換/量子化するTUに関する量子化パラメータqp(例えばQP=2pq/6)の値から、直前にDCT変換/量子化したTUに関する量子化パラメータqp’の値を減算して得られる差分値のことを指す。
(Transformation / quantization unit 22)
The transform / quantization unit 22 (1) performs DCT transform (Discrete Cosine Transform) on the prediction residual D obtained by subtracting the predicted image Pred from the encoding target image (original image) for each block (transform unit), and (2) The DCT coefficient obtained by the DCT transform is quantized, and (3) the quantized prediction residual QD obtained by the quantization is supplied to the variable length code encoder 27 and the inverse quantization / inverse transform unit 23. The transform / quantization unit 22 (1) selects a quantization step QP to be used for quantization for each TU, and (2) sets a quantization parameter difference Δqp indicating the size of the selected quantization step QP. This is supplied to the variable-length code encoding unit 27, and (3) the selected quantization step QP is supplied to the inverse quantization / inverse transform unit 23. Here, the quantization parameter difference Δqp is the quantization parameter qp for the TU that has been DCT transformed / quantized immediately before, based on the value of the quantization parameter qp (for example, QP = 2 pq / 6 ) for the TU to be DCT transformed / quantized. The difference value obtained by subtracting the value of '.
 (逆量子化・逆変換部23)
 逆量子化・逆変換部23は、(1)量子化予測残差QDを逆量子化し、(2)逆量子化によって得られたDCT係数を逆DCT(Discrete Cosine Transform)変換し、(3)逆DCT変換によって得られた予測残差Dを加算器24に供給する。量子化予測残差QDを逆量子化する際には、変換・量子化部22から供給された量子化ステップQPを利用する。なお、逆量子化・逆変換部23から出力される予測残差Dは、変換・量子化部22に入力される予測残差Dに量子化誤差が加わったものであるが、ここでは簡単のために共通の呼称を用いる。
(Inverse quantization / inverse transform unit 23)
The inverse quantization / inverse transform unit 23 (1) inversely quantizes the quantized prediction residual QD, (2) performs inverse DCT (Discrete Cosine Transform) transformation on the DCT coefficient obtained by the inverse quantization, and (3) The prediction residual D obtained by the inverse DCT transform is supplied to the adder 24. When the quantization prediction residual QD is inversely quantized, the quantization step QP supplied from the transform / quantization unit 22 is used. Note that the prediction residual D output from the inverse quantization / inverse transform unit 23 is obtained by adding a quantization error to the prediction residual D input to the transform / quantization unit 22. Common names are used for this purpose.
 (加算器24)
 加算器24は、予測方式制御部21dにて選択された予測画像Predを、逆量子化・逆変換部23にて生成された予測残差Dに加算することによって、(局所)復号画像Pを生成する。加算器24にて生成された(局所)復号画像Pは、ループフィルタ26に供給されると共にフレームメモリ25に格納され、イントラ予測における参照画像として利用される。
(Adder 24)
The adder 24 adds the predicted image Pred selected by the prediction scheme control unit 21d to the prediction residual D generated by the inverse quantization / inverse transform unit 23, thereby obtaining the (local) decoded image P. Generate. The (local) decoded image P generated by the adder 24 is supplied to the loop filter 26 and stored in the frame memory 25, and is used as a reference image in intra prediction.
 (可変長符号符号化部27)
 可変長符号符号化部27は、(1)変換・量子化部22から供給された量子化予測残差QD並びにΔqp、(2)予測方式制御部21dから供給された量子化パラメータPP(インター予測パラメータPP_Inter、および、イントラ予測パラメータPP_Intra)、および、(3)ループフィルタ26から供給されたフィルタパラメータFPを可変長符号化することによって、符号化データ#1を生成する。
(Variable-length code encoding unit 27)
The variable length code encoding unit 27 (1) the quantization prediction residual QD and Δqp supplied from the transform / quantization unit 22, and (2) the quantization parameter PP (inter prediction) supplied from the prediction scheme control unit 21d. The parameter PP_Inter and the intra prediction parameter PP_Intra) and (3) the filter parameter FP supplied from the loop filter 26 are variable-length-encoded to generate encoded data # 1.
 なお、可変長符号符号化部27による具体的な符号化方式としては、算術符号化/復号の一方式であるCABAC(Context-based Adaptive Binary Arithmetic Coding)、または、非算術符号化/復号の一方式であるCAVLC(Context-based Adaptive VLC)が用いられる。 As a specific encoding method by the variable length code encoding unit 27, CABAC (Context-based Adaptive Binary Arithmetic Coding) which is one arithmetic coding / decoding method, or one non-arithmetic encoding / decoding method is used. CAVLC (Context-based Adaptive VLC) which is a method is used.
 可変長符号符号化部27は、ピクチャ毎に、CABACおよびCAVLCの何れの符号化方式を用いるのかを決定し、決定した符号化方式により符号化すると共に、決定した符号化方式を指定する符号化モード情報(entropy_coding_mode_flag)を符号化データ#1のピクチャヘッダPHに含める。 The variable length code encoding unit 27 determines which encoding method of CABAC or CAVLC is to be used for each picture, performs encoding using the determined encoding method, and specifies the determined encoding method The mode information (entropy_coding_mode_flag) is included in the picture header PH of the encoded data # 1.
 (減算器28)
 減算器28は、予測方式制御部21dにて選択された予測画像Predを、符号化対象画像から減算することによって、予測残差Dを生成する。減算器28にて生成された予測残差Dは、変換・量子化部22によってDCT変換/量子化される。
(Subtractor 28)
The subtracter 28 generates the prediction residual D by subtracting the prediction image Pred selected by the prediction method control unit 21d from the encoding target image. The prediction residual D generated by the subtracter 28 is DCT transformed / quantized by the transform / quantization unit 22.
 (ループフィルタ26)
 ループフィルタ26は、フレームメモリ25から復号画像Pを読み出し、復号画像Pのパーティション境界およびブロック境界の何れか一方若しくは双方におけるブロックノイズ低減処理(デブロッキング処理)を施す。また、ループフィルタ26は、ブロックノイズ低減処理が施された復号画像に対して、適応的に算出されるフィルタパラメータFPを用いた適応的なフィルタ処理を施し、当該適応的なフィルタ処理が施された復号画像Pをフィルタ済復号画像P_ALFとしてフレームメモリ25に出力する。フィルタ済復号画像P_ALFは、主に、インター予測画像生成部21cにおいて、参照画像として用いられる。
(Loop filter 26)
The loop filter 26 reads the decoded image P from the frame memory 25 and performs block noise reduction processing (deblocking processing) at one or both of the partition boundary and the block boundary of the decoded image P. The loop filter 26 performs adaptive filter processing using the adaptively calculated filter parameter FP on the decoded image subjected to block noise reduction processing, and the adaptive filter processing is performed. The decoded image P is output to the frame memory 25 as a filtered decoded image P_ALF. The filtered decoded image P_ALF is mainly used as a reference image in the inter predicted image generation unit 21c.
 (応用例)
 上述した動画像復号装置1および動画像符号化装置2は、動画像の送信、受信、記録、再生を行う各種装置に搭載して利用することができる。なお、動画像は、カメラ等により撮像された自然動画像であってもよいし、コンピュータ等により生成された人工動画像(CGおよびGUIを含む)であってもよい。
(Application examples)
The moving picture decoding apparatus 1 and the moving picture encoding apparatus 2 described above can be used by being mounted on various apparatuses that perform moving picture transmission, reception, recording, and reproduction. The moving image may be a natural moving image captured by a camera or the like, or may be an artificial moving image (including CG and GUI) generated by a computer or the like.
 まず、上述した動画像復号装置1および動画像符号化装置2を、動画像の送信及び受信に利用できることを、図12を参照して説明する。 First, it will be described with reference to FIG. 12 that the moving picture decoding apparatus 1 and the moving picture encoding apparatus 2 described above can be used for transmission and reception of moving pictures.
 図12(a)は、動画像符号化装置2を搭載した送信装置Aの構成を示したブロック図である。図12(a)に示すように、送信装置Aは、動画像を符号化することによって符号化データを得る符号化部A1と、符号化部A1が得た符号化データで搬送波を変調することによって変調信号を得る変調部A2と、変調部A2が得た変調信号を送信する送信部A3と、を備えている。上述した動画像符号化装置2は、この符号化部A1として利用される。 FIG. 12A is a block diagram illustrating a configuration of the transmission apparatus A in which the moving picture encoding apparatus 2 is mounted. As shown in FIG. 12A, the transmitting apparatus A encodes a moving image, obtains encoded data, and modulates a carrier wave with the encoded data obtained by the encoding unit A1. A modulation unit A2 that obtains a modulation signal by the transmission unit A2 and a transmission unit A3 that transmits the modulation signal obtained by the modulation unit A2. The moving image encoding device 2 described above is used as the encoding unit A1.
 送信装置Aは、符号化部A1に入力する動画像の供給源として、動画像を撮像するカメラA4、動画像を記録した記録媒体A5、動画像を外部から入力するための入力端子A6、および画像を生成または加工する画像処理部A7を更に備えていてもよい。図12(a)においては、これら全てを送信装置Aが備えた構成を例示しているが、一部を省略しても構わない。 The transmission apparatus A has a camera A4 that captures a moving image, a recording medium A5 that records the moving image, an input terminal A6 for inputting the moving image from the outside, as a supply source of the moving image that is input to the encoding unit A1. You may further provide image processing part A7 which produces | generates or processes an image. FIG. 12A illustrates a configuration in which the transmission apparatus A includes all of these, but some of them may be omitted.
 なお、記録媒体A5は、符号化されていない動画像を記録したものであってもよいし、伝送用の符号化方式とは異なる記録用の符号化方式で符号化された動画像を記録したものであってもよい。後者の場合、記録媒体A5と符号化部A1との間に、記録媒体A5から読み出した符号化データを記録用の符号化方式に従って復号する復号部(不図示)を介在させるとよい。 The recording medium A5 may be a recording of a non-encoded moving image, or a recording of a moving image encoded using a recording encoding scheme different from the transmission encoding scheme. It may be a thing. In the latter case, a decoding unit (not shown) for decoding the encoded data read from the recording medium A5 according to the recording encoding method may be interposed between the recording medium A5 and the encoding unit A1.
 図12(b)は、動画像復号装置1を搭載した受信装置Bの構成を示したブロック図である。図12(b)に示すように、受信装置Bは、変調信号を受信する受信部B1と、受信部B1が受信した変調信号を復調することによって符号化データを得る復調部B2と、復調部B2が得た符号化データを復号することによって動画像を得る復号部B3と、を備えている。上述した動画像復号装置1は、この復号部B3として利用される。 FIG. 12B is a block diagram illustrating a configuration of the receiving device B on which the moving image decoding device 1 is mounted. As illustrated in FIG. 12B, the receiving device B includes a receiving unit B1 that receives a modulated signal, a demodulating unit B2 that obtains encoded data by demodulating the modulated signal received by the receiving unit B1, and a demodulating unit. And a decoding unit B3 that obtains a moving image by decoding the encoded data obtained by B2. The moving picture decoding apparatus 1 described above is used as the decoding unit B3.
 受信装置Bは、復号部B3が出力する動画像の供給先として、動画像を表示するディスプレイB4、動画像を記録するための記録媒体B5、及び、動画像を外部に出力するための出力端子B6を更に備えていてもよい。図12(b)においては、これら全てを受信装置Bが備えた構成を例示しているが、一部を省略しても構わない。 The receiving apparatus B has a display B4 for displaying a moving image, a recording medium B5 for recording the moving image, and an output terminal for outputting the moving image as a supply destination of the moving image output from the decoding unit B3. B6 may be further provided. FIG. 12B illustrates a configuration in which the receiving apparatus B includes all of these, but a part of the configuration may be omitted.
 なお、記録媒体B5は、符号化されていない動画像を記録するためのものであってもよいし、伝送用の符号化方式とは異なる記録用の符号化方式で符号化されたものであってもよい。後者の場合、復号部B3と記録媒体B5との間に、復号部B3から取得した動画像を記録用の符号化方式に従って符号化する符号化部(不図示)を介在させるとよい。 Note that the recording medium B5 may be for recording an unencoded moving image, or is encoded by a recording encoding method different from the transmission encoding method. May be. In the latter case, an encoding unit (not shown) that encodes the moving image acquired from the decoding unit B3 in accordance with the recording encoding method may be interposed between the decoding unit B3 and the recording medium B5.
 なお、変調信号を伝送する伝送媒体は、無線であってもよいし、有線であってもよい。また、変調信号を伝送する伝送態様は、放送(ここでは、送信先が予め特定されていない送信態様を指す)であってもよいし、通信(ここでは、送信先が予め特定されている送信態様を指す)であってもよい。すなわち、変調信号の伝送は、無線放送、有線放送、無線通信、及び有線通信の何れによって実現してもよい。 Note that the transmission medium for transmitting the modulation signal may be wireless or wired. Further, the transmission mode for transmitting the modulated signal may be broadcasting (here, a transmission mode in which the transmission destination is not specified in advance) or communication (here, transmission in which the transmission destination is specified in advance). Refers to the embodiment). That is, the transmission of the modulation signal may be realized by any of wireless broadcasting, wired broadcasting, wireless communication, and wired communication.
 例えば、地上デジタル放送の放送局(放送設備など)/受信局(テレビジョン受像機など)は、変調信号を無線放送で送受信する送信装置A/受信装置Bの一例である。また、ケーブルテレビ放送の放送局(放送設備など)/受信局(テレビジョン受像機など)は、変調信号を有線放送で送受信する送信装置A/受信装置Bの一例である。 For example, a terrestrial digital broadcast broadcasting station (such as broadcasting equipment) / receiving station (such as a television receiver) is an example of a transmitting apparatus A / receiving apparatus B that transmits and receives modulated signals by wireless broadcasting. A broadcasting station (such as broadcasting equipment) / receiving station (such as a television receiver) for cable television broadcasting is an example of a transmitting device A / receiving device B that transmits and receives a modulated signal by cable broadcasting.
 また、インターネットを用いたVOD(Video On Demand)サービスや動画共有サービスなどのサーバ(ワークステーションなど)/クライアント(テレビジョン受像機、パーソナルコンピュータ、スマートフォンなど)は、変調信号を通信で送受信する送信装置A/受信装置Bの一例である(通常、LANにおいては伝送媒体として無線又は有線の何れかが用いられ、WANにおいては伝送媒体として有線が用いられる)。ここで、パーソナルコンピュータには、デスクトップ型PC、ラップトップ型PC、及びタブレット型PCが含まれる。また、スマートフォンには、多機能携帯電話端末も含まれる。 Also, a server (workstation etc.) / Client (television receiver, personal computer, smart phone etc.) such as VOD (Video On Demand) service and video sharing service using the Internet is a transmitting device for transmitting and receiving modulated signals by communication. This is an example of A / reception device B (usually, either wireless or wired is used as a transmission medium in a LAN, and wired is used as a transmission medium in a WAN). Here, the personal computer includes a desktop PC, a laptop PC, and a tablet PC. The smartphone also includes a multi-function mobile phone terminal.
 なお、動画共有サービスのクライアントは、サーバからダウンロードした符号化データを復号してディスプレイに表示する機能に加え、カメラで撮像した動画像を符号化してサーバにアップロードする機能を有している。すなわち、動画共有サービスのクライアントは、送信装置A及び受信装置Bの双方として機能する。 In addition to the function of decoding the encoded data downloaded from the server and displaying it on the display, the video sharing service client has a function of encoding a moving image captured by the camera and uploading it to the server. That is, the client of the video sharing service functions as both the transmission device A and the reception device B.
 次に、上述した動画像復号装置1および動画像符号化装置2を、動画像の記録及び再生に利用できることを、図13を参照して説明する。 Next, the fact that the above-described moving picture decoding apparatus 1 and moving picture encoding apparatus 2 can be used for recording and reproduction of moving pictures will be described with reference to FIG.
 図13(a)は、上述した動画像復号装置1を搭載した記録装置Cの構成を示したブロック図である。図13(a)に示すように、記録装置Cは、動画像を符号化することによって符号化データを得る符号化部C1と、符号化部C1が得た符号化データを記録媒体Mに書き込む書込部C2と、を備えている。上述した動画像符号化装置2は、この符号化部C1として利用される。 FIG. 13A is a block diagram showing a configuration of a recording apparatus C on which the above-described moving picture decoding apparatus 1 is mounted. As shown in FIG. 13 (a), the recording device C encodes a moving image to obtain encoded data, and writes the encoded data obtained by the encoding unit C1 to the recording medium M. And a writing unit C2. The moving image encoding device 2 described above is used as the encoding unit C1.
 なお、記録媒体Mは、(1)HDD(Hard Disk Drive)やSSD(Solid State Drive)などのように、記録装置Cに内蔵されるタイプのものであってもよいし、(2)SDメモリカードやUSB(Universal Serial Bus)フラッシュメモリなどのように、記録装置Cに接続されるタイプのものであってもよいし、(3)DVD(Digital Versatile Disc)やBD(Blu-ray Disc:登録商標)などのように、記録装置Cに内蔵されたドライブ装置(不図示)に装填されるものであってもよい。 The recording medium M may be of a type built in the recording device C, such as (1) HDD (Hard Disk Drive) or SSD (Solid State Drive), or (2) SD memory. It may be of the type connected to the recording device C, such as a card or USB (Universal Serial Bus) flash memory, or (3) DVD (Digital Versatile Disc) or BD (Blu-ray Disc: registration) (Trademark) or the like may be mounted on a drive device (not shown) built in the recording apparatus C.
 また、記録装置Cは、符号化部C1に入力する動画像の供給源として、動画像を撮像するカメラC3、動画像を外部から入力するための入力端子C4、動画像を受信するための受信部C5、および、画像を生成または加工する画像処理部C6を更に備えていてもよい。図13(a)においては、これら全てを記録装置Cが備えた構成を例示しているが、一部を省略しても構わない。 The recording apparatus C also serves as a moving image supply source to be input to the encoding unit C1, a camera C3 that captures moving images, an input terminal C4 for inputting moving images from the outside, and reception for receiving moving images. A unit C5 and an image processing unit C6 that generates or processes an image may be further provided. FIG. 13A illustrates a configuration in which the recording apparatus C includes all of these, but a part of the configuration may be omitted.
 なお、受信部C5は、符号化されていない動画像を受信するものであってもよいし、記録用の符号化方式とは異なる伝送用の符号化方式で符号化された符号化データを受信するものであってもよい。後者の場合、受信部C5と符号化部C1との間に、伝送用の符号化方式で符号化された符号化データを復号する伝送用復号部(不図示)を介在させるとよい。 The receiving unit C5 may receive an unencoded moving image, or receives encoded data encoded by a transmission encoding method different from the recording encoding method. You may do. In the latter case, a transmission decoding unit (not shown) that decodes encoded data encoded by the transmission encoding method may be interposed between the reception unit C5 and the encoding unit C1.
 このような記録装置Cとしては、例えば、DVDレコーダ、BDレコーダ、HD(Hard Disk)レコーダなどが挙げられる(この場合、入力端子C4又は受信部C5が動画像の主な供給源となる)。また、カムコーダ(この場合、カメラC3が動画像の主な供給源となる)、パーソナルコンピュータ(この場合、受信部C5または画像処理部C6が動画像の主な供給源となる)、スマートフォン(この場合、カメラC3又は受信部C5が動画像の主な供給源となる)なども、このような記録装置Cの一例である。 Examples of such a recording device C include a DVD recorder, a BD recorder, and an HD (Hard Disk) recorder (in this case, the input terminal C4 or the receiving unit C5 is a main source of moving images). In addition, a camcorder (in this case, the camera C3 is a main source of moving images), a personal computer (in this case, the receiving unit C5 or the image processing unit C6 is a main source of moving images), a smartphone (this In this case, the camera C3 or the receiving unit C5 is a main source of moving images).
 図13(b)は、上述した動画像復号装置1を搭載した再生装置Dの構成を示したブロックである。図13(b)に示すように、再生装置Dは、記録媒体Mに書き込まれた符号化データを読み出す読出部D1と、読出部D1が読み出した符号化データを復号することによって動画像を得る復号部D2と、を備えている。上述した動画像復号装置1は、この復号部D2として利用される。 FIG. 13B is a block diagram showing the configuration of the playback device D equipped with the above-described video decoding device 1. As shown in FIG. 13 (b), the playback device D obtains a moving image by decoding the read data D1 that reads the encoded data written on the recording medium M and the read data read by the read unit D1. And a decoding unit D2. The moving picture decoding apparatus 1 described above is used as the decoding unit D2.
 なお、記録媒体Mは、(1)HDDやSSDなどのように、再生装置Dに内蔵されるタイプのものであってもよいし、(2)SDメモリカードやUSBフラッシュメモリなどのように、再生装置Dに接続されるタイプのものであってもよいし、(3)DVDやBDなどのように、再生装置Dに内蔵されたドライブ装置(不図示)に装填されるものであってもよい。 The recording medium M may be of a type built in the playback device D such as (1) HDD or SSD, or (2) such as an SD memory card or USB flash memory. It may be of a type connected to the playback device D, or (3) may be loaded into a drive device (not shown) built in the playback device D, such as DVD or BD. Good.
 また、再生装置Dは、復号部D2が出力する動画像の供給先として、動画像を表示するディスプレイD3、動画像を外部に出力するための出力端子D4、及び、動画像を送信する送信部D5を更に備えていてもよい。図13(b)においては、これら全てを再生装置Dが備えた構成を例示しているが、一部を省略しても構わない。 Further, the playback device D has a display D3 for displaying a moving image, an output terminal D4 for outputting the moving image to the outside, and a transmitting unit for transmitting the moving image as a supply destination of the moving image output by the decoding unit D2. D5 may be further provided. FIG. 13B illustrates a configuration in which the playback apparatus D includes all of these, but some of them may be omitted.
 なお、送信部D5は、符号化されていない動画像を送信するものであってもよいし、記録用の符号化方式とは異なる伝送用の符号化方式で符号化された符号化データを送信するものであってもよい。後者の場合、復号部D2と送信部D5との間に、動画像を伝送用の符号化方式で符号化する符号化部(不図示)を介在させるとよい。 The transmission unit D5 may transmit a non-encoded moving image, or transmits encoded data encoded by a transmission encoding method different from the recording encoding method. You may do. In the latter case, an encoding unit (not shown) that encodes a moving image with a transmission encoding method may be interposed between the decoding unit D2 and the transmission unit D5.
 このような再生装置Dとしては、例えば、DVDプレイヤ、BDプレイヤ、HDDプレイヤなどが挙げられる(この場合、テレビジョン受像機等が接続される出力端子D4が動画像の主な供給先となる)。また、テレビジョン受像機(この場合、ディスプレイD3が動画像の主な供給先となる)、デスクトップ型PC(この場合、出力端子D4又は送信部D5が動画像の主な供給先となる)、ラップトップ型又はタブレット型PC(この場合、ディスプレイD3又は送信部D5が動画像の主な供給先となる)、スマートフォン(この場合、ディスプレイD3又は送信部D5が動画像の主な供給先となる)、デジタルサイネージ(電子看板や電子掲示板等とも称され、ディスプレイD3又は送信部D5が動画像の主な供給先となる)なども、このような再生装置Dの一例である。 Examples of such a playback device D include a DVD player, a BD player, and an HDD player (in this case, an output terminal D4 to which a television receiver or the like is connected is a main moving image supply destination). . In addition, a television receiver (in this case, the display D3 is a main destination of moving images), a desktop PC (in this case, the output terminal D4 or the transmission unit D5 is a main destination of moving images), Laptop type or tablet type PC (in this case, display D3 or transmission unit D5 is the main video image supply destination), smartphone (in this case, display D3 or transmission unit D5 is the main video image supply destination) ), Digital signage (also referred to as an electronic signboard or an electronic bulletin board, and the display D3 or the transmission unit D5 is the main supply destination of moving images) is an example of such a playback device D.
 (ソフトウェアによる構成)
 最後に、動画像復号装置1、および動画像符号化装置2の各ブロック、特に可変長符号復号部11、予測画像生成部(動きベクトル復元部12a、インター予測画像生成部12b、イントラ予測画像生成部12c、予測方式決定部12d)12、逆量子化・逆変換部13、加算器14、フレームメモリ15、ループフィルタ16、予測画像生成部(イントラ予測画像生成部21a、動きベクトル検出部21b、インター予測画像生成部21c、予測方式制御部21d、動きベクトル冗長性削除部21e)21、変換・量子化部22、逆量子化・逆変換部23、加算器24、フレームメモリ25、ループフィルタ26、可変長符号符号化部27、および、減算器28は、集積回路(ICチップ)上に形成された論理回路によってハードウェア的に実現していてもよいし、CPU(central processing unit)を用いてソフトウェア的に実現してもよい。
(Configuration by software)
Finally, each block of the moving picture decoding apparatus 1 and the moving picture encoding apparatus 2, in particular, a variable length code decoding unit 11, a prediction image generation unit (motion vector restoration unit 12 a, inter prediction image generation unit 12 b, intra prediction image generation Unit 12c, prediction method determination unit 12d) 12, inverse quantization / inverse conversion unit 13, adder 14, frame memory 15, loop filter 16, prediction image generation unit (intra prediction image generation unit 21a, motion vector detection unit 21b, Inter prediction image generation unit 21c, prediction method control unit 21d, motion vector redundancy deletion unit 21e) 21, transform / quantization unit 22, inverse quantization / inverse transform unit 23, adder 24, frame memory 25, loop filter 26 The variable-length code encoding unit 27 and the subtracter 28 are implemented in hardware by a logic circuit formed on an integrated circuit (IC chip). It may be realized, or may be implemented in software using a CPU (central processing unit).
 後者の場合、動画像復号装置1、および動画像符号化装置2は、各機能を実現する制御プログラムの命令を実行するCPU、上記プログラムを格納したROM(read only memory)、上記プログラムを展開するRAM(random access memory)、上記プログラムおよび各種データを格納するメモリ等の記憶装置(記録媒体)などを備えている。そして、本発明の目的は、上述した機能を実現するソフトウェアである動画像復号装置1、および動画像符号化装置2の制御プログラムのプログラムコード(実行形式プログラム、中間コードプログラム、ソースプログラム)をコンピュータで読み取り可能に記録した記録媒体を、上記の動画像復号装置1、および動画像符号化装置2に供給し、そのコンピュータ(またはCPUやMPU(microprocessor unit))が記録媒体に記録されているプログラムコードを読み出し実行することによっても、達成可能である。 In the latter case, the video decoding device 1 and the video encoding device 2 develop a CPU that executes instructions of a control program that realizes each function, a ROM (read (only memory) that stores the program, and the program. A RAM (random access memory), a storage device (recording medium) such as a memory for storing the program and various data, and the like are provided. The object of the present invention is to provide a computer program code (execution format program, intermediate code program, source program) of a control program for the video decoding device 1 and the video encoding device 2 that is software that implements the functions described above. Is supplied to the above-described moving picture decoding apparatus 1 and moving picture encoding apparatus 2 and the computer (or CPU or MPU (microprocessor unit)) is recorded on the recording medium. It can also be achieved by reading and executing the code.
 上記記録媒体としては、例えば、磁気テープやカセットテープ等のテープ類、フロッピー(登録商標)ディスク/ハードディスク等の磁気ディスクやCD-ROM(compact disc read-only memory)/MO(magneto-optical)/MD(Mini Disc)/DVD(digital versatile disk)/CD-R(CD Recordable)等の光ディスクを含むディスク類、ICカード(メモリカードを含む)/光カード等のカード類、マスクROM/EPROM(erasable programmable read-only memory)/EEPROM(electrically erasable and programmable read-only memory)/フラッシュROM等の半導体メモリ類、あるいはPLD(Programmable logic device)やFPGA(Field Programmable Gate Array)等の論理回路類などを用いることができる。 Examples of the recording medium include tapes such as a magnetic tape and a cassette tape, a magnetic disk such as a floppy (registered trademark) disk / hard disk, a CD-ROM (compact disk-read-only memory) / MO (magneto-optical) / Discs including optical discs such as MD (Mini Disc) / DVD (digital versatile disc) / CD-R (CD Recordable), IC cards (including memory cards) / optical cards, mask ROM / EPROM (erasable) Programmable read-only memory) / EEPROM (electrically erasable and programmable programmable read-only memory) / semiconductor memory such as flash ROM, or logic circuits such as PLD (Programmable logic device) and FPGA (Field Programmable Gate Array) be able to.
 また、動画像復号装置1、および動画像符号化装置2を通信ネットワークと接続可能に構成し、上記プログラムコードを通信ネットワークを介して供給してもよい。この通信ネットワークは、プログラムコードを伝送可能であればよく、特に限定されない。例えば、インターネット、イントラネット、エキストラネット、LAN(local area network)、ISDN(integrated services digital network)、VAN(value-added network)、CATV(community antenna television)通信網、仮想専用網(virtual private network)、電話回線網、移動体通信網、衛星通信網等が利用可能である。また、この通信ネットワークを構成する伝送媒体も、プログラムコードを伝送可能な媒体であればよく、特定の構成または種類のものに限定されない。例えば、IEEE(institute of electrical and electronic engineers)1394、USB、電力線搬送、ケーブルTV回線、電話線、ADSL(asynchronous digital subscriber loop)回線等の有線でも、IrDA(infrared data association)やリモコンのような赤外線、Bluetooth(登録商標)、IEEE802.11無線、HDR(high data rate)、NFC(Near Field Communication)、DLNA(Digital Living Network Alliance)、携帯電話網、衛星回線、地上波デジタル網等の無線でも利用可能である。なお、本発明は、上記プログラムコードが電子的な伝送で具現化された、搬送波に埋め込まれたコンピュータデータ信号の形態でも実現され得る。 Further, the moving picture decoding apparatus 1 and the moving picture encoding apparatus 2 may be configured to be connectable to a communication network, and the program code may be supplied via the communication network. The communication network is not particularly limited as long as it can transmit the program code. For example, Internet, intranet, extranet, LAN (local area network), ISDN (integrated area services digital area), VAN (value-added area network), CATV (community area antenna television) communication network, virtual area private network (virtual area private network), A telephone line network, a mobile communication network, a satellite communication network, etc. can be used. The transmission medium constituting the communication network may be any medium that can transmit the program code, and is not limited to a specific configuration or type. For example, IEEE (institute of electrical and electronic engineers) 1394, USB, power line carrier, cable TV line, telephone line, ADSL (asynchronous digital subscriber loop) line, etc. wired such as IrDA (infrared data association) or remote control , Bluetooth (registered trademark), IEEE802.11 wireless, HDR (high data rate), NFC (Near field communication), DLNA (Digital Living Network Alliance), mobile phone network, satellite line, terrestrial digital network, etc. Is possible. The present invention can also be realized in the form of a computer data signal embedded in a carrier wave in which the program code is embodied by electronic transmission.
 (その他)
 なお、上述した実施の形態は、以下のように表現することもできる。
(Other)
The above-described embodiment can also be expressed as follows.
 本発明に係る画像復号装置は、予測画像を、予測単位ごとに、同じ画面内の別の位置にある参照画素の画素値から該予測画像内の画素の画素値を予測することにより生成し、生成した予測画像を符号化データから復号した予測残差に加算して復号画像を生成する画像復号装置であって、上記参照画素それぞれについて、該参照画素と隣接している他の画素との間で平滑化を行うか否かを決定する参照画素平滑化決定手段と、上記参照画素平滑化決定手段が平滑化を行うと決定した参照画素を上記他の画素との間で平滑化を行う平滑化手段と、上記参照画素のうち、上記参照画素平滑化決定手段が平滑化を行うと決定した参照画素については、上記平滑化手段が平滑化した後の参照画素を用いて上記予測画像を生成する予測画像生成手段と、を備えていることを特徴としている。 The image decoding apparatus according to the present invention generates a prediction image by predicting a pixel value of a pixel in the prediction image from a pixel value of a reference pixel at another position in the same screen for each prediction unit, An image decoding apparatus that generates a decoded image by adding a generated predicted image to a prediction residual decoded from encoded data, and for each reference pixel, between the reference pixel and another adjacent pixel The reference pixel smoothing determining means for determining whether or not to perform smoothing, and the smoothing for smoothing the reference pixel determined to be smoothed by the reference pixel smoothing determining means with the other pixels. Among the reference pixels and the reference pixels determined to be smoothed by the reference pixel smoothing determination unit, the predicted image is generated using the reference pixels smoothed by the smoothing unit. Predicted image generation means for It is characterized in that it comprises.
 上記の構成によれば、予測単位ごとに同じ画面内の別の位置にある参照画素の画素値から該予測画像内の画素の画素値を予測するイントラ予測に用いる参照画素を平滑化するか否かを決定し、平滑化を行うと決定した参照画素のみ平滑化を行って予測画像を生成する。 According to said structure, whether the reference pixel used for the intra prediction which predicts the pixel value of the pixel in the said prediction image from the pixel value of the reference pixel in another position in the same screen for every prediction unit is smoothed Is determined, and only the reference pixels determined to be smoothed are smoothed to generate a predicted image.
 よって、参照画素ごとに平滑化を行うか、行わないかを決定することができる。したがって、従来のように、予測単位ごとに平滑化の有無が決定されることにより、当該予測単位の参照画素全てについて平滑化が行われてしまうということ、または参照画素全てについて平滑化が行われないということを防止することができる。 Therefore, it is possible to determine whether to perform smoothing for each reference pixel. Therefore, as in the prior art, whether or not smoothing is determined for each prediction unit results in smoothing for all reference pixels of the prediction unit, or smoothing for all reference pixels. It is possible to prevent the absence.
 また、参照画素ごとに平滑化の有無を決定することができるので、平滑化が不要な参照画素まで平滑化をしてしまうということがなく、より高精度な予測画像を生成することができる。 In addition, since it is possible to determine whether or not smoothing is performed for each reference pixel, it is possible to generate a predicted image with higher accuracy without performing smoothing to reference pixels that do not require smoothing.
 したがって、イントラ予測に用いられる参照画素に対して、従来よりも適切に平滑化を実行することができる。 Therefore, it is possible to perform smoothing more appropriately for the reference pixels used for intra prediction than before.
 なお、予測単位は、実施形態において説明するPUであってもよいし、PUを分割したパーティションであってもよい。 Note that the prediction unit may be a PU described in the embodiment or a partition obtained by dividing the PU.
 本発明に係る画像復号装置では、上記参照画素平滑化決定手段は、決定対象の参照画素が、処理対象の予測単位に接している複数の単位領域間の境界近傍に存在するとき、該参照画素の平滑化を行うと決定するものであることが好ましい。 In the image decoding device according to the present invention, the reference pixel smoothing determining means is configured such that, when the reference pixel to be determined exists in the vicinity of a boundary between a plurality of unit regions in contact with the prediction unit to be processed, the reference pixel It is preferable to determine that smoothing is performed.
 ここで、単位領域とは、実施形態において説明するCU、TU、PU、PUを分割したパーティション、TUを分割したブロックのいずれであってもよい。また、境界近傍とは、境界から所定の距離(例えば2画素)の範囲をいう。 Here, the unit area may be any of a partition obtained by dividing a CU, a TU, a PU, and a PU described in the embodiment, and a block obtained by dividing a TU. Further, the vicinity of the boundary means a range of a predetermined distance (for example, two pixels) from the boundary.
 そして、単位領域間の境界は、高周波の歪が出やすい。よって、上記の構成によれば、境界近傍の参照画素について平滑化を行うと決定するので、この高周波の歪を適切に低減することができる。 And, the boundary between unit areas is likely to cause high-frequency distortion. Therefore, according to the above configuration, since it is determined that the reference pixels near the boundary are to be smoothed, this high-frequency distortion can be appropriately reduced.
 本発明に係る画像復号装置では、上記参照画素平滑化決定手段は、決定対象となる参照画素について、後の処理においてデブロッキングフィルタが適用される場合のブロック境界強度が閾値を超えると、該参照画素の平滑化を行うと決定するものであってもよい。 In the image decoding device according to the present invention, the reference pixel smoothing determining unit, when the block boundary strength when a deblocking filter is applied in a later process for a reference pixel to be determined exceeds a threshold, It may be determined that the pixel is smoothed.
 上記の構成によれば、参照画素に対し、後の処理においてデブロッキングフィルタが適用されるときのブロック境界強度が閾値を超えると、平滑化を行うと決定する。 According to the above configuration, when the block boundary strength when the deblocking filter is applied in the subsequent processing exceeds the threshold for the reference pixel, it is determined that smoothing is performed.
 ブロック境界強度の高いデブロッキングフィルタが適用される画素は、ブロック歪が大きい画素である。よって、ブロック歪が大きい参照画素について平滑化を行うことができる。 A pixel to which a deblocking filter having a high block boundary strength is applied is a pixel having a large block distortion. Therefore, it is possible to perform smoothing for a reference pixel having a large block distortion.
 本発明に係る画像復号装置では、上記参照画素平滑化決定手段は、決定対象の参照画素が含まれている単位領域および、該単位領域と上記境界で接している隣接単位領域の少なくともいずれか一方がイントラ予測によって予測画像が生成された単位領域であるとき、上記参照画素の平滑化を行うと決定するものであってもよい。 In the image decoding device according to the present invention, the reference pixel smoothing determining means includes at least one of a unit region including a reference pixel to be determined and an adjacent unit region in contact with the unit region at the boundary. May be determined to perform the smoothing of the reference pixel when the predicted region is generated by intra prediction.
 決定対象の参照画素が含まれている単位領域および、該単位領域と上記境界で接している隣接単位領域の少なくともいずれか一方がイントラ予測によって予測画像が生成された単位領域である場合、これらの単位領域間の境界にはブロック歪が出やすい。よって、上記の構成によれば、ブロック歪が出やすい参照画素を平滑化することができる。 When at least one of a unit region including a reference pixel to be determined and an adjacent unit region that is in contact with the unit region at the boundary is a unit region in which a prediction image is generated by intra prediction, these Block distortion tends to occur at the boundary between unit areas. Therefore, according to the above configuration, it is possible to smooth the reference pixels that are likely to cause block distortion.
 また、決定対象の参照画素が含まれている単位領域および、該単位領域と上記境界で接している隣接単位領域の少なくともいずれか一方がイントラ予測によって予測画像が生成された単位領域の境界近傍の参照画素は、デブロッキングフィルタが適用される場合、ブロック境界強度が高いデブロッキングフィルタが適用される。よって、上記の構成によれば、ブロック境界強度が高いデブロッキングフィルタが適用される参照画素を、ブロック境界強度を算出することなく平滑化することができる。 In addition, at least one of the unit region including the reference pixel to be determined and the adjacent unit region that is in contact with the unit region at the boundary is near the boundary of the unit region where the prediction image is generated by intra prediction. When a deblocking filter is applied to the reference pixel, a deblocking filter having a high block boundary strength is applied. Therefore, according to said structure, the reference pixel to which a deblocking filter with high block boundary intensity | strength is applied can be smoothed, without calculating block boundary intensity | strength.
 本発明に係る画像復号装置では、上記参照画素平滑化決定手段は、決定対象の参照画素が含まれている単位領域および、該単位領域と上記境界で接している隣接単位領域の双方がインター予測によって予測画像が生成された単位領域であるとき、上記それぞれの単位領域の割り付けられた動きベクトルがなす角度が閾値を超えると、上記参照画素の平滑化を行うと決定するものであってもよい。 In the image decoding apparatus according to the present invention, the reference pixel smoothing determining unit performs inter prediction on both the unit region including the reference pixel to be determined and the adjacent unit region in contact with the unit region at the boundary. If the angle formed by the motion vector assigned to each unit area exceeds a threshold when the predicted image is generated by the unit area, the reference pixel may be determined to be smoothed. .
 予測単位間の動きベクトルのなす角度が大きい場合、当該予測単位間の境界の画素は、歪が大きい可能性が高い。よって、上記の構成によれば、歪が大きいと考えられる参照画素を平滑化することができる。 When the angle formed by the motion vectors between the prediction units is large, the pixel at the boundary between the prediction units is likely to be distorted. Therefore, according to said structure, the reference pixel considered that distortion is large can be smoothed.
 本発明に係る画像復号装置では、上記参照画素平滑化決定手段は、決定対象の参照画素が含まれている単位領域および、該単位領域と上記境界で接している隣接単位領域の双方がインター予測によって予測画像が生成された単位領域であるとき、上記それぞれの単位領域に割り付けられた動きベクトルの成分の差が閾値を超えると、上記参照画素の平滑化を行うと決定するものであってもよい。 In the image decoding apparatus according to the present invention, the reference pixel smoothing determining unit performs inter prediction on both the unit region including the reference pixel to be determined and the adjacent unit region in contact with the unit region at the boundary. If the difference between the motion vector components assigned to each unit area exceeds a threshold when the predicted image is generated by the unit area, the reference pixel is determined to be smoothed. Good.
 単位領域間の動きベクトルの成分の差が大きい場合、当該単位領域間の境界の画素は、歪が大きい可能性が高い。よって、上記の構成によれば、歪が大きいと考えられる参照画素を平滑化することができる。 When there is a large difference in motion vector components between unit areas, the pixel at the boundary between the unit areas is likely to be distorted. Therefore, according to said structure, the reference pixel considered that distortion is large can be smoothed.
 本発明に係る画像復号装置では、上記参照画素平滑化決定手段は、決定対象の参照画素が含まれている単位領域および、該単位領域と上記境界で接している隣接単位領域の双方がインター予測によって予測画像が生成された単位領域である場合、上記それぞれの単位領域におけるインター予測に用いた参照画像枚数が異なるとき、上記参照画素の平滑化を行うと決定するものであってもよい。 In the image decoding apparatus according to the present invention, the reference pixel smoothing determining unit performs inter prediction on both the unit region including the reference pixel to be determined and the adjacent unit region in contact with the unit region at the boundary. In the case of the unit region in which the predicted image is generated by the above, when the number of reference images used for inter prediction in each unit region is different, it may be determined that the reference pixel is smoothed.
 インター予測に用いた参照画素の枚数が異なる単位領域間の境界の画素は、歪が大きい可能性が高い。よって、上記の構成によれば、歪が大きいと考えられる参照画素を平滑化することができる。 画素 Pixels at the boundary between unit areas with different numbers of reference pixels used for inter prediction are highly likely to be distorted. Therefore, according to said structure, the reference pixel considered that distortion is large can be smoothed.
 本発明に係る画像復号装置では、上記参照画素平滑化決定手段は、決定対象の参照画素が含まれている単位領域および、該単位領域と上記境界で接している隣接単位領域の双方がインター予測によって予測画像が生成された予測単位である場合、上記それぞれの単位領域におけるインター予測に用いた参照画像が異なるとき、上記参照画素の平滑化を行うと決定するものであってもよい。 In the image decoding apparatus according to the present invention, the reference pixel smoothing determining unit performs inter prediction on both the unit region including the reference pixel to be determined and the adjacent unit region in contact with the unit region at the boundary. In the case where the prediction unit is the prediction unit in which the prediction image is generated, the reference pixel may be determined to be smoothed when the reference image used for the inter prediction in each unit region is different.
 インター予測に用いた参照画像が異なる単位領域間の境界の画素は、歪が大きい可能性が高い。よって、上記の構成によれば、歪が大きいと考えられる参照画素を平滑化することができる。 画素 Pixels at the boundary between unit areas with different reference images used for inter prediction are likely to be distorted. Therefore, according to said structure, the reference pixel considered that distortion is large can be smoothed.
 本発明に係る画像復号装置では、上記参照画素平滑化決定手段は、決定対象の参照画素が含まれている単位領域および、該単位領域と上記境界で接している隣接単位領域の量子化値の差が閾値を超えると、参照画素の平滑化を行うと決定するものであってもよい。 In the image decoding apparatus according to the present invention, the reference pixel smoothing determining means includes a unit area including the reference pixel to be determined and quantization values of adjacent unit areas that are in contact with the unit area at the boundary. When the difference exceeds the threshold value, it may be determined that the reference pixel is smoothed.
 単位領域の量子化値の差が大きい場合、当該単位領域間の境界の画素は、歪が大きい可能性が高い。よって、上記の構成によれば、歪が大きいと考えられる参照画素を平滑化することができる。 When there is a large difference in quantization value between unit areas, the pixel at the boundary between the unit areas is likely to be distorted. Therefore, according to said structure, the reference pixel considered that distortion is large can be smoothed.
 本発明に係る画像復号装置では、上記参照画素平滑化決定手段は、決定対象の参照画素がエッジ画素のとき、該参照画素の平滑化を行うと決定するものであってもよい。 In the image decoding apparatus according to the present invention, the reference pixel smoothing determining means may determine that the reference pixel is smoothed when the reference pixel to be determined is an edge pixel.
 上記の構成によれば、決定対象の参照画素がエッジ画素のとき、該参照画素の平滑化を行う。エッジ画素は、ブロック歪が出ている可能性が非常に高い。よって、ブロック歪が出ている可能性が高い参照画素を平滑化することができる。 According to the above configuration, when the reference pixel to be determined is an edge pixel, the reference pixel is smoothed. Edge pixels are very likely to have block distortion. Therefore, it is possible to smooth the reference pixels that are likely to have block distortion.
 本発明に係る画像復号装置では、上記符号化データには、イントラ予測に用いる参照画素を平滑化すること示す平滑化情報が含まれており、上記平滑化手段は、上記平滑化情報を検出すると、処理対象の予測単位に接している複数の単位領域間の境界近傍に存在する参照画素を平滑化するものであってもよい。 In the image decoding apparatus according to the present invention, the encoded data includes smoothing information indicating that a reference pixel used for intra prediction is smoothed, and the smoothing unit detects the smoothing information. Alternatively, the reference pixel existing in the vicinity of the boundary between the plurality of unit regions that are in contact with the prediction unit to be processed may be smoothed.
 上記の構成によれば、平滑化情報によって、参照画素を平滑化する。よって、特定の条件を満たしているか否かの決定処理が不要となり、処理効率を向上させることができる。 According to the above configuration, the reference pixel is smoothed by the smoothing information. Therefore, it is not necessary to determine whether or not a specific condition is satisfied, and the processing efficiency can be improved.
 本発明に係る画像符号化装置は、予測画像を、予測単位ごとに、同じ画面内の別の位置にある参照画素の画素値から該予測画像内の画素の画素値を予測することにより生成し、生成した予測画像と原画像との予測残差を符号化することによって符号化データを生成する画像符号化装置であって、上記参照画素それぞれについて、該参照画素と隣接している他の画素との間で平滑化を行うか否かを決定する参照画素平滑化決定手段と、上記参照画素平滑化決定手段が平滑化を行うと決定した参照画素を上記他の画素との間で平滑化を行う平滑化手段と、上記参照画素のうち、上記参照画素平滑化決定手段が平滑化を行うと決定した参照画素については、上記平滑化手段が平滑化した後の参照画素を用いて上記予測画像を生成する予測画像生成手段と、を備えていることを特徴としている。 The image encoding device according to the present invention generates a prediction image by predicting a pixel value of a pixel in the prediction image from a pixel value of a reference pixel at another position in the same screen for each prediction unit. An image encoding device that generates encoded data by encoding a prediction residual between a generated predicted image and an original image, and for each of the reference pixels, another pixel adjacent to the reference pixel The reference pixel smoothing determining means for determining whether to perform smoothing between the reference pixel and the reference pixel determined to be smoothed by the reference pixel smoothing determining means between the other pixels. Among the reference pixels, the reference pixel determined to be smoothed by the reference pixel smoothing determining means is used for the prediction using the reference pixel after the smoothing means smoothes. Predictive image generation to generate images It is characterized in that it comprises a stage, a.
 上記の構成によれば、予測単位ごとに同じ画面内の別の位置にある参照画素の画素値から該予測画像内の画素の画素値を予測するイントラ予測に用いる参照画素を平滑化するか否かを決定し、平滑化を行うと決定した参照画素のみ平滑化を行って予測画像を生成する。 According to said structure, whether the reference pixel used for the intra prediction which predicts the pixel value of the pixel in the said prediction image from the pixel value of the reference pixel in another position in the same screen for every prediction unit is smoothed Is determined, and only the reference pixels determined to be smoothed are smoothed to generate a predicted image.
 よって、参照画素ごとに平滑化を行うか、行わないかを決定することができる。したがって、従来のように、予測単位ごとに平滑化の有無が決定されることにより、当該予測単位の参照画素全てについて平滑化が行われてしまうということ、または参照画素全てについて平滑化が行われないということを防止することができる。 Therefore, it is possible to determine whether to perform smoothing for each reference pixel. Therefore, as in the prior art, whether or not smoothing is determined for each prediction unit results in smoothing for all reference pixels of the prediction unit, or smoothing for all reference pixels. It is possible to prevent the absence.
 また、参照画素ごとに平滑化の有無を決定することができるので、平滑化が不要な参照画素まで平滑化をしてしまうということがなく、より高精度な予測画像を生成することができる。 In addition, since it is possible to determine whether or not smoothing is performed for each reference pixel, it is possible to generate a predicted image with higher accuracy without performing smoothing to reference pixels that do not require smoothing.
 したがって、イントラ予測に用いられる参照画素に対して、従来よりも適切に平滑化を実行することができる。 Therefore, it is possible to perform smoothing more appropriately for the reference pixels used for intra prediction than before.
 なお、予測単位は、実施形態において説明するPUであってもよいし、PUを分割したパーティションであってもよい。 Note that the prediction unit may be a PU described in the embodiment or a partition obtained by dividing the PU.
 本発明は上述した実施の形態に限定されるものではなく、請求項に示した範囲で種々の変更が可能である。すなわち、請求項に示した範囲で適宜変更した技術的手段を組み合わせて得られる実施形態についても本発明の技術的範囲に含まれる。 The present invention is not limited to the above-described embodiment, and various modifications can be made within the scope of the claims. That is, embodiments obtained by combining technical means appropriately modified within the scope of the claims are also included in the technical scope of the present invention.
 本発明は、符号化データを復号する復号装置、および、符号化データを生成する符号化装置に好適に適用することができる。また、符号化装置によって生成され、復号装置によって参照される符号化データのデータ構造に好適に適用することができる。 The present invention can be suitably applied to a decoding device that decodes encoded data and an encoding device that generates encoded data. Further, the present invention can be suitably applied to the data structure of encoded data generated by the encoding device and referenced by the decoding device.
  1  動画像復号装置(画像復号装置)
  2  動画像符号化装置(画像符号化装置)
 12  予測画像生成部
 12b インター予測画像生成部
 12c イントラ予測画像生成部(参照画素平滑化決定手段、平滑化手段、予測画像生成手段)
 21  予測画像生成部
 21a イントラ予測画像生成部(参照画素平滑化決定手段、平滑化手段、予測画像生成手段)
 21c インター予測画像生成部
1 video decoding device (image decoding device)
2 Video encoding device (image encoding device)
12 prediction image generation unit 12b inter prediction image generation unit 12c intra prediction image generation unit (reference pixel smoothing determination unit, smoothing unit, prediction image generation unit)
21 prediction image generation unit 21a intra prediction image generation unit (reference pixel smoothing determination unit, smoothing unit, prediction image generation unit)
21c Inter prediction image generation unit

Claims (12)

  1.  予測画像を、予測単位ごとに、同じ画面内の別の位置にある参照画素の画素値から該予測画像内の画素の画素値を予測することにより生成し、生成した予測画像を符号化データから復号した予測残差に加算して復号画像を生成する画像復号装置であって、
     上記参照画素それぞれについて、該参照画素と隣接している他の画素との間で平滑化を行うか否かを決定する参照画素平滑化決定手段と、
     上記参照画素平滑化決定手段が平滑化を行うと決定した参照画素を上記他の画素との間で平滑化を行う平滑化手段と、
     上記参照画素のうち、上記参照画素平滑化決定手段が平滑化を行うと決定した参照画素については、上記平滑化手段が平滑化した後の参照画素を用いて上記予測画像を生成する予測画像生成手段と、を備えていることを特徴とする画像復号装置。
    A prediction image is generated for each prediction unit by predicting a pixel value of a pixel in the prediction image from a pixel value of a reference pixel at another position in the same screen, and the generated prediction image is generated from encoded data. An image decoding device that generates a decoded image by adding to the decoded prediction residual,
    For each of the reference pixels, reference pixel smoothing determining means for determining whether or not to perform smoothing between other pixels adjacent to the reference pixel;
    Smoothing means for smoothing the reference pixel determined to be smoothed by the reference pixel smoothing determining means with the other pixels;
    Of the reference pixels, for the reference pixels determined to be smoothed by the reference pixel smoothing determination means, predicted image generation for generating the predicted image using the reference pixels smoothed by the smoothing means And an image decoding device.
  2.  上記参照画素平滑化決定手段は、決定対象の参照画素が、処理対象の予測単位に接している複数の単位領域間の境界近傍に存在するとき、該参照画素の平滑化を行うと決定することを特徴とする請求項1に記載の画像復号装置。 The reference pixel smoothing determining unit determines that the reference pixel is to be smoothed when the reference pixel to be determined exists in the vicinity of a boundary between a plurality of unit regions that are in contact with the prediction unit to be processed. The image decoding apparatus according to claim 1.
  3.  上記参照画素平滑化決定手段は、決定対象となる参照画素について、後の処理においてデブロッキングフィルタが適用される場合のブロック境界強度が閾値を超えると、該参照画素の平滑化を行うと決定することを特徴とする請求項2に記載の画像復号装置。 The reference pixel smoothing determining unit determines that the reference pixel is to be smoothed when the block boundary strength when the deblocking filter is applied in a later process exceeds a threshold for the reference pixel to be determined. The image decoding apparatus according to claim 2.
  4.  上記参照画素平滑化決定手段は、決定対象の参照画素が含まれている単位領域および、該単位領域と上記境界で接している隣接単位領域の少なくともいずれか一方がイントラ予測によって予測画像が生成された単位領域であるとき、上記参照画素の平滑化を行うと決定することを特徴とする請求項2または3に記載の画像復号装置。 The reference pixel smoothing determining unit generates a predicted image by intra prediction of at least one of a unit region including a reference pixel to be determined and an adjacent unit region that is in contact with the unit region at the boundary. The image decoding apparatus according to claim 2, wherein when it is a unit area, the reference pixel is determined to be smoothed.
  5.  上記参照画素平滑化決定手段は、決定対象の参照画素が含まれている単位領域および、該単位領域と上記境界で接している隣接単位領域の双方がインター予測によって予測画像が生成された単位領域であるとき、上記それぞれの単位領域に割り付けられた動きベクトルがなす角度が閾値を超えると、上記参照画素の平滑化を行うと決定することを特徴とする請求項2または3に記載の画像復号装置。 The reference pixel smoothing determining means includes a unit region in which a prediction image is generated by inter prediction of both a unit region including a reference pixel to be determined and an adjacent unit region that is in contact with the unit region at the boundary. 4. The image decoding according to claim 2, wherein when the angle formed by the motion vector assigned to each unit region exceeds a threshold, it is determined that the reference pixel is smoothed. 5. apparatus.
  6.  上記参照画素平滑化決定手段は、決定対象の参照画素が含まれている単位領域および、該単位領域と上記境界で接している隣接単位領域の双方がインター予測によって予測画像が生成された単位領域であるとき、上記それぞれの単位領域に割り付けられた動きベクトルの成分の差が閾値を超えると、上記参照画素の平滑化を行うと決定することを特徴とする請求項2または3に記載の画像復号装置。 The reference pixel smoothing determining means includes a unit region in which a prediction image is generated by inter prediction of both a unit region including a reference pixel to be determined and an adjacent unit region that is in contact with the unit region at the boundary. 4. The image according to claim 2, wherein when the difference between the motion vector components allocated to each of the unit regions exceeds a threshold, the reference pixel is determined to be smoothed. 5. Decoding device.
  7.  上記参照画素平滑化決定手段は、決定対象の参照画素が含まれている単位領域および、該単位領域と上記境界で接している隣接単位領域の双方がインター予測によって予測画像が生成された単位領域である場合、上記それぞれの単位領域におけるインター予測に用いた参照画像枚数が異なるとき、上記参照画素の平滑化を行うと決定することを特徴とする請求項2または3に記載の画像復号装置。 The reference pixel smoothing determining means includes a unit region in which a prediction image is generated by inter prediction of both a unit region including a reference pixel to be determined and an adjacent unit region that is in contact with the unit region at the boundary. The image decoding apparatus according to claim 2, wherein when the number of reference images used for inter prediction in each of the unit regions is different, the reference pixel is determined to be smoothed.
  8.  上記参照画素平滑化決定手段は、決定対象の参照画素が含まれている単位領域および、該単位領域と上記境界で接している隣接単位領域の双方がインター予測によって予測画像が生成された単位領域である場合、上記それぞれの単位領域におけるインター予測に用いた参照画像が異なるとき、上記参照画素の平滑化を行うと決定することを特徴とする請求項2または3に記載の画像復号装置。 The reference pixel smoothing determining means includes a unit region in which a prediction image is generated by inter prediction of both a unit region including a reference pixel to be determined and an adjacent unit region that is in contact with the unit region at the boundary. 4, when the reference images used for inter prediction in the respective unit regions are different, it is determined that the reference pixels are smoothed.
  9.  上記参照画素平滑化決定手段は、決定対象の参照画素が含まれている単位領域および、該単位領域と上記境界で接している隣接単位領域の量子化値の差が閾値を超えると、参照画素の平滑化を行うと決定することを特徴とする請求項2または3に記載の画像復号装置。 The reference pixel smoothing determining means, when the difference between the quantization values of the unit area including the reference pixel to be determined and the adjacent unit area in contact with the unit area at the boundary exceeds a threshold, The image decoding apparatus according to claim 2, wherein it is determined that smoothing is performed.
  10.  上記参照画素平滑化決定手段は、決定対象の参照画素がエッジ画素のとき、該参照画素の平滑化を行うと決定することを特徴とする請求項1に記載の画像復号装置。 The image decoding apparatus according to claim 1, wherein the reference pixel smoothing determining means determines that the reference pixel is smoothed when the reference pixel to be determined is an edge pixel.
  11.  上記符号化データには、イントラ予測に用いる参照画素を平滑化すること示す平滑化情報が含まれており、
     上記平滑化手段は、上記平滑化情報を検出すると、処理対象の予測単位に接している複数の単位領域間の境界近傍に存在する参照画素を平滑化することを特徴とする請求項1に記載の画像復号装置。
    The encoded data includes smoothing information indicating that the reference pixels used for intra prediction are smoothed,
    The said smoothing means smoothes the reference pixel which exists in the boundary vicinity between several unit area | regions which contact | connects the prediction unit of a process target, if the said smoothing information is detected. Image decoding apparatus.
  12.  予測画像を、予測単位ごとに、同じ画面内の別の位置にある参照画素の画素値から該予測画像内の画素の画素値を予測することにより生成し、生成した予測画像と原画像との予測残差を符号化することによって符号化データを生成する画像符号化装置であって、
     上記参照画素それぞれについて、該参照画素と隣接している他の画素との間で平滑化を行うか否かを決定する参照画素平滑化決定手段と、
     上記参照画素平滑化決定手段が平滑化を行うと決定した参照画素を上記他の画素との間で平滑化を行う平滑化手段と、
     上記参照画素のうち、上記参照画素平滑化決定手段が平滑化を行うと決定した参照画素については、上記平滑化手段が平滑化した後の参照画素を用いて上記予測画像を生成する予測画像生成手段と、を備えていることを特徴とする画像符号化装置。
    A prediction image is generated for each prediction unit by predicting a pixel value of a pixel in the prediction image from a pixel value of a reference pixel at another position in the same screen, and the generated prediction image and the original image An image encoding device that generates encoded data by encoding a prediction residual,
    For each of the reference pixels, reference pixel smoothing determining means for determining whether or not to perform smoothing between other pixels adjacent to the reference pixel;
    Smoothing means for smoothing the reference pixel determined to be smoothed by the reference pixel smoothing determining means with the other pixels;
    Of the reference pixels, for the reference pixels determined to be smoothed by the reference pixel smoothing determination means, predicted image generation for generating the predicted image using the reference pixels smoothed by the smoothing means And an image encoding device.
PCT/JP2011/078332 2010-12-09 2011-12-07 Image decoding device and image coding device WO2012077719A1 (en)

Applications Claiming Priority (2)

Application Number Priority Date Filing Date Title
JP2010-274668 2010-12-09
JP2010274668 2010-12-09

Publications (1)

Publication Number Publication Date
WO2012077719A1 true WO2012077719A1 (en) 2012-06-14

Family

ID=46207205

Family Applications (1)

Application Number Title Priority Date Filing Date
PCT/JP2011/078332 WO2012077719A1 (en) 2010-12-09 2011-12-07 Image decoding device and image coding device

Country Status (1)

Country Link
WO (1) WO2012077719A1 (en)

Cited By (8)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
WO2014045651A1 (en) * 2012-09-24 2014-03-27 株式会社エヌ・ティ・ティ・ドコモ Video prediction encoding device, video prediction encoding method, video prediction decoding device and video prediction decoding method
JP2014512129A (en) * 2011-04-01 2014-05-19 アイベックス・ピイティ・ホールディングス・カンパニー・リミテッド Video decoding method in intra prediction mode
JP2016012928A (en) * 2015-08-21 2016-01-21 株式会社Nttドコモ Moving image prediction encoding device, moving image prediction encoding method, moving image prediction decoding device, and moving image prediction decoding method
JP2016136786A (en) * 2016-04-28 2016-07-28 株式会社Nttドコモ Moving image prediction encoder, moving image prediction encoding method, moving image prediction decoder and moving image prediction decoding method
JP2017112629A (en) * 2017-02-03 2017-06-22 株式会社Nttドコモ Moving image prediction decoding device and moving image prediction decoding method
JP2018050320A (en) * 2017-11-07 2018-03-29 株式会社Nttドコモ Moving image prediction decoding method
JP2019013040A (en) * 2018-09-20 2019-01-24 株式会社Nttドコモ Moving image predicting and decoding method
US11812060B2 (en) 2016-05-13 2023-11-07 Interdigital Vc Holdings, Inc. Method and device for deblocking filtering a boundary within an intra predicted block

Citations (4)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
JP2004336705A (en) * 2003-02-21 2004-11-25 Matsushita Electric Ind Co Ltd Moving picture encoding method, moving picture decoding method, and program
JP2005512419A (en) * 2001-11-30 2005-04-28 ローベルト ボツシユ ゲゼルシヤフト ミツト ベシユレンクテル ハフツング Image block directivity prediction method
WO2007108254A1 (en) * 2006-03-17 2007-09-27 Sharp Kabushiki Kaisha Moving picture coding device and moving picture decoding device
WO2010061515A1 (en) * 2008-11-26 2010-06-03 株式会社日立製作所 Dynamic image encoding device, encoding method, dynamic image decoding device, and decoding method

Patent Citations (4)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
JP2005512419A (en) * 2001-11-30 2005-04-28 ローベルト ボツシユ ゲゼルシヤフト ミツト ベシユレンクテル ハフツング Image block directivity prediction method
JP2004336705A (en) * 2003-02-21 2004-11-25 Matsushita Electric Ind Co Ltd Moving picture encoding method, moving picture decoding method, and program
WO2007108254A1 (en) * 2006-03-17 2007-09-27 Sharp Kabushiki Kaisha Moving picture coding device and moving picture decoding device
WO2010061515A1 (en) * 2008-11-26 2010-06-03 株式会社日立製作所 Dynamic image encoding device, encoding method, dynamic image decoding device, and decoding method

Cited By (26)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
JP2017085592A (en) * 2011-04-01 2017-05-18 アイベックス・ピイティ・ホールディングス・カンパニー・リミテッド Video decoding method in intra-prediction mode
JP2018057030A (en) * 2011-04-01 2018-04-05 アイベックス・ピイティ・ホールディングス・カンパニー・リミテッド Method of decoding video in intra- prediction mode
JP2019033516A (en) * 2011-04-01 2019-02-28 アイベックス・ピイティ・ホールディングス・カンパニー・リミテッド Video encoding method in intra-prediction mode
JP2018057029A (en) * 2011-04-01 2018-04-05 アイベックス・ピイティ・ホールディングス・カンパニー・リミテッド Method of decoding video in intra- prediction mode
JP2014512129A (en) * 2011-04-01 2014-05-19 アイベックス・ピイティ・ホールディングス・カンパニー・リミテッド Video decoding method in intra prediction mode
JP2016123118A (en) * 2011-04-01 2016-07-07 アイベックス・ピイティ・ホールディングス・カンパニー・リミテッド Method of decoding video in intra- prediction mode
RU2602978C1 (en) * 2012-09-24 2016-11-20 Нтт Докомо, Инк. Video predictive encoding device, video predictive encoding method, video predictive decoding device and video predictive decoding method
JP2014064249A (en) * 2012-09-24 2014-04-10 Ntt Docomo Inc Moving image prediction encoder, moving image prediction encoding method, moving image prediction decoder, and moving image prediction decoding method
CN104604238A (en) * 2012-09-24 2015-05-06 株式会社Ntt都科摩 Video prediction encoding device, video prediction encoding method, video prediction decoding device and video prediction decoding method
KR101799846B1 (en) 2012-09-24 2017-11-22 가부시키가이샤 엔.티.티.도코모 Video prediction encoding device, video prediction encoding method, video prediction decoding device and video prediction decoding method
TWI562614B (en) * 2012-09-24 2016-12-11 Ntt Docomo Inc
CN104604238B (en) * 2012-09-24 2017-03-22 株式会社Ntt都科摩 Video prediction encoding device and method, and video prediction decoding device and method
KR101964171B1 (en) 2012-09-24 2019-04-01 가부시키가이샤 엔.티.티.도코모 Video prediction encoding device, video prediction encoding method, video prediction decoding device and video prediction decoding method
KR20180069128A (en) * 2012-09-24 2018-06-22 가부시키가이샤 엔.티.티.도코모 Video prediction encoding device, video prediction encoding method, video prediction decoding device and video prediction decoding method
WO2014045651A1 (en) * 2012-09-24 2014-03-27 株式会社エヌ・ティ・ティ・ドコモ Video prediction encoding device, video prediction encoding method, video prediction decoding device and video prediction decoding method
KR101869830B1 (en) 2012-09-24 2018-06-21 가부시키가이샤 엔.티.티.도코모 Video prediction encoding device, video prediction encoding method, video prediction decoding device and video prediction decoding method
KR20150060877A (en) * 2012-09-24 2015-06-03 가부시키가이샤 엔.티.티.도코모 Video prediction encoding device, video prediction encoding method, video prediction decoding device and video prediction decoding method
KR101662655B1 (en) 2012-09-24 2016-10-05 가부시키가이샤 엔.티.티.도코모 Video prediction encoding device, video prediction encoding method, video prediction decoding device and video prediction decoding method
KR101755363B1 (en) 2012-09-24 2017-07-07 가부시키가이샤 엔.티.티.도코모 Video prediction encoding device, video prediction encoding method, video prediction decoding device and video prediction decoding method
KR101764235B1 (en) 2012-09-24 2017-08-03 가부시키가이샤 엔.티.티.도코모 Video prediction encoding device, video prediction encoding method, video prediction decoding device and video prediction decoding method
JP2016012928A (en) * 2015-08-21 2016-01-21 株式会社Nttドコモ Moving image prediction encoding device, moving image prediction encoding method, moving image prediction decoding device, and moving image prediction decoding method
JP2016136786A (en) * 2016-04-28 2016-07-28 株式会社Nttドコモ Moving image prediction encoder, moving image prediction encoding method, moving image prediction decoder and moving image prediction decoding method
US11812060B2 (en) 2016-05-13 2023-11-07 Interdigital Vc Holdings, Inc. Method and device for deblocking filtering a boundary within an intra predicted block
JP2017112629A (en) * 2017-02-03 2017-06-22 株式会社Nttドコモ Moving image prediction decoding device and moving image prediction decoding method
JP2018050320A (en) * 2017-11-07 2018-03-29 株式会社Nttドコモ Moving image prediction decoding method
JP2019013040A (en) * 2018-09-20 2019-01-24 株式会社Nttドコモ Moving image predicting and decoding method

Similar Documents

Publication Publication Date Title
JP6957715B2 (en) Image filter device, filter method and moving image decoding device
JP7001768B2 (en) Arithmetic decoding device
US10547861B2 (en) Image decoding device
WO2012077719A1 (en) Image decoding device and image coding device
JP5551274B2 (en) Image filter device
WO2017068856A1 (en) Predictive image generation device, image decoding device, and image encoding device
WO2017134992A1 (en) Prediction image generation device, moving image decoding device, and moving image coding device
WO2017195532A1 (en) Image decoding device and image encoding device
JP7139144B2 (en) image filter device
TW202133613A (en) (無)
WO2013046990A1 (en) Offset decoding apparatus, offset encoding apparatus, image filter apparatus, and data structure
TW202114417A (en) (無)
JP2013192118A (en) Arithmetic decoder, image decoder, arithmetic encoder, and image encoder
JP2013141094A (en) Image decoding device, image encoding device, image filter device, and data structure of encoded data
WO2012121352A1 (en) Video decoding device, video coding device, and data structure
WO2012077795A1 (en) Image coding device, image decoding device, and data structure
JP2021150703A (en) Image decoding device and image encoding device
WO2012081706A1 (en) Image filter device, filter device, decoder, encoder, and data structure
WO2012147947A1 (en) Image decoding apparatus and image encoding apparatus
WO2012043676A1 (en) Decoding device, encoding device, and data structure
WO2012043766A1 (en) Image decoding device, image encoding device, and data structure for encoded data
JP2013251827A (en) Image filter device, image decoding device, image encoding device, and data structure
JP2024054362A (en) Video decoding device, video encoding device, and predicted image generating method
WO2012043678A1 (en) Image decoding device, image encoding device, and data structure

Legal Events

Date Code Title Description
121 Ep: the epo has been informed by wipo that ep was designated in this application

Ref document number: 11847866

Country of ref document: EP

Kind code of ref document: A1

NENP Non-entry into the national phase

Ref country code: DE

122 Ep: pct application non-entry in european phase

Ref document number: 11847866

Country of ref document: EP

Kind code of ref document: A1

NENP Non-entry into the national phase

Ref country code: JP