WO2015199040A1 - Dmm予測部、画像復号装置、および画像符号化装置 - Google Patents
Dmm予測部、画像復号装置、および画像符号化装置 Download PDFInfo
- Publication number
- WO2015199040A1 WO2015199040A1 PCT/JP2015/067921 JP2015067921W WO2015199040A1 WO 2015199040 A1 WO2015199040 A1 WO 2015199040A1 JP 2015067921 W JP2015067921 W JP 2015067921W WO 2015199040 A1 WO2015199040 A1 WO 2015199040A1
- Authority
- WO
- WIPO (PCT)
- Prior art keywords
- division pattern
- prediction
- size
- target
- unit
- Prior art date
Links
Images
Classifications
-
- H—ELECTRICITY
- H04—ELECTRIC COMMUNICATION TECHNIQUE
- H04N—PICTORIAL COMMUNICATION, e.g. TELEVISION
- H04N19/00—Methods or arrangements for coding, decoding, compressing or decompressing digital video signals
- H04N19/50—Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using predictive coding
- H04N19/597—Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using predictive coding specially adapted for multi-view video sequence encoding
-
- H—ELECTRICITY
- H04—ELECTRIC COMMUNICATION TECHNIQUE
- H04N—PICTORIAL COMMUNICATION, e.g. TELEVISION
- H04N19/00—Methods or arrangements for coding, decoding, compressing or decompressing digital video signals
- H04N19/50—Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using predictive coding
- H04N19/593—Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using predictive coding involving spatial prediction techniques
-
- H—ELECTRICITY
- H04—ELECTRIC COMMUNICATION TECHNIQUE
- H04N—PICTORIAL COMMUNICATION, e.g. TELEVISION
- H04N19/00—Methods or arrangements for coding, decoding, compressing or decompressing digital video signals
- H04N19/70—Methods or arrangements for coding, decoding, compressing or decompressing digital video signals characterised by syntax aspects related to video coding, e.g. related to compression standards
-
- H—ELECTRICITY
- H04—ELECTRIC COMMUNICATION TECHNIQUE
- H04N—PICTORIAL COMMUNICATION, e.g. TELEVISION
- H04N19/00—Methods or arrangements for coding, decoding, compressing or decompressing digital video signals
- H04N19/90—Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using coding techniques not provided for in groups H04N19/10-H04N19/85, e.g. fractals
- H04N19/96—Tree coding, e.g. quad-tree coding
-
- H—ELECTRICITY
- H04—ELECTRIC COMMUNICATION TECHNIQUE
- H04N—PICTORIAL COMMUNICATION, e.g. TELEVISION
- H04N13/00—Stereoscopic video systems; Multi-view video systems; Details thereof
- H04N13/10—Processing, recording or transmission of stereoscopic or multi-view image signals
- H04N13/106—Processing image signals
- H04N13/161—Encoding, multiplexing or demultiplexing different image signal components
Definitions
- the present invention relates to an image decoding apparatus that decodes encoded data representing an image, and an image encoding apparatus that generates encoded data by encoding an image.
- the multi-view image encoding technique includes a parallax predictive encoding that reduces the amount of information by predicting a parallax between images when encoding images of a plurality of viewpoints, and a decoding method corresponding to the encoding method.
- a vector representing the parallax between viewpoint images is called a displacement vector.
- the displacement vector is a two-dimensional vector having a horizontal element (x component) and a vertical element (y component), and is calculated for each block which is an area obtained by dividing one image.
- x component horizontal element
- y component vertical element
- each viewpoint image is encoded as a different layer in each of a plurality of layers.
- a method for encoding a moving image composed of a plurality of layers is generally referred to as scalable encoding or hierarchical encoding.
- scalable coding high coding efficiency is realized by performing prediction between layers.
- a reference layer without performing prediction between layers is called a base layer, and other layers are called enhancement layers.
- Scalable encoding in the case where a layer is composed of viewpoint images is referred to as view scalable encoding.
- the base layer is also called a base view
- the enhancement layer is also called a non-base view.
- scalable coding in the case of a texture layer (image layer) composed of texture (image) and a depth layer (distance image layer) composed of depth map (distance image) is a three-dimensional scalable code. It is called “Kake”.
- Non-Patent Document 1 there is a HEVC-based three-dimensional scalable coding technique disclosed in Non-Patent Document 1.
- DMM prediction also referred to as depth intra-prediction
- DMM prediction basically uses a depth model in which a target block (also referred to as a depth block) on a depth map is composed of two non-rectangular flat regions, and the depth value of each flat region is expressed by a fixed value. Is based.
- the depth model is composed of partition information indicating the region to which each pixel belongs, and depth value information of each region.
- Non-Patent Document 1 a division pattern (wedgelet pattern) of wedgelet division is held in a lookup table defined in advance for each block size, and a division designated by an identifier (division pattern index wedge_full_tab_idx) that designates the division pattern.
- This is a technique for selecting a pattern, dividing a depth block into two regions based on a selected division pattern, and predicting a depth value for each divided region.
- Non-Patent Document 1 DMM prediction based on wedgelet division in Non-Patent Document 1, it is necessary to previously store a division pattern for each block size (4 ⁇ 4 to 32 ⁇ 32) in a lookup table (division pattern list). There is a problem that the size of the uptable is very large. In particular, the size of the 32 ⁇ 32 divided pattern occupies 80% of the entire lookup table size. For this reason, it is difficult to place the lookup table in the cache, and an access to the external memory occurs every time the lookup table is accessed, resulting in a problem that the processing speed decreases.
- the present invention has been made in view of the above problems, and an object of the present invention is to delete a lookup table for holding a division pattern of the first size in DMM1 prediction, and to reduce the first size smaller than the first size.
- the purpose is to realize a decoding device and the like.
- a DMM prediction unit includes a DMM1 division pattern generation unit that derives a division pattern to be applied to a target PU, and a division pattern derived by the DMM1 division pattern generation unit A predicted DC value deriving unit for deriving a predicted value for each region in the target PU based on decoded pixels adjacent to the target PU and DC offset information for each region in the target PU specified by the division pattern
- the partition pattern generation unit is applied to the target PU based on a target PU size, a reference partition pattern size, a partition pattern index that specifies a partition pattern to be applied to the target PU, and a partition pattern list.
- a division pattern to be derived is derived.
- encoding is performed by scaling a division pattern having a second size smaller than the first size to a first size and generating a division pattern having the first size.
- FIG. 7 is a diagram showing a data configuration of encoded data generated by a video encoding device according to an embodiment of the present invention and decoded by the video decoding device, wherein (a) to (e) are sequences, respectively. It is a figure which shows a layer, a picture layer, a slice layer, a tree block layer, and a CU layer. It is a figure which shows the example of the syntax contained in a CU layer.
- (A) shows an example of a syntax table related to an intra CU
- (b) is an example of a syntax table related to intra prediction mode extension. It is an example of the syntax which concerns on the DC offset information contained in a CU layer. It is a figure which shows the example of the prediction mode number corresponding to the classification
- FIG. 1 It is a figure which shows PU setting order and PU contained in CU when an input image (for example, depth map) is YUV format of 4: 0: 0.
- A shows PUs in the CU when the size of the target CU is 8 ⁇ 8 pixels and the division type is N ⁇ N.
- B shows PUs in the CU when the size of the target CU is 16 ⁇ 16 pixels and the division type is 2N ⁇ 2N.
- (A) is an example showing an edge boundary of an object on a block
- (b) is a division pattern (wedgePattern) indicating that the block is divided into two regions (P1, P2) along the edge boundary of the object.
- (C) is an example in which a predicted value is assigned to each of the divided areas.
- DMM prediction it is a figure explaining the production
- (A) is an example of the start point S and the end point E on the block
- FIG. 1 It is a figure which shows an example of the division
- (A) shows an example of a division pattern when the reference division pattern size is 8 ⁇ 8, and (b) shows a division pattern obtained by scaling the division pattern shown in (a) to 16 ⁇ 16. It is an example. It is a figure explaining the effect which concerns on the DMM1 division
- (A) shows a transmitting apparatus equipped with a moving picture coding apparatus, and (b) shows a receiving apparatus equipped with a moving picture decoding apparatus. It is the figure shown about the structure of the recording device which mounts the said moving image encoder, and the reproducing
- (A) shows a recording apparatus equipped with a moving picture coding apparatus, and (b) shows a reproduction apparatus equipped with a moving picture decoding apparatus.
- FIG. 2 is a schematic diagram showing the configuration of the image transmission system 5 according to the present embodiment.
- the image transmission system 5 is a system that transmits a code obtained by encoding a plurality of layer images and displays an image obtained by decoding the transmitted code.
- the image transmission system 5 includes an image encoding device 2, a network 3, an image decoding device 2, and an image display device 4.
- the signal T indicating a plurality of layer images (also referred to as texture images) is input to the image encoding device 2.
- a layer image is an image that is viewed or photographed at a certain resolution and a certain viewpoint.
- each of the plurality of layer images is referred to as a viewpoint image.
- the viewpoint corresponds to the position or observation point of the photographing apparatus.
- the plurality of viewpoint images are images taken by the left and right photographing devices toward the subject.
- the image encoding device 2 encodes each of these signals to generate encoded data # 1. Details of the encoded data # 1 will be described later.
- a viewpoint image is a two-dimensional image (planar image) observed at a certain viewpoint.
- the viewpoint image is indicated by, for example, a luminance value or a color signal value for each pixel arranged in a two-dimensional plane.
- the plurality of layer images include a base layer image having a low resolution and an enhancement layer image having a high resolution.
- the plurality of layer images are composed of a base layer image with low image quality and an extended layer image with high image quality. Note that view scalable coding, spatial scalable coding, and SNR scalable coding may be arbitrarily combined.
- encoding and decoding of an image including at least a base layer image and an image other than the base layer image is handled as the plurality of layer images.
- the image on the reference side is referred to as a first layer image
- the image on the reference side is referred to as a second layer image.
- the enhancement layer image include an image of a viewpoint other than the base view and a depth image.
- the viewpoint image is indicated by, for example, a luminance value or a color signal value for each pixel arranged in a two-dimensional plane.
- the depth map (also called depth map, “depth image”, “depth image”, “distance image”) corresponds to the distance from the viewpoint (shooting device, etc.) of the subject or background included in the subject space.
- Signal values (referred to as “depth value”, “depth value”, “depth”, etc.), and are image signals composed of signal values (pixel values) for each pixel arranged on a two-dimensional plane.
- the pixels constituting the depth map correspond to the pixels constituting the viewpoint image. Therefore, the depth map is a clue for representing the three-dimensional object space by using the viewpoint image which is a reference image signal obtained by projecting the object space onto the two-dimensional plane.
- the network 3 transmits the encoded data # 1 generated by the image encoding device 2 to the image decoding device 1.
- the network 3 is the Internet, a wide area network (WAN: Wide Area Network), a small network (LAN: Local Area Network), or a combination thereof.
- the network 3 is not necessarily limited to a bidirectional communication network, and may be a unidirectional or bidirectional communication network that transmits broadcast waves such as terrestrial digital broadcasting and satellite broadcasting.
- the network 3 may be replaced with a storage medium that records encoded data # 1 such as a DVD (Digital Versatile Disc) or a BD (Blue-ray Disc).
- the image decoding apparatus 1 decodes each of the encoded data # 1 transmitted by the network 3, and generates and outputs a plurality of decoded layer images Td (decoded viewpoint image TexturePic and decoded depth map DepthPic), respectively.
- the image display device 4 displays all or part of the plurality of decoded layer images Td generated by the image decoding device 1. For example, in view scalable coding, a 3D image (stereoscopic image) and a free viewpoint image are displayed in all cases, and a 2D image is displayed in some cases.
- the image display device 4 includes a display device such as a liquid crystal display or an organic EL (Electro-luminescence) display.
- a display device such as a liquid crystal display or an organic EL (Electro-luminescence) display.
- an enhancement layer image with high image quality is displayed and only a lower processing capability is provided. Displays a base layer image that does not require higher processing capability and display capability as an extension layer.
- FIG. 3 is a functional block diagram illustrating a schematic configuration of the image decoding device 1.
- the moving image decoding apparatus 1 includes encoded data # 1 in which the moving image encoding apparatus 2 encodes a layer image (one or a plurality of viewpoint images TexturePic and a depth map DepthPic corresponding to the viewpoint image TexturePic). Entered.
- the moving image decoding apparatus 1 decodes the input encoded data # 1 and externally outputs the layer image # 2 (one or more viewpoint images TexturePic and a depth map DepthPic corresponding to the viewpoint image TexturePic). Output.
- the configuration of the encoded data # 1 will be described below.
- the encoded data # 1 exemplarily includes a sequence and a plurality of pictures constituting the sequence.
- Fig. 4 shows the structure of the hierarchy below the sequence layer in the encoded data # 1.
- 4A to 4E respectively show a sequence layer that defines a sequence SEQ, a picture layer that defines a picture PICT, a slice layer that defines a slice S, and a tree block (Tree block; coding tree unit, Coding). It is a figure which shows the CU layer which prescribes
- sequence layer In the sequence layer, a set of data referred to by the video decoding device 1 for decoding a sequence SEQ to be processed (hereinafter also referred to as a target sequence) is defined.
- the sequence SEQ includes a video parameter set (Sequence Parameter Set), a picture parameter set PPS (Picture Parameter Set), a picture PICT, and an additional extension.
- Information SEI Supplemental Enhancement Information
- # indicates the layer ID.
- FIG. 4 shows an example in which encoded data of # 0 and # 1, that is, layer 0 and layer 1, exists, but the type of layer and the number of layers are not dependent on this.
- the video parameter set VPS is a set of encoding parameters common to a plurality of moving images, a plurality of layers included in the moving image, and encoding parameters related to individual layers in a moving image composed of a plurality of layers.
- a set is defined.
- sequence parameter set SPS a set of encoding parameters referred to by the video decoding device 1 for decoding the target sequence is defined. For example, the width and height of the picture are defined.
- a set of encoding parameters referred to by the video decoding device 1 for decoding each picture in the target sequence is defined.
- a quantization width reference value (pic_init_qp_minus26) used for picture decoding and a flag (weighted_pred_flag) indicating application of weighted prediction are included.
- a plurality of PPS may exist. In that case, one of a plurality of PPSs is selected from each picture in the target sequence.
- Picture layer In the picture layer, a set of data referred to by the video decoding device 1 for decoding a picture PICT to be processed (hereinafter also referred to as a target picture) is defined. As shown in FIG. 4B, the picture PICT includes a picture header PH and slices S 1 to S NS (NS is the total number of slices included in the picture PICT).
- the picture header PH includes a coding parameter group referred to by the video decoding device 1 in order to determine a decoding method of the target picture.
- the reference value (pic_init_qp_minus26) in the picture in the quantization step of the prediction residual is an example of a coding parameter included in the picture header PH.
- picture header PH is also referred to as a picture parameter set (PPS).
- PPS picture parameter set
- slice layer In the slice layer, a set of data referred to by the video decoding device 1 for decoding the slice S to be processed (also referred to as a target slice) is defined. As shown in FIG. 4C, the slice S includes a slice header SH and tree blocks TBLK 1 to TBLK NC (where NC is the total number of tree blocks included in the slice S).
- the slice header SH includes a coding parameter group that the moving image decoding apparatus 1 refers to in order to determine a decoding method of the target slice.
- Slice type designation information (slice_type) for designating a slice type is an example of an encoding parameter included in the slice header SH.
- I slice that uses only intra prediction at the time of encoding (2) P slice that uses unidirectional prediction or intra prediction at the time of encoding, (3) B-slice using unidirectional prediction, bidirectional prediction, or intra prediction at the time of encoding may be used.
- Tree block layer In the tree block layer, a set of data referred to by the video decoding device 1 for decoding a processing target tree block TBLK (hereinafter also referred to as a target tree block) is defined.
- Tree block TBLK includes a tree block header TBLKH, and a coding unit information CU 1 ⁇ CU NL (NL is the total number of coding unit information included in the tree block TBLK).
- NL is the total number of coding unit information included in the tree block TBLK.
- the tree block TBLK is divided into units for specifying a block size for each process of intra prediction or inter prediction and conversion.
- the above unit of the tree block TBLK is divided by recursive quadtree partitioning.
- the tree structure obtained by this recursive quadtree partitioning is hereinafter referred to as a coding tree.
- a unit corresponding to a leaf that is a node at the end of the coding tree is referred to as a coding node.
- the encoding node is a basic unit of the encoding process, hereinafter, the encoding node is also referred to as an encoding unit (CU).
- CU encoding unit
- coding unit information (hereinafter referred to as CU information)
- CU 1 to CU NL is information corresponding to each coding node (coding unit) obtained by recursively dividing the tree block TBLK into quadtrees. is there.
- the root of the coding tree is associated with the tree block TBLK.
- the tree block TBLK is associated with the highest node of the tree structure of the quadtree partition that recursively includes a plurality of encoding nodes.
- each coding node is half the size of the coding node to which the coding node directly belongs (that is, the unit of the node one layer higher than the coding node).
- the size that each coding node can take depends on the size specification information of the coding node included in the size of the tree block and the sequence parameter set SPS of the coded data # 1. Since the tree block is the root of the encoding node, the maximum size of the encoding node is the size of the tree block. Since the maximum size of the tree block matches the maximum size of the coding node (CU), LCU (Largest CU) may be used as the name of the tree block. Regarding the minimum size, for example, the minimum encoding node size (log2_min_coding_block_size_minus3) and the difference between the maximum and minimum encoding node sizes (log2_diff_max_min_coding_block_size) are used as size designation information.
- size specification information of a coding node having a maximum coding node size of 64 ⁇ 64 pixels and a minimum coding node size of 8 ⁇ 8 pixels is used.
- the size of the encoding node and the encoding unit CU is 64 ⁇ 64 pixels, 32 ⁇ 32 pixels, 16 ⁇ 16 pixels, or 8 ⁇ 8 pixels.
- the tree block header TBLKH includes an encoding parameter referred to by the video decoding device 1 in order to determine a decoding method of the target tree block. Specifically, as shown in FIG. 4 (d), tree block division information SP_TBLK that specifies a division pattern of the target tree block into each CU, and a quantization parameter difference that specifies the size of the quantization step ⁇ qp (qp_delta) is included.
- the tree block division information SP_TBLK is information representing a coding tree for dividing the tree block. Specifically, the shape and size of each CU included in the target tree block, and the position in the target tree block Is information to specify.
- the tree block division information SP_TBLK may not explicitly include the shape or size of the CU.
- the tree block division information SP_TBLK may be a set of flags (split_coding_unit_flag) indicating whether or not the entire target tree block or a partial area of the tree block is divided into four.
- the shape and size of each CU can be specified by using the shape and size of the tree block together.
- the quantization parameter difference ⁇ qp is a difference qp ⁇ qp ′ between the quantization parameter qp in the target tree block and the quantization parameter qp ′ in the tree block encoded immediately before the target tree block.
- CU layer In the CU layer, a set of data referred to by the video decoding device 1 for decoding a CU to be processed (hereinafter also referred to as a target CU) is defined.
- the encoding node is a node at the root of a prediction tree (PT) and a transformation tree (TT).
- PT prediction tree
- TT transformation tree
- the encoding node is divided into one or a plurality of prediction blocks, and the position and size of each prediction block are defined.
- the prediction block is one or a plurality of non-overlapping areas constituting the encoding node.
- the prediction tree includes one or a plurality of prediction blocks obtained by the above division.
- Prediction processing is performed for each prediction block.
- a prediction block that is a unit of prediction is also referred to as a prediction unit (PU).
- intra prediction There are roughly two types of division in the prediction tree: intra prediction and inter prediction.
- inter prediction there are 2N ⁇ 2N (the same size as the encoding node), 2N ⁇ N, N ⁇ 2N, N ⁇ N, and the like.
- the encoding node is divided into one or a plurality of transform blocks, and the position and size of each transform block are defined.
- the transform block is one or a plurality of non-overlapping areas constituting the encoding node.
- the conversion tree includes one or a plurality of conversion blocks obtained by the above division.
- transform processing is performed for each conversion block.
- the transform block which is a unit of transform is also referred to as a transform unit (TU).
- the CU information CU specifically includes a skip flag SKIP, PT information PTI, and TT information TTI.
- the skip flag SKIP is a flag indicating whether or not the skip mode is applied to the target PU.
- the value of the skip flag SKIP is 1, that is, when the skip mode is applied to the target CU, PT information PTI and TT information TTI in the CU information CU are omitted. Note that the skip flag SKIP is omitted for the I slice.
- the PT information PTI is information regarding the PT included in the CU.
- the PT information PTI is a set of information related to each of one or more PUs included in the PT, and is referred to when the moving image decoding apparatus 1 generates a predicted image.
- the PT information PTI includes prediction type information PType and prediction information PInfo.
- Prediction type information PType is information that specifies whether intra prediction or inter prediction is used as a prediction image generation method for the target PU.
- the prediction information PInfo is composed of intra prediction information or inter prediction information depending on which prediction method is specified by the prediction type information PType.
- a PU to which intra prediction is applied is also referred to as an intra PU
- a PU to which inter prediction is applied is also referred to as an inter PU.
- the prediction information PInfo includes information specifying the shape, size, and position of the target PU. As described above, the generation of the predicted image is performed in units of PU. Details of the prediction information PInfo will be described later.
- TT information TTI is information related to TT included in the CU.
- the TT information TTI is a set of information regarding each of one or a plurality of TUs included in the TT, and is referred to when the moving image decoding apparatus 1 decodes residual data.
- a TU may be referred to as a conversion block.
- the TT information TTI includes TT division information SP_TU that designates a division pattern for each transform block of the target CU, and TU information TUI 1 to TUI NT (NT is the target CU). The total number of transform blocks included).
- TT division information SP_TU is information for determining the shape and size of each TU included in the target CU and the position in the target CU.
- the TT division information SP_TU can be realized from information (split_transform_unit_flag) indicating whether or not the target node is divided and information (trafoDepth) indicating the depth of the division.
- each TU obtained by the division can take a size from 32 ⁇ 32 pixels to 4 ⁇ 4 pixels.
- the TU partition information SP_TU includes information on whether or not a non-zero conversion coefficient exists in each TU. For example, non-zero coefficient existence information (CBP; Coded Block Flag) for each TU and non-zero coefficient existence information (no_residual_data_flag) for a plurality of TUs are included in the TU partition information SP_TU.
- CBP Coded Block Flag
- no_residual_data_flag non-zero coefficient existence information
- the TU information TUI 1 to TUI NT are individual information regarding one or more TUs included in the TT.
- the TU information TUI includes a quantized prediction residual.
- Each quantized prediction residual is encoded data generated by the video encoding device 2 performing the following processes 1 to 3 on a target block that is a processing target block.
- Process 1 DCT transform (Discrete Cosine Transform) of the prediction residual obtained by subtracting the prediction image from the encoding target image;
- Process 2 Quantize the transform coefficient obtained in Process 1;
- Process 3 Variable length coding is performed on the transform coefficient quantized in Process 2;
- prediction information PInfo As described above, there are two types of prediction information PInfo: inter prediction information and intra prediction information.
- the inter prediction information includes an encoding parameter that is referred to when the video decoding device 1 generates an inter predicted image by inter prediction. More specifically, the inter prediction information includes inter PU division information that specifies a division pattern of the target CU into each inter PU, and inter prediction parameters for each inter PU.
- the inter prediction parameters include a reference image index, an estimated motion vector index, and a motion vector residual.
- the intra prediction information includes an encoding parameter that is referred to when the video decoding device 1 generates an intra predicted image by intra prediction. More specifically, the intra prediction information includes intra PU division information that specifies a division pattern of the target CU into each intra PU, and intra prediction parameters for each intra PU.
- the intra prediction parameter is a parameter for restoring intra prediction (prediction mode) for each intra PU.
- Intra prediction parameters are flags related to MPM (Most Probable Mode, and so on).
- MPM Motion Probable Mode, and so on.
- a certain mpm_flag, mpm_idx that is an index for selecting an MPM, and rem_idx that is an index (residual prediction mode index) for specifying a prediction mode other than the MPM are included.
- MPM is an estimated prediction mode that is highly likely to be selected in the target partition.
- the MPM may include an estimated prediction mode estimated based on prediction modes assigned to partitions around the target partition, and a DC mode or Planar mode that generally has a high probability of occurrence.
- prediction mode when simply described as “prediction mode”, it indicates the luminance prediction mode.
- the color difference prediction mode is described as “color difference prediction mode” and is distinguished from the luminance prediction mode.
- the parameter for restoring the prediction mode includes chroma_mode that is a parameter for designating the color difference prediction mode. Note that mpm_flag and rem_idx correspond to “prev_intra_luma_pred_flag” (SYN02 in FIG.
- chroma_mode corresponds to “intra_chroma_pred_mode” (not shown).
- Parameters for restoring the prediction mode (intra extended mode (SYN01 in Fig. 5 (a))) for depth intra prediction (DMM prediction) used for coding depth maps (depth intra prediction parameters, DMM prediction mode information) Includes depth intra prediction presence / absence flag (depth intra prediction presence / absence flag) dim_not_present_flag (SYN01A in Fig. 5 (b)), depth intra prediction method (DMM1 prediction based on wedgelet division (INTRA_DMM_WFULL) and contour division based Flag for selecting DMM4 prediction (INTRA_DMM_CREDTEX) (depth intra mode flag) depth_intra_mode_flag (SYN01B in FIG. 5 (b)) and index (partition pattern index) for specifying a partition pattern in PU in DMM1 prediction wedge_full_tab_idex (FIG. 5) (b) SYN01C).
- the prediction parameters related to the depth intra prediction further include DC offset information for correcting the depth prediction values of the two divided areas in the PU, that is, a DC offset presence / absence flag depth_dc_flag (SYND1 in FIG. 6), a DC offset value.
- DC offset information for correcting the depth prediction values of the two divided areas in the PU, that is, a DC offset presence / absence flag depth_dc_flag (SYND1 in FIG. 6), a DC offset value.
- depth_dc_abs SYND02 in FIG. 6) indicating the absolute value
- depth_dc_sign_flag SYND03 in FIG. 6 indicating the sign of the DC offset value.
- the video decoding device 1 generates a predicted image for each PU, generates a decoded image # 2 by adding the generated predicted image and a prediction residual decoded from the encoded data # 1, and generates The decoded image # 2 is output to the outside.
- An encoding parameter is a parameter referred in order to generate a prediction image.
- the encoding parameters include PU size and shape, block size and shape, and original image and predicted image.
- residual data a set of all information excluding the residual data among the information included in the encoding parameter is referred to as side information.
- a picture (frame), a slice, a tree block, a CU, a block, and a PU to be decoded are a target picture, a target slice, a target tree block, a target CU, a target block, and a target PU, respectively. I will call it.
- the size of the tree block is, for example, 64 ⁇ 64 pixels
- the size of the CU is, for example, 64 ⁇ 64 pixels, 32 ⁇ 32 pixels, 16 ⁇ 16 pixels, and 8 ⁇ 8 pixels
- the size of the PU is For example, 64 ⁇ 64 pixels, 32 ⁇ 32 pixels, 16 ⁇ 16 pixels, 8 ⁇ 8 pixels, 4 ⁇ 4 pixels, and the like.
- these sizes are merely examples, and the sizes of the tree block, CU, and PU may be other than the sizes shown above.
- FIG. 3 is a functional block diagram illustrating a schematic configuration of the video decoding device 1.
- the moving picture decoding apparatus 1 includes a variable length decoding unit 11, an inverse quantization / inverse conversion unit 13, a predicted image generation unit 14, an adder 15, and a frame memory 16.
- variable length decoding unit 11 decodes various parameters included in the encoded data # 1 input from the video decoding device 1. In the following description, it is assumed that the variable length decoding unit 11 appropriately decodes a parameter encoded by an entropy encoding method such as CABAC. Specifically, the variable length decoding unit 11 decodes encoded data # 1 for one frame according to the following procedure.
- variable length decoding unit 11 separates the encoded data # 1 for one frame into various types of information included in the hierarchical structure shown in FIG. 4 by demultiplexing.
- the variable length decoding unit 11 refers to information included in various headers and sequentially separates the encoded data # 1 into slices and tree blocks.
- the various headers include (1) information about the method of dividing the target picture into slices, and (2) information about the size, shape, and position of the tree block belonging to the target slice. .
- variable length decoding unit 11 refers to the tree block division information SP_TBLK included in the tree block header TBLKH, and divides the target tree block into CUs. Further, the variable length decoding unit 11 decodes the TT information TTI related to the conversion tree obtained for the target CU and the PT information PTI related to the prediction tree obtained for the target CU.
- the variable length decoding unit 11 supplies the TT information TTI obtained for the target CU to the TU information decoding unit 12. Further, the variable length decoding unit 11 supplies the PT information PTI obtained for the target CU to the predicted image generation unit 14.
- the TT information TTI includes the TU information TUI corresponding to the TU included in the conversion tree as described above. Further, as described above, the PT information PTI includes PU information PUI (prediction information Info of each PU) corresponding to each PU included in the target prediction tree.
- the variable length decoding unit 11 decodes each syntax from the encoded data # 1 according to the syntax table of the intra prediction mode extension intra_mode_ext () shown in SYN01 of FIG.
- the intra prediction mode extension intra_mode_ext () is decoded when a flag (depth mode availability flag) vps_depth_modes_flag indicating the appropriateness of the depth encoding tool is 1 in the decoding target layer.
- depth mode availability flag is 1, it indicates that the depth encoding tool is applied in the decoding target layer, and when it is 0, it indicates that it is not applied.
- the depth mode enable / disable flag is decoded from a parameter set (video parameter set VPS, sequence parameter set SPS, picture parameter set PPS, slice header SH) and the like.
- the variable length decoding unit 11 decodes the depth intra prediction presence / absence flag dim_not_present_flag.
- the value of the flag is estimated as 1. This flag is a flag indicating the presence / absence of depth intra prediction.
- the depth intra prediction mode flag depth_intra_mode_flag for the target PU is not included in the encoded data, and the intra prediction mode numbers' 0 'to' This indicates that any intra prediction method of 34 ′ (DC prediction, Planar prediction, Angular prediction) is used for the target PU.
- the flag is 0, it indicates that the depth intra prediction mode depth_intra_mode_flag is in the encoded data.
- variable length decoding unit 11 derives the DMM flag DmmFlag by the following equation based on the decoded depth intra prediction presence / absence flag dim_not_present_flag.
- DmmFlag ! Dim_not_present_flag That is, the logical negation value of the depth intra prediction presence / absence flag is set in the DMM flag. When the DMM flag is 1, it indicates that depth intra prediction is used, and when the DMM flag is 0, it indicates that depth intra prediction is not used.
- the variable length decoding unit 11 further decodes the depth intra mode flag depth_intra_mode_flag.
- the flag is a flag related to selection of a depth intra prediction method. When the flag is 0, it indicates that the depth intra prediction is DMM1 prediction. When the flag is 1, it indicates that the depth intra prediction is DMM4 prediction.
- the variable length decoding unit 11 sets a prediction mode number indicating DMM1 prediction to the prediction mode predModeIntra when the depth intra mode flag depth_intra_mode_flag is 0, that is, when the depth intra prediction is DMM1 prediction. Furthermore, the division pattern index wedge_full_tab_idx that specifies the division pattern in the PU is decoded.
- the variable length decoding unit 11 sets a prediction mode number indicating DMM4 prediction to the prediction mode predModeIntra when the depth intra mode flag depth_intra_mode_flag is 1, that is, when the depth intra prediction is DMM4 prediction.
- the variable length decoding unit 11 decodes the MPM flag mpm_flag indicating whether or not the intra prediction mode of the target PU matches the estimated prediction mode MPM.
- the flag indicates that the intra prediction mode of the target PU matches the estimated prediction mode MPM.
- the prediction mode numbers “0” to “34” DC prediction, Planar prediction, Angular Among any prediction, this indicates any prediction mode except the estimated prediction mode MPM.
- variable length decoding unit 11 further decodes the MPM index mpm_idx designating the estimated prediction mode MPM, and sets the estimated prediction mode indicated by the mpm_idx to the prediction mode predModeIntra.
- variable length decoding unit 11 When the MPM flag is 0, the variable length decoding unit 11 further decodes an index rem_idx for designating a prediction mode other than the MPM, and is a prediction mode number excluding the estimated prediction mode MPM identified from the rem_idx. Any prediction mode number in 0 'to' 34 '(DC prediction, Planar prediction, or Angular prediction) is set to the prediction mode predModeIntra.
- variable length decoding unit 11 When the DC offset presence / absence flag depth_dc_flag is 1, the variable length decoding unit 11 further decodes depth_dc_abs [i] indicating the absolute value of the DC offset value (DC offset absolute value).
- variable length decoding unit 11 further decodes depth_dc_sign_flag [i] indicating the positive / negative sign of the DC offset value.
- the DC offset value DcOffset [i] to be derived is derived by the following equation.
- the inverse quantization / inverse transform unit 13 performs an inverse quantization / inverse transform process on each block included in the target CU based on the TT information TTI. Specifically, for each target TU, the inverse quantization / inverse transform unit 13 performs inverse quantization and inverse orthogonal transform on the quantized prediction residual included in the TU information TUI corresponding to the target TU, so that each pixel is per pixel. Is restored.
- the orthogonal transform refers to an orthogonal transform from the pixel region to the frequency region.
- the inverse orthogonal transform is a transform from the frequency domain to the pixel domain.
- Examples of inverse orthogonal transform include inverse DCT transform (Inverse Discrete Cosine Transform), inverse DST transform (Inverse Discrete Sine Transform), and the like.
- the inverse quantization / inverse transform unit 13 supplies the restored prediction residual D to the adder 15.
- the predicted image generation unit 14 generates a predicted image based on the PT information PTI for each PU included in the target CU. Specifically, the predicted image generation unit 14 is a decoded image by performing intra prediction or inter prediction for each target PU according to the parameters included in the PU information PUI (prediction information Info) corresponding to the target PU. A predicted image Pred is generated from the locally decoded image P ′. The predicted image generation unit 14 supplies the generated predicted image Pred to the adder 15. The configuration of the predicted image generation unit 14 will be described in more detail later.
- the adder 15 adds the predicted image Pred supplied from the predicted image generation unit 14 and the prediction residual D supplied from the inverse quantization / inverse transform unit 13, thereby obtaining the decoded image P for the target CU. Generate.
- the decoded image P that has been decoded is sequentially recorded in the frame memory 16.
- decoded images corresponding to all tree blocks decoded before the target tree block are stored. It is recorded.
- Decoded image # 2 corresponding to # 1 is output to the outside.
- the predicted image generation unit 14 generates and outputs a predicted image based on the PT information PTI.
- the PU information PTI input to the predicted image generation unit 14 includes a prediction mode (IntraPredMode) and a color difference prediction mode (IntraPredModeC).
- IntraPredMode a prediction mode
- IntraPredModeC a color difference prediction mode
- the definition of the prediction mode (brightness / color difference) will be described with reference to FIG. 7, FIG. 8, and FIG.
- FIG. 7 shows an example of prediction mode numbers corresponding to the classification of intra prediction schemes used in the video decoding device 1.
- Planar prediction (INTRA_PLANAR) is '0'
- DC prediction (INTRA_DC) is '1'
- Angular prediction (INTRA_ANGULAR) is '2' to '34'
- DMM1 prediction (INTRA_DMM_WFULL) is '35'
- DMM4 prediction INTRA_DMM_CREDTEX
- a prediction mode number of “36” is assigned.
- a prediction method assigned with a prediction mode number of “10” is also called horizontal prediction
- a prediction method assigned with a prediction mode number of “26” is also called vertical prediction.
- Horizontal prediction, vertical prediction, and Angular prediction are collectively referred to as direction prediction.
- the direction prediction is a prediction method for generating a prediction image by extrapolating adjacent pixel values around the target PU in a specific direction.
- the DMM1 prediction and the DMM4 prediction are collectively referred to as depth intra prediction.
- depth intra prediction a target block (also referred to as a depth block) on a depth map is basically composed of two non-rectangular flat regions, and the depth value of each flat region is expressed by a fixed value. Based on the model.
- the depth model is composed of partition information indicating the region to which each pixel belongs, and depth value information of each region.
- partition types that is, a wedgelet partition and a contour partition, as depth block partitioning methods. Details of the depth intra prediction will be described later.
- FIG. 8 shows prediction directions corresponding to the prediction mode identifiers for 33 types of prediction modes belonging to the direction prediction.
- the direction of the arrow in FIG. 8 represents the prediction direction, but more accurately indicates the direction of the vector from the prediction target pixel to the decoded pixel referred to by the prediction target pixel. In that sense, the prediction direction is also referred to as a reference direction.
- the identifier of each prediction mode is associated with a code indicating whether the main direction is the horizontal direction (HOR) or the vertical direction (VER) and an identifier composed of a combination of displacements with respect to the main direction.
- HOR is used for horizontal prediction
- VER is used for vertical prediction
- VER + 8 is used for a prediction mode that refers to surrounding pixels in the upper right 45-degree direction
- VER-8 is used for a prediction mode that refers to surrounding pixels in the 45-degree upper left direction
- 45 is used for lower left 45
- a prediction mode that refers to peripheral pixels in the direction of the degree is assigned a code of HOR + 8.
- 17 prediction directions of VER-8 to VER + 8 are defined as a vertical prediction mode
- 16 prediction directions of HOR-7 to HOR + 8 are defined as a prediction mode of horizontal prediction.
- FIG. 9 is a diagram illustrating an example of a prediction mode definition DEFPM1 that is a definition of correspondence between an intra prediction method and a prediction mode number.
- a prediction mode number of “0” is assigned to Planar prediction and “1” is assigned to DC prediction.
- Prediction mode numbers “2” to “18” are assigned Angular prediction modes VER-8 to VER + 8 in which the main direction is vertical prediction, respectively, and the main directions are assigned to prediction mode numbers “19” to “34”.
- FIG. 11 is a functional block diagram illustrating a configuration example of the predicted image generation unit 14.
- this structural example has illustrated the functional block which concerns on the prediction image production
- FIG. 11 is a functional block diagram illustrating a configuration example of the predicted image generation unit 14.
- this structural example has illustrated the functional block which concerns on the prediction image production
- the predicted image generation unit 14 includes a prediction unit setting unit 141, a reference pixel setting unit 142, a switch 143, a reference pixel filter unit 144, and a predicted image derivation unit 145.
- the prediction unit setting unit 141 sets the PU included in the target CU as the target PU in a prescribed setting order, and outputs information about the target PU (target PU information).
- the target PU information includes at least the size nS of the target PU, the position of the target PU in the CU, and the index (luminance color difference index cIdx) indicating the luminance or color difference plane of the target PU.
- the PU setting order is such that PUs corresponding to Y included in the target CU are set in raster scan order, and subsequently, PUs corresponding to the order of U and V are set in raster scan order. Use the order of setting.
- FIG. 12 illustrating PU setting order and PUs included in a CU when the input image is a viewpoint image TexturePic expressed in a 4: 2: 0 YUV format.
- FIG. 12 illustrates PUs in the CU when the size of the target CU is 8 ⁇ 8 pixels and the division type is N ⁇ N.
- four 4 ⁇ 4 pixel PUs corresponding to the luminance Y are set in the raster scan order (in the order of PU_Y0, PU_Y1, PU_Y2, PU_Y3).
- one 4 ⁇ 4 pixel PU (PU_U0) corresponding to the color difference U is set.
- PU_V0 one 4 ⁇ 4 pixel prediction unit corresponding to the color difference V is set.
- FIG. 12 illustrates PUs in the CU when the size of the target CU is 16 ⁇ 16 pixels and the division type is 2N ⁇ 2N.
- PU_Y0 16 ⁇ 16 pixel prediction unit
- PU_U0 one 8 ⁇ 8 pixel prediction unit
- PU_V0 one 8 ⁇ 8 pixel prediction unit
- FIG. 13 illustrating PU setting order and PUs included in the CU when the input image is a depth map DepthPic expressed in a 4: 0: 0 YUV format.
- FIG. 13 illustrates PUs in a CU when the size of the target CU is 8 ⁇ 8 pixels and the division type is N ⁇ N.
- Four 4 ⁇ 4 pixel PUs corresponding to the luminance Y are set in the raster scan order (the order of PU_Y0, PU_Y1, PU_Y2, PU_Y3).
- FIG. 13 illustrates PUs in the CU when the size of the target CU is 16 ⁇ 16 pixels and the division type is 2N ⁇ 2N.
- One prediction unit (PU_Y0) of 16 ⁇ 16 pixels corresponding to the luminance Y is set.
- the reference pixel setting unit 142 reads a pixel value (decoded pixel value) of a decoded image around the target PU recorded in the frame memory based on the input target PU information, and is a reference pixel referred to when a predicted image is generated Set.
- the reference pixel value p [x] [y] is set by the following equation using the decoded pixel value r [x] [y].
- a predetermined value for example, 1 ⁇ (BitDepth-1) may be used.
- BitDepth is the bit depth of the pixel.
- a referable decoded pixel value existing in the vicinity of the corresponding decoded pixel value may be used.
- the switch 143 outputs the reference pixel to the corresponding output destination based on the luminance / color difference index cIdx and the prediction mode predModeIntra among the input target PU information. More specifically, the luminance color difference index cIdx is 0 (the pixel to be processed is luminance), and the prediction mode predModeIntra is 0 to 34 (the prediction mode is Planar prediction or DC prediction, Alternatively, when Angular prediction is performed (predModeIntra ⁇ 35), the switch 143 outputs the input reference pixel to the reference pixel filter unit 144.
- the switch 143 outputs the input reference pixel to the predicted image deriving unit 145.
- the reference pixel filter unit 144 applies a filter to the input reference pixel value, and outputs a reference pixel value after the filter application. Specifically, the reference pixel filter unit 144 determines whether to apply the filter according to the target PU size and the prediction mode predModeIntra.
- the predicted image deriving unit 145 generates and outputs a predicted image predSamples of the target PU based on the input PU information (prediction mode predModeIntra, luminance color difference index cIdx, PU size nS) and the reference pixel p [x] [y]. To do. Detailed description of the predicted image deriving unit 145 will be described later.
- the prediction unit setting unit 141 sets one of the PUs included in the CU as the target PU according to a predetermined order, and sets the target PU information as the reference pixel setting unit 142 and the switch. It outputs to 143 (S11).
- the reference pixel setting unit 142 sets the reference pixel of the target PU using the decoded pixel value read from the external frame memory (S12).
- the switch 143 determines whether the target PU is luminance or color difference based on the input target PU information, or whether the prediction mode predModeIntra is DMM prediction, and switches the output according to the determination result ( S13).
- the output of the switch 143 is connected to the reference pixel filter unit 144. Subsequently, the reference pixel is input to the reference pixel filter unit 144, the reference pixel filter is applied according to the prediction mode separately input, and the reference pixel after the filter application is output to the predicted image derivation unit 145 (S14). ).
- the output of the switch 143 is output to the prediction image deriving unit 145. Connected.
- the predicted image derivation unit 145 generates predicted images predSamples in the target PU based on the input PU information (prediction mode predModeIntra, luminance color difference index cIdx, PU size nS) and the reference pixel p [x] [y]. And output (S15).
- the prediction unit setting unit 141 determines whether the prediction images of all the PUs in the target CU have been generated (S16). When the prediction image of some PUs in the target CU has not been generated (NO in S16), the process returns to S1, and the prediction image generation process for the next PU in the target CU is executed. If predicted images of all PUs in the target CU have been generated (YES in S16), the predicted images of the luminance and color difference of each PU in the target CU are combined and output as a predicted image of the target CU, and the process ends. To do.
- the predicted image derivation unit 145 further includes a DC prediction unit 145D, a Planar prediction unit 145P, an Angular prediction unit 145A, and a DMM prediction unit 145T.
- the predicted image derivation unit 145 selects a prediction method used for generating a predicted image based on the input prediction mode predModeIntra.
- the selection of the prediction method is realized by selecting a prediction method corresponding to the prediction mode number of the input prediction mode predModeIntra based on the definition of FIG.
- the predicted image deriving unit 145 derives a predicted image corresponding to the selection result of the prediction method. More specifically, the prediction image deriving unit 145, when the prediction method is Planar prediction, DC prediction, Angular prediction, and DMM prediction, respectively, Planar prediction unit 145P, DC prediction unit 145D, Angular prediction unit 145A, and A predicted image is derived by the DMM prediction unit 145T.
- the DC prediction unit 145D derives a DC prediction value corresponding to the average value of the pixel values of the input reference pixels, and outputs a prediction image using the derived DC prediction value as a pixel value.
- the Planar prediction unit 145P generates and outputs a prediction image based on pixel values derived by linearly adding a plurality of reference pixels according to the distance to the prediction target pixel.
- the pixel value predSamples [x] [y] of the predicted image can be derived by the following equation using the reference pixel value p [x] [y] and the size nS of the target PU.
- the Angular prediction unit 145A generates and outputs a prediction image corresponding to the target PU using reference pixels in a prediction direction (reference direction) corresponding to the input prediction mode predModeIntra.
- main reference pixels are set according to the value of the prediction mode predModeIntra, and the predicted image is generated with reference to the main reference pixels in units of lines or columns in the PU.
- the Angular prediction unit 145A derives a prediction direction (reference direction) associated with the input prediction mode predModeIntra.
- the derived prediction direction is expressed by a combination of a main direction flag bRefVer indicating whether or not the main direction is a vertical direction and a gradient (offset) intraPredAngle of the prediction direction with respect to the main direction.
- bRefVer indicating whether or not the main direction is a vertical direction
- the Angular prediction unit 145A derives a gradient intraPredAngle corresponding to the prediction mode predModeIntra with reference to the gradient definition table DEFANG1 shown in FIG.
- the gradient definition table DEFALG1 shown in FIG. 10 is a table showing the correspondence between the prediction mode number and the value of the gradient intraPredAngle.
- the value of the gradient intraPredAngle is a value representing the gradient in the prediction direction. More precisely, when the main direction is the vertical direction, the direction of the vector represented by (intraPredAngle, -32) is the prediction direction. When the main direction is the horizontal direction, the vector direction represented by ( ⁇ 32, intraPredAngle) is the prediction direction.
- the Angular prediction unit 145A sets the prediction image generation unit as a line, and sets the reference pixel above the target PU as the main reference pixel. To do. Specifically, the main reference pixel ref [x] is set using the value of the reference pixel p [x] [y] by the following procedure.
- invAngle corresponds to a value obtained by scaling (multiplying by 8192) the reciprocal of the displacement intraPredAngle in the prediction direction.
- the Angular prediction unit 145A represents the position of the main reference pixel used for generating the prediction target pixel calculated according to the distance (y + 1) between the prediction target line and the main reference pixel and the gradient intraPredAngle, and an integer precision position iIdx in pixel units.
- the position iFact of the decimal point precision in the pixel unit is derived by the following equation.
- & is an operator representing a logical bit operation
- the result of “A & 31” means a remainder obtained by dividing the integer A by 32. The same applies thereafter.
- the Angular prediction unit 145A derives the prediction image predSamples [x] [y] of the target PU according to the following equation in accordance with the derived variable iFact.
- the prediction image predSamples [x] [y] is derived by linear interpolation.
- the Angular prediction unit 145A represents an integer-accurate position iIdx in pixel units that represents the position of the main reference pixel used to generate the prediction reference pixel calculated according to the distance (x + 1) between the prediction target column and the main reference pixel and the gradient intraPredAngle.
- the position iFact of the decimal point precision in the pixel unit is derived by the following equation.
- Angular prediction unit 145A derives the predicted image predSamples [x] [y] of the target PU according to the following equation in accordance with the derived variable iFact.
- the prediction image predSamples [x] [y] is derived by linear interpolation.
- the DMM prediction unit 145T generates and outputs a prediction image corresponding to the target PU based on DMM prediction (also referred to as depth modeling mode or depth intra prediction) corresponding to the input prediction mode predModeIntra.
- FIG. 15 is a conceptual diagram of DMM prediction executed in the DMM prediction unit 145T.
- the depth map mainly has an edge region representing an object boundary and a flat region (substantially constant depth value) representing an object area.
- the target block is divided into two regions P1 and P2 along the edge of the object, and as shown in FIG.
- the division pattern WedgePattern [x] [y] is a matrix having a size corresponding to the width x height of the target block (target PU), and 0 or 1 is set for each element (x, y). Indicates which of the two regions P1 and P2 each pixel belongs to. In the example of FIG. 15B, if the element value is 0, it belongs to the region P1, and if it is 1, it belongs to the region P2. Next, as shown in FIG. 15C, a predicted image is generated by filling each of the regions P1 and P2 with depth prediction values.
- FIG. 1 is a functional block diagram showing a configuration example of the DMM prediction unit 145T.
- a DMM4 division pattern generation unit 145T1 As shown in FIG. 1, a DMM4 division pattern generation unit 145T1, a DMM1 division pattern generation unit 145T2, and a DC predicted value derivation unit 145T3 are provided.
- the DMM prediction unit 145T activates the division pattern derivation means (DMM1 division pattern derivation unit, DMM4 division pattern derivation unit) corresponding to the input prediction mode predModeIntra, and generates the division pattern wedgePattern [x] [y] of the target PU To do. More specifically, when the prediction mode predModeIntra is the prediction mode number ‘35’, that is, in the INTRA_DMM_WEDGEFULL mode, the DMM1 division pattern deriving unit 145T6 is activated. On the other hand, when the prediction mode predModeIntra is the prediction mode number “36”, that is, in the INTRA_DMM_CPCREDTEX mode, the DMM4 division pattern deriving unit 145T3 is activated.
- the prediction mode predModeIntra is the prediction mode number “36”
- the DMM4 division pattern deriving unit 145T3 is activated.
- the DMM4 division pattern generation unit 145T1 derives the division pattern wedgePattern [x] [y] of the target PU based on the decoded pixel value recTexPic of the luminance on the viewpoint image TexturePic corresponding to the target PU on the depth map DepthPic, and DC It outputs to the predicted value deriving unit 145T3.
- the DMM4 division pattern generation unit binarizes the two blocks P1 and P2 of the target PU on the depth map by the average value of the luminance of the target block on the corresponding viewpoint image TexturePic. It derives by that.
- the DMM4 division pattern generation unit 145T1 reads from the external frame memory 16 the luminance decoded pixel value recTextPic of the corresponding block on the viewpoint image TexturePic corresponding to the target PU from the external frame memory 16, and refers to the reference pixel refSamples [x]. Set to [y] by the following formula.
- the threshold value threshVals is derived by the following equation based on the sum sumRefVals and the target PU size nS. That is, the average pixel value of the corresponding block is derived.
- threshVal (sumRefVals >> (2 * log2 (nS))
- nS a value obtained by dividing the sumRefVals by the square number nS * nS of the target PU size nS.
- the DMM4 division pattern generation unit 145T1 refers to the derived threshold value threshVal and the reference pixel refSamples [x] [y] to derive the division pattern wedgePattern [x] [y] of the target PU by the following equation. Output.
- wedgePattern [x] [y] (refSamples [x] [y]> threshVal) That is, when the reference pixel refSamples [x] [y] is larger than the threshold value threshVal, 1 is set to the element (x, y) of the division pattern. When the reference pixel refSamples [x] [y] is equal to or less than the threshold value threshVal, 0 is set to the element (x, y) of the division pattern.
- the DMM1 division pattern generation unit 145T2 further includes a DMM1 division pattern derivation unit 145T6, a buffer 145T5, and a division pattern list generation unit 145T4.
- the DMM1 division pattern generation unit 145T2 activates the division pattern list generation unit 145T4 only at the first activation, and generates a division pattern list WedgePatternTable for each block size.
- the generated division pattern list is stored in the buffer 145T5.
- the DMM division pattern deriving unit 145T6 performs division from the division pattern list WedgePatternTable stored in the buffer 145T5 based on the input target PU size nS, the division pattern index wedge_full_tab_idx, and a preset reference division pattern size nBS.
- the pattern wedgePattern [x] [y] is derived and output to the DC predicted value deriving unit 145T3.
- FIG. 16A a division pattern in which all elements are 0 is generated.
- a start point S (xs, ys) and an end point E (xe, ye) are set in the division pattern.
- a line segment is drawn between the start point S and the end point E using Bresenham's algorithm (shaded elements in FIG. 16B).
- blocksize is the size (vertical width, horizontal width) of a block for generating a division pattern.
- the minimum block size for generating a division pattern is nMinS ⁇ nMinS and the maximum block size is nMaxS ⁇ nMaxS.
- wBlksize (1 ⁇ logBlkSize) is also used as the block size of the division pattern.
- the generated divided pattern wedgePattern [x] [y] is set to a corresponding element of the divided pattern list WedgePatternTable by the following equation.
- the overlapping divided pattern is the same divided pattern as the generated divided pattern, or a divided pattern obtained by inverting each element value of the generated divided pattern (for example, 0 in FIG. 15B is 1).
- the division pattern is the same as the division pattern in which 1 is replaced with 0).
- wBlksize indicates the size of the width and height of the block that generates the division pattern
- NumWedgePattern [] uses the logarithmic value of the block size (log2 (wBlkSize)) as an argument to divide by block size
- log2 (wBlkSize) log2 (wBlkSize)
- Generate subdivision pattern with wedge direction wedgeOri 2 in the same way as in Fig. 16 (a)-(d), repeatedly subtracting 1 from the X coordinate of the start point S and subtracting 1 from the Y coordinate of the end point E To do.
- WedgePatternTable [log2 (blocksize)] [NumWedgePattern [log2 (wBlkSize)]] of wBlkSize ⁇ wBlkSize.
- WedgePatternTable [log2 (wBlkSize)] [NumWedgePattern [log2 (wBlkSize)]] of wBlkSize ⁇ wBlkSize.
- the buffer 145T5 records the division pattern list WedgePatternTable for each block size supplied from the division pattern list generation unit 145T4.
- [DMM1 division pattern deriving unit 145T6] Based on the input target PU size nS, the division pattern index wedge_full_tab_idx, and the preset reference division pattern size nBS, the DMM1 division pattern deriving unit 145T6 uses the division pattern wedgePattern [] from the division pattern list WedgePatternTable stored in the buffer 145T5. x] [y] is derived and output to the DC predicted value deriving unit 145T3.
- log2 (nS) is a logarithmic value with 2 as the target PU size.
- the DMM1 division pattern deriving unit 145T6 derives the size ratio scale between the target PU size and the reference division pattern size nBS by the following equation (eq.1).
- log2 (nS / nBS) log2 (nS / nBS) (eq.1)
- log2 (nS / nBS) is a logarithmic value with 2 as the base value obtained by dividing the target PU size nS by the reference division pattern size nBS.
- the conversion value with respect to a certain number X defined beforehand is stored in the lookup table, and the lookup table is changed. You may ask for it.
- the DMM1 division pattern deriving unit 145T6 scales the division pattern specified by the reference division pattern size nBS and the division pattern index wedge_full_tab_idx according to the following formula to obtain the division pattern of the target PU. To derive.
- the DMM1 division pattern deriving unit 145T6 uses the division pattern specified by the division pattern index wedge_full_tab_idx from the division pattern list corresponding to the reference division pattern size nBS. , And the division pattern is scaled to the target PU size to derive the division pattern wedgePattern [x] [y] of the target PU. Therefore, the division pattern list generation unit 145T4 can omit the generation process of the division pattern list for a block size larger than the reference division pattern size nBS, and can reduce the memory size related to the division pattern to be held by the buffer 145T5.
- the reference division pattern size nBS in the DMM1 division pattern deriving unit 145T6 is set to a value common to the maximum block size nMaxS for generating the division pattern list in the division pattern list deriving unit 145T4.
- the conventional technique requires a total memory size of 1,935,072 bits (about 242 KB) in order to hold a division pattern of 4 ⁇ 4 to 32 ⁇ 32.
- the memory size required for the division pattern for 1503 modes reaches 80% of the total.
- the 32 ⁇ 32 division pattern is generated by scaling the 16 ⁇ 16 division pattern.
- the mode of the division pattern in the 32 ⁇ 32 block size It is possible to reduce the processing amount required for determination to about 10%.
- the inventors' experiments have shown that this setting can achieve almost the same coding efficiency as the prior art.
- the number of modes of the division pattern of 16 ⁇ 16 block size and 32 ⁇ 32 block size is the same as the number of modes of the division pattern of 8 ⁇ 8 (766). It is possible to reduce the processing amount required to determine the division pattern mode in the ⁇ 16 block size by about 44% and reduce the processing amount required to determine the division pattern mode in the 32 ⁇ 32 block size by about 50%.
- the DMM1 division pattern generation unit 145T2 is configured to activate the division pattern list generation unit 145T4 at the first activation, generate a division pattern list for each block size, and record it in the buffer 145T5.
- the division pattern list generation unit 145T4 may be removed from the components of the DMM1 division pattern generation unit 145T2, and a division pattern list for each block size may be recorded in the buffer 145T5 in advance. In this way, it is possible to omit the process of generating the division pattern list for each block size.
- the DC predicted value deriving unit 145T3 roughly divides the target PU into two regions based on the division pattern wedgePattern [x] [y] indicating the division pattern of the target PU (for example, FIG. 15C). Region P1, P2), the input PT information, and the reference pixel p [x] [y] based on the prediction value for the region P1 and the prediction value for the region P2, and the prediction value derived for each region Is set to the predicted image predSamples [x] [y].
- the DC predicted value deriving unit 145T3 determines the upper left element wedgePattern [0] [0] and the upper right element wedgePattern [nS ⁇ of the divided pattern.
- a vertical edge flag vertEdgeFlag and a horizontal edge flag horEdgeFlag are derived by the following equations, respectively.
- the vertical edge flag vertEdgeFlag 1, it means that there is a division boundary on the upper side of the target PU, and when it is 0, it means that there is no division boundary.
- the horizontal edge flag horEdgeFlag is set to 0, and if not equal, 1 is set. To do.
- the horizontal edge flag horEdgeFlag is 1, it means that there is a division boundary on the left side of the target PU, and when it is 0, it means that there is no division boundary.
- the DC predicted value deriving unit 145T3 derives a depth predicted value (DC predicted value) to be allocated to the two divided regions.
- a region composed of elements having the same value as the uppermost leftmost element wedgePattern [0] [0] of the divided pattern is a region P1
- a region composed of elements having a different value from the uppermost leftmost element wedgePattern [0] [0] is defined as a region P2. It is assumed that the DC predicted value for the region P1 is dcValLT and the DC predicted value for the region P2 is dcValBR.
- the DC predicted value deriving unit 145T3 derives DC predicted values dcValLT and dcValBR according to the derived vertical edge flag verEdgeFlag and horizontal edge flag horEdgeFlag.
- each DC prediction value is derived by the following procedure. .
- the DC predicted value derivation unit 145T3 calculates the average value of the reference pixel p [-1] [0] and the reference pixel p [0] [-1] adjacent to the left and upper left of the target PU by the following equation. To dcValLT.
- dcValLT (p [-1] [0] + p [0] [-1]) >> 1
- a DC predicted value dcValBR is derived according to the horizontal edge flag horEdgeFlag.
- the average value of the reference pixels p [nS-1] [-1] adjacent to the upper rightmost pixel is set to the DC predicted value dcValBR.
- dcValBR (p [-1] [nS-1] + p [nS-1] [-1]) >> 1
- the horizontal edge strength (pixel difference) horAbsDiff and the vertical edge strength (pixel difference) verAbsDiff of the reference pixel are set.
- each DC prediction value is derived by the following procedure. .
- the reference pixel p [-1] [(nS-1) >> 1] adjacent to the left of the central pixel on the left side of the target PU Is the DC predicted value dcValLT of the region P1
- the reference pixel p [-1] [nS-1] adjacent above the upper right pixel of the target PU is the DC predicted value dcValBR of the region P2.
- the DC predicted value deriving unit 145T3 the derived DC predicted values dcValBR, dcValLT of each region, the DC offset presence / absence flag depth_dc_flag of each region of the target PU supplied from the variable length decoding unit 11, and the DC offset value DcOffset Based on [], the prediction image predSamples [x] [y] of the target PU is derived.
- the target pixel DC offset value dcOffset is set with reference to the DC offset presence / absence flag depth_dc_flag and the DC offset value DcOffset [].
- dcOffset depth_dc_flag? DcOffset [wedgePattern [x] [y]]: 0 That is, when the DC offset presence / absence flag is 1, the DC offset value dcOffset [wedgePattern [x] [y]] corresponding to the value of the division pattern wedgePattern [x] [y] is set to the DC offset value dcOffset of the target pixel. Set. When the DC offset presence / absence flag is 0, 0 is set to the DC offset value dcOffset of the target pixel.
- the DC offset presence flag is 1, it indicates that there is a DC offset value, and when it is 0, it indicates that the DC offset value is 0.
- the sum of the derived target pixel DC predicted value predDcVal and the target pixel DC offset value dcOffset is set as the predicted value of the target pixel.
- predSamples [x] [y] predDcVal + dcOffset
- the DC predicted value deriving unit 145T3 can derive the predicted image predSamples [x] [y] of the target PU.
- the predicted image generation unit included in the video decoding device 1 has the target PU when the block size of the target PU is equal to or smaller than the reference division pattern size.
- the division pattern specified by the block size and the division pattern index wedge_full_tab_idx is read from the division pattern list, and the division pattern to be applied to the target PU is derived.
- the block size of the target PU is larger than the reference division pattern size
- the division pattern specified by the reference division pattern size and the division pattern index wedge_full_tab_idx is read from the division pattern list, and the division pattern is read from the target PU.
- a scaling pattern applied to the target PU is derived by scaling to the block size.
- the DMM1 prediction it is possible to omit the generation process of the division pattern list having a block size larger than the reference division pattern size and to reduce the memory size required to hold the division pattern list.
- the lookup table for holding the division pattern of the first size is deleted, the division pattern of the second size smaller than the first size is scaled to the first size, By generating the size division pattern, it is possible to significantly reduce the memory size for holding the division pattern while maintaining the encoding efficiency.
- the moving image encoding device 2 is a device that generates and outputs encoded data # 1 by encoding the input image # 10.
- the input image # 10 is a layer image including one or a plurality of viewpoint images TexturePic and a depth map DepthPic at the same time corresponding to the viewpoint image TexturePic.
- FIG. 21 is a functional block diagram showing the configuration of the moving image encoding device 2.
- the moving image encoding device 2 includes an encoding setting unit 21, an inverse quantization / inverse conversion unit 22, a predicted image generation unit 23, an adder 24, a frame memory 25, a subtractor 26, a conversion / A quantization unit 27 and an encoded data generation unit 29 are provided.
- the encoding setting unit 21 generates image data related to encoding and various setting information based on the input image # 10.
- the encoding setting unit 21 generates the next image data and setting information.
- the encoding setting unit 21 generates the CU image # 100 for the target CU by sequentially dividing the input image # 10 into slice units, tree block units, and CU units.
- the encoding setting unit 21 generates header information H ′ based on the result of the division process.
- the header information H ′ includes (1) information on the size and shape of the tree block belonging to the target slice and the position in the target slice, and (2) the size, shape and shape of the CU belonging to each tree block.
- the encoding setting unit 21 refers to the CU image # 100 and the CU information CU 'to generate PT setting information PTI'.
- the PT setting information PTI ' includes information on all combinations of (1) possible division patterns of the target CU for each PU and (2) prediction modes that can be assigned to each PU.
- the encoding setting unit 21 supplies the CU image # 100 to the subtractor 26. In addition, the encoding setting unit 21 supplies the header information H ′ to the encoded data generation unit 29. Also, the encoding setting unit 21 supplies the PT setting information PTI ′ to the predicted image generation unit 23.
- the inverse quantization / inverse transform unit 22 performs inverse quantization and inverse orthogonal transform on the quantized prediction residual for each block supplied from the transform / quantization unit 27, thereby predicting the prediction residual for each block. To restore.
- the inverse orthogonal transform has already been described with respect to the inverse quantization / inverse transform unit 13 shown in FIG. 3, and thus the description thereof is omitted here.
- the inverse quantization / inverse transform unit 22 integrates the prediction residual for each block according to the division pattern specified by the TT division information (described later), and generates the prediction residual D for the target CU.
- the inverse quantization / inverse transform unit 22 supplies the prediction residual D for the generated target CU to the adder 24.
- the predicted image generation unit 23 refers to the local decoded image P ′ and the PT setting information PTI ′ recorded in the frame memory 25 to generate a predicted image Pred for the target CU.
- the predicted image generation unit 23 sets the prediction parameter obtained by the predicted image generation process in the PT setting information PTI ′, and transfers the set PT setting information PTI ′ to the encoded data generation unit 29. Note that the predicted image generation process performed by the predicted image generation unit 23 is the same as that performed by the predicted image generation unit 14 included in the video decoding device 1, and thus description thereof is omitted here.
- the adder 24 adds the predicted image Pred supplied from the predicted image generation unit 23 and the prediction residual D supplied from the inverse quantization / inverse transform unit 22 to thereby obtain the decoded image P for the target CU. Generate.
- Decoded decoded image P is sequentially recorded in the frame memory 25.
- decoded images corresponding to all tree blocks decoded prior to the target tree block for example, all tree blocks preceding in the raster scan order
- the time of decoding the target tree block It is recorded.
- the subtractor 26 generates a prediction residual D for the target CU by subtracting the prediction image Pred from the CU image # 100.
- the subtractor 26 supplies the generated prediction residual D to the transform / quantization unit 27.
- the transform / quantization unit 27 generates a quantized prediction residual by performing orthogonal transform and quantization on the prediction residual D.
- the orthogonal transformation refers to transformation from the pixel region to the frequency region.
- Examples of inverse orthogonal transformation include DCT transformation (DiscretecreCosine Transform), DST transformation (Discrete Sine Transform), and the like.
- the transform / quantization unit 27 refers to the CU image # 100 and the CU information CU 'and determines a division pattern of the target CU into one or a plurality of blocks. Further, according to the determined division pattern, the prediction residual D is divided into prediction residuals for each block.
- the transform / quantization unit 27 generates a prediction residual in the frequency domain by orthogonally transforming the prediction residual for each block, and then quantizes the prediction residual in the frequency domain to Generate quantized prediction residuals.
- the transform / quantization unit 27 generates the quantization prediction residual for each block, TT division information that specifies the division pattern of the target CU, information about all possible division patterns for each block of the target CU, and TT setting information TTI ′ including is generated.
- the transform / quantization unit 27 supplies the generated TT setting information TTI ′ to the inverse quantization / inverse transform unit 22 and the encoded data generation unit 29.
- the encoded data generation unit 29 encodes header information H ′, TT setting information TTI ′, and PT setting information PTI ′, and multiplexes the encoded header information H, TT setting information TTI, and PT setting information PTI. Coded data # 1 is generated and output.
- the predicted image generation unit included in the video encoding device 2 is the target when the block size of the target PU is equal to or smaller than the reference division pattern size.
- the division pattern specified by the PU block size and the division pattern index wedge_full_tab_idx is read from the division pattern list, and the division pattern to be applied to the target PU is derived.
- the block size of the target PU is larger than the reference division pattern size
- the division pattern specified by the reference division pattern size and the division pattern index wedge_full_tab_idx is read from the division pattern list, and the division pattern is read from the target PU.
- a scaling pattern applied to the target PU is derived by scaling to the block size.
- the DMM1 prediction it is possible to omit the generation process of the division pattern list having a block size larger than the reference division pattern size and to reduce the memory size required to hold the division pattern list.
- the lookup table for holding the division pattern of the first size is deleted, the division pattern of the second size smaller than the first size is scaled to the first size, By generating the size division pattern, it is possible to significantly reduce the memory size for holding the division pattern while maintaining the encoding efficiency.
- the above-described moving image encoding device 2 and moving image decoding device 1 can be used by being mounted on various devices that perform transmission, reception, recording, and reproduction of moving images.
- the moving image may be a natural moving image captured by a camera or the like, or may be an artificial moving image (including CG and GUI) generated by a computer or the like.
- moving picture encoding apparatus 2 and moving picture decoding apparatus 1 can be used for transmission and reception of moving pictures.
- FIG. 22 is a block diagram illustrating a configuration of a transmission device PROD_A in which the moving image encoding device 2 is mounted.
- the transmission device PROD_A modulates a carrier wave with an encoding unit PROD_A1 that obtains encoded data by encoding a moving image and the encoded data obtained by the encoding unit PROD_A1.
- a modulation unit PROD_A2 that obtains a modulation signal and a transmission unit PROD_A3 that transmits the modulation signal obtained by the modulation unit PROD_A2 are provided.
- the moving image encoding apparatus 2 described above is used as the encoding unit PROD_A1.
- the transmission device PROD_A is a camera PROD_A4 that captures a moving image, a recording medium PROD_A5 that records the moving image, and an input terminal PROD_A6 for inputting the moving image from the outside as a supply source of the moving image input to the encoding unit PROD_A1. And an image processing unit A7 for generating or processing an image.
- the configuration in which the transmission apparatus PROD_A includes all of these is illustrated, but a part thereof may be omitted.
- the recording medium PROD_A5 may be a recording of a non-encoded moving image, or a recording of a moving image encoded by a recording encoding scheme different from the transmission encoding scheme. It may be a thing. In the latter case, a decoding unit (not shown) for decoding the encoded data read from the recording medium PROD_A5 according to the recording encoding method may be interposed between the recording medium PROD_A5 and the encoding unit PROD_A1.
- FIG. 22 is a block diagram illustrating a configuration of the receiving device PROD_B in which the moving image decoding device 1 is mounted.
- the reception device PROD_B includes a reception unit PROD_B1 that receives a modulated signal, a demodulation unit PROD_B2 that obtains encoded data by demodulating the modulation signal received by the reception unit PROD_B1, and a demodulation A decoding unit PROD_B3 that obtains a moving image by decoding the encoded data obtained by the unit PROD_B2.
- the moving picture decoding apparatus 1 described above is used as the decoding unit PROD_B3.
- the receiving device PROD_B has a display PROD_B4 for displaying a moving image, a recording medium PROD_B5 for recording the moving image, and an output terminal for outputting the moving image to the outside as a supply destination of the moving image output by the decoding unit PROD_B3.
- PROD_B6 may be further provided.
- FIG. 22B illustrates a configuration in which the reception apparatus PROD_B includes all of these, but a part of the configuration may be omitted.
- the recording medium PROD_B5 may be used for recording a non-encoded moving image, or may be encoded using a recording encoding method different from the transmission encoding method. May be. In the latter case, an encoding unit (not shown) for encoding the moving image acquired from the decoding unit PROD_B3 according to the recording encoding method may be interposed between the decoding unit PROD_B3 and the recording medium PROD_B5.
- the transmission medium for transmitting the modulation signal may be wireless or wired.
- the transmission mode for transmitting the modulated signal may be broadcasting (here, a transmission mode in which the transmission destination is not specified in advance) or communication (here, transmission in which the transmission destination is specified in advance). Refers to the embodiment). That is, the transmission of the modulation signal may be realized by any of wireless broadcasting, wired broadcasting, wireless communication, and wired communication.
- a terrestrial digital broadcast broadcasting station (broadcasting equipment or the like) / receiving station (such as a television receiver) is an example of a transmitting device PROD_A / receiving device PROD_B that transmits and receives a modulated signal by wireless broadcasting.
- a broadcasting station (such as broadcasting equipment) / receiving station (such as a television receiver) of cable television broadcasting is an example of a transmitting device PROD_A / receiving device PROD_B that transmits and receives a modulated signal by cable broadcasting.
- a server workstation etc.
- Client television receiver, personal computer, smart phone etc.
- VOD Video On Demand
- video sharing service using the Internet is a transmitting device for transmitting and receiving modulated signals by communication.
- PROD_A / reception device PROD_B usually, either a wireless or wired transmission medium is used in a LAN, and a wired transmission medium is used in a WAN.
- the personal computer includes a desktop PC, a laptop PC, and a tablet PC.
- the smartphone also includes a multi-function mobile phone terminal.
- the video sharing service client has a function of encoding a moving image captured by the camera and uploading it to the server. That is, the client of the video sharing service functions as both the transmission device PROD_A and the reception device PROD_B.
- moving image encoding device 2 and moving image decoding device 1 can be used for recording and reproduction of moving images.
- FIG. 23 is a block diagram showing a configuration of a recording apparatus PROD_C in which the above-described moving picture encoding apparatus 2 is mounted.
- the recording device PROD_C includes an encoding unit PROD_C1 that obtains encoded data by encoding a moving image, and the encoded data obtained by the encoding unit PROD_C1 on the recording medium PROD_M.
- the moving image encoding apparatus 2 described above is used as the encoding unit PROD_C1.
- the recording medium PROD_M may be of a type built in the recording device PROD_C, such as (1) HDD (Hard Disk Drive) or SSD (Solid State Drive), or (2) SD memory. It may be of the type connected to the recording device PROD_C, such as a card or USB (Universal Serial Bus) flash memory, or (3) DVD (Digital Versatile Disc) or BD (Blu-ray Disc: registration) Or a drive device (not shown) built in the recording device PROD_C.
- HDD Hard Disk Drive
- SSD Solid State Drive
- SD memory such as a card or USB (Universal Serial Bus) flash memory, or (3) DVD (Digital Versatile Disc) or BD (Blu-ray Disc: registration) Or a drive device (not shown) built in the recording device PROD_C.
- the recording device PROD_C receives a moving image as a supply source of a moving image to be input to the encoding unit PROD_C1, a camera PROD_C3 that captures a moving image, an input terminal PROD_C4 for inputting a moving image from the outside, and a moving image. May include a receiving unit PROD_C5 and an image processing unit C6 that generates or processes an image.
- FIG. 23A illustrates a configuration in which the recording apparatus PROD_C includes all of these, but a part of the configuration may be omitted.
- the receiving unit PROD_C5 may receive a non-encoded moving image, or may receive encoded data encoded by a transmission encoding scheme different from the recording encoding scheme. You may do. In the latter case, a transmission decoding unit (not shown) that decodes encoded data encoded by the transmission encoding method may be interposed between the reception unit PROD_C5 and the encoding unit PROD_C1.
- Examples of such a recording device PROD_C include a DVD recorder, a BD recorder, and an HD (Hard Disk) recorder (in this case, the input terminal PROD_C4 or the receiving unit PROD_C5 is a main source of moving images).
- a camcorder in this case, the camera PROD_C3 is a main source of moving images
- a personal computer in this case, the receiving unit PROD_C5 is a main source of moving images
- a smartphone in this case, the camera PROD_C3 or The receiving unit PROD_C5 or the image processing unit C6 is a main supply source of moving images
- a recording apparatus PROD_C is also an example of such a recording apparatus PROD_C.
- FIG. 23 is a block showing a configuration of the playback device PROD_D in which the above-described video decoding device 1 is mounted.
- the playback device PROD_D reads a moving image by decoding a read unit PROD_D1 that reads encoded data written to the recording medium PROD_M and a coded data read by the read unit PROD_D1. And a decoding unit PROD_D2 to be obtained.
- the moving picture decoding apparatus 1 described above is used as the decoding unit PROD_D2.
- the recording medium PROD_M may be of the type built into the playback device PROD_D, such as (1) HDD or SSD, or (2) such as an SD memory card or USB flash memory, It may be of a type connected to the playback device PROD_D, or (3) may be loaded into a drive device (not shown) built in the playback device PROD_D, such as DVD or BD. Good.
- the playback device PROD_D has a display PROD_D3 that displays a moving image, an output terminal PROD_D4 that outputs the moving image to the outside, and a transmission unit that transmits the moving image as a supply destination of the moving image output by the decoding unit PROD_D2.
- PROD_D5 may be further provided.
- FIG. 23B illustrates a configuration in which the playback apparatus PROD_D includes all of these, but a part of the configuration may be omitted.
- the transmission unit PROD_D5 may transmit an unencoded moving image, or transmits encoded data encoded by a transmission encoding method different from the recording encoding method. You may do. In the latter case, it is preferable to interpose an encoding unit (not shown) that encodes a moving image with an encoding method for transmission between the decoding unit PROD_D2 and the transmission unit PROD_D5.
- Examples of such a playback device PROD_D include a DVD player, a BD player, and an HDD player (in this case, an output terminal PROD_D4 to which a television receiver or the like is connected is a main supply destination of moving images).
- a television receiver in this case, the display PROD_D3 is a main supply destination of moving images
- a digital signage also referred to as an electronic signboard or an electronic bulletin board
- the display PROD_D3 or the transmission unit PROD_D5 is a main supply of moving images.
- Desktop PC (in this case, the output terminal PROD_D4 or the transmission unit PROD_D5 is the main video image supply destination), laptop or tablet PC (in this case, the display PROD_D3 or the transmission unit PROD_D5 is a moving image)
- a smartphone which is a main image supply destination
- a smartphone in this case, the display PROD_D3 or the transmission unit PROD_D5 is a main moving image supply destination
- the like are also examples of such a playback device PROD_D.
- Each block of the moving picture decoding apparatus 1 and the moving picture encoding apparatus 2 described above may be realized in hardware by a logic circuit formed on an integrated circuit (IC chip), or may be a CPU (Central Processing). Unit) may be implemented in software.
- IC chip integrated circuit
- CPU Central Processing
- each device includes a CPU that executes instructions of a program that realizes each function, a ROM (Read (Memory) that stores the program, a RAM (Random Memory) that expands the program, the program, and various types
- a storage device such as a memory for storing data is provided.
- An object of the present invention is to provide a recording medium in which a program code (execution format program, intermediate code program, source program) of a control program for each of the above devices, which is software that realizes the above-described functions, is recorded in a computer-readable manner. This can also be achieved by supplying each of the above devices and reading and executing the program code recorded on the recording medium by the computer (or CPU or MPU).
- Examples of the recording medium include tapes such as magnetic tapes and cassette tapes, magnetic disks such as floppy (registered trademark) disks / hard disks, CD-ROMs (Compact Disc-Read-Only Memory) / MO discs (Magneto-Optical discs).
- tapes such as magnetic tapes and cassette tapes
- magnetic disks such as floppy (registered trademark) disks / hard disks
- CD-ROMs Compact Disc-Read-Only Memory
- MO discs Magnetic-Optical discs
- IC cards including memory cards
- Cards such as optical cards
- each of the devices may be configured to be connectable to a communication network, and the program code may be supplied via the communication network.
- the communication network is not particularly limited as long as it can transmit the program code.
- Internet Intranet, Extranet, LAN (Local Area Network), ISDN (Integrated Services Digital Network), VAN (Value-Added Network), CATV (Community Antenna television / Cable Television) communication network, Virtual Private Network (Virtual Private Network) Network), telephone line network, mobile communication network, satellite communication network, and the like.
- the transmission medium constituting the communication network may be any medium that can transmit the program code, and is not limited to a specific configuration or type.
- IEEE Institute of Electrical and Electronic Engineers 1394, USB, power line carrier, cable TV line, telephone line, ADSL (Asymmetric Digital Subscriber Line) line, etc. wired, such as IrDA (Infrared Data Association) and remote control, Bluetooth (registered trademark), IEEE 80 2.11 Wireless, HDR (High Data Rate), NFC (Near Field Communication), DLNA (Digital Living Network Alliance), mobile phone network, satellite line, terrestrial digital network, etc.
- the present invention can also be realized in the form of a computer data signal embedded in a carrier wave in which the program code is embodied by electronic transmission.
- a DMM prediction unit includes a DMM1 division pattern generation unit that derives a division pattern to be applied to a target PU, a division pattern derived by the DMM1 division pattern generation unit, and a decoded pixel adjacent to the target PU
- the DMM prediction unit comprising a DC prediction value deriving unit for deriving a prediction value for each region in the target PU based on DC offset information for each region in the target PU specified by the partition pattern
- the pattern generation unit derives a division pattern to be applied to the target PU based on a target PU size, a reference division pattern size, a division pattern index that specifies a division pattern to be applied to the target PU, and a division pattern list.
- the DMM prediction unit is the division pattern list corresponding to the target PU size when the target PU size is equal to or smaller than a reference division pattern size. From the division pattern list corresponding to the reference division pattern size when the division pattern specified by the division pattern index is set as a division pattern to be applied to the target PU and the target PU size is larger than the reference division pattern size. The division pattern specified by the division pattern index is scaled to the target PU size to derive a division pattern to be applied to the target PU.
- the DMM prediction unit according to aspect 3 of the present invention is characterized in that, in the above aspect 2, a size ratio for scaling a divided pattern is derived from the reference divided pattern size and the target PU size.
- the DMM prediction unit according to Aspect 4 of the present invention is the DMM prediction unit according to Aspect 3, wherein the size ratio is a logarithmic value based on 2 of the division PU from a logarithmic value based on 2 of the target PU size. It is a difference.
- the DMM prediction unit according to aspect 5 of the present invention further includes the second coordinates on the division pattern having the reference division pattern size corresponding to the first coordinates on the division pattern having the target PU size,
- the first coordinate is a coordinate shifted right by the size ratio.
- the DMM prediction unit further includes a division pattern list generation unit that generates a division pattern list for each block size in the above aspects 1 to 5, and the division pattern list generation unit includes a minimum block A division pattern list from a size to a reference division pattern size is generated.
- the DMM prediction unit according to aspect 7 of the present invention is characterized in that, in aspect 6 above, the minimum block size is a 4 ⁇ 4 block size.
- the DMM prediction unit according to aspect 8 of the present invention is characterized in that, in the above aspect 6, the reference division pattern size is 8 ⁇ 8 block size.
- the DMM prediction unit according to aspect 9 of the present invention is characterized in that, in the above aspect 6, the reference division pattern size is 16 ⁇ 16 block size.
- An image decoding apparatus is an image decoding apparatus including the DMM prediction unit according to any one of the above aspects 1 to 9 and a DMM prediction mode information decoding unit that decodes prediction mode information related to DMM prediction.
- the DMM prediction unit performs DMM1 prediction when the DMM prediction mode information indicates DMM1 prediction.
- An image encoding apparatus includes an image encoding apparatus including the DMM prediction unit according to any one of the above aspects 1 to 9 and a DMM prediction mode information encoding unit that encodes prediction mode information related to DMM prediction.
- the present invention can be suitably applied to an image decoding apparatus that decodes encoded data obtained by encoding image data and an image encoding apparatus that generates encoded data obtained by encoding image data. Further, the present invention can be suitably applied to the data structure of encoded data generated by an image encoding device and referenced by the image decoding device.
Landscapes
- Engineering & Computer Science (AREA)
- Multimedia (AREA)
- Signal Processing (AREA)
- Compression Or Coding Systems Of Tv Signals (AREA)
- Testing, Inspecting, Measuring Of Stereoscopic Televisions And Televisions (AREA)
Abstract
Description
以下、図面を参照しながら本発明の実施形態について説明する。
図4を用いて、動画像符号化装置2によって生成され、動画像復号装置1によって復号される符号化データ#1の構成例について説明する。符号化データ#1は、例示的に、シーケンス、およびシーケンスを構成する複数のピクチャを含む。
シーケンスレイヤでは、処理対象のシーケンスSEQ(以下、対象シーケンスとも称する)を復号するために動画像復号装置1が参照するデータの集合が規定されている。シーケンスSEQは、図4の(a)に示すように、ビデオパラメータセット(Video Parameter Set)シーケンスパラメータセットSPS(Sequence Parameter Set)、ピクチャパラメータセットPPS(Picture Parameter Set)、ピクチャPICT、及び、付加拡張情報SEI(Supplemental Enhancement Information)を含んでいる。ここで#の後に示される値はレイヤIDを示す。図4では、#0と#1すなわちレイヤ0とレイヤ1の符号化データが存在する例を示すが、レイヤの種類およびレイヤの数はこれによらない。
ピクチャレイヤでは、処理対象のピクチャPICT(以下、対象ピクチャとも称する)を復号するために動画像復号装置1が参照するデータの集合が規定されている。ピクチャPICTは、図4の(b)に示すように、ピクチャヘッダPH、及び、スライスS1~SNSを含んでいる(NSはピクチャPICTに含まれるスライスの総数)。
スライスレイヤでは、処理対象のスライスS(対象スライスとも称する)を復号するために動画像復号装置1が参照するデータの集合が規定されている。スライスSは、図4の(c)に示すように、スライスヘッダSH、及び、ツリーブロックTBLK1~TBLKNC(NCはスライスSに含まれるツリーブロックの総数)を含んでいる。
ツリーブロックレイヤでは、処理対象のツリーブロックTBLK(以下、対象ツリーブロックとも称する)を復号するために動画像復号装置1が参照するデータの集合が規定されている。
ツリーブロックヘッダTBLKHには、対象ツリーブロックの復号方法を決定するために動画像復号装置1が参照する符号化パラメータが含まれる。具体的には、図4の(d)に示すように、対象ツリーブロックの各CUへの分割パターンを指定するツリーブロック分割情報SP_TBLK、および、量子化ステップの大きさを指定する量子化パラメータ差分Δqp(qp_delta)が含まれる。
CUレイヤでは、処理対象のCU(以下、対象CUとも称する)を復号するために動画像復号装置1が参照するデータの集合が規定されている。
続いて、図4の(e)を参照しながらCU情報CUに含まれるデータの具体的な内容について説明する。図4の(e)に示すように、CU情報CUは、具体的には、スキップフラグSKIP、PT情報PTI、および、TT情報TTIを含む。
処理2:処理1にて得られた変換係数を量子化する;
処理3:処理2にて量子化された変換係数を可変長符号化する;
なお、上述した量子化パラメータqpは、動画像符号化装置2が変換係数を量子化する際に用いた量子化ステップQPの大きさを表す(QP=2qp/6)。
上述のとおり、予測情報PInfoには、インター予測情報およびイントラ予測情報の2種類がある。
以下では、本実施形態に係る動画像復号装置1の構成について、図1~図20を参照して説明する。
動画像復号装置1は、PU毎に予測画像を生成し、生成された予測画像と、符号化データ#1から復号された予測残差とを加算することによって復号画像#2を生成し、生成された復号画像#2を外部に出力する。
再び、図3を参照して、動画像復号装置1の概略的構成について説明すると次のとおりである。図3は、動画像復号装置1の概略的構成について示した機能ブロック図である。
可変長復号部11は、動画像復号装置1から入力された符号化データ#1に含まれる各種のパラメータを復号する。以下の説明では、可変長復号部11が、CABAC等のエントロピー符号化方式により符号化されているパラメータの復号を適宜行うものとする。可変長復号部11は、具体的には、以下の手順により、1フレーム分の符号化データ#1を復号する。
すなわち、DMMフラグには、デプスイントラ予測有無フラグの論理否定の値が設定される。DMMフラグは1の場合、デプスイントラ予測が利用されることを示し、DMMフラグが0の場合、デプスイントラ予測が利用されないことを示す。
可変長復号部11は、デプスイントラ予測有無フラグdim_not_present_flagが1の場合、さらに、デプスイントラモードフラグdepth_intra_mode_flagを復号する。該フラグは、デプスイントラ予測方式の選択に係るフラグである。該フラグが0の場合、デプスイントラ予測は、DMM1予測であることを示す。該フラグが1の場合、デプスイントラ予測は、DMM4予測であることを示す。
可変長復号部11は、デプスイントラ予測有無フラグdim_not_present_flagが0場合、対象PUのイントラ予測モードが推定予測モードMPMと一致するか否かを示すMPMフラグmpm_flagを復号する。該フラグが1の場合、対象PUのイントラ予測モードが推定予測モードMPMと一致することを示し、該フラグが0の場合、予測モード番号‘0’~‘34’(DC予測、Planar予測、Angular予測のいずれか)の中で、推定予測モードMPMを除くいずれかの予測モードであることを示す。
[逆量子化・逆変換部]
逆量子化・逆変換部13は、対象CUに含まれる各ブロックについて、TT情報TTIに基づいて逆量子化・逆変換処理を実行する。具体的には、逆量子化・逆変換部13は、各対象TUについて、対象TUに対応するTU情報TUIに含まれる量子化予測残差を逆量子化および逆直交変換することによって、画素毎の予測残差Dを復元する。なお、ここで直交変換とは、画素領域から周波数領域への直交変換のことを指す。したがって、逆直交変換は、周波数領域から画素領域への変換である。また、逆直交変換の例としては、逆DCT変換(Inverse Discrete Cosine Transform)、および逆DST変換(Inverse Discrete Sine Transform)等が挙げられる。逆量子化・逆変換部13は、復元した予測残差Dを加算器15に供給する。
予測画像生成部14は、対象CUに含まれる各PUについて、PT情報PTIに基づいて予測画像を生成する。具体的には、予測画像生成部14は、各対象PUについて、対象PUに対応するPU情報PUI(予測情報Pinfo)に含まれるパラメータに従ってイントラ予測またはインター予測を行うことにより、復号済み画像である局所復号画像P’から予測画像Predを生成する。予測画像生成部14は、生成した予測画像Predを加算器15に供給する。なお、予測画像生成部14の構成については、後ほど、より詳しく説明する。
加算器15は、予測画像生成部14より供給される予測画像Predと、逆量子化・逆変換部13より供給される予測残差Dとを加算することによって、対象CUについての復号画像Pを生成する。
フレームメモリ16には、復号された復号画像Pが順次記録される。フレームメモリ16には、対象ツリーブロックを復号する時点において、当該対象ツリーブロックよりも先に復号された全てのツリーブロック(例えば、ラスタスキャン順で先行する全てのツリーブロック)に対応する復号画像が記録されている。
前述の通り、予測画像生成部14は、PT情報PTIに基づいて予測画像を生成して出力する。対象CUがイントラCUの場合、予測画像生成部14に入力されるPU情報PTIは、予測モード(IntraPredMode)と、色差予測モード(IntraPredModeC)を含む。以下、予測モード(輝度・色差)の定義について、図7、図8、及び図9を参照して説明する。
図7は、動画像復号装置1で利用されるイントラ予測方式の分類と対応する予測モード番号の例を示している。Planar予測(INTRA_PLANAR)に‘0’、DC予測(INTRA_DC)に‘1’、Angular予測(INTRA_ANGULAR)に‘2’~ ‘34’、DMM1予測(INTRA_DMM_WFULL)に‘35’、DMM4予測(INTRA_DMM_CREDTEX)に‘36’の予測モード番号がそれぞれ割り当てられている。なお、予測モード番号‘X’(X=2..34)のAngular予測は、INTRA_ANGULARXとも呼ぶ。また、Angular予測において、‘10’の予測モード番号が割り当てられた予測方式を水平予測、‘26’の予測モード番号が割り当てられた予測方式を垂直予測とも呼び、水平予測、垂直予測、Angular予測を総称して方向予測と呼ぶ。方向予測は、対象PU周辺の隣接画素値を特定の方向に外挿することで予測画像を生成する予測方式である。また、DMM1予測、及びDMM4予測を総称して、デプスイントラ予測とも呼ぶ。デプスイントラ予測とは、基本的に、デプスマップ上の対象ブロック(デプスブロックとも称する)は、2つの非矩形の平坦領域から構成され、各平坦領域のデプス値は固定値で表現されるというデプスモデルに基づいている。また、デプスモデルは、各画素が属する領域を表わすパーティション情報、及び各領域のデプス値情報から構成される。DMM予測において、デプスブロックの分割方法として、2種類の異なるパーティションタイプ、すなわち、ウェッジレット分割 (Wedgelet Partition)、及び輪郭分割(Contour Partition)がある。デプスイントラ予測の詳細については後述する。
次に、図11を用いて予測画像生成部14の構成についてさらに詳しく説明する。図11は予測画像生成部14の構成例について示す機能ブロック図である。なお、本構成例は、予測画像生成部14の機能のうち、イントラCUの予測画像生成に係る機能ブロックを図示している。
ここで、(xB,yB)は対象PU内左上画素の位置、nSは対象PUのサイズを表し、対象PUの幅または高さのうち大きい方の値を示す。上式では、基本的には、対象PUの上辺に隣接する復号画素のラインおよび対象PUの左辺に隣接する復号画素のコラムに含まれる復号画素値を対応する参照画素値にコピーしている。なお、特定の参照画素位置に対応する復号画素値が存在しない、または、参照できない場合には、既定の値、例えば、1<<(BitDepth-1)を利用してもよい。ここで、BitDepthは、画素のビット深度である。また、既定の値の代わりに、対応する復号画素値の近傍に存在する参照可能な復号画素値を利用してもよい。
次に、予測画像生成部14におけるCU単位の予測画像生成処理の概略を図14のフローチャートを用いて説明する。CU単位の予測画像生成処理が始まると、まず、予測単位設定部141がCU内に含まれるPUの一つを既定の順序に従って対象PUに設定して対象PU情報を参照画素設定部142およびスイッチ143に出力する(S11)。次に、参照画素設定部142は対象PUの参照画素を、外部のフレームメモリから読み出した復号画素値を用いて設定する(S12)。次に、スイッチ143が、入力された対象PU情報に基づいて対象PUが輝度か色差か、又は予測モードpredModeIntraがDMM予測であるか否かを判定し、当該判定結果に応じて出力を切り替える(S13)。
続いて、予測画像導出部145の詳細について説明する。図11に示すように、予測画像導出部145は、さらに、DC予測部145D、Planar予測部145P、Angular予測部145A、及びDMM予測部145Tを備える。
ここで、x, y = 0..nS-1であり、 k = log2( nS )と定義される。
Angular予測部145Aは、入力される予測モードpredModeIntraに対応する予測方向(参照方向)の参照画素を用いて対象PU内に対応する予測画像を生成して出力する。Angular予測による予測画像の生成処理では、予測モードpredModeIntraの値に応じて主参照画素を設定し、予測画像をPU内のラインまたはコラムの単位で主参照画素を参照して生成する。
Angular予測部145Aは、主方向フラグbRefVerが1のとき(主方向が垂直;predModeIntra>=18)、予測画像の生成単位をラインに設定し、対象PUの上方の参照画素を主参照画素に設定する。具体的には以下の手順で参照画素p[x][y]の値を用いて主参照画素ref[x]が設定される。
(2)予測モードpredModeIntraに対応する勾配intraPredAngleが0未満の場合、以下の式により、xを変数として、x=-1..(nS*intraPredAngle)>>5の範囲まで、以下の式により、対象PUの左辺に隣接する参照画素p[x][y]を主参照画素ref[x][y]へ設定する。
それ以外の場合(intraPredAngle>=0)、xを変数として、x=nS+1..2*nSの範囲まで、以下の式により、対象PUの上辺に隣接する参照画素p[x][y]を主参照画素ref[x][y]へ設定する。
なお、ここでinvAngleは予測方向の変位intraPredAngleの逆数をスケール(8192を乗算)した値に相当する。
iFact = ( ( y + 1 )*intraPredAngle ) & 31
ここで、‘&’は論理積のビット演算を表す演算子であり、“A&31”の結果は、整数Aを32で除算した余りを意味する。以降も同様である。
((32-iFact)*ref[x+iIdx+1] + iFact*ref[x+iIdx+2] + 16) >> 5
それ以外(iFact!=0)、主参照画素ref[x+iIdx+1]を予測画像predSamples[x][y]へ設定する
predSamples[x][y] = ref[x+iIdx+1]
(主方向が水平の場合)
主方向フラグbRefVerの値が0(主方向が水平方向; predModeIntra<18)の場合、予測画像の生成単位をコラムに設定するとともに対象PUの左側の参照画素を主参照画素に設定する。具体的には以下の手順により参照画素p[x][y]の値を用いて主参照画素ref[x]が設定される。
(2)予測モードpredModeIntraに対応する勾配intraPredAngleが0未満の場合、以下の式により、xを変数として、x=-1..(nS*intraPredAngle)>>5の範囲まで、以下の式により、対象PUの上辺に隣接する参照画素p[x][y]を主参照画素ref[x][y]へ設定する。
それ以外の場合(intraPredAngle>=0)、xを変数として、x=nS+1..2*nSの範囲まで、以下の式により、対象PUの左辺に隣接する参照画素p[x][y]を主参照画素ref[x][y]へ設定する。
Angular予測部145Aは、予測対象コラムと主参照画素の距離(x+1)と勾配intraPredAngleに応じて計算される予測参照画素の生成に用いる主参照画素の位置を表す、画素単位における整数精度の位置iIdx、画素単位における小数点精度の位置iFactを、以下の式で導出する。
iFact = ( ( x + 1 )*intraPredAngle ) & 31
Angular予測部145Aは、導出した変数iFactに応じて、対象PUの予測画像predSamples[x][y]を次式により導出する。
それ以外(iFact!=0)の場合、主参照画素ref[x+iIdx+1]を予測画像predSamples[x][y]へ設定する
predSamples[x][y] = ref[x+iIdx+1]
[DMM予測部145T]
DMM予測部145Tは、入力される予測モードpredModeIntraに対応するDMM予測(Depth Modeling Mode,デプスイントラ予測ともいう)に基づいて、対象PU内に対応する予測画像を生成して出力する。
DMM4分割パターン生成部145T1は、対象PUの分割パターンwedgePattern[x][y]を、デプスマップDepthPic上の対象PUに対応する視点画像TexturePic上の輝度の復号画素値recTexPicに基づいて導出し、DC予測値導出部145T3へ出力する。概略的には、DMM4分割パターン生成部は、デプスマップ上の対象PUの2つの領域P1、P2を、対応する視点画像TexturePic上の対象ブロックの輝度の平均値によって、対象ブロックを二値化することで導出する。
参照画素refSamples[x][y]に基づいて、対応ブロックの画素値の総和sumRefValsを下記式で導出する。
次に、総和sumRefValsと対象PUサイズnSに基づいて、閾値threshValsを下記式で導出する。すなわち、対応ブロックの平均画素値を導出する。
ここで、上記式の代わり、総和sumRefValsを対象PUサイズnSの二乗数nS*nSで除算した値を閾値threshValとしてもよい。
すなわち、参照画素refSamples[x][y]が、閾値threshValより大きい場合には、分割パターンの要素(x,y)に1を設定する。参照画素refSamples[x][y]が、閾値threshVal以下の場合には、分割パターンの要素(x,y)に0を設定する。
DMM1分割パターン生成部145T2は、さらに、DMM1分割パターン導出部145T6、バッファ145T5、及び分割パターンリスト生成部145T4を備える。概略的には、DMM1分割パターン生成部145T2は、初回起動時のみ、分割パターンリスト生成部145T4を起動させ、ブロックサイズ毎の分割パターンリストWedgePatternTableを生成する。次に、生成した分割パターンリストをバッファ145T5に格納する。続いて、DMM分割パターン導出部145T6は、入力される対象PUサイズnS、分割パターンインデックスwedge_full_tab_idx、及び予め設定された基準分割パターンサイズnBSに基づいて、バッファ145T5に格納された分割パターンリストWedgePatternTableから分割パターンwedgePattern[x][y]を導出して、DC予測値導出部145T3へ出力する。
分割パターンリスト生成部145T6における分割パターンリストの生成方法の説明に先だって、分割パターンの生成方法の概要について、図16を参照して説明する。まず、全要素が0の分割パターンを生成する。次に、分割パターン内に、始点S(xs,ys)と終点E(xe,ye)を設定する。図16の(a)の例では、始点S(xs,ys)=(3,blocksize-1)、終点E(xe,ye)=(blocksize-1,2)である。次に、始点Sと終点Eの間をBresenhamのアルゴリズムを用いて線分を引く(図16(b)の斜線の要素)。図16(c)の例では、続いて、図16(d)に示すように、その線分上及び線分より右側の座標に対応する要素を1に設定することで、分割パターンwedgePattern[x][y]が生成される。ここで、blocksizeは、分割パターンを生成するブロックのサイズ(縦幅、横幅)である。
wedgePattern[x][y], with x = 0..wBlkSize-1, y = 0..wBlkSize-1
なお、重複する分割パターンとは、生成された分割パターンと同一の分割パターン、又は、生成された分割パターンの各要素値を反転させた分割パターン(例えば、図15(b)において、0を1へ、1を0へ置換した分割パターン)と同一の分割パターンである。ここで、wBlksizeとは、分割パターンを生成するブロックの幅及び高さのサイズを示し、配列NumWedgePattern[]は、ブロックサイズの対数値(log2(wBlkSize))を引数とする、ブロックサイズ別の分割パターンリストに追加された分割パターンの個数(分割パターン数)を表わす。分割パターンリストへ分割パターンが追加される度に、分割パターン数NumWedgePattern[log2(wBlkSize)]は1加算される。なお、NumWedgePatternの初期値は0である。
バッファ145T5は、分割パターンリスト生成部145T4より供給されるブロックサイズ別の分割パターンリストWedgePatternTableを記録する。
DMM1分割パターン導出部145T6は、入力される対象PUサイズnS、分割パターンインデックスwedge_full_tab_idx、及び予め設定された基準分割パターンサイズnBSに基づいて、バッファ145T5に格納された分割パターンリストWedgePatternTableから分割パターンwedgePattern[x][y]を導出し、DC予測値導出部145T3へ出力する。
より具体的には、対象PUサイズnSが、基準分割パターンサイズnBS以下の場合には、対象PUサイズに対応する分割パターンリストから、分割パターンインデックスwedge_full_tab_idxで指定される分割パターンを読み出して出力する。すなわち、下記式によって、分割パターンwedgePattern[x][y]は導出される。
with x = 0..nS-1, y = 0..nS-1
ここで、log2(nS)は、対象PUサイズの2を底とする対数値である。
対象PUサイズnSが、基準分割パターンサイズnBSより大きい場合には、基準分割パターンサイズnBSに対応する分割パターンリストから、分割パターンインデックスwedge_full_tab_idxで指定される分割パターンを読み出し、該分割パターンを、対象PUサイズへスケーリングして、対象PUの分割パターンwedgePattern[x][y]を導出して出力する。
ここで、log2(nS/nBS)は、対象PUサイズnSを基準分割パターンサイズnBSで除算した値の2を底とする対数値である。なお、2を底とする対数値を求める場合には、演算子log2()の代わりに、予め定義された、ある数Xに対する変換値をルックアップテーブルに格納しておき、そのルックアップテーブルを参照して求めてもよい。
すなわち、対象PUサイズにおける分割パターン上の第1の座標(x1,y1)と対応する基準分割パターンサイズにおける分割パターン上の第2の座標(x2,y2)は、上記第1の座標を上記サイズ比scaleによって右シフトした座標である。
従って、対象PUサイズにおける分割パターン上の第1の座標(x1,y1)に設定される値は、対応する基準ウェッジ分割パターンサイズにおける分割パターン上の第2の座標(x2,y2)の値である。
なお、本実実施例に係るDMM1分割パターン生成部145T2は、初回起動時に、分割パターンリスト生成部145T4を起動させ、ブロックサイズ別の分割パターンリストを生成し、バッファ145T5へ記録する構成であったが、これに限定されない。例えば、DMM1分割パターン生成部145T2の構成要素から、分割パターンリスト生成部145T4を取り除き、予めブロックサイズ別の分割パターンリストをバッファ145T5に記録しておく構成としてもよい。このようにすれば、ブロックサイズ別の分割パターンリストの生成処理を省略することが可能である。
DC予測値導出部145T3は、概略的には、対象PUの分割パターンを示す分割パターンwedgePattern[x][y]に基づいて、対象PUを2つの領域に分割し(例えば、図15 (c)に示す領域P1、P2)、入力されたPT情報、及び参照画素p[x][y]に基づいて領域P1に関する予測値、及び領域P2に関する予測値を導出し、各領域に導出した予測値を予測画像predSamples[x][y]に設定して導出する。
horEdgeFlag = ( wedgePattern[0][0] != wedgePattern[0][nS-1] )
すなわち、最左上要素wedgePattern[0][0]と最左下要素wedgePattern[nS-1][0]の値が等しい場合、垂直エッジフラグvertEdgeFlagに0を設定し、等しくない場合、1を設定する。垂直エッジフラグvertEdgeFlagが1の場合、対象PUの上辺上に分割境界があることを意味し、0であれば、分割境界がないことを意味する。
続いて、水平エッジフラグhorEdgeFlagに応じて、DC予測値dcValBRを導出する。水平エッジフラグhorEdgeFlagが1(horEdgeFlag==1,図20上の(d))の場合、対象PUの最左下画素の左に隣接する参照画素p[-1][nS-1]と、対象PUの最右上画素の上に隣接する参照画素p[nS-1][-1]の平均値をDC予測値dcValBRへ設定する。
水平ヘッジフラグhorEdgeFlagが0(horEdgeFlag==0,図20上の(a))の場合、参照画素の水平方向のエッジ強度(画素差)horAbsDiffと、垂直方向のエッジ強度(画素差)verAbsDiffとを比較して、強度(画素差)の大きい方向の参照画素に基づいて、DC予測値を導出する。すなわち、水平方向のエッジ強度horAbsDiffが、垂直方向のエッジ強度verAbsDiffより大きい場合には、参照画素p[2*nS-1][-1]をDC予測値dcValBRへ設定する。それ以外の場合(horAbsDiff<=vertAbsDiff)、参照画素p[-1][2*nS-1]をDC予測値dcValBRへ設定する。
horAbsDiff = Abs(p[0][-1] - p[2*nS - 1][-1])
dcValBR = (horAbsDiff > verAbsDiff ) ? p[2*nS - 1][-1] : p[-1][2*nS-1]
(2)垂直エッジフラグvertEdgeFlagと水平エッジフラグhorEdgeFlagが異なる場合(vertEdgeFlag!=horEdgeFlag, 図20上の(b)、及び(c)の分割パターン)、以下の手順で、各DC予測値を導出する。
dcValBR = horEdgeFlag ? p[-1][nS - 1] : p[nS - 1][-1]
すなわち、水平エッジフラグhorEdgeFlagが1の場合(垂直エッジフラグvertEdgeFlagが0)には、対象PUの上辺の中央画素の上に隣接する参照画素p[(nS-1)>>1][-1]を領域P1のDC予測値dcValLTとし、対象PUの最左下画素の左に隣接する参照画素p[-1][nS-1]を領域P2のDC予測値dcValBRとする。一方、水平エッジフラグhorEdgeFlagが0の場合(垂直エッジフラグvertEdgeFlagが1)には、対象PUの左辺の中央画素の左に隣接する参照画素p[-1][(nS-1)>>1]を領域P1のDC予測値dcValLTとし、対象PUの最右上画素の上に隣接する参照画素p[-1][nS-1]を領域P2のDC予測値dcValBRとする。
次に、DCオフセット有無フラグdepth_dc_flag、及びDCオフセット値DcOffset[]を参照して、対象画素DCオフセット値dcOffsetを設定する。
すなわち、DCオフセット有無フラグが1の場合には、分割パターンwedgePattern[x][y]の値に対応するDCオフセット値dcOffset[wedgePattern[x][y]]を、対象画素のDCオフセット値dcOffsetへ設定する。DCオフセット有無フラグが0の場合、対象画素のDCオフセット値dcOffsetに0を設定する。ここで、DCオフセット有無フラグは1の場合、DCオフセット値があることを示し、0の場合には、DCオフセット値は0であることを示す。
以上のようにして、DC予測値導出部145T3は、対象PUの予測画像predSamples[x][y]を導出することができる。
以上説明した本実施形態に係る動画像復号装置1の備える予測画像生成部は、対象PUにおいて、DMM1予測が選択された場合、対象PUのブロックサイズが、基準分割パターンサイズ以下の場合、対象PUのブロックサイズと、分割パターンインデックスwedge_full_tab_idxにより指定される分割パターンを、分割パターンリストより読み出して、対象PUに適用する分割パターンを導出する。一方、対象PUのブロックサイズが、基準分割パターンサイズより大きい場合には、基準分割パターンサイズと、分割パターンインデックスwedge_full_tab_idxにより指定される分割パターンを、分割パターンリストより読み出し、該分割パターンを対象PUのブロックサイズへスケーリングすることで、対象PUに適用する分割パターンを導出する。
以下において、本実施形態に係る動画像符号化装置2について、図21を参照して説明する。
動画像符号化装置2は、概略的に言えば、入力画像#10を符号化することによって符号化データ#1を生成し、出力する装置である。ここで、入力画像#10は、1又は複数の視点画像TexturePic、及び視点画像TexturePicに対応する同時刻のデプスマップDepthPicからなるレイヤ画像である。
まず、図21を用いて、動画像符号化装置2の構成例について説明する。図21は、動画像符号化装置2の構成について示す機能ブロック図である。図21に示すように、動画像符号化装置2は、符号化設定部21、逆量子化・逆変換部22、予測画像生成部23、加算器24、フレームメモリ25、減算器26、変換・量子化部27、および符号化データ生成部29を備えている。
以上説明した本実施形態に係る動画像符号化装置2の備える予測画像生成部は、対象PUにおいて、DMM1予測が選択された場合、対象PUのブロックサイズが、基準分割パターンサイズ以下の場合、対象PUのブロックサイズと、分割パターンインデックスwedge_full_tab_idxにより指定される分割パターンを、分割パターンリストより読み出して、対象PUに適用する分割パターンを導出する。一方、対象PUのブロックサイズが、基準分割パターンサイズより大きい場合には、基準分割パターンサイズと、分割パターンインデックスwedge_full_tab_idxにより指定される分割パターンを、分割パターンリストより読み出し、該分割パターンを対象PUのブロックサイズへスケーリングすることで、対象PUに適用する分割パターンを導出する。
上述した動画像符号化装置2及び動画像復号装置1は、動画像の送信、受信、記録、再生を行う各種装置に搭載して利用することができる。なお、動画像は、カメラ等により撮像された自然動画像であってもよいし、コンピュータ等により生成された人工動画像(CGおよびGUIを含む)であってもよい。
また、上述した動画像復号装置1および動画像符号化装置2の各ブロックは、集積回路(ICチップ)上に形成された論理回路によってハードウェア的に実現してもよいし、CPU(Central Processing Unit)を用いてソフトウェア的に実現してもよい。
2.11無線、HDR(High Data Rate)、NFC(Near Field Communication)、DLNA(Digital Living Network Alliance)、携帯電話網、衛星回線、地上波デジタル網等の無線でも利用可能である。なお、本発明は、上記プログラムコードが電子的な伝送で具現化された、搬送波に埋め込まれたコンピュータデータ信号の形態でも実現され得る。
〔まとめ〕
11 可変長復号部(DMM予測モード情報復号部)
13 逆量子化・逆変換部
14 予測画像生成部
141 予測単位設定部
142 参照画素設定部
143 スイッチ
144 参照画素フィルタ部
145 予測画像導出部
145D DC予測部
145P Planar予測部
145A Angular予測部
145T DMM予測部
145T1 DMM4分割パターン生成部
145T2 DMM1分割パターン生成部
145T3 DC予測値導出部
145T5 分割パターンリスト生成部
145T5 バッファ
145T6 DMM1分割パターン導出部
15 加算器
16 フレームメモリ
2 動画像符号化装置
21 符号化設定部
22 逆量子化・逆変換部
23 予測画像生成部
24 加算器
25 フレームメモリ
26 減算器
27 変換・量子化部
29 符号化データ生成部(DMM予測モード情報符号化部)
Claims (11)
- 対象PUへ適用する分割パターンを導出するDMM1分割パターン生成部と、前記DMM1分割パターン生成部によって導出された分割パターン、対象PUに隣接する復号済画素、及び前記分割パターンによって特定される対象PU内の領域毎のDCオフセット情報に基づいて、前記対象PU内の領域毎の予測値を導出するDC予測値導出部を備えるDMM予測部において、
前記DMM1分割パターン生成部は、対象PUサイズ、基準分割パターンサイズ、対象PUへ適用する分割パターンを指定する分割パターンインデックス、及び分割パターンリストに基づいて、前記対象PUへ適用する分割パターンを導出することを特徴とするDMM予測部。 - 前記DMM1分割パターン生成部は、前記対象PUサイズが基準分割パターンサイズ以下の場合、前記対象PUサイズに対応する分割パターンリストから、前記分割パターンインデックスによって指定される分割パターンを前記対象PUへ適用する分割パターンとして設定し、
前記対象PUサイズが基準分割パターンサイズより大きい場合、前記基準分割パターンサイズに対応する分割パターンリストから、前記分割パターンインデックスによって指定される分割パターンを、前記対象PUサイズへスケーリングして前記対象PUへ適用する分割パターンを導出することを特徴とする請求項1に記載のDMM予測部。 - 前記基準分割パターンサイズと前記対象PUサイズから分割パターンをスケーリングするためのサイズ比を導出することを特徴とする請求項2に記載のDMM予測部。
- 前記サイズ比は、前記対象PUサイズの2を底とする対数値から、前記分割パターンの2を底とする対数値の差であることを特徴とする請求項3に記載のDMM予測部。
- 前記対象PUサイズの分割パターン上の第1の座標と対応する基準分割パターンサイズの分割パターン上の第2の座標は、前記第1の座標を前記サイズ比によって右シフトした座標であることを特徴とする請求項4に記載のDMM予測部。
- 前記DMM予測部は、さらに、ブロックサイズ別の分割パターンリストを生成する分割パターンリスト生成部を備え、
前記分割パターンリスト生成部は、最小ブロックサイズから基準分割パターンサイズまでの分割パターンリストを生成することを特徴とする請求項1から5に記載のDMM予測部。 - 前記最小ブロックサイズは、4×4ブロックサイズであることを特徴とする請求項6に記載のDMM予測部。
- 前記基準分割パターンサイズは、8×8ブロックサイズであることを特徴とする請求項6に記載のDMM予測部。
- 前記基準分割パターンサイズは、16×16ブロックサイズであることを特徴とする請求項6に記載のDMM予測部。
- 上記請求項1から請求項9の何れか一項に記載のDMM予測部と、
DMM予測に関するDMM予測モード情報を復号するDMM予測モード情報復号部を備える画像復号装置であって、
上記DMM予測部は、前記DMM予測モード情報が、DMM1予測を示す場合に、DMM1予測を行うことを特徴とする画像復号装置。 - 上記請求項1から請求項9の何れか一項に記載のDMM予測部と、
DMM予測に関するDMM予測モード情報を符号化するDMM予測モード情報符号化部を備える画像符号化装置であって、
上記DMM予測部は、前記DMM予測モード情報が、DMM1予測を示す場合に、DMM1予測を行うことを特徴とする画像符号化装置。
Priority Applications (3)
Application Number | Priority Date | Filing Date | Title |
---|---|---|---|
US15/321,083 US10136161B2 (en) | 2014-06-24 | 2015-06-22 | DMM prediction section, image decoding device, and image coding device |
JP2016529577A JPWO2015199040A1 (ja) | 2014-06-24 | 2015-06-22 | Dmm予測部、画像復号装置、および画像符号化装置 |
US16/134,676 US20190037242A1 (en) | 2014-06-24 | 2018-09-18 | Image decoding device, an image encoding device, and an image decoding method |
Applications Claiming Priority (2)
Application Number | Priority Date | Filing Date | Title |
---|---|---|---|
JP2014129099 | 2014-06-24 | ||
JP2014-129099 | 2014-06-24 |
Related Child Applications (2)
Application Number | Title | Priority Date | Filing Date |
---|---|---|---|
US15/321,083 A-371-Of-International US10136161B2 (en) | 2014-06-24 | 2015-06-22 | DMM prediction section, image decoding device, and image coding device |
US16/134,676 Continuation US20190037242A1 (en) | 2014-06-24 | 2018-09-18 | Image decoding device, an image encoding device, and an image decoding method |
Publications (1)
Publication Number | Publication Date |
---|---|
WO2015199040A1 true WO2015199040A1 (ja) | 2015-12-30 |
Family
ID=54938122
Family Applications (1)
Application Number | Title | Priority Date | Filing Date |
---|---|---|---|
PCT/JP2015/067921 WO2015199040A1 (ja) | 2014-06-24 | 2015-06-22 | Dmm予測部、画像復号装置、および画像符号化装置 |
Country Status (3)
Country | Link |
---|---|
US (2) | US10136161B2 (ja) |
JP (1) | JPWO2015199040A1 (ja) |
WO (1) | WO2015199040A1 (ja) |
Cited By (2)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
CN109937570A (zh) * | 2016-10-14 | 2019-06-25 | 世宗大学校产学协力团 | 影像编码方法/装置、影像解码方法/装置以及保存有比特流的记录介质 |
US11778184B1 (en) | 2016-10-14 | 2023-10-03 | Industry Academy Cooperation Foundation Of Sejong University | Image encoding method/device, image decoding method/device, and recording medium in which bitstream is stored |
Families Citing this family (10)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
EP3178229A4 (en) * | 2014-09-30 | 2018-03-14 | HFI Innovation Inc. | Method of lookup table size reduction for depth modelling mode in depth coding |
JP2018050091A (ja) * | 2015-02-02 | 2018-03-29 | シャープ株式会社 | 画像復号装置、画像符号化装置および予測ベクトル導出装置 |
CN113347446B (zh) * | 2016-02-09 | 2022-10-04 | 弗劳恩霍夫应用研究促进协会 | 解码器、方法以及可读存储介质 |
US11134235B2 (en) * | 2016-07-29 | 2021-09-28 | Sony Corporation | Image processing apparatus and image processing method |
JP6565885B2 (ja) * | 2016-12-06 | 2019-08-28 | 株式会社Jvcケンウッド | 画像符号化装置、画像符号化方法及び画像符号化プログラム、並びに画像復号化装置、画像復号化方法及び画像復号化プログラム |
CN115118992B (zh) * | 2017-08-22 | 2024-02-06 | 松下电器(美国)知识产权公司 | 图像编码器、图像解码器、和比特流生成设备 |
US20200195944A1 (en) * | 2018-12-14 | 2020-06-18 | Advanced Micro Devices, Inc. | Slice size map control of foveated coding |
JP2022068379A (ja) * | 2019-03-08 | 2022-05-10 | シャープ株式会社 | 画像復号装置 |
EP3922018A4 (en) * | 2019-03-12 | 2022-06-08 | Zhejiang Dahua Technology Co., Ltd. | SYSTEMS AND METHODS FOR IMAGE CODING |
WO2021110045A1 (en) * | 2019-12-03 | 2021-06-10 | Huawei Technologies Co., Ltd. | Coding method, device, system with merge mode |
Citations (1)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
WO2014005248A1 (en) * | 2012-07-02 | 2014-01-09 | Qualcomm Incorporated | Intra-coding of depth maps for 3d video coding |
Family Cites Families (4)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
US7728878B2 (en) * | 2004-12-17 | 2010-06-01 | Mitsubishi Electric Research Labortories, Inc. | Method and system for processing multiview videos for view synthesis using side information |
KR101807886B1 (ko) * | 2009-10-14 | 2017-12-11 | 돌비 인터네셔널 에이비 | 깊이 맵 처리를 위한 방법 및 디바이스 |
WO2015005750A1 (ko) * | 2013-07-12 | 2015-01-15 | 삼성전자 주식회사 | 변이 벡터 유도를 사용하는 비디오 부호화 방법 및 그 장치, 비디오 복호화 방법 및 그 장치 |
US9756359B2 (en) * | 2013-12-16 | 2017-09-05 | Qualcomm Incorporated | Large blocks and depth modeling modes (DMM'S) in 3D video coding |
-
2015
- 2015-06-22 WO PCT/JP2015/067921 patent/WO2015199040A1/ja active Application Filing
- 2015-06-22 JP JP2016529577A patent/JPWO2015199040A1/ja active Pending
- 2015-06-22 US US15/321,083 patent/US10136161B2/en not_active Expired - Fee Related
-
2018
- 2018-09-18 US US16/134,676 patent/US20190037242A1/en not_active Abandoned
Patent Citations (1)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
WO2014005248A1 (en) * | 2012-07-02 | 2014-01-09 | Qualcomm Incorporated | Intra-coding of depth maps for 3d video coding |
Non-Patent Citations (4)
Title |
---|
GERHARD TECH ET AL.: "3D-HEVC Draft Text 4", JOINT COLLABORATIVE TEAM ON 3D VIDEO CODING EXTENSIONS OF ITU-T SG 16 WP 3 AND ISO/IEC JTC 1/SC 29/WG 11, JCT3V-H1001-V2, 8TH MEETING, May 2014 (2014-05-01), Valencia, ES, pages i - x, 9-16, 22-43 * |
TAKESHI TSUKUBA ET AL.: "Lookup table size reduction in DMM1", JOINT COLLABORATIVE TEAM ON 3D VIDEO CODING EXTENSIONS OF ITU-T SG 16 WP 3 AND ISO/IEC JTC 1/SC 29/WG 11, JCT3V- I0110, 9TH MEETING, June 2014 (2014-06-01), Sapporo, JP, pages 1 - 9 * |
WEIRAN LI ET AL.: "CE5.h related: Simplification on DMM1 pattern generation mechanism", JOINT COLLABORATIVE TEAM ON 3D VIDEO CODING EXTENSIONS OF ITU-T SG 16 WP 3 AND ISO/IEC JTC 1/SC 29/WG 11, JCT3V-E0225R1, 5TH MEETING, July 2013 (2013-07-01), Vienna, AT, pages 1 - 8 * |
XIAOZHEN ZHENG ET AL.: "Reference samples sub- sampling for SDC and DMM", JOINT COLLABORATIVE TEAM ON 3D VIDEO CODING EXTENSION DEVELOPMENT OF ITU-T SG 16 WP 3 AND ISO/IEC JTC 1/SC 29/WG 11, JCT3V-C154_R1, 3RD MEETING, January 2013 (2013-01-01), Geneva, CH, pages 1 - 5 * |
Cited By (7)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
CN109937570A (zh) * | 2016-10-14 | 2019-06-25 | 世宗大学校产学协力团 | 影像编码方法/装置、影像解码方法/装置以及保存有比特流的记录介质 |
US11659178B2 (en) | 2016-10-14 | 2023-05-23 | Industry Academy Cooperation Foundation Of Sejong University | Image encoding method/device, image decoding method/device, and recording medium in which bitstream is stored |
CN109937570B (zh) * | 2016-10-14 | 2023-06-06 | 世宗大学校产学协力团 | 影像编码方法/装置、影像解码方法/装置以及保存有比特流的记录介质 |
US11778184B1 (en) | 2016-10-14 | 2023-10-03 | Industry Academy Cooperation Foundation Of Sejong University | Image encoding method/device, image decoding method/device, and recording medium in which bitstream is stored |
US11902521B2 (en) | 2016-10-14 | 2024-02-13 | Industry Academy Cooperation Foundation Of Sejong University | Image encoding method/device, image decoding method/device, and recording medium in which bitstream is stored |
US11909970B2 (en) | 2016-10-14 | 2024-02-20 | Industry Academy Cooperation Foundation Of Sejong University | Image encoding method/device, image decoding method/device, and recording medium in which bitstream is stored |
US11909971B2 (en) | 2016-10-14 | 2024-02-20 | Industry Academy Cooperation Foundation Of Sejong University | Image encoding method/device, image decoding method/device, and recording medium in which bitstream is stored |
Also Published As
Publication number | Publication date |
---|---|
US20170142442A1 (en) | 2017-05-18 |
US10136161B2 (en) | 2018-11-20 |
JPWO2015199040A1 (ja) | 2017-04-20 |
US20190037242A1 (en) | 2019-01-31 |
Similar Documents
Publication | Publication Date | Title |
---|---|---|
WO2015199040A1 (ja) | Dmm予測部、画像復号装置、および画像符号化装置 | |
JP6441418B2 (ja) | 画像復号装置、画像復号方法、画像符号化装置、および画像符号化方法 | |
US10142653B2 (en) | Image decoding apparatus, image coding apparatus, and coded data transformation apparatus | |
US20190075323A1 (en) | Image decoding device, image encoding device, and image decoding method | |
US10225567B2 (en) | Image decoder, image encoder, and encoded data converter | |
WO2017068856A1 (ja) | 予測画像生成装置、画像復号装置および画像符号化装置 | |
WO2014115283A1 (ja) | 画像復号装置、および画像符号化装置 | |
WO2016203881A1 (ja) | 算術復号装置及び算術符号化装置 | |
JP5995448B2 (ja) | 画像復号装置、および画像符号化装置 | |
WO2014162954A1 (ja) | 画像復号装置、および画像符号化装置 | |
WO2015182342A1 (ja) | 画像復号装置、および、画像符号化装置 | |
JPWO2013031575A1 (ja) | 画像処理装置および画像処理方法 | |
WO2019065488A1 (ja) | 画像復号装置および画像符号化装置 | |
WO2014104242A1 (ja) | 画像復号装置、および画像符号化装置 | |
JP2013118424A (ja) | 画像復号装置、画像符号化装置、および符号化データのデータ構造 | |
JP2015073213A (ja) | 画像復号装置、画像符号化装置、符号化データ変換装置、および、注目領域表示システム | |
JP6162289B2 (ja) | 画像復号装置および画像復号方法 | |
JP2016072941A (ja) | Dmm予測装置、画像復号装置、および画像符号化装置 | |
JP2016076904A (ja) | Dcオフセット情報復号装置、画像復号装置、およびdcオフセット情報符号化装置。 | |
JP2016143962A (ja) | 領域分割画像生成装置、画像復号装置、および符号化装置。 | |
JP2016066981A (ja) | Dmm予測部、画像復号装置、および画像符号化装置 | |
WO2014050554A1 (ja) | 画像復号装置、および画像符号化装置 | |
JP2013251827A (ja) | 画像フィルタ装置、画像復号装置、画像符号化装置、およびデータ構造 | |
JP2015076806A (ja) | 画像復号装置および画像符号化装置 | |
JP2014013976A (ja) | 画像復号装置、および画像符号化装置 |
Legal Events
Date | Code | Title | Description |
---|---|---|---|
121 | Ep: the epo has been informed by wipo that ep was designated in this application |
Ref document number: 15811323 Country of ref document: EP Kind code of ref document: A1 |
|
ENP | Entry into the national phase |
Ref document number: 2016529577 Country of ref document: JP Kind code of ref document: A |
|
WWE | Wipo information: entry into national phase |
Ref document number: 15321083 Country of ref document: US |
|
NENP | Non-entry into the national phase |
Ref country code: DE |
|
122 | Ep: pct application non-entry in european phase |
Ref document number: 15811323 Country of ref document: EP Kind code of ref document: A1 |