WO2019131778A1 - Dispositif de décodage d'image et dispositif de codage d'image - Google Patents

Dispositif de décodage d'image et dispositif de codage d'image Download PDF

Info

Publication number
WO2019131778A1
WO2019131778A1 PCT/JP2018/047887 JP2018047887W WO2019131778A1 WO 2019131778 A1 WO2019131778 A1 WO 2019131778A1 JP 2018047887 W JP2018047887 W JP 2018047887W WO 2019131778 A1 WO2019131778 A1 WO 2019131778A1
Authority
WO
WIPO (PCT)
Prior art keywords
tile
view
image
unit
decoding
Prior art date
Application number
PCT/JP2018/047887
Other languages
English (en)
Japanese (ja)
Inventor
知宏 猪飼
友子 青野
知典 橋本
中條 健
将伸 八杉
Original Assignee
シャープ株式会社
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Application filed by シャープ株式会社 filed Critical シャープ株式会社
Publication of WO2019131778A1 publication Critical patent/WO2019131778A1/fr

Links

Images

Classifications

    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04NPICTORIAL COMMUNICATION, e.g. TELEVISION
    • H04N19/00Methods or arrangements for coding, decoding, compressing or decompressing digital video signals
    • H04N19/50Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using predictive coding
    • H04N19/597Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using predictive coding specially adapted for multi-view video sequence encoding
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04NPICTORIAL COMMUNICATION, e.g. TELEVISION
    • H04N19/00Methods or arrangements for coding, decoding, compressing or decompressing digital video signals
    • H04N19/70Methods or arrangements for coding, decoding, compressing or decompressing digital video signals characterised by syntax aspects related to video coding, e.g. related to compression standards

Definitions

  • One aspect of the present invention relates to an image decoding device and an image coding device.
  • a moving picture coding apparatus that generates coded data by coding a moving picture to efficiently transmit or record a moving picture, and a moving picture that generates a decoded picture by decoding the coded data.
  • An image decoding device is used.
  • HEVC High-Efficiency Video Coding
  • an image (picture) constituting a moving picture is a slice obtained by dividing the image, a coding tree unit obtained by dividing the slice (CTU: Coding Tree Unit)
  • a coding unit obtained by dividing a coding tree unit (sometimes called a coding unit (CU))
  • a prediction unit which is a block obtained by dividing a coding unit It is managed by the hierarchical structure which consists of (PU) and a transform unit (TU), and is encoded / decoded per CU.
  • a predicted picture is usually generated based on a locally decoded picture obtained by coding / decoding an input picture, and the predicted picture is generated from the input picture (original picture).
  • the prediction residual obtained by subtraction (sometimes referred to as "difference image” or "residual image") is encoded.
  • inter prediction inter prediction
  • intra-screen prediction intra prediction
  • Non-Patent Document 1 can be cited as a technology for moving picture encoding and decoding in recent years.
  • Non-Patent Document 2 defines a technique called frame packing in which two view images are integrated into one picture. With frame packing, it is possible to encode stereo images without using scalable coding. Further, Non-Patent Document 3 discloses a technique for encoding a stereo image using intra block copying.
  • One aspect of the present invention has been made in view of the above problems, and can reduce the amount of processing and transmission data when decoding and transmitting a specific view image from a picture in which a plurality of view images are frame-packed.
  • An object of the present invention is to realize an image decoding apparatus.
  • Another object of the present invention is to realize an image coding apparatus that creates a coded stream capable of extracting a specific view image with a small amount of processing.
  • One aspect of the present invention is made in view of the above problems, and another object is to be able to efficiently decode a view image when decoding a view image included in a picture subjected to frame packing. It is to provide technology.
  • an image decoding apparatus is an image decoding apparatus that decodes a coded stream of an image, and includes a header information decoding unit and a picture decoding unit, and The header information decoding unit decodes view packing information which is view packing information included in the parameter set of the encoded stream and which indicates whether a target picture includes a plurality of views, and the picture decoding unit Each view is decoded by decoding a picture identified by the view packing information as a plurality of views being frame-packed in the vertical direction.
  • another image decoding apparatus is an image decoding apparatus that decodes a tile forming a view, and a target block of a first tile of the tiles.
  • the inter-tile reference information decoding unit decodes the inter-tile reference information indicating whether or not the second tile of the tiles can be referred to when predicting If it indicates that it is possible to refer to the tile, it predicts the tile decoding order information or the tile position information by generating a predicted image of the target block with reference to the reference picture including the second tile.
  • a header information decoding unit is an image decoding apparatus that decodes a tile forming a view, and a target block of a first tile of the tiles.
  • an image decoding apparatus is an image decoding apparatus that decodes a plurality of tiles forming a view, and is a first tile of the plurality of tiles.
  • An inter-tile reference information decoding unit that decodes inter-tile reference information indicating whether or not the second tile of the plurality of tiles can be referred to when predicting a target block, and the inter-tile reference information
  • a determination unit that determines whether to add the current picture to the reference picture list, and the determination unit determines that the current picture is to be added to the reference picture list, the current picture is determined to be the reference picture And an additional unit to be added to the list.
  • an image coding apparatus is an image coding apparatus that codes tiles that constitute a view, and is an object of a first tile of the tiles.
  • the inter-tile reference information encoding unit encodes the inter-tile reference information indicating whether or not the second tile of the tiles can be referred to when predicting a block, and the inter-tile reference information
  • a predicted image generation unit configured to generate a predicted image of the target block with reference to the second tile when it is indicated that the second tile can be referred to;
  • the prediction image of each target block of the tile is generated with reference to only the reference block included in the tile indicating that the inter-tile reference information can be referred to. .
  • an image coding apparatus is an image coding apparatus that codes tiles that constitute a view, and is an object of a first tile of the tiles.
  • the inter-tile reference information encoding unit encodes the inter-tile reference information indicating whether or not the second tile of the tiles can be referred to when predicting a block, and the inter-tile reference information
  • a predicted image generation unit configured to generate a predicted image of the target block with reference to the second tile when it is indicated that the second tile can be referred to;
  • a predicted image of each target block of the tile is generated in the coding order notified by the forward index.
  • the image decoding apparatus when extracting a specific view image from a picture in which a plurality of view images are frame-packed, the image decoding apparatus is a continuous range of a coded stream Since it is sufficient to decode the range corresponding to the end, the processing amount can be reduced.
  • Another image decoding apparatus can efficiently decode a view image when decoding a view image included in a frame-packed picture.
  • FIG. 1 It is a figure which shows the hierarchical structure of the data of the coding stream which concerns on this embodiment. It is a figure which shows the pattern of PU split mode. (A) to (h) show the partition shapes when the PU division mode is 2Nx2N, 2NxN, 2NxnU, 2NxnD, Nx2N, nLx2N, nRx2N, and NxN, respectively. It is a conceptual diagram which shows an example of a reference picture and a reference picture list. It is a block diagram which shows the structure of the image coding apparatus which concerns on this embodiment. It is the schematic which shows the structure of the image decoding apparatus which concerns on this embodiment.
  • (B) is a figure for demonstrating the image decoding method by the image decoding apparatus 31 which concerns on the specific example of this embodiment.
  • (A) is a figure which shows the structure of the coding data which concern on the specific example of this embodiment.
  • (B) is a figure for demonstrating the image decoding method by the image decoding apparatus 31 which concerns on the specific example of this embodiment.
  • It is a syntax table which shows the example of the production
  • generating a predicted image of a view image it is a figure showing an example of a view image which can be referred to. It is a figure which shows the example which clips the vector which points to the block referred in a prediction image generation process. It is a figure which shows an example of a padding process. It is a figure which shows an example of the block which can be referred to by the predicted image generation unit. It is a figure which shows an example of the block which can be referred to by the predicted image generation unit. It is a figure which shows an example of the block which can be referred to by the predicted image generation unit. It is a figure which shows an example of the block which can be referred to by the predicted image generation unit.
  • FIG. 1 It is a figure which shows an example of the clipping process of a motion vector, and a padding process. It is the figure shown about the composition of the transmitting device carrying the picture coding device concerning this embodiment, and the receiving device carrying a picture decoding device.
  • (A) shows a transmitting apparatus equipped with an image coding apparatus, and (b) shows a receiving apparatus equipped with an image decoding apparatus. It is the figure shown about the recording device carrying the picture coding device concerning this embodiment, and the composition of the reproduction device carrying a picture decoding device.
  • (A) shows a recording apparatus equipped with an image coding apparatus, and (b) shows a reproduction apparatus equipped with an image decoding apparatus. It is a figure which shows the example which restrict
  • FIG. 1 It is a figure which shows the example of the pixel reference process by the estimated image generation part which concerns on the specific example of this embodiment.
  • FIG. 1 (A) to (c) are diagrams showing more specific examples of padding processing by the predicted image generation unit according to the specific example of the present embodiment.
  • (A) to (c) are diagrams showing more specific examples of padding processing by the predicted image generation unit according to the specific example of the present embodiment.
  • (A) And (b) is a figure which shows the more specific example of the padding process by the estimated image generation part which concerns on the specific example of this embodiment, respectively. It is a flowchart figure explaining generation of a reference picture list concerning this embodiment. It is a schematic diagram showing composition of an image transmission system concerning this embodiment.
  • FIG. 53 is a schematic view showing the configuration of the image transmission system 1 according to the present embodiment.
  • the image transmission system 1 is a system that transmits a code obtained by coding an image to be coded, decodes the transmitted code, and displays the image.
  • the image transmission system 1 is configured to include an image encoding device (moving image encoding device) 11, a network 21, an image decoding device (moving image decoding device) 31, and an image display device 41.
  • the image coding apparatus 11 receives an image T of an image of a single layer or multiple layers.
  • a layer is a concept used to distinguish a plurality of pictures when there is one or more pictures that constitute a certain time. For example, if the same picture is encoded by a plurality of layers having different image quality and resolution, it becomes scalable coding, and if a picture of different viewpoints is encoded by a plurality of layers, it becomes view scalable coding.
  • prediction inter-layer prediction, inter-view prediction
  • encoded data can be summarized.
  • the network 21 transmits the encoded stream Te generated by the image encoding device 11 to the image decoding device 31.
  • the network 21 is the Internet, a wide area network (WAN), a small area network (LAN), or a combination of these.
  • the network 21 is not necessarily limited to a two-way communication network, and may be a one-way communication network for transmitting broadcast waves such as terrestrial digital broadcasting and satellite broadcasting.
  • the network 21 may be replaced by a storage medium recording a coded stream Te such as a DVD (Digital Versatile Disc) or a BD (Blue-ray Disc).
  • the image decoding apparatus 31 decodes each of the encoded streams Te transmitted by the network 21 and generates one or more decoded images Td which are respectively decoded.
  • the image display device 41 displays all or a part of one or more decoded images Td generated by the image decoding device 31.
  • the image display device 41 includes, for example, a display device such as a liquid crystal display or an organic EL (Electro-luminescence) display.
  • a display device such as a liquid crystal display or an organic EL (Electro-luminescence) display.
  • a display device such as a liquid crystal display or an organic EL (Electro-luminescence) display.
  • SNR scalable coding when the image decoding device 31 and the image display device 41 have high processing capabilities, they display enhancement layer images with high image quality and have only lower processing capabilities.
  • the base layer image which does not require the processing capability and the display capability as high as the enhancement layer.
  • X? Y: z is a ternary operator that takes y if x is true (other than 0) and z if x is false (0).
  • FIG. 1 is a diagram showing a hierarchical structure of data in a coded stream Te.
  • the coded stream Te illustratively includes a sequence and a plurality of pictures forming the sequence.
  • (A) to (f) in FIG. 1 respectively represent a coded video sequence defining the sequence SEQ, a coded picture defining the picture PICT, a coding slice defining the slice S, and a coding slice defining slice data.
  • It is a figure which shows a coding tree unit contained in data, coding slice data, and a coding unit (Coding Unit; CU) contained in a coding tree unit.
  • CU coding unit
  • the encoded video sequence In the encoded video sequence, a set of data to which the image decoding device 31 refers in order to decode the sequence SEQ to be processed is defined.
  • the sequence SEQ includes a video parameter set (Video Parameter Set), a sequence parameter set SPS (Sequence Parameter Set), a picture parameter set PPS (Picture Parameter Set), a picture PICT, and an addition. It includes supplemental information SEI (Supplemental Enhancement Information).
  • SEI Supplemental Enhancement Information
  • the value shown after # indicates a layer ID.
  • FIG. 1 shows an example in which coded data of # 0 and # 1, that is, layer 0 and layer 1 exist, the type of layer and the number of layers do not depend on this.
  • a video parameter set VPS is a set of coding parameters common to a plurality of moving pictures and a set of coding parameters related to the plurality of layers included in the moving picture and each layer in a moving picture composed of a plurality of layers.
  • a set is defined.
  • sequence parameter set SPS a set of coding parameters to be referred to by the image decoding device 31 for decoding the target sequence is defined.
  • the width and height of the picture are defined.
  • multiple SPS may exist. In that case, one of a plurality of SPSs is selected from PPS.
  • a set of coding parameters to which the image decoding device 31 refers to to decode each picture in the target sequence is defined. For example, a reference value of quantization width (pic_init_qp_minus 26) used for decoding a picture and a flag (weighted_pred_flag) indicating application of weighted prediction are included.
  • multiple PPS may exist. In that case, one of a plurality of PPSs is selected from each picture in the target sequence.
  • the picture PICT includes slices S0 to SNS-1 (NS is the total number of slices included in the picture PICT), as shown in (b) of FIG.
  • the slice S includes a slice header SH and slice data SDATA as shown in (c) of FIG.
  • the slice header SH includes a coding parameter group to which the image decoding device 31 refers in order to determine the decoding method of the target slice.
  • the slice type specification information (slice_type) for specifying a slice type is an example of a coding parameter included in the slice header SH.
  • slice types that can be designated by slice type designation information, (1) I slice using only intra prediction at the time of encoding, (2) P slice using unidirectional prediction at the time of encoding or intra prediction, (3) B-slice using uni-directional prediction, bi-directional prediction, or intra prediction at the time of encoding.
  • the slice header SH may include a reference (pic_parameter_set_id) to the picture parameter set PPS included in the encoded video sequence.
  • the slice data SDATA includes a coding tree unit (CTU: Coding Tree Unit), as shown in (d) of FIG.
  • the CTU is a block of a fixed size (for example, 64 ⁇ 64) that configures a slice, and may also be referred to as a largest coding unit (LCU: Largest Coding Unit).
  • Encoding tree unit As shown in (e) of FIG. 1, a set of data to which the image decoding device 31 refers in order to decode a coding tree unit to be processed is defined.
  • the coding tree unit is divided by recursive quadtree division.
  • a tree-structured node obtained by recursive quadtree division is called a coding node (CN).
  • the intermediate nodes of the quadtree are coding nodes, and the coding tree unit itself is also defined as the top coding node.
  • the CTU includes a split flag (cu_split_flag), and when cu_split_flag is 1, the CTU is split into four coding nodes CN.
  • the coding node CN is not split, and has one coding unit (CU: Coding Unit) as a node.
  • the coding unit CU is an end node of the coding node and is not further divided.
  • the coding unit CU is a basic unit of coding processing.
  • the size of the coding unit can be 64x64 pixels, 32x32 pixels, 16x16 pixels, or 8x8 pixels.
  • a set of data to which the image decoding device 31 refers in order to decode a coding unit to be processed is defined.
  • the coding unit is composed of a prediction tree, a transformation tree, and a CU header CUH.
  • a prediction mode, a division method (PU division mode), and the like are defined.
  • prediction information (reference picture index, motion vector, etc.) of each prediction unit (PU) obtained by dividing the coding unit into one or more is defined.
  • a prediction unit is one or more non-overlapping regions that make up a coding unit.
  • the prediction tree includes one or more prediction units obtained by the above-mentioned division.
  • segmented the prediction unit further is called a "subblock.”
  • the sub block is composed of a plurality of pixels. If the size of the prediction unit and the subblock is equal, there is one subblock in the prediction unit. If the prediction unit is larger than the size of the subblock, the prediction unit is divided into subblocks. For example, when the prediction unit is 8x8 and the subblock is 4x4, the prediction unit is divided into four subblocks, which are horizontally divided into two and vertically divided into two.
  • the prediction process may be performed for each prediction unit (sub block).
  • Intra prediction is prediction in the same picture
  • inter prediction refers to prediction processing performed between mutually different pictures (for example, between display times, between layer images).
  • the division method is encoded according to PU division mode (part_mode) of encoded data, 2Nx2N (the same size as the encoding unit), 2NxN, 2NxnU, 2NxnD, Nx2N, nLx2N, nRx2N, and There are NxN etc.
  • 2NxN and Nx2N indicate 1: 1 symmetric division
  • 2NxnU, 2NxnD and nLx2N and nRx2N indicate 1: 3 and 3: 1 asymmetric division.
  • the PUs included in the CU are expressed as PU0, PU1, PU2, PU3 in order.
  • FIG. 2 specifically illustrate the shapes of partitions (positions of boundaries of PU division) in respective PU division modes.
  • A) of FIG. 2 shows a 2Nx2N partition
  • (b) and (c) and (d) show 2NxN, 2NxnU, and 2NxnD partitions (horizontally long partitions), respectively.
  • (E), (f) and (g) show partitions (vertical partitions) in the case of Nx2N, nLx2N and nRx2N, respectively
  • (h) shows a partition of NxN. Note that the horizontally long partition and the vertically long partition are collectively referred to as a rectangular partition, and 2Nx2N and NxN are collectively referred to as a square partition.
  • the coding unit is divided into one or more transform units, and the position and size of each transform unit are defined.
  • a transform unit is one or more non-overlapping regions that make up a coding unit.
  • the transformation tree includes one or more transformation units obtained by the above-mentioned division.
  • Partitions in the transform tree may be allocated as a transform unit a region of the same size as the encoding unit, or may be based on recursive quadtree partitioning as in the case of CU partitioning described above.
  • a conversion process is performed for each conversion unit.
  • the prediction image of a prediction unit is derived by prediction parameters associated with PU.
  • the prediction parameters include intra prediction prediction parameters or inter prediction prediction parameters.
  • prediction parameters for inter prediction inter prediction (inter prediction parameters) will be described.
  • the inter prediction parameter includes prediction list use flags predFlagL0 and predFlagL1, reference picture indexes refIdxL0 and refIdxL1, and motion vectors mvL0 and mvL1.
  • the prediction list use flags predFlagL0 and predFlagL1 are flags indicating whether a reference picture list called an L0 list or an L1 list is used, respectively, and a reference picture list corresponding to a value of 1 is used.
  • a flag indicating whether or not it is XX if the flag is other than 0 (for example, 1) is XX, it is assumed that 0 is not XX; Treat 1 as true, 0 as false, and so on. However, in an actual apparatus or method, other values may be used as true values or false values.
  • Syntax elements for deriving inter prediction parameters included in encoded data include, for example, PU split mode part_mode, merge flag merge_flag, merge index merge_idx, inter prediction identifier inter_pred_idc, reference picture index refIdxLX, predicted vector index mvp_LX_idx, There is a difference vector mvdLX.
  • the reference picture list is a list of reference pictures stored in the reference picture memory 306.
  • FIG. 3 is a conceptual diagram showing an example of a reference picture and a reference picture list.
  • the rectangle is a picture
  • the arrow is a reference of the picture
  • the horizontal axis is time
  • I, P and B in the rectangle are intra pictures, uni-predicted pictures, bi-predicted pictures, and numbers in the rectangle are decoded. Show the order.
  • the decoding order of pictures is I0, P1, B2, B3, B4, and the display order is I0, B3, B2, B4, B1, P1.
  • FIG. 3B shows an example of the reference picture list.
  • the reference picture list is a list representing reference picture candidates, and one picture (slice) may have one or more reference picture lists.
  • the target picture B3 has two reference picture lists, an L0 list RefPicList0 and an L1 list RefPicList1.
  • Reference pictures when the target picture is B3 are I0, P1, and B2, and the reference pictures have these pictures as elements.
  • the reference picture index refIdxLX which picture in the reference picture list RefPicListX is actually referred to is designated by the reference picture index refIdxLX.
  • the figure shows an example in which reference pictures P1 and B2 are referenced by refIdxL0 and refIdxL1.
  • the prediction parameter decoding (encoding) method includes a merge prediction (merge) mode and an AMVP (Adaptive Motion Vector Prediction) mode.
  • the merge flag merge_flag is a flag for identifying these.
  • the merge prediction mode is a mode used to be derived from the prediction parameter of the already processed neighboring PU without including the prediction list use flag predFlagLX (or inter prediction identifier inter_pred_idc), the reference picture index refIdxLX, and the motion vector mvLX in the encoded data.
  • the AMVP mode is a mode in which the inter prediction identifier inter_pred_idc, the reference picture index refIdxLX, and the motion vector mvLX are included in the encoded data.
  • the motion vector mvLX is encoded as a prediction vector index mvp_LX_idx that identifies the prediction vector mvpLX and a difference vector mvdLX.
  • the inter prediction identifier inter_pred_idc is a value indicating the type and the number of reference pictures, and takes any one of PRED_L0, PRED_L1, and PRED_BI.
  • PRED_L0 and PRED_L1 indicate that reference pictures managed by reference pictures in the L0 list and the L1 list are used, respectively, and indicate that one reference picture is used (uniprediction).
  • PRED_BI indicates using two reference pictures (bi-prediction BiPred), and uses reference pictures managed by the L0 list and the L1 list.
  • the predicted vector index mvp_LX_idx is an index indicating a predicted vector
  • the reference picture index refIdxLX is an index indicating a reference picture managed in the reference picture list.
  • LX is a description method used when L0 prediction and L1 prediction are not distinguished, and parameters for L0 list and parameters for L1 list are distinguished by replacing LX with L0 and L1.
  • Merge index merge_idx is an index which shows whether any prediction parameter is used as a prediction parameter of decoding object PU among the prediction parameter candidates (merge candidate) derived
  • the motion vector mvLX indicates the amount of deviation between blocks on two different pictures.
  • the prediction vector and the difference vector relating to the motion vector mvLX are referred to as a prediction vector mvpLX and a difference vector mvdLX, respectively.
  • FIG. 5 is a schematic view showing the configuration of the image decoding device 31 according to the present embodiment.
  • the image decoding device 31 includes an entropy decoding unit 301, a prediction parameter decoding unit (prediction image decoding device) 302, a loop filter 305, a reference picture memory 306, a prediction parameter memory 307, a prediction image generation unit (prediction image generation device) 308, and an inverse
  • the quantization / inverse transform unit 311 and the addition unit 312 are included.
  • the prediction parameter decoding unit 302 is configured to include an inter prediction parameter decoding unit 303 and an intra prediction parameter decoding unit 304.
  • the predicted image generation unit 308 includes an inter predicted image generation unit 309 and an intra predicted image generation unit 310.
  • the entropy decoding unit 301 performs entropy decoding on the encoded stream Te input from the outside to separate and decode individual codes (syntax elements).
  • the separated codes include prediction information for generating a prediction image and residual information for generating a difference image.
  • the entropy decoding unit 301 outputs a part of the separated code to the prediction parameter decoding unit 302.
  • the part of the separated code is, for example, prediction mode predMode, PU division mode part_mode, merge flag merge_flag, merge index merge_idx, inter prediction identifier inter_pred_idc, reference picture index refIdxLX, prediction vector index mvp_LX_idx, difference vector mvdLX. Control of which code to decode is performed based on an instruction of the prediction parameter decoding unit 302.
  • the entropy decoding unit 301 outputs the quantization coefficient to the inverse quantization / inverse transform unit 311.
  • this quantization coefficient is applied to the residual signal by DCT (Discrete Cosine Transform, discrete cosine transform), DST (Discrete Sine Transform, discrete sine transform), KLT (Karyhnen Loeve Transform, Karhunen Loeve transform) Are coefficients obtained by performing frequency conversion such as.
  • DCT Discrete Cosine Transform, discrete cosine transform
  • DST Discrete Sine Transform, discrete sine transform
  • KLT Karyhnen Loeve Transform, Karhunen Loeve transform
  • the inter prediction parameter decoding unit 303 decodes the inter prediction parameter with reference to the prediction parameter stored in the prediction parameter memory 307 based on the code input from the entropy decoding unit 301.
  • the inter prediction parameter decoding unit 303 outputs the decoded inter prediction parameter to the prediction image generation unit 308, and stores the inter prediction parameter in the prediction parameter memory 307. Details of the inter prediction parameter decoding unit 303 will be described later.
  • the intra prediction parameter decoding unit 304 decodes the intra prediction parameter with reference to the prediction parameter stored in the prediction parameter memory 307 based on the code input from the entropy decoding unit 301.
  • the intra prediction parameter is a parameter used in a process of predicting a CU in one picture, for example, an intra prediction mode IntraPredMode.
  • the intra prediction parameter decoding unit 304 outputs the decoded intra prediction parameter to the prediction image generation unit 308, and stores it in the prediction parameter memory 307.
  • the intra prediction parameter decoding unit 304 may derive different intra prediction modes for luminance and chrominance.
  • the intra prediction parameter decoding unit 304 decodes a luminance prediction mode IntraPredModeY as a luminance prediction parameter and a chrominance prediction mode IntraPredModeC as a chrominance prediction parameter.
  • the luminance prediction mode IntraPredModeY is a 35 mode, which corresponds to planar prediction (0), DC prediction (1), and directional prediction (2 to 34).
  • the color difference prediction mode IntraPredModeC uses one of planar prediction (0), DC prediction (1), direction prediction (2 to 34), and LM mode (35).
  • the intra prediction parameter decoding unit 304 decodes a flag indicating whether IntraPredModeC is the same mode as the luminance mode, and if it indicates that the flag is the same mode as the luminance mode, IntraPredModeY is assigned to IntraPredModeC, and the flag indicates the luminance If intra mode is different from the mode, planar prediction (0), DC prediction (1), direction prediction (2 to 34), or LM mode (35) may be decoded as IntraPredModeC.
  • the loop filter 305 applies a filter such as a deblocking filter, a sample adaptive offset (SAO), or an adaptive loop filter (ALF) to the decoded image of the CU generated by the adding unit 312.
  • a filter such as a deblocking filter, a sample adaptive offset (SAO), or an adaptive loop filter (ALF)
  • the reference picture memory 306 stores the decoded image of the CU generated by the adding unit 312 in a predetermined position for each picture and CU to be decoded.
  • the prediction parameter memory 307 stores prediction parameters in a predetermined position for each picture to be decoded and each prediction unit (or sub block, fixed size block, pixel). Specifically, the prediction parameter memory 307 stores the inter prediction parameter decoded by the inter prediction parameter decoding unit 303, the intra prediction parameter decoded by the intra prediction parameter decoding unit 304, and the prediction mode predMode separated by the entropy decoding unit 301. .
  • the inter prediction parameters to be stored include, for example, a prediction list use flag predFlagLX (inter prediction identifier inter_pred_idc), a reference picture index refIdxLX, and a motion vector mvLX.
  • the prediction image generation unit 308 receives the prediction mode predMode input from the entropy decoding unit 301, and also receives a prediction parameter from the prediction parameter decoding unit 302. Further, the predicted image generation unit 308 reads the reference picture from the reference picture memory 306. The prediction image generation unit 308 generates a prediction image of a PU or a sub block using the input prediction parameter and the read reference picture (reference picture block) in the prediction mode indicated by the prediction mode predMode.
  • the inter prediction image generation unit 309 performs inter prediction using the inter prediction parameter input from the inter prediction parameter decoding unit 303 and the read reference picture (reference picture block). Generates a predicted image of PU or subblock according to.
  • the inter-predicted image generation unit 309 uses the reference picture index refIdxLX for the reference picture list (L0 list or L1 list) in which the prediction list use flag predFlagLX is 1, and the motion vector based on the PU to be decoded
  • the reference picture block at the position indicated by mvLX is read out from the reference picture memory 306.
  • the inter-prediction image generation unit 309 performs prediction based on the read reference picture block to generate a PU prediction image.
  • the inter prediction image generation unit 309 outputs the generated prediction image of PU to the addition unit 312.
  • the reference picture block is a set of pixels on the reference picture (usually referred to as a block because it is a rectangle), and is an area to be referenced to generate a predicted image of PU or sub block.
  • the intra prediction image generation unit 310 When the prediction mode predMode indicates the intra prediction mode, the intra prediction image generation unit 310 performs intra prediction using the intra prediction parameter input from the intra prediction parameter decoding unit 304 and the read reference picture. Specifically, the intra predicted image generation unit 310 reads, from the reference picture memory 306, neighboring PUs which are pictures to be decoded and which are in a predetermined range from the PU to be decoded among PUs already decoded.
  • the predetermined range is, for example, one of the left, upper left, upper, and upper right adjacent PUs when the decoding target PU sequentially moves in the so-called raster scan order, and varies depending on the intra prediction mode.
  • the order of raster scan is an order of sequentially moving from the left end to the right end for each row from the top to the bottom in each picture.
  • the intra prediction image generation unit 310 performs prediction in the prediction mode indicated by the intra prediction mode IntraPredMode based on the read adjacent PU, and generates a PU prediction image.
  • the intra predicted image generation unit 310 outputs the generated predicted image of PU to the addition unit 312.
  • the inverse quantization / inverse transform unit 311 inversely quantizes the quantization coefficient input from the entropy decoding unit 301 to obtain a transform coefficient.
  • the inverse quantization / inverse transform unit 311 performs inverse frequency transform such as inverse DCT, inverse DST, and inverse KLT on the obtained transform coefficient to calculate a residual signal.
  • the inverse quantization / inverse transform unit 311 outputs the calculated residual signal to the addition unit 312.
  • the addition unit 312 adds, for each pixel, the PU prediction image input from the inter prediction image generation unit 309 or the intra prediction image generation unit 310 and the residual signal input from the inverse quantization / inverse conversion unit 311, Generate a PU decoded image.
  • the addition unit 312 stores the generated PU decoded image in the reference picture memory 306, and externally outputs a decoded image Td in which the generated PU decoded image is integrated for each picture.
  • FIG. 12 is a schematic diagram showing a configuration of the inter prediction parameter decoding unit 303 according to the present embodiment.
  • the inter prediction parameter decoding unit 303 includes an inter prediction parameter decoding control unit 3031, an AMVP prediction parameter derivation unit 3032, an addition unit 3038, a merge prediction parameter derivation unit 3036, and a sub block prediction parameter derivation unit 3037.
  • the inter prediction parameter decoding control unit 3031 instructs the entropy decoding unit 301 to decode a code (syntax element) related to inter prediction, and a code (syntax element) included in the encoded data, for example, PU division mode part_mode , Merge flag merge_flag, merge index merge_idx, inter prediction identifier inter_pred_idc, reference picture index refIdxLX, prediction vector index mvp_LX_idx, difference vector mvdLX is extracted.
  • the inter prediction parameter decoding control unit 3031 first extracts the merge flag merge_flag. When the inter prediction parameter decoding control unit 3031 expresses that a syntax element is to be extracted, it instructs the entropy decoding unit 301 to decode a syntax element, which means that the corresponding syntax element is read out from the encoded data. Do.
  • the inter prediction parameter decoding control unit 3031 extracts an AMVP prediction parameter from the encoded data using the entropy decoding unit 301.
  • the AMVP prediction parameters for example, there are inter prediction identifier inter_pred_idc, reference picture index refIdxLX, prediction vector index mvp_LX_idx, difference vector mvdLX.
  • the AMVP prediction parameter derivation unit 3032 derives a prediction vector mvpLX from the prediction vector index mvp_LX_idx. Details will be described later.
  • the inter prediction parameter decoding control unit 3031 outputs the difference vector mvdLX to the addition unit 3038.
  • the addition unit 3038 adds the prediction vector mvpLX and the difference vector mvdLX to derive a motion vector.
  • the inter prediction parameter decoding control unit 3031 extracts a merge index merge_idx as a prediction parameter related to merge prediction.
  • the inter prediction parameter decoding control unit 3031 outputs the extracted merge index merge_idx to the merge prediction parameter derivation unit 3036 (details will be described later), and outputs the sub block prediction mode flag subPbMotionFlag to the sub block prediction parameter derivation unit 3037.
  • the sub-block prediction parameter derivation unit 3037 divides the PU into a plurality of sub-blocks according to the value of the sub-block prediction mode flag subPbMotionFlag, and derives a motion vector in units of sub-blocks.
  • a prediction block is predicted in small blocks of 4x4 or 8x8.
  • a method of dividing a CU into a plurality of partitions PUs such as 2NxN, Nx2N, NxN, etc.
  • PUs such as 2NxN, Nx2N, NxN, etc.
  • sub block prediction mode a plurality of sub-blocks are grouped into a set, and the syntax of the prediction parameter is encoded for each set, so that motion information of many sub-blocks can be encoded with a small code amount.
  • FIG. 7 is a schematic view showing the configuration of the merge prediction parameter derivation unit 3036 according to the present embodiment.
  • the merge prediction parameter derivation unit 3036 includes a merge candidate derivation unit 30361, a merge candidate selection unit 30362, and a merge candidate storage unit 30363.
  • the merge candidate storage unit 30363 stores the merge candidate input from the merge candidate derivation unit 30361.
  • the merge candidate is configured to include a prediction list use flag predFlagLX, a motion vector mvLX, and a reference picture index refIdxLX.
  • an index is assigned to the stored merge candidate according to a predetermined rule.
  • the merge candidate derivation unit 30361 derives merge candidates using the motion vector of the adjacent PU for which the decoding process has already been performed and the reference picture index refIdxLX as it is.
  • the merge candidate derived by the merge candidate derivation unit 30361 is stored in the merge candidate storage unit 30363.
  • the merge candidate selection unit 30362 selects, as the target PU, the merge candidate to which the index corresponding to the merge index merge_idx input from the inter prediction parameter decoding control unit 3031 is assigned. Select as inter prediction parameter of.
  • the merge candidate selection unit 30362 stores the selected merge candidate in the prediction parameter memory 307 and outputs the merge candidate to the predicted image generation unit 308.
  • FIG. 8 is a schematic diagram showing the configuration of the AMVP prediction parameter derivation unit 3032 according to the present embodiment.
  • the AMVP prediction parameter derivation unit 3032 includes a vector candidate derivation unit 3033, a vector candidate selection unit 3034, and a vector candidate storage unit 3035.
  • the vector candidate derivation unit 3033 derives a prediction vector candidate from the already processed PU motion vector mvLX stored in the prediction parameter memory 307 based on the reference picture index refIdx, and the vector candidate storage unit 3035 prediction vector candidate list mvpListLX [] Store in
  • the vector candidate selection unit 3034 selects a motion vector mvpListLX [mvp_LX_idx] indicated by the prediction vector index mvp_LX_idx among the prediction vector candidates of the prediction vector candidate list mvpListLX [] as a prediction vector mvpLX.
  • the vector candidate selection unit 3034 outputs the selected prediction vector mvpLX to the addition unit 3038.
  • the prediction vector candidate is a PU for which decoding processing has been completed, and is derived by scaling a motion vector of a PU (for example, an adjacent PU) in a predetermined range from the PU to be decoded.
  • the adjacent PU includes a PU spatially adjacent to the PU to be decoded, for example, a left PU, an upper PU, and an area temporally adjacent to the PU to be decoded, for example, the same position as the PU to be decoded It includes regions obtained from prediction parameters of PUs of different times.
  • the addition unit 3038 adds the prediction vector mvpLX input from the AMVP prediction parameter derivation unit 3032 and the difference vector mvdLX input from the inter prediction parameter decoding control unit 3031 to calculate a motion vector mvLX.
  • the addition unit 3038 outputs the calculated motion vector mvLX to the predicted image generation unit 308 and the prediction parameter memory 307.
  • FIG. 4 is a block diagram showing the configuration of the image coding apparatus 11 according to the present embodiment.
  • the image coding apparatus 11 includes a predicted image generation unit 101, a subtraction unit 102, a transform / quantization unit 103, an entropy coding unit 104, an inverse quantization / inverse transform unit 105, an addition unit 106, a loop filter 107, a prediction parameter memory (Prediction parameter storage unit, frame memory) 108, reference picture memory (reference image storage unit, frame memory) 109, coding parameter determination unit 110, and prediction parameter coding unit 111 are configured.
  • the prediction parameter coding unit 111 includes an inter prediction parameter coding unit 112 and an intra prediction parameter coding unit 113.
  • the prediction image generation unit 101 generates, for each picture of the image T, the prediction image P of the prediction unit PU for each coding unit CU, which is an area obtained by dividing the picture.
  • the predicted image generation unit 101 reads a decoded block from the reference picture memory 109 based on the prediction parameter input from the prediction parameter coding unit 111.
  • the prediction parameter input from the prediction parameter coding unit 111 is, for example, a motion vector in the case of inter prediction.
  • the predicted image generation unit 101 reads a block at a position on the reference image indicated by the motion vector starting from the target PU.
  • the prediction parameter is, for example, an intra prediction mode.
  • the pixel value of the adjacent PU used in the intra prediction mode is read from the reference picture memory 109, and a PU predicted image P is generated.
  • the prediction image generation unit 101 generates a PU prediction image P using one of a plurality of prediction methods for the read reference picture block.
  • the prediction image generation unit 101 outputs the generated prediction image P of PU to the subtraction unit 102.
  • the predicted image generation unit 101 performs the same operation as the predicted image generation unit 308 described above.
  • the prediction image generation unit 101 generates a PU prediction image P based on the pixel value of the reference block read from the reference picture memory, using the parameter input from the prediction parameter coding unit.
  • the predicted image generated by the predicted image generation unit 101 is output to the subtraction unit 102 and the addition unit 106.
  • the subtraction unit 102 subtracts the signal value of the predicted image P of the PU input from the predicted image generation unit 101 from the pixel value of the corresponding PU of the image T to generate a residual signal.
  • the subtraction unit 102 outputs the generated residual signal to the transformation / quantization unit 103.
  • the transform / quantization unit 103 performs frequency transform on the residual signal input from the subtraction unit 102 to calculate transform coefficients.
  • the transform / quantization unit 103 quantizes the calculated transform coefficient to obtain a quantization coefficient.
  • Transform / quantization section 103 outputs the obtained quantization coefficient to entropy coding section 104 and inverse quantization / inverse transform section 105.
  • the entropy coding unit 104 receives the quantization coefficient from the transform / quantization unit 103, and receives the coding parameter from the prediction parameter coding unit 111.
  • the coding parameters to be input include, for example, codes such as a reference picture index refIdxLX, a prediction vector index mvp_LX_idx, a difference vector mvdLX, a prediction mode predMode, and a merge index merge_idx.
  • the entropy coding unit 104 entropy-codes the input quantization coefficient and coding parameters to generate a coded stream Te, and outputs the generated coded stream Te to the outside.
  • the inverse quantization / inverse transform unit 105 inversely quantizes the quantization coefficient input from the transform / quantization unit 103 to obtain a transform coefficient.
  • the inverse quantization / inverse transform unit 105 performs inverse frequency transform on the obtained transform coefficient to calculate a residual signal.
  • the inverse quantization / inverse transform unit 105 outputs the calculated residual signal to the addition unit 106.
  • the addition unit 106 adds the signal value of the prediction image P of PU input from the prediction image generation unit 101 and the signal value of the residual signal input from the inverse quantization / inverse conversion unit 105 for each pixel, and decodes Generate an image.
  • the addition unit 106 stores the generated decoded image in the reference picture memory 109.
  • the loop filter 107 applies a deblocking filter, a sample adaptive offset (SAO), and an adaptive loop filter (ALF) to the decoded image generated by the adding unit 106.
  • a deblocking filter a sample adaptive offset (SAO)
  • ALF adaptive loop filter
  • the prediction parameter memory 108 stores the prediction parameter generated by the coding parameter determination unit 110 in a predetermined position for each picture and CU to be coded.
  • the reference picture memory 109 stores the decoded image generated by the loop filter 107 in a predetermined position for each picture and CU to be encoded.
  • the coding parameter determination unit 110 selects one of a plurality of sets of coding parameters.
  • the coding parameter is a prediction parameter described above or a parameter to be coded that is generated in association with the prediction parameter.
  • the prediction image generation unit 101 generates a PU prediction image P using each of these sets of coding parameters.
  • the coding parameter determination unit 110 calculates, for each of the plurality of sets, a cost value indicating the size of the information amount and the coding error.
  • the cost value is, for example, the sum of the code amount and a value obtained by multiplying the square error by the coefficient ⁇ .
  • the code amount is the information amount of the coded stream Te obtained by entropy coding the quantization error and the coding parameter.
  • the squared error is a sum between pixels with respect to the square value of the residual value of the residual signal calculated by the subtraction unit 102.
  • the factor ⁇ is a real number greater than a preset zero.
  • the coding parameter determination unit 110 selects a set of coding parameters that minimize the calculated cost value.
  • the entropy coding unit 104 externally outputs the set of selected coding parameters as the coded stream Te, and does not output the set of non-selected coding parameters.
  • the coding parameter determination unit 110 stores the determined coding parameters in the prediction parameter memory 108.
  • the prediction parameter coding unit 111 derives a format for coding from the parameters input from the coding parameter determination unit 110, and outputs the format to the entropy coding unit 104. Derivation of a form for encoding is, for example, derivation of a difference vector from a motion vector and a prediction vector. Further, the prediction parameter coding unit 111 derives parameters necessary to generate a prediction image from the parameters input from the coding parameter determination unit 110, and outputs the parameters to the prediction image generation unit 101.
  • the parameters required to generate a predicted image are, for example, motion vectors in units of subblocks.
  • the inter prediction parameter coding unit 112 derives inter prediction parameters such as a difference vector based on the prediction parameters input from the coding parameter determination unit 110.
  • the inter prediction parameter coding unit 112 derives the inter prediction parameter by the inter prediction parameter decoding unit 303 (refer to FIG. 5 and the like) as a configuration for deriving the parameters necessary for generating the prediction image to be output to the prediction image generation unit 101. Partially include the same configuration as the configuration.
  • the intra prediction parameter coding unit 113 derives a format (for example, MPM_idx, rem_intra_luma_pred_mode, etc.) for coding from the intra prediction mode IntraPredMode input from the coding parameter determination unit 110.
  • a format for example, MPM_idx, rem_intra_luma_pred_mode, etc.
  • the subtracting unit 1123 subtracts the prediction vector mvpLX input from the AMVP prediction parameter derivation unit 1122 from the motion vector mvLX input from the coding parameter determination unit 110 to generate a difference vector mvdLX.
  • the difference vector mvdLX is output to the entropy coding unit 104.
  • the image encoding device 11 and a part of the image decoding device 31 in the embodiment described above for example, the entropy decoding unit 301, the prediction parameter decoding unit 302, the loop filter 305, the prediction image generation unit 308, the inverse quantization / inverse transform Unit 311, addition unit 312, predicted image generation unit 101, subtraction unit 102, transform / quantization unit 103, entropy coding unit 104, inverse quantization / inverse transform unit 105, loop filter 107, coding parameter determination unit 110,
  • the prediction parameter coding unit 111 may be realized by a computer. In that case, a program for realizing the control function may be recorded in a computer readable recording medium, and the computer system may read and execute the program recorded in the recording medium.
  • the “computer system” is a computer system built in any of the image encoding device 11 and the image decoding device 31, and includes an OS and hardware such as peripheral devices.
  • the “computer-readable recording medium” means a portable medium such as a flexible disk, a magneto-optical disk, a ROM, a CD-ROM, or a storage device such as a hard disk built in a computer system.
  • the “computer-readable recording medium” is one that holds a program dynamically for a short time, like a communication line in the case of transmitting a program via a network such as the Internet or a communication line such as a telephone line.
  • a volatile memory in a computer system serving as a server or a client may be included, which holds a program for a predetermined time.
  • the program may be for realizing a part of the functions described above, or may be realized in combination with the program already recorded in the computer system.
  • FIG. 13 is a block diagram showing the configuration of the image decoding apparatus according to the present embodiment.
  • illustration of some members included in the block diagram shown in FIG. 13 is omitted.
  • the same reference numerals are added to members having the same functions as the members shown in FIG. 5, and the description thereof is omitted.
  • the image decoding device 31 includes a header information decoding unit 19, a decoding module 9, and a picture decoding unit 32.
  • the image decoding device 31 may further include a picture division unit 33.
  • the picture decoding unit 32 is a component that performs a process of decoding a picture, and includes a CT information decoding unit 10, a predicted image generation unit 308, an inverse quantization / inverse DCT unit 311, a reference picture memory 306, an addition unit 312, and a loop.
  • the filter 305 and the CU decoding unit 20 are provided.
  • the CU decoding unit 20 further includes a PU information decoding unit 12 and a TT information decoding unit 13, and the TT information decoding unit 13 further includes a TU decoding unit 22.
  • the decoding module 9 performs a decoding process to decode syntax values from binary data. More specifically, the decoding module 9 decodes and decodes the syntax value encoded by the entropy coding method such as CABAC based on the encoded data and the syntax type supplied from the supply source. Return syntax value to supplier.
  • the entropy coding method such as CABAC
  • the sources of the encoded data and the syntax type are the CT information decoding unit 10 and the CU decoding unit 20 (PU information decoding unit 12 and TT information decoding unit 13) (prediction unit decoding unit).
  • the header information decoding unit 19 decodes the video parameter set (VPS), the SPS, the PPS, and the slice header of the encoded data input from the image encoding device 11.
  • the CT information decoding unit 10 uses the decoding module 9 to decode the coding tree unit and the coding tree for the coded data input from the image coding device 11. Specifically, the CT information decoding unit 10 decodes the CTU information and the CT information from the encoded data according to the following procedure.
  • the CT information decoding unit 10 uses the decoding module 9 to decode the tree unit header CTUH from the CTU information included in the CTU.
  • the CT information decoding unit 10 decodes, from the CT information included in the CT, a QT division flag indicating whether or not to divide the target CT into QTs, and a BT division mode indicating a division method of BT division of the target CT.
  • the target CT is recursively divided and decoded until the QT division flag and the BT division mode do not notify of further division.
  • the tree unit footer CTUF is decoded from the CTU information.
  • the tree unit header CTUH and the tree unit footer CTUF include coding parameters to which the image decoding device 31 refers in order to determine the decoding method of the target coding tree unit.
  • the CT information may include parameters applied to the target CT and the lower coding node.
  • the CU decoding unit 20 includes a PU information decoding unit 12 and a TT information decoding unit 13, and decodes PUI information and TTI information of the lowest encoded tree CT (that is, CU).
  • PU information decryption unit In the PU information decoding unit 12, PU information (merge flag (merge_flag), merge index (merge_idx), motion vector index (mvp_idx), reference image index (ref_idx), inter prediction identifier (inter_pred_flag), and difference vector (PU) of each PU mvd) etc.) are decoded using the decoding module 9.
  • the TT information decoding unit 13 decodes each TTI (TU split flag SP_TU (split_transform_flag), CU residual flag CBP_TU (cbf_cb, cbf_cr, cbf_luma), etc., and TU) using the decoding module 9.
  • the TT information decoding unit 13 includes a TU decoding unit 22.
  • the TU decoding unit 22 decodes QP update information (quantization correction value) when a residual is included in the TU.
  • the QP update information is a value indicating a difference value from the quantization parameter prediction value qPpred which is a prediction value of the quantization parameter QP.
  • the TU decoding unit 22 decodes the quantized prediction residual (residual_coding).
  • the picture division unit 33 reads, as input, data indicating a picture in which a plurality of view images are frame-packed, divides the plurality of view images, and decodes the decoded image or encoded data corresponding to one or more view images. Process to output
  • each member included in the image decoding device 31 has a function of performing the processing described in the present example below.
  • the above-mentioned view image (view) is an image obtained by recording or generating a still image or a moving image of a subject or object from a certain viewpoint, and a multiview image is a still image of the subject or object from a plurality of viewpoints It is the image which recorded or produced the animation.
  • the frame-packed image does not necessarily have to be a view image.
  • the above-described frame packing is a configuration in which a plurality of view images are integrated into one picture, that is, a structure in which a plurality of view images are included in one picture.
  • One example is a configuration in which the left eye view image and the right eye view image are frame-packed.
  • FIG. 14 is a diagram illustrating an example of a picture in which a plurality of view images are frame-packed.
  • view 0 and view 1 are frame-packed in the vertical direction.
  • frame packing in the vertical direction means that, as shown in FIG. 14, a plurality of view images are vertically adjacent to each other across a horizontally extending boundary.
  • the plurality of view images naturally includes the case where there are three or more view images, and the same applies in the present specification.
  • the encoded stream Te generated by the image encoding device 11 and input to the image decoding device 31 is a parameter set of VSP (Video Parameter Set), SPS (Sequence Parameter Set), or other parameters which are parameter sets of the encoded stream Te.
  • the set includes view packing information indicating what kind of frame packing has been applied to the picture decoded from the encoded stream Te.
  • view packing information parameters such as view_width [id] and view_height [id] indicating the width and height of a view image frame-packed in a picture are used.
  • the id in the parameter is an index of the view image uniquely assigned to each view image frame-packed in the picture.
  • the header information decoding unit 19 decodes view packing information from the parameter set of the input coded stream Te.
  • the view packing information indicates that the picture to be decoded from the coded stream Te is that the view image is frame-packed in the vertical direction
  • the CT information decoding unit 10 of the picture decoding unit 32 applies Then, raster scan is performed in CTU units and decoding processing is performed.
  • view_packing_flag which is view packing information defined as an example
  • a picture decoded from the coded stream Te indicates that the view image is a picture with frame packing in the vertical direction.
  • view_packing_flag 1
  • the view_packing_flag may be called a view_vertical_packing_flag in order to clarify that frame packing in the vertical direction is performed.
  • the information indicating that the view is vertically packed may be indicated not by one syntax but by a combination of plural syntaxes. For example, a flag view_packing_flag indicating whether the view is packed and view_packing_type which is information on the packing method (showing decoding order) of the view may be divided and included in the encoded stream Te.
  • the picture division unit 33 divides the picture decoded by the picture decoding unit 32 into each view, and uses the picture as an output of the image decoding device 31.
  • the image decoding device 31 is the image decoding device 31 that decodes the coded stream Te of the image, and includes the header information decoding unit 19 and the picture decoding unit 32, and the header information decoding unit
  • the reference numeral 19 denotes view packing information included in the parameter set of the encoded stream Te, and decodes the view packing information indicating whether or not a plurality of views are included in the target picture, and the picture decoding unit 32
  • the view packing information decodes each view by decoding a picture specified that the plurality of views are vertically frame-packed pictures.
  • the image decoding apparatus 31 does not have to decode view 1 when extracting view 0 included in the picture 50.
  • the image decoding device 31 when extracting a specific view image from a picture in which a plurality of view images are frame-packed, the image decoding device 31 is a continuous range of the encoded stream Te, Since it is sufficient to decode the range corresponding to the end of the specific view image, the processing amount can be reduced. Similarly, coded data can be extracted without decoding subsequent views of a particular view image.
  • an image coding apparatus to be paired with the image decoding apparatus 31 is an image coding apparatus for coding the coded stream Te of the image, and comprises a header information coding unit and a picture coding unit,
  • the header information coding unit codes view packing information which is view packing information included in the parameter set of the coded stream Te and which indicates whether a target picture includes a plurality of views or not, and the picture code
  • the encoding unit encodes each view by encoding a picture specified by the view packing information as the picture in which the plurality of views are frame-packed in the vertical direction.
  • the encoded stream Te may have a syntax indicating the start position of each view image included in the picture in VPS, SPS, or another parameter set.
  • the parameter indicating the address of the start position in the horizontal direction of the id-th view image included in the picture may be defined as view_position_x [id]
  • the parameter indicating the address of the start position in the vertical direction may be defined as view_position_y [id].
  • the encoded stream Te may have a configuration in which the parameter is included as view packing information, or may be configured as another parameter set.
  • the header information decoding unit 19 decodes information indicating start positions of the plurality of views included in the parameter set of the encoded stream Te, in addition to the view packing information.
  • the picture decoding unit 32 refers to view_position_x [id] and view_position_y [id], which are parameters indicating the start position of the view image, and view_height [id], which is a parameter indicating the height of the view image, Decode the
  • FIG. 19 is pseudo code including at least view_packing_flag, view_position_x [id], view_position_y [id], view_height [id], and the like as view packing information of the encoded stream Te.
  • the picture decoding unit 32 decodes a view image included in a picture, only an arbitrary view image included in the picture can be decoded and extracted, so that an effect of reducing the amount of processing is achieved.
  • each member included in the image decoding device 31 has a function of performing the processing described in the present example below.
  • FIG. 15 is a diagram showing an example of a picture including a padding portion.
  • the view 0 and the view 1 are frame-packed in the vertical direction across the padding portion 52.
  • the image coding apparatus 11 performs padding with a vertical size of 72 below the view 0 and Frame-pack view 1 down.
  • the total size of the view 0 and the padding in the vertical direction is 1152, which is a multiple of 128 which is the size of the CTU. Also, there is no need to add padding below the view 1 located below the picture.
  • the image coding device 11 outputs a coded stream Te obtained by coding the picture 51, and the image decoding device 31 receives the coded stream Te and performs an image decoding process.
  • the CT information decoding unit 10 performs raster scan on the picture 51 in CTU units from the direction corresponding to the upper part in FIG.
  • FIG. 20 is a pseudo including view_packing_width [i], view_cropping_heigth [i], view_same_size_flag, etc. in addition to view_packing_flag, view_position_x [id], view_position_y [id], view_height [id] as view packing information of encoded stream Te. It is a code.
  • view_width [i] and view_height [i] are the width and height of the view with padding area
  • the width and height of the real area are respectively view_width [i]-view_cropping_width [i], view_height [i]-view_cropping_heigth [ i].
  • the present invention according to this example is not limited to the case in which a plurality of view images are frame-packed in the vertical direction, and is also applicable to the case in which frame-packing is performed in the horizontal direction.
  • the configuration may be such that padding is not performed when performing frame packing.
  • the image decoding device 31 is an image decoding device 31 that decodes the coded stream Te of the image, and includes a picture decoding unit 32 that decodes a picture, and the picture decoding unit 32 A packed picture, a portion padded so that the vertical or horizontal size of each view located other than the end of the picture is a multiple of a coding tree unit which is a predetermined coding unit, and a plurality of views Decode the picture it contains.
  • the CTU when referring to or extracting a frame-packed picture in CTU block units, the CTU is prevented from including data derived from a plurality of view images, and a predetermined view is obtained.
  • the effect is to decode and extract only.
  • each member included in the image decoding device 31 has a function of performing the processing described in the present example below.
  • the encoded stream Te generated by the image encoding device 11 and input to the image decoding device 31 is a parameter set of VSP (Video Parameter Set), SPS (Sequence Parameter Set), or other parameters which are parameter sets of the encoded stream Te.
  • VSP Video Parameter Set
  • SPS Sequence Parameter Set
  • the set includes view packing information indicating what kind of frame packing has been applied to the picture decoded from the encoded stream Te.
  • FIG. 16 is a diagram illustrating an example of a picture in which a plurality of view images are frame-packed.
  • the header information decoding unit 19 decodes the parameter set of the input coded stream Te, and refers to view packing information.
  • the view packing information indicates that the target picture to be decoded from the encoded stream Te is divided into a plurality of view images, and the target picture is a frame of the view image in the horizontal direction as shown in the picture 53 of FIG.
  • the picture decoding unit 32 decodes the target picture by scanning it for each view. In the example shown in FIG. 16, the picture decoding unit 32 starts the decoding process of view 1 after completing the decoding process of view 0 among view 0 and view 1.
  • the encoded stream Te may have a parameter set, eg, view_packing_type, for defining the decoding order of view images in the image decoding device 31.
  • view_packing_type 0
  • the header information decoding unit 19 interprets that the picture is subjected to frame packing in the vertical direction (top and bottom), and the CT information decoding unit 10 of the picture decoding unit 32 Scans the CTUs in raster scan order within the picture.
  • view_packing_type 1
  • the header information decoding unit 19 determines that the picture is frame-packed in the horizontal direction (side and side), and the picture decoding unit 32 executes the raster scan order for each view image.
  • the CTU may be scanned.
  • the present invention according to the present embodiment is not limited to the case of including a plurality of view images frame-packed in the horizontal direction, and is applied, for example, to the case of decoding a picture in which view images are arranged in a grid.
  • View_packing_type 2 etc. may be used).
  • the coded stream Te has view packing information indicating how many view images are frame-packed pictures in the horizontal direction and the vertical direction, for example, the view number in the horizontal direction and the view in the vertical direction. It has the number view_num_verical.
  • position view_position_x [i] and view_position_y [i] of each view and sizes view_width_minus1 [i] and view_height_minus1 [i] are used as view packing information. You may use.
  • the image decoding apparatus 31 is an image decoding apparatus 31 that decodes a coded stream Te of an image, and includes a header information decoding unit 19 and a picture decoding unit 32, and the header information decoding unit 19
  • the view packing information which is view packing information included in the parameter set of the encoded stream Te and which indicates whether a target picture includes a plurality of views, is decoded, and the picture decoding unit 32 performs the view packing.
  • the information indicates that the target picture is divided into a plurality of views, the target picture is decoded by scanning for each view.
  • the image decoding apparatus 31 can extract the specific view image only by decoding the continuous range including the specific view image from the beginning of the encoded stream Te, so the processing amount can be reduced. Play.
  • a view tile is composed of one or more tiles or slices.
  • symbol is appended and description is abbreviate
  • each member included in the image decoding device 31 has a function of performing the processing described in the present example below.
  • a tile refers to a picture divided into rectangles in CTU units.
  • the encoding and decoding order of each CTU included in the tile is the raster scan order in the tile.
  • FIG. 17 is a diagram showing an example of a picture divided into a plurality of view tiles.
  • one tile corresponds to one view tile.
  • one view tile may be composed of a plurality of tiles.
  • the coded stream Te generated by the image coding apparatus 11 has view packing information indicating, for example, view_tile_packing_flag indicating whether or not a view tile is used in a picture decoded from the coded stream Te.
  • view_tile_packing_flag indicating whether or not a view tile is used in a picture decoded from the coded stream Te.
  • the header information decoding unit 19 decodes view packing information which is view packing information included in the input coded stream Te and which indicates whether a view tile is used.
  • view packing information indicates that the view tile is used
  • the picture decoding unit 32 decodes the view image by decoding the target view tile to be decoded.
  • the image decoding device 31 is the image decoding device 31 that decodes the coded stream Te of the image, and is predicted in the header information decoding unit 19, the picture decoding unit 32, and the picture decoding unit 32.
  • the header information decoding unit 19 is view packing information included in the parameter set of the coded stream Te, and indicates whether or not a view tile is used. Decoding view packing information, the picture decoding unit 32 decodes a specified view image by decoding a designated view tile when the view packing information indicates that a view tile is used.
  • the encoded stream Te may have information indicating each tile included in the view tile, for example, tile_included_flag [view_tile_idx] [tile_idx].
  • tile_included_flag indicates whether or not the view tile indicated by the view tile index view_tile_idx includes a tile of a certain tile_idx.
  • the encoded stream Te may be configured to have information indicating the view tile index of each tile.
  • the image decoding device 31 may be configured to decode in parallel each tile included in the view tile. Note that, instead of the information tile_view_tile_idx [i] indicating the view tile index of each tile i, the information tile_view_idx [i] indicating an index for identifying a view may be used. Also, it may be information tile_view_order_idx [i] indicating a view order index which is an index for identifying the order of views.
  • a configuration may be adopted in which a view slice (view slice) which is a slice associated with one view image is defined.
  • a slice refers to a continuous CTU region when a picture is scanned in raster scan order. The end of the slice may be located in the middle of each row in raster scan. Such a continuous area may be called a view segment.
  • the coded stream Te has view packing information indicating, for example, a view_slice_packing_flag indicating whether or not a view slice is used in a picture decoded from the coded stream Te.
  • view packing information indicating, for example, a view_slice_packing_flag indicating whether or not a view slice is used in a picture decoded from the coded stream Te.
  • the image coding device 11 sets each of the view images to a view slice with a view_slice_packing_flag set to 1, and performs coding processing.
  • the header information decoding unit 19 decodes view packing information which is view packing information included in the input coded stream Te and which indicates whether or not a view slice is used.
  • view packing information indicates that the view slice is used
  • the picture decoding unit 32 decodes the view image by decoding the target view slice.
  • the predicted image generation unit 308 is restricted from referring to a block included in a tile other than the target tile.
  • the image decoding apparatus in the view tile, may be configured to be able to refer to a block included in a view tile other than the target view tile. That is, the image decoding device 31 may be configured to be able to refer to the pixels of the reference view tile other than the target view tile, the prediction mode, the intra prediction mode, and the motion vector. Specifically, the image decoding apparatus 31 may perform the reference illustrated below.
  • the area to be referred to by the predicted image generation unit 308 is limited to facilitate prefetching, thereby suppressing an increase in memory access.
  • coding efficiency can be improved because blocks of another view tile very similar to the target block can be referenced.
  • each different from the object view tile It may be configured to include inter-view referenceability information indicating whether or not the view tile can be referenced, for example, view_tile_dependency_flag [i] [j].
  • view_tile_dependency_flag [i] [j]
  • the first subscript (index) i of view_tile_dependency_flag [i] [j] represents the view tile of the reference source
  • the second subscript j represents the view tile of the reference destination.
  • i and j are integers of 0 or more, and i> j.
  • inter-view referenceability information is used to indicate a dependency between view tiles
  • inter-tile reference information is used to indicate a dependency between tiles. And both.
  • view_tile_dependency_flag [i] [j] the reference relation between the same views.
  • the image decoding apparatus does not use a picture that is not dependent, that is, a picture that can not be referred to, as a reference picture, and does not add it to the reference picture list.
  • the configuration using view_tile_dependency_flag [i] [j] is different from the configuration using layer_dependency_flag [i] [j] in that the region of the view image not having a dependency is not used as a reference region.
  • the view_tile_dependency_flag [] [] may be simply described as tile_dependency_flag [] [] as described later.
  • FIG. 18 is a diagram illustrating an example of inter-view referenceability information between view tiles in a picture.
  • the value of view_tile_dependency_flag [1] [0], which is inter-view referenceability information is 1.
  • the above values indicate that when the predicted image generation unit 308 generates a predicted image of the target block included in the view tile 1, it can refer to the reference block included in the view tile 0.
  • the above value is 0, it is included in the view tile 0 which is the reference tile of the reference destination when generating a predicted image of the target block included in the view tile 1 which is the view tile of the reference source. It means that reference to the block is not possible.
  • the configuration using inter-view reference information is also applicable to the case where a plurality of view tiles correspond to one view image as in the example shown in (b) of FIG. 18 (in this case) Use inter-tile reference information). That is, in the example shown in (b) of FIG. 18, view 0 consists of view tile 0 (tile 0) and view tile 1 (tile 1), and view 1 is from view tile 2 (tile 2) and view tile 3 ( become a tile 3). In the example shown in (b) of FIG. 18, the dependency between each view tile is defined by the syntax tile_dependency_flag [i] [j].
  • the view tile 0 and the view tile 1 to the view 0 are configured and independent of each other.
  • tile_dependency_flag [1] [0] 0 // View tile 1 does not refer to view tile 0 View 1 is constructed from view tile 2 and view 1 (view tile 2) refers to view 0 (view tile 0).
  • view tile 0 and view tile 1 constituting view 0 can be decoded independently.
  • the view tile 2 constituting the view 1 and the view tile 3 constituting the view 2 respectively have to wait for the decoding of the view tile 0 and the view tile 1 but can be decoded independently.
  • FIG. 21 is a syntax table showing an example of a generation procedure of parameter sets included in the coded stream Te according to the present embodiment.
  • ue (v) described in the item Descriptor in FIG. 21 and the other syntax table figures means that the size of the corresponding syntax is variable (when the code length is 1 bit or as a result, 1 bit).
  • u (1) means that the size of the corresponding syntax is 1, and is a flag that can normally take a value of 0 or 1.
  • some syntax names in FIG. 21 and other diagrams of syntax tables and the meanings of the syntaxes conform to the specification of H. 265 of the International Telecommunication Union-Telecommunication Standardization Sector (ITU-T).
  • FIG. 25 is a flowchart showing an example of a process of setting a parameter set included in the coded stream Te according to the present embodiment.
  • step S101 the image encoding device 11 sets a parameter set that the encoded stream Te has.
  • Lines 2 to 5 in FIG. 21 correspond to the process in step S101.
  • the fifth line view_packing_flag is a syntax indicating whether the picture to be encoded by the image encoding device 11 is a plurality of view images frame-packed.
  • step S102 it is determined whether the picture to be encoded is a picture including a plurality of view images.
  • the sixth line in FIG. 21 corresponds to the process in step S102. If the picture includes a plurality of view images, that is, if the view_packing_flag is 1, the process transitions to step S103. If not included, that is, if the view_packing_flag is 0, the process based on the flowchart in FIG.
  • step S103 the image encoding device 11 sets the number of rows and the number of columns of the view tile included in the picture.
  • the seventh to ninth lines in FIG. 21 correspond to the process in step S103.
  • view_same_size_flag in the ninth row is a syntax that indicates whether the picture and the view tile included in the picture have the same size.
  • view_same_size_flag is a syntax that indicates that the picture includes a view image of 1 when the value is 1.
  • step S104 the image encoding device 11 sets the dependency between view tiles included in the picture, that is, the inter-view referenceability information.
  • the tenth to fifteenth lines in FIG. 21 correspond to the process in step S104.
  • a unique index is set for each view tile in the process shown in the 12th line, and inter-view reference information is set in the process shown in the 15th line.
  • step S105 the position, size, and cropping area of each view tile included in the picture are set.
  • Lines 16 to 27 in FIG. 21 correspond to the process in step S105.
  • the above process is performed, and the process based on the flowchart of FIG. 25 is ended.
  • the image decoding apparatus 31 decodes the encoded stream Te having the inter-view reference information as described above as a parameter set.
  • the header information decoding unit 19 decodes the inter-view reference availability information
  • the predicted image generation unit 308 refers to the inter-view reference availability information
  • the view tile is different from the target view tile. Identify the referenced reference blocks to be included.
  • the predicted image generation unit 308 appropriately refers to the reference block and generates a predicted image of the target block included in the target view tile.
  • the header information decoding unit 19 decodes inter-view referenceability information indicating whether or not reference to each view tile different from the target view tile is possible, and the predicted image generation unit 308 detects the target view In predicted image generation processing for generating a predicted image of a target block included in a tile, a reference block is specified with reference to the inter-view reference information.
  • the predicted image generation unit 308 can easily specify the view tile that can be referred to by referring to the inter-view referenceability information, which is useful for the decoding process.
  • FIGS. 19, 20, and 22 to 24 are syntax tables showing an example of a generation procedure of parameter sets included in the coded stream Te.
  • the coded stream Te generated by the image coding device 11 does not have inter-view reference availability information.
  • the number of rows and the number of columns of view images constituting a picture may be set in the same loop, or may be set independently as shown in FIG.
  • FIG. 22 is an example of using the tile syntax in the case of using a view tile.
  • tile_view_flag on the eighth line is a syntax indicating whether or not the tile is treated as a view image.
  • tile_view_idx [i] indicates the view index of each tile i (it may be a view order index instead of a view index).
  • the image encoding device 11 separates the dependency in the picture (intra) and the dependency in the inter (picture) with respect to the dependency between the view tiles.
  • the configuration may be set in the syntax.
  • view_tile_dependency_type [i] [j] is information indicating the dependency type between views, and the value of dependency type is intra prediction (eg 1), inter prediction (eg 2), intra prediction and inter prediction (eg 3) It may be.
  • the process shown in FIG. 24 has a configuration in which the view tile is replaced with a tile in the process in FIG.
  • the view image to be encoded is composed of a plurality of view tiles (hereinafter referred to as tiles).
  • tiles A specific example in the case where there is a connection will be described with reference to FIG. 26, FIG. 27 and FIG.
  • FIG. 26 will be described for the tile order index tile_order_id [i] (tile decoding order index TileOrderId) used in this specific example.
  • (A) and (b) of FIG. 26 are diagrams for explaining the decoding order of each tile included in a certain picture.
  • the values in the rectangular tiles indicate the value of the tile order index TileOrderId of each tile.
  • the header information decoding unit 19 of the image decoding device 31 decodes the encoded data having the bit stream restriction.
  • the tile position index is an index that indicates the position of a tile in a picture, and is set in raster order in this example.
  • FIG. 26 (B) of FIG. 26 is a diagram for explaining the tile order index TileOrderId.
  • FIG. 26 (b) shows an example in which TileOrderId is assigned to each tile, such as the upper left tile, the lower left tile, the right tile,..., And the tile at the end of the second row and first column. .
  • the tile decoding order is not restricted to the raster order of the pictures.
  • the arrow in a figure shows the reference relation mentioned later.
  • tile order index tile_order_id [i] which is a syntax indicating a decoding order of a certain tile i, may be encoded in encoded data.
  • FIG. 27 is a flowchart for explaining the parameter set setting method according to this example.
  • FIG. 28 is a syntax table showing an example of a generation procedure of parameter sets according to this specific example.
  • the image encoding device 11 encodes tiles in a certain picture in the order indicated by TileOrderId.
  • step S110 the header information encoding unit of the image encoding device 11 determines whether the flag curr_pic_tile_referencing_flag indicates that another tile included in the same picture is to be referred to. When it is determined in step S110 that the header information encoding unit refers to another tile included in the same picture, the process proceeds to step S111. If not, the header information encoding unit ends the processing.
  • curr_pic_tile_referencing_flag The portion of if (curr_pic_tile_referencing_flag) in FIG. 28 corresponds to the process in step S110. As shown in (a) of FIG. 37, curr_pic_tile_referencing_flag may be explicitly encoded. In this case, the following processing is performed.
  • the image encoding device 11 encodes a flag curr_pic_tile_referencing_flag indicating whether to refer to another tile included in the same picture.
  • curr_pic_tile_referencing_flag may use tile_view_flag described above. Note that as shown in (b) of FIG. 37, curr_pic_tile_referencing_flag may not be explicitly coded, but may be derived from tile_dependency_flag [i] [j] (a flag in this case may be called curr_tile_dependency_flag).
  • step S111 the header information encoding unit (tile decoding order encoding unit) of the image encoding device 11 encodes the tile order index tile_order_id [i] described above.
  • i may be coded as the loop variable by the number of tiles NumTileMinus1 + 1. Encode a unique tile order index for each tile.
  • the tile order index tile_order_id [i] may be encoded in raster order with respect to the tiles constituting the inside of the picture.
  • step S112 the header information encoding unit of the image encoding device 11 encodes dependency relationships between view tiles included in a picture, that is, inter-tile reference availability information.
  • inter tile referenceability information tile_dependency_flag [i] [j] is set. If tile_dependency_flag [i] [j] is 1, it indicates that tile i (first tile) can refer to tile j (second tile). If tile_dependency_flag [i] [j] is 0, it indicates that tile i (first tile) can not refer to tile j (second tile).
  • the above image coding apparatus 11 is information indicating the tile decoding order when the flag indicating whether there is a dependency relationship between tiles (whether a tile references another tile or not) is 1.
  • the tile order index may be encoded. Unlike the conventional tiles that can be decoded in parallel with each other, in the present configuration in which reference is made to another tile, there is a restriction on the decoding order of the tiles, so information that indicates the tile decoding order is encoded.
  • the inter-tile reference availability information set in step S112 is the second of the tiles when predicting the target block of the first tile (the tile indicated by i) of the tiles constituting the view image. Indicates whether or not 2 tiles (the tile indicated by j) can be referred to.
  • the value of the first index (i mentioned above) specifying the first tile is higher than the value of the second index (j mentioned above) specifying the second tile. Limit that it is big. This has the merit that the number of tile_dependency_flag [] [] to be notified can be reduced and the code amount can be reduced.
  • the notification of the parameter means that the parameter is included in the encoded data (bit stream), the moving image coding apparatus encodes the parameter, and the moving image decoding apparatus decodes the parameter.
  • FIG. 29 is a flowchart for explaining a specific example of the tile decoding method based on the parameter set according to this specific example.
  • the header information decoding unit 19 of the image decoding device 31 determines whether the flag curr_pic_tile_referencing_flag indicates that another tile included in the same picture is referred to. When it is determined that the header information decoding unit references another tile included in the same picture (curr_pic_tile_referencing_flag is 1) (YES in step S119), the process proceeds to step S120. If not (NO in step S119), the header information decoding unit ends the process.
  • step S120 the header information decoding unit 19 of the image decoding device 31 decodes the tile order index (tile decoding order index) encoded by the image encoding device 11.
  • step S121 the header information decoding unit 19 of the image decoding device 31 decodes the inter-tile referenceability information encoded by the image encoding device 11.
  • the value of the first index (i mentioned above) specifying the first tile is higher than the value of the second index (j mentioned above) specifying the second tile. There is a restriction that is large.
  • step S122 the inter prediction parameter decoding unit 303 of the picture decoding unit 32 of the image decoding device 31 refers to the inter-tile referenceability information decoded by the header information decoding unit 19 and determines whether a referenceable tile exists.
  • the process proceeds to step S123. If the inter-tile reference information indicates that there is no tile that can be referred to, the process ends.
  • the presence of a referenceable tile indicates that the current tile (first tile) including the target block can refer to at least one other tile (second tile). Note that curr_pic_tile_referencing_flag may be explicitly decoded as shown in (a) of FIG.
  • curr_pic_tile_referencing_flag curr_tile_dependency_flag. Further, as shown in FIG. 34 described later, the current picture reference utilization flag pps_curr_pic_ref_enabled_flag may be used.
  • step S123 the inter prediction parameter decoding unit 303 of the picture decoding unit 32 of the image decoding device 31 adds the current picture to the reference picture list. As a result, the second tile is added to the reference picture list.
  • the above image decoding apparatus 31 is a tile that is information indicating a tile decoding order when the flag indicating whether there is a dependency relationship between tiles (whether a tile references another tile or not) is 1.
  • the order index may be decrypted.
  • the tile decoding order is constrained because the tile decoding order is constrained, and thus information indicating the tile decoding order is decoded.
  • the image decoding apparatus 31 can perform decoding with reference to the already decoded tile based on the inter-tile reference availability information in the decoding order specified by the tile order index, as a result of the above processing.
  • FIG. (A) and (b) of FIG. 30 is a figure for demonstrating the image decoding method by the image decoding apparatus 31 based on this specific example, respectively (The arrow has started the object block and has pointed to the reference block. ).
  • the inter-tile reference availability information decoded by the header information decoding unit 19 is expressed by the following equation.
  • i in tile_dependency_flag [i] [j] is the value of the first index indicating the tile decoding order (decoding order in FIG. 30)
  • j is the decoding order (decoding order in FIG. 30). It is the value of the 2nd index which shows.
  • the first index indicates the first tile that is the reference source, and the second index indicates the second tile that is the reference destination.
  • the order of the indexes of tile_dependency_flag [i] [j] may be tile_dependency_flag [j] [i]. That is, it may be described in the order of the second index j and the first index i.
  • the image decoding apparatus 31 decodes tiles in the order indicated by TileOrderId, that is, tile 0, tile 1, and tile 2 in this order.
  • the predicted image generation unit 308 refers to the inter-tile referenceability information tile_dependency_flag when decoding the target block included in the tile 1 (first tile), and the tile 1 (first If the flag tile_dependency_flag [1] [0] indicating whether or not tiles can be referenced to tile 0 (second tile) is 1, tile 1 (first tile) is used using tile 0 (second tile) Generate a predicted image of the target block of That is, a reference picture list including tile 0 (second tile) is used to specify a reference picture, and a predicted image of the target block is generated using the specified reference picture.
  • a predicted image of a target block of tile 2 (first tile) is generated. That is, a reference picture list including tile 0 (second tile) is used to specify a reference picture, and a predicted image of the target block is generated using the specified reference picture.
  • the image decoding device 31 decodes tiles in the order indicated by the tile order index TileOrderId. That is, the tiles are decoded in order from the tile with the smallest tile order index TileOrderId (here, in the order of tile 0, tile 1 and tile 2).
  • the predicted image generation unit 308 refers to the tile order index and the inter-tile referenceability information, and thus each object is in the order of tile 0, tile 1, and tile 2, which is the decoding order specified by the tile order index. Generate a predicted image of the block.
  • tile_dependency_flag [i] [j] indicated by the above three expressions i is the value of the first index indicating the tile order index, and j is the value of the second index indicating the tile order index.
  • the first index indicates a first tile and the second index indicates a second tile.
  • a predicted image of the target block of tile 1) is generated (at this point, tile 1 has been decoded). That is, a reference picture list including tile 1 (second tile) is used to specify a reference picture, and the specified reference picture is used to generate a predicted image of the target block.
  • the above configuration is an image decoding apparatus that decodes a tile that constitutes a view, and when predicting a target block of a first tile of the tiles, whether or not the second tile of the tiles can be referred to Inter-tile reference information decoding unit that decodes the inter-tile reference information and the reference picture including the second tile when the inter-tile reference information indicates that the second tile can be referred to; And a header information decoding unit for decoding a tile decoding order index indicating a decoding order of each tile.
  • the predicted image generation unit 308 refers to the tile order index and the inter-tile referenceability information, and thereby the tile order index is the decoding order specified by the tile order index. A predicted image of each target block of the tile can be generated.
  • the image decoding device 31 is an image decoding device that decodes a tile that configures a view, and when the target block of the first tile among the tiles is predicted, the tile Inter-tile reference information decoding unit (header information decoding unit 19) for decoding inter-tile reference information indicating whether or not the second tile in the second tile can be referred to;
  • a prediction image generation unit (prediction image generation unit 308) that generates a prediction image of the target block with reference to the second tile (reference tile) when it indicates that the tile can be referred to;
  • the image generation unit generates a predicted image of each target block of the tile in the decoding order notified by the tile order index (tile decoding order index).
  • the reference tile to which the target tile refers is only an index smaller than the target tile, it is possible to reduce the number of bits required for inter-tile reference information. Further, by referring to the inter-tile reference information, it is possible to easily identify the tile that can be referred to. Therefore, it is not necessary to refer to tiles other than the identified referenceable tiles, and a predicted image of each target block of the tile can be generated in a predetermined decoding order, and the view image can be efficiently decoded. Can.
  • an image coding device 11 corresponding to the above-described image decoding device 31 is also included in the present invention. More specifically, the image encoding device 11 is an image encoding device that encodes a tile that constitutes a view, and when predicting a target block of a first tile among the tiles, the image encoding device 11
  • the inter-tile reference information encoding unit (the above-mentioned header information encoding unit) that encodes inter-tile reference information indicating whether or not the second tile can be referred to;
  • a predicted image generation unit (predicted image generation unit 101) for generating a predicted image of the target block with reference to the second tile (reference tile) when indicating that the second tile can be referred to.
  • the predicted image generation unit generates a predicted image of each target block of the tile in the coding order notified by the tile order index.
  • the value of the index set to each tile is fixed (for example, in raster order), and the value of the first index specifying the first tile is the first A restriction is imposed that it is greater than the value of the second index specifying tile 2, so if the value of the first index is smaller than the value of the second index, it is possible to refer to the second tile
  • the image decoding apparatus 31 further includes a tile decoding order information decoding unit (header information decoding unit 19) that decodes a tile decoding order index that specifies the decoding order of the plurality of tiles, and the predicted image
  • the generation unit refers to the second tile in which the value of the tile decoding order index is specified as the value of the second index, and the tile decoding order is determined as the value of the first index.
  • the restriction is imposed that the value of the first index specifying the first tile is larger than the value of the second index specifying the second tile
  • the restriction Since it is sufficient to decode only the inter-tile reference information that is imposed, it is possible to generate predicted images of respective target blocks of a plurality of tiles in a predetermined decoding order while suppressing the bit amount of the flag. Images can be decoded efficiently.
  • FIG. 31 is a flowchart for explaining a specific example of the tile decoding method based on the parameter set according to this specific example.
  • the image decoding device 31 decodes a flag tile_view_flag indicating whether to refer to another tile included in the same picture.
  • step S129 The header information decoding unit 19 of the image decoding device 31 determines whether or not the flag tile_view_flag indicates that another tile included in the same picture is to be referred to. If it is determined that the header information decoding unit refers to another tile included in the same picture (YES in step S129), the process proceeds to step S130. If not (NO in step S129), the header information decoding unit ends the process. (S130) In step S130, the header information decoding unit 19 of the image decoding device 31 decodes the inter-tile reference availability information encoded by the image encoding device 11. Note that the inter-tile reference availability information includes the value of the first index specifying the first tile (i described above) and the value of the second index specifying the second tile (j described above). There is a restriction that is different.
  • step S131 the predicted image generation unit 308 of the picture decoding unit 32 of the image decoding device 31 derives a tile order from the inter-tile reference availability information decoded by the header information decoding unit 19.
  • step S132 the inter prediction parameter decoding unit 303 of the picture decoding unit 32 of the image decoding device 31 can refer to the second tile with reference to the inter-tile referenceability information decoded by the header information decoding unit 19. By determining whether or not there is a tile that can be referred to. If the inter-tile reference information indicates that the second tile can be referenced, the process proceeds to step S134. If the inter-tile reference information indicates that the second tile can not be referred to, the process ends.
  • the inter prediction parameter decoding unit 303 of the picture decoding unit 32 of the image decoding device 31 adds the current picture to the reference picture list. As a result, the second tile is added to the reference picture list.
  • FIG. 32 is a syntax table showing an example of a generation procedure of a parameter set according to this example.
  • (B) of FIG. 32 is a diagram for describing an image decoding method by the image decoding device 31 according to the present specific example (arrows start from the target block and point to reference blocks).
  • tile_order_idx [i] is deleted as compared with the syntax table shown in FIG. That is, the header information encoding unit of the image encoding device 11 does not encode the tile order index.
  • the predicted image generation unit 308 of the picture decoding unit 32 of the image decoding device 31 refers to the reference picture list generated in the process of step S133 to obtain the second tile.
  • the predicted image of the target block of the first tile designated by the value of the first index in the inter-tile reference information determined to be referable is preferentially generated.
  • the inter-tile referenceability information decoded by the header information decoding unit 19 is represented by the following equation.
  • the inter-tile reference availability information of combinations of (i, j) other than the above is 0.
  • the predicted image generation unit 308 of the picture decoding unit 32 of the image decoding device 31 designates the first index value 0 or 2 in the process following step S133 described above.
  • the predicted image using tile 1, which is the second tile, and the target block of the first tile specified by the value 3 or 5 of the first index, in the second tile A predicted image using a certain tile 4 is preferentially generated.
  • the prediction image generation unit 308 may generate prediction images of respective target blocks of the plurality of tiles in raster order (for example, tile 1, In the order of tiles 4).
  • the image decoding apparatus 31 specifies the first tile in the inter-tile reference information decoded by the inter-tile reference information decoding unit (header information decoding unit 19).
  • a restriction is imposed that the value of the first index is different from the value of the second index specifying the second tile, and the prediction image generation unit (prediction image generation unit 308) performs inter-tile reference
  • Each target block of the plurality of tiles in a decoding order in which the first tile designated by the first index value is preferentially referenced with reference to the second tile designated by the second index value in the permission information Generate a predicted image of
  • the code amount of the index can be reduced.
  • FIG. 33A is a view for explaining the configuration of encoded data and the tile position index TilePosID.
  • tile encoded data is arranged in tile order index order.
  • the tile position index TilePosID is information indicating where each tile should be displayed in the screen, and the picture information is determined by decoding the tile position index TilePosID.
  • FIG. 33 (b) shows the arrangement on the picture of the tile position index TilePosID, which is arranged in raster order in this example. Also, the lines in each figure show the reference relationship of each tile.
  • FIG. 34 is an example of a syntax table showing a configuration of encoded data of tile position index TilePosID.
  • a flag curr_tile_dependency_flag indicating whether there is a dependency between tiles is notified, and when curr_tile_dependency_flag is 1, tile position index TilePosID [i] of each tile in order of tile decoding order To notify.
  • tile_dependency_flag [] [] indicating individual dependencies of each tile may be notified next.
  • (B) of FIG. 34 notifies tile_dependency_flag [] [] indicating individual dependencies of each tile, and as described above, indicates whether or not there is a dependency relation between tiles from tile_dependency_flag [] [].
  • the flag curr_tile_dependency_flag is derived. When the derived curr_tile_dependency_flag is 1, the tile position index TilePosID [i] of each tile is notified in the order of tile decoding order.
  • the header information encoding unit (tile position information encoding unit) of the image encoding device 11 encodes tile position information when there is dependency between tiles.
  • the header information decoding unit 19 (tile position information decoding unit) of the image decoding device 31 decodes tile position information encoded by the image encoding device 11 when there is dependency between tiles.
  • FIG. 35 is a syntax table showing an example of notifying upper left coordinates tile_position_x, tile_position_y of tiles when there is a dependency between tiles.
  • a flag indicating that there is a dependency between tiles may be decoded ((a) in FIG. 35) or may be derived ((b) in FIG. 35).
  • the position may be in units of 2 pixels or 4 pixels instead of each pixel.
  • FIG. 36 shows a syntax configuration for notifying inter-tile reference availability information when pps_curr_pic_ref_enabled_flag is 1.
  • the header information encoding unit (tile position information encoding unit) of the image encoding device 11 encodes the inter-tile reference availability information.
  • the header information decoding unit 19 (tile position information decoding unit) of the image decoding device 31 decodes the inter-tile referenceability information encoded by the image encoding device 11.
  • FIG. 36 further shows a syntax configuration for notifying a tile order index when pps_curr_pic_ref_enabled_flag is 1.
  • (B) of FIG. 36 shows a syntax configuration for notifying of a tile position index when pps_curr_pic_ref_enabled_flag is further “1”.
  • the header information encoding unit (tile position information encoding unit) of the image encoding device 11 encodes tile position information when pps_curr_pic_ref_enabled_flag is 1.
  • the header information decoding unit 19 (tile position information decoding unit) of the image decoding device 31 decodes tile position information encoded by the image encoding device 11.
  • the tile position information may be a tile position index or tile upper left coordinates.
  • Example of reference block restriction (view image reference restriction)
  • the predicted image generation unit 308 restricts the view image to be referred to in the predicted image generation process of generating a predicted image of the target block.
  • symbol is appended and description is abbreviate
  • each member included in the image decoding device 31 has a function of performing the processing described in the present example below.
  • FIG. 38 is a diagram showing an example of a view image that can be referred to when generating a predicted image of a view image.
  • the block included in the view image here, the view 1 of the picture 57
  • the view i refers to the view image previously decoded in decoding order
  • h, i and j are integers of 0 or more.
  • a vector (motion vector) for specifying a reference block as shown by the arrow in FIG. 38 is used.
  • FIG. 39 is a diagram illustrating an example of clipping a vector pointing to a block to be referred to in a predicted image generation process.
  • mvLX [1] that is the y-direction (vertical direction) component of the clipped motion vector is shown below.
  • yPb + mvLX [1] + nPbH-1 view_position_y [view_id] + view_height [view_id]
  • view_position_y [view_id] and view_height [view_id] mean the start (upper left) y coordinate and height of the view image having view_id
  • yPb and nPbH mean the upper left coordinate and height of the target block, respectively.
  • Clip 3 (x, y, z) is a function that returns x when z ⁇ x, y when z> y, and z otherwise.
  • mvLX [0] Clip3 (0, view_position_x [view_id] + view_width [view_id]-xPb-nPbW + 1, mvLX [0])
  • mvLX [0] means the x direction component of the motion vector
  • view_position_x [view_id] and view_width [view_id] mean the start (upper left) x coordinate and width of the view image having view_id, respectively.
  • the said padding process means the process which clips the reference position of interpolation image generation in the utilization range.
  • the range (usable range) to which reference can be made when the predicted image generation unit 308 performs the predicted image generation process of the view i is as follows.
  • xMax min (pic_width_in_luma_samples-1, view_position_x [i] + view_width [i] -1)
  • yMax min (pic_height_in_luma_samples-1, view_position_y [i] + view_height [i] -1)
  • xMin and xMax are variables representing the reference restriction range (use range) in the x direction (horizontal direction)
  • yMin and yMax are
  • Y direction (vertical direction) is a variable representing the reference limit range (use range).
  • pic_width_in_luma_samples is a syntax which shows the width
  • pic_height_in_luma_samples is a syntax which shows the height of a brightness
  • the integer part of the motion vector mvLX is derived as xInt, yInt, and the following reference location xRef , yRef pixels are used.
  • xPb, yPb are upper left coordinates of PU (Prediction Unit)
  • mvshift is a constant indicating that the motion vector precision is 1 / (2 ⁇ mvshift) pel
  • xL, yL are within PU
  • the processing of replacing the reference position within the use range using the Clip 3 function is the substance of the padding processing.
  • FIG. 40 is a diagram showing an example of padding processing according to the present example.
  • motion image generation interpolated image generation
  • the predicted image generation unit 308 prohibits the reference-prohibited region as shown in a picture 61.
  • the blocks in the area are padded with the blocks in the area (white area) where reference is possible.
  • the image decoding apparatus 31 is the image decoding apparatus 31 that decodes encoded data of an image, and includes a predicted image generation unit 308 that generates a predicted image, and the predicted image generation unit 308
  • the motion vector is clipped (changed) within the usage range.
  • the reference block Restrict references to information on reference pictures included in a given view.
  • the predicted image generation unit 308 when performing inter prediction in the predicted image generation process, has an effect of preventing the reference region from being referred to.
  • each member included in the image decoding device 31 has a function of performing the processing described in the present example below.
  • the predicted image generation unit 308 included in the image decoding device 31 can generate a predicted image of a target block using intra block copying.
  • Intra block copying is a technique for generating a target block by referring to a reference block included in an already decoded area in a picture to which the target block to be generated belongs. Generating a predicted image using intra block copy is called intra block copy prediction.
  • Intra block copy prediction uses a vector to specify a reference block, as in inter prediction that refers to a picture different from the target block. Since intra block copying (IBC) can be realized by designating a target picture itself as a reference picture, it is also referred to as current picture reference (CPR).
  • IBC intra block copying
  • CPR current picture reference
  • FIG. 41 is a diagram illustrating an example of blocks to which the predicted image generation unit 308 can refer.
  • the view 0 is a decoded view image
  • the view 1 is a view image during decoding.
  • blocks to which the prediction image can be referenced are limited to reference blocks included in view 0.
  • the reference block 63b can be referred to but the reference to the block 63c is prohibited.
  • the predicted image generation unit 308 restricts the reference block to the reference block included in the decoded view.
  • the area to which the predicted image generation unit 308 refers is limited, and pre-reading becomes easy.
  • each member included in the image decoding device 31 has a function of performing the processing described in the present example below.
  • FIG. 42 is a diagram illustrating an example of blocks to which the predicted image generation unit 308 can refer.
  • a block that can be referred to by the predicted image generation unit 308 is a reference block included in view 0 and the target block It limits to the reference block which exists in a corresponding position, and the reference block located in the predetermined range centering on the said reference block.
  • the reference block corresponding to the target block 64a on the view 0 and the reference block are centered
  • the area 64b of the predetermined range can be referred to, but the block 64c on the view 0 and the block 64d on the view 1 are forbidden to refer to.
  • the area 64 b is a reference block in a view having a view_id smaller than the view in which the target pixel is present, and means within a predetermined range centered on the reference block present at a position corresponding to the target block. .
  • MVRange2 is a variable that indicates the size of the range that allows reference.
  • the above clipping is a reference block in view 0 that is view_id smaller than the view in which the target pixel is present, and the upper left position of the reference block present at the position corresponding to the target block is (0, -view_height [1])
  • the lower right position is limited within a range centered on (nPbW, -view_height [1] + nPbH).
  • MVrange2 Clip3 (xCtr-MVRange2, xCtr + MVRange2-1, mvLX [0])
  • mvLX [1] Clip3 (yCtr-MVRange2, yCtr + MVRange2-1, mvLX [1])
  • MVrange2U, MVrange2D, MVrange2L, and MVRange2R the same range MVrange2 is used in the upper, lower, left, and right directions, but different ranges may be used in upper, lower, left, and right (for example, MVrange2U, MVrange2D, MVrange2L, and MVRange2R).
  • the predicted image generation unit 308 sets the motion vector to a reference block in a view having a view_id smaller than the view in which the target pixel is present, and centers the reference block existing at a position corresponding to the target block.
  • the motion vector is restricted so that the reference pixel of the motion vector is located within the predetermined range.
  • the motion vector points to the area where the reference is prohibited
  • the above area is replaced with the area where the reference is not prohibited (in the use range), that is, padding It may be configured to perform processing.
  • pixels of the following reference positions xRef and yRef are used.
  • xMin max (0, xPb-MVRange2)
  • yMin max (0, yPb-view_height [1] -MVRange2)
  • xMax min (pic_width_in_luma_samples-1, xPb + nPbW + MVRange2-1)
  • yMax min (pic_height_in_luma_samples-1, yPb-view_height [1] + nPbH + MVRange2-1)
  • xRef Clip3 (xMin, xMax, xInt + i)
  • yRef Clip 3 (yMin, yMax, yInt + j)
  • the processing of replacing the reference position within the use range using the Clip 3 function is the substance of
  • the predicted image generation unit 308 sets the reference pixel as a reference block in a view having a view_id smaller than the view in which the target pixel exists, and centers the reference block existing at a position corresponding to the target block. And limit the reference pixels located within the predetermined range.
  • the area to be referred to by the predicted image generation unit 308 is limited to facilitate prefetching, thereby suppressing an increase in memory access.
  • each member included in the image decoding device 31 has a function of performing the processing described in the present example below.
  • FIG. 43 is a diagram illustrating an example of blocks to which the predicted image generation unit 308 can refer.
  • a block to which the prediction image can be referred is intra
  • the area is limited to the area described in the reference example 2 according to block copy and the reference block belonging to the same coding tree block (CTU) line of the same view as the target block.
  • CTU coding tree block
  • FIG. 44 is a diagram illustrating an example of motion vector clipping processing and padding processing according to this example.
  • xCtb is the upper left x coordinate of the CTU to which the target block belongs
  • yCtb is the upper left y coordinate of the CTU
  • ctb_height is the CTU width and height of view i
  • view_width [ i] mean the width and height of the view i, respectively.
  • (A) is a syntax table which shows an example of the clipping process of a motion vector based on this example.
  • the block may be replaced with the area where the reference is not prohibited, that is, the padding process may be performed.
  • (B) is a syntax table which shows an example of the padding process based on this example.
  • a reference pixel which is a reference destination pixel at a position shifted by i, j from a target pixel which is a reference source pixel
  • pixels of reference positions xRef and yRef shown in (b) are used.
  • the predicted image generation unit 308 is the same as the target block, which is the reference block belonging to the area described in the reference example 2 related to intra block copying and the view in which the target block exists. Restrict to the reference block belonging to the coding tree block line.
  • the predicted image generation unit 308 since the predicted image generation unit 308 refers to the reference block that can be stored in the line memory, the predicted image generation unit 308 has an effect of suppressing the delay of the decoding process.
  • the image decoding device 31 is an image decoding device 31 that decodes encoded data of an image, and the intra block copy that refers to a block in the same view
  • the prediction image generation unit 308 is used to generate a prediction image by using the prediction image generation unit 308.
  • the prediction image generation unit 308 generates a prediction image of a target block by referring to a reference block to be referred to in the prediction image generation processing.
  • the area to be referred to by the predicted image generation unit 308 is limited to facilitate prefetching, thereby suppressing an increase in memory access.
  • Example of restriction of reference block (reference restriction of tile image)
  • the predicted image generation unit 308 restricts tile images to be referred to in the predicted image generation process of generating a predicted image of a target block.
  • the configuration shown in the functional block diagram of FIG. 13 is used.
  • symbol is appended about the member which has the same function as the member mentioned above, and description is abbreviate
  • the predicted image generation unit 308 of the picture decoding unit 32 of the image decoding apparatus 31 performs step S122. It is determined whether or not the inter-tile reference information indicates that the second tile can be referred to. If the inter-tile reference information indicates that the second tile can be referred to, the current picture is added to the reference picture list to generate a reference picture list including the second tile that can be referred to.
  • the predicted image generation unit 308 (also serving as a clipping unit) is used to limit reference to tiles that can not be referred to when referring to the current picture.
  • the predicted image generation unit 308 generates a predicted image of the target block of the first tile with reference to the range of the second tile by replacing the reference position (the above-described intra block copy and the like).
  • FIG. 47 is a diagram showing an example of replacing a reference position when generating a predicted image of a tile image.
  • the inter-tile reference availability information in the example shown in FIG. 47 is shown below.
  • the inter-tile reference information indicates that tile 0 (second tile) can be referenced when predicting the target block of tile 2 (first tile). Therefore, the predicted image generation unit 308 replaces the reference position with the position in the referenceable range (bold frame in FIG. 47) before performing step S123 described above.
  • the predicted image generation unit 308 generates a predicted image of the target block of tile 2 with reference to the pixels of tile 0 that can be referred to.
  • the following is an expression (the above-described Clip 3 function) of reference position setting by the predicted image generation unit 308 in this specific example.
  • x Clip 3 (xTs, xTs + wT-1-blkW, x)
  • y Clip 3 (yTs, yTs + hT-1-blkH, y)
  • the meaning of each symbol is as follows.
  • NumTile (num_tile_columns_minus1 + 1) * (num_tile_rows_minus1 + 1)
  • the predicted image generation unit 308 sets the range of the reference position in the range of the second tile each time it determines that the inter-tile reference information indicates that the second tile can be referred to. . Then, in the image coding device 11, motion vector search may be performed in the referenceable range.
  • the range of loop variable j of the tile may be 0..NumTile-1 instead of 0..currTileOrderId-1.
  • the variation distTmp of the reference position in the case of limiting the reference range to each tile j is derived, and the reference position (bestX, bestY) in the case where the variation is the minimum is used as the final reference position (x, y And).
  • the amount of change of the reference position may use the following square distance or the like regardless of the absolute value distance.
  • distTmp
  • ⁇ 2 indicates the square.
  • the clipping unit clips the reference position to the range of the second tile (prediction (prediction)
  • the predicted image generation unit further refers to the range of the second tile clipped by the clipping unit, and the predicted image generation unit (predicted image generation unit 308) corresponds to the predicted image of the first tile.
  • the current picture As in the method of adding the current picture to the reference picture list described later, if there is a dependency between the current picture tiles, the current picture (currPic) is added to the reference picture list.
  • the current picture (currPic) is added to the reference picture list.
  • Tiles that are not dependent may not be decodable because they may not be held in reference memory.
  • decoding is performed also on the memory band, a load more than necessary is generated. Therefore, according to the configuration according to the above specific example, since the pixels of the tile that can be referred to can be referred to in advance, the tile having no dependency relationship is not erroneously referred to, and the image can be reliably obtained under a limited load. Can be decoded.
  • the predicted image generation unit 308 may or may not refer to between tiles
  • the pixels in the region where reference can be made are referred to instead.
  • reference is made to pixels of referenceable tiles.
  • FIG. 48 is a diagram illustrating an example of pixel reference processing by the predicted image generation unit 308 according to the present specific example.
  • the tile 0 corresponds to a second tile indicating that inter-tile reference information can be referred to.
  • the tile 1 and the tile 2 correspond to a second tile indicating that the inter-tile reference information can not be referred to.
  • the predicted image generation unit 308 performs pixel reference processing such as referring to the pixels of the referenceable area when referring to the area where the reference is prohibited by the inter-tile reference information. For example, reference is made to tile 0 which is a tile of a referenceable area. Then, the predicted image generation unit 308 stores the current picture currPic in the reference picture list stored in the reference picture memory 306 (the details of the reference picture list in the present embodiment will be described later).
  • the prediction image generation unit 308 when predicting a target block of tile 5 (first tile), the prediction image generation unit 308 (also serving as a replacement unit) If it indicates that the block of 2 or 3 (horizontal stripe tile) (second tile) can not be referred to, the pixel value of tile 1, 2 or 3 is replaced with a predetermined value (for example 128). Then, the predicted image generation unit 308 further generates a predicted image of the target block of the block 5 with reference to the reference block of the tile 1, 2 or 3 (horizontal stripe tile) in which the pixel values are replaced (more specifically, Before the step, store the current picture in the reference picture list).
  • blocks of tiles that are not dependent can be treated as reference blocks, and reference restrictions can be avoided. Therefore, the view image can be efficiently decoded.
  • the prediction image generation unit 308 determines whether the inter-tile reference availability information is tile 1, 2 or 3 ( A tile indicating that another inter-tile reference information can refer to the second tile instead of the pixel value of tile 1, 2 or 3 if it indicates that the block of the second tile can not be referred to The predicted image of the target block is further generated with reference to pixel values of 0 or 4 (another second tile) (more specifically, the current picture is stored in the reference picture list before the process).
  • FIG. (b) of FIG. 49 will be described in more detail with reference to FIG. (A) to (c) of FIG. 50 are diagrams showing more specific examples of pixel reference processing by the predicted image generation unit 308 according to the present specific example.
  • the predicted image generation unit 308 can refer to the pixel position currently being processed.
  • the pixel position of any tile (tile 0 (another second tile)) is referenced to search for the vicinity of that pixel.
  • the predicted image generation unit 308 may pad around the tile (tile 0 in the figure).
  • the predicted image generation unit 308 when predicting the target block of the tile 5 (first tile), the predicted image generation unit 308 performs a reference corresponding to the pixel position currently being processed.
  • the pixel position of any possible tile (tile 4 (another second tile)) is referenced to search for the vicinity of that pixel.
  • the predicted image generation unit 308 may pad around the tile (the tile 4 in the figure).
  • the predicted image generation unit 308 transmits the current picture currPic stored in the reference picture list. Prepare a memory separately from the memory of the current picture currPic currently being processed, and replace it with a copy of a tile (tile 0 or 4) to which reference can not be made (tile 1, 2, 3 or 5) May search near the current pixel.
  • the predicted image generation unit 308 when predicting the target block of the tile 5 (first tile), the predicted image generation unit 308 (also serving as the replacement unit) If the inter-tile reference information indicates that the block of tile 1, 2 or 3 (the second tile) can not be referred to, another inter-tile reference information indicates the pixel value of tile 1, 2 or 3 Replace with the average value of pixel values of tile 0 or 4 (another second tile) indicating that the reference block can be referenced. Then, the predicted image generation unit 308 further generates a predicted image of the target block with reference to the reference block of the tile 1, 2 or 3 in which the pixel value is replaced (more specifically, before the step, the reference picture Store the current picture in the list)
  • FIG. (c) of FIG. 49 An example shown in (c) of FIG. 49 will be described in more detail with reference to FIG. (A) and (b) of FIG. 51 is a figure which shows the more concrete example of the pixel reference process by the estimated image production
  • the predicted image generation unit 308 can refer to the pixel position currently being processed. Any tile (tile 0) may be replaced with the average of tiles 0 and 4 and then referenced to search around that pixel. At this time, the predicted image generation unit 308 pads the periphery of the tile (tile 0). Further, in the example illustrated in (b) of FIG. 51, for example, when predicting the target block of the tile 5 (first tile), the predicted image generation unit 308 transmits the current picture currPic stored in the reference picture list. Prepare a memory separately from the currPic memory currently being processed, and use the average of tiles (tiles 0 and 4) that can refer to non-referenceable tiles (tiles 1, 2, 3 and 5). May be used for the reference picture.
  • the image encoding device 11 and the image decoding device 31 described above can be mounted and used in various devices that transmit, receive, record, and reproduce moving images.
  • the moving image may be a natural moving image captured by a camera or the like, or an artificial moving image (including CG and GUI) generated by a computer or the like.
  • FIG. 45 is a block diagram showing a configuration of a transmission device PROD_A on which the image coding device 11 is mounted.
  • the transmission device PROD_A modulates the carrier wave with the coding unit PROD_A1 for obtaining the coding data by coding the moving image, and the coding data obtained by the coding unit PROD_A1.
  • the image coding apparatus 11 described above is used as the coding unit PROD_A1.
  • the transmission device PROD_A is a camera PROD_A4 for capturing a moving image, a recording medium PROD_A5 for recording the moving image, an input terminal PROD_A6 for externally inputting the moving image, and a transmission source of the moving image input to the encoding unit PROD_A1. , And may further include an image processing unit A7 that generates or processes an image. In (a) of FIG. 45, although the configuration in which the transmission device PROD_A includes all of these is illustrated, a part of the configuration may be omitted.
  • the recording medium PROD_A5 may be a recording of a non-coded moving image, or a moving image encoded by a recording encoding method different from the transmission encoding method. It may be one. In the latter case, it is preferable to interpose, between the recording medium PROD_A5 and the encoding unit PROD_A1, a decoding unit (not shown) that decodes the encoded data read from the recording medium PROD_A5 according to the encoding scheme for recording.
  • FIG. 45 is a block diagram showing a configuration of a reception device PROD_B on which the image decoding device 31 is mounted.
  • the receiver PROD_B demodulates the modulated signal received by the receiver PROD_B1, which receives the modulated signal, and the demodulator PROD_B2, which obtains encoded data by demodulating the modulated signal received by the receiver PROD_B1, and And a decoding unit PROD_B3 for obtaining a moving image by decoding encoded data obtained by the unit PROD_B2.
  • the image decoding device 31 described above is used as the decoding unit PROD_B3.
  • the receiving device PROD_B is a display PROD_B4 for displaying a moving image, a recording medium PROD_B5 for recording the moving image, and an output terminal for outputting the moving image to the outside as a supply destination of the moving image output by the decoding unit PROD_B3. It may further comprise PROD_B6. Although (b) of FIG. 45 exemplifies a configuration in which the reception device PROD_B includes all of them, a part may be omitted.
  • the recording medium PROD_B5 may be for recording a moving image which has not been encoded, or is encoded by a recording encoding method different from the transmission encoding method. May be In the latter case, an encoding unit (not shown) may be interposed between the decoding unit PROD_B3 and the recording medium PROD_B5 to encode the moving image acquired from the decoding unit PROD_B3 according to the encoding method for recording.
  • the transmission medium for transmitting the modulation signal may be wireless or wired.
  • the transmission mode for transmitting the modulation signal may be broadcast (here, a transmission mode in which the transmission destination is not specified in advance), or communication (in this case, transmission in which the transmission destination is specified in advance) (Refer to an aspect). That is, transmission of the modulation signal may be realized by any of wireless broadcast, wired broadcast, wireless communication, and wired communication.
  • a broadcasting station (broadcasting facility etc.) / Receiving station (television receiver etc.) of terrestrial digital broadcasting is an example of a transmitting device PROD_A / receiving device PROD_B which transmits and receives a modulated signal by wireless broadcasting.
  • a cable television broadcast station (broadcasting facility or the like) / receiving station (television receiver or the like) is an example of a transmitting device PROD_A / receiving device PROD_B which transmits and receives a modulated signal by cable broadcasting.
  • a server such as a workstation
  • client such as a VOD (Video On Demand) service or a video sharing service using the Internet
  • PROD_A / receiving device PROD_B
  • the personal computer includes a desktop PC, a laptop PC, and a tablet PC.
  • the smartphone also includes a multifunctional mobile phone terminal.
  • the client of the moving image sharing service has a function of encoding a moving image captured by a camera and uploading it to the server. That is, the client of the moving image sharing service functions as both the transmitting device PROD_A and the receiving device PROD_B.
  • FIG. 46 (a) is a block diagram showing a configuration of a recording device PROD_C on which the image coding device 11 described above is mounted.
  • the recording device PROD_C uses the encoding unit PROD_C1, which obtains encoded data by encoding a moving image, and the encoded data obtained by the encoding unit PROD_C1, to the recording medium PROD_M.
  • a writing unit PROD_C2 for writing.
  • the image coding device 11 described above is used as the coding unit PROD_C1.
  • the recording medium PROD_M may be (1) a type incorporated in the recording device PROD_C, such as a hard disk drive (HDD) or a solid state drive (SSD), or (2) an SD memory. It may be of a type connected to the recording device PROD_C, such as a card or a Universal Serial Bus (USB) flash memory, or (3) a DVD (Digital Versatile Disc) or a BD (Blu-ray Disc: Registration It may be loaded into a drive device (not shown) built in the recording device PROD_C, such as a trademark).
  • a type incorporated in the recording device PROD_C such as a hard disk drive (HDD) or a solid state drive (SSD), or (2) an SD memory. It may be of a type connected to the recording device PROD_C, such as a card or a Universal Serial Bus (USB) flash memory, or (3) a DVD (Digital Versatile Disc) or a BD (Blu-ray Disc: Registration It may be loaded into
  • the recording device PROD_C is a camera PROD_C3 for capturing a moving image as a supply source of the moving image input to the encoding unit PROD_C1, an input terminal PROD_C4 for inputting the moving image from the outside, and a reception for receiving the moving image
  • the image processing unit PROD_C5 may further include an image processing unit PROD_C6 that generates or processes an image.
  • FIG. 46 exemplifies the configuration in which the recording apparatus PROD_C includes all of them, a part of the configuration may be omitted.
  • the receiving unit PROD_C5 may receive an uncoded moving image, and receives encoded data encoded by a transmission encoding scheme different from the recording encoding scheme. It may be In the latter case, it is preferable to interpose a transmission decoding unit (not shown) that decodes encoded data encoded by the transmission encoding scheme between the reception unit PROD_C5 and the encoding unit PROD_C1.
  • Examples of such a recording device PROD_C include a DVD recorder, a BD recorder, an HDD (Hard Disk Drive) recorder, etc.
  • the input terminal PROD_C4 or the receiving unit PROD_C5 is a main supply source of moving images).
  • a camcorder in this case, the camera PROD_C3 is the main supply source of moving images
  • a personal computer in this case, the receiving unit PROD_C5 or the image processing unit C6 is the main supply source of moving images
  • a smartphone this In this case, the camera PROD_C3 or the receiving unit PROD_C5 is a main supply source of moving images
  • the like are also examples of such a recording device PROD_C.
  • FIG. 46 is a block showing the configuration of the playback device PROD_D on which the image decoding device 31 described above is mounted.
  • the playback device PROD_D decodes the moving image by decoding the encoded data read by the reading unit PROD_D1 that reads the encoded data written to the recording medium PROD_M and the reading unit PROD_D1. And a decryption unit PROD_D2 to be obtained.
  • the image decoding device 31 described above is used as the decoding unit PROD_D2.
  • the recording medium PROD_M may be (1) a type incorporated in the playback device PROD_D such as an HDD or an SSD, or (2) such as an SD memory card or a USB flash memory. It may be of a type connected to the playback device PROD_D, or (3) it may be loaded into a drive device (not shown) built in the playback device PROD_D, such as DVD or BD. Good.
  • the playback device PROD_D is a display PROD_D3 that displays a moving image as a supply destination of the moving image output by the decoding unit PROD_D2, an output terminal PROD_D4 that outputs the moving image to the outside, and a transmission unit that transmits the moving image. It may further comprise PROD_D5. Although (b) of FIG. 46 exemplifies a configuration in which the playback device PROD_D includes all of these, a part may be omitted.
  • the transmission unit PROD_D5 may transmit a non-encoded moving image, or transmit encoded data encoded by a transmission encoding method different from the recording encoding method. It may be In the latter case, an encoding unit (not shown) may be interposed between the decoding unit PROD_D2 and the transmission unit PROD_D5 for encoding moving pictures according to a transmission encoding scheme.
  • a playback device PROD_D for example, a DVD player, a BD player, an HDD player, etc. may be mentioned (in this case, the output terminal PROD_D4 to which a television receiver etc. is connected is the main supply destination of moving images) .
  • the display PROD_D3 is the main supply destination of moving images
  • digital signage also referred to as an electronic signboard or electronic bulletin board, etc.
  • the display PROD_D3 or the transmission unit PROD_D5 is the main supply of moving images.
  • desktop type PC in this case, output terminal PROD_D4 or transmission unit PROD_D5 is the main supply destination of moving images
  • laptop type or tablet type PC in this case, display PROD_D3 or transmission unit PROD_D5 is moving image
  • the main supply destination of the image the smartphone (in this case, the display PROD_D3 or the transmission unit PROD_D5 is the main supply destination of the moving image), and the like are also examples of such a reproduction device PROD_D.
  • the method of restricting the tile to be referred to when the image decoding device 31 according to the present embodiment generates a prediction image as described above based on the inter-tile referenceability information does not use the image decoding device 31. Even if it is, it is realizable by the bit stream restriction
  • the image encoding device 11 is an image encoding device that encodes a tile that constitutes a view, and when predicting a target block of a first tile among the tiles, the image encoding device 11
  • the inter-tile reference information encoding unit (header information encoding unit) which encodes inter-tile reference information indicating whether or not the second tile can be referred to, and the inter-tile reference information is the second
  • a predicted image generation unit (predicted image generation unit 101) that generates a predicted image of the target block with reference to the second tile, and indicates that the tile can be referred to.
  • the current picture is a reference picture, it refers only to the information on the first tile and the information on the tile indicating that the inter-tile reference information can be referred to. Generating a predicted image for each target block of the tile.
  • the predicted image is not referred to the data (pixels and coding parameters) of tiles other than the first tile to be processed. Can be generated. Therefore, the finally encoded encoded stream Te includes encoded data based on the predicted image, and the tile that can be referred to and the tile to be processed by decoding the encoded data by the image decoding device.
  • the decoded image generated can be generated with reference to only.
  • the image decoding apparatus 31 generates a predicted image of the target block (the block of the first tile) by referring to the block of the reference tile (the second tile) designated by the index in the inter-tile reference information Do. At this time, the image decoding device 31 reads a reference tile from the reference picture list stored in the reference picture memory 306.
  • generation of the reference picture list will be described with reference to FIG.
  • FIG. 52 is a flowchart illustrating generation of a reference picture list according to the present embodiment.
  • step S140 the header information decoding unit 19 of the image decoding device 31 decodes the above-described tile order index (tile decoding order index) that specifies the decoding order of a plurality of tiles.
  • step S141 the above-described inter-tile reference information indicating whether or not the header information decoding unit 19 of the image decoding device 31 can refer to the second tile when predicting the target block of the first tile.
  • step S142 the predicted image generation unit 308 (also serving as the determination unit) refers to the inter-tile referenceability information to determine whether the current picture including the first tile can refer to the second tile. Determine If the predicted image generation unit 308 determines that there is a tile that can refer to the current picture, the process proceeds to step S143. If the predicted image generation unit 308 determines that there is no tile to which the current picture can refer, the process ends.
  • step S143 the prediction image generation unit 308 adds the current picture to the reference picture list stored in the reference picture memory 306.
  • the image decoding apparatus 31 is an image decoding apparatus that decodes a plurality of tiles forming a view, and predicts a target block of a first tile among the plurality of tiles.
  • the inter-tile reference information decoding unit decodes the inter-tile reference information indicating whether or not the second tile of the plurality of tiles can be referred to, and the inter-tile reference information
  • a determination unit that determines whether to add the current picture to the reference picture list, and the addition unit adds the current picture to the reference picture list when the determination unit determines that the current picture is to be added to the reference picture list It has a department.
  • the inter-tile reference information by referring to the inter-tile reference information, it is possible to easily identify a referenceable picture, and to add the picture to the reference picture list.
  • the predicted image generation unit 308 (also serving as the determination unit) indicates that the inter-tile referenceability information refers to any tile j when generating the predicted image of the current tile currTileID in step S142 described above. If it is determined that there is, the current picture currPic is added to the reference picture list RefPicListTemp0 in step S143 described above.
  • the formula of the said structure is shown below.
  • the predicted image generation unit 308 adds the values of the inter-tile referenceability information tile_dependency_flag (inter-tile referenceability information) decoded by the header information decoding unit 19 to calculate the value of curr_tile_dependency_flag. If the value of curr_tile_dependency_flag is 1 or more, the first tile refers to any tile in the current picture as a second tile. Then, the predicted image generation unit 308 (also serving as the determination unit) derives a reference picture list represented by the following equation.
  • RefPicListTemp0 [list_entry_l0 [rIdx]]: RefPicListTemp0 [rIdx]
  • num_ref_idx_I0_active_minus1 indicates the number of active reference pictures-1
  • ref_pic_list_modification_flag_I0 indicates a flag indicating whether the reference picture list order is to be exchanged
  • list_entry_l0 [rIdx] indicates the order in the case of changing the reference picture list order. Furthermore, it is an index. Furthermore, the following process of adding the current picture to RefPicList0 may be performed. if (pps_curr_pic_ref_enabled_flag
  • the predicted image generation unit 308 may derive the reference picture list RefPicList1 of L1 in the above-described step S143, similarly to RefPicListTemp0. The following shows the formula for deriving the reference picture list.
  • RefPicListTemp1 [list_entry_l1 [rIdx]]: RefPicListTemp1 [rIdx] ⁇
  • the predicted image generation unit 308 adds the current picture currPic to the reference picture list if pps_curr_pic_ref_enabled_flag indicates 1 or if the final curr_tile_dependency_flag indicates a value of 1 or more.
  • the determination unit is configured such that, among the inter-tile reference information, at least one or more inter-tile reference information is the second information. It is determined that the current picture is to be added to the reference picture list if it indicates that the tile of can be referenced.
  • the original picture can be appropriately added to the reference picture list by referring to the inter-tile reference information. Therefore, when generating a predicted image, it is possible to refer to tiles that can be referred to by referring to the reference picture list. Thus, the view image can be efficiently decoded.
  • the predicted image generation unit 308 stores the current picture currPic at the top of the reference picture list in step S143 described above. You may The following shows an expression for deriving the reference picture list RefPicListTemp0 (RefPicList0) by the predicted image generation unit 308.
  • the prediction image generation unit 308 stores the current picture currPic at the top of the reference picture list.
  • RefPicListTemp1 RefPicList1 (RefPicList1) by the predicted image generation unit 308.
  • rIdx 0 while (rIdx ⁇ NumRpsCurrTempList1) ⁇ if (curr_tile_dependency_flag)
  • the adding unit adds the current picture to the top of the reference picture list.
  • the inter-tile reference information can be referred to, and if there is a tile that can be referred to, the current picture can be added to the top of the reference picture list.
  • the reference picture list Similar scenes exist in the current picture currPic in the multiview, so that coding efficiency can be improved by preferentially using it as a reference picture.
  • each block of the image decoding device 31 and the image encoding device 11 described above may be realized as hardware by a logic circuit formed on an integrated circuit (IC chip), or a CPU (Central Processing Unit) It may be realized as software using
  • each of the devices described above includes a CPU that executes instructions of a program that implements each function, a ROM (Read Only Memory) that stores the program, a RAM (Random Access Memory) that expands the program, the program, and various other methods.
  • a storage device such as a memory for storing data is provided.
  • the object of the embodiment of the present invention is to record computer program readable program codes (execution format program, intermediate code program, source program) of control programs of the above-mentioned respective devices which are software for realizing the functions described above.
  • the present invention can also be achieved by supplying a medium to each of the above-described devices, and a computer (or a CPU or an MPU) reading and executing a program code recorded on a recording medium.
  • Examples of the recording medium include tapes such as magnetic tapes and cassette tapes, magnetic disks such as floppy (registered trademark) disks / hard disks, CDs (Compact Disc Read-Only Memory) / MO disks (Magneto-Optical disc).
  • tapes such as magnetic tapes and cassette tapes
  • magnetic disks such as floppy (registered trademark) disks / hard disks
  • CDs Compact Disc Read-Only Memory
  • MO disks Magnetic-Optical disc
  • Disks including optical disks such as MD (Mini Disc) / DVD (Digital Versatile Disc) / CD-R (CD Recordable) / Blu-ray Disc (registered trademark), IC cards (including memory cards) Cards such as optical cards, mask ROMs / erasable programmable read-only memories (EPROMs) / electrically erasable and programmable read-only memories (EEPROMs) / semiconductor memories such as flash ROMs, or programmable logic devices (PLDs) And logic circuits such as FPGA (Field Programmable Gate Array) can be used.
  • MD Mini Disc
  • DVD Digital Versatile Disc
  • CD-R Compact Disc
  • Blu-ray Disc registered trademark
  • IC cards including memory cards
  • Cards such as optical cards
  • EPROMs erasable programmable read-only memories
  • EEPROMs electrically erasable and programmable read-only memories
  • semiconductor memories such as flash ROMs, or programmable logic devices (PLD
  • each device may be configured to be connectable to a communication network, and the program code may be supplied via the communication network.
  • This communication network is not particularly limited as long as the program code can be transmitted.
  • the Internet intranet, extranet, LAN (Local Area Network), ISDN (Integrated Services Digital Network), VAN (Value-Added Network), CATV (Community Antenna television / Cable Television) communication network, virtual private network (Virtual Private) Network), telephone network, mobile communication network, satellite communication network, etc.
  • the transmission medium that constitutes this communication network may be any medium that can transmit the program code, and is not limited to a specific configuration or type.
  • the embodiment of the present invention may also be realized in the form of a computer data signal embedded in a carrier wave, in which the program code is embodied by electronic transmission.
  • Embodiments of the present invention are not limited to the above-described embodiments, and various modifications are possible within the scope of the claims. That is, an embodiment obtained by combining technical means appropriately modified within the scope of the claims is also included in the technical scope of the present invention.
  • ⁇ Reference to related documents> The present application relates to priority claims based on Japanese Patent Applications (Japanese Patent Application No. 2017-250224 (filed on December 26, 2017)) and (Japanese Patent Application No. 2018-065877 (filed on March 29, 2018)). These documents are incorporated herein by reference.
  • An embodiment of the present invention is suitably applied to an image decoding apparatus that decodes encoded data obtained by encoding image data, and an image encoding apparatus that generates encoded data obtained by encoding image data. it can. Further, the present invention can be suitably applied to the data structure of encoded data generated by the image encoding device and referenced by the image decoding device.
  • Image Transmission System 9 Decoding Modules 10, 12, 13 Information Decoding Unit 11 Image Encoding Device 19 Header Information Decoding Unit (Tile Decoding Order Information Decoding Unit) 20, 22 Decoding Unit 21 Network 31 Image Decoding Device 32 Picture Decoding Unit 33 Picture Division Unit 41 Image Display Device 101, 308 Predicted Image Generation Unit 102, 1123 Subtraction Unit 103 Quantization Unit 104 Entropy Encoding Unit 105, 311 Inverse Converter 106, 312, 3038 Addition unit 107, 305 Loop filter 108, 307 Prediction parameter memory 109, 306 Reference picture memory Te Encoded stream

Landscapes

  • Engineering & Computer Science (AREA)
  • Multimedia (AREA)
  • Signal Processing (AREA)
  • Compression Or Coding Systems Of Tv Signals (AREA)

Abstract

L'invention concerne une caractéristique avec laquelle, lors du décodage d'une image de visualisation incluse dans une image à laquelle un bourrage de trame a été appliqué, il est possible de décoder efficacement l'image de visualisation. Un dispositif de décodage d'image pour décoder un flux codé est pourvu d'une unité de décodage d'informations d'en-tête et d'une unité de décodage d'image. L'unité de décodage d'informations d'en-tête décode des informations de bourrage de visualisation (visualisation_bourrage_drapeau, etc.) Inclus dans un ensemble de paramètres du flux codé, les informations de bourrage de visualisation indiquant si une pluralité de vues sont incluses dans une image à décoder ou non. L'unité de décodage d'image décode une image spécifiée par les informations de bourrage de visualisation en tant qu'image dans laquelle une pluralité de vues sont regroupées en trame dans la direction verticale et décode ainsi chacune des vues.
PCT/JP2018/047887 2017-12-26 2018-12-26 Dispositif de décodage d'image et dispositif de codage d'image WO2019131778A1 (fr)

Applications Claiming Priority (4)

Application Number Priority Date Filing Date Title
JP2017250224 2017-12-26
JP2017-250224 2017-12-26
JP2018-065877 2018-03-29
JP2018065877 2018-03-29

Publications (1)

Publication Number Publication Date
WO2019131778A1 true WO2019131778A1 (fr) 2019-07-04

Family

ID=67066440

Family Applications (1)

Application Number Title Priority Date Filing Date
PCT/JP2018/047887 WO2019131778A1 (fr) 2017-12-26 2018-12-26 Dispositif de décodage d'image et dispositif de codage d'image

Country Status (1)

Country Link
WO (1) WO2019131778A1 (fr)

Cited By (1)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US20210385437A1 (en) * 2019-02-02 2021-12-09 Beijing Bytedance Network Technology Co., Ltd. Prediction using extra-buffer samples for intra block copy in video coding

Citations (3)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
JP2012516114A (ja) * 2009-01-26 2012-07-12 トムソン ライセンシング ビデオ符号化のためのフレーム・パッキング
JP2013534782A (ja) * 2010-06-28 2013-09-05 シズベル テクノロジー エス.アール.エル. 2次元にも対応する立体映像フローのデコード方法
US20150098507A1 (en) * 2013-10-04 2015-04-09 Ati Technologies Ulc Motion estimation apparatus and method for multiview video

Patent Citations (3)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
JP2012516114A (ja) * 2009-01-26 2012-07-12 トムソン ライセンシング ビデオ符号化のためのフレーム・パッキング
JP2013534782A (ja) * 2010-06-28 2013-09-05 シズベル テクノロジー エス.アール.エル. 2次元にも対応する立体映像フローのデコード方法
US20150098507A1 (en) * 2013-10-04 2015-04-09 Ati Technologies Ulc Motion estimation apparatus and method for multiview video

Non-Patent Citations (4)

* Cited by examiner, † Cited by third party
Title
ALIREZA, AMINLOU ET AL.: "AHG8: Requirements and proposed method for viewport-adaptive quality assessment", JOINT VIDEO EXPLORATION TEAM (JVET) OF ITU-T SG 16 WP 3- JVET-E0107, no. JVET-E0107, 12 January 2017 (2017-01-12) - 20 January 2017 (2017-01-20), Geneva, CH, pages 1 - 11, XP030150604 *
JILL, BOYCE ET AL.: "AHG8: Subjective testing of 360° video projection/packing formats", JOINT VIDEO EXPLORATION TEAM (JVET) OF ITU-T SG 16 WP 3- JVET-F0021, no. JVET-F0021, 22 February 2017 (2017-02-22) - 7 April 2017 (2017-04-07), Hobart, AU, pages 1 - 12, XP030150668 *
R. SJOBERG , T. RUSERT: "High-Level Syntax for Bitstream Extraction", JOINT COLLABORATIVE TEAM ON VIDEO CODING (JCT-VC) OF ITU- T SG 16 WP3 JCTVC-G607, no. JCTVC-G607, 21 November 2011 (2011-11-21) - 30 November 2011 (2011-11-30), Geneva, CH, pages 1 - 14, XP030110591 *
YAN, YE ET AL.: "SEI message: independently decodable regions based on tiles", JOINT COLLABORATIVE TEAM ON VIDEO CODING (JCT-VC) OF ITU- T SG 16 WP 3- JCTVC-L0049, no. JCTVC-L0049, 14 January 2013 (2013-01-14) - 23 January 2013 (2013-01-23), Geneva, CH, pages 1 - 8, XP055623298 *

Cited By (2)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US20210385437A1 (en) * 2019-02-02 2021-12-09 Beijing Bytedance Network Technology Co., Ltd. Prediction using extra-buffer samples for intra block copy in video coding
US12003745B2 (en) 2019-02-02 2024-06-04 Beijing Bytedance Network Technology Co., Ltd Buffer updating for intra block copy in video coding

Similar Documents

Publication Publication Date Title
JP7223886B2 (ja) 画像復号方法
US11234011B2 (en) Image decoding apparatus and image coding apparatus
CN110574374B (zh) 图像解码装置
WO2018221368A1 (fr) Dispositif de décodage d'image animée et dispositif de codage d'image animée
JP2021010046A (ja) 画像符号化装置及び画像復号装置
US11722690B2 (en) Motion vector generation device, a prediction image generation device, a video decoding device and a video coding device
WO2018110203A1 (fr) Appareil de décodage d'images animées et appareil de codage d'images animées
WO2018116802A1 (fr) Dispositif de décodage d'images, dispositif de codage d'images, et dispositif de prédiction d'images
WO2019221072A1 (fr) Dispositif de codage d'images, dispositif d'extraction de flux codés et dispositif de décodage d'images
WO2019054300A1 (fr) Dispositif de codage d'image et dispositif de décodage d'image
JP7213689B2 (ja) 画像復号装置及び画像符号化装置
JPWO2018061522A1 (ja) 復号装置、符号化装置、復号方法、および、符号化方法
US20150334411A1 (en) Image decoding device, and image encoding device
WO2019131400A1 (fr) Dispositif de filtrage d'image, dispositif de décodage d'image et dispositif de codage d'image
CN108605138B (zh) 预测图像生成装置、运动图像解码装置、以及运动图像编码装置
JP7241153B2 (ja) 画像復号装置
CN112956206A (zh) 运动图像解码装置以及运动图像编码装置
WO2019230904A1 (fr) Dispositif de décodage d'image et dispositif de codage d'image
WO2019159820A1 (fr) Dispositif de décodage d'image mobile et dispositif de décodage d'image mobile
JP7238173B2 (ja) ビデオ/映像コーディングシステムにおけるシンタックスシグナリング方法及び装置
WO2019131349A1 (fr) Dispositif de décodage d'image et dispositif de codage d'image
WO2019131778A1 (fr) Dispositif de décodage d'image et dispositif de codage d'image
WO2018061550A1 (fr) Dispositif de décodage d'image et dispositif de codage d'image
JP2020036101A (ja) 画像復号装置および画像符号化装置
WO2019065537A1 (fr) Dispositif de filtre de compensation de mouvement, dispositif de décodage d'image et dispositif de codage d'image animée

Legal Events

Date Code Title Description
121 Ep: the epo has been informed by wipo that ep was designated in this application

Ref document number: 18896410

Country of ref document: EP

Kind code of ref document: A1

NENP Non-entry into the national phase

Ref country code: DE

122 Ep: pct application non-entry in european phase

Ref document number: 18896410

Country of ref document: EP

Kind code of ref document: A1

NENP Non-entry into the national phase

Ref country code: JP