WO2013137047A1 - 画像処理装置および方法 - Google Patents
画像処理装置および方法 Download PDFInfo
- Publication number
- WO2013137047A1 WO2013137047A1 PCT/JP2013/055884 JP2013055884W WO2013137047A1 WO 2013137047 A1 WO2013137047 A1 WO 2013137047A1 JP 2013055884 W JP2013055884 W JP 2013055884W WO 2013137047 A1 WO2013137047 A1 WO 2013137047A1
- Authority
- WO
- WIPO (PCT)
- Prior art keywords
- tile
- unit
- quantization parameter
- image
- initial value
- Prior art date
Links
Images
Classifications
-
- H—ELECTRICITY
- H04—ELECTRIC COMMUNICATION TECHNIQUE
- H04N—PICTORIAL COMMUNICATION, e.g. TELEVISION
- H04N19/00—Methods or arrangements for coding, decoding, compressing or decompressing digital video signals
- H04N19/70—Methods or arrangements for coding, decoding, compressing or decompressing digital video signals characterised by syntax aspects related to video coding, e.g. related to compression standards
-
- H—ELECTRICITY
- H04—ELECTRIC COMMUNICATION TECHNIQUE
- H04N—PICTORIAL COMMUNICATION, e.g. TELEVISION
- H04N19/00—Methods or arrangements for coding, decoding, compressing or decompressing digital video signals
- H04N19/10—Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using adaptive coding
- H04N19/102—Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using adaptive coding characterised by the element, parameter or selection affected or controlled by the adaptive coding
- H04N19/124—Quantisation
-
- H—ELECTRICITY
- H04—ELECTRIC COMMUNICATION TECHNIQUE
- H04N—PICTORIAL COMMUNICATION, e.g. TELEVISION
- H04N19/00—Methods or arrangements for coding, decoding, compressing or decompressing digital video signals
- H04N19/10—Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using adaptive coding
- H04N19/102—Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using adaptive coding characterised by the element, parameter or selection affected or controlled by the adaptive coding
- H04N19/124—Quantisation
- H04N19/126—Details of normalisation or weighting functions, e.g. normalisation matrices or variable uniform quantisers
-
- H—ELECTRICITY
- H04—ELECTRIC COMMUNICATION TECHNIQUE
- H04N—PICTORIAL COMMUNICATION, e.g. TELEVISION
- H04N19/00—Methods or arrangements for coding, decoding, compressing or decompressing digital video signals
- H04N19/44—Decoders specially adapted therefor, e.g. video decoders which are asymmetric with respect to the encoder
-
- H—ELECTRICITY
- H04—ELECTRIC COMMUNICATION TECHNIQUE
- H04N—PICTORIAL COMMUNICATION, e.g. TELEVISION
- H04N19/00—Methods or arrangements for coding, decoding, compressing or decompressing digital video signals
- H04N19/10—Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using adaptive coding
- H04N19/169—Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using adaptive coding characterised by the coding unit, i.e. the structural portion or semantic portion of the video signal being the object or the subject of the adaptive coding
- H04N19/17—Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using adaptive coding characterised by the coding unit, i.e. the structural portion or semantic portion of the video signal being the object or the subject of the adaptive coding the unit being an image region, e.g. an object
- H04N19/172—Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using adaptive coding characterised by the coding unit, i.e. the structural portion or semantic portion of the video signal being the object or the subject of the adaptive coding the unit being an image region, e.g. an object the region being a picture, frame or field
-
- H—ELECTRICITY
- H04—ELECTRIC COMMUNICATION TECHNIQUE
- H04N—PICTORIAL COMMUNICATION, e.g. TELEVISION
- H04N19/00—Methods or arrangements for coding, decoding, compressing or decompressing digital video signals
- H04N19/10—Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using adaptive coding
- H04N19/169—Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using adaptive coding characterised by the coding unit, i.e. the structural portion or semantic portion of the video signal being the object or the subject of the adaptive coding
- H04N19/17—Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using adaptive coding characterised by the coding unit, i.e. the structural portion or semantic portion of the video signal being the object or the subject of the adaptive coding the unit being an image region, e.g. an object
- H04N19/174—Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using adaptive coding characterised by the coding unit, i.e. the structural portion or semantic portion of the video signal being the object or the subject of the adaptive coding the unit being an image region, e.g. an object the region being a slice, e.g. a line of blocks or a group of blocks
-
- H—ELECTRICITY
- H04—ELECTRIC COMMUNICATION TECHNIQUE
- H04N—PICTORIAL COMMUNICATION, e.g. TELEVISION
- H04N19/00—Methods or arrangements for coding, decoding, compressing or decompressing digital video signals
- H04N19/10—Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using adaptive coding
- H04N19/169—Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using adaptive coding characterised by the coding unit, i.e. the structural portion or semantic portion of the video signal being the object or the subject of the adaptive coding
- H04N19/17—Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using adaptive coding characterised by the coding unit, i.e. the structural portion or semantic portion of the video signal being the object or the subject of the adaptive coding the unit being an image region, e.g. an object
- H04N19/176—Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using adaptive coding characterised by the coding unit, i.e. the structural portion or semantic portion of the video signal being the object or the subject of the adaptive coding the unit being an image region, e.g. an object the region being a block, e.g. a macroblock
Definitions
- the present disclosure relates to an image processing apparatus and method, and more particularly, to an image processing apparatus and method capable of efficiently performing decoding processing on a tiled picture.
- MPEG2 (ISO / IEC 13818-2) is defined as a general-purpose image encoding system, and is a standard that covers both interlaced scanning images and progressive scanning images, as well as standard resolution images and high-definition images.
- MPEG2 is currently widely used in a wide range of applications for professional and consumer applications.
- a code amount (bit rate) of 4 to 8 Mbps is assigned to an interlaced scanned image having a standard resolution of 720 ⁇ 480 pixels.
- a high resolution interlaced scanned image having 1920 ⁇ 1088 pixels is assigned a code amount (bit rate) of 18 to 22 Mbps.
- bit rate code amount
- MPEG2 was mainly intended for high-quality encoding suitable for broadcasting, but it did not support encoding methods with a lower code amount (bit rate) than MPEG1, that is, a higher compression rate. With the widespread use of mobile terminals, the need for such an encoding system is expected to increase in the future, and the MPEG4 encoding system has been standardized accordingly. Regarding the image coding system, the standard was approved as an international standard in December 1998 as ISO / IEC 14496-2.
- the standardization schedule is H.03 in March 2003. H.264 and MPEG-4 Part 10 (Advanced Video Coding, hereinafter referred to as AVC format).
- this AVC format extension includes RGB, 4: 2: 2, 4: 4: 4 coding tools required for business use, 8x8DCT and quantization matrix defined by MPEG-2.
- FRExt Full State Image Coding
- the macroblock size of 16 pixels ⁇ 16 pixels is optimal for large image frames such as UHD (Ultra High Definition: 4000 pixels ⁇ 2000 pixels) that are the targets of the next generation encoding method. There was no fear.
- HEVC High Efficiency Video Video Coding
- JCTVC Joint Collaboration Collaboration Team Video Coding
- a coding unit (Coding Unit) is defined as a processing unit similar to the macroblock in the AVC system.
- the CU is not fixed to a size of 16 ⁇ 16 pixels like the AVC macroblock, and is specified in the image compression information in each sequence.
- the CU is divided into prediction units (Prediction Unit (PU)) that are regions (partial regions of images in units of pictures) that are processing units of intra or inter prediction, and regions (pictures) that are processing units of orthogonal transformation
- PU Prediction Unit
- TU Transform Unit
- information on quantization parameters is sent to the decoding side.
- an initial quantization parameter for a picture is set in a picture parameter set (Picture parameter Set: PPS) in image compression information. At this time, it is set by which granularity (size) CU the quantization parameter is sent to the decoding side.
- PPS Picture parameter Set
- the initial quantization parameter in the slice is calculated on the decoding side.
- the difference between the initial quantization parameter in the slice and the quantization parameter in the CU is set at the head of the CU having a size set in the picture parameter set or larger. Thereby, the quantization parameter in the CU is calculated on the decoding side.
- a tile is defined as a tool for parallel processing.
- a tile is one of picture division units.
- the row (row) size and the column (column) size of each tile are specified in units of LCU in the sequence parameter set (Sequence Parameter Set: SPS) or the picture parameter set in the image compression information.
- SPS Sequence Parameter Set
- the LCU included in each tile is processed in the raster scan order, and the tile included in each picture is processed in the raster scan order.
- the slice may include a tile, and there may be a slice boundary in the middle of the tile.
- Non-Patent Document 2 proposes that information indicating where a tile starts is specified in the image compression information and transmitted to the decoding side.
- JCTVC-F335 Joint Collaborative Team on Video Coding (JCT-VC) of ITU-T SG16 WP3 andTCSC29 / WG11 VCEG) 6th Meeting: Torino, IT, 14-22 July, 2011 Kiran Misra, Andrew Segall, "New results for parallel decoding for Tiles", JCTVC-F594, Joint Collaborative Team on Video Coding (JCT-VC) of ITU-T SG16 WP3 and ISO / IEC JTC1 / SC29 / WG11th (VCEG) Meeting: Torino, IT, 14-22 July, 2011
- the slice may include a tile, and there may be a slice boundary in the middle of the tile.
- the picture is divided into two slices 0 and 1, Consider the case where the boundary of two slices 0 and 1 is in tile 1.
- the header of slice 1 near the boundary between slice 0 and 1 in the middle of tile 1 is transmitted to the decoding side. I have to access it.
- the tile 2 cannot be processed unless the processing of the slice header in the tile 1 is finished, which hinders parallel processing for each tile.
- the present disclosure has been made in view of such a situation, and can efficiently perform decoding processing on tile-divided pictures.
- An image processing apparatus includes a quantizer for a tile set to ensure independence of an encoded stream and a quantization parameter between tiles in a picture divided into a plurality of tiles of the image.
- a receiving unit for receiving an initial value of a quantization parameter, and a quantization parameter for reconstructing a quantization parameter of a coding unit constituting the tile using the initial value of the quantization parameter for the tile received by the receiving unit
- a reconstruction unit for reconstructing a decodes the encoded stream received by the reception unit using the quantization parameter reconstructed by the quantization parameter reconstruction unit and generates the image.
- the receiving unit can receive a value set at a predetermined position in the encoded stream as an initial value of the quantization parameter of the tile.
- the receiving unit can receive a value set in a picture parameter set in the encoded stream as an initial value of a quantization parameter of the tile.
- the receiving unit can receive a value set in the first slice header of the picture in the encoded stream as an initial value of the quantization parameter of the tile.
- the receiving unit can receive a value set near the top of the tile in the encoded stream as an initial value of the quantization parameter of the tile.
- the receiving unit can receive a value set in slice data in the encoded stream as an initial value of a quantization parameter of the tile.
- the receiving unit receives information on a quantization parameter of a coding unit constituting the tile
- the quantization parameter reconstructing unit is information on a quantization parameter of a coding unit constituting the tile received by the receiving unit. Can be used to reconstruct the quantization parameters of the coding units that make up the tile.
- An image processing method is configured such that an image processing apparatus ensures independence of processing related to quantization parameters between tiles in an encoded stream and a picture divided into a plurality of tiles of an image.
- the initial value of the quantization parameter for the tile that has been received, and the received initial value of the quantization parameter for the tile is used to reconstruct and reconstruct the quantization parameter of the coding unit that constitutes the tile
- the received encoded stream is decoded using the quantization parameter to generate the image.
- An image processing apparatus includes an encoding unit that encodes an image to generate an encoded stream, and a process related to a quantization parameter between tiles in a picture divided into a plurality of tiles of the image An initial value setting unit that sets an initial value of a quantization parameter for the tile, an initial value of a quantization parameter for the tile set by the initial value setting unit, and the encoding unit And a transmission unit for transmitting the encoded stream generated by.
- the initial value setting unit can set the initial value of the quantization parameter of the tile at a predetermined position in the encoded stream.
- the initial value setting unit can set the initial value of the quantization parameter of the tile in the picture parameter set in the encoded stream.
- the initial value setting unit can set the initial value of the quantization parameter of the tile in the first slice header of the picture in the encoded stream.
- the initial value setting unit can set the initial value of the quantization parameter of the tile near the head of the tile in the encoded stream.
- the initial value setting unit can set an initial value of a quantization parameter of the tile in slice data in the encoded stream.
- the transmission unit may transmit information on the quantization parameter of the coding unit constituting the tile generated using the initial value of the quantization parameter of the tile.
- An image processing method relates to a quantization parameter between tiles in a picture in which an image processing apparatus encodes an image to generate an encoded stream and is divided into a plurality of tiles of the image.
- the initial value of the quantization parameter for the tile is set so as to ensure the independence of the process, and the initial value of the quantization parameter for the set tile and the generated encoded stream are transmitted.
- the quantization parameter for a tile that is set to ensure independence of an encoded stream and a quantization parameter between tiles in a picture divided into a plurality of tiles of an image.
- An initial value is received.
- the quantization parameter of the coding unit constituting the tile is reconstructed, and the received encoding is performed using the reconstructed quantization parameter.
- the stream is decoded and the image is generated.
- an image is encoded and an encoded stream is generated to ensure processing independence regarding quantization parameters between tiles in a picture divided into a plurality of tiles of the image.
- the initial value of the quantization parameter for the tile is set.
- the initial value of the quantization parameter for the set tile and the generated encoded stream are transmitted.
- the above-described image processing apparatus may be an independent apparatus, or may be an internal block constituting one image encoding apparatus or image decoding apparatus.
- an image can be decoded.
- it is possible to improve decoding processing efficiency in a tile-divided picture.
- an image can be encoded.
- it is possible to improve decoding processing efficiency in a tile-divided picture.
- FIG. 20 is a block diagram illustrating a main configuration example of a computer. It is a block diagram which shows an example of a schematic structure of a television apparatus. It is a block diagram which shows an example of a schematic structure of a mobile telephone. It is a block diagram which shows an example of a schematic structure of a recording / reproducing apparatus. It is a block diagram which shows an example of a schematic structure of an imaging device. It is a block diagram which shows an example of scalable encoding utilization. It is a block diagram which shows the other example of scalable encoding utilization. It is a block diagram which shows the further another example of scalable encoding utilization.
- FIG. 1 is a block diagram illustrating a main configuration example of an image encoding device.
- the image encoding device 100 shown in FIG. 1 encodes image data using a prediction process based on, for example, HEVC (High Efficiency Video Coding).
- HEVC High Efficiency Video Coding
- the image encoding device 100 includes an A / D conversion unit 101, a screen rearrangement buffer 102, a calculation unit 103, an orthogonal transformation unit 104, a quantization unit 105, a lossless encoding unit 106, and a storage buffer 107. , An inverse quantization unit 108, and an inverse orthogonal transform unit 109.
- the image coding apparatus 100 includes a calculation unit 110, a deblock filter 111, a frame memory 112, a selection unit 113, an intra prediction unit 114, a motion prediction / compensation unit 115, a predicted image selection unit 116, and a rate control unit 117. Have.
- the image encoding device 100 further includes a tile dividing unit 121 and a quantization initial value setting unit 122.
- the A / D conversion unit 101 performs A / D conversion on the input image data, and supplies the converted image data (digital data) to the screen rearrangement buffer 102 for storage.
- the screen rearrangement buffer 102 rearranges the images of the frames in the stored display order in the order of frames for encoding in accordance with GOP (Group Of Picture). Then, the screen rearrangement buffer 102 supplies the image with the rearranged frame order to the arithmetic unit 103. This image is also supplied to the tile dividing unit 121. The screen rearrangement buffer 102 also supplies the image with the rearranged frame order to the intra prediction unit 114 and the motion prediction / compensation unit 115.
- GOP Group Of Picture
- the calculation unit 103 subtracts the prediction image supplied from the intra prediction unit 114 or the motion prediction / compensation unit 115 via the prediction image selection unit 116 from the image read from the screen rearrangement buffer 102, and the difference information Is output to the orthogonal transform unit 104.
- the calculation unit 103 subtracts the predicted image supplied from the motion prediction / compensation unit 115 from the image read from the screen rearrangement buffer 102.
- the orthogonal transform unit 104 performs orthogonal transform such as discrete cosine transform and Karhunen-Loeve transform on the difference information supplied from the computation unit 103. Note that this orthogonal transformation method is arbitrary.
- the orthogonal transform unit 104 supplies the transform coefficient to the quantization unit 105.
- the quantization unit 105 quantizes the transform coefficient supplied from the orthogonal transform unit 104 using the quantization parameter supplied from the rate control unit 117. Note that this quantization method is arbitrary.
- the quantization unit 105 supplies the quantized transform coefficient to the lossless encoding unit 106.
- the lossless encoding unit 106 encodes the transform coefficient quantized by the quantization unit 105 using an arbitrary encoding method. Since the coefficient data is quantized under the control of the rate control unit 117, the code amount becomes a target value set by the rate control unit 117 (or approximates the target value).
- the lossless encoding unit 106 acquires information indicating an intra prediction mode from the intra prediction unit 114, and acquires information indicating an inter prediction mode, differential motion vector information, and the like from the motion prediction / compensation unit 115. Further, the lossless encoding unit 106 acquires an address of a tile boundary as tile division information from the tile division unit 121. The lossless encoding unit 106 acquires information regarding the initial value of the quantization parameter from the quantization initial value setting unit 122.
- the lossless encoding unit 106 encodes these various types of information by an arbitrary encoding method, and uses (multiplexes) the information as a part of header information of encoded data (also referred to as an encoded stream).
- the lossless encoding unit 106 supplies the encoded data obtained by encoding to the accumulation buffer 107 for accumulation.
- Examples of the encoding method of the lossless encoding unit 106 include variable length encoding or arithmetic encoding.
- Examples of the variable length coding include CAVLC (Context-Adaptive Variable Length Coding) defined by the AVC method.
- Examples of arithmetic coding include CABAC (Context-Adaptive Binary Arithmetic Coding).
- the accumulation buffer 107 temporarily holds the encoded data supplied from the lossless encoding unit 106.
- the accumulation buffer 107 outputs the stored encoded data to, for example, a recording device (recording medium) (not shown) or a transmission path (not shown) at a predetermined timing at a predetermined timing. That is, the accumulation buffer 107 is also a transmission unit that transmits encoded data.
- the transform coefficient quantized by the quantization unit 105 is also supplied to the inverse quantization unit 108.
- the inverse quantization unit 108 inversely quantizes the quantized transform coefficient by a method corresponding to the quantization by the quantization unit 105.
- the inverse quantization method may be any method as long as it is a method corresponding to the quantization processing by the quantization unit 105.
- the inverse quantization unit 108 supplies the obtained transform coefficient to the inverse orthogonal transform unit 109.
- the inverse orthogonal transform unit 109 performs inverse orthogonal transform on the transform coefficient supplied from the inverse quantization unit 108 by a method corresponding to the orthogonal transform process by the orthogonal transform unit 104.
- the inverse orthogonal transform method may be any method as long as it corresponds to the orthogonal transform processing by the orthogonal transform unit 104.
- the inversely orthogonal transformed output (restored difference information) is supplied to the calculation unit 110.
- the computing unit 110 adds the restored difference information, which is the inverse orthogonal transformation result supplied from the inverse orthogonal transformation unit 109, to the prediction from the intra prediction unit 114 or the motion prediction / compensation unit 115 via the prediction image selection unit 116.
- the images are added to obtain a locally decoded image (decoded image).
- the decoded image is supplied to the deblock filter 111 or the frame memory 112.
- the deblock filter 111 appropriately performs a deblock filter process on the decoded image supplied from the calculation unit 110.
- the deblocking filter 111 removes block distortion of the decoded image by performing a deblocking filter process on the decoded image.
- the deblock filter 111 supplies the filter processing result (decoded image after the filter processing) to the frame memory 112. As described above, the decoded image output from the calculation unit 110 can be supplied to the frame memory 112 without passing through the deblocking filter 111. That is, the filtering process by the deblocking filter 111 can be omitted.
- the frame memory 112 stores the supplied decoded image, and supplies the stored decoded image as a reference image to the selection unit 113 at a predetermined timing.
- the selection unit 113 selects a supply destination of the reference image supplied from the frame memory 112. For example, in the case of inter prediction, the selection unit 113 supplies the reference image supplied from the frame memory 112 to the motion prediction / compensation unit 115.
- the intra prediction unit 114 basically uses the pixel value in the processing target picture, which is a reference image supplied from the frame memory 112 via the selection unit 113, to generate a prediction image using a prediction unit (PU) as a processing unit. Perform intra prediction (intra-screen prediction) to be generated. The intra prediction unit 114 performs this intra prediction in a plurality of intra prediction modes prepared in advance.
- the intra prediction unit 114 generates predicted images in all candidate intra prediction modes, evaluates the cost function value of each predicted image using the input image supplied from the screen rearrangement buffer 102, and selects the optimum mode. select. When the intra prediction unit 114 selects the optimal intra prediction mode, the intra prediction unit 114 supplies the predicted image generated in the optimal mode to the predicted image selection unit 116.
- the intra prediction unit 114 appropriately supplies the intra prediction mode information indicating the adopted intra prediction mode to the lossless encoding unit 106 and causes the encoding to be performed.
- the motion prediction / compensation unit 115 basically uses the input image supplied from the screen rearrangement buffer 102 and the reference image supplied from the frame memory 112 via the selection unit 113 as a processing unit. Perform motion prediction (inter prediction).
- the motion prediction / compensation unit 115 performs a motion compensation process according to the detected motion vector, and generates a prediction image (inter prediction image information).
- the motion prediction / compensation unit 115 performs such inter prediction in a plurality of inter prediction modes prepared in advance.
- the motion prediction / compensation unit 115 evaluates the cost function value of each predicted image using the input image and motion vector information supplied from the screen rearrangement buffer 102, and selects an optimal mode. When the optimal inter prediction mode is selected, the motion prediction / compensation unit 115 generates a prediction image in the optimal mode, and supplies the generated prediction image to the prediction image selection unit 116.
- the motion prediction / compensation unit 115 supplies information indicating the employed inter prediction mode, information necessary for performing processing in the inter prediction mode, and the like to the lossless encoding unit 106 when decoding the encoded data. And encoding.
- the predicted image selection unit 116 selects a supply source of a predicted image to be supplied to the calculation unit 103 or the calculation unit 110. For example, in the case of inter coding, the prediction image selection unit 116 selects the motion prediction / compensation unit 115 as a supply source of the prediction image, and calculates the prediction image supplied from the motion prediction / compensation unit 115 as the calculation unit 103 or the calculation unit. To the unit 110.
- the rate control unit 117 controls the quantization operation rate of the quantization unit 105 based on the code amount of the encoded data stored in the storage buffer 107 so that overflow or underflow does not occur.
- the rate control unit 117 sets a quantization parameter based on the information regarding the code amount target value, and supplies the quantization parameter to the quantization unit 105.
- the rate control unit 117 sends the initial quantization parameter for the picture (hereinafter referred to as the picture initial QP) and the initial quantization parameter for each tile (hereinafter referred to as the tile initial QP) to the quantization initial value setting unit 122. Supply.
- the tile division unit 121 is input with settings relating to how the picture is tiled and processed, which is designated by the user, via an operation input unit (not shown).
- the tile division unit 121 refers to the input image from the screen rearrangement buffer 102, and determines tile division based on the setting relating to how the picture is tiled and processed.
- the tile division unit 121 supplies the determined tile boundary address and the like to the quantization initial value setting unit 122 and the lossless encoding unit 106 as tile division information.
- the quantization initial value setting unit 122 refers to the tile division information from the tile dividing unit 121, and sets the initial value of the quantization parameter for each tile using the information supplied from the rate control unit 117.
- the initial value of the quantization parameter for each tile is the difference between the quantization parameter of the first CU in the tile on the encoding side and the reconstruction of the quantization parameter of the first CU in the tile on the decoding side. It is a value used as an initial value.
- the initial value of the quantization parameter for the tile is set so as to ensure the independence of the process regarding the quantization parameter between tiles in the picture.
- the quantization initial value setting unit 122 supplies the initial value of the quantization parameter for the tile, information regarding the quantization parameter of the first CU in the tile, and the like to the lossless encoding unit 106 as information regarding the initial value of the quantization parameter. To do.
- a hierarchical structure is defined by macroblocks and sub-macroblocks.
- a coding unit (Coding Unit)
- FIG. ing a coding unit
- CU is also called Coding Tree Block (CTB) and is a partial area of a picture unit image that plays the same role as a macroblock in the AVC method.
- CTB Coding Tree Block
- the latter is fixed to a size of 16 ⁇ 16 pixels, whereas the size of the former is not fixed, and is specified in the image compression information in each sequence.
- SPS Sequence Coding Unit
- LCU Large Coding Unit
- SCU Smallest Coding Unit
- the LCU size is 128 and the maximum hierarchical depth is 5.
- split_flag the value of split_flag is “1”
- the 2N ⁇ 2N size CU is divided into N ⁇ N size CUs that are one level below.
- the CU is divided into prediction units (Prediction Unit (PU)) that are regions (partial regions of images in units of pictures) that are processing units for intra or inter prediction.
- the PU is also divided into transform units (Transform Unit (TU)) that are regions (partial regions of images in units of pictures) that are processing units of orthogonal transformation.
- PU Prediction Unit
- TU Transform Unit
- inter _4 ⁇ 4_enable_flag is defined, and by setting this value to 0, it is possible to prohibit the use of an inter CU having a 4 ⁇ 4 block size.
- a macroblock in the AVC method corresponds to an LCU
- a block (subblock) corresponds to a CU. Then you can think.
- a motion compensation block in the AVC method can be considered to correspond to a PU.
- the size of the LCU of the highest hierarchy is generally set larger than that of the AVC macroblock, for example, 128 ⁇ 128 pixels.
- Non-Patent Document 1 defines tiles for parallel processing.
- a tile is one of picture division units.
- Tile is a sequence parameter set (SPS) or picture parameter set (PPS) in image compression information. It is specified.
- SPS sequence parameter set
- PPS picture parameter set
- FIG. 3 a picture divided into LCUs is shown.
- a rectangle with a number represents an LCU, and a picture is further divided into tiles T0 to T8 divided by a row boundary and a column boundary indicated by bold lines.
- the tile T0 located in the upper left is configured to include an LCU with a row of 3 and a column of 4 designated with 1 to 12.
- the tile T1 adjacent to the right side of the tile T0 is designated by a row of 3 and a column of 6 and includes LCUs having 13 to 30 attached thereto.
- the tile T2 adjacent to the right side of the tile T1 is designated by 3 as a row and 3 as a column, and is configured to include LCUs having 31 to 39.
- the LCUs included in each tile are processed in the raster scan order, which is the order of the attached numbers shown in FIG. Also, the tiles included in the picture are processed in the raster scan order which is the order of the tiles T0 to T8 shown in FIG.
- FIG. 4 there may be a slice boundary in the middle of the tile. That is, in the example of FIG. 4, two column boundaries that divide the horizontal direction of the picture into three tiles T11 to T13 are shown. A slice boundary is shown between the LCU to which 9 is attached and the LCU to which 10 is attached in the middle of the tile T11.
- 5 and 6 are diagrams showing examples of syntax of the picture parameter set of the HEVC method.
- the number at the left end of each line is the line number given for explanation.
- the initial quantization parameter for each picture is set in pic_init_qp_minus26 in the 10th row.
- max_cu_qp_delta_depth is set with which granularity (depth) CU the quantization parameter is transmitted.
- FIGS. 7 to 10 are diagrams showing examples of syntax of the HEVC system sequence parameter set. The number at the left end of each line is the line number given for explanation.
- slice_qp_delta on the 62nd line is difference information between the initial quantization parameter of the slice and the initial quantization parameter of the picture. Thereby, the initial quantization parameter in the slice is calculated on the decoding side.
- FIG. 11 is a diagram showing an example of the syntax of the HEVC transform unit. The number at the left end of each line is the line number given for explanation.
- cu_qp_delta in the fourth row is difference information between the initial quantization parameter of the slice and the quantization parameter of the first CU in the slice. Thereby, the quantization parameter of the first CU is calculated on the decoding side.
- This cu_qp_delta is transmitted to the decoding side in the first transform unit of the CU having a size specified by max_cu_qp_delta_depth in FIG. 5 or larger.
- the quantization parameter of the CU having a size smaller than the size specified by max_cu_qp_delta_depth in FIG. 5 is, for example, the quantization parameter of the adjacent region having a correlation among the adjacent regions, and thus needs to be transmitted to the decoding side. Absent.
- initial quantization parameters in a picture are set in a picture parameter set in image compression information. At that time, it is set by which granularity CU the quantization parameter is sent to the decoding side.
- the initial quantization parameter in the slice is calculated on the decoding side.
- the difference between the initial quantization parameter in the slice and the quantization parameter in the CU is set at the head of the CU having a size set in the picture parameter set or larger. Thereby, the quantization parameter in the CU is calculated on the decoding side.
- the horizontal direction of the picture is divided into three tiles T0 to T2.
- the picture is divided into a slice 0 that is not hatched and a slice 1 that is hatched, and the boundary is included in the tile T1.
- FIG. 12B shows an encoded stream (image compression information) transmitted from the encoding side in the case of FIG. 12A.
- the triangle represents each slice header, and the circle represents the first CU at the tile T2.
- the slice header of slice 0 is arranged near the beginning of slice 0 and near the beginning of tile T0.
- the slice header of slice 1 is arranged near the beginning of slice 1 and in the middle of tile T1.
- the tile T2 is included in the slice T1. Therefore, in order to reconstruct the quantization parameter in the first CU of the tile T2, it is necessary to access the slice header of the slice 1 arranged in the middle of the tile T1.
- the tile T2 cannot be processed unless the processing of the slice header of the slice 1 in the tile T1 is finished, which hinders parallel processing for each tile.
- the quantization parameter for the tiles in order to ensure the independence of the processing related to the quantization parameter between tiles, that is, to eliminate the dependency on the processing related to the quantization parameter between tiles, the quantization parameter for the tiles.
- the initial value of is set.
- the process regarding the quantization parameter between tiles is a process performed using a quantization parameter, for example, although a quantization process is mention
- FIG. 13 and FIG. 13 and 14 conceptually show an encoded stream of a picture divided into tiles T0 to T2.
- a triangle represents a slice header
- a square represents a picture parameter set
- a circle represents each CU at the beginning of the tile T1 and the tile T2.
- the initial value of the quantization parameter for the tile referred to by the first CU of each tile is set to information arranged outside the previous tile (before the tile).
- the quantization initial value setting unit 122 sets, as an initial value of a quantization parameter for each tile, a picture parameter set arranged before the first tile. To do. Specifically, pic_init_qp_minus26 (initial quantization parameter in an existing picture) of the picture parameter set shown in FIG. 5 is set as the initial value of the quantization parameter for each tile.
- the difference between the initial value of the quantization parameter for each tile (initial QP in the picture) and the initial QP in the tile is set as CU_delta_QP in the transform unit of the first CU of the tile.
- the quantization initial value setting unit 122 for example, as shown in FIG. 13B, is the first value in the picture that is arranged before the first tile as the initial value of the quantization parameter for each tile.
- the first slice is also referred to as a first slice.
- slice_qp_delta (initial quantization parameter in the existing first slice) in the first slice header shown in FIG. 8 is set as the initial value of the quantization parameter for each tile.
- the difference between the initial value of the quantization parameter for each tile (initial QP in the first slice) and the initial QP in the tile is set as CU_delta_QP in the transform unit of the first CU of the tile.
- the first method A information on the quantization parameter set in the slice header is skipped, so that the amount of calculation is reduced and the processing is shortened.
- the process for the quantization parameter is changed when the picture is tiled or not tiled.
- the picture is tiled or not tiled, and there is no change in the processing process for the quantization parameter.
- the initial value (TILE_int_QP) of the quantization parameter for each tile which is referred to by the CU at the head of each tile, is updated near the head of each tile.
- the difference between the initial value (TILE_int_QP) of the quantization parameter for each tile and the QP of the first CU constituting the tile is set as CU_delta_QP in the transform unit of the first CU of the tile.
- the initial value of the quantization parameter for each tile may be included in information (entry point) indicating the head of the tile proposed in Non-Patent Document 2, for example.
- the first method stores the initial value in the buffer at the beginning of the picture and refers to this at the beginning of each tile. This increases the capacity of the picture buffer, but the second method Such an increase in the circuit does not occur.
- the second method needs to transmit the initial value at the head of each tile, the encoding efficiency is lower than that of the first method. In the first method, such an increase in code amount does not occur.
- FIG. 15 is a block diagram illustrating a main configuration example of the quantization initial value setting unit 122 and the lossless encoding unit 106 that execute the second method described above with reference to FIG. 14.
- the 15 is configured to include a tile initial QP buffer 151, a picture initial QP buffer 152, and a differential QP generation unit 153.
- the lossless encoding unit 106 is configured to include at least a syntax setting unit 161.
- the tile dividing unit 121 supplies the tile boundary address to the tile initial QP buffer 151 and the syntax setting unit 161.
- the tile initial QP buffer 151 requests and acquires the tile initial QP from the rate control unit 117 when the tile boundary address is supplied.
- the tile initial QP is, for example, a quantization parameter of the first CU constituting the tile.
- the tile initial QP may be, for example, a value close to the quantization parameter of the first CU constituting the tile, or any value.
- the tile initial QP buffer 151 supplies the acquired tile initial QP to the differential QP generation unit 153.
- the picture initial QP buffer 152 is supplied with the picture initial QP from the rate control unit 117.
- the picture initial QP is, for example, a quantization parameter of the first CU constituting the picture.
- the picture initial QP may be, for example, a value close to the quantization parameter of the first CU constituting the picture, or any value.
- the picture initial QP buffer 152 supplies the picture initial QP to the differential QP generation unit 153 and the syntax setting unit 161 in response to a request from the differential QP generation unit 153.
- the difference QP generation unit 153 requests and acquires the picture initial QP from the picture initial QP buffer 152.
- the difference QP generation unit 153 generates a difference QP between the picture initial QP and the tile initial QP, and supplies the generated difference QP to the syntax setting unit 161.
- the syntax setting unit 161 uses the picture initial QP from the picture initial QP buffer 152 and the difference QP between the picture initial QP and the tile initial QP from the difference QP generation unit 153 as information on the initial value of the quantization parameter. Set.
- the syntax setting unit 161 sets the picture initial QP from the picture initial QP buffer 152 to pic_init_qp_minus26 of the picture parameter set in the encoded stream. In addition, the syntax setting unit 161 sets the difference QP between the picture initial QP and the tile initial QP as the initial value (TILE_int_QP) of the quantization parameter for the tile near the head of the corresponding tile in the encoded stream.
- the difference between the quantization parameter initial value for the tile and the quantization parameter of the first CU in the tile is, for example, It shall be calculated.
- the initial value of the quantization parameter for the tile is supplied to the quantization unit 105.
- the difference between the initial value of the quantization parameter for the tile and the quantization parameter of the first CU in the tile is supplied to the syntax setting unit 161 as information regarding the quantization parameter of the CU.
- the syntax setting unit 161 the information regarding the supplied CU quantization parameter is set in the transform unit of the first CU in the tile.
- the tile division information is also supplied to the syntax setting unit 161.
- This tile division information is also transmitted to the decoding side. At this time, for example, based on the tile division information, as shown in Non-Patent Document 2, information indicating a tile boundary is set near the head of the tile. You may make it do.
- the first method A shown in FIG. 13A can also be performed by the functional block of FIG. .
- the syntax setting unit 161 uses the picture initial QP from the picture initial QP buffer 152 and the difference QP from the difference QP generation unit 153 as information on the initial value of the quantization parameter. Set.
- the syntax setting unit 161 is different from the second method in that the picture initial QP is set in the pic_init_qp_minus26 of the picture parameter set in the encoded stream as the initial value of the quantization parameter for the tile. Also, the syntax setting unit 161 sets the difference QP between the picture initial QP and the tile initial QP as the difference QP of the first CU in the tile in the transform unit of the first CU of the corresponding tile in the encoded stream. This is different from the second method.
- the tile division unit 121 is input with settings relating to how the picture is tiled and processed, which is designated by the user, via an operation input unit (not shown).
- the tile division unit 121 refers to the input image from the screen rearrangement buffer 102, and determines tile division based on the setting relating to how the picture is tiled and processed.
- the tile division unit 121 supplies tile division information such as the determined tile boundary address to the tile initial QP buffer 151 and the lossless encoding unit 106 of the quantization initial value setting unit 122.
- step S102 the quantization initial value setting unit 122 performs quantization parameter initial value setting processing for each tile.
- the process of setting the quantization parameter initial value of each tile in step S102 will be described later with reference to FIG. 17.
- the syntax setting unit 161 sends the picture initial QP and the tile. Information about the initial QP is provided.
- step S ⁇ b> 103 the syntax setting unit 161 performs the supplied parameter encoding process as the encoding process for the information related to the quantization parameter initial value.
- the syntax setting unit 161 sets the picture initial QP from the picture initial QP buffer 152 to pic_init_qp_minus26 of the picture parameter set in the encoded stream as the encoding process.
- the syntax setting unit 161 sets information about the tile initial QP as the initial value (TILE_int_QP) of the quantization parameter for the tile near the head of the corresponding tile in the encoded stream.
- each unit of the image encoding device 100 performs encoding processing of each tile by parallel processing.
- the encoding process of each tile is an encoding process such as a DCT coefficient and a motion vector for each tile below the slice header.
- the encoding process for each tile will be described later with reference to FIG.
- step S104 In the encoding process of each tile in step S104, the information below each tile is encoded and transmitted to the image decoding apparatus 200 described later.
- step S111 the quantization initial value setting unit 122 sets a new quantization parameter initial value (TILE_int_QP), which is the second method described above with reference to FIG. 14, as the quantization parameter initial value.
- TILE_int_QP a new quantization parameter initial value
- the tile initial QP buffer 151 requests and obtains the tile initial QP from the rate control unit 117 in step S112.
- the tile initial QP buffer 151 supplies the acquired tile initial QP to the differential QP generation unit 153.
- the difference QP generation unit 153 requests and obtains the picture initial QP from the picture initial QP buffer 152 in step S113.
- step S114 the difference QP generation unit 153 generates a difference QP between the picture initial QP and the tile initial QP.
- the difference QP generation unit 153 supplies the generated difference QP to the syntax setting unit 161 as information on the tile initial QP.
- step S121 the A / D conversion unit 101 performs A / D conversion on the input image.
- step S122 the screen rearrangement buffer 102 stores the A / D converted image, and rearranges the picture from the display order to the encoding order.
- step S123 the intra prediction unit 114 performs an intra prediction process in the intra prediction mode.
- step S124 the motion prediction / compensation unit 115 performs an inter motion prediction process for performing motion prediction and motion compensation in the inter prediction mode.
- step S124 the motion vector of the luminance signal of the PU is searched, the cost function value is calculated, and the optimal inter prediction mode is determined from all the inter prediction modes. Then, a prediction image in the optimal inter prediction mode is generated.
- the predicted image and cost function value of the determined optimal inter prediction mode are supplied from the motion prediction / compensation unit 115 to the predicted image selection unit 116. Further, the information on the determined optimal inter prediction mode and the information on the motion vector are also supplied to the lossless encoding unit 106 and are losslessly encoded in step S134 described later.
- step S125 the predicted image selection unit 116 determines an optimal mode based on the cost function values output from the intra prediction unit 114 and the motion prediction / compensation unit 115. That is, the predicted image selection unit 116 selects one of the predicted image generated by the intra prediction unit 114 and the predicted image generated by the motion prediction / compensation unit 115.
- step S126 the calculation unit 103 calculates a difference between the image rearranged by the process of step S122 and the predicted image selected by the process of step S125.
- the data amount of the difference data is reduced compared to the original image data. Therefore, the data amount can be compressed as compared with the case where the image is encoded as it is.
- step S127 the orthogonal transform unit 104 orthogonally transforms the difference information generated by the process in step S126. Specifically, orthogonal transformation such as discrete cosine transformation and Karhunen-Loeve transformation is performed, and transformation coefficients are output.
- orthogonal transformation such as discrete cosine transformation and Karhunen-Loeve transformation is performed, and transformation coefficients are output.
- step S128 the quantization unit 105 quantizes the orthogonal transform coefficient obtained by the process in step S127, using the quantization parameter from the rate control unit 117.
- this quantization method is arbitrary, but for example, a quantization parameter for a CU having a specified size or larger is supplied from the rate control unit 117.
- a quantization parameter of a CU having a size smaller than the size for example, a quantization parameter of an adjacent region having a correlation among adjacent regions is used.
- step S128 The difference information quantized by the process of step S128 is locally decoded as follows. That is, in step S129, the inverse quantization unit 108 inversely quantizes the quantized orthogonal transform coefficient (also referred to as quantization coefficient) generated by the process in step S128 with characteristics corresponding to the characteristics of the quantization unit 105. To do. In step S ⁇ b> 130, the inverse orthogonal transform unit 109 performs inverse orthogonal transform on the orthogonal transform coefficient obtained by the process of step S ⁇ b> 129 with characteristics corresponding to the characteristics of the orthogonal transform unit 104.
- the quantized orthogonal transform coefficient also referred to as quantization coefficient
- step S131 the calculation unit 110 adds the predicted image to the locally decoded difference information, and generates a locally decoded image (an image corresponding to the input to the calculation unit 103).
- step S132 the deblocking filter 111 appropriately performs a deblocking filter process on the locally decoded image obtained by the process of step S131.
- step S133 the frame memory 112 stores the decoded image that has been subjected to the deblocking filter process by the process of step S132. It should be noted that an image that has not been filtered by the deblocking filter 111 is also supplied from the computing unit 110 and stored in the frame memory 112.
- step S134 the lossless encoding unit 106 encodes the transform coefficient quantized by the process in step S128. That is, lossless encoding such as variable length encoding or arithmetic encoding is performed on the difference image.
- the lossless encoding unit 106 encodes information about the prediction mode of the prediction image selected by the process of step S125, and adds the encoded information to the encoded data obtained by encoding the difference image. That is, the lossless encoding unit 106 also encodes and encodes the optimal intra prediction mode information supplied from the intra prediction unit 114 or information according to the optimal inter prediction mode supplied from the motion prediction / compensation unit 115, and the like. Append to data.
- step S135 the accumulation buffer 107 accumulates the encoded data obtained by the process in step S134.
- the encoded data stored in the storage buffer 107 is appropriately read and transmitted to the decoding side via a transmission path or a recording medium.
- step S136 the rate control unit 117 causes the quantization unit 105 to prevent overflow or underflow from occurring based on the code amount (generated code amount) of the encoded data accumulated in the accumulation buffer 107 by the process of step S135. Controls the rate of quantization operation.
- step S136 ends, the encoding process ends.
- FIG. 19 is a block diagram illustrating a main configuration example of the quantization initial value setting unit 122 and the lossless encoding unit 106 that execute the first B method described above with reference to FIG. 13B.
- the quantization initial value setting unit 122 in the example of FIG. 19 includes a tile initial QP buffer 181, a picture initial QP buffer 182, a first slice initial QP buffer 183, a differential QP generation unit 184, and a differential QP generation unit 185. Composed.
- the lossless encoding unit 106 is configured to include at least a syntax setting unit 191.
- the tile dividing unit 121 supplies the tile boundary address to the tile initial QP buffer 181 and the syntax setting unit 191.
- the tile initial QP buffer 181 requests and acquires the tile initial QP from the rate control unit 117 when the tile boundary address is supplied.
- the tile initial QP buffer 181 supplies the acquired tile initial QP to the differential QP generation unit 185.
- the picture initial QP buffer 182 is supplied with the picture initial QP from the rate control unit 117.
- the picture initial QP buffer 182 supplies the picture initial QP to the differential QP generation unit 184 and the syntax setting unit 191 in response to a request from the differential QP generation unit 184.
- the initial quantization parameter in the first slice (hereinafter referred to as the first slice initial QP) is supplied from the rate control unit 117 to the first slice initial QP buffer 183.
- the first slice initial QP is, for example, a quantization parameter of the first CU configuring the first slice.
- the first slice initial QP may be, for example, a value close to the quantization parameter of the first CU constituting the first slice, or any value.
- the first slice initial QP buffer 183 supplies the first slice initial QP to the differential QP generation unit 184 and the differential QP generation unit 185 in response to a request from the differential QP generation unit 185.
- the difference QP generation unit 184 requests and acquires the picture initial QP from the picture initial QP buffer 182.
- the difference QP generation unit 184 generates a first slice difference QP that is a difference between the picture initial QP and the first slice initial QP, and sets the generated first slice difference QP as information regarding the first slice initial QP. To the unit 191.
- the difference QP generation unit 185 requests and acquires the first slice initial QP from the first slice initial QP buffer 183.
- the difference QP generation unit 185 generates a tile difference QP that is a difference between the first slice initial QP and the tile initial QP, and supplies the generated tile difference QP to the syntax setting unit 191 as information on the tile initial QP.
- the syntax setting unit 191 uses the initial value of the quantization parameter such as the picture initial QP from the picture initial QP buffer 182, the first slice difference QP from the difference QP generation unit 184, and the tile difference QP from the difference QP generation unit 185.
- the syntax is set as information related to.
- the syntax setting unit 191 sets the picture initial QP from the picture initial QP buffer 182 in the picture parameter set pic_init_qp_minus26 in the encoded stream. Also, the syntax setting unit 191 sets the first slice difference QP from the difference QP generation unit 184 as slice_qp_delta in the first slice header as an initial value of the quantization parameter for the tile. The syntax setting unit 191 sets the tile difference QP between the first slice initial QP and the tile initial QP in the transform unit of the first CU of the corresponding tile in the encoded stream as information regarding the quantization parameter of the CU. .
- the tile setting information is supplied to the syntax setting unit 191 as in the case of the syntax setting unit 161.
- This tile division information is also transmitted to the decoding side. At this time, for example, based on the tile division information, as shown in Non-Patent Document 2, information indicating a tile boundary is set near the head of the tile. You may make it do.
- the difference between the configuration of the second method shown in FIG. 15 and the configuration of the first B method shown in FIG. 19 is as follows. That is, in the configuration of the second method in FIG. 15, the difference between the tile initial QP and the picture initial QP is encoded at the head of each tile. On the other hand, in the configuration of the first B method shown in FIG. 19, the difference between the tile initial QP and the first slice initial QP is encoded.
- step S102 of FIG. 16 in the case of the first B method will be described next.
- step S151 the quantization initial value setting unit 122 sets slice_qp_delta in the first slice header, which is the first B method described above with reference to FIG. 13B, as the quantization parameter initial value.
- the tile initial QP buffer 181 requests and acquires the tile initial QP from the rate control unit 117 in step S152.
- the tile initial QP buffer 181 supplies the acquired tile initial QP to the differential QP generation unit 185.
- step S153 the difference QP generation unit 184 requests and obtains the picture initial QP from the picture initial QP buffer 152. At this time, the picture initial QP is also supplied to the syntax setting unit 191.
- step S154 the difference QP generation unit 184 and the difference QP generation unit 185 request and acquire the first slice initial QP from the first slice initial QP buffer 183.
- the difference QP generation unit 184 generates a first slice difference QP between the picture initial QP and the first slice initial QP in step S155.
- the difference QP generation unit 153 supplies the generated first slice difference QP to the syntax setting unit 161 as the initial value of the quantization parameter for the tile.
- step S156 the difference QP generation unit 185 generates a tile difference QP between the first slice initial QP and the tile initial QP.
- the difference QP generation unit 185 supplies the generated tile difference QP to the syntax setting unit 161 as information regarding the quantization parameter of the CU.
- the syntax setting unit 161 executes an encoding process of information regarding the quantization parameter initial value in step S103 of FIG.
- the syntax setting unit 191 sets the picture initial QP from the picture initial QP buffer 182 in the picture parameter set pic_init_qp_minus26 in the encoded stream.
- the syntax setting unit 191 sets the first slice difference QP from the difference QP generation unit 184 as slice_qp_delta in the first slice header as an initial value of the quantization parameter for the tile.
- the syntax setting unit 191 sets the tile difference QP between the first slice initial QP and the tile initial QP in the transform unit of the first CU of the corresponding tile in the encoded stream as information regarding the quantization parameter of the CU. .
- the initial value of the quantization parameter for the tile depends on the process regarding the quantization parameter between the tiles so as to ensure the independence of the process regarding the quantization parameter between the tiles. It was set to eliminate the sex.
- FIGS. 21 to 23 are diagrams illustrating an example of syntax of a sequence parameter set
- FIGS. 24 and 25 are diagrams illustrating an example of syntax of a picture parameter set. The number at the left end of each line is the line number given for explanation.
- tile_init_delta_qp which is information regarding the initial value of the quantization parameter for each tile, is set. .
- tile_init_delta_qp is a difference value between the initial value of the quantization parameter for each tile and the initial value of the quantization parameter for the picture. In this case, since information for all tiles is written here at once, tile_init_delta_qp is an array value.
- the slice data may be changed as shown below.
- FIG. 26 is a diagram illustrating an example of the syntax of slice data. The number at the left end of each line is the line number given for explanation.
- Entry_point_marker_two_3bytes is set in the 26th line in FIG. Entry_point_marker_two_3bytes is information (entry point marker) indicating the top of a tile proposed in Non-Patent Document 2.
- tile_init_delta_qp which is information relating to the initial value of the quantization parameter for each tile.
- tile_init_delta_qp is a difference value between the initial value of the quantization parameter for each tile and the initial value of the quantization parameter for the picture.
- information regarding the initial value of the quantization parameter is written for each tile, and therefore the information regarding the initial value of the quantization parameter of the tile is a scalar value.
- the marker on the 26th line serves as an entry point. Therefore, information about the initial value of the quantization parameter for each tile can be set at the entry point (after) included in the slice data and transmitted to the decoding side.
- FIG. 27 is a block diagram illustrating a main configuration example of an image decoding apparatus corresponding to the image encoding apparatus 100 of FIG.
- the image decoding device 200 shown in FIG. 27 decodes the encoded data generated by the image encoding device 100 in FIG. 1 by a decoding method corresponding to the encoding method. Note that, similarly to the image encoding device 100, the image decoding device 200 performs inter prediction for each prediction unit (PU).
- PU prediction unit
- tile division information which is information on how to perform tile division, is transmitted together with an input encoded stream, and parallel processing is performed based on this. Done.
- the image decoding apparatus 200 includes a storage buffer 201, a lossless decoding unit 202, an inverse quantization unit 203, an inverse orthogonal transform unit 204, a calculation unit 205, a deblock filter 206, a screen rearrangement buffer 207, and A D / A converter 208 is included.
- the image decoding apparatus 200 includes a frame memory 209, a selection unit 210, an intra prediction unit 211, a motion prediction / compensation unit 212, and a selection unit 213.
- the image decoding device 200 includes a tile division decoding unit 221 and a quantization initial value decoding unit 222.
- the accumulation buffer 201 is also a receiving unit that receives transmitted encoded data.
- the accumulation buffer 201 receives and accumulates the transmitted encoded data, and supplies the encoded data to the lossless decoding unit 202 at a predetermined timing.
- the lossless decoding unit 202 acquires information about the quantization parameter initial value of the picture added to the encoded data, tile division information, and the like.
- the lossless decoding unit 202 supplies information related to the quantization parameter initial value of the picture to the quantization initial value decoding unit 222 and supplies tile division information to the tile division decoding unit 221.
- the lossless decoding unit 202 decodes the information supplied from the accumulation buffer 201 and encoded by the lossless encoding unit 106 in FIG. 1 by a method corresponding to the encoding method of the lossless encoding unit 106.
- the lossless decoding unit 202 supplies the quantized coefficient data of the difference image obtained by decoding to the inverse quantization unit 203.
- the lossless decoding unit 202 determines whether the intra prediction mode or the inter prediction mode is selected as the optimal prediction mode.
- the lossless decoding unit 202 supplies information regarding the optimal prediction mode to the mode determined to be selected from the intra prediction unit 211 and the motion prediction / compensation unit 212. That is, for example, when the inter prediction mode is selected as the optimal prediction mode in the image encoding device 100, information regarding the optimal prediction mode is supplied to the motion prediction / compensation unit 212.
- the inverse quantization unit 203 uses the initial value of the quantization parameter for the tile from the quantization initial value decoding unit 222 to convert the quantized coefficient data obtained by decoding by the lossless decoding unit 202 in FIG. Inverse quantization is performed by a method corresponding to the quantization method of the quantization unit 105, and the obtained coefficient data is supplied to the inverse orthogonal transform unit 204.
- the inverse orthogonal transform unit 204 performs inverse orthogonal transform on the coefficient data supplied from the inverse quantization unit 203 in a method corresponding to the orthogonal transform method of the orthogonal transform unit 104 in FIG.
- the inverse orthogonal transform unit 204 obtains decoded residual data corresponding to the residual data before being orthogonally transformed in the image coding apparatus 100 by the inverse orthogonal transform process.
- the decoded residual data obtained by the inverse orthogonal transform is supplied to the calculation unit 205.
- a prediction image is supplied to the calculation unit 205 from the intra prediction unit 211 or the motion prediction / compensation unit 212 via the selection unit 213.
- the calculation unit 205 adds the decoded residual data and the prediction image, and obtains decoded image data corresponding to the image data before the prediction image is subtracted by the calculation unit 103 of the image encoding device 100.
- the arithmetic unit 205 supplies the decoded image data to the deblock filter 206.
- the deblock filter 206 performs deblock filter processing on the supplied decoded image as appropriate, and supplies it to the screen rearrangement buffer 207.
- the deblocking filter 206 removes block distortion of the decoded image by performing a deblocking filter process on the decoded image.
- the deblock filter 206 supplies the filter processing result (the decoded image after the filter processing) to the screen rearrangement buffer 207 and the frame memory 209. Note that the decoded image output from the calculation unit 205 can be supplied to the screen rearrangement buffer 207 and the frame memory 209 without going through the deblocking filter 206. That is, the filtering process by the deblocking filter 206 can be omitted.
- the screen rearrangement buffer 207 rearranges images. Although not shown in FIG. 27, field coding information is supplied from the tile division decoding unit 221 or the like, and the screen rearrangement buffer 207 rearranges images based on the field coding information. . That is, the order of frames rearranged for the encoding order by the screen rearrangement buffer 102 in FIG. 1 is rearranged in the original display order.
- the D / A conversion unit 208 D / A converts the image supplied from the screen rearrangement buffer 207, outputs it to a display (not shown), and displays it.
- the frame memory 209 stores the supplied decoded image, and the stored decoded image is referred to as a reference image at a predetermined timing or based on an external request such as the intra prediction unit 211 or the motion prediction / compensation unit 212. To the selection unit 210.
- the selection unit 210 selects the supply destination of the reference image supplied from the frame memory 209.
- the selection unit 210 supplies the reference image supplied from the frame memory 209 to the intra prediction unit 211 when decoding an intra-coded image.
- the selection unit 210 also supplies the reference image supplied from the frame memory 209 to the motion prediction / compensation unit 212 when decoding an inter-coded image.
- the intra prediction unit 211 is appropriately supplied from the lossless decoding unit 202 with information indicating the intra prediction mode obtained by decoding the header information.
- the intra prediction unit 211 performs intra prediction using the reference image acquired from the frame memory 209 in the intra prediction mode used in the intra prediction unit 114 in FIG. 1, and generates a predicted image.
- the intra prediction unit 211 supplies the generated predicted image to the selection unit 213.
- the motion prediction / compensation unit 212 acquires information (optimum prediction mode information, motion vector information, reference image information, etc.) obtained by decoding the header information from the lossless decoding unit 202.
- the motion prediction / compensation unit 212 performs inter prediction using the reference image acquired from the frame memory 209 in the inter prediction mode indicated by the optimal prediction mode information acquired from the lossless decoding unit 202, and generates a predicted image.
- the selection unit 213 supplies the prediction image from the intra prediction unit 211 or the prediction image from the motion prediction / compensation unit 212 to the calculation unit 205.
- the arithmetic unit 205 adds the predicted image generated using the motion vector and the decoded residual data (difference image information) from the inverse orthogonal transform unit 204 to decode the original image. That is, the motion prediction / compensation unit 212, the lossless decoding unit 202, the inverse quantization unit 203, the inverse orthogonal transform unit 204, and the calculation unit 205 decode the encoded data using the motion vector to generate the original image. It is also a decryption unit.
- the tile division decoding unit 221 acquires tile division information from the lossless decoding unit 202 and supplies the tile division information to the quantization initial value decoding unit 222.
- the quantization initial value decoding unit 222 uses the tile division information from the tile division decoding unit 221 and the information on the quantization parameter initial value of the picture from the lossless decoding unit 202 to use the initial value of the quantization parameter for the tile. To rebuild.
- the quantization initial value decoding unit 222 supplies the quantization parameter initial value for the reconstructed tile to the inverse quantization unit 203.
- FIG. 28 is a block diagram illustrating a main configuration example of the lossless decoding unit 202 and the quantized initial value decoding unit 222 that execute the second method described above with reference to FIG. That is, the decoding block shown in FIG. 28 corresponds to the encoding block shown in FIG.
- the lossless decoding unit 202 is configured to include a syntax receiving unit 251.
- the tile division decoding unit 221 is configured to include a tile difference QP buffer 261, a picture initial QP buffer 262, and a tile initial QP reconstruction unit 263.
- the syntax receiving unit 251 acquires tile division information from the encoded stream, and supplies the acquired tile division information to the tile division decoding unit 221.
- the syntax receiving unit 251 acquires a picture initial QP, a tile initial difference QP, and the like as information on the quantization parameter initial value of the picture, and supplies the acquired picture initial QP to the quantization initial value decoding unit 222.
- the syntax receiving unit 251 acquires the picture initial QP from the sequence parameter set of the encoded stream, and supplies the acquired picture initial QP to the picture initial QP buffer 262. Also, the syntax receiving unit 251 acquires a tile initial difference QP (difference QP in FIG. 15) that is information on the tile initial QP from the vicinity of the head of each tile of the encoded stream, and the acquired tile initial difference QP is a tile difference.
- the QP buffer 261 is supplied.
- the tile difference QP buffer 261 refers to the tile division information from the tile division decoding unit 221 and transmits the tile initial difference QP from the syntax receiving unit 251 to the tile initial QP reconfiguration unit 263 at the timing of the tile boundary address. Supply.
- the picture initial QP buffer 262 supplies the picture initial QP from the syntax receiving unit 251 to the tile initial QP reconstruction unit 263.
- the tile initial QP reconstruction unit 263 reconstructs the tile initial QP using the tile initial difference QP from the tile difference QP buffer 261 and the picture initial QP from the picture initial QP buffer 262, and reconstructs the tile initial The QP is supplied to the inverse quantization unit 203.
- the difference between the initial value of the quantization parameter for the tile and the quantization parameter of the first CU in the tile is the information regarding the quantization parameter of the CU, as in the past.
- Set to the transform unit of the first CU in the tile Therefore, for example, in the inverse quantization unit 203, the quantization parameter of the first CU in the tile is reconstructed and used by using the tile initial QP from the tile initial QP reconstruction unit 263.
- the syntax receiving unit 251 acquires the initial value of the quantization parameter for the tile, the QP difference of the first CU of the tile, and the like as information on the initial value of the quantization parameter of the picture. .
- the syntax receiving unit 251 is different from the second method in that the picture initial QP is received from the picture parameter set in the encoded stream as the initial value of the quantization parameter for the tile.
- the syntax receiving unit 251 uses the tile initial difference QP as the QP difference of the first CU of the tile from the transform unit of the first CU of the corresponding tile in the encoded stream. Receiving is different from the second method.
- step S201 the quantization initial value decoding unit 222 receives information related to tile division of the picture from the syntax receiving unit 251 via the tile division decoding unit 221.
- step S202 the quantization initial value decoding unit 222 receives information on the quantization parameter initial value of the picture from the syntax reception unit 251.
- the tile difference QP buffer 261 receives the tile initial difference QP as one piece of information related to the quantization parameter initial value of the picture from the syntax receiving unit 251.
- the picture initial QP buffer 262 receives the picture initial QP from the syntax receiving unit 251 as one piece of information regarding the quantization parameter initial value of the picture.
- the tile difference QP buffer 261 supplies the received tile initial difference QP to the tile initial QP reconstruction unit 263.
- the picture initial QP buffer 262 supplies the received tile initial difference QP to the tile initial QP reconstruction unit 263.
- step S203 the tile initial QP reconstruction unit 263 reconstructs the quantization parameter initial value. That is, the tile initial QP reconstruction unit 263 reconstructs the tile initial QP by adding the tile initial difference QP and the picture initial QP. The tile initial QP reconstruction unit 263 supplies the reconstructed tile initial QP to the inverse quantization unit 203.
- each unit of the image decoding device 200 performs decoding processing of each tile by parallel processing.
- the decoding process of each tile is a decoding process such as a DCT coefficient and a motion vector for each tile below the slice header.
- the decoding process of each tile will be described later with reference to FIG.
- step S204 the information below each tile is decoded, the image encoded by the image encoding device 100 is reconstructed, and the process ends.
- step S221 the accumulation buffer 201 receives and accumulates the transmitted encoded stream.
- step S222 the lossless decoding unit 202 decodes the encoded stream (encoded difference image information) supplied from the accumulation buffer 201. That is, the I picture, P picture, and B picture encoded by the lossless encoding unit 106 in FIG. 1 are decoded.
- the lossless decoding unit 202 acquires, for example, prediction mode information, motion vector information, and the like. The lossless decoding unit 202 supplies the acquired information to the corresponding unit.
- step S223 the inverse quantization unit 203 inversely quantizes the quantized orthogonal transform coefficient obtained by the process in step S222.
- the inverse quantization unit 203 uses the tile initial QP from the quantization initial value decoding unit 222 as the quantization parameter of the first CU of the tile or the first tile calculated from the tile initial QP. Used as quantization parameter of CU. Note that, as described above with reference to FIG. 11, for example, the quantization parameter of the adjacent region having a correlation among the adjacent regions is used as the quantization parameter of the CU having a size smaller than the designated size.
- the inverse orthogonal transform unit 204 performs inverse orthogonal transform on the orthogonal transform coefficient inversely quantized in step S223.
- step S225 the lossless decoding unit 202 determines whether or not the encoded data to be processed is intra-encoded based on the information regarding the optimal prediction mode decoded in step S222. If it is determined that intra-encoding has been performed, the process proceeds to step S226.
- step S226 the intra prediction unit 211 acquires intra prediction mode information.
- step S227 the intra prediction unit 211 performs intra prediction using the intra prediction mode information acquired in step S226, and generates a predicted image.
- step S226 If it is determined in step S226 that the encoded data to be processed is not intra-encoded, that is, inter-encoded, the process proceeds to step S228.
- the motion prediction / compensation unit 212 acquires inter prediction mode information in step S228, and acquires motion vector information in step S229.
- step S230 the motion prediction / compensation unit 212 generates a prediction image using the motion vector acquired in step S229 in the inter prediction mode acquired in step S228.
- the generated predicted image is supplied to the selection unit 213.
- step S231 the selection unit 213 selects the predicted image generated in step S227 or step S230.
- step S232 the calculation unit 205 adds the predicted image selected in step S231 to the difference image information obtained by the inverse orthogonal transform in step S224.
- the original image is decoded. That is, a motion vector is used to generate a predicted image, and the generated predicted image and the difference image information from the inverse orthogonal transform unit 204 are added to decode the original image.
- step S233 the deblock filter 206 appropriately performs deblock filter processing on the decoded image obtained in step S232.
- step S234 the screen rearrangement buffer 207 rearranges the images filtered in step S233. That is, the order of frames rearranged for encoding by the screen rearrangement buffer 102 of the image encoding device 100 is rearranged to the original display order.
- step S235 the D / A converter 208 D / A converts the image in which the frame order is rearranged in step S234. This image is output to a display (not shown), and the image is displayed.
- step S236 the frame memory 209 stores the image filtered in step S235.
- step S236 ends, the decryption process ends.
- FIG. 31 is a block diagram illustrating a main configuration example of the lossless decoding unit 202 and the quantized initial value decoding unit 222 that execute the first B method described above with reference to FIG. 13B. That is, the decoding side block shown in FIG. 31 corresponds to the encoding side block shown in FIG.
- the lossless decoding unit 202 is configured to include a syntax receiving unit 281.
- the quantization initial value decoding unit 222 includes a tile difference QP buffer 291, a first slice difference QP buffer 292, a picture initial QP buffer 293, a first slice initial QP reconstruction unit 294, and a tile initial QP reconstruction unit 295. Configured.
- the syntax receiving unit 281 acquires tile division information from the encoded stream, and supplies the acquired tile division information to the tile division decoding unit 221.
- the syntax receiving unit 281 acquires the first slice difference QP as the initial value of the quantization parameter for the tile from the first slice header of the encoded stream, and uses the acquired first slice difference QP as the first slice difference QP buffer. 292.
- the syntax receiving unit 281 acquires the picture initial QP from the picture parameter set of the encoded stream as information for reconstructing the initial value of the quantization parameter for the tile, and supplies the picture initial QP to the picture initial QP buffer 293. Also, the syntax receiving unit 281 receives, from the transform unit of the first CU of each tile of the encoded stream, the tile initial difference QP (the tile of FIG. 19) as information on the tile initial QP as the QP difference of the first CU of the tile. Difference QP) is acquired. The syntax receiving unit 281 supplies the acquired tile initial difference QP to the tile difference QP buffer 291.
- the tile difference QP buffer 291 refers to the tile division information from the tile division decoding unit 221 and sends the tile initial difference QP from the syntax receiving unit 281 to the tile initial QP reconstruction unit 295 at the timing of the tile boundary address. Supply.
- the first slice difference QP buffer 292 supplies the first slice difference QP from the syntax receiving unit 281 to the first slice initial QP reconstruction unit 294.
- the picture initial QP buffer 293 supplies the picture initial QP from the syntax receiving unit 281 to the first slice initial QP reconstructing unit 294.
- the first slice initial QP reconstruction unit 294 reconstructs the first slice initial QP using the first slice difference QP from the first slice difference QP buffer 292 and the picture initial QP from the picture initial QP buffer 293. To do.
- the tile initial QP reconstruction unit 295 supplies the reconstructed first slice initial QP to the inverse quantization unit 203.
- the tile initial QP reconstruction unit 295 reconstructs the tile initial QP using the tile initial difference QP from the tile difference QP buffer 291 and the first slice initial QP from the first slice initial QP reconstruction unit 294.
- the reconstructed tile initial QP is supplied to the inverse quantization unit 203.
- each tile initial QP (of the tile) is based on the initial QP of the first slice in the picture that is the initial value of the quantization parameter for the tile.
- the first CU's quantization parameter) is reconstructed.
- step S251 the quantization initial value decoding unit 222 receives information related to tile division of the picture from the syntax reception unit 281 via the tile division decoding unit 221.
- step S252 the quantization initial value decoding unit 222 receives information on the quantization parameter initial value of the picture from the syntax reception unit 281.
- the tile difference QP buffer 291 receives, from the syntax receiving unit 281, a tile initial difference QP (quantization parameter of the first CU of the tile) as one piece of information regarding the quantization parameter initial value of the picture.
- the first slice difference QP buffer 292 receives the first slice difference QP from the syntax receiving unit 281 as one piece of information regarding the quantization parameter initial value of the picture.
- the picture initial QP buffer 293 receives the picture initial QP from the syntax receiving unit 251 as one piece of information regarding the quantization parameter initial value of the picture.
- the tile difference QP buffer 261 supplies the received tile initial difference QP to the tile initial QP reconstruction unit 295.
- the first slice difference QP buffer 292 supplies the received first slice difference QP to the first slice initial QP reconstruction unit 294.
- the picture initial QP buffer 262 supplies the received first slice difference QP to the first slice initial QP reconstruction unit 294.
- the first slice initial QP reconstruction unit 294 and the tile initial QP reconstruction unit 295 reconstruct the quantization parameter initial value in steps S253 and S254.
- the first slice initial QP reconstructing unit 294 reconstructs the first slice initial QP by adding the first slice difference QP and the picture initial QP.
- the first slice initial QP reconstruction unit 294 supplies the reconstructed first slice initial QP to the tile initial QP reconstruction unit 295.
- step S254 the tile initial QP reconstruction unit 295 reconstructs the tile initial QP by adding the tile initial difference QP and the first slice initial QP.
- the tile initial QP reconstruction unit 263 supplies the reconstructed tile initial QP to the inverse quantization unit 203 as the quantization parameter of the first CU of the tile.
- step S255 each unit of the image decoding apparatus 200 performs the decoding process of each tile by the parallel processing described above with reference to FIG.
- step S255 the information below each tile is decoded, the image encoded by the image encoding device 100 is reconstructed, and the process ends.
- the image decoding apparatus 200 can correctly decode the encoded data encoded by the image encoding apparatus 100, and can realize improvement in encoding efficiency.
- the initial value of the quantization parameter for the tile is set so as to ensure the independence of the processing regarding the quantization parameter between tiles, and information on the initial value of the quantization parameter is transmitted. It will be.
- the image decoding apparatus 200 can perform processing related to the quantization parameter between tiles in parallel. . Thereby, the decoding processing efficiency is improved.
- the present technology can also be applied to apparatuses using other encoding methods as long as the apparatus performs encoding processing and decoding processing that performs tile division and performs parallel processing based on tiles.
- the present technology is, for example, MPEG, H.264.
- image information bitstream
- orthogonal transform such as discrete cosine transform and motion compensation, such as 26x
- network media such as satellite broadcasting, cable television, the Internet, or mobile phones.
- the present invention can be applied to an image encoding device and an image decoding device used in the above.
- the present technology can be applied to an image encoding device and an image decoding device that are used when processing is performed on a storage medium such as an optical disk, a magnetic disk, and a flash memory.
- the present technology can also be applied to motion prediction / compensation devices included in such image encoding devices and image decoding devices.
- FIG. 33 shows an example of a multi-view image encoding method.
- the multi-viewpoint image includes a plurality of viewpoint images, and a predetermined one viewpoint image among the plurality of viewpoints is designated as the base view image.
- Each viewpoint image other than the base view image is treated as a non-base view image.
- parameters such as the initial value (information on the quantization parameter) and tile division information described above can be set in each view (same view).
- tile division information is omitted as an example of parameters, and only initial values of quantization parameters are described.
- the initial value of the quantization parameter set in another view can be shared.
- the initial value of the quantization parameter set in the base view is used in at least one non-base view.
- FIG. 34 is a diagram illustrating a multi-view image encoding apparatus that performs the multi-view image encoding described above. As illustrated in FIG. 34, the multi-view image encoding apparatus 600 includes an encoding unit 601, an encoding unit 602, and a multiplexing unit 603.
- the encoding unit 601 encodes the base view image and generates a base view image encoded stream.
- the encoding unit 602 encodes the non-base view image and generates a non-base view image encoded stream.
- the multiplexing unit 603 multiplexes the base view image encoded stream generated by the encoding unit 601 and the non-base view image encoded stream generated by the encoding unit 602 to generate a multi-view image encoded stream. To do.
- the image encoding device 100 (FIG. 1) can be applied to the encoding unit 601 and the encoding unit 602 of the multi-view image encoding device 600.
- multi-view image encoding apparatus 600 sets and transmits the initial value of the quantization parameter set by encoding unit 601 and the initial value of the quantization parameter set by encoding unit 602.
- the initial value of the quantization parameter set by the encoding unit 601 as described above may be set and transmitted so as to be shared by the encoding unit 601 and the encoding unit 602.
- the initial value of the quantization parameter set collectively by the encoding unit 602 may be set and transmitted so as to be shared by the encoding unit 601 and the encoding unit 602.
- FIG. 35 is a diagram illustrating a multi-view image decoding apparatus that performs the above-described multi-view image decoding.
- the multi-view image decoding apparatus 610 includes a demultiplexing unit 611, a decoding unit 612, and a decoding unit 613.
- the demultiplexing unit 611 demultiplexes the multi-view image encoded stream in which the base view image encoded stream and the non-base view image encoded stream are multiplexed, and the base view image encoded stream and the non-base view image The encoded stream is extracted.
- the decoding unit 612 decodes the base view image encoded stream extracted by the demultiplexing unit 611 to obtain a base view image.
- the decoding unit 613 decodes the non-base view image encoded stream extracted by the demultiplexing unit 611 to obtain a non-base view image.
- the image decoding device 200 (FIG. 27) can be applied to the decoding unit 612 and the decoding unit 613 of the multi-view image decoding device 610.
- the multi-view image decoding apparatus 610 includes an initial value of a quantization parameter set by the encoding unit 601 and decoded by the decoding unit 612, and a quantization parameter set by the encoding unit 602 and decoded by the decoding unit 613. Processing is performed using the initial value of.
- the initial value of the quantization parameter set by the encoding unit 601 (or the encoding unit 602) is set and transmitted so as to be shared by the encoding unit 601 and the encoding unit 602. There may be.
- processing is performed using the initial value of the quantization parameter set by encoding unit 601 (or encoding unit 602) and decoded by decoding unit 612 (or decoding unit 613). Is done.
- FIG. 36 shows an example of a multi-view image encoding method.
- a hierarchical image includes images of a plurality of layers (resolutions), and an image of a predetermined one layer among the plurality of resolutions is designated as a base layer image. Images in each layer other than the base layer image are treated as non-base layer images.
- the initial value of the quantization parameter described above can be set in each layer (same layer).
- each layer different layers can share the initial value of the quantization parameter set in another layer.
- the initial value of the quantization parameter set in the base layer is used in at least one non-base layer.
- FIG. 37 is a diagram illustrating a hierarchical image encoding apparatus that performs the above-described hierarchical image encoding.
- the hierarchical image encoding device 620 includes an encoding unit 621, an encoding unit 622, and a multiplexing unit 623.
- the encoding unit 621 encodes the base layer image and generates a base layer image encoded stream.
- the encoding unit 622 encodes the non-base layer image and generates a non-base layer image encoded stream.
- the multiplexing unit 623 multiplexes the base layer image encoded stream generated by the encoding unit 621 and the non-base layer image encoded stream generated by the encoding unit 622 to generate a hierarchical image encoded stream. .
- the image encoding device 100 (FIG. 1) can be applied to the encoding unit 621 and the encoding unit 622 of the hierarchical image encoding device 620.
- the hierarchical image encoding device 620 sets and transmits the initial value of the quantization parameter set by the encoding unit 621 and the initial value of the quantization parameter set by the encoding unit 602.
- the initial value of the quantization parameter set by the encoding unit 621 as described above may be set and transmitted so as to be shared by the encoding unit 621 and the encoding unit 622.
- the initial value of the quantization parameter set by the encoding unit 622 may be set and transmitted so as to be shared by the encoding unit 621 and the encoding unit 622.
- FIG. 38 is a diagram illustrating a hierarchical image decoding apparatus that performs the above-described hierarchical image decoding.
- the hierarchical image decoding device 630 includes a demultiplexing unit 631, a decoding unit 632, and a decoding unit 633.
- the demultiplexing unit 631 demultiplexes the hierarchical image encoded stream in which the base layer image encoded stream and the non-base layer image encoded stream are multiplexed, and the base layer image encoded stream and the non-base layer image code Stream.
- the decoding unit 632 decodes the base layer image encoded stream extracted by the demultiplexing unit 631 to obtain a base layer image.
- the decoding unit 633 decodes the non-base layer image encoded stream extracted by the demultiplexing unit 631 to obtain a non-base layer image.
- the image decoding device 200 (FIG. 27) can be applied to the decoding unit 632 and the decoding unit 633 of the hierarchical image decoding device 630.
- the hierarchical image decoding apparatus 630 includes an initial value of the quantization parameter set by the encoding unit 621 and decoded by the decoding unit 632, and the quantization parameter set by the encoding unit 622 and decoded by the decoding unit 633. Processing is performed using the initial value.
- the initial value of the quantization parameter set by the encoding unit 621 (or the encoding unit 622) is set and transmitted so as to be shared by the encoding unit 621 and the encoding unit 622. There may be.
- processing is performed using the initial value of the quantization parameter set by encoding unit 621 (or encoding unit 622) and decoded by decoding unit 632 (or decoding unit 633). Is done.
- the series of processes described above can be executed by hardware or can be executed by software.
- a program constituting the software is installed in the computer.
- the computer includes a computer incorporated in dedicated hardware, a general-purpose personal computer capable of executing various functions by installing various programs, and the like.
- FIG. 39 is a block diagram illustrating a configuration example of hardware of a computer that executes the series of processes described above according to a program.
- a CPU Central Processing Unit
- ROM Read Only Memory
- RAM Random Access Memory
- an input / output interface 805 is connected to the bus 804.
- An input unit 806, an output unit 807, a storage unit 808, a communication unit 809, and a drive 810 are connected to the input / output interface 805.
- the input unit 806 includes a keyboard, a mouse, a microphone, and the like.
- the output unit 807 includes a display, a speaker, and the like.
- the storage unit 808 includes a hard disk, a nonvolatile memory, and the like.
- the communication unit 809 includes a network interface or the like.
- the drive 810 drives a removable medium 811 such as a magnetic disk, an optical disk, a magneto-optical disk, or a semiconductor memory.
- the CPU 801 loads the program stored in the storage unit 808 to the RAM 803 via the input / output interface 805 and the bus 804 and executes the program, for example. Is performed.
- the program executed by the computer 800 can be provided by being recorded in, for example, a removable medium 811 as a package medium or the like.
- the program can be provided via a wired or wireless transmission medium such as a local area network, the Internet, or digital satellite broadcasting.
- the program can be installed in the storage unit 808 via the input / output interface 805 by attaching the removable medium 811 to the drive 810.
- the program can be received by the communication unit 809 via a wired or wireless transmission medium and installed in the storage unit 808.
- the program can be installed in the ROM 802 or the storage unit 808 in advance.
- the program executed by the computer may be a program that is processed in time series in the order described in this specification, or in parallel or at a necessary timing such as when a call is made. It may be a program for processing.
- the step of describing the program recorded on the recording medium is not limited to the processing performed in chronological order according to the described order, but may be performed in parallel or It also includes processes that are executed individually.
- system represents the entire apparatus composed of a plurality of devices (apparatuses).
- the configuration described as one device (or processing unit) may be divided and configured as a plurality of devices (or processing units).
- the configurations described above as a plurality of devices (or processing units) may be combined into a single device (or processing unit).
- a configuration other than that described above may be added to the configuration of each device (or each processing unit).
- a part of the configuration of a certain device (or processing unit) may be included in the configuration of another device (or other processing unit). . That is, the present technology is not limited to the above-described embodiment, and various modifications can be made without departing from the gist of the present technology.
- An image encoding device and an image decoding device include a transmitter or a receiver in optical broadcasting, satellite broadcasting, cable broadcasting such as cable TV, distribution on the Internet, and distribution to terminals by cellular communication, etc.
- the present invention can be applied to various electronic devices such as a recording device that records an image on a medium such as a magnetic disk and a flash memory, or a playback device that reproduces an image from these storage media.
- a recording device that records an image on a medium such as a magnetic disk and a flash memory
- a playback device that reproduces an image from these storage media.
- FIG. 40 illustrates an example of a schematic configuration of a television device to which the above-described embodiment is applied.
- the television apparatus 900 includes an antenna 901, a tuner 902, a demultiplexer 903, a decoder 904, a video signal processing unit 905, a display unit 906, an audio signal processing unit 907, a speaker 908, an external interface 909, a control unit 910, a user interface 911, And a bus 912.
- Tuner 902 extracts a signal of a desired channel from a broadcast signal received via antenna 901, and demodulates the extracted signal. Then, the tuner 902 outputs the encoded bit stream obtained by the demodulation to the demultiplexer 903. In other words, the tuner 902 serves as a transmission unit in the television apparatus 900 that receives an encoded stream in which an image is encoded.
- the demultiplexer 903 separates the video stream and audio stream of the viewing target program from the encoded bit stream, and outputs each separated stream to the decoder 904. Further, the demultiplexer 903 extracts auxiliary data such as EPG (Electronic Program Guide) from the encoded bit stream, and supplies the extracted data to the control unit 910. Note that the demultiplexer 903 may perform descrambling when the encoded bit stream is scrambled.
- EPG Electronic Program Guide
- the decoder 904 decodes the video stream and audio stream input from the demultiplexer 903. Then, the decoder 904 outputs the video data generated by the decoding process to the video signal processing unit 905. In addition, the decoder 904 outputs audio data generated by the decoding process to the audio signal processing unit 907.
- the video signal processing unit 905 reproduces the video data input from the decoder 904 and causes the display unit 906 to display the video.
- the video signal processing unit 905 may cause the display unit 906 to display an application screen supplied via a network.
- the video signal processing unit 905 may perform additional processing such as noise removal on the video data according to the setting.
- the video signal processing unit 905 may generate a GUI (Graphical User Interface) image such as a menu, a button, or a cursor, and superimpose the generated image on the output image.
- GUI Graphic User Interface
- the display unit 906 is driven by a drive signal supplied from the video signal processing unit 905, and displays an image on a video screen of a display device (for example, a liquid crystal display, a plasma display, or an OELD (Organic ElectroLuminescence Display) (organic EL display)). Or an image is displayed.
- a display device for example, a liquid crystal display, a plasma display, or an OELD (Organic ElectroLuminescence Display) (organic EL display)). Or an image is displayed.
- the audio signal processing unit 907 performs reproduction processing such as D / A conversion and amplification on the audio data input from the decoder 904, and outputs audio from the speaker 908.
- the audio signal processing unit 907 may perform additional processing such as noise removal on the audio data.
- the external interface 909 is an interface for connecting the television apparatus 900 to an external device or a network.
- a video stream or an audio stream received via the external interface 909 may be decoded by the decoder 904. That is, the external interface 909 also has a role as a transmission unit in the television apparatus 900 that receives an encoded stream in which an image is encoded.
- the control unit 910 includes a processor such as a CPU and memories such as a RAM and a ROM.
- the memory stores a program executed by the CPU, program data, EPG data, data acquired via a network, and the like.
- the program stored in the memory is read and executed by the CPU when the television apparatus 900 is activated.
- the CPU executes the program to control the operation of the television device 900 according to an operation signal input from the user interface 911, for example.
- the user interface 911 is connected to the control unit 910.
- the user interface 911 includes, for example, buttons and switches for the user to operate the television device 900, a remote control signal receiving unit, and the like.
- the user interface 911 detects an operation by the user via these components, generates an operation signal, and outputs the generated operation signal to the control unit 910.
- the bus 912 connects the tuner 902, the demultiplexer 903, the decoder 904, the video signal processing unit 905, the audio signal processing unit 907, the external interface 909, and the control unit 910 to each other.
- the decoder 904 has the function of the image decoding apparatus according to the above-described embodiment. As a result, when the image is decoded by the television device 900, the decoding process can be efficiently performed on the tile-divided picture.
- FIG. 41 shows an example of a schematic configuration of a mobile phone to which the above-described embodiment is applied.
- a mobile phone 920 includes an antenna 921, a communication unit 922, an audio codec 923, a speaker 924, a microphone 925, a camera unit 926, an image processing unit 927, a demultiplexing unit 928, a recording / reproducing unit 929, a display unit 930, a control unit 931, an operation A portion 932 and a bus 933.
- the antenna 921 is connected to the communication unit 922.
- the speaker 924 and the microphone 925 are connected to the audio codec 923.
- the operation unit 932 is connected to the control unit 931.
- the bus 933 connects the communication unit 922, the audio codec 923, the camera unit 926, the image processing unit 927, the demultiplexing unit 928, the recording / reproducing unit 929, the display unit 930, and the control unit 931 to each other.
- the mobile phone 920 has various operation modes including a voice call mode, a data communication mode, a shooting mode, and a videophone mode, and is used for sending and receiving voice signals, sending and receiving e-mail or image data, taking images, and recording data. Perform the action.
- the analog voice signal generated by the microphone 925 is supplied to the voice codec 923.
- the audio codec 923 converts an analog audio signal into audio data, A / D converts the compressed audio data, and compresses it. Then, the audio codec 923 outputs the compressed audio data to the communication unit 922.
- the communication unit 922 encodes and modulates the audio data and generates a transmission signal. Then, the communication unit 922 transmits the generated transmission signal to a base station (not shown) via the antenna 921. In addition, the communication unit 922 amplifies a radio signal received via the antenna 921 and performs frequency conversion to acquire a received signal.
- the communication unit 922 demodulates and decodes the received signal to generate audio data, and outputs the generated audio data to the audio codec 923.
- the audio codec 923 decompresses the audio data and performs D / A conversion to generate an analog audio signal. Then, the audio codec 923 supplies the generated audio signal to the speaker 924 to output audio.
- the control unit 931 generates character data constituting the e-mail in response to an operation by the user via the operation unit 932.
- the control unit 931 causes the display unit 930 to display characters.
- the control unit 931 generates e-mail data in response to a transmission instruction from the user via the operation unit 932, and outputs the generated e-mail data to the communication unit 922.
- the communication unit 922 encodes and modulates email data and generates a transmission signal. Then, the communication unit 922 transmits the generated transmission signal to a base station (not shown) via the antenna 921.
- the communication unit 922 amplifies a radio signal received via the antenna 921 and performs frequency conversion to acquire a received signal.
- the communication unit 922 demodulates and decodes the received signal to restore the email data, and outputs the restored email data to the control unit 931.
- the control unit 931 displays the content of the electronic mail on the display unit 930 and stores the electronic mail data in the storage medium of the recording / reproducing unit 929.
- the recording / reproducing unit 929 has an arbitrary readable / writable storage medium.
- the storage medium may be a built-in storage medium such as a RAM or a flash memory, or an externally mounted type such as a hard disk, magnetic disk, magneto-optical disk, optical disk, USB (Universal Serial Bus) memory, or memory card. It may be a storage medium.
- the camera unit 926 images a subject to generate image data, and outputs the generated image data to the image processing unit 927.
- the image processing unit 927 encodes the image data input from the camera unit 926 and stores the encoded stream in the storage medium of the storage / playback unit 929.
- the demultiplexing unit 928 multiplexes the video stream encoded by the image processing unit 927 and the audio stream input from the audio codec 923, and the multiplexed stream is the communication unit 922. Output to.
- the communication unit 922 encodes and modulates the stream and generates a transmission signal. Then, the communication unit 922 transmits the generated transmission signal to a base station (not shown) via the antenna 921.
- the communication unit 922 amplifies a radio signal received via the antenna 921 and performs frequency conversion to acquire a received signal.
- These transmission signal and reception signal may include an encoded bit stream.
- the communication unit 922 demodulates and decodes the received signal to restore the stream, and outputs the restored stream to the demultiplexing unit 928.
- the demultiplexing unit 928 separates the video stream and the audio stream from the input stream, and outputs the video stream to the image processing unit 927 and the audio stream to the audio codec 923.
- the image processing unit 927 decodes the video stream and generates video data.
- the video data is supplied to the display unit 930, and a series of images is displayed on the display unit 930.
- the audio codec 923 decompresses the audio stream and performs D / A conversion to generate an analog audio signal. Then, the audio codec 923 supplies the generated audio signal to the speaker 924 to output audio.
- the image processing unit 927 has the functions of the image encoding device and the image decoding device according to the above-described embodiment. Accordingly, when encoding and decoding an image with the mobile phone 920, decoding processing can be efficiently performed on a tile-divided picture.
- FIG. 42 shows an example of a schematic configuration of a recording / reproducing apparatus to which the above-described embodiment is applied.
- the recording / reproducing device 940 encodes audio data and video data of a received broadcast program and records the encoded data on a recording medium.
- the recording / reproducing device 940 may encode audio data and video data acquired from another device and record them on a recording medium, for example.
- the recording / reproducing device 940 reproduces data recorded on the recording medium on a monitor and a speaker, for example, in accordance with a user instruction. At this time, the recording / reproducing device 940 decodes the audio data and the video data.
- the recording / reproducing apparatus 940 includes a tuner 941, an external interface 942, an encoder 943, an HDD (Hard Disk Drive) 944, a disk drive 945, a selector 946, a decoder 947, an OSD (On-Screen Display) 948, a control unit 949, and a user interface. 950.
- Tuner 941 extracts a signal of a desired channel from a broadcast signal received via an antenna (not shown), and demodulates the extracted signal. Then, the tuner 941 outputs the encoded bit stream obtained by the demodulation to the selector 946. That is, the tuner 941 has a role as a transmission unit in the recording / reproducing apparatus 940.
- the external interface 942 is an interface for connecting the recording / reproducing apparatus 940 to an external device or a network.
- the external interface 942 may be, for example, an IEEE1394 interface, a network interface, a USB interface, or a flash memory interface.
- video data and audio data received via the external interface 942 are input to the encoder 943. That is, the external interface 942 serves as a transmission unit in the recording / reproducing device 940.
- the encoder 943 encodes video data and audio data when the video data and audio data input from the external interface 942 are not encoded. Then, the encoder 943 outputs the encoded bit stream to the selector 946.
- the HDD 944 records an encoded bit stream in which content data such as video and audio is compressed, various programs, and other data on an internal hard disk. Further, the HDD 944 reads out these data from the hard disk when reproducing video and audio.
- the disk drive 945 performs recording and reading of data to and from the mounted recording medium.
- the recording medium mounted on the disk drive 945 is, for example, a DVD disk (DVD-Video, DVD-RAM, DVD-R, DVD-RW, DVD + R, DVD + RW, etc.) or a Blu-ray (registered trademark) disk. It may be.
- the selector 946 selects an encoded bit stream input from the tuner 941 or the encoder 943 when recording video and audio, and outputs the selected encoded bit stream to the HDD 944 or the disk drive 945. In addition, the selector 946 outputs the encoded bit stream input from the HDD 944 or the disk drive 945 to the decoder 947 during video and audio reproduction.
- the decoder 947 decodes the encoded bit stream and generates video data and audio data. Then, the decoder 947 outputs the generated video data to the OSD 948. The decoder 904 outputs the generated audio data to an external speaker.
- OSD 948 reproduces the video data input from the decoder 947 and displays the video. Further, the OSD 948 may superimpose a GUI image such as a menu, a button, or a cursor on the video to be displayed.
- the control unit 949 includes a processor such as a CPU and memories such as a RAM and a ROM.
- the memory stores a program executed by the CPU, program data, and the like.
- the program stored in the memory is read and executed by the CPU when the recording / reproducing apparatus 940 is activated, for example.
- the CPU controls the operation of the recording / reproducing apparatus 940 in accordance with an operation signal input from the user interface 950, for example, by executing the program.
- the user interface 950 is connected to the control unit 949.
- the user interface 950 includes, for example, buttons and switches for the user to operate the recording / reproducing device 940, a remote control signal receiving unit, and the like.
- the user interface 950 detects an operation by the user via these components, generates an operation signal, and outputs the generated operation signal to the control unit 949.
- the encoder 943 has the function of the image encoding apparatus according to the above-described embodiment.
- the decoder 947 has the function of the image decoding apparatus according to the above-described embodiment.
- FIG. 43 shows an example of a schematic configuration of an imaging apparatus to which the above-described embodiment is applied.
- the imaging device 960 images a subject to generate an image, encodes the image data, and records it on a recording medium.
- the imaging device 960 includes an optical block 961, an imaging unit 962, a signal processing unit 963, an image processing unit 964, a display unit 965, an external interface 966, a memory 967, a media drive 968, an OSD 969, a control unit 970, a user interface 971, and a bus. 972.
- the optical block 961 is connected to the imaging unit 962.
- the imaging unit 962 is connected to the signal processing unit 963.
- the display unit 965 is connected to the image processing unit 964.
- the user interface 971 is connected to the control unit 970.
- the bus 972 connects the image processing unit 964, the external interface 966, the memory 967, the media drive 968, the OSD 969, and the control unit 970 to each other.
- the optical block 961 includes a focus lens and a diaphragm mechanism.
- the optical block 961 forms an optical image of the subject on the imaging surface of the imaging unit 962.
- the imaging unit 962 includes an image sensor such as a CCD (Charge-Coupled Device) or a CMOS (Complementary Metal-Oxide Semiconductor), and converts an optical image formed on the imaging surface into an image signal as an electrical signal by photoelectric conversion. Then, the imaging unit 962 outputs the image signal to the signal processing unit 963.
- CCD Charge-Coupled Device
- CMOS Complementary Metal-Oxide Semiconductor
- the signal processing unit 963 performs various camera signal processing such as knee correction, gamma correction, and color correction on the image signal input from the imaging unit 962.
- the signal processing unit 963 outputs the image data after the camera signal processing to the image processing unit 964.
- the image processing unit 964 encodes the image data input from the signal processing unit 963 and generates encoded data. Then, the image processing unit 964 outputs the generated encoded data to the external interface 966 or the media drive 968. The image processing unit 964 also decodes encoded data input from the external interface 966 or the media drive 968 to generate image data. Then, the image processing unit 964 outputs the generated image data to the display unit 965. In addition, the image processing unit 964 may display the image by outputting the image data input from the signal processing unit 963 to the display unit 965. Further, the image processing unit 964 may superimpose display data acquired from the OSD 969 on an image output to the display unit 965.
- the OSD 969 generates a GUI image such as a menu, a button, or a cursor, and outputs the generated image to the image processing unit 964.
- the external interface 966 is configured as a USB input / output terminal, for example.
- the external interface 966 connects the imaging device 960 and a printer, for example, when printing an image.
- a drive is connected to the external interface 966 as necessary.
- a removable medium such as a magnetic disk or an optical disk is attached to the drive, and a program read from the removable medium can be installed in the imaging device 960.
- the external interface 966 may be configured as a network interface connected to a network such as a LAN or the Internet. That is, the external interface 966 has a role as a transmission unit in the imaging device 960.
- the recording medium mounted on the media drive 968 may be any readable / writable removable medium such as a magnetic disk, a magneto-optical disk, an optical disk, or a semiconductor memory.
- a recording medium may be fixedly mounted on the media drive 968, and a non-portable storage unit such as an internal hard disk drive or an SSD (Solid State Drive) may be configured.
- the control unit 970 includes a processor such as a CPU and memories such as a RAM and a ROM.
- the memory stores a program executed by the CPU, program data, and the like.
- the program stored in the memory is read and executed by the CPU when the imaging device 960 is activated, for example.
- the CPU controls the operation of the imaging device 960 according to an operation signal input from the user interface 971 by executing the program.
- the user interface 971 is connected to the control unit 970.
- the user interface 971 includes, for example, buttons and switches for the user to operate the imaging device 960.
- the user interface 971 detects an operation by the user via these components, generates an operation signal, and outputs the generated operation signal to the control unit 970.
- the image processing unit 964 has the functions of the image encoding device and the image decoding device according to the above-described embodiment. Accordingly, when encoding and decoding an image by the imaging device 960, decoding processing can be efficiently performed on the tile-divided picture.
- Scalable encoding is used for selection of data to be transmitted, for example, as in the example shown in FIG.
- the distribution server 1002 reads the scalable encoded data stored in the scalable encoded data storage unit 1001, and via the network 1003, the personal computer 1004, the AV device 1005, the tablet This is distributed to the terminal device such as the device 1006 and the mobile phone 1007.
- the distribution server 1002 selects and transmits encoded data of appropriate quality according to the capability of the terminal device, the communication environment, and the like. Even if the distribution server 1002 transmits high-quality data unnecessarily, a high-quality image is not always obtained in the terminal device, which may cause a delay or an overflow. Moreover, there is a possibility that the communication band is unnecessarily occupied or the load on the terminal device is unnecessarily increased. On the other hand, even if the distribution server 1002 transmits unnecessarily low quality data, there is a possibility that an image with sufficient image quality cannot be obtained in the terminal device. Therefore, the distribution server 1002 appropriately reads and transmits the scalable encoded data stored in the scalable encoded data storage unit 1001 as encoded data having an appropriate quality with respect to the capability and communication environment of the terminal device. .
- the scalable encoded data storage unit 1001 stores scalable encoded data (BL + EL) 1011 encoded in a scalable manner.
- the scalable encoded data (BL + EL) 1011 is encoded data including both a base layer and an enhancement layer, and is a data that can be decoded to obtain both a base layer image and an enhancement layer image. It is.
- the distribution server 1002 selects an appropriate layer according to the capability of the terminal device that transmits data, the communication environment, and the like, and reads the data of the layer. For example, the distribution server 1002 reads high-quality scalable encoded data (BL + EL) 1011 from the scalable encoded data storage unit 1001 and transmits it to the personal computer 1004 and the tablet device 1006 with high processing capability as they are. . On the other hand, for example, the distribution server 1002 extracts base layer data from the scalable encoded data (BL + EL) 1011 for the AV device 1005 and the cellular phone 1007 having a low processing capability, and performs scalable encoding. Although it is data of the same content as the data (BL + EL) 1011, it is transmitted as scalable encoded data (BL) 1012 having a lower quality than the scalable encoded data (BL + EL) 1011.
- BL scalable encoded data
- scalable encoded data By using scalable encoded data in this way, the amount of data can be easily adjusted, so that the occurrence of delays and overflows can be suppressed, and unnecessary increases in the load on terminal devices and communication media can be suppressed. be able to.
- scalable encoded data (BL + EL) 1011 since scalable encoded data (BL + EL) 1011 has reduced redundancy between layers, the amount of data can be reduced as compared with the case where encoded data of each layer is used as individual data. . Therefore, the storage area of the scalable encoded data storage unit 1001 can be used more efficiently.
- the hardware performance of the terminal device varies depending on the device.
- the application which a terminal device performs is also various, the capability of the software is also various.
- the network 1003 serving as a communication medium can be applied to any communication network including wired, wireless, or both, such as the Internet and a LAN (Local Area Network), and has various data transmission capabilities. Furthermore, there is a risk of change due to other communications.
- the distribution server 1002 communicates with the terminal device that is the data transmission destination before starting data transmission, and the hardware performance of the terminal device, the performance of the application (software) executed by the terminal device, etc. Information regarding the capability of the terminal device and information regarding the communication environment such as the available bandwidth of the network 1003 may be obtained. The distribution server 1002 may select an appropriate layer based on the information obtained here.
- the layer extraction may be performed by the terminal device.
- the personal computer 1004 may decode the transmitted scalable encoded data (BL + EL) 1011 and display a base layer image or an enhancement layer image. Further, for example, the personal computer 1004 extracts the base layer scalable encoded data (BL) 1012 from the transmitted scalable encoded data (BL + EL) 1011 and stores it or transfers it to another device. The base layer image may be displayed after decoding.
- the numbers of the scalable encoded data storage unit 1001, the distribution server 1002, the network 1003, and the terminal devices are arbitrary.
- the example in which the distribution server 1002 transmits data to the terminal device has been described, but the usage example is not limited to this.
- the data transmission system 1000 may be any system as long as it transmits a scalable encoded data to a terminal device by selecting an appropriate layer according to the capability of the terminal device or a communication environment. Can be applied to the system.
- the present technology is applied in the same manner as the application to the hierarchical encoding / decoding described above with reference to FIGS. Effects similar to those described above with reference to FIGS. 36 to 38 can be obtained.
- scalable coding is used for transmission via a plurality of communication media, for example, as in the example shown in FIG.
- a broadcasting station 1101 transmits base layer scalable encoded data (BL) 1121 by terrestrial broadcasting 1111.
- the broadcast station 1101 transmits enhancement layer scalable encoded data (EL) 1122 via an arbitrary network 1112 including a wired or wireless communication network or both (for example, packetized transmission).
- BL base layer scalable encoded data
- EL enhancement layer scalable encoded data
- the terminal apparatus 1102 has a reception function of the terrestrial broadcast 1111 broadcast by the broadcast station 1101 and receives base layer scalable encoded data (BL) 1121 transmitted via the terrestrial broadcast 1111.
- the terminal apparatus 1102 further has a communication function for performing communication via the network 1112, and receives enhancement layer scalable encoded data (EL) 1122 transmitted via the network 1112.
- BL base layer scalable encoded data
- EL enhancement layer scalable encoded data
- the terminal device 1102 decodes the base layer scalable encoded data (BL) 1121 acquired via the terrestrial broadcast 1111 according to, for example, a user instruction, and obtains or stores a base layer image. Or transmit to other devices.
- BL base layer scalable encoded data
- the terminal device 1102 for example, in response to a user instruction, the base layer scalable encoded data (BL) 1121 acquired via the terrestrial broadcast 1111 and the enhancement layer scalable encoded acquired via the network 1112 Data (EL) 1122 is combined to obtain scalable encoded data (BL + EL), or decoded to obtain an enhancement layer image, stored, or transmitted to another device.
- BL base layer scalable encoded data
- EL enhancement layer scalable encoded acquired via the network 1112 Data
- the scalable encoded data can be transmitted via a communication medium that is different for each layer, for example. Therefore, the load can be distributed, and the occurrence of delay and overflow can be suppressed.
- the communication medium used for transmission may be selected for each layer. For example, scalable encoded data (BL) 1121 of a base layer having a relatively large amount of data is transmitted via a communication medium having a wide bandwidth, and scalable encoded data (EL) 1122 having a relatively small amount of data is transmitted. You may make it transmit via a communication medium with a narrow bandwidth. Further, for example, the communication medium for transmitting the enhancement layer scalable encoded data (EL) 1122 is switched between the network 1112 and the terrestrial broadcast 1111 according to the available bandwidth of the network 1112. May be. Of course, the same applies to data of an arbitrary layer.
- the number of layers is arbitrary, and the number of communication media used for transmission is also arbitrary.
- the number of terminal devices 1102 serving as data distribution destinations is also arbitrary.
- broadcasting from the broadcasting station 1101 has been described as an example, but the usage example is not limited to this.
- the data transmission system 1100 can be applied to any system as long as it is a system that divides scalable encoded data into a plurality of layers and transmits them through a plurality of lines.
- the present technology is applied in the same manner as the application to the hierarchical encoding / decoding described above with reference to FIGS.
- the same effect as described above with reference to FIG. 38 can be obtained.
- scalable encoding is used for storing encoded data as in the example shown in FIG. 46, for example.
- the imaging device 1201 performs scalable coding on image data obtained by imaging the subject 1211, and as scalable coded data (BL + EL) 1221, a scalable coded data storage device 1202. To supply.
- the scalable encoded data storage device 1202 stores the scalable encoded data (BL + EL) 1221 supplied from the imaging device 1201 with quality according to the situation. For example, in the normal case, the scalable encoded data storage device 1202 extracts base layer data from the scalable encoded data (BL + EL) 1221, and the base layer scalable encoded data ( BL) 1222. On the other hand, for example, in the case of attention, the scalable encoded data storage device 1202 stores scalable encoded data (BL + EL) 1221 with high quality and a large amount of data.
- the scalable encoded data storage device 1202 can store an image with high image quality only when necessary, so that an increase in the amount of data can be achieved while suppressing a reduction in the value of the image due to image quality degradation. And the use efficiency of the storage area can be improved.
- the imaging device 1201 is a surveillance camera.
- the monitoring target for example, an intruder
- the content of the captured image is likely to be unimportant. Data
- the image quality is given priority and the image data (scalable) (Encoded data) is stored with high quality.
- whether it is normal time or attention time may be determined by the scalable encoded data storage device 1202 analyzing an image, for example.
- the imaging apparatus 1201 may make a determination, and the determination result may be transmitted to the scalable encoded data storage device 1202.
- the criterion for determining whether the time is normal or noting is arbitrary, and the content of the image as the criterion is arbitrary. Of course, conditions other than the contents of the image can also be used as the criterion. For example, it may be switched according to the volume or waveform of the recorded sound, may be switched at every predetermined time, or may be switched by an external instruction such as a user instruction.
- the number of states is arbitrary, for example, normal, slightly attention, attention, very attention, etc.
- three or more states may be switched.
- the upper limit number of states to be switched depends on the number of layers of scalable encoded data.
- the imaging apparatus 1201 may determine the number of scalable coding layers according to the state. For example, in a normal case, the imaging apparatus 1201 may generate base layer scalable encoded data (BL) 1222 with low quality and a small amount of data, and supply the scalable encoded data storage apparatus 1202 to the scalable encoded data storage apparatus 1202. Further, for example, when attention is paid, the imaging device 1201 generates scalable encoded data (BL + EL) 1221 having a high quality and a large amount of data, and supplies the scalable encoded data storage device 1202 to the scalable encoded data storage device 1202. May be.
- BL base layer scalable encoded data
- BL + EL scalable encoded data
- the monitoring camera has been described as an example.
- the use of the imaging system 1200 is arbitrary and is not limited to the monitoring camera.
- the present technology is applied in the same manner as the application to the hierarchical encoding / decoding described above with reference to FIGS.
- the effect similar to the effect mentioned above with reference to 38 can be acquired.
- various types of information such as information on quantization parameter initial values, tile division information, motion vector information, and prediction mode information are multiplexed into an encoded stream from the encoding side to the decoding side.
- An example of transmission has been described.
- the method for transmitting such information is not limited to such an example.
- these pieces of information may be transmitted or recorded as separate data associated with the encoded bitstream without being multiplexed into the encoded bitstream.
- the term “associate” means that an image (which may be a part of an image such as a slice or a block) included in the bitstream and information corresponding to the image can be linked at the time of decoding. Means.
- information may be transmitted on a transmission path different from that of the image (or bit stream).
- Information may be recorded on a recording medium (or another recording area of the same recording medium) different from the image (or bit stream).
- the information and the image (or bit stream) may be associated with each other in an arbitrary unit such as a plurality of frames, one frame, or a part of the frame.
- this technique can also take the following structures.
- Receive an encoded stream and an initial value of a quantization parameter for a tile set so as to ensure processing independence regarding the quantization parameter between tiles in a picture divided into a plurality of tiles of an image A receiving part;
- a quantization parameter reconstructing unit that reconstructs a quantization parameter of a coding unit constituting the tile using an initial value of a quantization parameter for the tile received by the receiving unit;
- An image processing apparatus comprising: a decoding unit that decodes the encoded stream received by the receiving unit using the quantization parameter reconstructed by the quantization parameter reconstructing unit and generates the image.
- the image processing device according to (1) or (2), wherein a value set in slice data in the encoded stream is received as an initial value of a quantization parameter of the tile.
- the reception unit receives information on a quantization parameter of a coding unit constituting the tile,
- the quantization parameter reconstructing unit reconstructs the quantization parameter of the coding unit constituting the tile using the information regarding the quantization parameter of the coding unit constituting the tile received by the receiving unit.
- the image processing apparatus according to any one of 1) to (6).
- the image processing apparatus is Receiving an encoded stream and an initial value of a quantization parameter for a tile set to ensure processing independence regarding the quantization parameter between tiles in a picture divided into a plurality of tiles of an image; Reconstructing the quantization parameters of the coding units constituting the tile using the received initial values of the quantization parameters for the tile; An image processing method for generating the image by decoding a received encoded stream using the reconstructed quantization parameter.
- an encoding unit that encodes an image and generates an encoded stream
- An initial value setting unit for setting an initial value of a quantization parameter for the tile so as to ensure independence of processing regarding the quantization parameter between tiles in a picture divided into a plurality of tiles of the image
- An image processing apparatus comprising: a transmission unit that transmits an initial value of a quantization parameter for a tile set by the initial value setting unit and an encoded stream generated by the encoding unit.
- the initial value setting unit sets an initial value of a quantization parameter of the tile at a predetermined position in the encoded stream.
- the transmission unit transmits information on a quantization parameter of a coding unit constituting the tile generated using an initial value of a quantization parameter of the tile. Any one of (9) to (14) An image processing apparatus according to 1. (16) The image processing apparatus Encode the image to generate an encoded stream, Setting an initial value of the quantization parameter for the tile so as to ensure independence of the processing regarding the quantization parameter between tiles in the picture divided into a plurality of tiles of the image; An image processing method for transmitting an initial value of a quantization parameter for a set tile and a generated encoded stream.
Landscapes
- Engineering & Computer Science (AREA)
- Multimedia (AREA)
- Signal Processing (AREA)
- Compression Or Coding Systems Of Tv Signals (AREA)
Abstract
Description
1.第1の実施の形態(画像符号化装置)
2.第2の実施の形態(画像復号装置)
3.第3の実施の形態(多視点画像符号化・多視点画像復号装置)
4.第4の実施の形態(階層画像符号化・階層画像復号装置)
5.第3の実施の形態(コンピュータ)
6.応用例
7.スケーラブル符号化の応用例
[画像符号化装置]
図1は、画像符号化装置の主な構成例を示すブロック図である。
次に、HEVC方式で規定されているコーディングユニットについて説明する。マクロブロックサイズを16画素×16画素とするのは、次世代符号化方式の対象となるような、UHD(Ultra High Definition;4000画素×2000画素)といった大きな画枠に対しては、最適ではない。
HEVC方式においては、非特許文献1において、タイルが、並列処理のために規定されている。タイルは、ピクチャの分割単位の1つである。タイルは、画像圧縮情報中のシーケンスパラメータセット(Sequence Parameter Set:SPS)またはピクチャパラメータセット(Picture Parameter Set:PPS)において、各タイルのロウ(行)サイズおよびコラム(列)サイズがLCU を単位に指定されている。
次に、HEVC方式のシンタクスを参照して、量子化パラメータに関する情報について説明する。
例えば、図12のAにおいて、ピクチャは、その横方向が、3つのタイルT0乃至タイルT2で分割されている。また、ピクチャは、ハッチがされていないスライス0と、ハッチがなされているスライス1に分割されており、その境界が、タイルT1に含まれている。
そこで、画像符号化装置100においては、タイル間の量子化パラメータに関する処理の独立性を確保するように、すなわち、タイル間の量子化パラメータに関する処理に依存性を解消すべく、タイルに対する量子化パラメータの初期値が設定される。
図15は、図14を参照して上述した第2の方法を実行する量子化初期値設定部122および可逆符号化部106の主な構成例を示すブロック図である。
次に、以上のような画像符号化装置100により実行される各処理の流れについて説明する。最初に、図16のフローチャートを参照して、第2の方法の場合の符号化処理の流れの例を説明する。
次に、図17のフローチャートを参照して、図16のステップS102における各タイルの量子化パラメータ初期値の設定処理について説明する。
次に、図18のフローチャートを参照して、図16のステップS104の各タイルの符号化処理について説明する。
図19は、図13のBを参照して上述した第1のBの方法を実行する量子化初期値設定部122および可逆符号化部106の主な構成例を示すブロック図である。
次に、図20のフローチャートを参照して、図16のステップS102における各タイルの量子化パラメータ初期値の設定処理の他の例について説明する。
図21乃至図23は、シーケンスパラメータセットのシンタクスの例を示す図であり、図24および図25は、ピクチャパラメータセットのシンタクスの例を示す図である。各行の左端の数字は説明のために付した行番号である。
図26は、スライスデータのシンタクスの例を示す図である。各行の左端の数字は説明のために付した行番号である。
[画像復号装置]
次に、以上のように符号化された符号化データ(符号化ストリーム)の復号について説明する。図27は、図1の画像符号化装置100に対応する画像復号装置の主な構成例を示すブロック図である。
図28は、図14を参照して上述した第2の方法を実行する可逆復号部202および量子化初期値復号部222の主な構成例を示すブロック図である。すなわち、図28に示される復号側のブロックは、図15に示された符号化側のブロックに対応する。
次に、以上のような画像復号装置200により実行される各処理の流れについて説明する。最初に、図29のフローチャートを参照して、第2の方法の場合の復号処理の流れの例を説明する。
次に、図30のフローチャートを参照して、図29のステップS204におけるタイルの復号処理について説明する。
図31は、図13のBを参照して上述した第1のBの方法を実行する可逆復号部202および量子化初期値復号部222の主な構成例を示すブロック図である。すなわち、図31に示される復号側のブロックは、図19に示された符号化側のブロックに対応する。
次に、図32のフローチャートを参照して、第1のBの方法の場合の復号処理の流れの例を説明する。
[多視点画像符号化・多視点画像復号への適用]
上述した一連の処理は、多視点画像符号化・多視点画像復号に適用することができる。図33は、多視点画像符号化方式の一例を示す。
図34は、上述した多視点画像符号化を行う多視点画像符号化装置を示す図である。図34に示されるように、多視点画像符号化装置600は、符号化部601、符号化部602、および多重化部603を有する。
図35は、上述した多視点画像復号を行う多視点画像復号装置を示す図である。図35に示されるように、多視点画像復号装置610は、逆多重化部611、復号部612、および復号部613を有する。
[階層画像符号化・階層画像復号への適用]
上述した一連の処理は、階層画像符号化・階層画像復号に適用することができる。図36は、多視点画像符号化方式の一例を示す。
図37は、上述した階層画像符号化を行う階層画像符号化装置を示す図である。図37に示されるように、階層画像符号化装置620は、符号化部621、符号化部622、および多重化部623を有する。
図38は、上述した階層画像復号を行う階層画像復号装置を示す図である。図38に示されるように、階層画像復号装置630は、逆多重化部631、復号部632、および復号部633を有する。
[コンピュータ]
上述した一連の処理は、ハードウエアにより実行することもできるし、ソフトウエアにより実行することもできる。一連の処理をソフトウエアにより実行する場合には、そのソフトウエアを構成するプログラムが、コンピュータにインストールされる。ここで、コンピュータには、専用のハードウエアに組み込まれているコンピュータや、各種のプログラムをインストールすることで、各種の機能を実行することが可能な汎用のパーソナルコンピュータなどが含まれる。
[第1の応用例:テレビジョン受像機]
図40は、上述した実施形態を適用したテレビジョン装置の概略的な構成の一例を示している。テレビジョン装置900は、アンテナ901、チューナ902、デマルチプレクサ903、デコーダ904、映像信号処理部905、表示部906、音声信号処理部907、スピーカ908、外部インタフェース909、制御部910、ユーザインタフェース911、及びバス912を備える。
図41は、上述した実施形態を適用した携帯電話機の概略的な構成の一例を示している。携帯電話機920は、アンテナ921、通信部922、音声コーデック923、スピーカ924、マイクロホン925、カメラ部926、画像処理部927、多重分離部928、記録再生部929、表示部930、制御部931、操作部932、及びバス933を備える。
図42は、上述した実施形態を適用した記録再生装置の概略的な構成の一例を示している。記録再生装置940は、例えば、受信した放送番組の音声データ及び映像データを符号化して記録媒体に記録する。また、記録再生装置940は、例えば、他の装置から取得される音声データ及び映像データを符号化して記録媒体に記録してもよい。また、記録再生装置940は、例えば、ユーザの指示に応じて、記録媒体に記録されているデータをモニタ及びスピーカ上で再生する。このとき、記録再生装置940は、音声データ及び映像データを復号する。
図43は、上述した実施形態を適用した撮像装置の概略的な構成の一例を示している。撮像装置960は、被写体を撮像して画像を生成し、画像データを符号化して記録媒体に記録する。
[第1のシステム]
次に、図44を参照して、図36乃至図38を参照して上述したスケーラブル符号化(階層符号化)されたスケーラブル符号化データの具体的な利用例について説明する。スケーラブル符号化は、例えば、図44に示される例のように、伝送するデータの選択のために利用される。
また、スケーラブル符号化は、例えば、図45に示される例のように、複数の通信媒体を介する伝送のために利用される。
また、スケーラブル符号化は、例えば、図46に示される例のように、符号化データの記憶に利用される。
(1) 符号化ストリームと、画像の複数のタイルに分割されているピクチャにおいてタイル間の量子化パラメータに関する処理の独立性を確保するように設定されたタイルに対する量子化パラメータの初期値とを受け取る受け取り部と、
前記受け取り部により受け取られた前記タイルに対する量子化パラメータの初期値を用いて、前記タイルを構成するコーディングユニットの量子化パラメータを再構築する量子化パラメータ再構築部と、
前記量子化パラメータ再構築部により再構築された量子化パラメータを用いて、前記受け取り部により受け取られた符号化ストリームを復号して、前記画像を生成する復号部と
を備える画像処理装置。
(2) 前記受け取り部は、前記符号化ストリームにおける所定の位置に設定されている値を、前記タイルの量子化パラメータの初期値として受け取る
前記(1)に記載の画像処理装置。
(3) 前記受け取り部は、前記符号化ストリームにおけるピクチャパラメータセットにおいて設定されている値を、前記タイルの量子化パラメータの初期値として受け取る
前記(1)または(2)に記載の画像処理装置。
(4) 前記受け取り部は、前記符号化ストリームにおける前記ピクチャの最初のスライスヘッダにおいて設定されている値を、前記タイルの量子化パラメータの初期値として受け取る
前記(1)または(2)に記載の画像処理装置。
(5) 前記受け取り部は、前記符号化ストリームにおける前記タイル先頭付近において設定されている値を、前記タイルの量子化パラメータの初期値として受け取る
前記(1)または(2)に記載の画像処理装置。
(6) 前記符号化ストリームにおけるスライスデータにおいて設定されている値を、前記タイルの量子化パラメータの初期値として受け取る
前記(1)または(2)に記載の画像処理装置。
(7) 前記受け取り部は、前記タイルを構成するコーディングユニットの量子化パラメータに関する情報を受け取り、
前記量子化パラメータ再構築部は、前記受け取り部により受け取られた前記タイルを構成するコーディングユニットの量子化パラメータに関する情報を用いて、前記タイルを構成するコーディングユニットの量子化パラメータを再構築する
前記(1)乃至(6)のいずれかに記載の画像処理装置。
(8) 画像処理装置が、
符号化ストリームと、画像の複数のタイルに分割されているピクチャにおいてタイル間の量子化パラメータに関する処理の独立性を確保するように設定されたタイルに対する量子化パラメータの初期値とを受け取り、
受け取られた前記タイルに対する量子化パラメータの初期値を用いて、前記タイルを構成するコーディングユニットの量子化パラメータを再構築し、
再構築された量子化パラメータを用いて、受け取られた符号化ストリームを復号して、前記画像を生成する
画像処理方法。
(9) 画像を符号化して、符号化ストリームを生成する符号化部と、
前記画像の複数のタイルに分割されているピクチャにおいてタイル間の量子化パラメータに関する処理の独立性を確保するように、前記タイルに対する量子化パラメータの初期値を設定する初期値設定部と、
前記初期値設定部により設定されたタイルに対する量子化パラメータの初期値と、前記符号化部により生成された符号化ストリームとを伝送する伝送部と
を備える画像処理装置。
(10) 前記初期値設定部は、前記符号化ストリームにおける所定の位置に、前記タイルの量子化パラメータの初期値を設定する
前記(9)に記載の画像処理装置。
(11) 前記初期値設定部は、前記符号化ストリームにおけるピクチャパラメータセットに、前記タイルの量子化パラメータの初期値を設定する
前記(9)または(10)に記載の画像処理装置。
(12) 前記初期値設定部は、前記符号化ストリームにおける前記ピクチャの最初のスライスヘッダに、前記タイルの量子化パラメータの初期値を設定する
前記(9)または(10)に記載の画像処理装置。
(13) 前記初期値設定部は、前記符号化ストリームにおけるタイルの先頭付近に、前記タイルの量子化パラメータの初期値を設定する
前記(9)または(10)に記載の画像処理装置。
(14) 前記初期値設定部は、前記符号化ストリームにおけるスライスデータに、前記タイルの量子化パラメータの初期値を設定する
前記(9)または(10)に記載の画像処理装置。
(15) 前記伝送部は、前記タイルの量子化パラメータの初期値を用いて生成される前記タイルを構成するコーディングユニットの量子化パラメータに関する情報を伝送する
前記(9)乃至(14)のいずれかに記載の画像処理装置。
(16) 画像処理装置が、
画像を符号化して、符号化ストリームを生成し、
前記画像の複数のタイルに分割されているピクチャにおいてタイル間の量子化パラメータに関する処理の独立性を確保するように、前記タイルに対する量子化パラメータの初期値を設定し、
設定されたタイルに対する量子化パラメータの初期値と、生成された符号化ストリームとを伝送する
画像処理方法。
Claims (16)
- 符号化ストリームと、画像の複数のタイルに分割されているピクチャにおいてタイル間の量子化パラメータに関する処理の独立性を確保するように設定されたタイルに対する量子化パラメータの初期値とを受け取る受け取り部と、
前記受け取り部により受け取られた前記タイルに対する量子化パラメータの初期値を用いて、前記タイルを構成するコーディングユニットの量子化パラメータを再構築する量子化パラメータ再構築部と、
前記量子化パラメータ再構築部により再構築された量子化パラメータを用いて、前記受け取り部により受け取られた符号化ストリームを復号して、前記画像を生成する復号部と
を備える画像処理装置。 - 前記受け取り部は、前記符号化ストリームにおける所定の位置に設定されている値を、前記タイルの量子化パラメータの初期値として受け取る
請求項1に記載の画像処理装置。 - 前記受け取り部は、前記符号化ストリームにおけるピクチャパラメータセットにおいて設定されている値を、前記タイルの量子化パラメータの初期値として受け取る
請求項2に記載の画像処理装置。 - 前記受け取り部は、前記符号化ストリームにおける前記ピクチャの最初のスライスヘッダにおいて設定されている値を、前記タイルの量子化パラメータの初期値として受け取る
請求項2に記載の画像処理装置。 - 前記受け取り部は、前記符号化ストリームにおける前記タイル先頭付近において設定されている値を、前記タイルの量子化パラメータの初期値として受け取る
請求項2に記載の画像処理装置。 - 前記受け取り部は、前記符号化ストリームにおけるスライスデータにおいて設定されている値を、前記タイルの量子化パラメータの初期値として受け取る
請求項2に記載の画像処理装置。 - 前記受け取り部は、前記タイルを構成するコーディングユニットの量子化パラメータに関する情報を受け取り、
前記量子化パラメータ再構築部は、前記受け取り部により受け取られた前記タイルを構成するコーディングユニットの量子化パラメータに関する情報を用いて、前記タイルを構成するコーディングユニットの量子化パラメータを再構築する
請求項1に記載の画像処理装置。 - 画像処理装置が、
符号化ストリームと、画像の複数のタイルに分割されているピクチャにおいてタイル間の量子化パラメータに関する処理の独立性を確保するように設定されたタイルに対する量子化パラメータの初期値とを受け取り、
受け取られた前記タイルに対する量子化パラメータの初期値を用いて、前記タイルを構成するコーディングユニットの量子化パラメータを再構築し、
再構築された量子化パラメータを用いて、受け取られた符号化ストリームを復号して、前記画像を生成する
画像処理方法。 - 画像を符号化して、符号化ストリームを生成する符号化部と、
前記画像の複数のタイルに分割されているピクチャにおいてタイル間の量子化パラメータに関する処理の独立性を確保するように、前記タイルに対する量子化パラメータの初期値を設定する初期値設定部と、
前記初期値設定部により設定されたタイルに対する量子化パラメータの初期値と、前記符号化部により生成された符号化ストリームとを伝送する伝送部と
を備える画像処理装置。 - 前記初期値設定部は、前記符号化ストリームにおける所定の位置に、前記タイルの量子化パラメータの初期値を設定する
請求項9に記載の画像処理装置。 - 前記初期値設定部は、前記符号化ストリームにおけるピクチャパラメータセットに、前記タイルの量子化パラメータの初期値を設定する
請求項10に記載の画像処理装置。 - 前記初期値設定部は、前記符号化ストリームにおける前記ピクチャの最初のスライスヘッダに、前記タイルの量子化パラメータの初期値を設定する
請求項10に記載の画像処理装置。 - 前記初期値設定部は、前記符号化ストリームにおけるタイルの先頭付近に、前記タイルの量子化パラメータの初期値を設定する
請求項10に記載の画像処理装置。 - 前記初期値設定部は、前記符号化ストリームにおけるスライスデータに、前記タイルの量子化パラメータの初期値を設定する
請求項10に記載の画像処理装置。 - 前記伝送部は、前記タイルの量子化パラメータの初期値を用いて生成される前記タイルを構成するコーディングユニットの量子化パラメータに関する情報を伝送する
請求項9に記載の画像処理装置。 - 画像処理装置が、
画像を符号化して、符号化ストリームを生成し、
前記画像の複数のタイルに分割されているピクチャにおいてタイル間の量子化パラメータに関する処理の独立性を確保するように、前記タイルに対する量子化パラメータの初期値を設定し、
設定されたタイルに対する量子化パラメータの初期値と、生成された符号化ストリームとを伝送する
画像処理方法。
Priority Applications (2)
Application Number | Priority Date | Filing Date | Title |
---|---|---|---|
US14/373,691 US20150030068A1 (en) | 2012-03-15 | 2013-03-04 | Image processing device and method |
CN201380012781.XA CN104170383A (zh) | 2012-03-15 | 2013-03-04 | 图像处理装置和方法 |
Applications Claiming Priority (4)
Application Number | Priority Date | Filing Date | Title |
---|---|---|---|
JP2012058680 | 2012-03-15 | ||
JP2012-058680 | 2012-03-15 | ||
JP2012076222 | 2012-03-29 | ||
JP2012-076222 | 2012-03-29 |
Publications (1)
Publication Number | Publication Date |
---|---|
WO2013137047A1 true WO2013137047A1 (ja) | 2013-09-19 |
Family
ID=49160960
Family Applications (1)
Application Number | Title | Priority Date | Filing Date |
---|---|---|---|
PCT/JP2013/055884 WO2013137047A1 (ja) | 2012-03-15 | 2013-03-04 | 画像処理装置および方法 |
Country Status (4)
Country | Link |
---|---|
US (1) | US20150030068A1 (ja) |
JP (1) | JPWO2013137047A1 (ja) |
CN (1) | CN104170383A (ja) |
WO (1) | WO2013137047A1 (ja) |
Families Citing this family (9)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
WO2017023829A1 (en) * | 2015-07-31 | 2017-02-09 | Stc.Unm | System and methods for joint and adaptive control of rate, quality, and computational complexity for video coding and video delivery |
US10848761B2 (en) | 2017-06-29 | 2020-11-24 | Qualcomm Incorporated | Reducing seam artifacts in 360-degree video |
CN111771379B (zh) * | 2018-01-25 | 2022-10-11 | 弗劳恩霍夫应用研究促进协会 | 有效子图片提取 |
US10666863B2 (en) * | 2018-05-25 | 2020-05-26 | Microsoft Technology Licensing, Llc | Adaptive panoramic video streaming using overlapping partitioned sections |
US10764494B2 (en) | 2018-05-25 | 2020-09-01 | Microsoft Technology Licensing, Llc | Adaptive panoramic video streaming using composite pictures |
CN112789862B (zh) * | 2018-10-02 | 2024-07-23 | 瑞典爱立信有限公司 | 使用图块上进行的循环以信号传送的图片图块属性 |
WO2020135409A1 (zh) * | 2018-12-24 | 2020-07-02 | 华为技术有限公司 | 视频解码方法、装置及解码设备 |
CN111355951B (zh) | 2018-12-24 | 2023-11-10 | 华为技术有限公司 | 视频解码方法、装置及解码设备 |
GB201902831D0 (en) * | 2019-03-01 | 2019-04-17 | Canon Kk | Method and apparatus for encoding and decoding a video bitstream for merging regions of interest |
Family Cites Families (1)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
US20130083845A1 (en) * | 2011-09-30 | 2013-04-04 | Research In Motion Limited | Methods and devices for data compression using a non-uniform reconstruction space |
-
2013
- 2013-03-04 US US14/373,691 patent/US20150030068A1/en not_active Abandoned
- 2013-03-04 JP JP2014504801A patent/JPWO2013137047A1/ja active Pending
- 2013-03-04 WO PCT/JP2013/055884 patent/WO2013137047A1/ja active Application Filing
- 2013-03-04 CN CN201380012781.XA patent/CN104170383A/zh active Pending
Non-Patent Citations (5)
Title |
---|
ARILD FULDSETH ET AL.: "Tiles, Joint Collaborative Team on Video Coding (JCT-VC) of ITU-T SG16 WP3 and ISO/IEC JTC1/SC29/WG11", JCTVC-F335, 6TH MEETING, July 2011 (2011-07-01), TORINO, IT, pages 1 - 15 * |
HISAO SASAI ET AL.: "Constrained Tile for parallel decoding", JOINT COLLABORATIVE TEAM ON VIDEO CODING (JCT-VC) OF ITU-T SG16 WP3 AND ISO/IEC JTC1/SC29/WG11, JCTVC-H0345RL, 8TH MEETING, February 2012 (2012-02-01), SAN JOSE, CA, pages 1 - 4 * |
KIRAN MISRA ET AL.: "New results for parallel decoding for Tiles", JOINT COLLABORATIVE TEAM ON VIDEO CODING (JCT-VC) OF ITU-T SG16 WP3 AND ISO/IEC JTC1/SC29/WG11, JCTVC-F594_R1, 6TH MEETING, July 2011 (2011-07-01), TORINO, IT, pages 1 - 6 * |
KIRAN MISRA ET AL.: "Tiles for parallel decoding", JOINT COLLABORATIVE TEAM ON VIDEO CODING (JCT-VC) OF ITU-T SG16 WP3 AND ISO/IEC JTC1/SC29/WG11, JCTVC-E412-RL, 5TH MEETING, March 2011 (2011-03-01), GENEVA, CH, pages 1 - 5 * |
TAMMY LEE ET AL.: "Simplification on tiles and slices", JOINT COLLABORATIVE TEAM ON VIDEO CODING (JCT-VC) OF ITU-T SG16 WP3 AND ISO/IEC JTC1/SC29/WG11, JCTVC-H0348-R2, 8TH MEETING, February 2012 (2012-02-01), SAN JOSE, CA, pages 1 - 10 * |
Also Published As
Publication number | Publication date |
---|---|
US20150030068A1 (en) | 2015-01-29 |
CN104170383A (zh) | 2014-11-26 |
JPWO2013137047A1 (ja) | 2015-08-03 |
Similar Documents
Publication | Publication Date | Title |
---|---|---|
JP6508554B2 (ja) | 画像処理装置および方法、並びに、プログラム | |
RU2641261C2 (ru) | Устройство и способ обработки изображений | |
JP6358475B2 (ja) | 画像復号装置および方法、並びに、画像符号化装置および方法 | |
WO2013137047A1 (ja) | 画像処理装置および方法 | |
JPWO2014002896A1 (ja) | 符号化装置および符号化方法、復号装置および復号方法 | |
WO2014050676A1 (ja) | 画像処理装置および方法 | |
JP6300046B2 (ja) | 画像処理装置および方法 | |
WO2014050731A1 (ja) | 画像処理装置および方法 | |
WO2013108688A1 (ja) | 画像処理装置および方法 | |
WO2016147836A1 (ja) | 画像処理装置および方法 | |
JPWO2015098561A1 (ja) | 復号装置および復号方法、並びに、符号化装置および符号化方法 | |
US9930353B2 (en) | Image decoding device and method | |
WO2014103764A1 (ja) | 画像処理装置および方法 | |
WO2013051453A1 (ja) | 画像処理装置および方法 | |
WO2015098563A1 (ja) | 画像符号化装置および方法、並びに画像復号装置および方法 | |
WO2014162916A1 (ja) | 画像符号化装置および方法、並びに、画像復号装置および方法 | |
JP2013150124A (ja) | 画像処理装置および方法 | |
WO2014156707A1 (ja) | 画像符号化装置および方法、並びに、画像復号装置および方法 | |
JP2013012996A (ja) | 画像処理装置および方法 |
Legal Events
Date | Code | Title | Description |
---|---|---|---|
121 | Ep: the epo has been informed by wipo that ep was designated in this application |
Ref document number: 13761826 Country of ref document: EP Kind code of ref document: A1 |
|
ENP | Entry into the national phase |
Ref document number: 2014504801 Country of ref document: JP Kind code of ref document: A |
|
WWE | Wipo information: entry into national phase |
Ref document number: 14373691 Country of ref document: US |
|
NENP | Non-entry into the national phase |
Ref country code: DE |
|
122 | Ep: pct application non-entry in european phase |
Ref document number: 13761826 Country of ref document: EP Kind code of ref document: A1 |