WO2018159526A1 - Dispositif de codage d'image animée et dispositif de décodage d'image animée - Google Patents

Dispositif de codage d'image animée et dispositif de décodage d'image animée Download PDF

Info

Publication number
WO2018159526A1
WO2018159526A1 PCT/JP2018/006937 JP2018006937W WO2018159526A1 WO 2018159526 A1 WO2018159526 A1 WO 2018159526A1 JP 2018006937 W JP2018006937 W JP 2018006937W WO 2018159526 A1 WO2018159526 A1 WO 2018159526A1
Authority
WO
WIPO (PCT)
Prior art keywords
unit
encoding
decoding
last
flag
Prior art date
Application number
PCT/JP2018/006937
Other languages
English (en)
Japanese (ja)
Inventor
友子 青野
知宏 猪飼
Original Assignee
シャープ株式会社
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Application filed by シャープ株式会社 filed Critical シャープ株式会社
Publication of WO2018159526A1 publication Critical patent/WO2018159526A1/fr

Links

Images

Classifications

    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04NPICTORIAL COMMUNICATION, e.g. TELEVISION
    • H04N19/00Methods or arrangements for coding, decoding, compressing or decompressing digital video signals
    • H04N19/60Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using transform coding
    • H04N19/61Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using transform coding in combination with predictive coding
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04NPICTORIAL COMMUNICATION, e.g. TELEVISION
    • H04N19/00Methods or arrangements for coding, decoding, compressing or decompressing digital video signals
    • H04N19/70Methods or arrangements for coding, decoding, compressing or decompressing digital video signals characterised by syntax aspects related to video coding, e.g. related to compression standards
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04NPICTORIAL COMMUNICATION, e.g. TELEVISION
    • H04N19/00Methods or arrangements for coding, decoding, compressing or decompressing digital video signals
    • H04N19/90Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using coding techniques not provided for in groups H04N19/10-H04N19/85, e.g. fractals
    • H04N19/91Entropy coding, e.g. variable length coding [VLC] or arithmetic coding

Definitions

  • One embodiment of the present invention relates to an image decoding device and an image encoding device.
  • an image encoding device that generates encoded data by encoding the moving image, and image decoding that generates a decoded image by decoding the encoded data The device is used.
  • the moving image encoding method include a method proposed in H.264 / AVC and HEVC (High-Efficiency Video Coding).
  • an image (picture) constituting a moving image is a slice obtained by dividing the image, a coding unit obtained by dividing the slice (coding unit (Coding Unit : CU)), and a hierarchical structure consisting of a prediction unit (PU) and a transform unit (TU) that are obtained by dividing a coding unit. Decrypted.
  • a predicted image is usually generated based on a local decoded image obtained by encoding / decoding an input image, and the predicted image is generated from the input image (original image).
  • a prediction residual obtained by subtraction (sometimes referred to as “difference image” or “residual image”) is encoded. Examples of methods for generating a predicted image include inter-screen prediction (inter prediction) and intra-screen prediction (intra prediction).
  • the moving image encoding device encodes the quantized transform coefficient obtained by performing orthogonal transform and quantization on the prediction residual, and the moving image decoding device decodes the quantized transform coefficient from the encoded data, and performs inverse quantization and Inverse orthogonal transformation is performed to restore the prediction error (Non-Patent Document 1).
  • processing is divided into position information and level information of each quantized transform coefficient in the CU, thereby removing redundancy and reducing the amount of codes (Non-Patent Document) 2).
  • each quantized transform coefficient in the CU is a first flag indicating the presence or absence of a non-zero quantized transform coefficient in the CU, and indicates the position of the last non-zero quantized transform coefficient in the scan order.
  • LAST second flag indicating presence / absence of non-zero quantized transform coefficient for each sub-block obtained by dividing CU, third flag indicating whether each quantized transform coefficient in the sub-block is non-zero
  • each non-zero Encoding is performed by dividing into a plurality of syntaxes indicating levels (magnitudes) of quantized transform coefficients.
  • the quantization transform coefficient is not directly encoded, but is divided into a plurality of pieces of information, thereby reducing the amount of codes.
  • LAST indicating the position of the last non-zero quantized transform coefficient in the scan order is a single coordinate, it does not depend much on the quantization parameter and has a large code amount.
  • One aspect of the present invention has been made in view of the above-described problems, and its purpose is to reduce the code amount of LAST indicating the position of the last non-zero quantized transform coefficient in scan order,
  • An object of the present invention is to provide an image decoding apparatus and an image encoding apparatus capable of improving encoding efficiency.
  • An image encoding device includes a dividing unit that divides one screen of an input moving image into encoding units (CU) including a plurality of pixels, and performs predetermined conversion using the CU as a unit to perform conversion coefficients. And a variable-length coding unit that performs variable-length coding on the transform coefficient, and the variable-length coding unit indicates whether a non-zero transform coefficient exists in the CU.
  • a first determination unit for determining a value of one flag, and a second value for determining a second flag value indicating whether or not a non-zero conversion coefficient exists only in a limited region of the CU.
  • the derivation unit that derives the syntax indicating the farthest position (LAST) and non-zero coefficient value (LEVEL) by scanning the transform coefficient from the DC component in the scan order in the CU, and the encoding parameter Switching between the first flag and the second flag to be variable-length encoded.
  • the variable length coding is performed on the first encoding unit that performs variable length encoding and the first flag, and there is a non-zero transform coefficient in the CU
  • the syntax indicating LAST and LEVEL is encoded.
  • the variable length coding of the second encoding unit and the second flag is performed, if a non-zero transform coefficient exists only in a limited region of the CU, the syntax indicating LEVEL is encoded.
  • a third encoding unit that encodes syntax indicating LAST and LEVEL is provided.
  • An image decoding apparatus includes a variable length decoding unit, an output unit that performs variable length decoding of encoded data using a coding unit (CU) including a plurality of pixels as a processing unit, and outputs a syntax.
  • CU coding unit
  • variable length decoding unit refers to an encoding parameter and indicates whether or not a non-zero transform coefficient exists in the CU
  • first decoding unit that performs variable-length decoding by switching which of the second flags indicating whether or not non-zero transform coefficients exist only in a limited region of the CU
  • Variable length decoding the second flag and the CU limitation To indicate that non-zero transform coefficients exist only in the specified area, set the value indicating the highest frequency component in the limited area to LAST, decode the variable length code indicating LEVEL, and limit the CU
  • a third decoding unit that performs variable-length decoding on the syntax indicating LAST and LEVEL is provided.
  • FIG. 1 is a schematic diagram illustrating a configuration of an image transmission system according to an embodiment of the present invention. It is a figure which shows the hierarchical structure of the data of the encoding stream which concerns on one Embodiment of this invention. It is a figure which shows the pattern of PU division
  • 10 is another flowchart showing the operation of transform coefficient decoding processing when the definition of cbf is changed. It is a flowchart which shows the operation
  • 10 is another flowchart showing the operation of transform coefficient encoding processing when the definition of cbf is switched. It is a figure explaining the expression method of LAST. It is a figure which shows the code amount required for encoding of LAST. It is a flowchart which shows the operation
  • FIG. 1 is a schematic diagram showing a configuration of an image transmission system 1 according to the present embodiment.
  • the image transmission system 1 is a system that transmits a code obtained by encoding an encoding target image, decodes the transmitted code, and displays an image.
  • the image transmission system 1 includes an image encoding device 11, a network 21, an image decoding device 31, and an image display device 41.
  • the image encoding device 11 receives an image T indicating a single layer image or a plurality of layers.
  • a layer is a concept used to distinguish a plurality of pictures when there are one or more pictures constituting a certain time. For example, when the same picture is encoded with a plurality of layers having different image quality and resolution, scalable encoding is performed, and when a picture of a different viewpoint is encoded with a plurality of layers, view scalable encoding is performed.
  • inter-layer prediction, inter-view prediction When prediction is performed between pictures of a plurality of layers (inter-layer prediction, inter-view prediction), encoding efficiency is greatly improved. Further, even when prediction is not performed (simultaneous casting), encoded data can be collected.
  • the network 21 transmits the encoded stream Te generated by the image encoding device 11 to the image decoding device 31.
  • the network 21 is the Internet, a wide area network (WAN: Wide Area Network), a small network (LAN: Local Area Network), or a combination thereof.
  • the network 21 is not necessarily limited to a bidirectional communication network, and may be a unidirectional communication network that transmits broadcast waves such as terrestrial digital broadcasting and satellite broadcasting.
  • the network 21 may be replaced with a storage medium that records an encoded stream Te such as a DVD (Digital Versatile Disc) or a BD (Blue-ray Disc).
  • the image decoding device 31 decodes each of the encoded streams Te transmitted by the network 21, and generates one or a plurality of decoded images Td decoded.
  • the image display device 41 displays all or part of one or more decoded images Td generated by the image decoding device 31.
  • the image display device 41 includes, for example, a display device such as a liquid crystal display or an organic EL (Electro-luminescence) display.
  • a display device such as a liquid crystal display or an organic EL (Electro-luminescence) display.
  • a high-quality enhancement layer image is displayed and only a lower processing capability is provided. Displays a base layer image that does not require higher processing capability and display capability as an extension layer.
  • X? Y: z is a ternary operator that takes y when x is true (non-zero) and takes z when x is false (0).
  • FIG. 2 is a diagram showing a hierarchical structure of data in the encoded stream Te.
  • the encoded stream Te illustratively includes a sequence and a plurality of pictures constituting the sequence.
  • (A) to (f) of FIG. 2 respectively show an encoded video sequence defining a sequence SEQ, an encoded picture defining a picture PICT, an encoded slice defining a slice S, and an encoded slice defining a slice data
  • the encoded video sequence In the encoded video sequence, a set of data referred to by the image decoding device 31 for decoding the sequence SEQ to be processed is defined. As shown in FIG. 2A, the sequence SEQ includes a video parameter set (Video Parameter Set), a sequence parameter set SPS (Sequence Parameter Set), a picture parameter set PPS (Picture Parameter Set), a picture PICT, and an addition. Includes SEI (Supplemental Enhancement Information). Here, the value indicated after # indicates the layer ID.
  • FIG. 2 shows an example in which encoded data of # 0 and # 1, that is, layer 0 and layer 1, exists, but the type of layer and the number of layers are not dependent on this.
  • the video parameter set VPS is a set of coding parameters common to a plurality of moving images, a plurality of layers included in the moving image, and coding parameters related to individual layers in a moving image composed of a plurality of layers.
  • a set is defined.
  • the sequence parameter set SPS defines a set of encoding parameters that the image decoding device 31 refers to in order to decode the target sequence. For example, the width and height of the picture are defined. A plurality of SPSs may exist. In that case, one of a plurality of SPSs is selected from the PPS.
  • a set of encoding parameters referred to by the image decoding device 31 in order to decode each picture in the target sequence is defined.
  • a quantization width reference value (pic_init_qp_minus26) used for picture decoding and a flag (weighted_pred_flag) indicating application of weighted prediction are included.
  • the picture PICT includes slices S0 to S NS-1 (NS is the total number of slices included in the picture PICT).
  • the coded slice In the coded slice, a set of data referred to by the image decoding device 31 for decoding the slice S to be processed is defined. As shown in FIG. 2C, the slice S includes a slice header SH and slice data SDATA.
  • the slice header SH includes an encoding parameter group that is referred to by the image decoding device 31 in order to determine a decoding method of the target slice.
  • Slice type designation information (slice_type) for designating a slice type is an example of an encoding parameter included in the slice header SH.
  • I slice using only intra prediction at the time of encoding (2) P slice using unidirectional prediction or intra prediction at the time of encoding, (3) B-slice using unidirectional prediction, bidirectional prediction, or intra prediction at the time of encoding may be used.
  • inter prediction is not limited to single prediction and bi-prediction, and a predicted image may be generated using more reference pictures.
  • the P, PB slice refers to a slice including a block that can use inter prediction.
  • the slice header SH may include a reference (pic_parameter_set_id) to the picture parameter set PPS included in the encoded video sequence.
  • the slice data SDATA includes a coding tree unit (CTU).
  • a CTU is a block of a fixed size (for example, 64x64) that constitutes a slice, and is sometimes called a maximum coding unit (LCU: Large Coding Unit).
  • Encoding tree unit As shown in (e) of FIG. 2, a set of data referred to by the image decoding device 31 in order to decode the processing target coding tree unit is defined.
  • the coding tree unit is divided into coding units (CU: Coding Unit) which is a basic unit of coding processing by recursive quadtree division (QT division) or binary tree division (BT division). .
  • a tree structure obtained by recursive quadtree partitioning or binary tree partitioning is called a coding tree (CT), and a node of the tree structure is called a coding node (CN).
  • CT coding tree
  • CN coding node
  • the intermediate nodes of the quadtree and the binary tree are coding nodes, and the coding tree unit itself is defined as the highest coding node.
  • CT includes, as CT information, a QT split flag (cu_split_flag) indicating whether or not to perform QT split, and a BT split mode (split_bt_mode) indicating a split method of BT split.
  • cu_split_flag and / or split_bt_mode are transmitted for each coding node CN.
  • cu_split_flag is 1, the encoding node CN is divided into four encoding nodes CN.
  • split_bt_mode is 1, the encoding node CN is horizontally divided into two encoding nodes CN.
  • split_bt_mode When split_bt_mode is 2, the encoding node CN is vertically divided into two encoding nodes CN. When split_bt_mode is 0, the encoding node CN is not divided and has one encoding unit CU as a node.
  • the encoding unit CU is a terminal node (leaf node) of the encoding node and is not further divided.
  • the size of the coding unit is 64x64 pixels, 64x32 pixels, 32x64 pixels, 32x32 pixels, 64x16 pixels, 16x64 pixels, 32x16 Pixel, 16x32 pixel, 16x16 pixel, 64x8 pixel, 8x64 pixel, 32x8 pixel, 8x32 pixel, 16x8 pixel, 8x16 pixel, 8x8 pixel, 64x4 pixel, 4x64 pixel, 32x4 pixel, 4x32 pixel, 16x4 pixel, 4x16 pixel, 8x4 pixel, It can take either 4x8 pixels or 4x4 pixels.
  • the encoding unit As shown in (f) of FIG. 2, a set of data referred to by the image decoding device 31 in order to decode the encoding unit to be processed is defined.
  • the encoding unit includes a prediction tree, a conversion tree, and a CU header CUH.
  • the CU header defines a prediction mode, a division method (PU division mode), and the like.
  • the prediction parameters (reference picture index, motion vector, etc.) of each prediction unit (PU) obtained by dividing the coding unit into one or a plurality are defined.
  • the prediction unit is one or a plurality of non-overlapping areas constituting the encoding unit.
  • the prediction tree includes one or a plurality of prediction units obtained by the above-described division.
  • a prediction unit obtained by further dividing the prediction unit is referred to as a “sub-block”.
  • the sub block is composed of a plurality of pixels.
  • the number of sub-blocks in the prediction unit is one.
  • the prediction unit is larger than the size of the sub-block, the prediction unit is divided into sub-blocks. For example, when the prediction unit is 8 ⁇ 8 and the sub-block is 4 ⁇ 4, the prediction unit is divided into four sub-blocks that are divided into two horizontally and two vertically.
  • the prediction process may be performed for each prediction unit (sub block).
  • Intra prediction is prediction within the same picture
  • inter prediction refers to prediction processing performed between different pictures (for example, between display times and between layer images).
  • the division method is encoded by the PU division mode (part_mode) of encoded data, and 2Nx2N (the same size as the encoding unit), 2NxN, 2NxnU, 2NxnD, Nx2N, nLx2N, nRx2N, and NxN etc.
  • 2NxN and Nx2N indicate 1: 1 symmetrical division
  • nRx2N indicate 1: 3, 3: 1 asymmetric partitioning.
  • the PUs included in the CU are expressed as PU0, PU1, PU2, and PU3 in this order.
  • FIG. 3 specifically show the partition shape (the position of the boundary of the PU partition) in each PU partition mode.
  • 3A shows a 2Nx2N partition
  • FIGS. 3B, 3C, and 2D show 2NxN, 2NxnU, and 2NxnD partitions (horizontal partitions), respectively.
  • E), (f), and (g) show partitions (vertical partitions) in the case of Nx2N, nLx2N, and nRx2N, respectively, and (h) shows an NxN partition.
  • the horizontal partition and the vertical partition are collectively referred to as a rectangular partition
  • 2Nx2N and NxN are collectively referred to as a square partition.
  • the encoding unit is divided into one or a plurality of conversion units, and the position and size of each conversion unit are defined.
  • a transform unit is one or more non-overlapping areas that make up a coding unit.
  • the conversion tree includes one or a plurality of conversion units obtained by the above division.
  • the division in the conversion tree includes a case where an area having the same size as that of the encoding unit is assigned as a conversion unit, and a case where recursive quadtree division is used, as in the case of the CU division described above.
  • Conversion processing is performed for each conversion unit.
  • the prediction parameter includes prediction list use flags predFlagL0 and predFlagL1, reference picture indexes refIdxL0 and refIdxL1, and motion vectors mvL0 and mvL1.
  • the prediction list use flags predFlagL0 and predFlagL1 are flags indicating whether or not reference picture lists called L0 list and L1 list are used, respectively, and a reference picture list corresponding to a value of 1 is used.
  • flag indicating whether or not it is XX when “flag indicating whether or not it is XX” is described, when the flag is not 0 (for example, 1) is XX, 0 is not XX, and logical negation, logical product, etc. 1 is treated as true and 0 is treated as false (the same applies hereinafter).
  • flag when the flag is not 0 (for example, 1) is XX, 0 is not XX, and logical negation, logical product, etc. 1 is treated as true and 0 is treated as false (the same applies hereinafter).
  • other values can be used as true values and false values in an actual apparatus or method.
  • Syntax elements for deriving inter prediction parameters included in the encoded data include, for example, PU partition mode part_mode, merge flag merge_flag, merge index merge_idx, inter prediction identifier inter_pred_idc, reference picture index refIdxLX, prediction vector index mvp_LX_idx, There is a difference vector mvdLX.
  • the reference picture list is a list including reference pictures stored in the reference picture memory 306.
  • FIG. 4 is a conceptual diagram illustrating an example of a reference picture and a reference picture list.
  • a rectangle is a picture
  • an arrow is a reference relationship of the picture
  • a horizontal axis is time
  • I, P, and B in the rectangle are an intra picture
  • a single prediction picture a bi-prediction picture
  • numbers in the rectangle are Indicates the decoding order.
  • the decoding order of pictures is I0, P1, B2, B3, and B4
  • the display order is I0, B3, B2, B4, and P1.
  • FIG. 4B shows an example of the reference picture list.
  • the reference picture list is a list representing candidate reference pictures, and one picture (slice) may have one or more reference picture lists.
  • the target picture B3 has two reference picture lists, an L0 list RefPicList0 and an L1 list RefPicList1.
  • the reference pictures are I0, P1, and B2, and the reference picture has these pictures as elements.
  • refIdxLX the reference picture index
  • the figure shows an example in which reference pictures P1 and B2 are referred to by refIdxL0 and refIdxL1.
  • the prediction parameter decoding (encoding) method includes a merge prediction (merge) mode and an AMVP (Adaptive Motion Vector Prediction) mode.
  • the merge flag merge_flag is a flag for identifying these.
  • the merge mode is a mode in which the prediction list use flag predFlagLX (or inter prediction identifier inter_pred_idc), the reference picture index refIdxLX, and the motion vector mvLX are not included in the encoded data and are derived from the prediction parameters of already processed neighboring PUs.
  • the AMVP mode is a mode in which the inter prediction identifier inter_pred_idc, the reference picture index refIdxLX, and the motion vector mvLX are included in the encoded data.
  • the motion vector mvLX is encoded as a prediction vector index mvp_LX_idx for identifying the prediction vector mvpLX and a difference vector mvdLX.
  • the inter prediction identifier inter_pred_idc is a value indicating the type and number of reference pictures, and takes one of PRED_L0, PRED_L1, and PRED_BI.
  • PRED_L0 and PRED_L1 indicate that reference pictures managed by the reference picture lists of the L0 list and the L1 list are used, respectively, and that one reference picture is used (single prediction).
  • PRED_BI indicates that two reference pictures are used (bi-prediction BiPred), and reference pictures managed by the L0 list and the L1 list are used.
  • the prediction vector index mvp_LX_idx is an index indicating a prediction vector
  • the reference picture index refIdxLX is an index indicating a reference picture managed in the reference picture list.
  • LX is a description method used when L0 prediction and L1 prediction are not distinguished from each other. By replacing LX with L0 and L1, parameters for the L0 list and parameters for the L1 list are distinguished.
  • the merge index merge_idx is an index that indicates whether one of the prediction parameter candidates (merge candidates) derived from the processed PU is used as the prediction parameter of the decoding target PU.
  • the motion vector mvLX indicates a shift amount between blocks on two different pictures.
  • a prediction vector and a difference vector related to the motion vector mvLX are referred to as a prediction vector mvpLX and a difference vector mvdLX, respectively.
  • Inter prediction identifier inter_pred_idc and prediction list use flag predFlagLX The relationship between the inter prediction identifier inter_pred_idc and the prediction list use flags predFlagL0 and predFlagL1 is as follows and can be converted into each other.
  • the flag biPred as to whether it is a bi-prediction BiPred can be derived depending on whether the two prediction list use flags are both 1. For example, it can be derived by the following formula.
  • the flag biPred can also be derived depending on whether or not the inter prediction identifier is a value indicating that two prediction lists (reference pictures) are used. For example, it can be derived by the following formula.
  • the luminance intra prediction mode IntraPredModeY is 67 mode and corresponds to planar prediction (0), DC prediction (1), and direction prediction (2 to 66).
  • the color difference intra prediction mode IntraPredModeC is a 68 mode obtained by adding a Color Component Linear Mode (CCLM) to the above 67 mode.
  • CCLM is a mode in which the pixel value of the target pixel in the target color component is derived by linear prediction with reference to the pixel value of another color component encoded before the target color component.
  • the color component includes luminance Y, color difference Cb, and color difference Cr.
  • Different intra prediction modes may be assigned depending on luminance and color difference, and the prediction mode is encoded and decoded in units of CU or PU.
  • FIG. 5 is a schematic diagram illustrating a configuration of the image decoding device 31 according to the present embodiment.
  • the image decoding device 31 includes an entropy decoding unit 301, a prediction parameter decoding unit (prediction image decoding device) 302, a loop filter 305, a reference picture memory 306, a prediction parameter memory 307, a prediction image generation unit (prediction image generation device) 308, and inversely.
  • a quantization / inverse DCT unit 311 and an addition unit 312 are included.
  • the prediction parameter decoding unit 302 includes an inter prediction parameter decoding unit 303 and an intra prediction parameter decoding unit 304.
  • the predicted image generation unit 308 includes an inter predicted image generation unit 309 and an intra predicted image generation unit 310.
  • the entropy decoding unit 301 performs entropy decoding on the coded stream Te input from the outside, and separates and decodes individual codes (syntax elements).
  • the separated code includes a prediction parameter for generating a prediction image and residual information for generating a difference image.
  • the entropy decoding unit 301 outputs a part of the separated code to the prediction parameter decoding unit 302. Some of the separated codes are, for example, the prediction mode predMode, the PU partition mode part_mode, the merge flag merge_flag, the merge index merge_idx, the inter prediction identifier inter_pred_idc, the reference picture index ref_Idx_lX, the prediction vector index mvp_LX_idx, and the difference vector mvdLX. Control of which code is decoded is performed based on an instruction from the prediction parameter decoding unit 302.
  • the entropy decoding unit 301 outputs the quantization coefficient to the inverse quantization / inverse DCT unit 311.
  • the quantization coefficient is a coefficient obtained by performing quantization by performing DCT (Discrete Cosine Transform) on the residual signal in the encoding process.
  • the entropy decoding unit 301 includes a header decoding unit 1001, a CT information decoding unit 1002, a CU decoding unit 1003, and a decoding module 1004.
  • the decoding module 1004 performs a decoding process for decoding the syntax value from the encoded data. Based on the encoded data and syntax type supplied from the header decoding unit 1001, CT information decoding unit 1002, and CU decoding unit 1003, the decoding module 1004 uses a fixed-length encoding method or an entropy encoding method such as CABAC. Decodes the encoded syntax value and returns the decoded syntax value to the supplier.
  • the header decoding unit 1001 uses the decoding module 1004 to decode the VPS, SPS, PPS, and slice header of the encoded data input from the image encoding device 11.
  • the CT information decoding unit 1002 uses the decoding module 1004 to perform decoding processing of the encoding tree unit and the encoding tree from the encoded data input from the image encoding device 11.
  • the CT information decoding unit 1002 uses the decoding module 1004 to decode the tree unit header CTUH as CTU information included in the CTU.
  • the CT information decoding unit 1002 indicates, as CT information, a QT division flag indicating whether or not the target CT is QT-divided, and whether or not the target CT is BT-divided, and in the case of BT division, indicates a BT division method
  • the BT division mode is decoded, and the target CT is recursively divided and decoded until the QT division flag and the BT division mode no longer notify further division.
  • the tree unit footer CTUF is further decoded as CTU information.
  • the tree unit header CTUH and the tree unit footer CTUF include coding parameters referred to by the image decoding device 31 in order to determine a decoding method of the target coding tree unit.
  • the CT information may include parameters applied in the target CT and lower coding nodes.
  • the CU decoding unit 1003 includes PUI information (merge flag (merge_flag), merge index (merge_idx), prediction motion vector index (mvp_idx), reference image index (ref_idx_lX), and inter prediction of the lowest coding node CN (ie, CU) Identifier (inter_pred_flag), difference vector (mvdLX, etc.), quantization transform coefficient (residual_coding), and TTI information (TU partition flag SP_TU (split_transform_flag), CU residual flag CBP_TU (cbf_cb, cbf_cr, cbf_luma), etc.) Decryption is performed using the decryption module 1004.
  • PUI information merge flag (merge_flag), merge index (merge_idx), prediction motion vector index (mvp_idx), reference image index (ref_idx_lX), and inter prediction of the lowest coding node CN (i
  • Quantized residuals are CU residual flags CBP_TU (cbf_luma, cbf_cb, cbf_cr) and non-zero quantized transform coefficient positions (last_sig_coeff_x_prefix, last_sig_coeff_y_prefix, last_sig_coeff_x_suffix, last_sig_coefficient_coded_coefficient_coded_coefficients It is expressed by the syntax to be expressed (coeff_abs_level_greater1, coeff_abs_level_greater2, coeff_abs_level_remaining, coeff_sign_flag).
  • the quantized transform coefficient is referred to as a transform coefficient.
  • CBP_TU is a flag indicating whether a non-zero conversion coefficient is included in the luminance component and color difference component (Cb, Cr) of a certain CU.
  • cbf_luma 1, a non-zero conversion coefficient is included in the luminance component of the CU.
  • cbf_luma is 0, a non-zero conversion coefficient is not included.
  • cbf_cb and cbf_cr are each 1
  • a non-zero conversion coefficient is included in the Cb component and Cr component of the CU
  • cbf_cb and cbf_cr are each 0, the Cb component and Cr component of the CU are each non-zero. Does not include conversion factor.
  • the CU size is 128x128 to 4x4, and there are three types of scan directions (diagonal direction, horizontal direction, vertical direction), but the same processing is performed for different CU sizes and scan directions.
  • the luminance component and the color difference component are not distinguished.
  • LAST is the position of the last non-zero transform coefficient when the upper left coordinate of the CU is (0,0) and scanned in the specified scan direction, and is the four variable-length encoded syntaxes (last_sig_coeff_x_prefix, last_sig_coeff_y_prefix , Last_sig_coeff_y_suffix, last_sig_coeff_y_suffix).
  • variable length code table is FIG.
  • the part represented by “1” or “0” is a prefix
  • the part represented by “X” is a suffix.
  • “X” represents “1” or “0”.
  • the CU is divided into fixed-size sub-blocks (for example, 4x4).
  • the CU decoding unit 1003 sets all transform coefficients in the CU to 0.
  • the CU decoding unit 1003 decodes the cbf using the decoding module 1004.
  • the CU decoding unit 1003 checks whether cbf is 1. When cbf is not 1, the CU decoding unit 1003 ends the process. When cbf is 1, the process proceeds to S1304, where the CU decoding unit 1003 decodes the syntax representing LAST and derives the LAST position.
  • the CU decoding unit 1003 decodes the coded_sub_block_flag of each subblock using the subblock including the position of LAST as a starting point.
  • the CU decoding unit 1003 performs S1306 to S1308 for all subblocks before the subblock including LAST.
  • the CU decoding unit 1003 refers to coded_sub_block_flag to check whether each subblock has a non-zero transform coefficient.
  • coded_sub_block_flag 1
  • the process proceeds to S1307
  • the CU decoding unit 1003 decodes all the sig_coeff_flags of the subblock.
  • coded_sub_block_flag 0, the process proceeds to S1308, and the CU decoding unit 1003 sets all sig_coeff_flags of the subblock to 0.
  • the CU decoding unit 1003 performs S1309 to S1311 for all sig_coeff_flags.
  • sig_coeff_flag 1
  • the CU decoding unit 1003 decodes the syntax (coeff_abs_level_greater1, coeff_abs_level_greater2, coeff_abs_level_remaining, coeff_sign_flag) expressing the coefficient level in S1310, and derives the transform coefficient value by referring to the decoding result in S1311.
  • the inter prediction parameter decoding unit 303 decodes the inter prediction parameter with reference to the prediction parameter stored in the prediction parameter memory 307 based on the code input from the entropy decoding unit 301.
  • the inter prediction parameter decoding unit 303 outputs the decoded inter prediction parameter to the prediction image generation unit 308 and stores it in the prediction parameter memory 307.
  • the intra prediction parameter decoding unit 304 refers to the prediction parameter stored in the prediction parameter memory 307 on the basis of the code input from the entropy decoding unit 301 and decodes the intra prediction parameter.
  • the intra prediction parameter is a parameter used in a process of predicting a CU within one picture, for example, an intra prediction mode IntraPredMode.
  • the intra prediction parameter decoding unit 304 outputs the decoded intra prediction parameter to the prediction image generation unit 308 and stores it in the prediction parameter memory 307.
  • the loop filter 305 applies filters such as a deblocking filter, a sample adaptive offset (SAO), and an adaptive loop filter (ALF) to the decoded image of the CU generated by the adding unit 312.
  • filters such as a deblocking filter, a sample adaptive offset (SAO), and an adaptive loop filter (ALF) to the decoded image of the CU generated by the adding unit 312.
  • the reference picture memory 306 stores the decoded image of the CU generated by the adding unit 312 at a predetermined position for each decoding target picture and CU.
  • the prediction parameter memory 307 stores the prediction parameter in a predetermined position for each decoding target picture and prediction unit (or sub-block, fixed-size block, pixel). Specifically, the prediction parameter memory 307 stores the inter prediction parameter decoded by the inter prediction parameter decoding unit 303, the intra prediction parameter decoded by the intra prediction parameter decoding unit 304, and the prediction mode predMode separated by the entropy decoding unit 301. .
  • the stored inter prediction parameters include, for example, a prediction list utilization flag predFlagLX (inter prediction identifier inter_pred_idc), a reference picture index refIdxLX, and a motion vector mvLX.
  • the prediction image generation unit 308 receives the prediction mode predMode input from the entropy decoding unit 301 and the prediction parameter from the prediction parameter decoding unit 302. Further, the predicted image generation unit 308 reads a reference picture from the reference picture memory 306. The prediction image generation unit 308 generates a prediction image of the PU using the input prediction parameter and the read reference picture in the prediction mode indicated by the prediction mode predMode.
  • the inter prediction image generation unit 309 uses the inter prediction parameter input from the inter prediction parameter decoding unit 303 and the read reference picture to perform prediction of the PU by inter prediction. Is generated.
  • the inter prediction image generation unit 309 performs a motion vector on the basis of the decoding target PU from the reference picture indicated by the reference picture index refIdxLX for a reference picture list (L0 list or L1 list) having a prediction list use flag predFlagLX of 1.
  • the reference picture block at the position indicated by mvLX is read from the reference picture memory 306.
  • the inter prediction image generation unit 309 performs prediction based on the read reference picture block to generate a prediction image of the PU.
  • the inter prediction image generation unit 309 outputs the generated prediction image of the PU to the addition unit 312.
  • the intra predicted image generation unit 310 When the prediction mode predMode indicates the intra prediction mode, the intra predicted image generation unit 310 performs intra prediction using the intra prediction parameter input from the intra prediction parameter decoding unit 304 and the read reference picture. Specifically, the intra predicted image generation unit 310 reads, from the reference picture memory 306, neighboring PUs that are pictures to be decoded and are in a predetermined range from the decoding target PUs among the PUs that have already been decoded.
  • the predetermined range is, for example, one of the left, upper left, upper, and upper right adjacent PUs when the decoding target PU sequentially moves in the so-called raster scan order, and differs depending on the intra prediction mode.
  • the raster scan order is an order in which each row is sequentially moved from the left end to the right end in each picture from the upper end to the lower end.
  • the intra predicted image generation unit 310 performs prediction in the prediction mode indicated by the intra prediction mode IntraPredMode for the read adjacent PU, and generates a predicted image of the PU.
  • the intra predicted image generation unit 310 outputs the generated predicted image of the PU to the adding unit 312.
  • the intra prediction image generation unit 310 performs planar prediction (0), DC prediction (1), direction according to the luminance prediction mode IntraPredModeY.
  • Prediction image of luminance PU is generated by any one of prediction (2 to 66), and planar prediction (0), DC prediction (1), direction prediction (2 to 66), LM mode according to color difference prediction mode IntraPredModeC
  • a prediction image of a color difference PU is generated according to any one of (67).
  • the inverse quantization / inverse DCT unit 311 inversely quantizes the quantization coefficient input from the entropy decoding unit 301 to obtain a DCT coefficient.
  • the inverse quantization / inverse DCT unit 311 performs inverse DCT (Inverse Discrete Cosine Transform) on the obtained DCT coefficient to calculate a residual signal.
  • the inverse quantization / inverse DCT unit 311 outputs the calculated residual signal to the addition unit 312.
  • the addition unit 312 adds the prediction image of the PU input from the inter prediction image generation unit 309 or the intra prediction image generation unit 310 and the residual signal input from the inverse quantization / inverse DCT unit 311 for each pixel, Generate a decoded PU image.
  • the adding unit 312 stores the generated decoded image of the PU in the reference picture memory 306, and outputs a decoded image Td in which the generated decoded image of the PU is integrated for each picture to the outside.
  • FIG. 6 is a block diagram illustrating a configuration of the image encoding device 11 according to the present embodiment.
  • the image encoding device 11 includes a prediction image generation unit 101, a subtraction unit 102, a DCT / quantization unit 103, an entropy encoding unit 104, an inverse quantization / inverse DCT unit 105, an addition unit 106, a loop filter 107, and a prediction parameter memory.
  • the prediction parameter encoding unit 111 includes an inter prediction parameter encoding unit 112 and an intra prediction parameter encoding unit 113.
  • the predicted image generation unit 101 generates, for each picture of the image T, a predicted image P of the prediction unit PU for each encoding unit CU that is an area obtained by dividing the picture.
  • the predicted image generation unit 101 reads a decoded block from the reference picture memory 109 based on the prediction parameter input from the prediction parameter encoding unit 111.
  • the prediction parameter input from the prediction parameter encoding unit 111 is, for example, a motion vector in the case of inter prediction.
  • the predicted image generation unit 101 reads a block at a position on the reference image indicated by the motion vector with the target PU as a starting point.
  • the prediction parameter is, for example, an intra prediction mode.
  • a pixel value of an adjacent PU used in the intra prediction mode is read from the reference picture memory 109, and a predicted image P of the PU is generated.
  • the predicted image generation unit 101 generates a predicted image P of the PU using one prediction method among a plurality of prediction methods for the read reference picture block.
  • the predicted image generation unit 101 outputs the generated predicted image P of the PU to the subtraction unit 102.
  • predicted image generation unit 101 has the same operation as that of the predicted image generation unit 308 already described, and therefore description thereof is omitted here.
  • the prediction image generation unit 101 generates a prediction image P of the PU based on the pixel value of the reference block read from the reference picture memory, using the parameter input from the prediction parameter encoding unit.
  • the predicted image generated by the predicted image generation unit 101 is output to the subtraction unit 102 and the addition unit 106.
  • the subtraction unit 102 subtracts the signal value of the predicted image P of the PU input from the predicted image generation unit 101 from the pixel value of the corresponding PU of the image T, and generates a residual signal.
  • the subtraction unit 102 outputs the generated residual signal to the DCT / quantization unit 103.
  • the DCT / quantization unit 103 performs DCT on the residual signal input from the subtraction unit 102 and calculates a DCT coefficient.
  • the DCT / quantization unit 103 quantizes the calculated DCT coefficient to obtain a quantization coefficient.
  • the DCT / quantization unit 103 outputs the obtained quantization coefficient to the entropy coding unit 104 and the inverse quantization / inverse DCT unit 105.
  • the entropy encoding unit 104 receives a quantization coefficient from the DCT / quantization unit 103 and receives a prediction parameter from the prediction parameter encoding unit 111.
  • the input prediction parameters include, for example, codes such as a reference picture index ref_Idx_lX, a prediction vector index mvp_LX_idx, a difference vector mvdLX, a prediction mode pred_mode_flag, and a merge index merge_idx.
  • the entropy encoding unit 104 entropy-encodes the input division information, prediction parameters, quantization transform coefficients, and the like to generate an encoded stream Te, and outputs the generated encoded stream Te to the outside.
  • the entropy encoding unit 104 includes a header encoding unit 1101, a CT information encoding unit 1102, a CU encoding unit 1103, and an encoding module 1104.
  • the entropy encoding unit 104 encodes the header information supplied from the prediction parameter encoding unit 111, the prediction parameter, and the quantized transform coefficient supplied from the DCT / quantization unit 103, and outputs encoded data.
  • the header encoding unit 1101 encodes the VPS, SPS, PPS, and slice header using the encoding module 1104.
  • the CT information encoding unit 1102 uses the encoding module 1104 to perform CTU and CT encoding processing.
  • the CT information encoding unit 1102 uses the encoding module 1104 to encode the tree unit header CTUH as CTU information included in the CTU.
  • the CT information encoding unit 1102 uses, as CT information, a QT division flag indicating whether or not to subject the target CT to QT division, whether or not to subject the target CT to BT division, and a division method in the case of BT division.
  • the BT division mode shown is encoded, and the target CT is recursively divided and encoded until the QT division flag and the BT division mode no longer notify further division.
  • the tree unit footer CTUF is further encoded as CTU information.
  • the tree unit header CTUH and the tree unit footer CTUF include coding parameters referred to by the image decoding device 31 in order to determine a decoding method of the target coding tree unit.
  • the CT information may include parameters applied in the target CT and lower coding nodes.
  • the CU encoding unit 1103 includes PUI information (merge flag (merge_flag), merge index (merge_idx), prediction motion vector index (mvp_idx), reference image index (ref_idx_lX), interlaced encoding node CN (ie, CU) Prediction identifier (inter_pred_flag), difference vector (mvdLX, etc.), quantization transform coefficient (residual_coding), and TTI information (TU partition flag SP_TU (split_transform_flag), CU residual flag CBP_TU (cbf_cb, cbf_cr, cbf_luma), etc.) Encoding is performed using the encoding module 1104.
  • PUI information merge flag (merge_flag), merge index (merge_idx), prediction motion vector index (mvp_idx), reference image index (ref_idx_lX), interlaced encoding node CN (ie, CU
  • the encoding module 1104 performs an encoding process for encoding various prediction parameters, quantization transform coefficients, and the like in a fixed-length encoding method or entropy encoding. More specifically, the encoding module 1104 encodes the header encoding unit 1101, the CTU information encoding unit 1102, and the CU encoding unit 1103 using a fixed-length encoding or an entropy encoding scheme such as CABAC, and performs encoding. Output data.
  • the operation of the encoding process of the quantized transform coefficient (transform coefficient) will be described using the flowchart of FIG. In S1401, the CU encoding unit 1103 counts the number of non-zero transform coefficients in the CU.
  • the CU encoding unit 1103 derives the syntax representing LAST in S1407 and encodes the syntax representing LAST in S1408. Next, CU encoding section 1103 performs S1409 to S1413 for the subblocks before the subblock including LAST. In S1409, the CU encoding unit 1103 counts the number of non-zero transform coefficients in the sub-block.
  • coded_sub_block_flag 1
  • the process proceeds to S1415, and S1415 to S1418 are performed on each transform coefficient in the subblock.
  • coded_sub_block_flag 1
  • the CU encoding unit 1103 proceeds to S1419 and sets all sig_coeff_flags in the sub-block to 0.
  • the CU encoding unit 1103 checks whether or not the transform coefficient value is 0 in S1415. If the transform coefficient value is 0, the process proceeds to S1416, sets sig_coeff_flag to 0, and if the transform coefficient value is 1, the process proceeds to S1417. Then, sig_coeff_flag is set to 1, and sig_coeff_flag is encoded in S1418.
  • the CU encoding unit 1103 performs S1420 to S1421 for all transform coefficients.
  • the inverse quantization / inverse DCT unit 105 inversely quantizes the quantization coefficient input from the DCT / quantization unit 103 to obtain a DCT coefficient.
  • the inverse quantization / inverse DCT unit 105 performs inverse DCT on the obtained DCT coefficient to calculate a residual signal.
  • the inverse quantization / inverse DCT unit 105 outputs the calculated residual signal to the addition unit 106.
  • the addition unit 106 adds the signal value of the prediction image P of the PU input from the prediction image generation unit 101 and the signal value of the residual signal input from the inverse quantization / inverse DCT unit 105 for each pixel, and performs decoding. Generate an image.
  • the adding unit 106 stores the generated decoded image in the reference picture memory 109.
  • the loop filter 107 performs a deblocking filter, a sample adaptive offset (SAO), and an adaptive loop filter (ALF) on the decoded image generated by the adding unit 106.
  • SAO sample adaptive offset
  • ALF adaptive loop filter
  • the prediction parameter memory 108 stores the prediction parameter generated by the encoding parameter determination unit 110 at a predetermined position for each encoding target picture and CU.
  • the reference picture memory 109 stores the decoded image generated by the loop filter 107 at a predetermined position for each picture to be encoded and each CU.
  • the encoding parameter determination unit 110 selects one set from among a plurality of sets of encoding parameters.
  • the encoding parameter is a parameter to be encoded that is generated in association with the above-described QTBT division parameter and prediction parameter.
  • the predicted image generation unit 101 generates a predicted image P of the PU using each of these encoding parameter sets.
  • the encoding parameter determination unit 110 calculates an RD cost value indicating the amount of information and the encoding error for each of a plurality of sets.
  • the RD cost value is, for example, the sum of a code amount and a square error multiplied by a coefficient ⁇ .
  • the code amount is the information amount of the encoded stream Te obtained by entropy encoding the quantization error and the encoding parameter.
  • the square error is the sum between pixels regarding the square value of the residual value of the residual signal calculated by the subtracting unit 102.
  • the coefficient ⁇ is a real number larger than a preset zero.
  • the encoding parameter determination unit 110 selects a set of encoding parameters that minimizes the calculated RD cost value.
  • the entropy encoding unit 104 outputs the selected set of encoding parameters to the outside as the encoded stream Te, and does not output the set of unselected encoding parameters.
  • the encoding parameter determination unit 110 stores the determined encoding parameter in the prediction parameter memory 108.
  • the prediction parameter encoding unit 111 derives a format for encoding from the parameters input from the encoding parameter determination unit 110 and outputs the format to the entropy encoding unit 104. Deriving the format for encoding is, for example, deriving a difference vector from a motion vector and a prediction vector. Also, the prediction parameter encoding unit 111 derives parameters necessary for generating a prediction image from the parameters input from the encoding parameter determination unit 110 and outputs the parameters to the prediction image generation unit 101.
  • the parameter necessary for generating the predicted image is, for example, a motion vector in units of sub-blocks.
  • the inter prediction parameter encoding unit 112 derives an inter prediction parameter such as a difference vector based on the prediction parameter input from the encoding parameter determination unit 110.
  • the inter prediction parameter encoding unit 112 derives parameters necessary for generating a prediction image to be output to the prediction image generating unit 101, and an inter prediction parameter decoding unit 303 (see FIG. 5 and the like) derives inter prediction parameters.
  • Some of the configurations are the same as the configuration to be performed.
  • the intra prediction parameter encoding unit 113 derives a prediction parameter necessary for generating a prediction image to be output to the prediction image generation unit 101, and the intra prediction parameter decoding unit 304 (see FIG. 5 and the like) And a part of the same structure as that for deriving.
  • the intra prediction parameter encoding unit 113 derives a format (for example, MPM_idx, rem_intra_luma_pred_mode) for encoding from the intra prediction mode IntraPredMode input from the encoding parameter determination unit 110.
  • a format for example, MPM_idx, rem_intra_luma_pred_mode
  • CBP_TU and LAST In the image coding apparatus and the image decoding apparatus according to Embodiment 1, the transform coefficient level is limited by using CBP_TU (cbf_luma, cbf_cb, cbf_cr), LAST, coded_sub_block_flag, and sig_coeff_flag as the transform coefficient, and the code amount Have reduced.
  • FIG. 15 is a graph showing a breakdown of the code amount. From FIG.
  • FIG. 16 shows the result of aggregating the average value of the number of non-zero transform coefficients for each CU size and quantization parameter in 5 types of HD resolution sequences.
  • width and height are the width and height of the CU, and 22 and 27 are quantization parameters.
  • (1) is luminance
  • (2) is color difference (Cb)
  • (3) is the average number of non-zero transform coefficients included in 1CU in intra coding and inter coding of color difference (Cr).
  • the average number of non-zero coefficients per CU is often less than 3. Since the energy is concentrated on the low frequency component by the conversion, the non-zero conversion coefficient is considered to be concentrated on the DC component and one or two AC low frequency components.
  • cbf_luma, cbf_cb, and cbf_cr are represented as cbf.
  • FIGS. 17 (1) to (4) respectively show codes and code amounts of the syntaxes related to decoding.
  • the coded_sub_block_flag and the syntax representing the transform coefficient level are omitted because there is no change in the code amount in (1) to (4), and only cbf, LAST, and sig_coeff_flag are shown.
  • the variable length code table of LAST uses FIG.
  • cbf 2 bits
  • Fig. 17 (1) shows the syntax and code amount for (1) above.
  • the 1-bit code amount is increased as compared with the above (1). Therefore, the code amount can be efficiently reduced if the non-zero transform coefficient is used under the condition where the majority of the non-zero transform coefficients are concentrated near the DC component.
  • one of the above methods (2) to (4) is used for the inter coding of the color difference component
  • the above method (1) is used for the intra / inter coding of the luminance component and the intra coding of the color difference component.
  • Which of the above (2) to (4) is used in the inter-coding of the color difference component may be defined in advance between the image encoding device and the image decoding device, or among the SPS, PPS, and slice headers. You may be notified of how to use.
  • FIGS. 18 and 19 are flowcharts showing operations of the entropy decoding unit 301 (CU decoding unit 1003) and the entropy encoding unit 104 (CU encoding unit 1103) when the methods (2) to (4) are used. It is.
  • FIG. 18 (2) is a flowchart showing the decoding process of (2) above.
  • the CU decoding unit 1003 sets all transform coefficients in the CU to 0.
  • the CU decoding unit 1003 decodes cbf.
  • S1805 is the same as S1304 in FIG. 13
  • S1806 is the same as S1305 in FIG. 13
  • S1807 is the same as S1306 to S1308 in FIG.
  • the process proceeds to S1808, and the CU decoding unit 1003 sets (0, 0) to LAST.
  • the CU decoding unit 1003 sets 1 to sig_coeff_flag of the DC component and 0 to sig_coeff_flag of other AC components.
  • the CU decoding unit 1003 decodes the transform coefficient level, but this process is the same as S1309 to S1311 in FIG.
  • FIG. 18 (3) is a flowchart showing the decoding process (3) above.
  • FIG. 18 (2) is the same as FIG. 18 (2) except that the processing of S1808 and S1809 is changed to S18081 and S18091.
  • the CU decoding unit 1003 sets (0, 1) to LAST.
  • the CU decoding unit 1003 decodes sig_coeff_flag of the DC component and the AC0 component, and sets 0 to sig_coeff_flag of the other AC components.
  • FIG. 18 (4) is a flowchart showing the decoding process (4) above.
  • 18 (2) is the same as FIG. 18 (2) except that the processing of S1808 and S1809 in FIG. 18 (2) is changed to S18082 and S18092.
  • the CU decoding unit 1003 sets (1,0) to LAST.
  • the CU decoding unit 1003 decodes the DC component, sig_coeff_flag of the AC0 and AC1 components, and sets 0 to the sig_coeff_flag of the other AC components.
  • FIG. 19 (2) is a flowchart showing the encoding process of (2) above.
  • the CU encoding unit 1103 counts the number of non-zero transform coefficients in the CU.
  • the CU encoding unit 1103 checks whether or not there is a non-zero transform coefficient in the CU. When there is no non-zero transform coefficient, the process proceeds to S1903, and the CU encoding unit 1103 sets cbf to 0. If there is a non-zero transform coefficient, the process proceeds to S1904, where the CU encoder 1103 checks whether the non-zero transform coefficient is only a DC component. If the non-zero transform coefficient is only the DC component, the process proceeds to S1905, where cbf is set to 10.
  • the process proceeds to S1906, and cbf is set to 11.
  • the CU encoding unit 1103 encodes cbf.
  • S1911 is the same as S1409 to S1413 in FIG. 14, and S1912 is the same as S1414 to S1419 in FIG.
  • cbf 11
  • 1 is set to sig_coeff_flag of the DC component and 0 is set to sig_coeff_flag of the other AC components in S1914.
  • the CU encoding unit 1103 encodes a syntax expressing the coefficient level of the non-zero transform coefficient.
  • S1913 is the same as S1420 to S1421 in FIG.
  • FIG. 19 (3) is a flowchart showing the encoding process (3) above.
  • FIG. 19 (2) is the same as FIG. 19 (2) except that the processing of S1904 and S1914 is changed to S19041 and S19141.
  • the CU encoding unit 1103 checks whether there are non-zero transform coefficients other than DC and AC0 components.
  • the CU encoding unit 1103 encodes sig_coeff_flag of the DC component and the AC0 component, and sets 0 to sig_coeff_flag of the other AC components.
  • FIG. 19 (4) is a flowchart showing the encoding process (4) above.
  • FIG. 19 (2) is the same as FIG. 19 (2) except that the processes of S1904 and S1914 in FIG. 19 (2) are changed to S19042 and S19142.
  • the CU encoding unit 1103 checks whether there are non-zero transform coefficients other than DC, AC0, and AC1 components.
  • the CU encoding unit 1103 encodes the DC component, sig_coeff_flag of the AC0 and AC1 components, and sets sig_coeff_flag of the other AC components to 0.
  • the code amount of LAST is reduced by changing the definition of cbf.
  • the number of non-zero transform coefficients depends on the CU size and the quantization parameter (QP) in addition to intra coding and inter coding.
  • CUmax max (CUwidth, CUheight).
  • (1) of Embodiment 1 is used. Note that THQ1 ⁇ THQ2 and THC1> THC2> THC3.
  • color difference Cr as shown in FIG. 20 (4), different threshold values may be set for Cb (example in FIG. 20 (3)) and Cr.
  • 21 and 22 are flowcharts showing operations of the entropy decoding unit 301 (CU decoding unit 1003) and the entropy encoding unit 104 (CU encoding unit 1103) when the method of the second embodiment is used. Steps denoted by the same numbers as in FIGS. 18 and 19 are the same processes as in FIGS.
  • the CU decoding unit 1003 determines the definition of cbf from the CU size and the quantization parameter.
  • the definition of cbf is as shown in FIG. 21 (2).
  • LAST and sig_coeff_flag are extracted from the encoded data, so no non-zero transform coefficient is set.
  • Fig. 23 shows an example where LAST is at the position (14, 6). If the LAST coordinates in the CU are expressed directly (in one step), they are expressed in 17 bits (1111111010, 1111100) using the variable length code table of FIG.
  • the LAST coordinates in the CU in addition to directly expressing the LAST coordinates in the CU, it can also be expressed in stages using the LAST sub-block position and the position in the sub-block.
  • the position of the subblock included in LAST is (3,1)
  • the position of LAST in the subblock at position (3,1) is (2,2). It is expressed by 12 bits (1110,10) and (110,110) using a long code table. Therefore, if the position of LAST is far away from (0,0), rather than directly encoding the coordinates of the LAST in the CU, it is divided into 2 sub-block positions containing the LAST and 2 LAST positions in the sub-block. The amount of code can be reduced by encoding in stages.
  • LAST is expressed in two-dimensional coordinates, but the encoding method is the same in both the horizontal and vertical directions, so the following description is based on one dimension (either horizontal or vertical).
  • FIG. 24 is a diagram showing the LAST position and the code amount necessary to express the LAST position.
  • the code table shown in FIG. 12 was used.
  • Fig. 24 (1) shows the case where the LAST coordinates in the CU are directly encoded
  • Fig. 24 (2) shows the case where the encoding is performed in two stages: the position of the LAST in the sub-block and the position of the LAST in the sub-block. It is.
  • the code amount is less if the LAST coordinates in the CU are directly encoded, but in other cases, the LAST is included It can be seen that the amount of code is smaller when encoding is performed in two stages, the sub-block position and the LAST position in the sub-block.
  • FIG. 25 is a flowchart showing the operation when S1304 in FIG. 13 and S1805 in FIG. 18 are replaced by the method of the third embodiment.
  • the CU decoding unit 1003 decodes the position of the sub block including the LAST and the position of the LAST in the sub block. In S2502, the CU decoding unit 1003 derives the coordinates of the LAST in the CU from the position of the subblock including the decoded LAST and the position of the LAST in the subblock.
  • FIG. 26 is a flowchart showing the operation when S1407 to S1408 in FIG. 14 and S1910 in FIG. 19 are replaced by the method of the third embodiment.
  • the CU encoding unit 1103 derives the LAST position in the CU.
  • the CU encoding unit 1103 derives the position of the subblock including LAST in S2602, derives the position of LAST in the subblock in S2603, and positions of the subblock including LAST and the position of LAST in the subblock in S2604. Is encoded.
  • the amount of LAST code can be reduced by encoding LAST in two stages in intra encoding and directly encoding LAST in inter encoding.
  • different threshold values may be set for Cb and Cr.
  • encoding is performed in two stages: the position of the subblock including LAST and the position of LAST within the subblock, only when there is a high possibility that LAST is included in other than the upper left subblock of CU. As a result, the LAST code amount reduction effect can be enhanced.
  • the variable length code table A in FIG. 12 is used.
  • the variable-length code table A in FIG. 12 has an advantage that the code amount is small when LAST is small. However, since the rate of increase in the code amount increases as LAST increases, there is a disadvantage when there are many large values of LAST.
  • the code amount when the LAST is small is larger than the variable length code of FIG. A variable length code table that does not become too large will be described.
  • variable length code tables 28 (1) to (4) are examples of variable length code tables that are effective when LAST is large (the code amount does not become so large).
  • the portion indicated by “1” and “0” is last_sig_coeff_Z_prefix (Z is x or y)
  • the portion indicated by “X” is last_sig_coeff_Z_suffix (Z is x or y).
  • “X” is “1” or “0”.
  • the code amount when LAST is “3” is 4 bits when using the variable length code table A of FIG. 12, 4 bits when using the variable length code table B of FIG. 28 (1), and the variable length of FIG. 28 (2).
  • the code table C is 3 bits
  • the variable length code table D in FIG. 28 (3) is 3 bits
  • the variable length code table E in FIG. 28 (4) is 4 bits.
  • the code amount when LAST is “7” is 7 bits when using the variable length code table A in FIG. 12, 6 bits when using the variable length code table B in FIG. 28 (1), and the variable length code in FIG. 28 (2). If Table C is used, it will be 5 bits, if variable length code table D of FIG. 28 (3) is used, it will be 4 bits, and if variable length code table E of FIG. 28 (4) is used, it will be 4 bits.
  • the code amount when LAST is “15” is 10 bits when using the variable length code table A of FIG. 12, 8 bits when using the variable length code table B of FIG. 28 (1), and the variable length code of FIG. 28 (2). If Table C is used, it will be 8 bits, if variable length code table D of FIG. 28 (3) is used, it will be 6 bits, and if variable length code table E of FIG. 28 (4) is used, it will be 5 bits.
  • variable length code tables of FIGS. 28 (1) to (4) are used, and in other cases (inter coding and By encoding the color difference component using the variable length code table of FIG. 12, the amount of LAST code can be reduced.
  • the intra coding of the luminance component has been described as a condition that the non-zero transform coefficient exists in the high frequency region and the LAST coordinate becomes large.
  • the CU size (CUwidth: CU width, CUheight: CU height) and quantization parameter (QP) are set under the condition that the non-zero transform coefficient exists in the high frequency region and the LAST coordinate becomes large.
  • QP quantization parameter
  • FIG. 29 is a table showing a code amount necessary for LAST encoding when the variable length code tables A to E of FIGS. 12 and 28 are used.
  • 29A is a variable length code table of FIG. 12
  • B is a variable length code table of FIG. 28 (1)
  • C is FIG. 28 (2)
  • D is FIG. 28 (3)
  • E is FIG.
  • a variable length code table B close to fixed length coding is preferable.
  • the code amount of LAST can be further reduced by switching the variable-length code tables C, D, and E depending on the concentrated location. For example, when a large number of non-zero transform coefficients exist in high frequency components, the variable length code table E is used.
  • the long code table D (FIG. 28 (3)) is used.
  • Code table D (FIG. 28 (3)) is used.
  • A, D, and E are used for the variable length code table, but B and C may be used instead of D and E.
  • FIG. 31 (1) is a flowchart showing a process of determining a variable length code table based on a quantization parameter and a CU size, and encoding or decoding LAST.
  • the CU decoding unit 1003 refers to the quantization parameter and the CU size to determine a variable length code table used for LAST encoding or decoding.
  • the CU decoding unit 1003 encodes or decodes LAST using the variable length code table determined in S3101.
  • the encoding process of LAST is the same as the process shown in FIG. 14 and the decoding process is the same as the process shown in FIG.
  • variable length code table used for the LAST encoding As described above, when the non-zero transform coefficient exists in the high frequency region and the LAST coordinate becomes large, by switching the variable length code table used for the LAST encoding according to the CU size and the quantization parameter, The code amount can be further reduced by taking advantage of the characteristics of the variable length code table.
  • Modification 2 In the first modification, the variable length code table to be used is switched depending on the quantization parameter and the CU size. In the second modification, a method for switching the variable length code table used for LAST encoding according to the scan direction will be described.
  • FIG. 32 shows the scan order (reverse order) of diagonal scan, horizontal scan, and vertical scan when the CU size is 16x16.
  • the number of bits required for LAST encoding increases at an earlier stage of the scan order than in the oblique scan.
  • FIG. 33 (1) shows the code amount of LAST at the time of oblique scanning and horizontal / vertical scanning when the variable length code table A of FIG. 12 is used. In the horizontal and vertical scans, it can be seen that the code amount of LAST increases at an early stage of the scan.
  • FIG. 33 (2) shows an example of using the variable length code table of FIG. 28 in the horizontal direction and the variable length code table of FIG. (3) is an example in which the variable length code table of FIG.
  • FIG. 33 (4) shows the code amount of LAST in this case. It can be seen that the code amount in the high frequency region can be reduced as compared with the case where only the variable length code table of FIG. 12 shown in FIG. 33 (1) is used.
  • FIG. 31 (2) is a flowchart showing a process of determining a variable length code table according to the scan direction and encoding or decoding LAST.
  • the difference from FIG. 31 (1) is that S3101 in FIG. 31 (1) is changed to S31011 in FIG. 31 (2), and the other processes are the same, and thus the description thereof is omitted.
  • the CU decoding unit 1003 determines a variable length code table to be used for LAST encoding or decoding with reference to the scan direction.
  • variable length code table is switched by switching the variable length code table used for LAST encoding according to the scan direction.
  • the amount of codes can be further reduced by taking advantage of the above characteristics.
  • Modification 3 a method of switching the variable-length code table used for LAST encoding according to the number of non-zero transform coefficients of adjacent CUs of the target CU will be described.
  • the number of non-zero transform coefficients included in the upper adjacent CU (CU_A) and left adjacent CU (CU_L) of the target CU (CU_C) shown in FIG. 34 is NA and NL, and the estimated values of the non-zero transform coefficients of the target CU N is derived.
  • the target CU is inter-prediction
  • the number of non-zero transform coefficients of adjacent CUs of inter prediction is the number of transform coefficients N of the target CU, and both CU_A and CU_L are inter-predicted.
  • the average value of NA and NL is N.
  • the number of non-zero transform coefficients may be scaled. If the areas of CU_A, CU_B, and CU_C are AA, AB, and AC, scaling according to the following formula is performed first.
  • an estimated value N of the number of non-zero conversion coefficients of the target CU is derived according to the above description.
  • LAST may be encoded using the variable length code table E.
  • FIG. 31 (3) is a flowchart showing a process of determining a variable-length coding table based on the estimated value of the number of non-zero transform coefficients of the target CU and coding or decoding LAST.
  • the difference from FIG. 31 (1) is that S3101 in FIG. 31 (1) is changed to S31012 in FIG. 31 (2), and the other processes are the same, and the description thereof is omitted.
  • the CU decoding unit 1003 refers to the estimated value of the number of non-zero transform coefficients of the target CU to determine a variable length code table used for LAST encoding or decoding.
  • variable-length code table used for LAST encoding is switched depending on the number of non-zero coefficients of adjacent CUs.
  • the code amount can be further reduced by taking advantage of the characteristics of the variable-length code table.
  • An image encoding apparatus includes a unit that divides one screen of an input moving image into encoding units (CU) including a plurality of pixels, and performs a predetermined conversion using the CU as a unit to obtain a conversion coefficient.
  • Means for determining the value of the flag and non-zero conversion only within a limited region of the CU Means for determining the value of the second flag indicating whether or not a coefficient exists, and the most distant position (LAST) and non-zero coefficient value (LEVEL) by scanning the conversion coefficient from the DC component in the scan order in the CU
  • Means for deriving the syntax indicating the encoding parameter prediction mode (intra Or inter), quantization parameter, CU size), variable length encoding means by switching which one of the first flag and the second flag is variable length encoded, and the first When variable-length coding is performed on the flag, and there is a non-zero transform coefficient in the CU, means for encoding syntax indicating LAST and LEVEL, and variable-length coding of the second flag If the non-zero transform coefficient exists only in the limited region of CU, the LEVEL syntax is encoded.If the non-zero transform coefficient exists outside the
  • An image decoding apparatus includes a unit that performs variable length decoding of encoded data using a coding unit (CU) including a plurality of pixels as a processing unit and outputs a syntax, and a transform coefficient from the syntax.
  • the variable length decoding means refers to a coding parameter (prediction mode (intra or inter), quantization parameter, CU size), and whether a non-zero transform coefficient exists in the CU.
  • Non-zero only in the first flag indicating whether or not in a limited region of the CU (DC component only, DC component and first AC component, or DC component and first and second AC components) Means for variable length decoding by switching which of the second flags indicating whether or not a transform coefficient exists, variable length decoding; variable length decoding of the first flag; wherein the first flag is CU Must have nonzero transform coefficients
  • An image encoding apparatus includes a unit that divides one screen of the input moving image into encoding units (CU) including a plurality of pixels, and performs a predetermined conversion using the CU as a unit to perform a conversion coefficient.
  • variable length coding means for variable length coding the transform coefficient indicates whether or not there is a non-zero transform coefficient in the CU.
  • Means for determining the value of the first flag means for variable-length coding the first flag, the farthest position (CU LAST) and the non-zero position in the CU by scanning the transform coefficient from the DC component in the scan order
  • Means for deriving syntax indicating coefficient value (LEVEL) means for dividing CU into sub-blocks, position of sub-block including LAST, and position of LAST within sub-block (LAST of sub-block) And the first flag is non-zero in the CU.
  • variable-length encoding Whether to perform variable-length encoding of CU LAST or sub-block LAST with reference to encoding parameter prediction mode (intra or inter), quantization parameter, CU size) And a variable length coding means for switching and a means for coding a syntax indicating LEVEL.
  • An image decoding apparatus includes a unit that performs variable length decoding of encoded data using a coding unit (CU) including a plurality of pixels as a processing unit and outputs a syntax, and a transform coefficient from the syntax.
  • a means for decoding one of LAST of CU and LAST of sub-block with reference to encoding parameter prediction mode (intra or inter), quantization parameter, CU size), and sub-block In the case of decoding LAST, a means for deriving LAST of CU and a means for decoding syntax indicating LEVEL are provided.
  • An image encoding apparatus includes a unit that divides one screen of the input moving image into encoding units (CU) including a plurality of pixels, and performs a predetermined conversion using the CU as a unit to perform a conversion coefficient.
  • variable length coding means for variable length coding the transform coefficient indicates whether or not there is a non-zero transform coefficient in the CU.
  • An image decoding apparatus includes a unit that performs variable length decoding of encoded data using a coding unit (CU) including a plurality of pixels as a processing unit and outputs a syntax, and a transform coefficient from the syntax.
  • coding parameters prediction mode (intra or inter), quantization parameter, CU size, number of non-zero coefficients of adjacent CUs, scan direction
  • variable used for LAST decoding It is characterized by comprising means for variable length decoding by switching the long code table and means for decoding the syntax indicating LEVEL.
  • a part of the image encoding device 11 and the image decoding device 31 in the above-described embodiment for example, the entropy decoding unit 301, the prediction parameter decoding unit 302, the loop filter 305, the predicted image generation unit 308, the inverse quantization / inverse DCT.
  • the prediction parameter encoding unit 111 may be realized by a computer.
  • the program for realizing the control function may be recorded on a computer-readable recording medium, and the program recorded on the recording medium may be read by a computer system and executed.
  • the “computer system” is a computer system built in either the image encoding device 11 or the image decoding device 31 and includes hardware such as an OS and peripheral devices.
  • the “computer-readable recording medium” refers to a storage device such as a portable medium such as a flexible disk, a magneto-optical disk, a ROM, a CD-ROM, or a hard disk built in a computer system.
  • the “computer-readable recording medium” is a medium that dynamically holds a program for a short time, such as a communication line when transmitting a program via a network such as the Internet or a communication line such as a telephone line,
  • a volatile memory inside a computer system serving as a server or a client may be included and a program that holds a program for a certain period of time.
  • the program may be a program for realizing a part of the functions described above, and may be a program capable of realizing the functions described above in combination with a program already recorded in a computer system.
  • part or all of the image encoding device 11 and the image decoding device 31 in the above-described embodiment may be realized as an integrated circuit such as an LSI (Large Scale Integration).
  • LSI Large Scale Integration
  • Each functional block of the image encoding device 11 and the image decoding device 31 may be individually made into a processor, or a part or all of them may be integrated into a processor.
  • the method of circuit integration is not limited to LSI, and may be realized by a dedicated circuit or a general-purpose processor. Further, in the case where an integrated circuit technology that replaces LSI appears due to progress in semiconductor technology, an integrated circuit based on the technology may be used.
  • the image encoding device 11 and the image decoding device 31 described above can be used by being mounted on various devices that perform transmission, reception, recording, and reproduction of moving images.
  • the moving image may be a natural moving image captured by a camera or the like, or may be an artificial moving image (including CG and GUI) generated by a computer or the like.
  • FIG. 8 is a block diagram showing a configuration of a transmission device PROD_A in which the image encoding device 11 is mounted.
  • the transmission apparatus PROD_A modulates a carrier wave with an encoding unit PROD_A1 that obtains encoded data by encoding a moving image, and with the encoded data obtained by the encoding unit PROD_A1.
  • a modulation unit PROD_A2 that obtains a modulation signal and a transmission unit PROD_A3 that transmits the modulation signal obtained by the modulation unit PROD_A2 are provided.
  • the above-described image encoding device 11 is used as the encoding unit PROD_A1.
  • Transmission device PROD_A as a source of moving images to be input to the encoding unit PROD_A1, a camera PROD_A4 that captures moving images, a recording medium PROD_A5 that records moving images, an input terminal PROD_A6 for inputting moving images from the outside, and An image processing unit A7 that generates or processes an image may be further provided.
  • FIG. 8A illustrates a configuration in which the transmission apparatus PROD_A includes all of these, but some of them may be omitted.
  • the recording medium PROD_A5 may be a recording of a non-encoded moving image, or a recording of a moving image encoded by a recording encoding scheme different from the transmission encoding scheme. It may be a thing. In the latter case, a decoding unit (not shown) for decoding the encoded data read from the recording medium PROD_A5 in accordance with the recording encoding method may be interposed between the recording medium PROD_A5 and the encoding unit PROD_A1.
  • FIG. 8 is a block diagram showing a configuration of a receiving device PROD_B in which the image decoding device 31 is mounted.
  • the receiving device PROD_B includes a receiving unit PROD_B1 that receives the modulated signal, a demodulating unit PROD_B2 that obtains encoded data by demodulating the modulated signal received by the receiving unit PROD_B1, and a demodulator.
  • a decoding unit PROD_B3 that obtains a moving image by decoding the encoded data obtained by the unit PROD_B2.
  • the above-described image decoding device 31 is used as the decoding unit PROD_B3.
  • the receiving device PROD_B is a display destination PROD_B4 for displaying a moving image, a recording medium PROD_B5 for recording a moving image, and an output terminal for outputting the moving image to the outside as a supply destination of the moving image output by the decoding unit PROD_B3 PROD_B6 may be further provided.
  • FIG. 8B illustrates a configuration in which all of these are provided in the receiving device PROD_B, but some of them may be omitted.
  • the recording medium PROD_B5 may be used for recording a non-encoded moving image, or is encoded using a recording encoding method different from the transmission encoding method. May be. In the latter case, an encoding unit (not shown) for encoding the moving image acquired from the decoding unit PROD_B3 according to the recording encoding method may be interposed between the decoding unit PROD_B3 and the recording medium PROD_B5.
  • the transmission medium for transmitting the modulation signal may be wireless or wired.
  • the transmission mode for transmitting the modulated signal may be broadcasting (here, a transmission mode in which the transmission destination is not specified in advance) or communication (here, transmission in which the transmission destination is specified in advance). Refers to the embodiment). That is, the transmission of the modulation signal may be realized by any of wireless broadcasting, wired broadcasting, wireless communication, and wired communication.
  • a terrestrial digital broadcast broadcasting station (broadcasting equipment, etc.) / Receiving station (such as a television receiver) is an example of a transmitting device PROD_A / receiving device PROD_B that transmits and receives a modulated signal by wireless broadcasting.
  • a broadcasting station (such as broadcasting equipment) / receiving station (such as a television receiver) of cable television broadcasting is an example of a transmitting device PROD_A / receiving device PROD_B that transmits and receives a modulated signal by cable broadcasting.
  • a server workstation, etc.
  • Client television receiver, personal computer, smartphone, etc.
  • VOD Video On Demand
  • video sharing service using the Internet is a transmission device that transmits and receives modulated signals via communication.
  • PROD_A / receiving device PROD_B normally, either a wireless or wired transmission medium is used in a LAN, and a wired transmission medium is used in a WAN.
  • the personal computer includes a desktop PC, a laptop PC, and a tablet PC.
  • the smartphone also includes a multi-function mobile phone terminal.
  • the video sharing service client has a function of encoding a moving image captured by the camera and uploading it to the server. That is, the client of the video sharing service functions as both the transmission device PROD_A and the reception device PROD_B.
  • FIG. 9A is a block diagram showing a configuration of a recording apparatus PROD_C equipped with the image encoding device 11 described above.
  • the recording apparatus PROD_C includes an encoding unit PROD_C1 that obtains encoded data by encoding a moving image, and the encoded data obtained by the encoding unit PROD_C1 on a recording medium PROD_M.
  • the above-described image encoding device 11 is used as the encoding unit PROD_C1.
  • the recording medium PROD_M may be of a type built into the recording device PROD_C, such as (1) HDD (Hard Disk Drive) or SSD (Solid State Drive), or (2) SD memory. It may be of the type connected to the recording device PROD_C, such as a card or USB (Universal Serial Bus) flash memory, or (3) DVD (Digital Versatile Disc) or BD (Blu-ray Disc: registration) Or a drive device (not shown) built in the recording device PROD_C.
  • HDD Hard Disk Drive
  • SSD Solid State Drive
  • SD memory such as a card or USB (Universal Serial Bus) flash memory, or (3) DVD (Digital Versatile Disc) or BD (Blu-ray Disc: registration) Or a drive device (not shown) built in the recording device PROD_C.
  • the recording device PROD_C is a camera PROD_C3 that captures moving images as a source of moving images to be input to the encoding unit PROD_C1, an input terminal PROD_C4 for inputting moving images from the outside, and a reception for receiving moving images
  • a unit PROD_C5 and an image processing unit PROD_C6 for generating or processing an image may be further provided.
  • FIG. 9A illustrates a configuration in which the recording apparatus PROD_C includes all of these, but some of them may be omitted.
  • the receiving unit PROD_C5 may receive a non-encoded moving image, or may receive encoded data encoded by a transmission encoding scheme different from the recording encoding scheme. You may do. In the latter case, a transmission decoding unit (not shown) that decodes encoded data encoded by the transmission encoding method may be interposed between the reception unit PROD_C5 and the encoding unit PROD_C1.
  • Examples of such a recording device PROD_C include a DVD recorder, a BD recorder, an HDD (Hard Disk Drive) recorder, and the like (in this case, the input terminal PROD_C4 or the receiver PROD_C5 is a main source of moving images). .
  • a camcorder in this case, the camera PROD_C3 is a main source of moving images
  • a personal computer in this case, the receiving unit PROD_C5 or the image processing unit C6 is a main source of moving images
  • a smartphone this In this case, the camera PROD_C3 or the reception unit PROD_C5 is a main source of moving images
  • the like is also an example of such a recording apparatus PROD_C.
  • FIG. 9 is a block showing a configuration of a playback device PROD_D in which the above-described image decoding device 31 is mounted.
  • the playback device PROD_D reads a moving image by decoding a read unit PROD_D1 that reads encoded data written to the recording medium PROD_M and a read unit PROD_D1 that reads the encoded data. And a decoding unit PROD_D2 to obtain.
  • the above-described image decoding device 31 is used as the decoding unit PROD_D2.
  • the recording medium PROD_M may be of the type built into the playback device PROD_D, such as (1) HDD or SSD, or (2) such as an SD memory card or USB flash memory. It may be of the type connected to the playback device PROD_D, or (3) may be loaded into a drive device (not shown) built in the playback device PROD_D, such as a DVD or BD. Good.
  • the playback device PROD_D has a display unit PROD_D3 that displays a moving image as a supply destination of the moving image output by the decoding unit PROD_D2, an output terminal PROD_D4 that outputs the moving image to the outside, and a transmission unit that transmits the moving image.
  • PROD_D5 may be further provided.
  • FIG. 9B illustrates a configuration in which the playback apparatus PROD_D includes all of these, but some of them may be omitted.
  • the transmission unit PROD_D5 may transmit a non-encoded moving image, or transmits encoded data encoded by a transmission encoding scheme different from the recording encoding scheme. You may do. In the latter case, it is preferable to interpose an encoding unit (not shown) that encodes a moving image using a transmission encoding method between the decoding unit PROD_D2 and the transmission unit PROD_D5.
  • Examples of such a playback device PROD_D include a DVD player, a BD player, and an HDD player (in this case, an output terminal PROD_D4 to which a television receiver or the like is connected is a main moving image supply destination). .
  • a television receiver in this case, the display PROD_D3 is a main supply destination of moving images
  • a digital signage also referred to as an electronic signboard or an electronic bulletin board
  • the display PROD_D3 or the transmission unit PROD_D5 is the main supply of moving images
  • Desktop PC in this case, output terminal PROD_D4 or transmission unit PROD_D5 is the main video source
  • laptop or tablet PC in this case, display PROD_D3 or transmission unit PROD_D5 is video
  • a smartphone which is a main image supply destination
  • a smartphone in this case, the display PROD_D3 or the transmission unit PROD_D5 is a main moving image supply destination
  • the like are also examples of such a playback device PROD_D.
  • Each block of the image decoding device 31 and the image encoding device 11 described above may be realized in hardware by a logic circuit formed on an integrated circuit (IC chip), or may be a CPU (Central Processing Unit). You may implement
  • IC chip integrated circuit
  • CPU Central Processing Unit
  • each of the above devices includes a CPU that executes instructions of a program that realizes each function, a ROM (Read Memory) that stores the program, a RAM (RandomAccess Memory) that expands the program, the program, and various data
  • a storage device such as a memory for storing the.
  • the object of the embodiment of the present invention is a record in which the program code (execution format program, intermediate code program, source program) of the control program for each of the above devices, which is software that realizes the above-described functions, is recorded in a computer-readable manner This can also be achieved by supplying a medium to each of the above devices, and reading and executing the program code recorded on the recording medium by the computer (or CPU or MPU).
  • Examples of the recording medium include tapes such as magnetic tapes and cassette tapes, magnetic disks such as floppy (registered trademark) disks / hard disks, CD-ROMs (Compact Disc Read-Only Memory) / MO discs (Magneto-Optical discs).
  • tapes such as magnetic tapes and cassette tapes
  • magnetic disks such as floppy (registered trademark) disks / hard disks
  • CD-ROMs Compact Disc Read-Only Memory
  • MO discs Magnetic-Optical discs
  • IC cards including memory cards
  • Cards such as optical cards
  • Semiconductor memories such as flash ROM, or PLD (Programmable logic device ) Or FPGA (Field Programmable Gate Gate Array) or the like.
  • each of the above devices may be configured to be connectable to a communication network, and the program code may be supplied via the communication network.
  • the communication network is not particularly limited as long as it can transmit the program code.
  • Internet intranet, extranet, LAN (Local Area Network), ISDN (Integrated Services Digital Network), VAN (Value-Added Network), CATV (Community Area Antenna / television / Cable Television), Virtual Private Network (Virtual Private Network) Network), telephone line network, mobile communication network, satellite communication network, and the like.
  • the transmission medium constituting the communication network may be any medium that can transmit the program code, and is not limited to a specific configuration or type.
  • IEEE Institute of Electrical and Electronic Engineers 1394, USB, power line carrier, cable TV line, telephone line, ADSL (Asymmetric Digital Subscriber Line) line, etc. wired such as IrDA (Infrared Data Association) or remote control , BlueTooth (registered trademark), IEEE802.11 wireless, HDR (High Data Rate), NFC (Near Field Communication), DLNA (Digital Living Network Alliance: registered trademark), mobile phone network, satellite line, terrestrial digital broadcasting network, etc. It can also be used wirelessly.
  • the embodiment of the present invention can also be realized in the form of a computer data signal embedded in a carrier wave in which the program code is embodied by electronic transmission.
  • Embodiments of the present invention are preferably applied to an image decoding apparatus that decodes encoded data in which image data is encoded, and an image encoding apparatus that generates encoded data in which image data is encoded. it can. Further, the present invention can be suitably applied to the data structure of encoded data generated by an image encoding device and referenced by the image decoding device.
  • CT information decoding unit 11 Image encoding device 20
  • CU decoding unit 31 Image decoding device 41

Landscapes

  • Engineering & Computer Science (AREA)
  • Multimedia (AREA)
  • Signal Processing (AREA)
  • Compression Or Coding Systems Of Tv Signals (AREA)

Abstract

Selon l'invention, la définition d'un indicateur qui indique le nombre de coefficients de conversion non nuls dans une unité de codage (CU) est modifiée pour réduire la longueur du code pour LAST qui indique la position du dernier coefficient de conversion non nul dans la CU.
PCT/JP2018/006937 2017-03-03 2018-02-26 Dispositif de codage d'image animée et dispositif de décodage d'image animée WO2018159526A1 (fr)

Applications Claiming Priority (2)

Application Number Priority Date Filing Date Title
JP2017-040322 2017-03-03
JP2017040322A JP2020072277A (ja) 2017-03-03 2017-03-03 動画像符号化装置及び動画像復号装置

Publications (1)

Publication Number Publication Date
WO2018159526A1 true WO2018159526A1 (fr) 2018-09-07

Family

ID=63370019

Family Applications (1)

Application Number Title Priority Date Filing Date
PCT/JP2018/006937 WO2018159526A1 (fr) 2017-03-03 2018-02-26 Dispositif de codage d'image animée et dispositif de décodage d'image animée

Country Status (2)

Country Link
JP (1) JP2020072277A (fr)
WO (1) WO2018159526A1 (fr)

Cited By (4)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN112753221A (zh) * 2018-09-28 2021-05-04 松下电器(美国)知识产权公司 编码装置、解码装置、编码方法和解码方法
CN113039794A (zh) * 2018-11-16 2021-06-25 索尼集团公司 图像处理装置和方法
CN113196776A (zh) * 2018-12-20 2021-07-30 夏普株式会社 预测图像生成装置、运动图像解码装置、运动图像编码装置以及预测图像生成方法
CN114979641A (zh) * 2021-02-21 2022-08-30 腾讯科技(深圳)有限公司 视频编解码方法、装置、计算机可读介质及电子设备

Families Citing this family (1)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
EP4088465A4 (fr) * 2020-02-04 2023-11-29 Huawei Technologies Co., Ltd. Codeur, décodeur et procédés correspondants pour signaler une syntaxe de haut niveau

Citations (3)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
JP2013098988A (ja) * 2011-11-01 2013-05-20 Research In Motion Ltd 符号化および復号のためのマルチレベル有効性写像
WO2013099892A1 (fr) * 2011-12-28 2013-07-04 シャープ株式会社 Dispositif de décodage arithmétique, dispositif de décodage d'image et dispositif de codage arithmétique
US20160080749A1 (en) * 2013-01-16 2016-03-17 Blackberry Limited Transform coefficient coding for context-adaptive binary entropy coding of video

Patent Citations (3)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
JP2013098988A (ja) * 2011-11-01 2013-05-20 Research In Motion Ltd 符号化および復号のためのマルチレベル有効性写像
WO2013099892A1 (fr) * 2011-12-28 2013-07-04 シャープ株式会社 Dispositif de décodage arithmétique, dispositif de décodage d'image et dispositif de codage arithmétique
US20160080749A1 (en) * 2013-01-16 2016-03-17 Blackberry Limited Transform coefficient coding for context-adaptive binary entropy coding of video

Non-Patent Citations (1)

* Cited by examiner, † Cited by third party
Title
BENJAMIN, B. ET AL.: "High efficiency video coding(HEVC) text specification draft 7", JOINT COLLABORATIVE TEAM ON VIDEO CODING(JCT-VC) OF ITU- T SG 16 WP3 AND ISO/IEC JTCL/SC29/WG119TH MEETING, 27 April 2012 (2012-04-27), Geneva, Switzerland, pages 88 - 91 *

Cited By (7)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN112753221A (zh) * 2018-09-28 2021-05-04 松下电器(美国)知识产权公司 编码装置、解码装置、编码方法和解码方法
CN112753221B (zh) * 2018-09-28 2024-04-05 松下电器(美国)知识产权公司 编码装置、解码装置、编码方法和解码方法
CN113039794A (zh) * 2018-11-16 2021-06-25 索尼集团公司 图像处理装置和方法
CN113039794B (zh) * 2018-11-16 2024-05-24 索尼集团公司 图像处理装置和方法
CN113196776A (zh) * 2018-12-20 2021-07-30 夏普株式会社 预测图像生成装置、运动图像解码装置、运动图像编码装置以及预测图像生成方法
CN113196776B (zh) * 2018-12-20 2023-12-19 夏普株式会社 预测图像生成装置、运动图像解码装置、运动图像编码装置以及预测图像生成方法
CN114979641A (zh) * 2021-02-21 2022-08-30 腾讯科技(深圳)有限公司 视频编解码方法、装置、计算机可读介质及电子设备

Also Published As

Publication number Publication date
JP2020072277A (ja) 2020-05-07

Similar Documents

Publication Publication Date Title
JP7223886B2 (ja) 画像復号方法
WO2018037853A1 (fr) Appareil de décodage d'images et appareil de codage d'images
WO2018116925A1 (fr) Dispositif de génération d'image de prédiction intra, dispositif de décodage d'image et dispositif de codage d'image
WO2018199001A1 (fr) Dispositif de décodage d'image et dispositif de codage d'image
WO2018116802A1 (fr) Dispositif de décodage d'images, dispositif de codage d'images, et dispositif de prédiction d'images
WO2018221368A1 (fr) Dispositif de décodage d'image animée et dispositif de codage d'image animée
JP2021010046A (ja) 画像符号化装置及び画像復号装置
EP3557873B1 (fr) Dispositif de decodage d'image et dispositif de codage d'image
WO2018037896A1 (fr) Appareil de décodage d'image, appareil de codage d'image, procédé de décodage d'image et procédé de codage d'image
WO2019221072A1 (fr) Dispositif de codage d'images, dispositif d'extraction de flux codés et dispositif de décodage d'images
WO2018159526A1 (fr) Dispositif de codage d'image animée et dispositif de décodage d'image animée
WO2018110203A1 (fr) Appareil de décodage d'images animées et appareil de codage d'images animées
JPWO2019031410A1 (ja) 画像フィルタ装置、画像復号装置、および画像符号化装置
WO2018199002A1 (fr) Dispositif de codage d'image animée et dispositif de décodage d'image animée
JP7241153B2 (ja) 画像復号装置
WO2018216688A1 (fr) Dispositif d'encodage vidéo, dispositif de décodage vidéo et filtre
JP2019201255A (ja) 画像フィルタ装置、画像復号装置、および画像符号化装置
WO2019230904A1 (fr) Dispositif de décodage d'image et dispositif de codage d'image
WO2018143289A1 (fr) Dispositif de codage d'image et dispositif de décodage d'image
WO2019131349A1 (fr) Dispositif de décodage d'image et dispositif de codage d'image
WO2018061550A1 (fr) Dispositif de décodage d'image et dispositif de codage d'image
WO2018173862A1 (fr) Dispositif de décodage d'image et dispositif de codage d'image
JP2020036101A (ja) 画像復号装置および画像符号化装置
JP2019201256A (ja) 画像フィルタ装置
WO2020004277A1 (fr) Dispositif de décodage d'image et dispositif de codage d'image

Legal Events

Date Code Title Description
121 Ep: the epo has been informed by wipo that ep was designated in this application

Ref document number: 18760461

Country of ref document: EP

Kind code of ref document: A1

NENP Non-entry into the national phase

Ref country code: DE

122 Ep: pct application non-entry in european phase

Ref document number: 18760461

Country of ref document: EP

Kind code of ref document: A1

NENP Non-entry into the national phase

Ref country code: JP