WO2014049982A1 - Video encoding device, video decoding device, video encoding method and video decoding method - Google Patents

Video encoding device, video decoding device, video encoding method and video decoding method Download PDF

Info

Publication number
WO2014049982A1
WO2014049982A1 PCT/JP2013/005285 JP2013005285W WO2014049982A1 WO 2014049982 A1 WO2014049982 A1 WO 2014049982A1 JP 2013005285 W JP2013005285 W JP 2013005285W WO 2014049982 A1 WO2014049982 A1 WO 2014049982A1
Authority
WO
WIPO (PCT)
Prior art keywords
block
prediction
encoding
unit
coding
Prior art date
Application number
PCT/JP2013/005285
Other languages
French (fr)
Japanese (ja)
Inventor
彰 峯澤
杉本 和夫
関口 俊一
Original Assignee
三菱電機株式会社
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Application filed by 三菱電機株式会社 filed Critical 三菱電機株式会社
Priority to JP2014538129A priority Critical patent/JPWO2014049982A1/en
Publication of WO2014049982A1 publication Critical patent/WO2014049982A1/en

Links

Images

Classifications

    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04NPICTORIAL COMMUNICATION, e.g. TELEVISION
    • H04N19/00Methods or arrangements for coding, decoding, compressing or decompressing digital video signals
    • H04N19/10Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using adaptive coding
    • H04N19/169Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using adaptive coding characterised by the coding unit, i.e. the structural portion or semantic portion of the video signal being the object or the subject of the adaptive coding
    • H04N19/18Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using adaptive coding characterised by the coding unit, i.e. the structural portion or semantic portion of the video signal being the object or the subject of the adaptive coding the unit being a set of transform coefficients
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04NPICTORIAL COMMUNICATION, e.g. TELEVISION
    • H04N19/00Methods or arrangements for coding, decoding, compressing or decompressing digital video signals
    • H04N19/10Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using adaptive coding
    • H04N19/102Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using adaptive coding characterised by the element, parameter or selection affected or controlled by the adaptive coding
    • H04N19/13Adaptive entropy coding, e.g. adaptive variable length coding [AVLC] or context adaptive binary arithmetic coding [CABAC]
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04NPICTORIAL COMMUNICATION, e.g. TELEVISION
    • H04N19/00Methods or arrangements for coding, decoding, compressing or decompressing digital video signals
    • H04N19/10Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using adaptive coding
    • H04N19/134Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using adaptive coding characterised by the element, parameter or criterion affecting or controlling the adaptive coding
    • H04N19/157Assigned coding mode, i.e. the coding mode being predefined or preselected to be further used for selection of another element or parameter
    • H04N19/16Assigned coding mode, i.e. the coding mode being predefined or preselected to be further used for selection of another element or parameter for a given display mode, e.g. for interlaced or progressive display mode
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04NPICTORIAL COMMUNICATION, e.g. TELEVISION
    • H04N19/00Methods or arrangements for coding, decoding, compressing or decompressing digital video signals
    • H04N19/50Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using predictive coding
    • H04N19/593Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using predictive coding involving spatial prediction techniques

Definitions

  • the present invention relates to a moving image encoding apparatus and moving image encoding method for encoding a moving image with high efficiency, a moving image decoding apparatus and a moving image decoding method for decoding a moving image encoded with high efficiency, and It is about.
  • an input video frame is divided into macroblock units each composed of 16 ⁇ 16 pixel blocks, and after motion compensation prediction is performed, a prediction difference signal is orthogonally converted into block units.
  • Information compression is performed by quantization.
  • FIG. 23 shows MPEG-4 AVC / H. 1 is a configuration diagram illustrating a H.264 video encoding device.
  • the block division unit 101 when the block division unit 101 receives an image signal to be encoded, the block division unit 101 divides the image signal into macroblock units, and outputs the macroblock unit image signal to the prediction unit 102 as a divided image signal. To do.
  • the prediction unit 102 receives the divided image signal from the block division unit 101, the prediction unit 102 predicts the image signal of each color component in the macroblock within a frame or between frames, and calculates a prediction difference signal.
  • a motion vector is searched in units of a macroblock itself or a subblock obtained by further dividing a macroblock. Then, using the motion vector, a motion compensation prediction image is generated by performing motion compensation prediction on the reference image signal stored in the memory 107, and a prediction signal indicating the motion compensation prediction image and the divided image signal are generated. A prediction difference signal is calculated by obtaining the difference.
  • Non-Patent Document 1 can select one prediction mode from a plurality of prediction modes for each block as a luminance intra prediction mode.
  • FIG. 24 is an explanatory diagram illustrating an intra prediction mode when the luminance block size is 4 ⁇ 4 pixels.
  • white circles in the block represent pixels to be encoded, and black circles represent encoded pixels that are pixels used for prediction.
  • nine intra prediction modes from mode 0 to mode 8 are defined.
  • mode 2 is a mode in which average value (DC) prediction is performed, and the pixels in the block are predicted based on the average value of the adjacent pixels above and to the left of the block.
  • Modes other than mode 2 are modes in which directionality prediction is performed.
  • Mode 0 is vertical prediction, in which a prediction image is generated by repeating adjacent pixels on the block in the vertical direction. For example, mode 0 is selected for a vertical stripe pattern.
  • Mode 1 is horizontal prediction, in which a prediction image is generated by repeating the adjacent pixel on the left of the block in the horizontal direction. For example, mode 1 is selected for a horizontal stripe pattern.
  • mode 3 to mode 8 a predicted image is generated by generating an interpolation pixel in a predetermined direction (direction indicated by an arrow) using the encoded pixels on the upper or left side of the block.
  • the luminance block size to which the intra prediction is applied can be selected from 4 ⁇ 4 pixels, 8 ⁇ 8 pixels, and 16 ⁇ 16 pixels, and in the case of 8 ⁇ 8 pixels, 4 ⁇ 4 pixels.
  • nine intra prediction modes are defined.
  • the pixels used for prediction are not encoded pixels themselves, but those obtained by performing filter processing on these pixels.
  • 16 ⁇ 16 pixels in addition to intra prediction modes related to average value prediction, vertical direction prediction, and horizontal direction prediction, four intra prediction modes called “plane prediction” are defined.
  • the intra prediction mode related to the Plane prediction is a mode in which a pixel generated by interpolating the upper and left encoded adjacent pixels in the block in an oblique direction is a prediction value.
  • the prediction unit 102 outputs the prediction signal generation parameter determined when obtaining the prediction signal to the variable length coding unit 108.
  • the prediction signal generation parameter includes, for example, information such as an intra prediction mode indicating how to perform spatial prediction within a frame and a motion vector indicating a motion amount between frames.
  • the compression unit 103 When the compression unit 103 receives the prediction difference signal from the prediction unit 102, the compression unit 103 performs DCT (discrete cosine transform) processing on the prediction difference signal to remove signal correlation, and then obtains compressed data by quantization.
  • DCT discrete cosine transform
  • the local decoding unit 104 receives the compressed data from the compression unit 103, the local decoding unit 104 inversely quantizes the compressed data and performs inverse DCT processing, whereby the prediction differential signal corresponding to the prediction differential signal output from the prediction unit 102 Is calculated.
  • the adder 105 When the adder 105 receives the prediction difference signal from the local decoding unit 104, the adder 105 adds the prediction difference signal and the prediction signal output from the prediction unit 102 to generate a local decoded image.
  • the loop filter 106 removes block distortion superimposed on the locally decoded image signal indicating the locally decoded image generated by the adder 105, and stores the locally decoded image signal after distortion removal in the memory 107 as a reference image signal. .
  • variable length coding unit 108 When the variable length coding unit 108 receives the compressed data from the compression unit 103, the variable length coding unit 108 performs entropy coding on the compressed data and outputs a bit stream that is a result of the coding. Note that the variable length coding unit 108 multiplexes the prediction signal generation parameter output from the prediction unit 102 into the bit stream and outputs the bit stream.
  • Non-Patent Document 1 in order to efficiently encode an interlace signal, various functions such as a function of adaptively switching whether the interlace signal is encoded as a frame or a field in units of pictures or macroblocks. A simple encoding tool is incorporated.
  • MPEG-4 AVC ISO / IEC 14496-10
  • Non-Patent Document 2 a special encoding tool for improving the encoding efficiency of interlaced signals is not prepared.
  • a filter process for increasing the continuity of block boundaries is performed on a predicted image in a specific intra prediction mode as shown in FIG. 27 during intra prediction.
  • the spatial correlation in the vertical direction is lowered, so that the effect of the filtering process on the upper end of the block may be greatly reduced.
  • Non-Patent Document 2 divides the orthogonal transform block into 4 ⁇ 4 pixel unit blocks (orthogonal transform sub-blocks) called Coefficients Group (CG) as an orthogonal transform coefficient encoding method, and encodes coefficients in CG units.
  • FIG. 28 shows the coding order (scan order) of the coefficients in the 16 ⁇ 16 pixel orthogonal transform block. In this way, 16 CGs in units of 4 ⁇ 4 pixels are encoded in order from the lower right CG, and each CG encodes 16 coefficients in the CG in order from the lower right coefficient.
  • flag information indicating whether or not a significant (nonzero) coefficient exists in 16 coefficients in the CG is encoded, and then a significant (nonzero) coefficient exists in the CG. Only whether each coefficient in the CG is a significant (non-zero) coefficient is encoded in the above order, and finally, coefficient value information is encoded in order for a significant (non-zero) coefficient. This is performed in the above order in units of CG. In this case, the encoding efficiency by entropy encoding can be increased by using a biased scan order so that significant (non-zero) coefficients are generated as continuously as possible. Since progressive video and interlaced video have different distributions of significant (non-zero) coefficients, they cannot be efficiently encoded in the scan order of FIG.
  • Non-Patent Document 2 Since the moving image encoding apparatus of Non-Patent Document 2 is configured as described above, when the interlace signal is field-encoded, the vertical spatial resolution becomes 1/2 and the inter-pixel correlation decreases. There has been a problem that the prediction efficiency of intra prediction and the encoding efficiency of orthogonal transform coefficients are reduced.
  • the present invention has been made to solve the above-described problems, and even when an interlace signal is field-encoded, it is possible to improve encoding efficiency when the interlace signal is field-encoded. It is an object to obtain a moving image encoding device, a moving image decoding device, a moving image encoding method, and a moving image decoding method.
  • the moving picture coding apparatus comprises variable length coding means for generating a coded bitstream in which compressed data and coding modes are multiplexed, and the variable length coding means orthogonally transforms orthogonal transform blocks. Based on whether or not a flag based on information indicating whether or not to perform field encoding is valid and divided into transform sub-blocks, the coding order of the transform coefficients after quantization that is compressed data is changed to the orthogonal transform block. The unit is switched in units of the orthogonal transform sub-block.
  • an orthogonal transform block is divided into orthogonal transform sub-blocks, and a quantized transform that is compressed data is based on whether or not a flag based on information indicating whether or not field coding is valid. Since the coding order of the coefficients is configured to be switched in units of the orthogonal transform block and the orthogonal transform sub-block, it is possible to realize efficient prediction processing and coding processing according to the characteristics of the field signal, There is an effect of improving the encoding efficiency.
  • (A) shows the distribution of the encoding block and prediction block after a division
  • (b) is explanatory drawing which shows the condition where encoding mode m ( Bn ) is allocated by hierarchy division
  • the upper left pixels in the prediction block P i n is an explanatory diagram showing a relative coordinate whose origin.
  • FIG. 1 is a block diagram showing a moving picture coding apparatus according to Embodiment 1 of the present invention.
  • the slice dividing unit 14 when a video signal is input as an input image, the slice dividing unit 14 performs a process of dividing the input image into one or more partial images called “slices” according to the slice division information determined by the encoding control unit 2. carry out.
  • the slice division unit can be finely divided to a coding block unit described later.
  • the slice division unit 14 constitutes a slice division unit.
  • the slice dividing unit 1 divides the slice into maximum coding blocks which are coding blocks of the maximum size determined by the coding control unit 2, and Until the upper limit number of hierarchies determined by the encoding control unit 2 is reached, a process of dividing the maximum encoded block hierarchically into each encoded block is performed. That is, the block dividing unit 1 divides the slice into each encoded block in accordance with the division determined by the encoding control unit 2, and performs a process of outputting the encoded block.
  • Each coding block is divided into one or a plurality of prediction blocks which are prediction processing units.
  • the block dividing unit 1 constitutes a block dividing unit.
  • the encoding control unit 2 determines the maximum size of the encoded block that is a processing unit when the encoding process is performed, and sets the upper limit number of layers when the encoded block of the maximum size is hierarchically divided. By determining, the process of determining the size of each encoded block is performed.
  • the encoding control unit 2 also includes one or more selectable encoding modes (one or more intra encoding modes having different prediction block sizes indicating prediction processing units, one or more inter codes having different prediction block sizes, and the like).
  • the coding mode to be applied to the coding block output from the block dividing unit 1 is selected from among the coding modes.
  • the selection method there is a method of selecting a coding mode having the highest coding efficiency for the coding block output from the block dividing unit 1 from one or more selectable coding modes.
  • the coding control unit 2 sets the intra prediction parameters used when performing the intra prediction processing on the coding block in the intra coding mode.
  • the coding mode having the highest coding efficiency is the inter coding mode, determined for each prediction block that is the prediction processing unit indicated by the intra coding mode, the inter prediction process for the coding block is performed in the inter coding mode.
  • the process which determines the inter prediction parameter used when implementing for every prediction block which is a prediction process unit which the said inter coding mode shows is implemented.
  • the encoding control unit 2 performs a process of determining a prediction difference encoding parameter to be given to the transform / quantization unit 7 and the inverse quantization / inverse transform unit 8.
  • the prediction difference encoding parameter includes orthogonal transform block division information indicating the division information of the orthogonal transform block, which is an orthogonal transform processing unit in the encoded block, and a quantum that specifies a quantization step size when the transform coefficient is quantized. Parameters are included.
  • the encoding control unit 2 constitutes an encoding control unit.
  • the changeover switch 3 outputs the coded block output from the block dividing unit 1 to the intra prediction unit 4 and the coding control unit 2. If the coding mode determined by the above is the inter coding mode, a process of outputting the coding block output from the block dividing unit 1 to the motion compensation prediction unit 5 is performed.
  • the intra prediction unit 4 When the intra control mode is selected by the encoding control unit 2 as the encoding mode corresponding to the encoded block output from the changeover switch 3, the intra prediction unit 4 performs the prediction process for the encoded block. For each prediction block which is a prediction processing unit, an intra prediction process (intraframe prediction) using an intra prediction parameter determined by the encoding control unit 2 while referring to a locally decoded image stored in the intra prediction memory 10. Process) to generate an intra-predicted image.
  • intra prediction process intraframe prediction
  • the motion compensation prediction unit 5 and the motion compensation prediction frame memory 12 A motion vector is searched by comparing a locally decoded image of one frame or more stored in the image with a prediction block unit that is a prediction processing unit, and the motion vector and a frame number to be referenced determined by the encoding control unit 2 Using the inter prediction parameters, an inter prediction process (motion compensation prediction process) for the encoded block is performed for each prediction block, and an inter prediction image is generated.
  • the intra prediction unit 4, the intra prediction memory 10, the motion compensation prediction unit 5, and the motion compensation prediction frame memory 12 constitute a prediction unit.
  • the subtraction unit 6 subtracts the intra prediction image generated by the intra prediction unit 4 or the inter prediction image generated by the motion compensated prediction unit 5 from the encoded block output from the block division unit 1 and performs the subtraction.
  • the process which outputs the prediction difference signal which shows the difference image which is a result to the conversion and quantization part 7 is implemented.
  • the subtracting unit 6 constitutes a difference image generating unit.
  • the transform / quantization unit 7 refers to the orthogonal transform block division information included in the prediction difference coding parameter determined by the coding control unit 2 and performs orthogonal transform processing on the prediction difference signal output from the subtraction unit 6 (for example, , DCT (discrete cosine transform), DST (discrete sine transform), orthogonal transform processing such as KL transform in which a base design is made in advance for a specific learning sequence is performed in units of orthogonal transform blocks to calculate transform coefficients
  • the transform coefficient of the orthogonal transform block unit is quantized, and the quantized compressed data that is the transform coefficient is inversely quantized / inversely transformed. 8 and the process of outputting to the variable length encoding unit 13 is performed.
  • the transform / quantization unit 7 constitutes an image compression unit.
  • FIG. 10 is an explanatory diagram illustrating an example of a 4 ⁇ 4 DCT quantization matrix.
  • the numbers in the figure indicate the scaling value of the quantization step size of each transform coefficient. For example, in order to suppress the encoding bit rate, as shown in FIG. 10, by scaling the quantization step size to a larger value for the higher frequency transform coefficient, the higher frequency generated in a complex image region or the like. It is possible to perform encoding without dropping information on low-frequency coefficients that greatly affect subjective quality while suppressing the conversion coefficient and suppressing the code amount.
  • a quantization matrix may be used.
  • the quantization matrix can use an independent matrix for each color signal and coding mode (intra coding or inter coding) at each orthogonal transform size. It is possible to select either a quantization matrix that is commonly prepared in the moving picture decoding apparatus or an already encoded quantization matrix or a new quantization matrix. Accordingly, the transform / quantization unit 7 sets flag information indicating whether or not to use a new quantization matrix for each orthogonal transform size for each color signal and coding mode, in a quantization matrix parameter to be encoded. . Furthermore, when a new quantization matrix is used, each scaling value of the quantization matrix as shown in FIG. 10 is set as a quantization matrix parameter to be encoded.
  • a quantization matrix prepared in advance by the moving picture coding apparatus and the moving picture decoding apparatus as an initial value or a quantization matrix that has already been coded.
  • an index for specifying a matrix to be used is set as a quantization matrix parameter to be encoded.
  • only a quantization matrix prepared in advance by the moving picture coding apparatus and the moving picture decoding apparatus can be selected.
  • the inverse quantization / inverse transform unit 8 refers to the quantization parameter and the orthogonal transform block division information included in the prediction difference coding parameter determined by the coding control unit 2, and transforms / quantizes the orthogonal transform block unit. 7, the quantized data output from the subtractor 6 is inversely quantized, and the inverse orthogonal transform process is performed on the transform coefficient that is the compressed data after the inverse quantization, and the local decoding corresponding to the prediction difference signal output from the subtractor 6 A process of calculating the prediction difference signal is performed.
  • the transform / quantization unit 7 uses the quantization matrix to perform the quantization process, the corresponding inverse quantization can be performed by referring to the quantization matrix even during the inverse quantization process. Perform the process.
  • the addition unit 9 includes the local decoded prediction difference signal calculated by the inverse quantization / inverse conversion unit 8, the intra prediction image generated by the intra prediction unit 4, or the inter prediction image generated by the motion compensation prediction unit 5. Are added to calculate a locally decoded image corresponding to the encoded block output from the block dividing unit 1.
  • the inverse quantization / inverse transform unit 8 and the addition unit 9 constitute a local decoded image generation unit.
  • the intra prediction memory 10 is a recording medium that stores the locally decoded image calculated by the adding unit 9.
  • the loop filter unit 11 performs a predetermined filtering process on the local decoded image calculated by the adding unit 9 and performs a process of outputting the local decoded image after the filter process. Specifically, filter (deblocking filter) processing that reduces distortion occurring at the boundaries of orthogonal transform blocks and prediction blocks, processing for adaptively adding an offset in units of pixels (pixel adaptive offset), Wiener filter, etc.
  • the adaptive filter processing that performs the filter processing by adaptively switching the linear filter is performed.
  • the loop filter unit 11 determines whether or not to perform each of the deblocking filter process, the pixel adaptive offset process, and the adaptive filter process, and performs variable-length coding using the valid flag of each process as header information. To the unit 13. When a plurality of the above filter processes are used, each filter process is performed in order.
  • FIG. 11 shows a configuration example of the loop filter unit 11 when a plurality of filter processes are used.
  • the more types of filter processing that are used the better the image quality, but the higher the processing load. That is, image quality and processing load are in a trade-off relationship.
  • the image quality improvement effect of each filter process varies depending on the characteristics of the image to be filtered. Therefore, the filter processing to be used may be determined according to the processing load allowed by the moving image encoding device and the characteristics of the encoding target image.
  • the loop filter unit 11 constitutes filtering means.
  • the deblocking filter process various parameters used for selecting the filter strength applied to the block boundary can be changed from the initial values.
  • the parameter is output to the variable length coding unit 13 as header information.
  • the pixel adaptive offset process first, an image is divided into a plurality of blocks, and when the offset process is not performed for each block, it is defined as one of the class classification methods, and a plurality of class classifications prepared in advance are used. One classification method is selected from the methods. Next, each pixel in the block is classified by the selected class classification method, and an offset value for compensating the coding distortion is calculated for each class. Finally, the image quality of the locally decoded image is improved by performing a process of adding the offset value to the luminance value of the locally decoded image. Therefore, in the pixel adaptive offset processing, block division information, an index indicating the class classification method of each block, and offset information for specifying an offset value of each class in block units are output to the variable length coding unit 13 as header information.
  • the filter designed for each class is output to the variable length encoding unit 13 as header information.
  • the class classification method there are a simple method for spatially dividing an image at equal intervals, and a method for classifying an image according to local characteristics (dispersion, etc.) of each block.
  • the number of classes used in the adaptive filter processing may be set in advance as a value common to the moving image encoding device and the moving image decoding device, or may be a parameter to be encoded. Compared to the former, the latter can set the number of classes to be used freely, so the image quality improvement effect will be improved, but on the other hand, the amount of code will be increased to encode the number of classes. To do.
  • the motion compensated prediction frame memory 12 is a recording medium that stores a locally decoded image after the filter processing of the loop filter unit 11.
  • the variable length coding unit 13 outputs the compressed data output from the transform / quantization unit 7 and the output signal of the coding control unit 2 (block division information in the largest coding block, coding mode, prediction difference coding parameter, Intra prediction parameters or inter prediction parameters) and motion vectors output from the motion compensation prediction unit 5 (when the encoding mode is the inter encoding mode) are variable length encoded to generate encoded data. Further, as illustrated in FIG. 13, the variable length encoding unit 13 encodes a sequence level header and a picture level header as header information of the encoded bit stream, and generates an encoded bit stream together with the picture data.
  • the variable length encoding unit 13 constitutes variable length encoding means.
  • picture data is composed of one or more slice data
  • each slice data is a combination of a slice level header and the encoded data in the slice.
  • the sequence level header includes the image size, the color signal format, the bit depth of the signal value of the luminance signal and the color difference signal, and each filter process (adaptive filter process, pixel adaptive offset process, deblocking filter process) in the loop filter unit 11 in sequence units.
  • Effective flag information Effective flag information, quantization matrix effective flag information, and a flag indicating whether or not to perform field encoding.
  • the picture level header is a collection of header information set in units of pictures, such as an index of a sequence level header to be referenced, the number of reference pictures at the time of motion compensation, an entropy coding probability table initialization flag, and a quantization matrix parameter. .
  • the slice level header includes position information indicating where the slice is located in the picture, an index indicating which picture level header is referred to, a slice coding type (all-intra coding, inter coding, etc.), and a loop.
  • This is a summary of parameters in units of slices such as flag information indicating whether or not to perform each filter process (adaptive filter process, pixel adaptive offset process, deblocking filter process) in the filter unit 11.
  • a block division unit 1 an encoding control unit 2, a changeover switch 3, an intra prediction unit 4, a motion compensation prediction unit 5, a subtraction unit 6, transform / quantization, which are components of the moving image encoding device.
  • FIG. 2 is a flowchart showing the processing contents (moving image coding method) of the moving image coding apparatus according to Embodiment 1 of the present invention.
  • FIG. 3 is a block diagram showing a moving picture decoding apparatus according to Embodiment 1 of the present invention.
  • the variable length decoding unit 31 receives the encoded bit stream generated by the moving image encoding apparatus of FIG. 1, each header information such as a sequence level header, a picture level header, a slice level header, and the like from the bit stream.
  • the block division information indicating the division status of each encoded block divided hierarchically is variable-length decoded from the bitstream.
  • the quantization matrix is specified from the quantization matrix parameter variable-length decoded by the variable-length decoding unit 31.
  • specify the quantization matrix by referring to the index information for specifying which quantization matrix among the matrices.
  • the quantization matrix parameter indicates that a new quantization matrix is used, it is specified as a quantization matrix that uses the quantization matrix included in the quantization matrix parameter.
  • variable length decoding unit 31 refers to each header information to specify the slice division state, and the maximum decoding block included in the slice data of each slice (the “maximum encoding” of the video encoding device in FIG. 1).
  • a block corresponding to a “block” is identified, and the decoding block (the “code” of the moving picture encoding device in FIG. 1) is a unit for performing decoding processing by hierarchically dividing the maximum decoded block with reference to block division information.
  • the variable length decoding unit 31 constitutes a variable length decoding unit.
  • the inverse quantization / inverse transform unit 32 refers to the quantization parameter and the orthogonal transform block division information included in the prediction difference encoding parameter variable length decoded by the variable length decoding unit 31, and the variable length decoding unit 31 performs variable length decoding.
  • the decoded compressed data is inversely quantized in units of orthogonal transform blocks, and inverse orthogonal transform processing is performed on transform coefficients that are compressed data after inverse quantization, so that the inverse quantization / inverse transform unit 8 in FIG. A process of calculating the same decoded prediction difference signal as the output local decoding prediction difference signal is performed.
  • the inverse quantization / inverse transform unit 32 constitutes a difference image generation unit.
  • each header information variable-length decoded by the variable-length decoding unit 31 indicates that the inverse quantization process is performed using the quantization matrix in the slice
  • the header information is inverted using the quantization matrix.
  • Quantization processing is performed. Specifically, inverse quantization processing is performed using a quantization matrix specified from each header information.
  • the changeover switch 33 outputs the intra-prediction parameter variable-length decoded by the variable-length decoding unit 31 to the intra-prediction unit 34 if the coding mode variable-length decoded by the variable-length decoding unit 31 is the intra-coding mode. If the encoding mode variable-length decoded by the variable-length decoding unit 31 is an inter-coding mode, a process of outputting the inter prediction parameters and motion vectors variable-length decoded by the variable-length decoding unit 31 to the motion compensation unit 35 carry out.
  • the intra prediction unit 34 When the encoding mode related to the decoded block specified from the block division information variable-length decoded by the variable-length decoding unit 31 is the intra-coding mode, the intra prediction unit 34 performs prediction when performing the prediction process of the decoded block For each prediction block that is a processing unit, an intra prediction process (intraframe prediction process) using the intra prediction parameter output from the changeover switch 33 is performed with reference to the decoded image stored in the intra prediction memory 37. Then, a process for generating an intra-predicted image is performed.
  • an intra prediction process intraframe prediction process
  • the motion compensation unit 35 performs prediction when performing the prediction process of the decoded block when the coding mode related to the decoded block specified from the block division information subjected to variable length decoding by the variable length decoding unit 31 is the inter coding mode. For each prediction block that is a processing unit, while referring to the decoded image stored in the motion compensated prediction frame memory 39, the inter prediction process (motion compensated prediction) using the motion vector output from the changeover switch 33 and the inter prediction parameter is used. Process) to generate an inter prediction image.
  • the intra prediction unit 34, the intra prediction memory 37, the motion compensation unit 35, and the motion compensated prediction frame memory 39 constitute a prediction unit.
  • the addition unit 36 adds the decoded prediction difference signal calculated by the inverse quantization / inverse conversion unit 32 and the intra prediction image generated by the intra prediction unit 34 or the inter prediction image generated by the motion compensation unit 35. Then, a process of calculating the same decoded image as the local decoded image output from the adding unit 9 in FIG. 1 is performed.
  • the adding unit 36 constitutes a decoded image generating unit.
  • the intra prediction memory 37 is a recording medium that stores the decoded image calculated by the adding unit 36 as a reference image used in the intra prediction process.
  • the loop filter unit 38 performs a predetermined filter process on the decoded image calculated by the adding unit 36 and performs a process of outputting the decoded image after the filter process. Specifically, filter (deblocking filter) processing that reduces distortion occurring at the boundaries of orthogonal transform blocks and prediction blocks, processing for adaptively adding an offset in units of pixels (pixel adaptive offset), Wiener filter, etc.
  • the adaptive filter processing that performs the filter processing by adaptively switching the linear filter is performed.
  • the loop filter unit 38 performs each of the above deblocking filter processing, pixel adaptive offset processing, and adaptive filter processing with reference to each header information variable-length decoded by the variable-length decoding unit 31 in the corresponding slice. Specify whether or not. At this time, when two or more filter processes are performed, if the loop filter unit 11 of the moving picture coding apparatus is configured as shown in FIG. 11, a loop filter unit 38 is configured as shown in FIG. .
  • the loop filter unit 38 constitutes filtering means.
  • deblocking filter processing with reference to the header information that has been subjected to variable length decoding by the variable length decoding unit 31, there is information for changing various parameters used for selecting the filter strength applied to the block boundary from the initial value. Based on the change information, deblocking filter processing is performed. When there is no change information, it is performed according to a predetermined method.
  • the decoded image is divided based on the block division information of the pixel adaptive offset processing variable-length decoded by the variable-length decoding unit 31, and the variable-length decoding unit 31 performs variable-length decoding on the block basis. If the index indicating the block classification method is not an index indicating that “offset processing is not performed”, each pixel in the block is classified according to the class classification method indicated by the index. To do.
  • the same class classification method candidate as the pixel classification method candidate of the pixel adaptive offset process of the loop filter unit 11 is prepared in advance as a class classification method candidate. Then, a process of adding the offset to the luminance value of the decoded image is performed with reference to the offset information specifying the offset value of each class in block units.
  • the block division information is not encoded, and the image is always divided into fixed-size block units (for example, maximum encoded block units),
  • the loop filter unit 38 also applies pixel adaptation to a block unit of the same fixed size as the loop filter unit 11. Perform offset processing.
  • the motion compensation prediction frame memory 39 is a recording medium that stores the decoded image after the filter processing of the loop filter unit 38 as a reference image used in the inter prediction processing (motion compensation prediction processing).
  • variable length decoding unit 31 the inverse quantization / inverse conversion unit 32, the changeover switch 33, the intra prediction unit 34, the motion compensation unit 35, the addition unit 36, and the intra prediction, which are components of the video decoding device.
  • each of the memory 37, the loop filter unit 38, and the motion compensation prediction frame memory 39 is configured by dedicated hardware (for example, a semiconductor integrated circuit on which a CPU is mounted, a one-chip microcomputer, or the like).
  • FIG. 4 is a flowchart showing the processing contents (moving image decoding method) of the moving image decoding apparatus according to Embodiment 1 of the present invention.
  • each frame image of a video is used as an input image, intra prediction from encoded neighboring pixels or motion compensation prediction between adjacent frames is performed, and an obtained prediction difference signal is obtained.
  • a video encoding device that performs compression processing by orthogonal transform / quantization and then performs variable length encoding to generate an encoded bit stream, and decodes the encoded bit stream output from the video encoding device
  • a moving picture decoding apparatus will be described.
  • the moving picture coding apparatus in FIG. 1 performs intra-frame / inter-frame adaptive coding by dividing a video signal into blocks of various sizes in response to local changes in the spatial and temporal directions of the video signal. It is characterized by that.
  • a video signal has a characteristic that the complexity of the signal changes locally in space and time.
  • a small image such as a picture with a uniform signal characteristic in a relatively wide image area such as the sky or a wall, or a picture containing a person or fine texture, on a video frame.
  • a pattern having a complicated texture pattern in the region may be mixed.
  • Even when viewed temporally, the change in the pattern of the sky and the wall locally in the time direction is small, but because the outline of the moving person or object moves rigidly or non-rigidly in time, the temporal change Is big.
  • a prediction difference signal with small signal power and entropy is generated by temporal and spatial prediction to reduce the overall code amount.
  • the parameters used for the prediction are set as large as possible in the image signal region. If it can be applied uniformly, the code amount of the parameter can be reduced.
  • the same prediction parameter is applied to a large image region with respect to an image signal pattern having a large temporal and spatial change, the number of prediction differential signals increases because prediction errors increase. . Therefore, in a region where the temporal and spatial changes are large, the block size for performing the prediction process by applying the same prediction parameter is reduced, the amount of parameter data used for prediction is increased, and the power and entropy of the prediction difference signal are increased. It is desirable to reduce
  • the first embodiment in order to perform coding adapted to the general characteristics of such a video signal, first, prediction processing or the like is started from a predetermined maximum block size, and the video signal region is divided hierarchically. In addition, the prediction process and the encoding process of the prediction difference are adapted for each divided area.
  • the video signal format to be processed by the moving image encoding apparatus of FIG. 1 is a color video signal in an arbitrary color space such as a YUV signal composed of a luminance signal and two color difference signals, or an RGB signal output from a digital image sensor.
  • the video frame is an arbitrary video signal including a horizontal / vertical two-dimensional digital sample (pixel) sequence, such as a monochrome image signal or an infrared image signal.
  • the gradation of each pixel may be 8 bits, or a gradation of 10 bits or 12 bits.
  • the video signal of the input image is a YUV signal
  • the two color difference components U and V are subsampled with respect to the luminance component Y 4: 2: 0.
  • the case of handling format signals will be described.
  • the format of the color difference signal may be other than the 4: 2: 0 format of the YUV signal, or may be the 4: 2: 2 format, 4: 4: 4 format of the YUV signal, or the RGB signal.
  • a processing data unit corresponding to each frame of the video signal is referred to as a “picture”. Note that “picture” represents a frame signal when encoded in frame units, and represents a field signal when encoded in field units.
  • the encoding control unit 2 determines the slice division state of a picture to be encoded (current picture), and also determines the size of the maximum encoding block used for encoding the picture and the hierarchy for dividing the maximum encoding block into layers. The upper limit of the number is determined (step ST1 in FIG. 2).
  • the same size may be determined for all the pictures according to the resolution of the video signal of the input image, or the local motion of the video signal of the input image
  • the size difference may be quantified as a parameter, and a small size may be determined for a picture with high motion, while a large size may be determined for a picture with little motion.
  • the upper limit of the number of division layers can be determined by, for example, determining the same number of layers for all pictures according to the resolution of the video signal of the input image, or when the motion of the video signal of the input image is severe There is a method in which the number of hierarchies is increased so that finer movements can be detected, and when there are few movements, the number of hierarchies is set to be suppressed.
  • the size of the maximum coding block and the upper limit of the number of layers into which the maximum coding block is divided may be encoded in a sequence level header or the like. Processing may be performed.
  • the former increases the code amount of the header information, it is not necessary to perform the determination process on the video decoding device side, so the processing load on the video decoding device can be suppressed and the optimum on the video encoding device side. You can search for and send a new value. On the contrary, since the determination process is performed on the video decoding device side, the processing load on the video decoding device increases, but the code amount of the header information does not increase.
  • the encoding control unit 2 selects an encoding mode corresponding to each encoding block divided hierarchically from one or more available encoding modes (step ST2). That is, the encoding control unit 2 divides the image area of the maximum encoding block size into encoded blocks having the encoding block size hierarchically until reaching the upper limit of the number of division layers defined above. A coding mode for each coding block is determined. There are one or more intra coding modes (collectively referred to as “INTRA”) and one or more inter coding modes (collectively referred to as “INTER”). The coding control unit 2 selects a coding mode corresponding to each coding block from all coding modes available for the picture or a subset thereof.
  • ITRA intra coding modes
  • INTER inter coding modes
  • each coding block that is hierarchically divided by the block division unit 1 to be described later is further divided into one or a plurality of prediction blocks, which are units for performing prediction processing, and the division state of the prediction block is also coded mode.
  • the coding mode is an index for identifying what kind of prediction block division the intra or inter coding mode is. Since the encoding mode selection method by the encoding control unit 2 is a known technique, detailed description thereof is omitted. For example, an encoding process for an encoding block is performed using any available encoding mode. There is a method in which coding efficiency is verified by performing and a coding mode having the best coding efficiency is selected from among a plurality of available coding modes.
  • the encoding control unit 2 determines a quantization parameter and an orthogonal transform block division state that are used when the difference image is compressed for each encoding block, and is used when the prediction process is performed.
  • Prediction parameter (intra prediction parameter or inter prediction parameter) is determined.
  • a prediction parameter (intra prediction parameter or inter prediction parameter) can be selected for each prediction block.
  • the selectable transform block size is limited to the size of the prediction block or less.
  • the encoding control unit 2 outputs the prediction difference encoding parameter including the quantization parameter and the transform block size to the transform / quantization unit 7, the inverse quantization / inverse transform unit 8, and the variable length coding unit 13. Also, the encoding control unit 2 outputs intra prediction parameters to the intra prediction unit 4 as necessary. Also, the encoding control unit 2 outputs inter prediction parameters to the motion compensation prediction unit 5 as necessary.
  • the slice dividing unit 14 divides the input image into slices that are one or more partial images according to the slice division information determined by the encoding control unit 2. Each time the slice dividing unit 1 inputs each slice from the slice dividing unit 14, the slice dividing unit 1 divides the slice into the maximum coding block size determined by the coding control unit 2, and further encodes the divided maximum coding block.
  • the coding block is hierarchically divided into coding blocks determined by the coding control unit 2, and the coding blocks are output.
  • FIG. 5 is an explanatory diagram showing an example in which the maximum coding block is hierarchically divided into a plurality of coding blocks.
  • the maximum coding block is a coding block whose luminance component described as “0th layer” has a size of (L 0 , M 0 ).
  • the encoding block is obtained by performing hierarchical division to a predetermined depth separately defined by a quadtree structure.
  • the coding block is an image area of size (L n , M n ).
  • the coding block of the n hierarchy expressed in B n denote the encoding modes selectable by the coding block B n with m (B n).
  • the encoding mode m (B n ) may be configured to use an individual mode for each color component, or common to all color components. It may be configured to use a mode.
  • description will be made assuming that it indicates a coding mode for a luminance component of a coding block of a YUV signal and 4: 2: 0 format.
  • the encoded block B n is divided by the block dividing unit 1 into one or a plurality of prediction blocks representing a prediction processing unit.
  • a prediction block belonging to the coding block B n is denoted as P i n (i is a prediction block number in the n-th layer).
  • FIG. 5 shows an example of P 0 0 and P 1 0 .
  • How the prediction block is divided in the coding block Bn is included as information in the coding mode m ( Bn ). All the prediction blocks P i n are subjected to prediction processing according to the encoding mode m (B n ), and it is possible to select individual prediction parameters (intra prediction parameters or inter prediction parameters) for each prediction block P i n. it can.
  • the encoding control unit 2 generates a block division state as illustrated in FIG. 6 for the maximum encoding block, and identifies the encoding block.
  • a rectangle surrounded by a dotted line in FIG. 6A represents each coding block, and a block painted with diagonal lines in each coding block represents a division state of each prediction block.
  • FIG. 6B shows, in a quadtree graph, a situation in which the encoding mode m (B n ) is assigned by hierarchical division in the example of FIG. 6A. Nodes surrounded by squares in FIG. 6B are nodes (encoding blocks) to which the encoding mode m (B n ) is assigned.
  • Information of the quadtree graph is output from the encoding control unit 2 to the variable length encoding unit 13 together with the encoding mode m (B n ), and is multiplexed into the bit stream.
  • the changeover switch 3 is output from the block dividing unit 1 when the encoding mode m (B n ) determined by the encoding control unit 2 is an intra encoding mode (when m (B n ) ⁇ INTRA).
  • the encoded block B n is output to the intra prediction unit 4.
  • the encoding mode m (B n ) determined by the encoding control unit 2 is the inter encoding mode (when m (B n ) ⁇ INTER)
  • the encoded block output from the block dividing unit 1 B n is output to the motion compensation prediction unit 5.
  • the coding mode m (B n ) determined by the coding control unit 2 is the intra coding mode (when m (B n ) ⁇ INTRA), and the coding block B is changed from the changeover switch 3 to the coding block B.
  • n step ST3, using the intra prediction parameters determined by the encoding control unit 2 while referring to the local decoded image stored in the intra prediction memory 10, the encoding block B n and implementing intra prediction process for each of the prediction block P i n in, it generates an intra prediction image P INTRAi n (step ST4).
  • intra prediction parameters used for generating the intra prediction image P INTRAi n is from encoding control unit 2
  • the data is output to the variable length encoding unit 13 and multiplexed into the bit stream. Details of processing contents of the intra prediction unit 4 will be described later.
  • the motion-compensated prediction unit 5 has the coding mode m (B n ) determined by the coding control unit 2 in the inter coding mode (when m (B n ) ⁇ INTER), and the coding block is switched from the changeover switch 3 to the coding block.
  • the motion vector Upon receiving the B n (step ST3), the motion vector by comparing the locally decoded image after the filtering process stored in the prediction block P i n and the motion compensated prediction frame memory 12 of the encoding block B n Using the motion vector and the inter prediction parameter determined by the encoding control unit 2, the inter prediction process for each prediction block P i n in the encoding block B n is performed, and the inter prediction image P generating a INTERi n (step ST5).
  • inter prediction parameters used for generating the inter prediction image P INTERi n is from encoding control unit 2
  • the data is output to the variable length encoding unit 13 and multiplexed into the bit stream.
  • the motion vector searched by the motion compensation prediction unit 5 is also output to the variable length encoding unit 13 and multiplexed into the bit stream.
  • Subtraction unit 6 upon receiving the encoded block B n from the block dividing unit 1 from its prediction block P i n the coded block B n, the intra prediction image P INTRAi n generated by the intra prediction unit 4 or, , by subtracting one of the inter prediction image P INTERi n generated by the motion compensation prediction unit 5, and outputs the prediction difference signal e i n representing a difference image is the subtraction result to the transform and quantization unit 7 (Step ST6).
  • the transform / quantization unit 7 When the transform / quantization unit 7 receives the prediction difference signal e i n from the subtraction unit 6, the transform / quantization unit 7 refers to the orthogonal transform block division information included in the prediction difference encoding parameter determined by the encoding control unit 2, and orthogonal transform processing for the predicted differential signal e i n (e.g., DCT (discrete cosine transform) or DST (discrete sine transform), the orthogonal transform for KL conversion and the base design have been made in advance to the particular learning sequence) This is performed for each orthogonal transform block, and transform coefficients are calculated.
  • DCT discrete cosine transform
  • DST discrete sine transform
  • the transform / quantization unit 7 refers to the quantization parameter included in the prediction differential encoding parameter, quantizes the transform coefficient of the orthogonal transform block unit, and compresses the compressed data that is the quantized transform coefficient.
  • the data is output to the inverse quantization / inverse transform unit 8 and the variable length coding unit 13 (step ST7).
  • the quantization process may be performed using a quantization matrix that scales the quantization step size calculated from the quantization parameter for each transform coefficient.
  • the quantization matrix an independent matrix can be used for each color signal and coding mode (intra coding or inter coding) at each orthogonal transform size. It is possible to select either a quantization matrix that is commonly prepared in the moving picture decoding apparatus or an already encoded quantization matrix or a new quantization matrix. Accordingly, the transform / quantization unit 7 sets flag information indicating whether or not to use a new quantization matrix for each orthogonal transform size for each color signal and coding mode, in a quantization matrix parameter to be encoded. . Furthermore, when a new quantization matrix is used, each scaling value of the quantization matrix as shown in FIG. 10 is set as a quantization matrix parameter to be encoded.
  • a quantization matrix prepared in advance by the moving picture coding apparatus and the moving picture decoding apparatus as an initial value or a quantization matrix that has already been coded.
  • an index for specifying a matrix to be used is set as a quantization matrix parameter to be encoded.
  • the transform / quantization unit 7 outputs the set quantization matrix parameter to the variable length coding unit 13.
  • the inverse quantization / inverse transform unit 8 When the inverse quantization / inverse transform unit 8 receives the compressed data from the transform / quantization unit 7, the inverse quantization / inverse transform unit 8 displays the quantization parameter and the orthogonal transform block division information included in the prediction difference coding parameter determined by the coding control unit 2. With reference, the compressed data is inversely quantized in units of orthogonal transform blocks.
  • the transform / quantization unit 7 uses a quantization matrix for the quantization process, the corresponding inverse quantization process is performed with reference to the quantization matrix even during the inverse quantization process.
  • the inverse quantization / inverse transform unit 8 performs inverse orthogonal transform processing (for example, inverse DCT, inverse DST, inverse KL transform, etc.) on transform coefficients that are compressed data after inverse quantization in units of orthogonal transform blocks. Then, a local decoded prediction difference signal corresponding to the prediction difference signal e i n output from the subtraction unit 6 is calculated and output to the addition unit 9 (step ST8).
  • inverse orthogonal transform processing for example, inverse DCT, inverse DST, inverse KL transform, etc.
  • the adding unit 9 Upon receiving the local decoded prediction difference signal from the inverse quantization / inverse transform unit 8, the adding unit 9 receives the local decoded prediction difference signal and the intra predicted image P INTRAi n generated by the intra prediction unit 4 or motion compensation.
  • a local decoded image is calculated by adding one of the inter predicted images P INTERIn n generated by the prediction unit 5 (step ST9).
  • the adding unit 9 outputs the locally decoded image to the loop filter unit 11 and stores the locally decoded image in the intra prediction memory 10. This locally decoded image becomes an encoded image signal used in the subsequent intra prediction processing.
  • Step ST10 When the loop filter unit 11 receives the local decoded image from the adding unit 9, the loop filter unit 11 performs a predetermined filter process on the local decoded image, and stores the filtered local decoded image in the motion compensated prediction frame memory 12.
  • filter deblocking filter
  • the adaptive filter processing that performs the filter processing by adaptively switching the linear filter is performed.
  • the loop filter unit 11 determines whether or not to perform processing for each of the above deblocking filter processing, pixel adaptive offset processing, and adaptive filter processing, and sets a valid flag of each processing as a part of the sequence level header and The data is output to the variable length coding unit 13 as a part of the slice level header.
  • each filter process is performed in order.
  • FIG. 11 shows a configuration example of the loop filter unit 11 when a plurality of filter processes are used.
  • the more types of filter processing that are used the better the image quality, but the higher the processing load. That is, image quality and processing load are in a trade-off relationship.
  • the image quality improvement effect of each filter process varies depending on the characteristics of the image to be filtered. Therefore, the filter processing to be used may be determined according to the processing load allowed by the moving image encoding device and the characteristics of the encoding target image.
  • the parameter is output to the variable length coding unit 13 as header information.
  • an image is divided into a plurality of blocks, and when the offset process is not performed for each block, it is defined as one of the class classification methods, and a plurality of class classifications prepared in advance are provided.
  • One classification method is selected from the methods.
  • each pixel in the block is classified by the selected class classification method, and an offset value for compensating the coding distortion is calculated for each class.
  • the image quality of the locally decoded image is improved by performing a process of adding the offset value to the luminance value of the locally decoded image.
  • a classifying method a method of classifying by the size of the luminance value of the locally decoded image (referred to as a BO method), or a classification according to the situation around each pixel (whether it is an edge portion or the like) for each edge direction.
  • EO technique a technique of classifying by the size of the luminance value of the locally decoded image
  • EO technique a technique of classifying by the size of the luminance value of the locally decode
  • the pixel adaptive offset processing outputs the block division information, the index indicating the block class classification method, and the block offset information to the variable length coding unit 13 as header information.
  • a filter that compensates for superimposed distortion is designed for each region (local decoded image) belonging to each class. Then, the local decoded image is filtered. Then, the filter designed for each class is output to the variable length encoding unit 13 as header information.
  • a class classification method there are a simple method for spatially dividing an image at equal intervals, and a method for classifying an image according to local characteristics (dispersion, etc.) of the image in units of blocks.
  • the number of classes used in the adaptive filter process may be set in advance to a common value in the video encoding device and the video decoding device, or may be one of the parameters to be encoded. Compared to the former, the latter can set the number of classes to be used freely, so the image quality improvement effect will be improved, but on the other hand, the amount of code will be increased to encode the number of classes. To do.
  • steps ST3 to ST9 are repeated until the processes for all the coding blocks Bn divided hierarchically are completed, and when the processes for all the coding blocks Bn are completed, the process proceeds to the process of step ST13. (Steps ST11 and ST12).
  • the variable length encoding unit 13 uses the compressed data output from the transform / quantization unit 7 and the block division information (FIG. 6B) in the maximum encoding block output from the encoding control unit 2 as an example.
  • the parameters (when the encoding mode is the inter encoding mode) and the motion vector (when the encoding mode is the inter encoding mode) output from the motion compensated prediction unit 5 are variable-length encoded, Encoded data indicating the encoding result is generated (step ST13).
  • the orthogonal transform block is further divided into blocks of 4 ⁇ 4 pixel units (encoding sub-blocks) called “Coefficient Group (CG)”.
  • the coefficient encoding process is performed on FIG. 28 shows the coding order (scan order) of the coefficients in the 16 ⁇ 16 pixel orthogonal transform block.
  • 16 CGs in units of 4 ⁇ 4 pixels are encoded in this order from the lower right CG, and each CG encodes 16 coefficients in the CG in order from the lower right coefficient.
  • flag information indicating whether or not a significant (nonzero) coefficient exists in 16 coefficients in the CG is encoded, and then a significant (nonzero) coefficient exists in the CG. Only whether each coefficient in the CG is a significant (non-zero) coefficient is encoded in the above order, and finally, coefficient value information is encoded in order for a significant (non-zero) coefficient. This is performed in the above order in units of CG. In this case, the encoding efficiency by entropy encoding can be increased by using a biased scan order so that significant (non-zero) coefficients are generated as continuously as possible.
  • the coefficient after the orthogonal transformation represents the lower coefficient of the lower frequency component as it approaches the upper left, starting with the DC component located at the upper left, generally, in the progressive video as shown in FIG. Since many non-zero) coefficients are generated, efficient encoding can be performed by encoding sequentially from the lower right as shown in FIG.
  • the vertical spatial efficiency decreases, so the vertical prediction efficiency is
  • the frequency coefficient in the vertical direction also increases for the transform coefficient, which is a result of the orthogonal transformation of the prediction difference signal e i n , and the significant (non-zero) coefficient generation distribution is progressive as shown in FIG.
  • the shape is biased to the left side of the orthogonal transform block rather than the video. Accordingly, since the encoding cannot be efficiently performed in the encoding order shown in FIG. 28, for example, switching is performed in the encoding order shown in FIG.
  • the 16 ⁇ 16 pixel orthogonal transform block has been described.
  • encoding processing in units of CG is also performed in block sizes other than 16 ⁇ 16 pixels, such as a 32 ⁇ 32 pixel orthogonal transform block.
  • the encoding order is switched according to whether the flag indicating whether or not the field encoding of the sequence level header is valid, as in the case of the 16 ⁇ 16 pixel orthogonal transform block.
  • the coding order shown in FIG. 17 (coding block unit (coding order in a coding block of 16 ⁇ 16 pixels), code
  • the sub-block unit (coding order within the CG of 4 ⁇ 4 pixels) is changed, but the shape of the CG is changed from the 4 ⁇ 4 pixel block to the 8 ⁇ 2 pixel block as shown in FIG. It may be. Even in this way, encoding of significant (non-zero) coefficients is continuously processed by the CG in the rear in the encoding order, and the encoding efficiency by entropy encoding can be improved.
  • the coding order is as shown in FIG. 28, and in the case of FIG. 17, the coding block unit and the coding subblock unit are used. Since the coding order is switched, the coding efficiency can be improved. Further, in the case of FIG. 18, in addition to switching the coding order in units of coding blocks and coding sub-blocks, the shape of the coding sub-block is also changed. Since the change is made, the encoding efficiency can be further increased.
  • the case where the coding order is switched in both the coding block unit and the coding subblock unit has been described. However, only the coding block unit or the coding subblock unit may be switched. .
  • the encoding order shown in FIG. 19 may be used.
  • the encoding of more significant (non-zero) coefficients can be performed in the encoding order. Processing can be continuously performed in the rear direction, and the encoding efficiency by entropy encoding can be further increased.
  • a flag indicating whether or not the sequence level header field coding is prepared in the picture level header, and the coding order of the coefficients when coding the compressed data that is the quantized orthogonal transform coefficients is adapted to each picture. May be switched automatically. By doing so, control adaptive to each picture can be realized, and the encoding efficiency can be improved. Note that in the case of realizing coding that adaptively switches between frame coding and field coding on a picture-by-picture basis, the flag needs to be prepared in the picture level header.
  • the coding order, the shape, and the like are switched based on the flag indicating whether or not the field coding of the sequence level header or the picture level header has been described, but the sequence level header or the picture level header
  • a flag indicating whether or not to perform the switching process is defined, and based on the flag indicating whether or not the switching process is performed, the encoding order, the CG shape, The scanning order in the CG may be switched.
  • FIGS. 17, 18, and 19 are illustrated as examples of the coding order, the shape of the CG, and the scan order in the CG.
  • the processing order is not limited to this, and the encoding order other than those shown in FIGS. 17, 18, and 19, the CG shape, and the scan order within the CG may be used.
  • the combination of the CG shape and the scan order within the CG it is not limited to FIG. 17, FIG. 18, and FIG.
  • the CG may be 1 ⁇ 2, 1 ⁇ 4, 1 ⁇ 8, 1 ⁇ 16, 2 ⁇ 2, 2 ⁇ 4, 4 ⁇ 8 pixels, or the like.
  • the case of field encoding has been described with respect to the case of any one of FIGS.
  • FIG. 17, 18, and 19 cannot be selected, but a plurality of candidates (FIG. 17, FIG. 18, One of them may be selected from FIG. In that case, a flag indicating which of a plurality of candidates has been selected is prepared in the header. This flag may be shared with a flag indicating whether or not to perform field encoding or a flag indicating whether or not to perform this switching process.
  • variable length encoding unit 13 encodes a sequence level header and a picture level header as header information of the encoded bit stream, and generates an encoded bit stream together with the picture data.
  • picture data is composed of one or more slice data, and each slice data is a combination of a slice level header and the encoded data in the slice.
  • the sequence level header includes the image size, the color signal format, the bit depth of the signal value of the luminance signal and the color difference signal, and each filter process (adaptive filter process, pixel adaptive offset process, deblocking filter process) in the loop filter unit 11 in sequence units. ) Effective flag information, quantization matrix effective flag information, and a flag indicating whether or not to perform field encoding.
  • the picture level header is a collection of header information set in units of pictures such as an index of a sequence level header to be referenced, the number of reference pictures at the time of motion compensation, an entropy encoding probability table initialization flag, and the like.
  • the slice level header includes position information indicating where the slice is located in the picture, an index indicating which picture level header is referred to, a slice coding type (all-intra coding, inter coding, etc.), and a loop.
  • This is a summary of parameters in units of slices such as flag information indicating whether or not to perform each filter process (adaptive filter process, pixel adaptive offset process, deblocking filter process) in the filter unit 11.
  • FIG. 7 is an explanatory diagram showing an example of the intra prediction mode is an intra prediction parameters each prediction block P i n is selectable within a coding block B n.
  • N I represents the number of intra prediction modes.
  • FIG. 7 shows an index value of an intra prediction mode and a prediction direction vector indicated by the intra prediction mode. In the example of FIG. 7, as the number of selectable intra prediction modes increases, Designed to reduce the angle.
  • Intra prediction unit 4 as described above, with reference to the intra prediction parameters of the prediction block P i n, to implement intra prediction processing for the prediction block P i n, but to generate an intra prediction image P INTRAi n , it will be described here intra process of generating an intra prediction signal of a prediction block P i n in the luminance signal.
  • the size of the prediction block P i n is set to l i n ⁇ m i n pixels.
  • the encoded pixels (2 ⁇ l i n +1) and the left encoded pixels (2 ⁇ m i n ) on the prediction block P i n are used as prediction pixels.
  • the number of pixels used for prediction may be more or less than the pixels shown in FIG.
  • it is used to predict one line or pixel of one column in the vicinity of the predicted block P i n, 2 rows or two columns, or may be used more pixels in the prediction.
  • the encoded pixel adjacent on the prediction block P i n is adjacent to the left of the prediction block P i n. using encoded pixels that generates a predicted image interpolated value according to the distance of the pixel and the prediction pixel in the prediction block P i n as the predicted value.
  • the encoded pixels adjacent on the prediction block P i n and the prediction block P i n are placed to the left of the prediction block P i n .
  • the prediction block P i n in the upper end and the area A of Figure 20 positioned at the left end, B, and final prediction image by performing a filtering process for smoothing the block boundary relative to the C. For example, filter processing is performed using the following filter coefficients in the reference pixel arrangement of the filter of FIG.
  • the region A may be subjected to the same filtering process as the region C. .
  • the region A may not be filtered, and only the region C may be filtered.
  • For prediction block P i index value of the intra prediction mode for n is 26 (vertical prediction), and generates a prediction image by calculating the predicted values of the pixels in the prediction block P i n from the following equation (1).
  • the coordinates (x, y) is the relative coordinates with the origin at the upper left pixel in the prediction block P i n (see Figure 9)
  • S '(x, y) is the predicted value at the coordinates (x, y)
  • S (x, y) is the luminance value (decoded luminance value) of the encoded pixel at the coordinates (x, y).
  • the calculated predicted value exceeds the range of values that the luminance value can take, the value is rounded so that the predicted value falls within the range.
  • Equation (1) shows the filter processing in the vertical direction prediction of FIG. 27.
  • the equation in the first row of Equation (1) is MPEG-4 AVC / H.
  • the block boundary is obtained by adding a value obtained by halving the amount of change in the luminance value of the adjacent encoded pixel in the vertical direction to S (x, ⁇ 1) that is the predicted value of the vertical direction prediction in H.264.
  • S (x, ⁇ 1) that is the predicted value of the vertical direction prediction in H.264.
  • This means that filtering is performed so as to be smoothed
  • the expression in the second row of Expression (1) is MPEG-4 AVC / H.
  • the same prediction formula as the vertical direction prediction in H.264 is shown.
  • Equation (2) shows the filter processing in the horizontal prediction of FIG. 27.
  • the equation in the first row of Equation (2) is MPEG-4 AVC / H.
  • the block boundary is obtained by adding a value obtained by halving the amount of change in the horizontal luminance value of the adjacent encoded pixel to S ( ⁇ 1, y), which is the predicted value of the horizontal direction prediction in H.264.
  • S ( ⁇ 1, y) which is the predicted value of the horizontal direction prediction in H.264.
  • This means that filtering is performed so as to be smoothed
  • the expression in the second line of Expression (2) is MPEG-4 AVC / H.
  • the same prediction formula as the horizontal prediction in H.264 is shown.
  • the equation (3) is used instead of the equation (2) in the horizontal prediction. That is, as shown in FIG. 22, the filter processing is not performed on the upper end of the prediction block (in the case of average value prediction and vertical direction prediction, the filter processing is performed only on the left end of the prediction block, and in the case of horizontal direction prediction, the filter processing is performed. Do not process).
  • the filter processing is not performed on the upper end of the prediction block (in the case of average value prediction and vertical direction prediction, the filter processing is performed only on the left end of the prediction block, and in the case of horizontal direction prediction, the filter processing is performed. Do not process).
  • the filter processing is not performed on the upper end of the prediction block (in the case of average value prediction and vertical direction prediction, the filter processing is performed only on the left end of the prediction block, and in the case of horizontal direction prediction, the filter processing is performed. Do not process).
  • the filter processing is not performed on the upper end of the prediction block (in the case of average value prediction and vertical direction prediction, the filter processing is performed only on the left end of the prediction block
  • a flag indicating whether or not the field coding of the sequence level header is prepared in the picture level header and a prediction block of average value (DC) prediction and horizontal prediction according to the correlation between pixels in the vertical direction of each picture
  • the upper end filter processing may be switched ON / OFF. By doing so, control adaptive to each picture can be realized, and prediction efficiency can be improved. Note that in the case of realizing encoding that adaptively switches between frame encoding and field encoding on a picture-by-picture basis, the flag needs to be prepared in the picture level header.
  • the filtering process at the upper end of the prediction block is switched on / off based on the flag indicating whether or not the field encoding of the sequence level header or the picture level header is performed.
  • a flag indicating whether or not to perform the main switching process is defined separately from the flag indicating whether or not the field coding of the picture level header is performed, and the upper end of the prediction block is determined based on the flag indicating whether or not to perform the switching process. You may make it switch ON / OFF of a filter process.
  • the flag indicating whether or not to perform the switching process is a flag based on the flag indicating whether or not the field encoding is performed.
  • the coding order switching described above and the filter processing switching described above are separately described. However, these may be set in combination.
  • the block size to be subjected to the filtering process may be limited, for example, the block boundary filtering process of average value (DC) prediction, vertical direction prediction, and horizontal direction prediction is limited to, for example, blocks of 16 ⁇ 16 pixels or less. By doing in this way, the amount of calculations required for filter processing can be reduced.
  • DC average value
  • the prediction direction vector ⁇ p ( dx, on the basis of the dy), and it generates the predicted values of the pixels in the prediction block P i n.
  • the origin at the upper left pixel of the prediction block P i n setting the relative coordinates of the prediction block P i n (x, y) and, the position of the reference pixels used for prediction, the following L And the intersection of adjacent pixels.
  • k is a negative scalar value.
  • the integer pixel When the reference pixel is at the integer pixel position, the integer pixel is set as the prediction value of the prediction target pixel. When the reference pixel is not at the integer pixel position, an interpolation pixel generated from the integer pixel adjacent to the reference pixel is selected. Estimated value. In the example of FIG. 8, since the reference pixel is not located at the integer pixel position, a value interpolated from two pixels adjacent to the reference pixel is set as the predicted value. Note that an interpolation pixel may be generated not only from two adjacent pixels but also from two or more adjacent pixels, and used as a predicted value.
  • the processing described above to generate a predicted pixel for all the pixels of the luminance signal in the prediction block P i n, and outputs an intra prediction image P INTRAi n.
  • the intra prediction parameters used for generating the intra prediction image P INTRAi n is output to the variable length coding unit 13 for multiplexing the bitstream.
  • the MPEG-4 AVC / H Similar to the smoothing process performed on the reference image at the time of 8 ⁇ 8 pixel block intra prediction of the 264, the intra prediction unit 4, the reference pixels in generating the intermediate prediction image predicted block P i n even when configured as a prediction block P i n pixels smoothed the encoded pixels adjacent to, it is possible to perform the filtering for the same intermediate predicted image and the above example.
  • the intra prediction processing based on the intra prediction parameters (intra prediction mode) performed the intra prediction parameters used for generating the intra prediction image Is output to the variable length encoding unit 13.
  • the intra prediction parameter (intra prediction mode) that can be selected by the color difference signal may be different from that of the luminance signal.
  • the MPEG-4 AVC / H it is good also as the prediction method similar to H.264.
  • the color difference signal (U, V signal) is a signal obtained by reducing the resolution to 1/2 in both the horizontal direction and the vertical direction with respect to the luminance signal (Y signal).
  • the complexity of the image signal is low and prediction is easy, so that the number of intra prediction parameters that can be selected is smaller than that of the luminance signal, and the amount of code required to encode the intra prediction parameters can be reduced. May be reduced.
  • variable length decoding unit 31 receives the encoded bitstream generated by the moving picture encoding apparatus in FIG. 1, the variable length decoding unit 31 performs variable length decoding processing on the bitstream (step ST21 in FIG. 4), and performs field encoding.
  • Header information (sequence level header) composed of one or more pictures, such as a flag indicating whether or not, frame size information, and header information (picture level header) in units of pictures, used in the loop filter unit 38
  • the filter parameters and quantization matrix parameters to be decoded are decoded.
  • the quantization matrix is specified with reference to the quantization matrix parameter variable-length decoded by the variable-length decoding unit 31.
  • the quantization matrix is specified, and the quantization matrix parameter indicates that a new quantization matrix is to be used, it is specified as a quantization matrix that uses the quantization matrix included in the quantization matrix parameter.
  • slice unit header information such as slice division information is decoded from slice data constituting picture unit data, and encoded data of each slice is decoded.
  • variable length decoding unit 31 determines the maximum coding block size and the upper limit of the number of divided layers determined by the coding control unit 2 of the moving image coding device in FIG. 1 in the same procedure as the moving image coding device. (Step ST22). For example, when the maximum encoding block size and the upper limit of the number of division layers are determined according to the resolution of the video signal, the maximum encoding is performed in the same procedure as the moving image encoding apparatus based on the decoded frame size information. Determine the block size. When the maximum encoding block size and the upper limit of the number of divided layers are multiplexed on a sequence level header or the like on the moving image encoding device side, values decoded from the header are used.
  • the maximum encoded block size is referred to as a maximum decoded block size
  • the maximum encoded block is referred to as a maximum decoded block.
  • the variable length decoding unit 31 decodes the division state of the maximum decoding block as shown in FIG. 6 for each determined maximum decoding block. Based on the decoded division state, a decoded block (a block corresponding to the “encoded block” of the moving image encoding apparatus in FIG. 1) is identified hierarchically (step ST23).
  • variable length decoding unit 31 decodes the encoding mode assigned to the decoding block. Based on the information included in the decoded coding mode, the decoded block is further divided into one or more prediction blocks which are prediction processing units, and the prediction parameters assigned to the prediction block units are decoded (step ST24).
  • the variable length decoding unit 31 is included in the decoding block and is intra for each of one or more prediction blocks serving as a prediction processing unit. Decode prediction parameters.
  • the coding mode assigned to the decoding block is the inter coding mode
  • the inter prediction parameter and the motion vector are decoded for each one or more prediction blocks included in the decoding block and serving as a prediction processing unit. (Step ST24).
  • variable length decoding unit 31 decodes the compressed data (transformed / transformed transform coefficients) for each orthogonal transform block based on the orthogonal transform block division information included in the prediction difference encoding parameter (step ST24).
  • the variable length decoding unit 31 decodes 16 CGs in units of 4 ⁇ 4 pixels in order from the lower right CG, and each CG decodes 16 coefficients in the CG in order from the lower right coefficient. It will be done.
  • flag information indicating whether or not a significant (non-zero) coefficient exists in 16 coefficients in the CG is decoded, and then the decoded flag information is significant (non-zero) in the CG. Only when it indicates that a coefficient exists, whether each coefficient in the CG is a significant (non-zero) coefficient is decoded in the order described above, and finally, coefficient value information is sequentially displayed for the coefficient indicating the significant (non-zero) coefficient.
  • Decrypt This is performed in the above order in units of CG.
  • the flag indicating whether or not the field encoding of the sequence level header decoded by the variable length decoding unit 31 is valid, the moving picture encoding apparatus of FIG. Decoding processing is performed in the same order as the processing order determined by the variable length encoding unit 13. In this way, it is possible to generate the same compressed data as the stream generated by the moving image encoding apparatus in FIG.
  • variable length decoding unit 31 similarly applies the decoding order of compressed data in units of pictures according to the flag. To switch automatically.
  • variable-length decoded by the variable-length decoding unit 31 is an intra-encoding mode (when m (B n ) ⁇ INTRA)
  • the changeover switch 33 is changed by the variable-length decoding unit 31.
  • the intra-prediction parameter for each prediction block subjected to variable length decoding is output to the intra-prediction unit 34.
  • variable length decoded coding mode m (B n) is if the inter coding mode by the variable length decoding unit 31, variable length decoding by the variable length decoding unit 31
  • the predicted inter prediction parameters and motion vectors in units of prediction blocks are output to the motion compensation unit 35.
  • the intra prediction unit 34 selects the changeover switch 33. 1 is received, and the intra prediction parameter is obtained by referring to the decoded image stored in the intra prediction memory 37 in the same procedure as the intra prediction unit 4 in FIG. and implementing intra prediction process to generate an intra prediction image P INTRAi n for each of the prediction block P i n of the decoded block B n using (step ST26).
  • the average value (DC) prediction and the horizontal direction are performed in the same manner as the moving picture coding apparatus of FIG.
  • the filtering process at the upper end of the prediction block of the prediction is not performed. In this way, it is possible to generate a predicted image that is the same as the stream generated by the moving image encoding device in FIG.
  • the moving picture encoding apparatus when a flag indicating whether or not the sequence level header field encoding is provided in the picture level header, whether or not the picture level header field encoding is performed. Filter processing at the upper end of the prediction block of average value (DC) prediction and horizontal direction prediction is switched ON / OFF in units of pictures according to the flag value shown. In this way, it is possible to generate the same predicted image as the stream generated by the moving picture coding apparatus according to Embodiment 1 configured as described above.
  • DC prediction block of average value
  • the motion compensation unit 35 performs the changeover switch 33.
  • the motion vector and the inter prediction parameter for each prediction block output from the above are received, and the motion vector and the inter prediction parameter are used while referring to the decoded image after filtering stored in the motion compensated prediction frame memory 39. by carrying out inter-prediction processing for each of the prediction block P i n of the decoded block B n to generate an inter prediction image P INTERi n (step ST27).
  • the inverse quantization / inverse conversion unit 32 When receiving the compressed data and the prediction difference encoding parameter from the variable length decoding unit 31, the inverse quantization / inverse conversion unit 32 performs the prediction difference encoding in the same procedure as the inverse quantization / inverse conversion unit 8 of FIG. With reference to the quantization parameter and orthogonal transform block division information included in the parameters, the compressed data is inversely quantized in units of orthogonal transform blocks. At this time, when referring to each header information variable-length decoded by the variable-length decoding unit 31, each header information indicates that the inverse quantization process is performed using the quantization matrix in the slice. Inverse quantization processing is performed using a quantization matrix.
  • variable-length decoding unit 31 a quantization matrix used for each color signal and coding mode (intra coding or inter coding) with each orthogonal transform size Is identified. Further, the inverse quantization / inverse transform unit 32 performs an inverse orthogonal transform process on transform coefficients that are compressed data after inverse quantization in units of orthogonal transform blocks, and the inverse quantization / inverse transform unit 8 in FIG. A decoded prediction difference signal identical to the output local decoded prediction difference signal is calculated (step ST28).
  • Addition unit 36 decodes the prediction difference signal calculated by the inverse quantization and inverse transform unit 32, an intra prediction image P INTRAi n generated by the intra prediction unit 34 or, inter prediction generated by the motion compensation unit 35 by adding one of the image P INTERi n calculates a decoded image, and outputs the decoded image to the loop filter unit 38, and stores the decoded image to the intra prediction memory 37 (step ST29).
  • This decoded image becomes a decoded image signal used in the subsequent intra prediction processing.
  • the loop filter unit 38 performs a predetermined filtering process on the decoded image output from the adding unit 36, and filters
  • the decoded image after processing is stored in the motion compensated prediction frame memory 39 (step ST31).
  • filter deblocking filter
  • processing for adaptively adding an offset in units of pixels pixel adaptive offset
  • Wiener filter etc.
  • the adaptive filter processing that performs the filter processing by adaptively switching the linear filter is performed.
  • the loop filter unit 38 processes each of the above-described deblocking filter processing, pixel adaptive offset processing, and adaptive filter processing with reference to each header information variable-length decoded by the variable-length decoding unit 31 and processing in the corresponding slice. Specify whether or not to perform. At this time, when two or more filter processes are performed, if the loop filter unit 11 of the moving picture coding apparatus is configured as shown in FIG. 11, the loop filter unit 38 is configured as shown in FIG. Is done.
  • the deblocking filter processing when there is information for referring to the header information that has been variable-length decoded by the variable-length decoding unit 31 and changing various parameters used for selecting the filter strength applied to the block boundary from the initial value. Performs a deblocking filter process based on the change information. When there is no change information, it is performed according to a predetermined method.
  • each pixel in the block is classified into blocks in accordance with the class classification method indicated by the index.
  • the same class classification method candidate as the pixel classification method candidate of the pixel adaptive offset process of the loop filter unit 11 is prepared in advance as a class classification method candidate.
  • the loop filter unit 38 refers to the offset information that has been variable-length decoded by the variable-length decoding unit 31 that identifies the offset value of each class in block units, and performs processing to add the offset to the luminance value of the decoded image .
  • the filter is based on the class classification information.
  • the decoded image after the filter processing by the loop filter unit 38 becomes a reference image for motion compensation prediction and also becomes a reproduced image.
  • the intra prediction unit 4 when the flag indicating that the input video signal is encoded in the field unit is valid, the intra prediction unit 4 performs the average value prediction or the horizontal value prediction.
  • the intra prediction unit 34 A configuration in which the filtering process at the upper end of the prediction block when performing intra prediction processing by value prediction or horizontal prediction is not performed, and a configuration in which the inverse quantization / inverse transform unit 32 changes the decoding order of transform coefficients Are implemented separately or in combination, so that efficient prediction processing and coding processing according to the characteristics of the field signal can be realized, and coding efficiency can be improved. There is an effect that the bitstream encoded by the moving image encoding apparatus of Embodiment 1 can be correctly decoded.
  • the moving image encoding device, the moving image decoding device, the moving image encoding method, and the moving image decoding method according to the present invention include a moving image encoding device that performs encoding and decoding processing with high encoding efficiency, This is useful for a moving picture decoding device or the like.

Landscapes

  • Engineering & Computer Science (AREA)
  • Multimedia (AREA)
  • Signal Processing (AREA)
  • Compression Or Coding Systems Of Tv Signals (AREA)

Abstract

The present invention addresses the problem of intra prediction efficiency or orthogonal transformation coefficient encoding efficiency decreasing due to spatial resolution in the vertical direction dropping to 1/2, and correlation between pixels decreasing when subjecting an interlace signal to field encoding. The present invention is provided with a variable-length encoding means that splits orthogonal transformation blocks into orthogonal transformation sub-blocks, and switches the encoding order of post-quantization transformation coefficients, which comprise compressed data, between orthogonal transformation block units and orthogonal transformation sub-block units, on the basis of whether a flag based on information indicating whether field encoding is to be performed is enabled.

Description

動画像符号化装置、動画像復号装置、動画像符号化方法及び動画像復号方法Moving picture encoding apparatus, moving picture decoding apparatus, moving picture encoding method, and moving picture decoding method
 この発明は、動画像を高効率で符号化を行う動画像符号化装置及び動画像符号化方法と、高効率で符号化されている動画像を復号する動画像復号装置及び動画像復号方法とに関するものである。 The present invention relates to a moving image encoding apparatus and moving image encoding method for encoding a moving image with high efficiency, a moving image decoding apparatus and a moving image decoding method for decoding a moving image encoded with high efficiency, and It is about.
 従来、MPEGやITU-T H.26x等の国際標準映像符号化方式では、入力映像フレームを、16×16画素ブロックからなるマクロブロックの単位に分割して、動き補償予測を実施した後、予測差分信号をブロック単位に直交変換・量子化することによって情報圧縮を行うようにしている。 Conventionally, MPEG and ITU-T H.264. In an international standard video coding scheme such as 26x, an input video frame is divided into macroblock units each composed of 16 × 16 pixel blocks, and after motion compensation prediction is performed, a prediction difference signal is orthogonally converted into block units. Information compression is performed by quantization.
 ここで、図23は非特許文献1に開示されているMPEG-4 AVC/H.264の動画像符号化装置を示す構成図である。
 この動画像符号化装置では、ブロック分割部101が符号化対象の画像信号を入力すると、その画像信号をマクロブロック単位に分割し、マクロブロック単位の画像信号を分割画像信号として予測部102に出力する。
 予測部102は、ブロック分割部101から分割画像信号を受けると、マクロブロック内の各色成分の画像信号をフレーム内又はフレーム間で予測して、予測差分信号を算出する。
Here, FIG. 23 shows MPEG-4 AVC / H. 1 is a configuration diagram illustrating a H.264 video encoding device.
In this moving image encoding apparatus, when the block division unit 101 receives an image signal to be encoded, the block division unit 101 divides the image signal into macroblock units, and outputs the macroblock unit image signal to the prediction unit 102 as a divided image signal. To do.
When the prediction unit 102 receives the divided image signal from the block division unit 101, the prediction unit 102 predicts the image signal of each color component in the macroblock within a frame or between frames, and calculates a prediction difference signal.
 特に、フレーム間で動き補償予測を実施する場合、マクロブロック自体、または、マクロブロックをさらに細かく分割したサブブロックの単位で動きベクトルを探索する。
 そして、その動きベクトルを用いて、メモリ107により格納されている参照画像信号に対する動き補償予測を実施することで動き補償予測画像を生成し、その動き補償予測画像を示す予測信号と分割画像信号の差分を求めることで予測差分信号を算出する。
In particular, when motion compensation prediction is performed between frames, a motion vector is searched in units of a macroblock itself or a subblock obtained by further dividing a macroblock.
Then, using the motion vector, a motion compensation prediction image is generated by performing motion compensation prediction on the reference image signal stored in the memory 107, and a prediction signal indicating the motion compensation prediction image and the divided image signal are generated. A prediction difference signal is calculated by obtaining the difference.
一方、フレーム内予測を行う場合、非特許文献1では輝度のイントラ予測モードとしてブロック単位に、複数の予測モードの中から1つの予測モードを選択することができる。
 図24は輝度のブロックサイズが4×4画素の場合のイントラ予測モードを示す説明図である。
 図24では、ブロック内の白丸が符号化対象の画素を表し、黒丸は予測に用いる画素である符号化済みの画素を表している。輝度のブロックサイズが4×4画素の場合には、モード0からモード8の9つのイントラ予測モードが規定されている。
On the other hand, when performing intra-frame prediction, Non-Patent Document 1 can select one prediction mode from a plurality of prediction modes for each block as a luminance intra prediction mode.
FIG. 24 is an explanatory diagram illustrating an intra prediction mode when the luminance block size is 4 × 4 pixels.
In FIG. 24, white circles in the block represent pixels to be encoded, and black circles represent encoded pixels that are pixels used for prediction. When the luminance block size is 4 × 4 pixels, nine intra prediction modes from mode 0 to mode 8 are defined.
 図24において、モード2は平均値(DC)予測を行うモードであり、ブロックの上と左の隣接画素の平均値で、ブロック内の画素を予測するものである。
 モード2以外のモードは方向性予測を行うモードである。モード0は垂直方向予測であり、ブロックの上の隣接画素を垂直方向に繰り返すことで予測画像を生成するものである。例えば、縦縞模様のときにはモード0が選択される。
 モード1は水平方向予測であり、ブロックの左の隣接画素を水平方向に繰り返すことで予測画像を生成するものである。例えば、横縞模様のときにはモード1が選択される。
 モード3からモード8は、ブロックの上又は左の符号化済みの画素を用いて、所定の方向(矢印が示す方向)に補間画素を生成して予測画像を生成するものである。
In FIG. 24, mode 2 is a mode in which average value (DC) prediction is performed, and the pixels in the block are predicted based on the average value of the adjacent pixels above and to the left of the block.
Modes other than mode 2 are modes in which directionality prediction is performed. Mode 0 is vertical prediction, in which a prediction image is generated by repeating adjacent pixels on the block in the vertical direction. For example, mode 0 is selected for a vertical stripe pattern.
Mode 1 is horizontal prediction, in which a prediction image is generated by repeating the adjacent pixel on the left of the block in the horizontal direction. For example, mode 1 is selected for a horizontal stripe pattern.
In mode 3 to mode 8, a predicted image is generated by generating an interpolation pixel in a predetermined direction (direction indicated by an arrow) using the encoded pixels on the upper or left side of the block.
 ここで、イントラ予測を適用する輝度のブロックサイズは、4×4画素、8×8画素、16×16画素の中から選択することができ、8×8画素の場合には、4×4画素と同様に、9つのイントラ予測モードが規定されている。ただし、予測に用いる画素については、符号化済みの画素そのものではなく、これらの画素に対してフィルタ処理を施したものを用いている。
 これに対し、16×16画素の場合には、平均値予測、垂直方向予測及び水平方向予測に係るイントラ予測モードに加えて、Plane予測と呼ばれる4つのイントラ予測モードが規定されている。
 Plane予測に係るイントラ予測モードは、ブロックの上と左の符号化済みの隣接画素を斜め方向に内挿補間して生成された画素を予測値とするモードである。
Here, the luminance block size to which the intra prediction is applied can be selected from 4 × 4 pixels, 8 × 8 pixels, and 16 × 16 pixels, and in the case of 8 × 8 pixels, 4 × 4 pixels. Similarly, nine intra prediction modes are defined. However, the pixels used for prediction are not encoded pixels themselves, but those obtained by performing filter processing on these pixels.
On the other hand, in the case of 16 × 16 pixels, in addition to intra prediction modes related to average value prediction, vertical direction prediction, and horizontal direction prediction, four intra prediction modes called “plane prediction” are defined.
The intra prediction mode related to the Plane prediction is a mode in which a pixel generated by interpolating the upper and left encoded adjacent pixels in the block in an oblique direction is a prediction value.
 また、予測部102は、予測信号を得る際に決定した予測信号生成用パラメータを可変長符号化部108に出力する。
 なお、予測信号生成用パラメータには、例えば、フレーム内での空間予測をどのように行うかを示すイントラ予測モードや、フレーム間の動き量を示す動きベクトル等の情報が含まれる。
Further, the prediction unit 102 outputs the prediction signal generation parameter determined when obtaining the prediction signal to the variable length coding unit 108.
Note that the prediction signal generation parameter includes, for example, information such as an intra prediction mode indicating how to perform spatial prediction within a frame and a motion vector indicating a motion amount between frames.
 圧縮部103は、予測部102から予測差分信号を受けると、その予測差分信号に対するDCT(離散コサイン変換)処理を実施することで信号相関を除去した後、量子化することで圧縮データを得る。
 局所復号部104は、圧縮部103から圧縮データを受けると、その圧縮データを逆量子化して、逆DCT処理を実施することで、予測部102から出力された予測差分信号に相当する予測差分信号を算出する。
When the compression unit 103 receives the prediction difference signal from the prediction unit 102, the compression unit 103 performs DCT (discrete cosine transform) processing on the prediction difference signal to remove signal correlation, and then obtains compressed data by quantization.
When the local decoding unit 104 receives the compressed data from the compression unit 103, the local decoding unit 104 inversely quantizes the compressed data and performs inverse DCT processing, whereby the prediction differential signal corresponding to the prediction differential signal output from the prediction unit 102 Is calculated.
 加算器105は、局所復号部104から予測差分信号を受けると、その予測差分信号と予測部102から出力された予測信号を加算して、局所復号画像を生成する。
 ループフィルタ106は、加算器105により生成された局所復号画像を示す局所復号画像信号に重畳されているブロック歪みを除去し、歪み除去後の局所復号画像信号を参照画像信号としてメモリ107に格納する。
When the adder 105 receives the prediction difference signal from the local decoding unit 104, the adder 105 adds the prediction difference signal and the prediction signal output from the prediction unit 102 to generate a local decoded image.
The loop filter 106 removes block distortion superimposed on the locally decoded image signal indicating the locally decoded image generated by the adder 105, and stores the locally decoded image signal after distortion removal in the memory 107 as a reference image signal. .
 可変長符号化部108は、圧縮部103から圧縮データを受けると、その圧縮データをエントロピー符号化し、その符号化結果であるビットストリームを出力する。
 なお、可変長符号化部108は、ビットストリームを出力する際、予測部102から出力された予測信号生成用パラメータをビットストリームに多重化して出力する。
When the variable length coding unit 108 receives the compressed data from the compression unit 103, the variable length coding unit 108 performs entropy coding on the compressed data and outputs a bit stream that is a result of the coding.
Note that the variable length coding unit 108 multiplexes the prediction signal generation parameter output from the prediction unit 102 into the bit stream and outputs the bit stream.
 ここで、一般に符号化対象となる映像信号フォーマットとしては図25に示すフレームが全て同一時刻の信号で構成されるプログレッシブ信号と図26に示すフレームが2つの異なる時刻の信号(フィールド)で構成されるインタレース信号がある。非特許文献1ではインタレース信号を効率的に符号化するために、インタレース信号をピクチャ単位やマクロブロック単位にフレームとして符号化するかフィールドとして符号化するかを適応的に切り替える機能等の様々な符号化ツールが組み込まれている。 Here, in general, as a video signal format to be encoded, a progressive signal in which all the frames shown in FIG. 25 are composed of signals at the same time and a frame shown in FIG. 26 are composed of signals (fields) at two different times. There is an interlace signal. In Non-Patent Document 1, in order to efficiently encode an interlace signal, various functions such as a function of adaptively switching whether the interlace signal is encoded as a frame or a field in units of pictures or macroblocks. A simple encoding tool is incorporated.
 一方、非特許文献2に開示されている方式では、インタレース信号の符号化効率改善のための特別な符号化ツールは用意されていない。非特許文献2ではイントラ予測時に図27に示すように特定のイントラ予測モードにおいて予測画像に対してブロック境界の連続性を高めるためのフィルタ処理を実施する。しかし、フィールド単位に符号化を実施する場合、垂直方向の空間的相関が低下するためにブロックの上端に対するフィルタ処理の効果が大きく低下する可能性がある。 On the other hand, in the method disclosed in Non-Patent Document 2, a special encoding tool for improving the encoding efficiency of interlaced signals is not prepared. In Non-Patent Document 2, a filter process for increasing the continuity of block boundaries is performed on a predicted image in a specific intra prediction mode as shown in FIG. 27 during intra prediction. However, when encoding is performed on a field basis, the spatial correlation in the vertical direction is lowered, so that the effect of the filtering process on the upper end of the block may be greatly reduced.
また、非特許文献2は直交変換係数の符号化手法として直交変換ブロックをさらにCoefficient Group(CG)と呼ばれる4x4画素単位のブロック(直交変換サブブロック)に分割してCG単位に係数の符号化処理を実施する。図28に16×16画素の直交変換ブロックにおける係数の符号化順(スキャン順)を示す。このように4×4画素単位の16個のCGを右下のCGから順に符号化処理し、さらに各CGはCG内の16個の係数を右下の係数から順に符号化する。具体的には、まずCG内の16個の係数の中に有意(非零)係数が存在するか否かのフラグ情報を符号化し、次にCG内に有意(非零)係数が存在する場合のみCG内の各係数が有意(非零)係数であるかを上記順に符号化し、最後に有意(非零)係数に対してその係数値情報を順に符号化する。これをCG単位に上記順に行う。その際、有意(非零)係数がなるべく連続で発生するように偏るスキャン順とした方がエントロピー符号化による符号化効率を高めることができる。プログレッシブ映像とインタレース映像とでは、有意(非零)係数発生分布が異なるため、図28のスキャン順では効率的に符号化できなくなってしまう。 Further, Non-Patent Document 2 divides the orthogonal transform block into 4 × 4 pixel unit blocks (orthogonal transform sub-blocks) called Coefficients Group (CG) as an orthogonal transform coefficient encoding method, and encodes coefficients in CG units. To implement. FIG. 28 shows the coding order (scan order) of the coefficients in the 16 × 16 pixel orthogonal transform block. In this way, 16 CGs in units of 4 × 4 pixels are encoded in order from the lower right CG, and each CG encodes 16 coefficients in the CG in order from the lower right coefficient. Specifically, first, flag information indicating whether or not a significant (nonzero) coefficient exists in 16 coefficients in the CG is encoded, and then a significant (nonzero) coefficient exists in the CG. Only whether each coefficient in the CG is a significant (non-zero) coefficient is encoded in the above order, and finally, coefficient value information is encoded in order for a significant (non-zero) coefficient. This is performed in the above order in units of CG. In this case, the encoding efficiency by entropy encoding can be increased by using a biased scan order so that significant (non-zero) coefficients are generated as continuously as possible. Since progressive video and interlaced video have different distributions of significant (non-zero) coefficients, they cannot be efficiently encoded in the scan order of FIG.
 非特許文献2の動画像符号化装置は以上のように構成されているので、インタレース信号をフィールド符号化する際に垂直方向の空間解像度が1/2となり画素間相関が低下することによって、イントラ予測の予測効率や直交変換係数の符号化効率が低下してしまう課題があった。 Since the moving image encoding apparatus of Non-Patent Document 2 is configured as described above, when the interlace signal is field-encoded, the vertical spatial resolution becomes 1/2 and the inter-pixel correlation decreases. There has been a problem that the prediction efficiency of intra prediction and the encoding efficiency of orthogonal transform coefficients are reduced.
 この発明は上記のような課題を解決するためになされたもので、インタレース信号をフィールド符号化する場合であっても、インタレース信号をフィールド符号化する際の符号化効率を高めることができる動画像符号化装置、動画像復号装置、動画像符号化方法及び動画像復号方法を得ることを目的とする。 The present invention has been made to solve the above-described problems, and even when an interlace signal is field-encoded, it is possible to improve encoding efficiency when the interlace signal is field-encoded. It is an object to obtain a moving image encoding device, a moving image decoding device, a moving image encoding method, and a moving image decoding method.
 この発明に係る動画像符号化装置は、圧縮データ及び符号化モードが多重化された符号化ビットストリームを生成する可変長符号化手段を備え、上記可変長符号化手段は、直交変換ブロックを直交変換サブブロックに分割し、フィールド符号化か否かを示す情報に基づくフラグが有効であるか否かに基づいて、圧縮データである量子化後の変換係数の符号化順を、前記直交変換ブロック単位、前記直交変換サブブロック単位で切り替えるようにしたものである。 The moving picture coding apparatus according to the present invention comprises variable length coding means for generating a coded bitstream in which compressed data and coding modes are multiplexed, and the variable length coding means orthogonally transforms orthogonal transform blocks. Based on whether or not a flag based on information indicating whether or not to perform field encoding is valid and divided into transform sub-blocks, the coding order of the transform coefficients after quantization that is compressed data is changed to the orthogonal transform block. The unit is switched in units of the orthogonal transform sub-block.
 この発明によれば、直交変換ブロックを直交変換サブブロックに分割し、フィールド符号化か否かを示す情報に基づくフラグが有効であるか否かに基づいて、圧縮データである量子化後の変換係数の符号化順を、前記直交変換ブロック単位、前記直交変換サブブロック単位で切り替えるように構成したので、フィールド信号の特性に応じた効率的な予測処理及び符号化処理を実現することができ、符号化効率を高めることができる効果がある。 According to the present invention, an orthogonal transform block is divided into orthogonal transform sub-blocks, and a quantized transform that is compressed data is based on whether or not a flag based on information indicating whether or not field coding is valid. Since the coding order of the coefficients is configured to be switched in units of the orthogonal transform block and the orthogonal transform sub-block, it is possible to realize efficient prediction processing and coding processing according to the characteristics of the field signal, There is an effect of improving the encoding efficiency.
この発明の実施の形態1による動画像符号化装置を示す構成図である。It is a block diagram which shows the moving image encoder by Embodiment 1 of this invention. この発明の実施の形態1による動画像符号化装置の処理内容(動画像符号化方法)を示すフローチャートである。It is a flowchart which shows the processing content (moving image encoding method) of the moving image encoding device by Embodiment 1 of this invention. この発明の実施の形態1による動画像復号装置を示す構成図である。It is a block diagram which shows the moving image decoding apparatus by Embodiment 1 of this invention. この発明の実施の形態1による動画像復号装置の処理内容(動画像復号方法)を示すフローチャートである。It is a flowchart which shows the processing content (moving image decoding method) of the moving image decoding apparatus by Embodiment 1 of this invention. 最大符号化ブロックが階層的に複数の符号化ブロックに分割される例を示す説明図である。It is explanatory drawing which shows the example by which the largest encoding block is divided | segmented into a some encoding block hierarchically. (a)は分割後の符号化ブロック及び予測ブロックの分布を示し、(b)は階層分割によって符号化モードm(B)が割り当てられる状況を示す説明図である。(A) shows the distribution of the encoding block and prediction block after a division | segmentation, (b) is explanatory drawing which shows the condition where encoding mode m ( Bn ) is allocated by hierarchy division | segmentation. 符号化ブロックB内の各予測ブロックP が選択可能なイントラ予測パラメータ(イントラ予測モード)の一例を示す説明図である。Is an explanatory diagram showing an example of the prediction block P i n-selectable intra prediction parameter coding block B n (intra prediction mode). =m =4の場合の予測ブロックP 内の画素の予測値を生成する際に用いる画素の一例を示す説明図である。It is an explanatory diagram showing an example of a pixel used for generating the predicted values of the pixels in the prediction block P i n in the case of l i n = m i n = 4. 予測ブロックP 内の左上画素を原点とする相対座標を示す説明図である。The upper left pixels in the prediction block P i n is an explanatory diagram showing a relative coordinate whose origin. 量子化マトリクスの一例を示す説明図である。It is explanatory drawing which shows an example of a quantization matrix. この発明の実施の形態1による動画像符号化装置のループフィルタ部で複数のループフィルタ処理を用いる場合の構成例を示す説明図である。It is explanatory drawing which shows the structural example in the case of using a some loop filter process in the loop filter part of the moving image encoder by Embodiment 1 of this invention. この発明の実施の形態1による動画像復号装置のループフィルタ部で複数のループフィルタ処理を用いる場合の構成例を示す説明図である。It is explanatory drawing which shows the structural example in the case of using a several loop filter process in the loop filter part of the moving image decoding apparatus by Embodiment 1 of this invention. 符号化ビットストリームの一例を示す説明図である。It is explanatory drawing which shows an example of an encoding bit stream. 画素適応オフセット処理のクラス分類手法のインデックスを示す説明図である。It is explanatory drawing which shows the index of the class classification method of a pixel adaptive offset process. 16×16画素のサイズの直交変換における変換係数の分布の一例を示す説明図である。It is explanatory drawing which shows an example of distribution of the transformation coefficient in the orthogonal transformation of the size of 16x16 pixels. フィールド信号における16×16画素のサイズの直交変換における変換係数の分布の一例を示す説明図である。It is explanatory drawing which shows an example of distribution of the transformation coefficient in the orthogonal transformation of the size of 16x16 pixels in the field signal. フィールド信号における16×16画素のサイズの直交変換における変換係数の符号化順を示す説明図である。It is explanatory drawing which shows the encoding order of the conversion factor in the orthogonal transformation of the size of 16x16 pixels in a field signal. フィールド信号における16×16画素のサイズの直交変換における変換係数の符号化順を示す説明図である。It is explanatory drawing which shows the encoding order of the conversion factor in the orthogonal transformation of the size of 16x16 pixels in a field signal. フィールド信号における16×16画素のサイズの直交変換における変換係数の符号化順を示す説明図である。It is explanatory drawing which shows the encoding order of the conversion factor in the orthogonal transformation of the size of 16x16 pixels in a field signal. 平均値予測時のフィルタ処理におけるフィルタの切り替え領域を示す説明図である。It is explanatory drawing which shows the switching area | region of the filter in the filter process at the time of average value prediction. 平均値予測時のフィルタ処理の参照画素配置を示す説明図である。It is explanatory drawing which shows the reference pixel arrangement | positioning of the filter process at the time of average value prediction. フィールド符号化の際のイントラ予測画像に対するフィルタ処理を示す説明図である。It is explanatory drawing which shows the filter process with respect to the intra estimated image in the case of field coding. 非特許文献1に開示されている動画像符号化装置を示す構成図である。It is a block diagram which shows the moving image encoder disclosed by the nonpatent literature 1. 輝度のブロックサイズが4×4画素の場合のイントラ予測モードを示す説明図である。It is explanatory drawing which shows intra prediction mode in case the block size of a brightness | luminance is 4x4 pixel. プログレッシブ映像信号を示す説明図である。It is explanatory drawing which shows a progressive video signal. インタレース映像信号を示す説明図である。It is explanatory drawing which shows an interlace video signal. イントラ予測におけるフィルタ処理を示す説明図である。It is explanatory drawing which shows the filter process in intra prediction. 16×16画素のサイズの直交変換における変換係数の符号化順を示す説明図である。It is explanatory drawing which shows the encoding order of the conversion factor in the orthogonal transformation of the size of 16x16 pixels.
実施の形態1.
 図1はこの発明の実施の形態1による動画像符号化装置を示す構成図である。
 図1において、スライス分割部14は入力画像として映像信号を入力すると、その入力画像を符号化制御部2により決定されたスライス分割情報にしたがって1以上の“スライス”という部分画像に分割する処理を実施する。スライスの分割単位は、後述する符号化ブロック単位まで細かくすることができる。なお、スライス分割部14はスライス分割手段を構成している。
Embodiment 1 FIG.
1 is a block diagram showing a moving picture coding apparatus according to Embodiment 1 of the present invention.
In FIG. 1, when a video signal is input as an input image, the slice dividing unit 14 performs a process of dividing the input image into one or more partial images called “slices” according to the slice division information determined by the encoding control unit 2. carry out. The slice division unit can be finely divided to a coding block unit described later. The slice division unit 14 constitutes a slice division unit.
 ブロック分割部1はスライス分割部14により分割されたスライスを入力する毎に、そのスライスを符号化制御部2により決定された最大サイズの符号化ブロックである最大符号化ブロックに分割するとともに、符号化制御部2により決定された上限の階層数に至るまで、その最大符号化ブロックを階層的に各符号化ブロックへ分割する処理を実施する。
 即ち、ブロック分割部1はスライスを符号化制御部2により決定された分割に応じて各符号化ブロックに分割して、その符号化ブロックを出力する処理を実施する。また、各符号化ブロックは予測処理単位となる1つないし複数の予測ブロックに分割される。
 なお、ブロック分割部1はブロック分割手段を構成している。
Each time the block dividing unit 1 inputs the slice divided by the slice dividing unit 14, the slice dividing unit 1 divides the slice into maximum coding blocks which are coding blocks of the maximum size determined by the coding control unit 2, and Until the upper limit number of hierarchies determined by the encoding control unit 2 is reached, a process of dividing the maximum encoded block hierarchically into each encoded block is performed.
That is, the block dividing unit 1 divides the slice into each encoded block in accordance with the division determined by the encoding control unit 2, and performs a process of outputting the encoded block. Each coding block is divided into one or a plurality of prediction blocks which are prediction processing units.
The block dividing unit 1 constitutes a block dividing unit.
 符号化制御部2は符号化処理が実施される際の処理単位となる符号化ブロックの最大サイズを決定するとともに、最大サイズの符号化ブロックが階層的に分割される際の上限の階層数を決定することで、各々の符号化ブロックのサイズを決定する処理を実施する。
 また、符号化制御部2は選択可能な1以上の符号化モード(予測処理単位を示す予測ブロックのサイズなどが異なる1以上のイントラ符号化モード、予測ブロックのサイズなどが異なる1以上のインター符号化モード)の中から、ブロック分割部1から出力される符号化ブロックに適用する符号化モードを選択する処理を実施する。選択手法の例としては、選択可能な1以上の符号化モードの中から、ブロック分割部1から出力される符号化ブロックに対する符号化効率が最も高い符号化モードを選択する手法がある。
The encoding control unit 2 determines the maximum size of the encoded block that is a processing unit when the encoding process is performed, and sets the upper limit number of layers when the encoded block of the maximum size is hierarchically divided. By determining, the process of determining the size of each encoded block is performed.
The encoding control unit 2 also includes one or more selectable encoding modes (one or more intra encoding modes having different prediction block sizes indicating prediction processing units, one or more inter codes having different prediction block sizes, and the like). The coding mode to be applied to the coding block output from the block dividing unit 1 is selected from among the coding modes. As an example of the selection method, there is a method of selecting a coding mode having the highest coding efficiency for the coding block output from the block dividing unit 1 from one or more selectable coding modes.
 また、符号化制御部2は符号化効率が最も高い符号化モードがイントラ符号化モードである場合、そのイントラ符号化モードで符号化ブロックに対するイントラ予測処理を実施する際に用いるイントラ予測パラメータを上記イントラ符号化モードが示す予測処理単位である予測ブロック毎に決定し、符号化効率が最も高い符号化モードがインター符号化モードである場合、そのインター符号化モードで符号化ブロックに対するインター予測処理を実施する際に用いるインター予測パラメータを上記インター符号化モードが示す予測処理単位である予測ブロック毎に決定する処理を実施する。
 さらに、符号化制御部2は変換・量子化部7及び逆量子化・逆変換部8に与える予測差分符号化パラメータを決定する処理を実施する。予測差分符号化パラメータには、符号化ブロックにおける直交変換処理単位となる直交変換ブロックの分割情報を示す直交変換ブロック分割情報や、変換係数の量子化を行う際の量子化ステップサイズを規定する量子化パラメータなどが含まれる。
 なお、符号化制御部2は符号化制御手段を構成している。
In addition, when the coding mode having the highest coding efficiency is the intra coding mode, the coding control unit 2 sets the intra prediction parameters used when performing the intra prediction processing on the coding block in the intra coding mode. When the coding mode having the highest coding efficiency is the inter coding mode, determined for each prediction block that is the prediction processing unit indicated by the intra coding mode, the inter prediction process for the coding block is performed in the inter coding mode. The process which determines the inter prediction parameter used when implementing for every prediction block which is a prediction process unit which the said inter coding mode shows is implemented.
Further, the encoding control unit 2 performs a process of determining a prediction difference encoding parameter to be given to the transform / quantization unit 7 and the inverse quantization / inverse transform unit 8. The prediction difference encoding parameter includes orthogonal transform block division information indicating the division information of the orthogonal transform block, which is an orthogonal transform processing unit in the encoded block, and a quantum that specifies a quantization step size when the transform coefficient is quantized. Parameters are included.
The encoding control unit 2 constitutes an encoding control unit.
 切換スイッチ3は符号化制御部2により決定された符号化モードがイントラ符号化モードであれば、ブロック分割部1から出力された符号化ブロックをイントラ予測部4に出力し、符号化制御部2により決定された符号化モードがインター符号化モードであれば、ブロック分割部1から出力された符号化ブロックを動き補償予測部5に出力する処理を実施する。 If the coding mode determined by the coding control unit 2 is the intra coding mode, the changeover switch 3 outputs the coded block output from the block dividing unit 1 to the intra prediction unit 4 and the coding control unit 2. If the coding mode determined by the above is the inter coding mode, a process of outputting the coding block output from the block dividing unit 1 to the motion compensation prediction unit 5 is performed.
 イントラ予測部4は切換スイッチ3から出力された符号化ブロックに対応する符号化モードとして、符号化制御部2によりイントラ符号化モードが選択された場合、その符号化ブロックの予測処理を行う際の予測処理単位である予測ブロック毎に、イントラ予測用メモリ10に格納されている局所復号画像を参照しながら、符号化制御部2により決定されたイントラ予測パラメータを用いたイントラ予測処理(フレーム内予測処理)を実施してイントラ予測画像を生成する処理を実施する。 When the intra control mode is selected by the encoding control unit 2 as the encoding mode corresponding to the encoded block output from the changeover switch 3, the intra prediction unit 4 performs the prediction process for the encoded block. For each prediction block which is a prediction processing unit, an intra prediction process (intraframe prediction) using an intra prediction parameter determined by the encoding control unit 2 while referring to a locally decoded image stored in the intra prediction memory 10. Process) to generate an intra-predicted image.
 動き補償予測部5は切換スイッチ3から出力された符号化ブロックに対応する符号化モードとして、符号化制御部2によりインター符号化モードが選択された場合、符号化ブロックと動き補償予測フレームメモリ12に格納されている1フレーム以上の局所復号画像を予測処理単位である予測ブロック単位に比較して動きベクトルを探索し、その動きベクトルと符号化制御部2により決定された参照するフレーム番号などのインター予測パラメータを用いて、その符号化ブロックに対するインター予測処理(動き補償予測処理)を予測ブロック単位に実施してインター予測画像を生成する処理を実施する。
 なお、イントラ予測部4、イントラ予測用メモリ10、動き補償予測部5及び動き補償予測フレームメモリ12から予測手段が構成されている。
When the inter coding mode is selected by the coding control unit 2 as the coding mode corresponding to the coding block output from the changeover switch 3, the motion compensation prediction unit 5 and the motion compensation prediction frame memory 12 A motion vector is searched by comparing a locally decoded image of one frame or more stored in the image with a prediction block unit that is a prediction processing unit, and the motion vector and a frame number to be referenced determined by the encoding control unit 2 Using the inter prediction parameters, an inter prediction process (motion compensation prediction process) for the encoded block is performed for each prediction block, and an inter prediction image is generated.
The intra prediction unit 4, the intra prediction memory 10, the motion compensation prediction unit 5, and the motion compensation prediction frame memory 12 constitute a prediction unit.
 減算部6はブロック分割部1より出力された符号化ブロックから、イントラ予測部4により生成されたイントラ予測画像、または、動き補償予測部5により生成されたインター予測画像を減算して、その減算結果である差分画像を示す予測差分信号を変換・量子化部7に出力する処理を実施する。なお、減算部6は差分画像生成手段を構成している。
 変換・量子化部7は符号化制御部2により決定された予測差分符号化パラメータに含まれる直交変換ブロック分割情報を参照して、減算部6から出力された予測差分信号に対する直交変換処理(例えば、DCT(離散コサイン変換)やDST(離散サイン変換)、予め特定の学習系列に対して基底設計がなされているKL変換等の直交変換処理)を直交変換ブロック単位に実施して変換係数を算出するとともに、その予測差分符号化パラメータに含まれる量子化パラメータを参照して、その直交変換ブロック単位の変換係数を量子化し、量子化後の変換係数である圧縮データを逆量子化・逆変換部8及び可変長符号化部13に出力する処理を実施する。
 なお、変換・量子化部7は画像圧縮手段を構成している。
The subtraction unit 6 subtracts the intra prediction image generated by the intra prediction unit 4 or the inter prediction image generated by the motion compensated prediction unit 5 from the encoded block output from the block division unit 1 and performs the subtraction. The process which outputs the prediction difference signal which shows the difference image which is a result to the conversion and quantization part 7 is implemented. The subtracting unit 6 constitutes a difference image generating unit.
The transform / quantization unit 7 refers to the orthogonal transform block division information included in the prediction difference coding parameter determined by the coding control unit 2 and performs orthogonal transform processing on the prediction difference signal output from the subtraction unit 6 (for example, , DCT (discrete cosine transform), DST (discrete sine transform), orthogonal transform processing such as KL transform in which a base design is made in advance for a specific learning sequence is performed in units of orthogonal transform blocks to calculate transform coefficients In addition, with reference to the quantization parameter included in the prediction differential encoding parameter, the transform coefficient of the orthogonal transform block unit is quantized, and the quantized compressed data that is the transform coefficient is inversely quantized / inversely transformed. 8 and the process of outputting to the variable length encoding unit 13 is performed.
The transform / quantization unit 7 constitutes an image compression unit.
 変換・量子化部7は変換係数を量子化する際、上記量子化パラメータから算出される量子化ステップサイズを変換係数毎にスケーリングする量子化マトリクスを用いて、変換係数の量子化処理を実施するようにしてもよい。
 ここで、図10は4×4DCTの量子化マトリクスの一例を示す説明図である。
 図中の数字は、各変換係数の量子化ステップサイズのスケーリング値を示している。
 例えば、符号化ビットレートを抑制するために、図10に示すように、高域の変換係数程、量子化ステップサイズを大きな値にスケーリングすることで、複雑な画像領域等で発生する高域の変換係数を抑制して符号量を抑えつつ、主観品質に大きく影響する低域の係数の情報を落とさずに符号化することができる。
 このように、変換係数毎の量子化ステップサイズを制御したい場合には量子化マトリクスを用いればよい。
When the transform / quantization unit 7 quantizes the transform coefficient, the transform / quantization unit 7 performs a quantization process of the transform coefficient using a quantization matrix that scales the quantization step size calculated from the quantization parameter for each transform coefficient. You may do it.
Here, FIG. 10 is an explanatory diagram illustrating an example of a 4 × 4 DCT quantization matrix.
The numbers in the figure indicate the scaling value of the quantization step size of each transform coefficient.
For example, in order to suppress the encoding bit rate, as shown in FIG. 10, by scaling the quantization step size to a larger value for the higher frequency transform coefficient, the higher frequency generated in a complex image region or the like. It is possible to perform encoding without dropping information on low-frequency coefficients that greatly affect subjective quality while suppressing the conversion coefficient and suppressing the code amount.
Thus, when it is desired to control the quantization step size for each transform coefficient, a quantization matrix may be used.
 また、量子化マトリクスは、各直交変換サイズで色信号や符号化モード(イントラ符号化かインター符号化か)毎に独立したマトリクスを使用することができ、初期値として予め動画像符号化装置及び動画像復号装置で共通に用意されている量子化マトリクスや既に符号化された量子化マトリクスの中から選択するか、新しい量子化マトリクスを用いるかをそれぞれ選択することができる。
 したがって、変換・量子化部7は、各直交変換サイズに対して色信号や符号化モード毎に、新しい量子化マトリクスを用いるか否かを示すフラグ情報を符号化すべき量子化マトリクスパラメータに設定する。
 さらに、新しい量子化マトリクスを用いる場合には、図10に示すような量子化マトリクスの各スケーリング値を符号化すべき量子化マトリクスパラメータに設定する。一方、新しい量子化マトリクスを用いない場合には、初期値として予め動画像符号化装置及び動画像復号装置で共通に用意されている量子化マトリクス、または、既に符号化された量子化マトリクスの中から、使用するマトリクスを特定するインデックスを符号化すべき量子化マトリクスパラメータに設定する。ただし、参照可能な既に符号化された量子化マトリクスが存在しない場合、予め動画像符号化装置及び動画像復号装置で共通に用意されている量子化マトリクスのみ選択可能となる。
The quantization matrix can use an independent matrix for each color signal and coding mode (intra coding or inter coding) at each orthogonal transform size. It is possible to select either a quantization matrix that is commonly prepared in the moving picture decoding apparatus or an already encoded quantization matrix or a new quantization matrix.
Accordingly, the transform / quantization unit 7 sets flag information indicating whether or not to use a new quantization matrix for each orthogonal transform size for each color signal and coding mode, in a quantization matrix parameter to be encoded. .
Furthermore, when a new quantization matrix is used, each scaling value of the quantization matrix as shown in FIG. 10 is set as a quantization matrix parameter to be encoded. On the other hand, when a new quantization matrix is not used, a quantization matrix prepared in advance by the moving picture coding apparatus and the moving picture decoding apparatus as an initial value or a quantization matrix that has already been coded. Thus, an index for specifying a matrix to be used is set as a quantization matrix parameter to be encoded. However, when there is no already-encoded quantization matrix that can be referred to, only a quantization matrix prepared in advance by the moving picture coding apparatus and the moving picture decoding apparatus can be selected.
 逆量子化・逆変換部8は符号化制御部2により決定された予測差分符号化パラメータに含まれる量子化パラメータ及び直交変換ブロック分割情報を参照して、直交変換ブロック単位に変換・量子化部7から出力された圧縮データを逆量子化するとともに、逆量子化後の圧縮データである変換係数に対する逆直交変換処理を実施して、減算部6から出力された予測差分信号に相当する局所復号予測差分信号を算出する処理を実施する。なお、変換・量子化部7が量子化マトリクスを用いて、量子化処理を実施している場合には、逆量子化処理時においても、その量子化マトリクスを参照して、対応する逆量子化処理を実施する。
 加算部9は逆量子化・逆変換部8により算出された局所復号予測差分信号と、イントラ予測部4により生成されたイントラ予測画像、または、動き補償予測部5により生成されたインター予測画像とを加算して、ブロック分割部1から出力された符号化ブロックに相当する局所復号画像を算出する処理を実施する。
 なお、逆量子化・逆変換部8及び加算部9から局所復号画像生成手段が構成されている。
The inverse quantization / inverse transform unit 8 refers to the quantization parameter and the orthogonal transform block division information included in the prediction difference coding parameter determined by the coding control unit 2, and transforms / quantizes the orthogonal transform block unit. 7, the quantized data output from the subtractor 6 is inversely quantized, and the inverse orthogonal transform process is performed on the transform coefficient that is the compressed data after the inverse quantization, and the local decoding corresponding to the prediction difference signal output from the subtractor 6 A process of calculating the prediction difference signal is performed. When the transform / quantization unit 7 uses the quantization matrix to perform the quantization process, the corresponding inverse quantization can be performed by referring to the quantization matrix even during the inverse quantization process. Perform the process.
The addition unit 9 includes the local decoded prediction difference signal calculated by the inverse quantization / inverse conversion unit 8, the intra prediction image generated by the intra prediction unit 4, or the inter prediction image generated by the motion compensation prediction unit 5. Are added to calculate a locally decoded image corresponding to the encoded block output from the block dividing unit 1.
The inverse quantization / inverse transform unit 8 and the addition unit 9 constitute a local decoded image generation unit.
 イントラ予測用メモリ10は加算部9により算出された局所復号画像を格納する記録媒体である。
 ループフィルタ部11は加算部9により算出された局所復号画像に対して、所定のフィルタ処理を実施して、フィルタ処理後の局所復号画像を出力する処理を実施する。
 具体的には、直交変換ブロックの境界や予測ブロックの境界に発生する歪みを低減するフィルタ(デブロッキングフィルタ)処理、画素単位に適応的にオフセットを加算する(画素適応オフセット)処理、ウィーナフィルタ等の線形フィルタを適応的に切り替えてフィルタ処理する適応フィルタ処理などを行う。
The intra prediction memory 10 is a recording medium that stores the locally decoded image calculated by the adding unit 9.
The loop filter unit 11 performs a predetermined filtering process on the local decoded image calculated by the adding unit 9 and performs a process of outputting the local decoded image after the filter process.
Specifically, filter (deblocking filter) processing that reduces distortion occurring at the boundaries of orthogonal transform blocks and prediction blocks, processing for adaptively adding an offset in units of pixels (pixel adaptive offset), Wiener filter, etc. The adaptive filter processing that performs the filter processing by adaptively switching the linear filter is performed.
 ただし、ループフィルタ部11は、上記のデブロッキングフィルタ処理、画素適応オフセット処理及び適応フィルタ処理のそれぞれについて、処理を行うか否かを決定し、各処理の有効フラグをヘッダ情報として可変長符号化部13に出力する。なお、上記のフィルタ処理を複数使用する際は、各フィルタ処理を順番に実施する。図11は複数のフィルタ処理を用いる場合のループフィルタ部11の構成例を示している。
 一般に使用するフィルタ処理の種類が多いほど、画像品質は向上するが、一方で処理負荷は高くなる。即ち、画像品質と処理負荷はトレードオフの関係にある。また、各フィルタ処理の画像品質改善効果はフィルタ処理対象画像の特性によって異なる。したがって、動画像符号化装置が許容する処理負荷や符号化処理対象画像の特性にしたがって使用するフィルタ処理を決めればよい。
 なお、ループフィルタ部11はフィルタリング手段を構成している。
However, the loop filter unit 11 determines whether or not to perform each of the deblocking filter process, the pixel adaptive offset process, and the adaptive filter process, and performs variable-length coding using the valid flag of each process as header information. To the unit 13. When a plurality of the above filter processes are used, each filter process is performed in order. FIG. 11 shows a configuration example of the loop filter unit 11 when a plurality of filter processes are used.
Generally, the more types of filter processing that are used, the better the image quality, but the higher the processing load. That is, image quality and processing load are in a trade-off relationship. In addition, the image quality improvement effect of each filter process varies depending on the characteristics of the image to be filtered. Therefore, the filter processing to be used may be determined according to the processing load allowed by the moving image encoding device and the characteristics of the encoding target image.
The loop filter unit 11 constitutes filtering means.
 ここで、デブロッキングフィルタ処理では、ブロック境界にかけるフィルタ強度の選択に用いる各種パラメータを初期値から変更することができる。変更する場合には、そのパラメータをヘッダ情報として可変長符号化部13に出力する。
 画素適応オフセット処理では、最初に、画像を複数のブロックに分割し、そのブロック単位に、オフセット処理を行わない場合もクラス分類手法の一つとして定義して、予め用意されている複数のクラス分類手法の中から、1つのクラス分類手法を選択する。
 次に、選択したクラス分類手法によって、ブロック内の各画素をクラス分類し、クラス毎に符号化歪みを補償するオフセット値を算出する。
 最後に、局所復号画像の輝度値に対して、そのオフセット値を加算する処理を行うことで局所復号画像の画像品質を改善する。
 したがって、画素適応オフセット処理では、ブロック分割情報、各ブロックのクラス分類手法を示すインデックス、ブロック単位の各クラスのオフセット値を特定するオフセット情報をヘッダ情報として可変長符号化部13に出力する。
Here, in the deblocking filter process, various parameters used for selecting the filter strength applied to the block boundary can be changed from the initial values. When changing, the parameter is output to the variable length coding unit 13 as header information.
In the pixel adaptive offset process, first, an image is divided into a plurality of blocks, and when the offset process is not performed for each block, it is defined as one of the class classification methods, and a plurality of class classifications prepared in advance are used. One classification method is selected from the methods.
Next, each pixel in the block is classified by the selected class classification method, and an offset value for compensating the coding distortion is calculated for each class.
Finally, the image quality of the locally decoded image is improved by performing a process of adding the offset value to the luminance value of the locally decoded image.
Therefore, in the pixel adaptive offset processing, block division information, an index indicating the class classification method of each block, and offset information for specifying an offset value of each class in block units are output to the variable length coding unit 13 as header information.
 適応フィルタ処理では、局所復号画像を所定の手法でクラス分類し、各クラスに属する領域(局所復号画像)毎に、重畳されている歪みを補償するフィルタを設計し、そのフィルタを用いて、当該局所復号画像のフィルタ処理を実施する。
 そして、クラス毎に設計したフィルタをヘッダ情報として可変長符号化部13に出力する。
 クラス分類手法としては、画像を空間的に等間隔に区切る簡易な手法や、ブロック単位に画像の局所的な特性(分散など)に応じて分類する手法がある。
 また、適応フィルタ処理で使用するクラス数は、予め動画像符号化装置及び動画像復号装置に共通の値として設定してもよいし、符号化すべきパラメータとしてもよい。
 前者と比較して後者の方が、使用するクラス数を自由に設定することができるため、画像品質改善効果が上がるが、一方でクラス数を符号化するために、その分の符号量が増加する。
In adaptive filter processing, local decoded images are classified by a predetermined method, and a filter that compensates for the distortion that is superimposed is designed for each region (local decoded image) belonging to each class. Filter the local decoded image.
Then, the filter designed for each class is output to the variable length encoding unit 13 as header information.
As the class classification method, there are a simple method for spatially dividing an image at equal intervals, and a method for classifying an image according to local characteristics (dispersion, etc.) of each block.
Further, the number of classes used in the adaptive filter processing may be set in advance as a value common to the moving image encoding device and the moving image decoding device, or may be a parameter to be encoded.
Compared to the former, the latter can set the number of classes to be used freely, so the image quality improvement effect will be improved, but on the other hand, the amount of code will be increased to encode the number of classes. To do.
 なお、画素適応オフセット処理及び適応フィルタ処理を行う場合には、映像信号をループフィルタ部11で参照する必要があるため、映像信号がループフィルタ部11に入力されるように、図1の動画像符号化装置を変更する必要がある。 When performing the pixel adaptive offset process and the adaptive filter process, it is necessary to refer to the video signal by the loop filter unit 11, so that the video image of FIG. 1 is input so that the video signal is input to the loop filter unit 11. It is necessary to change the encoding device.
 動き補償予測フレームメモリ12はループフィルタ部11のフィルタ処理後の局所復号画像を格納する記録媒体である。
 可変長符号化部13は変換・量子化部7から出力された圧縮データと、符号化制御部2の出力信号(最大符号化ブロック内のブロック分割情報、符号化モード、予測差分符号化パラメータ、イントラ予測パラメータ又はインター予測パラメータ)と、動き補償予測部5から出力された動きベクトル(符号化モードがインター符号化モードである場合)とを可変長符号化して符号化データを生成する。
 また、可変長符号化部13は、図13に例示するように、符号化ビットストリームのヘッダ情報として、シーケンスレベルヘッダ、ピクチャレベルヘッダを符号化し、ピクチャデータと共に符号化ビットストリームを生成する。
 なお、可変長符号化部13は可変長符号化手段を構成している。
The motion compensated prediction frame memory 12 is a recording medium that stores a locally decoded image after the filter processing of the loop filter unit 11.
The variable length coding unit 13 outputs the compressed data output from the transform / quantization unit 7 and the output signal of the coding control unit 2 (block division information in the largest coding block, coding mode, prediction difference coding parameter, Intra prediction parameters or inter prediction parameters) and motion vectors output from the motion compensation prediction unit 5 (when the encoding mode is the inter encoding mode) are variable length encoded to generate encoded data.
Further, as illustrated in FIG. 13, the variable length encoding unit 13 encodes a sequence level header and a picture level header as header information of the encoded bit stream, and generates an encoded bit stream together with the picture data.
The variable length encoding unit 13 constitutes variable length encoding means.
 ただし、ピクチャデータは1以上のスライスデータから構成され、各スライスデータはスライスレベルヘッダと当該スライス内にある上記符号化データをまとめたものである。
 シーケンスレベルヘッダは、画像サイズ、色信号フォーマット、輝度信号や色差信号の信号値のビット深度、シーケンス単位でのループフィルタ部11における各フィルタ処理(適応フィルタ処理、画素適応オフセット処理、デブロッキングフィルタ処理)の有効フラグ情報、量子化マトリクスの有効フラグ情報、フィールド符号化か否かを示すフラグなど、一般的にシーケンス単位に共通となるヘッダ情報をまとめたものである。
 ピクチャレベルヘッダは、参照するシーケンスレベルヘッダのインデックスや動き補償時の参照ピクチャ数、エントロピー符号化の確率テーブル初期化フラグ、量子化マトリクスパラメータなど、ピクチャ単位で設定するヘッダ情報をまとめたものである。
However, picture data is composed of one or more slice data, and each slice data is a combination of a slice level header and the encoded data in the slice.
The sequence level header includes the image size, the color signal format, the bit depth of the signal value of the luminance signal and the color difference signal, and each filter process (adaptive filter process, pixel adaptive offset process, deblocking filter process) in the loop filter unit 11 in sequence units. ) Effective flag information, quantization matrix effective flag information, and a flag indicating whether or not to perform field encoding.
The picture level header is a collection of header information set in units of pictures, such as an index of a sequence level header to be referenced, the number of reference pictures at the time of motion compensation, an entropy coding probability table initialization flag, and a quantization matrix parameter. .
 スライスレベルヘッダは、当該スライスがピクチャのどの位置にあるかを示す位置情報、どのピクチャレベルヘッダを参照するかを示すインデックス、スライスの符号化タイプ(オールイントラ符号化、インター符号化など)、ループフィルタ部11における各フィルタ処理(適応フィルタ処理、画素適応オフセット処理、デブロッキングフィルタ処理)を行うか否かを示すフラグ情報などのスライス単位のパラメータをまとめたものである。 The slice level header includes position information indicating where the slice is located in the picture, an index indicating which picture level header is referred to, a slice coding type (all-intra coding, inter coding, etc.), and a loop. This is a summary of parameters in units of slices such as flag information indicating whether or not to perform each filter process (adaptive filter process, pixel adaptive offset process, deblocking filter process) in the filter unit 11.
 図1の例では、動画像符号化装置の構成要素であるブロック分割部1、符号化制御部2、切換スイッチ3、イントラ予測部4、動き補償予測部5、減算部6、変換・量子化部7、逆量子化・逆変換部8、加算部9、イントラ予測用メモリ10、ループフィルタ部11、動き補償予測フレームメモリ12及び可変長符号化部13のそれぞれが専用のハードウェア(例えば、CPUを実装している半導体集積回路や、ワンチップマイコンなど)で構成されているものを想定しているが、動画像符号化装置がコンピュータで構成される場合、ブロック分割部1、符号化制御部2、切換スイッチ3、イントラ予測部4、動き補償予測部5、減算部6、変換・量子化部7、逆量子化・逆変換部8、加算部9、ループフィルタ部11及び可変長符号化部13の処理内容を記述しているプログラムをコンピュータのメモリに格納し、当該コンピュータのCPUが当該メモリに格納されているプログラムを実行するようにしてもよい。
 図2はこの発明の実施の形態1による動画像符号化装置の処理内容(動画像符号化方法)を示すフローチャートである。
In the example of FIG. 1, a block division unit 1, an encoding control unit 2, a changeover switch 3, an intra prediction unit 4, a motion compensation prediction unit 5, a subtraction unit 6, transform / quantization, which are components of the moving image encoding device. Unit 7, inverse quantization / inverse transform unit 8, addition unit 9, intra prediction memory 10, loop filter unit 11, motion compensated prediction frame memory 12, and variable length coding unit 13, each of which has dedicated hardware (for example, It is assumed that the CPU is configured by a semiconductor integrated circuit or a one-chip microcomputer). However, when the moving image encoding apparatus is configured by a computer, the block dividing unit 1, encoding control Unit 2, changeover switch 3, intra prediction unit 4, motion compensation prediction unit 5, subtraction unit 6, transform / quantization unit 7, inverse quantization / inverse transform unit 8, addition unit 9, loop filter unit 11, and variable length code Chemical unit 1 The processing contents stored programs describing the the memory of the computer, may execute a program that the CPU of the computer is stored in the memory.
FIG. 2 is a flowchart showing the processing contents (moving image coding method) of the moving image coding apparatus according to Embodiment 1 of the present invention.
 図3はこの発明の実施の形態1による動画像復号装置を示す構成図である。
 図3において、可変長復号部31は図1の動画像符号化装置により生成された符号化ビットストリームを入力すると、そのビットストリームからシーケンスレベルヘッダ、ピクチャレベルヘッダ、スライスレベルヘッダなどの各ヘッダ情報を復号するとともに、そのビットストリームから、階層的に分割されている各々の符号化ブロックの分割状況を示すブロック分割情報を可変長復号する。
 このとき、可変長復号部31によって可変長復号された量子化マトリクスパラメータから、量子化マトリクスを特定する。具体的には、各直交変換サイズの色信号や符号化モード毎に、量子化マトリクスパラメータが初期値として予め動画像符号化装置及び動画像復号装置で共通に用意されている量子化マトリクス、または、既に復号された量子化マトリクスである(新しい量子化マトリクスでない)ことを示す場合は、上記マトリクスの内のどの量子化マトリクスであるかを特定するインデックス情報を参照して量子化マトリクスを特定し、量子化マトリクスパラメータが新しい量子化マトリクスを用いることを示す場合は、量子化マトリクスパラメータに含まれる量子化マトリクスを使用する量子化マトリクスとして特定する。
 また、可変長復号部31は、各ヘッダ情報を参照して、スライス分割状態を特定するとともに、各スライスのスライスデータに含まれる最大復号ブロック(図1の動画像符号化装置の「最大符号化ブロック」に相当するブロック)を特定し、ブロック分割情報を参照して、最大復号ブロックを階層的に分割して復号処理を行う単位である復号ブロック(図1の動画像符号化装置の「符号化ブロック」に相当するブロック)を特定し、各々の復号ブロックに係る圧縮データ、符号化モード、イントラ予測パラメータ(符号化モードがイントラ符号化モードである場合)、インター予測パラメータ(符号化モードがインター符号化モードである場合)、予測差分符号化パラメータ及び動きベクトル(符号化モードがインター符号化モードである場合)を可変長復号する処理を実施する。なお、可変長復号部31は可変長復号手段を構成している。
FIG. 3 is a block diagram showing a moving picture decoding apparatus according to Embodiment 1 of the present invention.
In FIG. 3, when the variable length decoding unit 31 receives the encoded bit stream generated by the moving image encoding apparatus of FIG. 1, each header information such as a sequence level header, a picture level header, a slice level header, and the like from the bit stream. In addition, the block division information indicating the division status of each encoded block divided hierarchically is variable-length decoded from the bitstream.
At this time, the quantization matrix is specified from the quantization matrix parameter variable-length decoded by the variable-length decoding unit 31. Specifically, for each color signal and encoding mode of each orthogonal transform size, a quantization matrix that is prepared in advance by the moving image encoding device and the moving image decoding device in advance as the initial value of the quantization matrix parameter, or When indicating that the quantization matrix is already decoded (not a new quantization matrix), specify the quantization matrix by referring to the index information for specifying which quantization matrix among the matrices. When the quantization matrix parameter indicates that a new quantization matrix is used, it is specified as a quantization matrix that uses the quantization matrix included in the quantization matrix parameter.
In addition, the variable length decoding unit 31 refers to each header information to specify the slice division state, and the maximum decoding block included in the slice data of each slice (the “maximum encoding” of the video encoding device in FIG. 1). A block corresponding to a “block” is identified, and the decoding block (the “code” of the moving picture encoding device in FIG. 1) is a unit for performing decoding processing by hierarchically dividing the maximum decoded block with reference to block division information. Block) corresponding to each decoded block, the compressed data, the encoding mode, the intra prediction parameter (when the encoding mode is the intra encoding mode), the inter prediction parameter (the encoding mode is Inter-coding mode), predictive differential coding parameters and motion vectors (if the coding mode is inter-coding mode) ) Carries out a process of variable length decoding the. The variable length decoding unit 31 constitutes a variable length decoding unit.
 逆量子化・逆変換部32は可変長復号部31により可変長復号された予測差分符号化パラメータに含まれる量子化パラメータ及び直交変換ブロック分割情報を参照して、可変長復号部31により可変長復号された圧縮データを直交変換ブロック単位に逆量子化するとともに、逆量子化後の圧縮データである変換係数に対する逆直交変換処理を実施して、図1の逆量子化・逆変換部8から出力された局所復号予測差分信号と同一の復号予測差分信号を算出する処理を実施する。なお、逆量子化・逆変換部32は差分画像生成手段を構成している。 The inverse quantization / inverse transform unit 32 refers to the quantization parameter and the orthogonal transform block division information included in the prediction difference encoding parameter variable length decoded by the variable length decoding unit 31, and the variable length decoding unit 31 performs variable length decoding. The decoded compressed data is inversely quantized in units of orthogonal transform blocks, and inverse orthogonal transform processing is performed on transform coefficients that are compressed data after inverse quantization, so that the inverse quantization / inverse transform unit 8 in FIG. A process of calculating the same decoded prediction difference signal as the output local decoding prediction difference signal is performed. The inverse quantization / inverse transform unit 32 constitutes a difference image generation unit.
 ここで、可変長復号部31により可変長復号された各ヘッダ情報が、当該スライスで量子化マトリクスを用いて、逆量子化処理を実施することを示している場合、量子化マトリクスを用いて逆量子化処理を行う。
 具体的には、各ヘッダ情報から特定される量子化マトリクスを用いて逆量子化処理を行う。
Here, when each header information variable-length decoded by the variable-length decoding unit 31 indicates that the inverse quantization process is performed using the quantization matrix in the slice, the header information is inverted using the quantization matrix. Quantization processing is performed.
Specifically, inverse quantization processing is performed using a quantization matrix specified from each header information.
 切換スイッチ33は可変長復号部31により可変長復号された符号化モードがイントラ符号化モードであれば、可変長復号部31により可変長復号されたイントラ予測パラメータをイントラ予測部34に出力し、可変長復号部31により可変長復号された符号化モードがインター符号化モードであれば、可変長復号部31により可変長復号されたインター予測パラメータ及び動きベクトルを動き補償部35に出力する処理を実施する。 The changeover switch 33 outputs the intra-prediction parameter variable-length decoded by the variable-length decoding unit 31 to the intra-prediction unit 34 if the coding mode variable-length decoded by the variable-length decoding unit 31 is the intra-coding mode. If the encoding mode variable-length decoded by the variable-length decoding unit 31 is an inter-coding mode, a process of outputting the inter prediction parameters and motion vectors variable-length decoded by the variable-length decoding unit 31 to the motion compensation unit 35 carry out.
 イントラ予測部34は可変長復号部31により可変長復号されたブロック分割情報から特定される復号ブロックに係る符号化モードがイントラ符号化モードである場合、その復号ブロックの予測処理を行う際の予測処理単位である予測ブロック毎に、イントラ予測用メモリ37に格納されている復号画像を参照しながら、切換スイッチ33から出力されたイントラ予測パラメータを用いたイントラ予測処理(フレーム内予測処理)を実施してイントラ予測画像を生成する処理を実施する。 When the encoding mode related to the decoded block specified from the block division information variable-length decoded by the variable-length decoding unit 31 is the intra-coding mode, the intra prediction unit 34 performs prediction when performing the prediction process of the decoded block For each prediction block that is a processing unit, an intra prediction process (intraframe prediction process) using the intra prediction parameter output from the changeover switch 33 is performed with reference to the decoded image stored in the intra prediction memory 37. Then, a process for generating an intra-predicted image is performed.
 動き補償部35は可変長復号部31により可変長復号されたブロック分割情報から特定される復号ブロックに係る符号化モードがインター符号化モードである場合、上記復号ブロックの予測処理を行う際の予測処理単位である予測ブロック毎に、動き補償予測フレームメモリ39に格納されている復号画像を参照しながら、切換スイッチ33から出力された動きベクトルとインター予測パラメータを用いたインター予測処理(動き補償予測処理)を実施してインター予測画像を生成する処理を実施する。
 なお、イントラ予測部34、イントラ予測用メモリ37、動き補償部35及び動き補償予測フレームメモリ39から予測手段が構成されている。
The motion compensation unit 35 performs prediction when performing the prediction process of the decoded block when the coding mode related to the decoded block specified from the block division information subjected to variable length decoding by the variable length decoding unit 31 is the inter coding mode. For each prediction block that is a processing unit, while referring to the decoded image stored in the motion compensated prediction frame memory 39, the inter prediction process (motion compensated prediction) using the motion vector output from the changeover switch 33 and the inter prediction parameter is used. Process) to generate an inter prediction image.
The intra prediction unit 34, the intra prediction memory 37, the motion compensation unit 35, and the motion compensated prediction frame memory 39 constitute a prediction unit.
 加算部36は逆量子化・逆変換部32により算出された復号予測差分信号と、イントラ予測部34により生成されたイントラ予測画像、または、動き補償部35により生成されたインター予測画像とを加算して、図1の加算部9から出力された局所復号画像と同一の復号画像を算出する処理を実施する。なお、加算部36は復号画像生成手段を構成している。 The addition unit 36 adds the decoded prediction difference signal calculated by the inverse quantization / inverse conversion unit 32 and the intra prediction image generated by the intra prediction unit 34 or the inter prediction image generated by the motion compensation unit 35. Then, a process of calculating the same decoded image as the local decoded image output from the adding unit 9 in FIG. 1 is performed. The adding unit 36 constitutes a decoded image generating unit.
 イントラ予測用メモリ37は加算部36により算出された復号画像をイントラ予測処理で用いる参照画像として格納する記録媒体である。
 ループフィルタ部38は加算部36により算出された復号画像に対して、所定のフィルタ処理を実施して、フィルタ処理後の復号画像を出力する処理を実施する。
 具体的には、直交変換ブロックの境界や予測ブロックの境界に発生する歪みを低減するフィルタ(デブロッキングフィルタ)処理、画素単位に適応的にオフセットを加算する(画素適応オフセット)処理、ウィーナフィルタ等の線形フィルタを適応的に切り替えてフィルタ処理する適応フィルタ処理などを行う。
 ただし、ループフィルタ部38は、上記のデブロッキングフィルタ処理、画素適応オフセット処理、適応フィルタ処理のそれぞれについて、可変長復号部31により可変長復号された各ヘッダ情報を参照して、当該スライスで行うか否かを特定する。
 このとき、2つ以上のフィルタ処理を行う場合において、動画像符号化装置のループフィルタ部11が図11のように構成されていれば、図12に示すようにループフィルタ部38が構成される。
 なお、ループフィルタ部38はフィルタリング手段を構成している。
The intra prediction memory 37 is a recording medium that stores the decoded image calculated by the adding unit 36 as a reference image used in the intra prediction process.
The loop filter unit 38 performs a predetermined filter process on the decoded image calculated by the adding unit 36 and performs a process of outputting the decoded image after the filter process.
Specifically, filter (deblocking filter) processing that reduces distortion occurring at the boundaries of orthogonal transform blocks and prediction blocks, processing for adaptively adding an offset in units of pixels (pixel adaptive offset), Wiener filter, etc. The adaptive filter processing that performs the filter processing by adaptively switching the linear filter is performed.
However, the loop filter unit 38 performs each of the above deblocking filter processing, pixel adaptive offset processing, and adaptive filter processing with reference to each header information variable-length decoded by the variable-length decoding unit 31 in the corresponding slice. Specify whether or not.
At this time, when two or more filter processes are performed, if the loop filter unit 11 of the moving picture coding apparatus is configured as shown in FIG. 11, a loop filter unit 38 is configured as shown in FIG. .
The loop filter unit 38 constitutes filtering means.
 ここで、デブロッキングフィルタ処理では、可変長復号部31により可変長復号されたヘッダ情報を参照し、ブロック境界にかけるフィルタ強度の選択に用いる各種パラメータを初期値から変更する情報が存在する場合、その変更情報に基づいて、デブロッキングフィルタ処理を実施する。変更情報がない場合は、予め定められた手法に従って行う。 Here, in the deblocking filter processing, with reference to the header information that has been subjected to variable length decoding by the variable length decoding unit 31, there is information for changing various parameters used for selecting the filter strength applied to the block boundary from the initial value. Based on the change information, deblocking filter processing is performed. When there is no change information, it is performed according to a predetermined method.
 画素適応オフセット処理では、可変長復号部31により可変長復号された画素適応オフセット処理のブロック分割情報に基づいて復号画像を分割し、そのブロック単位に、可変長復号部31により可変長復号されたブロック単位のクラス分類手法を示すインデックスを参照して、そのインデックスが“オフセット処理を行わない”ことを示すインデックスでない場合、ブロック単位にブロック内の各画素を上記インデックスが示すクラス分類手法に従ってクラス分類する。
 なお、クラス分類手法の候補として、ループフィルタ部11の画素適応オフセット処理のクラス分類手法の候補と同一のものが予め用意されている。
 そして、ブロック単位の各クラスのオフセット値を特定するオフセット情報を参照して、復号画像の輝度値にオフセットを加算する処理を行う。
In the pixel adaptive offset processing, the decoded image is divided based on the block division information of the pixel adaptive offset processing variable-length decoded by the variable-length decoding unit 31, and the variable-length decoding unit 31 performs variable-length decoding on the block basis. If the index indicating the block classification method is not an index indicating that “offset processing is not performed”, each pixel in the block is classified according to the class classification method indicated by the index. To do.
In addition, the same class classification method candidate as the pixel classification method candidate of the pixel adaptive offset process of the loop filter unit 11 is prepared in advance as a class classification method candidate.
Then, a process of adding the offset to the luminance value of the decoded image is performed with reference to the offset information specifying the offset value of each class in block units.
 ただし、動画像符号化装置のループフィルタ部11の画素適応オフセット処理において、ブロック分割情報は符号化せずに、常に画像を固定サイズのブロック単位(例えば、最大符号化ブロック単位)に分割し、そのブロック毎にクラス分類手法を選択して、クラス毎の適応オフセット処理を行うように構成されている場合、ループフィルタ部38においても、ループフィルタ部11と同一の固定サイズのブロック単位に画素適応オフセット処理を実施する。 However, in the pixel adaptive offset processing of the loop filter unit 11 of the moving image encoding device, the block division information is not encoded, and the image is always divided into fixed-size block units (for example, maximum encoded block units), When a class classification method is selected for each block and adaptive offset processing is performed for each class, the loop filter unit 38 also applies pixel adaptation to a block unit of the same fixed size as the loop filter unit 11. Perform offset processing.
 適応フィルタ処理では、可変長復号部31により可変長復号されたクラス毎のフィルタを用いて、図1の動画像符号化装置と同一の手法でクラス分類した後に、そのクラス分類情報に基づいてフィルタ処理を行う。
 動き補償予測フレームメモリ39はループフィルタ部38のフィルタ処理後の復号画像をインター予測処理(動き補償予測処理)で用いる参照画像として格納する記録媒体である。
In the adaptive filter processing, after classifying by the same method as the moving picture encoding apparatus of FIG. 1 using the filter for each class variable-length decoded by the variable-length decoding unit 31, the filter is based on the class classification information. Process.
The motion compensation prediction frame memory 39 is a recording medium that stores the decoded image after the filter processing of the loop filter unit 38 as a reference image used in the inter prediction processing (motion compensation prediction processing).
 図3の例では、動画像復号装置の構成要素である可変長復号部31、逆量子化・逆変換部32、切換スイッチ33、イントラ予測部34、動き補償部35、加算部36、イントラ予測用メモリ37、ループフィルタ部38及び動き補償予測フレームメモリ39のそれぞれが専用のハードウェア(例えば、CPUを実装している半導体集積回路や、ワンチップマイコンなど)で構成されているものを想定しているが、動画像復号装置がコンピュータで構成される場合、可変長復号部31、逆量子化・逆変換部32、切換スイッチ33、イントラ予測部34、動き補償部35、加算部36及びループフィルタ部38の処理内容を記述しているプログラムをコンピュータのメモリに格納し、当該コンピュータのCPUが当該メモリに格納されているプログラムを実行するようにしてもよい。
 図4はこの発明の実施の形態1による動画像復号装置の処理内容(動画像復号方法)を示すフローチャートである。
In the example of FIG. 3, the variable length decoding unit 31, the inverse quantization / inverse conversion unit 32, the changeover switch 33, the intra prediction unit 34, the motion compensation unit 35, the addition unit 36, and the intra prediction, which are components of the video decoding device. It is assumed that each of the memory 37, the loop filter unit 38, and the motion compensation prediction frame memory 39 is configured by dedicated hardware (for example, a semiconductor integrated circuit on which a CPU is mounted, a one-chip microcomputer, or the like). However, when the moving picture decoding apparatus is configured by a computer, the variable length decoding unit 31, the inverse quantization / inverse conversion unit 32, the changeover switch 33, the intra prediction unit 34, the motion compensation unit 35, the addition unit 36, and the loop A program describing the processing contents of the filter unit 38 is stored in the memory of a computer, and the CPU of the computer is stored in the memory. It is also possible to run the program.
FIG. 4 is a flowchart showing the processing contents (moving image decoding method) of the moving image decoding apparatus according to Embodiment 1 of the present invention.
 次に動作について説明する。
 この実施の形態1では、映像の各フレーム画像を入力画像として、符号化済みの近傍画素からのイントラ予測又は近接フレーム間での動き補償予測を実施して、得られた予測差分信号に対して直交変換・量子化による圧縮処理を施し、その後、可変長符号化を行って符号化ビットストリームを生成する動画像符号化装置と、その動画像符号化装置から出力される符号化ビットストリームを復号する動画像復号装置について説明する。
Next, the operation will be described.
In the first embodiment, each frame image of a video is used as an input image, intra prediction from encoded neighboring pixels or motion compensation prediction between adjacent frames is performed, and an obtained prediction difference signal is obtained. A video encoding device that performs compression processing by orthogonal transform / quantization and then performs variable length encoding to generate an encoded bit stream, and decodes the encoded bit stream output from the video encoding device A moving picture decoding apparatus will be described.
 図1の動画像符号化装置は、映像信号の空間・時間方向の局所的な変化に適応して、映像信号を多様なサイズのブロックに分割して、フレーム内・フレーム間適応符号化を行うことを特徴としている。
 一般的に、映像信号は、空間・時間的に信号の複雑さが局所的に変化する特性を有している。空間的に見ると、ある映像フレーム上では、例えば、空や壁などのような比較的広い画像領域中で均一な信号特性を有する絵柄もあれば、人物や細かいテクスチャを含む絵画など、小さい画像領域内で複雑なテクスチャパターンを有する絵柄も混在することがある。
 時間的に見ても、空や壁は局所的に時間方向の絵柄の変化は小さいが、動く人物や物体は、その輪郭が時間的に剛体・非剛体の運動をするため、時間的な変化が大きい。
The moving picture coding apparatus in FIG. 1 performs intra-frame / inter-frame adaptive coding by dividing a video signal into blocks of various sizes in response to local changes in the spatial and temporal directions of the video signal. It is characterized by that.
In general, a video signal has a characteristic that the complexity of the signal changes locally in space and time. When viewed spatially, a small image, such as a picture with a uniform signal characteristic in a relatively wide image area such as the sky or a wall, or a picture containing a person or fine texture, on a video frame. A pattern having a complicated texture pattern in the region may be mixed.
Even when viewed temporally, the change in the pattern of the sky and the wall locally in the time direction is small, but because the outline of the moving person or object moves rigidly or non-rigidly in time, the temporal change Is big.
 符号化処理は、時間・空間的な予測によって、信号電力やエントロピーの小さい予測差分信号を生成して、全体の符号量を削減する処理を行うが、予測に用いるパラメータをできるだけ大きな画像信号領域に均一に適用できれば、当該パラメータの符号量を小さくすることができる。
 一方、時間的・空間的に変化の大きい画像信号パターンに対して、同一の予測パラメータを大きな画像領域に適用すると、予測の誤りが増えてしまうため、予測差分信号の符号量が増加してしまう。
 したがって、時間的・空間的に変化が大きい領域では、同一の予測パラメータを適用して予測処理を行うブロックサイズを小さくして、予測に用いるパラメータのデータ量を増やし、予測差分信号の電力・エントロピーを低減する方が望ましい。
In the encoding process, a prediction difference signal with small signal power and entropy is generated by temporal and spatial prediction to reduce the overall code amount. However, the parameters used for the prediction are set as large as possible in the image signal region. If it can be applied uniformly, the code amount of the parameter can be reduced.
On the other hand, if the same prediction parameter is applied to a large image region with respect to an image signal pattern having a large temporal and spatial change, the number of prediction differential signals increases because prediction errors increase. .
Therefore, in a region where the temporal and spatial changes are large, the block size for performing the prediction process by applying the same prediction parameter is reduced, the amount of parameter data used for prediction is increased, and the power and entropy of the prediction difference signal are increased. It is desirable to reduce
 この実施の形態1では、このような映像信号の一般的な性質に適応した符号化を行うため、最初に所定の最大ブロックサイズから予測処理等を開始し、階層的に映像信号の領域を分割し、分割した領域毎に予測処理や、その予測差分の符号化処理を適応化させる構成をとるようにしている。 In the first embodiment, in order to perform coding adapted to the general characteristics of such a video signal, first, prediction processing or the like is started from a predetermined maximum block size, and the video signal region is divided hierarchically. In addition, the prediction process and the encoding process of the prediction difference are adapted for each divided area.
 図1の動画像符号化装置が処理対象とする映像信号フォーマットは、輝度信号と2つの色差信号からなるYUV信号や、ディジタル撮像素子から出力されるRGB信号等の任意の色空間のカラー映像信号のほか、モノクロ画像信号や赤外線画像信号など、映像フレームが水平・垂直2次元のディジタルサンプル(画素)列から構成される任意の映像信号とする。
 ただし、各画素の階調は、8ビットでもよいし、10ビットや12ビットなどの階調でもよい。
The video signal format to be processed by the moving image encoding apparatus of FIG. 1 is a color video signal in an arbitrary color space such as a YUV signal composed of a luminance signal and two color difference signals, or an RGB signal output from a digital image sensor. In addition to the above, it is assumed that the video frame is an arbitrary video signal including a horizontal / vertical two-dimensional digital sample (pixel) sequence, such as a monochrome image signal or an infrared image signal.
However, the gradation of each pixel may be 8 bits, or a gradation of 10 bits or 12 bits.
 以下の説明では、便宜上、特に断らない限り、入力画像の映像信号はYUV信号であるとし、かつ、2つの色差成分U,Vが輝度成分Yに対して、サブサンプルされた4:2:0フォーマットの信号を扱う場合について述べる。
 また、色差信号のフォーマットはYUV信号の4:2:0フォーマット以外でもよく、YUV信号の4:2:2フォーマットや4:4:4フォーマット、あるいはRGB信号などでもよい。
 また、映像信号の各フレームに対応する処理データ単位を「ピクチャ」と称する。
 なお、「ピクチャ」はフレーム単位に符号化する場合はフレーム信号を表し、フィールド単位に符号化する場合はフィールド信号を表す。
In the following description, for convenience, unless otherwise specified, it is assumed that the video signal of the input image is a YUV signal, and the two color difference components U and V are subsampled with respect to the luminance component Y 4: 2: 0. The case of handling format signals will be described.
Also, the format of the color difference signal may be other than the 4: 2: 0 format of the YUV signal, or may be the 4: 2: 2 format, 4: 4: 4 format of the YUV signal, or the RGB signal.
A processing data unit corresponding to each frame of the video signal is referred to as a “picture”.
Note that “picture” represents a frame signal when encoded in frame units, and represents a field signal when encoded in field units.
 最初に、図1の動画像符号化装置の処理内容を説明する。
 まず、符号化制御部2は、符号化対象となるピクチャ(カレントピクチャ)のスライス分割状態を決めると共に、ピクチャの符号化に用いる最大符号化ブロックのサイズと、最大符号化ブロックを階層分割する階層数の上限を決定する(図2のステップST1)。
 最大符号化ブロックのサイズの決め方としては、例えば、入力画像の映像信号の解像度に応じて、全てのピクチャに対して同一のサイズを定めてもよいし、入力画像の映像信号の局所的な動きの複雑さの違いをパラメータとして定量化して、動きの激しいピクチャには、小さいサイズを定める一方、動きが少ないピクチャには、大きいサイズを定めるようにしてもよい。
First, the processing contents of the moving picture encoding apparatus in FIG. 1 will be described.
First, the encoding control unit 2 determines the slice division state of a picture to be encoded (current picture), and also determines the size of the maximum encoding block used for encoding the picture and the hierarchy for dividing the maximum encoding block into layers. The upper limit of the number is determined (step ST1 in FIG. 2).
As a method of determining the size of the maximum coding block, for example, the same size may be determined for all the pictures according to the resolution of the video signal of the input image, or the local motion of the video signal of the input image The size difference may be quantified as a parameter, and a small size may be determined for a picture with high motion, while a large size may be determined for a picture with little motion.
 分割階層数の上限の決め方としては、例えば、入力画像の映像信号の解像度に応じて、全てのピクチャに対して同一の階層数を定める方法や、入力画像の映像信号の動きが激しい場合には、階層数を深くして、より細かい動きが検出できるように設定し、動きが少ない場合には、階層数を抑えるように設定する方法などがある。
 なお、上記最大符号化ブロックのサイズと、最大符号化ブロックを階層分割する階層数の上限はシーケンスレベルヘッダなどに符号化してもよいし、符号化せずに動画像復号装置側も同一の決定処理を行うようにしてもよい。前者はヘッダ情報の符号量が増加するものの、動画像復号装置側で上記決定処理を行わずに済むため、動画像復号装置の処理負荷を抑えることができる上、動画像符号化装置側で最適な値を探索して送ることができる。後者は反対に、動画像復号装置側で上記決定処理を行うため、動画像復号装置の処理負荷が増加するものの、ヘッダ情報の符号量は増加しない。
For example, the upper limit of the number of division layers can be determined by, for example, determining the same number of layers for all pictures according to the resolution of the video signal of the input image, or when the motion of the video signal of the input image is severe There is a method in which the number of hierarchies is increased so that finer movements can be detected, and when there are few movements, the number of hierarchies is set to be suppressed.
Note that the size of the maximum coding block and the upper limit of the number of layers into which the maximum coding block is divided may be encoded in a sequence level header or the like. Processing may be performed. Although the former increases the code amount of the header information, it is not necessary to perform the determination process on the video decoding device side, so the processing load on the video decoding device can be suppressed and the optimum on the video encoding device side. You can search for and send a new value. On the contrary, since the determination process is performed on the video decoding device side, the processing load on the video decoding device increases, but the code amount of the header information does not increase.
 また、符号化制御部2は、利用可能な1以上の符号化モードの中から、階層的に分割される各々の符号化ブロックに対応する符号化モードを選択する(ステップST2)。
 即ち、符号化制御部2は、最大符号化ブロックサイズの画像領域毎に、先に定めた分割階層数の上限に至るまで、階層的に符号化ブロックサイズを有する符号化ブロックに分割して、各々の符号化ブロックに対する符号化モードを決定する。
 符号化モードには、1つないし複数のイントラ符号化モード(総称して「INTRA」と称する)と、1つないし複数のインター符号化モード(総称して、「INTER」と称する)とがあり、符号化制御部2は、当該ピクチャで利用可能な全ての符号化モード、または、そのサブセットの中から、各々の符号化ブロックに対応する符号化モードを選択する。
Also, the encoding control unit 2 selects an encoding mode corresponding to each encoding block divided hierarchically from one or more available encoding modes (step ST2).
That is, the encoding control unit 2 divides the image area of the maximum encoding block size into encoded blocks having the encoding block size hierarchically until reaching the upper limit of the number of division layers defined above. A coding mode for each coding block is determined.
There are one or more intra coding modes (collectively referred to as “INTRA”) and one or more inter coding modes (collectively referred to as “INTER”). The coding control unit 2 selects a coding mode corresponding to each coding block from all coding modes available for the picture or a subset thereof.
 ただし、後述するブロック分割部1により階層的に分割される各々の符号化ブロックは、さらに予測処理を行う単位である1つないし複数の予測ブロックに分割され、予測ブロックの分割状態も符号化モードの中に情報として含まれる。即ち、符号化モードは、どのような予測ブロック分割を持つイントラまたはインター符号化モードかを識別するインデックスである。
 符号化制御部2による符号化モードの選択方法は、公知の技術であるため詳細な説明を省略するが、例えば、利用可能な任意の符号化モードを用いて、符号化ブロックに対する符号化処理を実施して符号化効率を検証し、利用可能な複数の符号化モードの中で、最も符号化効率がよい符号化モードを選択する方法などがある。
However, each coding block that is hierarchically divided by the block division unit 1 to be described later is further divided into one or a plurality of prediction blocks, which are units for performing prediction processing, and the division state of the prediction block is also coded mode. Is included as information. That is, the coding mode is an index for identifying what kind of prediction block division the intra or inter coding mode is.
Since the encoding mode selection method by the encoding control unit 2 is a known technique, detailed description thereof is omitted. For example, an encoding process for an encoding block is performed using any available encoding mode. There is a method in which coding efficiency is verified by performing and a coding mode having the best coding efficiency is selected from among a plurality of available coding modes.
 また、符号化制御部2は、各々の符号化ブロック毎に、差分画像が圧縮される際に用いられる量子化パラメータ及び直交変換ブロック分割状態を決定するとともに、予測処理が実施される際に用いられる予測パラメータ(イントラ予測パラメータ又はインター予測パラメータ)を決定する。
 ただし、符号化ブロックがさらに予測処理を行う予測ブロック単位に分割される場合は、予測ブロック毎に予測パラメータ(イントラ予測パラメータ又はインター予測パラメータ)を選択することができる。
 さらに、符号化モードがイントラ符号化モードである符号化ブロックにおいては、詳細は後述するが、イントラ予測処理を行う際に予測ブロックに隣接する符号化済みの画素を用いることから、予測ブロック単位に符号化を行う必要があるため、選択可能な変換ブロックサイズは予測ブロックのサイズ以下に制限される。
In addition, the encoding control unit 2 determines a quantization parameter and an orthogonal transform block division state that are used when the difference image is compressed for each encoding block, and is used when the prediction process is performed. Prediction parameter (intra prediction parameter or inter prediction parameter) is determined.
However, when the encoded block is further divided into prediction block units for which prediction processing is performed, a prediction parameter (intra prediction parameter or inter prediction parameter) can be selected for each prediction block.
Furthermore, in a coding block whose coding mode is the intra coding mode, the details will be described later. Since it is necessary to perform encoding, the selectable transform block size is limited to the size of the prediction block or less.
 符号化制御部2は、量子化パラメータ及び変換ブロックサイズを含む予測差分符号化パラメータを変換・量子化部7、逆量子化・逆変換部8及び可変長符号化部13に出力する。
 また、符号化制御部2は、イントラ予測パラメータを必要に応じてイントラ予測部4に出力する。
 また、符号化制御部2は、インター予測パラメータを必要に応じて動き補償予測部5に出力する。
The encoding control unit 2 outputs the prediction difference encoding parameter including the quantization parameter and the transform block size to the transform / quantization unit 7, the inverse quantization / inverse transform unit 8, and the variable length coding unit 13.
Also, the encoding control unit 2 outputs intra prediction parameters to the intra prediction unit 4 as necessary.
Also, the encoding control unit 2 outputs inter prediction parameters to the motion compensation prediction unit 5 as necessary.
 スライス分割部14は、入力画像として映像信号を入力すると、その入力画像を符号化制御部2により決定されたスライス分割情報にしたがって1以上の部分画像であるスライスに分割する。
 ブロック分割部1は、スライス分割部14から各スライスを入力する毎に、そのスライスを符号化制御部2により決定された最大符号化ブロックサイズに分割し、さらに、分割した最大符号化ブロックを符号化制御部2により決定された符号化ブロックへ階層的に分割して、その符号化ブロックを出力する。
When a video signal is input as an input image, the slice dividing unit 14 divides the input image into slices that are one or more partial images according to the slice division information determined by the encoding control unit 2.
Each time the slice dividing unit 1 inputs each slice from the slice dividing unit 14, the slice dividing unit 1 divides the slice into the maximum coding block size determined by the coding control unit 2, and further encodes the divided maximum coding block. The coding block is hierarchically divided into coding blocks determined by the coding control unit 2, and the coding blocks are output.
 ここで、図5は最大符号化ブロックが階層的に複数の符号化ブロックに分割される例を示す説明図である。
 図5において、最大符号化ブロックは、「第0階層」と記されている輝度成分が(L,M)のサイズを有する符号化ブロックである。
 最大符号化ブロックを出発点として、4分木構造で別途定める所定の深さまで、階層的に分割を行うことによって符号化ブロックを得るようにしている。
 深さnにおいては、符号化ブロックはサイズ(L,M)の画像領域である。
 ただし、LとMは、同じであってもよいし、異なっていてもよいが、図5では、L=Mのケースを示している。
Here, FIG. 5 is an explanatory diagram showing an example in which the maximum coding block is hierarchically divided into a plurality of coding blocks.
In FIG. 5, the maximum coding block is a coding block whose luminance component described as “0th layer” has a size of (L 0 , M 0 ).
Starting from the maximum encoding block, the encoding block is obtained by performing hierarchical division to a predetermined depth separately defined by a quadtree structure.
At depth n, the coding block is an image area of size (L n , M n ).
However, L n and M n may be the same or different, but FIG. 5 shows a case of L n = M n .
 以降、符号化制御部2により決定される符号化ブロックサイズは、符号化ブロックの輝度成分におけるサイズ(L,M)と定義する。
 4分木分割を行うため、常に、(Ln+1,Mn+1)=(L/2,M/2)が成立する。
 なお、RGB信号など、全ての色成分が同一サンプル数を有するカラー映像信号(4:4:4フォーマット)では、全ての色成分のサイズが(L,M)になるが、4:2:0フォーマットを扱う場合、対応する色差成分の符号化ブロックサイズは(L/2,M/2)になる。
Hereinafter, the coding block size determined by the coding control unit 2 is defined as the size (L n , M n ) in the luminance component of the coding block.
Since quadtree partitioning is performed, (L n + 1 , M n + 1 ) = (L n / 2, M n / 2) always holds.
Note that in a color video signal (4: 4: 4 format) in which all color components have the same number of samples, such as RGB signals, the size of all color components is (L n , M n ), but 4: 2. : When the 0 format is handled, the encoding block size of the corresponding color difference component is (L n / 2, M n / 2).
 以降、第n階層の符号化ブロックをBで表し、符号化ブロックBで選択可能な符号化モードをm(B)で表すものとする。
 複数の色成分からなるカラー映像信号の場合、符号化モードm(B)は、色成分毎に、それぞれ個別のモードを用いるように構成されてもよいし、全ての色成分に対し共通のモードを用いるように構成されてもよい。以降、特に断らない限り、YUV信号、4:2:0フォーマットの符号化ブロックの輝度成分に対する符号化モードを指すものとして説明を行う。
Hereinafter, the coding block of the n hierarchy expressed in B n, denote the encoding modes selectable by the coding block B n with m (B n).
In the case of a color video signal composed of a plurality of color components, the encoding mode m (B n ) may be configured to use an individual mode for each color component, or common to all color components. It may be configured to use a mode. Hereinafter, unless otherwise specified, description will be made assuming that it indicates a coding mode for a luminance component of a coding block of a YUV signal and 4: 2: 0 format.
 符号化ブロックBは、図5に示すように、ブロック分割部1によって、予測処理単位を表す1つないし複数の予測ブロックに分割される。
 以降、符号化ブロックBに属する予測ブロックをP (iは、第n階層における予測ブロック番号)と表記する。図5にはP とP の例を示している。
 符号化ブロックB内の予測ブロックの分割が、どのようになされているかは、符号化モードm(B)の中に情報として含まれる。
 予測ブロックP は、全て符号化モードm(B)に従って予測処理が行われるが、予測ブロックP 毎に、個別の予測パラメータ(イントラ予測パラメータ又はインター予測パラメータ)を選択することができる。
As shown in FIG. 5, the encoded block B n is divided by the block dividing unit 1 into one or a plurality of prediction blocks representing a prediction processing unit.
Hereinafter, a prediction block belonging to the coding block B n is denoted as P i n (i is a prediction block number in the n-th layer). FIG. 5 shows an example of P 0 0 and P 1 0 .
How the prediction block is divided in the coding block Bn is included as information in the coding mode m ( Bn ).
All the prediction blocks P i n are subjected to prediction processing according to the encoding mode m (B n ), and it is possible to select individual prediction parameters (intra prediction parameters or inter prediction parameters) for each prediction block P i n. it can.
 符号化制御部2は、最大符号化ブロックに対して、例えば、図6に示すようなブロック分割状態を生成して、符号化ブロックを特定する。
 図6(a)の点線で囲まれた矩形が各符号化ブロックを表し、各符号化ブロック内にある斜線で塗られたブロックが各予測ブロックの分割状態を表している。
 図6(b)は、図6(a)の例について、階層分割によって符号化モードm(B)が割り当てられる状況を4分木グラフで示したものである。図6(b)の□で囲まれているノードは、符号化モードm(B)が割り当てられたノード(符号化ブロック)である。
 この4分木グラフの情報は符号化モードm(B)と共に符号化制御部2から可変長符号化部13に出力されて、ビットストリームに多重化される。
For example, the encoding control unit 2 generates a block division state as illustrated in FIG. 6 for the maximum encoding block, and identifies the encoding block.
A rectangle surrounded by a dotted line in FIG. 6A represents each coding block, and a block painted with diagonal lines in each coding block represents a division state of each prediction block.
FIG. 6B shows, in a quadtree graph, a situation in which the encoding mode m (B n ) is assigned by hierarchical division in the example of FIG. 6A. Nodes surrounded by squares in FIG. 6B are nodes (encoding blocks) to which the encoding mode m (B n ) is assigned.
Information of the quadtree graph is output from the encoding control unit 2 to the variable length encoding unit 13 together with the encoding mode m (B n ), and is multiplexed into the bit stream.
 切換スイッチ3は、符号化制御部2により決定された符号化モードm(B)がイントラ符号化モードである場合(m(B)∈INTRAの場合)、ブロック分割部1から出力された符号化ブロックBをイントラ予測部4に出力する。
 一方、符号化制御部2により決定された符号化モードm(B)がインター符号化モードである場合(m(B)∈INTERの場合)、ブロック分割部1から出力された符号化ブロックBを動き補償予測部5に出力する。
The changeover switch 3 is output from the block dividing unit 1 when the encoding mode m (B n ) determined by the encoding control unit 2 is an intra encoding mode (when m (B n ) ∈INTRA). The encoded block B n is output to the intra prediction unit 4.
On the other hand, when the encoding mode m (B n ) determined by the encoding control unit 2 is the inter encoding mode (when m (B n ) εINTER), the encoded block output from the block dividing unit 1 B n is output to the motion compensation prediction unit 5.
 イントラ予測部4は、符号化制御部2により決定された符号化モードm(B)がイントラ符号化モードであり(m(B)∈INTRAの場合)、切換スイッチ3から符号化ブロックBを受けると(ステップST3)、イントラ予測用メモリ10に格納されている局所復号画像を参照しながら、符号化制御部2により決定されたイントラ予測パラメータを用いて、その符号化ブロックB内の各予測ブロックP に対するイントラ予測処理を実施して、イントラ予測画像PINTRAi を生成する(ステップST4)。
 なお、動画像復号装置がイントラ予測画像PINTRAi と全く同じイントラ予測画像を生成する必要があるため、イントラ予測画像PINTRAi の生成に用いられたイントラ予測パラメータは、符号化制御部2から可変長符号化部13に出力されて、ビットストリームに多重化される。
 イントラ予測部4の処理内容の詳細は後述する。
In the intra prediction unit 4, the coding mode m (B n ) determined by the coding control unit 2 is the intra coding mode (when m (B n ) ∈INTRA), and the coding block B is changed from the changeover switch 3 to the coding block B. n (step ST3), using the intra prediction parameters determined by the encoding control unit 2 while referring to the local decoded image stored in the intra prediction memory 10, the encoding block B n and implementing intra prediction process for each of the prediction block P i n in, it generates an intra prediction image P INTRAi n (step ST4).
Incidentally, since it is necessary to video decoding device generates exactly the same intra prediction image and the intra prediction image P INTRAi n, intra prediction parameters used for generating the intra prediction image P INTRAi n is from encoding control unit 2 The data is output to the variable length encoding unit 13 and multiplexed into the bit stream.
Details of processing contents of the intra prediction unit 4 will be described later.
 動き補償予測部5は、符号化制御部2により決定された符号化モードm(B)がインター符号化モードであり(m(B)∈INTERの場合)、切換スイッチ3から符号化ブロックBを受けると(ステップST3)、その符号化ブロックB内の各予測ブロックP と動き補償予測フレームメモリ12に格納されているフィルタ処理後の局所復号画像を比較して動きベクトルを探索し、その動きベクトルと符号化制御部2により決定されたインター予測パラメータを用いて、その符号化ブロックB内の各予測ブロックP に対するインター予測処理を実施して、インター予測画像PINTERi を生成する(ステップST5)。
 なお、動画像復号装置がインター予測画像PINTERi と全く同じインター予測画像を生成する必要があるため、インター予測画像PINTERi の生成に用いられたインター予測パラメータは、符号化制御部2から可変長符号化部13に出力されて、ビットストリームに多重化される。
 また、動き補償予測部5により探索された動きベクトルも可変長符号化部13に出力されて、ビットストリームに多重化される。
The motion-compensated prediction unit 5 has the coding mode m (B n ) determined by the coding control unit 2 in the inter coding mode (when m (B n ) ∈ INTER), and the coding block is switched from the changeover switch 3 to the coding block. Upon receiving the B n (step ST3), the motion vector by comparing the locally decoded image after the filtering process stored in the prediction block P i n and the motion compensated prediction frame memory 12 of the encoding block B n Using the motion vector and the inter prediction parameter determined by the encoding control unit 2, the inter prediction process for each prediction block P i n in the encoding block B n is performed, and the inter prediction image P generating a INTERi n (step ST5).
Incidentally, since it is necessary to video decoding device generates exactly the same inter prediction image and the inter-predicted image P INTERi n, inter prediction parameters used for generating the inter prediction image P INTERi n is from encoding control unit 2 The data is output to the variable length encoding unit 13 and multiplexed into the bit stream.
In addition, the motion vector searched by the motion compensation prediction unit 5 is also output to the variable length encoding unit 13 and multiplexed into the bit stream.
 減算部6は、ブロック分割部1から符号化ブロックBを受けると、その符号化ブロックB内の予測ブロックP から、イントラ予測部4により生成されたイントラ予測画像PINTRAi 、または、動き補償予測部5により生成されたインター予測画像PINTERi のいずれか一方を減算して、その減算結果である差分画像を示す予測差分信号e を変換・量子化部7に出力する(ステップST6)。 Subtraction unit 6, upon receiving the encoded block B n from the block dividing unit 1 from its prediction block P i n the coded block B n, the intra prediction image P INTRAi n generated by the intra prediction unit 4 or, , by subtracting one of the inter prediction image P INTERi n generated by the motion compensation prediction unit 5, and outputs the prediction difference signal e i n representing a difference image is the subtraction result to the transform and quantization unit 7 (Step ST6).
 変換・量子化部7は、減算部6から予測差分信号e を受けると、符号化制御部2により決定された予測差分符号化パラメータに含まれる直交変換ブロック分割情報を参照して、その予測差分信号e に対する直交変換処理(例えば、DCT(離散コサイン変換)やDST(離散サイン変換)、予め特定の学習系列に対して基底設計がなされているKL変換等の直交変換処理)を直交変換ブロック単位に実施して、変換係数を算出する。
 また、変換・量子化部7は、その予測差分符号化パラメータに含まれる量子化パラメータを参照して、その直交変換ブロック単位の変換係数を量子化し、量子化後の変換係数である圧縮データを逆量子化・逆変換部8及び可変長符号化部13に出力する(ステップST7)。このとき、上記量子化パラメータから算出される量子化ステップサイズを変換係数毎にスケーリングする量子化マトリクスを用いて量子化処理を実施するようにしてもよい。
When the transform / quantization unit 7 receives the prediction difference signal e i n from the subtraction unit 6, the transform / quantization unit 7 refers to the orthogonal transform block division information included in the prediction difference encoding parameter determined by the encoding control unit 2, and orthogonal transform processing for the predicted differential signal e i n (e.g., DCT (discrete cosine transform) or DST (discrete sine transform), the orthogonal transform for KL conversion and the base design have been made in advance to the particular learning sequence) This is performed for each orthogonal transform block, and transform coefficients are calculated.
Further, the transform / quantization unit 7 refers to the quantization parameter included in the prediction differential encoding parameter, quantizes the transform coefficient of the orthogonal transform block unit, and compresses the compressed data that is the quantized transform coefficient. The data is output to the inverse quantization / inverse transform unit 8 and the variable length coding unit 13 (step ST7). At this time, the quantization process may be performed using a quantization matrix that scales the quantization step size calculated from the quantization parameter for each transform coefficient.
 量子化マトリクスは、各直交変換サイズで色信号や符号化モード(イントラ符号化かインター符号化か)毎に独立しているマトリクスを使用することができ、初期値として予め動画像符号化装置及び動画像復号装置で共通に用意されている量子化マトリクスや既に符号化された量子化マトリクスの中から選択するか、新しい量子化マトリクスを用いるかをそれぞれ選択することができる。
 したがって、変換・量子化部7は、各直交変換サイズに対して色信号や符号化モード毎に、新しい量子化マトリクスを用いるか否かを示すフラグ情報を符号化すべき量子化マトリクスパラメータに設定する。
 さらに、新しい量子化マトリクスを用いる場合には、図10に示すような量子化マトリクスの各スケーリング値を符号化すべき量子化マトリクスパラメータに設定する。一方、新しい量子化マトリクスを用いない場合には、初期値として予め動画像符号化装置及び動画像復号装置で共通に用意されている量子化マトリクス、または、既に符号化された量子化マトリクスの中から、使用するマトリクスを特定するインデックスを符号化すべき量子化マトリクスパラメータに設定する。ただし、参照可能な既に符号化された量子化マトリクスが存在しない場合、予め動画像符号化装置及び動画像復号装置で共通に用意されている量子化マトリクスのみ選択可能となる。
 そして、変換・量子化部7は、設定した量子化マトリクスパラメータを可変長符号化部13に出力する。
As the quantization matrix, an independent matrix can be used for each color signal and coding mode (intra coding or inter coding) at each orthogonal transform size. It is possible to select either a quantization matrix that is commonly prepared in the moving picture decoding apparatus or an already encoded quantization matrix or a new quantization matrix.
Accordingly, the transform / quantization unit 7 sets flag information indicating whether or not to use a new quantization matrix for each orthogonal transform size for each color signal and coding mode, in a quantization matrix parameter to be encoded. .
Furthermore, when a new quantization matrix is used, each scaling value of the quantization matrix as shown in FIG. 10 is set as a quantization matrix parameter to be encoded. On the other hand, when a new quantization matrix is not used, a quantization matrix prepared in advance by the moving picture coding apparatus and the moving picture decoding apparatus as an initial value or a quantization matrix that has already been coded. Thus, an index for specifying a matrix to be used is set as a quantization matrix parameter to be encoded. However, when there is no already-encoded quantization matrix that can be referred to, only a quantization matrix prepared in advance by the moving picture coding apparatus and the moving picture decoding apparatus can be selected.
Then, the transform / quantization unit 7 outputs the set quantization matrix parameter to the variable length coding unit 13.
 逆量子化・逆変換部8は、変換・量子化部7から圧縮データを受けると、符号化制御部2により決定された予測差分符号化パラメータに含まれる量子化パラメータ及び直交変換ブロック分割情報を参照して、直交変換ブロック単位にその圧縮データを逆量子化する。
 変換・量子化部7が量子化処理に量子化マトリクスを用いている場合には、逆量子化処理時においても、その量子化マトリクスを参照して、対応した逆量子化処理を実施する。
 また、逆量子化・逆変換部8は、直交変換ブロック単位に逆量子化後の圧縮データである変換係数に対する逆直交変換処理(例えば、逆DCT、逆DST、逆KL変換など)を実施して、減算部6から出力された予測差分信号e に相当する局所復号予測差分信号を算出して加算部9に出力する(ステップST8)。
When the inverse quantization / inverse transform unit 8 receives the compressed data from the transform / quantization unit 7, the inverse quantization / inverse transform unit 8 displays the quantization parameter and the orthogonal transform block division information included in the prediction difference coding parameter determined by the coding control unit 2. With reference, the compressed data is inversely quantized in units of orthogonal transform blocks.
When the transform / quantization unit 7 uses a quantization matrix for the quantization process, the corresponding inverse quantization process is performed with reference to the quantization matrix even during the inverse quantization process.
Further, the inverse quantization / inverse transform unit 8 performs inverse orthogonal transform processing (for example, inverse DCT, inverse DST, inverse KL transform, etc.) on transform coefficients that are compressed data after inverse quantization in units of orthogonal transform blocks. Then, a local decoded prediction difference signal corresponding to the prediction difference signal e i n output from the subtraction unit 6 is calculated and output to the addition unit 9 (step ST8).
 加算部9は、逆量子化・逆変換部8から局所復号予測差分信号を受けると、その局所復号予測差分信号と、イントラ予測部4により生成されたイントラ予測画像PINTRAi 、または、動き補償予測部5により生成されたインター予測画像PINTERi のいずれか一方を加算することで、局所復号画像を算出する(ステップST9)。
 なお、加算部9は、その局所復号画像をループフィルタ部11に出力するとともに、その局所復号画像をイントラ予測用メモリ10に格納する。
 この局所復号画像が、以降のイントラ予測処理の際に用いられる符号化済みの画像信号になる。
Upon receiving the local decoded prediction difference signal from the inverse quantization / inverse transform unit 8, the adding unit 9 receives the local decoded prediction difference signal and the intra predicted image P INTRAi n generated by the intra prediction unit 4 or motion compensation. A local decoded image is calculated by adding one of the inter predicted images P INTERIn n generated by the prediction unit 5 (step ST9).
The adding unit 9 outputs the locally decoded image to the loop filter unit 11 and stores the locally decoded image in the intra prediction memory 10.
This locally decoded image becomes an encoded image signal used in the subsequent intra prediction processing.
 ループフィルタ部11は、加算部9から局所復号画像を受けると、その局所復号画像に対して、所定のフィルタ処理を実施して、フィルタ処理後の局所復号画像を動き補償予測フレームメモリ12に格納する(ステップST10)。
 具体的には、直交変換ブロックの境界や予測ブロックの境界に発生する歪みを低減するフィルタ(デブロッキングフィルタ)処理、画素単位に適応的にオフセットを加算する(画素適応オフセット)処理、ウィーナフィルタ等の線形フィルタを適応的に切り替えてフィルタ処理する適応フィルタ処理などを行う。
When the loop filter unit 11 receives the local decoded image from the adding unit 9, the loop filter unit 11 performs a predetermined filter process on the local decoded image, and stores the filtered local decoded image in the motion compensated prediction frame memory 12. (Step ST10).
Specifically, filter (deblocking filter) processing that reduces distortion occurring at the boundaries of orthogonal transform blocks and prediction blocks, processing for adaptively adding an offset in units of pixels (pixel adaptive offset), Wiener filter, etc. The adaptive filter processing that performs the filter processing by adaptively switching the linear filter is performed.
 ただし、ループフィルタ部11は、上記のデブロッキングフィルタ処理、画素適応オフセット処理、適応フィルタ処理のそれぞれについて、処理を行うか否かを決定し、各処理の有効フラグをシーケンスレベルヘッダの一部及びスライスレベルヘッダの一部として可変長符号化部13に出力する。なお、上記のフィルタ処理を複数使用する際は、各フィルタ処理を順番に実施する。図11は複数のフィルタ処理を用いる場合のループフィルタ部11の構成例を示している。
 一般に使用するフィルタ処理の種類が多いほど、画像品質は向上するが、一方で処理負荷は高くなる。即ち、画像品質と処理負荷はトレードオフの関係にある。また、各フィルタ処理の画像品質改善効果はフィルタ処理対象画像の特性によって異なる。したがって、動画像符号化装置が許容する処理負荷や符号化処理対象画像の特性にしたがって使用するフィルタ処理を決めればよい。
However, the loop filter unit 11 determines whether or not to perform processing for each of the above deblocking filter processing, pixel adaptive offset processing, and adaptive filter processing, and sets a valid flag of each processing as a part of the sequence level header and The data is output to the variable length coding unit 13 as a part of the slice level header. When a plurality of the above filter processes are used, each filter process is performed in order. FIG. 11 shows a configuration example of the loop filter unit 11 when a plurality of filter processes are used.
Generally, the more types of filter processing that are used, the better the image quality, but the higher the processing load. That is, image quality and processing load are in a trade-off relationship. In addition, the image quality improvement effect of each filter process varies depending on the characteristics of the image to be filtered. Therefore, the filter processing to be used may be determined according to the processing load allowed by the moving image encoding device and the characteristics of the encoding target image.
 ここで、デブロッキングフィルタ処理では、ブロック境界にかけるフィルタ強度の選択に用いる各種パラメータを初期値から変更することができる。変更する場合には、そのパラメータをヘッダ情報として可変長符号化部13に出力する。 Here, in the deblocking filter process, various parameters used for selecting the filter strength applied to the block boundary can be changed from the initial values. When changing, the parameter is output to the variable length coding unit 13 as header information.
 画素適応オフセット処理では、最初に、画像を複数のブロックに分割し、そのブロック単位に、オフセット処理を行わない場合もクラス分類手法の一つとして定義して、予め用意している複数のクラス分類手法の中から、1つのクラス分類手法を選択する。
 次に、選択したクラス分類手法によってブロック内の各画素をクラス分類し、クラス毎に符号化歪みを補償するオフセット値を算出する。
 最後に、局所復号画像の輝度値に対して、そのオフセット値を加算する処理を行うことで局所復号画像の画像品質を改善する。
In the pixel adaptive offset process, first, an image is divided into a plurality of blocks, and when the offset process is not performed for each block, it is defined as one of the class classification methods, and a plurality of class classifications prepared in advance are provided. One classification method is selected from the methods.
Next, each pixel in the block is classified by the selected class classification method, and an offset value for compensating the coding distortion is calculated for each class.
Finally, the image quality of the locally decoded image is improved by performing a process of adding the offset value to the luminance value of the locally decoded image.
 クラス分類手法としては、局所復号画像の輝度値の大きさで分類する手法(BO手法と呼ぶ)や、エッジの方向毎に各画素の周囲の状況(エッジ部か否か等)に応じて分類する手法(EO手法と呼ぶ)がある。
 これらの手法は、予め動画像符号化装置及び動画像復号装置で共通に用意されており、例えば図14に示すように、オフセット処理を行わない場合もクラス分類手法の一つとして定義して、これらの手法のうち、どの手法でクラス分類を行うかを示すインデックスを上記ブロック単位に選択する。
As a classifying method, a method of classifying by the size of the luminance value of the locally decoded image (referred to as a BO method), or a classification according to the situation around each pixel (whether it is an edge portion or the like) for each edge direction. There is a technique (referred to as EO technique).
These methods are prepared in advance by the moving image encoding device and the moving image decoding device in advance. For example, as shown in FIG. 14, when no offset processing is performed, it is defined as one of the class classification methods, Among these methods, an index indicating which method is used for class classification is selected for each block.
 したがって、画素適応オフセット処理は、ブロックの分割情報、ブロック単位のクラス分類手法を示すインデックス、ブロック単位のオフセット情報をヘッダ情報として可変長符号化部13に出力する。 Therefore, the pixel adaptive offset processing outputs the block division information, the index indicating the block class classification method, and the block offset information to the variable length coding unit 13 as header information.
 また、適応フィルタ処理では、局所復号画像を所定の手法でクラス分類し、各クラスに属する領域(局所復号画像)毎に、重畳されている歪みを補償するフィルタを設計し、そのフィルタを用いて、当該局所復号画像のフィルタ処理を実施する。
 そして、クラス毎に設計したフィルタをヘッダ情報として可変長符号化部13に出力する。
 ここで、クラス分類手法としては、画像を空間的に等間隔に区切る簡易な手法や、ブロック単位に画像の局所的な特性(分散など)に応じて分類する手法がある。また、適応フィルタ処理で使用するクラス数は、予め動画像符号化装置及び動画像復号装置で共通の値に設定してもよいし、符号化すべきパラメータの一つとしてもよい。
 前者と比較して後者の方が、使用するクラス数を自由に設定することができるため、画像品質改善効果が上がるが、一方でクラス数を符号化するために、その分の符号量が増加する。
Also, in adaptive filter processing, local decoded images are classified by a predetermined method, and a filter that compensates for superimposed distortion is designed for each region (local decoded image) belonging to each class. Then, the local decoded image is filtered.
Then, the filter designed for each class is output to the variable length encoding unit 13 as header information.
Here, as a class classification method, there are a simple method for spatially dividing an image at equal intervals, and a method for classifying an image according to local characteristics (dispersion, etc.) of the image in units of blocks. In addition, the number of classes used in the adaptive filter process may be set in advance to a common value in the video encoding device and the video decoding device, or may be one of the parameters to be encoded.
Compared to the former, the latter can set the number of classes to be used freely, so the image quality improvement effect will be improved, but on the other hand, the amount of code will be increased to encode the number of classes. To do.
 ステップST3~ST9の処理は、階層的に分割された全ての符号化ブロックBに対する処理が完了するまで繰り返し実施され、全ての符号化ブロックBに対する処理が完了すると、ステップST13の処理に移行する(ステップST11,ST12)。 The processes in steps ST3 to ST9 are repeated until the processes for all the coding blocks Bn divided hierarchically are completed, and when the processes for all the coding blocks Bn are completed, the process proceeds to the process of step ST13. (Steps ST11 and ST12).
 可変長符号化部13は、変換・量子化部7から出力された圧縮データと、符号化制御部2から出力された最大符号化ブロック内のブロック分割情報(図6(b)を例とする4分木情報)、符号化モードm(B)及び予測差分符号化パラメータと、符号化制御部2から出力されたイントラ予測パラメータ(符号化モードがイントラ符号化モードである場合)又はインター予測パラメータ(符号化モードがインター符号化モードである場合)と、動き補償予測部5から出力された動きベクトル(符号化モードがインター符号化モードである場合)とを可変長符号化して、それらの符号化結果を示す符号化データを生成する(ステップST13)。
 その際、量子化された直交変換係数である圧縮データの符号化手法として直交変換ブロックをさらにCoefficient Group(CG)と呼ばれる4×4画素単位のブロック(符号化サブブロック)に分割してCG単位に係数の符号化処理を実施する。図28に16×16画素の直交変換ブロックにおける係数の符号化順(スキャン順)を示す。非引用文献2では、このように4×4画素単位の16個のCGを右下のCGから順に符号化処理し、さらに各CGはCG内の16個の係数を右下の係数から順に符号化する。具体的には、まずCG内の16個の係数の中に有意(非零)係数が存在するか否かのフラグ情報を符号化し、次にCG内に有意(非零)係数が存在する場合のみCG内の各係数が有意(非零)係数であるかを上記順に符号化し、最後に有意(非零)係数に対してその係数値情報を順に符号化する。これをCG単位に上記順に行う。その際、有意(非零)係数がなるべく連続で発生するように偏るスキャン順とした方がエントロピー符号化による符号化効率を高めることができる。直交変換後の係数は左上に位置する直流成分をはじめとして左上に近い程低い周波数成分の低い係数を表すことから、図15に示す例のようにプログレッシブ映像では一般的に左上に近いほど有意(非零)係数が多く発生するために図28に示すように右下から順に符号化することで効率的に符号化できる。一方、シーケンスレベルヘッダのフィールド符号化か否かを示すフラグが有効である場合、すなわち入力信号をフィールド単位に符号化する場合は垂直方向の空間的相関が低下するために垂直方向の予測効率が低下し、予測差分信号e を直交変換した結果である変換係数についても垂直方向の周波数成分が多く発生するようになり図16に示す例のように有意(非零)係数発生分布がプログレッシブ映像よりも直交変換ブロックの左側に偏った形となる傾向がある。したがって、図28に示す符号化順では効率的に符号化できなくなってしまうため、例えば、図17に示す符号化順に切り替えるようにする。このようにすることで、有意(非零)係数の符号化を符号化順の後ろの方で連続して処理するようになり、エントロピー符号化による符号化効率を高めることができる。
 なお、上記では16×16画素の直交変換ブロックについて説明したが、32×32画素の直交変換ブロック等、16×16画素以外のブロックサイズにおいてもCG(符号化サブブロック)単位の符号化処理を実施するものとし、その符号化順は16×16画素の直交変換ブロックと同様にシーケンスレベルヘッダのフィールド符号化か否かを示すフラグが有効かどうかによって切り替えるようにする。
The variable length encoding unit 13 uses the compressed data output from the transform / quantization unit 7 and the block division information (FIG. 6B) in the maximum encoding block output from the encoding control unit 2 as an example. (Quadrant tree information), encoding mode m (B n ) and prediction differential encoding parameter, and intra prediction parameter output from the encoding control unit 2 (when the encoding mode is an intra encoding mode) or inter prediction The parameters (when the encoding mode is the inter encoding mode) and the motion vector (when the encoding mode is the inter encoding mode) output from the motion compensated prediction unit 5 are variable-length encoded, Encoded data indicating the encoding result is generated (step ST13).
At that time, as an encoding method of compressed data which is a quantized orthogonal transform coefficient, the orthogonal transform block is further divided into blocks of 4 × 4 pixel units (encoding sub-blocks) called “Coefficient Group (CG)”. The coefficient encoding process is performed on FIG. 28 shows the coding order (scan order) of the coefficients in the 16 × 16 pixel orthogonal transform block. In non-cited document 2, 16 CGs in units of 4 × 4 pixels are encoded in this order from the lower right CG, and each CG encodes 16 coefficients in the CG in order from the lower right coefficient. Turn into. Specifically, first, flag information indicating whether or not a significant (nonzero) coefficient exists in 16 coefficients in the CG is encoded, and then a significant (nonzero) coefficient exists in the CG. Only whether each coefficient in the CG is a significant (non-zero) coefficient is encoded in the above order, and finally, coefficient value information is encoded in order for a significant (non-zero) coefficient. This is performed in the above order in units of CG. In this case, the encoding efficiency by entropy encoding can be increased by using a biased scan order so that significant (non-zero) coefficients are generated as continuously as possible. Since the coefficient after the orthogonal transformation represents the lower coefficient of the lower frequency component as it approaches the upper left, starting with the DC component located at the upper left, generally, in the progressive video as shown in FIG. Since many non-zero) coefficients are generated, efficient encoding can be performed by encoding sequentially from the lower right as shown in FIG. On the other hand, when the flag indicating whether or not the field coding of the sequence level header is valid, that is, when the input signal is coded on a field basis, the vertical spatial efficiency decreases, so the vertical prediction efficiency is The frequency coefficient in the vertical direction also increases for the transform coefficient, which is a result of the orthogonal transformation of the prediction difference signal e i n , and the significant (non-zero) coefficient generation distribution is progressive as shown in FIG. There is a tendency that the shape is biased to the left side of the orthogonal transform block rather than the video. Accordingly, since the encoding cannot be efficiently performed in the encoding order shown in FIG. 28, for example, switching is performed in the encoding order shown in FIG. By doing so, encoding of significant (non-zero) coefficients is processed continuously in the rear of the encoding order, and the encoding efficiency by entropy encoding can be improved.
In the above description, the 16 × 16 pixel orthogonal transform block has been described. However, encoding processing in units of CG (coding sub-block) is also performed in block sizes other than 16 × 16 pixels, such as a 32 × 32 pixel orthogonal transform block. The encoding order is switched according to whether the flag indicating whether or not the field encoding of the sequence level header is valid, as in the case of the 16 × 16 pixel orthogonal transform block.
 上記ではシーケンスレベルヘッダのフィールド符号化か否かを示すフラグが有効である場合、図17に示す符号化順(符号化ブロック単位(16×16画素の符号化ブロック内の符号化順)、符号化サブブロック単位(4×4画素のCG内の符号化順))に切り替えるようにしたが、図18に示すようにCGの形状を4×4画素ブロックから8×2画素ブロックに変更するようにしてもよい。このようにすることでも、有意(非零)係数の符号化を符号化順で後ろの方のCGで連続して処理するようになり、エントロピー符号化による符号化効率を高めることができる。つまり、シーケンスレベルヘッダのフィールド符号化か否かを示すフラグが無効である場合は、図28のような符号化順であるため、図17の場合は符号化ブロック単位、符号化サブブロック単位で符号化順を切り替えるため、符号化効率を高めることができ、更に図18の場合は符号化ブロック単位、符号化サブブロック単位で符号化順を切り替えることに加えて、符号化サブブロックの形状も変更するようにしたため、更に符号化効率を高めることができる。なお、上記では、符号化順を符号化ブロック単位および符号化サブブロック単位の両方で切り替える場合について説明したが、符号化ブロック単位あるいは符号化サブブロック単位のいずれかのみを切り替えるようにしてもよい。
 または、シーケンスレベルヘッダのフィールド符号化か否かを示すフラグが有効である場合、図19に示す符号化順とするようにしてもよい。このように、CGの形状だけでなくCG内のスキャン順もブロックの右側の係数を優先して符号化するように変更することで、より有意(非零)係数の符号化を符号化順の後ろの方で連続して処理できるようになり、一層エントロピー符号化による符号化効率を高めることができる。
In the above, when the flag indicating whether or not the field coding of the sequence level header is valid, the coding order shown in FIG. 17 (coding block unit (coding order in a coding block of 16 × 16 pixels), code The sub-block unit (coding order within the CG of 4 × 4 pixels) is changed, but the shape of the CG is changed from the 4 × 4 pixel block to the 8 × 2 pixel block as shown in FIG. It may be. Even in this way, encoding of significant (non-zero) coefficients is continuously processed by the CG in the rear in the encoding order, and the encoding efficiency by entropy encoding can be improved. That is, when the flag indicating whether or not the field coding of the sequence level header is invalid, the coding order is as shown in FIG. 28, and in the case of FIG. 17, the coding block unit and the coding subblock unit are used. Since the coding order is switched, the coding efficiency can be improved. Further, in the case of FIG. 18, in addition to switching the coding order in units of coding blocks and coding sub-blocks, the shape of the coding sub-block is also changed. Since the change is made, the encoding efficiency can be further increased. In the above description, the case where the coding order is switched in both the coding block unit and the coding subblock unit has been described. However, only the coding block unit or the coding subblock unit may be switched. .
Alternatively, when the flag indicating whether or not the field encoding of the sequence level header is valid, the encoding order shown in FIG. 19 may be used. In this way, by changing not only the shape of the CG but also the scan order in the CG so that the coefficient on the right side of the block is preferentially encoded, the encoding of more significant (non-zero) coefficients can be performed in the encoding order. Processing can be continuously performed in the rear direction, and the encoding efficiency by entropy encoding can be further increased.
 なお、シーケンスレベルヘッダのフィールド符号化か否かを示すフラグをピクチャレベルヘッダに用意して、量子化された直交変換係数である圧縮データの符号化時の係数の符号化順をピクチャ単位に適応的に切り替えるようにしてもよい。そのようにすることで、ピクチャ単位適応的な制御が実現でき、符号化効率を高めることができる。なお、ピクチャ単位にフレーム符号化とフィールド符号化を適応的に切り替える符号化を実現する場合には、上記フラグはピクチャレベルヘッダに用意する必要がある。
 また、本実施の形態1では、シーケンスレベルヘッダあるいはピクチャレベルヘッダのフィールド符号化か否かを示すフラグに基づいて符号化順や形状等を切り替える場合について説明したが、シーケンスレベルヘッダあるいはピクチャレベルヘッダのフィールド符号化か否かを示すフラグとは別に本切り替え処理を行うか否かを示すフラグを定義し、この切り替え処理を行うか否かを示すフラグに基づいて符号化順やCGの形状やCG内のスキャン順等を切り替えるようにしてもよい。
 また、符号化順やCGの形状やCG内のスキャン順の例として図17、図18、図19を例示したが、有意(非零)係数の符号化を符号化順で後ろの方で連続して処理できれば、これに限られず、図17、図18、図19以外の符号化順やCGの形状やCG内のスキャン順であってもよく、CGの形状とCG内のスキャン順の組み合わせも図17、図18、図19に限られない。例えば、CGが1×2、1×4、1×8、1×16、2×2、2×4、4×8画素等であっても良い。
 また、本実施の形態1では、フィールド符号化の場合、図17、図18、図19のいずれか1つ(選択不可)である場合について説明したが、複数の候補(図17、図18、図19その他)の中から1つを選択するようにしてもよい。その場合、複数候補のうちどれを選択したかの示すフラグを上記ヘッダに用意する。このフラグは、フィールド符号化か否かを示すフラグや、本切り替え処理を行うか否かを示すフラグと共用してもよい。
In addition, a flag indicating whether or not the sequence level header field coding is prepared in the picture level header, and the coding order of the coefficients when coding the compressed data that is the quantized orthogonal transform coefficients is adapted to each picture. May be switched automatically. By doing so, control adaptive to each picture can be realized, and the encoding efficiency can be improved. Note that in the case of realizing coding that adaptively switches between frame coding and field coding on a picture-by-picture basis, the flag needs to be prepared in the picture level header.
Further, in the first embodiment, the case where the coding order, the shape, and the like are switched based on the flag indicating whether or not the field coding of the sequence level header or the picture level header has been described, but the sequence level header or the picture level header In addition to the flag indicating whether or not the field encoding is performed, a flag indicating whether or not to perform the switching process is defined, and based on the flag indicating whether or not the switching process is performed, the encoding order, the CG shape, The scanning order in the CG may be switched.
In addition, FIGS. 17, 18, and 19 are illustrated as examples of the coding order, the shape of the CG, and the scan order in the CG. However, encoding of significant (non-zero) coefficients is continued in the rear in the coding order. However, the processing order is not limited to this, and the encoding order other than those shown in FIGS. 17, 18, and 19, the CG shape, and the scan order within the CG may be used. The combination of the CG shape and the scan order within the CG Also, it is not limited to FIG. 17, FIG. 18, and FIG. For example, the CG may be 1 × 2, 1 × 4, 1 × 8, 1 × 16, 2 × 2, 2 × 4, 4 × 8 pixels, or the like.
In the first embodiment, the case of field encoding has been described with respect to the case of any one of FIGS. 17, 18, and 19 (cannot be selected), but a plurality of candidates (FIG. 17, FIG. 18, One of them may be selected from FIG. In that case, a flag indicating which of a plurality of candidates has been selected is prepared in the header. This flag may be shared with a flag indicating whether or not to perform field encoding or a flag indicating whether or not to perform this switching process.
 また、可変長符号化部13は、図13に例示するように、符号化ビットストリームのヘッダ情報として、シーケンスレベルヘッダ、ピクチャレベルヘッダを符号化し、ピクチャデータと共に符号化ビットストリームを生成する。
 ただし、ピクチャデータは1以上のスライスデータから構成され、各スライスデータはスライスレベルヘッダと当該スライス内にある上記符号化データをまとめたものである。
Further, as illustrated in FIG. 13, the variable length encoding unit 13 encodes a sequence level header and a picture level header as header information of the encoded bit stream, and generates an encoded bit stream together with the picture data.
However, picture data is composed of one or more slice data, and each slice data is a combination of a slice level header and the encoded data in the slice.
 シーケンスレベルヘッダは、画像サイズ、色信号フォーマット、輝度信号や色差信号の信号値のビット深度、シーケンス単位でのループフィルタ部11における各フィルタ処理(適応フィルタ処理、画素適応オフセット処理、デブロッキングフィルタ処理)の有効フラグ情報、量子化マトリクスの有効フラグ情報、フィールド符号化か否かを示すフラグなど、一般的にシーケンス単位に共通となるヘッダ情報をまとめたものである。
 ピクチャレベルヘッダは、参照するシーケンスレベルヘッダのインデックスや動き補償時の参照ピクチャ数、エントロピー符号化の確率テーブル初期化フラグ等のピクチャ単位で設定するヘッダ情報をまとめたものである。
 スライスレベルヘッダは、当該スライスがピクチャのどの位置にあるかを示す位置情報、どのピクチャレベルヘッダを参照するかを示すインデックス、スライスの符号化タイプ(オールイントラ符号化、インター符号化など)、ループフィルタ部11における各フィルタ処理(適応フィルタ処理、画素適応オフセット処理、デブロッキングフィルタ処理)を行うか否かを示すフラグ情報などといったスライス単位のパラメータをまとめたものである。
The sequence level header includes the image size, the color signal format, the bit depth of the signal value of the luminance signal and the color difference signal, and each filter process (adaptive filter process, pixel adaptive offset process, deblocking filter process) in the loop filter unit 11 in sequence units. ) Effective flag information, quantization matrix effective flag information, and a flag indicating whether or not to perform field encoding.
The picture level header is a collection of header information set in units of pictures such as an index of a sequence level header to be referenced, the number of reference pictures at the time of motion compensation, an entropy encoding probability table initialization flag, and the like.
The slice level header includes position information indicating where the slice is located in the picture, an index indicating which picture level header is referred to, a slice coding type (all-intra coding, inter coding, etc.), and a loop. This is a summary of parameters in units of slices such as flag information indicating whether or not to perform each filter process (adaptive filter process, pixel adaptive offset process, deblocking filter process) in the filter unit 11.
 次に、イントラ予測部4の処理内容を詳細に説明する。
 図7は符号化ブロックB内の各予測ブロックP が選択可能なイントラ予測パラメータであるイントラ予測モードの一例を示す説明図である。ただし、Nはイントラ予測モード数を表している。
 図7では、イントラ予測モードのインデックス値と、そのイントラ予測モードが示す予測方向ベクトルを示しており、図7の例では、選択可能なイントラ予測モードの個数が増えるに従って、予測方向ベクトル同士の相対角度が小さくなるように設計されている。
Next, the processing content of the intra estimation part 4 is demonstrated in detail.
Figure 7 is an explanatory diagram showing an example of the intra prediction mode is an intra prediction parameters each prediction block P i n is selectable within a coding block B n. However, N I represents the number of intra prediction modes.
FIG. 7 shows an index value of an intra prediction mode and a prediction direction vector indicated by the intra prediction mode. In the example of FIG. 7, as the number of selectable intra prediction modes increases, Designed to reduce the angle.
 イントラ予測部4は、上述したように、予測ブロックP のイントラ予測パラメータを参照して、その予測ブロックP に対するイントラ予測処理を実施して、イントラ予測画像PINTRAi を生成するが、ここでは、輝度信号における予測ブロックP のイントラ予測信号を生成するイントラ処理について説明する。 Intra prediction unit 4, as described above, with reference to the intra prediction parameters of the prediction block P i n, to implement intra prediction processing for the prediction block P i n, but to generate an intra prediction image P INTRAi n , it will be described here intra process of generating an intra prediction signal of a prediction block P i n in the luminance signal.
 予測ブロックP のサイズをl ×m 画素とする。
 図8はl =m =4の場合の予測ブロックP 内の画素の予測値を生成する際に用いる画素の一例を示す説明図である。
 図8では、予測ブロックP の上の符号化済みの画素(2×l +1)個と、左の符号化済みの画素(2×m )個を予測に用いる画素としているが、予測に用いる画素は、図8に示す画素より多くても少なくてもよい。
 また、図8では、予測ブロックP の近傍の1行又は1列分の画素を予測に用いているが、2行又は2列、あるいは、それ以上の画素を予測に用いてもよい。
The size of the prediction block P i n is set to l i n × m i n pixels.
FIG. 8 is an explanatory diagram illustrating an example of a pixel used when generating a prediction value of a pixel in the prediction block P i n when l i n = m i n = 4.
In FIG. 8, the encoded pixels (2 × l i n +1) and the left encoded pixels (2 × m i n ) on the prediction block P i n are used as prediction pixels. However, the number of pixels used for prediction may be more or less than the pixels shown in FIG.
Further, in FIG. 8, it is used to predict one line or pixel of one column in the vicinity of the predicted block P i n, 2 rows or two columns, or may be used more pixels in the prediction.
 予測ブロックP に対するイントラ予測モードのインデックス値が0(平面(Planar)予測)の場合には、予測ブロックP の上に隣接する符号化済み画素と予測ブロックP の左に隣接する符号化済み画素を用いて、これら画素と予測ブロックP 内の予測対象画素との距離に応じて内挿した値を予測値として予測画像を生成する。 When the index value of the intra prediction mode for the prediction block P i n is 0 (planar prediction), the encoded pixel adjacent on the prediction block P i n is adjacent to the left of the prediction block P i n. using encoded pixels that generates a predicted image interpolated value according to the distance of the pixel and the prediction pixel in the prediction block P i n as the predicted value.
 予測ブロックP に対するイントラ予測モードのインデックス値が2(平均値(DC)予測)の場合には、予測ブロックP の上に隣接する符号化済み画素と予測ブロックP の左に隣接する符号化済み画素の平均値を予測ブロックP 内の画素の予測値として予測画像を生成する。
さらに、予測ブロックP の上端及び左端に位置する図20の領域A、B、Cに対してブロック境界を平滑化するフィルタ処理を行って最終的な予測画像とする。例えば、図21のフィルタの参照画素配置で下記のフィルタ係数を用いてフィルタ処理を実施する。
・領域A(パーティションP の左上の画素)
   a=1/2,a=1/4,a=1/4
・領域B(領域A以外のパーティションP の上端の画素)
   a=3/4,a=1/4,(a=0)
・領域C(領域A以外のパーティションP の左端の画素)
   a=3/4,a=1/4,(a=0)
 ただし、シーケンスレベルヘッダのフィールド符号化か否かを示すフラグが有効である場合、図22に示すとおり予測ブロックの上端にフィルタ処理を実施しないようにする。フィールド符号化の場合、垂直方向の画素間の相関が低いことから図27の水平方向予測におけるフィルタ処理によって予測効率が悪化してしまう可能性がある。したがって、領域A、Cのみフィルタ処理を行い、領域Bはフィルタ処理を行わないようにすることで予測効率の低下を抑制しつつ演算量を削減することができる。
When the index value of the intra prediction mode for the prediction block P i n is 2 (average value (DC) prediction), the encoded pixels adjacent on the prediction block P i n and the prediction block P i n are placed to the left of the prediction block P i n . generating a prediction image the mean value of the encoded adjacent pixel as a prediction value of the pixels in the prediction block P i n.
Furthermore, the prediction block P i n in the upper end and the area A of Figure 20 positioned at the left end, B, and final prediction image by performing a filtering process for smoothing the block boundary relative to the C. For example, filter processing is performed using the following filter coefficients in the reference pixel arrangement of the filter of FIG.
· Area A (upper left pixel of the partition P i n)
a 0 = 1/2, a 1 = ¼, a 2 = ¼
- region B (the upper end of the pixel partition P i n other than the region A)
a 0 = 3/4, a 2 = ¼, (a 1 = 0)
· Area C (the leftmost pixel of the partition P i n other than the region A)
a 0 = 3/4, a 1 = ¼, (a 2 = 0)
However, when the flag indicating whether or not the field encoding of the sequence level header is valid, the filtering process is not performed on the upper end of the prediction block as shown in FIG. In the case of field coding, since the correlation between the pixels in the vertical direction is low, the prediction efficiency may be deteriorated by the filter processing in the horizontal direction prediction of FIG. Therefore, only the areas A and C are filtered, and the area B is not filtered, so that the amount of calculation can be reduced while suppressing a decrease in prediction efficiency.
ただし、上記ではシーケンスレベルヘッダのフィールド符号化か否かを示すフラグが有効である場合、領域A、Cのみフィルタ処理を行うようにしたが、領域Aも領域Cと同じフィルタ処理してもよい。このように、画素間の相関が低い垂直方向の画素を用いないようにすることでさらに予測効率悪化の可能性を下げつつ、フィルタ処理に要する演算量を削減することができる。あるいはより演算量を削減することを重視する場合は、領域Aもフィルタ処理を行わないようにし、領域Cのみフィルタ処理をおこなうようにしてもよい。 However, in the above, when the flag indicating whether or not the field encoding of the sequence level header is valid, only the regions A and C are subjected to the filtering process. However, the region A may be subjected to the same filtering process as the region C. . Thus, by not using the pixels in the vertical direction where the correlation between the pixels is low, it is possible to reduce the amount of computation required for the filter processing while further reducing the possibility of deterioration in prediction efficiency. Alternatively, when it is more important to reduce the amount of calculation, the region A may not be filtered, and only the region C may be filtered.
予測ブロックP に対するイントラ予測モードのインデックス値が26(垂直方向予測)の場合、下記の式(1)から予測ブロックP 内の画素の予測値を算出して予測画像を生成する。
Figure JPOXMLDOC01-appb-I000001

 ただし、座標(x,y)は予測ブロックP 内の左上画素を原点とする相対座標(図9を参照)であり、S’(x,y)は座標(x,y)における予測値、S(x,y)は座標(x,y)における符号化済み画素の輝度値(復号された輝度値)である。また、算出した予測値が輝度値の取り得る値の範囲を超えている場合、予測値がその範囲内に収まるように値を丸めるようにする。
 なお、式(1)は図27の垂直方向予測におけるフィルタ処理を示しており、具体的には、式(1)の1行目の式はMPEG-4 AVC/H.264における垂直方向予測の予測値であるS(x,-1)に対し、隣接する符号化済み画素の垂直方向の輝度値の変化量を1/2にした値を加算することでブロック境界が平滑化されるようにフィルタ処理することを意味しており、式(1)の2行目の式はMPEG-4 AVC/H.264における垂直方向予測と同じ予測式を示している。
For prediction block P i index value of the intra prediction mode for n is 26 (vertical prediction), and generates a prediction image by calculating the predicted values of the pixels in the prediction block P i n from the following equation (1).
Figure JPOXMLDOC01-appb-I000001

However, the coordinates (x, y) is the relative coordinates with the origin at the upper left pixel in the prediction block P i n (see Figure 9), S '(x, y) is the predicted value at the coordinates (x, y) , S (x, y) is the luminance value (decoded luminance value) of the encoded pixel at the coordinates (x, y). Further, when the calculated predicted value exceeds the range of values that the luminance value can take, the value is rounded so that the predicted value falls within the range.
Equation (1) shows the filter processing in the vertical direction prediction of FIG. 27. Specifically, the equation in the first row of Equation (1) is MPEG-4 AVC / H. The block boundary is obtained by adding a value obtained by halving the amount of change in the luminance value of the adjacent encoded pixel in the vertical direction to S (x, −1) that is the predicted value of the vertical direction prediction in H.264. This means that filtering is performed so as to be smoothed, and the expression in the second row of Expression (1) is MPEG-4 AVC / H. The same prediction formula as the vertical direction prediction in H.264 is shown.
予測ブロックP に対するイントラ予測モードのインデックス値が10(水平方向予測)の場合、下記の式(2)から予測ブロックP 内の画素の予測値を算出して予測画像を生成する。
Figure JPOXMLDOC01-appb-I000002

 ただし、座標(x,y)は予測ブロックP 内の左上画素を原点とする相対座標(図9を参照)であり、S’(x,y)は座標(x,y)における予測値、S(x,y)は座標(x,y)における符号化済み画素の輝度値(復号された輝度値)である。また、算出した予測値が輝度値の取り得る値の範囲を超えている場合、予測値がその範囲内に収まるように値を丸めるようにする。
 なお、式(2)は図27の水平方向予測におけるフィルタ処理を示しており、具体的には、式(2)の1行目の式はMPEG-4 AVC/H.264における水平方向予測の予測値であるS(-1,y)に対し、隣接する符号化済み画素の水平方向の輝度値の変化量を1/2にした値を加算することでブロック境界が平滑化されるようにフィルタ処理することを意味しており、式(2)の2行目の式はMPEG-4 AVC/H.264における水平方向予測と同じ予測式を示している。
If the index value of the intra prediction mode for prediction block P i n is 10 (horizontal prediction), and generates a prediction image by calculating the predicted values of the pixels in the prediction block P i n from the following equation (2).
Figure JPOXMLDOC01-appb-I000002

However, the coordinates (x, y) is the relative coordinates with the origin at the upper left pixel in the prediction block P i n (see Figure 9), S '(x, y) is the predicted value at the coordinates (x, y) , S (x, y) is the luminance value (decoded luminance value) of the encoded pixel at the coordinates (x, y). Further, when the calculated predicted value exceeds the range of values that the luminance value can take, the value is rounded so that the predicted value falls within the range.
Equation (2) shows the filter processing in the horizontal prediction of FIG. 27. Specifically, the equation in the first row of Equation (2) is MPEG-4 AVC / H. The block boundary is obtained by adding a value obtained by halving the amount of change in the horizontal luminance value of the adjacent encoded pixel to S (−1, y), which is the predicted value of the horizontal direction prediction in H.264. This means that filtering is performed so as to be smoothed, and the expression in the second line of Expression (2) is MPEG-4 AVC / H. The same prediction formula as the horizontal prediction in H.264 is shown.
 ただし、シーケンスレベルヘッダのフィールド符号化か否かを示すフラグが有効である場合、水平方向予測は式(2)の代わりに式(3)を用いる。
Figure JPOXMLDOC01-appb-I000003
すなわち、図22に示すとおり予測ブロックの上端にフィルタ処理を実施しないようにする(平均値予測及び垂直方向予測の場合は予測ブロックの左端のみにフィルタ処理を実施し、水平方向予測の場合はフィルタ処理を実施しない)。フィールド符号化の場合、垂直方向の画素間の相関が低いことから図27の水平方向予測におけるフィルタ処理によるブロック境界の連続性を高めることによる予測効率が悪化してしまう可能性がある。したがって、上記フィルタ処理を行わないことで予測効率の低下を抑制しつつ演算量を削減することができる。
However, when the flag indicating whether or not the field encoding of the sequence level header is valid, the equation (3) is used instead of the equation (2) in the horizontal prediction.
Figure JPOXMLDOC01-appb-I000003
That is, as shown in FIG. 22, the filter processing is not performed on the upper end of the prediction block (in the case of average value prediction and vertical direction prediction, the filter processing is performed only on the left end of the prediction block, and in the case of horizontal direction prediction, the filter processing is performed. Do not process). In the case of field coding, since the correlation between pixels in the vertical direction is low, there is a possibility that the prediction efficiency by increasing the continuity of the block boundary by the filter processing in the horizontal direction prediction in FIG. 27 may deteriorate. Therefore, by not performing the filtering process, it is possible to reduce the amount of computation while suppressing a decrease in prediction efficiency.
 なお、シーケンスレベルヘッダのフィールド符号化か否かを示すフラグをピクチャレベルヘッダに用意して各ピクチャの、垂直方向の画素間の相関に応じて平均値(DC)予測及び水平方向予測の予測ブロック上端のフィルタ処理をON/OFF切り替えるようにしてもよい。そのようにすることで、ピクチャ単位適応的な制御が実現でき、予測効率を高めることができる。なお、ピクチャ単位にフレーム符号化とフィールド符号化を適応的に切り替える符号化を実現する場合には、上記フラグはピクチャレベルヘッダに用意する必要がある。
 また、本実施の形態1では、シーケンスレベルヘッダあるいはピクチャレベルヘッダのフィールド符号化か否かを示すフラグに基づいて予測ブロック上端のフィルタ処理のON/OFFを切り替える場合について説明したが、シーケンスレベルヘッダあるいはピクチャレベルヘッダのフィールド符号化か否かを示すフラグとは別に本切り替え処理を行うか否かを示すフラグを定義し、この切り替え処理を行うか否かを示すフラグに基づいて予測ブロック上端のフィルタ処理のON/OFFを切り替えるようにしてもよい。ただし、この切り替え処理を行うか否かを示すフラグは、フィールド符号化か否かを示すフラグに基づいたフラグである。
 また、本実施の形態1では、先に説明した符号化順の切り替えと、上記フィルタ処理の切り替えとを、それぞれ別々に説明したが、これらを組み合わせて設定するようにしてもよい。
It should be noted that a flag indicating whether or not the field coding of the sequence level header is prepared in the picture level header, and a prediction block of average value (DC) prediction and horizontal prediction according to the correlation between pixels in the vertical direction of each picture The upper end filter processing may be switched ON / OFF. By doing so, control adaptive to each picture can be realized, and prediction efficiency can be improved. Note that in the case of realizing encoding that adaptively switches between frame encoding and field encoding on a picture-by-picture basis, the flag needs to be prepared in the picture level header.
In the first embodiment, the case has been described in which the filtering process at the upper end of the prediction block is switched on / off based on the flag indicating whether or not the field encoding of the sequence level header or the picture level header is performed. Alternatively, a flag indicating whether or not to perform the main switching process is defined separately from the flag indicating whether or not the field coding of the picture level header is performed, and the upper end of the prediction block is determined based on the flag indicating whether or not to perform the switching process. You may make it switch ON / OFF of a filter process. However, the flag indicating whether or not to perform the switching process is a flag based on the flag indicating whether or not the field encoding is performed.
In the first embodiment, the coding order switching described above and the filter processing switching described above are separately described. However, these may be set in combination.
 また、平均値(DC)予測、垂直方向予測、水平方向予測のブロック境界のフィルタ処理を例えば16×16画素以下のブロックのみとする等、フィルタ処理を行うブロックサイズを限定してもよい。このようにすることで、フィルタ処理に要する演算量を削減することができる。 Further, the block size to be subjected to the filtering process may be limited, for example, the block boundary filtering process of average value (DC) prediction, vertical direction prediction, and horizontal direction prediction is limited to, for example, blocks of 16 × 16 pixels or less. By doing in this way, the amount of calculations required for filter processing can be reduced.
 イントラ予測モードのインデックス値が0(平面予測)、2(平均値予測)、26(垂直方向予測)、10(水平方向予測)以外の場合には、インデックス値が示す予測方向ベクトルυ=(dx,dy)に基づいて、予測ブロックP 内の画素の予測値を生成する。
 図9に示すように、予測ブロックP の左上画素を原点として、予測ブロックP 内の相対座標を(x,y)と設定すると、予測に用いる参照画素の位置は、下記のLと隣接画素の交点になる。
Figure JPOXMLDOC01-appb-I000004
ただし、kは負のスカラ値である。
When the index value of the intra prediction mode is other than 0 (plane prediction), 2 (average value prediction), 26 (vertical direction prediction), and 10 (horizontal direction prediction), the prediction direction vector υ p = ( dx, on the basis of the dy), and it generates the predicted values of the pixels in the prediction block P i n.
As shown in FIG. 9, as the origin at the upper left pixel of the prediction block P i n, setting the relative coordinates of the prediction block P i n (x, y) and, the position of the reference pixels used for prediction, the following L And the intersection of adjacent pixels.
Figure JPOXMLDOC01-appb-I000004
Where k is a negative scalar value.
 参照画素が整数画素位置にある場合には、その整数画素を予測対象画素の予測値とし、参照画素が整数画素位置にない場合には、参照画素に隣接する整数画素から生成される補間画素を予測値とする。
 図8の例では、参照画素は整数画素位置にないので、参照画素に隣接する2画素から内挿したものを予測値とする。なお、隣接する2画素のみではなく、隣接する2画素以上の画素から補間画素を生成して予測値としてもよい。
 補間処理に用いる画素を多くすることで補間画素の補間精度を向上させる効果がある一方、補間処理に要する演算の複雑度が増加することから、演算負荷が大きくても高い符号化性能を要求する動画像符号化装置の場合には、より多くの画素から補間画素を生成するようにした方がよい。
When the reference pixel is at the integer pixel position, the integer pixel is set as the prediction value of the prediction target pixel. When the reference pixel is not at the integer pixel position, an interpolation pixel generated from the integer pixel adjacent to the reference pixel is selected. Estimated value.
In the example of FIG. 8, since the reference pixel is not located at the integer pixel position, a value interpolated from two pixels adjacent to the reference pixel is set as the predicted value. Note that an interpolation pixel may be generated not only from two adjacent pixels but also from two or more adjacent pixels, and used as a predicted value.
While increasing the number of pixels used in the interpolation process has the effect of improving the interpolation accuracy of the interpolated pixels, it increases the complexity of the calculation required for the interpolation process, requiring high coding performance even when the calculation load is large. In the case of a video encoding device, it is better to generate interpolation pixels from more pixels.
 以上に述べた処理によって、予測ブロックP 内の輝度信号の全ての画素に対する予測画素を生成して、イントラ予測画像PINTRAi を出力する。
 なお、イントラ予測画像PINTRAi の生成に用いられたイントラ予測パラメータ(イントラ予測モード)は、ビットストリームに多重化するために可変長符号化部13に出力される。
The processing described above, to generate a predicted pixel for all the pixels of the luminance signal in the prediction block P i n, and outputs an intra prediction image P INTRAi n.
Incidentally, the intra prediction parameters used for generating the intra prediction image P INTRAi n (intra prediction mode) is output to the variable length coding unit 13 for multiplexing the bitstream.
 なお、先に説明したMPEG-4 AVC/H.264における8×8画素のブロックのイントラ予測時に参照画像に対して施される平滑化処理と同様に、イントラ予測部4において、予測ブロックP の中間予測画像を生成する際の参照画素を、予測ブロックP に隣接する符号化済み画素を平滑化処理した画素とするように構成した場合であっても、上述の例と同様の中間予測画像に対するフィルタ処理を行うことができる。 Note that the MPEG-4 AVC / H. Similar to the smoothing process performed on the reference image at the time of 8 × 8 pixel block intra prediction of the 264, the intra prediction unit 4, the reference pixels in generating the intermediate prediction image predicted block P i n even when configured as a prediction block P i n pixels smoothed the encoded pixels adjacent to, it is possible to perform the filtering for the same intermediate predicted image and the above example.
 予測ブロックP の色差信号に対しても、輝度信号と同様の手順で、イントラ予測パラメータ(イントラ予測モード)に基づくイントラ予測処理を実施し、イントラ予測画像の生成に用いられたイントラ予測パラメータを可変長符号化部13に出力する。
 ただし、色差信号で選択可能なイントラ予測パラメータ(イントラ予測モード)は輝度信号と異なっていてもよい。例えば、演算量を削減するために、色差信号の垂直方向予測及び水平方向予測についてはブロック境界のフィルタ処理を行わずにMPEG-4 AVC/H.264と同様の予測手法としてもよい。YUV信号4:2:0フォーマットの場合、色差信号(U、V信号)は、輝度信号(Y信号)に対して解像度を水平方向、垂直方向共に1/2に縮小した信号であり、輝度信号に比べて画像信号の複雑性が低く予測が容易であることから、選択可能なイントラ予測パラメータは輝度信号よりも少ない数としてイントラ予測パラメータを符号化するのに要する符号量の削減や、予測処理の低演算化を図ってもよい。
Even for the color difference signal of the prediction block P i n, in the same procedure as the luminance signal, the intra prediction processing based on the intra prediction parameters (intra prediction mode) performed, the intra prediction parameters used for generating the intra prediction image Is output to the variable length encoding unit 13.
However, the intra prediction parameter (intra prediction mode) that can be selected by the color difference signal may be different from that of the luminance signal. For example, in order to reduce the amount of calculation, the MPEG-4 AVC / H. It is good also as the prediction method similar to H.264. In the case of the YUV signal 4: 2: 0 format, the color difference signal (U, V signal) is a signal obtained by reducing the resolution to 1/2 in both the horizontal direction and the vertical direction with respect to the luminance signal (Y signal). Compared to, the complexity of the image signal is low and prediction is easy, so that the number of intra prediction parameters that can be selected is smaller than that of the luminance signal, and the amount of code required to encode the intra prediction parameters can be reduced. May be reduced.
 次に、図3の動画像復号装置の処理内容を具体的に説明する。
 可変長復号部31は、図1の動画像符号化装置により生成された符号化ビットストリームを入力すると、そのビットストリームに対する可変長復号処理を実施して(図4のステップST21)、フィールド符号化か否かを示すフラグやフレームサイズの情報などの1フレーム以上のピクチャから構成されるシーケンス単位のヘッダ情報(シーケンスレベルヘッダ)及びピクチャ単位のヘッダ情報(ピクチャレベルヘッダ)、ループフィルタ部38で使用するフィルタパラメータや量子化マトリクスパラメータを復号する。
 このとき、可変長復号部31によって可変長復号された量子化マトリクスパラメータを参照して量子化マトリクスを特定する。具体的には、各直交変換サイズの色信号や符号化モード毎に、量子化マトリクスパラメータが初期値として予め動画像符号化装置及び動画像復号装置で共通に用意されている量子化マトリクス、または、既に復号された量子化マトリクスである(新しい量子化マトリクスでない)ことを示す場合は、量子化マトリクスパラメータに含まれる上記マトリクスの内のどの量子化マトリクスであるかを特定するインデックス情報を参照して量子化マトリクスを特定し、量子化マトリクスパラメータが新しい量子化マトリクスを用いることを示す場合は、量子化マトリクスパラメータに含まれる量子化マトリクスを使用する量子化マトリクスとして特定する。
 そして、ピクチャ単位のデータを構成するスライスデータから、スライス分割情報等のスライス単位のヘッダ情報(スライスレベルヘッダ)を復号し、各スライスの符号化データを復号する。
Next, the processing contents of the moving picture decoding apparatus in FIG. 3 will be specifically described.
When the variable length decoding unit 31 receives the encoded bitstream generated by the moving picture encoding apparatus in FIG. 1, the variable length decoding unit 31 performs variable length decoding processing on the bitstream (step ST21 in FIG. 4), and performs field encoding. Header information (sequence level header) composed of one or more pictures, such as a flag indicating whether or not, frame size information, and header information (picture level header) in units of pictures, used in the loop filter unit 38 The filter parameters and quantization matrix parameters to be decoded are decoded.
At this time, the quantization matrix is specified with reference to the quantization matrix parameter variable-length decoded by the variable-length decoding unit 31. Specifically, for each color signal and encoding mode of each orthogonal transform size, a quantization matrix that is prepared in advance by the moving image encoding device and the moving image decoding device in advance as the initial value of the quantization matrix parameter, or When indicating that the quantization matrix is already decoded (not a new quantization matrix), refer to the index information for specifying which quantization matrix among the matrices included in the quantization matrix parameter. When the quantization matrix is specified, and the quantization matrix parameter indicates that a new quantization matrix is to be used, it is specified as a quantization matrix that uses the quantization matrix included in the quantization matrix parameter.
Then, slice unit header information (slice level header) such as slice division information is decoded from slice data constituting picture unit data, and encoded data of each slice is decoded.
 また、可変長復号部31は、図1の動画像符号化装置の符号化制御部2により決定された最大符号化ブロックサイズ及び分割階層数の上限を動画像符号化装置と同様の手順で決定する(ステップST22)。
 例えば、最大符号化ブロックサイズや分割階層数の上限が映像信号の解像度に応じて決められた場合には、復号したフレームサイズ情報に基づいて、動画像符号化装置と同様の手順で最大符号化ブロックサイズを決定する。
 最大符号化ブロックサイズ及び分割階層数の上限が、動画像符号化装置側でシーケンスレベルヘッダなどに多重化されている場合には、上記ヘッダから復号した値を用いる。
 以降、動画像復号装置では、上記最大符号化ブロックサイズを最大復号ブロックサイズと称し、最大符号化ブロックを最大復号ブロックと称する。
 可変長復号部31は、決定された最大復号ブロック単位に、図6で示されるような最大復号ブロックの分割状態を復号する。復号された分割状態に基づき、階層的に復号ブロック(図1の動画像符号化装置の「符号化ブロック」に相当するブロック)を特定する(ステップST23)。
Further, the variable length decoding unit 31 determines the maximum coding block size and the upper limit of the number of divided layers determined by the coding control unit 2 of the moving image coding device in FIG. 1 in the same procedure as the moving image coding device. (Step ST22).
For example, when the maximum encoding block size and the upper limit of the number of division layers are determined according to the resolution of the video signal, the maximum encoding is performed in the same procedure as the moving image encoding apparatus based on the decoded frame size information. Determine the block size.
When the maximum encoding block size and the upper limit of the number of divided layers are multiplexed on a sequence level header or the like on the moving image encoding device side, values decoded from the header are used.
Hereinafter, in the video decoding apparatus, the maximum encoded block size is referred to as a maximum decoded block size, and the maximum encoded block is referred to as a maximum decoded block.
The variable length decoding unit 31 decodes the division state of the maximum decoding block as shown in FIG. 6 for each determined maximum decoding block. Based on the decoded division state, a decoded block (a block corresponding to the “encoded block” of the moving image encoding apparatus in FIG. 1) is identified hierarchically (step ST23).
 次に、可変長復号部31は、復号ブロックに割り当てられている符号化モードを復号する。復号した符号化モードに含まれる情報に基づき、復号ブロックをさらに1つないし複数の予測処理単位である予測ブロックに分割し、予測ブロック単位に割り当てられている予測パラメータを復号する(ステップST24)。 Next, the variable length decoding unit 31 decodes the encoding mode assigned to the decoding block. Based on the information included in the decoded coding mode, the decoded block is further divided into one or more prediction blocks which are prediction processing units, and the prediction parameters assigned to the prediction block units are decoded (step ST24).
 即ち、可変長復号部31は、復号ブロックに割り当てられている符号化モードがイントラ符号化モードである場合、復号ブロックに含まれており、予測処理単位となる1つ以上の予測ブロック毎にイントラ予測パラメータを復号する。
 一方、復号ブロックに割り当てられている符号化モードがインター符号化モードである場合、復号ブロックに含まれており、予測処理単位となる1つ以上の予測ブロック毎にインター予測パラメータ及び動きベクトルを復号する(ステップST24)。
That is, when the encoding mode assigned to the decoding block is the intra encoding mode, the variable length decoding unit 31 is included in the decoding block and is intra for each of one or more prediction blocks serving as a prediction processing unit. Decode prediction parameters.
On the other hand, when the coding mode assigned to the decoding block is the inter coding mode, the inter prediction parameter and the motion vector are decoded for each one or more prediction blocks included in the decoding block and serving as a prediction processing unit. (Step ST24).
 さらに、可変長復号部31は、予測差分符号化パラメータに含まれる直交変換ブロック分割情報に基づき、直交変換ブロック毎に圧縮データ(変換・量子化後の変換係数)を復号する(ステップST24)。
 その際、図1の動画像符号化装置の可変長符号化部13での圧縮データの符号化処理と同様に、CG単位の係数の復号処理を実施する。したがって、通常は図28に示すように4×4画素単位の16個のCGを右下のCGから順に復号処理し、さらに各CGはCG内の16個の係数を右下の係数から順に復号していくこととなる。具体的には、まずCG内の16個の係数の中に有意(非零)係数が存在するか否かのフラグ情報を復号し、次に復号したフラグ情報がCG内に有意(非零)係数が存在することを示す場合のみCG内の各係数が有意(非零)係数であるかを上記順に復号し、最後に有意(非零)係数を示す係数に対してその係数値情報を順に復号する。これをCG単位に上記順に行う。ただし、可変長復号部31によって復号されたシーケンスレベルヘッダのフィールド符号化か否かを示すフラグが有効である場合、図17、図18、図19の内、図1の動画像符号化装置の可変長符号化部13で決められた処理順と同一の順番で復号処理が行われる。このようにすることで、図1の動画像符号化装置で生成したストリームと同一の圧縮データを生成することができる。
Furthermore, the variable length decoding unit 31 decodes the compressed data (transformed / transformed transform coefficients) for each orthogonal transform block based on the orthogonal transform block division information included in the prediction difference encoding parameter (step ST24).
At that time, similarly to the encoding process of the compressed data in the variable length encoding unit 13 of the moving image encoding apparatus of FIG. Therefore, normally, as shown in FIG. 28, 16 CGs in units of 4 × 4 pixels are decoded in order from the lower right CG, and each CG decodes 16 coefficients in the CG in order from the lower right coefficient. It will be done. Specifically, first, flag information indicating whether or not a significant (non-zero) coefficient exists in 16 coefficients in the CG is decoded, and then the decoded flag information is significant (non-zero) in the CG. Only when it indicates that a coefficient exists, whether each coefficient in the CG is a significant (non-zero) coefficient is decoded in the order described above, and finally, coefficient value information is sequentially displayed for the coefficient indicating the significant (non-zero) coefficient. Decrypt. This is performed in the above order in units of CG. However, when the flag indicating whether or not the field encoding of the sequence level header decoded by the variable length decoding unit 31 is valid, the moving picture encoding apparatus of FIG. Decoding processing is performed in the same order as the processing order determined by the variable length encoding unit 13. In this way, it is possible to generate the same compressed data as the stream generated by the moving image encoding apparatus in FIG.
 なお、シーケンスレベルヘッダのフィールド符号化か否かを示すフラグをピクチャレベルヘッダに用意して、量子化された直交変換係数である圧縮データの符号化時の係数の符号化順をピクチャ単位に適応的に切り替えるように図1の動画像符号化装置の可変長符号化部13が構成されている場合、可変長復号部31も同様に上記フラグに応じて圧縮データの復号順をピクチャ単位に適応的に切り替えるようにする。 In addition, a flag indicating whether or not the sequence level header field coding is prepared in the picture level header, and the coding order of the coefficients when coding the compressed data that is the quantized orthogonal transform coefficient is adapted to each picture. When the variable length coding unit 13 of the moving picture coding apparatus in FIG. 1 is configured so as to be switched, the variable length decoding unit 31 similarly applies the decoding order of compressed data in units of pictures according to the flag. To switch automatically.
 切換スイッチ33は、可変長復号部31により可変長復号された符号化モードm(B)がイントラ符号化モードであれば(m(B)∈INTRAの場合)、可変長復号部31により可変長復号された予測ブロック単位のイントラ予測パラメータをイントラ予測部34に出力する。
 一方、可変長復号部31により可変長復号された符号化モードm(B)がインター符号化モードであれば(m(B)∈INTERの場合)、可変長復号部31により可変長復号された予測ブロック単位のインター予測パラメータ及び動きベクトルを動き補償部35に出力する。
If the encoding mode m (B n ) variable-length decoded by the variable-length decoding unit 31 is an intra-encoding mode (when m (B n ) ∈INTRA), the changeover switch 33 is changed by the variable-length decoding unit 31. The intra-prediction parameter for each prediction block subjected to variable length decoding is output to the intra-prediction unit 34.
On the other hand, (the case of m (B n) ∈INTER) variable length decoded coding mode m (B n) is if the inter coding mode by the variable length decoding unit 31, variable length decoding by the variable length decoding unit 31 The predicted inter prediction parameters and motion vectors in units of prediction blocks are output to the motion compensation unit 35.
 イントラ予測部34は、可変長復号部31により可変長復号された符号化モードm(B)がイントラ符号化モード(m(B)∈INTRA)である場合(ステップST25)、切換スイッチ33から出力された予測ブロック単位のイントラ予測パラメータを受け取って、図1のイントラ予測部4と同様の手順で、イントラ予測用メモリ37に格納されている復号画像を参照しながら、上記イントラ予測パラメータを用いた復号ブロックB内の各予測ブロックP に対するイントラ予測処理を実施して、イントラ予測画像PINTRAi を生成する(ステップST26)。
 ただし、可変長復号部31によって復号されたシーケンスレベルヘッダのフィールド符号化か否かを示すフラグが有効である場合、図1の動画像符号化装置と同様に平均値(DC)予測及び水平方向予測の予測ブロック上端のフィルタ処理を行わないようにする。このようにすることで、図1の動画像符号化装置で生成したストリームと同一の予測画像を生成することができる。
When the coding mode m (B n ) variable-length decoded by the variable-length decoding unit 31 is the intra coding mode (m (B n ) ∈INTRA) (step ST25), the intra prediction unit 34 selects the changeover switch 33. 1 is received, and the intra prediction parameter is obtained by referring to the decoded image stored in the intra prediction memory 37 in the same procedure as the intra prediction unit 4 in FIG. and implementing intra prediction process to generate an intra prediction image P INTRAi n for each of the prediction block P i n of the decoded block B n using (step ST26).
However, when the flag indicating whether or not the field coding of the sequence level header decoded by the variable length decoding unit 31 is valid, the average value (DC) prediction and the horizontal direction are performed in the same manner as the moving picture coding apparatus of FIG. The filtering process at the upper end of the prediction block of the prediction is not performed. In this way, it is possible to generate a predicted image that is the same as the stream generated by the moving image encoding device in FIG.
 なお、実施の形態1の動画像符号化装置において、シーケンスレベルヘッダのフィールド符号化か否かを示すフラグをピクチャレベルヘッダに用意している場合、このピクチャレベルヘッダのフィールド符号化か否かを示すフラグの値に応じてピクチャ単位に平均値(DC)予測及び水平方向予測の予測ブロック上端のフィルタ処理をON/OFF切り替えるようにする。このようにすることで、上記のように構成した実施の形態1の動画像符号化装置で生成したストリームと同一の予測画像を生成することができる。 In the moving picture encoding apparatus according to the first embodiment, when a flag indicating whether or not the sequence level header field encoding is provided in the picture level header, whether or not the picture level header field encoding is performed. Filter processing at the upper end of the prediction block of average value (DC) prediction and horizontal direction prediction is switched ON / OFF in units of pictures according to the flag value shown. In this way, it is possible to generate the same predicted image as the stream generated by the moving picture coding apparatus according to Embodiment 1 configured as described above.
 動き補償部35は、可変長復号部31により可変長復号された符号化モードm(B)がインター符号化モード(m(B)∈INTER)である場合(ステップST25)、切換スイッチ33から出力された予測ブロック単位の動きベクトルとインター予測パラメータを受け取って、動き補償予測フレームメモリ39に格納されているフィルタ処理後の復号画像を参照しながら、その動きベクトルとインター予測パラメータを用いた復号ブロックB内の各予測ブロックP に対するインター予測処理を実施してインター予測画像PINTERi を生成する(ステップST27)。 When the coding mode m (B n ) variable-length decoded by the variable-length decoding unit 31 is the inter coding mode (m (B n ) ∈INTER) (step ST25), the motion compensation unit 35 performs the changeover switch 33. The motion vector and the inter prediction parameter for each prediction block output from the above are received, and the motion vector and the inter prediction parameter are used while referring to the decoded image after filtering stored in the motion compensated prediction frame memory 39. by carrying out inter-prediction processing for each of the prediction block P i n of the decoded block B n to generate an inter prediction image P INTERi n (step ST27).
 逆量子化・逆変換部32は、可変長復号部31から圧縮データ及び予測差分符号化パラメータを受けると、図1の逆量子化・逆変換部8と同様の手順で、その予測差分符号化パラメータに含まれる量子化パラメータ及び直交変換ブロック分割情報を参照して、直交変換ブロック単位にその圧縮データを逆量子化する。
 このとき、可変長復号部31により可変長復号された各ヘッダ情報を参照し、各ヘッダ情報が、当該スライスで量子化マトリクスを用いて、逆量子化処理を実施することを示している場合は、量子化マトリクスを用いて逆量子化処理を行う。
When receiving the compressed data and the prediction difference encoding parameter from the variable length decoding unit 31, the inverse quantization / inverse conversion unit 32 performs the prediction difference encoding in the same procedure as the inverse quantization / inverse conversion unit 8 of FIG. With reference to the quantization parameter and orthogonal transform block division information included in the parameters, the compressed data is inversely quantized in units of orthogonal transform blocks.
At this time, when referring to each header information variable-length decoded by the variable-length decoding unit 31, each header information indicates that the inverse quantization process is performed using the quantization matrix in the slice. Inverse quantization processing is performed using a quantization matrix.
 この際、可変長復号部31により可変長復号された各ヘッダ情報を参照して、各直交変換サイズで色信号や符号化モード(イントラ符号化かインター符号化か)毎に使用する量子化マトリクスを特定する。
 また、逆量子化・逆変換部32は、直交変換ブロック単位に逆量子化後の圧縮データである変換係数に対する逆直交変換処理を実施して、図1の逆量子化・逆変換部8から出力された局所復号予測差分信号と同一の復号予測差分信号を算出する(ステップST28)。
At this time, referring to each header information variable-length decoded by the variable-length decoding unit 31, a quantization matrix used for each color signal and coding mode (intra coding or inter coding) with each orthogonal transform size Is identified.
Further, the inverse quantization / inverse transform unit 32 performs an inverse orthogonal transform process on transform coefficients that are compressed data after inverse quantization in units of orthogonal transform blocks, and the inverse quantization / inverse transform unit 8 in FIG. A decoded prediction difference signal identical to the output local decoded prediction difference signal is calculated (step ST28).
 加算部36は、逆量子化・逆変換部32により算出された復号予測差分信号と、イントラ予測部34により生成されたイントラ予測画像PINTRAi 、または、動き補償部35により生成されたインター予測画像PINTERi のいずれか一方を加算して復号画像を算出し、その復号画像をループフィルタ部38に出力するとともに、その復号画像をイントラ予測用メモリ37に格納する(ステップST29)。
 この復号画像が、以降のイントラ予測処理の際に用いられる復号済みの画像信号になる。
Addition unit 36, decodes the prediction difference signal calculated by the inverse quantization and inverse transform unit 32, an intra prediction image P INTRAi n generated by the intra prediction unit 34 or, inter prediction generated by the motion compensation unit 35 by adding one of the image P INTERi n calculates a decoded image, and outputs the decoded image to the loop filter unit 38, and stores the decoded image to the intra prediction memory 37 (step ST29).
This decoded image becomes a decoded image signal used in the subsequent intra prediction processing.
 ループフィルタ部38は、全ての復号ブロックBに対するステップST23~ST29の処理が完了すると(ステップST30)、加算部36から出力された復号画像に対して、所定のフィルタ処理を実施して、フィルタ処理後の復号画像を動き補償予測フレームメモリ39に格納する(ステップST31)。
 具体的には、直交変換ブロックの境界や予測ブロックの境界に発生する歪みを低減するフィルタ(デブロッキングフィルタ)処理、画素単位に適応的にオフセットを加算する(画素適応オフセット)処理、ウィーナフィルタ等の線形フィルタを適応的に切り替えてフィルタ処理する適応フィルタ処理などを行う。
 ただし、ループフィルタ部38は、上記のデブロッキングフィルタ処理、画素適応オフセット処理、適応フィルタ処理のそれぞれについて、可変長復号部31により可変長復号された各ヘッダ情報を参照して、当該スライスで処理を行うか否かを特定する。
 このとき、2つ以上のフィルタ処理を行う場合に、動画像符号化装置のループフィルタ部11が図11のように構成されている場合には、図12に示すようにループフィルタ部38が構成される。
When the processing of steps ST23 to ST29 for all the decoding blocks Bn is completed (step ST30), the loop filter unit 38 performs a predetermined filtering process on the decoded image output from the adding unit 36, and filters The decoded image after processing is stored in the motion compensated prediction frame memory 39 (step ST31).
Specifically, filter (deblocking filter) processing that reduces distortion occurring at the boundaries of orthogonal transform blocks and prediction blocks, processing for adaptively adding an offset in units of pixels (pixel adaptive offset), Wiener filter, etc. The adaptive filter processing that performs the filter processing by adaptively switching the linear filter is performed.
However, the loop filter unit 38 processes each of the above-described deblocking filter processing, pixel adaptive offset processing, and adaptive filter processing with reference to each header information variable-length decoded by the variable-length decoding unit 31 and processing in the corresponding slice. Specify whether or not to perform.
At this time, when two or more filter processes are performed, if the loop filter unit 11 of the moving picture coding apparatus is configured as shown in FIG. 11, the loop filter unit 38 is configured as shown in FIG. Is done.
 ここで、デブロッキングフィルタ処理では、可変長復号部31により可変長復号されたヘッダ情報を参照し、ブロック境界にかけるフィルタ強度の選択に用いる各種パラメータを初期値から変更する情報が存在する場合には、その変更情報に基づいて、デブロッキングフィルタ処理を実施する。変更情報がない場合は、予め定められた手法に従って行う。 Here, in the deblocking filter processing, when there is information for referring to the header information that has been variable-length decoded by the variable-length decoding unit 31 and changing various parameters used for selecting the filter strength applied to the block boundary from the initial value. Performs a deblocking filter process based on the change information. When there is no change information, it is performed according to a predetermined method.
 画素適応オフセット処理では、可変長復号部31により可変長復号された画素適応オフセット処理のブロック分割情報に基づいて分割し、そのブロック単位に、可変長復号部31により可変長復号されたブロック単位のクラス分類手法を示すインデックスを参照して、そのインデックスが“オフセット処理を行わない”ことを示すインデックスでない場合、ブロック単位にブロック内の各画素を上記インデックスが示すクラス分類手法に従ってクラス分類する。
 なお、クラス分類手法の候補として、ループフィルタ部11の画素適応オフセット処理のクラス分類手法の候補と同一のものが予め用意されている。
In the pixel adaptive offset processing, division is performed based on the block division information of the pixel adaptive offset processing that has been variable length decoded by the variable length decoding unit 31, and the block unit of variable length decoding by the variable length decoding unit 31 is divided into the blocks. When an index indicating a class classification method is referred to and the index is not an index indicating that “offset processing is not performed”, each pixel in the block is classified into blocks in accordance with the class classification method indicated by the index.
In addition, the same class classification method candidate as the pixel classification method candidate of the pixel adaptive offset process of the loop filter unit 11 is prepared in advance as a class classification method candidate.
 そして、ループフィルタ部38は、ブロック単位の各クラスのオフセット値を特定する可変長復号部31により可変長復号されたオフセット情報を参照して、復号画像の輝度値にオフセットを加算する処理を行う。 Then, the loop filter unit 38 refers to the offset information that has been variable-length decoded by the variable-length decoding unit 31 that identifies the offset value of each class in block units, and performs processing to add the offset to the luminance value of the decoded image .
 適応フィルタ処理では、可変長復号部31により可変長復号されたクラス毎のフィルタを用いて、図1の動画像符号化装置と同一の手法でクラス分類した後に、そのクラス分類情報に基づいてフィルタ処理を行う。
 このループフィルタ部38によるフィルタ処理後の復号画像が、動き補償予測用の参照画像となり、また、再生画像となる。
In the adaptive filter processing, after classifying by the same method as the moving picture encoding apparatus of FIG. 1 using the filter for each class variable-length decoded by the variable-length decoding unit 31, the filter is based on the class classification information. Process.
The decoded image after the filter processing by the loop filter unit 38 becomes a reference image for motion compensation prediction and also becomes a reproduced image.
 以上で明らかなように、この実施の形態1によれば、入力映像信号をフィールド単位に符号化することを示すフラグが有効となっている場合に、イントラ予測部4が、平均値予測または水平方向予測によるイントラ予測処理を実施する時の予測ブロックの上端のフィルタ処理を実施しないようにする構成と、変換・量子化部7が、変換係数の符号化順を変更する構成とを、それぞれ単独に実施、あるいは組み合わせて実施するように構成したので、フィールド信号の特性に応じた効率的な予測処理、符号化処理を実現することができ、符号化効率を高めることができる効果を奏する。 As apparent from the above, according to the first embodiment, when the flag indicating that the input video signal is encoded in the field unit is valid, the intra prediction unit 4 performs the average value prediction or the horizontal value prediction. A configuration in which the filter processing of the upper end of the prediction block when performing intra prediction processing by direction prediction is not performed, and a configuration in which the transform / quantization unit 7 changes the encoding order of transform coefficients, respectively. Therefore, it is possible to realize efficient prediction processing and encoding processing according to the characteristics of the field signal, and to increase the encoding efficiency.
 また、この実施の形態1によれば、可変長復号部31によって復号された入力映像信号をフィールド単位に符号化することを示すフラグが有効となっている場合に、イントラ予測部34が、平均値予測または水平方向予測によるイントラ予測処理を実施する時の予測ブロックの上端のフィルタ処理を実施しないようにする構成と、逆量子化・逆変換部32が、変換係数の復号順を変更する構成とを、それぞれ単独に実施、あるいは組み合わせて実施するように構成したので、フィールド信号の特性に応じた効率的な予測処理、符号化処理を実現することができ、符号化効率を高めることができる実施の形態1の動画像符号化装置で符号化されたビットストリームを正しく復号することができる効果を奏する。 Further, according to the first embodiment, when the flag indicating that the input video signal decoded by the variable length decoding unit 31 is encoded in units of fields is valid, the intra prediction unit 34 A configuration in which the filtering process at the upper end of the prediction block when performing intra prediction processing by value prediction or horizontal prediction is not performed, and a configuration in which the inverse quantization / inverse transform unit 32 changes the decoding order of transform coefficients Are implemented separately or in combination, so that efficient prediction processing and coding processing according to the characteristics of the field signal can be realized, and coding efficiency can be improved. There is an effect that the bitstream encoded by the moving image encoding apparatus of Embodiment 1 can be correctly decoded.
 以上のように、本発明にかかる動画像符号化装置、動画像復号装置、動画像符号化方法及び動画像復号方法は、符号化効率の高い符号化、復号処理を行う動画像符号化装置、動画像復号装置等に有用である。 As described above, the moving image encoding device, the moving image decoding device, the moving image encoding method, and the moving image decoding method according to the present invention include a moving image encoding device that performs encoding and decoding processing with high encoding efficiency, This is useful for a moving picture decoding device or the like.
 1 ブロック分割部(ブロック分割手段)、2 符号化制御部(符号化制御手段)、3 切換スイッチ、4 イントラ予測部(予測手段)、5 動き補償予測部(予測手段)、6 減算部(差分画像生成手段)、7 変換・量子化部(画像圧縮手段)、8 逆量子化・逆変換部(局所復号画像生成手段)、9 加算部(局所復号画像生成手段)、10 イントラ予測用メモリ(予測手段)、11 ループフィルタ部(フィルタリング手段)、12 動き補償予測フレームメモリ(予測手段)、13 可変長符号化部(可変長符号化手段)、14 スライス分割部(スライス分割手段)、31 可変長復号部(可変長復号手段)、32 逆量子化・逆変換部(差分画像生成手段)、33 切換スイッチ、34 イントラ予測部(予測手段)、35 動き補償部(予測手段)、36 加算部(復号画像生成手段)、37 イントラ予測用メモリ(予測手段)、38 ループフィルタ部(フィルタリング手段)、39 動き補償予測フレームメモリ(予測手段)、101 ブロック分割部、102 予測部、103 圧縮部、104 局所復号部、105 加算器、106 ループフィルタ、107 メモリ、108 可変長符号化部。 1 block division unit (block division unit), 2 encoding control unit (encoding control unit), 3 changeover switch, 4 intra prediction unit (prediction unit), 5 motion compensation prediction unit (prediction unit), 6 subtraction unit (difference) (Image generation means), 7 transform / quantization section (image compression means), 8 inverse quantization / inverse transform section (local decoded image generation means), 9 addition section (local decoded image generation means), 10 intra prediction memory ( Prediction means), 11 loop filter section (filtering means), 12 motion compensated prediction frame memory (prediction means), 13 variable length coding section (variable length coding means), 14 slice division section (slice division means), 31 variable Long decoding unit (variable length decoding unit), 32 inverse quantization / inverse transform unit (difference image generation unit), 33 changeover switch, 34 intra prediction unit (prediction unit), 5 motion compensation unit (prediction unit), 36 addition unit (decoded image generation unit), 37 intra prediction memory (prediction unit), 38 loop filter unit (filtering unit), 39 motion compensation prediction frame memory (prediction unit), 101 Block division unit, 102 prediction unit, 103 compression unit, 104 local decoding unit, 105 adder, 106 loop filter, 107 memory, 108 variable length coding unit.

Claims (6)

  1.  圧縮データ及び符号化モードが多重化された符号化ビットストリームを生成する可変長符号化手段を備え、
     上記可変長符号化手段は、直交変換ブロックを直交変換サブブロックに分割し、フィールド符号化か否かを示す情報に基づくフラグが有効であるか否かに基づいて、圧縮データである量子化後の変換係数の符号化順を、前記直交変換ブロック単位、前記直交変換サブブロック単位で切り替えることを特徴とする動画像符号化装置。
    Variable length encoding means for generating an encoded bitstream in which compressed data and encoding mode are multiplexed,
    The variable length coding means divides the orthogonal transform block into orthogonal transform sub-blocks, and after quantization, which is compressed data, based on whether or not a flag based on information indicating whether or not field coding is valid. The moving picture coding apparatus is characterized in that the coding order of the transform coefficients is switched in units of the orthogonal transform block and the orthogonal transform sub-block.
  2.  入力画像を複数の部分画像であるスライスに分割するスライス分割手段と、符号化処理が実施される際の処理単位となる符号化ブロックの最大サイズを決定するとともに、最大サイズの符号化ブロックが階層的に分割される際の上限の階層数を決定し、利用可能な1以上の符号化モードの中から、階層的に分割される各々の符号化ブロックに対応する符号化モードを選択する符号化制御手段と、上記スライス分割手段により分割されたスライスを上記符号化制御手段により決定された最大サイズの符号化ブロックに分割するとともに、上記符号化制御手段により決定された上限の階層数に至るまで、上記符号化ブロックを階層的に分割するブロック分割手段と、上記ブロック分割手段により分割された符号化ブロックと上記イントラ予測手段により生成された予測画像との差分画像を生成する差分画像生成手段と、上記差分画像生成手段により生成された差分画像の変換処理を実施して、上記差分画像の変換係数を量子化し、量子化後の変換係数を圧縮データとして出力する画像圧縮手段と、上記画像圧縮手段から出力された圧縮データから差分画像を復号し、復号後の差分画像と上記予測手段により生成された予測画像を加算して局所復号画像を生成する局所復号画像生成手段と、符号化ビットストリームを生成する可変長符号化手段とを備え、
     上記可変長符号化手段は、上記画像圧縮手段から出力された圧縮データ、上記符号化制御手段により選択された符号化モード、フィールド符号化か否かを示すフラグを可変長符号化して、上記圧縮データ、上記符号化モード、上記フラグの符号化データが多重化された符号化ビットストリームを生成することを特徴とする請求項1記載の動画像符号化装置。
    A slice dividing unit that divides an input image into slices that are a plurality of partial images, and determines the maximum size of an encoding block that is a processing unit when encoding processing is performed, and the encoding block of the maximum size is hierarchical Coding for determining the upper limit number of layers when divided and selecting a coding mode corresponding to each coding block divided hierarchically from one or more available coding modes The control unit and the slice divided by the slice dividing unit are divided into coding blocks of the maximum size determined by the encoding control unit, and the upper limit number of hierarchies determined by the encoding control unit is reached. Block dividing means for hierarchically dividing the coded block, coded blocks divided by the block dividing means, and intra prediction means A difference image generation unit that generates a difference image with the predicted image generated by the method, and a conversion process of the difference image generated by the difference image generation unit, quantizes the conversion coefficient of the difference image, and performs quantization An image compression unit that outputs the subsequent transform coefficient as compressed data, a differential image is decoded from the compressed data output from the image compression unit, and the decoded differential image and the prediction image generated by the prediction unit are added. A local decoded image generating means for generating a local decoded image and a variable length encoding means for generating an encoded bitstream,
    The variable-length encoding unit performs variable-length encoding on the compressed data output from the image compression unit, the encoding mode selected by the encoding control unit, and a flag indicating whether or not the field encoding is performed. 2. The moving picture coding apparatus according to claim 1, wherein a coded bit stream in which the coded data of the data, the coding mode, and the flag is multiplexed is generated.
  3.  階層的に分割されている各々の符号化ブロックに係る圧縮データ及び符号化モードを可変長復号する可変長復号手段を備え、
     上記可変長復号手段は、直交変換ブロックを直交変換サブブロックに分割し、フィールド符号化か否かを示す情報に基づくフラグが有効であるか否かに基づいて、圧縮データである量子化後の変換係数の復号順を、前記直交変換ブロック単位、前記直交変換サブブロック単位で切り替えることを特徴とする動画像復号装置。
    Variable length decoding means for variable length decoding the compressed data and the coding mode for each coding block divided hierarchically;
    The variable length decoding means divides the orthogonal transform block into orthogonal transform sub-blocks, and determines whether or not the quantized compressed data is based on whether or not a flag based on information indicating whether or not field coding is valid. The moving picture decoding apparatus, wherein the decoding order of transform coefficients is switched in units of the orthogonal transform block and the orthogonal transform sub-block.
  4.  符号化ビットストリームに多重化された符号化データからフィールド符号化か否かを示すフラグを含むヘッダ情報を可変長復号し、上記符号化データから階層的に分割されている各々の符号化ブロックに係る圧縮データ及び符号化モードを可変長復号する可変長復号手段と、上記可変長復号手段により可変長復号された符号化ブロックに係る符号化モードに応じた予測処理を実施して予測画像を生成する予測手段と、上記可変長復号手段により可変長復号された符号化ブロックに係る圧縮データである変換係数を逆量子化し、逆量子化後の変換係数を逆変換して、圧縮前の差分画像を生成する差分画像生成手段と、上記差分画像生成手段により生成された差分画像と上記予測手段により生成された予測画像とを加算して復号画像を生成する復号画像生成手段とを備えていることを特徴とする請求項3記載の動画像復号装置。 Variable length decoding is performed on header information including a flag indicating whether or not field encoding is performed from encoded data multiplexed in an encoded bitstream, and each encoded block hierarchically divided from the encoded data is encoded. A variable length decoding unit that performs variable length decoding on the compressed data and the encoding mode, and a prediction process according to the encoding mode related to the encoding block that has been variable length decoded by the variable length decoding unit, to generate a prediction image A difference image before compression by inversely quantizing a transform coefficient that is compressed data relating to a coding block that has been variable length decoded by the variable length decoding means, and by performing a reverse transform on the transform coefficient after the inverse quantization A decoded image is generated by adding the difference image generation means for generating the difference image, the difference image generated by the difference image generation means, and the prediction image generated by the prediction means Video decoding apparatus according to claim 3, characterized in that it comprises a No. image generating means.
  5.  圧縮データ及び符号化モードが多重化された符号化ビットストリームを生成する可変長符号化ステップを備え、
     上記可変長符号化ステップは、直交変換ブロックを直交変換サブブロックに分割し、フィールド符号化か否かを示す情報に基づくフラグが有効であるか否かに基づいて、圧縮データである量子化後の変換係数の符号化順を、前記直交変換ブロック単位、前記直交変換サブブロック単位で切り替えることを特徴とする動画像符号化方法。
    A variable length encoding step for generating an encoded bitstream in which compressed data and an encoding mode are multiplexed,
    The variable-length coding step divides the orthogonal transform block into orthogonal transform sub-blocks, and after quantization, which is compressed data, based on whether or not a flag based on information indicating whether or not field coding is valid. A moving picture coding method characterized by switching the coding order of transform coefficients in units of the orthogonal transform block and the orthogonal transform sub-block.
  6.  階層的に分割されている各々の符号化ブロックに係る圧縮データ及び符号化モードを可変長復号する可変長復号ステップを備え、
     上記可変長復号ステップは、直交変換ブロックを直交変換サブブロックに分割し、フィールド符号化か否かを示す情報に基づくフラグが有効であるか否かに基づいて、圧縮データである量子化後の変換係数の復号順を、前記直交変換ブロック単位、前記直交変換サブブロック単位で切り替えることを特徴とする動画像復号方法。
    A variable-length decoding step for variable-length decoding the compressed data and the encoding mode associated with each encoding block that is hierarchically divided;
    The variable length decoding step divides the orthogonal transform block into orthogonal transform sub-blocks, and determines whether or not the quantized compressed data is based on whether or not a flag based on information indicating whether or not field coding is valid. A moving picture decoding method, wherein the decoding coefficient decoding order is switched in units of the orthogonal transform block and the orthogonal transform sub-block.
PCT/JP2013/005285 2012-09-28 2013-09-06 Video encoding device, video decoding device, video encoding method and video decoding method WO2014049982A1 (en)

Priority Applications (1)

Application Number Priority Date Filing Date Title
JP2014538129A JPWO2014049982A1 (en) 2012-09-28 2013-09-06 Moving picture encoding apparatus, moving picture decoding apparatus, moving picture encoding method, and moving picture decoding method

Applications Claiming Priority (2)

Application Number Priority Date Filing Date Title
JP2012215380 2012-09-28
JP2012-215380 2012-09-28

Publications (1)

Publication Number Publication Date
WO2014049982A1 true WO2014049982A1 (en) 2014-04-03

Family

ID=50387429

Family Applications (1)

Application Number Title Priority Date Filing Date
PCT/JP2013/005285 WO2014049982A1 (en) 2012-09-28 2013-09-06 Video encoding device, video decoding device, video encoding method and video decoding method

Country Status (2)

Country Link
JP (1) JPWO2014049982A1 (en)
WO (1) WO2014049982A1 (en)

Cited By (2)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
WO2019194485A1 (en) * 2018-04-01 2019-10-10 김기백 Method and apparatus for encoding/decoding image
CN115315958A (en) * 2020-03-30 2022-11-08 Kddi 株式会社 Image decoding device, image decoding method, and program

Citations (3)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
JP2006054846A (en) * 2004-07-12 2006-02-23 Sony Corp Coding method and device, decoding method and device, and program thereof
WO2011128268A1 (en) * 2010-04-13 2011-10-20 Fraunhofer-Gesellschaft zur Förderung der angewandten Forschung e.V. Probability interval partioning encoder and decoder
WO2011128303A2 (en) * 2010-04-13 2011-10-20 Fraunhofer-Gesellschaft zur Förderung der angewandten Forschung e.V. Coding of significance maps and transform coefficient blocks

Patent Citations (3)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
JP2006054846A (en) * 2004-07-12 2006-02-23 Sony Corp Coding method and device, decoding method and device, and program thereof
WO2011128268A1 (en) * 2010-04-13 2011-10-20 Fraunhofer-Gesellschaft zur Förderung der angewandten Forschung e.V. Probability interval partioning encoder and decoder
WO2011128303A2 (en) * 2010-04-13 2011-10-20 Fraunhofer-Gesellschaft zur Förderung der angewandten Forschung e.V. Coding of significance maps and transform coefficient blocks

Non-Patent Citations (1)

* Cited by examiner, † Cited by third party
Title
YUE YU ET AL.: "Adaptive Scan for Large Blocks for HEVC", JOINT COLLABORATIVE TEAM ON VIDEO CODING (JCT-VC) OF ITU-T SG16 WP3 AND ISO/IEC JTC1/SC29/WG11 JCTVC-F569_R2, ITU-T, 16 July 2011 (2011-07-16), pages 1 - 6 *

Cited By (3)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
WO2019194485A1 (en) * 2018-04-01 2019-10-10 김기백 Method and apparatus for encoding/decoding image
US11297309B2 (en) 2018-04-01 2022-04-05 B1 Institute Of Image Technology, Inc. Method and apparatus for encoding/decoding image
CN115315958A (en) * 2020-03-30 2022-11-08 Kddi 株式会社 Image decoding device, image decoding method, and program

Also Published As

Publication number Publication date
JPWO2014049982A1 (en) 2016-08-22

Similar Documents

Publication Publication Date Title
JP6573689B2 (en) Image coding apparatus and image coding method
JP6226863B2 (en) Image encoding device, image decoding device, image encoding method, and image decoding method
JP6082073B2 (en) Image decoding apparatus, image decoding method, image encoding apparatus, image encoding method, and bit stream
JP6716836B2 (en) Video coded data
KR101563834B1 (en) Image decoding device, image decoding method, image encoding device, and image encoding method
KR101565228B1 (en) Image encoding apparatus, image decoding apparatus, image encoding method, image decoding method, and image prediction device
WO2014049981A1 (en) Video encoding device, video decoding device, video encoding method and video decoding method
WO2014163200A1 (en) Color image encoding apparatus, color image decoding apparatus, color image encoding method, and color image decoding method
JP2014090327A (en) Moving image encoder, moving image decoder, moving image encoding method and moving image decoding method
JP2014090326A (en) Moving image encoder, moving image decoder, moving image encoding method and moving image decoding method
WO2014049982A1 (en) Video encoding device, video decoding device, video encoding method and video decoding method
JP2014204311A (en) Color image encoding device, color image decoding device, color image encoding method and color image decoding method
WO2013108882A1 (en) Video encoding apparatus, video decoding apparatus, video encoding method, and video decoding method

Legal Events

Date Code Title Description
121 Ep: the epo has been informed by wipo that ep was designated in this application

Ref document number: 13840454

Country of ref document: EP

Kind code of ref document: A1

ENP Entry into the national phase

Ref document number: 2014538129

Country of ref document: JP

Kind code of ref document: A

NENP Non-entry into the national phase

Ref country code: DE

122 Ep: pct application non-entry in european phase

Ref document number: 13840454

Country of ref document: EP

Kind code of ref document: A1