WO2011089865A1 - Image encoding method, image decoding method, device therefor, program, and integrated circuit - Google Patents

Image encoding method, image decoding method, device therefor, program, and integrated circuit Download PDF

Info

Publication number
WO2011089865A1
WO2011089865A1 PCT/JP2011/000085 JP2011000085W WO2011089865A1 WO 2011089865 A1 WO2011089865 A1 WO 2011089865A1 JP 2011000085 W JP2011000085 W JP 2011000085W WO 2011089865 A1 WO2011089865 A1 WO 2011089865A1
Authority
WO
WIPO (PCT)
Prior art keywords
filter coefficient
removal filter
coefficient
image
filter
Prior art date
Application number
PCT/JP2011/000085
Other languages
French (fr)
Japanese (ja)
Inventor
マティアス ナロスキ
寿郎 笹井
ヴィルジニー ドリゥジョーン
Original Assignee
パナソニック株式会社
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Application filed by パナソニック株式会社 filed Critical パナソニック株式会社
Publication of WO2011089865A1 publication Critical patent/WO2011089865A1/en

Links

Images

Classifications

    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04NPICTORIAL COMMUNICATION, e.g. TELEVISION
    • H04N19/00Methods or arrangements for coding, decoding, compressing or decompressing digital video signals
    • H04N19/46Embedding additional information in the video signal during the compression process
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04NPICTORIAL COMMUNICATION, e.g. TELEVISION
    • H04N19/00Methods or arrangements for coding, decoding, compressing or decompressing digital video signals
    • H04N19/60Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using transform coding
    • H04N19/61Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using transform coding in combination with predictive coding
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04NPICTORIAL COMMUNICATION, e.g. TELEVISION
    • H04N19/00Methods or arrangements for coding, decoding, compressing or decompressing digital video signals
    • H04N19/85Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using pre-processing or post-processing specially adapted for video compression
    • H04N19/86Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using pre-processing or post-processing specially adapted for video compression involving reduction of coding artifacts, e.g. of blockiness

Definitions

  • Non-Patent Document 1 an image encoding method that predictively encodes a moving image for each block and an image decoding method that decodes an image that is predictively encoded for each block have been proposed (see, for example, Non-Patent Document 1).
  • Non-Patent Document 1 a difference between a block of a moving image (original image) and a block of a predicted image predicted for the block is frequency-converted and quantized to obtain a plurality of differences. Quantized values are generated, and the plurality of quantized values are entropy encoded.
  • the above-described moving image is encoded, and as a result, an encoded bit stream is generated.
  • a difference image is generated by dequantizing the plurality of quantized values and inverse frequency transforming, and the difference image and the predicted image are added for each block to generate a reconstructed image.
  • a deblocking filter is used to remove the distortion.
  • Non-Patent Document 1 a plurality of quantized values are extracted from the encoded bitstream by entropy decoding the encoded bitstream.
  • the extracted quantized values are inversely quantized and inverse frequency transformed to generate a difference image, and the difference image and the predicted image are added for each block to generate a reconstructed image. Is done.
  • Such a reconstructed image is used to generate a predicted image, as in the image encoding method.
  • block boundary distortion in the reconstructed image is removed by filtering using the above-described deblocking filter.
  • the encoded bit stream is decoded and output as a decoded image. Further, the reconstructed image from which the block boundary distortion has been removed in this way is used to generate the predicted image described above.
  • the image coding method of Non-Patent Document 1 has a problem in that block distortion cannot be appropriately removed because there is a limit to filter coefficient selection.
  • a plurality of types of filter coefficients that can be used are prepared in advance, and based on the characteristics of the reconstructed image, the block coefficients to be processed are set. For the same filter coefficient is selected.
  • the image encoding method can be freely based on information other than the features of the reconstructed image.
  • An appropriate filter coefficient cannot be selected.
  • an appropriate filter coefficient cannot be selected even in the image decoding method. Therefore, there is a problem that block distortion cannot be removed appropriately, encoding efficiency is lowered, and image quality is deteriorated.
  • the present invention has been made in view of such problems, and an object thereof is to provide an image encoding method, an image decoding method, and the like that improve encoding efficiency and image quality.
  • an image encoding method for encoding a moving image, and encodes the moving image for each block using a predicted image.
  • a coefficient for deriving according to the feature and performing filtering using the derived distortion removal filter coefficient on the block boundary in order to generate a predicted image, and identifying the derived distortion removal filter coefficient Specific information is inserted into the encoded bitstream.
  • coefficient specifying information for specifying the distortion removal filter coefficient is inserted into the encoded bitstream, so that the same distortion removal filter coefficient is applied to the block boundary to be processed at the time of image encoding and image decoding.
  • the appropriate filter coefficient can be derived freely based on information other than the features of the reconstructed image. Therefore, block distortion can be removed appropriately, and encoding efficiency and image quality can be improved.
  • the distortion removal filter coefficient is derived by calculating the distortion removal filter coefficient, and in the insertion of the coefficient specifying information into the encoded bitstream, the derived A distortion removal filter coefficient is inserted into the coded bitstream as the coefficient specifying information.
  • distortion removal filter coefficients such as deblocking filter coefficients are calculated according to the characteristics of the block boundary, so there is no restriction on the selection of filter coefficients, and an appropriate distortion removal filter coefficient is derived for the block boundary. can do.
  • the distortion removal filter coefficients derived in this way By performing filtering using the distortion removal filter coefficients derived in this way on the block boundary (pixels on the block boundary), the distortion of the block boundary can be appropriately removed, and the original moving image Predictive images closer to each other can be generated, and as a result, encoding efficiency and image quality can be improved.
  • the derived distortion removal filter coefficient is inserted into the encoded bitstream, when the image decoding apparatus decodes the encoded bitstream, the distortion removal filter coefficient is extracted from the encoded bitstream. As in the case of encoding a moving image, filtering using the distortion removal filter coefficient can be performed on the block boundary. As a result, the image quality of the decoded image can be improved.
  • the image encoding method further determines whether or not to derive the distortion removal filter coefficient according to a feature of a block boundary, and determines the distortion removal filter coefficient in the determination of the distortion removal filter derivation. Is derived in order to derive an internal noise removal filter coefficient for removing noise in the processing target block corresponding to the block boundary, and to generate the predicted image.
  • filtering using the internal noise removal filter coefficient is performed on the processing target block, and the derived internal noise removal filter coefficient is inserted into the encoded bitstream.
  • the distortion removal filter coefficient is derived in the processing target block together with the distortion of the block boundary.
  • the filter coefficient for removing the noise is derived as the distortion removal filter coefficient, and in the filtering using the distortion removal filter coefficient, the filtering using the filter coefficient is performed on the block boundary and the processing target block.
  • a filter coefficient for removing noise in the processing target block is derived as a distortion removal filter coefficient together with distortion at the block boundary. Therefore, the characteristic as the distortion removal filter coefficient and the characteristic as the internal noise removal filter coefficient are obtained.
  • Combined filter coefficients can be derived. Since filtering using the filter coefficient is performed on the block boundary and the processing target block, it is necessary to perform filtering on the block boundary and filtering on the processing target block sequentially (sequentially) by switching the filter coefficient. And filtering on them at once. For example, a deblocking filter can be applied at once with a winner filter. As a result, filtering can be performed easily and quickly.
  • the filter coefficient is derived.
  • the image encoding method further includes, for each block boundary in the image region, a filter mode corresponding to the block boundary according to the feature of the block boundary, from among a plurality of predetermined filter modes.
  • the common filter coefficient is derived for a plurality of block boundaries corresponding to the filter mode included in the image region for each filter mode.
  • a common filter coefficient is derived for a plurality of block boundaries corresponding to the filter mode. For example, a separate filter is applied to each block boundary included in an image region such as a frame. There is no need to derive coefficients, and the amount of calculation for deriving filter coefficients can be reduced.
  • the image encoding method further counts the number of block boundaries corresponding to the filter mode included in the image area for each filter mode, and determines the filter mode with the smallest counted number as the number. Is changed to the filter mode having the second smallest number, and the filter mode having the smallest number is combined with the filter mode having the second smallest number.
  • the common filter coefficient is derived for a plurality of block boundaries corresponding to the filter mode.
  • the derivation of the filter coefficient for the filter mode with the smallest number can be omitted.
  • the amount of calculation for deriving the coefficient can be further reduced.
  • the derivation of the filter coefficients is omitted in the filter mode having the smallest number of block boundaries, so that the influence of the image quality on the image region can be minimized.
  • the filter mode having the fewest number and the filter mode having the second smallest number are repeatedly combined, and the insertion of the coefficient specifying information into the coded bitstream is further repeated. Is inserted into the encoded bitstream.
  • the image encoding method further includes changing the filter mode of any one of the plurality of predetermined filter modes to another filter mode of the plurality of filter modes. Combining a filter mode with the other filter mode, and deriving the distortion removal filter coefficient, the common filter coefficient is derived for a plurality of block boundaries corresponding to the two combined filter modes, In the insertion of the coefficient specifying information into the encoded bitstream, a combined index for specifying the combined two filter modes is further inserted into the encoded bitstream.
  • the derivation of the filter coefficient for one filter mode can be omitted, and the amount of calculation for deriving the filter coefficient can be further reduced.
  • a combined index for specifying the two combined filter modes is inserted into the encoded bitstream, when the image decoding apparatus decodes the encoded bitstream, the combined index is combined from the encoded bitstream. An index is extracted, and it is possible to grasp which filter mode is combined based on the combined index. As a result, the image decoding apparatus can perform filtering on the same filter mode as the filter mode used for filtering when encoding a moving image.
  • the filtering using the distortion removal filter coefficient for each pixel at the block boundary, it is determined whether or not the pixel should be filtered based on the difference in pixel value between the pixels, and the filtering is performed. Filtering is performed on pixels determined to be performed.
  • filtering is performed only on the necessary pixels among the pixels at the block boundary, so that it is possible to reduce the amount of filtering calculation and to further improve the encoding efficiency and image quality. it can.
  • the image encoding method further derives a predicted value for the distortion removal filter coefficient by predicting the derived distortion removal filter coefficient, and inserts the coefficient identification information into the encoded bitstream. The difference between the distortion removal filter coefficient and the predicted value is inserted into the encoded bitstream.
  • the difference between the distortion removal filter coefficient and the predicted value is inserted into the encoded bit stream, the code amount of the encoded bit stream can be suppressed.
  • the distortion removal filter coefficient is derived for each color component and filtering direction.
  • the distortion removal filter coefficient is derived using an arithmetic expression used for derivation of the filter coefficient of the Wiener filter.
  • the predicted image can be brought closer to the original moving image, and the encoding efficiency and the image quality can be further improved.
  • an image decoding method for decoding an encoded bitstream, in which encoded blocks included in the encoded bitstream are sequentially re-executed.
  • Coefficient specifying information for specifying a distortion removal filter coefficient for removing distortion of the block boundary according to the feature of the block boundary, which is a boundary between the configured and reconstructed blocks, from the coded bitstream Extraction is performed, and filtering using the distortion removal filter coefficient specified by the extracted coefficient specifying information is performed on the block boundary.
  • the coefficient specifying information for specifying the distortion removing filter coefficient is extracted from the encoded bitstream according to the feature of the block boundary, and the filtering using the distortion removing filter coefficient specified by the coefficient specifying information is performed. Is performed on the block boundary, so that the distortion removal filter coefficient used for the filtering at the time of encoding the moving image can also be used for the filtering at the time of decoding. As a result, the image quality of the decoded image can be improved. Can do.
  • a filter coefficient for removing noise in the processing target block corresponding to the block boundary is extracted as the distortion removing filter coefficient together with the distortion of the block boundary, and the distortion removing filter coefficient
  • filtering using the filter coefficient is performed on the block boundary and the processing target block.
  • the filter coefficient for removing noise in the processing target block is extracted as the distortion removal filter coefficient together with the distortion at the block boundary. Therefore, the characteristic as the distortion removal filter coefficient and the characteristic as the internal noise removal filter coefficient are obtained. Combined filter coefficients can be extracted. Since filtering using the filter coefficient is performed on the block boundary and the processing target block, it is necessary to perform filtering on the block boundary and filtering on the processing target block sequentially (sequentially) by switching the filter coefficient. And filtering on them at once. For example, a deblocking filter can be applied at once with a winner filter. As a result, filtering can be performed easily and quickly.
  • the present invention can be realized not only as such an image encoding method and image decoding method, but also in an image encoding device, an image decoding device, an integrated circuit, and a computer that perform image processing according to those methods.
  • the present invention can also be realized as a program for executing image processing according to the method and a recording medium for storing the program. Moreover, you may combine how to solve the above subjects how.
  • FIG. 1A is a block diagram of an image encoding device according to an aspect of the present invention.
  • FIG. 1B is a flowchart illustrating an image encoding method according to an aspect of the present invention.
  • FIG. 2A is a block diagram of an image decoding apparatus according to an aspect of the present invention.
  • FIG. 2B is a flowchart illustrating an image decoding method according to an aspect of the present invention.
  • FIG. 3 is a block diagram of the image coding apparatus according to Embodiment 1 of the present invention.
  • FIG. 4 is a diagram for explaining block boundary specification and filter mode determination according to Embodiment 1 of the present invention.
  • FIG. 5 is a diagram showing pixels on the block boundary in the first embodiment of the present invention.
  • FIG. 6 is a diagram for explaining the filter mode in the first embodiment of the present invention.
  • FIG. 8 is a flowchart showing processing for calculating the filter coefficient by switching the weighting coefficient ⁇ for each filter mode according to Embodiment 1 of the present invention.
  • FIG. 9 is a diagram showing filter coefficients for each filter mode according to Embodiment 1 of the present invention.
  • FIG. 10 is a diagram for explaining filter mode coupling in Embodiment 1 of the present invention.
  • FIG. 11 is a diagram illustrating a syntax regarding a noise removal filter of an encoded bitstream according to Embodiment 1 of the present invention.
  • FIG. 11 is a diagram illustrating a syntax regarding a noise removal filter of an encoded bitstream according to Embodiment 1 of the present invention.
  • FIG. 12 is a block diagram of the image decoding apparatus according to Embodiment 1 of the present invention.
  • FIG. 13 is a diagram for explaining pixels to be filtered according to the first modification of the first embodiment of the present invention.
  • FIG. 14 is a diagram showing a filter mode combination table according to the second modification of the first embodiment of the present invention.
  • FIG. 15 is an explanatory diagram for describing filter coefficient designation values according to the third modification of the first embodiment of the present invention.
  • FIG. 16 is a schematic diagram illustrating an example of the overall configuration of a content supply system that implements a content distribution service.
  • FIG. 17 is a diagram illustrating an appearance of a mobile phone.
  • FIG. 18 is a block diagram illustrating a configuration example of a mobile phone.
  • FIG. 19 is a schematic diagram illustrating an example of the overall configuration of a digital broadcasting system.
  • FIG. 20 is a block diagram illustrating a configuration example of a television.
  • FIG. 21 is a block diagram illustrating a configuration example of an information reproducing / recording unit that reads and writes information from and on a recording medium that is an optical disk.
  • FIG. 22 is a diagram illustrating a structure example of a recording medium that is an optical disk.
  • FIG. 23 is a block diagram illustrating a configuration example of an integrated circuit that realizes the image encoding method and the image decoding method according to each embodiment.
  • FIG. 1A is a block diagram of an image encoding device according to an aspect of the present invention.
  • the image encoding apparatus 10 is an apparatus that encodes a moving image, and includes an encoding unit 11, a reconstruction unit 12, a filter coefficient derivation unit 13, a filtering unit 14, and an insertion unit 15.
  • the encoding unit 11 generates an encoded bitstream by encoding a moving image for each block using a predicted image.
  • the reconstruction unit 12 sequentially reconstructs the encoded blocks.
  • the filter coefficient deriving unit 13 derives a distortion removal filter coefficient for removing distortion at a block boundary, which is a boundary between reconstructed blocks, according to the feature of the block boundary.
  • the filtering unit 14 performs filtering on the block boundary using the derived distortion removal filter coefficient in order to generate the predicted image.
  • the insertion unit 15 inserts coefficient specifying information for specifying the derived distortion removal filter coefficient into the encoded bitstream.
  • FIG. 1B is a flowchart showing processing of the image encoding device 10 according to an aspect of the present invention.
  • the image encoding device 10 first generates an encoded bitstream by encoding a moving image block by block using a predicted image (step S11). Next, the image encoding device 10 sequentially reconstructs the encoded blocks (step S12). Next, the image coding apparatus 10 derives a distortion removal filter coefficient for removing distortion at the block boundary, which is a boundary between the reconstructed blocks, according to the feature of the block boundary (step S13). Next, the image encoding device 10 performs filtering on the block boundary using the derived distortion removal filter coefficient in order to generate the above-described predicted image (step S14). Further, the image encoding device 10 inserts coefficient specifying information for specifying the derived distortion removal filter coefficient into the encoded bitstream (step S15).
  • coefficient specifying information for specifying the distortion removal filter coefficient is inserted into the encoded bitstream, so that at the time of image encoding and image decoding An appropriate filter coefficient can be derived freely based on information other than the characteristics of the reconstructed image (reconstructed block) while using the same distortion removal filter coefficient for the block boundary to be processed. Therefore, block distortion can be removed appropriately, and encoding efficiency and image quality can be improved.
  • the filter coefficient deriving unit 13 derives the distortion removal filter coefficient by calculating the distortion removal filter coefficient, and the insertion unit 15 encodes the encoded bitstream using the derived distortion removal filter coefficient as coefficient specifying information. May be inserted.
  • distortion removal filter coefficients such as the filter coefficient of the deblocking filter are derived (calculated) according to the feature of the block boundary.
  • the distortion removal filter coefficient appropriate for the block boundary can be derived.
  • the distortion of the block boundary can be appropriately removed, and the original moving image Predictive images closer to each other can be generated, and as a result, encoding efficiency and image quality can be improved.
  • the derived distortion removal filter coefficient is inserted into the encoded bitstream, when the image decoding apparatus decodes the encoded bitstream, the distortion removal filter coefficient is extracted from the encoded bitstream.
  • filtering using the distortion removal filter coefficient can be performed on the block boundary. As a result, the image quality of the decoded image can be improved.
  • the unit for reconfiguration and filtering may be one slice, but may be a smaller unit.
  • the information amount of the filter coefficient can be reduced.
  • the unit is smaller (for example, several blocks), the delay time required for encoding can be shortened.
  • the filter coefficient deriving unit 13 derives the distortion removal filter coefficient by selecting the above-described distortion removal filter coefficient from among a plurality of distortion removal filter coefficient candidates, and the insertion unit 15 derives the distortion removal filter coefficient.
  • An index indicating the distortion removal filter coefficient may be inserted into the encoded bitstream as coefficient specifying information. Further, the filter coefficient deriving unit 13 may calculate a distortion removal filter coefficient and select a distortion removal filter coefficient closest to the calculated distortion removal filter coefficient from the plurality of candidates described above.
  • FIG. 2A is a block diagram of an image decoding apparatus according to an aspect of the present invention.
  • the image decoding device 20 is a device that decodes an encoded bit stream, and includes a reconstruction unit 21, an extraction unit 22, and a filtering unit 23.
  • the reconstruction unit 21 sequentially reconstructs the encoded blocks included in the encoded bitstream.
  • the extraction unit 22 generates coefficient specifying information for specifying a distortion removal filter coefficient for removing distortion at the block boundary in accordance with the feature of the block boundary that is a boundary between the reconstructed blocks. Extract from The filtering unit 23 performs filtering using the distortion removal filter coefficient specified by the extracted coefficient specifying information on the block boundary.
  • FIG. 2B is a flowchart showing processing of the image decoding device 20 according to an aspect of the present invention.
  • the image decoding device 20 sequentially reconstructs the encoded blocks included in the encoded bitstream (step S21). Next, the image decoding device 20 generates coefficient specifying information for specifying a distortion removing filter coefficient for removing distortion at the block boundary according to the feature of the block boundary that is a boundary between the reconstructed blocks. Extract from the encoded bit stream (step S22). Next, the image decoding device 20 performs filtering on the block boundary using the distortion removal filter coefficient specified by the extracted coefficient specifying information (step S23).
  • coefficient specifying information for specifying the distortion removal filter coefficient is extracted from the encoded bitstream according to the feature of the block boundary, and the coefficient specifying Since the filtering using the distortion removal filter coefficient specified by the information is performed on the block boundary, the distortion removal filter coefficient used for the filtering at the time of encoding the moving image may be used for the filtering at the time of decoding. As a result, the image quality of the decoded image can be improved.
  • the extraction unit 22 may extract the above-described distortion removal filter coefficient from the encoded bitstream as coefficient specifying information.
  • the extraction unit 22 extracts an index indicating the distortion removal filter coefficient from the encoded bitstream as coefficient specifying information, and the filtering unit 23 converts the distortion removal filter coefficient indicated by the extracted index into a plurality of distortions. Filtering using the selected distortion removal filter coefficient may be performed on the block boundary by selecting from the candidates for the removal filter coefficient.
  • FIG. 3 is a block diagram of the image coding apparatus according to Embodiment 1 of the present invention.
  • An image encoding device 100 is a device that generates an encoded bitstream by predictively encoding an original image S that is a moving image for each block, and includes a subtracter 101 and a transform quantization unit 102. , An entropy coding unit 103, an inverse quantization inverse transform unit 104, an adder 105, a noise removal filter 106, and a prediction unit 107.
  • the subtracter 101 acquires the original image S and the predicted image S ⁇ for each block, and calculates a prediction error image e that is the difference between them.
  • the transform quantizing unit 102 generates a coefficient block including a plurality of frequency coefficients by performing orthogonal transform (frequency transform) such as discrete cosine transform on the prediction error image e. Further, the transform quantization unit 102 generates a quantization block including a plurality of quantization values by performing quantization on each of the plurality of frequency coefficients included in the coefficient block.
  • orthogonal transform frequency transform
  • quantization unit 102 generates a quantization block including a plurality of quantization values by performing quantization on each of the plurality of frequency coefficients included in the coefficient block.
  • the entropy encoding unit 103 performs entropy encoding on a plurality of quantization values included in the quantization block. By performing such entropy encoding, an encoded bit stream is generated.
  • the inverse quantization inverse transform unit 104 generates a coefficient block including a plurality of frequency coefficients by performing inverse quantization on each of the plurality of quantization values included in the quantization block. Further, the inverse quantization inverse transform unit 104 generates a prediction error image e ′ in block units by performing inverse orthogonal transform (inverse frequency transform) such as inverse discrete cosine transform on the coefficient block.
  • inverse orthogonal transform inverse frequency transform
  • the adder 105 acquires the prediction error image e ′ and the prediction image S ⁇ for each block, and generates the reconstructed image S ′ by adding them. As a result, the encoded blocks are sequentially reconstructed.
  • the noise removal filter 106 acquires the reconstructed image S ′ and the predicted image S ⁇ , and based on these images, the distortion of the block boundary of the reconstructed image S ′ and the processing target block corresponding to the block boundary. Remove the included noise. That is, the noise removal filter 106 functions as a deblocking filter and a Wiener filter. Further, the noise removal filter 106 calculates a plurality of filter coefficients used for filtering for each image region such as a frame based on the original image S.
  • the prediction unit 107 generates the prediction image S ⁇ in units of blocks by performing intra prediction or inter prediction using the reconstructed image S ′ from which noise has been removed by the noise removal filter 106.
  • the image encoding device 10 in FIG. 1A corresponds to the image encoding device 100 in the present embodiment
  • the encoding unit 11 in FIG. 1A includes the subtractor 101 and the transform quantization unit 102 in the present embodiment.
  • An entropy encoding unit 103 and a prediction unit 107. 1A includes the inverse quantization inverse transform unit 104 and the adder 105 in the present embodiment, and the filter coefficient deriving unit 13 and the filtering unit 14 in FIG. This corresponds to the removal filter 106.
  • the insertion part 15 of FIG. 1A consists of a part of function of the entropy encoding part 103 in this Embodiment.
  • the noise removal filter 106 performs the following processes (1) to (6).
  • FIG. 4 is a diagram for explaining block boundary specification and filter mode determination.
  • the noise removal filter 106 identifies block boundaries on the left side and the upper side of the processing target block for each processing target block in the image area. Further, the noise removal filter 106 determines (selects) a filter mode for the identified block boundary. For example, the noise removal filter 106 determines the filter mode m1 corresponding to the feature of the block boundary for the left block boundary, and the filter mode m2 corresponding to the feature of the block boundary for the upper block boundary. To decide. That is, the noise removal filter 106 determines the filter mode m1 as the horizontal filter mode for the processing target block, and determines the filter mode m2 as the vertical filter mode for the processing target block. Details of the filter mode will be described later.
  • the noise removal filter 106 specifies four pixels arranged in the horizontal direction with the left block boundary in the center as candidates for filtering by the filter mode m1. Since there are a plurality of such pixel groups composed of four pixels along the block boundary, each of these pixel groups is specified as a candidate for filtering. Similarly, the noise removal filter 106 specifies four pixels arranged in the vertical direction across the upper block boundary as candidates for filtering by the filter mode m2. In addition, since there are a plurality of such pixel groups including four pixels along the block boundary, each of these pixel groups is specified as a filtering target candidate. Here, the noise removal filter 106 determines whether or not to perform filtering on each of the pixels specified as candidates by performing an operation described later.
  • the noise removal filter 106 calculates a filter coefficient corresponding to the filter mode by performing an operation to be described later for each filter mode determined as described above for each of the block boundaries in the image region. That is, for each filter mode, the noise removal filter 106 calculates a common filter coefficient (distortion removal filter coefficient) for one or more block boundaries corresponding to the fill mode. The noise removal filter 106 derives a predicted value by predicting the filter coefficient thus calculated, and quantizes the difference between the predicted value and the filter coefficient. The noise removal filter 106 outputs the quantized difference to the entropy encoding unit 103. The entropy encoding unit 103 performs entropy encoding on the quantized difference and inserts the difference into the encoded bitstream. Further, the noise removal filter 106 performs filtering using the filter coefficient calculated in the above (4) on the pixel determined to be filtered in the above (3).
  • FIG. 5 is a diagram showing pixels on the block boundary.
  • the block p and the block q are adjacent to each other.
  • the six pixels p 2 , p 1 , p 0 , q 0 , q 1 , and q 2 are each a sample, and the horizontal direction with the boundary (block boundary) between the block p and the block q in the middle Or they are arranged along the vertical direction.
  • a pixel p 0 , a pixel p 1 , and a pixel p 2 are arranged in order from the block boundary side.
  • the pixel q 0 , the pixel q 1 , and the pixel q 2 are sequentially arranged from the block boundary side. Are arranged.
  • the pixel values of the pixels p 2 , p 1 , p 0 , q 0 , q 1 , and q 2 in the reconstructed image S ′ are p 2, s ′ , p 1, s ′ , p 0, s ' , q0 , s' , q1 , s ' , and q2 , s' .
  • the pixel values of the pixels p 2 , p 1 , p 0 , q 0 , q 1 , and q 2 in the predicted image S ⁇ are p 2, s ⁇ , p 1, s ⁇ , p 0, s ⁇ , Q 0, s ⁇ , q 1, s ⁇ and q 2, s ⁇ .
  • the noise removal filter 106 determines that filtering should be performed on each of the four pixels p 1 , p 0 , q 0 , and q 1 on the block boundary of the reconstructed image S ′.
  • filtering is performed on each of those pixels (filtering target pixels).
  • the pixel values (p ′ 1 , p ′ 0 , q ′ 0 , and q ′ 1 ) of the filtered pixels p 1 , p 0 , q 0 , and q 1 are calculated.
  • a 1, m, ⁇ , a 4, m, b 1, m, ⁇ , b 4, m, c 1, m, ⁇ , c 4 , M , d 1, m ,..., D 4, m , o 1, m , and o 2, m are filter coefficients calculated for each filter mode for an image region such as a frame. . That is, the noise removal filter 106 adaptively calculates these filter coefficients according to the feature of the block boundary. These filter coefficients are calculated independently for each of the color components (luminance component Y, color difference component U, and color difference component V), and are also independent for each of the filtering directions (horizontal direction and vertical direction). Is calculated. Of the luminance component Y, the color difference component U, and the color difference component V, the filter coefficients of the color difference component U and the color difference component V may be made common.
  • FIG. 6 is a diagram for explaining the filter mode.
  • the filter mode m0 (DBF_SKIP) is a so-called skip mode, and is a mode that is applied when a specific condition that does not require filtering for removing distortion at the block boundary is satisfied.
  • This specific condition is a condition that a block boundary is a boundary of an object in an image, or a condition that distortion generated by encoding at the block boundary is assumed to be inconspicuous.
  • the specific condition is that the motion vector used for motion compensation and the reference picture are the same in adjacent blocks, and that no frequency coefficient (quantized value) is transmitted, or adjacent to each other.
  • This is a condition that the edge vector detected by using the pixel of the block to be processed has a norm larger than the threshold and passes through the block boundary.
  • the filter mode m1 (DBF_INTRA_QUANT) is a so-called intra mode, and is a mode applied when one of the adjacent blocks is intra-coded.
  • the filter mode m2 (DBF_PRED_SIGNIF) is applied when both adjacent blocks are inter-coded and the sum of the number of DCT coding coefficients (non-zero frequency coefficients) of both blocks is greater than the threshold. Mode.
  • both adjacent blocks are inter-coded with the same (or substantially the same) motion vector, and the total number of DCT coding coefficients of both blocks is 1 or more. This mode is applied to.
  • both adjacent blocks are inter-coded with different motion vectors or inter-template matching coded, and the total number of DCT coding coefficients of both blocks is 1. This mode is applied when the above is true.
  • inter template matching encoding is performed by using already encoded and decoded pixels adjacent to the encoding target block (for example, adjacent pixels on the left and above) from among encoded and decoded frames. In this encoding, the positions of pixels similar to those pixels are found, and the pixel corresponding to the lower right block of the similar pixels is used as a predicted image.
  • the filter mode m5 (DBF_MOT_DISC) is applied when both adjacent blocks are inter-coded with different motion vectors or inter-template matching coded, and both blocks have no DCT coding coefficient. Mode.
  • Filter mode m6 is a mode applied when both adjacent blocks are inter-coded with different motion vectors and two intensity-related parameters are different between both blocks.
  • the intensity related parameter is a parameter for correcting the pixel value, and includes a scale Sc and an offset value Of.
  • Filter mode m7 (DBF_IC_INTERMED) is a mode applied when both adjacent blocks are inter-coded with the same motion vector and two intensity related parameters are different between both blocks.
  • Filter mode m8 (DBF_IC_WEAK) is a mode applied when both adjacent blocks are inter-coded with different motion vectors and at least one intensity-related parameter is different between both blocks.
  • the filter mode m9 (DBF_BS_PLUS) is a mode applied when both adjacent blocks are inter-coded with different motion vectors and the merge mode is used for two pixels at the block boundary.
  • the merge mode is to move a predetermined area (for example, 2 pixels) in a block using motion information (motion vector and reference image) of the block in the direction (for example, left or upward direction) indicated by the index. This is a method for compensation (motion prediction).
  • the noise removal filter 106 has a predetermined filter mode corresponding to a block boundary for each block boundary in the image area according to the feature (condition) of the block boundary.
  • a filter coefficient is selected from the types of filter modes, and for each filter mode, a common filter coefficient is derived for a plurality of block boundaries corresponding to the filter mode included in the image region.
  • a common filter coefficient is derived for a plurality of block boundaries corresponding to the filter mode. Therefore, for example, each of the block boundaries included in the image region such as a frame. Therefore, it is not necessary to derive individual filter coefficients, and the amount of calculation for deriving filter coefficients can be reduced.
  • the noise removal filter 106 calculates a filter coefficient using two evaluation values for each filter mode as described above.
  • One of the two evaluation values is an evaluation value E1 for objective image quality, and is a value corresponding to the mean square of the difference between the original image S and the deconstructed reconstructed image S ′.
  • Another of the two evaluation values is an evaluation value E2 for subjective image quality, which is a value indicating the smoothness of the block boundary.
  • the noise removal filter 106 calculates the filter coefficient so that the evaluation value E1 becomes small and the evaluation value E2 becomes small.
  • the evaluation value E1 is calculated by the following (Formula 5).
  • p 1, org , p 0, org , q 0, org , and q 1, org are the pixels p 1 , p 0 , q 0 , and q 1 in the original image S, respectively. Value.
  • calculation using four pixels at the block boundary is shown, but the evaluation value E1 is included in an image region such as a frame for each filter mode, not only for the four pixels. It is calculated for all pixels on all block boundaries corresponding to the filter mode.
  • the evaluation value E1 indicated by (Equation 5) is also used for calculating the filter coefficient of the Wiener filter. That is, in the Wiener filter, the filter coefficient is calculated so that the evaluation value E1 is the smallest.
  • Evaluation value E2 is calculated by the following (formula 6). Further, in (Equation 6), calculation using four pixels at the block boundary is shown, but the evaluation value E2 is included in an image region such as a frame for each filter mode, not only for the four pixels. It is calculated for all pixels on all block boundaries corresponding to the filter mode.
  • the noise removal filter 106 calculates the filter coefficient for each filter mode by performing the following (Equation 7).
  • is a weighting coefficient set by the filter mode.
  • the noise removal filter 106 calculates the filter coefficient so that the sum of the product of the evaluation value E1 and the weighting coefficient ⁇ and the product of the evaluation value E2 and (1 ⁇ weighting coefficient ⁇ ) is minimized.
  • the filter mode is the skip mode, that is, the filter mode m1 (DBF_SKIP)
  • the noise removal filter 106 derives the distortion removal filter coefficient using the arithmetic expression (Formula 5) used for derivation of the filter coefficient of the Wiener filter.
  • the encoding efficiency and the image quality can be further improved.
  • the noise removal filter 106 sets the weighting factor ⁇ to 1 when calculating the skip mode filter coefficient.
  • the noise removal filter 106 has 4 ⁇ 4 pixels whose block boundary between the blocks p and q is located at the center. Is processed as a processing target block, and a filter coefficient for removing noise in the processing target block is calculated. That is, the noise removal filter 106 calculates a filter coefficient (internal noise removal filter coefficient) for applying the Wiener filter to the processing target block.
  • FIG. 8 is a flowchart showing processing for calculating the filter coefficient by switching the weighting coefficient ⁇ for each filter mode.
  • the noise removal filter 106 determines whether or not the processing target filter mode selected in step S100 is the intra mode (DBF_INTRA_QUANT). (Step S106).
  • the noise removal filter 106 determines that the mode is the intra mode (DBF_INTRA_QUANT) (Y in step S106)
  • a filter coefficient is calculated such that the Wiener filter strength is greater than the deblocking filter strength for the filter modes other than the skip mode and the intra mode.
  • a filter coefficient having both deblocking filter properties and Wiener filter properties is calculated.
  • the noise removal filter 106 determines whether or not there is an unselected filter mode among the above ten filter modes m0 to m9 (step S114). Here, if it is determined that there is an unselected filter mode (Y in step S114), the noise removal filter 106 repeatedly executes the processing from step S100, and if it is determined that there is no unselected filter mode (in step S114). N) The filter coefficient calculation process for the image area such as a frame is terminated.
  • filtering on the block boundary using distortion removal filter coefficients such as filter coefficients of a deblocking filter and internal noise removal such as filter coefficients of a Wiener filter, for example, according to the characteristics of the block boundary.
  • the filtering for the processing target block using the filter coefficient is performed by switching. Therefore, it is possible to select and apply appropriate filtering according to the state of the image, and further improve the encoding efficiency and the image quality.
  • the noise removal filter 106 determines the ratio between the strength for suppressing the distortion of the block boundary and the strength for suppressing the noise in the processing target block as the weighting factor ⁇ according to the feature of the block boundary. (Steps S104, S108, and S110), and filter coefficients are derived according to the ratio (Step S112).
  • FIG. 9 is a diagram showing filter coefficients for each filter mode.
  • the noise removal filter 106 applies the filter coefficient a 1 to each filter mode for an image region such as one frame as shown in FIG. , M ,..., A 4, m , b 1, m ,..., B 4, m , c 1, m ,..., C 4, m , d 1, m ,. 4, m 1 , o 1, m , and o 2, m are calculated. Further, the filter coefficient for each filter mode is calculated independently for each of the filtering direction (horizontal direction and vertical direction) and color components (luminance component Y, color difference component U, and color difference component V).
  • the noise removal filter 106 performs filtering using the filter coefficient calculated as described above, the pixels p 1 , p 0 , q 0 that are candidates for filtering in the reconstructed image S ′ in advance. , Q 1 , whether or not to perform filtering is determined by performing operations shown in the following (Expression 8) to (Expression 11).
  • the noise removal filter 106 when 'the pixel value p 2 of the pixel p 2 and p 0 in, s' reconstructed image S and p 0, s' is satisfies the following conditions of (Equation 10), It is determined that filtering should be performed on the pixel p1 of the reconstructed image S ′.
  • the noise removal filter 106 may adaptively determine the offset values (Offset A and Offset B ).
  • the noise removal filter 106 depends on whether the blocks p and q adjacent to each other are motion-compensated at different positions, that is, whether motion compensation is performed with different motion vectors or reference images. Determine the offset value. If the noise removal filter 106 determines that motion compensation is performed at different positions, the noise removal filter 106 increases the offset value from a predetermined value so that a candidate pixel to be filtered is selected as a filtering target. On the other hand, if it is determined that motion compensation is performed at the same position, the noise removal filter 106 decreases the offset value from a predetermined value so that a candidate pixel to be filtered is not selected as a filtering target.
  • the noise removal filter 106 may determine an offset value depending on whether or not the blocks p and q adjacent to each other are encoded in different prediction modes.
  • the noise removal filter 106 determines that the blocks p and q are encoded in different prediction modes when the block p is encoded in the intra prediction mode and the block q is encoded in the inter prediction mode. To do. Further, the noise removal filter 106 encodes the blocks p and q in different prediction modes when the block p is encoded in the vertical intra prediction mode and the block q is encoded in the horizontal intra prediction mode. Is determined.
  • the offset value is determined from a predetermined value so that a candidate pixel to be filtered is selected as a filtering target. increase.
  • the offset value is set to a predetermined value so that the candidate pixel to be filtered is not selected as the filtering target. Decrease from.
  • the offset value determined in this way is determined according to the feature of the image, it is determined to be the same value in the image encoding device 100 and the image decoding device that decodes the encoded bitstream. Therefore, the image encoding device 100 does not need to transmit the offset value to the image decoding device.
  • the image encoding apparatus 100 may determine an arbitrary offset value and transmit the offset value included in the encoded bitstream.
  • the noise removal filter 106 of the image encoding device 100 can adjust the image quality in consideration of the code amount, the quantization error, and the like. Further, the noise removal filter 106 of the image encoding device 100 calculates a difference between an arbitrary offset value and an offset value determined by the motion compensation or prediction mode, and transmits the difference included in the encoded bitstream. May be. In this case, the image quality can be adjusted with an arbitrary offset value, and an increase in the code amount of the encoded bit stream can be suppressed.
  • the noise removal filter 106 has filter coefficients a 1, m ,..., A 4, m , b 1, m ,..., B 4, m , c 1, m ,. .., C 4, m , d 1, m ,..., D 4, m , o 1, m , o 2, m are predicted to predict the filter coefficients (prediction target filter coefficients). Is derived. Furthermore, the noise removal filter 106 calculates the difference between the prediction target filter coefficient and the prediction value for each prediction target filter coefficient. Then, the noise removal filter 106 quantizes the difference (filter prediction error coefficient) and outputs the result to the entropy encoding unit 103. The entropy encoding unit 103 acquires the quantized filter prediction error coefficient output from the noise removal filter 106, performs entropy encoding, and inserts it into the encoded bitstream.
  • Fixed noise removal filter 106 for example, the filter coefficients a 1, m, ⁇ , a 4, m, c 1, m, ⁇ , and with respect to c 4, each m is a predetermined Use the value as the predicted value. Moreover, the noise removal filter 106, the filter coefficients b 1, m, ⁇ , b 4, m, d 1, m, ⁇ , and with respect to d 4, each of m, the filter coefficient a 1, m, ⁇ , a 4, m , c 1, m, is used., and c 4, m as the predicted value.
  • the noise removal filter 106 derives a prediction value a 1, m that is the filter coefficient a 1, m by predicting the filter coefficient b 1, m, and predicts the filter coefficient d 1, m , A prediction value c 1, m that is a filter coefficient c 1, m is derived.
  • the noise removal filter 106 quantizes the difference between the prediction target filter coefficient b 1, m and the prediction value a 1, m as a filter prediction error coefficient, and the prediction target filter coefficient d 1, m and the prediction value c 1, m. Is quantized as a filter prediction error coefficient.
  • the noise removal filter 106 the filter coefficients a 1, m, ⁇ , a 4, m, c 1, m, ⁇ , and c 4, the value constant is multiplied by each of the m , filter coefficients b 1, m, ⁇ , b 4, m, d 1, m, ⁇ , and d 4, may be used as the predicted value for each m.
  • the noise removal filter 106 can use twelve different values as quantization step sizes used for quantization of the filter prediction error coefficient. Such a quantization step size is the same value for all filter prediction error coefficients (filter coefficients) corresponding to one color component and one filtering direction, and is entropy encoded and included in the encoded bitstream. And transmitted to the image decoding apparatus.
  • the noise removal filter 106 may select a quantization step size used for quantization from among 12 values of quantization step size by RD (Rate-Distortion) optimization. That is, the noise removal filter 106 has a quantization step size so that the difference between the original image S and the filtered reconstructed image S ′ is small and the trade-off balance between code amount and distortion is optimal. Select.
  • FIG. 10 is a diagram for explaining the combination of the filter modes.
  • the noise removal filter 106 counts the number of boundaries corresponding to each of the filter modes m0 to m6 in the image area. As shown in FIG. 10, when the number of boundaries of the filter mode m5 is the smallest and the number of boundaries of the filter mode m6 is the second smallest, the noise removal filter 106 couples the filter mode m5 to the filter mode m6. In other words, the noise removal filter 106 changes the filter mode m5 to the filter mode m6. As a result, the noise removal filter 106 does not need to derive and encode a filter coefficient group including a plurality of filter coefficients individually for the filter mode m5 and the filter mode m6, and the filter mode (the filter mode m5 and the filter mode m6).
  • the filter coefficient group described above the filter coefficients a 1, m, ⁇ , a 4, m, b 1, m, ⁇ , b 4, m, c 1, m, ⁇ , c 4 , M , d 1, m ,..., D 4, m , o 1, m , and o 2, m .
  • RD Rate-Distortion
  • the noise removal filter 106 outputs a flag (0 or 1) indicating whether or not the filter mode is applied to the image area to the entropy encoding unit 103 for each filter mode. Further, when at least one filter mode is applied, the noise removal filter 106 is a filter for each of the number of times the filter mode is combined (number of times of combination), the quantization step size, and the applied filter mode.
  • the coefficient group (a plurality of quantized filter prediction error coefficients) is output to the entropy coding unit 103. As a result, the number of combinations, the quantization step size, and the filter coefficient group are entropy-coded, inserted into the encoded bitstream, and transmitted to the image decoding apparatus.
  • NUM_LOOP_MODES_DBF indicates a predetermined number of filter modes (for example, 10).
  • yuv indicates a color component
  • hv is a flag indicating a filtering direction (horizontal direction or vertical direction).
  • Mode_on [yuv] [hv] [mode] is a flag indicating whether or not filtering is performed according to the color component, filtering direction, and filter mode indicated by [yuv], [hv], and [mode]. That is, mode_on [yuv] [hv] [mode] is inserted into the encoded bitstream for each combination of color component, filtering method, and filter mode.
  • fixed_filter [yuv] [hv] [mode] may be used as an implicit filter coefficient group.
  • fixed_filter [yuv] [hv] [mode] may be replaced with diff_fixed_filter [yuv] [hv] [mode].
  • the diff_fixed_filter [yuv] [hv] [mode] is, for example, the difference between the fixed_filter [yuv] [hv] [mode] to be processed and the fixed_filter [yuv] [hv] [mode] processed immediately before. .
  • the image coding apparatus 100 and the image decoding apparatus derive fixed_filter [yuv] [hv] [mode] using diff_fixed_filter [yuv] [hv] [mode], respectively.
  • the code amount of the encoded bit stream can be reduced.
  • FIG. 12 is a block diagram of the image decoding apparatus in the present embodiment.
  • the image decoding apparatus 200 is an apparatus that decodes the encoded bitstream generated by the image encoding apparatus 100, and includes an entropy decoding unit 201, an inverse quantization inverse transform unit 202, an adder 203, and a prediction A unit 204 and a noise removal filter 206 are provided.
  • the entropy decoding unit 201 sequentially generates quantized blocks including a plurality of quantized values by performing entropy decoding on the encoded bit stream. Further, the entropy decoding unit 201 extracts a filter coefficient group (distortion removal filter coefficient) corresponding to the block boundary (filter mode) from the encoded bitstream in accordance with the feature (filter mode) of the block boundary, and the filter The coefficient group is output to the noise removal filter 206.
  • each filter coefficient included in the filter coefficient group is inserted into the encoded bitstream as a filter prediction error coefficient that is quantized and entropy-coded.
  • the entropy decoding unit 201 when extracting the filter coefficient from the encoded bit stream, performs entropy decoding on the filter prediction error coefficient and extracts the filter prediction error coefficient from the encoded bit stream. Furthermore, the entropy decoding unit 201 extracts the number of times of combination and the quantization step size from the encoded bitstream, and outputs them to the noise removal filter 206 together with the above-described filter coefficient (quantized filter prediction error coefficient).
  • the inverse quantization inverse transform unit 202 generates a coefficient block including a plurality of frequency coefficients by performing inverse quantization on each of the plurality of quantization values included in the quantization block. Further, the inverse quantization inverse transform unit 202 performs the inverse orthogonal transform (inverse frequency transform) such as inverse discrete cosine transform on the coefficient block, thereby generating the prediction error image e ′ in units of blocks.
  • inverse orthogonal transform inverse frequency transform
  • the adder 203 acquires the prediction error image e ′ and the prediction image S ⁇ for each block, and generates a reconstructed image S ′ by adding them. As a result, the encoded blocks included in the encoded bitstream are sequentially reconstructed.
  • the noise removal filter 206 acquires the reconstructed image S ′ and the predicted image S ⁇ from the adder 203 and the prediction unit 204, and further acquires the filter coefficient, the number of combinations, and the quantization step size from the entropy decoding unit 201.
  • the filter coefficient is acquired from the entropy decoding unit 201 as the filter prediction error coefficient quantized as described above. Therefore, the noise removal filter 206 dequantizes the filter prediction error coefficient using the quantization step size. Further, as described above, the noise removal filter 206 derives a prediction value by predicting the filter coefficient, and restores the filter coefficient by adding the prediction value and the dequantized filter prediction error coefficient. To do. Further, the noise removal filter 206 repeatedly executes the filter mode combination for the number of times of combination.
  • the noise removal filter 206 like the noise removal filter 106 of the image encoding device 100, is based on the reconstructed image S ′, the predicted image S ⁇ , and the filter coefficient group, and the block boundary of the reconstructed image S ′, Filtering is performed on the processing target block corresponding to the block boundary.
  • the noise removal filter 206 removes block boundary distortion and noise included in the processing target block. That is, the noise removal filter 206 functions as a deblocking filter and a Wiener filter.
  • the prediction unit 204 generates a prediction image S ⁇ in units of blocks by performing intra prediction or inter prediction using the reconstructed image S 'from which noise has been removed by the noise removal filter 206.
  • the reconstruction unit 21 in FIG. 2A corresponds to the image decoding device 200 in the present embodiment
  • the reconstruction unit 21 in FIG. 2A is derived from the inverse quantization inverse transform unit 202 and the adder 203 in the present embodiment.
  • the extraction part 22 of FIG. 2A consists of a part of function of the entropy decoding part 201 in this Embodiment
  • the filtering part 23 of FIG. 2A is equivalent to the noise removal filter 206 in this Embodiment.
  • the filter coefficient (distortion removal filter coefficient) is extracted from the encoded bitstream in accordance with the feature of the block boundary, and filtering using the filter coefficient is performed on the block. Since it is performed on the boundary, the filter coefficient used for filtering at the time of encoding a moving image can also be used for filtering at the time of decoding, and as a result, the image quality of the decoded image can be improved.
  • the noise removal filter 106 in the above embodiment performs filtering on pixels at a block boundary (a plurality of pixels arranged on the left and right or top and bottom with the block boundary in the middle).
  • the removal filter 106 also performs filtering on pixels on the block boundary and on pixels in the processing target block q.
  • FIG. 13 is a diagram for explaining pixels to be filtered in the present modification.
  • the noise removal filter 106 includes pixels p 0 , p 1 , q 0 , and q 1 arranged on the left and right with a block boundary in the center, and a processing target block q Filtering is performed on the pixels q 2 to q n ⁇ 3 inside.
  • the processing target block q is composed of n ⁇ n pixels.
  • the noise removal filter 106 performs filtering on the pixels q 2 to q n ⁇ 3 in the processing countermeasure block q by performing the following calculations (Equation 13) to (Equation 16).
  • the noise removal filter 106 calculates the filter coefficient by (Expression 5) to (Expression 7), and substitutes the calculated filter coefficient into (Expression 13) to (Expression 16), thereby processing the processing target block q. Filtering may be performed on the pixels q 2 to q n ⁇ 3 inside.
  • the noise removal filter 206 of the image decoding apparatus 200 may perform filtering in the same manner as the noise removal filter 106 according to this modification.
  • the noise removal filter 106 may calculate the evaluation value E1 by the following (Expression 17) or (Expression 18) instead of the above (Expression 12).
  • the weighting factor ⁇ corresponding to the filter mode is set.
  • the calculation method of the evaluation value E1 and the evaluation value E2 may be set or selected according to the filter mode.
  • a calculation formula for the evaluation value E1 is selected from (Formula 5), (Formula 12), (Formula 17), and (Formula 18) according to the filter mode, and the evaluation value E1 is based on the selected calculation formula. May be calculated.
  • Modification 2 Here, a second modification of the present embodiment will be described.
  • the number of block boundaries (boundary number) corresponding to the filter mode in the image area is counted, and the filter modes are combined according to the number of boundaries. Modes may be combined.
  • the noise removal filter 106 combines filter modes having similar conditions shown in FIG. 6, that is, filter modes having similar block boundary characteristics. Then, the noise removal filter 106 refers to the filter mode combination table indicating the combination relationship of the filter modes and an index (combination index) for specifying the combination relationship, and determines the combination index corresponding to the combined filter mode.
  • the data is output to the entropy encoding unit 103.
  • the combined index is included in the encoded bitstream and transmitted to the image decoding apparatus 200.
  • the noise removal filter 206 of the image decoding apparatus 200 holds the same table as the filter mode combination table referred to by the noise removal filter 106. Therefore, the noise removal filter 206 can grasp which filter mode is combined by acquiring the combination index included in the encoded bitstream and referring to the filter mode combination table. .
  • FIG. 14 is a diagram showing a filter mode combination table.
  • the combination index Idxm0 indicating that no filter mode is combined
  • the filter mode m3 is combined with the filter mode m2.
  • a combined index Idxm1 indicating that the filter mode m3 is combined with the filter mode m2
  • a combined index Idxm2 indicating that the filter mode m6 is combined with the filter mode m5.
  • the noise removal filter 106 first determines whether or not at least one of these filter modes should be combined with another filter mode by RD optimization or the like. Here, if it is determined that the noise removal filter 106 should not be combined, the noise removal filter 106 performs filtering by calculating a filter coefficient for each filter mode. Then, the noise removal filter 106 outputs the combined index Idxm0 to the entropy encoding unit 103, thereby inserting the combined index Idxm0 into the encoded bitstream. Thereby, the noise removal filter 206 of the image decoding apparatus 200 acquires the combined index Idxm0, and can grasp that the filter mode is not combined based on the combined index Idxm0. As a result, the noise removal filter 206 of the image decoding apparatus 200 performs filtering using the filter coefficient corresponding to each of the seven filter modes.
  • the noise removal filter 106 when it is determined that the noise removal filter 106 should be combined, at least one of the filter modes is combined with another filter mode. For example, if the block boundary feature corresponding to the filter mode m3 is similar to the block boundary feature corresponding to the filter mode m2, the noise removal filter 106 couples the filter mode m3 to the filter mode m2. In other words, the noise removal filter 106 applies the filtering of the filter mode m2 to the block boundary that satisfies the conditions of the filter mode m3. Then, the noise removal filter 106 outputs the combined index Idxm1 indicating the content of the combination to the entropy encoding unit 103, thereby including the combined index Idxm1 in the encoded bitstream.
  • the noise removal filter 206 of the image decoding apparatus 200 acquires the combined index Idxm1 of the encoded bitstream, and the filter mode m3 is combined with the filter mode m2 based on the combined index Idxm1 and the filter mode combination table. I can know that. As a result, the noise removal filter 206 of the image decoding apparatus 200 performs filtering using the filter coefficient corresponding to the filter mode for each filter mode except the filter mode m3.
  • the filter mode is combined according to the feature of the block boundary, so that the code amount of the filter coefficient can be reduced, and more appropriate filtering is performed to improve the encoding efficiency and the image quality. can do.
  • Modification 3 a third modification of the present embodiment will be described.
  • the filter modes are combined and a common filter coefficient group is inserted into the encoded bit stream for the plurality of filter modes.
  • the overhead may be reduced by other methods.
  • overhead is reduced by using a list and a filter coefficient designation value.
  • FIG. 15 is an explanatory diagram for explaining a list and filter coefficient designation values according to this modification.
  • the noise removal filter 106 first outputs a list of filter coefficient groups used for filtering to the entropy encoding unit 103.
  • This list includes a plurality of filter coefficient groups (gc1, gc2,..., Gck) used for filtering, and each filter coefficient included in the filter coefficient group is a filter prediction quantized with the above-described quantization step size. Expressed as an error coefficient. Therefore, the noise removal filter 106 outputs the quantization step size together with the above list. As a result, the list and the quantization step size are inserted into the encoded bit stream.
  • the noise removal filter 106 outputs a filter coefficient designated value corresponding to the filter mode for each filter mode.
  • the filter coefficient designation value corresponding to the predetermined filter mode indicates 0.
  • the filter coefficient designation value corresponding to the filter mode indicates a value of 1 or more.
  • the filter coefficient designation value corresponding to the predetermined filter mode indicates a value for specifying a filter coefficient group (for example, a group of 18 filter coefficients) corresponding to the filter mode from the list. For example, when the filter coefficient designated value corresponding to the filter mode is 1, the filter coefficient designated value indicates that the filter coefficient group gc1 transmitted first in the list is used for the filter mode. When the filter coefficient designation value corresponding to another filter mode is 2, the filter coefficient group gc2 transmitted second in the list is used for the other filter mode. Indicates.
  • the noise removal filter 106 specifies, for each filter mode, a filter coefficient group used for filtering by the filter mode for one color component and one filtering direction.
  • the filter coefficient designation value which is a value, is output to the entropy encoding unit 103. Further, the number of filter coefficient groups included in the above list is equal to the maximum value of the filter coefficient designated value.
  • the entropy encoding unit 103 entropy-encodes the filter coefficient designation value, quantization step size, and list output as described above and inserts them into the encoded bitstream.
  • the entropy decoding unit 201 of the image decoding apparatus 200 performs entropy decoding on the above-described filter coefficient designation value, quantization step size, and list included in the encoded bitstream. Then, the noise removal filter 206 performs inverse quantization using a quantization step size for each of the quantized filter prediction error coefficients included in the list. Further, the noise removal filter 206 restores a filter coefficient (filter coefficient group) by adding a prediction value to each of the inversely quantized filter prediction error coefficients. For each filter mode, the noise removal filter 206 selects a filter coefficient group from the list based on the filter coefficient designation value corresponding to the filter mode, and performs filtering using the filter coefficient group.
  • the weighting factor ⁇ corresponding to the filter mode is set for each filter mode, but the tap length corresponding to the filter mode may be set for each filter mode.
  • filtering is performed using four pixels (4 taps) as shown in (Expression 1) to (Expression 4).
  • filtering may be performed using six pixels (6 taps).
  • these weighting factors are merely examples, and other weighting factors may be set.
  • the filter mode is the skip mode (DBF_SKIP)
  • a common filter Wiener filter
  • Wiener filters having different strengths or properties to each processing target block or each block boundary corresponding to the skip mode.
  • Wiener filters having different filter coefficients and tap lengths may be applied.
  • the image encoding apparatus 100 inserts an index indicating what strength or property of the Wiener filter has been applied to the encoded bitstream.
  • the image decoding apparatus 200 can also apply a Wiener filter having the same strength or property as the Wiener filter applied by the image encoding apparatus 100.
  • ten types of filter modes are used, but eleven or more types of filter modes or nine or less types of filter modes may be used.
  • a frame is used as an example of an image area.
  • the image area may be a slice or a sequence.
  • the noise removal filter 106 of the image coding apparatus 100 calculates a filter coefficient group by performing the calculation of (Equation 7), and calculates the filter coefficient group from among at least one candidate filter coefficient group held in advance. A candidate closest to the set of filter coefficients is selected.
  • the noise removal filter 106 performs filtering using the selected candidate filter coefficient group, and outputs an index for identifying or specifying the filter coefficient group to the entropy encoding unit 103.
  • the entropy encoding unit 103 performs entropy encoding on the index and inserts the index into the encoded bit stream.
  • the entropy decoding unit 201 of the image decoding apparatus 200 extracts an index from the encoded bitstream by entropy decoding the encoded bitstream, and outputs the index to the noise removal filter 206.
  • the noise removal filter 206 selects a filter coefficient group identified or specified by the extracted index from among at least one candidate filter coefficient group held in advance, and uses the selected filter coefficient group. Filtering.
  • the filter coefficient group held in advance may be a filter coefficient group that has already been calculated.
  • the noise coefficient filter 106 assigns an index to the calculated filter coefficient group, and holds the filter coefficient group to which the index is attached.
  • the noise removal filter 106 performs filtering using the filter coefficient group to which the index is attached, and outputs the filter coefficient group to which the index is attached to the entropy coding unit 103.
  • the entropy encoding unit 103 performs entropy encoding on the filter coefficient group to which the index is attached and inserts the filter coefficient group into the encoded bitstream.
  • the entropy decoding unit 201 of the image decoding device 200 performs entropy decoding on the encoded bit stream, extracts a filter coefficient group with an index from the encoded bit stream, and outputs the filter coefficient group to the noise removal filter 206.
  • the noise removal filter 206 holds the extracted filter coefficient group with an index and performs filtering using the filter coefficient group.
  • the storage medium may be any medium that can record a program, such as a magnetic disk, an optical disk, a magneto-optical disk, an IC card, and a semiconductor memory.
  • FIG. 16 is a diagram showing an overall configuration of a content supply system ex100 that realizes a content distribution service.
  • the communication service providing area is divided into desired sizes, and base stations ex106 to ex110, which are fixed radio stations, are installed in each cell.
  • the content supply system ex100 includes a computer ex111, a PDA (Personal Digital Assistant) ex112, a camera ex113, a mobile phone ex114, a game machine via the Internet ex101, the Internet service provider ex102, the telephone network ex104, and the base stations ex106 to ex110. Each device such as ex115 is connected.
  • PDA Personal Digital Assistant
  • each device may be directly connected to the telephone network ex104 without going through the base stations ex106 to ex110 which are fixed wireless stations.
  • the devices may be directly connected to each other via short-range wireless or the like.
  • the camera ex113 is a device that can shoot moving images such as a digital video camera
  • the camera ex116 is a device that can shoot still images and movies such as a digital camera.
  • the mobile phone ex114 is a GSM (Global System for Mobile Communications) method, a CDMA (Code Division Multiple Access) method, a W-CDMA (Wideband-Code Division Multiple Access L (Semiconductor Access) method, a W-CDMA (Wideband-Code Division Multiple Access L method, or a high access rate).
  • GSM Global System for Mobile Communications
  • CDMA Code Division Multiple Access
  • W-CDMA Wideband-Code Division Multiple Access L (Semiconductor Access) method
  • W-CDMA Wideband-Code Division Multiple Access L method
  • a high access rate A High Speed Packet Access
  • PHS Personal Handyphone System
  • the camera ex113 and the like are connected to the streaming server ex103 through the base station ex109 and the telephone network ex104, thereby enabling live distribution and the like.
  • the content for example, music live video
  • the streaming server ex103 streams the content data transmitted to the requested client.
  • the client include a computer ex111, a PDA ex112, a camera ex113, a mobile phone ex114, a game machine ex115, and the like that can decode the encoded data.
  • Each device that has received the distributed data decodes and reproduces the received data.
  • the encoded processing of the captured data may be performed by the camera ex113, the streaming server ex103 that performs the data transmission processing, or may be performed in a shared manner.
  • the decryption processing of the distributed data may be performed by the client, the streaming server ex103, or may be performed in a shared manner.
  • still images and / or moving image data captured by the camera ex116 may be transmitted to the streaming server ex103 via the computer ex111.
  • the encoding process in this case may be performed by any of the camera ex116, the computer ex111, and the streaming server ex103, or may be performed in a shared manner.
  • these encoding processing and decoding processing are generally performed in a computer ex111 and an LSI (Large Scale Integration) ex500 included in each device.
  • the LSI ex500 may be configured as a single chip or a plurality of chips.
  • image encoding and image decoding software is incorporated into some recording medium (CD-ROM, flexible disk, hard disk, etc.) that can be read by the computer ex111 and the like, and the encoding processing and decoding processing are performed using the software. May be.
  • moving image data acquired by the camera may be transmitted. The moving image data at this time is data encoded by the LSI ex500 included in the mobile phone ex114.
  • the streaming server ex103 may be a plurality of servers or a plurality of computers, and may process, record, and distribute data in a distributed manner.
  • the encoded data can be received and reproduced by the client.
  • the information transmitted by the user can be received, decrypted and reproduced in real time by the client, and even a user who does not have special rights or facilities can realize personal broadcasting.
  • the image encoding method or the image decoding method shown in the above embodiment may be used for encoding and decoding of each device constituting the content supply system.
  • FIG. 17 is a diagram illustrating the mobile phone ex114 using the image encoding method and the image decoding method described in the above embodiment.
  • the cellular phone ex114 includes an antenna ex601 for transmitting and receiving radio waves to and from the base station ex110, a video from a CCD camera, a camera unit ex603 capable of taking a still image, a video shot by the camera unit ex603, and an antenna ex601.
  • a display unit ex602 such as a liquid crystal display that displays data obtained by decoding received video and the like, a main body unit composed of a group of operation keys ex604, an audio output unit ex608 such as a speaker for outputting audio, and a voice input Audio input unit ex605 such as a microphone, recorded moving image or still image data, received mail data, moving image data or still image data, etc., for storing encoded data or decoded data
  • Recording media ex607 can be attached to media ex607 and mobile phone ex114 And a slot unit ex606 for.
  • the recording medium ex607 stores a flash memory element, which is a kind of EEPROM, which is a nonvolatile memory that can be electrically rewritten and erased, in a plastic case such as an SD card.
  • the mobile phone ex114 has a power supply circuit ex710, an operation input control unit ex704, an image encoding unit, and a main control unit ex711 configured to control the respective units of the main body unit including the display unit ex602 and the operation key ex604.
  • Unit ex712, camera interface unit ex703, LCD (Liquid Crystal Display) control unit ex702, image decoding unit ex709, demultiplexing unit ex708, recording / reproducing unit ex707, modulation / demodulation circuit unit ex706, and audio processing unit ex705 are connected to each other via a synchronization bus ex713. It is connected.
  • the power supply circuit ex710 activates the camera-equipped digital mobile phone ex114 by supplying power to each unit from the battery pack. .
  • the cellular phone ex114 converts the audio signal collected by the audio input unit ex605 in the audio call mode into digital audio data by the audio processing unit ex705 based on the control of the main control unit ex711 including a CPU, a ROM, a RAM, and the like.
  • the modulation / demodulation circuit unit ex706 performs spread spectrum processing, the transmission / reception circuit unit ex701 performs digital analog conversion processing and frequency conversion processing, and then transmits the result via the antenna ex601.
  • the cellular phone ex114 amplifies the received data received by the antenna ex601 in the voice call mode, performs frequency conversion processing and analog-digital conversion processing, performs spectrum despreading processing by the modulation / demodulation circuit unit ex706, and performs analog speech by the voice processing unit ex705. After the data is converted, it is output via the audio output unit ex608.
  • text data of the e-mail input by operating the operation key ex604 on the main body is sent to the main control unit ex711 via the operation input control unit ex704.
  • the main control unit ex711 performs spread spectrum processing on the text data in the modulation / demodulation circuit unit ex706, performs digital analog conversion processing and frequency conversion processing in the transmission / reception circuit unit ex701, and then transmits the text data to the base station ex110 via the antenna ex601.
  • the image data captured by the camera unit ex603 is supplied to the image encoding unit ex712 via the camera interface unit ex703.
  • the image data captured by the camera unit ex603 can be directly displayed on the display unit ex602 via the camera interface unit ex703 and the LCD control unit ex702.
  • the image encoding unit ex712 is configured to include the image encoding device described in the present invention, and an encoding method using the image data supplied from the camera unit ex603 in the image encoding device described in the above embodiment. Is converted into encoded image data by compression encoding and sent to the demultiplexing unit ex708. At the same time, the mobile phone ex114 sends the sound collected by the sound input unit ex605 during imaging by the camera unit ex603 to the demultiplexing unit ex708 via the sound processing unit ex705 as digital sound data.
  • the demultiplexing unit ex708 multiplexes the encoded image data supplied from the image encoding unit ex712 and the audio data supplied from the audio processing unit ex705 by a predetermined method, and the resulting multiplexed data is a modulation / demodulation circuit unit Spread spectrum processing is performed in ex706, digital analog conversion processing and frequency conversion processing are performed in the transmission / reception circuit unit ex701, and then transmission is performed via the antenna ex601.
  • the received data received from the base station ex110 via the antenna ex601 is subjected to spectrum despreading processing by the modulation / demodulation circuit unit ex706, and the resulting multiplexing is obtained.
  • Data is sent to the demultiplexing unit ex708.
  • the demultiplexing unit ex708 separates the multiplexed data into a bit stream of image data and a bit stream of audio data, and a synchronization bus
  • the encoded image data is supplied to the image decoding unit ex709 via ex713 and the audio data is supplied to the audio processing unit ex705.
  • the image decoding unit ex709 is configured to include the image decoding device described in the present application, and is reproduced by decoding the bit stream of the image data with a decoding method corresponding to the encoding method described in the above embodiment.
  • Moving image data is generated and supplied to the display unit ex602 via the LCD control unit ex702, thereby displaying, for example, moving image data included in a moving image file linked to a home page.
  • the audio processing unit ex705 converts the audio data into analog audio data, and then supplies the analog audio data to the audio output unit ex608.
  • the audio data included in the moving image file linked to the home page is reproduced.
  • a decoding device can be incorporated. Specifically, in the broadcasting station ex201, audio data, video data, or a bit stream in which those data are multiplexed is transmitted to a communication or broadcasting satellite ex202 via radio waves. In response, the broadcasting satellite ex202 transmits a radio wave for broadcasting, and a home antenna ex204 having a satellite broadcasting receiving facility receives the radio wave, and the television (receiver) ex300 or the set top box (STB) ex217 or the like. The device decodes the bitstream and reproduces it.
  • audio data, video data recorded on a recording medium ex215 such as DVD or BD, or an encoded bit stream in which those data are multiplexed are read and decoded, or audio data, video data or these are recorded on the recording medium ex215.
  • the image decoding apparatus or the image encoding apparatus described in the above embodiment can also be mounted on the reader / recorder ex218 that encodes the data and records the multiplexed data as multiplexed data.
  • the reproduced video signal is displayed on the monitor ex219.
  • the recording medium ex215 on which the encoded bit stream is recorded allows other devices and systems to reproduce the video signal.
  • the other reproduction device ex212 can reproduce the video signal on the monitor ex213 using the recording medium ex214 on which the encoded bitstream is copied.
  • an image decoding device may be mounted in the set-top box ex217 connected to the cable ex203 for cable television or the antenna ex204 for satellite / terrestrial broadcasting and displayed on the monitor ex219 of the television.
  • the image decoding apparatus may be incorporated in the television instead of the set top box.
  • FIG. 20 is a diagram illustrating a television (receiver) ex300 that uses the image decoding method and the image encoding method described in the above embodiment.
  • the television ex300 obtains or outputs a bit stream of video information via the antenna ex204 or the cable ex203 that receives the broadcast, and a tuner ex301 that outputs or outputs the encoded data that is received or demodulated.
  • Modulation / demodulation unit ex302 that modulates data for transmission to the outside, and multiplexing / separation unit ex303 that separates demodulated video data and audio data, or multiplexes encoded video data and audio data Is provided.
  • the television ex300 decodes each of the audio data and the video data, or encodes the respective information, the audio signal processing unit ex304, the signal processing unit ex306 including the video signal processing unit ex305, and the decoded audio signal. And an output unit ex309 including a display unit ex308 such as a display for displaying the decoded video signal.
  • the television ex300 includes an interface unit ex317 including an operation input unit ex312 that receives an input of a user operation.
  • the television ex300 includes a control unit ex310 that controls each unit in an integrated manner, and a power supply circuit unit ex311 that supplies power to each unit.
  • the interface unit ex317 includes a bridge ex313 connected to an external device such as a reader / recorder ex218, a slot unit ex314 for enabling recording media ex216 such as an SD card, and an external recording such as a hard disk
  • a driver ex315 for connecting to a medium, a modem ex316 for connecting to a telephone network, and the like may be included.
  • the recording medium ex216 is capable of electrically recording information by using a nonvolatile / volatile semiconductor memory element to be stored.
  • Each part of the television ex300 is connected to each other via a synchronous bus.
  • the television ex300 receives a user operation from the remote controller ex220 or the like, and demultiplexes the video data and audio data demodulated by the modulation / demodulation unit ex302 by the multiplexing / separation unit ex303 based on the control of the control unit ex310 having a CPU or the like. . Furthermore, the television ex300 decodes the separated audio data by the audio signal processing unit ex304, and the separated video data is decoded by the video signal processing unit ex305 using the decoding method described in the above embodiment. The decoded audio signal and video signal are output to the outside from the output unit ex309.
  • these signals may be temporarily stored in the buffers ex318, ex319, etc. so that the audio signal and the video signal are reproduced in synchronization.
  • the television ex300 may read the encoded bitstream encoded from the recording media ex215 and ex216 such as a magnetic / optical disk and an SD card, not from a broadcast or the like. Next, a configuration will be described in which the television ex300 encodes an audio signal and a video signal and transmits them to the outside or writes them to a recording medium or the like.
  • the television ex300 receives a user operation from the remote controller ex220 or the like, and encodes an audio signal with the audio signal processing unit ex304 based on the control of the control unit ex310, and the video signal with the video signal processing unit ex305 in the above embodiment. Encoding is performed using the described encoding method.
  • the encoded audio signal and video signal are multiplexed by the multiplexing / demultiplexing unit ex303 and output to the outside. When multiplexing, these signals may be temporarily stored in the buffers ex320 and ex321 so that the audio signal and the video signal are synchronized.
  • a plurality of buffers ex318 to ex321 may be provided as shown in the figure, or a configuration in which one or more buffers are shared may be used. Further, in addition to the illustrated example, data may be stored in the buffer as a buffer material that prevents system overflow and underflow even between the modulation / demodulation unit ex302 and the multiplexing / demultiplexing unit ex303, for example.
  • the television ex300 In addition to acquiring audio data and video data from broadcast and recording media, the television ex300 has a configuration for receiving AV input of a microphone and a camera, and even if encoding processing is performed on the data acquired therefrom Good.
  • the television ex300 has been described as a configuration capable of the above-described encoding processing, multiplexing, and external output. However, all of these processing cannot be performed, and the above reception, decoding processing, and external
  • the configuration may be such that only one of the outputs is possible.
  • the decoding process or the encoding process may be performed by either the television ex300 or the reader / recorder ex218.
  • the television ex300 and the reader / recorder ex218 may be shared with each other.
  • FIG. 21 shows the configuration of the information reproducing / recording unit ex400 when data is read from or written to an optical disk.
  • the information reproducing / recording unit ex400 includes elements ex401 to ex407 described below.
  • the optical head ex401 irradiates a laser spot on the recording surface of the recording medium ex215 that is an optical disc to write information, and detects information reflected from the recording surface of the recording medium ex215 to read the information.
  • the modulation recording unit ex402 electrically drives a semiconductor laser built in the optical head ex401 and modulates the laser beam according to the recording data.
  • the reproduction demodulator ex403 amplifies a reproduction signal obtained by electrically detecting reflected light from the recording surface by a photodetector built in the optical head ex401, separates and demodulates a signal component recorded on the recording medium ex215, and is necessary. To play back information.
  • the buffer ex404 temporarily holds information to be recorded on the recording medium ex215 and information reproduced from the recording medium ex215.
  • the disk motor ex405 rotates the recording medium ex215.
  • the servo control unit ex406 moves the optical head ex401 to a predetermined information track while controlling the rotational drive of the disk motor ex405, and performs a laser spot tracking process.
  • the system control unit ex407 controls the entire information reproduction / recording unit ex400.
  • the system control unit ex407 uses various types of information held in the buffer ex404, and generates and adds new information as necessary. This is realized by recording / reproducing information through the optical head ex401 while the unit ex403 and the servo control unit ex406 are operated cooperatively.
  • the system control unit ex407 includes, for example, a microprocessor, and executes these processes by executing a read / write program.
  • the optical head ex401 has been described as irradiating a laser spot, but it may be configured to perform higher-density recording using near-field light.
  • FIG. 22 shows a schematic diagram of a recording medium ex215 that is an optical disk.
  • Guide grooves grooves
  • address information indicating the absolute position on the disc is recorded in advance on the information track ex230 by changing the shape of the groove.
  • This address information includes information for specifying the position of the recording block ex231 which is a unit for recording data, and the recording and reproducing apparatus specifies the recording block by reproducing the information track ex230 and reading the address information. be able to.
  • the recording medium ex215 includes a data recording area ex233, an inner peripheral area ex232, and an outer peripheral area ex234.
  • the area used for recording user data is the data recording area ex233, and the inner circumference area ex232 and the outer circumference area ex234 arranged on the inner circumference or outer circumference of the data recording area ex233 are used for specific purposes other than recording user data. Used.
  • the information reproducing / recording unit ex400 reads / writes encoded audio data, video data, or encoded data obtained by multiplexing these data, with respect to the data recording area ex233 of the recording medium ex215.
  • an optical disk such as a single-layer DVD or BD has been described as an example.
  • the present invention is not limited to these, and an optical disk having a multilayer structure and capable of recording other than the surface may be used. It also has a structure that performs multidimensional recording / reproduction, such as recording information using light of various different wavelengths at the same location on the disc, and recording different layers of information from various angles. It may be an optical disk.
  • the car ex210 having the antenna ex205 can receive data from the satellite ex202 and the like, and the moving image can be reproduced on a display device such as the car navigation ex211 that the car ex210 has.
  • the configuration of the car navigation ex211 may be, for example, a configuration in which a GPS receiving unit is added in the configuration illustrated in FIG.
  • the mobile phone ex114 and the like can be used in three ways: a transmitting terminal having only an encoder and a receiving terminal having only a decoder. The implementation form of can be considered.
  • the image encoding method or the image decoding method described in the above embodiment can be used in any of the above-described devices and systems, and by doing so, the effects described in the above embodiment can be obtained. be able to.
  • FIG. 23 shows a configuration of an LSI ex500 that is made into one chip.
  • the LSI ex500 includes elements ex501 to ex509 described below, and each element is connected via a bus ex510.
  • the power supply circuit unit ex505 starts up to an operable state by supplying power to each unit when the power supply is in an on state.
  • the LSI ex500 when performing the encoding process, inputs an AV signal from the microphone ex117, the camera ex113, and the like by the AV I / Oex 509 based on the control of the control unit ex501 having the CPU ex502, the memory controller ex503, the stream controller ex504, and the like. Accept.
  • the input AV signal is temporarily stored in an external memory ex511 such as SDRAM.
  • the accumulated data is appropriately divided into a plurality of times according to the processing amount and the processing speed, and sent to the signal processing unit ex507.
  • the signal processing unit ex507 performs encoding of an audio signal and / or encoding of a video signal.
  • the encoding process of the video signal is the encoding process described in the above embodiment.
  • the signal processing unit ex507 further performs processing such as multiplexing the encoded audio data and the encoded video data according to circumstances, and outputs the result from the stream I / Oex 506 to the outside.
  • the output bit stream is transmitted to the base station ex107 or written to the recording medium ex215. It should be noted that data should be temporarily stored in the buffer ex508 so that the data is synchronized when multiplexed.
  • the LSI ex500 is obtained by reading from the encoded data obtained via the base station ex107 by the stream I / Oex 506 or the recording medium ex215 based on the control of the control unit ex501.
  • the encoded data is temporarily stored in the memory ex511 or the like.
  • the accumulated data is appropriately divided into a plurality of times according to the processing amount and the processing speed and sent to the signal processing unit ex507.
  • the signal processing unit ex507 performs decoding of audio data and / or decoding of video data.
  • the decoding process of the video signal is the decoding process described in the above embodiment.
  • each signal may be temporarily stored in the buffer ex508 or the like so that the decoded audio signal and the decoded video signal can be reproduced in synchronization.
  • the decoded output signal is output from each output unit such as the mobile phone ex114, the game machine ex115, and the television ex300 through the memory ex511 or the like as appropriate.
  • the memory ex511 has been described as an external configuration of the LSI ex500.
  • a configuration included in the LSI ex500 may be used.
  • the buffer ex508 is not limited to one, and a plurality of buffers may be provided.
  • the LSI ex500 may be made into one chip or a plurality of chips.
  • LSI LSI
  • IC system LSI
  • super LSI ultra LSI depending on the degree of integration
  • the method of circuit integration is not limited to LSI, and implementation with a dedicated circuit or a general-purpose processor is also possible.
  • An FPGA that can be programmed after manufacturing the LSI or a reconfigurable processor that can reconfigure the connection and setting of circuit cells inside the LSI may be used.
  • the present invention is not limited to these embodiments. Unless it deviates from the meaning of the present invention, various forms conceived by those skilled in the art are applied to the embodiment, and other forms constructed by combining components and steps in different embodiments are also included in the present invention. It is included in the range.
  • the image encoding method and the image decoding method according to the present invention have the effect of being able to improve the encoding efficiency and the image quality.

Abstract

Provided is an image encoding method intended to improve encoding efficiency and image quality. The image encoding method comprises generating an encoded bit stream by encoding the moving image per block using a prediction image (S11); sequentially reconfiguring the encoded blocks (S12); deriving a distortion removing filter coefficient for removing distortion at a block boundary which is a boundary between the reconfigured blocks, in accordance with a feature of the block boundary (S13); filtering the block boundary using the derived distortion removing filter coefficient in order to generate a prediction image (S14), and inserting coefficient specifying information for specifying the derived distortion removing filter coefficient into the encoded bit stream (S15).

Description

画像符号化方法、画像復号方法、それらの装置、プログラムおよび集積回路Image encoding method, image decoding method, apparatus thereof, program, and integrated circuit
 本発明は、画像符号化方法および画像復号方法等に関し、特に、フィルタリングを行う画像符号化方法および画像復号方法等に関する。 The present invention relates to an image encoding method and an image decoding method, and more particularly to an image encoding method and an image decoding method for performing filtering.
 従来、動画像をブロックごとに予測符号化する画像符号化方法と、そのブロックごとに予測符号化された画像を復号する画像復号方法とが提案されている(例えば、非特許文献1参照)。 Conventionally, an image encoding method that predictively encodes a moving image for each block and an image decoding method that decodes an image that is predictively encoded for each block have been proposed (see, for example, Non-Patent Document 1).
 上記非特許文献1の画像符号化方法では、動画像(原画像)のブロックと、そのブロックに対して予測された予測画像のブロックとの差分が周波数変換されて量子化されることによって、複数の量子化値が生成され、それらの複数の量子化値はエントロピー符号化される。このようなエントロピー符号化によって、上述の動画像が符号化され、その結果、符号化ビットストリームが生成される。さらに、その複数の量子化値が逆量子化されて逆周波数変換されることによって差分画像が生成され、その差分画像と予測画像とがブロックごとに加算されることによって、再構成画像が生成される。 In the image encoding method of Non-Patent Document 1, a difference between a block of a moving image (original image) and a block of a predicted image predicted for the block is frequency-converted and quantized to obtain a plurality of differences. Quantized values are generated, and the plurality of quantized values are entropy encoded. By such entropy encoding, the above-described moving image is encoded, and as a result, an encoded bit stream is generated. Furthermore, a difference image is generated by dequantizing the plurality of quantized values and inverse frequency transforming, and the difference image and the predicted image are added for each block to generate a reconstructed image. The
 ここで、再構成画像はブロックごとに生成されるため、再構成画像において、ブロックとそのブロックに隣接するブロックとの間(ブロック境界)には歪みが生じる。そこで、その歪を取り除くために、デブロッキングフィルタが用いられる。 Here, since the reconstructed image is generated for each block, in the reconstructed image, distortion occurs between the block and a block adjacent to the block (block boundary). Therefore, a deblocking filter is used to remove the distortion.
 デブロッキングフィルタは、ブロック境界を挟んで水平方向または垂直方向に配置されている複数の画素の画素値のそれぞれに対して予め定められた係数(フィルタ係数)を乗算し、係数が乗算された画素値を積算することなどによって、そのブロック境界を挟んで隣接する画素の画素値を導出する。これにより、ブロック境界の歪が取り除かれる。そして、このようにブロック境界の歪が取り除かれた再構成画像は上述の予測画像の生成に利用される。 The deblocking filter multiplies each pixel value of a plurality of pixels arranged in the horizontal direction or the vertical direction across the block boundary by a predetermined coefficient (filter coefficient), and the pixel multiplied by the coefficient The pixel values of pixels adjacent to each other across the block boundary are derived by integrating the values. This eliminates block boundary distortion. Then, the reconstructed image from which the block boundary distortion is removed in this way is used to generate the predicted image.
 また、上記非特許文献1の画像復号方法では、符号化ビットストリームがエントロピー復号されることによって、その符号化ビットストリームから複数の量子化値が抽出される。その抽出された複数の量子化値は逆量子化されて逆周波数変換されることによって差分画像が生成され、その差分画像と予測画像とがブロックごとに加算されることによって、再構成画像が生成される。このような再構成画像は、画像符号化方法と同様、予測画像の生成に利用される。そして、再構成画像におけるブロック境界の歪は上述のデブロッキングフィルタによるフィルタリングによって取り除かれる。その結果、符号化ビットストリームは復号されて復号画像として出力される。また、このようにブロック境界の歪が取り除かれた再構成画像は上述の予測画像の生成に利用される。 In the image decoding method of Non-Patent Document 1, a plurality of quantized values are extracted from the encoded bitstream by entropy decoding the encoded bitstream. The extracted quantized values are inversely quantized and inverse frequency transformed to generate a difference image, and the difference image and the predicted image are added for each block to generate a reconstructed image. Is done. Such a reconstructed image is used to generate a predicted image, as in the image encoding method. Then, block boundary distortion in the reconstructed image is removed by filtering using the above-described deblocking filter. As a result, the encoded bit stream is decoded and output as a decoded image. Further, the reconstructed image from which the block boundary distortion has been removed in this way is used to generate the predicted image described above.
 しかしながら、上記非特許文献1の画像符号化方法では、フィルタ係数の選択に制限があるため、ブロック歪みを適切に取り除くことができないという問題がある。具体的には、画像符号化方法および画像復号方法のそれぞれでは、使用可能な複数種のフィルタ係数が予め用意されており、その中から再構成画像の特徴に基づいて、処理対象のブロック境界に対して同一のフィルタ係数を選択する。つまり、画像符号化方法および画像復号方法のそれぞれにおいて処理対象のブロック境界に対して同一のフィルタ係数を選択するためには、画像符号化方法において再構成画像の特徴以外の情報に基づいて自由に適切なフィルタ係数を選択することができない。その結果、画像復号方法においても適切なフィルタ係数を選択することができない。したがって、ブロック歪を適切に取り除くことができず、符号化効率が低下すると共に画質が劣化してしまうという問題がある。 However, the image coding method of Non-Patent Document 1 has a problem in that block distortion cannot be appropriately removed because there is a limit to filter coefficient selection. Specifically, in each of the image encoding method and the image decoding method, a plurality of types of filter coefficients that can be used are prepared in advance, and based on the characteristics of the reconstructed image, the block coefficients to be processed are set. For the same filter coefficient is selected. In other words, in order to select the same filter coefficient for the block boundary to be processed in each of the image encoding method and the image decoding method, the image encoding method can be freely based on information other than the features of the reconstructed image. An appropriate filter coefficient cannot be selected. As a result, an appropriate filter coefficient cannot be selected even in the image decoding method. Therefore, there is a problem that block distortion cannot be removed appropriately, encoding efficiency is lowered, and image quality is deteriorated.
 そこで、本発明は、かかる問題に鑑みてなされたものであって、符号化効率および画質の向上を図った画像符号化方法および画像復号方法等を提供することを目的とする。 Therefore, the present invention has been made in view of such problems, and an object thereof is to provide an image encoding method, an image decoding method, and the like that improve encoding efficiency and image quality.
 上記目的を達成するために、本発明の一態様に係る画像符号化方法は、動画像を符号化する画像符号化方法であって、予測画像を用いて前記動画像をブロックごとに符号化することによって符号化ビットストリームを生成し、符号化された前記ブロックを順次再構成し、再構成されたブロック間の境界であるブロック境界の歪を取り除くための歪除去フィルタ係数を、前記ブロック境界の特徴に応じて導出し、予測画像を生成するために、導出された前記歪除去フィルタ係数を用いたフィルタリングを前記ブロック境界に対して行い、導出された前記歪除去フィルタ係数を特定するための係数特定情報を前記符号化ビットストリームに挿入する。 In order to achieve the above object, an image encoding method according to an aspect of the present invention is an image encoding method for encoding a moving image, and encodes the moving image for each block using a predicted image. Generating a coded bitstream, sequentially reconstructing the coded blocks, and removing distortion distortion filter coefficients for removing distortion at a block boundary, which is a boundary between the reconstructed blocks. A coefficient for deriving according to the feature and performing filtering using the derived distortion removal filter coefficient on the block boundary in order to generate a predicted image, and identifying the derived distortion removal filter coefficient Specific information is inserted into the encoded bitstream.
 これにより、歪除去フィルタ係数を特定するための係数特定情報が符号化ビットストリームに挿入されるため、画像符号化時と画像復号化時において処理対象のブロック境界に対して同一の歪除去フィルタ係数を用いながら、再構成画像の特徴以外の情報に基づいて自由に適切なフィルタ係数を導出することができる。したがって、ブロック歪を適切に取り除くことができ、符号化効率および画質を向上することができる。 As a result, coefficient specifying information for specifying the distortion removal filter coefficient is inserted into the encoded bitstream, so that the same distortion removal filter coefficient is applied to the block boundary to be processed at the time of image encoding and image decoding. The appropriate filter coefficient can be derived freely based on information other than the features of the reconstructed image. Therefore, block distortion can be removed appropriately, and encoding efficiency and image quality can be improved.
 また、前記歪除去フィルタ係数の導出では、前記歪除去フィルタ係数を算出することによって、当該歪除去フィルタ係数を導出し、前記係数特定情報の前記符号化ビットストリームへの挿入では、導出された前記歪除去フィルタ係数を前記係数特定情報として前記符号化ビットストリームに挿入する。 Further, in the derivation of the distortion removal filter coefficient, the distortion removal filter coefficient is derived by calculating the distortion removal filter coefficient, and in the insertion of the coefficient specifying information into the encoded bitstream, the derived A distortion removal filter coefficient is inserted into the coded bitstream as the coefficient specifying information.
 これにより、例えばデブロッキングフィルタのフィルタ係数などの歪除去フィルタ係数がブロック境界の特徴に応じて算出されるため、フィルタ係数の選択に制限がなく、そのブロック境界に適切な歪除去フィルタ係数を導出することができる。このように導出される歪除去フィルタ係数を用いたフィルタリングがそのブロック境界(ブロック境界にある画素)に対して行われることによって、そのブロック境界の歪を適切に取り除くことができ、元の動画像により近い予測画像を生成することができ、その結果、符号化効率および画質を向上することができる。また、その導出された歪除去フィルタ係数は符号化ビットストリームに挿入されるため、画像復号装置がその符号化ビットストリームを復号する際には、その符号化ビットストリームから歪除去フィルタ係数を抽出して、動画像の符号化時と同様に、その歪除去フィルタ係数を用いたフィルタリングをブロック境界に対して行うことができる。その結果、復号画像の画質を向上することができる。 As a result, distortion removal filter coefficients such as deblocking filter coefficients are calculated according to the characteristics of the block boundary, so there is no restriction on the selection of filter coefficients, and an appropriate distortion removal filter coefficient is derived for the block boundary. can do. By performing filtering using the distortion removal filter coefficients derived in this way on the block boundary (pixels on the block boundary), the distortion of the block boundary can be appropriately removed, and the original moving image Predictive images closer to each other can be generated, and as a result, encoding efficiency and image quality can be improved. Since the derived distortion removal filter coefficient is inserted into the encoded bitstream, when the image decoding apparatus decodes the encoded bitstream, the distortion removal filter coefficient is extracted from the encoded bitstream. As in the case of encoding a moving image, filtering using the distortion removal filter coefficient can be performed on the block boundary. As a result, the image quality of the decoded image can be improved.
 また、前記画像符号化方法は、さらに、ブロック境界の特徴に応じて、前記歪除去フィルタ係数を導出すべきか否かを判別し、前記歪除去フィルタの導出の判別において、前記歪除去フィルタ係数を導出すべきでないと判別された際には、前記ブロック境界に対応する処理対象ブロック内のノイズを取り除くための内部ノイズ除去フィルタ係数を導出し、前記予測画像を生成するために、導出された前記内部ノイズ除去フィルタ係数を用いたフィルタリングを前記処理対象ブロックに対して行い、導出された前記内部ノイズ除去フィルタ係数を前記符号化ビットストリームに挿入し、前記歪除去フィルタの導出の判別において、前記歪除去フィルタ係数を導出すべきと判別された際には、前記歪除去フィルタ係数の導出、前記歪除去フィルタ係数を用いたフィルタリング、および前記係数特定情報の前記符号化ビットストリームへの挿入を実行する。 The image encoding method further determines whether or not to derive the distortion removal filter coefficient according to a feature of a block boundary, and determines the distortion removal filter coefficient in the determination of the distortion removal filter derivation. Is derived in order to derive an internal noise removal filter coefficient for removing noise in the processing target block corresponding to the block boundary, and to generate the predicted image. In the determination of the derivation of the distortion removal filter, filtering using the internal noise removal filter coefficient is performed on the processing target block, and the derived internal noise removal filter coefficient is inserted into the encoded bitstream. When it is determined that a distortion removal filter coefficient should be derived, derivation of the distortion removal filter coefficient, Performing filtering using the coefficients, and the insertion into the coded bit stream of the coefficient specifying information.
 これにより、ブロック境界の特徴に応じて、例えばデブロッキングフィルタのフィルタ係数などの歪除去フィルタ係数を用いたブロック境界に対するフィルタリングと、例えばウィナーフィルタのフィルタ係数などの内部ノイズ除去フィルタ係数を用いた処理対象ブロックに対するフィルタリングとが切り替えて行われる。したがって、画像の状況に応じた適切なフィルタリングを選択して適用することができ、符号化効率および画質のさらなる向上を図ることができる。 Thus, according to the characteristics of the block boundary, for example, filtering on the block boundary using the distortion removal filter coefficient such as the filter coefficient of the deblocking filter, and processing using the internal noise removal filter coefficient such as the filter coefficient of the Wiener filter, for example. The filtering for the target block is switched. Therefore, it is possible to select and apply appropriate filtering according to the state of the image, and further improve the encoding efficiency and the image quality.
 また、前記歪除去フィルタの導出の判別において、前記歪除去フィルタ係数を導出すべきと判別された際には、前記歪除去フィルタ係数の導出では、前記ブロック境界の歪とともに、前記処理対象ブロック内のノイズを取り除くためのフィルタ係数を前記歪除去フィルタ係数として導出し、前記歪除去フィルタ係数を用いたフィルタリングでは、前記フィルタ係数を用いたフィルタリングを前記ブロック境界および前記処理対象ブロックに対して行う。 In the derivation determination of the distortion removal filter, when it is determined that the distortion removal filter coefficient should be derived, the distortion removal filter coefficient is derived in the processing target block together with the distortion of the block boundary. The filter coefficient for removing the noise is derived as the distortion removal filter coefficient, and in the filtering using the distortion removal filter coefficient, the filtering using the filter coefficient is performed on the block boundary and the processing target block.
 これにより、ブロック境界の歪とともに、処理対象ブロック内のノイズを取り除くためのフィルタ係数が歪除去フィルタ係数として導出されるため、歪除去フィルタ係数としての性質と内部ノイズ除去フィルタ係数としての性質とを兼ね備えたフィルタ係数を導出することができる。そして、このフィルタ係数を用いたフィルタリングがブロック境界と処理対象ブロックとに対して行われるため、ブロック境界に対するフィルタリングと処理対象ブロックに対するフィルタリングとを、フィルタ係数を切り替えて順に(シーケンシャルに)行う必要がなく、それらに対して一度にフィルタリングを行うことができる。例えば、ウィナーフィルタとでデブロッキングフィルタとを一度にかけることができる。その結果、フィルタリングを簡単に且つ迅速に行うことができる。 As a result, a filter coefficient for removing noise in the processing target block is derived as a distortion removal filter coefficient together with distortion at the block boundary. Therefore, the characteristic as the distortion removal filter coefficient and the characteristic as the internal noise removal filter coefficient are obtained. Combined filter coefficients can be derived. Since filtering using the filter coefficient is performed on the block boundary and the processing target block, it is necessary to perform filtering on the block boundary and filtering on the processing target block sequentially (sequentially) by switching the filter coefficient. And filtering on them at once. For example, a deblocking filter can be applied at once with a winner filter. As a result, filtering can be performed easily and quickly.
 また、前記歪除去フィルタ係数の導出では、前記ブロック境界の歪を抑える強度と、前記処理対象ブロック内のノイズを抑える強度との比率を、前記ブロック境界の特徴に応じて決定し、前記比率に応じて前記フィルタ係数を導出する。 In the derivation of the distortion removal filter coefficient, a ratio between the strength for suppressing the distortion at the block boundary and the strength for suppressing the noise in the processing target block is determined according to the feature of the block boundary, and the ratio is determined. Accordingly, the filter coefficient is derived.
 これにより、例えばブロック境界の歪が目立つ場合には、その歪を抑える強度を強くし、処理対象ブロック内のノイズを抑える強度を弱くしたフィルタ係数を導出することができ、逆に、ブロック境界の歪が目立たない場合には、その歪を抑える強度を弱くし、処理対象ブロック内のノイズを抑える強度を強くしたフィルタ係数を導出することができる。その結果、画像に応じた適切なフィルタリングを行うことができ、符号化効率および画質のさらなる向上を図ることができる。 As a result, for example, when block boundary distortion is conspicuous, it is possible to derive a filter coefficient that increases the strength for suppressing the distortion and decreases the strength for suppressing noise in the processing target block. When the distortion is not conspicuous, it is possible to derive a filter coefficient that reduces the strength for suppressing the distortion and increases the strength for suppressing noise in the processing target block. As a result, appropriate filtering according to the image can be performed, and the encoding efficiency and the image quality can be further improved.
 また、前記画像符号化方法は、さらに、画像領域内のブロック境界ごとに、当該ブロック境界の特徴に応じて、当該ブロック境界に対応するフィルタモードを、予め定められた複数のフィルタモードの中から選択し、前記歪除去フィルタ係数の導出では、フィルタモードごとに、前記画像領域に含まれる当該フィルタモードに対応する複数のブロック境界に対して、共通の前記フィルタ係数を導出する。 Further, the image encoding method further includes, for each block boundary in the image region, a filter mode corresponding to the block boundary according to the feature of the block boundary, from among a plurality of predetermined filter modes. In the derivation of the distortion removal filter coefficient, the common filter coefficient is derived for a plurality of block boundaries corresponding to the filter mode included in the image region for each filter mode.
 これにより、フィルタモードごとに、そのフィルタモードに対応する複数のブロック境界に対して共通のフィルタ係数が導出されるため、例えばフレームなどの画像領域に含まれるブロック境界のそれぞれに対して個別のフィルタ係数を導出する必要がなく、フィルタ係数を導出する演算量を削減することができる。 As a result, for each filter mode, a common filter coefficient is derived for a plurality of block boundaries corresponding to the filter mode. For example, a separate filter is applied to each block boundary included in an image region such as a frame. There is no need to derive coefficients, and the amount of calculation for deriving filter coefficients can be reduced.
 また、前記画像符号化方法は、さらに、フィルタモードごとに、前記画像領域に含まれる当該フィルタモードに対応するブロック境界の数をカウントし、カウントされた前記数が最も少ないフィルタモードを、前記数が2番目に少ないフィルタモードに変更することにより、前記数が最も少ないフィルタモードと、前記数が2番目に少ないフィルタモードとを結合し、前記歪除去フィルタ係数の導出では、結合された2つのフィルタモードに対応する複数のブロック境界に対して、共通の前記フィルタ係数を導出する。 The image encoding method further counts the number of block boundaries corresponding to the filter mode included in the image area for each filter mode, and determines the filter mode with the smallest counted number as the number. Is changed to the filter mode having the second smallest number, and the filter mode having the smallest number is combined with the filter mode having the second smallest number. In the derivation of the distortion removal filter coefficient, The common filter coefficient is derived for a plurality of block boundaries corresponding to the filter mode.
 これにより、カウントされた数が最も少ないフィルタモードが、数が2番目に少ないフィルタモードに変更されて結合されるため、その数が最も少ないフィルタモードに対するフィルタ係数の導出を省くことができ、フィルタ係数を導出する演算量をより削減することができる。また、フィルタ係数の導出が省かれるのは、ブロック境界の数が最も少ないフィルタモードであるため、画像領域に対する画質の影響を最小限に抑えることができる。 As a result, since the filter mode with the smallest number is changed to the filter mode with the second smallest number and combined, the derivation of the filter coefficient for the filter mode with the smallest number can be omitted. The amount of calculation for deriving the coefficient can be further reduced. Also, the derivation of the filter coefficients is omitted in the filter mode having the smallest number of block boundaries, so that the influence of the image quality on the image region can be minimized.
 また、前記結合では、前記数が少ないフィルタモードと、前記数が2番目に少ないフィルタモードとを結合することを繰り返し行い、前記係数特定情報の前記符号化ビットストリームへの挿入では、さらに、繰り返して行われた結合の回数を前記符号化ビットストリームに挿入する。 In the combining, the filter mode having the fewest number and the filter mode having the second smallest number are repeatedly combined, and the insertion of the coefficient specifying information into the coded bitstream is further repeated. Is inserted into the encoded bitstream.
 これにより、フィルタモードの結合が繰り返して行われることによって、フィルタ係数を導出する演算量をさらに削減することができる。また、結合の回数が符号化ビットストリームに挿入されるため、画像復号装置がその符号化ビットストリームを復号する際には、その符号化ビットストリームから結合の回数を抽出し、その回数に基づいてどのフィルタモードが結合されたかを把握することができる。その結果、画像復号装置では、動画像の符号化時のフィルタリングに用いられたフィルタモードと同じフィルタモードに対してフィルタリングを行うことができる。 This makes it possible to further reduce the amount of calculation for deriving the filter coefficients by repeatedly combining the filter modes. Further, since the number of times of combination is inserted into the encoded bit stream, when the image decoding apparatus decodes the encoded bit stream, the number of times of combination is extracted from the encoded bit stream, and based on the number of times It is possible to grasp which filter mode is combined. As a result, the image decoding apparatus can perform filtering on the same filter mode as the filter mode used for filtering when encoding a moving image.
 また、前記画像符号化方法は、さらに、前記予め定められた複数のフィルタモードの中の何れか1つのフィルタモードを、前記複数のフィルタモードの中の他のフィルタモードに変更することにより、前記フィルタモードと前記他のフィルタモードとを結合し、前記歪除去フィルタ係数の導出では、結合された2つのフィルタモードに対応する複数のブロック境界に対して、共通の前記フィルタ係数を導出し、前記係数特定情報の前記符号化ビットストリームへの挿入では、さらに、結合された前記2つのフィルタモードを特定するための結合インデックスを前記符号化ビットストリームに挿入する。 Further, the image encoding method further includes changing the filter mode of any one of the plurality of predetermined filter modes to another filter mode of the plurality of filter modes. Combining a filter mode with the other filter mode, and deriving the distortion removal filter coefficient, the common filter coefficient is derived for a plurality of block boundaries corresponding to the two combined filter modes, In the insertion of the coefficient specifying information into the encoded bitstream, a combined index for specifying the combined two filter modes is further inserted into the encoded bitstream.
 これにより、一方のフィルタモードが他方のフィルタモードに変更されて結合されるため、一方のフィルタモードに対するフィルタ係数の導出を省くことができ、フィルタ係数を導出する演算量をより削減することができる。また、結合された2つのフィルタモードを特定するための結合インデックスが符号化ビットストリームに挿入されるため、画像復号装置がその符号化ビットストリームを復号する際には、その符号化ビットストリームから結合インデックスを抽出し、その結合インデックスに基づいてどのフィルタモードが結合されたかを把握することができる。その結果、画像復号装置では、動画像の符号化時のフィルタリングに用いられたフィルタモードと同じフィルタモードに対してフィルタリングを行うことができる。 Thereby, since one filter mode is changed to the other filter mode and combined, the derivation of the filter coefficient for one filter mode can be omitted, and the amount of calculation for deriving the filter coefficient can be further reduced. . In addition, since a combined index for specifying the two combined filter modes is inserted into the encoded bitstream, when the image decoding apparatus decodes the encoded bitstream, the combined index is combined from the encoded bitstream. An index is extracted, and it is possible to grasp which filter mode is combined based on the combined index. As a result, the image decoding apparatus can perform filtering on the same filter mode as the filter mode used for filtering when encoding a moving image.
 また、前記歪除去フィルタ係数を用いたフィルタリングでは、前記ブロック境界にある画素ごとに、画素間の画素値の差分に基づいて、当該画素に対してフィルタリングを行うべきか否かを判別し、フィルタリングを行うべきと判別された画素に対してフィルタリングを行う。 Further, in the filtering using the distortion removal filter coefficient, for each pixel at the block boundary, it is determined whether or not the pixel should be filtered based on the difference in pixel value between the pixels, and the filtering is performed. Filtering is performed on pixels determined to be performed.
 これにより、ブロック境界にある画素の中でも、必要とされる画素のみに対してフィルタリングが行われるため、フィルタリングの演算量を削減することができるとともに、符号化効率および画質のさらなる向上を図ることができる。 As a result, filtering is performed only on the necessary pixels among the pixels at the block boundary, so that it is possible to reduce the amount of filtering calculation and to further improve the encoding efficiency and image quality. it can.
 また、前記画像符号化方法は、さらに、導出された前記歪除去フィルタ係数を予測することによって前記歪除去フィルタ係数に対する予測値を導出し、前記係数特定情報の前記符号化ビットストリームへの挿入では、前記歪除去フィルタ係数と前記予測値との差分を、前記符号化ビットストリームに挿入する。 Further, the image encoding method further derives a predicted value for the distortion removal filter coefficient by predicting the derived distortion removal filter coefficient, and inserts the coefficient identification information into the encoded bitstream. The difference between the distortion removal filter coefficient and the predicted value is inserted into the encoded bitstream.
 これにより、歪除去フィルタ係数と予測値との差分が符号化ビットストリームに挿入されるため、符号化ビットストリームの符号量を抑えることができる。 Thereby, since the difference between the distortion removal filter coefficient and the predicted value is inserted into the encoded bit stream, the code amount of the encoded bit stream can be suppressed.
 また、前記歪除去フィルタ係数の導出では、色成分およびフィルタリング方向ごとに、前記歪除去フィルタ係数を導出する。 In the derivation of the distortion removal filter coefficient, the distortion removal filter coefficient is derived for each color component and filtering direction.
 これにより、より適切なフィルタリングを行うことができ、符号化効率および画質のさらなる向上を図ることができる。 This makes it possible to perform more appropriate filtering and to further improve the encoding efficiency and image quality.
 また、前記歪除去フィルタ係数の導出では、ウィナーフィルタのフィルタ係数の導出に用いられる演算式を用いて前記歪除去フィルタ係数を導出する。 In the derivation of the distortion removal filter coefficient, the distortion removal filter coefficient is derived using an arithmetic expression used for derivation of the filter coefficient of the Wiener filter.
 これにより、予測画像を元の動画像により近づけることができ、符号化効率および画質のさらなる向上を図ることができる。 Thereby, the predicted image can be brought closer to the original moving image, and the encoding efficiency and the image quality can be further improved.
 上記目的を達成するために、本発明の一態様に係る画像復号方法は、符号化ビットストリームを復号する画像復号方法であって、前記符号化ビットストリームに含まれる符号化されたブロックを順次再構成し、再構成されたブロック間の境界であるブロック境界の特徴に応じて、前記ブロック境界の歪を取り除くための歪除去フィルタ係数を特定するための係数特定情報を、前記符号化ビットストリームから抽出し、抽出された前記係数特定情報によって特定される前記歪除去フィルタ係数を用いたフィルタリングを前記ブロック境界に対して行う。 In order to achieve the above object, an image decoding method according to an aspect of the present invention is an image decoding method for decoding an encoded bitstream, in which encoded blocks included in the encoded bitstream are sequentially re-executed. Coefficient specifying information for specifying a distortion removal filter coefficient for removing distortion of the block boundary according to the feature of the block boundary, which is a boundary between the configured and reconstructed blocks, from the coded bitstream Extraction is performed, and filtering using the distortion removal filter coefficient specified by the extracted coefficient specifying information is performed on the block boundary.
 これにより、ブロック境界の特徴に応じて、その符号化ビットストリームから、歪除去フィルタ係数を特定するための係数特定情報が抽出され、その係数特定情報によって特定される歪除去フィルタ係数を用いたフィルタリングがそのブロック境界に対して行われるため、動画像の符号化時のフィルタリングに用いられた歪除去フィルタ係数を復号時のフィルタリングにも用いることができ、その結果、復号画像の画質を向上することができる。 As a result, the coefficient specifying information for specifying the distortion removing filter coefficient is extracted from the encoded bitstream according to the feature of the block boundary, and the filtering using the distortion removing filter coefficient specified by the coefficient specifying information is performed. Is performed on the block boundary, so that the distortion removal filter coefficient used for the filtering at the time of encoding the moving image can also be used for the filtering at the time of decoding. As a result, the image quality of the decoded image can be improved. Can do.
 また、前記係数特定情報の抽出では、前記ブロック境界の歪とともに、前記ブロック境界に対応する処理対象ブロック内のノイズを取り除くためのフィルタ係数を前記歪除去フィルタ係数として抽出し、前記歪除去フィルタ係数を用いたフィルタリングでは、前記フィルタ係数を用いたフィルタリングを前記ブロック境界および前記処理対象ブロックに対して行う。 Further, in the extraction of the coefficient specifying information, a filter coefficient for removing noise in the processing target block corresponding to the block boundary is extracted as the distortion removing filter coefficient together with the distortion of the block boundary, and the distortion removing filter coefficient In the filtering using, filtering using the filter coefficient is performed on the block boundary and the processing target block.
 これにより、ブロック境界の歪とともに、処理対象ブロック内のノイズを取り除くためのフィルタ係数が歪除去フィルタ係数として抽出されるため、歪除去フィルタ係数としての性質と内部ノイズ除去フィルタ係数としての性質とを兼ね備えたフィルタ係数を抽出することができる。そして、このフィルタ係数を用いたフィルタリングがブロック境界と処理対象ブロックとに対して行われるため、ブロック境界に対するフィルタリングと処理対象ブロックに対するフィルタリングとを、フィルタ係数を切り替えて順に(シーケンシャルに)行う必要がなく、それらに対して一度にフィルタリングを行うことができる。例えば、ウィナーフィルタとでデブロッキングフィルタとを一度にかけることができる。その結果、フィルタリングを簡単に且つ迅速に行うことができる。 As a result, the filter coefficient for removing noise in the processing target block is extracted as the distortion removal filter coefficient together with the distortion at the block boundary. Therefore, the characteristic as the distortion removal filter coefficient and the characteristic as the internal noise removal filter coefficient are obtained. Combined filter coefficients can be extracted. Since filtering using the filter coefficient is performed on the block boundary and the processing target block, it is necessary to perform filtering on the block boundary and filtering on the processing target block sequentially (sequentially) by switching the filter coefficient. And filtering on them at once. For example, a deblocking filter can be applied at once with a winner filter. As a result, filtering can be performed easily and quickly.
 なお、本発明は、このような画像符号化方法および画像復号方法として実現することができるだけでなく、それらの方法にしたがって画像処理する画像符号化装置、画像復号装置、集積回路、コンピュータにそれらの方法にしたがった画像処理を実行させるプログラム、および、そのプログラムを格納する記録媒体としても実現することができる。また、上述のような課題を解決するための手段をどのように組み合わせてもよい。 The present invention can be realized not only as such an image encoding method and image decoding method, but also in an image encoding device, an image decoding device, an integrated circuit, and a computer that perform image processing according to those methods. The present invention can also be realized as a program for executing image processing according to the method and a recording medium for storing the program. Moreover, you may combine how to solve the above subjects how.
 本発明の画像符号化方法および画像復号方法では、符号化効率および画質の向上を図ることができる。 In the image encoding method and the image decoding method of the present invention, it is possible to improve encoding efficiency and image quality.
図1Aは、本発明の一態様に係る画像符号化装置のブロック図である。FIG. 1A is a block diagram of an image encoding device according to an aspect of the present invention. 図1Bは、本発明の一態様に係る画像符号化方法を示すフローチャートである。FIG. 1B is a flowchart illustrating an image encoding method according to an aspect of the present invention. 図2Aは、本発明の一態様に係る画像復号装置のブロック図である。FIG. 2A is a block diagram of an image decoding apparatus according to an aspect of the present invention. 図2Bは、本発明の一態様に係る画像復号方法を示すフローチャートである。FIG. 2B is a flowchart illustrating an image decoding method according to an aspect of the present invention. 図3は、本発明の実施の形態1における画像符号化装置のブロック図である。FIG. 3 is a block diagram of the image coding apparatus according to Embodiment 1 of the present invention. 図4は、本発明の実施の形態1におけるブロック境界の特定とフィルタモードの決定を説明するための図である。FIG. 4 is a diagram for explaining block boundary specification and filter mode determination according to Embodiment 1 of the present invention. 図5は、本発明の実施の形態1におけるブロック境界にある画素を示す図である。FIG. 5 is a diagram showing pixels on the block boundary in the first embodiment of the present invention. 図6は、本発明の実施の形態1におけるフィルタモードを説明するための図である。FIG. 6 is a diagram for explaining the filter mode in the first embodiment of the present invention. 図7は、本発明の実施の形態1における重み係数ω=1の場合のフィルタリングを説明するための図である。FIG. 7 is a diagram for explaining filtering when the weighting factor ω = 1 in the first embodiment of the present invention. 図8は、本発明の実施の形態1におけるフィルタモードごとに重み係数ωを切り替えてフィルタ係数を算出する処理を示すフローチャートである。FIG. 8 is a flowchart showing processing for calculating the filter coefficient by switching the weighting coefficient ω for each filter mode according to Embodiment 1 of the present invention. 図9は、本発明の実施の形態1におけるフィルタモードごとのフィルタ係数を示す図である。FIG. 9 is a diagram showing filter coefficients for each filter mode according to Embodiment 1 of the present invention. 図10は、本発明の実施の形態1におけるフィルタモードの結合を説明するための図である。FIG. 10 is a diagram for explaining filter mode coupling in Embodiment 1 of the present invention. 図11は、本発明の実施の形態1における符号化ビットストリームのノイズ除去フィルタに関するシンタックスを示す図である。FIG. 11 is a diagram illustrating a syntax regarding a noise removal filter of an encoded bitstream according to Embodiment 1 of the present invention. 図12は、本発明の実施の形態1における画像復号装置のブロック図である。FIG. 12 is a block diagram of the image decoding apparatus according to Embodiment 1 of the present invention. 図13は、本発明の実施の形態1の変形例1に係るフィルタリング対象の画素を説明するための図である。FIG. 13 is a diagram for explaining pixels to be filtered according to the first modification of the first embodiment of the present invention. 図14は、本発明の実施の形態1の変形例2に係るフィルタモード結合テーブルを示す図である。FIG. 14 is a diagram showing a filter mode combination table according to the second modification of the first embodiment of the present invention. 図15は、本発明の実施の形態1の変形例3に係るフィルタ係数指定値を説明するための説明図である。FIG. 15 is an explanatory diagram for describing filter coefficient designation values according to the third modification of the first embodiment of the present invention. 図16は、コンテンツ配信サービスを実現するコンテンツ供給システムの全体構成の一例を示す模式図である。FIG. 16 is a schematic diagram illustrating an example of the overall configuration of a content supply system that implements a content distribution service. 図17は、携帯電話の外観を示す図である。FIG. 17 is a diagram illustrating an appearance of a mobile phone. 図18は、携帯電話の構成例を示すブロック図である。FIG. 18 is a block diagram illustrating a configuration example of a mobile phone. 図19は、デジタル放送用システムの全体構成の一例を示す模式図である。FIG. 19 is a schematic diagram illustrating an example of the overall configuration of a digital broadcasting system. 図20は、テレビの構成例を示すブロック図である。FIG. 20 is a block diagram illustrating a configuration example of a television. 図21は、光ディスクである記録メディアに情報の読み書きを行う情報再生記録部の構成例を示すブロック図である。FIG. 21 is a block diagram illustrating a configuration example of an information reproducing / recording unit that reads and writes information from and on a recording medium that is an optical disk. 図22は、光ディスクである記録メディアの構造例を示す図である。FIG. 22 is a diagram illustrating a structure example of a recording medium that is an optical disk. 図23は、各実施の形態に係る画像符号化方法および画像復号方法を実現する集積回路の構成例を示すブロック図である。FIG. 23 is a block diagram illustrating a configuration example of an integrated circuit that realizes the image encoding method and the image decoding method according to each embodiment.
 以下、本発明について図面を参照しながら説明する。 Hereinafter, the present invention will be described with reference to the drawings.
 図1Aは、本発明の一態様に係る画像符号化装置のブロック図である。 FIG. 1A is a block diagram of an image encoding device according to an aspect of the present invention.
 この画像符号化装置10は、動画像を符号化する装置であって、符号化部11、再構成部12、フィルタ係数導出部13、フィルタリング部14、および挿入部15を備える。符号化部11は、予測画像を用いて動画像をブロックごとに符号化することによって符号化ビットストリームを生成する。再構成部12は、符号化されたブロックを順次再構成する。フィルタ係数導出部13は、再構成されたブロック間の境界であるブロック境界の歪を取り除くための歪除去フィルタ係数を、そのブロック境界の特徴に応じて導出する。フィルタリング部14は、上述の予測画像を生成するために、導出された歪除去フィルタ係数を用いたフィルタリングをブロック境界に対して行う。挿入部15は、導出された歪除去フィルタ係数を特定するための係数特定情報を符号化ビットストリームに挿入する。 The image encoding apparatus 10 is an apparatus that encodes a moving image, and includes an encoding unit 11, a reconstruction unit 12, a filter coefficient derivation unit 13, a filtering unit 14, and an insertion unit 15. The encoding unit 11 generates an encoded bitstream by encoding a moving image for each block using a predicted image. The reconstruction unit 12 sequentially reconstructs the encoded blocks. The filter coefficient deriving unit 13 derives a distortion removal filter coefficient for removing distortion at a block boundary, which is a boundary between reconstructed blocks, according to the feature of the block boundary. The filtering unit 14 performs filtering on the block boundary using the derived distortion removal filter coefficient in order to generate the predicted image. The insertion unit 15 inserts coefficient specifying information for specifying the derived distortion removal filter coefficient into the encoded bitstream.
 図1Bは、本発明の一態様に係る画像符号化装置10の処理を示すフローチャートである。 FIG. 1B is a flowchart showing processing of the image encoding device 10 according to an aspect of the present invention.
 画像符号化装置10は、まず、予測画像を用いて動画像をブロックごとに符号化することによって符号化ビットストリームを生成する(ステップS11)。次に、画像符号化装置10は、符号化されたブロックを順次再構成する(ステップS12)。次に、画像符号化装置10は、再構成されたブロック間の境界であるブロック境界の歪を取り除くための歪除去フィルタ係数を、そのブロック境界の特徴に応じて導出する(ステップS13)。次に、画像符号化装置10は、上述の予測画像を生成するために、導出された歪除去フィルタ係数を用いたフィルタリングをブロック境界に対して行う(ステップS14)。さらに、画像符号化装置10は、その導出された歪除去フィルタ係数を特定するための係数特定情報を符号化ビットストリームに挿入する(ステップS15)。 The image encoding device 10 first generates an encoded bitstream by encoding a moving image block by block using a predicted image (step S11). Next, the image encoding device 10 sequentially reconstructs the encoded blocks (step S12). Next, the image coding apparatus 10 derives a distortion removal filter coefficient for removing distortion at the block boundary, which is a boundary between the reconstructed blocks, according to the feature of the block boundary (step S13). Next, the image encoding device 10 performs filtering on the block boundary using the derived distortion removal filter coefficient in order to generate the above-described predicted image (step S14). Further, the image encoding device 10 inserts coefficient specifying information for specifying the derived distortion removal filter coefficient into the encoded bitstream (step S15).
 これにより、本願発明の一態様に係る画像符号化装置10では、歪除去フィルタ係数を特定するための係数特定情報が符号化ビットストリームに挿入されるため、画像符号化時と画像復号化時において処理対象のブロック境界に対して同一の歪除去フィルタ係数を用いながら、再構成画像(再構成されたブロック)の特徴以外の情報に基づいて自由に適切なフィルタ係数を導出することができる。したがって、ブロック歪を適切に取り除くことができ、符号化効率および画質を向上することができる。 As a result, in the image encoding device 10 according to one aspect of the present invention, coefficient specifying information for specifying the distortion removal filter coefficient is inserted into the encoded bitstream, so that at the time of image encoding and image decoding An appropriate filter coefficient can be derived freely based on information other than the characteristics of the reconstructed image (reconstructed block) while using the same distortion removal filter coefficient for the block boundary to be processed. Therefore, block distortion can be removed appropriately, and encoding efficiency and image quality can be improved.
 また、フィルタ係数導出部13は、歪除去フィルタ係数を算出することによって、その歪除去フィルタ係数を導出し、挿入部15は、導出されたその歪除去フィルタ係数を係数特定情報として符号化ビットストリームに挿入してもよい。 The filter coefficient deriving unit 13 derives the distortion removal filter coefficient by calculating the distortion removal filter coefficient, and the insertion unit 15 encodes the encoded bitstream using the derived distortion removal filter coefficient as coefficient specifying information. May be inserted.
 これにより、本発明の一態様に係る画像符号化装置10では、例えばデブロッキングフィルタのフィルタ係数などの歪除去フィルタ係数がブロック境界の特徴に応じて導出(算出)されるため、フィルタ係数の選択に制限がなく、そのブロック境界に適切な歪除去フィルタ係数を導出することができる。このように導出される歪除去フィルタ係数を用いたフィルタリングがそのブロック境界(ブロック境界にある画素)に対して行われることによって、そのブロック境界の歪を適切に取り除くことができ、元の動画像により近い予測画像を生成することができ、その結果、符号化効率および画質を向上することができる。また、その導出された歪除去フィルタ係数は符号化ビットストリームに挿入されるため、画像復号装置がその符号化ビットストリームを復号する際には、その符号化ビットストリームから歪除去フィルタ係数を抽出して、動画像の符号化時と同様に、その歪除去フィルタ係数を用いたフィルタリングをブロック境界に対して行うことができる。その結果、復号画像の画質を向上することができる。 Thereby, in the image encoding device 10 according to an aspect of the present invention, for example, distortion removal filter coefficients such as the filter coefficient of the deblocking filter are derived (calculated) according to the feature of the block boundary. The distortion removal filter coefficient appropriate for the block boundary can be derived. By performing filtering using the distortion removal filter coefficients derived in this way on the block boundary (pixels on the block boundary), the distortion of the block boundary can be appropriately removed, and the original moving image Predictive images closer to each other can be generated, and as a result, encoding efficiency and image quality can be improved. Since the derived distortion removal filter coefficient is inserted into the encoded bitstream, when the image decoding apparatus decodes the encoded bitstream, the distortion removal filter coefficient is extracted from the encoded bitstream. As in the case of encoding a moving image, filtering using the distortion removal filter coefficient can be performed on the block boundary. As a result, the image quality of the decoded image can be improved.
 なお、前記再構成し、フィルタリングする単位は1スライスを単位としてもよいが、さらに小さい単位としてもよい。1スライスを単位とした場合、フィルタ係数の情報量を削減することができる。また、さらに小さい単位(例えば数ブロック単位)とした場合、符号化時に必要な遅延時間を短くすることができる。 Note that the unit for reconfiguration and filtering may be one slice, but may be a smaller unit. When one slice is used as a unit, the information amount of the filter coefficient can be reduced. Further, when the unit is smaller (for example, several blocks), the delay time required for encoding can be shortened.
 また、フィルタ係数導出部13は、複数の歪除去フィルタ係数の候補の中から、上述の歪除去フィルタ係数を選択することによって、その歪除去フィルタ係数を導出し、挿入部15は、導出されたその歪除去フィルタ係数を示すインデックスを係数特定情報として符号化ビットストリームに挿入してもよい。また、フィルタ係数導出部13は、歪除去フィルタ係数を算出し、その算出された歪除去フィルタ係数に最も近い歪除去フィルタ係数を上述の複数の候補の中から選択してもよい。 The filter coefficient deriving unit 13 derives the distortion removal filter coefficient by selecting the above-described distortion removal filter coefficient from among a plurality of distortion removal filter coefficient candidates, and the insertion unit 15 derives the distortion removal filter coefficient. An index indicating the distortion removal filter coefficient may be inserted into the encoded bitstream as coefficient specifying information. Further, the filter coefficient deriving unit 13 may calculate a distortion removal filter coefficient and select a distortion removal filter coefficient closest to the calculated distortion removal filter coefficient from the plurality of candidates described above.
 図2Aは、本発明の一態様に係る画像復号装置のブロック図である。 FIG. 2A is a block diagram of an image decoding apparatus according to an aspect of the present invention.
 この画像復号装置20は、符号化ビットストリームを復号する装置であって、再構成部21、抽出部22およびフィルタリング部23を備える。再構成部21は、その符号化ビットストリームに含まれる符号化されたブロックを順次再構成する。抽出部22は、再構成されたブロック間の境界であるブロック境界の特徴に応じて、そのブロック境界の歪を取り除くための歪除去フィルタ係数を特定するための係数特定情報を、符号化ビットストリームから抽出する。フィルタリング部23は、抽出されたその係数特定情報によって特定される歪除去フィルタ係数を用いたフィルタリングをそのブロック境界に対して行う。 The image decoding device 20 is a device that decodes an encoded bit stream, and includes a reconstruction unit 21, an extraction unit 22, and a filtering unit 23. The reconstruction unit 21 sequentially reconstructs the encoded blocks included in the encoded bitstream. The extraction unit 22 generates coefficient specifying information for specifying a distortion removal filter coefficient for removing distortion at the block boundary in accordance with the feature of the block boundary that is a boundary between the reconstructed blocks. Extract from The filtering unit 23 performs filtering using the distortion removal filter coefficient specified by the extracted coefficient specifying information on the block boundary.
 図2Bは、本発明の一態様に係る画像復号装置20の処理を示すフローチャートである。 FIG. 2B is a flowchart showing processing of the image decoding device 20 according to an aspect of the present invention.
 画像復号装置20は、まず、符号化ビットストリームに含まれる符号化されたブロックを順次再構成する(ステップS21)。次に、画像復号装置20は、再構成されたブロック間の境界であるブロック境界の特徴に応じて、そのブロック境界の歪を取り除くための歪除去フィルタ係数を特定するための係数特定情報を、符号化ビットストリームから抽出する(ステップS22)。次に、画像復号装置20は、抽出されたその係数特定情報によって特定される歪除去フィルタ係数を用いたフィルタリングをそのブロック境界に対して行う(ステップS23)。 First, the image decoding device 20 sequentially reconstructs the encoded blocks included in the encoded bitstream (step S21). Next, the image decoding device 20 generates coefficient specifying information for specifying a distortion removing filter coefficient for removing distortion at the block boundary according to the feature of the block boundary that is a boundary between the reconstructed blocks. Extract from the encoded bit stream (step S22). Next, the image decoding device 20 performs filtering on the block boundary using the distortion removal filter coefficient specified by the extracted coefficient specifying information (step S23).
 これにより、本発明の一態様に係る画像復号装置20では、ブロック境界の特徴に応じて、その符号化ビットストリームから、歪除去フィルタ係数を特定するための係数特定情報が抽出され、その係数特定情報によって特定される歪除去フィルタ係数を用いたフィルタリングがそのブロック境界に対して行われるため、動画像の符号化時のフィルタリングに用いられた歪除去フィルタ係数を復号時のフィルタリングにも用いることができ、その結果、復号画像の画質を向上することができる。 Thereby, in the image decoding apparatus 20 according to an aspect of the present invention, coefficient specifying information for specifying the distortion removal filter coefficient is extracted from the encoded bitstream according to the feature of the block boundary, and the coefficient specifying Since the filtering using the distortion removal filter coefficient specified by the information is performed on the block boundary, the distortion removal filter coefficient used for the filtering at the time of encoding the moving image may be used for the filtering at the time of decoding. As a result, the image quality of the decoded image can be improved.
 また、抽出部22は、上述の歪除去フィルタ係数を係数特定情報として符号化ビットストリームから抽出してもよい。または、抽出部22は、その歪除去フィルタ係数を示すインデックスを係数特定情報として前記符号化ビットストリームから抽出し、フィルタリング部23は、抽出されたインデックスによって示される歪除去フィルタ係数を、複数の歪除去フィルタ係数の候補の中から選択し、選択された歪除去フィルタ係数を用いたフィルタリングをブロック境界に対して行ってもよい。 Also, the extraction unit 22 may extract the above-described distortion removal filter coefficient from the encoded bitstream as coefficient specifying information. Alternatively, the extraction unit 22 extracts an index indicating the distortion removal filter coefficient from the encoded bitstream as coefficient specifying information, and the filtering unit 23 converts the distortion removal filter coefficient indicated by the extracted index into a plurality of distortions. Filtering using the selected distortion removal filter coefficient may be performed on the block boundary by selecting from the candidates for the removal filter coefficient.
 このような本発明について、以下、実施の形態を用いて図面を参照しながら詳細に説明する。 Hereinafter, the present invention will be described in detail with reference to the drawings using embodiments.
 (実施の形態1)
 図3は、本発明の実施の形態1における画像符号化装置のブロック図である。
(Embodiment 1)
FIG. 3 is a block diagram of the image coding apparatus according to Embodiment 1 of the present invention.
 本実施の形態における画像符号化装置100は、動画像である原画像Sをブロックごとに予測符号化することによって符号化ビットストリームを生成する装置であって、減算器101、変換量子化部102、エントロピー符号化部103、逆量子化逆変換部104、加算器105、ノイズ除去フィルタ106および予測部107を備える。 An image encoding device 100 according to the present embodiment is a device that generates an encoded bitstream by predictively encoding an original image S that is a moving image for each block, and includes a subtracter 101 and a transform quantization unit 102. , An entropy coding unit 103, an inverse quantization inverse transform unit 104, an adder 105, a noise removal filter 106, and a prediction unit 107.
 減算器101は、ブロックごとに、原画像Sと予測画像S^とを取得し、それらの差である予測誤差画像eを算出する。 The subtracter 101 acquires the original image S and the predicted image S ^ for each block, and calculates a prediction error image e that is the difference between them.
 変換量子化部102は、予測誤差画像eに対して離散コサイン変換などの直交変換(周波数変換)を行うことによって、複数の周波数係数からなる係数ブロックを生成する。さらに、変換量子化部102は、その係数ブロックに含まれる複数の周波数係数のそれぞれに対して量子化を行うことによって、複数の量子化値からなる量子化ブロックを生成する。 The transform quantizing unit 102 generates a coefficient block including a plurality of frequency coefficients by performing orthogonal transform (frequency transform) such as discrete cosine transform on the prediction error image e. Further, the transform quantization unit 102 generates a quantization block including a plurality of quantization values by performing quantization on each of the plurality of frequency coefficients included in the coefficient block.
 エントロピー符号化部103は、その量子化ブロックに含まれる複数の量子化値に対してエントロピー符号化を行う。このようなエントロピー符号化が行われることによって、符号化ビットストリームが生成される。 The entropy encoding unit 103 performs entropy encoding on a plurality of quantization values included in the quantization block. By performing such entropy encoding, an encoded bit stream is generated.
 逆量子化逆変換部104は、量子化ブロックに含まれる複数の量子化値のそれぞれに対して逆量子化を行うことによって、複数の周波数係数からなる係数ブロックを生成する。さらに、逆量子化逆変換部104は、係数ブロックに対して逆離散コサイン変換などの逆直交変換(逆周波数変換)を行うことによって、予測誤差画像e’をブロック単位で生成する。 The inverse quantization inverse transform unit 104 generates a coefficient block including a plurality of frequency coefficients by performing inverse quantization on each of the plurality of quantization values included in the quantization block. Further, the inverse quantization inverse transform unit 104 generates a prediction error image e ′ in block units by performing inverse orthogonal transform (inverse frequency transform) such as inverse discrete cosine transform on the coefficient block.
 加算器105は、ブロックごとに、予測誤差画像e’と予測画像S^とをそれぞれ取得し、それらを加算することによって再構成画像S’を生成する。これにより、符号化されたブロックが順次再構成される。 The adder 105 acquires the prediction error image e ′ and the prediction image S ^ for each block, and generates the reconstructed image S ′ by adding them. As a result, the encoded blocks are sequentially reconstructed.
 ノイズ除去フィルタ106は、再構成画像S’と予測画像S^とを取得し、それらの画像に基づいて、再構成画像S’のブロック境界の歪と、そのブロック境界に対応する処理対象ブロックに含まれるノイズとを取り除く。つまり、このノイズ除去フィルタ106は、デブロッキングフィルタおよびウィナーフィルタとしての機能を有する。また、このノイズ除去フィルタ106は、原画像Sに基づいて、フィルタリングに用いられる複数のフィルタ係数を、例えばフレームなどの画像領域ごとに算出する。 The noise removal filter 106 acquires the reconstructed image S ′ and the predicted image S ^, and based on these images, the distortion of the block boundary of the reconstructed image S ′ and the processing target block corresponding to the block boundary. Remove the included noise. That is, the noise removal filter 106 functions as a deblocking filter and a Wiener filter. Further, the noise removal filter 106 calculates a plurality of filter coefficients used for filtering for each image region such as a frame based on the original image S.
 予測部107は、ノイズ除去フィルタ106によってノイズ除去された再構成画像S’を用いてイントラ予測またはインター予測を行うことにより、予測画像S^をブロック単位で生成する。 The prediction unit 107 generates the prediction image S ^ in units of blocks by performing intra prediction or inter prediction using the reconstructed image S ′ from which noise has been removed by the noise removal filter 106.
 ここで、図1Aの画像符号化装置10は、本実施の形態における画像符号化装置100に相当し、図1Aの符号化部11は、本実施の形態における減算器101、変換量子化部102、エントロピー符号化部103および予測部107からなる。また、図1Aの再構成部12は、本実施の形態における逆量子化逆変換部104および加算器105からなり、図1Aのフィルタ係数導出部13およびフィルタリング部14は、本実施の形態におけるノイズ除去フィルタ106に相当する。また、図1Aの挿入部15は、本実施の形態におけるエントロピー符号化部103の一部の機能からなる。 Here, the image encoding device 10 in FIG. 1A corresponds to the image encoding device 100 in the present embodiment, and the encoding unit 11 in FIG. 1A includes the subtractor 101 and the transform quantization unit 102 in the present embodiment. , An entropy encoding unit 103 and a prediction unit 107. 1A includes the inverse quantization inverse transform unit 104 and the adder 105 in the present embodiment, and the filter coefficient deriving unit 13 and the filtering unit 14 in FIG. This corresponds to the removal filter 106. Moreover, the insertion part 15 of FIG. 1A consists of a part of function of the entropy encoding part 103 in this Embodiment.
 以下、本実施の形態におけるノイズ除去フィルタ106の詳細について説明する。ノイズ除去フィルタ106は、以下の(1)~(6)の処理を行う。 Hereinafter, details of the noise removal filter 106 in the present embodiment will be described. The noise removal filter 106 performs the following processes (1) to (6).
(1)水平方向および垂直方向のそれぞれのブロック境界を特定する
(2)水平方向および垂直方向のそれぞれのブロック境界に対してフィルタモードを決定する
(3)ブロック境界にある画素のそれぞれに対してフィルタリングを行うべきか否かを判別する
(4)決定されたフィルタモードごとにフィルタ係数を導出(算出)する
(5)導出されたフィルタ係数を量子化して符号化する
(6)導出されたフィルタ係数を用いて、フィルタリングを行うべきと判別された画素のそれぞれに対して水平方向または垂直方向のフィルタリングを行う
(1) Identify each block boundary in the horizontal and vertical directions (2) Determine the filter mode for each block boundary in the horizontal and vertical directions (3) For each pixel in the block boundary (4) Deriving (calculating) filter coefficients for each determined filter mode (5) Quantizing and encoding the derived filter coefficients (6) Derived filter Perform horizontal or vertical filtering for each pixel that is determined to be filtered using coefficients.
 ここで、上述の(1)のブロック境界の特定と、(2)のフィルタモードの決定について説明する。 Here, the identification of the block boundary (1) and the determination of the filter mode (2) will be described.
 図4は、ブロック境界の特定とフィルタモードの決定を説明するための図である。 FIG. 4 is a diagram for explaining block boundary specification and filter mode determination.
 図4に示すように、ノイズ除去フィルタ106は、画像領域における処理対象ブロックごとに、処理対象ブロックの左側と上側にあるブロック境界を特定する。さらに、ノイズ除去フィルタ106は、特定されたブロック境界に対してフィルタモードを決定(選択)する。例えば、ノイズ除去フィルタ106は、左側のブロック境界に対して、そのブロック境界の特徴に応じたフィルタモードm1を決定し、上側のブロック境界に対して、そのブロック境界の特徴に応じたフィルタモードm2を決定する。つまり、ノイズ除去フィルタ106は、処理対象ブロックに対する水平方向のフィルタモードとしてフィルタモードm1を決定し、処理対象ブロックに対する垂直方向のフィルタモードとしてフィルタモードm2を決定する。なお、フィルタモードの詳細については後述する。 As shown in FIG. 4, the noise removal filter 106 identifies block boundaries on the left side and the upper side of the processing target block for each processing target block in the image area. Further, the noise removal filter 106 determines (selects) a filter mode for the identified block boundary. For example, the noise removal filter 106 determines the filter mode m1 corresponding to the feature of the block boundary for the left block boundary, and the filter mode m2 corresponding to the feature of the block boundary for the upper block boundary. To decide. That is, the noise removal filter 106 determines the filter mode m1 as the horizontal filter mode for the processing target block, and determines the filter mode m2 as the vertical filter mode for the processing target block. Details of the filter mode will be described later.
 次に、上述の(3)のフィルタリングを行うべきか否かの判別について説明する。 Next, the determination of whether or not to perform the above filtering (3) will be described.
 ノイズ除去フィルタ106は、左側のブロック境界を中央に挟んで水平方向に配列されている4つの画素をフィルタモードm1によるフィルタリングの対象の候補として特定する。このような4つの画素からなる画素群は、ブロック境界に沿って複数存在するため、それらの画素群のそれぞれがフィルタリングの対象の候補として特定される。同様に、ノイズ除去フィルタ106は、上側のブロック境界を挟んで垂直方向に配列されている4つの画素をフィルタモードm2によるフィルタリングの対象の候補として特定する。また、このような4つの画素からなる画素群は、ブロック境界に沿って複数存在するため、それらの画素群のそれぞれがフィルタリングの対象の候補として特定される。ここで、ノイズ除去フィルタ106は、後述する演算を行うことにより、その候補として特定された画素のそれぞれに対してフィルタリングを行うべきか否かを判別する。 The noise removal filter 106 specifies four pixels arranged in the horizontal direction with the left block boundary in the center as candidates for filtering by the filter mode m1. Since there are a plurality of such pixel groups composed of four pixels along the block boundary, each of these pixel groups is specified as a candidate for filtering. Similarly, the noise removal filter 106 specifies four pixels arranged in the vertical direction across the upper block boundary as candidates for filtering by the filter mode m2. In addition, since there are a plurality of such pixel groups including four pixels along the block boundary, each of these pixel groups is specified as a filtering target candidate. Here, the noise removal filter 106 determines whether or not to perform filtering on each of the pixels specified as candidates by performing an operation described later.
 次に、上述の(4)のフィルタ係数の導出と、(5)の量子化および符号化と、(6)のフィルタリングについて説明する。 Next, the derivation of the filter coefficient (4), the quantization and coding (5), and the filtering (6) will be described.
 ノイズ除去フィルタ106は、画像領域におけるブロック境界のそれぞれに対して上述のように決定されたフィルタモードごとに、後述する演算を行うことにより、そのフィルタモードに対応するフィルタ係数を算出する。つまり、ノイズ除去フィルタ106は、フィルタモードごとに、そのフィルモードに対応する1つ以上のブロック境界に対して共通のフィルタ係数(歪除去フィルタ係数)を算出する。ノイズ除去フィルタ106は、このように算出されたフィルタ係数を予測することにより予測値を導出し、その予測値とフィルタ係数との差分を量子化する。ノイズ除去フィルタ106は、その量子化された差分をエントロピー符号化部103に出力する。エントロピー符号化部103は、その量子化された差分をエントロピー符号化して符号化ビットストリームに挿入する。また、ノイズ除去フィルタ106は、上述の(3)によってフィルタリングを行うべきと判別された画素に対して、上述の(4)によって算出されたフィルタ係数を用いたフィルタリングを行う。 The noise removal filter 106 calculates a filter coefficient corresponding to the filter mode by performing an operation to be described later for each filter mode determined as described above for each of the block boundaries in the image region. That is, for each filter mode, the noise removal filter 106 calculates a common filter coefficient (distortion removal filter coefficient) for one or more block boundaries corresponding to the fill mode. The noise removal filter 106 derives a predicted value by predicting the filter coefficient thus calculated, and quantizes the difference between the predicted value and the filter coefficient. The noise removal filter 106 outputs the quantized difference to the entropy encoding unit 103. The entropy encoding unit 103 performs entropy encoding on the quantized difference and inserts the difference into the encoded bitstream. Further, the noise removal filter 106 performs filtering using the filter coefficient calculated in the above (4) on the pixel determined to be filtered in the above (3).
 図5は、ブロック境界にある画素を示す図である。 FIG. 5 is a diagram showing pixels on the block boundary.
 例えば、図5に示すように、ブロックpとブロックqは隣接している。6つの画素p、p、p、q、q、およびqは、それぞれサンプルであって、ブロックpとブロックqと間の境界(ブロック境界)を中央に挟んで、水平方向または垂直方向に沿って配列されている。なお、ブロックp内には、ブロック境界側から順に画素p、画素p、画素pが配列され、ブロックq内には、ブロック境界側から順に画素q、画素q、画素qが配列されている。 For example, as shown in FIG. 5, the block p and the block q are adjacent to each other. The six pixels p 2 , p 1 , p 0 , q 0 , q 1 , and q 2 are each a sample, and the horizontal direction with the boundary (block boundary) between the block p and the block q in the middle Or they are arranged along the vertical direction. In the block p, a pixel p 0 , a pixel p 1 , and a pixel p 2 are arranged in order from the block boundary side. In the block q, the pixel q 0 , the pixel q 1 , and the pixel q 2 are sequentially arranged from the block boundary side. Are arranged.
 ここで、再構成画像S’における画素p、p、p、q、q、およびqのそれぞれの画素値は、p2,s’、p1,s’、p0,s’、q0,s’、q1,s’、およびq2,s’として定義される。また、予測画像S^における画素p、p、p、q、q、およびqのそれぞれの画素値は、p2,s^、p1,s^、p0,s^、q0,s^、q1,s^、およびq2,s^として定義される。 Here, the pixel values of the pixels p 2 , p 1 , p 0 , q 0 , q 1 , and q 2 in the reconstructed image S ′ are p 2, s ′ , p 1, s ′ , p 0, s ' , q0 , s' , q1 , s ' , and q2 , s' . Further, the pixel values of the pixels p 2 , p 1 , p 0 , q 0 , q 1 , and q 2 in the predicted image S ^ are p 2, s ^ , p 1, s ^ , p 0, s ^ , Q 0, s ^ , q 1, s ^ and q 2, s ^ .
 このような場合、ノイズ除去フィルタ106は、再構成画像S’のブロック境界にある4つの画素p、p、q、およびqのそれぞれに対してフィルタリングを行うべきと判別すると、以下の(式1)~(式4)の演算を行うことによって、それらの画素(フィルタリング対象の画素)のそれぞれに対してフィルタリングを行う。その結果、フィルタリングされた画素p、p、q、およびqの画素値(p’、p’、q’、およびq’)が算出される。 In such a case, when the noise removal filter 106 determines that filtering should be performed on each of the four pixels p 1 , p 0 , q 0 , and q 1 on the block boundary of the reconstructed image S ′, By performing the operations of (Expression 1) to (Expression 4), filtering is performed on each of those pixels (filtering target pixels). As a result, the pixel values (p ′ 1 , p ′ 0 , q ′ 0 , and q ′ 1 ) of the filtered pixels p 1 , p 0 , q 0 , and q 1 are calculated.
Figure JPOXMLDOC01-appb-M000001
Figure JPOXMLDOC01-appb-M000001
Figure JPOXMLDOC01-appb-M000002
Figure JPOXMLDOC01-appb-M000002
Figure JPOXMLDOC01-appb-M000003
Figure JPOXMLDOC01-appb-M000003
Figure JPOXMLDOC01-appb-M000004
Figure JPOXMLDOC01-appb-M000004
 (式1)~(式4)において、a1,m、・・・、a4,m、b1,m、・・・、b4,m、c1,m、・・・、c4,m、d1,m、・・・、d4,m、o1,m、およびo2,mはそれぞれ、フレームなどの画像領域に対して、フィルタモードごとに算出されるフィルタ係数である。つまり、ノイズ除去フィルタ106は、これらのフィルタ係数をブロック境界の特徴に応じて適応的に算出する。また、これらのフィルタ係数は、色成分(輝度成分Y、色差成分Uおよび色差成分V)のそれぞれに対して独立して算出され、フィルタリング方向(水平方向および垂直方向)のそれぞれに対しても独立して算出される。なお、輝度成分Y、色差成分U、および色差成分Vのうち、色差成分Uと色差成分Vのそれぞれのフィルタ係数を共通にしてもよい。 In formulas (1) to (4), a 1, m, ···, a 4, m, b 1, m, ···, b 4, m, c 1, m, ···, c 4 , M , d 1, m ,..., D 4, m , o 1, m , and o 2, m are filter coefficients calculated for each filter mode for an image region such as a frame. . That is, the noise removal filter 106 adaptively calculates these filter coefficients according to the feature of the block boundary. These filter coefficients are calculated independently for each of the color components (luminance component Y, color difference component U, and color difference component V), and are also independent for each of the filtering directions (horizontal direction and vertical direction). Is calculated. Of the luminance component Y, the color difference component U, and the color difference component V, the filter coefficients of the color difference component U and the color difference component V may be made common.
 図6は、フィルタモードを説明するための図である。ブロック境界の特徴(条件)に応じて、例えば10種類のフィルタモードm0~m9がある。 FIG. 6 is a diagram for explaining the filter mode. There are, for example, ten types of filter modes m0 to m9 depending on the feature (condition) of the block boundary.
 フィルタモードm0(DBF_SKIP)は、いわゆるスキップモードであり、ブロック境界の歪を取り除くためのフィルタリングを要しない特定の条件を満たしている場合に適用されるモードである。 The filter mode m0 (DBF_SKIP) is a so-called skip mode, and is a mode that is applied when a specific condition that does not require filtering for removing distortion at the block boundary is satisfied.
 この特定の条件は、ブロック境界が画像中の物体の境界であるという条件、または、ブロック境界において符号化によって発生する歪みが目立たないと想定されるという条件である。具体的には、その特定の条件は、互いに隣接するブロックにおいて、動き補償に用いられる動きベクトルおよび参照ピクチャが同じであり、周波数係数(量子化値)を全く伝送しないという条件、または、互いに隣接するブロックの画素を用いてエッジ検出されたエッジベクトルが閾値よりも大きなノルムを有し、ブロック境界を通るという条件である。 This specific condition is a condition that a block boundary is a boundary of an object in an image, or a condition that distortion generated by encoding at the block boundary is assumed to be inconspicuous. Specifically, the specific condition is that the motion vector used for motion compensation and the reference picture are the same in adjacent blocks, and that no frequency coefficient (quantized value) is transmitted, or adjacent to each other. This is a condition that the edge vector detected by using the pixel of the block to be processed has a norm larger than the threshold and passes through the block boundary.
 フィルタモードm1(DBF_INTRA_QUANT)は、いわゆるイントラモードであり、互いに隣接するブロックのうち、どちらか一方のブロックがイントラ符号化されている場合に適用されるモードである。 The filter mode m1 (DBF_INTRA_QUANT) is a so-called intra mode, and is a mode applied when one of the adjacent blocks is intra-coded.
 フィルタモードm2(DBF_PRED_SIGNIF)は、互いに隣接する両方のブロックがインター符号化され、且つ、両方のブロックのDCT符号化係数(非0の周波数係数)の数の合計が閾値よりも大きい場合に適用されるモードである。 The filter mode m2 (DBF_PRED_SIGNIF) is applied when both adjacent blocks are inter-coded and the sum of the number of DCT coding coefficients (non-zero frequency coefficients) of both blocks is greater than the threshold. Mode.
 フィルタモードm3(DBF_PRED_QUANT)は、互いに隣接する両方のブロックが同一(または略同一)の動きベクトルでインター符号化され、且つ、両方のブロックのDCT符号化係数の数の合計が1以上である場合に適用されるモードである。 In the filter mode m3 (DBF_PRED_QUANT), both adjacent blocks are inter-coded with the same (or substantially the same) motion vector, and the total number of DCT coding coefficients of both blocks is 1 or more. This mode is applied to.
 フィルタモードm4(DBF_PRED_MOT)は、互いに隣接する両方のブロックが互いに異なる動きベクトルでインター符号化され、または、インターテンプレートマッチング符号化され、且つ、両方のブロックのDCT符号化係数の数の合計が1以上である場合に適用されるモードである。 In the filter mode m4 (DBF_PRED_MOT), both adjacent blocks are inter-coded with different motion vectors or inter-template matching coded, and the total number of DCT coding coefficients of both blocks is 1. This mode is applied when the above is true.
 なお、インターテンプレートマッチング符号化とは、符号化対象ブロックに隣接する既に符号化および復号化済みの画素(例えば左および上にある隣接画素)を用いて、符号化および復号済みのフレームの中からそれらの画素と類似する画素の位置を見つけ、その類似する画素の右下のブロックに対応する画素を予測画像として用いる符号化である。 Note that inter template matching encoding is performed by using already encoded and decoded pixels adjacent to the encoding target block (for example, adjacent pixels on the left and above) from among encoded and decoded frames. In this encoding, the positions of pixels similar to those pixels are found, and the pixel corresponding to the lower right block of the similar pixels is used as a predicted image.
 フィルタモードm5(DBF_MOT_DISC)は、互いに隣接する両方のブロックが互いに異なる動きベクトルでインター符号化され、または、インターテンプレートマッチング符号化され、且つ、両方のブロックにDCT符号化係数がない場合に適用されるモードである。 The filter mode m5 (DBF_MOT_DISC) is applied when both adjacent blocks are inter-coded with different motion vectors or inter-template matching coded, and both blocks have no DCT coding coefficient. Mode.
 フィルタモードm6(DBF_IC_STRONG)は、互いに隣接する両方のブロックが互いに異なる動きベクトルでインター符号化され、且つ、両方のブロック間で2つのインテンシティ関連パラメータが互いに異なる場合に適用されるモードである。なお、インテンシティ関連パラメータは、画素値を補正するためのパラメータであって、スケールScとオフセット値Ofとからなる。予測画像S^の画素値をp^とした場合、その画素値p^は、p^’=p^×Sc+Ofによって、画素値p^’に補正される。 Filter mode m6 (DBF_IC_STRONG) is a mode applied when both adjacent blocks are inter-coded with different motion vectors and two intensity-related parameters are different between both blocks. The intensity related parameter is a parameter for correcting the pixel value, and includes a scale Sc and an offset value Of. When the pixel value of the predicted image S ^ is p ^, the pixel value p ^ is corrected to the pixel value p ^ 'by p ^' = p ^ xSc + Of.
 フィルタモードm7(DBF_IC_INTERMED)は、互いに隣接する両方のブロックが同一の動きベクトルでインター符号化され、且つ、両方のブロック間で2つのインテンシティ関連パラメータが互いに異なる場合に適用されるモードである。 Filter mode m7 (DBF_IC_INTERMED) is a mode applied when both adjacent blocks are inter-coded with the same motion vector and two intensity related parameters are different between both blocks.
 フィルタモードm8(DBF_IC_WEAK)は、互いに隣接する両方のブロックが互いに異なる動きベクトルでインター符号化され、且つ、両方のブロック間で少なくとも1つのインテンシティ関連パラメータが異なる場合に適用されるモードである。 Filter mode m8 (DBF_IC_WEAK) is a mode applied when both adjacent blocks are inter-coded with different motion vectors and at least one intensity-related parameter is different between both blocks.
 フィルタモードm9(DBF_BS_PLUS)は、互いに隣接する両方のブロックが互いに異なる動きベクトルでインター符号化され、且つ、ブロック境界の2画素に対してマージモードが用いられる場合に適用されるモードである。マージモードとは、インデックスで示された方向(例えば左または上方向)にあるブロックの動き情報(動きベクトルおよび参照画像)を用いて、ブロック内の予め決められた(例えば2画素)領域を動き補償(動き予測)する方式である。 The filter mode m9 (DBF_BS_PLUS) is a mode applied when both adjacent blocks are inter-coded with different motion vectors and the merge mode is used for two pixels at the block boundary. The merge mode is to move a predetermined area (for example, 2 pixels) in a block using motion information (motion vector and reference image) of the block in the direction (for example, left or upward direction) indicated by the index. This is a method for compensation (motion prediction).
 このように本実施の形態のノイズ除去フィルタ106は、画像領域内のブロック境界ごとに、そのブロック境界の特徴(条件)に応じて、そのブロック境界に対応するフィルタモードを、予め定められた10種類のフィルタモードの中から選択し、フィルタモードごとに、画像領域に含まれるそのフィルタモードに対応する複数のブロック境界に対して共通のフィルタ係数を導出する。これにより、本実施の形態では、フィルタモードごとに、そのフィルタモードに対応する複数のブロック境界に対して共通のフィルタ係数が導出されるため、例えばフレームなどの画像領域に含まれるブロック境界のそれぞれに対して個別のフィルタ係数を導出する必要がなく、フィルタ係数を導出する演算量を削減することができる。 As described above, the noise removal filter 106 according to the present embodiment has a predetermined filter mode corresponding to a block boundary for each block boundary in the image area according to the feature (condition) of the block boundary. A filter coefficient is selected from the types of filter modes, and for each filter mode, a common filter coefficient is derived for a plurality of block boundaries corresponding to the filter mode included in the image region. As a result, in this embodiment, for each filter mode, a common filter coefficient is derived for a plurality of block boundaries corresponding to the filter mode. Therefore, for example, each of the block boundaries included in the image region such as a frame. Therefore, it is not necessary to derive individual filter coefficients, and the amount of calculation for deriving filter coefficients can be reduced.
 ノイズ除去フィルタ106は、上述のようなフィルタモードごとに、2つの評価値を用いてフィルタ係数を算出する。2つの評価値のうちの1つは、客観的画質のための評価値E1であり、原画像Sとノイズ除去された再構成画像S’との差の平均二乗に対応する値である。2つの評価値のうちのもう1つは、主観的画質のための評価値E2であり、ブロック境界の滑らかさを示す値である。ノイズ除去フィルタ106は、その評価値E1が小さくなるように、且つ、評価値E2が小さくなるようにフィルタ係数を算出する。 The noise removal filter 106 calculates a filter coefficient using two evaluation values for each filter mode as described above. One of the two evaluation values is an evaluation value E1 for objective image quality, and is a value corresponding to the mean square of the difference between the original image S and the deconstructed reconstructed image S ′. Another of the two evaluation values is an evaluation value E2 for subjective image quality, which is a value indicating the smoothness of the block boundary. The noise removal filter 106 calculates the filter coefficient so that the evaluation value E1 becomes small and the evaluation value E2 becomes small.
 評価値E1は、以下の(式5)によって算出される。なお、(式5)において、p1,org、p0,org、q0,org、およびq1,orgは、それぞれ原画像Sにおける画素p、p、q、およびqの画素値である。また、(式5)では、ブロック境界における4画素を用いた演算が示されているが、評価値E1は、その4画素だけでなく、フィルタモードごとに、フレームなどの画像領域に含まれる、そのフィルタモードに対応する全てのブロック境界の全ての画素に対して算出される。また、(式5)によって示される評価値E1は、ウィナーフィルタのフィルタ係数を算出するためにも用いられる。つまり、ウィナーフィルタでは、この評価値E1が最も小さくなるように、フィルタ係数が算出される。 The evaluation value E1 is calculated by the following (Formula 5). In (Equation 5), p 1, org , p 0, org , q 0, org , and q 1, org are the pixels p 1 , p 0 , q 0 , and q 1 in the original image S, respectively. Value. Further, in (Equation 5), calculation using four pixels at the block boundary is shown, but the evaluation value E1 is included in an image region such as a frame for each filter mode, not only for the four pixels. It is calculated for all pixels on all block boundaries corresponding to the filter mode. Further, the evaluation value E1 indicated by (Equation 5) is also used for calculating the filter coefficient of the Wiener filter. That is, in the Wiener filter, the filter coefficient is calculated so that the evaluation value E1 is the smallest.
Figure JPOXMLDOC01-appb-M000005
Figure JPOXMLDOC01-appb-M000005
 評価値E2は、以下の(式6)によって算出される。また、(式6)では、ブロック境界における4画素を用いた演算が示されているが、評価値E2は、その4画素だけでなく、フィルタモードごとに、フレームなどの画像領域に含まれる、そのフィルタモードに対応する全てのブロック境界の全ての画素に対して算出される。 Evaluation value E2 is calculated by the following (formula 6). Further, in (Equation 6), calculation using four pixels at the block boundary is shown, but the evaluation value E2 is included in an image region such as a frame for each filter mode, not only for the four pixels. It is calculated for all pixels on all block boundaries corresponding to the filter mode.
Figure JPOXMLDOC01-appb-M000006
Figure JPOXMLDOC01-appb-M000006
 つまり、ノイズ除去フィルタ106は、以下の(式7)の演算を行うことによって、各フィルタモードに対するフィルタ係数を算出する。なお、(式7)においてωは、フィルタモードによって設定される重み係数である。言い換えれば、ノイズ除去フィルタ106は、評価値E1と重み係数ωの積と、評価値E2と(1-重み係数ω)の積との和が最小になるようにフィルタ係数を算出する。例えば、フィルタモードがスキップモード、つまりフィルタモードm1(DBF_SKIP)である場合には、ノイズ除去フィルタ106は重み係数ωをω=1に設定する。また、フィルタモードがスキップモード以外のモードである場合には、ノイズ除去フィルタ106は重み係数ωをω=0.7、0.5または0に設定する。 That is, the noise removal filter 106 calculates the filter coefficient for each filter mode by performing the following (Equation 7). In (Expression 7), ω is a weighting coefficient set by the filter mode. In other words, the noise removal filter 106 calculates the filter coefficient so that the sum of the product of the evaluation value E1 and the weighting coefficient ω and the product of the evaluation value E2 and (1−weighting coefficient ω) is minimized. For example, when the filter mode is the skip mode, that is, the filter mode m1 (DBF_SKIP), the noise removal filter 106 sets the weight coefficient ω to ω = 1. When the filter mode is a mode other than the skip mode, the noise removal filter 106 sets the weighting coefficient ω to ω = 0.7, 0.5, or 0.
Figure JPOXMLDOC01-appb-M000007
Figure JPOXMLDOC01-appb-M000007
 このように本実施の形態におけるノイズ除去フィルタ106は、ウィナーフィルタのフィルタ係数の導出に用いられる演算式(式5)を用いて歪除去フィルタ係数を導出するため、予測画像を元の動画像により近づけることができ、符号化効率および画質のさらなる向上を図ることができる。 As described above, the noise removal filter 106 according to the present embodiment derives the distortion removal filter coefficient using the arithmetic expression (Formula 5) used for derivation of the filter coefficient of the Wiener filter. The encoding efficiency and the image quality can be further improved.
 図7は、重み係数ω=1の場合のフィルタリングを説明するための図である。 FIG. 7 is a diagram for explaining filtering when the weighting factor ω = 1.
 ノイズ除去フィルタ106は、スキップモードのフィルタ係数を算出するときには、重み係数ωを1に設定する。その結果、例えばブロックpおよびqがそれぞれ4×4画素からなる場合、ノイズ除去フィルタ106は、図7に示すように、ブロックpとブロックqの間のブロック境界が中央に位置する4×4画素からなるブロックを、処理対象ブロックとして扱い、その処理対象ブロック内のノイズを取り除くためのフィルタ係数を算出する。つまり、ノイズ除去フィルタ106は、処理対象ブロックに対してウィナーフィルタをかけるためのフィルタ係数(内部ノイズ除去フィルタ係数)を算出する。 The noise removal filter 106 sets the weighting factor ω to 1 when calculating the skip mode filter coefficient. As a result, for example, when the blocks p and q are each composed of 4 × 4 pixels, as shown in FIG. 7, the noise removal filter 106 has 4 × 4 pixels whose block boundary between the blocks p and q is located at the center. Is processed as a processing target block, and a filter coefficient for removing noise in the processing target block is calculated. That is, the noise removal filter 106 calculates a filter coefficient (internal noise removal filter coefficient) for applying the Wiener filter to the processing target block.
 図8は、フィルタモードごとに重み係数ωを切り替えてフィルタ係数を算出する処理を示すフローチャートである。 FIG. 8 is a flowchart showing processing for calculating the filter coefficient by switching the weighting coefficient ω for each filter mode.
 ノイズ除去フィルタ106は、まず、10種類のフィルタモードm0~m9の中から1つのフィルタモードを処理対象のフィルタモードとして選択する(ステップS100)。そして、ノイズ除去フィルタ106は、その処理対象のフィルタモードがスキップモード(DBF_SKIP)であるか否かを判別する(ステップS102)。ここで、ノイズ除去フィルタ106は、スキップモード(DBF_SKIP)であると判別すると(ステップS102のY)、重み係数をω=1に設定する(ステップS104)。 The noise removal filter 106 first selects one filter mode from among the ten filter modes m0 to m9 as the filter mode to be processed (step S100). Then, the noise removal filter 106 determines whether or not the processing target filter mode is the skip mode (DBF_SKIP) (step S102). If the noise removal filter 106 determines that the skip mode (DBF_SKIP) is set (Y in step S102), the weighting coefficient is set to ω = 1 (step S104).
 一方、ノイズ除去フィルタ106は、スキップモード(DBF_SKIP)でないと判別すると(ステップS102のN)、さらに、ステップS100で選択された処理対象のフィルタモードがイントラモード(DBF_INTRA_QUANT)であるか否かを判別する(ステップS106)。ここで、ノイズ除去フィルタ106は、イントラモード(DBF_INTRA_QUANT)であると判別すると(ステップS106のY)、重み係数をω=0.5に設定する(ステップS108)。一方、ノイズ除去フィルタ106は、イントラモード(DBF_INTRA_QUANT)でないと判別すると(ステップS106のN)、重み係数をω=0.7に設定する(ステップS110)。 On the other hand, if the noise removal filter 106 determines that it is not the skip mode (DBF_SKIP) (N in step S102), it further determines whether or not the processing target filter mode selected in step S100 is the intra mode (DBF_INTRA_QUANT). (Step S106). Here, when the noise removal filter 106 determines that the mode is the intra mode (DBF_INTRA_QUANT) (Y in step S106), the noise removal filter 106 sets ω = 0.5 (step S108). On the other hand, if the noise removal filter 106 determines that the mode is not the intra mode (DBF_INTRA_QUANT) (N in step S106), it sets the weighting coefficient to ω = 0.7 (step S110).
 ステップS104、S108およびS110で重み係数ωが設定されると、ノイズ除去フィルタ106は、その設定された重み係数ωを用いて、処理対象のフィルタモードに対応するフィルタ係数を算出する(ステップS112)。つまり、重み係数ω=1の場合には、スキップモードに対して、デブロッキングフィルタの強度が0でありウィナーフィルタの強度が最大となるようなフィルタ係数、つまりウィナーフィルタそのもののフィルタ係数が算出される。また、重み係数ω=0.5の場合には、イントラモードに対して、デブロッキングフィルタの強度とウィナーフィルタの強度とが等しくなるようなフィルタ係数が算出される。また、重み係数ω=0.7の場合には、スキップモードおよびイントラモード以外のフィルタモードに対して、デブロッキングフィルタの強度よりもウィナーフィルタの強度の方が大きくなるようなフィルタ係数が算出される。言い換えれば、スキップモード以外のフィルタモードに対しては、デブロッキングフィルタの性質とウィナーフィルタの性質とを兼ね備えたフィルタ係数が算出される。 When the weighting coefficient ω is set in steps S104, S108, and S110, the noise removal filter 106 calculates a filter coefficient corresponding to the filter mode to be processed using the set weighting coefficient ω (step S112). . That is, when the weighting factor ω = 1, for the skip mode, a filter coefficient with which the strength of the deblocking filter is 0 and the strength of the Wiener filter is maximum, that is, the filter coefficient of the Wiener filter itself is calculated. The Further, when the weighting factor ω = 0.5, a filter coefficient is calculated such that the strength of the deblocking filter and the strength of the Wiener filter are equal to each other in the intra mode. Further, when the weighting factor ω = 0.7, a filter coefficient is calculated such that the Wiener filter strength is greater than the deblocking filter strength for the filter modes other than the skip mode and the intra mode. The In other words, for filter modes other than the skip mode, a filter coefficient having both deblocking filter properties and Wiener filter properties is calculated.
 そして、ノイズ除去フィルタ106は、上述の10種類のフィルタモードm0~m9の中に未選択のフィルタモードがあるか否かを判別する(ステップS114)。ここで、ノイズ除去フィルタ106は、未選択のフィルタモードがあると判別すると(ステップS114のY)、ステップS100からの処理を繰り返し実行し、未選択のフィルタモードがないと判別すると(ステップS114のN)、フレームなどの画像領域に対するフィルタ係数の算出の処理を終了する。 Then, the noise removal filter 106 determines whether or not there is an unselected filter mode among the above ten filter modes m0 to m9 (step S114). Here, if it is determined that there is an unselected filter mode (Y in step S114), the noise removal filter 106 repeatedly executes the processing from step S100, and if it is determined that there is no unselected filter mode (in step S114). N) The filter coefficient calculation process for the image area such as a frame is terminated.
 このように本実施の形態では、ノイズ除去フィルタ106は、ステップS102において、ブロック境界の歪を取り除くための歪除去フィルタ係数を導出すべきか否かを判別している。ここで、ノイズ除去フィルタ106は、その歪除去フィルタ係数を導出すべきでないと判別した際には(ステップS102のY)、そのブロック境界に対応する処理対象ブロック内のノイズを取り除くための内部ノイズ除去フィルタ係数を導出する(ステップS104、S112)。このように導出された内部ノイズ除去フィルタ係数を用いたフィルタリングは、予測画像S^を生成するために処理対象ブロックに対して行われ、その導出された内部ノイズ除去フィルタ係数は符号化ビットストリームに挿入される。また、ノイズ除去フィルタ106は、歪除去フィルタ係数を導出すべきと判別した際には(ステップS102のN)、ブロック境界の歪とともに、処理対象ブロック内のノイズを取り除くためのフィルタ係数を歪除去フィルタ係数として導出する(ステップS108、S110、S112)。このように導出されたフィルタ係数(歪除去フィルタ係数)を用いたフィルタリングは、予測画像S^を生成するためにブロック境界および処理対象ブロックに対して行われ、その導出されたフィルタ係数は符号化ビットストリームに挿入される。 As described above, in this embodiment, the noise removal filter 106 determines whether or not a distortion removal filter coefficient for removing distortion at the block boundary should be derived in step S102. Here, when the noise removal filter 106 determines that the distortion removal filter coefficient should not be derived (Y in step S102), the internal noise for removing the noise in the processing target block corresponding to the block boundary. A removal filter coefficient is derived (steps S104 and S112). The filtering using the internal noise removal filter coefficient derived in this way is performed on the processing target block to generate the predicted image S ^, and the derived internal noise removal filter coefficient is converted into the encoded bitstream. Inserted. When determining that the distortion removal filter coefficient should be derived (N in step S102), the noise removal filter 106 removes the filter coefficient for removing noise in the processing target block along with the distortion of the block boundary. Derived as filter coefficients (steps S108, S110, S112). Filtering using the filter coefficient (distortion removal filter coefficient) derived in this way is performed on the block boundary and the block to be processed in order to generate the prediction image S ^, and the derived filter coefficient is encoded. Inserted into the bitstream.
 これにより、本実施の形態では、ブロック境界の特徴に応じて、例えばデブロッキングフィルタのフィルタ係数などの歪除去フィルタ係数を用いたブロック境界に対するフィルタリングと、例えばウィナーフィルタのフィルタ係数などの内部ノイズ除去フィルタ係数を用いた処理対象ブロックに対するフィルタリングとが切り替えて行われる。したがって、画像の状況に応じた適切なフィルタリングを選択して適用することができ、符号化効率および画質のさらなる向上を図ることができる。 Thus, in the present embodiment, filtering on the block boundary using distortion removal filter coefficients such as filter coefficients of a deblocking filter and internal noise removal such as filter coefficients of a Wiener filter, for example, according to the characteristics of the block boundary. The filtering for the processing target block using the filter coefficient is performed by switching. Therefore, it is possible to select and apply appropriate filtering according to the state of the image, and further improve the encoding efficiency and the image quality.
 さらに、本実施の形態では、ブロック境界の歪とともに、処理対象ブロック内のノイズを取り除くためのフィルタ係数が歪除去フィルタ係数として導出されるため、歪除去フィルタ係数としての性質と内部ノイズ除去フィルタ係数としての性質とを兼ね備えたフィルタ係数を導出することができる。そして、このフィルタ係数を用いたフィルタリングがブロック境界と処理対象ブロックとに対して行われるため、ブロック境界に対するフィルタリングと処理対象ブロックに対するフィルタリングとを、フィルタ係数を切り替えて順に(シーケンシャルに)行う必要がなく、それらに対して一度にフィルタリングを行うことができる。例えば、ウィナーフィルタとでデブロッキングフィルタとを一度にかけることができる。その結果、フィルタリングを簡単に且つ迅速に行うことができる。 Furthermore, in this embodiment, since the filter coefficient for removing the noise in the processing target block is derived as the distortion removal filter coefficient together with the distortion of the block boundary, the characteristics as the distortion removal filter coefficient and the internal noise removal filter coefficient It is possible to derive a filter coefficient having both the characteristics of Since filtering using the filter coefficient is performed on the block boundary and the processing target block, it is necessary to perform filtering on the block boundary and filtering on the processing target block sequentially (sequentially) by switching the filter coefficient. And filtering on them at once. For example, a deblocking filter can be applied at once with a winner filter. As a result, filtering can be performed easily and quickly.
 また、本実施の形態では、ノイズ除去フィルタ106は、ブロック境界の歪を抑える強度と、処理対象ブロック内のノイズを抑える強度との比率を、重み係数ωとして、ブロック境界の特徴に応じて決定し(ステップS104、S108、S110)、その比率に応じてフィルタ係数を導出している(ステップS112)。これにより、本実施の形態では、例えばブロック境界の歪が目立つ場合には、その歪を抑える強度を強くし、処理対象ブロック内のノイズを抑える強度を弱くしたフィルタ係数を導出することができ、逆に、ブロック境界の歪が目立たない場合には、その歪を抑える強度を弱くし、処理対象ブロック内のノイズを抑える強度を強くしたフィルタ係数を導出することができる。その結果、画像に応じた適切なフィルタリングを行うことができ、符号化効率および画質のさらなる向上を図ることができる。 In the present embodiment, the noise removal filter 106 determines the ratio between the strength for suppressing the distortion of the block boundary and the strength for suppressing the noise in the processing target block as the weighting factor ω according to the feature of the block boundary. (Steps S104, S108, and S110), and filter coefficients are derived according to the ratio (Step S112). Thereby, in this embodiment, for example, when block boundary distortion is conspicuous, it is possible to increase the strength to suppress the distortion and to derive the filter coefficient with reduced strength to suppress noise in the processing target block. On the contrary, when the distortion at the block boundary is not conspicuous, it is possible to derive a filter coefficient that reduces the strength for suppressing the distortion and increases the strength for suppressing noise in the processing target block. As a result, appropriate filtering according to the image can be performed, and the encoding efficiency and the image quality can be further improved.
 図9は、フィルタモードごとのフィルタ係数を示す図である。 FIG. 9 is a diagram showing filter coefficients for each filter mode.
 上述のように10種類のフィルタモードm0~m9がある場合、ノイズ除去フィルタ106は、1つのフレームなどの画像領域に対して、図9に示すように、それらのフィルタモードごとにフィルタ係数a1,m、・・・、a4,m、b1,m、・・・、b4,m、c1,m、・・・、c4,m、d1,m、・・・、d4,m、o1,m、およびo2,mを算出する。また、このような各フィルタモードに対するフィルタ係数は、フィルタリング方向(水平方向および垂直方向)および色成分(輝度成分Y、色差成分Uおよび色差成分V)のそれぞれに対して独立して算出される。 As described above, when there are ten types of filter modes m0 to m9, the noise removal filter 106 applies the filter coefficient a 1 to each filter mode for an image region such as one frame as shown in FIG. , M ,..., A 4, m , b 1, m ,..., B 4, m , c 1, m ,..., C 4, m , d 1, m ,. 4, m 1 , o 1, m , and o 2, m are calculated. Further, the filter coefficient for each filter mode is calculated independently for each of the filtering direction (horizontal direction and vertical direction) and color components (luminance component Y, color difference component U, and color difference component V).
 ノイズ除去フィルタ106は、上述のように算出されたフィルタ係数を用いたフィルタリングを行う際には、事前に、再構成画像S’においてフィルタリング対象の候補とされる画素p,p,q,qのそれぞれに対してフィルタリングを行うべきか否かを、以下の(式8)~(式11)に示す演算を行うことによって判別する。 When the noise removal filter 106 performs filtering using the filter coefficient calculated as described above, the pixels p 1 , p 0 , q 0 that are candidates for filtering in the reconstructed image S ′ in advance. , Q 1 , whether or not to perform filtering is determined by performing operations shown in the following (Expression 8) to (Expression 11).
 つまり、ノイズ除去フィルタ106は、再構成画像S’において画素p、p、q、およびqの画素値p1,s’、p0,s’、q0,s’、およびq1,s’が、以下の(式8)の条件を満たす場合には、再構成画像S’の画素pおよびqに対してフィルタリングを行うべきと判別する。 That is, the noise removal filter 106 has pixel values p 1, s ′ , p 0, s ′ , q 0, s ′ , and q of the pixels p 1 , p 0 , q 0 , and q 1 in the reconstructed image S ′. If 1, s ′ satisfies the following condition (Equation 8), it is determined that filtering should be performed on the pixels p 0 and q 0 of the reconstructed image S ′.
Figure JPOXMLDOC01-appb-M000008
Figure JPOXMLDOC01-appb-M000008
 なお、(式8)において、QPは量子化パラメータであり、OffsetおよびOffsetはそれぞれ予め定められたオフセット値(詳細については後述する)である。また、x=QP+OffsetまたはOffsetである場合、α(x)およびβ(x)は、以下の(式9)による演算によって算出される。 In (Equation 8), QP is a quantization parameter, and Offset A and Offset B are predetermined offset values (details will be described later). Further, when x = QP + Offset A or Offset B , α (x) and β (x) are calculated by the following calculation (Equation 9).
Figure JPOXMLDOC01-appb-M000009
Figure JPOXMLDOC01-appb-M000009
 また、ノイズ除去フィルタ106は、再構成画像S’において画素pおよびpの画素値p2,s’およびp0,s’が、以下の(式10)の条件を満たす場合には、再構成画像S’の画素p1に対してフィルタリングを行うべきと判別する。 The noise removal filter 106, when 'the pixel value p 2 of the pixel p 2 and p 0 in, s' reconstructed image S and p 0, s' is satisfies the following conditions of (Equation 10), It is determined that filtering should be performed on the pixel p1 of the reconstructed image S ′.
Figure JPOXMLDOC01-appb-M000010
Figure JPOXMLDOC01-appb-M000010
 また、ノイズ除去フィルタ106は、再構成画像S’において画素qおよびqの画素値q2,s’およびq0,s’が、以下の(式11)の条件を満たす場合には、再構成画像S’の画素qに対してフィルタリングを行うべきと判別する。 The noise removal filter 106, when 'pixel value q 2 of the pixel q 2 and q 0 in, s' reconstructed image S and q 0, s' is satisfies the following conditions of Equation (11), It determines that should perform filtering for pixel q 1 of a reconstructed image S '.
Figure JPOXMLDOC01-appb-M000011
Figure JPOXMLDOC01-appb-M000011
 ここで、ノイズ除去フィルタ106は、オフセット値(OffsetおよびOffset)を適応的に決定してもよい。 Here, the noise removal filter 106 may adaptively determine the offset values (Offset A and Offset B ).
 例えば、ノイズ除去フィルタ106は、互いに隣接しているブロックpおよびqがそれぞれ異なる位置において動き補償されるか否か、つまり、それぞれ異なる動きベクトルまたは参照画像で動き補償されるか否かに応じて、オフセット値を決定する。ここで、ノイズ除去フィルタ106は、それぞれ異なる位置において動き補償されると判別すると、フィルタリング対象の候補の画素がフィルタリング対象に選ばれるように、オフセット値を予め定められた値から増加させる。一方、ノイズ除去フィルタ106は、それぞれ同じ位置において動き補償されると判別すると、フィルタリング対象の候補の画素がフィルタリング対象に選ばれないように、オフセット値を予め定められた値から減少させる。 For example, the noise removal filter 106 depends on whether the blocks p and q adjacent to each other are motion-compensated at different positions, that is, whether motion compensation is performed with different motion vectors or reference images. Determine the offset value. If the noise removal filter 106 determines that motion compensation is performed at different positions, the noise removal filter 106 increases the offset value from a predetermined value so that a candidate pixel to be filtered is selected as a filtering target. On the other hand, if it is determined that motion compensation is performed at the same position, the noise removal filter 106 decreases the offset value from a predetermined value so that a candidate pixel to be filtered is not selected as a filtering target.
 また、ノイズ除去フィルタ106は、互いに隣接しているブロックpおよびqがそれぞれ異なる予測モードで符号化されるか否かに応じて、オフセット値を決定してもよい。 Further, the noise removal filter 106 may determine an offset value depending on whether or not the blocks p and q adjacent to each other are encoded in different prediction modes.
 例えば、ノイズ除去フィルタ106は、ブロックpがイントラ予測モードで符号化され、ブロックqがインター予測モードで符号化される場合には、ブロックpおよびqがそれぞれ異なる予測モードで符号化されると判別する。また、ノイズ除去フィルタ106は、ブロックpが垂直イントラ予測モードで符号化され、ブロックqが水平イントラ予測モードで符号化される場合には、ブロックpおよびqがそれぞれ異なる予測モードで符号化されると判別する。 For example, the noise removal filter 106 determines that the blocks p and q are encoded in different prediction modes when the block p is encoded in the intra prediction mode and the block q is encoded in the inter prediction mode. To do. Further, the noise removal filter 106 encodes the blocks p and q in different prediction modes when the block p is encoded in the vertical intra prediction mode and the block q is encoded in the horizontal intra prediction mode. Is determined.
 そして、ノイズ除去フィルタ106は、ブロックpおよびqがそれぞれ異なる予測モードで符号化されると判別すると、フィルタリング対象の候補の画素がフィルタリング対象に選ばれるように、オフセット値を予め定められた値から増加させる。一方、ノイズ除去フィルタ106は、ブロックpおよびqがそれぞれ同じ予測モードで符号化されると判別すると、フィルタリング対象の候補の画素がフィルタリング対象に選ばれないように、オフセット値を予め定められた値から減少させる。 When the noise removal filter 106 determines that the blocks p and q are encoded in different prediction modes, the offset value is determined from a predetermined value so that a candidate pixel to be filtered is selected as a filtering target. increase. On the other hand, when the noise removal filter 106 determines that the blocks p and q are each encoded in the same prediction mode, the offset value is set to a predetermined value so that the candidate pixel to be filtered is not selected as the filtering target. Decrease from.
 このように決定されるオフセット値は、画像の特徴に応じて決定されるため、画像符号化装置100と、符号化ビットストリームを復号する画像復号装置とにおいてそれぞれ同一の値に決定される。したがって、画像符号化装置100は、そのオフセット値を画像復号装置に送信する必要がない。 Since the offset value determined in this way is determined according to the feature of the image, it is determined to be the same value in the image encoding device 100 and the image decoding device that decodes the encoded bitstream. Therefore, the image encoding device 100 does not need to transmit the offset value to the image decoding device.
 なお、画像符号化装置100は任意のオフセット値を決定し、そのオフセット値を符号化ビットストリームに含めて送信してもよい。この場合には、画像符号化装置100のノイズ除去フィルタ106は、符号量や量子化誤差などを考慮して画質を調整することができる。また、画像符号化装置100のノイズ除去フィルタ106は、任意のオフセット値と、動き補償または予測モードによって決定されるオフセット値との差分を算出し、その差分を符号化ビットストリームに含めて送信してもよい。この場合には、任意のオフセット値によって画質を調整することができるとともに、符号化ビットストリームの符号量の増加を抑えることができる。 Note that the image encoding apparatus 100 may determine an arbitrary offset value and transmit the offset value included in the encoded bitstream. In this case, the noise removal filter 106 of the image encoding device 100 can adjust the image quality in consideration of the code amount, the quantization error, and the like. Further, the noise removal filter 106 of the image encoding device 100 calculates a difference between an arbitrary offset value and an offset value determined by the motion compensation or prediction mode, and transmits the difference included in the encoded bitstream. May be. In this case, the image quality can be adjusted with an arbitrary offset value, and an increase in the code amount of the encoded bit stream can be suppressed.
 ノイズ除去フィルタ106は、上述のように算出されたフィルタ係数a1,m、・・・、a4,m、b1,m、・・・、b4,m、c1,m、・・・、c4,m、d1,m、・・・、d4,m、o1,m、o2,mをそれぞれ予測することによって、それらのフィルタ係数(予測対象フィルタ係数)に対する予測値を導出する。さらに、ノイズ除去フィルタ106は、予測対象フィルタ係数ごとに、その予測対象フィルタ係数と予測値との差分を算出する。そして、ノイズ除去フィルタ106は、その差分(フィルタ予測誤差係数)を量子化してエントロピー符号化部103に出力する。エントロピー符号化部103は、ノイズ除去フィルタ106から出力される、量子化されたフィルタ予測誤差係数を取得してエントロピー符号化し、符号化ビットストリームに挿入する。 The noise removal filter 106 has filter coefficients a 1, m ,..., A 4, m , b 1, m ,..., B 4, m , c 1, m ,. .., C 4, m , d 1, m ,..., D 4, m , o 1, m , o 2, m are predicted to predict the filter coefficients (prediction target filter coefficients). Is derived. Furthermore, the noise removal filter 106 calculates the difference between the prediction target filter coefficient and the prediction value for each prediction target filter coefficient. Then, the noise removal filter 106 quantizes the difference (filter prediction error coefficient) and outputs the result to the entropy encoding unit 103. The entropy encoding unit 103 acquires the quantized filter prediction error coefficient output from the noise removal filter 106, performs entropy encoding, and inserts it into the encoded bitstream.
 ノイズ除去フィルタ106は、例えば、フィルタ係数a1,m、・・・、a4,m、c1,m、・・・、およびc4,mのそれぞれに対しては、予め定められた固定値を予測値として用いる。さらに、ノイズ除去フィルタ106は、フィルタ係数b1,m、・・・、b4,m、d1,m、・・・、およびd4,mのそれぞれに対しては、フィルタ係数a1,m、・・・、a4,m、c1,m、・・・、およびc4,mを予測値として用いる。例えば、ノイズ除去フィルタ106は、フィルタ係数b1,mを予測することにより、フィルタ係数a1,mである予測値a1,mを導出し、フィルタ係数d1,mを予測することにより、フィルタ係数c1,mである予測値c1,mを導出する。そして、ノイズ除去フィルタ106は、予測対象フィルタ係数b1,mと予測値a1,mとの差分をフィルタ予測誤差係数として量子化し、予測対象フィルタ係数d1,mと予測値c1,mとの差分をフィルタ予測誤差係数として量子化する。 Fixed noise removal filter 106, for example, the filter coefficients a 1, m, ···, a 4, m, c 1, m, ···, and with respect to c 4, each m is a predetermined Use the value as the predicted value. Moreover, the noise removal filter 106, the filter coefficients b 1, m, ···, b 4, m, d 1, m, ···, and with respect to d 4, each of m, the filter coefficient a 1, m, ···, a 4, m , c 1, m, is used., and c 4, m as the predicted value. For example, the noise removal filter 106 derives a prediction value a 1, m that is the filter coefficient a 1, m by predicting the filter coefficient b 1, m, and predicts the filter coefficient d 1, m , A prediction value c 1, m that is a filter coefficient c 1, m is derived. The noise removal filter 106 quantizes the difference between the prediction target filter coefficient b 1, m and the prediction value a 1, m as a filter prediction error coefficient, and the prediction target filter coefficient d 1, m and the prediction value c 1, m. Is quantized as a filter prediction error coefficient.
 なお、ノイズ除去フィルタ106は、フィルタ係数a1,m、・・・、a4,m、c1,m、・・・、およびc4,mのそれぞれに対して定数が乗算された値を、フィルタ係数b1,m、・・・、b4,m、d1,m、・・・、およびd4,mのそれぞれに対する予測値として用いてもよい。 Incidentally, the noise removal filter 106, the filter coefficients a 1, m, ···, a 4, m, c 1, m, ···, and c 4, the value constant is multiplied by each of the m , filter coefficients b 1, m, ···, b 4, m, d 1, m, ···, and d 4, may be used as the predicted value for each m.
 また、ノイズ除去フィルタ106は、フィルタ予測誤差係数の量子化に用いられる量子化ステップサイズとして、互いに異なる12個の値を用いることができる。このような量子化ステップサイズは、1つの色成分および1つのフィルタリング方向に対応する全てのフィルタ予測誤差係数(フィルタ係数)に対して同じ値であり、エントロピー符号化され、符号化ビットストリームに含められて画像復号装置に送信される。ここで、ノイズ除去フィルタ106は、RD(Rate-Distortion)最適化によって、12個の値の量子化ステップサイズの中から、量子化に用いられる量子化ステップサイズを選択してもよい。つまり、ノイズ除去フィルタ106は、原画像Sと、フィルタリングされた再構成画像S’との間の差分が小さく、符号量と歪みとのトレードオフのバランスが最適となるように、量子化ステップサイズを選択する。 Also, the noise removal filter 106 can use twelve different values as quantization step sizes used for quantization of the filter prediction error coefficient. Such a quantization step size is the same value for all filter prediction error coefficients (filter coefficients) corresponding to one color component and one filtering direction, and is entropy encoded and included in the encoded bitstream. And transmitted to the image decoding apparatus. Here, the noise removal filter 106 may select a quantization step size used for quantization from among 12 values of quantization step size by RD (Rate-Distortion) optimization. That is, the noise removal filter 106 has a quantization step size so that the difference between the original image S and the filtered reconstructed image S ′ is small and the trade-off balance between code amount and distortion is optimal. Select.
 ここで、ノイズ除去フィルタ106は、図6に示すフィルタモードを結合してもよい。これにより、フィルタモードの数が減るため、フィルタ係数を算出するための演算量を削減することができ、さらに、符号化されたフィルタ係数の符号量(符号化ビットストリームの符号量)を抑えることができる。例えば、ノイズ除去フィルタ106は、フィルタモードごとに、フレームなどの画像領域においてそのフィルタモードに対応するブロック境界(ブロック)の数を境界数としてカウントする。そして、ノイズ除去フィルタ106は、その境界数が最も少ないフィルタモードを、その境界数が2番目に少ないフィルタモードに結合する。 Here, the noise removal filter 106 may combine the filter modes shown in FIG. Thereby, since the number of filter modes is reduced, the amount of calculation for calculating the filter coefficient can be reduced, and further, the code amount of the encoded filter coefficient (code amount of the encoded bit stream) is suppressed. Can do. For example, for each filter mode, the noise removal filter 106 counts the number of block boundaries (blocks) corresponding to the filter mode in the image region such as a frame as the number of boundaries. Then, the noise removal filter 106 combines the filter mode having the smallest number of boundaries with the filter mode having the second smallest number of boundaries.
 図10は、フィルタモードの結合を説明するための図である。 FIG. 10 is a diagram for explaining the combination of the filter modes.
 例えば、ノイズ除去フィルタ106は、画像領域においてフィルタモードm0~m6のそれぞれに対応する境界数をカウントする。図10に示すように、フィルタモードm5の境界数が最も少なく、フィルタモードm6の境界数が2番目に少ない場合、ノイズ除去フィルタ106は、フィルタモードm5をフィルタモードm6に結合する。言い換えれば、ノイズ除去フィルタ106は、フィルタモードm5をフィルタモードm6に変更する。その結果、ノイズ除去フィルタ106は、フィルタモードm5とフィルタモードm6に対して個別に、複数のフィルタ係数からなるフィルタ係数群を導出して符号化する必要がなく、それらのフィルタモード(それらのフィルタモードに対応する複数のブロック境界)に対して共通の1つのフィルタ係数群を導出して符号化するだけでよい。これにより、符号化ビットストリームに挿入されるフィルタ係数の符号量を抑えることができる。なお、上述のフィルタ係数群は、フィルタ係数a1,m、・・・、a4,m、b1,m、・・・、b4,m、c1,m、・・・、c4,m、d1,m、・・・、d4,m、o1,m、およびo2,mからなる。 For example, the noise removal filter 106 counts the number of boundaries corresponding to each of the filter modes m0 to m6 in the image area. As shown in FIG. 10, when the number of boundaries of the filter mode m5 is the smallest and the number of boundaries of the filter mode m6 is the second smallest, the noise removal filter 106 couples the filter mode m5 to the filter mode m6. In other words, the noise removal filter 106 changes the filter mode m5 to the filter mode m6. As a result, the noise removal filter 106 does not need to derive and encode a filter coefficient group including a plurality of filter coefficients individually for the filter mode m5 and the filter mode m6, and the filter mode (the filter mode m5 and the filter mode m6). It is only necessary to derive and encode one common filter coefficient group for a plurality of block boundaries corresponding to the mode. Thereby, the code amount of the filter coefficient inserted into the encoded bit stream can be suppressed. The filter coefficient group described above, the filter coefficients a 1, m, ···, a 4, m, b 1, m, ···, b 4, m, c 1, m, ···, c 4 , M , d 1, m ,..., D 4, m , o 1, m , and o 2, m .
 ノイズ除去フィルタ106は、例えばRD(Rate-Distortion)最適化によって、フィルタモードを結合すべきか否かを判断する。つまり、ノイズ除去フィルタ106は、RD最適化によってコストが最適となるまで、上述のようなフィルタモードの結合を繰り返す。例えば、図10に示すように、フィルタモードm5がフィルタモードm6に結合されると、フィルタモードm6の境界数は3(=1+2)となる。その結果、フィルタモードm6の境界数が最も少なく、フィルタモードm2の境界数が2番目に少ない状態となる。この状態でRD最適化のコストが最適とならない場合、ノイズ除去フィルタ106は、さらに、フィルタモード6をフィルタモード2に結合する。 The noise removal filter 106 determines whether or not to combine the filter modes by, for example, RD (Rate-Distortion) optimization. That is, the noise removal filter 106 repeats the combination of the filter modes as described above until the cost is optimized by the RD optimization. For example, as shown in FIG. 10, when the filter mode m5 is combined with the filter mode m6, the number of boundaries of the filter mode m6 is 3 (= 1 + 2). As a result, the number of boundaries of the filter mode m6 is the smallest, and the number of boundaries of the filter mode m2 is the second smallest. If the RD optimization cost is not optimal in this state, the noise removal filter 106 further couples the filter mode 6 to the filter mode 2.
 また、ノイズ除去フィルタ106は、フィルタモードごとに、画像領域に対してそのフィルタモードが適用されているか否かを示すフラグ(0または1)をエントロピー符号化部103に出力する。さらに、ノイズ除去フィルタ106は、少なくとも1つのフィルタモードが適用されている場合には、フィルタモードが結合された回数(結合回数)と、量子化ステップサイズと、適用されたフィルタモードのそれぞれに対するフィルタ係数群(複数の量子化されたフィルタ予測誤差係数)とをエントロピー符号化部103に出力する。これにより、その結合回数、量子化ステップサイズ、およびフィルタ係数群がエントロピー符号化されて符号化ビットストリームに挿入され、画像復号装置に送信される。 Further, the noise removal filter 106 outputs a flag (0 or 1) indicating whether or not the filter mode is applied to the image area to the entropy encoding unit 103 for each filter mode. Further, when at least one filter mode is applied, the noise removal filter 106 is a filter for each of the number of times the filter mode is combined (number of times of combination), the quantization step size, and the applied filter mode. The coefficient group (a plurality of quantized filter prediction error coefficients) is output to the entropy coding unit 103. As a result, the number of combinations, the quantization step size, and the filter coefficient group are entropy-coded, inserted into the encoded bitstream, and transmitted to the image decoding apparatus.
 図11は、符号化ビットストリームのノイズ除去フィルタ106に関するシンタックスを示す図である。 FIG. 11 is a diagram illustrating a syntax related to the noise removal filter 106 of the encoded bit stream.
 NUM_LOOP_MODES_DBFは、予め定められたフィルタモードの数(例えば10)を示す。yuvは色成分を示し、hvはフィルタリング方向(水平方向または垂直方向)を示すフラグである。 NUM_LOOP_MODES_DBF indicates a predetermined number of filter modes (for example, 10). yuv indicates a color component, and hv is a flag indicating a filtering direction (horizontal direction or vertical direction).
 mode_on[yuv][hv][mode]は、[yuv]、[hv]および[mode]によって示される色成分、フィルタリング方向およびフィルタモードによるフィルタリングが行われるか否かを示すフラグである。つまり、符号化ビットストリームには、色成分、フィルタリング方法、およびフィルタモードの組み合わせごとに、mode_on[yuv][hv][mode]が挿入される。 Mode_on [yuv] [hv] [mode] is a flag indicating whether or not filtering is performed according to the color component, filtering direction, and filter mode indicated by [yuv], [hv], and [mode]. That is, mode_on [yuv] [hv] [mode] is inserted into the encoded bitstream for each combination of color component, filtering method, and filter mode.
 at_least_one_mode_on[yuv][hor_ver]は、mode_on[yuv][hv][mode]の合計値である。number_merges[yuv][hor_ver]は結合回数を示す値である。quant_step_size[yuv][hor_ver]はフィルタ予測誤差係数の量子化に用いられる量子化ステップサイズを示す。つまり、画像領域において1つのフィルタモードでもフィルタリングが行われるのであれば、その符号化ビットストリームには、結合回数を示すnumber_merges[yuv][hor_ver]と、量子化ステップサイズを示すquant_step_size[yuv][hor_ver]とが挿入される。 At_least_one_mode_on [yuv] [hor_ver] is the total value of mode_on [yuv] [hv] [mode]. number_merges [yuv] [hor_ver] is a value indicating the number of times of combining. quant_step_size [yuv] [hor_ver] indicates a quantization step size used for quantization of the filter prediction error coefficient. That is, if filtering is performed even in one filter mode in the image region, number_merges [yuv] [hor_ver] indicating the number of times of coupling and quant_step_size [yuv] [quantity] indicating the quantization step size are included in the encoded bitstream. hor_ver] is inserted.
 fixed_filter[yuv][hv][mode]は、0または1のフラグであって、1である場合には、固定されたフィルタ係数によるフィルタリングが行われて、フィルタ係数群が出力されないことを示す。つまり、mode_on[yuv][hv][mode]=1である場合には、そのmode_on[yuv][hv][mode]に対応する色成分、フィルタリング方法、およびフィルタモードの組み合わせに対して、フィルタ係数が固定されているか否かを、fixed_filter[yuv][hv][mode]は示している。このような、fixed_filter[yuv][hv][mode]は、1を示すmode_on[yuv][hv][mode]ごとに、符号化ビットストリームに挿入される。 Fixed_filter [yuv] [hv] [mode] is a flag of 0 or 1, and when it is 1, it indicates that filtering by a fixed filter coefficient is performed and a filter coefficient group is not output. That is, when mode_on [yuv] [hv] [mode] = 1, a filter is applied to a combination of a color component, a filtering method, and a filter mode corresponding to the mode_on [yuv] [hv] [mode]. Fixed_filter [yuv] [hv] [mode] indicates whether the coefficient is fixed or not. Such fixed_filter [yuv] [hv] [mode] is inserted into the encoded bitstream every mode_on [yuv] [hv] [mode] indicating 1.
 filter_coeff[yuv][hv][mode][i]は、フィルタ係数群に含まれるi番目のフィルタ係数の量子化されたフィルタ予測誤差係数を示す。つまり、fixed_filter[yuv][hv][mode]=0であって、上述の組み合わせのフィルタリングに対してフィルタ係数が固定されていない場合には、その組み合わせのフィルタリングに用いられるフィルタ係数(量子化されたフィルタ予測誤差係数)が順に、filter_coeff[yuv][hv][mode][i]として符号化ビットストリームに挿入される。 Filter_coeff [yuv] [hv] [mode] [i] indicates the quantized filter prediction error coefficient of the i-th filter coefficient included in the filter coefficient group. That is, when fixed_filter [yuv] [hv] [mode] = 0 and the filter coefficient is not fixed for the above-described combination filtering, the filter coefficient (quantized) used for filtering the combination is filtered. Filter prediction error coefficients) are sequentially inserted into the encoded bitstream as filter_coeff [yuv] [hv] [mode] [i].
 なお、fixed_filter[yuv][hv][mode]を、暗示的なフィルタ係数群として用いてもよい。さらに、fixed_filter[yuv][hv][mode]をdiff_fixed_filter[yuv][hv][mode]に置き換えてもよい。このdiff_fixed_filter[yuv][hv][mode]は、例えば、処理対象のfixed_filter[yuv][hv][mode]と、直前に処理されたfixed_filter[yuv][hv][mode]との差分である。この場合、画像符号化装置100および画像復号装置はそれぞれdiff_fixed_filter[yuv][hv][mode]を用いてfixed_filter[yuv][hv][mode]を導出する。その結果、符号化ビットストリームの符号量を削減することができる。 Note that fixed_filter [yuv] [hv] [mode] may be used as an implicit filter coefficient group. Furthermore, fixed_filter [yuv] [hv] [mode] may be replaced with diff_fixed_filter [yuv] [hv] [mode]. The diff_fixed_filter [yuv] [hv] [mode] is, for example, the difference between the fixed_filter [yuv] [hv] [mode] to be processed and the fixed_filter [yuv] [hv] [mode] processed immediately before. . In this case, the image coding apparatus 100 and the image decoding apparatus derive fixed_filter [yuv] [hv] [mode] using diff_fixed_filter [yuv] [hv] [mode], respectively. As a result, the code amount of the encoded bit stream can be reduced.
 このように、本実施の形態における画像符号化装置100では、フィルタ係数(歪除去フィルタ係数)がブロック境界の特徴に応じて算出されるため、フィルタ係数の選択に制限がなく、そのブロック境界に適切な歪除去フィルタ係数を導出することができる。このように算出されるフィルタ係数を用いたフィルタリングがそのブロック境界(ブロック境界にある画素)に対して行われることによって、そのブロック境界の歪を適切に取り除くことができ、元の動画像により近い予測画像S^を生成することができ、その結果、符号化効率および画質を向上することができる。また、その算出されたフィルタ係数は符号化ビットストリームに挿入されるため、画像復号装置がその符号化ビットストリームを復号する際には、その符号化ビットストリームからフィルタ係数を抽出して、動画像の符号化時と同様に、そのフィルタ係数を用いたフィルタリングをブロック境界に対して行うことができる。その結果、復号画像の画質を向上することができる。 Thus, in image coding apparatus 100 according to the present embodiment, filter coefficients (distortion removal filter coefficients) are calculated according to the characteristics of block boundaries, so there is no restriction on the selection of filter coefficients, and the block boundaries are not limited. Appropriate distortion removal filter coefficients can be derived. By performing filtering using the filter coefficient calculated in this way on the block boundary (pixels on the block boundary), distortion of the block boundary can be appropriately removed and closer to the original moving image. The predicted image S ^ can be generated, and as a result, encoding efficiency and image quality can be improved. Further, since the calculated filter coefficient is inserted into the encoded bit stream, when the image decoding apparatus decodes the encoded bit stream, the filter coefficient is extracted from the encoded bit stream, and a moving image is extracted. Similarly to the encoding of, filtering using the filter coefficient can be performed on the block boundary. As a result, the image quality of the decoded image can be improved.
 図12は、本実施の形態における画像復号装置のブロック図である。 FIG. 12 is a block diagram of the image decoding apparatus in the present embodiment.
 本実施の形態における画像復号装置200は、画像符号化装置100によって生成された符号化ビットストリームを復号する装置であって、エントロピー復号部201、逆量子化逆変換部202、加算器203、予測部204およびノイズ除去フィルタ206を備える。 The image decoding apparatus 200 according to the present embodiment is an apparatus that decodes the encoded bitstream generated by the image encoding apparatus 100, and includes an entropy decoding unit 201, an inverse quantization inverse transform unit 202, an adder 203, and a prediction A unit 204 and a noise removal filter 206 are provided.
 エントロピー復号部201は、符号化ビットストリームに対してエントロピー復号を行うことによって、複数の量子化値からなる量子化ブロックを順次生成する。また、エントロピー復号部201は、ブロック境界の特徴(フィルタモード)に応じて、そのブロック境界(フィルタモード)に対応するフィルタ係数群(歪除去フィルタ係数)を符号化ビットストリームから抽出し、そのフィルタ係数群をノイズ除去フィルタ206に出力する。ここで、フィルタ係数群に含まれる各フィルタ係数は、上述のように、量子化されてエントロピー符号化されたフィルタ予測誤差係数として符号化ビットストリームに挿入されている。したがって、エントロピー復号部201は、そのフィルタ係数を符号化ビットストリームから抽出する際には、そのフィルタ予測誤差係数をエントロピー復号して符号化ビットストリームから抽出する。さらに、エントロピー復号部201は、符号化ビットストリームから結合回数および量子化ステップサイズを抽出し、上述のフィルタ係数(量子化されたフィルタ予測誤差係数)とともにノイズ除去フィルタ206に出力する。 The entropy decoding unit 201 sequentially generates quantized blocks including a plurality of quantized values by performing entropy decoding on the encoded bit stream. Further, the entropy decoding unit 201 extracts a filter coefficient group (distortion removal filter coefficient) corresponding to the block boundary (filter mode) from the encoded bitstream in accordance with the feature (filter mode) of the block boundary, and the filter The coefficient group is output to the noise removal filter 206. Here, as described above, each filter coefficient included in the filter coefficient group is inserted into the encoded bitstream as a filter prediction error coefficient that is quantized and entropy-coded. Therefore, when extracting the filter coefficient from the encoded bit stream, the entropy decoding unit 201 performs entropy decoding on the filter prediction error coefficient and extracts the filter prediction error coefficient from the encoded bit stream. Furthermore, the entropy decoding unit 201 extracts the number of times of combination and the quantization step size from the encoded bitstream, and outputs them to the noise removal filter 206 together with the above-described filter coefficient (quantized filter prediction error coefficient).
 逆量子化逆変換部202は、量子化ブロックに含まれる複数の量子化値のそれぞれに対して逆量子化を行うことによって、複数の周波数係数からなる係数ブロックを生成する。さらに、逆量子化逆変換部202は、係数ブロックに対して逆離散コサイン変換などの逆直交変換(逆周波数変換)を行うことによって、予測誤差画像e’をブロック単位で生成する。 The inverse quantization inverse transform unit 202 generates a coefficient block including a plurality of frequency coefficients by performing inverse quantization on each of the plurality of quantization values included in the quantization block. Further, the inverse quantization inverse transform unit 202 performs the inverse orthogonal transform (inverse frequency transform) such as inverse discrete cosine transform on the coefficient block, thereby generating the prediction error image e ′ in units of blocks.
 加算器203は、ブロックごとに、予測誤差画像e’と予測画像S^とをそれぞれ取得し、それらを加算することによって再構成画像S’を生成する。これにより、符号化ビットストリームに含まれる符号化されたブロックが順次再構成される。 The adder 203 acquires the prediction error image e ′ and the prediction image S ^ for each block, and generates a reconstructed image S ′ by adding them. As a result, the encoded blocks included in the encoded bitstream are sequentially reconstructed.
 ノイズ除去フィルタ206は、加算器203および予測部204から再構成画像S’および予測画像S^を取得し、さらに、エントロピー復号部201からフィルタ係数、結合回数および量子化ステップサイズを取得する。ここで、フィルタ係数は、上述のように量子化されたフィルタ予測誤差係数としてエントロピー復号部201から取得される。したがって、ノイズ除去フィルタ206は、量子化ステップサイズを用いてそのフィルタ予測誤差係数を逆量子化する。さらに、ノイズ除去フィルタ206は、上述と同様に、フィルタ係数を予測することによって予測値を導出し、その予測値と逆量子化されたフィルタ予測誤差係数とを加算することによって、フィルタ係数を復元する。また、ノイズ除去フィルタ206は、結合回数だけフィルタモードの結合を繰り返し実行する。 The noise removal filter 206 acquires the reconstructed image S ′ and the predicted image S ^ from the adder 203 and the prediction unit 204, and further acquires the filter coefficient, the number of combinations, and the quantization step size from the entropy decoding unit 201. Here, the filter coefficient is acquired from the entropy decoding unit 201 as the filter prediction error coefficient quantized as described above. Therefore, the noise removal filter 206 dequantizes the filter prediction error coefficient using the quantization step size. Further, as described above, the noise removal filter 206 derives a prediction value by predicting the filter coefficient, and restores the filter coefficient by adding the prediction value and the dequantized filter prediction error coefficient. To do. Further, the noise removal filter 206 repeatedly executes the filter mode combination for the number of times of combination.
 さらに、ノイズ除去フィルタ206は、画像符号化装置100のノイズ除去フィルタ106と同様、再構成画像S’、予測画像S^、およびフィルタ係数群に基づいて、再構成画像S’のブロック境界と、そのブロック境界に対応する処理対象ブロックとに対してフィルタリングを行う。これにより、ノイズ除去フィルタ206は、ブロック境界の歪と、処理対象ブロックに含まれるノイズとを取り除く。つまり、このノイズ除去フィルタ206は、デブロッキングフィルタおよびウィナーフィルタとしての機能を有する。 Further, the noise removal filter 206, like the noise removal filter 106 of the image encoding device 100, is based on the reconstructed image S ′, the predicted image S ^, and the filter coefficient group, and the block boundary of the reconstructed image S ′, Filtering is performed on the processing target block corresponding to the block boundary. Thus, the noise removal filter 206 removes block boundary distortion and noise included in the processing target block. That is, the noise removal filter 206 functions as a deblocking filter and a Wiener filter.
 このように、本実施の形態では、ブロック境界の歪とともに、処理対象ブロック内のノイズを取り除くためのフィルタ係数が歪除去フィルタ係数として抽出されるため、歪除去フィルタ係数としての性質と内部ノイズ除去フィルタ係数としての性質とを兼ね備えたフィルタ係数を抽出することができる。そして、このフィルタ係数を用いたフィルタリングがブロック境界と処理対象ブロックとに対して行われるため、ブロック境界に対するフィルタリングと処理対象ブロックに対するフィルタリングとを、フィルタ係数を切り替えて順に(シーケンシャルに)行う必要がなく、それらに対して一度にフィルタリングを行うことができる。例えば、ウィナーフィルタとでデブロッキングフィルタとを一度にかけることができる。その結果、フィルタリングを簡単に且つ迅速に行うことができる。 As described above, in this embodiment, the filter coefficient for removing noise in the processing target block is extracted as the distortion removal filter coefficient together with the distortion of the block boundary. Therefore, the characteristics as the distortion removal filter coefficient and the internal noise removal are extracted. It is possible to extract a filter coefficient having both properties as a filter coefficient. Since filtering using the filter coefficient is performed on the block boundary and the processing target block, it is necessary to perform filtering on the block boundary and filtering on the processing target block sequentially (sequentially) by switching the filter coefficient. And filtering on them at once. For example, a deblocking filter can be applied at once with a winner filter. As a result, filtering can be performed easily and quickly.
 予測部204は、ノイズ除去フィルタ206によってノイズ除去された再構成画像S’を用いてイントラ予測またはインター予測を行うことにより、予測画像S^をブロック単位で生成する。 The prediction unit 204 generates a prediction image S ^ in units of blocks by performing intra prediction or inter prediction using the reconstructed image S 'from which noise has been removed by the noise removal filter 206.
 なお、図2Aの画像復号装置20は、本実施の形態における画像復号装置200に相当し、図2Aの再構成部21は、本実施の形態における逆量子化逆変換部202および加算器203からなる。また、図2Aの抽出部22は、本実施の形態におけるエントロピー復号部201の一部の機能からなり、図2Aのフィルタリング部23は、本実施の形態におけるノイズ除去フィルタ206に相当する。 2A corresponds to the image decoding device 200 in the present embodiment, and the reconstruction unit 21 in FIG. 2A is derived from the inverse quantization inverse transform unit 202 and the adder 203 in the present embodiment. Become. Moreover, the extraction part 22 of FIG. 2A consists of a part of function of the entropy decoding part 201 in this Embodiment, and the filtering part 23 of FIG. 2A is equivalent to the noise removal filter 206 in this Embodiment.
 このように、本実施の形態における画像復号装置200では、ブロック境界の特徴に応じて、符号化ビットストリームからフィルタ係数(歪除去フィルタ係数)が抽出され、そのフィルタ係数を用いたフィルタリングがそのブロック境界に対して行われるため、動画像の符号化時のフィルタリングに用いられたフィルタ係数を復号時のフィルタリングにも用いることができ、その結果、復号画像の画質を向上することができる。 As described above, in the image decoding apparatus 200 according to the present embodiment, the filter coefficient (distortion removal filter coefficient) is extracted from the encoded bitstream in accordance with the feature of the block boundary, and filtering using the filter coefficient is performed on the block. Since it is performed on the boundary, the filter coefficient used for filtering at the time of encoding a moving image can also be used for filtering at the time of decoding, and as a result, the image quality of the decoded image can be improved.
 (変形例1)
 ここで、上記実施の形態における第1の変形例について説明する。上記実施の形態におけるノイズ除去フィルタ106は、ブロック境界にある画素(ブロック境界を中央に挟んで左右または上下に配列された複数の画素)に対してフィルタリングを行ったが、本変形例に係るノイズ除去フィルタ106は、ブロック境界にある画素と、処理対象ブロックq内にある画素に対してもフィルタリングを行う。
(Modification 1)
Here, a first modification of the above embodiment will be described. The noise removal filter 106 in the above embodiment performs filtering on pixels at a block boundary (a plurality of pixels arranged on the left and right or top and bottom with the block boundary in the middle). The removal filter 106 also performs filtering on pixels on the block boundary and on pixels in the processing target block q.
 図13は、本変形例におけるフィルタリング対象の画素を説明するための図である。 FIG. 13 is a diagram for explaining pixels to be filtered in the present modification.
 本変形例に係るノイズ除去フィルタ106は、例えば水平方向のフィルタリングの場合、ブロック境界を中央に挟んで左右に配列された画素p、p、q、およびqと、処理対象ブロックq内にある画素q~qn-3とに対してフィルタリングを行う。なお、処理対象ブロックqはn×n画素からなる。 For example, in the case of horizontal filtering, the noise removal filter 106 according to the present modification includes pixels p 0 , p 1 , q 0 , and q 1 arranged on the left and right with a block boundary in the center, and a processing target block q Filtering is performed on the pixels q 2 to q n−3 inside. Note that the processing target block q is composed of n × n pixels.
 この場合、ノイズ除去フィルタ106は、上記(式5)の代わりに、以下の(式12)によって評価値E1を算出する。つまり、ノイズ除去フィルタ106は、(式12)、(式6)および(式7)の演算を行うことによってフィルタ係数を算出する。これにより、上記(式7)において重み係数ω=1の場合には、5×5画素以上の画素からなる処理対処ブロックqに対してウィナーフィルタによるフィルタリングを行うことができる。 In this case, the noise removal filter 106 calculates the evaluation value E1 by the following (Formula 12) instead of the above (Formula 5). That is, the noise removal filter 106 calculates the filter coefficient by performing the calculations of (Expression 12), (Expression 6), and (Expression 7). Thereby, when the weighting factor ω = 1 in the above (Equation 7), filtering by the Wiener filter can be performed on the processing countermeasure block q composed of 5 × 5 pixels or more.
Figure JPOXMLDOC01-appb-M000012
Figure JPOXMLDOC01-appb-M000012
 つまり、ノイズ除去フィルタ106は、処理対処ブロックq内にある画素q~qn-3に対しては、以下の(式13)~(式16)の演算を行うことによってフィルタリングを行う。なお、(式13)~(式16)において、jを0以上の整数とすると、kは、k=3+4×jによって示され、3≦k≦(n-5)の範囲の整数である。 That is, the noise removal filter 106 performs filtering on the pixels q 2 to q n−3 in the processing countermeasure block q by performing the following calculations (Equation 13) to (Equation 16). In (Expression 13) to (Expression 16), if j is an integer of 0 or more, k is represented by k = 3 + 4 × j and is an integer in the range of 3 ≦ k ≦ (n−5).
Figure JPOXMLDOC01-appb-M000013
Figure JPOXMLDOC01-appb-M000013
Figure JPOXMLDOC01-appb-M000014
Figure JPOXMLDOC01-appb-M000014
Figure JPOXMLDOC01-appb-M000015
Figure JPOXMLDOC01-appb-M000015
Figure JPOXMLDOC01-appb-M000016
Figure JPOXMLDOC01-appb-M000016
 なお、ノイズ除去フィルタ106は、(式5)~(式7)によってフィルタ係数を算出し、その算出されたフィルタ係数を(式13)~(式16)に代入することによって、処理対象ブロックq内にある画素q~qn-3に対してフィルタリングを行ってもよい。 The noise removal filter 106 calculates the filter coefficient by (Expression 5) to (Expression 7), and substitutes the calculated filter coefficient into (Expression 13) to (Expression 16), thereby processing the processing target block q. Filtering may be performed on the pixels q 2 to q n−3 inside.
 また、画像復号装置200のノイズ除去フィルタ206も、本変形例に係るノイズ除去フィルタ106と同様にフィルタリングを行ってもよい。 Also, the noise removal filter 206 of the image decoding apparatus 200 may perform filtering in the same manner as the noise removal filter 106 according to this modification.
 また、ノイズ除去フィルタ106は、上記(式12)の代わりに、以下の(式17)または(式18)によって評価値E1を算出してもよい。 Further, the noise removal filter 106 may calculate the evaluation value E1 by the following (Expression 17) or (Expression 18) instead of the above (Expression 12).
Figure JPOXMLDOC01-appb-M000017
Figure JPOXMLDOC01-appb-M000017
Figure JPOXMLDOC01-appb-M000018
Figure JPOXMLDOC01-appb-M000018
 また、上記実施の形態ではフィルタモードに応じた重み係数ωを設定したが、重み係数ωだけでなく、評価値E1および評価値E2の算出方法もフィルタモードに応じて設定または選択してもよい。例えば、評価値E1の算出式を、(式5)、(式12)、(式17)および(式18)からフィルタモードに応じて選択し、その選択された算出式に基づいて評価値E1を算出してもよい。 In the above embodiment, the weighting factor ω corresponding to the filter mode is set. However, not only the weighting factor ω but also the calculation method of the evaluation value E1 and the evaluation value E2 may be set or selected according to the filter mode. . For example, a calculation formula for the evaluation value E1 is selected from (Formula 5), (Formula 12), (Formula 17), and (Formula 18) according to the filter mode, and the evaluation value E1 is based on the selected calculation formula. May be calculated.
 (変形例2)
 ここで、本実施の形態における第2の変形例について説明する。上記実施の形態では、フィルタモードごとに、画像領域においてそのフィルタモードに対応するブロック境界の数(境界数)をカウントし、その境界数に応じてフィルタモードを結合したが、他の方法によりフィルタモードを結合してもよい。
(Modification 2)
Here, a second modification of the present embodiment will be described. In the above embodiment, for each filter mode, the number of block boundaries (boundary number) corresponding to the filter mode in the image area is counted, and the filter modes are combined according to the number of boundaries. Modes may be combined.
 本変形例に係るノイズ除去フィルタ106は、図6に示す条件が類似するフィルタモード、つまり、ブロック境界の特徴が類似するフィルタモードを結合する。そして、ノイズ除去フィルタ106は、フィルタモードの結合関係と、その結合関係を特定するためのインデックス(結合インデックス)とを示すフィルタモード結合テーブルを参照し、結合されたフィルタモードに対応する結合インデックスをエントロピー符号化部103に出力する。これにより、その結合インデックスが符号化ビットストリームに含められて画像復号装置200に送信される。画像復号装置200のノイズ除去フィルタ206は、ノイズ除去フィルタ106に参照されるフィルタモード結合テーブルと同一のテーブルを保持している。したがって、ノイズ除去フィルタ206は、その符号化ビットストリームに含められた結合インデックスを取得し、フィルタモード結合テーブルを参照することにより、何れのフィルタモードがどのように結合されたかを把握することができる。 The noise removal filter 106 according to this modification combines filter modes having similar conditions shown in FIG. 6, that is, filter modes having similar block boundary characteristics. Then, the noise removal filter 106 refers to the filter mode combination table indicating the combination relationship of the filter modes and an index (combination index) for specifying the combination relationship, and determines the combination index corresponding to the combined filter mode. The data is output to the entropy encoding unit 103. As a result, the combined index is included in the encoded bitstream and transmitted to the image decoding apparatus 200. The noise removal filter 206 of the image decoding apparatus 200 holds the same table as the filter mode combination table referred to by the noise removal filter 106. Therefore, the noise removal filter 206 can grasp which filter mode is combined by acquiring the combination index included in the encoded bitstream and referring to the filter mode combination table. .
 図14は、フィルタモード結合テーブルを示す図である。 FIG. 14 is a diagram showing a filter mode combination table.
 フィルタモード結合テーブルは、例えば7つのフィルタモードm0、m2、m3、…、m6のうち、何れのフィルタモードも結合されていないことを示す結合インデックスIdxm0と、フィルタモードm3がフィルタモードm2に結合されていることを示す結合インデックスIdxm1と、フィルタモードm3がフィルタモードm2に結合され、さらに、フィルタモードm6がフィルタモードm5に結合されていることを示す結合インデックスIdxm2などを示す。 In the filter mode combination table, for example, among the seven filter modes m0, m2, m3,..., M6, the combination index Idxm0 indicating that no filter mode is combined, and the filter mode m3 is combined with the filter mode m2. A combined index Idxm1 indicating that the filter mode m3 is combined with the filter mode m2, and a combined index Idxm2 indicating that the filter mode m6 is combined with the filter mode m5.
 ノイズ除去フィルタ106は、まず、RD最適化などによって、それらのフィルタモードのうちの何れか少なくとも1つのフィルタモードを他のフィルタモードに結合すべきか否かを判別する。ここで、ノイズ除去フィルタ106は、結合すべきでないと判別すると、それぞれのフィルタモードごとにフィルタ係数を算出してフィルタリングを行う。そして、ノイズ除去フィルタ106は、結合インデックスIdxm0をエントロピー符号化部103に出力することにより、その結合インデックスIdxm0を符号化ビットストリームに挿入する。これにより、画像復号装置200のノイズ除去フィルタ206は、その結合インデックスIdxm0を取得し、その結合インデックスIdxm0に基づいて、フィルタモードが結合されていないことを把握することができる。その結果、画像復号装置200のノイズ除去フィルタ206は、7つのフィルタモードごとに、そのフィルタモードに対応するフィルタ係数を用いてフィルタリングを行う。 The noise removal filter 106 first determines whether or not at least one of these filter modes should be combined with another filter mode by RD optimization or the like. Here, if it is determined that the noise removal filter 106 should not be combined, the noise removal filter 106 performs filtering by calculating a filter coefficient for each filter mode. Then, the noise removal filter 106 outputs the combined index Idxm0 to the entropy encoding unit 103, thereby inserting the combined index Idxm0 into the encoded bitstream. Thereby, the noise removal filter 206 of the image decoding apparatus 200 acquires the combined index Idxm0, and can grasp that the filter mode is not combined based on the combined index Idxm0. As a result, the noise removal filter 206 of the image decoding apparatus 200 performs filtering using the filter coefficient corresponding to each of the seven filter modes.
 一方、ノイズ除去フィルタ106は、結合すべきと判別すると、それらのフィルタモードのうちの何れか少なくとも1つのフィルタモードを他のフィルタモードに結合する。例えば、フィルタモードm3に対応するブロック境界の特徴が、フィルタモードm2に対応するブロック境界の特徴に類似していれば、ノイズ除去フィルタ106は、フィルタモードm3をフィルタモードm2に結合する。言い換えれば、ノイズ除去フィルタ106は、フィルタモードm3の条件を満たすブロック境界に対してフィルタモードm2のフィルタリングを適用する。そして、ノイズ除去フィルタ106は、その結合の内容を示す結合インデックスIdxm1をエントロピー符号化部103に出力することにより、その結合インデックスIdxm1を符号化ビットストリームに含める。これにより、画像復号装置200のノイズ除去フィルタ206は、符号化ビットストリームの結合インデックスIdxm1を取得し、その結合インデックスIdxm1とフィルタモード結合テーブルとに基づいて、フィルタモードm3がフィルタモードm2に結合されていることを知ることができる。その結果、画像復号装置200のノイズ除去フィルタ206は、フィルタモードm3を除くフィルタモードごとに、そのフィルタモードに対応するフィルタ係数を用いてフィルタリングを行う。 On the other hand, when it is determined that the noise removal filter 106 should be combined, at least one of the filter modes is combined with another filter mode. For example, if the block boundary feature corresponding to the filter mode m3 is similar to the block boundary feature corresponding to the filter mode m2, the noise removal filter 106 couples the filter mode m3 to the filter mode m2. In other words, the noise removal filter 106 applies the filtering of the filter mode m2 to the block boundary that satisfies the conditions of the filter mode m3. Then, the noise removal filter 106 outputs the combined index Idxm1 indicating the content of the combination to the entropy encoding unit 103, thereby including the combined index Idxm1 in the encoded bitstream. Thereby, the noise removal filter 206 of the image decoding apparatus 200 acquires the combined index Idxm1 of the encoded bitstream, and the filter mode m3 is combined with the filter mode m2 based on the combined index Idxm1 and the filter mode combination table. I can know that. As a result, the noise removal filter 206 of the image decoding apparatus 200 performs filtering using the filter coefficient corresponding to the filter mode for each filter mode except the filter mode m3.
 このように、本変形例では、ブロック境界の特徴に応じてフィルタモードを結合するため、フィルタ係数の符号量を低減することができるとともに、より適切なフィルタリングを行い、符号化効率および画質を向上することができる。 As described above, in this modification, the filter mode is combined according to the feature of the block boundary, so that the code amount of the filter coefficient can be reduced, and more appropriate filtering is performed to improve the encoding efficiency and the image quality. can do.
 (変形例3)
 ここで、本実施の形態における第3の変形例について説明する。上記実施の形態では、符号化ビットストリームのオーバーヘッド(符号量)を低減するために、フィルタモードを結合し、複数のフィルタモードに対して共通のフィルタ係数群を符号化ビットストリームに挿入したが、他の方法によってそのオーバーヘッドを低減してもよい。本変形例では、リストとフィルタ係数指定値を用いることによってオーバーヘッドを低減する。
(Modification 3)
Here, a third modification of the present embodiment will be described. In the above embodiment, in order to reduce the overhead (code amount) of the encoded bit stream, the filter modes are combined and a common filter coefficient group is inserted into the encoded bit stream for the plurality of filter modes. The overhead may be reduced by other methods. In this modification, overhead is reduced by using a list and a filter coefficient designation value.
 図15は、本変形例に係るリストおよびフィルタ係数指定値を説明するための説明図である。 FIG. 15 is an explanatory diagram for explaining a list and filter coefficient designation values according to this modification.
 本変形例に係るノイズ除去フィルタ106は、まず、フィルタリングに用いられるフィルタ係数群のリストをエントロピー符号化部103に出力する。このリストは、フィルタリングに用いられる複数のフィルタ係数群(gc1、gc2、…、gck)を含み、そのフィルタ係数群に含まれる各フィルタ係数は、上述の量子化ステップサイズで量子化されたフィルタ予測誤差係数として表現されている。したがって、ノイズ除去フィルタ106は、上述のリストと共に量子化ステップサイズも出力する。これにより、符号化ビットストリームには、そのリストと量子化ステップサイズが挿入される。 The noise removal filter 106 according to the present modification first outputs a list of filter coefficient groups used for filtering to the entropy encoding unit 103. This list includes a plurality of filter coefficient groups (gc1, gc2,..., Gck) used for filtering, and each filter coefficient included in the filter coefficient group is a filter prediction quantized with the above-described quantization step size. Expressed as an error coefficient. Therefore, the noise removal filter 106 outputs the quantization step size together with the above list. As a result, the list and the quantization step size are inserted into the encoded bit stream.
 さらに、ノイズ除去フィルタ106は、フィルタモードごとに、そのフィルタモードに対応するフィルタ係数指定値を出力する。所定のフィルタモードによるフィルタリングが行われない場合には、その所定のフィルタモードに対応するフィルタ係数指定値は0を示す。逆に、所定のフィルタモードによるフィルタリングが行われる場合には、そのフィルタモードに対応するフィルタ係数指定値は1以上の値を示す。この場合、その所定のフィルタモードに対応するフィルタ係数指定値は、そのフィルタモードに対応するフィルタ係数群(例えば、18個のフィルタ係数からなる群)をリストから特定するための値を示す。例えば、フィルタモードに対応するフィルタ係数指定値が1の場合には、そのフィルタ係数指定値は、上記リストにおいて1番目に送信されたフィルタ係数群gc1がそのフィルタモードに使用されることを示す。また、他のフィルタモードに対応するフィルタ係数指定値が2の場合には、そのフィルタ係数指定値は、上記リストにおいて2番目に送信されたフィルタ係数群gc2が他のフィルタモードに使用されることを示す。 Furthermore, the noise removal filter 106 outputs a filter coefficient designated value corresponding to the filter mode for each filter mode. When filtering by a predetermined filter mode is not performed, the filter coefficient designation value corresponding to the predetermined filter mode indicates 0. On the contrary, when filtering by a predetermined filter mode is performed, the filter coefficient designation value corresponding to the filter mode indicates a value of 1 or more. In this case, the filter coefficient designation value corresponding to the predetermined filter mode indicates a value for specifying a filter coefficient group (for example, a group of 18 filter coefficients) corresponding to the filter mode from the list. For example, when the filter coefficient designated value corresponding to the filter mode is 1, the filter coefficient designated value indicates that the filter coefficient group gc1 transmitted first in the list is used for the filter mode. When the filter coefficient designation value corresponding to another filter mode is 2, the filter coefficient group gc2 transmitted second in the list is used for the other filter mode. Indicates.
 このように、本変形例に係るノイズ除去フィルタ106は、1つの色成分および1つのフィルタリング方向に対して、フィルタモードごとに、そのフィルタモードによるフィルタリングに使用されるフィルタ係数群を特定するための値であるフィルタ係数指定値を、エントロピー符号化部103に出力する。また、上述のリストに含めらられるフィルタ係数群の数は、フィルタ係数指定値の最大値に等しい。 As described above, the noise removal filter 106 according to the present modification example specifies, for each filter mode, a filter coefficient group used for filtering by the filter mode for one color component and one filtering direction. The filter coefficient designation value, which is a value, is output to the entropy encoding unit 103. Further, the number of filter coefficient groups included in the above list is equal to the maximum value of the filter coefficient designated value.
 エントロピー符号化部103は、上述のように出力されるフィルタ係数指定値、量子化ステップサイズおよびリストをエントロピー符号化して符号化ビットストリームに挿入する。 The entropy encoding unit 103 entropy-encodes the filter coefficient designation value, quantization step size, and list output as described above and inserts them into the encoded bitstream.
 画像復号装置200のエントロピー復号部201は、符号化ビットストリームに含まれる上述のフィルタ係数指定値、量子化ステップサイズおよびリストをエントロピー復号する。そして、ノイズ除去フィルタ206は、リストに含まれる量子化されたフィルタ予測誤差係数のそれぞれに対して量子化ステップサイズを用いた逆量子化を行う。さらに、ノイズ除去フィルタ206は、逆量子化されたフィルタ予測誤差係数のそれぞれに予測値を加算することによってフィルタ係数(フィルタ係数群)を復元する。そして、ノイズ除去フィルタ206は、フィルタモードごとに、そのフィルタモードに対応するフィルタ係数指定値に基づいてリストからフィルタ係数群を選択し、そのフィルタ係数群を用いたフィルタリングを行う。 The entropy decoding unit 201 of the image decoding apparatus 200 performs entropy decoding on the above-described filter coefficient designation value, quantization step size, and list included in the encoded bitstream. Then, the noise removal filter 206 performs inverse quantization using a quantization step size for each of the quantized filter prediction error coefficients included in the list. Further, the noise removal filter 206 restores a filter coefficient (filter coefficient group) by adding a prediction value to each of the inversely quantized filter prediction error coefficients. For each filter mode, the noise removal filter 206 selects a filter coefficient group from the list based on the filter coefficient designation value corresponding to the filter mode, and performs filtering using the filter coefficient group.
 これにより、同じフィルタ係数群を重複して符号化ビットストリームに含めることを防ぐことができ、符号化ビットストリームの符号量を抑えることができる。 Thereby, it is possible to prevent the same filter coefficient group from being duplicated and included in the encoded bit stream, and to suppress the code amount of the encoded bit stream.
 本発明の画像符号化方法および画像復号方法について、上記実施の形態1およびその変形例を用いて説明したが、本発明はそれらに限定されるものではない。 Although the image encoding method and the image decoding method of the present invention have been described using the first embodiment and the modifications thereof, the present invention is not limited to these.
 例えば、上記実施の形態1およびその変形例では、フィルタモードごとにそのフィルタモードに応じた重み係数ωを設定したが、フィルタモードモードごとにそのフィルタモードに応じたタップ長さを設定してもよい。例えば、上記実施の形態では、何れのフィルタモードに対しても、(式1)~(式4)に示すように4つの画素(4タップ)を用いてフィルタリングを行うが、所定のフィルタモードに対しては6つの画素(6タップ)を用いてフィルタリングを行ってもよい。 For example, in Embodiment 1 and the modification thereof, the weighting factor ω corresponding to the filter mode is set for each filter mode, but the tap length corresponding to the filter mode may be set for each filter mode. Good. For example, in the above embodiment, for any filter mode, filtering is performed using four pixels (4 taps) as shown in (Expression 1) to (Expression 4). On the other hand, filtering may be performed using six pixels (6 taps).
 また、上記実施の形態1およびその変形例では、フィルタモードがスキップモード(DBF_SKIP)の場合には、重み係数をω=1に設定し、フィルタモードがイントラモード(DBF_INTRA_QUANT)の場合には、重み係数をω=0.5に設定し、フィルタモードがそれらのモード以外の場合には、重み係数をω=0.7に設定した。しかし、これらの重み係数は一例であって、他の重み係数に設定してもよい。 In the first embodiment and the modification thereof, when the filter mode is the skip mode (DBF_SKIP), the weighting factor is set to ω = 1, and when the filter mode is the intra mode (DBF_INTRA_QUANT), the weight is set. When the coefficient was set to ω = 0.5 and the filter mode was other than those modes, the weighting coefficient was set to ω = 0.7. However, these weighting factors are merely examples, and other weighting factors may be set.
 また、上記実施の形態1およびその変形例では、フィルタモードがスキップモード(DBF_SKIP)のときには、重み係数をω=1に設定し、そのスキップモードに対応する何れの処理対象ブロックまたはブロック境界に対しても共通のフィルタ(ウィナーフィルタ)をかけたが、そのスキップモードに対応する各処理対象ブロックまたは各ブロック境界に対して互いに異なる強度または性質のウィナーフィルタをかけてもよい。例えば、フィルタ係数やタップ長さなどが互いに異なるウィナーフィルタをかけてもよい。この場合には、画像符号化装置100は、どのような強度または性質のウィナーフィルタがかけられたかを示すインデックスを符号化ビットストリームに挿入する。これにより、画像復号装置200でも、画像符号化装置100でかけられたウィナーフィルタと同一の強度または性質のウィナーフィルタをかけることができる。 Further, in the first embodiment and the modification thereof, when the filter mode is the skip mode (DBF_SKIP), the weighting factor is set to ω = 1, and any processing target block or block boundary corresponding to the skip mode is set. Although a common filter (Wiener filter) is applied, it is also possible to apply Wiener filters having different strengths or properties to each processing target block or each block boundary corresponding to the skip mode. For example, Wiener filters having different filter coefficients and tap lengths may be applied. In this case, the image encoding apparatus 100 inserts an index indicating what strength or property of the Wiener filter has been applied to the encoded bitstream. As a result, the image decoding apparatus 200 can also apply a Wiener filter having the same strength or property as the Wiener filter applied by the image encoding apparatus 100.
 また、上記実施の形態1およびその変形例では、10種類のフィルタモードを用いたが、11種類以上のフィルタモード、または9種類以下のフィルタモードを用いてもよい。 Further, in the first embodiment and the modifications thereof, ten types of filter modes are used, but eleven or more types of filter modes or nine or less types of filter modes may be used.
 また、上記実施の形態1およびその変形例では、画像領域の一例としてフレームをあげたが、画像領域はスライスまたはシーケンスなどであってもよい。 In the first embodiment and its modification, a frame is used as an example of an image area. However, the image area may be a slice or a sequence.
 また、上記実施の形態1およびその変形例では、デブロッキングフィルタおよびウィナーフィルタのためのフィルタ係数群を算出することによってそのフィルタ係数群を導出したが、予め保持されている複数の候補の中から1つの候補を選択することによってそのフィルタ係数を導出してもよい。 In the first embodiment and the modification thereof, the filter coefficient group is derived by calculating the filter coefficient group for the deblocking filter and the Wiener filter. From among a plurality of candidates held in advance, The filter coefficient may be derived by selecting one candidate.
 例えば、画像符号化装置100のノイズ除去フィルタ106は、(式7)の演算を行うことによってフィルタ係数群を算出し、予め保持されている少なくとも1つのフィルタ係数群の候補の中から、その算出されたフィルタ係数群に最も近い候補を選択する。そして、ノイズ除去フィルタ106は、その選択された候補であるフィルタ係数群を用いたフィルタリングを行うとともに、そのフィルタ係数群を識別または特定するためのインデックスをエントロピー符号化部103に出力する。エントロピー符号化部103は、そのインデックスをエントロピー符号化して符号化ビットストリームに挿入する。一方、画像復号装置200のエントロピー復号部201は、符号化ビットストリームをエントロピー復号することによってその符号化ビットストリームからインデックスを抽出してノイズ除去フィルタ206に出力する。ノイズ除去フィルタ206は、その抽出されたインデックスによって識別また特定されるフィルタ係数群を、予め保持されている少なくとも1つのフィルタ係数群の候補の中から選択し、その選択されたフィルタ係数群を用いたフィルタリングを行う。 For example, the noise removal filter 106 of the image coding apparatus 100 calculates a filter coefficient group by performing the calculation of (Equation 7), and calculates the filter coefficient group from among at least one candidate filter coefficient group held in advance. A candidate closest to the set of filter coefficients is selected. The noise removal filter 106 performs filtering using the selected candidate filter coefficient group, and outputs an index for identifying or specifying the filter coefficient group to the entropy encoding unit 103. The entropy encoding unit 103 performs entropy encoding on the index and inserts the index into the encoded bit stream. On the other hand, the entropy decoding unit 201 of the image decoding apparatus 200 extracts an index from the encoded bitstream by entropy decoding the encoded bitstream, and outputs the index to the noise removal filter 206. The noise removal filter 206 selects a filter coefficient group identified or specified by the extracted index from among at least one candidate filter coefficient group held in advance, and uses the selected filter coefficient group. Filtering.
 なお、予め保持されているフィルタ係数群は、既に算出されたフィルタ係数群であってもよい。この場合、画像符号化装置100のノイズ除去フィルタ106は、フィルタ係数群を算出すると、その算出されたフィルタ係数群に対してインデックスを付し、そのインデックスが付されたフィルタ係数群を保持する。さらに、ノイズ除去フィルタ106は、そのインデックスが付されたフィルタ係数群を用いてフィルタリングを行うとともに、そのインデックスが付されたフィルタ係数群をエントロピー符号化部103に出力する。エントロピー符号化部103は、そのインデックスが付されたフィルタ係数群をエントロピー符号化して符号化ビットストリームに挿入する。一方、画像復号装置200のエントロピー復号部201は、符号化ビットストリームをエントロピー復号することによってその符号化ビットストリームから、インデックスが付されたフィルタ係数群を抽出してノイズ除去フィルタ206に出力する。ノイズ除去フィルタ206は、その抽出された、インデックスが付されたフィルタ係数群を保持しておくとともに、そのフィルタ係数群を用いたフィルタリングを行う。 It should be noted that the filter coefficient group held in advance may be a filter coefficient group that has already been calculated. In this case, when the noise removal filter 106 of the image encoding device 100 calculates the filter coefficient group, the noise coefficient filter 106 assigns an index to the calculated filter coefficient group, and holds the filter coefficient group to which the index is attached. Furthermore, the noise removal filter 106 performs filtering using the filter coefficient group to which the index is attached, and outputs the filter coefficient group to which the index is attached to the entropy coding unit 103. The entropy encoding unit 103 performs entropy encoding on the filter coefficient group to which the index is attached and inserts the filter coefficient group into the encoded bitstream. On the other hand, the entropy decoding unit 201 of the image decoding device 200 performs entropy decoding on the encoded bit stream, extracts a filter coefficient group with an index from the encoded bit stream, and outputs the filter coefficient group to the noise removal filter 206. The noise removal filter 206 holds the extracted filter coefficient group with an index and performs filtering using the filter coefficient group.
 (実施の形態2)
 上記実施の形態で示した画像符号化方法または画像復号方法の構成を実現するためのプログラムを記憶メディアに記録することにより、上記実施の形態で示した処理を独立したコンピュータシステムにおいて簡単に実施することが可能となる。記憶メディアは、磁気ディスク、光ディスク、光磁気ディスク、ICカード、半導体メモリ等、プログラムを記録できるものであればよい。
(Embodiment 2)
By recording a program for realizing the configuration of the image encoding method or the image decoding method described in the above embodiment on a storage medium, the processing described in the above embodiment can be easily performed in an independent computer system. It becomes possible. The storage medium may be any medium that can record a program, such as a magnetic disk, an optical disk, a magneto-optical disk, an IC card, and a semiconductor memory.
 さらにここで、上記実施の形態で示した画像符号化方法および画像復号方法の応用例とそれを用いたシステムを説明する。 Further, application examples of the image encoding method and the image decoding method shown in the above embodiment and a system using the same will be described here.
 図16は、コンテンツ配信サービスを実現するコンテンツ供給システムex100の全体構成を示す図である。通信サービスの提供エリアを所望の大きさに分割し、各セル内にそれぞれ固定無線局である基地局ex106~ex110が設置されている。 FIG. 16 is a diagram showing an overall configuration of a content supply system ex100 that realizes a content distribution service. The communication service providing area is divided into desired sizes, and base stations ex106 to ex110, which are fixed radio stations, are installed in each cell.
 このコンテンツ供給システムex100は、インターネットex101にインターネットサービスプロバイダex102および電話網ex104、および、基地局ex106~ex110を介して、コンピュータex111、PDA(Personal Digital Assistant)ex112、カメラex113、携帯電話ex114、ゲーム機ex115などの各機器が接続される。 The content supply system ex100 includes a computer ex111, a PDA (Personal Digital Assistant) ex112, a camera ex113, a mobile phone ex114, a game machine via the Internet ex101, the Internet service provider ex102, the telephone network ex104, and the base stations ex106 to ex110. Each device such as ex115 is connected.
 しかし、コンテンツ供給システムex100は図16のような構成に限定されず、いずれかの要素を組み合わせて接続するようにしてもよい。また、固定無線局である基地局ex106~ex110を介さずに、各機器が電話網ex104に直接接続されてもよい。また、各機器が近距離無線等を介して直接相互に接続されていてもよい。 However, the content supply system ex100 is not limited to the configuration shown in FIG. 16, and may be connected by combining any of the elements. Further, each device may be directly connected to the telephone network ex104 without going through the base stations ex106 to ex110 which are fixed wireless stations. In addition, the devices may be directly connected to each other via short-range wireless or the like.
 カメラex113はデジタルビデオカメラ等の動画撮影が可能な機器であり、カメラex116はデジタルカメラ等の静止画撮影、動画撮影が可能な機器である。また、携帯電話ex114は、GSM(Global System for Mobile Communications)方式、CDMA(Code Division Multiple Access)方式、W-CDMA(Wideband-Code Division Multiple Access)方式、もしくはLTE(Long Term Evolution)方式、HSPA(High Speed Packet Access)の携帯電話機、または、PHS(Personal Handyphone System)等であり、いずれでも構わない。 The camera ex113 is a device that can shoot moving images such as a digital video camera, and the camera ex116 is a device that can shoot still images and movies such as a digital camera. In addition, the mobile phone ex114 is a GSM (Global System for Mobile Communications) method, a CDMA (Code Division Multiple Access) method, a W-CDMA (Wideband-Code Division Multiple Access L (Semiconductor Access) method, a W-CDMA (Wideband-Code Division Multiple Access L method, or a high access rate). A High Speed Packet Access) mobile phone or a PHS (Personal Handyphone System) may be used.
 コンテンツ供給システムex100では、カメラex113等が基地局ex109、電話網ex104を通じてストリーミングサーバex103に接続されることで、ライブ配信等が可能になる。ライブ配信では、ユーザがカメラex113を用いて撮影するコンテンツ(例えば、音楽ライブの映像等)に対して上記実施の形態で説明したように符号化処理を行い、ストリーミングサーバex103に送信する。一方、ストリーミングサーバex103は要求のあったクライアントに対して送信されたコンテンツデータをストリーム配信する。クライアントとしては、上記符号化処理されたデータを復号することが可能な、コンピュータex111、PDAex112、カメラex113、携帯電話ex114、ゲーム機ex115等がある。配信されたデータを受信した各機器では、受信したデータを復号処理して再生する。 In the content supply system ex100, the camera ex113 and the like are connected to the streaming server ex103 through the base station ex109 and the telephone network ex104, thereby enabling live distribution and the like. In the live distribution, the content (for example, music live video) captured by the user using the camera ex113 is encoded as described in the above embodiment and transmitted to the streaming server ex103. On the other hand, the streaming server ex103 streams the content data transmitted to the requested client. Examples of the client include a computer ex111, a PDA ex112, a camera ex113, a mobile phone ex114, a game machine ex115, and the like that can decode the encoded data. Each device that has received the distributed data decodes and reproduces the received data.
 なお、撮影したデータの符号化処理はカメラex113で行っても、データの送信処理をするストリーミングサーバex103で行ってもよいし、互いに分担して行ってもよい。同様に配信されたデータの復号処理はクライアントで行っても、ストリーミングサーバex103で行ってもよいし、互いに分担して行ってもよい。また、カメラex113に限らず、カメラex116で撮影した静止画像および/または動画像データを、コンピュータex111を介してストリーミングサーバex103に送信してもよい。この場合の符号化処理はカメラex116、コンピュータex111、ストリーミングサーバex103のいずれで行ってもよいし、互いに分担して行ってもよい。 Note that the encoded processing of the captured data may be performed by the camera ex113, the streaming server ex103 that performs the data transmission processing, or may be performed in a shared manner. Similarly, the decryption processing of the distributed data may be performed by the client, the streaming server ex103, or may be performed in a shared manner. In addition to the camera ex113, still images and / or moving image data captured by the camera ex116 may be transmitted to the streaming server ex103 via the computer ex111. The encoding process in this case may be performed by any of the camera ex116, the computer ex111, and the streaming server ex103, or may be performed in a shared manner.
 また、これら符号化処理および復号処理は、一般的にコンピュータex111および各機器が有するLSI(Large Scale Integration)ex500において処理する。LSIex500は、ワンチップであっても複数チップからなる構成であってもよい。なお、画像符号化用および画像復号用のソフトウェアをコンピュータex111等で読み取り可能な何らかの記録メディア(CD-ROM、フレキシブルディスク、ハードディスクなど)に組み込み、そのソフトウェアを用いて符号化処理および復号処理を行ってもよい。さらに、携帯電話ex114がカメラ付きである場合には、そのカメラで取得した動画データを送信してもよい。このときの動画データは携帯電話ex114が有するLSIex500で符号化処理されたデータである。 In addition, these encoding processing and decoding processing are generally performed in a computer ex111 and an LSI (Large Scale Integration) ex500 included in each device. The LSI ex500 may be configured as a single chip or a plurality of chips. Note that image encoding and image decoding software is incorporated into some recording medium (CD-ROM, flexible disk, hard disk, etc.) that can be read by the computer ex111 and the like, and the encoding processing and decoding processing are performed using the software. May be. Furthermore, when the mobile phone ex114 is equipped with a camera, moving image data acquired by the camera may be transmitted. The moving image data at this time is data encoded by the LSI ex500 included in the mobile phone ex114.
 また、ストリーミングサーバex103は複数のサーバまたは複数のコンピュータであって、データを分散して処理したり記録したり配信するものであってもよい。 Further, the streaming server ex103 may be a plurality of servers or a plurality of computers, and may process, record, and distribute data in a distributed manner.
 以上のようにして、コンテンツ供給システムex100では、符号化されたデータをクライアントが受信して再生することができる。このようにコンテンツ供給システムex100では、ユーザが送信した情報をリアルタイムでクライアントが受信して復号し、再生することができ、特別な権利または設備を有さないユーザでも個人放送を実現できる。 As described above, in the content supply system ex100, the encoded data can be received and reproduced by the client. As described above, in the content supply system ex100, the information transmitted by the user can be received, decrypted and reproduced in real time by the client, and even a user who does not have special rights or facilities can realize personal broadcasting.
 このコンテンツ供給システムを構成する各機器の符号化、復号には上記実施の形態で示した画像符号化方法あるいは画像復号方法を用いるようにすればよい。 The image encoding method or the image decoding method shown in the above embodiment may be used for encoding and decoding of each device constituting the content supply system.
 その一例として携帯電話ex114について説明する。 As an example, a mobile phone ex114 will be described.
 図17は、上記実施の形態で説明した画像符号化方法と画像復号方法を用いた携帯電話ex114を示す図である。携帯電話ex114は、基地局ex110との間で電波を送受信するためのアンテナex601、CCDカメラ等の映像、静止画を撮ることが可能なカメラ部ex603、カメラ部ex603で撮影した映像、アンテナex601で受信した映像等が復号されたデータを表示する液晶ディスプレイ等の表示部ex602、操作キーex604群から構成される本体部、音声出力をするためのスピーカ等の音声出力部ex608、音声入力をするためのマイク等の音声入力部ex605、撮影した動画もしくは静止画のデータ、受信したメールのデータ、動画のデータもしくは静止画のデータ等、符号化されたデータまたは復号されたデータを保存するための記録メディアex607、携帯電話ex114に記録メディアex607を装着可能とするためのスロット部ex606を有している。記録メディアex607はSDカード等のプラスチックケース内に電気的に書換えおよび消去が可能な不揮発性メモリであるEEPROMの一種であるフラッシュメモリ素子を格納したものである。 FIG. 17 is a diagram illustrating the mobile phone ex114 using the image encoding method and the image decoding method described in the above embodiment. The cellular phone ex114 includes an antenna ex601 for transmitting and receiving radio waves to and from the base station ex110, a video from a CCD camera, a camera unit ex603 capable of taking a still image, a video shot by the camera unit ex603, and an antenna ex601. A display unit ex602 such as a liquid crystal display that displays data obtained by decoding received video and the like, a main body unit composed of a group of operation keys ex604, an audio output unit ex608 such as a speaker for outputting audio, and a voice input Audio input unit ex605 such as a microphone, recorded moving image or still image data, received mail data, moving image data or still image data, etc., for storing encoded data or decoded data Recording media ex607 can be attached to media ex607 and mobile phone ex114 And a slot unit ex606 for. The recording medium ex607 stores a flash memory element, which is a kind of EEPROM, which is a nonvolatile memory that can be electrically rewritten and erased, in a plastic case such as an SD card.
 さらに、携帯電話ex114について図18を用いて説明する。携帯電話ex114は表示部ex602および操作キーex604を備えた本体部の各部を統括的に制御するようになされた主制御部ex711に対して、電源回路部ex710、操作入力制御部ex704、画像符号化部ex712、カメラインターフェース部ex703、LCD(Liquid Crystal Display)制御部ex702、画像復号部ex709、多重分離部ex708、記録再生部ex707、変復調回路部ex706および音声処理部ex705が同期バスex713を介して互いに接続されている。 Further, the cellular phone ex114 will be described with reference to FIG. The mobile phone ex114 has a power supply circuit ex710, an operation input control unit ex704, an image encoding unit, and a main control unit ex711 configured to control the respective units of the main body unit including the display unit ex602 and the operation key ex604. Unit ex712, camera interface unit ex703, LCD (Liquid Crystal Display) control unit ex702, image decoding unit ex709, demultiplexing unit ex708, recording / reproducing unit ex707, modulation / demodulation circuit unit ex706, and audio processing unit ex705 are connected to each other via a synchronization bus ex713. It is connected.
 電源回路部ex710は、ユーザの操作により終話および電源キーがオン状態にされると、バッテリパックから各部に対して電力を供給することによりカメラ付デジタル携帯電話ex114を動作可能な状態に起動する。 When the end of call and the power key are turned on by a user operation, the power supply circuit ex710 activates the camera-equipped digital mobile phone ex114 by supplying power to each unit from the battery pack. .
 携帯電話ex114は、CPU、ROMおよびRAM等でなる主制御部ex711の制御に基づいて、音声通話モード時に音声入力部ex605で集音した音声信号を音声処理部ex705によってデジタル音声データに変換し、これを変復調回路部ex706でスペクトラム拡散処理し、送受信回路部ex701でデジタルアナログ変換処理および周波数変換処理を施した後にアンテナex601を介して送信する。また携帯電話ex114は、音声通話モード時にアンテナex601で受信した受信データを増幅して周波数変換処理およびアナログデジタル変換処理を施し、変復調回路部ex706でスペクトラム逆拡散処理し、音声処理部ex705によってアナログ音声データに変換した後、音声出力部ex608を介してこれを出力する。 The cellular phone ex114 converts the audio signal collected by the audio input unit ex605 in the audio call mode into digital audio data by the audio processing unit ex705 based on the control of the main control unit ex711 including a CPU, a ROM, a RAM, and the like. The modulation / demodulation circuit unit ex706 performs spread spectrum processing, the transmission / reception circuit unit ex701 performs digital analog conversion processing and frequency conversion processing, and then transmits the result via the antenna ex601. Further, the cellular phone ex114 amplifies the received data received by the antenna ex601 in the voice call mode, performs frequency conversion processing and analog-digital conversion processing, performs spectrum despreading processing by the modulation / demodulation circuit unit ex706, and performs analog speech by the voice processing unit ex705. After the data is converted, it is output via the audio output unit ex608.
 さらに、データ通信モード時に電子メールを送信する場合、本体部の操作キーex604の操作によって入力された電子メールのテキストデータは操作入力制御部ex704を介して主制御部ex711に送出される。主制御部ex711は、テキストデータを変復調回路部ex706でスペクトラム拡散処理し、送受信回路部ex701でデジタルアナログ変換処理および周波数変換処理を施した後にアンテナex601を介して基地局ex110へ送信する。 Further, when an e-mail is transmitted in the data communication mode, text data of the e-mail input by operating the operation key ex604 on the main body is sent to the main control unit ex711 via the operation input control unit ex704. The main control unit ex711 performs spread spectrum processing on the text data in the modulation / demodulation circuit unit ex706, performs digital analog conversion processing and frequency conversion processing in the transmission / reception circuit unit ex701, and then transmits the text data to the base station ex110 via the antenna ex601.
 データ通信モード時に画像データを送信する場合、カメラ部ex603で撮像された画像データを、カメラインターフェース部ex703を介して画像符号化部ex712に供給する。また、画像データを送信しない場合には、カメラ部ex603で撮像した画像データをカメラインターフェース部ex703およびLCD制御部ex702を介して表示部ex602に直接表示することも可能である。 When transmitting image data in the data communication mode, the image data captured by the camera unit ex603 is supplied to the image encoding unit ex712 via the camera interface unit ex703. When image data is not transmitted, the image data captured by the camera unit ex603 can be directly displayed on the display unit ex602 via the camera interface unit ex703 and the LCD control unit ex702.
 画像符号化部ex712は、本願発明で説明した画像符号化装置を備えた構成であり、カメラ部ex603から供給された画像データを上記実施の形態で示した画像符号化装置に用いた符号化方法によって圧縮符号化することにより符号化画像データに変換し、これを多重分離部ex708に送出する。また、このとき同時に携帯電話ex114は、カメラ部ex603で撮像中に音声入力部ex605で集音した音声を、音声処理部ex705を介してデジタルの音声データとして多重分離部ex708に送出する。 The image encoding unit ex712 is configured to include the image encoding device described in the present invention, and an encoding method using the image data supplied from the camera unit ex603 in the image encoding device described in the above embodiment. Is converted into encoded image data by compression encoding and sent to the demultiplexing unit ex708. At the same time, the mobile phone ex114 sends the sound collected by the sound input unit ex605 during imaging by the camera unit ex603 to the demultiplexing unit ex708 via the sound processing unit ex705 as digital sound data.
 多重分離部ex708は、画像符号化部ex712から供給された符号化画像データと音声処理部ex705から供給された音声データとを所定の方式で多重化し、その結果得られる多重化データを変復調回路部ex706でスペクトラム拡散処理し、送受信回路部ex701でデジタルアナログ変換処理および周波数変換処理を施した後にアンテナex601を介して送信する。 The demultiplexing unit ex708 multiplexes the encoded image data supplied from the image encoding unit ex712 and the audio data supplied from the audio processing unit ex705 by a predetermined method, and the resulting multiplexed data is a modulation / demodulation circuit unit Spread spectrum processing is performed in ex706, digital analog conversion processing and frequency conversion processing are performed in the transmission / reception circuit unit ex701, and then transmission is performed via the antenna ex601.
 データ通信モード時にホームページ等にリンクされた動画像ファイルのデータを受信する場合、アンテナex601を介して基地局ex110から受信した受信データを変復調回路部ex706でスペクトラム逆拡散処理し、その結果得られる多重化データを多重分離部ex708に送出する。 When data of a moving image file linked to a home page or the like is received in the data communication mode, the received data received from the base station ex110 via the antenna ex601 is subjected to spectrum despreading processing by the modulation / demodulation circuit unit ex706, and the resulting multiplexing is obtained. Data is sent to the demultiplexing unit ex708.
 また、アンテナex601を介して受信された多重化データを復号するには、多重分離部ex708は、多重化データを分離することにより画像データのビットストリームと音声データのビットストリームとに分け、同期バスex713を介して当該符号化画像データを画像復号部ex709に供給すると共に当該音声データを音声処理部ex705に供給する。 In addition, in order to decode multiplexed data received via the antenna ex601, the demultiplexing unit ex708 separates the multiplexed data into a bit stream of image data and a bit stream of audio data, and a synchronization bus The encoded image data is supplied to the image decoding unit ex709 via ex713 and the audio data is supplied to the audio processing unit ex705.
 次に、画像復号部ex709は、本願で説明した画像復号装置を備えた構成であり、画像データのビットストリームを上記実施の形態で示した符号化方法に対応した復号方法で復号することにより再生動画像データを生成し、これを、LCD制御部ex702を介して表示部ex602に供給し、これにより、例えばホームページにリンクされた動画像ファイルに含まれる動画データが表示される。このとき同時に音声処理部ex705は、音声データをアナログ音声データに変換した後、これを音声出力部ex608に供給し、これにより、例えばホームページにリンクされた動画像ファイルに含まれる音声データが再生される。 Next, the image decoding unit ex709 is configured to include the image decoding device described in the present application, and is reproduced by decoding the bit stream of the image data with a decoding method corresponding to the encoding method described in the above embodiment. Moving image data is generated and supplied to the display unit ex602 via the LCD control unit ex702, thereby displaying, for example, moving image data included in a moving image file linked to a home page. At the same time, the audio processing unit ex705 converts the audio data into analog audio data, and then supplies the analog audio data to the audio output unit ex608. Thus, for example, the audio data included in the moving image file linked to the home page is reproduced. The
 なお、上記システムの例に限られず、最近は衛星、地上波によるデジタル放送が話題となっており、図19に示すようにデジタル放送用システムにも上記実施の形態の少なくとも画像符号化装置または画像復号装置を組み込むことができる。具体的には、放送局ex201では音声データ、映像データまたはそれらのデータが多重化されたビットストリームが電波を介して通信または放送衛星ex202に伝送される。これを受けた放送衛星ex202は、放送用の電波を発信し、衛星放送受信設備をもつ家庭のアンテナex204はこの電波を受信し、テレビ(受信機)ex300またはセットトップボックス(STB)ex217などの装置はビットストリームを復号してこれを再生する。また、記録媒体であるCDおよびDVD等の記録メディアex215、ex216に記録した画像データと、音声データが多重化されたビットストリームを読み取り、復号するリーダ/レコーダex218にも上記実施の形態で示した画像復号装置を実装することが可能である。この場合、再生された映像信号はモニタex219に表示される。また、ケーブルテレビ用のケーブルex203または衛星/地上波放送のアンテナex204に接続されたセットトップボックスex217内に画像復号装置を実装し、これをテレビのモニタex219で再生する構成も考えられる。このときセットトップボックスではなく、テレビ内に画像復号装置を組み込んでも良い。また、アンテナex205を有する車ex210で、衛星ex202または基地局等から信号を受信し、車ex210が有するカーナビゲーションex211等の表示装置に動画を再生することも可能である。 Note that the present invention is not limited to the above-described system, and recently, digital broadcasting using satellites and terrestrial waves has become a hot topic. As shown in FIG. A decoding device can be incorporated. Specifically, in the broadcasting station ex201, audio data, video data, or a bit stream in which those data are multiplexed is transmitted to a communication or broadcasting satellite ex202 via radio waves. In response, the broadcasting satellite ex202 transmits a radio wave for broadcasting, and a home antenna ex204 having a satellite broadcasting receiving facility receives the radio wave, and the television (receiver) ex300 or the set top box (STB) ex217 or the like. The device decodes the bitstream and reproduces it. The reader / recorder ex218 that reads and decodes a bitstream in which image data and audio data recorded on recording media ex215 and ex216 such as CD and DVD as recording media are multiplexed is also shown in the above embodiment. It is possible to implement an image decoding device. In this case, the reproduced video signal is displayed on the monitor ex219. Further, a configuration in which an image decoding device is mounted in a set-top box ex217 connected to a cable ex203 for cable television or an antenna ex204 for satellite / terrestrial broadcasting, and this is reproduced on the monitor ex219 of the television is also conceivable. At this time, the image decoding apparatus may be incorporated in the television instead of the set top box. In addition, a car ex210 having an antenna ex205 can receive a signal from a satellite ex202 or a base station and reproduce a moving image on a display device such as a car navigation ex211 included in the car ex210.
 また、DVD、BD等の記録メディアex215に記録した音声データ、映像データまたはそれらのデータが多重化された符号化ビットストリームを読み取り復号する、または、記録メディアex215に、音声データ、映像データまたはそれらのデータを符号化し、多重化データとして記録するリーダ/レコーダex218にも上記実施の形態で示した画像復号装置または画像符号化装置を実装することが可能である。この場合、再生された映像信号はモニタex219に表示される。また、符号化ビットストリームが記録された記録メディアex215により、他の装置およびシステム等は、映像信号を再生することができる。例えば、他の再生装置ex212は、符号化ビットストリームがコピーされた記録メディアex214を用いて、モニタex213に映像信号を再生することができる。 Also, audio data, video data recorded on a recording medium ex215 such as DVD or BD, or an encoded bit stream in which those data are multiplexed are read and decoded, or audio data, video data or these are recorded on the recording medium ex215. The image decoding apparatus or the image encoding apparatus described in the above embodiment can also be mounted on the reader / recorder ex218 that encodes the data and records the multiplexed data as multiplexed data. In this case, the reproduced video signal is displayed on the monitor ex219. In addition, the recording medium ex215 on which the encoded bit stream is recorded allows other devices and systems to reproduce the video signal. For example, the other reproduction device ex212 can reproduce the video signal on the monitor ex213 using the recording medium ex214 on which the encoded bitstream is copied.
 また、ケーブルテレビ用のケーブルex203または衛星/地上波放送のアンテナex204に接続されたセットトップボックスex217内に画像復号装置を実装し、これをテレビのモニタex219で表示してもよい。このときセットトップボックスではなく、テレビ内に画像復号装置を組み込んでもよい。 Also, an image decoding device may be mounted in the set-top box ex217 connected to the cable ex203 for cable television or the antenna ex204 for satellite / terrestrial broadcasting and displayed on the monitor ex219 of the television. At this time, the image decoding apparatus may be incorporated in the television instead of the set top box.
 図20は、上記実施の形態で説明した画像復号方法および画像符号化方法を用いたテレビ(受信機)ex300を示す図である。テレビex300は、上記放送を受信するアンテナex204またはケーブルex203等を介して映像情報のビットストリームを取得、または、出力するチューナex301と、受信した符号化データを復調する、または、生成された符号化データを外部に送信するために変調する変調/復調部ex302と、復調した映像データと音声データとを分離する、または、符号化された映像データと音声データとを多重化する多重/分離部ex303を備える。また、テレビex300は、音声データ、映像データそれぞれを復号する、または、それぞれの情報を符号化する音声信号処理部ex304、映像信号処理部ex305を有する信号処理部ex306と、復号された音声信号を出力するスピーカex307、復号された映像信号を表示するディスプレイ等の表示部ex308を有する出力部ex309とを有する。さらに、テレビex300は、ユーザ操作の入力を受け付ける操作入力部ex312等を有するインターフェース部ex317を有する。さらに、テレビex300は、各部を統括的に制御する制御部ex310、各部に電力を供給する電源回路部ex311を有する。インターフェース部ex317は、操作入力部ex312以外に、リーダ/レコーダex218等の外部機器と接続されるブリッジex313、SDカード等の記録メディアex216を装着可能とするためのスロット部ex314、ハードディスク等の外部記録メディアと接続するためのドライバex315、電話網と接続するモデムex316等を有していてもよい。なお記録メディアex216は、格納する不揮発性/揮発性の半導体メモリ素子により電気的に情報の記録を可能としたものである。テレビex300の各部は同期バスを介して互いに接続されている。 FIG. 20 is a diagram illustrating a television (receiver) ex300 that uses the image decoding method and the image encoding method described in the above embodiment. The television ex300 obtains or outputs a bit stream of video information via the antenna ex204 or the cable ex203 that receives the broadcast, and a tuner ex301 that outputs or outputs the encoded data that is received or demodulated. Modulation / demodulation unit ex302 that modulates data for transmission to the outside, and multiplexing / separation unit ex303 that separates demodulated video data and audio data, or multiplexes encoded video data and audio data Is provided. In addition, the television ex300 decodes each of the audio data and the video data, or encodes the respective information, the audio signal processing unit ex304, the signal processing unit ex306 including the video signal processing unit ex305, and the decoded audio signal. And an output unit ex309 including a display unit ex308 such as a display for displaying the decoded video signal. Furthermore, the television ex300 includes an interface unit ex317 including an operation input unit ex312 that receives an input of a user operation. Furthermore, the television ex300 includes a control unit ex310 that controls each unit in an integrated manner, and a power supply circuit unit ex311 that supplies power to each unit. In addition to the operation input unit ex312, the interface unit ex317 includes a bridge ex313 connected to an external device such as a reader / recorder ex218, a slot unit ex314 for enabling recording media ex216 such as an SD card, and an external recording such as a hard disk A driver ex315 for connecting to a medium, a modem ex316 for connecting to a telephone network, and the like may be included. The recording medium ex216 is capable of electrically recording information by using a nonvolatile / volatile semiconductor memory element to be stored. Each part of the television ex300 is connected to each other via a synchronous bus.
 まず、テレビex300がアンテナex204等により外部から取得したデータを復号し、再生する構成について説明する。テレビex300は、リモートコントローラex220等からのユーザ操作を受け、CPU等を有する制御部ex310の制御に基づいて、変調/復調部ex302で復調した映像データ、音声データを多重/分離部ex303で分離する。さらにテレビex300は、分離した音声データを音声信号処理部ex304で復号し、分離した映像データを映像信号処理部ex305で上記実施の形態で説明した復号方法を用いて復号する。復号した音声信号、映像信号は、それぞれ出力部ex309から外部に向けて出力される。出力する際には、音声信号と映像信号が同期して再生するよう、バッファex318、ex319等に一旦これらの信号を蓄積するとよい。また、テレビex300は、放送等からではなく、磁気/光ディスク、SDカード等の記録メディアex215、ex216から符号化された符号化ビットストリームを読み出してもよい。次に、テレビex300が音声信号および映像信号を符号化し、外部に送信または記録メディア等に書き込む構成について説明する。テレビex300は、リモートコントローラex220等からのユーザ操作を受け、制御部ex310の制御に基づいて、音声信号処理部ex304で音声信号を符号化し、映像信号処理部ex305で映像信号を上記実施の形態で説明した符号化方法を用いて符号化する。符号化した音声信号、映像信号は多重/分離部ex303で多重化され外部に出力される。多重化する際には、音声信号と映像信号が同期するように、バッファex320、ex321等に一旦これらの信号を蓄積するとよい。なお、バッファex318~ex321は図示しているように複数備えていてもよいし、一つ以上のバッファを共有する構成であってもよい。さらに、図示している以外に、例えば変調/復調部ex302と多重/分離部ex303との間等でもシステムのオーバフローおよびアンダーフローを避ける緩衝材としてバッファにデータを蓄積することとしてもよい。 First, a configuration in which the television ex300 decodes and reproduces data acquired from the outside by the antenna ex204 and the like will be described. The television ex300 receives a user operation from the remote controller ex220 or the like, and demultiplexes the video data and audio data demodulated by the modulation / demodulation unit ex302 by the multiplexing / separation unit ex303 based on the control of the control unit ex310 having a CPU or the like. . Furthermore, the television ex300 decodes the separated audio data by the audio signal processing unit ex304, and the separated video data is decoded by the video signal processing unit ex305 using the decoding method described in the above embodiment. The decoded audio signal and video signal are output to the outside from the output unit ex309. When outputting, these signals may be temporarily stored in the buffers ex318, ex319, etc. so that the audio signal and the video signal are reproduced in synchronization. Further, the television ex300 may read the encoded bitstream encoded from the recording media ex215 and ex216 such as a magnetic / optical disk and an SD card, not from a broadcast or the like. Next, a configuration will be described in which the television ex300 encodes an audio signal and a video signal and transmits them to the outside or writes them to a recording medium or the like. The television ex300 receives a user operation from the remote controller ex220 or the like, and encodes an audio signal with the audio signal processing unit ex304 based on the control of the control unit ex310, and the video signal with the video signal processing unit ex305 in the above embodiment. Encoding is performed using the described encoding method. The encoded audio signal and video signal are multiplexed by the multiplexing / demultiplexing unit ex303 and output to the outside. When multiplexing, these signals may be temporarily stored in the buffers ex320 and ex321 so that the audio signal and the video signal are synchronized. Note that a plurality of buffers ex318 to ex321 may be provided as shown in the figure, or a configuration in which one or more buffers are shared may be used. Further, in addition to the illustrated example, data may be stored in the buffer as a buffer material that prevents system overflow and underflow even between the modulation / demodulation unit ex302 and the multiplexing / demultiplexing unit ex303, for example.
 また、テレビex300は、放送および記録メディア等から音声データおよび映像データを取得する以外に、マイクおよびカメラのAV入力を受け付ける構成を備え、それらから取得したデータに対して符号化処理を行ってもよい。なお、ここではテレビex300は、上記の符号化処理、多重化、および、外部出力ができる構成として説明したが、これらのすべての処理を行うことはできず、上記受信、復号処理、および、外部出力のうちいずれかのみが可能な構成であってもよい。 In addition to acquiring audio data and video data from broadcast and recording media, the television ex300 has a configuration for receiving AV input of a microphone and a camera, and even if encoding processing is performed on the data acquired therefrom Good. Here, the television ex300 has been described as a configuration capable of the above-described encoding processing, multiplexing, and external output. However, all of these processing cannot be performed, and the above reception, decoding processing, and external The configuration may be such that only one of the outputs is possible.
 また、リーダ/レコーダex218で記録メディアから符号化ビットストリームを読み出す、または、書き込む場合には、上記復号処理または符号化処理はテレビex300およびリーダ/レコーダex218のうちいずれかで行ってもよいし、テレビex300とリーダ/レコーダex218とが互いに分担して行ってもよい。 When the encoded bitstream is read from or written to the recording medium by the reader / recorder ex218, the decoding process or the encoding process may be performed by either the television ex300 or the reader / recorder ex218. The television ex300 and the reader / recorder ex218 may be shared with each other.
 一例として、光ディスクからデータの読み込みまたは書き込みをする場合の情報再生/記録部ex400の構成を図21に示す。情報再生/記録部ex400は、以下に説明する要素ex401~ex407を備える。光ヘッドex401は、光ディスクである記録メディアex215の記録面にレーザスポットを照射して情報を書き込み、記録メディアex215の記録面からの反射光を検出して情報を読み込む。変調記録部ex402は、光ヘッドex401に内蔵された半導体レーザを電気的に駆動し記録データに応じてレーザ光の変調を行う。再生復調部ex403は、光ヘッドex401に内蔵されたフォトディテクタにより記録面からの反射光を電気的に検出した再生信号を増幅し、記録メディアex215に記録された信号成分を分離して復調し、必要な情報を再生する。バッファex404は、記録メディアex215に記録するための情報および記録メディアex215から再生した情報を一時的に保持する。ディスクモータex405は記録メディアex215を回転させる。サーボ制御部ex406は、ディスクモータex405の回転駆動を制御しながら光ヘッドex401を所定の情報トラックに移動させ、レーザスポットの追従処理を行う。システム制御部ex407は、情報再生/記録部ex400全体の制御を行う。上記の読み出しおよび書き込みの処理は、システム制御部ex407が、バッファex404に保持された各種情報を利用し、また必要に応じて新たな情報の生成および追加を行うと共に、変調記録部ex402、再生復調部ex403およびサーボ制御部ex406を協調動作させながら、光ヘッドex401を通して、情報の記録再生を行うことにより実現される。システム制御部ex407は、例えばマイクロプロセッサで構成され、読み出し書き込みのプログラムを実行することでそれらの処理を実行する。 As an example, FIG. 21 shows the configuration of the information reproducing / recording unit ex400 when data is read from or written to an optical disk. The information reproducing / recording unit ex400 includes elements ex401 to ex407 described below. The optical head ex401 irradiates a laser spot on the recording surface of the recording medium ex215 that is an optical disc to write information, and detects information reflected from the recording surface of the recording medium ex215 to read the information. The modulation recording unit ex402 electrically drives a semiconductor laser built in the optical head ex401 and modulates the laser beam according to the recording data. The reproduction demodulator ex403 amplifies a reproduction signal obtained by electrically detecting reflected light from the recording surface by a photodetector built in the optical head ex401, separates and demodulates a signal component recorded on the recording medium ex215, and is necessary. To play back information. The buffer ex404 temporarily holds information to be recorded on the recording medium ex215 and information reproduced from the recording medium ex215. The disk motor ex405 rotates the recording medium ex215. The servo control unit ex406 moves the optical head ex401 to a predetermined information track while controlling the rotational drive of the disk motor ex405, and performs a laser spot tracking process. The system control unit ex407 controls the entire information reproduction / recording unit ex400. In the reading and writing processes described above, the system control unit ex407 uses various types of information held in the buffer ex404, and generates and adds new information as necessary. This is realized by recording / reproducing information through the optical head ex401 while the unit ex403 and the servo control unit ex406 are operated cooperatively. The system control unit ex407 includes, for example, a microprocessor, and executes these processes by executing a read / write program.
 以上では、光ヘッドex401はレーザスポットを照射するとして説明したが、近接場光を用いてより高密度な記録を行う構成であってもよい。 In the above, the optical head ex401 has been described as irradiating a laser spot, but it may be configured to perform higher-density recording using near-field light.
 図22に光ディスクである記録メディアex215の模式図を示す。記録メディアex215の記録面には案内溝(グルーブ)がスパイラル状に形成され、情報トラックex230には、あらかじめグルーブの形状の変化によってディスク上の絶対位置を示す番地情報が記録されている。この番地情報はデータを記録する単位である記録ブロックex231の位置を特定するための情報を含み、記録および再生を行う装置は、情報トラックex230を再生し番地情報を読み取ることで記録ブロックを特定することができる。また、記録メディアex215は、データ記録領域ex233、内周領域ex232、外周領域ex234を含んでいる。ユーザデータを記録するために用いる領域がデータ記録領域ex233であり、データ記録領域ex233の内周または外周に配置されている内周領域ex232と外周領域ex234は、ユーザデータの記録以外の特定用途に用いられる。情報再生/記録部ex400は、このような記録メディアex215のデータ記録領域ex233に対して、符号化された音声データ、映像データまたはそれらのデータを多重化した符号化データの読み書きを行う。 FIG. 22 shows a schematic diagram of a recording medium ex215 that is an optical disk. Guide grooves (grooves) are formed in a spiral shape on the recording surface of the recording medium ex215, and address information indicating the absolute position on the disc is recorded in advance on the information track ex230 by changing the shape of the groove. This address information includes information for specifying the position of the recording block ex231 which is a unit for recording data, and the recording and reproducing apparatus specifies the recording block by reproducing the information track ex230 and reading the address information. be able to. Further, the recording medium ex215 includes a data recording area ex233, an inner peripheral area ex232, and an outer peripheral area ex234. The area used for recording user data is the data recording area ex233, and the inner circumference area ex232 and the outer circumference area ex234 arranged on the inner circumference or outer circumference of the data recording area ex233 are used for specific purposes other than recording user data. Used. The information reproducing / recording unit ex400 reads / writes encoded audio data, video data, or encoded data obtained by multiplexing these data, with respect to the data recording area ex233 of the recording medium ex215.
 以上では、1層のDVD、BD等の光ディスクを例に挙げ説明したが、これらに限ったものではなく、多層構造であって表面以外にも記録可能な光ディスクであってもよい。また、ディスクの同じ場所にさまざまな異なる波長の色の光を用いて情報を記録したり、さまざまな角度から異なる情報の層を記録したりするなど、多次元的な記録/再生を行う構造の光ディスクであってもよい。 In the above description, an optical disk such as a single-layer DVD or BD has been described as an example. However, the present invention is not limited to these, and an optical disk having a multilayer structure and capable of recording other than the surface may be used. It also has a structure that performs multidimensional recording / reproduction, such as recording information using light of various different wavelengths at the same location on the disc, and recording different layers of information from various angles. It may be an optical disk.
 また、デジタル放送用システムex200において、アンテナex205を有する車ex210で衛星ex202等からデータを受信し、車ex210が有するカーナビゲーションex211等の表示装置に動画を再生することも可能である。なお、カーナビゲーションex211の構成は例えば図20に示す構成のうち、GPS受信部を加えた構成が考えられ、同様なことがコンピュータex111および携帯電話ex114等でも考えられる。また、上記携帯電話ex114等の端末は、テレビex300と同様に、符号化器および復号器を両方持つ送受信型端末の他に、符号化器のみの送信端末、復号器のみの受信端末という3通りの実装形式が考えられる。 Also, in the digital broadcasting system ex200, the car ex210 having the antenna ex205 can receive data from the satellite ex202 and the like, and the moving image can be reproduced on a display device such as the car navigation ex211 that the car ex210 has. The configuration of the car navigation ex211 may be, for example, a configuration in which a GPS receiving unit is added in the configuration illustrated in FIG. In addition to the transmission / reception terminal having both an encoder and a decoder, the mobile phone ex114 and the like can be used in three ways: a transmitting terminal having only an encoder and a receiving terminal having only a decoder. The implementation form of can be considered.
 このように、上記実施の形態で示した画像符号化方法あるいは画像復号方法を上述したいずれの機器およびシステムに用いることは可能であり、そうすることで、上記実施の形態で説明した効果を得ることができる。 As described above, the image encoding method or the image decoding method described in the above embodiment can be used in any of the above-described devices and systems, and by doing so, the effects described in the above embodiment can be obtained. be able to.
 また、本発明は、上記実施の形態に限定されるものではなく、本発明の範囲を逸脱することなく種々の変形または修正が可能である。 Further, the present invention is not limited to the above-described embodiment, and various changes or modifications can be made without departing from the scope of the present invention.
 (実施の形態3)
 上記各実施の形態で示した画像符号化方法および装置、画像復号方法および装置は、典型的には集積回路であるLSIで実現される。一例として、図23に1チップ化されたLSIex500の構成を示す。LSIex500は、以下に説明する要素ex501~ex509を備え、各要素はバスex510を介して接続している。電源回路部ex505は電源がオン状態の場合に各部に対して電力を供給することで動作可能な状態に起動する。
(Embodiment 3)
The image encoding method and apparatus and the image decoding method and apparatus described in the above embodiments are typically realized by an LSI that is an integrated circuit. As an example, FIG. 23 shows a configuration of an LSI ex500 that is made into one chip. The LSI ex500 includes elements ex501 to ex509 described below, and each element is connected via a bus ex510. The power supply circuit unit ex505 starts up to an operable state by supplying power to each unit when the power supply is in an on state.
 例えば符号化処理を行う場合には、LSIex500は、CPUex502、メモリコントローラex503およびストリームコントローラex504等を有する制御部ex501の制御に基づいて、AV I/Oex509によりマイクex117およびカメラex113等からAV信号の入力を受け付ける。入力されたAV信号は、一旦SDRAM等の外部のメモリex511に蓄積される。制御部ex501の制御に基づいて、蓄積したデータは、処理量および処理速度に応じて適宜複数回に分けるなどされ、信号処理部ex507に送られる。信号処理部ex507は、音声信号の符号化および/または映像信号の符号化を行う。ここで映像信号の符号化処理は、上記実施の形態で説明した符号化処理である。信号処理部ex507ではさらに、場合により符号化された音声データと符号化された映像データを多重化するなどの処理を行い、ストリームI/Oex506から外部に出力する。この出力されたビットストリームは、基地局ex107に向けて送信されたり、または、記録メディアex215に書き込まれたりする。なお、多重化する際には同期するよう、一旦バッファex508にデータを蓄積するとよい。 For example, when performing the encoding process, the LSI ex500 inputs an AV signal from the microphone ex117, the camera ex113, and the like by the AV I / Oex 509 based on the control of the control unit ex501 having the CPU ex502, the memory controller ex503, the stream controller ex504, and the like. Accept. The input AV signal is temporarily stored in an external memory ex511 such as SDRAM. Based on the control of the control unit ex501, the accumulated data is appropriately divided into a plurality of times according to the processing amount and the processing speed, and sent to the signal processing unit ex507. The signal processing unit ex507 performs encoding of an audio signal and / or encoding of a video signal. Here, the encoding process of the video signal is the encoding process described in the above embodiment. The signal processing unit ex507 further performs processing such as multiplexing the encoded audio data and the encoded video data according to circumstances, and outputs the result from the stream I / Oex 506 to the outside. The output bit stream is transmitted to the base station ex107 or written to the recording medium ex215. It should be noted that data should be temporarily stored in the buffer ex508 so that the data is synchronized when multiplexed.
 また、例えば復号処理を行う場合には、LSIex500は、制御部ex501の制御に基づいて、ストリームI/Oex506によって基地局ex107を介して得た符号化データ、または、記録メディアex215から読み出して得た符号化データを一旦メモリex511等に蓄積する。制御部ex501の制御に基づいて、蓄積したデータは、処理量および処理速度に応じて適宜複数回に分けるなどされ信号処理部ex507に送られる。信号処理部ex507は、音声データの復号および/または映像データの復号を行う。ここで映像信号の復号処理は、上記実施の形態で説明した復号処理である。さらに、場合により復号された音声信号と復号された映像信号を同期して再生できるようそれぞれの信号を一旦バッファex508等に蓄積するとよい。復号された出力信号は、メモリex511等を適宜介しながら、携帯電話ex114、ゲーム機ex115およびテレビex300等の各出力部から出力される。 Further, for example, when performing decoding processing, the LSI ex500 is obtained by reading from the encoded data obtained via the base station ex107 by the stream I / Oex 506 or the recording medium ex215 based on the control of the control unit ex501. The encoded data is temporarily stored in the memory ex511 or the like. Based on the control of the control unit ex501, the accumulated data is appropriately divided into a plurality of times according to the processing amount and the processing speed and sent to the signal processing unit ex507. The signal processing unit ex507 performs decoding of audio data and / or decoding of video data. Here, the decoding process of the video signal is the decoding process described in the above embodiment. Further, in some cases, each signal may be temporarily stored in the buffer ex508 or the like so that the decoded audio signal and the decoded video signal can be reproduced in synchronization. The decoded output signal is output from each output unit such as the mobile phone ex114, the game machine ex115, and the television ex300 through the memory ex511 or the like as appropriate.
 なお、上記では、メモリex511がLSIex500の外部の構成として説明したが、LSIex500の内部に含まれる構成であってもよい。バッファex508も一つに限ったものではなく、複数のバッファを備えていてもよい。また、LSIex500は1チップ化されてもよいし、複数チップ化されてもよい。 In the above description, the memory ex511 has been described as an external configuration of the LSI ex500. However, a configuration included in the LSI ex500 may be used. The buffer ex508 is not limited to one, and a plurality of buffers may be provided. The LSI ex500 may be made into one chip or a plurality of chips.
 なお、ここでは、LSIとしたが、集積度の違いにより、IC、システムLSI、スーパーLSI、ウルトラLSIと呼称されることもある。 In addition, although it was set as LSI here, it may be called IC, system LSI, super LSI, and ultra LSI depending on the degree of integration.
 また、集積回路化の手法はLSIに限るものではなく、専用回路または汎用プロセッサで実現してもよい。LSI製造後に、プログラムすることが可能なFPGA、または、LSI内部の回路セルの接続および設定を再構成可能なリコンフィギュラブル・プロセッサを利用してもよい。 Further, the method of circuit integration is not limited to LSI, and implementation with a dedicated circuit or a general-purpose processor is also possible. An FPGA that can be programmed after manufacturing the LSI or a reconfigurable processor that can reconfigure the connection and setting of circuit cells inside the LSI may be used.
 さらには、半導体技術の進歩または派生する別技術によりLSIに置き換わる集積回路化の技術が登場すれば、当然、その技術を用いて機能ブロックの集積化を行ってもよい。バイオ技術の適応等が可能性としてありえる。 Furthermore, if integrated circuit technology that replaces LSI emerges as a result of advances in semiconductor technology or other derived technology, it is naturally also possible to integrate functional blocks using this technology. Biotechnology can be applied.
 以上、本発明に係る符号化方法、符号化装置、復号方法および復号装置について、実施の形態に基づいて説明したが、本発明は、これらの実施の形態に限定されるものではない。本発明の趣旨を逸脱しない限り、当業者が思いつく各種変形を当該実施の形態に施した形態、および、異なる実施の形態における構成要素およびステップ等を組み合わせて構築される別の形態も、本発明の範囲内に含まれる。 As mentioned above, although the encoding method, the encoding device, the decoding method, and the decoding device according to the present invention have been described based on the embodiments, the present invention is not limited to these embodiments. Unless it deviates from the meaning of the present invention, various forms conceived by those skilled in the art are applied to the embodiment, and other forms constructed by combining components and steps in different embodiments are also included in the present invention. It is included in the range.
 本発明にかかる画像符号化方法および画像復号方法は、符号化効率および画質の向上を図ることができるという効果を奏し、例えば、ビデオカメラ、動画の撮影および再生機能を有する携帯電話、パーソナルコンピュータ、または録画再生装置などに適用することができる。 The image encoding method and the image decoding method according to the present invention have the effect of being able to improve the encoding efficiency and the image quality. For example, a video camera, a mobile phone having a video shooting and playback function, a personal computer, Alternatively, it can be applied to a recording / playback apparatus.
 10  画像符号化装置
 11  符号化部
 12  再構成部
 13  フィルタ係数導出部
 14  フィルタリング部
 15  挿入部
 20  画像復号装置
 21  再構成部
 22  抽出部
 23  フィルタリング部
 100  画像符号化装置
 101  減算器
 102  変換量子化部
 103  エントロピー符号化部
 104  逆量子化逆変換部
 105  加算器
 106  ノイズ除去フィルタ
 107  予測部
 200  画像復号装置
 201  エントロピー復号部
 202  逆量子化逆変換部
 204  予測部
 206  ノイズ除去フィルタ
 ex100 コンテンツ供給システム
 ex101 インターネット
 ex102 インターネットサービスプロバイダ
 ex103 ストリーミングサーバ
 ex104 電話網
 ex106,ex107,ex108,ex109,ex110 基地局
 ex111 コンピュータ
 ex112 PDA
 ex113,ex116 カメラ
 ex114 カメラ付デジタル携帯電話(携帯電話)
 ex115 ゲーム機
 ex117 マイク
 ex200 デジタル放送用システム
 ex201 放送局
 ex202 放送衛星(衛星)
 ex203 ケーブル
 ex204,ex205,ex601 アンテナ
 ex210 車
 ex211 カーナビゲーション(カーナビ)
 ex212 再生装置
 ex213,ex219 モニタ
 ex214,ex215,ex216,ex607 記録メディア
 ex217 セットトップボックス(STB)
 ex218 リーダ/レコーダ
 ex220 リモートコントローラ
 ex230 情報トラック
 ex231 記録ブロック
 ex232 内周領域
 ex233 データ記録領域
 ex234 外周領域
 ex300 テレビ
 ex301 チューナ
 ex302 変調/復調部
 ex303 多重/分離部
 ex304 音声信号処理部
 ex305 映像信号処理部
 ex306,ex507 信号処理部
 ex307 スピーカ
 ex308,ex602 表示部
 ex309 出力部
 ex310,ex501 制御部
 ex311,ex505,ex710 電源回路部
 ex312 操作入力部
 ex313 ブリッジ
 ex314,ex606 スロット部
 ex315 ドライバ
 ex316 モデム
 ex317 インターフェース部
 ex318,ex319,ex320,ex321,ex404,ex508 バッファ
 ex400 情報再生/記録部
 ex401 光ヘッド
 ex402 変調記録部
 ex403 再生復調部
 ex405 ディスクモータ
 ex406 サーボ制御部
 ex407 システム制御部
 ex500 LSI
 ex502 CPU
 ex503 メモリコントローラ
 ex504 ストリームコントローラ
 ex506 ストリームI/O
 ex509 AV I/O
 ex510 バス
 ex603 カメラ部
 ex604 操作キー
 ex605 音声入力部
 ex608 音声出力部
 ex701 送受信回路部
 ex702 LCD制御部
 ex703 カメラインターフェース部(カメラI/F部)
 ex704 操作入力制御部
 ex705 音声処理部
 ex706 変復調回路部
 ex707 記録再生部
 ex708 多重分離部
 ex709 画像復号部
 ex711 主制御部
 ex712 画像符号化部
 ex713 同期バス
DESCRIPTION OF SYMBOLS 10 Image coding apparatus 11 Encoding part 12 Reconstruction part 13 Filter coefficient derivation part 14 Filtering part 15 Insertion part 20 Image decoding apparatus 21 Reconstruction part 22 Extraction part 23 Filtering part 100 Image coding apparatus 101 Subtractor 102 Transform quantization Unit 103 entropy encoding unit 104 inverse quantization inverse transform unit 105 adder 106 noise removal filter 107 prediction unit 200 image decoding device 201 entropy decoding unit 202 inverse quantization inverse transform unit 204 prediction unit 206 noise removal filter ex100 content supply system ex101 Internet ex102 Internet service provider ex103 Streaming server ex104 Telephone network ex106, ex107, ex108, ex109, ex110 Base station ex1 1 computer ex112 PDA
ex113, ex116 Camera ex114 Digital mobile phone with camera (mobile phone)
ex115 Game console ex117 Microphone ex200 Digital broadcasting system ex201 Broadcast station ex202 Broadcast satellite (satellite)
ex203 Cable ex204, ex205, ex601 Antenna ex210 Car ex211 Car navigation (car navigation system)
ex212 Playback device ex213, ex219 Monitor ex214, ex215, ex216, ex607 Recording media ex217 Set-top box (STB)
ex218 reader / recorder ex220 remote controller ex230 information track ex231 recording block ex232 inner circumference area ex233 data recording area ex234 outer circumference area ex300 television ex301 tuner ex302 modulation / demodulation section ex303 multiplexing / separation section ex304 audio signal processing section ex305 video signal processing section ex306, ex507 signal processing unit ex307 speaker ex308, ex602 display unit ex309 output unit ex310, ex501 control unit ex311, ex505, ex710 power supply circuit unit ex312 operation input unit ex313 bridge ex314, ex606 slot unit ex315 driver ex316 modem ex317, interface unit 3318 , Ex321 x404, ex508 Buffer ex400 Information reproducing / recording unit ex401 optical head ex402 modulation recording unit ex403 reproducing demodulating portion ex405 Disk motor ex406 Servo control unit ex407 System control unit EX500 LSI
ex502 CPU
ex503 Memory controller ex504 Stream controller ex506 Stream I / O
ex509 AV I / O
ex510 bus ex603 camera unit ex604 operation key ex605 audio input unit ex608 audio output unit ex701 transmission / reception circuit unit ex702 LCD control unit ex703 camera interface unit (camera I / F unit)
ex704 operation input control unit ex705 audio processing unit ex706 modulation / demodulation circuit unit ex707 recording / playback unit ex708 demultiplexing unit ex709 image decoding unit ex711 main control unit ex712 image encoding unit ex713 synchronization bus

Claims (24)

  1.  動画像を符号化する画像符号化方法であって、
     予測画像を用いて前記動画像をブロックごとに符号化することによって符号化ビットストリームを生成し、
     符号化された前記ブロックを順次再構成し、
     再構成されたブロック間の境界であるブロック境界の歪を取り除くための歪除去フィルタ係数を、前記ブロック境界の特徴に応じて導出し、
     予測画像を生成するために、導出された前記歪除去フィルタ係数を用いたフィルタリングを前記ブロック境界に対して行い、
     導出された前記歪除去フィルタ係数を特定するための係数特定情報を前記符号化ビットストリームに挿入する
     画像符号化方法。
    An image encoding method for encoding a moving image,
    Generating an encoded bitstream by encoding the moving image block by block using a predicted image;
    Sequentially reconstructing the encoded blocks;
    Deriving a distortion removal filter coefficient for removing distortion at a block boundary, which is a boundary between reconstructed blocks, according to the feature of the block boundary;
    In order to generate a predicted image, filtering using the derived distortion removal filter coefficient is performed on the block boundary,
    An image encoding method for inserting coefficient specifying information for specifying the derived distortion removal filter coefficient into the encoded bitstream.
  2.  前記歪除去フィルタ係数の導出では、前記歪除去フィルタ係数を算出することによって、当該歪除去フィルタ係数を導出し、
     前記係数特定情報の前記符号化ビットストリームへの挿入では、導出された前記歪除去フィルタ係数を前記係数特定情報として前記符号化ビットストリームに挿入する
     請求項1記載の画像符号化方法。
    In the derivation of the distortion removal filter coefficient, the distortion removal filter coefficient is derived by calculating the distortion removal filter coefficient,
    The image coding method according to claim 1, wherein in the insertion of the coefficient specifying information into the encoded bitstream, the derived distortion removal filter coefficient is inserted into the encoded bitstream as the coefficient specifying information.
  3.  前記画像符号化方法は、さらに、
     ブロック境界の特徴に応じて、前記歪除去フィルタ係数を導出すべきか否かを判別し、
     前記歪除去フィルタの導出の判別において、前記歪除去フィルタ係数を導出すべきでないと判別された際には、
     前記ブロック境界に対応する処理対象ブロック内のノイズを取り除くための内部ノイズ除去フィルタ係数を導出し、
     前記予測画像を生成するために、導出された前記内部ノイズ除去フィルタ係数を用いたフィルタリングを前記処理対象ブロックに対して行い、
     導出された前記内部ノイズ除去フィルタ係数を前記符号化ビットストリームに挿入し、
     前記歪除去フィルタの導出の判別において、前記歪除去フィルタ係数を導出すべきと判別された際には、
     前記歪除去フィルタ係数の導出、前記歪除去フィルタ係数を用いたフィルタリング、および前記係数特定情報の前記符号化ビットストリームへの挿入を実行する
     請求項2記載の画像符号化方法。
    The image encoding method further includes:
    According to the feature of the block boundary, it is determined whether or not the distortion removal filter coefficient should be derived,
    In the determination of the derivation of the distortion removal filter, when it is determined that the distortion removal filter coefficient should not be derived,
    Deriving an internal noise removal filter coefficient for removing noise in the processing target block corresponding to the block boundary,
    In order to generate the predicted image, filtering is performed on the processing target block using the derived internal noise removal filter coefficient,
    Inserting the derived internal noise removal filter coefficients into the encoded bitstream;
    In the determination of the derivation of the distortion removal filter, when it is determined that the distortion removal filter coefficient should be derived,
    The image coding method according to claim 2, wherein derivation of the distortion removal filter coefficient, filtering using the distortion removal filter coefficient, and insertion of the coefficient specifying information into the encoded bitstream are executed.
  4.  前記歪除去フィルタの導出の判別において、前記歪除去フィルタ係数を導出すべきと判別された際には、
     前記歪除去フィルタ係数の導出では、前記ブロック境界の歪とともに、前記処理対象ブロック内のノイズを取り除くためのフィルタ係数を前記歪除去フィルタ係数として導出し、
     前記歪除去フィルタ係数を用いたフィルタリングでは、前記フィルタ係数を用いたフィルタリングを前記ブロック境界および前記処理対象ブロックに対して行う
     請求項3記載の画像符号化方法。
    In the determination of the derivation of the distortion removal filter, when it is determined that the distortion removal filter coefficient should be derived,
    In the derivation of the distortion removal filter coefficient, together with the distortion of the block boundary, a filter coefficient for removing noise in the processing target block is derived as the distortion removal filter coefficient,
    The image coding method according to claim 3, wherein in filtering using the distortion removal filter coefficient, filtering using the filter coefficient is performed on the block boundary and the processing target block.
  5.  前記歪除去フィルタ係数の導出では、前記ブロック境界の歪を抑える強度と、前記処理対象ブロック内のノイズを抑える強度との比率を、前記ブロック境界の特徴に応じて決定し、前記比率に応じて前記フィルタ係数を導出する
     請求項4記載の画像符号化方法。
    In the derivation of the distortion removal filter coefficient, a ratio between the strength that suppresses distortion at the block boundary and the strength that suppresses noise in the processing target block is determined according to the feature of the block boundary, and according to the ratio The image encoding method according to claim 4, wherein the filter coefficient is derived.
  6.  前記画像符号化方法は、さらに、
     画像領域内のブロック境界ごとに、当該ブロック境界の特徴に応じて、当該ブロック境界に対応するフィルタモードを、予め定められた複数のフィルタモードの中から選択し、
     前記歪除去フィルタ係数の導出では、フィルタモードごとに、前記画像領域に含まれる当該フィルタモードに対応する複数のブロック境界に対して、共通の前記フィルタ係数を導出する
     請求項4記載の画像符号化方法。
    The image encoding method further includes:
    For each block boundary in the image area, according to the feature of the block boundary, select a filter mode corresponding to the block boundary from a plurality of predetermined filter modes,
    5. The image encoding according to claim 4, wherein, in the derivation of the distortion removal filter coefficient, the common filter coefficient is derived for a plurality of block boundaries corresponding to the filter mode included in the image region for each filter mode. Method.
  7.  前記画像符号化方法は、さらに、
     フィルタモードごとに、前記画像領域に含まれる当該フィルタモードに対応するブロック境界の数をカウントし、
     カウントされた前記数が最も少ないフィルタモードを、前記数が2番目に少ないフィルタモードに変更することにより、前記数が最も少ないフィルタモードと、前記数が2番目に少ないフィルタモードとを結合し、
     前記歪除去フィルタ係数の導出では、結合された2つのフィルタモードに対応する複数のブロック境界に対して、共通の前記フィルタ係数を導出する
     請求項6記載の画像符号化方法。
    The image encoding method further includes:
    For each filter mode, count the number of block boundaries corresponding to the filter mode included in the image area,
    The filter mode with the smallest number is combined with the filter mode with the second smallest number by changing the filter mode with the smallest number to the filter mode with the second smallest number,
    The image coding method according to claim 6, wherein in the derivation of the distortion removal filter coefficient, the common filter coefficient is derived for a plurality of block boundaries corresponding to the two combined filter modes.
  8.  前記結合では、前記数が少ないフィルタモードと、前記数が2番目に少ないフィルタモードとを結合することを繰り返し行い、
     前記係数特定情報の前記符号化ビットストリームへの挿入では、さらに、繰り返して行われた結合の回数を前記符号化ビットストリームに挿入する
     請求項7記載の画像符号化方法。
    In the combination, repeatedly combining the filter mode with the smallest number and the filter mode with the second smallest number,
    The image coding method according to claim 7, wherein the insertion of the coefficient specifying information into the coded bitstream further inserts the number of times of repetitive combination into the coded bitstream.
  9.  前記画像符号化方法は、さらに、
     前記予め定められた複数のフィルタモードの中の何れか1つのフィルタモードを、前記複数のフィルタモードの中の他のフィルタモードに変更することにより、前記フィルタモードと前記他のフィルタモードとを結合し、
     前記歪除去フィルタ係数の導出では、結合された2つのフィルタモードに対応する複数のブロック境界に対して、共通の前記フィルタ係数を導出し、
     前記係数特定情報の前記符号化ビットストリームへの挿入では、さらに、結合された前記2つのフィルタモードを特定するための結合インデックスを前記符号化ビットストリームに挿入する
     請求項6記載の画像符号化方法。
    The image encoding method further includes:
    The filter mode and the other filter mode are combined by changing any one of the plurality of predetermined filter modes to another filter mode in the plurality of filter modes. And
    In the derivation of the distortion removal filter coefficient, the common filter coefficient is derived for a plurality of block boundaries corresponding to the combined two filter modes,
    The image coding method according to claim 6, wherein in the insertion of the coefficient specifying information into the encoded bitstream, a combined index for specifying the combined two filter modes is further inserted into the encoded bitstream. .
  10.  前記歪除去フィルタ係数を用いたフィルタリングでは、
     前記ブロック境界にある画素ごとに、画素間の画素値の差分に基づいて、当該画素に対してフィルタリングを行うべきか否かを判別し、
     フィルタリングを行うべきと判別された画素に対してフィルタリングを行う
     請求項2記載の画像符号化方法。
    In filtering using the distortion removal filter coefficient,
    For each pixel at the block boundary, based on the difference in pixel value between the pixels, determine whether to perform filtering on the pixel,
    The image encoding method according to claim 2, wherein filtering is performed on pixels that are determined to be filtered.
  11.  前記画像符号化方法は、さらに、
     導出された前記歪除去フィルタ係数を予測することによって前記歪除去フィルタ係数に対する予測値を導出し、
     前記係数特定情報の前記符号化ビットストリームへの挿入では、
     前記歪除去フィルタ係数と前記予測値との差分を、前記符号化ビットストリームに挿入する
     請求項2記載の画像符号化方法。
    The image encoding method further includes:
    Deriving a predicted value for the distortion removal filter coefficient by predicting the derived distortion removal filter coefficient;
    In inserting the coefficient specifying information into the encoded bitstream,
    The image coding method according to claim 2, wherein a difference between the distortion removal filter coefficient and the predicted value is inserted into the coded bitstream.
  12.  前記歪除去フィルタ係数の導出では、色成分およびフィルタリング方向ごとに、前記歪除去フィルタ係数を導出する
     請求項2記載の画像符号化方法。
    The image coding method according to claim 2, wherein, in deriving the distortion removal filter coefficient, the distortion removal filter coefficient is derived for each color component and filtering direction.
  13.  前記歪除去フィルタ係数の導出では、ウィナーフィルタのフィルタ係数の導出に用いられる演算式を用いて前記歪除去フィルタ係数を導出する
     請求項2記載の画像符号化方法。
    The image encoding method according to claim 2, wherein in the derivation of the distortion removal filter coefficient, the distortion removal filter coefficient is derived using an arithmetic expression used to derive a filter coefficient of a Wiener filter.
  14.  前記歪除去フィルタ係数の導出では、複数の歪除去フィルタ係数の候補の中から、前記歪除去フィルタ係数を選択することによって、当該歪除去フィルタ係数を導出し、
     前記係数特定情報の前記符号化ビットストリームへの挿入では、導出された前記歪除去フィルタ係数を示すインデックスを前記係数特定情報として前記符号化ビットストリームに挿入する
     請求項1記載の画像符号化方法。
    In deriving the distortion removal filter coefficient, the distortion removal filter coefficient is derived by selecting the distortion removal filter coefficient from among a plurality of distortion removal filter coefficient candidates,
    The image coding method according to claim 1, wherein, in inserting the coefficient specifying information into the encoded bitstream, an index indicating the derived distortion removal filter coefficient is inserted into the encoded bitstream as the coefficient specifying information.
  15.  符号化ビットストリームを復号する画像復号方法であって、
     前記符号化ビットストリームに含まれる符号化されたブロックを順次再構成し、
     再構成されたブロック間の境界であるブロック境界の特徴に応じて、前記ブロック境界の歪を取り除くための歪除去フィルタ係数を特定するための係数特定情報を、前記符号化ビットストリームから抽出し、
     抽出された前記係数特定情報によって特定される前記歪除去フィルタ係数を用いたフィルタリングを前記ブロック境界に対して行う
     画像復号方法。
    An image decoding method for decoding an encoded bitstream, comprising:
    Sequentially reconstructing the encoded blocks included in the encoded bitstream;
    Coefficient specifying information for specifying a distortion removal filter coefficient for removing distortion at the block boundary is extracted from the encoded bitstream according to the feature of the block boundary that is a boundary between the reconstructed blocks.
    An image decoding method in which filtering using the distortion removal filter coefficient specified by the extracted coefficient specifying information is performed on the block boundary.
  16.  前記係数特定情報の抽出では、前記歪除去フィルタ係数を前記係数特定情報として前記符号化ビットストリームから抽出する
     請求項15記載の画像復号方法。
    The image decoding method according to claim 15, wherein in the extraction of the coefficient specifying information, the distortion removal filter coefficient is extracted from the coded bitstream as the coefficient specifying information.
  17.  前記係数特定情報の抽出では、前記ブロック境界の歪とともに、前記ブロック境界に対応する処理対象ブロック内のノイズを取り除くためのフィルタ係数を前記歪除去フィルタ係数として抽出し、
     前記歪除去フィルタ係数を用いたフィルタリングでは、前記フィルタ係数を用いたフィルタリングを前記ブロック境界および前記処理対象ブロックに対して行う
     請求項16記載の画像復号方法。
    In the extraction of the coefficient specifying information, together with the distortion of the block boundary, a filter coefficient for removing noise in the processing target block corresponding to the block boundary is extracted as the distortion removal filter coefficient,
    The image decoding method according to claim 16, wherein in the filtering using the distortion removal filter coefficient, filtering using the filter coefficient is performed on the block boundary and the processing target block.
  18.  前記係数特定情報の抽出では、前記歪除去フィルタ係数を示すインデックスを前記係数特定情報として前記符号化ビットストリームから抽出し、
     前記歪除去フィルタ係数を用いたフィルタリングでは、抽出された前記インデックスによって示される前記歪除去フィルタ係数を、複数の歪除去フィルタ係数の候補の中から選択し、選択された前記歪除去フィルタ係数を用いたフィルタリングを前記ブロック境界に対して行う
     請求項15記載の画像復号方法。
    In the extraction of the coefficient specifying information, an index indicating the distortion removal filter coefficient is extracted from the coded bitstream as the coefficient specifying information,
    In the filtering using the distortion removal filter coefficient, the distortion removal filter coefficient indicated by the extracted index is selected from a plurality of distortion removal filter coefficient candidates, and the selected distortion removal filter coefficient is used. The image decoding method according to claim 15, wherein the received filtering is performed on the block boundary.
  19.  動画像を符号化する画像符号化装置であって、
     予測画像を用いて前記動画像をブロックごとに符号化することによって符号化ビットストリームを生成する符号化部と、
     符号化された前記ブロックを順次再構成する再構成部と、
     再構成されたブロック間の境界であるブロック境界の歪を取り除くための歪除去フィルタ係数を、前記ブロック境界の特徴に応じて導出するフィルタ係数導出部と、
     予測画像を生成するために、導出された前記歪除去フィルタ係数を用いたフィルタリングを前記ブロック境界に対して行うフィルタリング部と、
     導出された前記歪除去フィルタ係数を特定するための係数特定情報を前記符号化ビットストリームに挿入する挿入部と
     を備える画像符号化装置。
    An image encoding device for encoding a moving image,
    An encoding unit that generates an encoded bitstream by encoding the moving image for each block using a predicted image;
    A reconstruction unit that sequentially reconstructs the encoded blocks;
    A filter coefficient derivation unit for deriving a distortion removal filter coefficient for removing distortion at a block boundary, which is a boundary between reconstructed blocks, according to the feature of the block boundary;
    A filtering unit that performs filtering using the derived distortion removal filter coefficient on the block boundary to generate a predicted image;
    An image encoding apparatus comprising: an insertion unit that inserts coefficient specifying information for specifying the derived distortion removal filter coefficient into the encoded bitstream.
  20.  符号化ビットストリームを復号する画像復号装置であって、
     前記符号化ビットストリームに含まれる符号化されたブロックを順次再構成する再構成部と、
     再構成されたブロック間の境界であるブロック境界の特徴に応じて、前記ブロック境界の歪を取り除くための歪除去フィルタ係数を特定するための係数特定情報を、前記符号化ビットストリームから抽出する抽出部と、
     抽出された前記係数特定情報によって特定される前記歪除去フィルタ係数を用いたフィルタリングを前記ブロック境界に対して行うフィルタリング部と
     を備える画像復号装置。
    An image decoding device for decoding an encoded bitstream,
    A reconstructing unit for sequentially reconstructing encoded blocks included in the encoded bitstream;
    Extraction that extracts coefficient specifying information for specifying a distortion removing filter coefficient for removing distortion of the block boundary from the coded bitstream in accordance with a feature of the block boundary that is a boundary between reconstructed blocks And
    An image decoding device comprising: a filtering unit that performs filtering on the block boundary using the distortion removal filter coefficient specified by the extracted coefficient specifying information.
  21.  動画像を符号化するためのプログラムであって、
     予測画像を用いて前記動画像をブロックごとに符号化することによって符号化ビットストリームを生成し、
     符号化された前記ブロックを順次再構成し、
     再構成されたブロック間の境界であるブロック境界の歪を取り除くための歪除去フィルタ係数を、前記ブロック境界の特徴に応じて導出し、
     予測画像を生成するために、導出された前記歪除去フィルタ係数を用いたフィルタリングを前記ブロック境界に対して行い、
     導出された前記歪除去フィルタ係数を特定するための係数特定情報を前記符号化ビットストリームに挿入する
     ことをコンピュータに実行させるプログラム。
    A program for encoding a moving image,
    Generating an encoded bitstream by encoding the moving image block by block using a predicted image;
    Sequentially reconstructing the encoded blocks;
    Deriving a distortion removal filter coefficient for removing distortion at a block boundary, which is a boundary between reconstructed blocks, according to the feature of the block boundary;
    In order to generate a predicted image, filtering using the derived distortion removal filter coefficient is performed on the block boundary,
    A program that causes a computer to execute insertion of coefficient specifying information for specifying the derived distortion removal filter coefficient into the encoded bitstream.
  22.  符号化ビットストリームを復号するためのプログラムであって、
     前記符号化ビットストリームに含まれる符号化されたブロックを順次再構成し、
     再構成されたブロック間の境界であるブロック境界の特徴に応じて、前記ブロック境界の歪を取り除くための歪除去フィルタ係数を特定するための係数特定情報を、前記符号化ビットストリームから抽出し、
     抽出された前記係数特定情報によって特定される前記歪除去フィルタ係数を用いたフィルタリングを前記ブロック境界に対して行う
     ことをコンピュータに実行させるプログラム。
    A program for decoding an encoded bitstream,
    Sequentially reconstructing the encoded blocks included in the encoded bitstream;
    Coefficient specifying information for specifying a distortion removal filter coefficient for removing distortion at the block boundary is extracted from the encoded bitstream according to the feature of the block boundary that is a boundary between the reconstructed blocks.
    A program that causes a computer to execute filtering on the block boundary using the distortion removal filter coefficient specified by the extracted coefficient specifying information.
  23.  動画像を符号化する集積回路であって、
     予測画像を用いて前記動画像をブロックごとに符号化することによって符号化ビットストリームを生成する符号化部と、
     符号化された前記ブロックを順次再構成する再構成部と、
     再構成されたブロック間の境界であるブロック境界の歪を取り除くための歪除去フィルタ係数を、前記ブロック境界の特徴に応じて導出するフィルタ係数導出部と、
     予測画像を生成するために、導出された前記歪除去フィルタ係数を用いたフィルタリングを前記ブロック境界に対して行うフィルタリング部と、
     導出された前記歪除去フィルタ係数を特定するための係数特定情報を前記符号化ビットストリームに挿入する挿入部と
     を備える集積回路。
    An integrated circuit for encoding a moving image,
    An encoding unit that generates an encoded bitstream by encoding the moving image for each block using a predicted image;
    A reconstruction unit that sequentially reconstructs the encoded blocks;
    A filter coefficient derivation unit for deriving a distortion removal filter coefficient for removing distortion at a block boundary, which is a boundary between reconstructed blocks, according to the feature of the block boundary;
    A filtering unit that performs filtering using the derived distortion removal filter coefficient on the block boundary to generate a predicted image;
    An integrated circuit comprising: an insertion unit that inserts coefficient specifying information for specifying the derived distortion removal filter coefficient into the encoded bitstream.
  24.  符号化ビットストリームを復号する集積回路であって、
     前記符号化ビットストリームに含まれる符号化されたブロックを順次再構成する再構成部と、
     再構成されたブロック間の境界であるブロック境界の特徴に応じて、前記ブロック境界の歪を取り除くための歪除去フィルタ係数を特定するための係数特定情報を、前記符号化ビットストリームから抽出する抽出部と、
     抽出された前記係数特定情報によって特定される前記歪除去フィルタ係数を用いたフィルタリングを前記ブロック境界に対して行うフィルタリング部と
     を備える集積回路。
    An integrated circuit for decoding an encoded bitstream comprising:
    A reconstructing unit for sequentially reconstructing encoded blocks included in the encoded bitstream;
    Extraction that extracts coefficient specifying information for specifying a distortion removing filter coefficient for removing distortion of the block boundary from the coded bitstream in accordance with a feature of the block boundary that is a boundary between reconstructed blocks And
    An integrated circuit comprising: a filtering unit that performs filtering on the block boundary using the distortion removal filter coefficient specified by the extracted coefficient specifying information.
PCT/JP2011/000085 2010-01-21 2011-01-12 Image encoding method, image decoding method, device therefor, program, and integrated circuit WO2011089865A1 (en)

Applications Claiming Priority (2)

Application Number Priority Date Filing Date Title
US29698910P 2010-01-21 2010-01-21
US61/296,989 2010-01-21

Publications (1)

Publication Number Publication Date
WO2011089865A1 true WO2011089865A1 (en) 2011-07-28

Family

ID=44306668

Family Applications (1)

Application Number Title Priority Date Filing Date
PCT/JP2011/000085 WO2011089865A1 (en) 2010-01-21 2011-01-12 Image encoding method, image decoding method, device therefor, program, and integrated circuit

Country Status (1)

Country Link
WO (1) WO2011089865A1 (en)

Cited By (1)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
WO2013030902A1 (en) * 2011-08-26 2013-03-07 株式会社 東芝 Moving image encoding method, moving image decoding method, moving image encoding apparatus and moving image decoding apparatus

Citations (4)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
WO2008075247A1 (en) * 2006-12-18 2008-06-26 Koninklijke Philips Electronics N.V. Image compression and decompression
WO2008084745A1 (en) * 2007-01-09 2008-07-17 Panasonic Corporation Image coding apparatus and image decoding apparatus
WO2009110559A1 (en) * 2008-03-07 2009-09-11 株式会社 東芝 Dynamic image encoding/decoding device
WO2010001999A1 (en) * 2008-07-04 2010-01-07 株式会社 東芝 Dynamic image encoding/decoding method and device

Patent Citations (8)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
WO2008075247A1 (en) * 2006-12-18 2008-06-26 Koninklijke Philips Electronics N.V. Image compression and decompression
KR20090100402A (en) * 2006-12-18 2009-09-23 코닌클리케 필립스 일렉트로닉스 엔.브이. Image compression and decompression
CN101563926A (en) * 2006-12-18 2009-10-21 皇家飞利浦电子股份有限公司 Image compression and decompression
US20100027686A1 (en) * 2006-12-18 2010-02-04 Koninklijke Philips Electronics N.V. Image compression and decompression
JP2010514246A (en) * 2006-12-18 2010-04-30 コーニンクレッカ フィリップス エレクトロニクス エヌ ヴィ Image compression and decompression
WO2008084745A1 (en) * 2007-01-09 2008-07-17 Panasonic Corporation Image coding apparatus and image decoding apparatus
WO2009110559A1 (en) * 2008-03-07 2009-09-11 株式会社 東芝 Dynamic image encoding/decoding device
WO2010001999A1 (en) * 2008-07-04 2010-01-07 株式会社 東芝 Dynamic image encoding/decoding method and device

Cited By (1)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
WO2013030902A1 (en) * 2011-08-26 2013-03-07 株式会社 東芝 Moving image encoding method, moving image decoding method, moving image encoding apparatus and moving image decoding apparatus

Similar Documents

Publication Publication Date Title
JP5588438B2 (en) Image encoding method and image encoding apparatus
JP5485983B2 (en) Moving picture coding method, moving picture decoding method, moving picture coding apparatus, and moving picture decoding apparatus
JP5574345B2 (en) Encoding method, error detection method, decoding method, encoding device, error detection device, and decoding device
WO2010061607A1 (en) Method for decoding moving images, method for encoding moving images, moving image decoder, moving image encoder, program and integrated circuit
KR101746251B1 (en) Filter positioning and selection
JP5479470B2 (en) Moving picture coding method, moving picture coding apparatus, program, and integrated circuit
WO2010026770A1 (en) Image coding method, image decoding method, image coding device, image decoding device, system, program, and integrated circuit
WO2010087157A1 (en) Image coding method and image decoding method
WO2010050156A1 (en) Image encoding method, image decoding method, image encoding device, image decoding device, integrated circuit, and program
JP5659160B2 (en) Moving picture decoding apparatus, moving picture encoding apparatus, moving picture decoding circuit, and moving picture decoding method
EP2464018A1 (en) Encoding method, decoding method, encoding device and decoding device
EP2464017A1 (en) Encoding method, decoding method, encoding device and decoding device
JP2013505647A (en) Image encoding apparatus, image decoding apparatus, image encoding method, and image decoding method
KR101863397B1 (en) Efficient decisions for deblocking
KR20120046726A (en) Encoding method, decoding method, encoding device and decoding device
KR20120089690A (en) Decoding method, decoder apparatus, encoding method, and encoder apparatus
KR20120099418A (en) Image decoding method, image encoding method, image decoding device, image encoding device, programs, and integrated circuits
JP5499035B2 (en) Image coding method, image coding apparatus, program, and integrated circuit
WO2011052216A1 (en) Image encoding method, image decoding method, image encoding device and image decoding device
WO2010131422A1 (en) Image decoding apparatus, integrated circuit, image decoding method, and image decoding system
WO2011089865A1 (en) Image encoding method, image decoding method, device therefor, program, and integrated circuit
WO2011128055A1 (en) Efficient storing of picture representation using adaptive non-uniform quantizer

Legal Events

Date Code Title Description
121 Ep: the epo has been informed by wipo that ep was designated in this application

Ref document number: 11734473

Country of ref document: EP

Kind code of ref document: A1

NENP Non-entry into the national phase

Ref country code: DE

122 Ep: pct application non-entry in european phase

Ref document number: 11734473

Country of ref document: EP

Kind code of ref document: A1

NENP Non-entry into the national phase

Ref country code: JP