WO2012035640A1 - Moving picture encoding method and moving picture decoding method - Google Patents

Moving picture encoding method and moving picture decoding method Download PDF

Info

Publication number
WO2012035640A1
WO2012035640A1 PCT/JP2010/066102 JP2010066102W WO2012035640A1 WO 2012035640 A1 WO2012035640 A1 WO 2012035640A1 JP 2010066102 W JP2010066102 W JP 2010066102W WO 2012035640 A1 WO2012035640 A1 WO 2012035640A1
Authority
WO
WIPO (PCT)
Prior art keywords
prediction
image signal
unit
intra
pixel
Prior art date
Application number
PCT/JP2010/066102
Other languages
French (fr)
Japanese (ja)
Inventor
昭行 谷沢
太一郎 塩寺
Original Assignee
株式会社 東芝
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Application filed by 株式会社 東芝 filed Critical 株式会社 東芝
Priority to PCT/JP2010/066102 priority Critical patent/WO2012035640A1/en
Publication of WO2012035640A1 publication Critical patent/WO2012035640A1/en

Links

Images

Classifications

    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04NPICTORIAL COMMUNICATION, e.g. TELEVISION
    • H04N19/00Methods or arrangements for coding, decoding, compressing or decompressing digital video signals
    • H04N19/10Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using adaptive coding
    • H04N19/102Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using adaptive coding characterised by the element, parameter or selection affected or controlled by the adaptive coding
    • H04N19/103Selection of coding mode or of prediction mode
    • H04N19/11Selection of coding mode or of prediction mode among a plurality of spatial predictive coding modes
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04NPICTORIAL COMMUNICATION, e.g. TELEVISION
    • H04N19/00Methods or arrangements for coding, decoding, compressing or decompressing digital video signals
    • H04N19/10Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using adaptive coding
    • H04N19/169Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using adaptive coding characterised by the coding unit, i.e. the structural portion or semantic portion of the video signal being the object or the subject of the adaptive coding
    • H04N19/17Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using adaptive coding characterised by the coding unit, i.e. the structural portion or semantic portion of the video signal being the object or the subject of the adaptive coding the unit being an image region, e.g. an object
    • H04N19/176Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using adaptive coding characterised by the coding unit, i.e. the structural portion or semantic portion of the video signal being the object or the subject of the adaptive coding the unit being an image region, e.g. an object the region being a block, e.g. a macroblock
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04NPICTORIAL COMMUNICATION, e.g. TELEVISION
    • H04N19/00Methods or arrangements for coding, decoding, compressing or decompressing digital video signals
    • H04N19/50Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using predictive coding
    • H04N19/593Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using predictive coding involving spatial prediction techniques
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04NPICTORIAL COMMUNICATION, e.g. TELEVISION
    • H04N19/00Methods or arrangements for coding, decoding, compressing or decompressing digital video signals
    • H04N19/70Methods or arrangements for coding, decoding, compressing or decompressing digital video signals characterised by syntax aspects related to video coding, e.g. related to compression standards
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04NPICTORIAL COMMUNICATION, e.g. TELEVISION
    • H04N19/00Methods or arrangements for coding, decoding, compressing or decompressing digital video signals
    • H04N19/90Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using coding techniques not provided for in groups H04N19/10-H04N19/85, e.g. fractals
    • H04N19/96Tree coding, e.g. quad-tree coding

Definitions

  • Embodiments of the present invention relate to an intra-screen prediction method, a video encoding method, and a video decoding method in video encoding and decoding.
  • H.264 achieves higher prediction efficiency than in-screen prediction in ISO / IEC MPEG-1, 2 and 4 (hereinafter referred to as intra prediction) by incorporating direction prediction in the spatial region (pixel region).
  • intra prediction in-screen prediction in ISO / IEC MPEG-1, 2 and 4
  • JCT-VC Joint Collaborative Team on Video Coding
  • Non-Patent Document 1 since a prediction value is generated at an individual prediction angle for each of a plurality of types of prediction modes and copied in the prediction direction, a texture having a luminance gradient that smoothly changes within a pixel block. Such a video or a video with gradation cannot be predicted efficiently, and the prediction error may increase.
  • an object of the present embodiment is to provide a moving image encoding device and a moving image decoding device including a prediction image generating device capable of improving encoding efficiency.
  • the moving image encoding method divides an input image signal into pixel blocks represented by hierarchical depths according to quadtree division, and generates prediction error signals for these divided pixel blocks. Then, in the moving picture encoding method for encoding the transform coefficient, a step of setting a first prediction direction from a plurality of prediction direction sets to create a first prediction image signal, and the plurality of prediction direction sets A step of creating a second prediction image signal by setting a second prediction direction different from the first prediction direction, and a pixel to be predicted and reference for each of the first and second prediction directions Deriving a relative distance from the pixel, deriving a difference value of the relative distance, deriving a predetermined weight component according to the difference value, and according to the weight component, a first unit Direction A step of weighted averaging the prediction image and the second unidirectional intra prediction image to generate a third prediction image signal, a step of generating a prediction error signal from the third prediction image signal, and encoding the prediction error signal And
  • the moving picture decoding method divides an input image signal into pixel blocks expressed by hierarchical depth according to quadtree division, and performs a decoding process on the divided pixel blocks.
  • the measurement image and the second unidirectional intra-prediction image and the weighted average comprises the steps of generating a third predictive image signal, and generating a decoded image signal from said third prediction image signal.
  • FIG. 1 is a block diagram illustrating a moving image encoding apparatus according to a first embodiment.
  • Explanatory drawing of the prediction encoding order of a pixel block Explanatory drawing of an example of pixel block size.
  • Explanatory drawing of another example of pixel block size Explanatory drawing of another example of pixel block size.
  • Explanatory drawing of an example of the pixel block in a coding tree unit Explanatory drawing of another example of the pixel block in a coding tree unit.
  • Explanatory drawing of another example of the pixel block in a coding tree unit Explanatory drawing of another example of the pixel block in a coding tree unit.
  • (A) is explanatory drawing of intra prediction mode
  • (b) is explanatory drawing of the reference pixel and prediction pixel of intra prediction mode
  • (c) is explanatory drawing of the horizontal prediction mode of intra prediction mode
  • ( d) Explanatory drawing of the orthogonal lower right prediction mode of intra prediction mode.
  • the block diagram which illustrates the intra prediction part concerning a 1st embodiment. Explanatory drawing of the number of unidirectional intra predictions and the number of bidirectional intra predictions which concern on 1st Embodiment. Explanatory drawing which illustrates the prediction direction which concerns on 1st Embodiment.
  • the table figure which illustrates the relationship of prediction mode, prediction type, bidirectional
  • the table figure which illustrates the relationship of prediction mode, prediction type, bidirectional
  • the table figure which illustrates the relationship of prediction mode, prediction type, bidirectional
  • the table figure which illustrates a response
  • the table figure which illustrates a response
  • the block diagram which shows an example of the calculation method of a city area distance based on 1st Embodiment.
  • the block diagram which shows another example of the calculation method of the city area distance based on 1st Embodiment.
  • the block diagram which shows another example of the calculation method of the city area distance based on 1st Embodiment.
  • the block diagram which shows another example of the calculation method of the city area distance based on 1st Embodiment.
  • the table which illustrates the relationship between the prediction mode and the distance of a prediction pixel position based on 1st Embodiment.
  • the table which illustrates the mapping of prediction mode and a distance table based on 1st Embodiment.
  • the table which illustrates the relationship between relative distance and a weight component based on 1st Embodiment. 6 is another table illustrating the relationship between the relative distance and the weight component according to the first embodiment.
  • Explanatory drawing of a syntax structure Explanatory drawing of a slice header syntax.
  • Explanatory drawing which shows an example of a prediction unit syntax.
  • Explanatory drawing which shows another example of a prediction unit syntax.
  • Explanatory drawing which shows another example of a prediction unit syntax.
  • Explanatory drawing which shows another example of a prediction unit syntax.
  • the flowchart which shows an example of the calculation method of intraPredModeN.
  • the table which shows the relationship at the time of predicting prediction mode The table which shows another example of the relationship at the time of predicting prediction mode.
  • Explanatory drawing which shows an example of the prediction unit syntax in the 1st modification based on 1st Embodiment.
  • Explanatory drawing which shows another example of the prediction unit syntax in the 1st modification based on 1st Embodiment.
  • Explanatory drawing which shows an example of the predicted value generation method of a pixel level.
  • the block diagram which shows an example of the composite intra estimated image generation part based on 1st Embodiment.
  • Explanatory drawing which shows an example according to Prediction unit syntax based on 1st Embodiment.
  • Explanatory drawing which shows another example of the prediction unit syntax based on 1st Embodiment.
  • the block diagram which illustrates the moving picture coding device concerning a 2nd embodiment.
  • the block diagram which illustrates the orthogonal transformation part based on 2nd Embodiment.
  • the block diagram which illustrates the inverse orthogonal transformation part based on 2nd Embodiment.
  • the table which shows the relationship between prediction mode and a conversion index based on 2nd Embodiment.
  • the block diagram which illustrates the coefficient order control part concerning a 2nd embodiment.
  • the block diagram which illustrates another coefficient order control part concerning a 2nd embodiment.
  • Explanatory drawing which shows an example of the transform unit syntax based on 2nd Embodiment.
  • the block diagram which shows another example of the orthogonal transformation part based on 3rd Embodiment.
  • the block diagram which shows an example of the inverse orthogonal transformation part based on 3rd Embodiment.
  • Explanatory drawing which shows an example of the transform unit syntax based on 3rd Embodiment.
  • Explanatory drawing which shows an example of the unidirectional intra prediction mode, prediction type, and prediction angle parameter
  • the block diagram which shows an example of the moving image decoding apparatus based on 4th Embodiment.
  • the block diagram which shows an example of the moving image decoding apparatus based on 5th Embodiment.
  • the block diagram which illustrates the coefficient order restoration part concerning a 5th embodiment.
  • the block diagram which shows another example of the coefficient order decompression
  • the first embodiment relates to an image encoding device.
  • a moving picture decoding apparatus corresponding to the picture encoding apparatus according to the present embodiment will be described in a fourth embodiment.
  • This image encoding device can be realized by hardware such as an LSI (Large-Scale Integration) chip, a DSP (Digital Signal Processor), or an FPGA (Field Programmable Gate Array).
  • the image encoding apparatus can also be realized by causing a computer to execute an image encoding program.
  • the image coding apparatus includes a subtraction unit 101, an orthogonal transformation unit 102, a quantization unit 103, an inverse quantization unit 104, an inverse orthogonal transformation unit 105, an addition unit 106, and a loop filter. 107, a reference image memory 108, an intra prediction unit 109, an inter prediction unit 110, a prediction selection switch 111, a prediction selection unit 112, an entropy encoding unit 113, an output buffer 114, and an encoding control unit 115.
  • the image coding apparatus in FIG. 1 divides each frame or each field constituting the input image signal 116 into a plurality of pixel blocks, performs predictive coding on the divided pixel blocks, and generates coded data 127. Output.
  • pixel blocks are predictively encoded from the upper left to the lower right as shown in FIG. 2A.
  • the encoded pixel block p is located on the left side and the upper side of the encoding target pixel block c in the encoding processing target frame f.
  • the pixel block refers to a unit for processing an image such as an M ⁇ N size block (N and M are natural numbers), a coding tree unit, a macro block, a sub block, and one pixel.
  • N and M are natural numbers
  • the pixel block is basically used in the meaning of the coding tree unit.
  • the pixel block can be interpreted in the above-described meaning by appropriately replacing the description.
  • the coding tree unit is typically a 16 ⁇ 16 pixel block shown in FIG. 2B, for example, but may be a 32 ⁇ 32 pixel block shown in FIG. 2C or a 64 ⁇ 64 pixel block shown in FIG. 3D. It may be an 8 ⁇ 8 pixel block (not shown) or a 4 ⁇ 4 pixel block.
  • the coding tree unit need not necessarily be square.
  • the encoding target block or coding tree unit of the input image signal 116 may be referred to as a “prediction target block”.
  • the coding unit is not limited to a pixel block such as a coding tree unit, and a frame, a field, a slice, or a combination thereof can be used.
  • FIG. 3A to 3D are diagrams showing specific examples of coding tree units.
  • N represents the size of the reference coding tree unit.
  • the coding tree unit has a quadtree structure, and when divided, the four pixel blocks are indexed in the Z-scan order.
  • FIG. 3B shows an example in which the 64 ⁇ 64 pixel block in FIG. 3A is divided into quadtrees.
  • the numbers shown in the figure represent the Z scan order.
  • the depth of division is defined by Depth.
  • a unit having the largest coding tree unit is called a large coding tree unit, and an input image signal is encoded in the order of raster scanning in this unit.
  • the image encoding apparatus in FIG. 1 performs intra prediction (also referred to as intra-frame prediction, intra-frame prediction, etc.) or inter prediction (inter-screen prediction) for a pixel block based on the encoding parameter input from the encoding control unit 115. Prediction, inter-frame prediction, motion compensation prediction, etc.) is performed to generate the predicted image signal 126.
  • This image coding apparatus orthogonally transforms and quantizes the prediction error signal 117 between the pixel block (input image signal 116) and the predicted image signal 126, performs entropy coding, generates coded data 127, and outputs it. To do.
  • the image encoding device in FIG. 1 performs encoding by selectively applying a plurality of prediction modes having different block sizes and generation methods of the predicted image signal 126.
  • the generation method of the predicted image signal 126 can be broadly divided into two types: intra prediction in which prediction is performed within the encoding target frame and inter prediction in which prediction is performed using one or a plurality of reference frames that are temporally different. is there.
  • the subtraction unit 101 subtracts the corresponding prediction image signal 126 from the encoding target block of the input image signal 116 to obtain a prediction error signal 117.
  • the subtraction unit 101 inputs the prediction error signal 117 to the orthogonal transformation unit 102.
  • the orthogonal transform unit 102 performs an orthogonal transform such as discrete cosine transform (DCT) on the prediction error signal 117 from the subtraction unit 101 to obtain a transform coefficient 118.
  • the orthogonal transform unit 102 inputs the transform coefficient 118 to the quantization unit 103.
  • the quantization unit 103 performs quantization on the transform coefficient from the orthogonal transform unit 102 to obtain a quantized transform coefficient 119. Specifically, the quantization unit 103 performs quantization according to quantization information such as a quantization parameter and a quantization matrix specified by the encoding control unit 115.
  • the quantization parameter indicates the fineness of quantization.
  • the quantization matrix is used for weighting the fineness of quantization for each component of the transform coefficient.
  • the quantization unit 103 inputs the quantized transform coefficient 119 to the entropy encoding unit 113 and the inverse quantization unit 104.
  • the entropy encoding unit 113 performs various encoding parameters such as the quantized transform coefficient 119 from the quantization unit 103, the prediction information 125 from the prediction selection unit 112, and the quantization information specified by the encoding control unit 115.
  • Entropy encoding (for example, Huffman encoding, arithmetic encoding, etc.) is performed to generate encoded data.
  • the encoding parameter is a parameter necessary for decoding, such as prediction information 125, information on transform coefficients, information on quantization, and the like.
  • the encoding control unit 115 has an internal memory (not shown), the encoding parameter is held in this memory, and the encoding parameter of an already encoded pixel block adjacent when encoding the prediction target block is stored. It is good also as a structure to use. For example, H.M. In the H.264 intra prediction, the prediction value of the prediction mode of the prediction target block can be derived from the prediction mode information of the encoded adjacent block.
  • the encoded data generated by the entropy encoding unit 113 is temporarily accumulated, for example, in the output buffer 114 through multiplexing, and is output as encoded data 127 according to an appropriate output timing managed by the encoding control unit 115. .
  • the encoded data 127 is output to, for example, a storage system (storage medium) or a transmission system (communication line) not shown.
  • the inverse quantization unit 104 performs inverse quantization on the quantized transform coefficient 119 from the quantization unit 103 to obtain a restored transform coefficient 120. Specifically, the inverse quantization unit 104 performs inverse quantization according to the quantization information used in the quantization unit 103. The quantization information used in the quantization unit 103 is loaded from the internal memory of the encoding control unit 115. The inverse quantization unit 104 inputs the restored transform coefficient 120 to the inverse orthogonal transform unit 105.
  • the inverse orthogonal transform unit 105 performs an inverse orthogonal transform corresponding to the orthogonal transform performed in the orthogonal transform unit 102 such as an inverse discrete cosine transform on the restored transform coefficient 120 from the inverse quantization unit 104, A restoration prediction error signal 121 is obtained.
  • the inverse orthogonal transform unit 105 inputs the restored prediction error signal 121 to the addition unit 106.
  • the addition unit 106 adds the restored prediction error signal 121 and the corresponding prediction image signal 126 to generate a local decoded image signal 122.
  • the decoded image signal 122 is input to the loop filter 107.
  • the loop filter 107 performs a deblocking filter, a Wiener filter, or the like on the input decoded image signal 122 to generate a filtered image signal 123.
  • the generated filtered image signal 123 is input to the reference image memory 108.
  • the reference image memory 108 stores the filtered image signal 123 after local decoding in the memory, and when the predicted image is generated as necessary by the intra prediction unit 109 and the inter prediction unit 110, the reference image signal 124 is used. Referenced each time.
  • the intra prediction unit 109 performs intra prediction using the reference image signal 124 stored in the reference image memory 108.
  • H.M. In H.264, an intra prediction image is obtained by performing pixel interpolation (copying or copying after interpolation) along a prediction direction such as a vertical direction or a horizontal direction using an encoded reference pixel value adjacent to a prediction target block. Generate. FIG. The prediction direction of intra prediction in H.264 is shown. Further, in FIG. 2 shows an arrangement relationship between reference pixels and encoding target pixels in H.264.
  • FIG. 5C illustrates a predicted image generation method in mode 1 (horizontal prediction)
  • FIG. 5D illustrates a predicted image generation method in mode 4 (diagonal lower right prediction).
  • H.264 In non-patent literature, H. The prediction direction of H.264 is further expanded to 34 directions to increase the number of prediction modes. A predicted pixel value is created by performing linear interpolation with 32-pixel accuracy in accordance with the predicted angle, and is copied in the predicted direction. Details of the intra prediction unit 109 used in the present embodiment of the present invention will be described later.
  • the inter prediction unit 110 performs inter prediction using the reference image signal 124 stored in the reference image memory 108. Specifically, the inter prediction unit 110 performs block matching processing between the prediction target block and the reference image signal 124 to derive a motion shift amount (motion vector). The inter prediction unit 110 performs an interpolation process (motion compensation) based on the motion vector to generate an inter prediction image. H. With H.264, interpolation processing up to 1/4 pixel accuracy is possible.
  • the derived motion vector is entropy encoded as part of the prediction information 125.
  • the prediction selection switch 111 selects the output terminal of the intra prediction unit 109 or the output terminal of the inter prediction unit 110 according to the prediction information 125 from the prediction selection unit 112, and subtracts the intra prediction image or the inter prediction image as the prediction image signal 126. 101 and the adder 106.
  • the prediction selection switch 111 connects a switch to the output terminal from the intra prediction unit 109.
  • the prediction selection switch 111 connects a switch to the output terminal from the inter prediction unit 110.
  • the prediction selection unit 112 has a function of setting the prediction information 125 according to the prediction mode controlled by the encoding control unit 115. As described above, intra prediction or inter prediction can be selected for generating the predicted image signal 126, but a plurality of modes can be further selected for each of intra prediction and inter prediction.
  • the encoding control unit 115 determines one of a plurality of prediction modes of intra prediction and inter prediction as the optimal prediction mode, and the prediction selection unit 112 sets the prediction information 125 according to the determined optimal prediction mode. .
  • prediction mode information is designated by the intra prediction unit 109 from the encoding control unit 115, and the intra prediction unit 109 generates a predicted image signal 126 according to the prediction mode information.
  • the encoding control unit 115 may specify a plurality of prediction mode information in order from the smallest prediction mode number, or may specify a plurality of prediction mode information in order from the largest.
  • the encoding control unit 115 may limit the prediction mode according to the characteristics of the input image.
  • the encoding control unit 115 does not necessarily specify all prediction modes, and may specify at least one prediction mode information for the encoding target block.
  • the encoding control unit 115 determines an optimal prediction mode using a cost function represented by the following mathematical formula (1).
  • Equation (1) (hereinafter referred to as simple encoding cost), OH indicates a code amount relating to prediction information 125 (for example, motion vector information and prediction block size information), and SAD is a prediction target block, a prediction image signal 126, and The difference absolute value sum (ie, the cumulative sum of the absolute values of the prediction error signal 117) is shown. Further, ⁇ represents a Lagrange undetermined multiplier determined based on the value of quantization information (quantization parameter), and K represents an encoding cost. When Expression (1) is used, the prediction mode that minimizes the coding cost K is determined as the optimum prediction mode from the viewpoint of the generated code amount and the prediction error. As a modification of Equation (1), the encoding cost may be estimated from OH alone or SAD alone, or the encoding cost may be estimated using a value obtained by subjecting SAD to Hadamard transform or an approximation thereof.
  • the encoding control unit 115 determines an optimal prediction mode using a cost function expressed by the following formula (2).
  • Equation (2) D represents a square error sum (ie, encoding distortion) between the prediction target block and the local decoded image, and R represents a prediction between the prediction target block and the prediction image signal 126 in the prediction mode.
  • An error amount indicates a code amount estimated by provisional encoding
  • J indicates an encoding cost.
  • provisional encoding processing and local decoding processing are required for each prediction mode, so that the circuit scale or the amount of calculation increases. .
  • the encoding cost J is derived based on more accurate encoding distortion and code amount, it is easy to determine the optimal prediction mode with high accuracy and maintain high encoding efficiency.
  • the encoding cost may be estimated from only R or D, or the encoding cost may be estimated using an approximate value of R or D. These costs may be used hierarchically.
  • the encoding control unit 115 performs determination using Expression (1) or Expression (2) based on information obtained in advance regarding the prediction target block (prediction mode of surrounding pixel blocks, image analysis result, and the like). The number of prediction mode candidates may be narrowed down in advance.
  • the number of prediction mode candidates can be further reduced while maintaining coding performance by performing two-stage mode determination combining Formula (1) and Formula (2). It becomes possible.
  • the simple encoding cost represented by the formula (1) does not require a local decoding process, and can be calculated at high speed.
  • H.264 is used. Since the number of prediction modes is large even when compared with H.264, mode determination using the detailed coding cost is not realistic. Therefore, as a first step, mode determination using the simple coding cost is performed on the prediction modes available in the pixel block, and prediction mode candidates are derived.
  • the number of prediction mode candidates is changed using the property that the correlation between the simple coding cost and the detailed coding cost increases as the value of the quantization parameter that determines the roughness of quantization increases.
  • FIG. 4 shows the number of prediction mode candidates selected in the first step.
  • PuSize is an index indicating the size of a pixel block (sometimes referred to as a prediction unit) that performs prediction described later.
  • QP represents a quantization parameter, and the number of prediction mode candidates changes depending on the quotient divided by 5. Since the detailed coding cost only has to be derived for the number of prediction mode candidates narrowed down in this way, the number of local decoding processes can be greatly reduced.
  • the encoding control unit 115 controls each element of the image encoding device in FIG. Specifically, the encoding control unit 115 performs various controls for the encoding process including the above-described operation.
  • the intra prediction unit 109 illustrated in FIG. 6 includes a unidirectional intra predicted image generation unit 601, a bidirectional intra predicted image generation unit 602, a prediction mode information setting unit 603, and a selection switch 604.
  • the reference image signal 124 is input from the reference image memory 108 to the unidirectional intra predicted image generation unit 601 and the bidirectional intra predicted image generation unit 602.
  • the prediction mode information setting unit 603 determines the prediction mode generated by the unidirectional intra prediction image generation unit 601 or the bidirectional intra prediction image generation unit 602. Set and output prediction mode 605.
  • the selection switch 604 has a function of switching the output ends of the respective intra predicted image generation units according to the prediction mode 605. If the input prediction mode 605 is the unidirectional intra prediction mode, the output terminal of the unidirectional intra prediction image generation unit 601 is connected to the switch, and if the prediction mode 605 is the bidirectional intra prediction mode, the bidirectional intra prediction is performed. The output terminal of the image generation unit 602 is connected. On the other hand, each of the intra predicted image generation units 601 and 602 generates the predicted image signal 126 according to the prediction mode 605. The generated prediction image signal 126 (also referred to as a fifth prediction image signal) is output from the intra prediction unit 109. The output signal of the unidirectional intra predicted image generation unit 601 is also called a fourth predicted image signal, and the output signal of the bidirectional intra predicted image generation unit 602 is also called a third predicted image signal.
  • FIG. 7 shows the number of prediction modes according to the block size according to the present embodiment of the present invention.
  • PuSize indicates the pixel block (prediction unit) size to be predicted, and seven types of sizes from PU_2x2 to PU_128x128 are defined.
  • IntraUniModeNum represents the number of prediction modes for unidirectional intra prediction
  • IntraBiModeNum represents the number of prediction modes for bidirectional intra prediction.
  • Number of modes is the total number of prediction modes for each pixel block (prediction unit) size.
  • FIG. 9 shows the relationship between the prediction mode and the prediction method when PuSize is PU_8x8, PU_16x16, and PU_32x32.
  • FIG. 10 shows a case where PuSize is PU_4 ⁇ 4
  • FIG. 11 shows a case where PU_64 ⁇ 64 or PU_128 ⁇ 128.
  • IntraPredMode indicates a prediction mode number
  • IntraBipredFlag is a flag indicating whether or not bidirectional intra prediction. When the flag is 1, it indicates that the prediction mode is the bidirectional intra prediction mode. When the flag is 0, it indicates that the prediction mode is a unidirectional intra prediction mode.
  • IntraPredTypeLX indicates the prediction type of intra prediction.
  • Intra_Vertical means that the vertical direction is the reference for prediction
  • Intra_Horizontal means that the horizontal direction is the reference for prediction. Note that 0 or 1 is applied to X in IntraPredTypeLX.
  • IntraPredTypeL0 indicates the first prediction mode of unidirectional intra prediction or bidirectional intra prediction.
  • IntraPredTypeL1 indicates the second prediction mode of bidirectional intra prediction.
  • IntraPred AngleID is an index indicating an index of a prediction angle. The prediction angle actually used in the generation of the predicted value is shown in FIG.
  • puPartIdx represents an index of the prediction unit that is divided in the quadtree division described with reference to FIG. 3B.
  • IntraPredMode 4
  • IntraPredTypeL0 Intra_Vertical
  • the prediction mode information setting unit 603 converts the above-described prediction information corresponding to the designated prediction mode 605 to the unidirectional intra prediction image generation unit 601 and the bidirectional intra prediction image generation unit 602 under the control of the encoding control unit 115. And the prediction mode 605 is output to the selection switch 604.
  • the unidirectional intra predicted image generation unit 601 has a function of generating the predicted image signal 126 for a plurality of prediction directions shown in FIG. In FIG. 8, there are 33 different prediction directions for the vertical and horizontal coordinates indicated by the bold lines. H.
  • the direction of a typical prediction angle indicated by H.264 is indicated by an arrow.
  • 33 kinds of prediction directions are prepared in a direction in which a line is drawn from the origin to a mark indicated by a diamond.
  • IntraPredMode 4
  • IntraPredAngleIDL0 4
  • An arrow indicated by a dotted line in FIG. 8 indicates a prediction mode whose prediction type is Intra_Vertical
  • an arrow indicated by a solid line indicates a prediction mode whose prediction type is Intra_Horizontal.
  • FIG. 12 shows the relationship between IntraPredAngleIDLX and intraPredAngle used for predictive image value generation.
  • intraPredAngle indicates a prediction angle that is actually used when a predicted value is generated.
  • a prediction value generation method is expressed by Expression (3).
  • BLK_SIZE indicates the size of the pixel block (prediction unit)
  • ref [] indicates an array in which reference image signals are stored.
  • Pred (k, m) indicates the generated predicted image signal 126.
  • a predicted value can be generated by a similar method according to the table of FIG.
  • the above is description of the unidirectional intra estimated image generation part 601 in this Embodiment of this invention.
  • FIG. 13 shows a block diagram of the bidirectional intra-predicted image generation unit 602.
  • the bidirectional intra predicted image generation unit 602 includes a first unidirectional intra predicted image generation unit 1301, a second unidirectional intra predicted image generation unit 1302, and a weighted average unit 1303, and is based on the input reference image signal 124. Two unidirectional intra-predicted images are generated, and a function of generating a predicted image signal 126 by weighting and averaging them is provided.
  • the functions of the first unidirectional intra predicted image generation unit 1301 and the second unidirectional intra predicted image generation unit 1302 are the same. In either case, a prediction image signal corresponding to a prediction mode given according to prediction mode information controlled by the encoding control unit 115 is generated.
  • a first predicted image signal 1304 is output from the first unidirectional intra predicted image generation unit 1301, and a second predicted image signal 1305 is output from the second unidirectional intra predicted image generation unit 1302.
  • Each predicted image signal is input to the weighted average unit 1303, and weighted average processing is performed.
  • the output signal of the weighted average unit 1303 is also called a third predicted image signal.
  • the table in FIG. 14 is a table for deriving two unidirectional intra prediction modes from the bidirectional intra prediction mode.
  • BiPredIdx is derived using Equation (4).
  • the first predicted image signal 1304 and the second predicted image signal 1305 generated by the first unidirectional intra predicted image generation unit 1301 and the second unidirectional intra predicted image generation unit 1302 are sent to the weighted average unit 1303. Entered.
  • the weighted average unit 1303 calculates the Euclidean distance or the city area distance (Manhattan distance) based on the prediction directions of IntraPredModeL0 and IntraPredModeL1, and derives a weight component used in the weighted average process.
  • the weight component of each pixel is represented by the Euclidean distance from the reference pixel used for prediction or the reciprocal of the urban distance, and is generalized by the following equation.
  • ⁇ L is expressed by the following equation.
  • ⁇ L is expressed by the following equation.
  • the weight table for each prediction mode is generalized by the following equation.
  • ⁇ L0 (n) represents the weight component of the pixel position n in IntraPredModeL0
  • ⁇ L1 (n) represents the weight component of the pixel position n in IntraPredModeL1. Therefore, the final prediction signal at the pixel position n is expressed by the following equation.
  • BiPred (n) represents a predicted image signal at the pixel position n
  • PredL0 (n) and PredL1 (n) are predicted image signals of IntraPredModeL0 and IntraPredModeL1, respectively.
  • the prediction signal is generated by selecting two prediction modes for generating the prediction pixel.
  • a prediction value may be generated by selecting three or more prediction modes.
  • the ratio of the reciprocal of the spatial distance from the reference pixel to the prediction pixel may be set as the weighting factor.
  • the Euclidean distance from the reference pixel used in the prediction mode or the reciprocal of the urban area distance is directly used as a weight component.
  • the Euclidean distance and the urban area distance from the reference pixel are variables.
  • the weight component may be set using the distribution model.
  • the distribution model uses at least one of a linear model, an M-order function (M ⁇ 1), a nonlinear function such as a one-sided Laplace distribution or a one-sided Gaussian distribution, and a fixed value that is a fixed value regardless of the distance from the reference pixel.
  • M ⁇ 1 M-order function
  • a nonlinear function such as a one-sided Laplace distribution or a one-sided Gaussian distribution
  • the weight component is expressed by the following equation.
  • ⁇ (n) is a weight component at the position n of the predicted pixel
  • ⁇ 2 is variance
  • A is a constant (A> 0).
  • the weight component is expressed by the following equation.
  • is a standard deviation
  • B is a constant (B> 0).
  • an isotropic correlation model obtained by modeling an autocorrelation function, an elliptic correlation model, a generalized Gaussian model obtained by generalizing a Laplace function or a Gaussian function may be used as the weight component model.
  • the circuit scale required for the said calculation can be reduced by calculating a weight component beforehand according to the relative distance for every prediction mode, and hold
  • the relative distance is a distance between a prediction target pixel and a reference pixel with respect to a certain prediction direction.
  • the city area distance ⁇ L L0 of IntraPredMode L0 and the city area distance ⁇ L L1 of IntraPredMode L1 are calculated from Equation (7).
  • the relative distance varies depending on the prediction direction of the two prediction modes.
  • the distance can be derived using Expression (6) or Expression (7) according to each prediction mode.
  • the table sizes of these distance tables may increase.
  • FIG. 17 shows the mapping of IntraPredModeLX used for distance table derivation.
  • a table of only the prediction mode corresponding to the prediction mode corresponding to the prediction mode and the DC prediction in 45 degrees is prepared, and other prediction angles are mapped closer to the prepared reference prediction mode. ing.
  • the index is mapped to the smaller one.
  • the prediction mode shown in “MappedIntraPredMode” is referred to from FIG. 17, and a distance table can be derived.
  • the relative distance for each pixel in the two prediction modes is calculated using the following equation.
  • BLK_WIDTH and BLK_HEIGHT indicate the width and height of the pixel block (prediction unit), respectively, and DistDiff (n) indicates the relative distance between the two prediction modes at the pixel position n.
  • DistDiff (n) indicates the relative distance between the two prediction modes at the pixel position n.
  • SHIFT indicates the calculation accuracy of the decimal point calculation of the weight component, and an optimal combination may be selected by balancing the coding performance and the circuit scale at the time of hardware implementation.
  • FIGS. 18A and 18B show examples in which weight components using the one-sided Laplace distribution model in the present embodiment of the present invention are tabulated.
  • Other PuSizes can also be derived using Equation (5), Equation (8), Equation (10), and Equation (11). The above is the details of the intra prediction unit 109 according to the present embodiment of the present invention.
  • the internal configuration of the intra prediction unit 109 may be the configuration shown in FIG.
  • an image buffer 1901 is added, and the bidirectional intra predicted image generation unit 602 is replaced with a weighted average unit 1303.
  • the primary image buffer 1901 has a function of temporarily storing the prediction image signal 126 for each prediction mode generated by the unidirectional intra prediction image generation unit 601 in the buffer, and the prediction controlled by the encoding control unit 115.
  • the prediction image signal 126 corresponding to the necessary prediction mode is output to the weighted average unit 1303. This eliminates the need for the bidirectional intra predicted image generation unit 602 to hold the first unidirectional intra predicted image generation unit 1301 and the second unidirectional intra predicted image generation unit 1302, thereby reducing the hardware scale. It becomes possible.
  • the syntax indicates the structure of encoded data (for example, encoded data 127 in FIG. 1) when the image encoding device encodes moving image data.
  • the moving picture decoding apparatus interprets the syntax with reference to the same syntax structure.
  • FIG. 20 shows an example of syntax 2000 used by the moving picture encoding apparatus of FIG.
  • the syntax 2000 includes three parts: a high level syntax 2001, a slice level syntax 2002, and a coding tree level syntax 2003.
  • the high level syntax 2001 includes syntax information of a layer higher than the slice.
  • a slice refers to a rectangular area or a continuous area included in a frame or a field.
  • the slice level syntax 2002 includes information necessary for decoding each slice.
  • the coding tree level syntax 2003 includes information necessary for decoding each coding tree (ie, each coding tree unit). Each of these parts includes more detailed syntax.
  • the high level syntax 2001 includes sequence and picture level syntaxes such as a sequence parameter set syntax 2004 and a picture parameter set syntax 2005.
  • the slice level syntax 2002 includes a slice header syntax 2006, a slice data syntax 2007, and the like.
  • the coding tree level syntax 2003 includes a coding tree unit syntax 2008, a prediction unit syntax 2009, and the like.
  • the coding tree unit syntax 2008 can have a quadtree structure. Specifically, the coding tree unit syntax 2008 can be recursively called as a syntax element of the coding tree unit syntax 2008. That is, one coding tree unit can be subdivided with a quadtree.
  • the coding tree unit syntax 2008 includes a transform unit syntax 2010.
  • the transform unit syntax 2010 is called at each coding tree unit syntax 2008 at the extreme end of the quadtree.
  • the transform unit syntax 2010 describes information related to inverse orthogonal transformation and quantization.
  • FIG. 21 illustrates the slice header syntax 2006 according to the present embodiment.
  • the slice_bipred_intra_flag shown in FIG. 21 is a syntax element indicating, for example, validity / invalidity of bidirectional intra prediction according to the present embodiment for the slice.
  • the orthogonal transform unit 102 and the inverse orthogonal transform unit 105 perform only unidirectional intra prediction.
  • unidirectional intra prediction prediction in which IntraBipredFlag [] in FIGS. Intra prediction specified in H.264 may be performed.
  • the bidirectional intra prediction according to the present embodiment is effective in the entire area in the slice.
  • slice_bipred_intra_flag 1
  • the prediction validity / efficiency according to the present embodiment is determined for each local region in the slice. Invalidity may be specified.
  • FIG. 22A shows an example of the prediction unit syntax.
  • Pred_mode in the figure indicates the prediction type of the prediction unit.
  • MODE_INTRA indicates that the prediction type is intra prediction.
  • intra_split_flag is a flag indicating whether or not the prediction unit is further divided into four prediction units. When intra_split_flag is 1, a prediction unit is obtained by dividing a prediction unit into four in half in the vertical and horizontal sizes. When intra_split_flag is 0, the prediction unit is not divided.
  • Intra_luma_bipred_flag [i] is a flag indicating whether the prediction mode IntraPredMode applied to the prediction unit is a unidirectional intra prediction mode or a bidirectional intra prediction mode. i indicates the position of the divided prediction unit. When the intra_split_flag is 0, 0 is set, and when the intra_split_flag is 1, 0 to 3 are set. The flag is set with the value of IntraBiprededFlag of the prediction unit shown in FIGS.
  • intra_luma_bipred_flag [i] When intra_luma_bipred_flag [i] is 1, this indicates that the prediction unit is bi-directional intra prediction, and is information that identifies the used bi-directional intra prediction mode among a plurality of prepared bi-directional intra prediction modes.
  • Intra_luma_bipred_mode [i] is encoded.
  • intra_luma_bipred_mode [i] may be encoded with the isometric length according to the number of bidirectional intra prediction modes IntraBiModeNum shown in FIG. 7, or may be encoded using a predetermined code table.
  • intra_luma_bipred_flag [i] When intra_luma_bipred_flag [i] is 0, it indicates that the prediction unit is unidirectional intra prediction, and predictive encoding is performed from adjacent blocks.
  • prev_intra_luma_unipred_flag [i] is a flag indicating whether or not the prediction value MostProbable of the prediction mode calculated from the adjacent block and the intra prediction mode of the prediction unit are the same. Details of the MostProbable calculation method will be described later. When prev_intra_luma_unipred_flag [i] is 1, it indicates that the MostProbable and the intra prediction mode IntraPredMode are equal.
  • prev_intra_luma_unipred_flag [i] When prev_intra_luma_unipred_flag [i] is 0, it indicates that the MostProbable and the intra prediction mode IntraPredMode are different, and the information rem_intraiprecoded_code that specifies whether the intra prediction mode IntraPredMode is a mode other than MostProbable. . rem_intra_luma_unipred_mode [i] may be encoded with the isometric length according to the bidirectional intra prediction mode number IntraUniModeNum shown in FIG. 7, or may be encoded using a predetermined code table. From the intra prediction mode IntraPredMode, rem_intra_luma_unipred_mode [i] is calculated using the following equation.
  • MostProbable is calculated according to the following equation.
  • Min (x, y) is a parameter for outputting the smaller one of the inputs x and y.
  • intraPredModeA and intraPredModeB indicate intra prediction modes of prediction units adjacent to the left and above the encoded prediction unit.
  • intraPredModeA and intraPredModeB are collectively expressed as intraPredModeN.
  • N is set to A or B.
  • a method of calculating intraPredModeN will be described using the flowchart shown in FIG. First, it is determined whether a coding tree unit to which an adjacent prediction unit belongs can be used (step S2301). If the coding tree unit cannot be used (NO in S2301), reference to intraPredModeN is not possible. “ ⁇ 1” indicating “” is set.
  • intraPredModeN is calculated using the following equation.
  • IntraUniModeNum is the number of unidirectional intra prediction modes determined by the size of the adjacent prediction unit, and an example thereof is shown in FIG.
  • “MappedBi2Uni (List, idx)” is a table for converting the bidirectional intra prediction mode into the unidirectional intra prediction mode.
  • List is the unidirectional intra prediction mode of List0 (corresponding to IntraPredTypeL0 [] shown in FIGS. 9, 10, and 11) of the two unidirectional intra prediction modes constituting the bidirectional intra prediction mode. Is a flag indicating whether to use the unidirectional intra prediction mode of List1 or the IntraPredTypeL1 [] shown in FIGS. 9, 10, and 11).
  • List1 is used for conversion to the unidirectional intra prediction mode.
  • FIG. 14 shows an example of the conversion table. The numerical values in the figure correspond to IntraPredMode shown in FIGS.
  • MappedMostProble () is a table for converting MostProbable, and an example is shown in FIG. 24.
  • luma_pred_mode_code_type [i] indicates the type of the prediction mode IntraPredMode applied to the prediction unit, where 0 (IntraUnifiedMostProb) is unidirectional intra prediction and the intra prediction mode is the same as MostProbable, 1 (IntraUnipre intrareprediction) The intra prediction mode is different from MostProbable, and 2 (IntraBipred) indicates a bidirectional intra prediction mode.
  • 0 IntraUnifiedMostProb
  • 1 IntraUnipre intrareprediction
  • the intra prediction mode is different from MostProbable
  • 2 IntraBipred indicates a bidirectional intra prediction mode.
  • FIG. 24 shows bidirectional intra prediction modes, 1 and unidirectional intra prediction (whether or not the same prediction mode as MostProbable).
  • FIG. 25 shows an example of assignment of the number of modes according to the meaning corresponding to luma_pred_mode_code_type, bin, and the mode configuration shown in FIG.
  • luma_pred_mode_code_type [i] When luma_pred_mode_code_type [i] is 0, the intra prediction mode is the MostProbable mode, so no further information encoding is necessary.
  • luma_pred_mode_code_type [i] is 1, information rem_intra_luma_unipred_mode [i] that specifies which mode other than MostProbable is the intra prediction mode IntraPredMode is encoded.
  • rem_intra_luma_unipred_mode [i] may be encoded with the isometric length according to the bidirectional intra prediction mode number IntraUniModeNum shown in FIG. 7, or may be encoded using a predetermined code table. From the intra prediction mode IntraPredMode, rem_intra_luma_unipred_mode [i] is calculated using Equation (16). Further, when luma_pred_mode_code_type [i] is 2, it indicates that the prediction unit is bidirectional intra prediction, and information that identifies the used bidirectional intra prediction mode among the prepared bidirectional intra prediction modes. Intra_luma_bipred_mode [i] is encoded.
  • intra_luma_bipred_mode [i] may be encoded with the isometric length according to the number of bidirectional intra prediction modes IntraBiModeNum shown in FIG. 7, or may be encoded using a predetermined code table.
  • the above is the syntax configuration according to the present embodiment.
  • FIG. 22D Yet another example of the prediction unit syntax is shown in FIG. 22D.
  • the table shown in FIG. 43 may be used instead of FIG. 9, and the IntraPredMode in FIG. The table may be ignored.
  • FIG. 43 is a table in which IntraPredTypeL1 and IntraPredAngleIdL1 indicating information related to the second prediction mode at the time of bidirectional intra prediction are deleted from FIG.
  • IntraPredMode 33 or more tables are deleted.
  • FIG. 42 and FIG. 9 are applicable also regarding FIG. 10, FIG.
  • pred_mode and intra_split_flag are the same as the syntax example described above, and thus description thereof is omitted.
  • Intra_bipred_flag is a flag indicating whether or not bi-directional intra prediction can be used in the encoded prediction unit. When intra_bipred_flag is 0, it indicates that bi-directional intra prediction is not used in the encoded prediction unit. Even when intra_split_flag is 1, that is, when the encoded prediction unit is further divided into four, bi-directional intra prediction is not used in all prediction units, and only uni-directional intra prediction is effective.
  • intra_bipred_flag When intra_bipred_flag is 1, it indicates that bi-directional intra prediction can be used in the encoded prediction unit. Even when intra_split_flag is 1, that is, when the encoded prediction unit is further divided into four, in all prediction units, bidirectional intra prediction can be selected in addition to unidirectional intra prediction.
  • intra_bipred_flag is encoded as 0 to disable bi-directional intra prediction. Since the amount of codes necessary for encoding can be reduced, encoding efficiency is improved.
  • FIG. 22E Still another example relating to the prediction unit syntax is shown in FIG. 22E.
  • intra_bipred_flag is a flag indicating whether or not bi-directional intra prediction can be used in the encoding prediction unit, and is the same as the above-described intra_bipred_flag, and thus the description thereof is omitted.
  • FIG. 26 shows the intra prediction unit 109 when adaptive reference pixel filtering is used. 6 is different from the intra prediction unit 109 shown in FIG. 6 in that a reference pixel filter unit 2601 is added.
  • the reference pixel filter unit 2601 receives the reference image signal 124 and the prediction mode 605, performs an adaptive filtering process described later, and outputs a filtered reference image signal 2602.
  • the filtered reference image signal 2602 is input to the unidirectional intra predicted image generation unit 601 and the bidirectional intra predicted image generation unit 602.
  • the configuration and processing other than the reference pixel filter unit 2601 are the same as those of the intra prediction unit 109 shown in FIG.
  • the reference pixel filter unit 2601 determines whether or not to filter reference pixels used for intra prediction according to the reference pixel filter flag and the intra prediction mode included in the prediction mode 605.
  • the reference pixel filter flag is a flag indicating whether or not reference pixels are filtered when the intra prediction mode IntraPredMode is a value other than “Intra_DC”.
  • IntraPredMode is “Intra_DC”
  • the reference pixel is not filtered and the reference pixel filter flag is set to 0.
  • a filtered reference image signal 2602 is calculated by the following filtering.
  • p [x, y] indicates a reference pixel before filtering
  • pf [x, y] indicates a reference pixel in filter terms.
  • PuPartSize indicates the size (pixel) of the prediction unit.
  • FIG. 27A and 27B show a prediction unit syntax structure when performing adaptive reference pixel filtering.
  • FIG. 27A adds the syntax intra_luma_filter_flag [i] related to the adaptive reference pixel filter to FIG. 22A.
  • FIG. 27B adds syntax intra_luma_filter_flag [i] related to the adaptive reference pixel filter to FIG. 22C.
  • intra_luma_filter_flag [i] is further encoded when the intra prediction mode IntraPredMode [i] is other than Intra_DC. When the flag is 0, it indicates that the reference pixel is not filtered. Further, when intra_luma_filter_flag [i] is 1, it indicates that the reference pixel filtering is applied.
  • intra_luma_filter_flag [i] is encoded when the intra prediction mode IntraPredMode [i] is other than Intra_DC.
  • IntraPredMode [i] is 0 to 2
  • intra_luma_filter_flag [i ] May not be encoded. In this case, intra_luma_filter_flag [i] is set to 0.
  • intra_luma_filter_flag [i] described above may be added in the same meaning for the other syntax structures shown in FIGS. 22B, 22D, and 22E.
  • the decoded image signal 122 is calculated in the moving image decoding device 4400 or the image encoding device 100, it is possible to use decoded pixels as pixels adjacent to the left, upper, and upper left.
  • the input image signal 116 is used as a pixel adjacent to the left, upper, and upper left.
  • FIG. 28 shows positions of adjacent decoded pixels A (left), B (upper), and C (upper left) used for prediction of the prediction target pixel X. Therefore, composite intra prediction is a so-called open-loop prediction method in which prediction values differ between the image encoding device 100 and the moving image decoding device 4400.
  • FIG. 30 shows a block diagram of the intra prediction unit 4408 (109) when combined with composite intra prediction. A difference is that a composite intra predicted image generation unit 2901, a selection switch 2902, and a decoded image buffer 3001 are added to the intra prediction unit 109 shown in FIG.
  • the selection switch 604 When the bidirectional intra prediction and the composite intra prediction are combined, first, in the selection switch 604, the unidirectional intra prediction image generation unit 601 or the bidirectional intra prediction image generation unit according to the prediction mode information controlled by the encoding control unit 115. The output terminal of 602 is switched.
  • the output predicted image signal 126 is referred to as a direction predicted image signal 126.
  • the direction prediction image signal is input to the composite intra prediction image generation unit 2901, and a prediction image signal 4420 (126) in the composite intra prediction is generated.
  • the description of the composite intra predicted image generation unit 2901 will be described later.
  • the selection switch 2902 which one of the prediction image signal 4420 (126) and the direction prediction image signal in the composite intra prediction is used according to the composite intra prediction application flag in the prediction mode information controlled by the encoding control unit 115.
  • the final predicted image signal 4420 (126) in the intra prediction unit 4408 (109) is output.
  • the composite intra prediction application flag is 1
  • the predicted image signal 4420 (126) output from the composite intra predicted image generation unit 2901 is the final predicted image signal 4420 (126).
  • the direction prediction image signal 4420 (126) is the prediction image signal 4420 (126) that is finally output.
  • the predicted image signal output from the composite intra predicted image generation unit 2901 is also called a sixth predicted image signal.
  • the decoded prediction error signal 4416 separately decoded by the addition unit 4405 is added to the pixel unit to generate a decoded image signal 4417 for each pixel. And stored in the decoded pixel buffer 3001.
  • the stored decoded image signal 4417 in units of pixels is input to the composite intra predicted image generation unit 2901 as the reference pixel 3002, and is used for pixel level prediction described later as the adjacent pixel 3104 shown in FIG.
  • the composite intra prediction image generation unit 2901 includes a pixel level prediction signal generation unit 3101 and a composite intra prediction calculation unit 3102.
  • the pixel level prediction signal generation unit 3101 receives the reference pixel 3002 as the adjacent pixel 3104, and outputs the pixel level prediction signal 3103 by predicting the prediction target pixel X from the adjacent pixel.
  • the pixel level prediction signal 3103 (X) of the prediction target pixel is calculated from A, B, and C indicating the adjacent pixel 3104 using Expression (21).
  • coefficients related to A, B, and C may be other values.
  • the composite intra prediction calculation unit 3102 performs a weighted average of the direction prediction image signal 126 (X ′) and the pixel level prediction signal 3103 (X), and outputs a final prediction image signal 126 (P). Specifically, the following formula is used.
  • the decoded image signal 122 may have different values in encoding and decoding. . Therefore, after all the decoded image signals 122 in the encoded prediction syntax are generated, the above-described combined intra prediction is performed again using the decoded image signal 122 as an adjacent pixel, so that the same predicted image signal 126 as that in the decoding is obtained. Is further added to the prediction error signal 117, and the decoded image signal 122 identical to the decoding can be generated.
  • the weighting factor W may be switched according to the position of the prediction pixel in the prediction unit.
  • a prediction image signal generated using unidirectional intra prediction and bidirectional intra prediction generates a prediction value from spatially adjacent reference pixels positioned on the left or above already encoded.
  • the absolute value of the prediction error tends to increase as the distance from the reference pixel increases. Therefore, the weighting coefficient of the direction prediction image signal 126 and the pixel level prediction signal 3103 is increased when the weight coefficient of the direction prediction image signal 126 is close to the reference pixel, and is decreased when the distance is far away, thereby improving the prediction accuracy. It becomes possible.
  • a prediction error signal is generated using an input image signal at the time of encoding.
  • the pixel level prediction signal 3103 becomes an input image signal, even if the spatial distance between the reference pixel position and the prediction pixel position is increased, the prediction of the pixel level prediction signal 3103 is compared with the direction prediction image signal 126. High accuracy.
  • the weighting coefficient of the direction prediction image signal 126 and the pixel level prediction signal 3103 is simply increased when the weight coefficient of the direction prediction image signal 126 is close to the reference pixel, and is decreased when the distance is small.
  • the prediction error is reduced, there is a problem that the prediction accuracy at the time of encoding and the prediction value at the time of local decoding are different and the prediction accuracy is lowered. Therefore, especially when the value of the quantization parameter is large, as the spatial distance between the reference pixel position and the predicted pixel position becomes large, the difference generated in the case of such an open loop is set by setting the value of W small. A decrease in coding efficiency due to the phenomenon can be suppressed.
  • FIG. 32A and 32B show a prediction unit syntax structure when performing composite intra prediction.
  • FIG. 32A is different from FIG. 22A in that a syntax combined_intra_pred_flag for switching presence / absence of composite intra prediction is added. This is equivalent to the above-described composite intra prediction application flag.
  • FIG. 32B adds a syntax combined_intra_pred_flag for switching presence / absence of composite intra prediction to FIG. 22C.
  • the selection switch 2902 shown in FIG. 30 is connected to the output terminal of the composite intra prediction image generation unit 2901.
  • the selection switch 2902 shown in FIG. 30 is connected to the output terminal of either the unidirectional intra prediction image generation unit 601 or the bidirectional intra prediction image generation unit 602 to which the selection switch 604 is connected. .
  • intra_luma_filter_flag [i] described above may be added in the same meaning for the other syntax structures shown in FIGS. 22B, 22D, and 22E.
  • the video encoding apparatus according to the second embodiment differs from the above-described image encoding apparatus according to the first embodiment in the details of orthogonal transform and inverse orthogonal transform.
  • the same parts as those in the first embodiment are denoted by the same indexes, and different parts will be mainly described.
  • a moving picture decoding apparatus corresponding to the picture encoding apparatus according to the present embodiment will be described in a fifth embodiment.
  • FIG. 33 is a block diagram showing a video encoding apparatus according to the second embodiment.
  • the change from the moving picture encoding apparatus according to the first embodiment is that a transformation selection unit 3301 and a coefficient order control unit 3302 are added. Also, the internal structures of the orthogonal transform unit 102 and the inverse orthogonal transform unit 105 are different.
  • the process of FIG. 33 will be described.
  • ⁇ Orthogonal transform unit 102> 34 includes a first orthogonal transform unit 3401, a second orthogonal transform unit 3402, an Nth orthogonal transform unit 3403, and a transform selection switch 3404.
  • N types of orthogonal transform units there may be a plurality of transform sizes using the same orthogonal transform method, or there may be a plurality of orthogonal transform units performing different orthogonal transform methods. .
  • each may be mixed.
  • the first orthogonal transform unit 3401 can be set to 4 ⁇ 4 size DCT
  • the second orthogonal transform unit 3402 can be set to 8 ⁇ 8 size DCT
  • the Nth orthogonal transform unit 3403 can be set to 16 ⁇ 16 size DCT.
  • the first orthogonal transform unit 3401 is 4 ⁇ 4 size DCT
  • the second orthogonal transform unit 3402 is 4 ⁇ 4 size DST (discrete sine transform)
  • the Nth orthogonal transform unit 3403 is 8 ⁇ 8 size KLT (Karunen-Labe transform). ) Can also be set.
  • N 1 is considered.
  • the conversion selection switch 3404 has a function of selecting the output terminal of the subtraction unit 101 according to the conversion selection information 3303.
  • the transformation selection information 3303 is one piece of information controlled by the encoding control unit 115 and is set by the transformation selection unit 3301 according to the prediction information 125.
  • the transformation selection information 3303 indicates the first orthogonal transformation
  • the output terminal of the switch is connected to the first orthogonal transformation unit 3401.
  • the transformation selection information 3303 is the second orthogonal transformation
  • the output end is connected to the second orthogonal transformation unit 3402.
  • the first orthogonal transform unit 3401 performs DCT
  • the other orthogonal transform units 3402 and 3403 perform KLT (Carhunen-Labe transform).
  • the inverse orthogonal transform unit 105 in FIG. 35 includes a first inverse orthogonal transform unit 3501, a second inverse orthogonal transform unit 3502, an Nth inverse orthogonal transform unit 3503, and a transform selection switch 3504.
  • the conversion selection switch 3504 will be described.
  • the conversion selection switch 3504 has a function of selecting the output terminal of the inverse quantization unit 104 according to the input conversion selection information 3303.
  • the transformation selection information 3303 is one piece of information controlled by the encoding control unit 115 and is set by the transformation selection unit 3301 according to the prediction information 125.
  • the output terminal of the switch is connected to the first inverse orthogonal transform unit 3501.
  • the transformation selection information 3303 is the second orthogonal transformation
  • the output end is connected to the second inverse orthogonal transformation unit 3502.
  • the transform selection information 3303 is the Nth orthogonal transform
  • the output terminal is connected to the Nth inverse orthogonal transform unit 3503.
  • the transform selection information 3303 set in the orthogonal transform unit 102 and the transform selection information 3303 set in the inverse orthogonal transform unit 105 are the same, and the inverse orthogonal transform corresponding to the transform performed in the orthogonal transform unit 102 is performed.
  • the first inverse orthogonal transform unit 3501 performs inverse discrete cosine transform (hereinafter referred to as IDCT), and the second inverse orthogonal transform unit 3502 and the Nth inverse orthogonal transform unit 3503 are based on KLT (Karunen-Labe transform).
  • IDCT inverse discrete cosine transform
  • KLT Karunen-Labe transform
  • Inverse transformation is performed.
  • orthogonal transformation such as Hadamard transformation or discrete sine transformation may be used, or non-orthogonal transformation may be used.
  • the corresponding inverse conversion is performed in conjunction with the conversion unit 102.
  • Prediction information 125 controlled by the encoding control unit 115 and including the prediction mode set by the prediction selection unit 112 is input to the transform selection unit 3301.
  • the transform selection unit 3301 has a function of setting MapdTransformIdx information indicating which orthogonal transform is used for which prediction mode.
  • FIG. 36 shows conversion selection information 3303 (MappedTransformIdx) in intra prediction.
  • FIG. 37 shows a block diagram of the coefficient order control unit 3302.
  • the coefficient order control unit 3302 includes a coefficient order selection switch 3704, a first coefficient order conversion unit 3701, a second coefficient order conversion unit 3702, and an Nth coefficient order conversion unit 3703.
  • the coefficient order selection switch 3704 has a function of switching the output terminal of the switch and the coefficient order conversion units 3701 to 3703 in accordance with, for example, the mapped transform IDx shown in FIG.
  • the N types of coefficient forward conversion units 3701 to 3703 have a function of converting the two-dimensional data of the quantized conversion coefficient 119 quantized by the quantization unit 103 into one-dimensional data. For example, H.M. In H.264, two-dimensional data is converted into one-dimensional data using a zigzag scan.
  • the quantized transform coefficient 119 obtained by performing the quantization process on the transform coefficient 118 subjected to the orthogonal transform has a characteristic that the tendency of generating non-zero transform coefficients in the block is biased. have.
  • the tendency of occurrence of this non-zero transform coefficient has different properties for each prediction direction of intra prediction.
  • the generation tendency of non-zero transform coefficients in the same prediction direction has a similar property. Therefore, when transforming two-dimensional data into one-dimensional data (2D-1D conversion), entropy coding is performed preferentially from transform coefficients at positions where the occurrence probability of non-zero transform coefficients is high, thereby encoding transform coefficients.
  • the coefficient order control unit 3302 may dynamically update the scan order in 2D-1D conversion.
  • the coefficient order control unit 3302 that performs such an operation is illustrated in FIG.
  • the coefficient order control unit 3302 includes an occurrence frequency counting unit 3801 and an updating unit 3802 in addition to the configuration of FIG.
  • the coefficient order conversion units 3701,..., 3703 are the same except that the scan order is updated by the coefficient order update unit 3802.
  • the occurrence frequency counting unit 3801 creates a histogram 3804 of the number of occurrences of non-zero coefficients in each element of the quantized transform coefficient sequence 3304 for each prediction mode.
  • the occurrence frequency counting unit 3801 inputs the created histogram 3804 to the update unit 3802.
  • the update unit 3802 updates the coefficient order based on the histogram 3804 at a predetermined timing.
  • the timing is, for example, the timing when the coding process of the coding tree unit is finished, the timing when the coding process for one line in the coding tree unit is finished, or the like.
  • the update unit 3802 refers to the histogram 3804 and updates the coefficient order with respect to a prediction mode having an element in which the number of occurrences of non-zero coefficients is counted more than a threshold. For example, the update unit 3802 updates the prediction mode having an element in which the occurrence of a non-zero coefficient is counted 16 times or more. By providing a threshold value for the number of occurrences, the coefficient order is updated globally, so that it is difficult to converge to a local optimum solution.
  • the update unit 3802 sorts the elements in descending order of the occurrence frequency of the non-zero coefficient with respect to the prediction mode to be updated. Sorting can be realized by existing algorithms such as bubble sort and quick sort. Then, the update unit 3802 inputs the update coefficient order 3803 indicating the order of the sorted elements to the coefficient order conversion units 3701 to 3703 corresponding to the prediction mode to be updated.
  • each conversion unit When the update coefficient order 3803 is input, each conversion unit performs 2D-1D conversion according to the updated scan order.
  • the initial scan order of each 2D-1D conversion unit needs to be determined in advance. In this way, by dynamically updating the scan order, the tendency of occurrence of non-zero coefficients in the quantized transform coefficients 119 changes according to the influence of the properties of the predicted image, quantization information (quantization parameters), and the like. Even in this case, high encoding efficiency can be expected stably. Specifically, the generated code amount of run-length encoding in the entropy encoding unit 113 can be suppressed.
  • the syntax configuration in this embodiment is the same as that in the first embodiment.
  • the conversion selection unit 3301 can select the mapped transform IDx separately from the prediction information 125.
  • information indicating which nine types of orthogonal transforms or inverse orthogonal transforms are used is set in the entropy encoding unit 113 and encoded together with the quantized transform coefficient sequence 3304.
  • FIG. 39 shows an example of syntax in this modification.
  • Directional_transform_idx indicated in the syntax indicates information indicating which of N orthogonal transforms has been selected.
  • FIG. 40 shows a block diagram of the orthogonal transform unit 102 according to the present embodiment.
  • the orthogonal transform unit 102 includes new processing units such as a first rotation transform unit 4001, a second rotation transform unit 4002, an Nth rotation transform unit 4003, and a discrete cosine transform unit 4004, and has an existing transform selection switch 3404.
  • the discrete cosine transform unit 4004 performs DCT, for example.
  • the conversion coefficient after DCT is input to the conversion selection switch 3404.
  • the conversion selection switch 3404 connects the output end of the switch to one of the first rotation conversion unit 4001, the second rotation conversion unit 4002, and the Nth rotation conversion unit 4003 according to the conversion selection information 3303.
  • the switches are sequentially switched according to the control of the encoding control unit 115.
  • the rotation conversion units 4001 to 4003 perform rotation conversion for each conversion coefficient using a predetermined rotation matrix.
  • the conversion coefficient 118 after the rotation conversion is output. This conversion is a reversible conversion.
  • rotation matrix it may be determined which rotation matrix is to be used by using the encoding cost as shown in Equation (1) and Equation (2).
  • the rotation conversion unit may be applied to the quantization conversion coefficient 119 after the quantization process.
  • the orthogonal transform unit 102 performs only DCT.
  • FIG. 41 is a block diagram of the inverse orthogonal transform unit 105 according to the present embodiment.
  • the inverse orthogonal transform unit 105 includes new processing units such as a first inverse rotation transform unit 4101, a second inverse rotation transform unit 4102, an Nth inverse rotation transform unit 4103, and an inverse discrete cosine transform unit 4104, and an existing transform selection switch 3504.
  • the conversion selection switch 3504 connects the output terminal of the switch to one of the first reverse rotation conversion unit 4101, the second reverse rotation conversion unit 4102, and the Nth reverse rotation conversion unit 4103 according to the conversion selection information 3303.
  • the same inverse rotation transform used in the orthogonal transform unit 102 is subjected to inverse rotation transform processing by any one of the inverse rotation transform units 4101 to 4103, and is output to the inverse discrete cosine transform unit 4104.
  • the inverse discrete cosine transform unit 4104 performs, for example, IDCT on the input signal to restore the restored prediction error signal 121.
  • IDCT an example using IDCT is shown here as an example, orthogonal transform such as Hadamard transform or discrete sine transform may be used, or non-orthogonal transform may be used. In any case, the corresponding inverse conversion is performed in conjunction with the conversion unit 102.
  • FIG. 42 shows the syntax in the present embodiment.
  • the rotation_transform_idx shown in the syntax means the number of the rotation matrix to be used.
  • the fourth embodiment relates to a moving picture decoding apparatus.
  • the video encoding device corresponding to the video decoding device according to the present embodiment is as described in the first embodiment. That is, the moving picture decoding apparatus according to the present embodiment decodes encoded data generated by, for example, the moving picture encoding apparatus according to the first embodiment.
  • the moving picture decoding apparatus includes an input buffer 4401, an entropy decoding unit 4402, an inverse quantization unit 4403, an inverse orthogonal transform unit 4404, an addition unit 4405, and a loop filter 4406.
  • An image memory 4407, an intra prediction unit 4408, an inter prediction unit 4409, a prediction selection switch 4410, and an output buffer 4411 are included.
  • the encoded data 4413 decodes the encoded data 4413 stored in the input buffer 4401, stores the decoded image 4422 in the output buffer 4411, and outputs it as an output image.
  • the encoded data 4413 is output from, for example, the moving image encoding apparatus shown in FIG. 1, and is temporarily stored in the input buffer 4401 via a storage system or transmission system (not shown).
  • the entropy decoding unit 4402 performs decoding based on the syntax for each frame or field for decoding the encoded data 4413.
  • the entropy decoding unit 4402 sequentially entropy-decodes the code string of each syntax, and reproduces the encoding parameters of the encoding target block such as the prediction information 4421 including the prediction mode information and the quantization transform coefficient 4414.
  • the encoding parameter is a parameter necessary for decoding such as prediction information 4421, information on transform coefficients, information on quantization, and the like.
  • the inverse quantization unit 4403 performs inverse quantization on the quantized transform coefficient 4414 from the entropy decoding unit 4402 to obtain a restored transform coefficient 4415. Specifically, the inverse quantization unit 4403 performs inverse quantization according to the information regarding the quantization decoded by the entropy decoding unit 4402. The inverse quantization unit 4403 inputs the restored transform coefficient 4415 to the inverse orthogonal transform unit 4404.
  • the inverse orthogonal transform unit 4404 performs inverse orthogonal transform corresponding to the orthogonal transform performed on the encoding side, on the reconstruction transform coefficient 4415 from the inverse quantization unit 4403, and obtains a reconstruction prediction error signal 4416.
  • the inverse orthogonal transform unit 4404 inputs the restored prediction error signal 4416 to the adder 4405.
  • the addition unit 4405 adds the restored prediction error signal 4416 and the corresponding predicted image signal 4420 to generate a decoded image signal 4417.
  • the decoded image signal 4417 is input to the loop filter 4406.
  • the loop filter 4406 performs a deblocking filter, a Wiener filter, or the like on the input decoded image signal 4417 to generate a filtered image signal 4418.
  • the generated filtered image signal 4418 is temporarily stored in the output buffer 4411 for the output image, and is also stored in the reference image memory 4407 for the reference image signal 4419.
  • the filtered image signal 4418 stored in the reference image memory 4407 is referenced as a reference image signal 4419 by the intra prediction unit 4408 and the inter prediction unit 4409 as necessary in units of frames or fields.
  • the filtered image signal 4418 temporarily accumulated in the output buffer 4411 is output according to the output timing managed by the decoding control unit 4412.
  • the intra prediction unit 4408, the inter prediction unit 4409, and the selection switch 4410 are substantially the same or similar elements as the intra prediction unit 109, the inter prediction unit 110, and the selection switch 111 in FIG.
  • the intra prediction unit 4408 (109) performs intra prediction using the reference image signal 4419 stored in the reference image memory 4407.
  • H.M. In H.264, an intra prediction image is obtained by performing pixel interpolation (copying or copying after interpolation) along a prediction direction such as a vertical direction or a horizontal direction using an encoded reference pixel value adjacent to a prediction target block. Generate. FIG. The prediction direction of intra prediction in H.264 is shown. Further, in FIG. 2 shows an arrangement relationship between reference pixels and encoding target pixels in H.264. 5C shows a predicted image generation method in mode 1 (horizontal prediction), and FIG. 5D shows a predicted image generation method in mode 4 (diagonal lower right prediction; Intra_NxN_Diagonal_Down_Right in FIG. 4A). Yes.
  • the inter prediction unit 4409 (110) performs inter prediction using the reference image signal 4419 stored in the reference image memory 4407. Specifically, the inter prediction unit 4409 (110) obtains a motion shift amount (motion vector) between the prediction target block and the reference image signal 124 from the entropy decoding unit 4402, and based on this motion vector. Inter prediction processing (motion compensation) is performed to generate an inter predicted image. H. With H.264, interpolation processing up to 1/4 pixel accuracy is possible.
  • the prediction selection switch 4410 selects the output terminal of the intra prediction unit 4408 or the output terminal of the inter prediction unit 4409 according to the decoded prediction information 4421, and inputs the intra predicted image or the inter predicted image as the predicted image signal 4420 to the adding unit 4405. .
  • the prediction selection switch 4410 connects a switch to the output terminal from the intra prediction unit 4408.
  • the prediction selection switch 4410 connects a switch to the output terminal from the inter prediction unit 4409.
  • the decoding control unit 4412 controls each element of the moving picture decoding apparatus in FIG. Specifically, the decoding control unit 4412 performs various controls for decoding processing including the above-described operation.
  • the intra prediction unit 4408 has the same configuration and processing content as the intra prediction unit 109 described in the first embodiment.
  • the intra prediction unit 4408 (109) shown in FIG. 6 includes a unidirectional intra predicted image generation unit 601, a bidirectional intra predicted image generation unit 602, a prediction mode information setting unit 603, and a selection switch 604.
  • a reference image signal 4419 (124) is input from the reference image memory 4407 to the unidirectional intra predicted image generation unit 601 and the bidirectional intra predicted image generation unit 602.
  • the prediction mode information setting unit 603 selects the prediction mode generated by the unidirectional intra prediction image generation unit 601 or the bidirectional intra prediction image generation unit 602.
  • the selection switch 604 has a function of switching the output ends of the respective intra predicted image generation units according to the prediction mode 605.
  • the output terminal of the unidirectional intra prediction image generation unit 601 is connected to the switch, and if the prediction mode 605 is the bidirectional intra prediction mode, the bidirectional intra prediction is performed.
  • the output terminal of the image generation unit 602 is connected.
  • generation part 601 and 602 produces
  • FIG. The generated predicted image signal 4420 (126) is output from the intra prediction unit 109.
  • FIG. 7 shows the number of prediction modes according to the block size according to the present embodiment of the present invention.
  • PuSize indicates the pixel block (prediction unit) size to be predicted, and seven types of sizes from PU_2x2 to PU_128x128 are defined.
  • IntraUniModeNum represents the number of prediction modes for unidirectional intra prediction
  • IntraBiModeNum represents the number of prediction modes for bidirectional intra prediction.
  • Number of modes is the total number of prediction modes for each pixel block (prediction unit) size.
  • FIG. 9 shows the relationship between the prediction mode and the prediction method when PuSize is PU_8x8, PU_16x16, and PU_32x32.
  • FIG. 10 shows a case where PuSize is PU_4 ⁇ 4
  • FIG. 11 shows a case where PU_64 ⁇ 64 or PU_128 ⁇ 128.
  • IntraPredMode indicates a prediction mode number
  • IntraBipredFlag is a flag indicating whether or not bidirectional intra prediction. When the flag is 1, it indicates that the prediction mode is the bidirectional intra prediction mode. When the flag is 0, it indicates that the prediction mode is a unidirectional intra prediction mode.
  • IntraPredTypeLX indicates the prediction type of intra prediction.
  • Intra_Vertical means that the vertical direction is the reference for prediction
  • Intra_Horizontal means that the horizontal direction is the reference for prediction. Note that 0 or 1 is applied to X in IntraPredTypeLX.
  • IntraPredTypeL0 indicates the first prediction mode of unidirectional intra prediction or bidirectional intra prediction.
  • IntraPredTypeL1 indicates the second prediction mode of bidirectional intra prediction.
  • IntraPred AngleID is an index indicating an index of a prediction angle. The prediction angle actually used in the generation of the predicted value is shown in FIG.
  • puPartIdx represents the index of the divided block in the quadtree division described with reference to FIG. 3B.
  • IntraPredMode 4
  • IntraPredTypeL0 Intra_Vertical
  • the prediction mode information setting unit 603 converts the above-described prediction information corresponding to the designated prediction mode 605 to the unidirectional intra prediction image generation unit 601 and the bidirectional intra prediction image generation unit 602 under the control of the decoding control unit 4412. And the prediction mode 605 is output to the selection switch.
  • the unidirectional intra predicted image generation unit 601 has a function of generating a predicted image signal 4420 (126) for a plurality of prediction directions shown in FIG. In FIG. 8, there are 33 different prediction directions for the vertical and horizontal coordinates indicated by the bold lines.
  • the direction of a typical prediction angle indicated by H.264 is indicated by an arrow.
  • 33 kinds of prediction directions are prepared in a direction in which a line is drawn from the origin to a mark indicated by a diamond.
  • IntraPredMode 4
  • IntraPredAngleIDL0 4
  • An arrow indicated by a dotted line in FIG. 8 indicates a prediction mode whose prediction type is Intra_Vertical, and an arrow indicated by a solid line indicates a prediction mode whose prediction type is Intra_Horizontal.
  • FIG. 12 shows the relationship between IntraPredAngleIDLX and intraPredAngle used for predictive image value generation.
  • intraPredAngle indicates a prediction angle that is actually used when a predicted value is generated.
  • a prediction value generation method is expressed by Expression (3).
  • BLK_SIZE indicates the size of the pixel block (prediction unit)
  • ref [] indicates an array in which reference image signals are stored.
  • pred (k, m) indicates the generated predicted image signal 4420 (126).
  • a predicted value can be generated by a similar method according to the table of FIG.
  • the above is description of the unidirectional intra estimated image generation part 601 in this Embodiment of this invention.
  • FIG. 13 shows a block diagram of the bidirectional intra-predicted image generation unit 602.
  • the bidirectional intra predicted image generation unit 602 includes a first unidirectional intra predicted image generation unit 1301, a second unidirectional intra predicted image generation unit 1302, and a weighted average unit 1303.
  • the input reference image signal 4419 (124 ) Two unidirectional intra prediction images are generated, and a weighted average of these is generated to generate a prediction image signal 4420 (126).
  • the functions of the first unidirectional intra predicted image generation unit 1301 and the second unidirectional intra predicted image generation unit 1302 are the same. In either case, a prediction image signal corresponding to a prediction mode given according to prediction mode information controlled by the encoding control unit 115 is generated.
  • a first predicted image signal 1304 is output from the first unidirectional intra predicted image generation unit 1301, and a second predicted image signal 1305 is output from the second unidirectional intra predicted image generation unit 1302.
  • Each predicted image signal is input to the weighted average unit 1303, and weighted average processing is performed.
  • the table in FIG. 14 is a table for deriving two unidirectional intra prediction modes from the bidirectional intra prediction mode.
  • BiPredIdx is derived using Equation (4).
  • the first predicted image signal 1304 and the second predicted image signal 1305 generated by the first unidirectional intra predicted image generation unit 1301 and the second unidirectional intra predicted image generation unit 1302 are sent to the weighted average unit 1303. Entered.
  • the weighted average unit 1303 calculates a Euclidean distance or a city area distance (Manhattan distance) based on the prediction directions of IntraPredModeL0 and IntraPredModeL1, and derives a weight component used in the weighted average process.
  • the weight component of each pixel is represented by the reciprocal of the Euclidean distance or the city distance from the reference pixel used for prediction, and is generalized by Expression (5).
  • ⁇ L is expressed by Equation (6).
  • ⁇ L is expressed by Equation (7).
  • the weight table for each prediction mode is generalized to Equation (8). Therefore, the final prediction signal at the pixel position n is expressed by Equation (9).
  • the prediction signal is generated by selecting two prediction modes for generating the prediction pixel.
  • a prediction value may be generated by selecting three or more prediction modes.
  • the ratio of the reciprocal of the spatial distance from the reference pixel to the prediction pixel may be set as the weighting factor.
  • the Euclidean distance from the reference pixel used in the prediction mode or the reciprocal of the urban area distance is used as a weight component as it is, but as another embodiment, the Euclidean distance from the reference pixel and the urban area distance are variables.
  • the weight component may be set using the distributed model.
  • the distribution model uses at least one of a linear model, an M-order function (M ⁇ 1), a nonlinear function such as a one-sided Laplace distribution or a one-sided Gaussian distribution, or a fixed value that is a fixed value regardless of the distance from the reference pixel.
  • M ⁇ 1 M-order function
  • the weight component is expressed by Equation (10).
  • the weight component is expressed by Expression (11).
  • an isotropic correlation model obtained by modeling an autocorrelation function, an elliptic correlation model, a generalized Gaussian model obtained by generalizing a Laplace function or a Gaussian function may be used as the weight component model.
  • Equation (5), Equation (8), Equation (10), and Equation (11) are calculated each time the predicted image is generated, a plurality of multipliers are required, and the hardware scale increases. . For this reason, the circuit scale required for the said calculation can be reduced by calculating a weight component beforehand according to the relative distance for every prediction mode, and hold
  • a method for deriving the weight component when the city distance is used will be described.
  • the city area distance ⁇ L L0 of IntraPredMode L0 and the city area distance ⁇ L L1 of IntraPredMode L1 are calculated from Equation (7).
  • the relative distance varies depending on the prediction direction of the two prediction modes.
  • the distance can be derived using Expression (6) or Expression (7) according to each prediction mode.
  • the table sizes of these distance tables may increase.
  • FIG. 17 shows the mapping of IntraPredModeLX used for distance table derivation.
  • IntraPredModeLX used for distance table derivation.
  • an example is shown in which a table of only the prediction mode corresponding to the prediction mode corresponding to the prediction mode and the DC prediction in 45 degrees is prepared, and other prediction angles are mapped closer to the prepared reference prediction mode. ing.
  • the index is mapped to the smaller one.
  • the prediction mode shown in “MappedIntraPredMode” is referred to from FIG. 17, and a distance table can be derived.
  • Equation (12) the relative distance for each pixel in the two prediction modes is calculated using Equation (12).
  • Equation (12) the final prediction signal at the pixel position n is represented by Equation (13).
  • the weight component is scaled in advance and converted to integer arithmetic, it can be expressed by Equation (14).
  • WM 1024
  • Offset 512
  • SHIFT 10.
  • FIGS. 18A and 18B show examples in which weight components using the one-sided Laplace distribution model in the present embodiment of the present invention are tabulated.
  • Other PuSizes can also be derived using Equation (5), Equation (8), Equation (10), and Equation (11). The above is the details of the intra prediction unit 109 according to the present embodiment of the present invention.
  • the syntax indicates the structure of encoded data (for example, encoded data 127 in FIG. 1) when the moving image decoding apparatus 4400 decodes moving image data.
  • the image encoding apparatus represented by the first embodiment encodes this encoded data using the same syntax structure.
  • FIG. 20 shows an example of syntax 2000 used by the image coding apparatus in FIG. Since the syntax 2000 is the same as that of the first embodiment, detailed description thereof is omitted.
  • FIG. 22A shows an example of the prediction unit syntax.
  • Pred_mode in the figure indicates the prediction type of the prediction unit.
  • MODE_INTRA indicates that the prediction type is intra prediction.
  • intra_split_flag is a flag indicating whether or not the prediction unit is further divided into four prediction units. When intra_split_flag is 1, a prediction unit is obtained by dividing a prediction unit into four in half in the vertical and horizontal sizes. When intra_split_flag is 0, the prediction unit is not divided.
  • Intra_luma_bipred_flag [i] is a flag indicating whether the prediction mode IntraPredMode applied to the prediction unit is a unidirectional intra prediction mode or a bidirectional intra prediction mode. i indicates the position of the divided prediction unit. When the intra_split_flag is 0, 0 is set, and when the intra_split_flag is 1, 0 to 3 are set. The flag is set with the value of IntraBiprededFlag of the prediction unit shown in FIGS.
  • intra_luma_bipred_flag [i] When intra_luma_bipred_flag [i] is 1, this indicates that the prediction unit is bi-directional intra prediction, and is information that identifies the used bi-directional intra prediction mode among a plurality of prepared bi-directional intra prediction modes.
  • Intra_luma_bipred_mode [i] is decoded.
  • intra_luma_bipred_mode [i] may be decoded in equal length according to the bidirectional intra prediction mode number IntraBiModeNum shown in FIG. 7, or may be decoded using a predetermined code table.
  • intra_luma_bipred_flag [i] When intra_luma_bipred_flag [i] is 0, it indicates that the prediction unit is unidirectional intra prediction, and predictive decoding is performed from adjacent blocks.
  • Prev_intra_luma_unipred_flag [i] is a flag indicating whether or not the prediction value MostProbable of the prediction mode calculated from the adjacent block and the intra prediction mode of the prediction unit are the same. Details of the MostProbable calculation method will be described later. When prev_intra_luma_unipred_flag [i] is 1, it indicates that the MostProbable and the intra prediction mode IntraPredMode are equal.
  • prev_intra_luma_unipred_flag [i] When prev_intra_luma_unipred_flag [i] is 0, it indicates that the MostProbable and the intra prediction mode IntraPredMode are different, and the information rem_intraprelum decoding that further specifies the intra prediction mode IntraPredMode other than MostProbable. . rem_intra_luma_unipred_mode [i] may be decoded in equal length according to the bidirectional intra prediction mode number IntraUniModeNum shown in FIG. 7, or may be decoded using a predetermined code table. From the intra prediction mode IntraPredMode, rem_intra_luma_unipred_mode [i] is calculated using Equation (16).
  • MostProbable which is a predicted value in the prediction mode.
  • MostProbable is calculated according to Equation (17).
  • Min (x, y) is a parameter for outputting the smaller one of the inputs x and y.
  • intraPredModeA and intraPredModeB indicate intra prediction modes of prediction units adjacent to the left and above the decoded prediction unit.
  • intraPredModeA and intraPredModeB are collectively expressed as intraPredModeN.
  • N is set to A or B.
  • a method of calculating intraPredModeN will be described using the flowchart shown in FIG. First, it is determined whether a coding tree unit to which an adjacent prediction unit belongs can be used (step S2301). If the coding tree unit cannot be used (NO in S2301), reference to intraPredModeN is not possible. “ ⁇ 1” indicating “” is set.
  • step S2302 it is next determined whether or not intra prediction is applied to the adjacent prediction unit (step S2302).
  • the adjacent prediction unit is not intra prediction (NO in S2302)
  • “2” meaning “Intra_DC” is set in intraPredModeN.
  • the adjacent prediction unit is intra prediction (YES in S2302)
  • the adjacent prediction unit is not bidirectional intra prediction, that is, in the case of unidirectional intra prediction (NO in S2303), the prediction mode IntraPredMode of the adjacent prediction unit is set in intraPredModeN.
  • intraPredModeN is calculated using Equation (18).
  • IntraUniModeNum is the number of unidirectional intra prediction modes determined by the size of the adjacent prediction unit, and an example thereof is shown in FIG.
  • MappedBi2Uni (List, idx)” is a table for converting the bidirectional intra prediction mode into the unidirectional intra prediction mode.
  • List is the unidirectional intra prediction mode of List0 (corresponding to IntraPredTypeL0 [] shown in FIGS. 9, 10, and 11) of the two unidirectional intra prediction modes constituting the bidirectional intra prediction mode.
  • FIG. 14 shows an example of the conversion table. The numerical values in the figure correspond to IntraPredMode shown in FIGS.
  • MappedMostProble () is a table for converting MostProbable, and an example is shown in FIG. 24.
  • luma_pred_mode_code_type [i] indicates the type of the prediction mode IntraPredMode applied to the prediction unit, where 0 (IntraUnifiedMostProb) is unidirectional intra prediction and the intra prediction mode is the same as MostProbable, 1 (IntraUnipre intrareprediction) The intra prediction mode is different from MostProbable, and 2 (IntraBipred) indicates a bidirectional intra prediction mode.
  • 0 IntraUnifiedMostProb
  • 1 IntraUnipre intrareprediction
  • the intra prediction mode is different from MostProbable
  • 2 IntraBipred indicates a bidirectional intra prediction mode.
  • FIG. 24 shows bidirectional intra prediction modes, 1 and unidirectional intra prediction (whether or not the same prediction mode as MostProbable).
  • FIG. 25 shows an example of assignment of the number of modes according to the meaning corresponding to luma_pred_mode_code_type, bin, and the mode configuration shown in FIG.
  • luma_pred_mode_code_type [i] When luma_pred_mode_code_type [i] is 0, the intra prediction mode is the MostProbable mode, so no further information decoding is necessary.
  • luma_pred_mode_code_type [i] is 1, information rem_intra_luma_unipred_mode [i] that specifies which mode other than MostProbable is the intra prediction mode IntraPredMode is decoded.
  • rem_intra_luma_unipred_mode [i] may be decoded in equal length according to the bidirectional intra prediction mode number IntraUniModeNum shown in FIG. 7, or may be decoded using a predetermined code table. From the intra prediction mode IntraPredMode, rem_intra_luma_unipred_mode [i] is calculated using Equation (16). Further, when luma_pred_mode_code_type [i] is 2, it indicates that the prediction unit is bidirectional intra prediction, and information that identifies the used bidirectional intra prediction mode among the prepared bidirectional intra prediction modes. Intra_luma_bipred_mode [i] is decoded.
  • intra_luma_bipred_mode [i] may be decoded in equal length according to the bidirectional intra prediction mode number IntraBiModeNum shown in FIG. 7, or may be decoded using a predetermined code table.
  • the above is the syntax configuration according to the present embodiment.
  • FIG. 22D Yet another example of the prediction unit syntax is shown in FIG. 22D.
  • pred_mode and intra_split_flag are the same as the syntax example described above, and thus description thereof is omitted.
  • Intra_bipred_flag is a flag indicating whether or not bidirectional intra prediction can be used in the decoding prediction unit. When intra_bipred_flag is 0, it indicates that bi-directional intra prediction is not used in the decoding prediction unit. Even when intra_split_flag is 1, that is, when the decoded prediction unit is further divided into four, bi-directional intra prediction is not used in all prediction units, and only uni-directional intra prediction is effective.
  • intra_bipred_flag When intra_bipred_flag is 1, it indicates that bidirectional intra prediction can be used in the decoding prediction unit. Even when intra_split_flag is 1, that is, when the decoded prediction unit is further divided into four, in all prediction units, bidirectional intra prediction can be selected in addition to unidirectional intra prediction.
  • the intra-bipred_flag is decoded as 0 to disable bi-directional intra prediction. Since the amount of code required for decoding can be reduced, the coding efficiency is improved.
  • FIG. 22E Still another example relating to the prediction unit syntax is shown in FIG. 22E.
  • intra_bipred_flag is a flag indicating whether or not bi-directional intra prediction can be used in the decoding prediction unit, and is the same as the above-described intra_bipred_flag, and thus the description thereof is omitted.
  • FIG. 26 shows an intra prediction unit 4408 (109) when adaptive reference pixel filtering is used. It differs from the intra prediction unit 4408 (109) shown in FIG. 6 in that a reference pixel filter unit 2601 is added.
  • the reference pixel filter unit 2601 receives the reference image signal 4419 (124) and the prediction mode 605, performs adaptive filter processing described later, and outputs a filtered reference image signal 2602.
  • the filtered reference image signal 2602 is input to the unidirectional intra predicted image generation unit 601 and the bidirectional intra predicted image generation unit 602.
  • the configuration and processing other than the reference pixel filter unit 2601 are the same as those of the intra prediction unit 4408 (109) shown in FIG.
  • the reference pixel filter unit 2601 determines whether or not to filter reference pixels used for intra prediction according to the reference pixel filter flag and the intra prediction mode included in the prediction mode 605.
  • the reference pixel filter flag is a flag indicating whether or not reference pixels are filtered when the intra prediction mode IntraPredMode is a value other than “Intra_DC”.
  • IntraPredMode is “Intra_DC”
  • the reference pixel is not filtered and the reference pixel filter flag is set to 0.
  • a filtered reference image signal 2602 is calculated by filtering shown in Expression (20).
  • p [x, y] indicates a reference pixel before filtering
  • pf [x, y] indicates a reference pixel in filter terms.
  • PuPartSize indicates the size (pixel) of the prediction unit.
  • FIG. 27A and 27B show a prediction unit syntax structure when performing adaptive reference pixel filtering.
  • FIG. 27A adds the syntax intra_luma_filter_flag [i] related to the adaptive reference pixel filter to FIG. 22A.
  • FIG. 27B adds syntax intra_luma_filter_flag [i] related to the adaptive reference pixel filter to FIG. 22C.
  • intra_luma_filter_flag [i] is further decoded when the intra prediction mode IntraPredMode [i] is other than Intra_DC. When the flag is 0, it indicates that the reference pixel is not filtered. Further, when intra_luma_filter_flag [i] is 1, it indicates that the reference pixel filtering is applied.
  • intra_luma_filter_flag [i] is decoded when the intra prediction mode IntraPredMode [i] is other than Intra_DC.
  • IntraPredMode [i] is 0 to 2
  • intra_luma_filter_flag [i ] Need not be decrypted. In this case, intra_luma_filter_flag [i] is set to 0.
  • intra_luma_filter_flag [i] described above may be added in the same meaning for the other syntax structures shown in FIGS. 22B, 22D, and 22E.
  • FIG. 30 shows a block diagram of the intra prediction unit 4408 (109) when combined with composite intra prediction. The difference is that a composite intra predicted image generation unit 2901, a selection switch 2902, and a decoded pixel buffer 3001 are added to the intra prediction unit 4408 (109) shown in FIG.
  • the selection switch 604 When the bidirectional intra prediction and the composite intra prediction are combined, first, in the selection switch 604, the unidirectional intra prediction image generation unit 601 or the bidirectional intra prediction image generation unit according to the prediction mode information controlled by the decoding control unit 4412. The output terminal of 602 is switched.
  • the output predicted image signal 4420 (126) is referred to as a direction predicted image signal 4420 (126).
  • the direction prediction image signal is input to the composite intra prediction image generation unit 2901, and a prediction image signal 4420 (126) in the composite intra prediction is generated.
  • the description of the composite intra predicted image generation unit 2901 will be described later.
  • the selection switch 2902 which one of the prediction image signal 4420 (126) and the direction prediction image signal in the composite intra prediction is used according to the composite intra prediction application flag in the prediction mode information controlled by the decoding control unit 4412. And the final predicted image signal 4420 (126) in the intra prediction unit 4408 (109) is output.
  • the composite intra prediction application flag is 1
  • the predicted image signal 4420 (126) output from the composite intra predicted image generation unit 2901 is the final predicted image signal 4420 (126).
  • the direction prediction image signal 4420 (126) is the prediction image signal 126 that is finally output.
  • the composite intra prediction image generation unit 2901 includes a pixel level prediction signal generation unit 3101 and a composite intra prediction calculation unit 3102.
  • the pixel level prediction signal generation unit 3101 predicts the prediction target pixel X from adjacent pixels and outputs a pixel level prediction signal 3103.
  • the adjacent pixel indicates the decoded image signal 4417.
  • the pixel level prediction signal 3103 (X) of the prediction target pixel is calculated using the number (21).
  • the coefficients related to A, B, and C may be other values.
  • the composite intra prediction calculation unit 3102 performs a weighted average of the direction prediction image signal 4420 (126) (X ′) and the pixel level prediction signal 3103 (X), and outputs a final prediction image signal 4420 (126) (P). To do. Specifically, Formula (22) is used.
  • the weighting factor W may be switched according to the position of the prediction pixel in the prediction unit.
  • a prediction image signal generated using unidirectional intra prediction and bidirectional intra prediction generates a prediction value from spatially adjacent reference pixels positioned on the left or above already encoded.
  • the absolute value of the prediction error tends to increase as the distance from the reference pixel increases. Therefore, the weighting coefficient of the direction prediction image signal 126 and the pixel level prediction signal 3103 is increased when the weight coefficient of the direction prediction image signal 126 is close to the reference pixel, and is decreased when the distance is far away, thereby improving the prediction accuracy. It becomes possible.
  • a prediction error signal is generated using an input image signal at the time of encoding.
  • the pixel level prediction signal 3103 becomes an input image signal, even if the spatial distance between the reference pixel position and the prediction pixel position is increased, the prediction of the pixel level prediction signal 3103 is compared with the direction prediction image signal 126. High accuracy.
  • the weighting coefficient of the direction prediction image signal 126 and the pixel level prediction signal 3103 is simply increased when the weight coefficient of the direction prediction image signal 126 is close to the reference pixel, and is decreased when the distance is small.
  • the prediction error is reduced, there is a problem that the prediction accuracy at the time of encoding and the prediction value at the time of local decoding are different and the prediction accuracy is lowered. Therefore, especially when the value of the quantization parameter is large, as the spatial distance between the reference pixel position and the predicted pixel position becomes large, the difference generated in the case of such an open loop is set by setting the value of W small. A decrease in coding efficiency due to the phenomenon can be suppressed.
  • FIG. 32A and 32B show a prediction unit syntax structure when performing composite intra prediction.
  • FIG. 32A is different from FIG. 22A in that a syntax combined_intra_pred_flag for switching presence / absence of composite intra prediction is added. This is equivalent to the above-described composite intra prediction application flag.
  • FIG. 32B adds a syntax combined_intra_pred_flag for switching presence / absence of composite intra prediction to FIG. 22C.
  • the selection switch 2902 shown in FIG. 30 is connected to the output terminal of the composite intra prediction image generation unit 2901.
  • When combined_intra_pred_flag is 0, the selection switch 2902 shown in FIG. 29 is connected to the output terminal of either the unidirectional intra prediction image generation unit 601 or the bidirectional intra prediction image generation unit 602 to which the selection switch 604 is connected. .
  • intra_luma_filter_flag [i] described above may be added in the same meaning for the other syntax structures shown in FIGS. 22B, 22D, and 22E.
  • the same or similar intra prediction unit as that of the video encoding device according to the first embodiment is included, the same or the same as the video encoding device according to the first embodiment or Similar effects can be obtained.
  • the video decoding device differs from the video decoding device according to the above-described fourth embodiment in the details of inverse orthogonal transform.
  • the same parts as those in the fourth embodiment are denoted by the same reference numerals, and different parts will be mainly described.
  • the moving picture coding apparatus corresponding to the moving picture decoding apparatus according to the present embodiment is as described in the second embodiment.
  • FIG. 45 is a block diagram showing a moving picture decoding apparatus according to the fifth embodiment.
  • a change from the moving picture decoding apparatus according to the fifth embodiment is that a conversion selection unit 4502 and a coefficient order restoration unit 4501 are added. Also, the internal structure of the inverse orthogonal transform unit 4404 is different.
  • the inverse orthogonal transform unit 4404 will be described with reference to FIG. Note that the inverse orthogonal transform unit 4404 has the same configuration as the inverse orthogonal transform unit 105 according to the second embodiment. Therefore, in the present embodiment, the conversion selection information 3303 in FIG. 35 is replaced with the conversion selection information 4504, the restored transform coefficient 120 is replaced with the restored transform coefficient 4415, and the restored prediction error signal 121 is replaced with the restored prediction error signal 4416. .
  • the conversion selection switch 3504 has a function of selecting the output terminal of the inverse quantization unit 4403 according to the input conversion selection information 4504.
  • the conversion selection information 4504 is one of information controlled by the decoding control unit 4412, and is set by the conversion selection unit 4502 in accordance with the prediction information 4421 (125).
  • the output terminal of the switch is connected to the first inverse orthogonal transform unit 3501.
  • the transformation selection information 4504 is the second orthogonal transformation
  • the output end is connected to the second inverse orthogonal transformation unit 3502.
  • the transform selection information 4504 is the Nth orthogonal transform
  • the output terminal is connected to the Nth inverse orthogonal transform unit 3503.
  • Prediction information 4421 (125) which is controlled by the decoding control unit 4412 and decoded by the entropy decoding unit 4402, is input to the transformation selection unit 4502.
  • the transform selection unit 4502 has a function of setting MapdTransformIdx information indicating which inverse orthogonal transform is to be used for which prediction mode.
  • FIG. 36 shows conversion selection information 4504 (MappedTransformIdx) in intra prediction.
  • FIG. 46 shows a block diagram of the coefficient order restoration unit 4501.
  • the coefficient order restoration unit 4501 has a function of performing reverse scan order conversion with the coefficient order control unit 3302 according to the second embodiment.
  • the coefficient order restoration unit 4501 includes a coefficient order selection switch 4604, a first coefficient forward / reverse transform unit 4601, a second coefficient forward / reverse transform unit 4602, and an Nth coefficient forward / reverse transform unit 4603.
  • the coefficient order selection switch 4604 has a function of switching the output terminal of the switch and the coefficient order inverse conversion units 4601 to 4603 in accordance with, for example, the mapped transform IDx shown in FIG.
  • the N types of coefficient forward / inverse transform units 4601 to 4603 have a function of inversely transforming one-dimensional data into two-dimensional data with respect to the quantized transform coefficient sequence 4503 decoded by the entropy decoding unit 4402.
  • two-dimensional data is converted into one-dimensional data using a zigzag scan.
  • the quantized transform coefficient obtained by performing quantization processing on the transform coefficient that has been subjected to orthogonal transform has the property that the tendency of generating non-zero transform coefficients in the block is biased. Have. The tendency of occurrence of this non-zero transform coefficient has different properties for each prediction direction of intra prediction. However, when different videos are encoded, the generation tendency of non-zero transform coefficients in the same prediction direction has a similar property. Therefore, when transforming two-dimensional data into one-dimensional data (2D-1D conversion), entropy coding is performed preferentially from transform coefficients at positions where the occurrence probability of non-zero transform coefficients is high, thereby encoding transform coefficients. It is possible to reduce information. Conversely, on the decoding side, it is necessary to restore the one-dimensional data to the two-dimensional data. Here, the raster scan is restored as a one-dimensional reference scan.
  • the coefficient order restoration unit 4501 may dynamically update the scan order in the 1D-2D conversion.
  • the configuration of the coefficient order restoration unit 4501 that performs such an operation is illustrated in FIG.
  • the coefficient order restoration unit 4501 includes an occurrence frequency counting unit 4701 and an updating unit 4702 in addition to the configuration of FIG. .., 4603 are the same except that the 1D-2D scan order is updated by the updating unit 4702.
  • the occurrence frequency counting unit 4701 creates a histogram 4704 of the number of occurrences of non-zero coefficients in each element of the quantized transform coefficient sequence 4503 for each prediction mode.
  • the occurrence frequency counting unit 4701 inputs the created histogram 4704 to the update unit 4702.
  • the update unit 4702 updates the coefficient order based on the histogram 4704 at a predetermined timing.
  • the timing is, for example, the timing when the coding process of the coding tree unit is finished, the timing when the coding process for one line in the coding tree unit is finished, or the like.
  • the update unit 4702 refers to the histogram 4704 and updates the coefficient order with respect to a prediction mode having an element in which the number of occurrences of non-zero coefficients is counted more than a threshold. For example, the update unit 4702 updates the prediction mode having an element in which the occurrence of a non-zero coefficient is counted 16 times or more. By providing a threshold value for the number of occurrences, the coefficient order is updated globally, so that it is difficult to converge to a local optimum solution.
  • the update unit 4702 sorts the elements in descending order of the occurrence frequency of the non-zero coefficient regarding the prediction mode to be updated. Sorting can be realized by existing algorithms such as bubble sort and quick sort. Then, the update unit 4702 inputs the update coefficient order 4703 indicating the order of the sorted elements to the coefficient order inverse transform units 4601 to 4603 corresponding to the prediction mode to be updated.
  • each inverse conversion unit performs 1D-2D conversion in accordance with the updated scan order.
  • the initial scan order of each 1D-2D conversion unit needs to be determined in advance.
  • the initial scanning order is the same as that of the coefficient order control unit 3302 of the moving picture coding apparatus shown in FIG. In this way, when the scan order is dynamically updated, the tendency of occurrence of non-zero coefficients in the quantized transform coefficients changes according to the effect of the predicted image properties, quantization information (quantization parameters), etc. In addition, stable and high encoding efficiency can be expected. Specifically, the generated code amount of run-length encoding in the entropy encoding unit 113 can be suppressed.
  • the syntax configuration in the present embodiment is the same as that in the fourth embodiment.
  • the conversion selection unit 4502 can select the mapped transform idx separately from the prediction information 4421.
  • information indicating which nine types of orthogonal transforms or inverse orthogonal transforms are used is set in the decoding control unit 4412 and used by the inverse orthogonal transform unit 4404.
  • FIG. 39 shows an example of syntax in the present embodiment.
  • Directional_transform_idx indicated in the syntax indicates information indicating which of N orthogonal transforms has been selected.
  • the same or similar inverse orthogonal transform unit as that of the video encoding device according to the second embodiment is included, and thus the same as the video encoding device according to the second embodiment. Or a similar effect can be obtained.
  • the video decoding device differs from the video decoding device according to the above-described fourth embodiment in the details of inverse orthogonal transform.
  • the same parts as those in the fourth embodiment are denoted by the same reference numerals, and different parts will be mainly described.
  • the moving picture encoding apparatus corresponding to the moving picture decoding apparatus according to the present embodiment is as described in the third embodiment.
  • JCTVC-B205_draft002 As an embodiment related to the inverse orthogonal transform unit 105, JCTVC-B205_draft002, 5.3.5.2 section “Rotational transformation process”, JCT-VC 2nd Meeting Geneva, July, 2010, may be combined.
  • FIG. 41 is a block diagram of inverse orthogonal transform section 4404 (105) according to the present embodiment.
  • the inverse orthogonal transform unit 4404 (105) has new processing units such as a first inverse rotation transform unit 4101, a second inverse rotation transform unit 4102, an Nth inverse rotation transform unit 4103, and an inverse discrete cosine transform unit 4104.
  • a selection switch 3504 is included.
  • the restored transform coefficient 4415 (120) input after the inverse quantization processing is input to the transform selection switch 3504.
  • the conversion selection switch 3504 sets the output end of the switch to one of the first reverse rotation conversion unit 4101, the second reverse rotation conversion unit 4102, and the Nth reverse rotation conversion unit 4103 according to the conversion selection information 4504 (3303). Connecting. Thereafter, the same inverse rotation transform unit 4101 to 4103 as that used in the orthogonal transform unit 102 shown in FIG. 40 is subjected to the inverse rotation transform process, and the result is output to the inverse discrete cosine transform unit 4104.
  • the inverse discrete cosine transform unit 4104 performs, for example, IDCT on the input signal to restore the restored prediction error signal 4416 (121).
  • orthogonal transform such as Hadamard transform or discrete sine transform may be used, or non-orthogonal transform may be used.
  • corresponding inverse transformation is performed in conjunction with the orthogonal transformation unit 102 shown in FIG.
  • FIG. 42 shows the syntax in the present embodiment.
  • the rotation_transform_idx shown in the syntax means the number of the rotation matrix to be used.
  • the same or similar inverse orthogonal transform unit as that of the image encoding device according to the third embodiment is included, and therefore the same or similar as that of the image encoding device according to the third embodiment. The effect of can be obtained.
  • encoding and decoding may be performed sequentially from the lower right to the upper left, or encoding and decoding may be performed so as to draw a spiral from the center of the screen toward the screen end.
  • encoding and decoding may be performed in order from the upper right to the lower left, or encoding and decoding may be performed so as to draw a spiral from the screen edge toward the center of the screen.
  • the prediction target block is a uniform block. It does not have to be a shape.
  • the prediction target block (prediction unit) size may be a 16 ⁇ 8 pixel block, an 8 ⁇ 16 pixel block, an 8 ⁇ 4 pixel block, a 4 ⁇ 8 pixel block, or the like. Also, it is not necessary to unify all the block sizes within one coding tree unit, and a plurality of different block sizes may be mixed.
  • the amount of codes for encoding or decoding the division information increases as the number of divisions increases. Therefore, it is desirable to select the block size in consideration of the balance between the code amount of the division information and the quality of the locally decoded image or the decoded image.
  • the color signal component has been described without distinguishing between the luminance signal and the color difference signal.
  • the same or different prediction methods may be used. If different prediction methods are used between the luminance signal and the chrominance signal, the prediction method selected for the chrominance signal can be encoded or decoded in the same manner as the luminance signal.
  • the orthogonal transformation process is different between the luminance signal and the color difference signal
  • the same or different orthogonal transformation methods may be used. If different orthogonal transformation methods are used between the luminance signal and the color difference signal, the orthogonal transformation method selected for the color difference signal can be encoded or decoded in the same manner as the luminance signal.
  • syntax elements that are not defined in the present invention can be inserted between the rows of the table shown in the syntax configuration, and other conditional branch descriptions are included. It does not matter.
  • the syntax table can be divided and integrated into a plurality of tables. Moreover, it is not always necessary to use the same term, and it may be arbitrarily changed depending on the form to be used.
  • each embodiment can realize highly efficient orthogonal transformation and inverse orthogonal transformation while alleviating the difficulty in hardware implementation and software implementation. Therefore, according to each embodiment, the encoding efficiency is improved, and the subjective image quality is also improved.
  • the instructions shown in the processing procedure shown in the above embodiment can be executed based on a program that is software.
  • a general-purpose computer system stores this program in advance, and by reading this program, it is also possible to obtain the same effects as those obtained by the video encoding device and video decoding device of the above-described embodiment. is there.
  • the instructions described in the above-described embodiments are, as programs that can be executed by a computer, magnetic disks (flexible disks, hard disks, etc.), optical disks (CD-ROM, CD-R, CD-RW, DVD-ROM, DVD). ⁇ R, DVD ⁇ RW, etc.), semiconductor memory, or a similar recording medium. As long as the recording medium is readable by the computer or the embedded system, the storage format may be any form.
  • the computer reads the program from the recording medium and causes the CPU to execute instructions described in the program based on the program
  • the computer is similar to the video encoding device and video decoding device of the above-described embodiment. Operation can be realized.
  • the computer acquires or reads the program, it may be acquired or read through a network.
  • the OS operating system
  • database management software database management software
  • MW middleware
  • a part of each process for performing may be executed.
  • the recording medium in the present invention is not limited to a medium independent of a computer or an embedded system, but also includes a recording medium in which a program transmitted via a LAN or the Internet is downloaded and stored or temporarily stored.
  • the program for realizing the processing of each of the above embodiments may be stored on a computer (server) connected to a network such as the Internet and downloaded to the computer (client) via the network.
  • the number of recording media is not limited to one, and when the processing in the present embodiment is executed from a plurality of media, it is included in the recording media in the present invention, and the configuration of the media may be any configuration.
  • the computer or the embedded system in the present invention is for executing each process in the present embodiment based on a program stored in a recording medium, and includes a single device such as a personal computer or a microcomputer, Any configuration such as a system in which apparatuses are connected to a network may be used.
  • the computer in the embodiment of the present invention is not limited to a personal computer, but includes an arithmetic processing device, a microcomputer, and the like included in an information processing device, and a device capable of realizing the functions in the embodiment of the present invention by a program, The device is a general term.
  • Prediction error Signal 118, transform coefficient, 119 quantized transform coefficient, 120 restored transform coefficient, 121 restored restoration error signal, 122 decoded image signal, 123 filtered image signal, 124 reference image signal, 125 prediction information , 126 ... predicted image signal, 127 ... encoded data, 601 ... unidirectional intra predicted image generation unit, 602 ... bidirectional intra predicted image raw 603 ... Prediction mode information setting unit, 604 ... Selection switch, 605 ... Prediction mode, 1301 ... First unidirectional intra prediction image generation unit, 1302 ... Second unidirectional intra prediction image generation unit, 1303 ... Weighted average unit DESCRIPTION OF SYMBOLS 1304 ... 1st prediction image signal, 1305 ... 2nd prediction image signal, 1901 ...
  • Image buffer 2000 ... Syntax, 2001 ... High level syntax, 2002 ... Slice level syntax, 2003 ... Coding tree level syntax, 2004 ... Sequence parameter set Syntax: 2005 ... Picture parameter set syntax, 2006: Slice header syntax, 2007 ... Slice data syntax, 2008 ... Coding tree unit syntax, 2009 ... Prediction unit syntax , 2010 ... Transform unit syntax, 2601 ... Reference pixel filter unit, 2602 ... Filtered reference image signal, 2901 ... Composite intra prediction image generation unit, 2902 ... Selection switch, 3001 ... Decoded pixel buffer, 3001 ... Decoded image buffer, 3002 Reference pixels, 3101 ... Pixel level prediction signal generation unit, 3102 ... Composite intra prediction calculation unit, 3103 ... Pixel level prediction signal, 3104 ...
  • Occurrence frequency counting section 3802 ... Coefficient order Update unit, 3803 ... Update coefficient order, 3804 ... Histogram, 4001 ... Rotation conversion unit, 4001 ... First rotation conversion unit, 4002 ... Second rotation conversion unit, 4003 ... Nth rotation conversion unit, 4004 ... Discrete cosine conversion unit, 4101: First reverse rotation conversion unit, 4102 ... Second reverse rotation conversion unit, 4103 ... Nth reverse rotation conversion unit, 4104 ... Inverse discrete cosine conversion unit, 4401 ... Input buffer, 4402 ... Entropy decoding unit, 4403 ... Reverse Quantization unit, 4404 ... inverse orthogonal transform unit, 4405 ... addition unit, 4406 ... loop filter, 4407 ...
  • reference image memory 4408 ... in La prediction unit, 4409 ... inter prediction unit, 4410 ... prediction selection switch, 4411 ... output buffer, 4412 ... decoding control unit, 4413 ... encoded data, 4414 ... quantized transform coefficient, 4415 ... restoration transform coefficient, 4416 ... restoration Prediction error signal, 4417 ... decoded image signal, 4418 ... filtered image signal, 4419 ... reference image signal, 4420 ... predicted image signal, 4421 ... prediction information, 4422 ... decoded image, 4501 ... coefficient order restoration unit, 4502 ... conversion selection. , 4503... Quantized transform coefficient sequence, 4504... Transform selection information, 4601, 4602, 4603... Coefficient order inverse transform section, 4701... Occurrence frequency count section, 4702.

Landscapes

  • Engineering & Computer Science (AREA)
  • Multimedia (AREA)
  • Signal Processing (AREA)
  • Compression Or Coding Systems Of Tv Signals (AREA)

Abstract

A moving picture encoding method for dividing an input image signal into pixel blocks represented by the depth of hierarchies according to a quadtree segmentation, generating a prediction error signal for the pixel blocks obtained by the division, and encoding transform coefficients, the method including: a step of setting a first prediction direction from a set of a plurality of prediction directions and generating a first predictive picture signal; a step of setting a second prediction direction different from the first prediction direction from the set of prediction directions, and generating a second predictive picture signal; a step of deriving a relative distance between a pixel to be predicted and a reference pixel in each of the first and second prediction directions, and deriving a difference value between the relative distances; a step of deriving a predetermined weight component according to the difference value; a step of weighted averaging a first unidirectional intra predictive picture and a second unidirectional intra predictive picture according to the weight component to generate a third predictive picture signal; a step of generating a prediction error signal from the third predictive picture signal; and a step of encoding the prediction error signal.

Description

動画像符号化方法及び動画像復号化方法Video encoding method and video decoding method
 本発明の実施形態は、動画像の符号化及び復号化における画面内予測方法、動画像符号化方法及び動画像復号化方法に関する。 Embodiments of the present invention relate to an intra-screen prediction method, a video encoding method, and a video decoding method in video encoding and decoding.
 近年、大幅に符号化効率を向上させた画像符号化方法がITU-TとISO/IECとの共同で、ITU-T REC. H.26及びISO/IEC 14496-10(以下、「H.264」という。)として勧告されている。H.264では空間領域(画素領域)での方向予測を取り入れることにより、ISO/IEC MPEG-1, 2及び4における画面内予測(以下、イントラ予測と呼ぶ)と比較して高い予測効率を実現している。H.264の拡張として、最大34種類の予測角度や予測方法を導入してイントラ予測を行うことによって、さらに符号化効率を向上する手法が提案されている。 In recent years, an image coding method with greatly improved coding efficiency has been jointly developed by ITU-T and ISO / IEC. H. 26 and ISO / IEC 14496-10 (hereinafter referred to as “H.264”). H. H.264 achieves higher prediction efficiency than in-screen prediction in ISO / IEC MPEG-1, 2 and 4 (hereinafter referred to as intra prediction) by incorporating direction prediction in the spatial region (pixel region). Yes. H. As an extension of H.264, a method for further improving the coding efficiency by introducing a maximum of 34 types of prediction angles and prediction methods and performing intra prediction has been proposed.
 しかしながら、非特許文献1では複数種類の予測モードの夫々について個別の予測角度で予測値を生成し、当該予測方向にコピーするため、画素ブロック内で滑らかに変化するような輝度勾配を持ったテクスチャなどの映像や、グラデーションのある映像を効率よく予測することが出来ず、予測誤差が増大する場合がある。 However, in Non-Patent Document 1, since a prediction value is generated at an individual prediction angle for each of a plurality of types of prediction modes and copied in the prediction direction, a texture having a luminance gradient that smoothly changes within a pixel block. Such a video or a video with gradation cannot be predicted efficiently, and the prediction error may increase.
 従って、本実施形態は、符号化効率を向上可能な予測画像生成装置を含んだ動画像符号化装置及び動画像復号化装置を提供することを目的とする。 Therefore, an object of the present embodiment is to provide a moving image encoding device and a moving image decoding device including a prediction image generating device capable of improving encoding efficiency.
 実施形態によれば、動画像符号化方法は、入力画像信号を四分木分割に従って、階層の深さで表現される画素ブロックに分割し、これら分割した画素ブロックに対して予測誤差信号を生成し、変換係数を符号化する動画像符号化方法において、複数の予測方向セットの中から、第一の予測方向を設定して第一予測画像信号を作成するステップと、前記複数の予測方向セットの中から、第一の予測方向とは異なる第二の予測方向を設定して第二予測画像信号を作成するステップと、第一及び第二の予測方向のそれぞれに関して、予測対象の画素と参照画素との間の相対距離を導出し、該相対距離の差分値を導出するステップと、前記差分値に応じて、予め定められた重み成分を導出するステップと、前記重み成分に従って、第一単方向イントラ予測画像と第二単方向イントラ予測画像を重み付き平均し、第三予測画像信号を生成するステップと、前記第三予測画像信号から予測誤差信号を生成するステップと、前記予測誤差信号を符号化するステップと、を具備する。 According to the embodiment, the moving image encoding method divides an input image signal into pixel blocks represented by hierarchical depths according to quadtree division, and generates prediction error signals for these divided pixel blocks. Then, in the moving picture encoding method for encoding the transform coefficient, a step of setting a first prediction direction from a plurality of prediction direction sets to create a first prediction image signal, and the plurality of prediction direction sets A step of creating a second prediction image signal by setting a second prediction direction different from the first prediction direction, and a pixel to be predicted and reference for each of the first and second prediction directions Deriving a relative distance from the pixel, deriving a difference value of the relative distance, deriving a predetermined weight component according to the difference value, and according to the weight component, a first unit Direction A step of weighted averaging the prediction image and the second unidirectional intra prediction image to generate a third prediction image signal, a step of generating a prediction error signal from the third prediction image signal, and encoding the prediction error signal And the step of performing.
 実施形態によれば、動画像復号化方法は、入力画像信号を四分木分割に従って、階層の深さで表現される画素ブロックに分割し、これら分割した画素ブロックに対して復号処理を行う動画像復号化方法において、複数の予測方向セットの中から、復号された第一の予測方向を設定して第一予測画像信号を作成するステップと、前記複数の予測方向セットの中から、前記第一の予測方向と異なる復号された第二の予測方向を設定して第二予測画像信号を作成するステップと、前記第一及び第二の予測方向のそれぞれに関して、予測対象の画素と参照画素との間の相対距離を導出し、該相対距離の差分値を導出するステップと、前記差分値に応じて、予め定められた重み成分を導出するステップと、前記重み成分に従って、第一単方向イントラ予測画像と第二単方向イントラ予測画像を重み付き平均し、第三予測画像信号を生成するステップと、前記第三予測画像信号から復号画像信号を生成するステップと、を具備する。 According to the embodiment, the moving picture decoding method divides an input image signal into pixel blocks expressed by hierarchical depth according to quadtree division, and performs a decoding process on the divided pixel blocks. In the image decoding method, a step of creating a first predicted image signal by setting a decoded first prediction direction from among a plurality of prediction direction sets; and Creating a second predicted image signal by setting a decoded second prediction direction different from the one prediction direction; and for each of the first and second prediction directions, a prediction target pixel and a reference pixel, Deriving a relative distance between the two, deriving a difference value of the relative distance, deriving a predetermined weight component according to the difference value, and according to the weight component, a first unidirectional intra The measurement image and the second unidirectional intra-prediction image and the weighted average comprises the steps of generating a third predictive image signal, and generating a decoded image signal from said third prediction image signal.
第1の実施形態に係る動画像符号化装置を例示するブロック図。1 is a block diagram illustrating a moving image encoding apparatus according to a first embodiment. 画素ブロックの予測符号化順の説明図。Explanatory drawing of the prediction encoding order of a pixel block. 画素ブロックサイズの一例の説明図。Explanatory drawing of an example of pixel block size. 画素ブロックサイズの別の例の説明図。Explanatory drawing of another example of pixel block size. 画素ブロックサイズの別の例の説明図。Explanatory drawing of another example of pixel block size. コーディングツリーユニットにおける画素ブロックの一例の説明図。Explanatory drawing of an example of the pixel block in a coding tree unit. コーディングツリーユニットにおける画素ブロックの別の例の説明図。Explanatory drawing of another example of the pixel block in a coding tree unit. コーディングツリーユニットにおける画素ブロックの別の例の説明図。Explanatory drawing of another example of the pixel block in a coding tree unit. コーディングツリーユニットにおける画素ブロックの別の例の説明図。Explanatory drawing of another example of the pixel block in a coding tree unit. 第1の実施形態に係る高速モード判定時の予測モード候補数の説明図。Explanatory drawing of the prediction mode candidate number at the time of the high speed mode determination which concerns on 1st Embodiment. (a)はイントラ予測モードの説明図であり、(b)はイントラ予測モードの参照画素と予測画素の説明図であり、(c)はイントラ予測モードの水平予測モードの説明図であり、(d)はイントラ予測モードの直交右下予測モードの説明図。(A) is explanatory drawing of intra prediction mode, (b) is explanatory drawing of the reference pixel and prediction pixel of intra prediction mode, (c) is explanatory drawing of the horizontal prediction mode of intra prediction mode, ( d) Explanatory drawing of the orthogonal lower right prediction mode of intra prediction mode. 第1の実施形態に係るイントラ予測部を例示するブロック図。The block diagram which illustrates the intra prediction part concerning a 1st embodiment. 第1の実施形態に係る単方向イントラ予測数と双方向イントラ予測数の説明図。Explanatory drawing of the number of unidirectional intra predictions and the number of bidirectional intra predictions which concern on 1st Embodiment. 第1の実施形態に係る予測方向を例示する説明図。Explanatory drawing which illustrates the prediction direction which concerns on 1st Embodiment. 第1の実施形態に係る、予測モードと予測タイプと双方向イントラ予測、単方向イントラ予測の関係を例示するテーブル図。The table figure which illustrates the relationship of prediction mode, prediction type, bidirectional | two-way intra prediction, and unidirectional intra prediction based on 1st Embodiment. 第1の実施形態に係る、予測モードと予測タイプと双方向イントラ予測、単方向イントラ予測の関係を例示するテーブル図。The table figure which illustrates the relationship of prediction mode, prediction type, bidirectional | two-way intra prediction, and unidirectional intra prediction based on 1st Embodiment. 第1の実施形態に係る、予測モードと予測タイプと双方向イントラ予測、単方向イントラ予測の関係を例示するテーブル図。The table figure which illustrates the relationship of prediction mode, prediction type, bidirectional | two-way intra prediction, and unidirectional intra prediction based on 1st Embodiment. 第1の実施形態に係る、予測角度の指標と予測画像生成の際の予測角度との対応を例示するテーブル図。The table figure which illustrates a response | compatibility with the parameter | index of the prediction angle based on 1st Embodiment, and the prediction angle in the case of prediction image generation. 第1の実施形態に係る、双方向イントラ予測画像生成部のブロック図。The block diagram of the bidirectional | two-way intra estimated image generation part based on 1st Embodiment. 第1の実施形態に係る、双方向イントラ予測と2つの単方向イントラ予測との対応を例示するテーブル図。The table figure which illustrates a response | compatibility with bidirectional | two-way intra prediction and two unidirectional intra prediction based on 1st Embodiment. 第1の実施形態に係る、市街地距離の算出方法の一例を示すブロック図。The block diagram which shows an example of the calculation method of a city area distance based on 1st Embodiment. 第1の実施形態に係る、市街地距離の算出方法の別の一例を示すブロック図。The block diagram which shows another example of the calculation method of the city area distance based on 1st Embodiment. 第1の実施形態に係る、市街地距離の算出方法の別の一例を示すブロック図。The block diagram which shows another example of the calculation method of the city area distance based on 1st Embodiment. 第1の実施形態に係る、予測モードと予測画素位置の距離との関係を例示するテーブル。The table which illustrates the relationship between the prediction mode and the distance of a prediction pixel position based on 1st Embodiment. 第1の実施形態に係る、予測モードと距離テーブルのマッピングを例示するテーブル。The table which illustrates the mapping of prediction mode and a distance table based on 1st Embodiment. 第1の実施形態に係る、相対距離と重み成分との関係を例示するテーブル。The table which illustrates the relationship between relative distance and a weight component based on 1st Embodiment. 第1の実施形態に係る、相対距離と重み成分との関係を例示する別のテーブル。6 is another table illustrating the relationship between the relative distance and the weight component according to the first embodiment. 第1の実施形態に係る、イントラ予測部の別の実施形態を示すブロック図。The block diagram which shows another embodiment of the intra estimation part based on 1st Embodiment. シンタクス構造の説明図。Explanatory drawing of a syntax structure. スライスヘッダーシンタクスの説明図。Explanatory drawing of a slice header syntax. プレディクションユニットシンタクスの一例を示す説明図。Explanatory drawing which shows an example of a prediction unit syntax. プレディクションユニットシンタクスの別の例を示す説明図。Explanatory drawing which shows another example of a prediction unit syntax. プレディクションユニットシンタクスの別の例を示す説明図。Explanatory drawing which shows another example of a prediction unit syntax. プレディクションユニットシンタクスの別の例を示す説明図。Explanatory drawing which shows another example of a prediction unit syntax. プレディクションユニットシンタクスの別の例を示す説明図。Explanatory drawing which shows another example of a prediction unit syntax. intraPredModeNの計算方法の一例を示すフローチャート。The flowchart which shows an example of the calculation method of intraPredModeN. 予測モードを予測する際の関係を示すテーブル。The table which shows the relationship at the time of predicting prediction mode. 予測モードを予測する際の関係の別の例を示すテーブル。The table which shows another example of the relationship at the time of predicting prediction mode. 第1の実施形態に係る、イントラ予測部の第1の変形例を示すブロック図。The block diagram which shows the 1st modification of the intra estimation part based on 1st Embodiment. 第1の実施形態に係る、第1の変形例でのプレディクションユニットシンタクスの一例を示す説明図。Explanatory drawing which shows an example of the prediction unit syntax in the 1st modification based on 1st Embodiment. 第1の実施形態に係る、第1の変形例でのプレディクションユニットシンタクスの別の例を示す説明図。Explanatory drawing which shows another example of the prediction unit syntax in the 1st modification based on 1st Embodiment. 画素レベルの予測値生成方法の一例を示す説明図。Explanatory drawing which shows an example of the predicted value generation method of a pixel level. 第1の実施形態に係る、イントラ予測部の別の一例を示すブロック図。The block diagram which shows another example of the intra estimation part based on 1st Embodiment. 第1の実施形態に係る、イントラ予測部の別の一例を示すブロック図。The block diagram which shows another example of the intra estimation part based on 1st Embodiment. 第1の実施形態に係る、複合イントラ予測画像生成部の一例を示すブロック図。The block diagram which shows an example of the composite intra estimated image generation part based on 1st Embodiment. 第1の実施形態に係る、プレディクションユニットシンタクス別の一例を示す説明図。Explanatory drawing which shows an example according to Prediction unit syntax based on 1st Embodiment. 第1の実施形態に係る、プレディクションユニットシンタクスの別の一例を示す説明図。Explanatory drawing which shows another example of the prediction unit syntax based on 1st Embodiment. 第2の実施形態に係る、動画像符号化装置を例示するブロック図。The block diagram which illustrates the moving picture coding device concerning a 2nd embodiment. 第2の実施形態に係る、直交変換部を例示するブロック図。The block diagram which illustrates the orthogonal transformation part based on 2nd Embodiment. 第2の実施形態に係る、逆直交変換部を例示するブロック図。The block diagram which illustrates the inverse orthogonal transformation part based on 2nd Embodiment. 第2の実施形態に係る、予測モードと変換インデックスの関係を示すテーブル。The table which shows the relationship between prediction mode and a conversion index based on 2nd Embodiment. 第2の実施形態に係る、係数順制御部を例示するブロック図。The block diagram which illustrates the coefficient order control part concerning a 2nd embodiment. 第2の実施形態に係る、別の係数順制御部を例示するブロック図。The block diagram which illustrates another coefficient order control part concerning a 2nd embodiment. 第2の実施形態に係る、トランスフォームユニットシンタクスの一例を示す説明図。Explanatory drawing which shows an example of the transform unit syntax based on 2nd Embodiment. 第3の実施形態に係る、直交変換部の別の一例を示すブロック図。The block diagram which shows another example of the orthogonal transformation part based on 3rd Embodiment. 第3の実施形態に係る、逆直交変換部の一例を示すブロック図。The block diagram which shows an example of the inverse orthogonal transformation part based on 3rd Embodiment. 第3の実施形態に係る、トランスフォームユニットシンタクスの一例を示す説明図。Explanatory drawing which shows an example of the transform unit syntax based on 3rd Embodiment. 第1の実施形態に係る、単方向イントラ予測モードと予測タイプ、予測角度指標の一例を示す説明図。Explanatory drawing which shows an example of the unidirectional intra prediction mode, prediction type, and prediction angle parameter | index based on 1st Embodiment. 第4の実施形態に係る、動画像復号化装置の一例を示すブロック図。The block diagram which shows an example of the moving image decoding apparatus based on 4th Embodiment. 第5の実施形態に係る、動画像復号化装置の一例を示すブロック図。The block diagram which shows an example of the moving image decoding apparatus based on 5th Embodiment. 第5の実施形態に係る、係数順復元部を例示するブロック図。The block diagram which illustrates the coefficient order restoration part concerning a 5th embodiment. 第5の実施形態に係る、係数順復元部の別の一例を示すブロック図。The block diagram which shows another example of the coefficient order decompression | restoration part based on 5th Embodiment.
 以下、図面を参照して、各実施形態に係る動画像符号化装置及び動画像復号化装置について詳細に説明する。なお、以降の説明において、「画像」という用語は、「映像」「画素」「画像信号」、「画像データ」などの用語として適宜読み替えることができる。また、以下の実施の形態では、同一の番号を付した部分については同様の動作を行うものとして、重ねての説明を省略する。 
 (第1の実施形態) 
 第1の実施形態は画像符号化装置に関する。本実施形態に係る画像符号化装置に対応する動画像復号化装置は、第4の実施形態において説明する。この画像符号化装置は、LSI(Large-Scale Integration)チップやDSP(Digital Signal Processor)、FPGA(Field Programmable Gate Array)などのハードウェアにより実現可能である。また、この画像符号化装置は、コンピュータに画像符号化プログラムを実行させることによっても実現可能である。
Hereinafter, with reference to the drawings, a video encoding device and a video decoding device according to each embodiment will be described in detail. In the following description, the term “image” can be appropriately read as terms such as “video”, “pixel”, “image signal”, and “image data”. Further, in the following embodiments, the same numbered portions are assumed to perform the same operation, and repeated description is omitted.
(First embodiment)
The first embodiment relates to an image encoding device. A moving picture decoding apparatus corresponding to the picture encoding apparatus according to the present embodiment will be described in a fourth embodiment. This image encoding device can be realized by hardware such as an LSI (Large-Scale Integration) chip, a DSP (Digital Signal Processor), or an FPGA (Field Programmable Gate Array). The image encoding apparatus can also be realized by causing a computer to execute an image encoding program.
 図1に示すように、本実施形態に係る画像符号化装置は、減算部101、直交変換部102、量子化部103、逆量子化部104、逆直交変換部105、加算部106、ループフィルタ107、参照画像メモリ108、イントラ予測部109、インター予測部110、予測選択スイッチ111、予測選択部112、エントロピー符号化部113、出力バッファ114、及び符号化制御部115を含む。 As shown in FIG. 1, the image coding apparatus according to the present embodiment includes a subtraction unit 101, an orthogonal transformation unit 102, a quantization unit 103, an inverse quantization unit 104, an inverse orthogonal transformation unit 105, an addition unit 106, and a loop filter. 107, a reference image memory 108, an intra prediction unit 109, an inter prediction unit 110, a prediction selection switch 111, a prediction selection unit 112, an entropy encoding unit 113, an output buffer 114, and an encoding control unit 115.
 図1の画像符号化装置は、入力画像信号116を構成する各フレームまたは各フィールドを複数の画素ブロックに分割し、これら分割した画素ブロックに対して予測符号化を行って、符号化データ127を出力する。以降の説明では、簡単化のために、図2Aに示されるように左上から右下に向かって画素ブロックの予測符号化が行われることを仮定する。図2Aでは、符号化処理対象のフレームfにおいて、符号化対象画素ブロックcよりも左側及び上側に符号化済み画素ブロックpが位置している。 The image coding apparatus in FIG. 1 divides each frame or each field constituting the input image signal 116 into a plurality of pixel blocks, performs predictive coding on the divided pixel blocks, and generates coded data 127. Output. In the following description, for the sake of simplicity, it is assumed that pixel blocks are predictively encoded from the upper left to the lower right as shown in FIG. 2A. In FIG. 2A, the encoded pixel block p is located on the left side and the upper side of the encoding target pixel block c in the encoding processing target frame f.
 ここで、画素ブロックは、例えば、M×Nサイズのブロック(N及びMは自然数)、コーディングツリーユニット、マクロブロック、サブブロック、1画素などの画像を処理する単位を指す。なお、以降の説明では、画素ブロックをコーディングツリーユニットの意味で基本的に使用するが、説明を適宜読み替えることにより画素ブロックを上述した意味で解釈することも可能である。コーディングツリーユニットは、典型的には、例えば図2Bに示す16×16画素ブロックであるが、図2Cに示す32×32画素ブロック、図3Dに示す64×64画素ブロックであってもよいし、図示しない8×8画素ブロック、4×4画素ブロックであってもよい。また、コーディングツリーユニットは必ずしも正方形である必要はない。以下、入力画像信号116の符号化対象ブロックもしくはコーディングツリーユニットを「予測対象ブロック」と称することもある。また、符号化単位には、コーディングツリーユニットのような画素ブロックに限らず、フレームまたはフィールド、スライス、或いはこれらの組み合わせを用いることができる。 Here, the pixel block refers to a unit for processing an image such as an M × N size block (N and M are natural numbers), a coding tree unit, a macro block, a sub block, and one pixel. In the following description, the pixel block is basically used in the meaning of the coding tree unit. However, the pixel block can be interpreted in the above-described meaning by appropriately replacing the description. The coding tree unit is typically a 16 × 16 pixel block shown in FIG. 2B, for example, but may be a 32 × 32 pixel block shown in FIG. 2C or a 64 × 64 pixel block shown in FIG. 3D. It may be an 8 × 8 pixel block (not shown) or a 4 × 4 pixel block. Also, the coding tree unit need not necessarily be square. Hereinafter, the encoding target block or coding tree unit of the input image signal 116 may be referred to as a “prediction target block”. The coding unit is not limited to a pixel block such as a coding tree unit, and a frame, a field, a slice, or a combination thereof can be used.
 図3A~Dは、コーディングツリーユニットの具体例を示した図である。図3Aは、コーディングツリーユニットのサイズが64×64(N=32)の場合の例を示している。ここでNは、基準となるコーディングツリーユニットのサイズを表しており、分割された場合のサイズをNと定義し、分割されない場合を2Nと定義する。コーディングツリーユニットは四分木構造を持ち、分割された場合は、4つの画素ブロックに対してZスキャン順でインデックスが付される。図3Bに、図3Aの64x64画素ブロックを四分木分割した例を示す。図中に示される番号がZスキャンの順番を表している。また、コーディングツリーユニットの1つの四分木のインデックス内でさらに四分木分割することが可能である。分割の深さをDepthで定義する。つまり、図3AはDepth=0の例を示している。図3CにDepth=1の場合の32×32(N=16)サイズのコーディングツリーユニットの例を示す。このようなコーディングツリーユニットの最も大きいユニットをラージコーディングツリーユニットと呼び、この単位で入力画像信号がラスタースキャン順に符号化される。 3A to 3D are diagrams showing specific examples of coding tree units. FIG. 3A shows an example where the size of the coding tree unit is 64 × 64 (N = 32). Here, N represents the size of the reference coding tree unit. The size when divided is defined as N, and the case where it is not divided is defined as 2N. The coding tree unit has a quadtree structure, and when divided, the four pixel blocks are indexed in the Z-scan order. FIG. 3B shows an example in which the 64 × 64 pixel block in FIG. 3A is divided into quadtrees. The numbers shown in the figure represent the Z scan order. Further, it is possible to further perform quadtree division within the index of one quadtree of the coding tree unit. The depth of division is defined by Depth. That is, FIG. 3A shows an example in which Depth = 0. FIG. 3C shows an example of a 32 × 32 (N = 16) size coding tree unit in the case of Depth = 1. A unit having the largest coding tree unit is called a large coding tree unit, and an input image signal is encoded in the order of raster scanning in this unit.
 図1の画像符号化装置は、符号化制御部115から入力される符号化パラメータに基づいて、画素ブロックに対するイントラ予測(画面内予測、フレーム内予測などとも称される)またはインター予測(画面間予測、フレーム間予測、動き補償予測などとも称される)を行って、予測画像信号126を生成する。この画像符号化装置は、画素ブロック(入力画像信号116)と予測画像信号126との間の予測誤差信号117を直交変換及び量子化し、エントロピー符号化を行って符号化データ127を生成して出力する。 The image encoding apparatus in FIG. 1 performs intra prediction (also referred to as intra-frame prediction, intra-frame prediction, etc.) or inter prediction (inter-screen prediction) for a pixel block based on the encoding parameter input from the encoding control unit 115. Prediction, inter-frame prediction, motion compensation prediction, etc.) is performed to generate the predicted image signal 126. This image coding apparatus orthogonally transforms and quantizes the prediction error signal 117 between the pixel block (input image signal 116) and the predicted image signal 126, performs entropy coding, generates coded data 127, and outputs it. To do.
 図1の画像符号化装置は、ブロックサイズ及び予測画像信号126の生成方法の異なる複数の予測モードを選択的に適用して符号化を行う。予測画像信号126の生成方法は、大別すると、符号化対象フレーム内で予測を行うイントラ予測と、時間的に異なる1つまたは複数の参照フレームを用いて予測を行うインター予測との2種類である。 The image encoding device in FIG. 1 performs encoding by selectively applying a plurality of prediction modes having different block sizes and generation methods of the predicted image signal 126. The generation method of the predicted image signal 126 can be broadly divided into two types: intra prediction in which prediction is performed within the encoding target frame and inter prediction in which prediction is performed using one or a plurality of reference frames that are temporally different. is there.
 以下、図1の画像符号化装置に含まれる各要素を説明する。 
 減算部101は、入力画像信号116の符号化対象ブロックから、対応する予測画像信号126を減算して予測誤差信号117を得る。減算部101は、予測誤差信号117を直交変換部102に入力する。
Hereinafter, each element included in the image encoding device in FIG. 1 will be described.
The subtraction unit 101 subtracts the corresponding prediction image signal 126 from the encoding target block of the input image signal 116 to obtain a prediction error signal 117. The subtraction unit 101 inputs the prediction error signal 117 to the orthogonal transformation unit 102.
 直交変換部102は、減算部101からの予測誤差信号117に対して、例えば離散コサイン変換(DCT)のような直交変換を行い、変換係数118を得る。直交変換部102は、変換係数118を量子化部103に入力する。 The orthogonal transform unit 102 performs an orthogonal transform such as discrete cosine transform (DCT) on the prediction error signal 117 from the subtraction unit 101 to obtain a transform coefficient 118. The orthogonal transform unit 102 inputs the transform coefficient 118 to the quantization unit 103.
 量子化部103は、直交変換部102からの変換係数に対して量子化を行い、量子化変換係数119を得る。具体的には、量子化部103は、符号化制御部115によって指定される量子化パラメータ、量子化マトリクスなどの量子化情報に従って量子化を行う。量子化パラメータは、量子化の細かさを示す。量子化マトリクスは、量子化の細かさを変換係数の成分毎に重み付けするために使用される。量子化部103は、量子化変換係数119をエントロピー符号化部113及び逆量子化部104に入力する。 The quantization unit 103 performs quantization on the transform coefficient from the orthogonal transform unit 102 to obtain a quantized transform coefficient 119. Specifically, the quantization unit 103 performs quantization according to quantization information such as a quantization parameter and a quantization matrix specified by the encoding control unit 115. The quantization parameter indicates the fineness of quantization. The quantization matrix is used for weighting the fineness of quantization for each component of the transform coefficient. The quantization unit 103 inputs the quantized transform coefficient 119 to the entropy encoding unit 113 and the inverse quantization unit 104.
 エントロピー符号化部113は、量子化部103からの量子化変換係数119、予測選択部112からの予測情報125、符号化制御部115によって指定される量子化情報などの様々な符号化パラメータに対してエントロピー符号化(例えば、ハフマン符号化、算術符号化など)を行い、符号化データを生成する。なお、符号化パラメータとは、予測情報125、変換係数に関する情報、量子化に関する情報、などの復号に必要となるパラメータである。例えば、符号化制御部115が内部メモリ(図示しない)を持ち、このメモリに符号化パラメータが保持され、予測対象ブロックを符号化する際に隣接する既に符号化済みの画素ブロックの符号化パラメータを用いる構成としてもよい。例えば、H.264のイントラ予測では符号化済みの隣接ブロックの予測モード情報から、予測対象ブロックの予測モードの予測値を導出することが可能である。 The entropy encoding unit 113 performs various encoding parameters such as the quantized transform coefficient 119 from the quantization unit 103, the prediction information 125 from the prediction selection unit 112, and the quantization information specified by the encoding control unit 115. Entropy encoding (for example, Huffman encoding, arithmetic encoding, etc.) is performed to generate encoded data. The encoding parameter is a parameter necessary for decoding, such as prediction information 125, information on transform coefficients, information on quantization, and the like. For example, the encoding control unit 115 has an internal memory (not shown), the encoding parameter is held in this memory, and the encoding parameter of an already encoded pixel block adjacent when encoding the prediction target block is stored. It is good also as a structure to use. For example, H.M. In the H.264 intra prediction, the prediction value of the prediction mode of the prediction target block can be derived from the prediction mode information of the encoded adjacent block.
 エントロピー符号化部113によって生成された符号化データは、例えば多重化を経て出力バッファ114に一時的に蓄積され、符号化制御部115が管理する適切な出力タイミングに従って符号化データ127として出力される。符号化データ127は、例えば、図示しない蓄積系(蓄積メディア)または伝送系(通信回線)へ出力される。 The encoded data generated by the entropy encoding unit 113 is temporarily accumulated, for example, in the output buffer 114 through multiplexing, and is output as encoded data 127 according to an appropriate output timing managed by the encoding control unit 115. . The encoded data 127 is output to, for example, a storage system (storage medium) or a transmission system (communication line) not shown.
 逆量子化部104は、量子化部103からの量子化変換係数119に対して逆量子化を行い、復元変換係数120を得る。具体的には、逆量子化部104は、量子化部103において使用された量子化情報に従って逆量子化を行う。量子化部103において使用された量子化情報は、符号化制御部115の内部メモリからロードされる。逆量子化部104は、復元変換係数120を逆直交変換部105に入力する。 The inverse quantization unit 104 performs inverse quantization on the quantized transform coefficient 119 from the quantization unit 103 to obtain a restored transform coefficient 120. Specifically, the inverse quantization unit 104 performs inverse quantization according to the quantization information used in the quantization unit 103. The quantization information used in the quantization unit 103 is loaded from the internal memory of the encoding control unit 115. The inverse quantization unit 104 inputs the restored transform coefficient 120 to the inverse orthogonal transform unit 105.
 逆直交変換部105は、逆量子化部104からの復元変換係数120に対して、例えば逆離散コサイン変換などのような直交変換部102において行われた直交変換に対応する逆直交変換を行い、復元予測誤差信号121を得る。逆直交変換部105は、復元予測誤差信号121を加算部106に入力する。 The inverse orthogonal transform unit 105 performs an inverse orthogonal transform corresponding to the orthogonal transform performed in the orthogonal transform unit 102 such as an inverse discrete cosine transform on the restored transform coefficient 120 from the inverse quantization unit 104, A restoration prediction error signal 121 is obtained. The inverse orthogonal transform unit 105 inputs the restored prediction error signal 121 to the addition unit 106.
 加算部106は、復元予測誤差信号121と、対応する予測画像信号126とを加算し、局所的な復号画像信号122を生成する。復号画像信号122は、ループフィルタ107に入力される。ループフィルタ107は、入力された復号画像信号122にデブロッキングフィルタやウィナーフィルタなどを施し、被フィルタ画像信号123を生成する。生成した被フィルタ画像信号123は、参照画像メモリ108へと入力される。 The addition unit 106 adds the restored prediction error signal 121 and the corresponding prediction image signal 126 to generate a local decoded image signal 122. The decoded image signal 122 is input to the loop filter 107. The loop filter 107 performs a deblocking filter, a Wiener filter, or the like on the input decoded image signal 122 to generate a filtered image signal 123. The generated filtered image signal 123 is input to the reference image memory 108.
 参照画像メモリ108は、メモリに局部復号後の被フィルタ画像信号123を蓄積しており、イントラ予測部109及びインター予測部110によって必要に応じて予測画像を生成する際に、参照画像信号124として都度参照される。 The reference image memory 108 stores the filtered image signal 123 after local decoding in the memory, and when the predicted image is generated as necessary by the intra prediction unit 109 and the inter prediction unit 110, the reference image signal 124 is used. Referenced each time.
 イントラ予測部109は、参照画像メモリ108に保存されている参照画像信号124を利用してイントラ予測を行う。例えば、H.264では、予測対象ブロックに隣接する符号化済みの参照画素値を利用して、垂直方向、水平方向などの予測方向に沿って画素補填(コピーまたは補間後にコピー)を行うことによってイントラ予測画像を生成する。図5(a)にH.264におけるイントラ予測の予測方向を示す。また、図5(b)にH.264における参照画素と符号化対象画素との配置関係を示す。図5(c)はモード1(水平予測)の予測画像生成方法を示しており、図5(d)はモード4(対角右下予測)の予測画像生成方法を示している。 The intra prediction unit 109 performs intra prediction using the reference image signal 124 stored in the reference image memory 108. For example, H.M. In H.264, an intra prediction image is obtained by performing pixel interpolation (copying or copying after interpolation) along a prediction direction such as a vertical direction or a horizontal direction using an encoded reference pixel value adjacent to a prediction target block. Generate. FIG. The prediction direction of intra prediction in H.264 is shown. Further, in FIG. 2 shows an arrangement relationship between reference pixels and encoding target pixels in H.264. FIG. 5C illustrates a predicted image generation method in mode 1 (horizontal prediction), and FIG. 5D illustrates a predicted image generation method in mode 4 (diagonal lower right prediction).
 また、非特許文献では、H.264の予測方向をさらに34方向に拡張し、予測モード数を増やしている。予測角度に応じて32画素精度の線形補間を行うことで予測画素値を作成し、予測方向にコピーする。本発明の本実施の形態で用いるイントラ予測部109の詳細は後述する。 In non-patent literature, H. The prediction direction of H.264 is further expanded to 34 directions to increase the number of prediction modes. A predicted pixel value is created by performing linear interpolation with 32-pixel accuracy in accordance with the predicted angle, and is copied in the predicted direction. Details of the intra prediction unit 109 used in the present embodiment of the present invention will be described later.
 インター予測部110は、参照画像メモリ108に保存されている参照画像信号124を利用してインター予測を行う。具体的には、インター予測部110は、予測対象ブロックと参照画像信号124との間でブロックマッチング処理を行って動きのズレ量(動きベクトル)を導出する。インター予測部110は、この動きベクトルに基づいて補間処理(動き補償)を行ってインター予測画像を生成する。H.264では、1/4画素精度までの補間処理が可能である。導出された動きベクトルは予測情報125の一部としてエントロピー符号化される。 The inter prediction unit 110 performs inter prediction using the reference image signal 124 stored in the reference image memory 108. Specifically, the inter prediction unit 110 performs block matching processing between the prediction target block and the reference image signal 124 to derive a motion shift amount (motion vector). The inter prediction unit 110 performs an interpolation process (motion compensation) based on the motion vector to generate an inter prediction image. H. With H.264, interpolation processing up to 1/4 pixel accuracy is possible. The derived motion vector is entropy encoded as part of the prediction information 125.
 予測選択スイッチ111は、イントラ予測部109の出力端またはインター予測部110の出力端を予測選択部112からの予測情報125に従って選択し、イントラ予測画像またはインター予測画像を予測画像信号126として減算部101及び加算部106に入力する。予測情報125がイントラ予測を示唆する場合には、予測選択スイッチ111はイントラ予測部109からの出力端にスイッチを接続する。一方、予測情報125がインター予測を示唆する場合には、予測選択スイッチ111はインター予測部110からの出力端にスイッチを接続する。 The prediction selection switch 111 selects the output terminal of the intra prediction unit 109 or the output terminal of the inter prediction unit 110 according to the prediction information 125 from the prediction selection unit 112, and subtracts the intra prediction image or the inter prediction image as the prediction image signal 126. 101 and the adder 106. When the prediction information 125 suggests intra prediction, the prediction selection switch 111 connects a switch to the output terminal from the intra prediction unit 109. On the other hand, when the prediction information 125 suggests inter prediction, the prediction selection switch 111 connects a switch to the output terminal from the inter prediction unit 110.
 予測選択部112は、符号化制御部115が制御する予測モードに従って、予測情報125を設定する機能を有する。前述のように、予測画像信号126の生成のためにイントラ予測またはインター予測が選択可能であるが、イントラ予測及びインター予測の夫々に複数のモードがさらに選択可能である。符号化制御部115はイントラ予測及びインター予測の複数の予測モードのうち1つを最適な予測モードとして判定し、予測選択部112は判定された最適な予測モードに応じて予測情報125を設定する。 The prediction selection unit 112 has a function of setting the prediction information 125 according to the prediction mode controlled by the encoding control unit 115. As described above, intra prediction or inter prediction can be selected for generating the predicted image signal 126, but a plurality of modes can be further selected for each of intra prediction and inter prediction. The encoding control unit 115 determines one of a plurality of prediction modes of intra prediction and inter prediction as the optimal prediction mode, and the prediction selection unit 112 sets the prediction information 125 according to the determined optimal prediction mode. .
 例えば、イントラ予測に関して、符号化制御部115から予測モード情報がイントラ予測部109に指定され、イントラ予測部109はこの予測モード情報に従って予測画像信号126を生成する。符号化制御部115は、予測モードの番号が小さい方から順に複数の予測モード情報を指定してもよいし、大きい方から順に複数の予測モード情報を指定してもよい。また、符号化制御部115は、入力画像の特性に従って予測モードを限定してもよい。符号化制御部115は、必ずしも全ての予測モードを指定する必要はなく符号化対象ブロックに対して少なくとも1つの予測モード情報を指定すればよい。 For example, for intra prediction, prediction mode information is designated by the intra prediction unit 109 from the encoding control unit 115, and the intra prediction unit 109 generates a predicted image signal 126 according to the prediction mode information. The encoding control unit 115 may specify a plurality of prediction mode information in order from the smallest prediction mode number, or may specify a plurality of prediction mode information in order from the largest. The encoding control unit 115 may limit the prediction mode according to the characteristics of the input image. The encoding control unit 115 does not necessarily specify all prediction modes, and may specify at least one prediction mode information for the encoding target block.
 例えば、符号化制御部115は、次の数式(1)に示すコスト関数を用いて最適な予測モードを判定する。 
Figure JPOXMLDOC01-appb-M000001
For example, the encoding control unit 115 determines an optimal prediction mode using a cost function represented by the following mathematical formula (1).
Figure JPOXMLDOC01-appb-M000001
 数式(1)(以下、簡易符号化コストと呼ぶ)において、OHは予測情報125(例えば、動きベクトル情報、予測ブロックサイズ情報)に関する符号量を示し、SADは予測対象ブロックと予測画像信号126との間の差分絶対値和(即ち、予測誤差信号117の絶対値の累積和)を示す。また、λは量子化情報(量子化パラメータ)の値に基づいて決定されるラグランジュ未定乗数を示し、Kは符号化コストを示す。数式(1)を用いる場合には、符号化コストKを最小化する予測モードが発生符号量及び予測誤差の観点から最適な予測モードとして判定される。数式(1)の変形として、OHのみまたはSADのみから符号化コストを見積もってもよいし、SADにアダマール変換を施した値またはその近似値を利用して符号化コストを見積もってもよい。 In Equation (1) (hereinafter referred to as simple encoding cost), OH indicates a code amount relating to prediction information 125 (for example, motion vector information and prediction block size information), and SAD is a prediction target block, a prediction image signal 126, and The difference absolute value sum (ie, the cumulative sum of the absolute values of the prediction error signal 117) is shown. Further, λ represents a Lagrange undetermined multiplier determined based on the value of quantization information (quantization parameter), and K represents an encoding cost. When Expression (1) is used, the prediction mode that minimizes the coding cost K is determined as the optimum prediction mode from the viewpoint of the generated code amount and the prediction error. As a modification of Equation (1), the encoding cost may be estimated from OH alone or SAD alone, or the encoding cost may be estimated using a value obtained by subjecting SAD to Hadamard transform or an approximation thereof.
 また、図示しない仮符号化ユニットを用いることにより最適な予測モードを判定することも可能である。例えば、符号化制御部115は、次の数式(2)に示すコスト関数を用いて最適な予測モードを判定する。
Figure JPOXMLDOC01-appb-M000002
It is also possible to determine an optimal prediction mode by using a temporary encoding unit (not shown). For example, the encoding control unit 115 determines an optimal prediction mode using a cost function expressed by the following formula (2).
Figure JPOXMLDOC01-appb-M000002
 数式(2)において、Dは予測対象ブロックと局所復号画像との間の二乗誤差和(即ち、符号化歪)を示し、Rは予測対象ブロックと予測モードの予測画像信号126との間の予測誤差について仮符号化によって見積もられた符号量を示し、Jは符号化コストを示す。数式(2)の符号化コストJ(以後、詳細符号化コストと呼ぶ)を導出する場合には予測モード毎に仮符号化処理及び局部復号化処理が必要なので、回路規模または演算量が増大する。反面、より正確な符号化歪と符号量とに基づいて符号化コストJが導出されるので、最適な予測モードを高精度に判定して高い符号化効率を維持しやすい。なお、数式(2)の変形として、RのみまたはDのみから符号化コストを見積もってもよいし、RまたはDの近似値を利用して符号化コストを見積もってもよい。また、これらのコストを階層的に用いてもよい。符号化制御部115は、予測対象ブロックに関して事前に得られる情報(周囲の画素ブロックの予測モード、画像解析の結果など)に基づいて、数式(1)または数式(2)を用いた判定を行う予測モードの候補の数を、予め絞り込んでおいてもよい。 In Equation (2), D represents a square error sum (ie, encoding distortion) between the prediction target block and the local decoded image, and R represents a prediction between the prediction target block and the prediction image signal 126 in the prediction mode. An error amount indicates a code amount estimated by provisional encoding, and J indicates an encoding cost. In order to derive the encoding cost J (hereinafter referred to as the detailed encoding cost) of Equation (2), provisional encoding processing and local decoding processing are required for each prediction mode, so that the circuit scale or the amount of calculation increases. . On the other hand, since the encoding cost J is derived based on more accurate encoding distortion and code amount, it is easy to determine the optimal prediction mode with high accuracy and maintain high encoding efficiency. As a modification of Equation (2), the encoding cost may be estimated from only R or D, or the encoding cost may be estimated using an approximate value of R or D. These costs may be used hierarchically. The encoding control unit 115 performs determination using Expression (1) or Expression (2) based on information obtained in advance regarding the prediction target block (prediction mode of surrounding pixel blocks, image analysis result, and the like). The number of prediction mode candidates may be narrowed down in advance.
 本実施の形態の変形例として、数式(1)と数式(2)を組み合わせた二段階のモード判定を行うことで、符号化性能を維持しつつ、予測モードの候補数をさらに削減することが可能となる。ここで、数式(1)で示される簡易符号化コストは、数式(2)と異なり局部復号化処理が必要ないため、高速に演算が可能である。本実施形態の動画像符号化装置では、H.264と比較しても予測モード数が多いため、詳細符号化コストを用いたモード判定は現実的ではない。そこで、第一ステップとして、簡易符号化コストを用いたモード判定を、当該画素ブロックで利用可能な予測モードに対して行い、予測モード候補を導出する。 As a modification of the present embodiment, the number of prediction mode candidates can be further reduced while maintaining coding performance by performing two-stage mode determination combining Formula (1) and Formula (2). It becomes possible. Here, unlike the formula (2), the simple encoding cost represented by the formula (1) does not require a local decoding process, and can be calculated at high speed. In the moving picture coding apparatus according to the present embodiment, H.264 is used. Since the number of prediction modes is large even when compared with H.264, mode determination using the detailed coding cost is not realistic. Therefore, as a first step, mode determination using the simple coding cost is performed on the prediction modes available in the pixel block, and prediction mode candidates are derived.
 ここで、量子化の粗さを定めた量子化パラメータの値が大きくなるほど、簡易符号化コストと詳細符号化コストの相関が高くなる性質を利用して、予測モード候補数を変更する。 Here, the number of prediction mode candidates is changed using the property that the correlation between the simple coding cost and the detailed coding cost increases as the value of the quantization parameter that determines the roughness of quantization increases.
 図4は、第一ステップで選択する予測モード候補数を示している。PuSizeは、後述する予測を行う画素ブロック(プレディクションユニットとよぶ場合もある)サイズを示すインデックスである。QPは量子化パラメータを示しており、5で割った商によって予測モード候補数が変化する。このようにして絞られた予測モード候補数に対してだけ、詳細符号化コストを導出すればよいため、局部復号化処理の回数を大幅に削減することが可能となる。 FIG. 4 shows the number of prediction mode candidates selected in the first step. PuSize is an index indicating the size of a pixel block (sometimes referred to as a prediction unit) that performs prediction described later. QP represents a quantization parameter, and the number of prediction mode candidates changes depending on the quotient divided by 5. Since the detailed coding cost only has to be derived for the number of prediction mode candidates narrowed down in this way, the number of local decoding processes can be greatly reduced.
 符号化制御部115は、図1の画像符号化装置の各要素を制御する。具体的には、符号化制御部115は、上述の動作を含む符号化処理のための種々の制御を行う。 The encoding control unit 115 controls each element of the image encoding device in FIG. Specifically, the encoding control unit 115 performs various controls for the encoding process including the above-described operation.
 以下、図6を用いて本実施形態に係るイントラ予測部109の詳細を説明する。 
 <イントラ予測部109> 
 図6に示されるイントラ予測部109は、単方向イントラ予測画像生成部601、双方向イントラ予測画像生成部602、予測モード情報設定部603、選択スイッチ604を含む。先ず、参照画像メモリ108から参照画像信号124が、単方向イントラ予測画像生成部601及び双方向イントラ予測画像生成部602に入力される。ここで、符号化制御部115で制御されている予測モード情報に従って、予測モード情報設定部603は、単方向イントラ予測画像生成部601或いは双方向イントラ予測画像生成部602で生成される予測モードを設定し、予測モード605を出力する。選択スイッチ604は、この予測モード605に従って、それぞれのイントラ予測画像生成部の出力端を繋ぎかえる機能を有する。入力された予測モード605が単方向イントラ予測モードであれば、単方向イントラ予測画像生成部601の出力端とスイッチを接続し、予測モード605が双方向イントラ予測モードであれば、双方向イントラ予測画像生成部602の出力端を接続する。一方、それぞれのイントラ予測画像生成部601、602は、予測モード605に従って、予測画像信号126を生成する。生成された予測画像信号126(第五予測画像信号とも呼ぶ)がイントラ予測部109から出力される。単方向イントラ予測画像生成部601の出力信号を第四予測画像信号とも呼び、双方向イントラ予測画像生成部602の出力信号を第三予測画像信号とも呼ぶ。
Hereinafter, the details of the intra prediction unit 109 according to the present embodiment will be described with reference to FIG.
<Intra Prediction Unit 109>
The intra prediction unit 109 illustrated in FIG. 6 includes a unidirectional intra predicted image generation unit 601, a bidirectional intra predicted image generation unit 602, a prediction mode information setting unit 603, and a selection switch 604. First, the reference image signal 124 is input from the reference image memory 108 to the unidirectional intra predicted image generation unit 601 and the bidirectional intra predicted image generation unit 602. Here, according to the prediction mode information controlled by the encoding control unit 115, the prediction mode information setting unit 603 determines the prediction mode generated by the unidirectional intra prediction image generation unit 601 or the bidirectional intra prediction image generation unit 602. Set and output prediction mode 605. The selection switch 604 has a function of switching the output ends of the respective intra predicted image generation units according to the prediction mode 605. If the input prediction mode 605 is the unidirectional intra prediction mode, the output terminal of the unidirectional intra prediction image generation unit 601 is connected to the switch, and if the prediction mode 605 is the bidirectional intra prediction mode, the bidirectional intra prediction is performed. The output terminal of the image generation unit 602 is connected. On the other hand, each of the intra predicted image generation units 601 and 602 generates the predicted image signal 126 according to the prediction mode 605. The generated prediction image signal 126 (also referred to as a fifth prediction image signal) is output from the intra prediction unit 109. The output signal of the unidirectional intra predicted image generation unit 601 is also called a fourth predicted image signal, and the output signal of the bidirectional intra predicted image generation unit 602 is also called a third predicted image signal.
 先ず、予測モード情報設定部603について詳細に説明する。図7は、本発明の本実施の形態に関わる予測モードのブロックサイズ別の数を示している。PuSizeは、予測を行う画素ブロック(プレディクションユニット)サイズを示しており、PU_2x2~PU_128x128まで7種類のサイズが規定されている。IntraUniModeNumは単方向イントラ予測の予測モード数を表しており、IntraBiModeNumは双方向イントラ予測の予測モード数を表している。また、Number of modesは、画素ブロック(プレディクションユニット)サイズ毎の予測モード数の総数である。 First, the prediction mode information setting unit 603 will be described in detail. FIG. 7 shows the number of prediction modes according to the block size according to the present embodiment of the present invention. PuSize indicates the pixel block (prediction unit) size to be predicted, and seven types of sizes from PU_2x2 to PU_128x128 are defined. IntraUniModeNum represents the number of prediction modes for unidirectional intra prediction, and IntraBiModeNum represents the number of prediction modes for bidirectional intra prediction. Also, Number of modes is the total number of prediction modes for each pixel block (prediction unit) size.
 一方、図9にPuSizeがPU_8x8、PU_16x16及びPU_32x32の場合の、予測モードと予測方法の関係を示す。なお、図10にPuSizeがPU_4x4の場合を示し、図11にPU_64x64或いはPU_128x128の場合を示す。ここで、IntraPredModeは予測モード番号を示し、IntraBipredFlagは、双方向イントラ予測であるかを示すフラグである。当該フラグが1である場合は、当該予測モードが双方向イントラ予測モードであることを示している。当該フラグが0の場合は、当該予測モードが単方向イントラ予測モードであることを示している。IntraPredTypeLXはイントラ予測の予測タイプを示している。Intra_Verticalは、垂直方向を予測の基準とすることを意味し、Intra_Horizontalは水平方向を予測の基準とすることを意味する。なお、IntraPredTypeLX内のXには、0或いは1が適用される。IntraPredTypeL0は、単方向イントラ予測或いは双方向イントラ予測の最初の予測モードを示している。IntraPredTypeL1は、双方向イントラ予測の2番目の予測モードを示している。また、IntraPredAngleIDは、予測角度のインデックスを示す指標である。実際に予測値生成で使われる予測角度は図12に示されている。ここで、puPartIdxは図3Bで説明した四分木分割の際の分割したプレディクションユニットのインデックスを表している。 On the other hand, FIG. 9 shows the relationship between the prediction mode and the prediction method when PuSize is PU_8x8, PU_16x16, and PU_32x32. FIG. 10 shows a case where PuSize is PU_4 × 4, and FIG. 11 shows a case where PU_64 × 64 or PU_128 × 128. Here, IntraPredMode indicates a prediction mode number, and IntraBipredFlag is a flag indicating whether or not bidirectional intra prediction. When the flag is 1, it indicates that the prediction mode is the bidirectional intra prediction mode. When the flag is 0, it indicates that the prediction mode is a unidirectional intra prediction mode. IntraPredTypeLX indicates the prediction type of intra prediction. Intra_Vertical means that the vertical direction is the reference for prediction, and Intra_Horizontal means that the horizontal direction is the reference for prediction. Note that 0 or 1 is applied to X in IntraPredTypeLX. IntraPredTypeL0 indicates the first prediction mode of unidirectional intra prediction or bidirectional intra prediction. IntraPredTypeL1 indicates the second prediction mode of bidirectional intra prediction. IntraPred AngleID is an index indicating an index of a prediction angle. The prediction angle actually used in the generation of the predicted value is shown in FIG. Here, puPartIdx represents an index of the prediction unit that is divided in the quadtree division described with reference to FIG. 3B.
 例えば、IntraPredModeが4の場合、IntraPredTypeL0がIntra_Verticalであるため、垂直方向を予測の基準とすることが判る。本図から判る通り、IntraPredMode=0から32まで計33個が単方向イントラ予測モードを示しており、IntraPredMode=32から48まで計16個が双方向イントラ予測モードを示している。 For example, when IntraPredMode is 4, since IntraPredTypeL0 is Intra_Vertical, it can be seen that the vertical direction is used as a reference for prediction. As can be seen from the figure, a total of 33 from IntraPredMode = 0 to 32 indicate the unidirectional intra prediction mode, and a total of 16 from IntraPredMode = 32 to 48 indicate the bidirectional intra prediction mode.
 予測モード情報設定部603は、符号化制御部115における制御の下で、指定された予測モード605に対応する上述した予測情報を単方向イントラ予測画像生成部601及び双方向イントラ予測画像生成部602に設定し、選択スイッチ604へ予測モード605を出力する。 The prediction mode information setting unit 603 converts the above-described prediction information corresponding to the designated prediction mode 605 to the unidirectional intra prediction image generation unit 601 and the bidirectional intra prediction image generation unit 602 under the control of the encoding control unit 115. And the prediction mode 605 is output to the selection switch 604.
 次に、単方向イントラ予測画像生成部601について詳細に説明する。単方向イントラ予測画像生成部601は、図8に示される複数の予測方向に対して予測画像信号126を生成する機能を有する。図8では、太線で示される垂直方向、水平方向の座標に対して、33個の異なる予測方向を持つ。また、H.264で示される代表的な予測角度の方向を矢印で示している。本発明の本実施の形態では、原点から菱形で示されるマークまで線を引いた方向に33種類の予測方向が用意されている。また、H.264と同様、利用可能な参照画素の平均値で予測するDC予測が追加されており、合計で34個の予測モードが存在する。 Next, the unidirectional intra predicted image generation unit 601 will be described in detail. The unidirectional intra predicted image generation unit 601 has a function of generating the predicted image signal 126 for a plurality of prediction directions shown in FIG. In FIG. 8, there are 33 different prediction directions for the vertical and horizontal coordinates indicated by the bold lines. H. The direction of a typical prediction angle indicated by H.264 is indicated by an arrow. In the present embodiment of the present invention, 33 kinds of prediction directions are prepared in a direction in which a line is drawn from the origin to a mark indicated by a diamond. H. Similar to H.264, DC prediction for predicting with an average value of available reference pixels is added, and there are 34 prediction modes in total.
 IntraPredMode=4の場合、IntraPredAngleIDL0が-4であるため、図8におけるIntraPredMode=4で示される予測方向で予測画像信号126が生成される。図8における点線で示される矢印は、予測タイプがIntra_Verticalの予測モードを示しており、実線で示される矢印は予測タイプがIntra_Horizontalの予測モードを示している。 When IntraPredMode = 4, since IntraPredAngleIDL0 is −4, the prediction image signal 126 is generated in the prediction direction indicated by IntraPredMode = 4 in FIG. An arrow indicated by a dotted line in FIG. 8 indicates a prediction mode whose prediction type is Intra_Vertical, and an arrow indicated by a solid line indicates a prediction mode whose prediction type is Intra_Horizontal.
 次に単方向イントラ予測画像生成部601の予測画像生成方法について説明する。ここでは、入力された参照画像信号124に基づいて、予測画像値を生成し、上述した予測方向に画素をコピーする。予測画像値は、1/32画素精度で内挿補間を行うことによって生成する。図12は、IntraPredAngleIDLXと予測画像値生成に使われるintraPredAngleとの関係を示す。intraPredAngleは予測値生成の際に、実際に利用される予測角度を示している。例えば、予測タイプがIntra_Verticalで且つ、図12で示されるintraPredAngleが正の値の場合の予測値の生成方法を数式で書くと数式(3)で表される。ここでBLK_SIZEは、当該画素ブロック(プレディクションユニット)のサイズを示しており、ref[]は参照画像信号が格納された配列を示している。また、pred(k,m)は生成された予測画像信号126を示している。
Figure JPOXMLDOC01-appb-M000003
Next, a prediction image generation method of the unidirectional intra prediction image generation unit 601 will be described. Here, a predicted image value is generated based on the input reference image signal 124, and pixels are copied in the above-described prediction direction. The predicted image value is generated by performing interpolation with 1/32 pixel accuracy. FIG. 12 shows the relationship between IntraPredAngleIDLX and intraPredAngle used for predictive image value generation. intraPredAngle indicates a prediction angle that is actually used when a predicted value is generated. For example, when the prediction type is Intra_Vertical and intraPredAngle shown in FIG. 12 is a positive value, a prediction value generation method is expressed by Expression (3). Here, BLK_SIZE indicates the size of the pixel block (prediction unit), and ref [] indicates an array in which reference image signals are stored. Pred (k, m) indicates the generated predicted image signal 126.
Figure JPOXMLDOC01-appb-M000003
 上記条件以外に関しても、図12のテーブルに従って同様な方法で予測値が生成可能である。例えば、IntraPredMode=1で示される予測モードの予測値は、図5(c)で示されるH.264の水平予測と同一となる。以上が、本発明の本実施の形態における単方向イントラ予測画像生成部601の説明である。 Even for conditions other than the above, a predicted value can be generated by a similar method according to the table of FIG. For example, the prediction value of the prediction mode indicated by IntraPredMode = 1 is H.264 shown in FIG. This is the same as H.264 horizontal prediction. The above is description of the unidirectional intra estimated image generation part 601 in this Embodiment of this invention.
 次に、双方向イントラ予測画像生成部602について詳細に説明する。図13に双方向イントラ予測画像生成部602のブロック図を示す。双方向イントラ予測画像生成部602は、第一単方向イントラ予測画像生成部1301、第二単方向イントラ予測画像生成部1302、重み付き平均部1303を含み、入力された参照画像信号124に基づいて、2つの単方向イントラ予測画像を生成し、これらを重み付き平均することにより予測画像信号126を生成する機能を有する。 Next, the bidirectional intra predicted image generation unit 602 will be described in detail. FIG. 13 shows a block diagram of the bidirectional intra-predicted image generation unit 602. The bidirectional intra predicted image generation unit 602 includes a first unidirectional intra predicted image generation unit 1301, a second unidirectional intra predicted image generation unit 1302, and a weighted average unit 1303, and is based on the input reference image signal 124. Two unidirectional intra-predicted images are generated, and a function of generating a predicted image signal 126 by weighting and averaging them is provided.
 第一単方向イントラ予測画像生成部1301及び第二単方向イントラ予測画像生成部1302の機能は同一である。いずれも符号化制御部115で制御される予測モード情報に従って与えられた予測モードに対応する予測画像信号を生成する。第一単方向イントラ予測画像生成部1301から第一予測画像信号1304が、第二単方向イントラ予測画像生成部1302から第二予測画像信号1305が出力される。それぞれの予測画像信号が重み付き平均部1303に入力され、重み付き平均処理が行われる。重み付き平均部1303の出力信号は第三予測画像信号とも呼ばれる。 The functions of the first unidirectional intra predicted image generation unit 1301 and the second unidirectional intra predicted image generation unit 1302 are the same. In either case, a prediction image signal corresponding to a prediction mode given according to prediction mode information controlled by the encoding control unit 115 is generated. A first predicted image signal 1304 is output from the first unidirectional intra predicted image generation unit 1301, and a second predicted image signal 1305 is output from the second unidirectional intra predicted image generation unit 1302. Each predicted image signal is input to the weighted average unit 1303, and weighted average processing is performed. The output signal of the weighted average unit 1303 is also called a third predicted image signal.
 図14のテーブルは、双方向イントラ予測モードから、2つの単方向イントラ予測モードを導出するためのテーブルである。ここで数式(4)を用いてBiPredIdxが導出される。
Figure JPOXMLDOC01-appb-M000004
The table in FIG. 14 is a table for deriving two unidirectional intra prediction modes from the bidirectional intra prediction mode. Here, BiPredIdx is derived using Equation (4).
Figure JPOXMLDOC01-appb-M000004
例えば、PuSize=PU_8x8、IntraPredMode=33の場合、図7からIntraUniModeNum=33であることが判るので、BiPredIdx=0であることが判る。結果として、図14から、第一単方向イントラ予測モード(MappedBi2Uni(0,idx))が1、第二単方向イントラ予測モード(MappedBi2Uni(1,idx))が0であることが導出される。他のPuSize及びIntraPredModeにおいても同様の方法で2つの予測モードを導出することが可能である。なお、以後、第一単方向イントラ予測モードをIntraPredModeL0、第二単方向イントラ予測モードをIntraPredModeL1で表現する。 For example, when PuSize = PU — 8 × 8 and IntraPredMode = 33, it can be seen from FIG. 7 that IntraUniModeNum = 33, and thus BiPredIdx = 0. As a result, it is derived from FIG. 14 that the first unidirectional intra prediction mode (MappedBi2Uni (0, idx)) is 1 and the second unidirectional intra prediction mode (MappedBi2Uni (1, idx)) is 0. In other PuSize and IntraPredMode, it is possible to derive two prediction modes by the same method. Hereinafter, the first unidirectional intra prediction mode is expressed as IntraPredModeL0, and the second unidirectional intra prediction mode is expressed as IntraPredModeL1.
 このように第一単方向イントラ予測画像生成部1301及び第二単方向イントラ予測画像生成部1302で生成された第一予測画像信号1304、第二予測画像信号1305が、重み付き平均部1303へと入力される。 Thus, the first predicted image signal 1304 and the second predicted image signal 1305 generated by the first unidirectional intra predicted image generation unit 1301 and the second unidirectional intra predicted image generation unit 1302 are sent to the weighted average unit 1303. Entered.
 重み付き平均部1303では、IntraPredModeL0及びIntraPredModeL1の予測方向を元にユークリッド距離或いは市街地距離(マンハッタン距離)を計算し、重み付き平均処理で用いる重み成分を導出する。各画素の重み成分は予測に用いる参照画素からのユークリッド距離若しくは市街地距離の逆数で表され、以下の式で一般化される。
Figure JPOXMLDOC01-appb-M000005
The weighted average unit 1303 calculates the Euclidean distance or the city area distance (Manhattan distance) based on the prediction directions of IntraPredModeL0 and IntraPredModeL1, and derives a weight component used in the weighted average process. The weight component of each pixel is represented by the Euclidean distance from the reference pixel used for prediction or the reciprocal of the urban distance, and is generalized by the following equation.
Figure JPOXMLDOC01-appb-M000005
ここでユークリッド距離を用いる場合は、ΔLは以下の式で表される。
Figure JPOXMLDOC01-appb-M000006
Here, when the Euclidean distance is used, ΔL is expressed by the following equation.
Figure JPOXMLDOC01-appb-M000006
一方、市街地距離を用いる場合は、ΔLは以下の式で表される。
Figure JPOXMLDOC01-appb-M000007
On the other hand, when using the city distance, ΔL is expressed by the following equation.
Figure JPOXMLDOC01-appb-M000007
各予測モードに関する重みテーブルは、下式で一般化される。
Figure JPOXMLDOC01-appb-M000008
The weight table for each prediction mode is generalized by the following equation.
Figure JPOXMLDOC01-appb-M000008
ここで、ρL0(n)はIntraPredModeL0における画素位置nの重み成分、ρL1(n)はIntraPredModeL1における画素位置nの重み成分を表している。従って、画素位置nにおける最終的な予測信号は、下式で示される。
Figure JPOXMLDOC01-appb-M000009
Here, ρ L0 (n) represents the weight component of the pixel position n in IntraPredModeL0, and ρ L1 (n) represents the weight component of the pixel position n in IntraPredModeL1. Therefore, the final prediction signal at the pixel position n is expressed by the following equation.
Figure JPOXMLDOC01-appb-M000009
ここで、BiPred(n)は画素位置nにおける予測画像信号を表しており、PredL0(n)、PredL1(n)はそれぞれIntraPredModeL0、IntraPredModeL1の予測画像信号である。 Here, BiPred (n) represents a predicted image signal at the pixel position n, and PredL0 (n) and PredL1 (n) are predicted image signals of IntraPredModeL0 and IntraPredModeL1, respectively.
 本実施形態では、予測画素の生成に2つの予測モードを選択して予測信号を生成しているが、別の実施形態として3つ以上の予測モードを選択して予測値を生成してもよい。この場合、参照画素から予測画素の空間的距離の逆数の比を重み係数と設定すればよい。 In this embodiment, the prediction signal is generated by selecting two prediction modes for generating the prediction pixel. However, as another embodiment, a prediction value may be generated by selecting three or more prediction modes. . In this case, the ratio of the reciprocal of the spatial distance from the reference pixel to the prediction pixel may be set as the weighting factor.
 また、本実施形態では、予測モードが使用する参照画素からのユークリッド距離、或いは市街地距離の逆数をそのまま重み成分としているが、本実施形態の変形例として参照画素からのユークリッド距離、市街地距離を変数とした分布モデルを用いて重み成分を設定してもよい。分布モデルは線形モデルやM次関数(M≧1)、片側ラプラス分布や片側ガウス分布といった非線形関数、ある固定値を参照画素との距離に依らず固定値、のうち少なくとも一つを用いる。片側ガウス分布をモデルとして用いた場合、重み成分は下式で表される。
Figure JPOXMLDOC01-appb-M000010
In this embodiment, the Euclidean distance from the reference pixel used in the prediction mode or the reciprocal of the urban area distance is directly used as a weight component. However, as a modification of the present embodiment, the Euclidean distance and the urban area distance from the reference pixel are variables. The weight component may be set using the distribution model. The distribution model uses at least one of a linear model, an M-order function (M ≧ 1), a nonlinear function such as a one-sided Laplace distribution or a one-sided Gaussian distribution, and a fixed value that is a fixed value regardless of the distance from the reference pixel. When a one-sided Gaussian distribution is used as a model, the weight component is expressed by the following equation.
Figure JPOXMLDOC01-appb-M000010
ここで、ρ(n)は予測画素の位置nにおける重み成分、σは分散、Aは定数(A>0)である。 Here, ρ (n) is a weight component at the position n of the predicted pixel, σ 2 is variance, and A is a constant (A> 0).
 また、片側ラプラス分布をモデルとして用いた場合重み成分は下式で表される。
Figure JPOXMLDOC01-appb-M000011
When the one-sided Laplace distribution is used as a model, the weight component is expressed by the following equation.
Figure JPOXMLDOC01-appb-M000011
ここで、σは標準偏差、Bは定数(B>0)である。 Here, σ is a standard deviation, and B is a constant (B> 0).
 また、自己相関関数をモデル化した等方相関モデル、楕円相関モデル、ラプラス関数やガウス関数を一般化した一般化ガウスモデルを重み成分のモデルとして用いてもよい。 Further, an isotropic correlation model obtained by modeling an autocorrelation function, an elliptic correlation model, a generalized Gaussian model obtained by generalizing a Laplace function or a Gaussian function may be used as the weight component model.
 数式(5)、数式(8)、数式(10)、数式(11)で示される重み成分を予測画像生成の際に都度計算した場合、複数の乗算器が必要となり、ハードウェア規模が増大する。このため、予め予測モード毎の相対距離に応じて重み成分を計算し、メモリに保持することで上記演算に必要なる回路規模を削減することができる。ここでは、市街地距離を用いた場合の重み成分の導出方法について説明する。ここで相対距離は、ある予測方向に関して、予測対象の画素と参照画素との間の距離である。 When the weight components represented by Equation (5), Equation (8), Equation (10), and Equation (11) are calculated each time the predicted image is generated, a plurality of multipliers are required, and the hardware scale increases. . For this reason, the circuit scale required for the said calculation can be reduced by calculating a weight component beforehand according to the relative distance for every prediction mode, and hold | maintaining in a memory. Here, a method for deriving the weight component when the city distance is used will be described. Here, the relative distance is a distance between a prediction target pixel and a reference pixel with respect to a certain prediction direction.
 IntraPredModeL0の市街地距離ΔLL0とIntraPredModeL1の市街地距離ΔLL1は、数式(7)より計算される。ここで、相対距離は2つの予測モードの予測方向によって変化する。例として、PuSize=PU_4x4の場合の代表的な距離を図15A、B、Cに示す。図15Aは、IntraPredModeLX=0の場合の市街地距離を示している。図15Bは、IntraPredModeLX=1の場合の市街地距離を示している。また、図15Cは、IntraPredModeLX=3の場合の市街地距離を示している。同様に、それぞれの予測モードに応じて数式(6)或いは数式(7)を用いて距離を導出することができる。但し、IntraPredModeLX=2のDC予測の場合、全ての画素位置で距離を2としている。図16にPuSize=PU_4x4の場合の代表的な5つの予測モードにおける距離のテーブルを示す。IntraPredModeLXの数が多い場合、これらの距離テーブルのテーブルサイズが増加する場合がある。 The city area distance ΔL L0 of IntraPredMode L0 and the city area distance ΔL L1 of IntraPredMode L1 are calculated from Equation (7). Here, the relative distance varies depending on the prediction direction of the two prediction modes. As an example, typical distances in the case of PuSize = PU — 4 × 4 are shown in FIGS. FIG. 15A shows the city distance when IntraPredModeLX = 0. FIG. 15B shows the city distance in the case of IntraPredModeLX = 1. FIG. 15C shows the city distance in the case of IntraPredModeLX = 3. Similarly, the distance can be derived using Expression (6) or Expression (7) according to each prediction mode. However, in the case of DC prediction with IntraPredModeLX = 2, the distance is 2 at all pixel positions. FIG. 16 shows a table of distances in five typical prediction modes when PuSize = PU — 4 × 4. When the number of IntraPredModeLX is large, the table sizes of these distance tables may increase.
 本実施の形態では、いくつかの予測角度の近い予測モードの距離テーブルを共有化することにより、必要なメモリ量を削減している。図17は、距離テーブル導出に用いるIntraPredModeLXのマッピングを示している。ここでは予測角度が45度刻みの予測モードとDC予測に対応する予測モードのみのテーブルを用意しており、それ以外の予測角度を、用意した基準の予測モードに近い方にマッピングする例を示している。なお、基準の予測モードとの距離が同じ場合は、インデックスが小さい方にマッピングしている。MappedIntraPredModeに示される予測モードが図17から参照され、距離テーブルが導出できる。 In the present embodiment, the required memory amount is reduced by sharing a distance table in a prediction mode with several prediction angles close to each other. FIG. 17 shows the mapping of IntraPredModeLX used for distance table derivation. Here, an example is shown in which a table of only the prediction mode corresponding to the prediction mode corresponding to the prediction mode and the DC prediction in 45 degrees is prepared, and other prediction angles are mapped closer to the prepared reference prediction mode. ing. When the distance from the reference prediction mode is the same, the index is mapped to the smaller one. The prediction mode shown in “MappedIntraPredMode” is referred to from FIG. 17, and a distance table can be derived.
 上記距離テーブルを利用することにより、次式を用いて2つの予測モードの画素毎の相対距離が算出される。
Figure JPOXMLDOC01-appb-M000012
By using the distance table, the relative distance for each pixel in the two prediction modes is calculated using the following equation.
Figure JPOXMLDOC01-appb-M000012
ここで、BLK_WIDTH及びBLK_HEIGHTは、それぞれ当該画素ブロック(プレディクションユニット)の幅及び高さを示しており、DistDiff(n)は画素位置nにおける2つの予測モードの相対距離を示している。数式(12)を用いて、画素位置nにおける最終的な予測信号は、下式で示される。
Figure JPOXMLDOC01-appb-M000013
Here, BLK_WIDTH and BLK_HEIGHT indicate the width and height of the pixel block (prediction unit), respectively, and DistDiff (n) indicates the relative distance between the two prediction modes at the pixel position n. Using Equation (12), the final prediction signal at the pixel position n is expressed by the following equation.
Figure JPOXMLDOC01-appb-M000013
ここで、小数点演算を用いることによるハードウェア規模増加を避けるため、重み成分を予めスケーリングし、整数演算に直すと次式で表せる。
Figure JPOXMLDOC01-appb-M000014
Here, in order to avoid an increase in hardware scale due to the use of decimal point arithmetic, when weight components are scaled in advance and converted to integer arithmetic, the following expression is obtained.
Figure JPOXMLDOC01-appb-M000014
ここで、例えば小数点部分を10ビット精度で表現した場合、WM=1024、Offset=512、SHIFT=10となる。これらは以下の関係を満たす。
Figure JPOXMLDOC01-appb-M000015
Here, for example, when the decimal part is expressed with 10-bit precision, WM = 1024, Offset = 512, and SHIFT = 10. These satisfy the following relationship.
Figure JPOXMLDOC01-appb-M000015
SHIFTは重み成分の小数点演算の演算精度を示しており、符号化性能とハードウェア実現時の回路規模のバランスを取って、最適な組み合わせを選べばよい。 SHIFT indicates the calculation accuracy of the decimal point calculation of the weight component, and an optimal combination may be selected by balancing the coding performance and the circuit scale at the time of hardware implementation.
 本発明の本実施の形態における片側ラプラス分布モデルを用いた重み成分をテーブル化した例を図18A、Bに示す。図18AはPuSize=PU_4x4の場合の重み成分テーブルを示している。また、図18BはPuSize=PU_8x8の場合の重み成分テーブルを示している。それ以外のPuSizeに関しても、数式(5)、数式(8)、数式(10)、数式(11)を用いることにより導出が可能である。 
 以上が、本発明の本実施の形態に関わるイントラ予測部109の詳細である。
FIGS. 18A and 18B show examples in which weight components using the one-sided Laplace distribution model in the present embodiment of the present invention are tabulated. FIG. 18A shows a weight component table in the case of PuSize = PU — 4 × 4. FIG. 18B shows a weight component table in the case of PuSize = PU_8 × 8. Other PuSizes can also be derived using Equation (5), Equation (8), Equation (10), and Equation (11).
The above is the details of the intra prediction unit 109 according to the present embodiment of the present invention.
 <イントラ予測部109の符号化部側の処理量削減の変形例> 
 本実施形態の変形例として、イントラ予測部109の内部構成を図19で示される構成としてもよい。この場合、図6で示されるイントラ予測部109の構成と比較すると、画像バッファ1901が追加されおり、双方向イントラ予測画像生成部602が、重み付き平均部1303に置き換わっている。一次画像バッファ1901は、単方向イントラ予測画像生成部601で生成された予測モード毎の予測画像信号126を一次的にバッファに保持する機能を有しており、符号化制御部115が制御する予測モードに応じて、必要な予測モードに対応する予測画像信号126を重み付き平均部1303に出力する。これにより、第一単方向イントラ予測画像生成部1301及び第二単方向イントラ予測画像生成部1302を、双方向イントラ予測画像生成部602が保持する必要がなくなるため、ハードウェア規模を削減することが可能となる。
<Modification of processing amount reduction on the encoding unit side of the intra prediction unit 109>
As a modification of the present embodiment, the internal configuration of the intra prediction unit 109 may be the configuration shown in FIG. In this case, compared with the configuration of the intra prediction unit 109 shown in FIG. 6, an image buffer 1901 is added, and the bidirectional intra predicted image generation unit 602 is replaced with a weighted average unit 1303. The primary image buffer 1901 has a function of temporarily storing the prediction image signal 126 for each prediction mode generated by the unidirectional intra prediction image generation unit 601 in the buffer, and the prediction controlled by the encoding control unit 115. Depending on the mode, the prediction image signal 126 corresponding to the necessary prediction mode is output to the weighted average unit 1303. This eliminates the need for the bidirectional intra predicted image generation unit 602 to hold the first unidirectional intra predicted image generation unit 1301 and the second unidirectional intra predicted image generation unit 1302, thereby reducing the hardware scale. It becomes possible.
 <シンタクス構成 1> 
 以下、図1の画像符号化装置100が利用するシンタクスについて説明する。 
 シンタクスは、画像符号化装置が動画像データを符号化する際の符号化データ(例えば、図1の符号化データ127)の構造を示している。この符号化データを復号化する際に、同じシンタクス構造を参照して動画像復号化装置がシンタクス解釈を行う。図1の動画像符号化装置が利用するシンタクス2000を図20に例示する。
<Syntax structure 1>
Hereinafter, the syntax used by the image coding apparatus 100 in FIG. 1 will be described.
The syntax indicates the structure of encoded data (for example, encoded data 127 in FIG. 1) when the image encoding device encodes moving image data. When decoding the encoded data, the moving picture decoding apparatus interprets the syntax with reference to the same syntax structure. FIG. 20 shows an example of syntax 2000 used by the moving picture encoding apparatus of FIG.
 シンタクス2000は、ハイレベルシンタクス2001、スライスレベルシンタクス2002及びコーディングツリーレベルシンタクス2003の3つのパートを含む。ハイレベルシンタクス2001は、スライスよりも上位のレイヤのシンタクス情報を含む。スライスとは、フレームまたはフィールドに含まれる矩形領域もしくは連続領域を指す。スライスレベルシンタクス2002は、各スライスを復号化するために必要な情報を含む。コーディングツリーレベルシンタクス2003は、各コーディングツリー(即ち、各コーディングツリーユニット)を復号化するために必要な情報を含む。これら各パートは、さらに詳細なシンタクスを含む。 The syntax 2000 includes three parts: a high level syntax 2001, a slice level syntax 2002, and a coding tree level syntax 2003. The high level syntax 2001 includes syntax information of a layer higher than the slice. A slice refers to a rectangular area or a continuous area included in a frame or a field. The slice level syntax 2002 includes information necessary for decoding each slice. The coding tree level syntax 2003 includes information necessary for decoding each coding tree (ie, each coding tree unit). Each of these parts includes more detailed syntax.
 ハイレベルシンタクス2001は、シーケンスパラメータセットシンタクス2004及びピクチャパラメータセットシンタクス2005などの、シーケンス及びピクチャレベルのシンタクスを含む。スライスレベルシンタクス2002は、スライスヘッダーシンタクス2006及びスライスデータシンタクス2007などを含む。コーディングツリーレベルシンタクス2003は、コーディングツリーユニットシンタクス2008及びプレディクションユニットシンタクス2009などを含む。 The high level syntax 2001 includes sequence and picture level syntaxes such as a sequence parameter set syntax 2004 and a picture parameter set syntax 2005. The slice level syntax 2002 includes a slice header syntax 2006, a slice data syntax 2007, and the like. The coding tree level syntax 2003 includes a coding tree unit syntax 2008, a prediction unit syntax 2009, and the like.
 コーディングツリーユニットシンタクス2008は、四分木構造を持つことができる。具体的には、コーディングツリーユニットシンタクス2008のシンタクス要素として、さらにコーディングツリーユニットシンタクス2008を再帰呼び出しすることができる。即ち、1つのコーディングツリーユニットを四分木で細分化することができる。また、コーディングツリーユニットシンタクス2008内にはトランスフォームユニットシンタクス2010が含まれている。トランスフォームユニットシンタクス2010は、四分木の最末端の各コーディングツリーユニットシンタクス2008において呼び出される。トランスフォームユニットシンタクス2010は、逆直交変換及び量子化などに関わる情報が記述されている。 The coding tree unit syntax 2008 can have a quadtree structure. Specifically, the coding tree unit syntax 2008 can be recursively called as a syntax element of the coding tree unit syntax 2008. That is, one coding tree unit can be subdivided with a quadtree. The coding tree unit syntax 2008 includes a transform unit syntax 2010. The transform unit syntax 2010 is called at each coding tree unit syntax 2008 at the extreme end of the quadtree. The transform unit syntax 2010 describes information related to inverse orthogonal transformation and quantization.
 図21は、本実施形態に係るスライスヘッダーシンタクス2006を例示する。図21に示されるslice_bipred_intra_flagは、例えば、当該スライスに関して本実施形態に係る双方向イントラ予測の有効/無効を示すシンタクス要素である。 FIG. 21 illustrates the slice header syntax 2006 according to the present embodiment. The slice_bipred_intra_flag shown in FIG. 21 is a syntax element indicating, for example, validity / invalidity of bidirectional intra prediction according to the present embodiment for the slice.
 slice_bipred_intra_flagが0である場合、当該スライス内での本実施形態に係る双方向イントラは無効である。故に、直交変換部102及び逆直交変換部105は、単方向イントラ予測のみを行う。単方向イントラ予測の一例として、図9,10,11中のIntraBipredFlag[]が0となる予測や、H.264において規定されるイントラ予測を行っても構わない。 When slice_bipred_intra_flag is 0, the bidirectional intra according to this embodiment in the slice is invalid. Therefore, the orthogonal transform unit 102 and the inverse orthogonal transform unit 105 perform only unidirectional intra prediction. As an example of unidirectional intra prediction, prediction in which IntraBipredFlag [] in FIGS. Intra prediction specified in H.264 may be performed.
 一例として、slice_bipred_intra_flagが1である場合には、当該スライス内全域で本実施形態に係る双方向イントラ予測が有効となる。 As an example, when slice_bipred_intra_flag is 1, the bidirectional intra prediction according to the present embodiment is effective in the entire area in the slice.
 また、別の例として、slice_bipred_intra_flagが1である場合には、より下位のレイヤ(コーディングツリーユニット、トランスフォームユニットなど)のシンタクスにおいて当該スライス内部の局所領域毎に本実施形態に係る予測の有効/無効が規定されてもよい。 As another example, when slice_bipred_intra_flag is 1, in the syntax of a lower layer (coding tree unit, transform unit, etc.), the prediction validity / efficiency according to the present embodiment is determined for each local region in the slice. Invalidity may be specified.
 図22Aに、プレディクションユニットシンタクスの一例を示す。図中のpred_modeは当該プレディクションユニットの予測タイプをしている。MODE_INTRAは予測タイプがイントラ予測であることを示す。intra_split_flagは当該プレディクションユニットをさらに4つのプレディクションユニットに分割するか否かを示すフラグである。intra_split_flagが1の場合、プレディクションユニットを、縦横のサイズ半分で4分割したものをプレディクションユニットとする。intra_split_flagが0の場合、プレディクションユニットを分割しない。 FIG. 22A shows an example of the prediction unit syntax. Pred_mode in the figure indicates the prediction type of the prediction unit. MODE_INTRA indicates that the prediction type is intra prediction. intra_split_flag is a flag indicating whether or not the prediction unit is further divided into four prediction units. When intra_split_flag is 1, a prediction unit is obtained by dividing a prediction unit into four in half in the vertical and horizontal sizes. When intra_split_flag is 0, the prediction unit is not divided.
 intra_luma_bipred_flag[i]は当該プレディクションユニットに適用した予測モードIntraPredModeが単方向イントラ予測モードか双方向イントラ予測モードであるかを示すフラグである。iは分割されたプレディクションユニットの位置を示しており、前記intra_split_flagが0の場合には0、前記intra_split_flagが1の場合には0~3が設定される。当該フラグは、図9、10、11に示される当該プレディクションユニットのIntraBipredFlagの値がセットされている。 Intra_luma_bipred_flag [i] is a flag indicating whether the prediction mode IntraPredMode applied to the prediction unit is a unidirectional intra prediction mode or a bidirectional intra prediction mode. i indicates the position of the divided prediction unit. When the intra_split_flag is 0, 0 is set, and when the intra_split_flag is 1, 0 to 3 are set. The flag is set with the value of IntraBiprededFlag of the prediction unit shown in FIGS.
 intra_luma_bipred_flag[i]が1の場合、当該プレディクションユニットは双方向イントラ予測であることを示し、用意された複数の双方向イントラ予測モードの内、使用した双方向イントラ予測モードを特定する情報であるintra_luma_bipred_mode[i]を符号化する。intra_luma_bipred_mode[i]は、図7に示される双方向イントラ予測モード数IntraBiModeNumに従って等長符号化されてもよいし、予め決定された符号表を用いて符号化されてもよい。intra_luma_bipred_flag[i]が0の場合、当該プレディクションユニットは単方向イントラ予測であることを示し、隣接ブロックから予測符号化を行う。 When intra_luma_bipred_flag [i] is 1, this indicates that the prediction unit is bi-directional intra prediction, and is information that identifies the used bi-directional intra prediction mode among a plurality of prepared bi-directional intra prediction modes. Intra_luma_bipred_mode [i] is encoded. intra_luma_bipred_mode [i] may be encoded with the isometric length according to the number of bidirectional intra prediction modes IntraBiModeNum shown in FIG. 7, or may be encoded using a predetermined code table. When intra_luma_bipred_flag [i] is 0, it indicates that the prediction unit is unidirectional intra prediction, and predictive encoding is performed from adjacent blocks.
 prev_intra_luma_unipred_flag[i]は、隣接ブロックから計算される予測モードの予測値MostProbableと当該プレディクションユニットのイントラ予測モードが同一であるか否かを示すフラグである。MostProbableの計算方法の詳細は後述する。prev_intra_luma_unipred_flag[i]が1の場合、前記MostProbableとイントラ予測モードIntraPredModeが等しいことを示す。prev_intra_luma_unipred_flag[i]が0の場合、前記MostProbableとイントラ予測モードIntraPredModeは異なることを示し、イントラ予測モードIntraPredModeがさらにMostProbable以外のいずれのモードであるかを特定する情報rem_intra_luma_unipred_mode[i]が符号化される。rem_intra_luma_unipred_mode[i]は、図7に示される双方向イントラ予測モード数IntraUniModeNumに従って等長符号化されてもよいし、予め決定された符号表を用いて符号化されてもよい。イントラ予測モードIntraPredModeからrem_intra_luma_unipred_mode[i]は下式を用いて計算される。
Figure JPOXMLDOC01-appb-M000016
prev_intra_luma_unipred_flag [i] is a flag indicating whether or not the prediction value MostProbable of the prediction mode calculated from the adjacent block and the intra prediction mode of the prediction unit are the same. Details of the MostProbable calculation method will be described later. When prev_intra_luma_unipred_flag [i] is 1, it indicates that the MostProbable and the intra prediction mode IntraPredMode are equal. When prev_intra_luma_unipred_flag [i] is 0, it indicates that the MostProbable and the intra prediction mode IntraPredMode are different, and the information rem_intraiprecoded_code that specifies whether the intra prediction mode IntraPredMode is a mode other than MostProbable. . rem_intra_luma_unipred_mode [i] may be encoded with the isometric length according to the bidirectional intra prediction mode number IntraUniModeNum shown in FIG. 7, or may be encoded using a predetermined code table. From the intra prediction mode IntraPredMode, rem_intra_luma_unipred_mode [i] is calculated using the following equation.
Figure JPOXMLDOC01-appb-M000016
 次に、予測モードの予測値であるMostProbableの計算方法について説明する。MostProbableは下式に従って計算される。
Figure JPOXMLDOC01-appb-M000017
Next, a method for calculating MostProbable that is a predicted value in the prediction mode will be described. MostProbable is calculated according to the following equation.
Figure JPOXMLDOC01-appb-M000017
なお、Min(x,y)は入力x、yのうち小さい方を出力するパラメータである。 Min (x, y) is a parameter for outputting the smaller one of the inputs x and y.
 また、intraPredModeA,intraPredModeBは、符号化プレディクションユニットに対して左に及び上に隣接するプレディクションユニットのイントラ予測モードを示している。以降、intraPredModeAとintraPredModeBをまとめてintraPredModeNと表現する。NにはA若しくはBが設定される。図23に示すフローチャートを用いてintraPredModeNの計算方法を説明する。まず、隣接するプレディクションユニットが属するコーディングツリーユニットが利用可能であるかを判定し(ステップS2301)、当該コーディングツリーユニットが利用不可能である場合には(S2301がNO)、intraPredModeNに参照不可能を示す"-1"をセットする。一方、当該コーディングツリーユニットが利用可能である場合には(S2301がYES)、次に当該隣接プレディクションユニットがイントラ予測が適用されたか否かを判定する(ステップS2302)。当該隣接プレディクションユニットがイントラ予測ではない場合(S2302がNO)、intraPredModeNには"Intra_DC"を意味する"2"をセットする。一方、当該隣接プレディクションユニットがイントラ予測である場合(S2302がYES)、次に、当該隣接プレディクションユニットが双方向イントラ予測あるか否かを判定する(ステップS2303)。当該隣接プレディクションユニットが双方向イントラ予測ではない、つまり単方向イントラ予測の場合(S2303がNO)、intraPredModeNには隣接プレディクションユニットの予測モードIntraPredModeをセットする。一方、当該隣接PUが双方向イントラ予測である場号(S2303がYES)、隣接ブロックの予測モードを単方向イントラ予測モードに変換する。具体的には下式を用いてintraPredModeNを計算する。
Figure JPOXMLDOC01-appb-M000018
Also, intraPredModeA and intraPredModeB indicate intra prediction modes of prediction units adjacent to the left and above the encoded prediction unit. Hereinafter, intraPredModeA and intraPredModeB are collectively expressed as intraPredModeN. N is set to A or B. A method of calculating intraPredModeN will be described using the flowchart shown in FIG. First, it is determined whether a coding tree unit to which an adjacent prediction unit belongs can be used (step S2301). If the coding tree unit cannot be used (NO in S2301), reference to intraPredModeN is not possible. “−1” indicating “” is set. On the other hand, if the coding tree unit is available (YES in S2301), it is next determined whether or not intra prediction is applied to the adjacent prediction unit (step S2302). When the adjacent prediction unit is not intra prediction (NO in S2302), “2” meaning “Intra_DC” is set in intraPredModeN. On the other hand, when the adjacent prediction unit is intra prediction (YES in S2302), it is next determined whether or not the adjacent prediction unit is bidirectional intra prediction (step S2303). When the adjacent prediction unit is not bidirectional intra prediction, that is, in the case of unidirectional intra prediction (NO in S2303), the prediction mode IntraPredMode of the adjacent prediction unit is set in intraPredModeN. On the other hand, when the adjacent PU is bidirectional intra prediction (S2303 is YES), the prediction mode of the adjacent block is converted to the unidirectional intra prediction mode. Specifically, intraPredModeN is calculated using the following equation.
Figure JPOXMLDOC01-appb-M000018
ここで、IntraUniModeNumは当該隣接プレディクションユニットのサイズによって決定される単方向イントラ予測モードの数であり、図7にその一例が示されている。また、MappedBi2Uni(List, idx)は双方向イントラ予測モードを単方向イントラ予測モードに変換するテーブルである。Listは双方向イントラ予測モードを構成する2種類の単方向イントラ予測モードのうち、List0の単方向イントラ予測モード(図9,10,11に示されるIntraPredTypeL0[]に相当)を単方向イントラ予測モードへの変換に使用するか、List1の単方向イントラ予測モード図9,10,11に示されるIntraPredTypeL1[]に相当)を使用するかを示すフラグであり、0であればList0、1であればList1を単方向イントラ予測モードへの変換に使用する。図14に、変換テーブルの例が示されている。図中の数値は、図9,10,11に示されるIntraPredModeに対応している。 Here, IntraUniModeNum is the number of unidirectional intra prediction modes determined by the size of the adjacent prediction unit, and an example thereof is shown in FIG. Also, “MappedBi2Uni (List, idx)” is a table for converting the bidirectional intra prediction mode into the unidirectional intra prediction mode. List is the unidirectional intra prediction mode of List0 (corresponding to IntraPredTypeL0 [] shown in FIGS. 9, 10, and 11) of the two unidirectional intra prediction modes constituting the bidirectional intra prediction mode. Is a flag indicating whether to use the unidirectional intra prediction mode of List1 or the IntraPredTypeL1 [] shown in FIGS. 9, 10, and 11). List1 is used for conversion to the unidirectional intra prediction mode. FIG. 14 shows an example of the conversion table. The numerical values in the figure correspond to IntraPredMode shown in FIGS.
 以上のように計算されたintraPredModeNと数式(18)を用いて計算されたMostProbableが-1の場合には、MostProbableを2(Intra_DC)に置き換える。また、MostProbableが符号化プレディクションユニットの単方向イントラ予測モード数IntraUniPredModeNumよりも大きい場合には、下式を用いてMostProbableを再計算する。
Figure JPOXMLDOC01-appb-M000019
When MostProbable calculated using intraPredModeN calculated as described above and Expression (18) is −1, MostProbable is replaced with 2 (Intra_DC). Further, when MostProbable is larger than the number of unidirectional intra prediction modes IntraUniPredModeNum of the encoded prediction unit, MostProbable is recalculated using the following equation.
Figure JPOXMLDOC01-appb-M000019
MappedMostProbable()は、MostProbableを変換するテーブルであり、図24に一例が示されている。 “MappedMostProble ()” is a table for converting MostProbable, and an example is shown in FIG. 24.
 <シンタクス構成 2> 
 次に、プレディクションユニットシンタクスに係る別の例を図22Cに示す。pred_mode、intra_split_flagは先に述べたシンタクスの一例と同様であるので説明を省略する。luma_pred_mode_code_type[i]は当該プレディクションユニットに適用した予測モードIntraPredModeの種類を示しており、0(IntraUnipredMostProb)は単方向イントラ予測でイントラ予測モードがMostProbableと等しい、1(IntraUnipredRem)は単方向イントラ予測でイントラ予測モードがMostProbableと異なる、2(IntraBipred)は双方向イントラ予測モードであることを夫々示している。図24に双方向イントラ予測モード、1は及び単方向イントラ予測(MostProbableと同一の予測モードであるか否か)を示す。図25にluma_pred_mode_code_typeと対応する意味、及びbin、図7に示すモード構成に従ったモード数の割当の一例を示している。luma_pred_mode_code_type[i]が0の場合、イントラ予測モードがMostProbableモードとなるため、これ以上の情報の符号化は必要無い。luma_pred_mode_code_type[i]が1の場合、イントラ予測モードIntraPredModeがさらにMostProbable以外のいずれのモードであるかを特定する情報rem_intra_luma_unipred_mode[i]が符号化される。rem_intra_luma_unipred_mode[i]は、図7に示される双方向イントラ予測モード数IntraUniModeNumに従って等長符号化されてもよいし、予め決定された符号表を用いて符号化されてもよい。イントラ予測モードIntraPredModeからrem_intra_luma_unipred_mode[i]は数式(16)を用いて計算される。また、luma_pred_mode_code_type[i]が2の場合、当該プレディクションユニットは双方向イントラ予測であることを示し、用意された複数の双方向イントラ予測モードの内、使用した双方向イントラ予測モードを特定する情報であるintra_luma_bipred_mode[i]を符号化する。intra_luma_bipred_mode[i]は、図7に示される双方向イントラ予測モード数IntraBiModeNumに従って等長符号化されてもよいし、予め決定された符号表を用いて符号化されてもよい。 
 以上が、本実施の形態に係るシンタクス構成である。
<Syntax structure 2>
Next, another example of the prediction unit syntax is shown in FIG. 22C. Since pred_mode and intra_split_flag are the same as the syntax example described above, description thereof is omitted. luma_pred_mode_code_type [i] indicates the type of the prediction mode IntraPredMode applied to the prediction unit, where 0 (IntraUnifiedMostProb) is unidirectional intra prediction and the intra prediction mode is the same as MostProbable, 1 (IntraUnipre intrareprediction) The intra prediction mode is different from MostProbable, and 2 (IntraBipred) indicates a bidirectional intra prediction mode. FIG. 24 shows bidirectional intra prediction modes, 1 and unidirectional intra prediction (whether or not the same prediction mode as MostProbable). FIG. 25 shows an example of assignment of the number of modes according to the meaning corresponding to luma_pred_mode_code_type, bin, and the mode configuration shown in FIG. When luma_pred_mode_code_type [i] is 0, the intra prediction mode is the MostProbable mode, so no further information encoding is necessary. When luma_pred_mode_code_type [i] is 1, information rem_intra_luma_unipred_mode [i] that specifies which mode other than MostProbable is the intra prediction mode IntraPredMode is encoded. rem_intra_luma_unipred_mode [i] may be encoded with the isometric length according to the bidirectional intra prediction mode number IntraUniModeNum shown in FIG. 7, or may be encoded using a predetermined code table. From the intra prediction mode IntraPredMode, rem_intra_luma_unipred_mode [i] is calculated using Equation (16). Further, when luma_pred_mode_code_type [i] is 2, it indicates that the prediction unit is bidirectional intra prediction, and information that identifies the used bidirectional intra prediction mode among the prepared bidirectional intra prediction modes. Intra_luma_bipred_mode [i] is encoded. intra_luma_bipred_mode [i] may be encoded with the isometric length according to the number of bidirectional intra prediction modes IntraBiModeNum shown in FIG. 7, or may be encoded using a predetermined code table.
The above is the syntax configuration according to the present embodiment.
 <シンタクス構成 3> 
 プレディクションユニットシンタクスに係るさらに別の例を図22Dに示す。本例では、図22Aで示されるプレディクションユニットシンタクスをベースに、双方向イントラ予測を使用可能とするか、双方向イントラ予測を使用不可能として従来の単方向イントラ予測のみを使用可能とするかを、符号化プレディクションユニット内で切り替える場合のシンタクスを示している。尚、双方向イントラ予測を使用不可能として従来の単方向イントラ予測のみを使用可能とする場合、図9に代わって図43で示すテーブルを使ってもよいし、図9のIntraPredModeが33以上のテーブルを無視する構成としてもよい。図43は、図9内から双方向イントラ予測時の第二の予測モードに関する情報を示すIntraPredTypeL1、IntraPredAngleIdL1を削除するとともに、不要なIntraPredModeが33以上のテーブルを削除したものである。図42と図9に関する同様な構成を、図10、図11に関しても適用可能である。 
 なお、pred_mode、intra_split_flagは先に述べたシンタクスの一例と同様であるので説明を省略する。
<Syntax structure 3>
Yet another example of the prediction unit syntax is shown in FIG. 22D. In this example, based on the prediction unit syntax shown in FIG. 22A, whether bidirectional intra prediction can be used or whether conventional intra unidirectional prediction can be used with bidirectional intra prediction disabled. Shows the syntax for switching within the encoding prediction unit. In the case where the bidirectional intra prediction cannot be used and only the conventional unidirectional intra prediction can be used, the table shown in FIG. 43 may be used instead of FIG. 9, and the IntraPredMode in FIG. The table may be ignored. FIG. 43 is a table in which IntraPredTypeL1 and IntraPredAngleIdL1 indicating information related to the second prediction mode at the time of bidirectional intra prediction are deleted from FIG. 9, and unnecessary IntraPredMode 33 or more tables are deleted. The same structure regarding FIG. 42 and FIG. 9 is applicable also regarding FIG. 10, FIG.
Note that pred_mode and intra_split_flag are the same as the syntax example described above, and thus description thereof is omitted.
 intra_bipred_flagは、符号化プレディクションユニット内で双方向イントラ予測を使用可能とするか否かを示すフラグである。intra_bipred_flagが0の場合、符号化プレディクションユニット内で双方向イントラ予測が使用されないことを示している。intra_split_flagが1、つまり符号化プレディクションユニットがさらに4分割されている場合においても、全てのプレディクションユニットにおいて双方向イントラ予測は使用されず、単方向イントラ予測のみが有効となる。 Intra_bipred_flag is a flag indicating whether or not bi-directional intra prediction can be used in the encoded prediction unit. When intra_bipred_flag is 0, it indicates that bi-directional intra prediction is not used in the encoded prediction unit. Even when intra_split_flag is 1, that is, when the encoded prediction unit is further divided into four, bi-directional intra prediction is not used in all prediction units, and only uni-directional intra prediction is effective.
 intra_bipred_flagが1の場合、符号化プレディクションユニット内で双方向イントラ予測が使用可能であることを示している。intra_split_flagが1、つまり符号化プレディクションユニットがさらに4分割されている場合においても、全てのプレディクションユニットにおいて、単方向イントラ予測に加え双方向イントラ予測が選択可能となる。 When intra_bipred_flag is 1, it indicates that bi-directional intra prediction can be used in the encoded prediction unit. Even when intra_split_flag is 1, that is, when the encoded prediction unit is further divided into four, in all prediction units, bidirectional intra prediction can be selected in addition to unidirectional intra prediction.
 双方向イントラ予測が不要である予測が比較的容易な領域(例えば、平坦領域)では、intra_bipred_flagを0として符号化して、双方向イントラ予測を使用不可能とすることにより、双方向イントラ予測モードの符号化に必要な符号量が削減可能となるため、符号化効率は改善する。 In a region where bi-directional intra prediction is not necessary (for example, a flat region), intra_bipred_flag is encoded as 0 to disable bi-directional intra prediction. Since the amount of codes necessary for encoding can be reduced, encoding efficiency is improved.
 <シンタクス構成 4> 
 プレディクションユニットシンタクスに係るさらに別の例を図22Eに示す。本例では、図22Cで示されるプレディクションユニットシンタクスをベースに、双方向イントラ予測を使用可能とするか、双方向イントラ予測を使用不可能として従来の単方向イントラ予測のみを使用可能とするかを、符号化プレディクションユニット内で切り替える場合のシンタクスを示している。intra_bipred_flagは、符号化プレディクションユニット内で双方向イントラ予測を使用可能とするか否かを示すフラグであり、前述のintra_bipred_flagと同様であるので説明を省略する。
<Syntax structure 4>
Still another example relating to the prediction unit syntax is shown in FIG. 22E. In this example, based on the prediction unit syntax shown in FIG. 22C, whether bidirectional intra prediction can be used or whether only conventional unidirectional intra prediction can be used with bidirectional intra prediction disabled. Shows the syntax for switching within the encoding prediction unit. intra_bipred_flag is a flag indicating whether or not bi-directional intra prediction can be used in the encoding prediction unit, and is the same as the above-described intra_bipred_flag, and thus the description thereof is omitted.
 (第1の変形例) 
 <イントラ予測部の第1の変形例> 
 イントラ予測部109に係る第1の変形例として、JCTVC-B205_draft002, 5.2.1節" Intra prediction process for luma samples ", JCT-VC 2nd Meeting Geneva, July, 2010に示される適応参照画素フィルタリングと組み合わせても構わない。図26に、適応参照画素フィルタリングを用いた際のイントラ予測部109を示している。図6で示されるイントラ予測部109とは、参照画素フィルタ部2601が追加されている点が異なる。参照画素フィルタ部2601では、参照画像信号124及び予測モード605を入力し、後述する適応的なフィルタ処理を行い、被フィルタ参照画像信号2602を出力する。被フィルタ参照画像信号2602は単方向イントラ予測画像生成部601と双方向イントラ予測画像生成部602に入力される。参照画素フィルタ部2601以外の構成及び処理に関しては、図6で示されるイントラ予測部109と同様であるので、説明を省略する。
(First modification)
<First Modification of Intra Prediction Unit>
As a first modification related to the intra prediction unit 109, in combination with adaptive reference pixel filtering shown in JCTVC-B205_draft002, section 5.2.1 “Intra prediction process for luma samples”, JCT-VC 2nd Meeting Geneva, July, 2010 It doesn't matter. FIG. 26 shows the intra prediction unit 109 when adaptive reference pixel filtering is used. 6 is different from the intra prediction unit 109 shown in FIG. 6 in that a reference pixel filter unit 2601 is added. The reference pixel filter unit 2601 receives the reference image signal 124 and the prediction mode 605, performs an adaptive filtering process described later, and outputs a filtered reference image signal 2602. The filtered reference image signal 2602 is input to the unidirectional intra predicted image generation unit 601 and the bidirectional intra predicted image generation unit 602. The configuration and processing other than the reference pixel filter unit 2601 are the same as those of the intra prediction unit 109 shown in FIG.
 次に参照画素フィルタ部2601について説明する。参照画素フィルタ部2601では、予測モード605に含まれる参照画素フィルタフラグ及びイントラ予測モードに従って、イントラ予測に使用する参照画素をフィルタリングするか否かを決定する。参照画素フィルタフラグはイントラ予測モードIntraPredModeが”Intra_DC”以外の値の場合に、参照画素をフィルタリングするか否かを示すフラグである。参照画素フィルタフラグが1の場合、参照画素をフィルタリングする。また、参照画素フィルタフラグ0の場合、参照画素をフィルタリングしない。なお、IntraPredModeが”Intra_DC”の場合には、参照画素はフィルタリングせず、参照画素フィルタフラグは0にセットされる。参照画素フィルタフラグが1の場合、以下のフィルタリングにより、被フィルタ参照画像信号2602が計算される。なお、p[x,y]はフィルタ前の参照画素、pf[x,y]はフィルタ的用語の参照画素を示している。また、x,yは当該プレディクションユニット内の左上画素位置をx=0,y=0とした場合の、参照画素の相対位置を示している。また、PuPartSizeは当該プレディクションユニットのサイズ(画素)を示している。
Figure JPOXMLDOC01-appb-M000020
Next, the reference pixel filter unit 2601 will be described. The reference pixel filter unit 2601 determines whether or not to filter reference pixels used for intra prediction according to the reference pixel filter flag and the intra prediction mode included in the prediction mode 605. The reference pixel filter flag is a flag indicating whether or not reference pixels are filtered when the intra prediction mode IntraPredMode is a value other than “Intra_DC”. When the reference pixel filter flag is 1, the reference pixel is filtered. In the case of the reference pixel filter flag 0, the reference pixel is not filtered. When IntraPredMode is “Intra_DC”, the reference pixel is not filtered and the reference pixel filter flag is set to 0. When the reference pixel filter flag is 1, a filtered reference image signal 2602 is calculated by the following filtering. Note that p [x, y] indicates a reference pixel before filtering, and pf [x, y] indicates a reference pixel in filter terms. Further, x and y indicate relative positions of the reference pixels when the upper left pixel position in the prediction unit is x = 0 and y = 0. PuPartSize indicates the size (pixel) of the prediction unit.
Figure JPOXMLDOC01-appb-M000020
 <シンタクス構成 5> 
 図27A,図27Bには、適応参照画素フィルタを行う際のプレディクションユニットシンタクス構造を示している。図27Aは、図22Aに適応参照画素フィルタに関するシンタクスintra_luma_filter_flag[i]を追加している。また、図27Bは、図22Cに適応参照画素フィルタに関するシンタクスintra_luma_filter_flag[i]を追加している。intra_luma_filter_flag[i]はイントラ予測モードIntraPredMode[i]がIntra_DC以外の場合に、さらに符号化される。当該フラグが0の場合、上記参照画素のフィルタリングは行われないことを示す。また、intra_luma_filter_flag[i]が1の場合、上記参照画素のフィルタリングを適用することを示す。
<Syntax structure 5>
27A and 27B show a prediction unit syntax structure when performing adaptive reference pixel filtering. FIG. 27A adds the syntax intra_luma_filter_flag [i] related to the adaptive reference pixel filter to FIG. 22A. In addition, FIG. 27B adds syntax intra_luma_filter_flag [i] related to the adaptive reference pixel filter to FIG. 22C. intra_luma_filter_flag [i] is further encoded when the intra prediction mode IntraPredMode [i] is other than Intra_DC. When the flag is 0, it indicates that the reference pixel is not filtered. Further, when intra_luma_filter_flag [i] is 1, it indicates that the reference pixel filtering is applied.
 上記の例では、イントラ予測モードIntraPredMode[i]がIntra_DC以外の場合に、intra_luma_filter_flag[i]を符号化したが、他の例として、IntraPredMode[i]が0~2の場合には、intra_luma_filter_flag[i]を符号化しなくても構わない。この場合、intra_luma_filter_flag[i]は0に設定される。 In the above example, intra_luma_filter_flag [i] is encoded when the intra prediction mode IntraPredMode [i] is other than Intra_DC. As another example, when IntraPredMode [i] is 0 to 2, intra_luma_filter_flag [i ] May not be encoded. In this case, intra_luma_filter_flag [i] is set to 0.
 また、図22B、図22D、図22Eに示される他のシンタクス構造について、上記で説明したintra_luma_filter_flag[i]を同様の意味で追加しても構わない。 In addition, the intra_luma_filter_flag [i] described above may be added in the same meaning for the other syntax structures shown in FIGS. 22B, 22D, and 22E.
 (第2の変形例) 
 <イントラ予測部の第二の変形例> 
 イントラ予測部109に係る第2の変形例として、JCTVC-B205_draft002, 9.6節"Combined Intra Prediction", JCT-VC 2nd Meeting Geneva, July, 2010に示される複合イントラ予測と組み合わせて使用しても構わない。この文献における復号イントラ予測は、前述の単方向イントラ予測の結果と、予測画素に対し左、上、左上に隣接する画素の平均値とを加重平均することにより、予測値を得る。動画像復号化装置4400や画像符号化装置100内において復号画像信号122を計算した場合には、前記左、上、左上に隣接する画素として、復号画素を用いることが可能である。一方、画像符号化装置100において、復号画像信号122を計算する前では、復号画素を用いることが不可能であるため、前記左、上、左上に隣接する画素として入力画像信号116を用いる。図28には、予測対象画素Xの予測に用いる隣接復号画素A(左),B(上),C(左上)の位置が示されている。従って、複合イントラ予測は画像符号化装置100と動画像復号化装置4400で予測値が異なる、所謂オープンループの予測方法である。
(Second modification)
<Second Modification of Intra Prediction Unit>
As a second modification of the intra prediction unit 109, it may be used in combination with the composite intra prediction shown in JCTVC-B205_draft002, section 9.6 “Combined Intra Prediction”, JCT-VC 2nd Meeting Geneva, July, 2010 . In the decoded intra prediction in this document, a prediction value is obtained by performing weighted averaging of the result of the above-described unidirectional intra prediction and the average value of pixels adjacent to the left, top, and top left with respect to the prediction pixel. When the decoded image signal 122 is calculated in the moving image decoding device 4400 or the image encoding device 100, it is possible to use decoded pixels as pixels adjacent to the left, upper, and upper left. On the other hand, since it is impossible to use decoded pixels before the decoded image signal 122 is calculated in the image encoding device 100, the input image signal 116 is used as a pixel adjacent to the left, upper, and upper left. FIG. 28 shows positions of adjacent decoded pixels A (left), B (upper), and C (upper left) used for prediction of the prediction target pixel X. Therefore, composite intra prediction is a so-called open-loop prediction method in which prediction values differ between the image encoding device 100 and the moving image decoding device 4400.
 図30に、複合イントラ予測と組み合わせた場合のイントラ予測部4408(109)のブロック図を示す。図6で示されるイントラ予測部109に複合イントラ予測画像生成部2901、選択スイッチ2902、および復号画像バッファ3001が追加されている点が異なる。 FIG. 30 shows a block diagram of the intra prediction unit 4408 (109) when combined with composite intra prediction. A difference is that a composite intra predicted image generation unit 2901, a selection switch 2902, and a decoded image buffer 3001 are added to the intra prediction unit 109 shown in FIG.
 双方向イントラ予測と複合イントラ予測を組み合わせた場合、まず選択スイッチ604において、符号化制御部115で制御されている予測モード情報に従って、単方向イントラ予測画像生成部601若しくは双方向イントラ予測画像生成部602の出力端を切り替える。以後、出力された当該予測画像信号126を方向予測画像信号126と呼ぶ。 When the bidirectional intra prediction and the composite intra prediction are combined, first, in the selection switch 604, the unidirectional intra prediction image generation unit 601 or the bidirectional intra prediction image generation unit according to the prediction mode information controlled by the encoding control unit 115. The output terminal of 602 is switched. Hereinafter, the output predicted image signal 126 is referred to as a direction predicted image signal 126.
 その後、前記方向予測画像信号を複合イントラ予測画像生成部2901に入力し、複合イントラ予測における予測画像信号4420(126)を生成する。複合イントラ予測画像生成部2901における説明は後述する。その後、選択スイッチ2902において、符号化制御部115で制御されている予測モード情報内の複合イントラ予測適用フラグに従って、複合イントラ予測における予測画像信号4420(126)と方向予測画像信号のいずれを用いるかを切り替えて、イントラ予測部4408(109)における最終的な予測画像信号4420(126)が出力される。複合イントラ予測適用フラグが1の場合には、複合イントラ予測画像生成部2901から出力される予測画像信号4420(126)が最終的な予測画像信号4420(126)となる。一方、複合イントラ予測適用フラグが0の場合には、方向予測画像信号4420(126)が最終的に出力される予測画像信号4420(126)となる。複合イントラ予測画像生成部2901から出力される予測画像信号は第六予測画像信号とも呼ばれる。 Then, the direction prediction image signal is input to the composite intra prediction image generation unit 2901, and a prediction image signal 4420 (126) in the composite intra prediction is generated. The description of the composite intra predicted image generation unit 2901 will be described later. Thereafter, in the selection switch 2902, which one of the prediction image signal 4420 (126) and the direction prediction image signal in the composite intra prediction is used according to the composite intra prediction application flag in the prediction mode information controlled by the encoding control unit 115. And the final predicted image signal 4420 (126) in the intra prediction unit 4408 (109) is output. When the composite intra prediction application flag is 1, the predicted image signal 4420 (126) output from the composite intra predicted image generation unit 2901 is the final predicted image signal 4420 (126). On the other hand, when the composite intra prediction application flag is 0, the direction prediction image signal 4420 (126) is the prediction image signal 4420 (126) that is finally output. The predicted image signal output from the composite intra predicted image generation unit 2901 is also called a sixth predicted image signal.
 複合イントラ予測画像生成部2901から予測画像信号4420(126)が生成された場合、加算部4405にて別途復号された復元予測誤差信号4416と画素単位に加算され画素単位の復号画像信号4417を生成し、復号画素バッファ3001に格納する。格納された画素単位の復号画像信号4417は、参照画素3002として複合イントラ予測画像生成部2901に入力され、図31に示される隣接画素3104として、後述する画素レベルの予測に使用される。 When the prediction image signal 4420 (126) is generated from the composite intra prediction image generation unit 2901, the decoded prediction error signal 4416 separately decoded by the addition unit 4405 is added to the pixel unit to generate a decoded image signal 4417 for each pixel. And stored in the decoded pixel buffer 3001. The stored decoded image signal 4417 in units of pixels is input to the composite intra predicted image generation unit 2901 as the reference pixel 3002, and is used for pixel level prediction described later as the adjacent pixel 3104 shown in FIG.
 次に、複合イントラ予測画像生成部2901について図31を用いて説明する。複合イントラ予測画像生成部2901は画素レベル予測信号生成部3101及び複合イントラ予測計算部3102を含む。画素レベル予測信号生成部3101では、隣接画素3104として参照画素3002を入力し、予測対象画素Xを隣接する画素から予測することによって画素レベル予測信号3103を出力する。具体的には、数式(21)を用いて隣接画素3104を示すA,B,Cから予測対象画素の画素レベル予測信号3103(X)を計算する。
Figure JPOXMLDOC01-appb-M000021
Next, the composite intra prediction image generation unit 2901 will be described with reference to FIG. The composite intra prediction image generation unit 2901 includes a pixel level prediction signal generation unit 3101 and a composite intra prediction calculation unit 3102. The pixel level prediction signal generation unit 3101 receives the reference pixel 3002 as the adjacent pixel 3104, and outputs the pixel level prediction signal 3103 by predicting the prediction target pixel X from the adjacent pixel. Specifically, the pixel level prediction signal 3103 (X) of the prediction target pixel is calculated from A, B, and C indicating the adjacent pixel 3104 using Expression (21).
Figure JPOXMLDOC01-appb-M000021
 なお、A,B,Cに係る係数は、他の値であっても構わない。 Note that the coefficients related to A, B, and C may be other values.
 複合イントラ予測計算部3102では、方向予測画像信号126(X’)と画素レベル予測信号3103(X)の加重平均を行い、最終的な予測画像信号126(P)を出力する。具体的には、下式を用いる。
Figure JPOXMLDOC01-appb-M000022
The composite intra prediction calculation unit 3102 performs a weighted average of the direction prediction image signal 126 (X ′) and the pixel level prediction signal 3103 (X), and outputs a final prediction image signal 126 (P). Specifically, the following formula is used.
Figure JPOXMLDOC01-appb-M000022
 なお、Wは方向予測画像信号126(X’)と画素レベル予測信号3103(X)との加重平均の重み係数(W=0~32の間の整数値)である。 Note that W is a weighted average weighting factor (an integer value between W = 0 and 32) of the direction prediction image signal 126 (X ′) and the pixel level prediction signal 3103 (X).
 当該複合イントラ予測を使用して予測画像信号126を生成し、さらに予測誤差信号117、復号画像信号122を生成した場合、符号化と復号化で前記復号画像信号122が異なる値を持つ場合がある。そこで、符号化プレディクションシンタクス内の全ての復号画像信号122が生成された後、復号画像信号122を隣接画素として再度上述の複合イントラ予測を実行することで、復号化と同一の予測画像信号126が生成され、さらに予測誤差信号117と加算することにより、復号化と同一の前記復号画像信号122を生成することが可能となる。 When the prediction image signal 126 is generated using the composite intra prediction, and the prediction error signal 117 and the decoded image signal 122 are further generated, the decoded image signal 122 may have different values in encoding and decoding. . Therefore, after all the decoded image signals 122 in the encoded prediction syntax are generated, the above-described combined intra prediction is performed again using the decoded image signal 122 as an adjacent pixel, so that the same predicted image signal 126 as that in the decoding is obtained. Is further added to the prediction error signal 117, and the decoded image signal 122 identical to the decoding can be generated.
 以上が、複合イントラ予測と組み合わせた場合の実施形態である。 The above is an embodiment when combined with composite intra prediction.
 (第3の変形例) 
 <イントラ予測部の第三の変形例> 
 また、上記重み係数Wはプレディクションユニット内の予測画素の位置に応じて、切り替えても構わない。一般に、単方向イントラ予測及び双方向イントラ予測を用いて生成された予測画像信号は、空間的に隣接している既に符号化済みの上或いは左に位置する参照画素から予測値が生成されるため、参照画素からの距離が離れるほど予測誤差の絶対値が増加する傾向にある。従って、方向予測画像信号126と画素レベル予測信号3103の重み係数を、参照画素に近い場合は方向予測画像信号126の重み係数を大きくし、離れた場合は、小さくすることで予測精度を向上させることが可能となる。
(Third Modification)
<Third Modification of Intra Prediction Unit>
The weighting factor W may be switched according to the position of the prediction pixel in the prediction unit. In general, a prediction image signal generated using unidirectional intra prediction and bidirectional intra prediction generates a prediction value from spatially adjacent reference pixels positioned on the left or above already encoded. The absolute value of the prediction error tends to increase as the distance from the reference pixel increases. Therefore, the weighting coefficient of the direction prediction image signal 126 and the pixel level prediction signal 3103 is increased when the weight coefficient of the direction prediction image signal 126 is close to the reference pixel, and is decreased when the distance is far away, thereby improving the prediction accuracy. It becomes possible.
 一方、当該複合イントラ予測では、符号化時に入力画像信号を利用した予測誤差信号の生成を行う。この際、画素レベル予測信号3103は入力画像信号となるため、参照画素位置と予測画素位置との空間距離が大きくなっても、方向予測画像信号126と比較して、画素レベル予測信号3103の予測精度が高い。しかし、単純に方向予測画像信号126と画素レベル予測信号3103の重み係数を、参照画素に近い場合は方向予測画像信号126の重み係数を大きくし、離れた場合は、小さくすると、離れた場合の予測誤差が小さくなるが、符号化時の予測値と局部復号化時の予測値に乖離が生まれ、予測精度が低下する問題が生じる。従って、特に量子化パラメータの値が大きい場合に、参照画素位置と予測画素位置との空間距離が大きくなるにつれ、Wの値を小さく設定することにより、このようなオープンループの場合に発生する乖離現象による符号化効率低下を抑えることが出来る。 On the other hand, in the complex intra prediction, a prediction error signal is generated using an input image signal at the time of encoding. At this time, since the pixel level prediction signal 3103 becomes an input image signal, even if the spatial distance between the reference pixel position and the prediction pixel position is increased, the prediction of the pixel level prediction signal 3103 is compared with the direction prediction image signal 126. High accuracy. However, the weighting coefficient of the direction prediction image signal 126 and the pixel level prediction signal 3103 is simply increased when the weight coefficient of the direction prediction image signal 126 is close to the reference pixel, and is decreased when the distance is small. Although the prediction error is reduced, there is a problem that the prediction accuracy at the time of encoding and the prediction value at the time of local decoding are different and the prediction accuracy is lowered. Therefore, especially when the value of the quantization parameter is large, as the spatial distance between the reference pixel position and the predicted pixel position becomes large, the difference generated in the case of such an open loop is set by setting the value of W small. A decrease in coding efficiency due to the phenomenon can be suppressed.
 <シンタクス構成 6> 
 図32A,図32Bには、複合イントラ予測を行う際のプレディクションユニットシンタクス構造を示している。図32Aは、図22Aと比較して複合イントラ予測の有無を切り替えるシンタクスcombined_intra_pred_flagを追加している点で異なる。これは、前述の複合イントラ予測適用フラグと等しい。また、図32Bは、図22Cに複合イントラ予測の有無を切り替えるシンタクスcombined_intra_pred_flagを追加している。combined_intra_pred_flagが1の場合、図30に示される選択スイッチ2902は複合イントラ予測画像生成部2901の出力端に接続される。combined_intra_pred_flagが0の場合、図30に示される選択スイッチ2902は選択スイッチ604が接続している単方向イントラ予測画像生成部601若しくは双方向イントラ予測画像生成部602のいずれかの出力端に接続される。
<Syntax structure 6>
32A and 32B show a prediction unit syntax structure when performing composite intra prediction. FIG. 32A is different from FIG. 22A in that a syntax combined_intra_pred_flag for switching presence / absence of composite intra prediction is added. This is equivalent to the above-described composite intra prediction application flag. In addition, FIG. 32B adds a syntax combined_intra_pred_flag for switching presence / absence of composite intra prediction to FIG. 22C. When combined_intra_pred_flag is 1, the selection switch 2902 shown in FIG. 30 is connected to the output terminal of the composite intra prediction image generation unit 2901. When combined_intra_pred_flag is 0, the selection switch 2902 shown in FIG. 30 is connected to the output terminal of either the unidirectional intra prediction image generation unit 601 or the bidirectional intra prediction image generation unit 602 to which the selection switch 604 is connected. .
 また、図22B、図22D、図22Eに示される他のシンタクス構造について、上記で説明したintra_luma_filter_flag[i]を同様の意味で追加しても構わない。 In addition, the intra_luma_filter_flag [i] described above may be added in the same meaning for the other syntax structures shown in FIGS. 22B, 22D, and 22E.
 さらに、イントラ予測部の第2の実施形態と組み合わせても構わない。以上が、イントラ予測部109の別の実施形態に関する説明である。 Furthermore, it may be combined with the second embodiment of the intra prediction unit. The above is the description regarding another embodiment of the intra prediction unit 109.
 以上の第1の実施形態によれば、高効率なイントラ予測を実現することができる。故に、符号化効率が向上し、ひいては主観画質も向上する。 According to the first embodiment described above, highly efficient intra prediction can be realized. Therefore, the coding efficiency is improved, and the subjective image quality is also improved.
 (第2の実施形態) 
 <動画像符号化装置-第二の実施形態> 
 第2の実施形態に係る動画像符号化装置は、前述の第1の実施形態に係る画像符号化装置と直交変換及び逆直交変換の詳細において異なる。以降の説明では、本実施形態において第1の実施形態と同一部分には同一のインデックスを付して示し、異なる部分を中心に説明する。本実施形態に係る画像符号化装置に対応する動画像復号化装置は、第5の実施形態において説明する。
(Second Embodiment)
<Moving Image Encoding Device—Second Embodiment>
The video encoding apparatus according to the second embodiment differs from the above-described image encoding apparatus according to the first embodiment in the details of orthogonal transform and inverse orthogonal transform. In the following description, in the present embodiment, the same parts as those in the first embodiment are denoted by the same indexes, and different parts will be mainly described. A moving picture decoding apparatus corresponding to the picture encoding apparatus according to the present embodiment will be described in a fifth embodiment.
 図33は、第2の実施形態に係る動画像符号化装置を示すブロック図である。第1の実施形態に係る動画像符号化装置からの変更点は、変換選択部3301、係数順制御部3302が追加されている点である。また、直交変換部102及び逆直交変換部105の内部構造も異なる。以下、図33の処理について説明する。 FIG. 33 is a block diagram showing a video encoding apparatus according to the second embodiment. The change from the moving picture encoding apparatus according to the first embodiment is that a transformation selection unit 3301 and a coefficient order control unit 3302 are added. Also, the internal structures of the orthogonal transform unit 102 and the inverse orthogonal transform unit 105 are different. Hereinafter, the process of FIG. 33 will be described.
 先ず、直交変換部102と逆直交変換部105をそれぞれ図34、図35で説明する。 First, the orthogonal transform unit 102 and the inverse orthogonal transform unit 105 will be described with reference to FIGS.
 <直交変換部102> 
 図34の直交変換部102は、第一直交変換部3401と第二直交変換部3402、第N直交変換部3403、変換選択スイッチ3404を有する。なお、ここでは、N種類の直交変換部を持つ例を示しているが、同じ直交変換方法で変換サイズが複数あってもよいし、異なる直交変換方法を行う直交変換部が複数あってもよい。また、それぞれが混ざっていてもよい。例えば、第一直交変換部3401が4×4サイズDCT、第二直交変換部3402が8×8サイズDCT、第N直交変換部3403が16×16サイズDCTと設定することも可能であるし、第一直交変換部3401が4×4サイズDCT、第二直交変換部3402が4×4サイズDST(離散サイン変換)、第N直交変換部3403が8×8サイズKLT(カルーネン・レーベ変換)と設定することも可能である。また、直交変換でない変換も選択することも可能であし、単一の変換であってもよい。この場合N=1と考える。
<Orthogonal transform unit 102>
34 includes a first orthogonal transform unit 3401, a second orthogonal transform unit 3402, an Nth orthogonal transform unit 3403, and a transform selection switch 3404. Although an example having N types of orthogonal transform units is shown here, there may be a plurality of transform sizes using the same orthogonal transform method, or there may be a plurality of orthogonal transform units performing different orthogonal transform methods. . Moreover, each may be mixed. For example, the first orthogonal transform unit 3401 can be set to 4 × 4 size DCT, the second orthogonal transform unit 3402 can be set to 8 × 8 size DCT, and the Nth orthogonal transform unit 3403 can be set to 16 × 16 size DCT. The first orthogonal transform unit 3401 is 4 × 4 size DCT, the second orthogonal transform unit 3402 is 4 × 4 size DST (discrete sine transform), and the Nth orthogonal transform unit 3403 is 8 × 8 size KLT (Karunen-Labe transform). ) Can also be set. In addition, it is possible to select a transform that is not orthogonal transform, or a single transform. In this case, N = 1 is considered.
 先ず、変換選択スイッチ3404について説明する。変換選択スイッチ3404は、減算部101の出力端を、変換選択情報3303に従って選択する機能を有する。変換選択情報3303は、符号化制御部115で制御されている情報の1つであり、予測情報125に従って変換選択部3301で設定される。例えば、H.264では、4×4画素ブロック(プレディクションユニット)のイントラ予測時には4×4DCTが、8×8画素ブロック(プレディクションユニット)のイントラ予測時には8×8DCTが設定される。本実施の形態では、変換選択情報3303が第一直交変換を指す場合には、スイッチの出力端を第一直交変換部3401に接続する。一方、変換選択情報3303が第二直交変換である場合は、出力端を第二直交変換部3402に接続する。 First, the conversion selection switch 3404 will be described. The conversion selection switch 3404 has a function of selecting the output terminal of the subtraction unit 101 according to the conversion selection information 3303. The transformation selection information 3303 is one piece of information controlled by the encoding control unit 115 and is set by the transformation selection unit 3301 according to the prediction information 125. For example, H.M. In H.264, 4 × 4 DCT is set for intra prediction of a 4 × 4 pixel block (prediction unit), and 8 × 8 DCT is set for intra prediction of an 8 × 8 pixel block (prediction unit). In the present embodiment, when the transformation selection information 3303 indicates the first orthogonal transformation, the output terminal of the switch is connected to the first orthogonal transformation unit 3401. On the other hand, when the transformation selection information 3303 is the second orthogonal transformation, the output end is connected to the second orthogonal transformation unit 3402.
 次に第一直交変換部3401ないし第N直交変換部3403の処理について説明する。本実施の形態では、N個ある直交変換部の中で1つがDCT、それ以外がKLT(カルーネン・レーベ変換)である例について説明する。ここでは、第一直交変換部3401がDCT、それ以外の直交変換部3402、3403ではKLT(カルーネン・レーベ変換)が行われる。 Next, processing of the first orthogonal transform unit 3401 to the Nth orthogonal transform unit 3403 will be described. In the present embodiment, an example in which one of N orthogonal transform units is DCT and the other is KLT (Karunen-Loeve transform) will be described. Here, the first orthogonal transform unit 3401 performs DCT, and the other orthogonal transform units 3402 and 3403 perform KLT (Carhunen-Labe transform).
 <逆直交変換部105> 
 図35の逆直交変換部105は、第一逆直交変換部3501、第二逆直交変換部3502、第N逆直交変換部3503、変換選択スイッチ3504を有する。先ず、変換選択スイッチ3504について説明する。変換選択スイッチ3504は、逆量子化部104の出力端を、入力された変換選択情報3303に従って選択する機能を有する。変換選択情報3303は、符号化制御部115で制御されている情報の1つであり、予測情報125に従って変換選択部3301で設定される。
<Inverse orthogonal transform unit 105>
The inverse orthogonal transform unit 105 in FIG. 35 includes a first inverse orthogonal transform unit 3501, a second inverse orthogonal transform unit 3502, an Nth inverse orthogonal transform unit 3503, and a transform selection switch 3504. First, the conversion selection switch 3504 will be described. The conversion selection switch 3504 has a function of selecting the output terminal of the inverse quantization unit 104 according to the input conversion selection information 3303. The transformation selection information 3303 is one piece of information controlled by the encoding control unit 115 and is set by the transformation selection unit 3301 according to the prediction information 125.
 変換選択情報3303が第一直交変換である場合は、スイッチの出力端を第一逆直交変換部3501に接続する。一方、変換選択情報3303が第二直交変換である場合は、出力端を第二逆直交変換部3502に接続する。同様に、変換選択情報3303が第N直交変換である場合は、出力端を第N逆直交変換部3503に接続する。ここで、直交変換部102に設定される変換選択情報3303と逆直交変換部105に設定される変換選択情報3303は同一であり、直交変換部102で行われた変換に対応する逆直交変換が逆直交変換部105で同期して行われる。つまり、第一逆直交変換部3501では逆離散コサイン変換(以下、IDCTという)が行われ、第二逆直交変換部3502、第N逆直交変換部3503ではKLT(カルーネン・レーベ変換)に基づいた逆変換が行われる。ここでは例としてIDCTなどを用いる例を示したが、アダマール変換や離散サイン変換などの直交変換を使ってもよいし、非直交変換を用いてもよい。いずれにしても変換部102と連動して対応する逆変換が執り行われる。 When the transform selection information 3303 is the first orthogonal transform, the output terminal of the switch is connected to the first inverse orthogonal transform unit 3501. On the other hand, when the transformation selection information 3303 is the second orthogonal transformation, the output end is connected to the second inverse orthogonal transformation unit 3502. Similarly, when the transform selection information 3303 is the Nth orthogonal transform, the output terminal is connected to the Nth inverse orthogonal transform unit 3503. Here, the transform selection information 3303 set in the orthogonal transform unit 102 and the transform selection information 3303 set in the inverse orthogonal transform unit 105 are the same, and the inverse orthogonal transform corresponding to the transform performed in the orthogonal transform unit 102 is performed. This is performed synchronously by the inverse orthogonal transform unit 105. That is, the first inverse orthogonal transform unit 3501 performs inverse discrete cosine transform (hereinafter referred to as IDCT), and the second inverse orthogonal transform unit 3502 and the Nth inverse orthogonal transform unit 3503 are based on KLT (Karunen-Labe transform). Inverse transformation is performed. Although an example using IDCT or the like is shown here as an example, orthogonal transformation such as Hadamard transformation or discrete sine transformation may be used, or non-orthogonal transformation may be used. In any case, the corresponding inverse conversion is performed in conjunction with the conversion unit 102.
 <変換選択部3301> 
 次に、図33に示される変換選択部3301について説明する。変換選択部3301には、符号化制御部115で制御され、予測選択部112で設定された予測モードなどを含む予測情報125が入力される。変換選択部3301は、この予測情報125に基づいて、どの予測モードに対して、どの直交変換を使うかどうかを示すMappedTransformIdx情報を設定する機能を有する。図36に、イントラ予測における変換選択情報3303(MappedTransformIdx)を示す。ここではN=9の例を示している。なお、IntraPredModeLX=2に対応するDC予測時には、第一直交変換部3401、及び対応する第一逆直交変換部3501が選択される。このように予測角度の近い基準予測モードへマッピングすることにより、全予測モードに対して直交変換器及び逆直交変換器を用意する場合と比較して、ハードウェア実現時の直交変換及び逆直交変換の回路規模を削減することが可能である。なお、双方向イントラ予測が選択されている場合、図14に従って、それぞれ2つのIntraPredModeL0とIntraPredModeL1を導出した後、IntraPredModeL1に対応する予測モードを利用して図36からMappedTransformIdxを導出する。本発明の本実施の形態では、N=9の例を示したが、Nの値は符号化性能とハードウェア実現時の回路規模のバランスを取って、最適な組み合わせを選べばよい。
<Conversion selection unit 3301>
Next, the conversion selection unit 3301 shown in FIG. 33 will be described. Prediction information 125 controlled by the encoding control unit 115 and including the prediction mode set by the prediction selection unit 112 is input to the transform selection unit 3301. Based on the prediction information 125, the transform selection unit 3301 has a function of setting MapdTransformIdx information indicating which orthogonal transform is used for which prediction mode. FIG. 36 shows conversion selection information 3303 (MappedTransformIdx) in intra prediction. Here, an example of N = 9 is shown. Note that the first orthogonal transform unit 3401 and the corresponding first inverse orthogonal transform unit 3501 are selected at the time of DC prediction corresponding to IntraPredModeLX = 2. By mapping to the reference prediction mode with a close prediction angle in this way, compared to the case of preparing an orthogonal transformer and an inverse orthogonal transformer for all prediction modes, orthogonal transformation and inverse orthogonal transformation at the time of hardware implementation It is possible to reduce the circuit scale. When bi-directional intra prediction is selected, two IntraPredModeL0 and IntraPredModeL1 are derived according to FIG. 14, respectively, and then the mapped transformIdx is derived from FIG. 36 using a prediction mode corresponding to IntraPredModeL1. In the present embodiment of the present invention, an example of N = 9 is shown, but the value of N may be selected in an optimal combination by balancing the coding performance and the circuit scale at the time of hardware implementation.
 <係数順制御部3302> 
 次に、係数順制御部3302を説明する。図37に係数順制御部3302のブロック図を示す。係数順制御部3302は、係数順選択スイッチ3704と、第一係数順変換部3701、第二係数順変換部3702、第N係数順変換部3703とを有する。係数順選択スイッチ3704は、例えば、図36に示されるMappedTransformIdxに従って、スイッチの出力端と係数順変換部3701ないし3703を切り替える機能を有する。N種類の係数順変換部3701~3703は、量子化部103で量子化処理された量子化変換係数119の2次元データを1次元データへと変換する機能を有する。例えば、H.264では、ジグザグスキャンを用いて2次元データを1次元データへと変換している。
<Coefficient order control unit 3302>
Next, the coefficient order control unit 3302 will be described. FIG. 37 shows a block diagram of the coefficient order control unit 3302. The coefficient order control unit 3302 includes a coefficient order selection switch 3704, a first coefficient order conversion unit 3701, a second coefficient order conversion unit 3702, and an Nth coefficient order conversion unit 3703. The coefficient order selection switch 3704 has a function of switching the output terminal of the switch and the coefficient order conversion units 3701 to 3703 in accordance with, for example, the mapped transform IDx shown in FIG. The N types of coefficient forward conversion units 3701 to 3703 have a function of converting the two-dimensional data of the quantized conversion coefficient 119 quantized by the quantization unit 103 into one-dimensional data. For example, H.M. In H.264, two-dimensional data is converted into one-dimensional data using a zigzag scan.
 イントラの予測方向を考慮した直交変換を用いる場合、直交変換を施した変換係数118に量子化処理を施した量子化変換係数119は、ブロック内の非ゼロとなる変換係数の発生傾向が偏る性質を持つ。この非ゼロ変換係数の発生傾向は、イントラ予測の予測方向毎に異なる性質がある。しかし、異なる映像を符号化した際に同じ予測方向における非ゼロ変換係数の発生傾向は似る性質を持つ。そこで、2次元データを1次元データへ変換(2D-1D変換)する際、非ゼロ変換係数の発生確率が高い位置の変換係数から優先的にエントロピー符号化することで、変換係数の符号化する情報を削減することが可能である。そこで、予測情報125に含まれる予測モードなどの予測方向を表す情報に基づいて、予め非ゼロ変換係数の発生確率を学習することによって、例えばH.264と比較して演算量の増加を引き起こすことなく、変換係数の符号量を削減することが可能となる。 When orthogonal transform in consideration of the intra prediction direction is used, the quantized transform coefficient 119 obtained by performing the quantization process on the transform coefficient 118 subjected to the orthogonal transform has a characteristic that the tendency of generating non-zero transform coefficients in the block is biased. have. The tendency of occurrence of this non-zero transform coefficient has different properties for each prediction direction of intra prediction. However, when different videos are encoded, the generation tendency of non-zero transform coefficients in the same prediction direction has a similar property. Therefore, when transforming two-dimensional data into one-dimensional data (2D-1D conversion), entropy coding is performed preferentially from transform coefficients at positions where the occurrence probability of non-zero transform coefficients is high, thereby encoding transform coefficients. It is possible to reduce information. Therefore, by learning the occurrence probability of a non-zero conversion coefficient in advance based on information representing a prediction direction such as a prediction mode included in the prediction information 125, for example, H.264. Compared with H.264, it is possible to reduce the code amount of the transform coefficient without causing an increase in the calculation amount.
 さらに別の例として、係数順制御部3302は、2D-1D変換におけるスキャン順を動的に更新してもよい。このような動作を行う係数順制御部3302は、図38に例示される。この係数順制御部3302は、図37の構成に加え、発生頻度カウント部3801と、更新部3802とを含む。係数順変換部3701、・・・、3703は、そのスキャン順が係数順更新部3802によって更新される点以外は同一である。 As yet another example, the coefficient order control unit 3302 may dynamically update the scan order in 2D-1D conversion. The coefficient order control unit 3302 that performs such an operation is illustrated in FIG. The coefficient order control unit 3302 includes an occurrence frequency counting unit 3801 and an updating unit 3802 in addition to the configuration of FIG. The coefficient order conversion units 3701,..., 3703 are the same except that the scan order is updated by the coefficient order update unit 3802.
 発生頻度カウント部3801は、予測モード毎に、量子化変換係数列3304の各要素における非零係数の発生回数のヒストグラム3804を作成する。発生頻度カウント部3801は、作成したヒストグラム3804を更新部3802に入力する。 The occurrence frequency counting unit 3801 creates a histogram 3804 of the number of occurrences of non-zero coefficients in each element of the quantized transform coefficient sequence 3304 for each prediction mode. The occurrence frequency counting unit 3801 inputs the created histogram 3804 to the update unit 3802.
 更新部3802は、予め定められたタイミングで、ヒストグラム3804に基づいて係数順の更新を行う。上記タイミングは、例えば、コーディングツリーユニットの符号化処理が終了したタイミング、コーディングツリーユニット内の1ライン分の符号化処理が終了したタイミングなどである。 The update unit 3802 updates the coefficient order based on the histogram 3804 at a predetermined timing. The timing is, for example, the timing when the coding process of the coding tree unit is finished, the timing when the coding process for one line in the coding tree unit is finished, or the like.
 具体的には、更新部3802は、ヒストグラム3804を参照して、非零係数の発生回数が閾値以上にカウントされた要素を持つ予測モードに関して係数順の更新を行う。例えば、更新部3802は、非零係数の発生が16回以上カウントされた要素を持つ予測モードに関して更新を行う。このような発生回数に閾値を設けることによって、係数順の更新が大域的に実施されるので、局所的な最適解に収束しにくくなる。 Specifically, the update unit 3802 refers to the histogram 3804 and updates the coefficient order with respect to a prediction mode having an element in which the number of occurrences of non-zero coefficients is counted more than a threshold. For example, the update unit 3802 updates the prediction mode having an element in which the occurrence of a non-zero coefficient is counted 16 times or more. By providing a threshold value for the number of occurrences, the coefficient order is updated globally, so that it is difficult to converge to a local optimum solution.
 更新部3802は、更新対象となる予測モードに関して、非零係数の発生頻度の降順に要素をソーティングする。ソーティングは、例えばバブルソート、クイックソートなどの既存のアルゴリズムによって実現できる。そして、更新部3802は、ソーティングされた要素の順序を示す更新係数順3803を、更新対象となる予測モードに対応する係数順変換部3701ないし3703に入力する。 The update unit 3802 sorts the elements in descending order of the occurrence frequency of the non-zero coefficient with respect to the prediction mode to be updated. Sorting can be realized by existing algorithms such as bubble sort and quick sort. Then, the update unit 3802 inputs the update coefficient order 3803 indicating the order of the sorted elements to the coefficient order conversion units 3701 to 3703 corresponding to the prediction mode to be updated.
 更新係数順3803が入力されると、各変換部は更新後のスキャン順に従って2D-1D変換を行う。なお、スキャン順を動的に更新する場合には、各2D-1D変換部の初期スキャン順を予め定めておく必要がある。このように、動的にスキャン順を更新することにより、予測画像の性質、量子化情報(量子化パラメータ)などの影響に応じて、量子化変換係数119における非零係数の発生傾向が変化する場合にも、安定的に高い符号化効率を期待できる。具体的には、エントロピー符号化部113におけるランレングス符号化の発生符号量を抑制できる。 
 なお、本実施の形態におけるシンタクス構成は、第一の実施の形態と同一である。
When the update coefficient order 3803 is input, each conversion unit performs 2D-1D conversion according to the updated scan order. When the scan order is dynamically updated, the initial scan order of each 2D-1D conversion unit needs to be determined in advance. In this way, by dynamically updating the scan order, the tendency of occurrence of non-zero coefficients in the quantized transform coefficients 119 changes according to the influence of the properties of the predicted image, quantization information (quantization parameters), and the like. Even in this case, high encoding efficiency can be expected stably. Specifically, the generated code amount of run-length encoding in the entropy encoding unit 113 can be suppressed.
The syntax configuration in this embodiment is the same as that in the first embodiment.
 本実施の形態の変形例として、変換選択部3301が予測情報125とは別にMappedTransformIdxを選択することも可能である。この場合、9種類のどの直交変換或いは逆直交変換を用いたかを示す情報をエントロピー符号化部113に設定し、量子化変換係数列3304とともに符号化される。図39に本変形例におけるシンタクスの例を示す。シンタクス中に示されるdirectional_transform_idxは、N個に対応する直交変換のいずれを選択したかの情報が示されている。 As a modification of the present embodiment, the conversion selection unit 3301 can select the mapped transform IDx separately from the prediction information 125. In this case, information indicating which nine types of orthogonal transforms or inverse orthogonal transforms are used is set in the entropy encoding unit 113 and encoded together with the quantized transform coefficient sequence 3304. FIG. 39 shows an example of syntax in this modification. Directional_transform_idx indicated in the syntax indicates information indicating which of N orthogonal transforms has been selected.
 以上の第2の実施形態によれば、ハードウェア実装及びソフトウェア実装における困難性を緩和しつつ、高効率な直交変換及び逆直交変換を実現することができる。故に、符号化効率が向上し、ひいては主観画質も向上する。 According to the second embodiment described above, highly efficient orthogonal transformation and inverse orthogonal transformation can be realized while alleviating the difficulty in hardware implementation and software implementation. Therefore, the coding efficiency is improved, and the subjective image quality is also improved.
 (第3の実施形態) 
 <動画像符号化装置-第三の実施形態> 
 直交変換部102に係る実施形態として、JCTVC-B205_draft002, 5.3.5.2節" Rotational transformation process ", JCT-VC 2nd Meeting Geneva, July, 2010に示される回転変換と組み合わせても構わない。回転変換は、DCTを用いた直交変換後にさらに回転変換を行うことで変換係数の係数集密度をさらに高める手法である。
(Third embodiment)
<Video Encoding Device—Third Embodiment>
As an embodiment related to the orthogonal transformation unit 102, it may be combined with the rotation transformation shown in JCTVC-B205_draft002, section 5.3.5.2 “Rotational transformation process”, JCT-VC 2nd Meeting Geneva, July, 2010. Rotational transformation is a technique for further increasing the coefficient density of transformation coefficients by further performing rotational transformation after orthogonal transformation using DCT.
 <直交変換部102> 
 図40に本実施の形態に関わる直交変換部102のブロック図を示す。直交変換部102は、第一回転変換部4001、第二回転変換部4002、第N回転変換部4003、離散コサイン変換部4004の新しい処理部を持ち、既存の変換選択スイッチ3404を有する。離散コサイン変換部4004は、例えばDCTを行う。DCT後の変換係数が変換選択スイッチ3404に入力される。ここで、変換選択スイッチ3404は、変換選択情報3303に従って、スイッチの出力端を第一回転変換部4001、第二回転変換部4002、第N回転変換部4003のいずれかに接続する。例えば、符号化制御部115の制御に従って、順番にスイッチを切り替える。回転変換部4001ないし4003は、それぞれの変換係数に対して、予め定められた回転行列を用いて、回転変換を行う。回転変換後の変換係数118が出力される。この変換は可逆変換である。
<Orthogonal transform unit 102>
FIG. 40 shows a block diagram of the orthogonal transform unit 102 according to the present embodiment. The orthogonal transform unit 102 includes new processing units such as a first rotation transform unit 4001, a second rotation transform unit 4002, an Nth rotation transform unit 4003, and a discrete cosine transform unit 4004, and has an existing transform selection switch 3404. The discrete cosine transform unit 4004 performs DCT, for example. The conversion coefficient after DCT is input to the conversion selection switch 3404. Here, the conversion selection switch 3404 connects the output end of the switch to one of the first rotation conversion unit 4001, the second rotation conversion unit 4002, and the Nth rotation conversion unit 4003 according to the conversion selection information 3303. For example, the switches are sequentially switched according to the control of the encoding control unit 115. The rotation conversion units 4001 to 4003 perform rotation conversion for each conversion coefficient using a predetermined rotation matrix. The conversion coefficient 118 after the rotation conversion is output. This conversion is a reversible conversion.
 ここでは、数式(1)、数式(2)に示されるような符号化コストを用いて、どの回転行列を使うかどうかを判定してもよい。また、予め図36に示したような予測モードと変換番号を対応付けたテーブルを用意して選択してもよい。また、ここでは、量子化部103の前に、回転変換部を適用する例を示したが、量子化処理後の量子化変換係数119に対して、回転変換部を適用してもよい。この場合、直交変換部102は、DCTのみを行う。 Here, it may be determined which rotation matrix is to be used by using the encoding cost as shown in Equation (1) and Equation (2). In addition, a table in which a prediction mode and a conversion number are associated with each other as shown in FIG. Here, an example in which the rotation conversion unit is applied before the quantization unit 103 is shown, but the rotation conversion unit may be applied to the quantization conversion coefficient 119 after the quantization process. In this case, the orthogonal transform unit 102 performs only DCT.
 <逆直交変換部105> 
 図41は本実施の形態に関わる逆直交変換部105のブロック図である。逆直交変換部105は、第一逆回転変換部4101、第二逆回転変換部4102、第N逆回転変換部4103、逆離散コサイン変換部4104の新しい処理部を持ち、既存の変換選択スイッチ3504を有する。逆量子化処理後に入力された復元変換係数120が変換選択スイッチ3504に入力される。ここで、変換選択スイッチ3504は、変換選択情報3303に従って、スイッチの出力端を第一逆回転変換部4101、第二逆回転変換部4102、第N逆回転変換部4103のいずれかに接続する。その後、直交変換部102で利用された回転変換と同じ、いずれかの逆回転変換部4101ないし4103で逆回転変換処理が施され、逆離散コサイン変換部4104へと出力する。逆離散コサイン変換部4104は、入力された信号に対して例えばIDCTを施し、復元予測誤差信号121を復元する。ここでは例としてIDCTを用いる例を示したが、アダマール変換や離散サイン変換などの直交変換を使ってもよいし、非直交変換を用いてもよい。いずれにしても変換部102と連動して対応する逆変換が執り行われる。
<Inverse orthogonal transform unit 105>
FIG. 41 is a block diagram of the inverse orthogonal transform unit 105 according to the present embodiment. The inverse orthogonal transform unit 105 includes new processing units such as a first inverse rotation transform unit 4101, a second inverse rotation transform unit 4102, an Nth inverse rotation transform unit 4103, and an inverse discrete cosine transform unit 4104, and an existing transform selection switch 3504. Have The restored transform coefficient 120 input after the inverse quantization process is input to the transform selection switch 3504. Here, the conversion selection switch 3504 connects the output terminal of the switch to one of the first reverse rotation conversion unit 4101, the second reverse rotation conversion unit 4102, and the Nth reverse rotation conversion unit 4103 according to the conversion selection information 3303. Thereafter, the same inverse rotation transform used in the orthogonal transform unit 102 is subjected to inverse rotation transform processing by any one of the inverse rotation transform units 4101 to 4103, and is output to the inverse discrete cosine transform unit 4104. The inverse discrete cosine transform unit 4104 performs, for example, IDCT on the input signal to restore the restored prediction error signal 121. Although an example using IDCT is shown here as an example, orthogonal transform such as Hadamard transform or discrete sine transform may be used, or non-orthogonal transform may be used. In any case, the corresponding inverse conversion is performed in conjunction with the conversion unit 102.
 本実施の形態におけるシンタクスが図42で示されている。シンタクス中に示されるrotational_transform_idxは利用する回転行列の番号を意味している。 FIG. 42 shows the syntax in the present embodiment. The rotation_transform_idx shown in the syntax means the number of the rotation matrix to be used.
 以上の第3の実施形態によれば、ハードウェア実装及びソフトウェア実装における困難性を緩和しつつ、高効率な直交変換及び逆直交変換を実現することができる。故に、符号化効率が向上し、ひいては主観画質も向上する。 According to the third embodiment described above, highly efficient orthogonal transformation and inverse orthogonal transformation can be realized while alleviating the difficulty in hardware implementation and software implementation. Therefore, the coding efficiency is improved, and the subjective image quality is also improved.
 (第4の実施形態) 
 第4の実施形態は動画像復号化装置に関する。本実施形態に係る動画像復号化装置に対応する動画像符号化装置は、第1の実施形態において説明した通りである。即ち、本実施形態に係る動画像復号化装置は、例えば第1の実施形態に係る動画像符号化装置によって生成された符号化データを復号化する。
(Fourth embodiment)
The fourth embodiment relates to a moving picture decoding apparatus. The video encoding device corresponding to the video decoding device according to the present embodiment is as described in the first embodiment. That is, the moving picture decoding apparatus according to the present embodiment decodes encoded data generated by, for example, the moving picture encoding apparatus according to the first embodiment.
 図44に示すように、本実施形態に係る動画像復号化装置は、入力バッファ4401、エントロピー復号化部4402、逆量子化部4403、逆直交変換部4404、加算部4405、ループフィルタ4406、参照画像メモリ4407、イントラ予測部4408、インター予測部4409、予測選択スイッチ4410、出力バッファ4411を含む。 As shown in FIG. 44, the moving picture decoding apparatus according to this embodiment includes an input buffer 4401, an entropy decoding unit 4402, an inverse quantization unit 4403, an inverse orthogonal transform unit 4404, an addition unit 4405, and a loop filter 4406. An image memory 4407, an intra prediction unit 4408, an inter prediction unit 4409, a prediction selection switch 4410, and an output buffer 4411 are included.
 図44の動画像復号化装置は、入力バッファ4401に蓄積される符号化データ4413を復号し、復号画像4422を出力バッファ4411に蓄積して出力画像として出力する。符号化データ4413は、例えば図1の動画像符号化装置などから出力され、図示しない蓄積系または伝送系を経て、入力バッファ4401に一時的に蓄積される。 44 decodes the encoded data 4413 stored in the input buffer 4401, stores the decoded image 4422 in the output buffer 4411, and outputs it as an output image. The encoded data 4413 is output from, for example, the moving image encoding apparatus shown in FIG. 1, and is temporarily stored in the input buffer 4401 via a storage system or transmission system (not shown).
 エントロピー復号化部4402は、符号化データ4413の復号化のために、1フレームまたは1フィールド毎にシンタクスに基づいて解読を行う。エントロピー復号化部4402は、各シンタクスの符号列を順次エントロピー復号化し、予測モード情報などを含む予測情報4421、量子化変換係数4414などの符号化対象ブロックの符号化パラメータを再生する。符号化パラメータとは、予測情報4421、変換係数に関する情報、量子化に関する情報、などの復号に必要となるパラメータである。 The entropy decoding unit 4402 performs decoding based on the syntax for each frame or field for decoding the encoded data 4413. The entropy decoding unit 4402 sequentially entropy-decodes the code string of each syntax, and reproduces the encoding parameters of the encoding target block such as the prediction information 4421 including the prediction mode information and the quantization transform coefficient 4414. The encoding parameter is a parameter necessary for decoding such as prediction information 4421, information on transform coefficients, information on quantization, and the like.
 逆量子化部4403は、エントロピー復号化部4402からの量子化変換係数4414に逆量子化を行って、復元変換係数4415を得る。具体的には、逆量子化部4403は、エントロピー復号化部4402によって復号化された量子化に関する情報に従って逆量子化を行う。逆量子化部4403は、復元変換係数4415を逆直交変換部4404に入力する。 The inverse quantization unit 4403 performs inverse quantization on the quantized transform coefficient 4414 from the entropy decoding unit 4402 to obtain a restored transform coefficient 4415. Specifically, the inverse quantization unit 4403 performs inverse quantization according to the information regarding the quantization decoded by the entropy decoding unit 4402. The inverse quantization unit 4403 inputs the restored transform coefficient 4415 to the inverse orthogonal transform unit 4404.
 逆直交変換部4404は、逆量子化部4403からの復元変換係数4415に対して、符号化側において行われた直交変換に対応する逆直交変換を行い、復元予測誤差信号4416を得る。逆直交変換部4404は、復元予測誤差信号4416を加算部4405に入力する。 The inverse orthogonal transform unit 4404 performs inverse orthogonal transform corresponding to the orthogonal transform performed on the encoding side, on the reconstruction transform coefficient 4415 from the inverse quantization unit 4403, and obtains a reconstruction prediction error signal 4416. The inverse orthogonal transform unit 4404 inputs the restored prediction error signal 4416 to the adder 4405.
 加算部4405は、復元予測誤差信号4416と、対応する予測画像信号4420とを加算し、復号画像信号4417を生成する。復号画像信号4417は、ループフィルタ4406へと入力される。ループフィルタ4406は、入力された復号画像信号4417にデブロッキングフィルタやウィナーフィルタなどを施し、被フィルタ画像信号4418を生成する。生成した被フィルタ画像信号4418は、出力画像のために出力バッファ4411に一時的に蓄積されると共に、参照画像信号4419のために参照画像メモリ4407にも保存される。参照画像メモリ4407に保存された被フィルタ画像信号4418は、参照画像信号4419としてイントラ予測部4408及びインター予測部4409によって必要に応じてフレーム単位またはフィールド単位で参照される。出力バッファ4411に一時的に蓄積された被フィルタ画像信号4418は、復号化制御部4412によって管理される出力タイミングに従って出力される。 The addition unit 4405 adds the restored prediction error signal 4416 and the corresponding predicted image signal 4420 to generate a decoded image signal 4417. The decoded image signal 4417 is input to the loop filter 4406. The loop filter 4406 performs a deblocking filter, a Wiener filter, or the like on the input decoded image signal 4417 to generate a filtered image signal 4418. The generated filtered image signal 4418 is temporarily stored in the output buffer 4411 for the output image, and is also stored in the reference image memory 4407 for the reference image signal 4419. The filtered image signal 4418 stored in the reference image memory 4407 is referenced as a reference image signal 4419 by the intra prediction unit 4408 and the inter prediction unit 4409 as necessary in units of frames or fields. The filtered image signal 4418 temporarily accumulated in the output buffer 4411 is output according to the output timing managed by the decoding control unit 4412.
 イントラ予測部4408、インター予測部4409及び選択スイッチ4410は、図1のイントラ予測部109、インター予測部110及び選択スイッチ111と実質的に同一または類似の要素である。イントラ予測部4408(109)は、参照画像メモリ4407に保存されている参照画像信号4419を利用してイントラ予測を行う。例えば、H.264では、予測対象ブロックに隣接する符号化済みの参照画素値を利用して、垂直方向、水平方向などの予測方向に沿って画素補填(コピーまたは補間後にコピー)を行うことによってイントラ予測画像を生成する。図5(a)にH.264におけるイントラ予測の予測方向を示す。また、図5(b)にH.264における参照画素と符号化対象画素との配置関係を示す。図5(c)はモード1(水平予測)の予測画像生成方法を示しており、図5(d)はモード4(対角右下予測;図4AのIntra_NxN_Diagonal_Down_Right)の予測画像生成方法を示している。 The intra prediction unit 4408, the inter prediction unit 4409, and the selection switch 4410 are substantially the same or similar elements as the intra prediction unit 109, the inter prediction unit 110, and the selection switch 111 in FIG. The intra prediction unit 4408 (109) performs intra prediction using the reference image signal 4419 stored in the reference image memory 4407. For example, H.M. In H.264, an intra prediction image is obtained by performing pixel interpolation (copying or copying after interpolation) along a prediction direction such as a vertical direction or a horizontal direction using an encoded reference pixel value adjacent to a prediction target block. Generate. FIG. The prediction direction of intra prediction in H.264 is shown. Further, in FIG. 2 shows an arrangement relationship between reference pixels and encoding target pixels in H.264. 5C shows a predicted image generation method in mode 1 (horizontal prediction), and FIG. 5D shows a predicted image generation method in mode 4 (diagonal lower right prediction; Intra_NxN_Diagonal_Down_Right in FIG. 4A). Yes.
 また、Jung-Hye Min, “Unification of the Directional Intra Prediction Methods in TMuC”, Joint Collaborative Team on Video Coding (JCT-VC) of ITU-T SG16 WP3 and ISO/IEC JTC1/SC29/WG11 Document, JCTVC-B100, July 2010では、H.264の予測方向をさらに34方向に拡張し、予測モード数を増やしている。予測角度に応じて32画素精度の線形補間を行うことで予測画素値を作成し、予測方向にコピーする。本発明の本実施の形態で用いるイントラ予測部109の詳細は後述する。 Also, Jung-Hye Min, “Unification of the Directional Intra Prediction Methods in TMuC”, Joint Collaborative Team on Video Coding (JCT-VC) of ITU-T SG16 WP3 and ISO / IEC JTC1 / SC29 / WG11-VCB , July 2010, H. The prediction direction of H.264 is further expanded to 34 directions to increase the number of prediction modes. A predicted pixel value is created by performing linear interpolation with 32-pixel accuracy in accordance with the predicted angle, and is copied in the predicted direction. Details of the intra prediction unit 109 used in the present embodiment of the present invention will be described later.
 インター予測部4409(110)は、参照画像メモリ4407に保存されている参照画像信号4419を利用してインター予測を行う。具体的には、インター予測部4409(110)は、予測対象ブロックと参照画像信号124との間の動きのズレ量(動きベクトル)をエントロピー復号化部4402から取得し、この動きベクトルに基づいて補間処理(動き補償)を行ってインター予測画像を生成する。H.264では、1/4画素精度までの補間処理が可能である。 The inter prediction unit 4409 (110) performs inter prediction using the reference image signal 4419 stored in the reference image memory 4407. Specifically, the inter prediction unit 4409 (110) obtains a motion shift amount (motion vector) between the prediction target block and the reference image signal 124 from the entropy decoding unit 4402, and based on this motion vector. Inter prediction processing (motion compensation) is performed to generate an inter predicted image. H. With H.264, interpolation processing up to 1/4 pixel accuracy is possible.
 予測選択スイッチ4410は、イントラ予測部4408の出力端またはインター予測部4409の出力端を復号した予測情報4421に従って選択し、イントラ予測画像またはインター予測画像を予測画像信号4420として加算部4405に入力する。予測情報4421がイントラ予測を示す場合には、予測選択スイッチ4410はイントラ予測部4408からの出力端にスイッチを接続する。一方、予測情報4421がインター予測を示す場合には、予測選択スイッチ4410はインター予測部4409からの出力端にスイッチを接続する。 The prediction selection switch 4410 selects the output terminal of the intra prediction unit 4408 or the output terminal of the inter prediction unit 4409 according to the decoded prediction information 4421, and inputs the intra predicted image or the inter predicted image as the predicted image signal 4420 to the adding unit 4405. . When the prediction information 4421 indicates intra prediction, the prediction selection switch 4410 connects a switch to the output terminal from the intra prediction unit 4408. On the other hand, when the prediction information 4421 indicates inter prediction, the prediction selection switch 4410 connects a switch to the output terminal from the inter prediction unit 4409.
 復号化制御部4412は、図44の動画像復号化装置の各要素を制御する。具体的には、復号化制御部4412は、上述の動作を含む復号化処理のための種々の制御を行う。 The decoding control unit 4412 controls each element of the moving picture decoding apparatus in FIG. Specifically, the decoding control unit 4412 performs various controls for decoding processing including the above-described operation.
 また、図44の動画像復号化装置は、図20、図21、図22A~E、図27A~B、及び図32A~Bに関して説明したシンタクスと同一または類似のシンタクスを利用するのでその詳細な説明を省略する。 44 uses the same or similar syntax as the syntax described with reference to FIGS. 20, 21, 22A to E, 27A to 27B, and 32A to 32B. Description is omitted.
 以下、イントラ予測部4408(109)の詳細について図6を用いて説明する。 Details of the intra prediction unit 4408 (109) will be described below with reference to FIG.
 本実施形態における、イントラ予測部4408は第1の実施形態で説明したイントラ予測部109とその構成、処理内容は同一である。 In this embodiment, the intra prediction unit 4408 has the same configuration and processing content as the intra prediction unit 109 described in the first embodiment.
 図6に示されるイントラ予測部4408(109)は、単方向イントラ予測画像生成部601、双方向イントラ予測画像生成部602、予測モード情報設定部603、選択スイッチ604を有する。先ず、参照画像メモリ4407から参照画像信号4419(124)が、単方向イントラ予測画像生成部601及び双方向イントラ予測画像生成部602に入力される。ここで、復号化制御部4412で制御されている予測モード情報に従って、予測モード情報設定部603は、単方向イントラ予測画像生成部601或いは双方向イントラ予測画像生成部602で生成される予測モードを設定し、予測モード605を出力する。選択スイッチ604は、この予測モード605に従って、それぞれのイントラ予測画像生成部の出力端を繋ぎかえる機能を有する。入力された予測モード605が単方向イントラ予測モードであれば、単方向イントラ予測画像生成部601の出力端とスイッチを接続し、予測モード605が双方向イントラ予測モードであれば、双方向イントラ予測画像生成部602の出力端を接続する。一方、それぞれのイントラ予測画像生成部601、602は、予測モード605に従って、予測画像信号4420(126)を生成する。生成された予測画像信号4420(126)がイントラ予測部109から出力される。 The intra prediction unit 4408 (109) shown in FIG. 6 includes a unidirectional intra predicted image generation unit 601, a bidirectional intra predicted image generation unit 602, a prediction mode information setting unit 603, and a selection switch 604. First, a reference image signal 4419 (124) is input from the reference image memory 4407 to the unidirectional intra predicted image generation unit 601 and the bidirectional intra predicted image generation unit 602. Here, according to the prediction mode information controlled by the decoding control unit 4412, the prediction mode information setting unit 603 selects the prediction mode generated by the unidirectional intra prediction image generation unit 601 or the bidirectional intra prediction image generation unit 602. Set and output prediction mode 605. The selection switch 604 has a function of switching the output ends of the respective intra predicted image generation units according to the prediction mode 605. If the input prediction mode 605 is the unidirectional intra prediction mode, the output terminal of the unidirectional intra prediction image generation unit 601 is connected to the switch, and if the prediction mode 605 is the bidirectional intra prediction mode, the bidirectional intra prediction is performed. The output terminal of the image generation unit 602 is connected. On the other hand, each intra prediction image production | generation part 601 and 602 produces | generates the prediction image signal 4420 (126) according to the prediction mode 605. FIG. The generated predicted image signal 4420 (126) is output from the intra prediction unit 109.
 先ず、予測モード情報設定部603について詳細に説明する。図7は、本発明の本実施の形態に関わる予測モードのブロックサイズ別の数を示している。PuSizeは、予測を行う画素ブロック(プレディクションユニット)サイズを示しており、PU_2x2~PU_128x128まで7種類のサイズが規定されている。IntraUniModeNumは単方向イントラ予測の予測モード数を表しており、IntraBiModeNumは双方向イントラ予測の予測モード数を表している。また、Number of modesは、各画素ブロック(プレディクションユニット)サイズ毎の予測モード数の総数である。 First, the prediction mode information setting unit 603 will be described in detail. FIG. 7 shows the number of prediction modes according to the block size according to the present embodiment of the present invention. PuSize indicates the pixel block (prediction unit) size to be predicted, and seven types of sizes from PU_2x2 to PU_128x128 are defined. IntraUniModeNum represents the number of prediction modes for unidirectional intra prediction, and IntraBiModeNum represents the number of prediction modes for bidirectional intra prediction. Also, Number of modes is the total number of prediction modes for each pixel block (prediction unit) size.
 一方、図9にPuSizeがPU_8x8、PU_16x16及びPU_32x32の場合の、予測モードと予測方法の関係を示す。なお、図10にPuSizeがPU_4x4の場合を示し、図11にPU_64x64或いはPU_128x128の場合を示す。ここで、IntraPredModeは予測モード番号を示し、IntraBipredFlagは、双方向イントラ予測であるかを示すフラグである。当該フラグが1である場合は、当該予測モードが双方向イントラ予測モードであることを示している。当該フラグが0の場合は、当該予測モードが単方向イントラ予測モードであることを示している。IntraPredTypeLXはイントラ予測の予測タイプを示している。Intra_Verticalは、垂直方向を予測の基準とすることを意味し、Intra_Horizontalは水平方向を予測の基準とすることを意味する。なお、IntraPredTypeLX内のXには、0或いは1が適用される。IntraPredTypeL0は、単方向イントラ予測或いは双方向イントラ予測の最初の予測モードを示している。IntraPredTypeL1は、双方向イントラ予測の2番目の予測モードを示している。また、IntraPredAngleIDは、予測角度のインデックスを示す指標である。実際に予測値生成で使われる予測角度は図12に示されている。ここで、puPartIdxは図3Bで説明した四分木分割の際の分割したブロックのインデックスを表している。 On the other hand, FIG. 9 shows the relationship between the prediction mode and the prediction method when PuSize is PU_8x8, PU_16x16, and PU_32x32. FIG. 10 shows a case where PuSize is PU_4 × 4, and FIG. 11 shows a case where PU_64 × 64 or PU_128 × 128. Here, IntraPredMode indicates a prediction mode number, and IntraBipredFlag is a flag indicating whether or not bidirectional intra prediction. When the flag is 1, it indicates that the prediction mode is the bidirectional intra prediction mode. When the flag is 0, it indicates that the prediction mode is a unidirectional intra prediction mode. IntraPredTypeLX indicates the prediction type of intra prediction. Intra_Vertical means that the vertical direction is the reference for prediction, and Intra_Horizontal means that the horizontal direction is the reference for prediction. Note that 0 or 1 is applied to X in IntraPredTypeLX. IntraPredTypeL0 indicates the first prediction mode of unidirectional intra prediction or bidirectional intra prediction. IntraPredTypeL1 indicates the second prediction mode of bidirectional intra prediction. IntraPred AngleID is an index indicating an index of a prediction angle. The prediction angle actually used in the generation of the predicted value is shown in FIG. Here, puPartIdx represents the index of the divided block in the quadtree division described with reference to FIG. 3B.
 例えば、IntraPredModeが4の場合、IntraPredTypeL0がIntra_Verticalであるため、垂直方向を予測の基準とすることが判る。本図から判る通り、IntraPredMode=0から32まで計33個が単方向イントラ予測モードを示しており、IntraPredMode=32から48まで計16個が双方向イントラ予測モードを示している。 For example, when IntraPredMode is 4, since IntraPredTypeL0 is Intra_Vertical, it can be seen that the vertical direction is used as a reference for prediction. As can be seen from the figure, a total of 33 from IntraPredMode = 0 to 32 indicate the unidirectional intra prediction mode, and a total of 16 from IntraPredMode = 32 to 48 indicate the bidirectional intra prediction mode.
 予測モード情報設定部603は、復号化制御部4412における制御の下で、指定された予測モード605に対応する上述した予測情報を単方向イントラ予測画像生成部601及び双方向イントラ予測画像生成部602に設定し、選択スイッチへ予測モード605を出力する。 The prediction mode information setting unit 603 converts the above-described prediction information corresponding to the designated prediction mode 605 to the unidirectional intra prediction image generation unit 601 and the bidirectional intra prediction image generation unit 602 under the control of the decoding control unit 4412. And the prediction mode 605 is output to the selection switch.
 次に、単方向イントラ予測画像生成部601について詳細に説明する。単方向イントラ予測画像生成部601は、図8に示される複数の予測方向に対して予測画像信号4420(126)を生成する機能を有する。図8では、太線で示される垂直方向、水平方向の座標に対して、33個の異なる予測方向を持つ。また、H.264で示される代表的な予測角度の方向を矢印で示している。本発明の本実施の形態では、原点から菱形で示されるマークまで線を引いた方向に33種類の予測方向が用意されている。また、H.264と同様、利用可能な参照画素の平均値で予測するDC予測が追加されており、合計で34個の予測モードが存在する。 Next, the unidirectional intra predicted image generation unit 601 will be described in detail. The unidirectional intra predicted image generation unit 601 has a function of generating a predicted image signal 4420 (126) for a plurality of prediction directions shown in FIG. In FIG. 8, there are 33 different prediction directions for the vertical and horizontal coordinates indicated by the bold lines. H. The direction of a typical prediction angle indicated by H.264 is indicated by an arrow. In the present embodiment of the present invention, 33 kinds of prediction directions are prepared in a direction in which a line is drawn from the origin to a mark indicated by a diamond. H. Similar to H.264, DC prediction for predicting with an average value of available reference pixels is added, and there are 34 prediction modes in total.
 IntraPredMode=4の場合、IntraPredAngleIDL0が‐4であるため、図8におけるIntraPredMode=4で示される予測方向で予測画像信号4420(126)が生成される。図8における点線で示される矢印は、予測タイプがIntra_Verticalの予測モードを示しており、実線で示される矢印は予測タイプがIntra_Horizontalの予測モードを示している。 In the case of IntraPredMode = 4, since IntraPredAngleIDL0 is −4, the prediction image signal 4420 (126) is generated in the prediction direction indicated by IntraPredMode = 4 in FIG. An arrow indicated by a dotted line in FIG. 8 indicates a prediction mode whose prediction type is Intra_Vertical, and an arrow indicated by a solid line indicates a prediction mode whose prediction type is Intra_Horizontal.
 <イントラ予測部4408(109)>
 次に単方向イントラ予測画像生成部601の予測画像生成方法について説明する。ここでは、入力された参照画像信号4419(124)に基づいて、予測画像値を生成し、上述した予測方向に画素をコピーする。予測画像値は、1/32画素精度で内挿補間を行うことによって生成する。図12は、IntraPredAngleIDLXと予測画像値生成に使われるintraPredAngleとの関係を示す。intraPredAngleは予測値生成の際に、実際に利用される予測角度を示している。例えば、予測タイプがIntra_Verticalで且つ、図12で示されるintraPredAngleが正の値の場合の予測値の生成方法を数式で書くと数式(3)で表される。ここでBLK_SIZEは、当該画素ブロック(プレディクションユニット)のサイズを示しており、ref[]は参照画像信号が格納された配列を示している。また、pred(k,m)は生成された予測画像信号4420(126)を示している。
<Intra Prediction Unit 4408 (109)>
Next, a prediction image generation method of the unidirectional intra prediction image generation unit 601 will be described. Here, a predicted image value is generated based on the input reference image signal 4419 (124), and a pixel is copied in the above-described prediction direction. The predicted image value is generated by performing interpolation with 1/32 pixel accuracy. FIG. 12 shows the relationship between IntraPredAngleIDLX and intraPredAngle used for predictive image value generation. intraPredAngle indicates a prediction angle that is actually used when a predicted value is generated. For example, when the prediction type is Intra_Vertical and intraPredAngle shown in FIG. 12 is a positive value, a prediction value generation method is expressed by Expression (3). Here, BLK_SIZE indicates the size of the pixel block (prediction unit), and ref [] indicates an array in which reference image signals are stored. Also, pred (k, m) indicates the generated predicted image signal 4420 (126).
 上記条件以外に関しても、図12のテーブルに従って同様な方法で予測値が生成可能である。例えば、IntraPredMode=1で示される予測モードの予測値は、図5(c)で示されるH.264の水平予測と同一となる。以上が、本発明の本実施の形態における単方向イントラ予測画像生成部601の説明である。 Even for conditions other than the above, a predicted value can be generated by a similar method according to the table of FIG. For example, the prediction value of the prediction mode indicated by IntraPredMode = 1 is H.264 shown in FIG. This is the same as H.264 horizontal prediction. The above is description of the unidirectional intra estimated image generation part 601 in this Embodiment of this invention.
 次に、双方向イントラ予測画像生成部602について詳細に説明する。図13に双方向イントラ予測画像生成部602のブロック図を示す。双方向イントラ予測画像生成部602は、第一単方向イントラ予測画像生成部1301、第二単方向イントラ予測画像生成部1302、重み付き平均部1303を有し、入力された参照画像信号4419(124)に基づいて、2つの単方向イントラ予測画像を生成し、これらを重み付き平均することにより予測画像信号4420(126)を生成する機能を有する。 Next, the bidirectional intra predicted image generation unit 602 will be described in detail. FIG. 13 shows a block diagram of the bidirectional intra-predicted image generation unit 602. The bidirectional intra predicted image generation unit 602 includes a first unidirectional intra predicted image generation unit 1301, a second unidirectional intra predicted image generation unit 1302, and a weighted average unit 1303. The input reference image signal 4419 (124 ), Two unidirectional intra prediction images are generated, and a weighted average of these is generated to generate a prediction image signal 4420 (126).
 第一単方向イントラ予測画像生成部1301及び第二単方向イントラ予測画像生成部1302の機能は同一である。いずれも符号化制御部115で制御される予測モード情報に従って与えられた予測モードに対応する予測画像信号を生成する。第一単方向イントラ予測画像生成部1301から第一予測画像信号1304が、第二単方向イントラ予測画像生成部1302から第二予測画像信号1305が出力される。それぞれの予測画像信号が重み付き平均部1303に入力され、重み付き平均処理が行われる。 The functions of the first unidirectional intra predicted image generation unit 1301 and the second unidirectional intra predicted image generation unit 1302 are the same. In either case, a prediction image signal corresponding to a prediction mode given according to prediction mode information controlled by the encoding control unit 115 is generated. A first predicted image signal 1304 is output from the first unidirectional intra predicted image generation unit 1301, and a second predicted image signal 1305 is output from the second unidirectional intra predicted image generation unit 1302. Each predicted image signal is input to the weighted average unit 1303, and weighted average processing is performed.
 図14のテーブルは、双方向イントラ予測モードから、2つの単方向イントラ予測モードを導出するためのテーブルである。ここで数式(4)を用いてBiPredIdxが導出される。 The table in FIG. 14 is a table for deriving two unidirectional intra prediction modes from the bidirectional intra prediction mode. Here, BiPredIdx is derived using Equation (4).
 例えば、PuSize=PU_8x8、IntraPredMode=33の場合、図7からIntraUniModeNum=33であることが判るので、BiPredIdx=0であることが判る。結果として、図14から、第一単方向イントラ予測モード(MappedBi2Uni(0,idx))が1、第二単方向イントラ予測モード(MappedBi2Uni(1,idx))が0であることが導出される。他のPuSize及びIntraPredModeにおいても同様の方法で2つの予測モードを導出することが可能である。なお、以後、第一単方向イントラ予測モードをIntraPredModeL0、第二単方向イントラ予測モードをIntraPredModeL1で表現する。 For example, in the case of PuSize = PU_8 × 8 and IntraPredMode = 33, it can be seen from FIG. 7 that IntraUniModeNum = 33, and BiPredIdx = 0. As a result, it is derived from FIG. 14 that the first unidirectional intra prediction mode (MappedBi2Uni (0, idx)) is 1 and the second unidirectional intra prediction mode (MappedBi2Uni (1, idx)) is 0. In other PuSize and IntraPredMode, it is possible to derive two prediction modes by the same method. Hereinafter, the first unidirectional intra prediction mode is expressed as IntraPredModeL0, and the second unidirectional intra prediction mode is expressed as IntraPredModeL1.
 このように第一単方向イントラ予測画像生成部1301及び第二単方向イントラ予測画像生成部1302で生成された第一予測画像信号1304、第二予測画像信号1305が、重み付き平均部1303へと入力される。 Thus, the first predicted image signal 1304 and the second predicted image signal 1305 generated by the first unidirectional intra predicted image generation unit 1301 and the second unidirectional intra predicted image generation unit 1302 are sent to the weighted average unit 1303. Entered.
 重み付き平均部1303では、IntraPredModeL0及びIntraPredModeL1の予測方向を元にユークリッド距離或いは市街地距離(マンハッタン距離)を計算し、重み付き平均処理で用いる重み成分を導出する。各画素の重み成分は予測に用いる参照画素からのユークリッド距離若しくは市街地距離の逆数で表され、数式(5)で一般化される。ここでユークリッド距離を用いる場合は、ΔLは数式(6)で表される。一方、市街地距離を用いる場合は、ΔLは数式(7)で表される。各予測モードに関する重みテーブルは、数式(8)に一般化される。従って、画素位置nにおける最終的な予測信号は、数式(9)で示される。 The weighted average unit 1303 calculates a Euclidean distance or a city area distance (Manhattan distance) based on the prediction directions of IntraPredModeL0 and IntraPredModeL1, and derives a weight component used in the weighted average process. The weight component of each pixel is represented by the reciprocal of the Euclidean distance or the city distance from the reference pixel used for prediction, and is generalized by Expression (5). Here, when using the Euclidean distance, ΔL is expressed by Equation (6). On the other hand, when using the city distance, ΔL is expressed by Equation (7). The weight table for each prediction mode is generalized to Equation (8). Therefore, the final prediction signal at the pixel position n is expressed by Equation (9).
 本実施形態では、予測画素の生成に2つの予測モードを選択して予測信号を生成しているが、別の実施形態として3つ以上の予測モードを選択して予測値を生成してもよい。この場合、参照画素から予測画素の空間的距離の逆数の比を重み係数と設定すればよい。 In this embodiment, the prediction signal is generated by selecting two prediction modes for generating the prediction pixel. However, as another embodiment, a prediction value may be generated by selecting three or more prediction modes. . In this case, the ratio of the reciprocal of the spatial distance from the reference pixel to the prediction pixel may be set as the weighting factor.
 また、本実施形態では、予測モードが使用する参照画素からのユークリッド距離、或いは市街地距離の逆数をそのまま重み成分としているが、別の一実施形態として参照画素からのユークリッド距離、市街地距離を変数とした分布モデルを用いて重み成分を設定してもよい。分布モデルは線形モデルやM次関数(M≧1)、片側ラプラス分布や片側ガウス分布といった非線形関数、ある固定値を参照画素との距離に依らず固定値、のうちの少なくとも一つを用いる。片側ガウス分布をモデルとして用いた場合、重み成分は数式(10)で表される。また、片側ラプラス分布をモデルとして用いた場合重み成分は数式(11)で表される。 In this embodiment, the Euclidean distance from the reference pixel used in the prediction mode or the reciprocal of the urban area distance is used as a weight component as it is, but as another embodiment, the Euclidean distance from the reference pixel and the urban area distance are variables. The weight component may be set using the distributed model. The distribution model uses at least one of a linear model, an M-order function (M ≧ 1), a nonlinear function such as a one-sided Laplace distribution or a one-sided Gaussian distribution, or a fixed value that is a fixed value regardless of the distance from the reference pixel. When the one-sided Gaussian distribution is used as a model, the weight component is expressed by Equation (10). Further, when the one-sided Laplace distribution is used as a model, the weight component is expressed by Expression (11).
 また、自己相関関数をモデル化した等方相関モデル、楕円相関モデル、ラプラス関数やガウス関数を一般化した一般化ガウスモデルを重み成分のモデルとして用いてもよい。 Further, an isotropic correlation model obtained by modeling an autocorrelation function, an elliptic correlation model, a generalized Gaussian model obtained by generalizing a Laplace function or a Gaussian function may be used as the weight component model.
 数式(5)、数式(8)、数式(10)、数式(11)で示される重み成分を予測画像生成の際に都度計算した場合、複数の乗算器が必要となり、ハードウェア規模が増大する。このため、予め予測モード毎の相対距離に応じて重み成分を計算し、メモリに保持することで上記演算に必要なる回路規模を削減することができる。ここでは、市街地距離を用いた場合の重み成分の導出方法について説明する。 When the weight components represented by Equation (5), Equation (8), Equation (10), and Equation (11) are calculated each time the predicted image is generated, a plurality of multipliers are required, and the hardware scale increases. . For this reason, the circuit scale required for the said calculation can be reduced by calculating a weight component beforehand according to the relative distance for every prediction mode, and hold | maintaining in a memory. Here, a method for deriving the weight component when the city distance is used will be described.
 IntraPredModeL0の市街地距離ΔLL0とIntraPredModeL1の市街地距離ΔLL1は、数式(7)より計算される。ここで、相対距離は2つの予測モードの予測方向によって変化する。例として、PuSize=PU_4x4の場合の代表的な距離を図14A、B、Cに示す。図15Aは、IntraPredModeLX=0の場合の市街地距離を示している。図15Bは、IntraPredModeLX=1の場合の市街地距離を示している。また、図15Cは、IntraPredModeLX=3の場合の市街地距離を示している。同様に、それぞれの予測モードに応じて数式(6)或いは数式(7)を用いて距離を導出することができる。但し、IntraPredModeLX=2のDC予測の場合、全ての画素位置で距離を2としている。図16にPuSize=PU_4x4の場合の代表的な5つの予測モードにおける距離のテーブルを示す。IntraPredModeLXの数が多い場合、これらの距離テーブルのテーブルサイズが増加する場合がある。 The city area distance ΔL L0 of IntraPredMode L0 and the city area distance ΔL L1 of IntraPredMode L1 are calculated from Equation (7). Here, the relative distance varies depending on the prediction direction of the two prediction modes. As an example, typical distances in the case of PuSize = PU — 4 × 4 are shown in FIGS. FIG. 15A shows the city distance when IntraPredModeLX = 0. FIG. 15B shows the city distance in the case of IntraPredModeLX = 1. FIG. 15C shows the city distance in the case of IntraPredModeLX = 3. Similarly, the distance can be derived using Expression (6) or Expression (7) according to each prediction mode. However, in the case of DC prediction with IntraPredModeLX = 2, the distance is 2 at all pixel positions. FIG. 16 shows a table of distances in five typical prediction modes when PuSize = PU — 4 × 4. When the number of IntraPredModeLX is large, the table sizes of these distance tables may increase.
 本発明の本実施の形態では、いくつかの予測角度の近い予測モードの距離テーブルを共有化することにより、必要なメモリ量を削減している。図17は、距離テーブル導出に用いるIntraPredModeLXのマッピングを示している。ここでは予測角度が45度刻みの予測モードとDC予測に対応する予測モードのみのテーブルを用意しており、それ以外の予測角度を、用意した基準の予測モードに近い方にマッピングする例を示している。なお、基準の予測モードとの距離が同じ場合は、インデックスが小さい方にマッピングしている。MappedIntraPredModeに示される予測モードが図17から参照され、距離テーブルが導出できる。 In the present embodiment of the present invention, the required memory amount is reduced by sharing the distance table of several prediction modes with close prediction angles. FIG. 17 shows the mapping of IntraPredModeLX used for distance table derivation. Here, an example is shown in which a table of only the prediction mode corresponding to the prediction mode corresponding to the prediction mode and the DC prediction in 45 degrees is prepared, and other prediction angles are mapped closer to the prepared reference prediction mode. ing. When the distance from the reference prediction mode is the same, the index is mapped to the smaller one. The prediction mode shown in “MappedIntraPredMode” is referred to from FIG. 17, and a distance table can be derived.
 上記距離テーブルを利用することにより、数式(12)を用いて2つの予測モードの画素毎の相対距離が算出される。数式(12)を用いて、画素位置nにおける最終的な予測信号は、数式(13)で示される。ここで、小数点演算を用いることによるハードウェア規模増加を避けるため、重み成分を予めスケーリングし、整数演算に直すと数式(14)で表せる。ここで、例えば小数点部分を10ビット精度で表現した場合、WM=1024、Offset=512、SHIFT=10となる。これらは数式(15)の関係を満たす。 By using the distance table, the relative distance for each pixel in the two prediction modes is calculated using Equation (12). Using Equation (12), the final prediction signal at the pixel position n is represented by Equation (13). Here, in order to avoid an increase in hardware scale due to the use of decimal point arithmetic, the weight component is scaled in advance and converted to integer arithmetic, it can be expressed by Equation (14). Here, for example, when the decimal part is expressed with 10-bit precision, WM = 1024, Offset = 512, and SHIFT = 10. These satisfy the relationship of Expression (15).
 本発明の本実施の形態における片側ラプラス分布モデルを用いた重み成分をテーブル化した例を図18A、Bに示す。図18AはPuSize=PU_4x4の場合の重み成分テーブルを示している。また、図18BはPuSize=PU_8x8の場合の重み成分テーブルを示している。それ以外のPuSizeに関しても、数式(5)、数式(8)、数式(10)、数式(11)を用いることにより導出が可能である。 
 以上が、本発明の本実施の形態に関わるイントラ予測部109の詳細である。
FIGS. 18A and 18B show examples in which weight components using the one-sided Laplace distribution model in the present embodiment of the present invention are tabulated. FIG. 18A shows a weight component table in the case of PuSize = PU — 4 × 4. FIG. 18B shows a weight component table in the case of PuSize = PU_8 × 8. Other PuSizes can also be derived using Equation (5), Equation (8), Equation (10), and Equation (11).
The above is the details of the intra prediction unit 109 according to the present embodiment of the present invention.
 <シンタクス構成 1> 
 以下、図44の動画像復号化装置4400が利用するシンタクスについて説明する。 
 シンタクスは、動画像復号化装置4400が動画像データを復号化する際の符号化データ(例えば、図1の符号化データ127)の構造を示している。第1の実施形態に代表される画像符号化装置は同じシンタクス構造を用いてこの符号化データを符号化する。図44の画像符号化装置が利用するシンタクス2000を図20に例示する。シンタクス2000については、第1の実施形態と同一であるため、詳細な説明は省略する。
<Syntax structure 1>
Hereinafter, the syntax used by the moving picture decoding apparatus 4400 in FIG. 44 will be described.
The syntax indicates the structure of encoded data (for example, encoded data 127 in FIG. 1) when the moving image decoding apparatus 4400 decodes moving image data. The image encoding apparatus represented by the first embodiment encodes this encoded data using the same syntax structure. FIG. 20 shows an example of syntax 2000 used by the image coding apparatus in FIG. Since the syntax 2000 is the same as that of the first embodiment, detailed description thereof is omitted.
 次に、本実施の形態に係るレディクションユニットシンタクスの一例について説明する。 Next, an example of the reflection unit syntax according to the present embodiment will be described.
 図22Aに、プレディクションユニットシンタクスの一例を示す。図中のpred_modeは当該プレディクションユニットの予測タイプをしている。MODE_INTRAは予測タイプがイントラ予測であることを示す。intra_split_flagは当該プレディクションユニットをさらに4つのプレディクションユニットに分割するか否かを示すフラグである。intra_split_flagが1の場合、プレディクションユニットを、縦横のサイズ半分で4分割したものをプレディクションユニットとする。intra_split_flagが0の場合、プレディクションユニットを分割しない。 FIG. 22A shows an example of the prediction unit syntax. Pred_mode in the figure indicates the prediction type of the prediction unit. MODE_INTRA indicates that the prediction type is intra prediction. intra_split_flag is a flag indicating whether or not the prediction unit is further divided into four prediction units. When intra_split_flag is 1, a prediction unit is obtained by dividing a prediction unit into four in half in the vertical and horizontal sizes. When intra_split_flag is 0, the prediction unit is not divided.
 intra_luma_bipred_flag[i]は当該プレディクションユニットに適用した予測モードIntraPredModeが単方向イントラ予測モードか双方向イントラ予測モードであるかを示すフラグである。iは分割されたプレディクションユニットの位置を示しており、前記intra_split_flagが0の場合には0、前記intra_split_flagが1の場合には0~3が設定される。当該フラグは、図9、10、11に示される当該プレディクションユニットのIntraBipredFlagの値がセットされている。 Intra_luma_bipred_flag [i] is a flag indicating whether the prediction mode IntraPredMode applied to the prediction unit is a unidirectional intra prediction mode or a bidirectional intra prediction mode. i indicates the position of the divided prediction unit. When the intra_split_flag is 0, 0 is set, and when the intra_split_flag is 1, 0 to 3 are set. The flag is set with the value of IntraBiprededFlag of the prediction unit shown in FIGS.
 intra_luma_bipred_flag[i]が1の場合、当該プレディクションユニットは双方向イントラ予測であることを示し、用意された複数の双方向イントラ予測モードの内、使用した双方向イントラ予測モードを特定する情報であるintra_luma_bipred_mode[i]を復号化する。intra_luma_bipred_mode[i]は、図7に示される双方向イントラ予測モード数IntraBiModeNumに従って等長復号化されてもよいし、予め決定された符号表を用いて復号化されてもよい。intra_luma_bipred_flag[i]が0の場合、当該プレディクションユニットは単方向イントラ予測であることを示し、隣接ブロックから予測復号化を行う。 When intra_luma_bipred_flag [i] is 1, this indicates that the prediction unit is bi-directional intra prediction, and is information that identifies the used bi-directional intra prediction mode among a plurality of prepared bi-directional intra prediction modes. Intra_luma_bipred_mode [i] is decoded. intra_luma_bipred_mode [i] may be decoded in equal length according to the bidirectional intra prediction mode number IntraBiModeNum shown in FIG. 7, or may be decoded using a predetermined code table. When intra_luma_bipred_flag [i] is 0, it indicates that the prediction unit is unidirectional intra prediction, and predictive decoding is performed from adjacent blocks.
 prev_intra_luma_unipred_flag[i]は、隣接ブロックから計算される予測モードの予測値MostProbableと当該プレディクションユニットのイントラ予測モードが同一であるか否かを示すフラグである。MostProbableの計算方法の詳細は後述する。prev_intra_luma_unipred_flag[i]が1の場合、前記MostProbableとイントラ予測モードIntraPredModeが等しいことを示す。prev_intra_luma_unipred_flag[i]が0の場合、前記MostProbableとイントラ予測モードIntraPredModeは異なることを示し、イントラ予測モードIntraPredModeがさらにMostProbable以外のいずれのモードであるかを特定する情報rem_intra_luma_unipred_mode[i]が復号化される。rem_intra_luma_unipred_mode[i]は、図7に示される双方向イントラ予測モード数IntraUniModeNumに従って等長復号化されてもよいし、予め決定された符号表を用いて復号化されてもよい。イントラ予測モードIntraPredModeからrem_intra_luma_unipred_mode[i]は数式(16)を用いて計算される。 Prev_intra_luma_unipred_flag [i] is a flag indicating whether or not the prediction value MostProbable of the prediction mode calculated from the adjacent block and the intra prediction mode of the prediction unit are the same. Details of the MostProbable calculation method will be described later. When prev_intra_luma_unipred_flag [i] is 1, it indicates that the MostProbable and the intra prediction mode IntraPredMode are equal. When prev_intra_luma_unipred_flag [i] is 0, it indicates that the MostProbable and the intra prediction mode IntraPredMode are different, and the information rem_intraprelum decoding that further specifies the intra prediction mode IntraPredMode other than MostProbable. . rem_intra_luma_unipred_mode [i] may be decoded in equal length according to the bidirectional intra prediction mode number IntraUniModeNum shown in FIG. 7, or may be decoded using a predetermined code table. From the intra prediction mode IntraPredMode, rem_intra_luma_unipred_mode [i] is calculated using Equation (16).
 次に、予測モードの予測値であるMostProbableの計算方法について説明する。MostProbableは数式(17)に従って計算される。なお、Min(x,y)は入力x、yのうち小さい方を出力するパラメータである。 Next, a method for calculating MostProbable, which is a predicted value in the prediction mode, will be described. MostProbable is calculated according to Equation (17). Min (x, y) is a parameter for outputting the smaller one of the inputs x and y.
 また、intraPredModeA,intraPredModeBは、復号化プレディクションユニットに対して左に及び上に隣接するプレディクションユニットのイントラ予測モードを示している。以降、intraPredModeAとintraPredModeBをまとめてintraPredModeNと表現する。NにはA若しくはBが設定される。図23に示すフローチャートを用いてintraPredModeNの計算方法を説明する。まず、隣接するプレディクションユニットが属するコーディングツリーユニットが利用可能であるかを判定し(ステップS2301)、当該コーディングツリーユニットが利用不可能である場合には(S2301がNO)、intraPredModeNに参照不可能を示す"-1"をセットする。一方、当該コーディングツリーユニットが利用可能である場合には(S2301がYES)、次に当該隣接プレディクションユニットがイントラ予測が適用されたか否かを判定する(ステップS2302)。当該隣接プレディクションユニットがイントラ予測ではない場合(S2302がNO)、intraPredModeNには"Intra_DC"を意味する"2"をセットする。一方、当該隣接プレディクションユニットがイントラ予測である場合(S2302がYES)、次に、当該隣接プレディクションユニットが双方向イントラ予測あるか否かを判定する(ステップS2303)。当該隣接プレディクションユニットが双方向イントラ予測ではない、つまり単方向イントラ予測の場合(S2303がNO)、intraPredModeNには隣接プレディクションユニットの予測モードIntraPredModeをセットする。一方、当該隣接PUが双方向イントラ予測である場号(S2303がYES)、隣接ブロックの予測モードを単方向イントラ予測モードに変換する。具体的には数式(18)を用いてintraPredModeNを計算する。ここで、IntraUniModeNumは当該隣接プレディクションユニットのサイズによって決定される単方向イントラ予測モードの数であり、図7にその一例が示されている。また、MappedBi2Uni(List, idx)は双方向イントラ予測モードを単方向イントラ予測モードに変換するテーブルである。Listは双方向イントラ予測モードを構成する2種類の単方向イントラ予測モードのうち、List0の単方向イントラ予測モード(図9,10,11に示されるIntraPredTypeL0[]に相当)を単方向イントラ予測モードへの変換に使用するか、List1の単方向イントラ予測モード図9,10,11に示されるIntraPredTypeL1[]に相当)をを使用するかを示すフラグであり、0であればList0、1であればList1を単方向イントラ予測モードへの変換に使用する。図14に、変換テーブルの例が示されている。図中の数値は、図9,10,11に示されるIntraPredModeに対応している。 In addition, intraPredModeA and intraPredModeB indicate intra prediction modes of prediction units adjacent to the left and above the decoded prediction unit. Hereinafter, intraPredModeA and intraPredModeB are collectively expressed as intraPredModeN. N is set to A or B. A method of calculating intraPredModeN will be described using the flowchart shown in FIG. First, it is determined whether a coding tree unit to which an adjacent prediction unit belongs can be used (step S2301). If the coding tree unit cannot be used (NO in S2301), reference to intraPredModeN is not possible. “−1” indicating “” is set. On the other hand, if the coding tree unit is available (YES in S2301), it is next determined whether or not intra prediction is applied to the adjacent prediction unit (step S2302). When the adjacent prediction unit is not intra prediction (NO in S2302), “2” meaning “Intra_DC” is set in intraPredModeN. On the other hand, when the adjacent prediction unit is intra prediction (YES in S2302), it is next determined whether or not the adjacent prediction unit is bidirectional intra prediction (step S2303). When the adjacent prediction unit is not bidirectional intra prediction, that is, in the case of unidirectional intra prediction (NO in S2303), the prediction mode IntraPredMode of the adjacent prediction unit is set in intraPredModeN. On the other hand, when the adjacent PU is bidirectional intra prediction (S2303 is YES), the prediction mode of the adjacent block is converted to the unidirectional intra prediction mode. Specifically, intraPredModeN is calculated using Equation (18). Here, IntraUniModeNum is the number of unidirectional intra prediction modes determined by the size of the adjacent prediction unit, and an example thereof is shown in FIG. Also, “MappedBi2Uni (List, idx)” is a table for converting the bidirectional intra prediction mode into the unidirectional intra prediction mode. List is the unidirectional intra prediction mode of List0 (corresponding to IntraPredTypeL0 [] shown in FIGS. 9, 10, and 11) of the two unidirectional intra prediction modes constituting the bidirectional intra prediction mode. Is a flag indicating whether to use the unidirectional intra prediction mode of List1 or equivalent to IntraPredTypeL1 [] shown in FIGS. 9, 10, and 11; For example, List1 is used for conversion to the unidirectional intra prediction mode. FIG. 14 shows an example of the conversion table. The numerical values in the figure correspond to IntraPredMode shown in FIGS.
 以上のように計算されたintraPredModeNと数式(18)を用いて計算されたMostProbableが-1の場合には、MostProbableを2(Intra_DC)に置き換える。また、MostProbableが復号化プレディクションユニットの単方向イントラ予測モード数IntraUniPredModeNumよりも大きい場合には、数式(19)下式を用いてMostProbableを再計算するする。MappedMostProbable()は、MostProbableを変換するテーブルであり、図24に一例が示されている。 When MostProbable calculated using intraPredModeN calculated as described above and Equation (18) is −1, MostProbable is replaced with 2 (Intra_DC). When MostProbable is larger than the number of unidirectional intra prediction modes IntraUniPredModeNum of the decoding prediction unit, MostProbable is recalculated using the following equation (19). “MappedMostProble ()” is a table for converting MostProbable, and an example is shown in FIG. 24.
 <シンタクス構成 2> 
 次に、プレディクションユニットシンタクスに係る別の例を図22Cに示す。pred_mode、intra_split_flagは先に述べたシンタクスの一例と同様であるので説明を省略する。luma_pred_mode_code_type[i]は当該プレディクションユニットに適用した予測モードIntraPredModeの種類を示しており、0(IntraUnipredMostProb)は単方向イントラ予測でイントラ予測モードがMostProbableと等しい、1(IntraUnipredRem)は単方向イントラ予測でイントラ予測モードがMostProbableと異なる、2(IntraBipred)は双方向イントラ予測モードであることを夫々示している。図24に双方向イントラ予測モード、1は及び単方向イントラ予測(MostProbableと同一の予測モードであるか否か)を示す。図25にluma_pred_mode_code_typeと対応する意味、及びbin、図7に示すモード構成に従ったモード数の割当の一例を示している。luma_pred_mode_code_type[i]が0の場合、イントラ予測モードがMostProbableモードとなるため、これ以上の情報の復号化は必要無い。luma_pred_mode_code_type[i]が1の場合、イントラ予測モードIntraPredModeがさらにMostProbable以外のいずれのモードであるかを特定する情報rem_intra_luma_unipred_mode[i]が復号化される。rem_intra_luma_unipred_mode[i]は、図7に示される双方向イントラ予測モード数IntraUniModeNumに従って等長復号化されてもよいし、予め決定された符号表を用いて復号化されてもよい。イントラ予測モードIntraPredModeからrem_intra_luma_unipred_mode[i]は数式(16)を用いて計算される。また、luma_pred_mode_code_type[i]が2の場合、当該プレディクションユニットは双方向イントラ予測であることを示し、用意された複数の双方向イントラ予測モードの内、使用した双方向イントラ予測モードを特定する情報であるintra_luma_bipred_mode[i]を復号化する。intra_luma_bipred_mode[i]は、図7に示される双方向イントラ予測モード数IntraBiModeNumに従って等長復号化されてもよいし、予め決定された符号表を用いて復号化されてもよい。 
 以上が、本実施の形態に係るシンタクス構成である。
<Syntax structure 2>
Next, another example of the prediction unit syntax is shown in FIG. 22C. Since pred_mode and intra_split_flag are the same as the syntax example described above, description thereof is omitted. luma_pred_mode_code_type [i] indicates the type of the prediction mode IntraPredMode applied to the prediction unit, where 0 (IntraUnifiedMostProb) is unidirectional intra prediction and the intra prediction mode is the same as MostProbable, 1 (IntraUnipre intrareprediction) The intra prediction mode is different from MostProbable, and 2 (IntraBipred) indicates a bidirectional intra prediction mode. FIG. 24 shows bidirectional intra prediction modes, 1 and unidirectional intra prediction (whether or not the same prediction mode as MostProbable). FIG. 25 shows an example of assignment of the number of modes according to the meaning corresponding to luma_pred_mode_code_type, bin, and the mode configuration shown in FIG. When luma_pred_mode_code_type [i] is 0, the intra prediction mode is the MostProbable mode, so no further information decoding is necessary. When luma_pred_mode_code_type [i] is 1, information rem_intra_luma_unipred_mode [i] that specifies which mode other than MostProbable is the intra prediction mode IntraPredMode is decoded. rem_intra_luma_unipred_mode [i] may be decoded in equal length according to the bidirectional intra prediction mode number IntraUniModeNum shown in FIG. 7, or may be decoded using a predetermined code table. From the intra prediction mode IntraPredMode, rem_intra_luma_unipred_mode [i] is calculated using Equation (16). Further, when luma_pred_mode_code_type [i] is 2, it indicates that the prediction unit is bidirectional intra prediction, and information that identifies the used bidirectional intra prediction mode among the prepared bidirectional intra prediction modes. Intra_luma_bipred_mode [i] is decoded. intra_luma_bipred_mode [i] may be decoded in equal length according to the bidirectional intra prediction mode number IntraBiModeNum shown in FIG. 7, or may be decoded using a predetermined code table.
The above is the syntax configuration according to the present embodiment.
 <シンタクス構成 3> 
 プレディクションユニットシンタクスに係るさらに別の例を図22Dに示す。本例では、図22Aで示されるプレディクションユニットシンタクスをベースに、双方向イントラ予測を使用可能とするか、双方向イントラ予測を使用不可能として従来の単方向イントラ予測のみを使用可能とするかを、復号化するプレディクションユニット内で切り替える場合のシンタクスを示している。 
 なお、pred_mode、intra_split_flagは先に述べたシンタクスの一例と同様であるので説明を省略する。
<Syntax structure 3>
Yet another example of the prediction unit syntax is shown in FIG. 22D. In this example, based on the prediction unit syntax shown in FIG. 22A, whether bidirectional intra prediction can be used or whether conventional intra unidirectional prediction can be used with bidirectional intra prediction disabled. Shows the syntax for switching within the prediction unit to be decoded.
Note that pred_mode and intra_split_flag are the same as the syntax example described above, and thus description thereof is omitted.
 intra_bipred_flagは、復号化プレディクションユニット内で双方向イントラ予測を使用可能とするか否かを示すフラグである。intra_bipred_flagが0の場合、復号化プレディクションユニット内で双方向イントラ予測が使用されないことを示している。intra_split_flagが1、つまり復号化プレディクションユニットがさらに4分割されている場合においても、全てのプレディクションユニットにおいて双方向イントラ予測は使用されず、単方向イントラ予測のみが有効となる。 Intra_bipred_flag is a flag indicating whether or not bidirectional intra prediction can be used in the decoding prediction unit. When intra_bipred_flag is 0, it indicates that bi-directional intra prediction is not used in the decoding prediction unit. Even when intra_split_flag is 1, that is, when the decoded prediction unit is further divided into four, bi-directional intra prediction is not used in all prediction units, and only uni-directional intra prediction is effective.
 intra_bipred_flagが1の場合、復号化プレディクションユニット内で双方向イントラ予測が使用可能であることを示している。intra_split_flagが1、つまり復号化プレディクションユニットがさらに4分割されている場合においても、全てのプレディクションユニットにおいて、単方向イントラ予測に加え双方向イントラ予測が選択可能となる。 When intra_bipred_flag is 1, it indicates that bidirectional intra prediction can be used in the decoding prediction unit. Even when intra_split_flag is 1, that is, when the decoded prediction unit is further divided into four, in all prediction units, bidirectional intra prediction can be selected in addition to unidirectional intra prediction.
 双方向イントラ予測が不要である予測が比較的容易な領域(例えば、平坦領域)では、intra_bipred_flagを0として復号化して、双方向イントラ予測を使用不可能とすることにより、双方向イントラ予測モードの復号化に必要な符号量が削減可能となるため、符号化効率は改善する。 In a region where bi-directional intra prediction is unnecessary (for example, a flat region), the intra-bipred_flag is decoded as 0 to disable bi-directional intra prediction. Since the amount of code required for decoding can be reduced, the coding efficiency is improved.
 <シンタクス構成 4> 
 プレディクションユニットシンタクスに係るさらに別の例を図22Eに示す。本例では、図22Cで示されるプレディクションユニットシンタクスをベースに、双方向イントラ予測を使用可能とするか、双方向イントラ予測を使用不可能として従来の単方向イントラ予測のみを使用可能とするかを、復号化プレディクションユニット内で切り替える場合のシンタクスを示している。intra_bipred_flagは、復号化プレディクションユニット内で双方向イントラ予測を使用可能とするか否かを示すフラグでああり、前述のintra_bipred_flagと同様であるので説明を省略する。
<Syntax structure 4>
Still another example relating to the prediction unit syntax is shown in FIG. 22E. In this example, based on the prediction unit syntax shown in FIG. 22C, whether bidirectional intra prediction can be used or whether only conventional unidirectional intra prediction can be used with bidirectional intra prediction disabled. Shows the syntax for switching in the decoding prediction unit. intra_bipred_flag is a flag indicating whether or not bi-directional intra prediction can be used in the decoding prediction unit, and is the same as the above-described intra_bipred_flag, and thus the description thereof is omitted.
 (第1の変形例) 
<イントラ予測部 第一の変形例> 
 イントラ予測部4408に係る第1の変形例として、JCTVC-B205_draft002, 5.2.1節" Intra prediction process for luma samples ", JCT-VC 2nd Meeting Geneva, July, 2010に示される適応参照画素フィルタリングと組み合わせても構わない。図26に、適応参照画素フィルタリングを用いた際のイントラ予測部4408(109)を示している。図6で示されるイントラ予測部4408(109)とは、参照画素フィルタ部2601が追加されている点が異なる。参照画素フィルタ部2601では、参照画像信号4419(124)及び予測モード605を入力し、後述する適応的なフィルタ処理を行い、被フィルタ参照画像信号2602を出力する。被フィルタ参照画像信号2602は単方向イントラ予測画像生成部601と双方向イントラ予測画像生成部602に入力される。参照画素フィルタ部2601以外の構成及び処理に関しては、図6で示されるイントラ予測部4408(109)と同様であるので、説明を省略する。
(First modification)
<Intra prediction unit first modification>
As a first modification related to the intra prediction unit 4408, in combination with adaptive reference pixel filtering shown in JCTVC-B205_draft002, section 5.2.1 “Intra prediction process for luma samples”, JCT-VC 2nd Meeting Geneva, July, 2010 It doesn't matter. FIG. 26 shows an intra prediction unit 4408 (109) when adaptive reference pixel filtering is used. It differs from the intra prediction unit 4408 (109) shown in FIG. 6 in that a reference pixel filter unit 2601 is added. The reference pixel filter unit 2601 receives the reference image signal 4419 (124) and the prediction mode 605, performs adaptive filter processing described later, and outputs a filtered reference image signal 2602. The filtered reference image signal 2602 is input to the unidirectional intra predicted image generation unit 601 and the bidirectional intra predicted image generation unit 602. The configuration and processing other than the reference pixel filter unit 2601 are the same as those of the intra prediction unit 4408 (109) shown in FIG.
 次に参照画素フィルタ部2601について説明する。参照画素フィルタ部2601では、予測モード605に含まれる参照画素フィルタフラグ及びイントラ予測モードに従って、イントラ予測に使用する参照画素をフィルタリングするか否かを決定する。参照画素フィルタフラグはイントラ予測モードIntraPredModeが”Intra_DC”以外の値の場合に、参照画素をフィルタリングするか否かを示すフラグである。参照画素フィルタフラグが1の場合、参照画素をフィルタリングする。また、参照画素フィルタフラグ0の場合、参照画素をフィルタリングしない。なお、IntraPredModeが”Intra_DC”の場合には、参照画素はフィルタリングせず、参照画素フィルタフラグは0にセットされる。参照画素フィルタフラグが1の場合、数式(20)に示されるフィルタリングにより、被フィルタ参照画像信号2602が計算される。なお、p[x,y]はフィルタ前の参照画素、pf[x,y]はフィルタ的用語の参照画素を示している。また、x,yは当該プレディクションユニット内の左上画素位置をx=0, y=0とした場合の、参照画素の相対位置を示している。また、PuPartSizeは当該プレディクションユニットのサイズ(画素)を示している。 Next, the reference pixel filter unit 2601 will be described. The reference pixel filter unit 2601 determines whether or not to filter reference pixels used for intra prediction according to the reference pixel filter flag and the intra prediction mode included in the prediction mode 605. The reference pixel filter flag is a flag indicating whether or not reference pixels are filtered when the intra prediction mode IntraPredMode is a value other than “Intra_DC”. When the reference pixel filter flag is 1, the reference pixel is filtered. In the case of the reference pixel filter flag 0, the reference pixel is not filtered. When IntraPredMode is “Intra_DC”, the reference pixel is not filtered and the reference pixel filter flag is set to 0. When the reference pixel filter flag is 1, a filtered reference image signal 2602 is calculated by filtering shown in Expression (20). Note that p [x, y] indicates a reference pixel before filtering, and pf [x, y] indicates a reference pixel in filter terms. X and y indicate the relative positions of the reference pixels when the upper left pixel position in the prediction unit is x = 0 and y = 0. PuPartSize indicates the size (pixel) of the prediction unit.
 <シンタクス構成 5> 
 図27A,図27Bには、適応参照画素フィルタを行う際のプレディクションユニットシンタクス構造を示している。図27Aは、図22Aに適応参照画素フィルタに関するシンタクスintra_luma_filter_flag[i]を追加している。また、図27Bは、図22Cに適応参照画素フィルタに関するシンタクスintra_luma_filter_flag[i]を追加している。intra_luma_filter_flag[i]はイントラ予測モードIntraPredMode[i]がIntra_DC以外の場合に、さらに復号化される。当該フラグが0の場合、上記参照画素のフィルタリングは行われないことを示す。また、intra_luma_filter_flag[i]が1の場合、上記参照画素のフィルタリングを適用することを示す。
<Syntax structure 5>
27A and 27B show a prediction unit syntax structure when performing adaptive reference pixel filtering. FIG. 27A adds the syntax intra_luma_filter_flag [i] related to the adaptive reference pixel filter to FIG. 22A. In addition, FIG. 27B adds syntax intra_luma_filter_flag [i] related to the adaptive reference pixel filter to FIG. 22C. intra_luma_filter_flag [i] is further decoded when the intra prediction mode IntraPredMode [i] is other than Intra_DC. When the flag is 0, it indicates that the reference pixel is not filtered. Further, when intra_luma_filter_flag [i] is 1, it indicates that the reference pixel filtering is applied.
 上記の例では、イントラ予測モードIntraPredMode[i]がIntra_DC以外の場合に、intra_luma_filter_flag[i]を復号化したが、他の例として、IntraPredMode[i]が0~2の場合には、intra_luma_filter_flag[i]を復号化しなくても構わない。この場合、intra_luma_filter_flag[i]は0に設定される。 In the above example, intra_luma_filter_flag [i] is decoded when the intra prediction mode IntraPredMode [i] is other than Intra_DC. However, as another example, when IntraPredMode [i] is 0 to 2, intra_luma_filter_flag [i ] Need not be decrypted. In this case, intra_luma_filter_flag [i] is set to 0.
 また、図22B、図22D、図22Eに示される他のシンタクス構造について、上記で説明したintra_luma_filter_flag[i]を同様の意味で追加しても構わない。 In addition, the intra_luma_filter_flag [i] described above may be added in the same meaning for the other syntax structures shown in FIGS. 22B, 22D, and 22E.
 (第2の変形例) 
 <イントラ予測部 第二の変形例> 
 イントラ予測部4408(109)に係る第2の変形例として、JCTVC-B205_draft002, 9.6節"Combined Intra Prediction", JCT-VC 2nd Meeting Geneva, July, 2010に示される複合イントラ予測と組み合わせて使用しても構わない。この文献における復号イントラ予測は、前述の単方向イントラ予測の結果と、予測画素に対し左、上、左上に隣接する画素の平均値とを加重平均することにより、予測値を得る。動画像復号化装置4400や画像符号化装置100内において復号画像信号4417を計算した場合には、前記左、上、左上に隣接する画素として、復号画素を用いることが可能である。
(Second modification)
<Intra prediction unit second modification>
As a second modification related to the intra prediction unit 4408 (109), it is used in combination with the composite intra prediction shown in JCTVC-B205_draft002, section 9.6 “Combined Intra Prediction”, JCT-VC 2nd Meeting Geneva, July, 2010. It doesn't matter. In the decoded intra prediction in this document, a prediction value is obtained by performing weighted averaging of the result of the above-described unidirectional intra prediction and the average value of pixels adjacent to the left, top, and top left with respect to the prediction pixel. When the decoded image signal 4417 is calculated in the moving image decoding device 4400 or the image encoding device 100, it is possible to use decoded pixels as pixels adjacent to the left, upper, and upper left.
 図30に、複合イントラ予測と組み合わせた場合のイントラ予測部4408(109)のブロック図を示す。図6で示されるイントラ予測部4408(109)に複合イントラ予測画像生成部2901、選択スイッチ2902、及び復号画素バッファ3001が追加されている点が異なる。 FIG. 30 shows a block diagram of the intra prediction unit 4408 (109) when combined with composite intra prediction. The difference is that a composite intra predicted image generation unit 2901, a selection switch 2902, and a decoded pixel buffer 3001 are added to the intra prediction unit 4408 (109) shown in FIG.
 双方向イントラ予測と複合イントラ予測を組み合わせた場合、まず選択スイッチ604において、復号化制御部4412で制御されている予測モード情報に従って、単方向イントラ予測画像生成部601若しくは双方向イントラ予測画像生成部602の出力端を切り替える。以後、出力された当該予測画像信号4420(126)を方向予測画像信号4420(126)と呼ぶ。 When the bidirectional intra prediction and the composite intra prediction are combined, first, in the selection switch 604, the unidirectional intra prediction image generation unit 601 or the bidirectional intra prediction image generation unit according to the prediction mode information controlled by the decoding control unit 4412. The output terminal of 602 is switched. Hereinafter, the output predicted image signal 4420 (126) is referred to as a direction predicted image signal 4420 (126).
 その後、前記方向予測画像信号を複合イントラ予測画像生成部2901に入力し、複合イントラ予測における予測画像信号4420(126)を生成する。複合イントラ予測画像生成部2901における説明は後述する。その後、選択スイッチ2902において、復号化制御部4412で制御されている予測モード情報内の複合イントラ予測適用フラグに従って、複合イントラ予測における予測画像信号4420(126)と方向予測画像信号のいずれを用いるかを切り替えて、イントラ予測部4408(109)における最終的な予測画像信号4420(126)が出力される。複合イントラ予測適用フラグが1の場合には、複合イントラ予測画像生成部2901から出力される予測画像信号4420(126)が最終的な予測画像信号4420(126)となる。一方、複合イントラ予測適用フラグが0の場合には、方向予測画像信号4420(126)が最終的に出力される予測画像信号126となる。 Then, the direction prediction image signal is input to the composite intra prediction image generation unit 2901, and a prediction image signal 4420 (126) in the composite intra prediction is generated. The description of the composite intra predicted image generation unit 2901 will be described later. Thereafter, in the selection switch 2902, which one of the prediction image signal 4420 (126) and the direction prediction image signal in the composite intra prediction is used according to the composite intra prediction application flag in the prediction mode information controlled by the decoding control unit 4412. And the final predicted image signal 4420 (126) in the intra prediction unit 4408 (109) is output. When the composite intra prediction application flag is 1, the predicted image signal 4420 (126) output from the composite intra predicted image generation unit 2901 is the final predicted image signal 4420 (126). On the other hand, when the composite intra prediction application flag is 0, the direction prediction image signal 4420 (126) is the prediction image signal 126 that is finally output.
 次に、複合イントラ予測画像生成部2901について図31を用いて説明する。複合イントラ予測画像生成部2901は画素レベル予測信号生成部3101及び複合イントラ予測計算部3102を含む。画素レベル予測信号生成部3101では、予測対象画素Xを隣接する画素から予測し、画素レベル予測信号3103を出力する。前述のように当該隣接画素は復号画像信号4417を示す。具体的には、数(21)を用いて予測対象画素の画素レベル予測信号3103(X)を計算する。なお、A,B,Cに係る係数は、他の値であっても構わない。 Next, the composite intra prediction image generation unit 2901 will be described with reference to FIG. The composite intra prediction image generation unit 2901 includes a pixel level prediction signal generation unit 3101 and a composite intra prediction calculation unit 3102. The pixel level prediction signal generation unit 3101 predicts the prediction target pixel X from adjacent pixels and outputs a pixel level prediction signal 3103. As described above, the adjacent pixel indicates the decoded image signal 4417. Specifically, the pixel level prediction signal 3103 (X) of the prediction target pixel is calculated using the number (21). The coefficients related to A, B, and C may be other values.
 複合イントラ予測計算部3102では、方向予測画像信号4420(126)(X’)と画素レベル予測信号3103(X)の加重平均を行い、最終的な予測画像信号4420(126)(P)を出力する。具体的には、数式(22)を用いる。 The composite intra prediction calculation unit 3102 performs a weighted average of the direction prediction image signal 4420 (126) (X ′) and the pixel level prediction signal 3103 (X), and outputs a final prediction image signal 4420 (126) (P). To do. Specifically, Formula (22) is used.
 なお、Wは方向予測画像信号4420(126)(X’)と画素レベル予測信号3103(X)との加重平均の重み係数(W=0~32の間の整数値)である。以上が、複合イントラ予測と組み合わせた場合の実施形態である。 Note that W is a weighted average weight coefficient (an integer value between W = 0 and 32) of the direction prediction image signal 4420 (126) (X ′) and the pixel level prediction signal 3103 (X). The above is an embodiment when combined with composite intra prediction.
 (第3の変形例) 
 <イントラ予測部 第三の変形例> 
 また、上記重み係数Wはプレディクションユニット内の予測画素の位置に応じて、切り替えても構わない。一般に、単方向イントラ予測及び双方向イントラ予測を用いて生成された予測画像信号は、空間的に隣接している既に符号化済みの上或いは左に位置する参照画素から予測値が生成されるため、参照画素からの距離が離れるほど予測誤差の絶対値が増加する傾向にある。従って、方向予測画像信号126と画素レベル予測信号3103の重み係数を、参照画素に近い場合は方向予測画像信号126の重み係数を大きくし、離れた場合は、小さくすることで予測精度を向上させることが可能となる。
(Third Modification)
<Third modification of intra prediction unit>
The weighting factor W may be switched according to the position of the prediction pixel in the prediction unit. In general, a prediction image signal generated using unidirectional intra prediction and bidirectional intra prediction generates a prediction value from spatially adjacent reference pixels positioned on the left or above already encoded. The absolute value of the prediction error tends to increase as the distance from the reference pixel increases. Therefore, the weighting coefficient of the direction prediction image signal 126 and the pixel level prediction signal 3103 is increased when the weight coefficient of the direction prediction image signal 126 is close to the reference pixel, and is decreased when the distance is far away, thereby improving the prediction accuracy. It becomes possible.
 一方、当該複合イントラ予測では、符号化時に入力画像信号を利用した予測誤差信号の生成を行う。この際、画素レベル予測信号3103は入力画像信号となるため、参照画素位置と予測画素位置との空間距離が大きくなっても、方向予測画像信号126と比較して、画素レベル予測信号3103の予測精度が高い。しかし、単純に方向予測画像信号126と画素レベル予測信号3103の重み係数を、参照画素に近い場合は方向予測画像信号126の重み係数を大きくし、離れた場合は、小さくすると、離れた場合の予測誤差が小さくなるが、符号化時の予測値と局部復号化時の予測値に乖離が生まれ、予測精度が低下する問題が生じる。従って、特に量子化パラメータの値が大きい場合に、参照画素位置と予測画素位置との空間距離が大きくなるにつれ、Wの値を小さく設定することにより、このようなオープンループの場合に発生する乖離現象による符号化効率低下を抑えることが出来る。 On the other hand, in the complex intra prediction, a prediction error signal is generated using an input image signal at the time of encoding. At this time, since the pixel level prediction signal 3103 becomes an input image signal, even if the spatial distance between the reference pixel position and the prediction pixel position is increased, the prediction of the pixel level prediction signal 3103 is compared with the direction prediction image signal 126. High accuracy. However, the weighting coefficient of the direction prediction image signal 126 and the pixel level prediction signal 3103 is simply increased when the weight coefficient of the direction prediction image signal 126 is close to the reference pixel, and is decreased when the distance is small. Although the prediction error is reduced, there is a problem that the prediction accuracy at the time of encoding and the prediction value at the time of local decoding are different and the prediction accuracy is lowered. Therefore, especially when the value of the quantization parameter is large, as the spatial distance between the reference pixel position and the predicted pixel position becomes large, the difference generated in the case of such an open loop is set by setting the value of W small. A decrease in coding efficiency due to the phenomenon can be suppressed.
 <シンタクス構成 6> 
 図32A,図32Bには、複合イントラ予測を行う際のプレディクションユニットシンタクス構造を示している。図32Aは、図22Aと比較して複合イントラ予測の有無を切り替えるシンタクスcombined_intra_pred_flagを追加している点で異なる。これは、前述の複合イントラ予測適用フラグと等しい。また、図32Bは、図22Cに複合イントラ予測の有無を切り替えるシンタクスcombined_intra_pred_flagを追加している。combined_intra_pred_flagが1の場合、図30に示される選択スイッチ2902は複合イントラ予測画像生成部2901の出力端に接続される。combined_intra_pred_flagが0の場合、図29に示される選択スイッチ2902は選択スイッチ604が接続している単方向イントラ予測画像生成部601若しくは双方向イントラ予測画像生成部602のいずれかの出力端に接続される。
<Syntax structure 6>
32A and 32B show a prediction unit syntax structure when performing composite intra prediction. FIG. 32A is different from FIG. 22A in that a syntax combined_intra_pred_flag for switching presence / absence of composite intra prediction is added. This is equivalent to the above-described composite intra prediction application flag. In addition, FIG. 32B adds a syntax combined_intra_pred_flag for switching presence / absence of composite intra prediction to FIG. 22C. When combined_intra_pred_flag is 1, the selection switch 2902 shown in FIG. 30 is connected to the output terminal of the composite intra prediction image generation unit 2901. When combined_intra_pred_flag is 0, the selection switch 2902 shown in FIG. 29 is connected to the output terminal of either the unidirectional intra prediction image generation unit 601 or the bidirectional intra prediction image generation unit 602 to which the selection switch 604 is connected. .
 また、図22B、図22D、図22Eに示される他のシンタクス構造について、上記で説明したintra_luma_filter_flag[i]を同様の意味で追加しても構わない。 In addition, the intra_luma_filter_flag [i] described above may be added in the same meaning for the other syntax structures shown in FIGS. 22B, 22D, and 22E.
 さらに、イントラ予測部の第2の実施形態と組み合わせても構わない。 Furthermore, it may be combined with the second embodiment of the intra prediction unit.
 以上が、イントラ予測部4408の別の実施形態に関する説明である。 This completes the description of another embodiment of the intra prediction unit 4408.
 以上の第4の実施形態によれば、第1の実施形態に係る動画像符号化装置と同一または類似のイントラ予測部を含むので、第1の実施形態に係る動画像符号化装置と同一または類似の効果を得ることができる。 According to the fourth embodiment described above, since the same or similar intra prediction unit as that of the video encoding device according to the first embodiment is included, the same or the same as the video encoding device according to the first embodiment or Similar effects can be obtained.
 (第5の実施形態) 
 <動画像復号化装置-第五の実施形態> 
 第5の実施形態に係る動画像復号化装置は、前述の第4の実施形態に係る動画像復号化装置と逆直交変換の詳細において異なる。以降の説明では、本実施形態において第4の実施形態と同一部分には同一符号を付して示し、異なる部分を中心に説明する。本実施形態に係る動画像復号化装置に対応する動画像符号化装置は、第2の実施形態において説明した通りである。
(Fifth embodiment)
<Video Decoding Device—Fifth Embodiment>
The video decoding device according to the fifth embodiment differs from the video decoding device according to the above-described fourth embodiment in the details of inverse orthogonal transform. In the following description, in this embodiment, the same parts as those in the fourth embodiment are denoted by the same reference numerals, and different parts will be mainly described. The moving picture coding apparatus corresponding to the moving picture decoding apparatus according to the present embodiment is as described in the second embodiment.
 図45は、第5の実施形態に係る動画像復号化装置を示すブロック図である。第5の実施形態に係る動画像復号化装置からの変更点は、変換選択部4502、係数順復元部4501が追加されている点である。また、逆直交変換部4404の内部構造も異なる。 FIG. 45 is a block diagram showing a moving picture decoding apparatus according to the fifth embodiment. A change from the moving picture decoding apparatus according to the fifth embodiment is that a conversion selection unit 4502 and a coefficient order restoration unit 4501 are added. Also, the internal structure of the inverse orthogonal transform unit 4404 is different.
 <逆直交変換部4404> 
 先ず、逆直交変換部4404を、図35を用いて説明する。なお、逆直交変換部4404は第2の実施形態に係る逆直交変換部105と同一の構成である。従って、本実施形態では、図35内の変換選択情報3303を変換選択情報4504と、復元変換係数120を復元変換係数4415、復元予測誤差信号121を復元予測誤差信号4416と、夫々置き換えて説明する。
<Inverse orthogonal transform unit 4404>
First, the inverse orthogonal transform unit 4404 will be described with reference to FIG. Note that the inverse orthogonal transform unit 4404 has the same configuration as the inverse orthogonal transform unit 105 according to the second embodiment. Therefore, in the present embodiment, the conversion selection information 3303 in FIG. 35 is replaced with the conversion selection information 4504, the restored transform coefficient 120 is replaced with the restored transform coefficient 4415, and the restored prediction error signal 121 is replaced with the restored prediction error signal 4416. .
 図35の逆直交変換部105(4404)は、第一逆直交変換部3501、第二逆直交変換部3502、第N逆直交変換部3503、変換選択スイッチ3504を有する。先ず、変換選択スイッチ3504について説明する。変換選択スイッチ3504は、逆量子化部4403の出力端を、入力された変換選択情報4504に従って選択する機能を有する。変換選択情報4504は、復号化制御部4412で制御されている情報の1つであり、予測情報4421(125)に従って変換選択部4502で設定される。 35 includes a first inverse orthogonal transform unit 3501, a second inverse orthogonal transform unit 3502, an Nth inverse orthogonal transform unit 3503, and a transform selection switch 3504. First, the conversion selection switch 3504 will be described. The conversion selection switch 3504 has a function of selecting the output terminal of the inverse quantization unit 4403 according to the input conversion selection information 4504. The conversion selection information 4504 is one of information controlled by the decoding control unit 4412, and is set by the conversion selection unit 4502 in accordance with the prediction information 4421 (125).
 変換選択情報4504が第一直交変換である場合は、スイッチの出力端を第一逆直交変換部3501に接続する。一方、変換選択情報4504が第二直交変換である場合は、出力端を第二逆直交変換部3502に接続する。同様に、変換選択情報4504が第N直交変換である場合は、出力端を第N逆直交変換部3503に接続する。 When the conversion selection information 4504 is the first orthogonal transform, the output terminal of the switch is connected to the first inverse orthogonal transform unit 3501. On the other hand, when the transformation selection information 4504 is the second orthogonal transformation, the output end is connected to the second inverse orthogonal transformation unit 3502. Similarly, when the transform selection information 4504 is the Nth orthogonal transform, the output terminal is connected to the Nth inverse orthogonal transform unit 3503.
 <変換選択部4502> 
 次に、図45に示される変換選択部4502について説明する。変換選択部4502には、復号化制御部4412で制御され、エントロピー復号化部4402で復号された予測情報4421(125)が入力される。変換選択部4502は、この予測情報4421(125)に基づいて、どの予測モードに対して、どの逆直交変換を使うかどうかを示すMappedTransformIdx情報を設定する機能を有する。図36に、イントラ予測における変換選択情報4504(MappedTransformIdx)を示す。ここではN=9の例を示している。なお、IntraPredModeLX=2に対応するDC予測時には、第一逆直交変換部3501が選択される。このように予測角度の近い基準予測モードへマッピングすることにより、全予測モードに対して直交変換器及び逆直交変換器を用意する場合と比較して、ハードウェア実現時の直交変換及び逆直交変換の回路規模を削減することが可能である。なお、双方向イントラ予測が選択されている場合、図14に従って、それぞれ2つのIntraPredModeL0とIntraPredModeL1を導出した後、IntraPredModeL1に対応する予測モードを利用して図36からMappedTransformIdxを導出する。本発明の本実施の形態では、N=9の例を示したが、Nの値は符号化性能とハードウェア実現時の回路規模のバランスを取って、最適な組み合わせを選べばよい。
<Conversion selection unit 4502>
Next, the conversion selection unit 4502 shown in FIG. 45 will be described. Prediction information 4421 (125), which is controlled by the decoding control unit 4412 and decoded by the entropy decoding unit 4402, is input to the transformation selection unit 4502. Based on the prediction information 4421 (125), the transform selection unit 4502 has a function of setting MapdTransformIdx information indicating which inverse orthogonal transform is to be used for which prediction mode. FIG. 36 shows conversion selection information 4504 (MappedTransformIdx) in intra prediction. Here, an example of N = 9 is shown. Note that the first inverse orthogonal transform unit 3501 is selected at the time of DC prediction corresponding to IntraPredModeLX = 2. By mapping to the reference prediction mode with a close prediction angle in this way, compared to the case of preparing an orthogonal transformer and an inverse orthogonal transformer for all prediction modes, orthogonal transformation and inverse orthogonal transformation at the time of hardware implementation It is possible to reduce the circuit scale. When bi-directional intra prediction is selected, two IntraPredModeL0 and IntraPredModeL1 are derived according to FIG. 14, respectively, and then the mapped transformIdx is derived from FIG. 36 using a prediction mode corresponding to IntraPredModeL1. In the present embodiment of the present invention, an example of N = 9 has been shown, but the value of N may be selected in an optimal combination by balancing the coding performance and the circuit scale at the time of hardware implementation.
 <係数順復元部4501> 
 次に、係数順復元部4501を説明する。図46に係数順復元部4501のブロック図を示す。なお、係数順復元部4501は第2の実施形態に係る係数順制御部3302と、逆のスキャン順変換を行う機能を有する。
<Coefficient order restoration unit 4501>
Next, the coefficient order restoration unit 4501 will be described. FIG. 46 shows a block diagram of the coefficient order restoration unit 4501. The coefficient order restoration unit 4501 has a function of performing reverse scan order conversion with the coefficient order control unit 3302 according to the second embodiment.
 係数順復元部4501は、係数順選択スイッチ4604と、第一係数順逆変換部4601、第二係数順逆変換部4602、第N係数順逆変換部4603とを有する。係数順選択スイッチ4604は、例えば、図36に示されるMappedTransformIdxに従って、スイッチの出力端と係数順逆変換部4601ないし4603を切り替える機能を有する。N種類の係数順逆変換部4601~4603は、エントロピー復号化部4402で復号された量子化変換係数列4503に対して1次元データを2次元データへと逆変換する機能を有する。例えば、H.264では、ジグザグスキャンを用いて2次元データを1次元データへと変換している。ここでは、例えば、ジグザグスキャンからラスタースキャンへの変換を行うことを意味する。 The coefficient order restoration unit 4501 includes a coefficient order selection switch 4604, a first coefficient forward / reverse transform unit 4601, a second coefficient forward / reverse transform unit 4602, and an Nth coefficient forward / reverse transform unit 4603. The coefficient order selection switch 4604 has a function of switching the output terminal of the switch and the coefficient order inverse conversion units 4601 to 4603 in accordance with, for example, the mapped transform IDx shown in FIG. The N types of coefficient forward / inverse transform units 4601 to 4603 have a function of inversely transforming one-dimensional data into two-dimensional data with respect to the quantized transform coefficient sequence 4503 decoded by the entropy decoding unit 4402. For example, H.M. In H.264, two-dimensional data is converted into one-dimensional data using a zigzag scan. Here, for example, it means that conversion from a zigzag scan to a raster scan is performed.
 イントラ予測の予測方向を考慮した直交変換を用いる場合、直交変換を施した変換係数に量子化処理を施した量子化変換係数は、ブロック内の非ゼロとなる変換係数の発生傾向が偏る性質を持つ。この非ゼロ変換係数の発生傾向は、イントラ予測の予測方向毎に異なる性質がある。しかし、異なる映像を符号化した際に同じ予測方向における非ゼロ変換係数の発生傾向は似る性質を持つ。そこで、2次元データを1次元データへ変換(2D-1D変換)する際、非ゼロ変換係数の発生確率が高い位置の変換係数から優先的にエントロピー符号化することで、変換係数の符号化する情報を削減することが可能である。復号側では、逆に1次元データを2次元データに復元する必要がある。ここでは、ラスタースキャンを一次元の基準スキャンとして復元を行う。 When using orthogonal transform in consideration of the prediction direction of intra prediction, the quantized transform coefficient obtained by performing quantization processing on the transform coefficient that has been subjected to orthogonal transform has the property that the tendency of generating non-zero transform coefficients in the block is biased. Have. The tendency of occurrence of this non-zero transform coefficient has different properties for each prediction direction of intra prediction. However, when different videos are encoded, the generation tendency of non-zero transform coefficients in the same prediction direction has a similar property. Therefore, when transforming two-dimensional data into one-dimensional data (2D-1D conversion), entropy coding is performed preferentially from transform coefficients at positions where the occurrence probability of non-zero transform coefficients is high, thereby encoding transform coefficients. It is possible to reduce information. Conversely, on the decoding side, it is necessary to restore the one-dimensional data to the two-dimensional data. Here, the raster scan is restored as a one-dimensional reference scan.
 さらに別の例として、係数順復元部4501は、1D-2D変換におけるスキャン順を動的に更新してもよい。このような動作を行う係数順復元部4501の構成は、図47に例示される。この係数順復元部4501は、図46の構成に加え、発生頻度カウント部4701と、更新部4702とを含む。係数順逆変換部4601、・・・,4603は、その1D-2Dスキャン順が更新部4702によって更新される点以外は同一である。 As yet another example, the coefficient order restoration unit 4501 may dynamically update the scan order in the 1D-2D conversion. The configuration of the coefficient order restoration unit 4501 that performs such an operation is illustrated in FIG. The coefficient order restoration unit 4501 includes an occurrence frequency counting unit 4701 and an updating unit 4702 in addition to the configuration of FIG. .., 4603 are the same except that the 1D-2D scan order is updated by the updating unit 4702.
 発生頻度カウント部4701は、予測モード毎に、量子化変換係数列4503の各要素における非零係数の発生回数のヒストグラム4704を作成する。発生頻度カウント部4701は、作成したヒストグラム4704を更新部4702に入力する。 The occurrence frequency counting unit 4701 creates a histogram 4704 of the number of occurrences of non-zero coefficients in each element of the quantized transform coefficient sequence 4503 for each prediction mode. The occurrence frequency counting unit 4701 inputs the created histogram 4704 to the update unit 4702.
 更新部4702は、予め定められたタイミングで、ヒストグラム4704に基づいて係数順の更新を行う。上記タイミングは、例えば、コーディングツリーユニットの符号化処理が終了したタイミング、コーディングツリーユニット内の1ライン分の符号化処理が終了したタイミングなどである。 The update unit 4702 updates the coefficient order based on the histogram 4704 at a predetermined timing. The timing is, for example, the timing when the coding process of the coding tree unit is finished, the timing when the coding process for one line in the coding tree unit is finished, or the like.
 具体的には、更新部4702は、ヒストグラム4704を参照して、非零係数の発生回数が閾値以上にカウントされた要素を持つ予測モードに関して係数順の更新を行う。例えば、更新部4702は、非零係数の発生が16回以上カウントされた要素を持つ予測モードに関して更新を行う。このような発生回数に閾値を設けることによって、係数順の更新が大域的に実施されるので、局所的な最適解に収束しにくくなる。 Specifically, the update unit 4702 refers to the histogram 4704 and updates the coefficient order with respect to a prediction mode having an element in which the number of occurrences of non-zero coefficients is counted more than a threshold. For example, the update unit 4702 updates the prediction mode having an element in which the occurrence of a non-zero coefficient is counted 16 times or more. By providing a threshold value for the number of occurrences, the coefficient order is updated globally, so that it is difficult to converge to a local optimum solution.
 更新部4702は、更新対象となる予測モードに関して、非零係数の発生頻度の降順に要素をソーティングする。ソーティングは、例えばバブルソート、クイックソートなどの既存のアルゴリズムによって実現できる。そして、更新部4702は、ソーティングされた要素の順序を示す更新係数順4703を、更新対象となる予測モードに対応する係数順逆変換部4601ないし4603に入力する。 The update unit 4702 sorts the elements in descending order of the occurrence frequency of the non-zero coefficient regarding the prediction mode to be updated. Sorting can be realized by existing algorithms such as bubble sort and quick sort. Then, the update unit 4702 inputs the update coefficient order 4703 indicating the order of the sorted elements to the coefficient order inverse transform units 4601 to 4603 corresponding to the prediction mode to be updated.
 更新係数順4703が入力されると、各逆変換部は更新後のスキャン順に従って1D-2D変換を行う。なお、スキャン順を動的に更新する場合には、各1D-2D変換部の初期スキャン順を予め定めておく必要がある。図33に示される動画像符号化装置の係数順制御部3302と同じ初期スキャン順に定める。このように、動的にスキャン順を更新することにより、予測画像の性質、量子化情報(量子化パラメータ)などの影響に応じて、量子化変換係数における非零係数の発生傾向が変化する場合にも、安定的に高い符号化効率を期待できる。具体的には、エントロピー符号化部113におけるランレングス符号化の発生符号量を抑制できる。 When the update coefficient order 4703 is input, each inverse conversion unit performs 1D-2D conversion in accordance with the updated scan order. When the scan order is dynamically updated, the initial scan order of each 1D-2D conversion unit needs to be determined in advance. The initial scanning order is the same as that of the coefficient order control unit 3302 of the moving picture coding apparatus shown in FIG. In this way, when the scan order is dynamically updated, the tendency of occurrence of non-zero coefficients in the quantized transform coefficients changes according to the effect of the predicted image properties, quantization information (quantization parameters), etc. In addition, stable and high encoding efficiency can be expected. Specifically, the generated code amount of run-length encoding in the entropy encoding unit 113 can be suppressed.
 なお、本実施の形態におけるシンタクス構成は、第4の実施形態と同一である。 The syntax configuration in the present embodiment is the same as that in the fourth embodiment.
 本実施形態の変形例として、変換選択部4502が予測情報4421とは別にMappedTransformIdxを選択することも可能である。この場合、9種類のどの直交変換或いは逆直交変換を用いたかを示す情報が復号化制御部4412に設定され、逆直交変換部4404で利用される。図39に本実施の形態におけるシンタクスの例を示す。シンタクス中に示されるdirectional_transform_idxは、N個に対応する直交変換のいずれを選択したかの情報が示されている。 As a modification of the present embodiment, the conversion selection unit 4502 can select the mapped transform idx separately from the prediction information 4421. In this case, information indicating which nine types of orthogonal transforms or inverse orthogonal transforms are used is set in the decoding control unit 4412 and used by the inverse orthogonal transform unit 4404. FIG. 39 shows an example of syntax in the present embodiment. Directional_transform_idx indicated in the syntax indicates information indicating which of N orthogonal transforms has been selected.
 以上の第5の実施形態によれば、第2の実施形態に係る動画像符号化装置と同一または類似の逆直交変換部を含むので、第2の実施形態に係る動画像符号化装置と同一または類似の効果を得ることができる。 According to the fifth embodiment described above, the same or similar inverse orthogonal transform unit as that of the video encoding device according to the second embodiment is included, and thus the same as the video encoding device according to the second embodiment. Or a similar effect can be obtained.
 (第6の実施形態) 
 <動画像復号化装置-第六の実施形態> 
 第6の実施形態に係る動画像復号化装置は、前述の第4の実施形態に係る動画像復号化装置と逆直交変換の詳細において異なる。以降の説明では、本実施形態において第4の実施形態と同一部分には同一符号を付して示し、異なる部分を中心に説明する。本実施形態に係る動画像復号化装置に対応する動画像符号化装置は、第3の実施形態において説明した通りである。
(Sixth embodiment)
<Video Decoding Device—Sixth Embodiment>
The video decoding device according to the sixth embodiment differs from the video decoding device according to the above-described fourth embodiment in the details of inverse orthogonal transform. In the following description, in this embodiment, the same parts as those in the fourth embodiment are denoted by the same reference numerals, and different parts will be mainly described. The moving picture encoding apparatus corresponding to the moving picture decoding apparatus according to the present embodiment is as described in the third embodiment.
 逆直交変換部105に係る実施形態として、JCTVC-B205_draft002, 5.3.5.2節" Rotational transformation process ", JCT-VC 2nd Meeting Geneva, July, 2010に示される回転変換と組み合わせても構わない。 As an embodiment related to the inverse orthogonal transform unit 105, JCTVC-B205_draft002, 5.3.5.2 section “Rotational transformation process”, JCT-VC 2nd Meeting Geneva, July, 2010, may be combined.
 <逆直交変換部4404(105)> 
 図41は本実施の形態に関わる逆直交変換部4404(105)のブロック図である。逆直交変換部4404(105)は、第一逆回転変換部4101、第二逆回転変換部4102、第N逆回転変換部4103、逆離散コサイン変換部4104の新しい処理部を持ち、既存の変換選択スイッチ3504を有する。逆量子化処理後に入力された復元変換係数4415(120)が変換選択スイッチ3504に入力される。ここで、変換選択スイッチ3504は、変換選択情報4504(3303)に従って、スイッチの出力端を第一逆回転変換部4101、第二逆回転変換部4102、第N逆回転変換部4103のいずれかに接続する。その後、図40に示される直交変換部102で利用された回転変換と同じ、いずれかの逆回転変換部4101ないし4103で逆回転変換処理が施され、逆離散コサイン変換部4104へと出力する。逆離散コサイン変換部4104は、入力された信号に対して例えばIDCTを施し、復元予測誤差信号4416(121)を復元する。ここでは例としてIDCTを用いる例を示したが、アダマール変換や離散サイン変換などの直交変換を使ってもよいし、非直交変換を用いてもよい。いずれにしても図40に示される直交変換部102と連動して対応する逆変換が執り行われる。
<Inverse orthogonal transform unit 4404 (105)>
FIG. 41 is a block diagram of inverse orthogonal transform section 4404 (105) according to the present embodiment. The inverse orthogonal transform unit 4404 (105) has new processing units such as a first inverse rotation transform unit 4101, a second inverse rotation transform unit 4102, an Nth inverse rotation transform unit 4103, and an inverse discrete cosine transform unit 4104. A selection switch 3504 is included. The restored transform coefficient 4415 (120) input after the inverse quantization processing is input to the transform selection switch 3504. Here, the conversion selection switch 3504 sets the output end of the switch to one of the first reverse rotation conversion unit 4101, the second reverse rotation conversion unit 4102, and the Nth reverse rotation conversion unit 4103 according to the conversion selection information 4504 (3303). Connecting. Thereafter, the same inverse rotation transform unit 4101 to 4103 as that used in the orthogonal transform unit 102 shown in FIG. 40 is subjected to the inverse rotation transform process, and the result is output to the inverse discrete cosine transform unit 4104. The inverse discrete cosine transform unit 4104 performs, for example, IDCT on the input signal to restore the restored prediction error signal 4416 (121). Although an example using IDCT is shown here as an example, orthogonal transform such as Hadamard transform or discrete sine transform may be used, or non-orthogonal transform may be used. In any case, corresponding inverse transformation is performed in conjunction with the orthogonal transformation unit 102 shown in FIG.
 本実施の形態におけるシンタクスが図42で示されている。シンタクス中に示されるrotational_transform_idxは利用する回転行列の番号を意味している。 FIG. 42 shows the syntax in the present embodiment. The rotation_transform_idx shown in the syntax means the number of the rotation matrix to be used.
 以上の第6の実施形態によれば、第3の実施形態に係る画像符号化装置と同一または類似の逆直交変換部を含むので、第3の実施形態に係る画像符号化装置と同一または類似の効果を得ることができる。 According to the sixth embodiment described above, the same or similar inverse orthogonal transform unit as that of the image encoding device according to the third embodiment is included, and therefore the same or similar as that of the image encoding device according to the third embodiment. The effect of can be obtained.
 以下、各実施形態の変形例を列挙して紹介する。 
 第1乃至第6の実施形態において、フレームを16×16画素サイズなどの矩形ブロックに分割し、画面左上のブロックから右下に向かって順に符号化/復号化を行う例について説明している(図2Aを参照)。しかしながら、符号化順序及び復号化順序はこの例に限定されない。例えば、右下から左上に向かって順に符号化及び復号化が行われてもよいし、画面中央から画面端に向かって渦巻を描くように符号化及び復号化が行われてもよい。さらに、右上から左下に向かって順に符号化及び復号化が行われてもよいし、画面端から画面中央に向かって渦巻きを描くように符号化及び復号化が行われてもよい。
Hereinafter, modifications of each embodiment will be listed and introduced.
In the first to sixth embodiments, an example is described in which a frame is divided into rectangular blocks of 16 × 16 pixel size and the like, and encoding / decoding is performed in order from the upper left block to the lower right side of the screen ( (See FIG. 2A). However, the encoding order and the decoding order are not limited to this example. For example, encoding and decoding may be performed sequentially from the lower right to the upper left, or encoding and decoding may be performed so as to draw a spiral from the center of the screen toward the screen end. Furthermore, encoding and decoding may be performed in order from the upper right to the lower left, or encoding and decoding may be performed so as to draw a spiral from the screen edge toward the center of the screen.
 第1乃至第6の実施形態において、4×4画素ブロック、8×8画素ブロック、16×16画素ブロックなどの予測対象ブロックサイズを例示して説明を行ったが、予測対象ブロックは均一なブロック形状でなくてもよい。例えば、予測対象ブロック(プレディクションユニット)サイズは、16×8画素ブロック、8×16画素ブロック、8×4画素ブロック、4×8画素ブロックなどであってもよい。また、1つのコーディングツリーユニット内で全てのブロックサイズを統一させる必要はなく、複数の異なるブロックサイズを混在させてもよい。1つのコーディングツリーユニット内で複数の異なるブロックサイズを混在させる場合、分割数の増加に伴って分割情報を符号化または復号化するための符号量も増加する。そこで、分割情報の符号量と局部復号画像または復号画像の品質との間のバランスを考慮して、ブロックサイズを選択することが望ましい。 In the first to sixth embodiments, the description has been given by exemplifying the prediction target block size such as the 4 × 4 pixel block, the 8 × 8 pixel block, and the 16 × 16 pixel block. However, the prediction target block is a uniform block. It does not have to be a shape. For example, the prediction target block (prediction unit) size may be a 16 × 8 pixel block, an 8 × 16 pixel block, an 8 × 4 pixel block, a 4 × 8 pixel block, or the like. Also, it is not necessary to unify all the block sizes within one coding tree unit, and a plurality of different block sizes may be mixed. When a plurality of different block sizes are mixed in one coding tree unit, the amount of codes for encoding or decoding the division information increases as the number of divisions increases. Therefore, it is desirable to select the block size in consideration of the balance between the code amount of the division information and the quality of the locally decoded image or the decoded image.
 第1乃至第6の実施形態において、簡単化のために、輝度信号と色差信号とを区別せず、色信号成分に関して包括的な説明を記述した。しかしながら、予測処理が輝度信号と色差信号との間で異なる場合には、同一または異なる予測方法が用いられてよい。輝度信号と色差信号との間で異なる予測方法が用いられるならば、色差信号に対して選択した予測方法を輝度信号と同様の方法で符号化または復号化できる。 In the first to sixth embodiments, for the sake of simplicity, a comprehensive description of the color signal component has been described without distinguishing between the luminance signal and the color difference signal. However, when the prediction process is different between the luminance signal and the color difference signal, the same or different prediction methods may be used. If different prediction methods are used between the luminance signal and the chrominance signal, the prediction method selected for the chrominance signal can be encoded or decoded in the same manner as the luminance signal.
 第1乃至第6の実施形態において、簡単化のために、輝度信号と色差信号とを区別せず、色信号成分に関して包括的な説明を記述した。しかしながら、直交変換処理が輝度信号と色差信号との間で異なる場合には、同一または異なる直交変換方法が用いられてよい。輝度信号と色差信号との間で異なる直交変換方法が用いられるならば、色差信号に対して選択した直交変換方法を輝度信号と同様の方法で符号化または復号化できる。 In the first to sixth embodiments, for the sake of simplicity, a comprehensive description of the color signal component has been described without distinguishing between the luminance signal and the color difference signal. However, when the orthogonal transformation process is different between the luminance signal and the color difference signal, the same or different orthogonal transformation methods may be used. If different orthogonal transformation methods are used between the luminance signal and the color difference signal, the orthogonal transformation method selected for the color difference signal can be encoded or decoded in the same manner as the luminance signal.
 第1乃至第6の実施形態において、シンタクス構成に示す表の行間には、本発明で規定していないシンタクス要素が挿入されることも可能であるし、それ以外の条件分岐に関する記述が含まれていても構わない。或いは、シンタクステーブルを複数のテーブルに分割、統合することも可能である。また、必ずしも同一の用語を用いる必要は無く、利用する形態によって任意に変更しても構わない。 In the first to sixth embodiments, syntax elements that are not defined in the present invention can be inserted between the rows of the table shown in the syntax configuration, and other conditional branch descriptions are included. It does not matter. Alternatively, the syntax table can be divided and integrated into a plurality of tables. Moreover, it is not always necessary to use the same term, and it may be arbitrarily changed depending on the form to be used.
 以上説明したように、各実施形態は、ハードウェア実装及びソフトウェア実装における困難性を緩和しつつ、高効率な直交変換及び逆直交変換を実現することができる。故に、各実施形態によれば、符号化効率が向上し、ひいては主観画質も向上する。 As described above, each embodiment can realize highly efficient orthogonal transformation and inverse orthogonal transformation while alleviating the difficulty in hardware implementation and software implementation. Therefore, according to each embodiment, the encoding efficiency is improved, and the subjective image quality is also improved.
 本発明のいくつかの実施形態を説明したが、これらの実施形態は、例として提示したものであり、発明の範囲を限定することは意図していない。これら新規な実施形態は、その他の様々な形態で実施されることが可能であり、発明の要旨を逸脱しない範囲で、種々の省略、置き換え、変更を行うことができる。これら実施形態やその変形は、発明の範囲や要旨に含まれるとともに、特許請求の範囲に記載された発明とその均等の範囲に含まれる。 Although several embodiments of the present invention have been described, these embodiments are presented as examples and are not intended to limit the scope of the invention. These novel embodiments can be implemented in various other forms, and various omissions, replacements, and changes can be made without departing from the scope of the invention. These embodiments and modifications thereof are included in the scope and gist of the invention, and are included in the invention described in the claims and the equivalents thereof.
 また、上述の実施形態の中で示した処理手順に示された指示は、ソフトウェアであるプログラムに基づいて実行されることが可能である。汎用の計算機システムが、このプログラムを予め記憶しておき、このプログラムを読み込むことにより、上述した実施形態の動画像符号化装置及び動画像復号化装置による効果と同様な効果を得ることも可能である。上述の実施形態で記述された指示は、コンピュータに実行させることのできるプログラムとして、磁気ディスク(フレキシブルディスク、ハードディスクなど)、光ディスク(CD-ROM、CD-R、CD-RW、DVD-ROM、DVD±R、DVD±RWなど)、半導体メモリ、またはこれに類する記録媒体に記録される。コンピュータまたは組み込みシステムが読み取り可能な記録媒体であれば、その記憶形式は何れの形態であってもよい。コンピュータは、この記録媒体からプログラムを読み込み、このプログラムに基づいてプログラムに記述されている指示をCPUで実行させれば、上述した実施形態の動画像符号化装置及び動画像復号化装置と同様な動作を実現することができる。もちろん、コンピュータがプログラムを取得する場合または読み込む場合はネットワークを通じて取得または読み込んでもよい。 
 また、記録媒体からコンピュータや組み込みシステムにインストールされたプログラムの指示に基づきコンピュータ上で稼働しているOS(オペレーティングシステム)や、データベース管理ソフト、ネットワーク等のMW(ミドルウェア)等が本実施形態を実現するための各処理の一部を実行してもよい。 
 さらに、本願発明における記録媒体は、コンピュータあるいは組み込みシステムと独立した媒体に限らず、LANやインターネット等により伝達されたプログラムをダウンロードして記憶または一時記憶した記録媒体も含まれる。また、上記各実施形態の処理を実現するプログラムを、インターネットなどのネットワークに接続されたコンピュータ(サーバ)上に格納し、ネットワーク経由でコンピュータ(クライアント)にダウンロードさせてもよい。 
 また、記録媒体は1つに限られず、複数の媒体から本実施形態における処理が実行される場合も、本発明における記録媒体に含まれ、媒体の構成は何れの構成であってもよい。
The instructions shown in the processing procedure shown in the above embodiment can be executed based on a program that is software. A general-purpose computer system stores this program in advance, and by reading this program, it is also possible to obtain the same effects as those obtained by the video encoding device and video decoding device of the above-described embodiment. is there. The instructions described in the above-described embodiments are, as programs that can be executed by a computer, magnetic disks (flexible disks, hard disks, etc.), optical disks (CD-ROM, CD-R, CD-RW, DVD-ROM, DVD). ± R, DVD ± RW, etc.), semiconductor memory, or a similar recording medium. As long as the recording medium is readable by the computer or the embedded system, the storage format may be any form. If the computer reads the program from the recording medium and causes the CPU to execute instructions described in the program based on the program, the computer is similar to the video encoding device and video decoding device of the above-described embodiment. Operation can be realized. Of course, when the computer acquires or reads the program, it may be acquired or read through a network.
In addition, the OS (operating system), database management software, MW (middleware) such as a network, etc. running on the computer based on the instructions of the program installed in the computer or embedded system from the recording medium implement this embodiment. A part of each process for performing may be executed.
Furthermore, the recording medium in the present invention is not limited to a medium independent of a computer or an embedded system, but also includes a recording medium in which a program transmitted via a LAN or the Internet is downloaded and stored or temporarily stored. Further, the program for realizing the processing of each of the above embodiments may be stored on a computer (server) connected to a network such as the Internet and downloaded to the computer (client) via the network.
Further, the number of recording media is not limited to one, and when the processing in the present embodiment is executed from a plurality of media, it is included in the recording media in the present invention, and the configuration of the media may be any configuration.
 なお、本願発明におけるコンピュータまたは組み込みシステムは、記録媒体に記憶されたプログラムに基づき、本実施形態における各処理を実行するためのものであって、パソコン、マイコン等の1つからなる装置、複数の装置がネットワーク接続されたシステム等の何れの構成であってもよい。 
 また、本願発明の実施形態におけるコンピュータとは、パソコンに限らず、情報処理機器に含まれる演算処理装置、マイコン等も含み、プログラムによって本発明の実施形態における機能を実現することが可能な機器、装置を総称している。
The computer or the embedded system in the present invention is for executing each process in the present embodiment based on a program stored in a recording medium, and includes a single device such as a personal computer or a microcomputer, Any configuration such as a system in which apparatuses are connected to a network may be used.
Further, the computer in the embodiment of the present invention is not limited to a personal computer, but includes an arithmetic processing device, a microcomputer, and the like included in an information processing device, and a device capable of realizing the functions in the embodiment of the present invention by a program, The device is a general term.
 本発明のいくつかの実施形態を説明したが、これらの実施形態は、例として提示したものであり、発明の範囲を限定することは意図していない。これら新規な実施形態は、その他の様々な形態で実施されることが可能であり、発明の要旨を逸脱しない範囲で、種々の省略、置き換え、変更を行うことができる。これら実施形態やその変形は、発明の範囲や要旨に含まれるとともに、特許請求の範囲に記載された発明とその均等の範囲に含まれる。 Although several embodiments of the present invention have been described, these embodiments are presented as examples and are not intended to limit the scope of the invention. These novel embodiments can be implemented in various other forms, and various omissions, replacements, and changes can be made without departing from the scope of the invention. These embodiments and modifications thereof are included in the scope and gist of the invention, and are included in the invention described in the claims and the equivalents thereof.
101…減算部、102…直交変換部、103…量子化部、104…逆量子化部、105…逆直交変換部、106…加算部、107…ループフィルタ、108…参照画像メモリ、109…イントラ予測部、110…インター予測部、111…予測選択スイッチ、112…予測選択部、113…エントロピー符号化部、114…出力バッファ、115…符号化制御部、116…入力画像信号、117…予測誤差信号、118…変換係数、119…量子化変換係数、120…復元変換係数、121…復元予測誤差信号、122…復号画像信号、123…被フィルタ画像信号、124…参照画像信号、125…予測情報、126…予測画像信号、127…符号化データ、601…単方向イントラ予測画像生成部、602…双方向イントラ予測画像生成部、603…予測モード情報設定部、604…選択スイッチ、605…予測モード、1301…第一単方向イントラ予測画像生成部、1302…第二単方向イントラ予測画像生成部、1303…重み付き平均部、1304…第一予測画像信号、1305…第二予測画像信号、1901…画像バッファ、2000…シンタクス、2001…ハイレベルシンタクス、2002…スライスレベルシンタクス、2003…コーディングツリーレベルシンタクス、2004…シーケンスパラメータセットシンタクス、2005…ピクチャパラメータセットシンタクス、2006…スライスヘッダーシンタクス、2007…スライスデータシンタクス、2008…コーディングツリーユニットシンタクス、2009…プレディクションユニットシンタクス、2010…トランスフォームユニットシンタクス、2601…参照画素フィルタ部、2602…被フィルタ参照画像信号、2901…複合イントラ予測画像生成部、2902…選択スイッチ、3001…復号画素バッファ、3001…復号画像バッファ、3002…参照画素、3101…画素レベル予測信号生成部、3102…複合イントラ予測計算部、3103…画素レベル予測信号、3104…隣接画素、3301…変換選択部、3302…係数順制御部、3303…変換選択情報、3304…量子化変換係数列、3401…第一直交変換部、3402…第二直交変換部、3403…第N直交変換部、3404…変換選択スイッチ、3501…第一逆直交変換部、3502…第二逆直交変換部、3503…第N逆直交変換部、3504…変換選択スイッチ、3701…第一係数順変換部、3702…第二係数順変換部、3703…第N係数順変換部、3704…係数順選択スイッチ、3801…発生頻度カウント部、3802…係数順更新部、3803…更新係数順、3804…ヒストグラム、4001…回転変換部、4001…第一回転変換部、4002…第二回転変換部、4003…第N回転変換部、4004…離散コサイン変換部、4101…第一逆回転変換部、4102…第二逆回転変換部、4103…第N逆回転変換部、4104…逆離散コサイン変換部、4401…入力バッファ、4402…エントロピー復号化部、4403…逆量子化部、4404…逆直交変換部、4405…加算部、4406…ループフィルタ、4407…参照画像メモリ、4408…イントラ予測部、4409…インター予測部、4410…予測選択スイッチ、4411…出力バッファ、4412…復号化制御部、4413…符号化データ、4414…量子化変換係数、4415…復元変換係数、4416…復元予測誤差信号、4417…復号画像信号、4418…被フィルタ画像信号、4419…参照画像信号、4420…予測画像信号、4421…予測情報、4422…復号画像、4501…係数順復元部、4502…変換選択部、4503…量子化変換係数列、4504…変換選択情報、4601、4602、4603…係数順逆変換部、4701…発生頻度カウント部、4702…更新部、4703…更新係数順、4704…ヒストグラム。 DESCRIPTION OF SYMBOLS 101 ... Subtraction part, 102 ... Orthogonal transformation part, 103 ... Quantization part, 104 ... Dequantization part, 105 ... Inverse orthogonal transformation part, 106 ... Addition part, 107 ... Loop filter, 108 ... Reference image memory, 109 ... Intra Prediction unit 110 ... Inter prediction unit 111 ... Prediction selection switch 112 ... Prediction selection unit 113 ... Entropy encoding unit 114 ... Output buffer 115 ... Encoding control unit 116 ... Input image signal 117 ... Prediction error Signal 118, transform coefficient, 119 quantized transform coefficient, 120 restored transform coefficient, 121 restored restoration error signal, 122 decoded image signal, 123 filtered image signal, 124 reference image signal, 125 prediction information , 126 ... predicted image signal, 127 ... encoded data, 601 ... unidirectional intra predicted image generation unit, 602 ... bidirectional intra predicted image raw 603 ... Prediction mode information setting unit, 604 ... Selection switch, 605 ... Prediction mode, 1301 ... First unidirectional intra prediction image generation unit, 1302 ... Second unidirectional intra prediction image generation unit, 1303 ... Weighted average unit DESCRIPTION OF SYMBOLS 1304 ... 1st prediction image signal, 1305 ... 2nd prediction image signal, 1901 ... Image buffer, 2000 ... Syntax, 2001 ... High level syntax, 2002 ... Slice level syntax, 2003 ... Coding tree level syntax, 2004 ... Sequence parameter set Syntax: 2005 ... Picture parameter set syntax, 2006: Slice header syntax, 2007 ... Slice data syntax, 2008 ... Coding tree unit syntax, 2009 ... Prediction unit syntax , 2010 ... Transform unit syntax, 2601 ... Reference pixel filter unit, 2602 ... Filtered reference image signal, 2901 ... Composite intra prediction image generation unit, 2902 ... Selection switch, 3001 ... Decoded pixel buffer, 3001 ... Decoded image buffer, 3002 Reference pixels, 3101 ... Pixel level prediction signal generation unit, 3102 ... Composite intra prediction calculation unit, 3103 ... Pixel level prediction signal, 3104 ... Adjacent pixel, 3301 ... Conversion selection unit, 3302 ... Coefficient order control unit, 3303 ... Conversion selection Information, 3304 ... Quantized transform coefficient sequence, 3401 ... First orthogonal transform unit, 3402 ... Second orthogonal transform unit, 3403 ... Nth orthogonal transform unit, 3404 ... Transform selection switch, 3501 ... First inverse orthogonal transform unit, 3502 ... Second inverse orthogonal transform unit, 3503 ... Nth inverse orthogonal transform unit, 3504 ... Conversion selection switch, 3701 ... First coefficient order conversion section, 3702 ... Second coefficient order conversion section, 3703 ... Nth coefficient order conversion section, 3704 ... Coefficient order selection switch, 3801 ... Occurrence frequency counting section, 3802 ... Coefficient order Update unit, 3803 ... Update coefficient order, 3804 ... Histogram, 4001 ... Rotation conversion unit, 4001 ... First rotation conversion unit, 4002 ... Second rotation conversion unit, 4003 ... Nth rotation conversion unit, 4004 ... Discrete cosine conversion unit, 4101: First reverse rotation conversion unit, 4102 ... Second reverse rotation conversion unit, 4103 ... Nth reverse rotation conversion unit, 4104 ... Inverse discrete cosine conversion unit, 4401 ... Input buffer, 4402 ... Entropy decoding unit, 4403 ... Reverse Quantization unit, 4404 ... inverse orthogonal transform unit, 4405 ... addition unit, 4406 ... loop filter, 4407 ... reference image memory, 4408 ... in La prediction unit, 4409 ... inter prediction unit, 4410 ... prediction selection switch, 4411 ... output buffer, 4412 ... decoding control unit, 4413 ... encoded data, 4414 ... quantized transform coefficient, 4415 ... restoration transform coefficient, 4416 ... restoration Prediction error signal, 4417 ... decoded image signal, 4418 ... filtered image signal, 4419 ... reference image signal, 4420 ... predicted image signal, 4421 ... prediction information, 4422 ... decoded image, 4501 ... coefficient order restoration unit, 4502 ... conversion selection. , 4503... Quantized transform coefficient sequence, 4504... Transform selection information, 4601, 4602, 4603... Coefficient order inverse transform section, 4701... Occurrence frequency count section, 4702.

Claims (20)

  1.  入力画像信号を四分木分割に従って、階層の深さで表現される画素ブロックに分割し、これら分割した画素ブロックに対して予測誤差信号を生成し、変換係数を符号化する動画像符号化方法において、
     複数の予測方向セットの中から、第一の予測方向を設定して第一予測画像信号を作成するステップと、
     前記複数の予測方向セットの中から、第一の予測方向とは異なる第二の予測方向を設定して第二予測画像信号を作成するステップと、
     前記第一及び第二の予測方向のそれぞれに関して、予測対象の画素と参照画素との間の相対距離を導出し、該相対距離の差分値を導出するステップと、
     前記差分値に応じて、予め定められた重み成分を導出するステップと、
     前記重み成分に従って、第一単方向イントラ予測画像と第二単方向イントラ予測画像を重み付き平均し、第三予測画像信号を生成するステップと、
     前記第三予測画像信号から予測誤差信号を生成するステップと、
     前記予測誤差信号を符号化するステップと、
    を具備する動画像符号化方法。
    A moving picture coding method that divides an input image signal into pixel blocks expressed by hierarchical depth according to quadtree division, generates a prediction error signal for these divided pixel blocks, and encodes transform coefficients In
    Creating a first predicted image signal by setting a first predicted direction from among a plurality of predicted direction sets;
    Creating a second prediction image signal by setting a second prediction direction different from the first prediction direction from the plurality of prediction direction sets;
    Deriving a relative distance between a pixel to be predicted and a reference pixel for each of the first and second prediction directions, and deriving a difference value of the relative distance;
    Deriving a predetermined weight component according to the difference value;
    In accordance with the weight component, a weighted average of the first unidirectional intra-predicted image and the second unidirectional intra-predicted image to generate a third predicted image signal;
    Generating a prediction error signal from the third predicted image signal;
    Encoding the prediction error signal;
    A video encoding method comprising:
  2.  前記複数の予測方向セットの中から、第四の予測方向を設定して第四予測画像信号を作成するステップと、
     前記第三予測画像信号と前記第四予測画像信号のうちの一つを第五予測画像信号として選択するステップと、
     前記第五予測画像信号から予測誤差信号を生成するステップと、
     予測誤差信号を符号化するステップと、
    をさらに具備する請求項1に記載の動画像符号化方法。
    Creating a fourth predicted image signal by setting a fourth prediction direction from the plurality of prediction direction sets;
    Selecting one of the third predicted image signal and the fourth predicted image signal as a fifth predicted image signal;
    Generating a prediction error signal from the fifth predicted image signal;
    Encoding a prediction error signal;
    The moving picture encoding method according to claim 1, further comprising:
  3.  前記第五予測画像信号を特定する予測モードを符号化するために、
     少なくとも一つの符号化済み画素ブロックに対応する前記予測モードから、参照予測モードを生成するステップと、
     前記参照予測モードを利用して前記予測モードを予測符号化するステップと、
    をさらに具備する請求項2に記載の動画像符号化方法。
    In order to encode a prediction mode that identifies the fifth predicted image signal,
    Generating a reference prediction mode from the prediction mode corresponding to at least one encoded pixel block;
    Predictively encoding the prediction mode using the reference prediction mode;
    The video encoding method according to claim 2, further comprising:
  4.  前記予測モードが前記第五予測画像信号として前記第四予測画像信号を特定し、さらに前記参照予測モードが前記第五予測画像信号として前記第三予測画像信号を特定する場合に、
     前記第三予測画像信号に含まれる前記第一予測画像信号及び前記第二予測画像信号のうち、どちらか一つを特定する前記予測モードを前記参照予測モードとして置き換えるステップをさらに具備する請求項3に記載の動画像符号化方法。
    When the prediction mode specifies the fourth prediction image signal as the fifth prediction image signal, and the reference prediction mode specifies the third prediction image signal as the fifth prediction image signal,
    The method further comprising: replacing the prediction mode for specifying one of the first prediction image signal and the second prediction image signal included in the third prediction image signal as the reference prediction mode. The moving image encoding method described in 1.
  5.  前記相対距離は、予測方向の始点に対応する参照画素を基準とし、ユークリッド距離或いは市街地距離に基づいて導出され、
     前記相対距離の差分値は、前記第一の予測方向と前記第二の予測方向とでそれぞれ導出された前記差分値と前記画素ブロックの幅若しくは高さに基づいて導出され、
     前記第一の予測方向の重み成分は、前記相対距離が近いほど大きく、遠いほど小さくなるように設計される請求項4に記載の動画像符号化方法。
    The relative distance is derived based on the Euclidean distance or the urban area distance, based on the reference pixel corresponding to the start point of the prediction direction,
    The difference value of the relative distance is derived based on the difference value derived in the first prediction direction and the second prediction direction, respectively, and the width or height of the pixel block,
    The moving picture coding method according to claim 4, wherein the weight component in the first prediction direction is designed to be larger as the relative distance is closer and smaller as it is farther.
  6.  前記予測誤差信号を直交変換するステップをさらに具備し、
     前記直交変換は、選択された第一若しくは第二のいずれか定められた予測方向に対して、予測誤差の絶対値が参照画素からの距離に応じて大きくなる傾向に基づいて、離散コサイン変換よりも変換後の係数集密度が高くなるように予測方向毎に予め複数設計された変換行列セットを用いて直交変換する請求項5に記載の動画像符号化方法。
    Further comprising orthogonally transforming the prediction error signal;
    The orthogonal transform is based on the tendency that the absolute value of the prediction error increases according to the distance from the reference pixel with respect to the selected first or second predetermined prediction direction. 6. The moving picture coding method according to claim 5, wherein orthogonal transformation is performed using a plurality of transformation matrix sets designed in advance for each prediction direction so that the coefficient density after transformation is high.
  7.  前記第五予測画像信号と前記入力画像信号の重み付き平均を行って局部予測画像信号を生成するステップと、
     前記局部予測画像信号に基づいて局部復号画像信号を生成するステップと、
     前記第五予測画像信号と前記局部復号画像信号から、前記同じ重み付き平均を行い、第六予測画像信号を生成するステップと、
    をさらに具備する請求項6に記載の動画像符号化方法。
    Performing a weighted average of the fifth predicted image signal and the input image signal to generate a local predicted image signal;
    Generating a locally decoded image signal based on the locally predicted image signal;
    Performing the same weighted average from the fifth predicted image signal and the locally decoded image signal to generate a sixth predicted image signal;
    The video encoding method according to claim 6, further comprising:
  8.  前記重み付き平均は、前記第六予測画像信号を作成した際の、前記第一或いは前記第二のいずれか定められた予測方向に対して、参照画像との相対距離を導出し、相対距離が近い画素の重み成分を大きくし、相対距離が遠い画素の重み成分を小さくする請求項7に記載の動画像符号化方法。 The weighted average derives a relative distance from a reference image with respect to the first or second predetermined prediction direction when the sixth predicted image signal is generated, and the relative distance is The moving image encoding method according to claim 7, wherein a weight component of a close pixel is increased and a weight component of a pixel having a long relative distance is decreased.
  9.  前記第一予測画像信号、前記第二予測画像信号、前記第四予測画像信号を作成する際に利用する既に符号化された参照画素に対し、
     前記参照画素と該参照画素に隣接する参照画素とのフィルタリングにより前記参照画素を置き換えるステップと、
     前記フィルタリングの適用有無を、前記画素ブロック毎に選択するステップと、
    をさらに具備する請求項7に記載の動画像符号化方法。
    For the already-encoded reference pixels used when creating the first predicted image signal, the second predicted image signal, and the fourth predicted image signal,
    Replacing the reference pixel by filtering the reference pixel and a reference pixel adjacent to the reference pixel;
    Selecting whether to apply the filtering for each pixel block;
    The moving picture coding method according to claim 7, further comprising:
  10.  前記予測誤差信号を直交変換するステップをさらに具備し、
     前記直交変換は、離散コサイン変換を行って変換係数を得るステップと、前記変換係数に回転変換を行って、さらに変換係数を変換するステップと、を含み、
     前記回転変換は、前記離散コサイン変換よりも回転変換後の係数集密度が高くなるように、予め複数設計された回転変換行列セットから1つを選択して回転変換し、前記選択した回転変換行列セットを示すインデックスを符号化する請求項9に記載の動画像符号化方法。
    Further comprising orthogonally transforming the prediction error signal;
    The orthogonal transform includes a step of obtaining a transform coefficient by performing a discrete cosine transform, a step of performing a rotational transform on the transform coefficient, and further transforming the transform coefficient,
    In the rotation transformation, one selected from a plurality of rotation transformation matrix sets designed in advance so as to have higher coefficient density after the rotation transformation than the discrete cosine transformation, and the rotation transformation matrix is selected. The moving image encoding method according to claim 9, wherein an index indicating a set is encoded.
  11.  入力画像信号を四分木分割に従って、階層の深さで表現される画素ブロックに分割し、これら分割した画素ブロックに対して復号処理を行う動画像復号化方法において、
     複数の予測方向セットの中から、復号された第一の予測方向を設定して第一予測画像信号を作成するステップと、
     前記複数の予測方向セットの中から、前記第一の予測方向と異なる復号された第二の予測方向を設定して第二予測画像信号を作成するステップと、
     第一及び第二の予測方向のそれぞれに関して、予測対象の画素と参照画素との間の相対距離を導出し、該相対距離の差分値を導出するステップと、
     前記差分値に応じて、予め定められた重み成分を導出するステップと、
     前記重み成分に従って、第一単方向イントラ予測画像と第二単方向イントラ予測画像を重み付き平均し、第三予測画像信号を生成するステップと、
     前記第三予測画像信号から復号画像信号を生成するステップと、
    を具備する動画像復号化方法。
    In the moving picture decoding method in which the input image signal is divided into pixel blocks represented by the depth of the hierarchy according to quadtree division, and decoding processing is performed on these divided pixel blocks.
    Creating a first predicted image signal by setting a decoded first prediction direction from among a plurality of prediction direction sets;
    Creating a second predicted image signal by setting a decoded second prediction direction different from the first prediction direction from the plurality of prediction direction sets;
    Deriving a relative distance between a pixel to be predicted and a reference pixel for each of the first and second prediction directions, and deriving a difference value of the relative distance;
    Deriving a predetermined weight component according to the difference value;
    In accordance with the weight component, a weighted average of the first unidirectional intra-predicted image and the second unidirectional intra-predicted image to generate a third predicted image signal;
    Generating a decoded image signal from the third predicted image signal;
    A video decoding method comprising:
  12.  複数の予測方向セットの中から、第四の予測方向を設定して第四予測画像信号を作成するステップと、
     前記第三予測画像信号と前記第四予測画像信号のうちの一つを第五予測画像信号として選択するステップと、
     前記第五予測画像信号から予測誤差信号を生成するステップと、
     予測誤差信号を復号化するステップと、
    をさらに具備する請求項11に記載の動画像復号化方法。
    Creating a fourth predicted image signal by setting a fourth predicted direction from among a plurality of predicted direction sets;
    Selecting one of the third predicted image signal and the fourth predicted image signal as a fifth predicted image signal;
    Generating a prediction error signal from the fifth predicted image signal;
    Decoding a prediction error signal;
    The moving picture decoding method according to claim 11, further comprising:
  13.  前記第五予測画像信号を特定する予測モードを復号化するために、
     少なくとも一つの復号化済み画素ブロックに対応する前記予測モードから、参照予測モードを生成するステップと、
     前記参照予測モードを利用して前記予測モードを予測復号化するステップと、
    をさらに具備する請求項12に記載の動画像復号化方法。
    In order to decode a prediction mode that identifies the fifth predicted image signal,
    Generating a reference prediction mode from the prediction mode corresponding to at least one decoded pixel block;
    Predictively decoding the prediction mode using the reference prediction mode;
    The moving picture decoding method according to claim 12, further comprising:
  14.  前記予測モードが前記第五予測画像信号として前記第四予測画像信号を特定し、さらに前記参照予測モードが前記第五予測画像信号として前記第三予測画像信号を特定する場合に、
     前記第三予測画像信号に含まれる前記第一予測画像信号及び前記第二予測画像信号のうち、どちらか一つを特定する前記予測モードを前記参照予測モードとして置き換えるステップをさらに具備する請求項13に記載の動画像復号化方法。
    When the prediction mode specifies the fourth prediction image signal as the fifth prediction image signal, and the reference prediction mode specifies the third prediction image signal as the fifth prediction image signal,
    The method of claim 13, further comprising: replacing the prediction mode for specifying one of the first prediction image signal and the second prediction image signal included in the third prediction image signal as the reference prediction mode. The moving picture decoding method described in 1.
  15.  前記相対距離は、予測方向の始点に対応する参照画素を基準とし、ユークリッド距離或いは市街地距離に基づいて導出され、
     前記相対距離の差分値は、前記第一の予測方向と前記第二の予測方向とでそれぞれ導出された前記差分値と前記画素ブロックの幅若しくは高さに基づいて導出され、
     前記第一の予測方向の重み成分は、前記相対距離が近いほど大きく、遠いほど小さくなるように設計される請求項14に記載の動画像復号化方法。
    The relative distance is derived based on the Euclidean distance or the urban area distance, based on the reference pixel corresponding to the start point of the prediction direction,
    The difference value of the relative distance is derived based on the difference value derived in the first prediction direction and the second prediction direction, respectively, and the width or height of the pixel block,
    The moving picture decoding method according to claim 14, wherein the weight component in the first prediction direction is designed to be larger as the relative distance is closer and smaller as it is farther.
  16.  復号された第一若しくは第二のいずれか定められた予測方向に対して、予測誤差の絶対値が参照画素からの距離に応じて大きくなる傾向に基づいて、離散コサイン変換よりも変換後の係数集密度が高くなるように予測方向毎に予め複数設計された変換行列セットを用いて逆直交変換するステップをさらに具備する請求項15に記載の動画像復号化方法。 Based on the tendency that the absolute value of the prediction error increases according to the distance from the reference pixel with respect to the decoded first or second determined prediction direction, the coefficient after the conversion is more than the discrete cosine conversion The moving picture decoding method according to claim 15, further comprising a step of performing an inverse orthogonal transformation using a plurality of transformation matrix sets designed in advance for each prediction direction so as to increase the density.
  17.  符号化データを復号し、局部復号画像信号を作成するステップと、
     前記第三予測画像信号と前記局部復号画像信号の重み付き平均を行って第四予測画像信号を生成するステップと、
    をさらに具備する請求項16に記載の動画像復号化方法。
    Decoding the encoded data to create a locally decoded image signal;
    Performing a weighted average of the third predicted image signal and the locally decoded image signal to generate a fourth predicted image signal;
    The video decoding method according to claim 16, further comprising:
  18.  前記重み付き平均は、前記第三予測画像信号を作成した際の、前記第一或いは前記第二のいずれか定められた予測方向に対して、参照画像との相対距離を導出し、相対距離が近い画素の重み成分を大きくし、相対距離が遠い画素の重み成分を小さくする請求項17に記載の動画像復号化方法。 The weighted average derives a relative distance from a reference image with respect to the first or second predetermined prediction direction when the third predicted image signal is generated, and the relative distance is The moving picture decoding method according to claim 17, wherein a weight component of a close pixel is increased and a weight component of a pixel having a relative distance is decreased.
  19.  前記第一予測画像信号、前記第二予測画像信号、前記第四予測画像信号を作成する際に利用する既に復号化された参照画素に対し、
     前記参照画素と該参照画素に隣接する参照画素とのフィルタリングにより前記参照画素を置き換えるステップと、
     前記フィルタリングの適用有無を、前記画素ブロック毎に選択するステップと、
    をさらに具備する請求項17に記載の動画像復号化方法。
    For the already decoded reference pixels used when creating the first predicted image signal, the second predicted image signal, and the fourth predicted image signal,
    Replacing the reference pixel by filtering the reference pixel and a reference pixel adjacent to the reference pixel;
    Selecting whether to apply the filtering for each pixel block;
    The video decoding method according to claim 17, further comprising:
  20.  前記回転変換行列セットを示すインデックスを復号するステップと、
     離散コサイン変換よりも回転変換後の係数集密度が高くなるように、予め複数設計された回転変換行列セットから復号されたインデックスに従って1つを選択して復号された変換係数に逆回転変換を行って、変換係数を逆変換するステップと、
     前記逆回転変換後の変換係数に対して逆離散コサイン変換を行って予測誤差を得るステップと、
    をさらに具備する請求項18に記載の動画像復号化方法。
    Decoding an index indicating the rotation transformation matrix set;
    In order to increase the coefficient density after the rotation transformation than the discrete cosine transformation, one is selected according to an index decoded from a plurality of rotation transformation matrix sets designed in advance, and the inverse rotation transformation is performed on the decoded transformation coefficient. Reversely transforming the transform coefficient;
    Performing inverse discrete cosine transform on the transform coefficient after the inverse rotation transform to obtain a prediction error;
    The video decoding method according to claim 18, further comprising:
PCT/JP2010/066102 2010-09-16 2010-09-16 Moving picture encoding method and moving picture decoding method WO2012035640A1 (en)

Priority Applications (1)

Application Number Priority Date Filing Date Title
PCT/JP2010/066102 WO2012035640A1 (en) 2010-09-16 2010-09-16 Moving picture encoding method and moving picture decoding method

Applications Claiming Priority (1)

Application Number Priority Date Filing Date Title
PCT/JP2010/066102 WO2012035640A1 (en) 2010-09-16 2010-09-16 Moving picture encoding method and moving picture decoding method

Publications (1)

Publication Number Publication Date
WO2012035640A1 true WO2012035640A1 (en) 2012-03-22

Family

ID=45831140

Family Applications (1)

Application Number Title Priority Date Filing Date
PCT/JP2010/066102 WO2012035640A1 (en) 2010-09-16 2010-09-16 Moving picture encoding method and moving picture decoding method

Country Status (1)

Country Link
WO (1) WO2012035640A1 (en)

Cited By (11)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
WO2017205701A1 (en) * 2016-05-25 2017-11-30 Arris Enterprises Llc Weighted angular prediction for intra coding
CN110063056A (en) * 2016-12-07 2019-07-26 株式会社Kt Method and apparatus for handling vision signal
US10542264B2 (en) 2017-04-04 2020-01-21 Arris Enterprises Llc Memory reduction implementation for weighted angular prediction
CN110807789A (en) * 2019-08-23 2020-02-18 腾讯科技(深圳)有限公司 Image processing method, model, device, electronic equipment and readable storage medium
US10575023B2 (en) 2017-10-09 2020-02-25 Arris Enterprises Llc Adaptive unequal weight planar prediction
US10616596B2 (en) 2016-12-28 2020-04-07 Arris Enterprises Llc Unequal weight planar prediction
US10645395B2 (en) 2016-05-25 2020-05-05 Arris Enterprises Llc Weighted angular prediction coding for intra coding
CN111108749A (en) * 2018-09-25 2020-05-05 北京大学 Encoding method, decoding method, encoding device, and decoding device
CN111279703A (en) * 2017-07-05 2020-06-12 艾锐势有限责任公司 Post-filtering for weighted angle prediction
JP2020522180A (en) * 2017-05-29 2020-07-27 オランジュ Method and device for encoding and decoding a data stream representing at least one image
US10944963B2 (en) 2016-05-25 2021-03-09 Arris Enterprises Llc Coding weighted angular prediction for intra coding

Citations (3)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
JP2004200991A (en) * 2002-12-18 2004-07-15 Nippon Telegr & Teleph Corp <Ntt> Image encoding method, image decoding method, image encoder, image decoder, image encoding program, image decoding program, recording medium with the image recording program, and recording medium with image decording program recorded thereon
WO2007063808A1 (en) * 2005-11-30 2007-06-07 Kabushiki Kaisha Toshiba Image encoding/image decoding method and image encoding/image decoding apparatus
WO2008084817A1 (en) * 2007-01-09 2008-07-17 Kabushiki Kaisha Toshiba Image encoding and decoding method and device

Patent Citations (3)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
JP2004200991A (en) * 2002-12-18 2004-07-15 Nippon Telegr & Teleph Corp <Ntt> Image encoding method, image decoding method, image encoder, image decoder, image encoding program, image decoding program, recording medium with the image recording program, and recording medium with image decording program recorded thereon
WO2007063808A1 (en) * 2005-11-30 2007-06-07 Kabushiki Kaisha Toshiba Image encoding/image decoding method and image encoding/image decoding apparatus
WO2008084817A1 (en) * 2007-01-09 2008-07-17 Kabushiki Kaisha Toshiba Image encoding and decoding method and device

Non-Patent Citations (1)

* Cited by examiner, † Cited by third party
Title
PENG ZHANG ET AL.: "Multiple modes intra-prediction in intra coding", 2004 IEEE INTERNATIONAL CONFERENCE ON MULTIMEDIA AND EXPO (ICME '04), vol. 1, 30 June 2004 (2004-06-30), pages 419 - 422 *

Cited By (32)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US10944963B2 (en) 2016-05-25 2021-03-09 Arris Enterprises Llc Coding weighted angular prediction for intra coding
US11997290B2 (en) 2016-05-25 2024-05-28 Arris Enterprises Llc Weighted angular prediction for intra coding
US10523949B2 (en) 2016-05-25 2019-12-31 Arris Enterprises Llc Weighted angular prediction for intra coding
US11917166B2 (en) 2016-05-25 2024-02-27 Arris Enterprises Llc Weighted angular prediction coding for intra coding
US11758153B2 (en) 2016-05-25 2023-09-12 Arris Enterprises Llc Weighted angular prediction for intra coding
WO2017205701A1 (en) * 2016-05-25 2017-11-30 Arris Enterprises Llc Weighted angular prediction for intra coding
US11627312B2 (en) 2016-05-25 2023-04-11 Arris Enterprises Llc Coding weighted angular prediction for intra coding
US10645395B2 (en) 2016-05-25 2020-05-05 Arris Enterprises Llc Weighted angular prediction coding for intra coding
US11553189B2 (en) 2016-05-25 2023-01-10 Arris Enterprises Llc Weighted angular prediction coding for intra coding
US11303906B2 (en) 2016-05-25 2022-04-12 Arris Enterprises Llc Weighted angular prediction for intra coding
US11153573B2 (en) 2016-05-25 2021-10-19 Arris Enterprises Llc Weighted angular prediction coding for intra coding
US20200322600A1 (en) 2016-05-25 2020-10-08 Arris Enterprises Llc Weighted angular prediction for intra coding
US10939097B2 (en) 2016-05-25 2021-03-02 Arris Enterprises Llc Weighted angular prediction for intra coding
US11716467B2 (en) 2016-12-07 2023-08-01 Kt Corporation Method and apparatus for processing video signal
CN110063056A (en) * 2016-12-07 2019-07-26 株式会社Kt Method and apparatus for handling vision signal
CN110063056B (en) * 2016-12-07 2023-09-12 株式会社Kt Method and apparatus for processing video signal
US11736686B2 (en) 2016-12-07 2023-08-22 Kt Corporation Method and apparatus for processing video signal
US11019353B2 (en) 2016-12-28 2021-05-25 Arris Enterprises Llc Unequal weight planar prediction
US10616596B2 (en) 2016-12-28 2020-04-07 Arris Enterprises Llc Unequal weight planar prediction
US10542264B2 (en) 2017-04-04 2020-01-21 Arris Enterprises Llc Memory reduction implementation for weighted angular prediction
US11575915B2 (en) 2017-04-04 2023-02-07 Arris Enterprises Llc Memory reduction implementation for weighted angular prediction
JP7274427B2 (en) 2017-05-29 2023-05-16 オランジュ Method and device for encoding and decoding data streams representing at least one image
JP2020522180A (en) * 2017-05-29 2020-07-27 オランジュ Method and device for encoding and decoding a data stream representing at least one image
CN111279703A (en) * 2017-07-05 2020-06-12 艾锐势有限责任公司 Post-filtering for weighted angle prediction
CN111279703B (en) * 2017-07-05 2023-08-04 艾锐势有限责任公司 Post-filtering for weighted angle prediction
US11627315B2 (en) 2017-07-05 2023-04-11 Arris Enterprises Llc Post-filtering for weighted angular prediction
US11902519B2 (en) 2017-07-05 2024-02-13 Arris Enterprises Llc Post-filtering for weighted angular prediction
US10992934B2 (en) 2017-07-05 2021-04-27 Arris Enterprises Llc Post-filtering for weighted angular prediction
US10575023B2 (en) 2017-10-09 2020-02-25 Arris Enterprises Llc Adaptive unequal weight planar prediction
US11159828B2 (en) 2017-10-09 2021-10-26 Arris Enterprises Llc Adaptive unequal weight planar prediction
CN111108749A (en) * 2018-09-25 2020-05-05 北京大学 Encoding method, decoding method, encoding device, and decoding device
CN110807789A (en) * 2019-08-23 2020-02-18 腾讯科技(深圳)有限公司 Image processing method, model, device, electronic equipment and readable storage medium

Similar Documents

Publication Publication Date Title
US11936858B1 (en) Constrained position dependent intra prediction combination (PDPC)
WO2012035640A1 (en) Moving picture encoding method and moving picture decoding method
US9392282B2 (en) Moving-picture encoding apparatus and moving-picture decoding apparatus
Han et al. Improved video compression efficiency through flexible unit representation and corresponding extension of coding tools
KR101677406B1 (en) Video codec architecture for next generation video
WO2011125256A1 (en) Image encoding method and image decoding method
WO2012148139A2 (en) Method for managing a reference picture list, and apparatus using same
US20120128064A1 (en) Image processing device and method
JP7124222B2 (en) Method and apparatus for color conversion in VVC
CN111819853A (en) Signaling residual symbols for prediction in transform domain
CN113163211B (en) Inter-frame prediction method and device based on merging mode
JP2024026317A (en) Method, device, and program for decoding and encoding encoded unit
JP2024056945A (en) Method, apparatus and program for decoding and encoding coding units
KR20220032620A (en) Method, apparatus and system for encoding and decoding a block of video samples
KR20220041940A (en) Method, apparatus and system for encoding and decoding a block of video samples
JP2017073598A (en) Moving image coding apparatus, moving image coding method, and computer program for moving image coding
WO2012090286A1 (en) Video image encoding method, and video image decoding method
JP5537695B2 (en) Image decoding apparatus, method and program
WO2012172667A1 (en) Video encoding method, video decoding method, and device
JP6042478B2 (en) Image decoding device
JP5367161B2 (en) Image encoding method, apparatus, and program
JP2013176110A (en) Image encoder
JP6871343B2 (en) Moving image coding method and moving image decoding method
JP5649701B2 (en) Image decoding method, apparatus, and program
JP6154588B2 (en) Image encoding apparatus, image decoding apparatus, and program

Legal Events

Date Code Title Description
121 Ep: the epo has been informed by wipo that ep was designated in this application

Ref document number: 10857274

Country of ref document: EP

Kind code of ref document: A1

NENP Non-entry into the national phase

Ref country code: DE

122 Ep: pct application non-entry in european phase

Ref document number: 10857274

Country of ref document: EP

Kind code of ref document: A1

NENP Non-entry into the national phase

Ref country code: JP