WO2013001939A1 - Dispositif de traitement d'image et procédé de traitement d'image - Google Patents

Dispositif de traitement d'image et procédé de traitement d'image Download PDF

Info

Publication number
WO2013001939A1
WO2013001939A1 PCT/JP2012/062925 JP2012062925W WO2013001939A1 WO 2013001939 A1 WO2013001939 A1 WO 2013001939A1 JP 2012062925 W JP2012062925 W JP 2012062925W WO 2013001939 A1 WO2013001939 A1 WO 2013001939A1
Authority
WO
WIPO (PCT)
Prior art keywords
prediction
unit
mode
layer
prediction mode
Prior art date
Application number
PCT/JP2012/062925
Other languages
English (en)
Japanese (ja)
Inventor
佐藤 数史
Original Assignee
ソニー株式会社
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Application filed by ソニー株式会社 filed Critical ソニー株式会社
Priority to CN201280030622.8A priority Critical patent/CN103636211A/zh
Priority to US14/110,984 priority patent/US20140037002A1/en
Publication of WO2013001939A1 publication Critical patent/WO2013001939A1/fr

Links

Images

Classifications

    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04NPICTORIAL COMMUNICATION, e.g. TELEVISION
    • H04N19/00Methods or arrangements for coding, decoding, compressing or decompressing digital video signals
    • H04N19/10Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using adaptive coding
    • H04N19/169Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using adaptive coding characterised by the coding unit, i.e. the structural portion or semantic portion of the video signal being the object or the subject of the adaptive coding
    • H04N19/17Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using adaptive coding characterised by the coding unit, i.e. the structural portion or semantic portion of the video signal being the object or the subject of the adaptive coding the unit being an image region, e.g. an object
    • H04N19/176Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using adaptive coding characterised by the coding unit, i.e. the structural portion or semantic portion of the video signal being the object or the subject of the adaptive coding the unit being an image region, e.g. an object the region being a block, e.g. a macroblock
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04NPICTORIAL COMMUNICATION, e.g. TELEVISION
    • H04N19/00Methods or arrangements for coding, decoding, compressing or decompressing digital video signals
    • H04N19/10Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using adaptive coding
    • H04N19/102Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using adaptive coding characterised by the element, parameter or selection affected or controlled by the adaptive coding
    • H04N19/103Selection of coding mode or of prediction mode
    • H04N19/11Selection of coding mode or of prediction mode among a plurality of spatial predictive coding modes
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04NPICTORIAL COMMUNICATION, e.g. TELEVISION
    • H04N19/00Methods or arrangements for coding, decoding, compressing or decompressing digital video signals
    • H04N19/10Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using adaptive coding
    • H04N19/134Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using adaptive coding characterised by the element, parameter or criterion affecting or controlling the adaptive coding
    • H04N19/157Assigned coding mode, i.e. the coding mode being predefined or preselected to be further used for selection of another element or parameter
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04NPICTORIAL COMMUNICATION, e.g. TELEVISION
    • H04N19/00Methods or arrangements for coding, decoding, compressing or decompressing digital video signals
    • H04N19/10Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using adaptive coding
    • H04N19/189Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using adaptive coding characterised by the adaptation method, adaptation tool or adaptation type used for the adaptive coding
    • H04N19/196Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using adaptive coding characterised by the adaptation method, adaptation tool or adaptation type used for the adaptive coding being specially adapted for the computation of encoding parameters, e.g. by averaging previously computed encoding parameters
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04NPICTORIAL COMMUNICATION, e.g. TELEVISION
    • H04N19/00Methods or arrangements for coding, decoding, compressing or decompressing digital video signals
    • H04N19/30Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using hierarchical techniques, e.g. scalability
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04NPICTORIAL COMMUNICATION, e.g. TELEVISION
    • H04N19/00Methods or arrangements for coding, decoding, compressing or decompressing digital video signals
    • H04N19/46Embedding additional information in the video signal during the compression process
    • H04N19/463Embedding additional information in the video signal during the compression process by compressing encoding parameters before transmission
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04NPICTORIAL COMMUNICATION, e.g. TELEVISION
    • H04N19/00Methods or arrangements for coding, decoding, compressing or decompressing digital video signals
    • H04N19/50Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using predictive coding
    • H04N19/593Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using predictive coding involving spatial prediction techniques

Definitions

  • the present disclosure relates to an image processing apparatus and an image processing method.
  • Intra prediction is a technique for reducing the amount of encoded information by using the correlation between adjacent blocks in an image and predicting the pixel values in a block from the pixel values of other adjacent blocks. .
  • intra prediction is possible for all pixel components.
  • intra prediction may be performed using a block of 4 ⁇ 4 pixels, 8 ⁇ 8 pixels, or 16 ⁇ 16 pixels as one processing unit (ie, prediction unit (PU)).
  • PU prediction unit
  • HEVC High Efficiency Video Coding
  • the size of the prediction unit is about to be expanded to 32 ⁇ 32 pixels and 64 ⁇ 64 pixels.
  • an optimal prediction mode for predicting a pixel value of a prediction target block is selected from a plurality of prediction modes.
  • the prediction mode can typically be distinguished by the prediction direction from the reference pixel to the prediction target pixel.
  • H.M For the prediction unit of 4 ⁇ 4 pixels and 8 ⁇ 8 pixels of the H.264 / AVC luminance component, there are 8 types of prediction directions (vertical, horizontal, diagonally lower left, diagonally lower right, vertical right, horizontal lower, vertical left, horizontal upper ) And nine prediction modes corresponding to DC (average value) prediction were selectable (see FIGS. 22 and 23).
  • Scalable encoding refers to a technique for hierarchically encoding a layer that transmits a coarse image signal and a layer that transmits a fine image signal.
  • Typical attributes hierarchized in scalable coding are mainly the following three types. Spatial scalability: Spatial resolution or image size is layered. -Time scalability: Frame rate is layered. -SNR (Signal to Noise Ratio) scalability: SN ratio is hierarchized.
  • bit depth scalability and chroma format scalability are also discussed, although not yet adopted by the standard.
  • the prediction mode candidate set is equal between the prediction unit of the lower layer and the corresponding prediction unit of the upper layer, the prediction mode set in the lower layer can be reused in the upper layer.
  • the prediction mode cannot be simply reused because the set of prediction mode candidates is different. Such a situation becomes more conspicuous in HEVC in which the range of block sizes is expanded and the candidate set of prediction modes is diversified.
  • the intra prediction mode for the first prediction unit in the first layer of the image to be scalable decoded including the first layer and the second layer higher than the first layer.
  • the prediction set in the first prediction unit when the number of candidates is different from the number of intra prediction mode candidates for the second prediction unit corresponding to the first prediction unit in the second layer A mode setting unit that sets a prediction mode selected based on a mode to the second prediction unit, and a prediction that generates a prediction image of the second prediction unit according to the prediction mode set by the mode setting unit And an image processing apparatus.
  • the image processing apparatus can typically be realized as an image decoding apparatus that decodes a scalable encoded image.
  • intra prediction for a first prediction unit in the first layer of a scalable decoded image including a first layer and a second layer higher than the first layer.
  • the number of mode candidates is different from the number of intra prediction mode candidates for the second prediction unit corresponding to the first prediction unit in the second layer, it is set to the first prediction unit.
  • An image including: setting a prediction mode selected based on the prediction mode as the second prediction unit; and generating a prediction image of the second prediction unit according to the set prediction mode.
  • the intra of the first prediction unit in the first layer of the image to be scalable encoded including the first layer and the second layer higher than the first layer.
  • a prediction mode selected based on the predicted prediction mode is set to the second prediction unit, and a prediction image of the second prediction unit is set according to the prediction mode set by the mode setting unit.
  • An image processing apparatus including a predicting unit to be generated is provided.
  • the image processing apparatus can be typically realized as an image encoding apparatus that performs scalable encoding of an image.
  • the intra of the first prediction unit in the first layer of the image to be scalable encoded including the first layer and the second layer higher than the first layer.
  • Set as the first prediction unit when the number of prediction mode candidates is different from the number of intra prediction mode candidates for the second prediction unit corresponding to the first prediction unit in the second layer Setting a prediction mode selected based on the set prediction mode to the second prediction unit, and generating a prediction image of the second prediction unit according to the set prediction mode.
  • a mechanism is provided that can more efficiently encode the prediction mode of intra prediction in scalable encoding.
  • H. 3 is an explanatory diagram showing a set of prediction mode candidates for a luminance component of a prediction unit of 4 ⁇ 4 pixels in H.264 / AVC.
  • FIG. It is explanatory drawing which shows the set of the prediction mode candidates about the luminance component of a prediction unit of 8x8 pixel. It is explanatory drawing which shows the set of the prediction mode candidate about the luminance component of a prediction unit of 16x16 pixels.
  • FIG. 1 is a block diagram illustrating an example of a configuration of an image encoding device 10 according to an embodiment.
  • an image encoding device 10 includes an A / D (Analogue to Digital) conversion unit 11, a rearrangement buffer 12, a subtraction unit 13, an orthogonal transformation unit 14, a quantization unit 15, a lossless encoding unit 16, The accumulation buffer 17, rate control unit 18, inverse quantization unit 21, inverse orthogonal transform unit 22, addition unit 23, deblock filter 24, frame memory 25, selectors 26 and 27, motion search unit 30, and intra prediction unit 40 Prepare.
  • a / D Analogue to Digital
  • the A / D converter 11 converts an image signal input in an analog format into image data in a digital format, and outputs a series of digital image data to the rearrangement buffer 12.
  • the rearrangement buffer 12 rearranges images included in a series of image data input from the A / D conversion unit 11.
  • the rearrangement buffer 12 rearranges the images according to the GOP (Group of Pictures) structure related to the encoding process, and then outputs the rearranged image data to the subtraction unit 13, the motion search unit 30, and the intra prediction unit 40. To do.
  • GOP Group of Pictures
  • the subtraction unit 13 is supplied with image data input from the rearrangement buffer 12 and predicted image data input from the motion search unit 30 or the intra prediction unit 40 described later.
  • the subtraction unit 13 calculates prediction error data that is the difference between the image data input from the rearrangement buffer 12 and the predicted image data, and outputs the calculated prediction error data to the orthogonal transform unit 14.
  • the orthogonal transform unit 14 performs orthogonal transform on the prediction error data input from the subtraction unit 13.
  • the orthogonal transformation performed by the orthogonal transformation part 14 may be discrete cosine transformation (Discrete Cosine Transform: DCT) or Karoonen-Labe transformation, for example.
  • the orthogonal transform unit 14 outputs transform coefficient data acquired by the orthogonal transform process to the quantization unit 15.
  • the quantization unit 15 is supplied with transform coefficient data input from the orthogonal transform unit 14 and a rate control signal from the rate control unit 18 described later.
  • the quantizing unit 15 quantizes the transform coefficient data and outputs the quantized transform coefficient data (hereinafter referred to as quantized data) to the lossless encoding unit 16 and the inverse quantization unit 21. Further, the quantization unit 15 changes the bit rate of the quantized data input to the lossless encoding unit 16 by switching the quantization parameter (quantization scale) based on the rate control signal from the rate control unit 18.
  • the lossless encoding unit 16 generates an encoded stream by performing a lossless encoding process on the quantized data input from the quantization unit 15.
  • the lossless encoding by the lossless encoding unit 16 may be variable length encoding or arithmetic encoding, for example. Further, the lossless encoding unit 16 multiplexes information related to intra prediction or information related to inter prediction input from the selector 27 in the header region of the encoded stream. Then, the lossless encoding unit 16 outputs the generated encoded stream to the accumulation buffer 17.
  • the accumulation buffer 17 temporarily accumulates the encoded stream input from the lossless encoding unit 16 using a storage medium such as a semiconductor memory. Then, the accumulation buffer 17 outputs the accumulated encoded stream to a transmission unit (not shown) (for example, a communication interface or a connection interface with a peripheral device) at a rate corresponding to the bandwidth of the transmission path.
  • a transmission unit for example, a communication interface or a connection interface with a peripheral device
  • the rate control unit 18 monitors the free capacity of the accumulation buffer 17. Then, the rate control unit 18 generates a rate control signal according to the free capacity of the accumulation buffer 17 and outputs the generated rate control signal to the quantization unit 15. For example, the rate control unit 18 generates a rate control signal for reducing the bit rate of the quantized data when the free capacity of the storage buffer 17 is small. For example, when the free capacity of the accumulation buffer 17 is sufficiently large, the rate control unit 18 generates a rate control signal for increasing the bit rate of the quantized data.
  • the inverse quantization unit 21 performs an inverse quantization process on the quantized data input from the quantization unit 15. Then, the inverse quantization unit 21 outputs transform coefficient data acquired by the inverse quantization process to the inverse orthogonal transform unit 22.
  • the inverse orthogonal transform unit 22 restores the prediction error data by performing an inverse orthogonal transform process on the transform coefficient data input from the inverse quantization unit 21. Then, the inverse orthogonal transform unit 22 outputs the restored prediction error data to the addition unit 23.
  • the adding unit 23 generates decoded image data by adding the restored prediction error data input from the inverse orthogonal transform unit 22 and the predicted image data input from the motion search unit 30 or the intra prediction unit 40. . Then, the addition unit 23 outputs the generated decoded image data to the deblock filter 24 and the frame memory 25.
  • the deblocking filter 24 performs a filtering process for reducing block distortion that occurs during image coding.
  • the deblocking filter 24 removes block distortion by filtering the decoded image data input from the adding unit 23, and outputs the decoded image data after filtering to the frame memory 25.
  • the frame memory 25 stores the decoded image data input from the adder 23 and the decoded image data after filtering input from the deblock filter 24 using a storage medium.
  • the selector 26 reads out the decoded image data after filtering used for inter prediction from the frame memory 25 and supplies the read out decoded image data to the motion search unit 30 as reference image data.
  • the selector 26 reads out decoded image data before filtering used for intra prediction from the frame memory 25 and supplies the read decoded image data to the intra prediction unit 40 as reference image data.
  • the selector 27 In the inter prediction mode, the selector 27 outputs the prediction image data as a result of the inter prediction output from the motion search unit 30 to the subtraction unit 13 and outputs information related to the inter prediction to the lossless encoding unit 16. Further, in the intra prediction mode, the selector 27 outputs predicted image data as a result of the intra prediction output from the intra prediction unit 40 to the subtraction unit 13 and outputs information related to the intra prediction to the lossless encoding unit 16. . The selector 27 switches between the inter prediction mode and the intra prediction mode according to the size of the cost function value output from the motion search unit 30 and the intra prediction unit 40.
  • the motion search unit 30 performs inter prediction processing (interframe prediction processing) based on the image data to be encoded (original image data) input from the rearrangement buffer 12 and the decoded image data supplied via the selector 26. )I do. For example, the motion search unit 30 evaluates the prediction result in each prediction mode using a predetermined cost function. Next, the motion search unit 30 selects the prediction mode with the smallest cost function value, that is, the prediction mode with the highest compression rate, as the optimum prediction mode. In addition, the motion search unit 30 generates predicted image data according to the optimal prediction mode. Then, the motion search unit 30 outputs prediction mode information indicating the selected optimal prediction mode, information on inter prediction including motion vector information and reference image information, a cost function value, and predicted image data to the selector 27.
  • inter prediction processing interframe prediction processing
  • the intra prediction unit 40 performs an intra prediction process for each block set in the image based on the original image data input from the rearrangement buffer 12 and the decoded image data as reference image data supplied from the frame memory 25. I do. Then, the intra prediction unit 40 outputs information related to intra prediction including prediction mode information representing the optimal prediction mode, a cost function value, and predicted image data to the selector 27.
  • the number of prediction mode candidates that can be selected by the intra prediction unit 40 varies depending on the block size of the prediction unit. For example, when the angle intra prediction method described above is employed, the number of prediction mode candidates for each block size is as shown in Table 1 below.
  • the number of prediction mode candidates (Possible Intra Prediction Modes) is 17.
  • the 16 prediction modes except the prediction mode corresponding to DC prediction correspond to 16 types of prediction direction candidates (possible prediction directions) from the reference pixel to the prediction target pixel, respectively.
  • the number of prediction mode candidates is 34.
  • the 33 prediction modes except for the prediction mode corresponding to DC prediction correspond to 33 types of prediction direction candidates from the reference pixel to the prediction target pixel, respectively.
  • the block size is 16 ⁇ 16 pixels or 32 ⁇ 32 pixels, there are 34 prediction mode candidates and 33 prediction direction candidates.
  • the block size is 64 ⁇ 64 pixels, the number of prediction mode candidates is three.
  • the two prediction modes other than the prediction mode corresponding to DC prediction correspond to two types of prediction direction candidates (vertical direction and horizontal direction) from the reference pixel to the prediction target pixel, respectively.
  • the image encoding device 10 repeats the series of encoding processes described here for each of a plurality of layers of an image to be scalable encoded.
  • the layer that is encoded first is a layer that expresses the coarsest image, called a base layer.
  • the base layer coded stream may be decoded independently without decoding the other layer coded streams.
  • the layers other than the base layer are layers that represent a finer image, called an enhancement layer.
  • the enhancement layer coded stream is coded using information included in the base layer coded stream in order to increase coding efficiency. Accordingly, in order to reproduce the enhancement layer image, both the base layer and enhancement layer encoded streams are decoded. There may be three or more layers handled in scalable coding.
  • the lowest layer is the base layer, and the remaining layers are enhancement layers.
  • the higher enhancement layer encoded stream may be encoded and decoded using information contained in the lower enhancement layer or base layer encoded stream.
  • the layer on the dependent side is referred to as a lower layer, and the layer on the dependent side is referred to as an upper layer.
  • the mode buffer 44 of the intra prediction unit 40 shown in FIG. 1 is provided for temporarily storing lower layer prediction mode information.
  • the same prediction mode as the prediction mode set for the prediction unit of the lower layer may be set as it is for the corresponding prediction unit of the higher layer.
  • the block sizes of two corresponding prediction units are different from each other, and therefore the number of intra prediction mode candidates varies between layers. Can occur.
  • FIG. 2 shows three layers L1, L2, and L3 that are scalable coded as an example of spatial scalability.
  • Layer L1 is a base layer
  • layers L2 and L3 are enhancement layers.
  • the ratio of the spatial resolution of the layer L2 to the layer L1 is 2: 1.
  • the ratio of the spatial resolution of layer L3 to layer L1 is 4: 1.
  • the block size of the prediction unit B2 of the layer L2 is twice (in one side) the block size of the corresponding prediction unit B1 of the layer L1.
  • the block size of the prediction unit B3 of the layer L3 is twice the block size of the corresponding prediction unit B2 of the layer L2, and is four times the block size of the corresponding prediction unit B1 of the layer L1.
  • the intra prediction unit 40 of the image encoding device 10 expands or aggregates the prediction modes based on the prediction mode of the lower layer as described in detail in the next section. Predict higher layer prediction mode.
  • the lower layer prediction unit corresponding to the upper layer prediction unit is, for example, a lower layer prediction unit having a pixel corresponding to a pixel at a predetermined position (for example, upper left) of the upper layer prediction unit. It's okay. With such a definition, even when there is an upper layer prediction unit that integrates a plurality of lower layer prediction units, the lower layer prediction unit corresponding to the upper layer prediction unit can be uniquely determined. .
  • the technology according to the present disclosure is not limited to such an example, and is generally applicable in a situation where the number of intra prediction mode candidates is different between layers during scalable coding.
  • FIG. 3 is a block diagram illustrating an example of a detailed configuration of the intra prediction unit 40 of the image encoding device 10 illustrated in FIG. 1.
  • the intra prediction unit 40 includes a mode setting unit 41, a prediction unit 42, a mode determination unit 43, a mode buffer 44, and a parameter generation unit 45.
  • the mode setting unit 41 sequentially sets each of a plurality of prediction mode candidates as one or more prediction units in a coding unit.
  • the prediction unit 42 generates a prediction image of each prediction unit using the reference image data input from the frame memory 25 according to the prediction mode candidates set by the mode setting unit 41.
  • the mode determination unit 43 calculates a cost function value for each prediction mode candidate based on the original image data input from the rearrangement buffer 12 and the predicted image data input from the prediction unit 42. And the mode determination part 43 determines arrangement
  • the mode buffer 44 temporarily stores prediction mode information representing the determined optimum prediction mode using a storage medium for processing in an upper layer.
  • the parameter generation unit 45 generates a parameter representing the arrangement of prediction units and the prediction mode determined to be optimal by the mode determination unit 43. Then, the mode determination unit 43 outputs the information related to intra prediction including the parameters generated by the parameter generation unit 45, the cost function value, and the predicted image data to the selector 27.
  • FIG. 4 is an explanatory diagram for explaining prediction direction candidates that can be selected when the angle intra prediction method is used in such intra prediction.
  • the pixel P1 illustrated in FIG. 4 is a prediction target pixel.
  • the shaded pixels around the block to which the pixel P1 belongs are reference pixels.
  • 17 types of prediction directions (corresponding to prediction modes) connecting the reference pixel and the prediction target pixel, which are indicated by solid lines (both thick and thin lines) in the figure, Can be selected (in addition to DC prediction).
  • the block size is 8 ⁇ 8 pixels, 16 ⁇ 16 pixels, or 32 ⁇ 32 pixels, prediction types corresponding to 33 types of prediction directions (shown by dotted lines and solid lines (both thick lines and thin lines)) are shown.
  • Mode can be selected (in addition to DC prediction and planar prediction).
  • the block size is 64 ⁇ 64 pixels, two types of prediction directions (corresponding to prediction modes) indicated by bold lines in the figure can be selected (in addition to DC prediction).
  • the mode setting unit 41 illustrated in FIG. 3 sets a plurality of prediction mode candidates in each prediction unit according to the size of each prediction unit.
  • the resolution of the angle in the prediction direction is high.
  • the prediction unit 42 first calculates a reference pixel value with 1/8 pixel accuracy as shown in FIG. 5, and calculates a predicted pixel value according to each candidate of the prediction mode using the calculated reference pixel value.
  • Enhancement layer intra prediction processing is mainly classified into three types: prediction direction reuse, prediction direction extension, and prediction direction aggregation.
  • the reuse of the prediction direction may be performed when the number of prediction mode candidates for the lower layer is equal to the number of prediction mode candidates for the upper layer.
  • the extension of the prediction direction may be performed when the number of prediction mode candidates in the lower layer is smaller than the number of prediction mode candidates in the upper layer.
  • the aggregation of the prediction directions may be performed when the number of prediction mode candidates in the lower layer is larger than the number of prediction mode candidates in the upper layer.
  • the present invention is not limited to this example. For example, when the number of lower layer prediction mode candidates is smaller than the number of upper layer prediction mode candidates, prediction direction reuse is performed instead of prediction direction extension. May be.
  • the mode setting unit 41 stores the prediction mode stored in the mode buffer 44. Reuse the prediction mode indicated by the information. That is, in this case, the mode setting unit 41 sets the same prediction mode as the prediction mode set for the prediction unit corresponding to the lower layer for each prediction unit of the upper layer.
  • the prediction unit 42 generates a prediction image of each prediction unit according to the one prediction mode set by the mode setting unit 41.
  • the determination of the optimal prediction mode based on the cost function value in the mode determination unit 43 is omitted (the cost function value can be calculated).
  • the mode buffer 44 stores prediction mode information indicating the prediction mode set by the mode setting unit 41 when there is a further upper layer.
  • the mode setting unit 41 adds the lower layer prediction mode to each prediction unit in the upper layer.
  • Each candidate of the prediction mode selected based on the prediction mode set to the corresponding prediction unit is set sequentially.
  • partial images at the same position are correlated with each other between corresponding blocks of two layers that differ only in spatial resolution. Therefore, there is a high possibility that the optimal prediction mode in a block in a lower layer is optimal in a corresponding block in an upper layer.
  • the optimal prediction mode in a block in a lower layer is optimal in a corresponding block in an upper layer.
  • the search range of the prediction mode may be limited to several prediction directions in the vicinity of the prediction direction set in the lower layer in order to reduce the processing cost.
  • the prediction unit 42 generates a prediction image of each prediction unit using the reference image data input from the frame memory 25 according to each prediction mode candidate set by the mode setting unit 41.
  • the mode determination unit 43 calculates a cost function value for each prediction mode candidate based on the original image data and the predicted image data input from the prediction unit 42. And the mode determination part 43 determines the optimal prediction mode based on the calculated cost function value.
  • the mode buffer 44 stores prediction mode information indicating the optimal prediction mode determined by the mode determination unit 43 when there is a further upper layer.
  • the parameter generation unit 45 is encoded in accordance with the prediction direction difference between the prediction mode set in the lower layer and the optimal prediction mode determined by the mode determination unit 43, as illustrated in FIG.
  • the parameter P1 is generated.
  • the prediction direction DL is the prediction direction of the prediction mode set in the prediction unit B1.
  • the prediction direction candidates of the prediction mode that can be set in the prediction unit B2 include prediction directions D U0 , D U1 , D U2 , D U3 , D U4 .
  • the difference in angle between two adjacent prediction direction candidates is ⁇ .
  • the parameter P1 is encoded with a smaller code number as the absolute value of the difference in the prediction direction is smaller. For example, if the optimal prediction mode set in the prediction unit B2 is the prediction mode representing the prediction direction DU0 , the angle difference is zero, and the parameter P1 is encoded with the code number “0”. If the optimal prediction mode set in the prediction unit B2 is a prediction mode representing the prediction direction D U1 or D U2 , the difference in angle is ⁇ or ⁇ , and the parameter P1 has a code number “1” or “2”. It is encoded with.
  • the optimal prediction mode set in the prediction unit B2 is the prediction mode representing the prediction direction D U3 or D U4
  • the angle difference is 2 ⁇ or ⁇ 2 ⁇
  • the parameter P1 has the code number “3” or “4”. It is encoded with.
  • the smaller code number is mapped to a shorter codeword in the lossless encoding unit 16. Therefore, with respect to the parameter P1, in this way, by using a smaller code number as the difference (in angle) in the prediction direction is smaller, a prediction mode having a higher occurrence frequency in the higher layer is mapped to a shorter codeword and encoded. Efficiency can be increased.
  • a smaller code number is assigned to the direction in which the prediction direction is rotated clockwise from the lower layer to the upper layer among the differences in the prediction direction that are different only in positive and negative.
  • a smaller code number may be assigned to one of the predefined prediction modes. Instead, as illustrated in FIGS. 7A and 7B, it is dynamic whether the prediction direction of the upper layer is closer to a specific direction (for example, the vertical direction or the horizontal direction) when any prediction mode is selected. A smaller code number may be assigned to a prediction direction closer to the specific direction.
  • prediction direction candidates D U0 , D U1 , D U2 ... That can be set as a prediction unit of a higher layer of the image Im1 are shown.
  • Prediction direction of the prediction mode set in the lower layer is a prediction direction D L.
  • the aspect ratio (vertical / horizontal) V / H of the image Im1 is smaller than 1 (that is, the size in the horizontal direction is larger than the size in the vertical direction). In such a horizontally long image, prediction accuracy tends to be higher when intra prediction is performed in a prediction direction closer to the horizontal direction.
  • the prediction direction D U1 is closer to the horizontal direction than the prediction direction D U2 . Therefore, in the table on the right side of FIG. 7A, the parameter P1 is encoded with the code number “1” for the prediction mode representing the prediction direction D U1, and the parameter P1 is the code number “2” for the prediction mode representing the prediction direction D U2. Encoded.
  • the example of FIG. 7A the prediction direction D U1 is closer to the horizontal direction than the prediction direction D U2 .
  • the aspect ratio V / H of the image Im2 is larger than 1 (that is, the horizontal size is smaller than the vertical size). Therefore, in this case, for two prediction modes having the same absolute value of the difference between prediction directions, it is desirable that a smaller code number is assigned to the higher layer in the direction in which the prediction direction of the higher layer is closer to the vertical direction. Therefore, in the table on the right side of FIG. 7B, the parameter P1 is encoded with the code number “1” for the prediction mode representing the prediction direction D U2, and the parameter P1 is the code number “2” for the prediction mode representing the prediction direction D U1. Encoded.
  • the mapping between the angle difference and the code number for the parameter P1 may be adaptively determined according to the aspect ratio of the image to be encoded.
  • the mode setting unit 41 assigns lower layer layer prediction units to each upper layer prediction unit.
  • the prediction mode selected based on the prediction mode set to the corresponding prediction unit is set.
  • the prediction mode that is optimal in a prediction unit having a lower layer among two layers that differ only in spatial resolution is likely to be optimal in the corresponding prediction unit of the upper layer.
  • the mode setting unit 41 does not simply reuse the prediction mode, but predicts the optimal prediction mode in the upper layer from the prediction mode set in the lower layer.
  • the prediction mode predicted to be optimal in this case is a prediction mode of the upper layer that represents the prediction direction closest to the prediction direction of the prediction mode set in the lower layer.
  • the mode setting unit 41 can set the prediction mode representing the prediction direction DU1 as the prediction unit B2.
  • the prediction directions D U1 and D U2 are equidistant from the prediction direction D L of the lower layer.
  • the mode setting part 41 can set the prediction mode showing average value (DC) prediction to the prediction unit B2 as one method.
  • the mode setting unit 41 should set the prediction unit of the upper layer according to a pre-defined condition instead of selecting the average value prediction as in the example of FIG.
  • a prediction mode may be selected.
  • the condition defined in advance may be, for example, a condition that the prediction direction is rotated in a predetermined rotation direction (clockwise or counterclockwise).
  • the prediction direction D U1 that is derived by rotating the prediction direction D L clockwise may be set as the prediction unit B2.
  • the predefined condition may be, for example, a condition that a prediction direction with a smaller code number is selected.
  • the image data of the upper layer to be scalable encoded can be specially specified. It is possible to correctly decode without requiring a large parameter.
  • the prediction unit 42 generates a prediction image of each prediction unit using the reference image data input from the frame memory 25 according to the prediction mode set by the mode setting unit 41. In this case, determination of the optimal prediction mode based on the cost function value in the mode determination unit 43 is omitted (a cost function value can be calculated).
  • the mode buffer 44 stores prediction mode information indicating the prediction mode set by the mode setting unit 41 when there is a further upper layer.
  • the optimum prediction mode may be searched even when the prediction modes are aggregated.
  • the mode setting unit 41 applies each prediction unit of the upper layer when there are a plurality of prediction modes representing the prediction direction closest to the prediction direction of the lower layer in the prediction mode candidates for the upper layer. Each of a plurality of (usually two) prediction mode candidates is sequentially set.
  • the prediction unit 42 generates a prediction image for each prediction unit using the reference image data input from the frame memory 25 according to each prediction mode candidate set by the mode setting unit 41.
  • the mode determination unit 43 calculates a cost function value for each prediction mode candidate based on the original image data and the predicted image data input from the prediction unit 42. And the mode determination part 43 determines the optimal prediction mode based on the calculated cost function value.
  • the mode buffer 44 stores prediction mode information indicating the optimal prediction mode determined by the mode determination unit 43 when there is a further upper layer.
  • the parameter generation unit 45 can generate the parameter P2 as illustrated in FIG. 10 that identifies the optimal prediction mode determined by the mode determination unit 43.
  • the prediction direction DL is the prediction direction of the prediction mode set in the prediction unit B1 of the lower layer.
  • the prediction direction candidates in the prediction mode that can be set in the prediction unit B2 include the prediction directions D Ua and D Ub and do not include the prediction direction D L.
  • the prediction directions D Ua and D Ub are equidistant from the prediction direction D L of the lower layer.
  • the parameter generation unit 45 can generate a 1-bit parameter P2 (encoded with the code number “0” or “1”) that represents the optimum prediction mode determined by the mode determination unit 43.
  • the parameters generated by the parameter generation unit 45 are each encoded by the lossless encoding unit 16 as one piece of information related to intra prediction, and are decoded in the header area of the encoded stream. Can be transmitted.
  • the mode setting unit 41 uses the prediction mode (prediction direction) optimum for the prediction target block from the prediction mode (prediction direction) set in the reference block in order to suppress an increase in the code amount due to encoding the prediction mode information. ) May be estimated.
  • the estimated prediction mode hereinafter referred to as the estimated prediction mode
  • only information indicating that the prediction mode can be estimated is predicted. It can be encoded as mode information.
  • the information indicating that the prediction mode can be estimated is, for example, H.264. This corresponds to “MostProbableMode” in H.264 / AVC.
  • the mode setting unit 41 can refer to, for example, the prediction unit of the lower layer corresponding to the prediction unit of the upper layer in determining the MostProbableMode.
  • the mode setting unit 41 determines the MostProbableMode after converting the prediction mode of the prediction unit of the lower layer into the prediction mode in the prediction mode of the upper layer according to the above-described concept of expansion and aggregation of the prediction mode. For example, as shown in FIG.
  • a parameter indicating that the prediction mode can be estimated is generated by the parameter generation unit 45, and the generated parameter is generated by the lossless encoding unit 16. Can be encoded.
  • the prediction mode of the lower layer is also referred to when determining the MostProbableMode, so that the prediction mode with high accuracy using the correlation of images between layers can be obtained. Estimation is possible.
  • FIG. 12 is a flowchart illustrating an example of the flow of an intra prediction process performed by the intra prediction unit 40 having the configuration illustrated in FIG.
  • FIG. 13 is a flowchart illustrating an example of a detailed flow of the prediction mode expansion process.
  • 14A and 14B are flowcharts illustrating a first example and a second example of a detailed flow of a prediction mode aggregation process, respectively.
  • the intra prediction unit 40 performs a base layer intra prediction process (step S100). As a result, the arrangement of the prediction units within each coding unit is determined, and an optimal prediction mode is set for each prediction unit in the lower layer.
  • the mode buffer 44 buffers prediction mode information representing the optimal prediction mode of each prediction unit.
  • steps S110 to S160 are enhancement layer intra prediction processes. Of these processes, the processes of steps S110 to S150 are repeated for each block (each prediction unit) of each enhancement layer.
  • “upper layer” is a prediction target layer
  • “lower layer” is a lower layer of the prediction target layer.
  • step S120 the mode setting unit 41 sets the same prediction mode as the prediction mode set for the corresponding PU of the lower layer as the attention PU (that is, the prediction mode is reused). And the prediction part 42 produces
  • step S130 the prediction mode extension process illustrated in FIG. 13 is performed.
  • step S140 prediction mode aggregation processing illustrated in FIGS. 14A and 14B is performed.
  • a prediction image of the attention PU is generated by the prediction unit 42 according to the prediction mode candidate set as the attention PU by the mode setting unit 41 (step S132).
  • the mode determination unit 43 calculates a cost function value using the predicted image data and the original image data (Step S133).
  • the mode determination unit 43 selects an optimal prediction mode by comparing the cost function values calculated for a plurality of prediction mode candidates (step S134).
  • the parameter generation unit 45 generates a parameter P1 corresponding to the difference in the prediction direction between layers for specifying the selected optimal prediction mode (step S135).
  • the mode setting unit 41 determines whether there are a plurality of prediction directions closest to the prediction direction of the corresponding PU of the lower layer in the prediction direction candidates of the upper layer. Is determined (step S141).
  • the mode setting unit 41 sets an average value (DC) prediction mode or a prediction mode selected according to a pre-defined condition as the attention PU (step). S142).
  • the mode setting unit 41 sets a prediction mode representing the one prediction direction as the attention PU (step S143).
  • the prediction part 42 produces
  • the mode setting unit 41 determines whether there are a plurality of prediction directions closest to the prediction direction of the corresponding PU of the lower layer in the prediction direction candidates of the upper layer. Is determined (step S141). The process performed when only one closest prediction direction exists is the same as in the first example of FIG. 14A (steps S143 and S144). On the other hand, when there are a plurality of closest prediction directions, the processes of steps S146 and S147 are repeated for each of the plurality of prediction directions (step S145). First, a prediction image of the attention PU is generated by the prediction unit 42 in accordance with a prediction mode candidate representing each prediction direction (step S146).
  • the mode determination unit 43 calculates a cost function value using the predicted image data and the original image data (step S147). When the loop ends, the mode determination unit 43 selects an optimal prediction mode by comparing the cost function values calculated for a plurality of prediction mode candidates (step S148). Then, the parameter generation unit 45 generates a parameter P2 for specifying the selected optimal prediction mode (step S149).
  • step S110 After the prediction mode is set for the attention PU in step S120, S130, or S140 and the prediction image is generated, if an unprocessed PU remains in the prediction target layer, the process returns to step S110 (step S110). S150). On the other hand, if there is no unprocessed PU remaining, it is further determined whether there is a remaining layer (higher layer) (step S160). Here, if there is a remaining layer, The processes after step S110 are repeated with the previous prediction target layer as the lower layer and the next layer as the upper layer. The prediction mode information is buffered by the mode buffer 44. If there are no remaining layers, the intra prediction process in FIG. 12 ends. Note that the prediction image data generated here and information on inter prediction (which may include parameters P1 and P2) are output from the mode determination unit 43 to the subtraction unit 13 and the lossless encoding unit 16 via the selector 27, respectively. obtain.
  • FIG. 15 is a block diagram illustrating an example of the configuration of the image decoding device 60 according to an embodiment.
  • the image decoding device 60 includes an accumulation buffer 61, a lossless decoding unit 62, an inverse quantization unit 63, an inverse orthogonal transform unit 64, an addition unit 65, a deblock filter 66, a rearrangement buffer 67, a D / A A (Digital to Analogue) conversion unit 68, a frame memory 69, selectors 70 and 71, a motion compensation unit 80, and an intra prediction unit 90 are provided.
  • the accumulation buffer 61 temporarily accumulates the encoded stream input via the transmission path using a storage medium.
  • the lossless decoding unit 62 decodes the encoded stream input from the accumulation buffer 61 according to the encoding method used at the time of encoding. In addition, the lossless decoding unit 62 decodes information multiplexed in the header area of the encoded stream.
  • the information multiplexed in the header area of the encoded stream can include, for example, the information related to inter prediction and the information related to intra prediction described above.
  • the lossless decoding unit 62 outputs information related to inter prediction to the motion compensation unit 80. Further, the lossless decoding unit 62 outputs information related to intra prediction to the intra prediction unit 90.
  • the inverse quantization unit 63 performs inverse quantization on the quantized data decoded by the lossless decoding unit 62.
  • the inverse orthogonal transform unit 64 generates prediction error data by performing inverse orthogonal transform on the transform coefficient data input from the inverse quantization unit 63 according to the orthogonal transform method used at the time of encoding. Then, the inverse orthogonal transform unit 64 outputs the generated prediction error data to the addition unit 65.
  • the addition unit 65 adds the prediction error data input from the inverse orthogonal transform unit 64 and the prediction image data input from the selector 71 to generate decoded image data. Then, the addition unit 65 outputs the generated decoded image data to the deblock filter 66 and the frame memory 69.
  • the deblock filter 66 removes block distortion by filtering the decoded image data input from the adder 65, and outputs the filtered decoded image data to the rearrangement buffer 67 and the frame memory 69.
  • the rearrangement buffer 67 generates a series of time-series image data by rearranging the images input from the deblocking filter 66. Then, the rearrangement buffer 67 outputs the generated image data to the D / A conversion unit 68.
  • the D / A converter 68 converts the digital image data input from the rearrangement buffer 67 into an analog image signal. Then, the D / A conversion unit 68 displays an image by outputting an analog image signal to a display (not shown) connected to the image decoding device 60, for example.
  • the frame memory 69 stores the decoded image data before filtering input from the adding unit 65 and the decoded image data after filtering input from the deblocking filter 66 using a storage medium.
  • the selector 70 switches the output destination of the image data from the frame memory 69 between the motion compensation unit 80 and the intra prediction unit 90 for each block in the image according to the mode information acquired by the lossless decoding unit 62. .
  • the selector 70 outputs the decoded image data after filtering supplied from the frame memory 69 to the motion compensation unit 80 as reference image data.
  • the selector 70 outputs the decoded image data before filtering supplied from the frame memory 69 to the intra prediction unit 90 as reference image data.
  • the selector 71 switches the output source of the predicted image data to be supplied to the addition unit 65 between the motion compensation unit 80 and the intra prediction unit 90 according to the mode information acquired by the lossless decoding unit 62. For example, when the inter prediction mode is designated, the selector 71 supplies the predicted image data output from the motion compensation unit 80 to the adding unit 65. In addition, when the intra prediction mode is designated, the selector 71 supplies the predicted image data output from the intra prediction unit 90 to the adding unit 65.
  • the motion compensation unit 80 performs motion compensation processing based on the inter prediction information input from the lossless decoding unit 62 and the reference image data from the frame memory 69 to generate predicted image data. Then, the motion compensation unit 80 outputs the generated predicted image data to the selector 71.
  • the intra prediction unit 90 performs intra prediction processing based on the information related to intra prediction input from the lossless decoding unit 62 and the reference image data from the frame memory 69, and generates predicted image data.
  • the number of prediction mode candidates that can be selected by the intra prediction unit 90 differs depending on the block size of the prediction unit. For example, when the angle intra prediction method described above is employed, the number of prediction mode candidates for each block size is as shown in Table 1 above. Then, the intra prediction unit 90 outputs the generated predicted image data to the selector 71. Such intra prediction processing by the intra prediction unit 90 will be described in detail later.
  • the image decoding device 60 repeats the series of decoding processes described here for each of a plurality of layers of a scalable encoded image.
  • the layer that is decoded first is the base layer. After the base layer is decoded, one or more enhancement layers are decoded.
  • decoding the enhancement layer information obtained by decoding the lower layer, which is the base layer or another enhancement layer, is used.
  • prediction of the prediction mode of the upper layer based on the prediction mode of the lower layer is performed in the prediction unit.
  • the prediction of the prediction mode here may include the reuse of the prediction mode, the expansion of the prediction mode, and the aggregation of the prediction modes described in detail above.
  • the mode buffer 93 of the intra prediction unit 90 illustrated in FIG. 15 is provided to temporarily store the prediction mode information of the lower layer in prediction mode prediction.
  • FIG. 16 is a block diagram illustrating an example of a detailed configuration of the intra prediction unit 90 of the image decoding device 60 illustrated in FIG. 15.
  • the intra prediction unit 90 includes a parameter acquisition unit 91, a mode setting unit 92, a mode buffer 93, and a prediction unit 94.
  • the parameter acquisition unit 91 acquires information about intra prediction decoded by the lossless decoding unit 62.
  • Information related to intra prediction for the base layer may include, for example, information for specifying the arrangement of prediction units in each coding unit and prediction mode information for each prediction unit.
  • the mode setting unit 92 arranges a prediction unit in each coding unit based on the information acquired by the parameter acquisition unit 91, and further sets a prediction mode for each prediction unit.
  • the mode buffer 93 temporarily stores prediction mode information indicating the prediction mode set for each prediction unit.
  • the prediction unit 94 generates a prediction image for each prediction unit using the reference image data input from the frame memory 69 according to the prediction mode set by the mode setting unit 92. Then, the prediction unit 94 outputs the predicted image data to the addition unit 65.
  • Enhancement layer intra prediction processing is mainly classified into three types: prediction direction reuse, prediction direction extension, and prediction direction aggregation.
  • the mode setting unit 92 reuses the prediction mode indicated by the prediction mode information stored in the mode buffer 93. That is, in this case, the mode setting unit 92 sets the same prediction mode as the prediction mode set in the corresponding prediction unit in the lower layer for each prediction unit in the upper layer.
  • the prediction unit 94 generates a prediction image of each prediction unit according to the prediction mode set by the mode setting unit 92.
  • the mode buffer 93 stores prediction mode information indicating the prediction mode set by the mode setting unit 92 when there is a further upper layer.
  • the parameter acquisition unit 91 selects the lower layer corresponding to the prediction unit of the upper layer.
  • the above-described parameter P1 that is encoded according to the difference in the prediction direction from the prediction unit is acquired.
  • the parameter P1 is a parameter encoded with a smaller code number as the absolute value of the prediction direction difference is smaller. For example, when the codeword corresponding to the parameter P1 is the shortest codeword, the codeword is mapped to code number “0” zero by the lossless decoding unit 62 shown in FIG. Then, according to the code number table illustrated in FIG. 6, FIG. 7A, or FIG.
  • the code number “0” is interpreted as indicating that the difference in the prediction direction is zero.
  • the mode setting unit 92 may set a prediction mode representing the same prediction direction as the prediction mode set in the corresponding prediction unit of the lower layer as the prediction unit of the upper layer.
  • the mode setting unit 92 sets the prediction mode representing the prediction direction selected according to the prediction direction difference corresponding to the code number to the prediction of the upper layer. Can be set to units.
  • the sign of the difference in the prediction direction may be interpreted according to the aspect ratio of the decoded image.
  • the prediction unit 94 generates a prediction image of each prediction unit according to the prediction mode set by the mode setting unit 92.
  • the mode buffer 93 stores prediction mode information indicating the prediction mode set by the mode setting unit 92 when there is a further upper layer.
  • the parameter acquisition unit 91 may acquire an additional parameter P2. Or additional parameters may not be obtained.
  • the mode setting unit 92 sets the prediction mode selected based on only the prediction mode set in the corresponding prediction unit in the lower layer as the prediction unit in the upper layer.
  • the prediction mode set for the prediction unit of the upper layer is a prediction mode that represents the prediction direction closest to the prediction direction of the corresponding prediction unit of the lower layer.
  • the mode setting unit 92 may set the prediction mode representing the average value prediction as the prediction unit of the upper layer. Such a method can be employed, for example, when the block size of the upper layer is 64 ⁇ 64 pixels.
  • the mode setting unit 92 may select a prediction mode that is set as a prediction unit of a higher layer according to a predefined condition.
  • the predefined condition may be, for example, a condition that the prediction direction is rotated in a predetermined rotation direction, or a condition that a smaller code number is selected.
  • the parameter acquisition unit 91 acquires the parameter P2.
  • the mode setting unit 92 predicts the prediction mode specified by the parameter P2 from the two prediction modes representing the prediction direction closest to the prediction direction of the prediction mode set in the corresponding prediction unit of the lower layer. Set to unit.
  • the prediction unit 94 generates a prediction image of each prediction unit according to the prediction mode set by the mode setting unit 92, as in the prediction direction expansion process.
  • the mode buffer 93 stores prediction mode information indicating the prediction mode set by the mode setting unit 92 when there is a further upper layer.
  • the mode setting unit 92 sets the prediction mode estimated by the above MostProbableMode to the prediction unit. May be.
  • MostProbableMode can be determined based not only on the left and upper reference blocks but also on the prediction mode set in the corresponding prediction unit of the lower layer.
  • the mode setting unit 92 determines the MostProbableMode after converting the prediction mode of the prediction unit of the lower layer into the prediction mode in the prediction mode candidate of the upper layer according to the above-described concept of expansion and aggregation of the prediction mode.
  • FIG. 17 is a flowchart illustrating an example of the flow of intra prediction processing by the intra prediction unit 90 having the configuration illustrated in FIG.
  • the intra prediction unit 90 performs a base layer intra prediction process (step S200). As a result, a prediction image of the base layer is generated, and prediction mode information indicating the prediction mode set for each prediction unit is buffered by the mode buffer 93.
  • steps S210 to S270 is an enhancement layer intra prediction process. Of these processes, the processes of steps S210 to S260 are repeated for each block (each prediction unit) of each enhancement layer.
  • “upper layer” is a prediction target layer
  • “lower layer” is a lower layer of the prediction target layer.
  • step S220 the mode setting unit 92 sets the same prediction mode as the prediction mode set in the corresponding PU of the lower layer as the attention PU (that is, the prediction mode is reused) (step S220).
  • step S230 the mode setting unit 92 sets the prediction mode selected based on the prediction mode set for the PU corresponding to the lower layer and the parameter P1 acquired by the parameter acquisition unit 91 as the attention PU (step S230). S230).
  • step S240 the mode setting unit 92 selects the prediction mode set for the PU corresponding to the lower layer, and the prediction mode selected based on the parameter P2 when the parameter P2 is encoded, The attention PU is set (step S240).
  • the prediction unit 94 generates a predicted image of the attention PU using the reference image data input from the frame memory 69 in accordance with the prediction mode set by the mode setting unit 92 (step S250).
  • step S260 If the unprocessed PU remains in the prediction target layer after the predicted image of the attention PU is generated, the process returns to step S210 (step S260). On the other hand, if there is no unprocessed PU remaining, it is determined whether there is a remaining layer (higher layer) (step S270). Here, if there is a remaining layer, The processes after step S210 are repeated with the previous prediction target layer as the lower layer and the next layer as the upper layer. The prediction mode information is buffered by the mode buffer 93. If there are no remaining layers, the intra prediction process in FIG. 17 ends. Note that the predicted image data generated here is output to the addition unit 65 via the selector 71.
  • the image encoding device 10 and the image decoding device 60 are a transmitter or a receiver in satellite broadcasting, cable broadcasting such as cable TV, distribution on the Internet, and distribution to terminals by cellular communication
  • the present invention can be applied to various electronic devices such as a recording device that records an image on a medium such as an optical disk, a magnetic disk, and a flash memory, or a playback device that reproduces an image from these storage media.
  • a recording device that records an image on a medium such as an optical disk, a magnetic disk, and a flash memory
  • a playback device that reproduces an image from these storage media.
  • FIG. 18 shows an example of a schematic configuration of a television apparatus to which the above-described embodiment is applied.
  • the television apparatus 900 includes an antenna 901, a tuner 902, a demultiplexer 903, a decoder 904, a video signal processing unit 905, a display unit 906, an audio signal processing unit 907, a speaker 908, an external interface 909, a control unit 910, a user interface 911, And a bus 912.
  • Tuner 902 extracts a signal of a desired channel from a broadcast signal received via antenna 901, and demodulates the extracted signal. Then, the tuner 902 outputs the encoded bit stream obtained by the demodulation to the demultiplexer 903. In other words, the tuner 902 serves as a transmission unit in the television apparatus 900 that receives an encoded stream in which an image is encoded.
  • the demultiplexer 903 separates the video stream and audio stream of the viewing target program from the encoded bit stream, and outputs each separated stream to the decoder 904. In addition, the demultiplexer 903 extracts auxiliary data such as EPG (Electronic Program Guide) from the encoded bit stream, and supplies the extracted data to the control unit 910. Note that the demultiplexer 903 may perform descrambling when the encoded bit stream is scrambled.
  • EPG Electronic Program Guide
  • the decoder 904 decodes the video stream and audio stream input from the demultiplexer 903. Then, the decoder 904 outputs the video data generated by the decoding process to the video signal processing unit 905. In addition, the decoder 904 outputs audio data generated by the decoding process to the audio signal processing unit 907.
  • the video signal processing unit 905 reproduces the video data input from the decoder 904 and causes the display unit 906 to display the video.
  • the video signal processing unit 905 may cause the display unit 906 to display an application screen supplied via a network.
  • the video signal processing unit 905 may perform additional processing such as noise removal on the video data according to the setting.
  • the video signal processing unit 905 may generate a GUI (Graphical User Interface) image such as a menu, a button, or a cursor, and superimpose the generated image on the output image.
  • GUI Graphic User Interface
  • the display unit 906 is driven by a drive signal supplied from the video signal processing unit 905, and displays a video or an image on a video screen of a display device (for example, a liquid crystal display, a plasma display, or an OLED).
  • a display device for example, a liquid crystal display, a plasma display, or an OLED.
  • the audio signal processing unit 907 performs reproduction processing such as D / A conversion and amplification on the audio data input from the decoder 904, and outputs audio from the speaker 908.
  • the audio signal processing unit 907 may perform additional processing such as noise removal on the audio data.
  • the external interface 909 is an interface for connecting the television apparatus 900 to an external device or a network.
  • a video stream or an audio stream received via the external interface 909 may be decoded by the decoder 904. That is, the external interface 909 also has a role as a transmission unit in the television apparatus 900 that receives an encoded stream in which an image is encoded.
  • the control unit 910 has a processor such as a CPU (Central Processing Unit) and a memory such as a RAM (Random Access Memory) and a ROM (Read Only Memory).
  • the memory stores a program executed by the CPU, program data, EPG data, data acquired via a network, and the like.
  • the program stored in the memory is read and executed by the CPU when the television device 900 is activated, for example.
  • the CPU controls the operation of the television device 900 according to an operation signal input from the user interface 911, for example, by executing the program.
  • the user interface 911 is connected to the control unit 910.
  • the user interface 911 includes, for example, buttons and switches for the user to operate the television device 900, a remote control signal receiving unit, and the like.
  • the user interface 911 detects an operation by the user via these components, generates an operation signal, and outputs the generated operation signal to the control unit 910.
  • the bus 912 connects the tuner 902, the demultiplexer 903, the decoder 904, the video signal processing unit 905, the audio signal processing unit 907, the external interface 909, and the control unit 910 to each other.
  • the decoder 904 has the function of the image decoding apparatus 60 according to the above-described embodiment. Thereby, the enhancement layer image data encoded more efficiently can be decoded in the scalable decoding of the image in the television apparatus 900.
  • FIG. 19 shows an example of a schematic configuration of a mobile phone to which the above-described embodiment is applied.
  • a mobile phone 920 includes an antenna 921, a communication unit 922, an audio codec 923, a speaker 924, a microphone 925, a camera unit 926, an image processing unit 927, a demultiplexing unit 928, a recording / reproducing unit 929, a display unit 930, a control unit 931, an operation A portion 932 and a bus 933.
  • the antenna 921 is connected to the communication unit 922.
  • the speaker 924 and the microphone 925 are connected to the audio codec 923.
  • the operation unit 932 is connected to the control unit 931.
  • the bus 933 connects the communication unit 922, the audio codec 923, the camera unit 926, the image processing unit 927, the demultiplexing unit 928, the recording / reproducing unit 929, the display unit 930, and the control unit 931 to each other.
  • the mobile phone 920 has various operation modes including a voice call mode, a data communication mode, a shooting mode, and a videophone mode, and is used for sending and receiving voice signals, sending and receiving e-mail or image data, taking images, and recording data. Perform the action.
  • the analog voice signal generated by the microphone 925 is supplied to the voice codec 923.
  • the audio codec 923 converts an analog audio signal into audio data, A / D converts the converted audio data, and compresses it. Then, the audio codec 923 outputs the compressed audio data to the communication unit 922.
  • the communication unit 922 encodes and modulates the audio data and generates a transmission signal. Then, the communication unit 922 transmits the generated transmission signal to a base station (not shown) via the antenna 921. In addition, the communication unit 922 amplifies a radio signal received via the antenna 921 and performs frequency conversion to acquire a received signal.
  • the communication unit 922 demodulates and decodes the received signal to generate audio data, and outputs the generated audio data to the audio codec 923.
  • the audio codec 923 expands the audio data and performs D / A conversion to generate an analog audio signal. Then, the audio codec 923 supplies the generated audio signal to the speaker 924 to output audio.
  • the control unit 931 generates character data constituting the e-mail in response to an operation by the user via the operation unit 932.
  • the control unit 931 causes the display unit 930 to display characters.
  • the control unit 931 generates e-mail data in response to a transmission instruction from the user via the operation unit 932, and outputs the generated e-mail data to the communication unit 922.
  • the communication unit 922 encodes and modulates email data and generates a transmission signal. Then, the communication unit 922 transmits the generated transmission signal to a base station (not shown) via the antenna 921.
  • the communication unit 922 amplifies a radio signal received via the antenna 921 and performs frequency conversion to acquire a received signal.
  • the communication unit 922 demodulates and decodes the received signal to restore the email data, and outputs the restored email data to the control unit 931.
  • the control unit 931 displays the content of the electronic mail on the display unit 930 and stores the electronic mail data in the storage medium of the recording / reproducing unit 929.
  • the recording / reproducing unit 929 has an arbitrary readable / writable storage medium.
  • the storage medium may be a built-in storage medium such as a RAM or a flash memory, or an externally mounted storage medium such as a hard disk, a magnetic disk, a magneto-optical disk, an optical disk, a USB memory, or a memory card. May be.
  • the camera unit 926 images a subject to generate image data, and outputs the generated image data to the image processing unit 927.
  • the image processing unit 927 encodes the image data input from the camera unit 926 and stores the encoded stream in the storage medium of the recording / playback unit 929.
  • the demultiplexing unit 928 multiplexes the video stream encoded by the image processing unit 927 and the audio stream input from the audio codec 923, and the multiplexed stream is the communication unit 922. Output to.
  • the communication unit 922 encodes and modulates the stream and generates a transmission signal. Then, the communication unit 922 transmits the generated transmission signal to a base station (not shown) via the antenna 921.
  • the communication unit 922 amplifies a radio signal received via the antenna 921 and performs frequency conversion to acquire a received signal.
  • These transmission signal and reception signal may include an encoded bit stream.
  • the communication unit 922 demodulates and decodes the received signal to restore the stream, and outputs the restored stream to the demultiplexing unit 928.
  • the demultiplexing unit 928 separates the video stream and the audio stream from the input stream, and outputs the video stream to the image processing unit 927 and the audio stream to the audio codec 923.
  • the image processing unit 927 decodes the video stream and generates video data.
  • the video data is supplied to the display unit 930, and a series of images is displayed on the display unit 930.
  • the audio codec 923 decompresses the audio stream and performs D / A conversion to generate an analog audio signal. Then, the audio codec 923 supplies the generated audio signal to the speaker 924 to output audio.
  • the image processing unit 927 has the functions of the image encoding device 10 and the image decoding device 60 according to the above-described embodiment. Accordingly, when scalable coding and decoding of an image is performed on the mobile phone 920, the enhancement layer image data can be more efficiently encoded and the image data can be decoded.
  • FIG. 20 shows an example of a schematic configuration of a recording / reproducing apparatus to which the above-described embodiment is applied.
  • the recording / reproducing device 940 encodes audio data and video data of a received broadcast program and records the encoded data on a recording medium.
  • the recording / reproducing device 940 may encode audio data and video data acquired from another device and record them on a recording medium, for example.
  • the recording / reproducing device 940 reproduces data recorded on the recording medium on a monitor and a speaker, for example, in accordance with a user instruction. At this time, the recording / reproducing device 940 decodes the audio data and the video data.
  • the recording / reproducing device 940 includes a tuner 941, an external interface 942, an encoder 943, an HDD (Hard Disk Drive) 944, a disk drive 945, a selector 946, a decoder 947, an OSD (On-Screen Display) 948, a control unit 949, and a user interface. 950.
  • Tuner 941 extracts a signal of a desired channel from a broadcast signal received via an antenna (not shown), and demodulates the extracted signal. Then, the tuner 941 outputs the encoded bit stream obtained by the demodulation to the selector 946. That is, the tuner 941 has a role as a transmission unit in the recording / reproducing apparatus 940.
  • the external interface 942 is an interface for connecting the recording / reproducing apparatus 940 to an external device or a network.
  • the external interface 942 may be, for example, an IEEE 1394 interface, a network interface, a USB interface, or a flash memory interface.
  • video data and audio data received via the external interface 942 are input to the encoder 943. That is, the external interface 942 serves as a transmission unit in the recording / reproducing device 940.
  • the encoder 943 encodes video data and audio data when the video data and audio data input from the external interface 942 are not encoded. Then, the encoder 943 outputs the encoded bit stream to the selector 946.
  • the HDD 944 records an encoded bit stream in which content data such as video and audio is compressed, various programs, and other data on an internal hard disk. Also, the HDD 944 reads out these data from the hard disk when playing back video and audio.
  • the disk drive 945 performs recording and reading of data to and from the mounted recording medium.
  • the recording medium loaded in the disk drive 945 may be, for example, a DVD disk (DVD-Video, DVD-RAM, DVD-R, DVD-RW, DVD + R, DVD + RW, etc.) or a Blu-ray (registered trademark) disk. .
  • the selector 946 selects an encoded bit stream input from the tuner 941 or the encoder 943 when recording video and audio, and outputs the selected encoded bit stream to the HDD 944 or the disk drive 945. In addition, the selector 946 outputs the encoded bit stream input from the HDD 944 or the disk drive 945 to the decoder 947 during video and audio reproduction.
  • the decoder 947 decodes the encoded bit stream and generates video data and audio data. Then, the decoder 947 outputs the generated video data to the OSD 948. The decoder 904 outputs the generated audio data to an external speaker.
  • the OSD 948 reproduces the video data input from the decoder 947 and displays the video. Further, the OSD 948 may superimpose a GUI image such as a menu, a button, or a cursor on the video to be displayed.
  • a GUI image such as a menu, a button, or a cursor
  • the control unit 949 includes a processor such as a CPU and memories such as a RAM and a ROM.
  • the memory stores a program executed by the CPU, program data, and the like.
  • the program stored in the memory is read and executed by the CPU when the recording / reproducing apparatus 940 is activated, for example.
  • the CPU controls the operation of the recording / reproducing device 940 according to an operation signal input from the user interface 950, for example, by executing the program.
  • the user interface 950 is connected to the control unit 949.
  • the user interface 950 includes, for example, buttons and switches for the user to operate the recording / reproducing device 940, a remote control signal receiving unit, and the like.
  • the user interface 950 detects an operation by the user via these components, generates an operation signal, and outputs the generated operation signal to the control unit 949.
  • the encoder 943 has the function of the image encoding apparatus 10 according to the above-described embodiment.
  • the decoder 947 has the function of the image decoding device 60 according to the above-described embodiment.
  • FIG. 21 illustrates an example of a schematic configuration of an imaging apparatus to which the above-described embodiment is applied.
  • the imaging device 960 images a subject to generate an image, encodes the image data, and records it on a recording medium.
  • the imaging device 960 includes an optical block 961, an imaging unit 962, a signal processing unit 963, an image processing unit 964, a display unit 965, an external interface 966, a memory 967, a media drive 968, an OSD 969, a control unit 970, a user interface 971, and a bus. 972.
  • the optical block 961 is connected to the imaging unit 962.
  • the imaging unit 962 is connected to the signal processing unit 963.
  • the display unit 965 is connected to the image processing unit 964.
  • the user interface 971 is connected to the control unit 970.
  • the bus 972 connects the image processing unit 964, the external interface 966, the memory 967, the media drive 968, the OSD 969, and the control unit 970 to each other.
  • the optical block 961 includes a focus lens and a diaphragm mechanism.
  • the optical block 961 forms an optical image of the subject on the imaging surface of the imaging unit 962.
  • the imaging unit 962 includes an image sensor such as a CCD or a CMOS, and converts an optical image formed on the imaging surface into an image signal as an electrical signal by photoelectric conversion. Then, the imaging unit 962 outputs the image signal to the signal processing unit 963.
  • the signal processing unit 963 performs various camera signal processing such as knee correction, gamma correction, and color correction on the image signal input from the imaging unit 962.
  • the signal processing unit 963 outputs the image data after the camera signal processing to the image processing unit 964.
  • the image processing unit 964 encodes the image data input from the signal processing unit 963 and generates encoded data. Then, the image processing unit 964 outputs the generated encoded data to the external interface 966 or the media drive 968. The image processing unit 964 also decodes encoded data input from the external interface 966 or the media drive 968 to generate image data. Then, the image processing unit 964 outputs the generated image data to the display unit 965. In addition, the image processing unit 964 may display the image by outputting the image data input from the signal processing unit 963 to the display unit 965. Further, the image processing unit 964 may superimpose display data acquired from the OSD 969 on an image output to the display unit 965.
  • the OSD 969 generates a GUI image such as a menu, a button, or a cursor, for example, and outputs the generated image to the image processing unit 964.
  • the external interface 966 is configured as a USB input / output terminal, for example.
  • the external interface 966 connects the imaging device 960 and a printer, for example, when printing an image.
  • a drive is connected to the external interface 966 as necessary.
  • a removable medium such as a magnetic disk or an optical disk is attached to the drive, and a program read from the removable medium can be installed in the imaging device 960.
  • the external interface 966 may be configured as a network interface connected to a network such as a LAN or the Internet. That is, the external interface 966 has a role as a transmission unit in the imaging device 960.
  • the recording medium mounted on the media drive 968 may be any readable / writable removable medium such as a magnetic disk, a magneto-optical disk, an optical disk, or a semiconductor memory. Further, a recording medium may be fixedly attached to the media drive 968, and a non-portable storage unit such as an internal hard disk drive or an SSD (Solid State Drive) may be configured.
  • a non-portable storage unit such as an internal hard disk drive or an SSD (Solid State Drive) may be configured.
  • the control unit 970 includes a processor such as a CPU and memories such as a RAM and a ROM.
  • the memory stores a program executed by the CPU, program data, and the like.
  • the program stored in the memory is read and executed by the CPU when the imaging device 960 is activated, for example.
  • the CPU controls the operation of the imaging device 960 according to an operation signal input from the user interface 971, for example, by executing the program.
  • the user interface 971 is connected to the control unit 970.
  • the user interface 971 includes, for example, buttons and switches for the user to operate the imaging device 960.
  • the user interface 971 detects an operation by the user via these components, generates an operation signal, and outputs the generated operation signal to the control unit 970.
  • the image processing unit 964 has the functions of the image encoding device 10 and the image decoding device 60 according to the above-described embodiment. Therefore, when the image encoding device 960 performs scalable encoding and decoding of an image, the enhancement layer image data can be more efficiently encoded and such image data can be decoded.
  • the image encoding device 10 and the image decoding device 60 according to an embodiment have been described with reference to FIGS. 1 to 21.
  • the prediction mode selected based on the prediction mode set as the prediction unit of the lower layer is set as the prediction unit of the upper layer. Therefore, it is possible to reduce the amount of code associated with encoding the prediction mode information of the higher layer.
  • the amount of code generated when the prediction mode information is encoded as it is is not small. The above-described mechanism that can omit most of is useful.
  • the parameter when the number of candidates for the prediction mode of the upper layer is larger than that of the lower layer, the parameter is set in the upper layer using the parameter encoded according to the difference in the prediction direction. A prediction mode is selected. While avoiding encoding of prediction mode information of the upper layer, introducing such an additional parameter with a small number of bits improves the prediction accuracy of the intra prediction of the upper layer, resulting in encoding efficiency. Can be increased.
  • the parameter is encoded with a smaller code number as the absolute value of the difference in prediction direction between layers is smaller. In general, between corresponding prediction units of two layers that differ only in spatial resolution, partial images at the same position are correlated with each other. Therefore, by encoding the above parameters with a smaller code number as the prediction direction difference is smaller, more codewords with shorter variable length coding can be used. As a result, the encoding efficiency is further improved.
  • the prediction mode indicating the prediction direction closest to the prediction direction of the lower layer is the prediction unit of the upper layer.
  • the method for transmitting such information is not limited to such an example.
  • these pieces of information may be transmitted or recorded as separate data associated with the encoded bitstream without being multiplexed into the encoded bitstream.
  • the term “associate” means that an image (which may be a part of an image such as a slice or a block) included in the bitstream and information corresponding to the image can be linked at the time of decoding. Means. That is, information may be transmitted on a transmission path different from that of the image (or bit stream).
  • Information may be recorded on a recording medium (or another recording area of the same recording medium) different from the image (or bit stream). Furthermore, the information and the image (or bit stream) may be associated with each other in an arbitrary unit such as a plurality of frames, one frame, or a part of the frame.
  • the number of intra prediction mode candidates for the first prediction unit in the first layer of the image to be scalable decoded including the first layer and the second layer higher than the first layer is the first layer.
  • the selection is made based on the prediction mode set for the first prediction unit.
  • a mode setting unit that sets the prediction mode to be the second prediction unit;
  • a prediction unit that generates a prediction image of the second prediction unit according to a prediction mode set by the mode setting unit;
  • An image processing apparatus comprising: (2) When the number of intra prediction mode candidates for the first prediction unit is smaller than the number of intra prediction mode candidates for the second prediction unit, the first prediction unit and the second prediction unit A parameter acquisition unit that acquires a first parameter encoded according to a difference in prediction direction between The mode setting unit selects a prediction mode set in the second prediction unit according to the first parameter acquired by the parameter acquisition unit.
  • the image processing apparatus according to (1).
  • (3) The image processing apparatus according to (2), wherein the first parameter is encoded with a smaller code number as the absolute value of the difference in the prediction direction is smaller.
  • the mode setting unit when the number of intra prediction mode candidates for the first prediction unit is larger than the number of intra prediction mode candidates for the second prediction unit,
  • the image processing apparatus according to any one of (1) to (6), wherein a prediction mode representing a prediction direction closest to the prediction direction is set as the second prediction unit.
  • the mode setting unit performs an average value prediction when there are a plurality of prediction modes representing prediction directions closest to the prediction direction of the first prediction unit in the prediction mode candidates for the second prediction unit.
  • the image processing device according to (7), wherein a prediction mode to be expressed is set to the second prediction unit.
  • the mode setting unit is defined in advance when there are a plurality of prediction modes representing prediction directions closest to the prediction direction of the first prediction unit in the prediction mode candidates for the second prediction unit.
  • the image processing device wherein one of a plurality of prediction modes representing the closest prediction direction is selected according to a condition.
  • the predefined condition is a condition that the prediction direction is rotated in a predetermined rotation direction.
  • the predefined condition is a condition that a smaller code number is selected.
  • a parameter acquisition unit for acquiring The mode setting unit selects one of a plurality of prediction modes representing the closest prediction direction according to the second parameter acquired by the parameter acquisition unit.
  • the mode setting unit is adjacent to a prediction mode to be set as the second prediction unit, a prediction mode set as the first prediction unit, and the second prediction unit in the second layer.
  • the image processing apparatus according to (1) wherein estimation is performed by using MostProbableMode based on at least a prediction mode set as a third prediction unit.
  • the mode setting unit determines the MostProbableMode after converting the prediction mode set in the first prediction unit to a prediction mode in a prediction mode candidate for the second prediction unit, (13) An image processing apparatus according to 1.
  • the number of intra prediction mode candidates for the first prediction unit in the first layer of the image to be scalable decoded including the first layer and the second layer higher than the first layer is the first layer.
  • the selection is made based on the prediction mode set for the first prediction unit.
  • An image processing method including: (17) The number of intra prediction mode candidates for the first prediction unit in the first layer of a scalable encoded image including the first layer and a second layer higher than the first layer is Selection based on the prediction mode set in the first prediction unit when different from the number of intra prediction mode candidates for the second prediction unit corresponding to the first prediction unit in the second layer A mode setting unit that sets the predicted mode to be the second prediction unit; A prediction unit that generates a prediction image of the second prediction unit according to a prediction mode set by the mode setting unit; An image processing apparatus comprising: (18) The number of intra prediction mode candidates for the first prediction unit in the first layer of a scalable encoded image including the first layer and a second layer higher than the first layer is Selection based on the prediction mode set in the first prediction unit when different from the number of intra prediction mode candidates for the second prediction unit corresponding to the first prediction unit in the second layer Setting the predicted mode to be the second
  • Image encoding device (image processing device) 41 mode setting unit 42 prediction unit 45 parameter generation unit 60 image decoding device (image processing device) 91 Parameter acquisition unit 92 Mode setting unit 94 Prediction unit

Landscapes

  • Engineering & Computer Science (AREA)
  • Multimedia (AREA)
  • Signal Processing (AREA)
  • Computing Systems (AREA)
  • Theoretical Computer Science (AREA)
  • Compression Or Coding Systems Of Tv Signals (AREA)

Abstract

L'invention a pour but de proposer un système au moyen duquel un mode de prédiction pour une prédiction intra peut être codé de manière plus efficace dans un codage échelonnable. A cet effet, l'invention concerne un dispositif de traitement d'image qui comprend : un module de configuration de mode pour, lorsque le nombre de candidats de mode de prédiction intra pour une première unité de prédiction à l'intérieur d'une première couche d'une image qui est soumise à un décodage échelonnable et comprend la première couche et une seconde couche d'un ordre supérieur à la première couche est différent du nombre de candidats de mode de prédiction intra pour une seconde unité de prédiction qui correspond à la première unité de prédiction à l'intérieur de la seconde couche, configurer la seconde unité de prédiction pour avoir un mode de prédiction qui est sélectionné sur la base du mode de prédiction qui a été configuré pour la première unité de prédiction ; et un module de prédiction qui génère une image de prédiction pour la seconde unité de prédiction conformément au mode de prédiction configuré par le module de configuration de mode.
PCT/JP2012/062925 2011-06-28 2012-05-21 Dispositif de traitement d'image et procédé de traitement d'image WO2013001939A1 (fr)

Priority Applications (2)

Application Number Priority Date Filing Date Title
CN201280030622.8A CN103636211A (zh) 2011-06-28 2012-05-21 图像处理设备以及图像处理方法
US14/110,984 US20140037002A1 (en) 2011-06-28 2012-05-21 Image processing apparatus and image processing method

Applications Claiming Priority (2)

Application Number Priority Date Filing Date Title
JP2011-143271 2011-06-28
JP2011143271A JP2013012846A (ja) 2011-06-28 2011-06-28 画像処理装置及び画像処理方法

Publications (1)

Publication Number Publication Date
WO2013001939A1 true WO2013001939A1 (fr) 2013-01-03

Family

ID=47423842

Family Applications (1)

Application Number Title Priority Date Filing Date
PCT/JP2012/062925 WO2013001939A1 (fr) 2011-06-28 2012-05-21 Dispositif de traitement d'image et procédé de traitement d'image

Country Status (4)

Country Link
US (1) US20140037002A1 (fr)
JP (1) JP2013012846A (fr)
CN (1) CN103636211A (fr)
WO (1) WO2013001939A1 (fr)

Families Citing this family (8)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US10645398B2 (en) 2011-10-25 2020-05-05 Texas Instruments Incorporated Sample-based angular intra-prediction in video coding
KR101979284B1 (ko) * 2011-10-26 2019-05-17 인텔렉추얼디스커버리 주식회사 인터 예측 모드 스케일러블 코딩 방법 및 장치
GB2509901A (en) * 2013-01-04 2014-07-23 Canon Kk Image coding methods based on suitability of base layer (BL) prediction data, and most probable prediction modes (MPMs)
US10270123B2 (en) * 2015-01-09 2019-04-23 GM Global Technology Operations LLC Prevention of cell-to-cell thermal propagation within a battery system using passive cooling
WO2019059107A1 (fr) * 2017-09-20 2019-03-28 パナソニック インテレクチュアル プロパティ コーポレーション オブ アメリカ Dispositif d'encodage, dispositif de décodage, procédé d'encodage et procédé de décodage
WO2019129166A1 (fr) * 2017-12-29 2019-07-04 Fg Innovation Ip Company Limited Dispositif et procédé de codage de données vidéo basé sur une liste de modes contenant différents groupes de modes
CN110049339B (zh) 2018-01-16 2023-02-17 腾讯科技(深圳)有限公司 图像编码中的预测方向选取方法、装置和存储介质
CN111418207B (zh) * 2018-11-06 2024-04-19 北京字节跳动网络技术有限公司 依赖块尺寸的对运动信息的存储

Citations (1)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
JP2009500981A (ja) * 2005-07-11 2009-01-08 トムソン ライセンシング マクロブロック適応型レイヤ間テクスチャ内予測の方法及び装置

Family Cites Families (7)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN100534192C (zh) * 2003-10-28 2009-08-26 松下电器产业株式会社 帧内预测编码方法
JP4763422B2 (ja) * 2004-12-03 2011-08-31 パナソニック株式会社 イントラ予測装置
KR100913088B1 (ko) * 2005-01-21 2009-08-21 엘지전자 주식회사 베이스 레이어의 내부모드 블록의 예측정보를 이용하여영상신호를 엔코딩/디코딩하는 방법 및 장치
KR100703748B1 (ko) * 2005-01-25 2007-04-05 삼성전자주식회사 다 계층 기반의 비디오 프레임을 효율적으로 예측하는 방법및 그 방법을 이용한 비디오 코딩 방법 및 장치
KR100725407B1 (ko) * 2005-07-21 2007-06-07 삼성전자주식회사 방향적 인트라 잔차 예측에 따라 비디오 신호를 인코딩하고디코딩하는 방법 및 장치
CN101860759B (zh) * 2009-04-07 2012-06-20 华为技术有限公司 一种编码方法和编码装置
CN105791875B (zh) * 2011-06-10 2018-12-11 联发科技股份有限公司 可伸缩视频编码方法及其装置

Patent Citations (1)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
JP2009500981A (ja) * 2005-07-11 2009-01-08 トムソン ライセンシング マクロブロック適応型レイヤ間テクスチャ内予測の方法及び装置

Non-Patent Citations (2)

* Cited by examiner, † Cited by third party
Title
"Test Model under Consideration", JOINT COLLABORATIVE TEAM ON VIDEO CODING (JCT-VC) OF ITU-T SG16 WP3 AND ISO/IEC JTC1/SC29/WG11 2ND MEETING, 9 October 2010 (2010-10-09), GENEVA, CH, pages 67 - 70 *
HE LI ET AL.: "Fast mode decision for spatial scalable video coding", 2006 IEEE INTERNATIONAL SYMPOSIUM ON CIRCUITS AND SYSTEMS (ISCAS2006), 21 May 2006 (2006-05-21) *

Also Published As

Publication number Publication date
US20140037002A1 (en) 2014-02-06
JP2013012846A (ja) 2013-01-17
CN103636211A (zh) 2014-03-12

Similar Documents

Publication Publication Date Title
US10623761B2 (en) Image processing apparatus and image processing method
TWI749530B (zh) 資訊處理設備及資訊處理方法
TWI620433B (zh) Image processing device and method
US20150043637A1 (en) Image processing device and method
WO2013001939A1 (fr) Dispositif de traitement d'image et procédé de traitement d'image
WO2014002896A1 (fr) Dispositif et procédé de codage, dispositif et procédé de décodage
WO2013164922A1 (fr) Dispositif de traitement d'image et procédé de traitement d'image
WO2012005099A1 (fr) Dispositif de traitement d'image, et procédé de traitement d'image
WO2015008652A1 (fr) Dispositif et procédé de traitement d'image
WO2013011738A1 (fr) Appareil de traitement d'images et procédé de traitement d'images
US20180338156A1 (en) Image processing apparatus, image processing method, and program
WO2013150838A1 (fr) Dispositif de traitement d'image et procédé de traitement d'image
JPWO2015005367A1 (ja) 画像復号装置および方法
WO2013088833A1 (fr) Dispositif de traitement d'image et procédé de traitement d'image
JP2015076861A (ja) 復号装置および復号方法、並びに、符号化装置および符号化方法
WO2014038330A1 (fr) Dispositif de traitement d'image et procédé de traitement d'image
WO2013073328A1 (fr) Appareil de traitement d'image et procédé de traitement d'image
WO2014148070A1 (fr) Dispositif de traitement d'image et procédé de traitement d'image
WO2014002900A1 (fr) Dispositif et procédé de traitement d'images
WO2014097703A1 (fr) Dispositif et procédé de traitement d'image
JP6265249B2 (ja) 画像処理装置及び画像処理方法
JP6048564B2 (ja) 画像処理装置及び画像処理方法
WO2014050311A1 (fr) Dispositif de traitement d'image et procédé de traitement d'image
WO2014156707A1 (fr) Dispositif et procédé de codage d'image et dispositif et procédé de décodage d'image
WO2013088838A1 (fr) Dispositif de traitement d'image et procédé de traitement d'image

Legal Events

Date Code Title Description
121 Ep: the epo has been informed by wipo that ep was designated in this application

Ref document number: 12803853

Country of ref document: EP

Kind code of ref document: A1

WWE Wipo information: entry into national phase

Ref document number: 14110984

Country of ref document: US

NENP Non-entry into the national phase

Ref country code: DE

122 Ep: pct application non-entry in european phase

Ref document number: 12803853

Country of ref document: EP

Kind code of ref document: A1