CN114788270A - Image encoding method and image decoding method - Google Patents

Image encoding method and image decoding method Download PDF

Info

Publication number
CN114788270A
CN114788270A CN202080086924.1A CN202080086924A CN114788270A CN 114788270 A CN114788270 A CN 114788270A CN 202080086924 A CN202080086924 A CN 202080086924A CN 114788270 A CN114788270 A CN 114788270A
Authority
CN
China
Prior art keywords
prediction
image
intra
block
predicted image
Prior art date
Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
Pending
Application number
CN202080086924.1A
Other languages
Chinese (zh)
Inventor
清水拓也
Current Assignee (The listed assignees may be inaccurate. Google has not performed a legal analysis and makes no representation or warranty as to the accuracy of the list.)
Maxell Ltd
Original Assignee
Maxell Ltd
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Application filed by Maxell Ltd filed Critical Maxell Ltd
Publication of CN114788270A publication Critical patent/CN114788270A/en
Pending legal-status Critical Current

Links

Images

Classifications

    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04NPICTORIAL COMMUNICATION, e.g. TELEVISION
    • H04N19/00Methods or arrangements for coding, decoding, compressing or decompressing digital video signals
    • H04N19/10Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using adaptive coding
    • H04N19/102Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using adaptive coding characterised by the element, parameter or selection affected or controlled by the adaptive coding
    • H04N19/103Selection of coding mode or of prediction mode
    • H04N19/105Selection of the reference unit for prediction within a chosen coding or prediction mode, e.g. adaptive choice of position and number of pixels used for prediction
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04NPICTORIAL COMMUNICATION, e.g. TELEVISION
    • H04N19/00Methods or arrangements for coding, decoding, compressing or decompressing digital video signals
    • H04N19/10Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using adaptive coding
    • H04N19/102Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using adaptive coding characterised by the element, parameter or selection affected or controlled by the adaptive coding
    • H04N19/103Selection of coding mode or of prediction mode
    • H04N19/107Selection of coding mode or of prediction mode between spatial and temporal predictive coding, e.g. picture refresh
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04NPICTORIAL COMMUNICATION, e.g. TELEVISION
    • H04N19/00Methods or arrangements for coding, decoding, compressing or decompressing digital video signals
    • H04N19/10Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using adaptive coding
    • H04N19/134Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using adaptive coding characterised by the element, parameter or criterion affecting or controlling the adaptive coding
    • H04N19/157Assigned coding mode, i.e. the coding mode being predefined or preselected to be further used for selection of another element or parameter
    • H04N19/159Prediction type, e.g. intra-frame, inter-frame or bidirectional frame prediction
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04NPICTORIAL COMMUNICATION, e.g. TELEVISION
    • H04N19/00Methods or arrangements for coding, decoding, compressing or decompressing digital video signals
    • H04N19/10Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using adaptive coding
    • H04N19/169Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using adaptive coding characterised by the coding unit, i.e. the structural portion or semantic portion of the video signal being the object or the subject of the adaptive coding
    • H04N19/17Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using adaptive coding characterised by the coding unit, i.e. the structural portion or semantic portion of the video signal being the object or the subject of the adaptive coding the unit being an image region, e.g. an object
    • H04N19/176Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using adaptive coding characterised by the coding unit, i.e. the structural portion or semantic portion of the video signal being the object or the subject of the adaptive coding the unit being an image region, e.g. an object the region being a block, e.g. a macroblock
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04NPICTORIAL COMMUNICATION, e.g. TELEVISION
    • H04N19/00Methods or arrangements for coding, decoding, compressing or decompressing digital video signals
    • H04N19/50Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using predictive coding
    • H04N19/503Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using predictive coding involving temporal prediction
    • H04N19/51Motion estimation or motion compensation
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04NPICTORIAL COMMUNICATION, e.g. TELEVISION
    • H04N19/00Methods or arrangements for coding, decoding, compressing or decompressing digital video signals
    • H04N19/50Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using predictive coding
    • H04N19/593Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using predictive coding involving spatial prediction techniques

Landscapes

  • Engineering & Computer Science (AREA)
  • Multimedia (AREA)
  • Signal Processing (AREA)
  • Compression Or Coding Systems Of Tv Signals (AREA)

Abstract

The invention provides an image encoding method and an image decoding method. An image encoding method for encoding an image includes: a predicted image generation step of performing a synthesis process of synthesizing the predicted image for the inter prediction and the predicted image for the intra prediction with respect to the block to be encoded to generate a synthesized predicted image; and an encoding step of encoding a difference between pixel values of the predicted image generated in the predicted image generation step and the image of the encoding target block, wherein in the combining step, weighting processing is performed on the inter-predicted image and the intra-predicted image, and weighting parameters for the weighting processing are determined according to a combination of a type of the intra-prediction of the encoding target block and prediction modes of a plurality of adjacent blocks adjacent to the encoding target block, and the type of the intra-prediction of the encoding target block includes plane prediction and DC prediction.

Description

Image encoding method and image decoding method
Technical Field
The present invention relates to an image encoding technique for encoding an image or an image decoding technique for decoding an image.
Background
As a method for digitizing, recording, and transmitting image and audio information, h.264/avc (advanced Video coding) and h.265/hevc (high Efficiency Video coding) standards have been conventionally established. Research is being conducted to realize a next-generation system called vvc (very Video coding) that further exceeds the compression ratio of ISO/IEC MPEG and ITU-T VCEG (see non-patent document 1).
Documents of the prior art
Non-patent literature
Non-patent document 1: xiaozhong Xu and Shan Liu, "Recent advances in video coding beyond the HEVC standard" SIP (2019), vol.8
Disclosure of Invention
Problems to be solved by the invention
As one of the technical candidates for VVC, studies have been made on using a prediction image obtained by synthesizing an inter prediction image and an intra prediction image for the same block to be encoded/decoded.
However, only a predicted image obtained by simply averaging and combining an inter-predicted image and an intra-predicted image is used, and the reduction in the amount of code is not sufficient.
The present invention has been made in view of the above problems, and an object of the present invention is to provide a more preferable image encoding technique and image decoding technique.
Means for solving the problems
In order to achieve the above object, one embodiment of the present invention may employ, for example, the following: the method comprises the following steps: a predicted image generation step of performing a synthesis process of synthesizing the predicted image for the inter prediction and the predicted image for the intra prediction with respect to the block to be encoded to generate a synthesized predicted image; and an encoding step of encoding a difference between pixel values of the predicted image generated in the predicted image generation step and the image of the block to be encoded, wherein in the combining process, weighting processing is performed on the predicted image for inter prediction and the predicted image for intra prediction, and weighting parameters for the weighting processing are determined according to a combination of a type of the intra prediction of the block to be encoded and prediction modes of a plurality of adjacent blocks adjacent to the block to be encoded, and the type of the intra prediction of the block to be encoded includes planar prediction and DC prediction.
Effects of the invention
According to the present invention, a more preferable image encoding technique and image decoding technique can be provided.
Drawings
Fig. 1 is an explanatory diagram of an example of an image coding apparatus according to embodiment 1 of the present invention.
Fig. 2 is an explanatory diagram of an example of the image decoding apparatus according to embodiment 2 of the present invention.
Fig. 3 is an explanatory diagram of an example of the image coding method according to embodiment 1 of the present invention.
Fig. 4 is an explanatory diagram of an example of the image decoding method according to embodiment 2 of the present invention.
FIG. 5 is an explanatory diagram of an example of the data recording medium according to embodiment 3 of the present invention.
Fig. 6 is an explanatory diagram of an example of the type of intra prediction according to an embodiment of the present invention.
Fig. 7 is an explanatory diagram of an example of plane prediction according to an embodiment of the present invention.
Fig. 8 is an explanatory diagram of an example of the process of the synthesis prediction according to the embodiment of the present invention.
Fig. 9 is an explanatory diagram of an example of the process of the synthesis prediction according to the embodiment of the present invention.
Fig. 10 is an explanatory diagram of an example of the process of the synthesis prediction according to the embodiment of the present invention.
Detailed Description
Hereinafter, embodiments of the present invention will be described with reference to the drawings.
In addition, in each drawing, components denoted by the same reference numerals have the same functions.
In the description and drawings of the present specification, "0 vec" or "0 vector" indicates a vector in which the value of each component is 0, or a vector obtained by conversion and setting.
In addition, "unreferenceable" in each description and each drawing of the present specification means that information of a block cannot be acquired because the block position is out of the range of the screen or the like. "referenceable" means that information of a block can be acquired, and the information of the block includes information such as a pixel value, a vector, a reference frame number, and/or a prediction mode.
Note that the expression "residual component" in each description of the present specification and each drawing also includes the same meaning as "prediction error".
Note that the expression "region" in each description of the present specification and each drawing also includes the same meaning as "image".
In addition, the expression "transmit with tag" in each description of the present specification and each drawing also includes the meaning of "transmit including tag".
(example 1)
First, embodiment 1 of the present invention is explained with reference to the drawings.
Fig. 1 shows an example of a block diagram of an image encoding device according to an embodiment of the present invention.
The image coding apparatus includes, for example, an image input unit 101, a block dividing unit 102, a mode management unit 103, an intra prediction unit 104, an inter prediction unit 105, a synthesis prediction unit 120, a block processing unit 106, a transformation/quantization unit 107, an inverse quantization/inverse transformation unit 108, an image synthesis/filtering unit 109, a coded image management unit 110, an entropy coding unit 111, and a data output unit 112.
The operation of each component of the image coding apparatus will be described in detail below.
The operation of each component of the image coding apparatus can be regarded as autonomous operation of each component, for example, as described below. For example, the present invention may be realized by cooperation of software stored in the control unit and the storage unit.
First, the image input unit 101 acquires and inputs an original image to be encoded. Next, the block dividing unit 102 divides the input original image into blocks of a certain size called CTUs (coding Tree unit), and further analyzes the input image, thereby dividing each CTU into more detailed blocks according to its characteristics. These blocks, which are units of coding, are called cu (coding unit). The CU partition by the CTU is managed by a tree structure such as a Quadtree (Quadtree), a Ternary tree (Ternary tree), or a Binary tree. The inside of a CU may be further divided into subblocks for prediction and tu (transform unit) for frequency transform, quantization, and the like.
The mode management unit 103 manages a mode for determining a coding method of each CU. The encoding process is performed using a plurality of intra prediction methods, inter prediction methods, and composite prediction methods, and a mode with the highest encoding efficiency is determined for the CU. The most efficient mode is a mode in which a coding error can be minimized for a certain amount of code. There are a plurality of optimum modes, and the mode may be selected appropriately according to the situation. The determination of which mode is efficient is made by combining prediction processing in a plurality of modes by the intra prediction unit 104, the inter prediction unit 105, or the composite prediction unit 120, code amount measurement of residual components and various flags by other processing units, and reproduced image error prediction at the time of decoding. In general, the mode is determined for each CU unit, but the CU may be divided into sub-blocks and the mode may be determined for each sub-block.
As a prediction method of a block to be encoded (CU or sub-block), there are intra (intra) prediction, inter (inter) prediction, and composite prediction in which intra prediction and inter prediction are combined, and these are performed by the intra prediction unit 104, the inter prediction unit 105, and the composite prediction unit 120, respectively. The intra-frame prediction uses information of the same frame encoded before the encoding target block, and the inter-frame prediction uses information of a frame before or after the reproduction time, which is encoded before the encoding target frame. Here, the intra prediction unit 104, the inter prediction unit 105, and the synthesis prediction unit 120 are described as 1 unit for the sake of description, but may be provided for each encoding mode and each frame.
The intra prediction unit 104 performs intra prediction processing. In the "prediction processing", a predicted image is generated. The intra-picture prediction process predicts pixels of a block to be encoded using information of the same frame encoded prior to the block to be encoded. The intra prediction includes DC prediction, angle prediction, plane prediction, matrix prediction, cross-component prediction, multi-line prediction, intra block copy, and the like. For the transmission of the intra prediction mode, the mode with the highest probability is estimated from the intra prediction mode of the encoded block.
The inter prediction unit 105 performs inter prediction processing. In addition, a prediction image is generated in the "prediction process". The inter-picture prediction process predicts the pixels of the encoding target block using information of a frame before or after the reproduction time, which is encoded before the encoding target frame. The inter prediction includes motion compensation prediction, merge mode prediction, affine transformation-based prediction, triangular block segmentation-based prediction, optical flow prediction, decoder-side motion prediction, and the like.
The synthesis prediction unit 120 performs synthesis prediction for generating a synthesized prediction image in which a prediction image obtained by the intra prediction and a prediction image obtained by the inter prediction are synthesized. The predicted image obtained by the intra prediction used by the synthesis prediction unit 120 may be a predicted image obtained by various intra prediction processes that can be performed by the intra prediction unit 104. As the predicted image obtained by the inter prediction used in the synthesis prediction unit 120, a predicted image obtained by various intra prediction processes that can be performed by the inter prediction unit 105 described above may be used. Details of the synthesis method are described later.
The block processing unit 106 obtains, for each block to be encoded, a difference (difference) between a predicted image generated by the intra prediction unit 104 through intra prediction, a predicted image generated by the inter prediction unit 105 through inter prediction, or a predicted image generated by the synthesis prediction unit 120 through synthesis prediction and an original image of the block to be encoded obtained from the block division unit 102, calculates a residual component, and outputs the residual component.
The transform/quantization unit 107 performs frequency transform and quantization processing on the residual component input from the block processing unit 106, and outputs a coefficient group. The frequency conversion may be performed using dct (discrete Cosine transform), dst (discrete Sine transform), or a form that can be processed by integer arithmetic. The coefficient set is transmitted to both a process for restoring an image to generate a decoded image for prediction and a process for outputting data. The transform and quantization may also be skipped according to the mode designation.
The inverse quantization/inverse transformation unit 108 performs inverse quantization and inverse transformation to generate a decoded image using the coefficient group obtained from the transformation/quantization unit 107 for prediction, and outputs a residual component obtained by restoration. The inverse quantization and the inverse transformation may be performed in opposite directions corresponding to quantization and transformation by the transformation/quantization unit, respectively. Inverse quantization and inverse transformation may also be skipped according to the mode designation.
The image synthesis/filtering unit 109 synthesizes the predicted image generated by the intra prediction unit 104, the predicted image generated by the inter prediction unit 105, or the predicted image generated by the synthesis prediction unit 120 with the residual component restored by the inverse quantization/inverse transformation unit 108, and performs processing such as loop filtering to generate a decoded image.
The decoded image management unit 110 holds a decoded image, and manages images, mode information, and the like that are referred to for intra prediction, inter prediction, or composite prediction.
The entropy encoding unit 111 performs entropy encoding processing on the mode information and the coefficient group information and outputs the result as a bit string. As the entropy encoding method, a cabac (context Adaptive Binary arithmetric code) method or the like may be used. Variable length codes, fixed length codes may be used in combination. The context determination may refer to a predetermined table.
The data output unit 112 outputs the encoded data to the recording medium and the transmission path.
Next, a flow of an encoding method in the image encoding device according to embodiment 1 of the present invention will be described with reference to fig. 3.
First, in step 301, an original image to be encoded is input, and the content of the image is analyzed to determine a division method, and the image is divided into blocks. The analysis of the image content may be performed for the entire image, or for a plurality of frames in combination, or may be performed for each block unit such as a slice (slice), tile (tile), brick (brick), CTU, or the like obtained by dividing the image. The block division is generally performed by dividing the block into CTUs of a predetermined size and then dividing the CTUs into CUs in a tree structure.
Next, in step 302, intra prediction is performed on the encoding target block of the original image acquired in step 301. The intra prediction mode is as described above. Prediction is performed for a plurality of modes for each intra prediction mode.
Next, in step 303, inter prediction is performed on the encoding target block of the original image acquired in step 301. The inter prediction mode is as described above. Prediction is performed for a plurality of modes for each inter prediction mode.
Next, in step 320, synthesis prediction is performed on the encoding target block of the original image acquired in step 301. Details of the synthetic prediction mode are described later.
Next, in step 304, residual components are separated from the images of the encoding target blocks after intra prediction, inter prediction, and composite prediction, and the transform processing, quantization processing, and entropy encoding processing of the residual components are performed, in accordance with the respective modes, to calculate encoded data.
Next, in step 305, inverse quantization and inverse transform processing are performed in accordance with each mode, and a residual component is synthesized with a predicted image, thereby generating a decoded image. The decoded image is managed together with prediction data and various kinds of encoded data in intra prediction and inter prediction, and is used for prediction of other blocks to be encoded.
Next, in step 306, the patterns are compared to determine the pattern that can be encoded with the highest efficiency. The modes include an intra prediction mode, an inter prediction mode, a synthesis prediction mode, and the like, and are collectively referred to as an encoding mode. The mode selection method is as described above.
In step 307, the encoded data of the block to be encoded is output in accordance with the determined encoding mode. The above-described encoding process for each block to be encoded is repeated for the entire image, and the image is encoded.
Next, the synthetic prediction mode of the present embodiment will be described. Here, the same processing is performed on the encoding side and the decoding side in the predicted image generation processing based on the combined prediction mode. Therefore, in the following description of the synthetic prediction mode of the present embodiment, the same processing is performed on the encoding side and the decoding side unless otherwise specified.
The synthesized prediction mode of the present embodiment generates a predicted image obtained by synthesized prediction by performing weighted average of an intra-predicted image obtained by intra-prediction and an inter-predicted image obtained by inter-prediction. Specifically, the inter-prediction pixel is Pinter and the intra-prediction pixel is Pintra, and each prediction pixel Pcom (hereinafter referred to as a synthesized prediction pixel) of the prediction image is calculated by the following expression 1.
Pcom ═ (w × Pintra + (4-w) × Pinter)/4 … … formula 1
That is, the process of expression 1 is a weighted average process of inter-prediction pixels and intra-prediction pixels using a weighting parameter w. In addition, this process may also be expressed as weighted addition.
As for the inter-prediction image used in the synthesized prediction mode of the present embodiment, a prediction image obtained by all inter-predictions that can be used in the above-described inter-prediction mode can be used as a candidate. In the inter prediction mode, only the prediction image obtained in the merge mode may be used.
As for the intra-prediction image used in the synthetic prediction mode of the present embodiment, a prediction image obtained by all intra-predictions usable in the above-described intra-prediction mode can be used as a candidate. In the intra prediction mode, only a prediction image obtained by angular prediction, DC prediction, or plane prediction may be used.
Here, in the synthesized prediction mode of the present embodiment, the weighting parameter w for calculating the synthesized prediction pixel is determined in accordance with the type of intra prediction for generating the intra prediction image used in the synthesized prediction mode. That is, the weight of the intra-prediction image in the weighted average of the intra-prediction image and the inter-prediction image is changed in accordance with the characteristics of the different intra-predictions, thereby realizing generation of a more preferable prediction image.
A specific example of the process of determining the weighting parameter w in the synthesized prediction mode according to the present embodiment will be described below.
First, intra prediction, that is, angular prediction, DC prediction, and planar prediction, which are included in candidates used for generating an intra prediction image used in the synthetic prediction mode according to the present embodiment, will be described with reference to fig. 6 and 7.
The prediction direction of the angle prediction is shown in fig. 6. The angle prediction is a prediction method for predicting a prediction target pixel by referring to adjacent pixels indicated by prediction directions shown in the drawing. This prediction is unidirectional prediction, and it is also possible to perform prediction in which adjacent pixel values are copied in the opposite direction to each prediction direction. The prediction directions can be identified by numbers, and are set as the prediction directions 2 to 66 shown in the figure, for example. In the figure, 2 (lower left direction), 18 (horizontal left direction), 34 (upper left direction), 50 (upper direction (vertical direction)), and 66 (upper right direction) are shown as representative prediction directions. In this figure, the description of the other prediction directions is omitted for the sake of simplicity of explanation, but the prediction direction is defined in the clockwise rotation direction from the prediction direction 2 to the prediction direction 66 as the number of prediction directions increases. The angle prediction is a unidirectional prediction based on neighboring pixel values.
Therefore, the prediction residual tends to be reduced for an image having a pattern such as a stripe pattern extending in a predetermined direction. The compression rate increases with decreasing prediction residual. When the encoding side selects angle prediction, when an image with reduced prediction residues in angle prediction is encoded, there is a high possibility that the image is an image in which a pattern extends in a predetermined direction. In this way, when the encoding side selects angle prediction for the block to be encoded, there is a high possibility that angle prediction using the same prediction direction as the block to be encoded or a prediction direction close to the prediction direction of the block to be encoded is also used for adjacent blocks. This is because, as long as the pattern in the image in which the pattern extends in the predetermined direction continues to extend, the prediction residual is highly likely to decrease in angular prediction in the same prediction direction as the block to be encoded or in a prediction direction close to the prediction direction of the block to be encoded. This can also be expressed as a high possibility of high prediction accuracy.
In this case, in a plurality of adjacent blocks, it is highly likely that angle prediction of the same prediction direction or angle prediction of a prediction direction close to the same direction is repeatedly selected. That is, the angle prediction may be considered to be intra prediction having a relatively high approximation with the prediction method of the adjacent block. In addition, "the prediction direction close to the prediction direction of the encoding target block" may be specifically defined as a prediction direction whose number of the prediction direction is ± 1 of the number of the prediction direction of the encoding target block. Alternatively, specifically, the prediction direction number may be defined as a prediction direction of ± 2 of the prediction direction number of the block to be encoded.
In fig. 6, the DC prediction is further indicated by 1. DC prediction is a prediction method for generating a prediction image corresponding to only the DC component in the frequency domain. That is, all the pixel values within the prediction image of the DC prediction are constant values. DC prediction is a prediction method that is not based on neighboring pixel values. However, an image having a tendency to reduce the prediction residual in the DC prediction is a planar pattern having a certain luminance and a wide range. In this way, when the encoding side selects DC prediction for the block to be encoded, there is a high possibility that the same DC prediction as that for the block to be encoded is also used for adjacent blocks. This is because the likelihood of a reduction in the prediction residual in the same DC prediction is high as long as the patterns of planes of a certain luminance are connected in the peripheral direction.
This can also be expressed as a high possibility of high prediction accuracy. In this case, the same DC prediction is likely to be selected repeatedly in a plurality of adjacent blocks. That is, although DC prediction is a prediction method not based on the neighboring pixel values, it can be considered as intra prediction having a relatively high approximation with the prediction method of the neighboring block.
In fig. 6, it is further shown that the number of the plane prediction is 0. Fig. 7 shows a specific prediction method of the plane prediction. In plane prediction, a reference pixel is generated at the right boundary of the encoding target block using the upper right neighboring pixel of the encoding target block (step 710) (step 711). Next, a first prediction image is generated by generating each prediction pixel by linear interpolation using the right-side reference pixel and a left-side adjacent pixel opposite to the right-side reference pixel (fig. 7(a)) (process 712).
In addition, a reference pixel is generated at the lower boundary of the encoding target block using the lower left neighboring pixel of the encoding target block (process 720) (process 721). Next, a second prediction image is generated by generating each prediction pixel by linear interpolation using the lower reference pixel and the upper adjacent pixel opposed to the lower reference pixel (fig. 7(B)) (process 722). Then, the first predicted image (fig. 7 a) and the second predicted image (fig. 7B) thus generated are averaged to generate a final predicted image. The plane prediction is an intra-frame prediction having a feature in that a first prediction image (fig. 7 a) and a second prediction image (fig. 7B) generated by linear interpolation are averaged to generate a prediction image of a nonlinear curved surface.
Planar prediction is prediction based on neighboring pixel values. However, since a predicted image generated by plane prediction is a characteristic nonlinear curved surface, the approximation to a prediction method of an adjacent block is not high. As described above, when the prediction residual decreases in the angular prediction and the DC prediction, there is a high possibility that the same prediction method as that of the adjacent block is repeated. In contrast, there are not many images of patterns that require repetition of a characteristic nonlinear curved surface of a predicted image for plane prediction.
Therefore, the plane prediction can be considered as an intra prediction having a relatively low approximation with a prediction method of a neighboring block. When the encoding side selects plane prediction for the block to be encoded, there is a high possibility that the image is an image in which the prediction residual is not sufficiently reduced in angular prediction and DC prediction. Since an image in which the prediction residual is not sufficiently reduced in the angle prediction and the DC prediction is not a simple pattern, a prediction residual obtained by plane prediction of an image in which the prediction residual is not reduced in the angle prediction and the DC prediction is more likely to be larger than a prediction residual obtained when the prediction residual is reduced in the angle prediction and the DC prediction. This can also be expressed as a high possibility that the prediction accuracy is low.
As described using fig. 6 and 7, the angular prediction, the DC prediction, and the plane prediction that can be used in the synthetic prediction mode of the present embodiment are intra predictions having different characteristics, respectively. In the combined prediction mode according to the present embodiment, the value of the weighting parameter w in equation 1 is controlled in accordance with the type of intra prediction, focusing on the characteristics of each type of intra prediction. The w determination process in the synthetic prediction mode according to the present embodiment will be described with reference to fig. 8.
Fig. 8 shows an example of w determination in the synthetic prediction mode according to the present embodiment. Fig. 8 shows 3 examples of w determination example 1, w determination example 2, and w determination example 3. These are examples of respectively different determination methods. In each determination method, the value of w is determined from 1, 2, and 3 for each condition. In equation 1, the larger the value of w, the greater the weighting of intra prediction in the composite prediction mode. Conversely, in equation 1, the smaller the value of w, the smaller the weight of intra prediction in the composite prediction mode. These determination examples are explained below.
In w determination example 1, when the intra prediction used in the composite prediction mode is the plane prediction, it is determined that w is 1. When intra prediction used in the combined prediction mode is angular prediction and the prediction direction is the horizontal direction (18) or the vertical direction (50), w is determined to be 3. The intra prediction used in the synthetic prediction mode is angle prediction, and when the prediction direction is 2-17 or 51-66, it is determined that w is 3. When intra prediction used in the synthesis prediction mode is angle prediction and the prediction direction is 19 to 49, w is determined to be 3. When the intra prediction used in the synthetic prediction mode is DC prediction, w is determined to be 3.
That is, in the w decision example 1, the decision method is to make the plane prediction, which is the intra prediction having a relatively low similarity to the prediction method of the adjacent block, have a smaller value of w than the angle prediction and the DC prediction. It can also be expressed that the predicted image of the plane prediction is a characteristic nonlinear curved surface and the prediction accuracy is highly likely to be low, and therefore it is preferable to apply the smallest weight in the combined prediction mode.
In contrast, in the w determination example 1, for the intra prediction, i.e., the angular prediction, which has relatively high similarity to the prediction method of the adjacent block, 3, which is a value of w larger than the value of w at the time of the plane prediction, is selected in any prediction direction. The determination method is to make the value of w larger than that of the plane prediction, which is intra prediction having a relatively high similarity to the prediction method of the adjacent block, i.e., angular prediction. It can also be expressed that since the predicted image of the angular prediction is a simple one-directional prediction and has a high possibility of high prediction accuracy, it is preferable to apply a relatively large weight to the synthesized prediction mode.
In addition, in w determination example 1, for DC prediction, which is intra prediction having a relatively high similarity to the prediction method of the adjacent block, 3, which is a value of w larger than the value of w in plane prediction, is selected. The method of determination is to make DC prediction, which is intra prediction having a relatively high similarity to the prediction method of the adjacent block, have a larger value of w than planar prediction. The DC prediction may be expressed as a prediction image that is a simple plane of constant value and has a high possibility of high prediction accuracy, and therefore, it is preferable to apply a relatively large weight to the prediction image in the combined prediction mode.
In this way, in w determination example 1, the value of w is determined according to the type of intra prediction. In this case, the value for which the plane prediction is the minimum is set. In the angular prediction, the value of w is set to be constant for any prediction direction. That is, in the w determination example 1, the value of w does not change according to the prediction direction of the angle prediction.
Next, in w determination example 2, when the intra prediction used in the composite prediction mode is the plane prediction, it is determined that w is 1. When intra prediction used in the synthetic prediction mode is angular prediction and the prediction direction is the horizontal direction (18) or the vertical direction (50), w is determined to be 2. The intra prediction used in the synthetic prediction mode is angle prediction, and when the prediction direction is 2-17 or 51-66, it is determined that w is 3. When intra prediction used in the synthesis prediction mode is angle prediction and the prediction direction is 19 to 49, w is determined to be 3. When the intra prediction used in the synthetic prediction mode is DC prediction, w is determined to be 3.
That is, in the w decision example 2, the decision method is such that the plane prediction, which is intra prediction having relatively low similarity to the prediction method of the adjacent block, takes a smaller value of w than the angle prediction and the DC prediction, as in the w decision example 1. The reason is the same as in determination example 1. In w determination example 2, in intra prediction, i.e., angular prediction, which is relatively high in the similarity to the prediction methods of adjacent blocks, 2 or 3, which is a value of w larger than that in plane prediction, is selected in any prediction direction. The determination method is to make the value of w larger than that of the plane prediction, which is intra prediction having a relatively high similarity to the prediction method of the adjacent block, i.e., angular prediction.
It can also be expressed that since the predicted image of the angular prediction is a simple one-directional prediction and has a high possibility of high prediction accuracy, it is preferable to apply a relatively large weight to the synthesized prediction mode. However, in w determination example 2, unlike w determination example 1, even if angle prediction is performed in the same manner, the value of w is changed according to the prediction direction.
Specifically, when the prediction direction is the horizontal direction (18) or the vertical direction (50), w is set to 2, and a lower weight is set than w of the other prediction direction in the angle prediction to 3. This is because the reference pixel in the prediction direction in the horizontal direction (18) is only the pixel of the left adjacent block, and the probability that the approximation with the prediction direction of the adjacent block is limited to the left adjacent block is high, and the overall approximation with the prediction direction of the adjacent block is considered to be lower than that with other prediction directions.
Similarly, since the reference pixels in the prediction direction in the vertical direction (50) include only the pixels of the upper neighboring blocks, the probability that the approximation to the prediction direction of the neighboring blocks is limited to the upper neighboring blocks is high, and the overall approximation to the prediction direction of the neighboring blocks is considered to be lower than that in other prediction directions. In addition, in w determination example 2, as in w determination example 1, for DC prediction, which is intra prediction having a relatively high similarity to the prediction direction of the adjacent block, 3, which is a value of w larger than the value of w in plane prediction, is selected. The reason is the same as in determination example 1.
In this way, in the w determination example 2, the value of w is determined according to the type of intra prediction. In addition, when intra prediction is angular prediction, the value of w is also determined in consideration of the prediction direction. In addition, the minimum value of the plane prediction is set.
w determination example 3 is different from w determination example 2 in that when intra prediction used in the synthetic prediction mode is angular prediction and the prediction directions are 2 to 17 or 51 to 66, w is determined to be 2, and the other points are common. In the case of angle prediction and prediction directions 2 to 17 or 51 to 66, since the angle of the prediction direction is not a completely horizontal direction or a completely vertical direction, it is considered that there is an approximation of the prediction method to any one of the upper neighboring block and the left neighboring block.
However, the pixel included in the reference pixel is either the pixel of the upper neighboring block or the pixel of the left neighboring block. In the case of the prediction directions 2 to 17 or 51 to 66, the overall similarity to the prediction methods of the neighboring blocks is likely to be lower than the prediction directions 19 to 49 in which the reference pixels include both the pixels of the upper neighboring block and the pixels of the left neighboring block. In view of this, in w determination example 3, in the case of prediction directions 2 to 17 or 51 to 66, w is determined to be 2, and the weight is made smaller than in the case of prediction directions 19 to 49.
According to the example of determining w in the combined prediction mode in fig. 8 described above, the weighting of the intra-prediction image in the combination of the intra-prediction image and the inter-prediction image can be determined more appropriately according to the characteristics of each type of intra-prediction used in the combined prediction mode.
As described above, the encoding process in one embodiment of the present application is performed.
According to the image encoding device and the image encoding method of embodiment 1 described above, a more preferable synthesis prediction mode can be realized.
In addition, the image encoding device and the image encoding method of embodiment 1 can be applied to a recording device, a mobile phone, a digital camera, and the like using them.
According to the image encoding device and the image encoding method according to embodiment 1 of the present invention described above, the amount of encoded data can be reduced, and the quality of a decoded image can be prevented from deteriorating when the encoded data is decoded. That is, a high compression ratio and a better image quality can be achieved.
That is, according to the image encoding device and the image encoding method of embodiment 1 of the present invention, a more preferable image encoding technique can be provided.
(example 2)
Next, fig. 2 shows an example of a block diagram of an image decoding apparatus according to embodiment 2 of the present invention.
The image decoding apparatus includes, for example, a stream analyzer 201, a block manager 202, a mode determiner 203, an intra predictor 204, an inter predictor 205, a synthesis predictor 220, a coefficient analyzer 206, an inverse quantization/inverse transformation unit 207, an image synthesis/filter unit 208, a decoded image manager 209, and an image output unit 210.
The operation of each component of the image decoding apparatus will be described in detail below.
The operation of each component of the image decoding apparatus can be regarded as autonomous operation of each component, for example, as described below. For example, the present invention may be realized by cooperation of software stored in the control unit and the storage unit.
First, the stream analysis unit 201 analyzes an input coded stream. Here, the flow analysis unit 201 also performs processing for extracting data from the packet and processing for acquiring information of various headers and flags.
In this case, the encoded stream input by the stream analysis unit 201 is, for example, an encoded stream generated by the image encoding device and the image encoding method according to embodiment 1. The method of production is as in example 1, and therefore, the description thereof is omitted. Or may be a coded stream read from the data recording medium shown in embodiment 3. The recording method is described later.
Next, the block management unit 202 manages the processing of the blocks according to the block-divided information analyzed by the flow analysis unit 201. Generally, an encoded image is divided into blocks, and each encoding target block is managed by a tree structure or the like. The processing order of the blocks is often performed in the raster scan order, but the processing may be performed in an arbitrarily determined order such as zigzag scan. The block division method is as described in embodiment 1.
Next, the mode determination unit 203 determines the coding mode specified by a flag or the like for each block to be coded. The decoding process performs a process corresponding to the coding mode of the discrimination result. The following describes processing for each coding mode.
First, when the encoding mode is intra-encoding, the intra-prediction unit 204 generates a prediction image based on intra-prediction. The intra prediction mode is as described in embodiment 1. The generation processing of the prediction image based on prediction is basically the same processing on the encoding side and the decoding side.
When the encoding mode is the encoding based on the inter prediction, the inter prediction unit 205 generates a predicted image based on the inter prediction. The inter prediction mode is as described in embodiment 1. The generation processing of the prediction image based on prediction is basically the same processing on the encoding side and the decoding side.
When the encoding mode is encoding based on the synthesis prediction, the synthesis prediction unit 220 generates a prediction image based on the synthesis prediction. The synthetic prediction mode is as described in example 1. The generation processing of the prediction image based on prediction is basically the same processing on the encoding side and the decoding side.
On the other hand, the coefficient analysis unit 206 analyzes the encoded data of each encoding target block included in the input encoded stream, decodes the entropy-encoded data, and outputs encoded data of a coefficient group including a residual component. At this time, processing corresponding to the coding mode of the determination result of the mode determination unit 203 is performed.
The inverse quantization/inverse transformation unit 207 performs inverse quantization processing and inverse transformation on the encoded data of the coefficient group including the residual component, and restores the residual component. The inverse quantization and inverse transformation methods are as described above. Inverse quantization and inverse transformation may also be skipped according to the mode designation.
The residual components restored as described above are synthesized by the image synthesis/filter unit 208 with the prediction image output from the intra prediction unit 204, the inter prediction unit 205, or the synthesis prediction unit 220, and are output as a decoded image by performing processing such as loop filtering.
The decoded image management unit 209 holds the decoded image, and manages images, mode information, and the like that are referred to for intra prediction, inter prediction, or composite prediction.
The finally decoded image is output by the image output unit 210, and the image is decoded.
Next, a flow of an image decoding method in the image decoding device according to embodiment 2 of the present invention will be described with reference to fig. 4.
First, in step 401, a coded stream to be decoded is acquired, and data is analyzed. In addition, the processing of the block is managed according to the block division information obtained by the analysis. The block division method is as described in embodiment 1.
Next, in step 402, the information of the coding mode analyzed in step 401 is used to determine the coding mode for 1 coding unit (block unit, pixel unit, or the like) included in the coded data. Here, the process proceeds to step 403 in the case of the intra coding mode, to step 404 in the case of the inter coding mode, and to step 420 in the case of the composite coding mode.
In step 403, a predicted image is generated by intra prediction according to the method specified by the encoding mode. The intra prediction mode is as described in embodiment 1.
In step 404, a predicted image is generated by inter prediction according to the method specified by the encoding mode. The inter prediction mode is as described in embodiment 1.
In step 420, a predicted image is generated by the synthetic prediction according to the method specified by the encoding mode. The synthetic prediction mode is as described in example 1.
In step 405, the encoded data of each block to be encoded is analyzed by a method specified by the encoding mode, the entropy-encoded data is decoded, and the encoded data of the coefficient group including the residual component is output. Further, the encoded data of the coefficient group including the residual component is subjected to inverse quantization and inverse transformation, and the residual component is restored. The inverse quantization and inverse transformation methods are as described above. Inverse quantization and inverse transformation may also be skipped according to the mode designation.
In step 406, a decoded image is generated by synthesizing a predicted image generated by intra prediction, inter prediction, or synthetic prediction with the residual component obtained by the restoration, and performing loop filtering or the like on each block to be encoded. The above-described decoding process is performed on the entire image in units of blocks to be encoded, thereby generating a decoded image.
In step 407, the generated decoded image is output and displayed.
The processing in the decoding-side synthetic prediction mode in this embodiment is substantially the same as the processing in the synthetic prediction mode described in embodiment 1 using expression 1, fig. 6, fig. 7, and fig. 8, and therefore, the repetitive description thereof will be omitted. In addition, the expression "encoding target block" in the description of the synthetic prediction mode in embodiment 1 may be replaced with "decoding target block" in the processing of the synthetic prediction mode on the decoding side.
In the decoding-side synthesized prediction mode of the present embodiment, as described above, weighted average processing using the inter prediction pixels and the intra prediction pixels in expression 1 is performed. In the decoding-side synthetic prediction mode of the present embodiment, the weighting parameter w is determined by the determination process shown in fig. 8 as described above. This makes it possible to more appropriately determine the weighting of the intra-predicted image in the synthesis of the intra-predicted image and the inter-predicted image, according to the characteristics of each type of intra-prediction used in the synthesis prediction mode.
In addition to the example shown in the present embodiment, the size of a block used in an encoding mode may be used as a parameter, and each encoding mode may be subdivided to use a predetermined encoded stream as a stream to be decoded.
As described above, the decoding process of one embodiment of the present application is performed.
The image decoding apparatus and the image decoding method of embodiment 2 can be applied to a playback apparatus, a mobile phone, a digital camera, and the like using them.
According to the image decoding apparatus and the image decoding method according to embodiment 2 of the present invention described above, encoded data with a small code amount can be decoded with a higher picture quality.
That is, according to the image decoding apparatus and the image decoding method of embodiment 2 of the present invention, a more preferable image decoding technique can be provided.
(example 3)
Next, fig. 5 shows an example of a data recording medium according to embodiment 3 of the present invention.
The encoded stream of the present embodiment of the present invention is an encoded stream generated by the image encoding apparatus or the image encoding method of embodiment 1. The method of production is as in example 1, and therefore, the description thereof is omitted.
Here, the encoded stream of the present embodiment is recorded as a data string 502 in the data recording medium 501, for example. The data string 502 is recorded as a coded stream conforming to a prescribed syntax, for example.
First, a bit string divided into units of a certain size called nal (network Abstraction layer) unit 503 is derived for a coded stream. Bit strings of NAL units are read according to a predetermined rule such as variable length coding and converted into rbsp (raw Byte Sequence payload). The data of the RBSP is composed of information such as a sequence parameter set 504, a picture parameter set 505, a decoding parameter set, a video parameter set, and the like, and slice data 506.
Inside each slice, information 507 about each block is included, for example. Within the information on the block, for example, an area in which the encoding mode is recorded for each block is present, and this area is set as the encoding mode flag 508.
According to the data recording medium of embodiment 3 of the present invention described above, the amount of code can be reduced, and image quality degradation can be prevented. That is, a data recording medium on which a coded stream having a high compression rate and a high image quality is recorded can be realized.
Further, the coded stream generated by the image coding apparatus or the image coding method in each embodiment described later can also be recorded in the data recording medium of the present embodiment with the configuration described above in the present embodiment.
(example 4)
Next, embodiment 4 of the present invention will be described with reference to the drawings.
In embodiment 4 of the present invention, the image encoding device and the image encoding method in embodiment 1 change the process of determining the weighting parameter w in the process of the synthesis prediction.
Specifically, the process of determining the weighting parameter w is changed from the process of determining the weighting parameter w shown in fig. 8 to the process of determining the weighting parameter w shown in fig. 9. In the image encoding device and the image encoding method according to embodiment 4, the configuration, operation, and processing other than the determination processing of the weighting parameter w shown in fig. 9 are the same as those of the image encoding device and the image encoding method according to embodiment 1. Therefore, the same points as those of the image encoding device and the image encoding method of embodiment 1 will not be described repeatedly.
Fig. 9 shows an example of the determination process of the weighting parameter w in the process of the synthesis prediction in the image coding apparatus and the image coding method according to embodiment 4 of the present invention. Since the same process of determining the weighting parameter w is performed also on the decoding side, the expression "encoding (decoding) target block" for the prediction target block will be described.
In the example of fig. 9, unlike the example of fig. 8, in the process of determining the weighting parameter w, not only the intra prediction mode of the block to be encoded (decoded) but also the combination of the prediction modes of a plurality of adjacent blocks are considered. Here, in the present embodiment, as an example of the plurality of adjacent blocks, a combination of prediction modes of 2 adjacent blocks, i.e., a block immediately above the block to be encoded (decoded) and a block immediately to the left of the block to be encoded (decoded), is considered.
That is, in the example of fig. 8, the weighting parameter w is determined only in accordance with the intra prediction mode of the block to be encoded (decoded), but in the example of fig. 9, even if the intra prediction mode of the block to be encoded (decoded) is the same, the weighting parameter w is changed if the combination of the prediction modes of the plurality of adjacent blocks is different.
In embodiment 1, it has been described that the approximation to the prediction method of the adjacent block differs depending on the kind of the intra prediction mode of the encoding (decoding) target block. Specifically, the characteristics of the similarity to the prediction method of the adjacent block are described for the plane prediction, the case where the prediction direction in the angle prediction is 18 or 50, the case where the prediction direction in the angle prediction is 2 to 17 or 51 to 66, the case where the prediction direction in the angle prediction is 19 to 49, and the DC prediction.
In fig. 9, the weighting parameter w is determined for each of the characteristics of the types of intra prediction modes in accordance with the combination of the prediction modes of the plurality of adjacent blocks.
Specific conditions for the combination of prediction modes of the plurality of adjacent blocks include (1) a case where 2 adjacent blocks, that is, a block adjacent to the block to be encoded (decoded) and a block adjacent to the block to be encoded (decoded) directly above the block to be encoded (decoded), are not available, (2) a case where one of the 2 adjacent blocks, that is, a block adjacent to the block to be encoded (decoded) directly above the block to be encoded (decoded) and a block adjacent to the block to be encoded (decoded) directly left, is unavailable and the other is inter-predicted, (3) a case where both of the 2 adjacent blocks, that is, a block adjacent to the block to be encoded (decoded) directly above the block to be encoded (decoded) and a block adjacent to the block to be encoded (decoded) directly left, are inter-predicted, (4) a case where one of the 2 adjacent blocks, that is, an adjacent to the block to be encoded (decoded) and a block adjacent to the block to be encoded (decoded) directly left is unavailable and the other is intra-predicted, (5) The conditions (1) to (6) are such that one of 2 adjacent blocks, which are a block immediately above the block to be encoded (decoded) and a block immediately to the left of the block to be encoded (decoded), is inter-predicted and the other is intra-predicted, and (6) both of 2 adjacent blocks, which are a block immediately above the block to be encoded (decoded) and a block immediately to the left of the block to be encoded (decoded), are intra-predicted.
Specifically, the conditions (1) to (6) are determined as follows.
First, when the intra prediction mode of the block to be encoded (decoded) is the plane prediction, w may be determined as shown in fig. 9. Specifically, w is set to 1 under the conditions (1) to (3), and w is set to 2 under the conditions (4) to (6). As illustrated in embodiment 1, the plane prediction is intra prediction in which the approximation of the prediction method with the neighboring block is relatively low.
However, even when there is an available intra prediction block in the adjacent block, there is a possibility that plane prediction is performed similarly to the encoding (decoding) target block, and therefore, the approximation to the prediction method of the adjacent block is slightly improved. Therefore, it is preferable that the conditions (4) to (6) be weighted more than the conditions (1) to (3). As a modification, w may be set to 1 under the conditions (1) to (5) and may be set to 2 only under the condition (6). Since plane prediction uses both the upper and left neighboring pixels of the target block in the processing, the approximation of the prediction method improves only when both the prediction modes of the plurality of neighboring blocks are intra prediction, and a modification uses this principle.
Next, when the intra prediction mode of the block to be encoded (decoded) is angular prediction and the prediction direction is 18 or 50, w may be determined as shown in fig. 9. Specifically, w is set to 1 under the conditions (1) to (3), and w is set to 2 under the conditions (4) to (6). When the intra prediction mode is angular prediction and the prediction direction is 18 or 50, the intra prediction has an approximation to a prediction method of an adjacent block in either the horizontal direction or the vertical direction.
Therefore, for the conditions (4) to (6) where there is an available intra prediction block, it is preferable to increase the weight as compared with the conditions (1) to (3). However, the effect of the proximity to the prediction method of the adjacent block is limited to either the horizontal direction or the vertical direction. Therefore, even in the case of the condition (6) in which both of the plurality of adjacent blocks are available, the effect of the approximation to the prediction method of the adjacent blocks is not changed as compared with the conditions (4) and (5) in which only one of the plurality of adjacent blocks is available. Accordingly, in the example of the present figure, w is also set to 2 under the condition (6).
Next, when the intra prediction mode of the block to be encoded (decoded) is angular prediction and the prediction direction is 2 to 17 or 51 to 66, and when the intra prediction mode of the block to be encoded (decoded) is angular prediction and the prediction direction is 2 to 17 or 51 to 66, w may be determined as shown in fig. 9. Specifically, w is 1 under the conditions (1) to (3), 2 under the conditions (4) and (5), and 3 under the condition (6). In the case where the intra prediction mode is angular prediction and the prediction direction is 2 to 17 or 51 to 66, or in the case where the intra prediction mode is angular prediction and the prediction direction is 2 to 17 or 51 to 66, since the prediction direction is diagonal, the intra prediction has an approximation of the prediction method to both the block adjacent to the upper side and the block adjacent to the left side. Thus, there is a tendency that the more intra prediction blocks available in the neighboring blocks are increased, the higher the approximation of the prediction method with the neighboring blocks. Therefore, it is preferable that the value of the weighting parameter w is set to be larger as the intra prediction block available in the neighboring block increases as described above.
Next, when the intra prediction mode of the block to be encoded (decoded) is DC prediction, w may be determined as shown in fig. 9. Specifically, w is set to 2 under the conditions (1) to (3), and w is set to 3 under the conditions (4) to (6). As described in embodiment 1, when DC prediction is selected in the encoding process, the encoding target image is likely to be a planar pattern having a large range of constant brightness. Since this probability is particularly high when available intra prediction blocks are included in the neighboring blocks, it is preferable to make w 3. However, in the example of the present figure, when no available intra prediction block is included in the neighboring blocks, since the possibility that the encoding target image is a pattern of a plane having a wide range of fixed luminance is slightly low, w is set to 2.
According to the above-described process of determining the weighting parameter w in fig. 9, the weighting parameter w can be determined more preferably by considering not only the intra prediction mode of the block to be encoded (decoded) but also the combination of the prediction modes of a plurality of adjacent blocks in the determination of the weighting parameter w in expression 1 in the process of synthesizing prediction.
Further, by using a combination of prediction modes of a plurality of adjacent blocks as a condition, it is possible to perform preferable condition setting without transmitting a new flag.
In the example of fig. 9, the weight w is set to be higher in (6) than in condition (1) regardless of the prediction mode of the block to be encoded (decoded). This is because, regardless of which prediction mode the prediction mode of the encoding (decoding) target block is, there is a tendency that if the intra prediction blocks available in the neighboring blocks are increased, at least the approximation with the prediction method of the neighboring block is improved.
In the example of fig. 9, in the case where neither plane prediction nor DC prediction is specified for intra prediction used in the synthetic prediction, the weight w is increased in the case where an available intra prediction block exists in a plurality of adjacent blocks, as compared with the case where no available intra prediction block exists in the plurality of adjacent blocks.
As a modification of fig. 9, w may be 2 under the conditions (4) and (5) of DC prediction as shown in fig. 10. In the example of fig. 10, w is only the condition (6) of 3 in the DC prediction. This is an example in which the possibility that the encoding target image is a pattern of a plane having a large range of constant brightness is high under the condition (6) that intra prediction is available only for both of a plurality of adjacent blocks. The example of fig. 10 is the same as that of fig. 9 except for the conditions (4) and (5) of DC prediction, and therefore, repetitive description thereof will be omitted.
According to the image encoding device and the image encoding method of embodiment 4 of the present invention described above, an image can be encoded more preferably.
(example 5)
Next, example 5 of the present invention will be described.
In embodiment 5 of the present invention, the image decoding apparatus and the image decoding method in embodiment 2 change the process of determining the weighting parameter w in the process of the synthesis prediction.
Specifically, the process of determining the weighting parameter w is changed from the process of determining the weighting parameter w shown in fig. 8 to the process of determining the weighting parameter w shown in fig. 9 or 10. The image decoding apparatus and the image decoding method according to embodiment 5 use the determination process of the weighting parameter w shown in fig. 9 or fig. 10, and the configuration, operation, and processing other than the determination process of the weighting parameter w shown in fig. 9 or fig. 10 are the same as those of the image decoding apparatus and the image decoding method according to embodiment 2. Therefore, the same points as those of the image decoding apparatus and the image decoding method of embodiment 2 will not be described repeatedly.
The details of the determination process of the weighting parameter w in fig. 9 or 10 are as described in example 4, and therefore, the repetitive description thereof will be omitted.
As shown in fig. 9 and 10, in the image decoding apparatus and the image decoding method according to embodiment 5 of the present invention described above, the weighting parameter w can be determined more preferably by considering not only the intra prediction mode of the block to be encoded (decoded) but also the combination of the prediction modes of a plurality of adjacent blocks in the determination of the weighting parameter w in expression 1 in the process of synthesizing prediction.
According to the image decoding apparatus and the image decoding method of embodiment 5 of the present invention described above, it is possible to decode an image more preferably.
In addition, any combination of the embodiments of the drawings, the methods, and the like described above may be an embodiment of the present invention.
According to the embodiments of the present invention described above, the amount of code can be reduced, and image quality degradation can be prevented. That is, a high compression ratio and a better image quality can be achieved.
Description of the reference numerals
101 … image input section, 102 … block division section, 103 … mode management section, 104 … intra prediction section, 105 … inter prediction section, 106 … block processing section, 107 … transformation/quantization section, 108 … inverse quantization/inverse transformation section, 109 … image synthesis/filter section, 110 … decoded image management section, 111 … entropy coding section, 112 … data output section, 201 … stream analysis section, 202 … block management section, 203 … mode determination section, 204 … intra prediction section, 205 … inter prediction section, 206 … coefficient analysis section, 207 … inverse quantization/inverse transformation section, 208 … image synthesis/filter section, 209 … decoded image management section, 210 … image output section, 120 … synthesis prediction section, 220 … synthesis prediction section.

Claims (4)

1. An image encoding method for encoding an image, comprising:
a predicted image generation step of performing a synthesis process of synthesizing the predicted image for the inter prediction and the predicted image for the intra prediction with respect to the block to be encoded to generate a synthesized predicted image; and
an encoding step of encoding a difference between pixel values of the prediction image generated in the prediction image generation step and the image of the encoding target block,
in the combining process, the weighting process is performed on the inter-predicted image and the intra-predicted image, and the weighting parameter of the intra-predicted image in the weighting process is determined according to a combination of the type of the intra-prediction of the encoding target block and prediction modes of a plurality of adjacent blocks adjacent to the encoding target block, and the type of the intra-prediction of the encoding target block includes planar prediction and DC prediction.
2. An image encoding method for encoding an image, comprising:
a predicted image generation step of performing a synthesis process of synthesizing the inter-predicted image and the intra-predicted image with respect to the block to be encoded to generate a synthesized predicted image; and
an encoding step of encoding a difference between pixel values of the prediction image generated in the prediction image generation step and the image of the encoding target block,
in the combining process, the inter-predicted image and the intra-predicted image are subjected to a weighting process of increasing a weighting of the intra-predicted image in the combining prediction in a case where neither planar prediction nor DC prediction is specified for the intra-prediction used for the combining prediction of the encoding target block, as compared with a case where no available intra-predicted block exists in a plurality of adjacent blocks adjacent to the encoding target block right to and above the encoding target block, in the plurality of adjacent blocks of the encoding target block.
3. An image decoding method for decoding a coded stream obtained by coding an image, comprising:
a predicted image generation step of performing a synthesis process of synthesizing the inter-predicted image and the intra-predicted image with respect to the block to be decoded to generate a synthesized predicted image; and
a decoded image generation step of generating a decoded image based on the prediction image generated in the prediction image generation step and a difference image of the decoding target block,
in the combining process, the weighting process is performed on the inter-predicted image and the intra-predicted image, and the weighting parameter of the intra-predicted image in the weighting process is determined based on a combination of a type of the intra-prediction of the decoding target block and prediction modes of a plurality of adjacent blocks adjacent to the decoding target block, and the type of the intra-prediction of the encoding target block includes plane prediction and DC prediction.
4. An image decoding method for decoding a coded stream obtained by coding an image, comprising:
a predicted image generation step of performing a synthesis process of synthesizing the inter-predicted image and the intra-predicted image with respect to the block to be decoded to generate a synthesized predicted image; and
a decoded image generation step of generating a decoded image based on the predicted image generated in the predicted image generation step and a difference image of the block to be decoded,
in the combining process, the inter-predicted image and the intra-predicted image are subjected to a weighting process of increasing a weighting of the intra-predicted image in the combining prediction in a case where neither planar prediction nor DC prediction is specified for the intra-prediction used for the combining prediction of the decoding target block, as compared with a case where no available intra-predicted block exists in a plurality of adjacent blocks adjacent to the encoding target block right to and above the decoding target block, in the plurality of adjacent blocks of the decoding target block.
CN202080086924.1A 2019-12-17 2020-12-11 Image encoding method and image decoding method Pending CN114788270A (en)

Applications Claiming Priority (3)

Application Number Priority Date Filing Date Title
JP2019226898A JP2021097309A (en) 2019-12-17 2019-12-17 Image coding method and image decoding method
JP2019-226898 2019-12-17
PCT/JP2020/046286 WO2021125083A1 (en) 2019-12-17 2020-12-11 Image encoding method and image decoding method

Publications (1)

Publication Number Publication Date
CN114788270A true CN114788270A (en) 2022-07-22

Family

ID=76431729

Family Applications (1)

Application Number Title Priority Date Filing Date
CN202080086924.1A Pending CN114788270A (en) 2019-12-17 2020-12-11 Image encoding method and image decoding method

Country Status (3)

Country Link
JP (1) JP2021097309A (en)
CN (1) CN114788270A (en)
WO (1) WO2021125083A1 (en)

Also Published As

Publication number Publication date
JP2021097309A (en) 2021-06-24
WO2021125083A1 (en) 2021-06-24

Similar Documents

Publication Publication Date Title
US11936858B1 (en) Constrained position dependent intra prediction combination (PDPC)
US11936856B2 (en) Template matching for JVET intra prediction
US11997305B2 (en) System and method of motion information storage for video coding and signaling
CN114830646A (en) Image encoding method and image decoding method
CN114830642A (en) Image encoding method and image decoding method
WO2021132568A1 (en) Image encoding method and image decoding method
CN114788270A (en) Image encoding method and image decoding method
CN114521325A (en) Image encoding method and image decoding method
CN114830650A (en) Image encoding method and image decoding method
CN114830641A (en) Image encoding method and image decoding method
CN114830645A (en) Image encoding method and image decoding method
CN114830643A (en) Image encoding method and image decoding method
US20230038870A1 (en) Image encoding method and image decoding method
CN114450949A (en) Image encoding method, image encoding device, image decoding method, and image decoding device

Legal Events

Date Code Title Description
PB01 Publication
PB01 Publication
SE01 Entry into force of request for substantive examination
SE01 Entry into force of request for substantive examination