WO2022077490A1 - Procédé de prédiction intra, codeur, décodeur et support de stockage - Google Patents

Procédé de prédiction intra, codeur, décodeur et support de stockage Download PDF

Info

Publication number
WO2022077490A1
WO2022077490A1 PCT/CN2020/121668 CN2020121668W WO2022077490A1 WO 2022077490 A1 WO2022077490 A1 WO 2022077490A1 CN 2020121668 W CN2020121668 W CN 2020121668W WO 2022077490 A1 WO2022077490 A1 WO 2022077490A1
Authority
WO
WIPO (PCT)
Prior art keywords
prediction
intra
filtering
block
reference pixel
Prior art date
Application number
PCT/CN2020/121668
Other languages
English (en)
Chinese (zh)
Inventor
谢志煌
Original Assignee
Oppo广东移动通信有限公司
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Application filed by Oppo广东移动通信有限公司 filed Critical Oppo广东移动通信有限公司
Priority to CN202310996326.3A priority Critical patent/CN116980596A/zh
Priority to MX2023000279A priority patent/MX2023000279A/es
Priority to CN202080106028.7A priority patent/CN116325727A/zh
Priority to PCT/CN2020/121668 priority patent/WO2022077490A1/fr
Publication of WO2022077490A1 publication Critical patent/WO2022077490A1/fr

Links

Images

Classifications

    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04NPICTORIAL COMMUNICATION, e.g. TELEVISION
    • H04N19/00Methods or arrangements for coding, decoding, compressing or decompressing digital video signals
    • H04N19/10Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using adaptive coding
    • H04N19/102Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using adaptive coding characterised by the element, parameter or selection affected or controlled by the adaptive coding
    • H04N19/117Filters, e.g. for pre-processing or post-processing
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04NPICTORIAL COMMUNICATION, e.g. TELEVISION
    • H04N19/00Methods or arrangements for coding, decoding, compressing or decompressing digital video signals
    • H04N19/50Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using predictive coding
    • H04N19/597Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using predictive coding specially adapted for multi-view video sequence encoding
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04NPICTORIAL COMMUNICATION, e.g. TELEVISION
    • H04N19/00Methods or arrangements for coding, decoding, compressing or decompressing digital video signals
    • H04N19/80Details of filtering operations specially adapted for video compression, e.g. for pixel interpolation
    • H04N19/82Details of filtering operations specially adapted for video compression, e.g. for pixel interpolation involving filtering within a prediction loop

Definitions

  • the embodiments of the present application relate to video coding technologies, and relate to, but are not limited to, an intra-frame prediction method, an encoder, a decoder, and a storage medium.
  • Intra-frame prediction may perform an intra-frame prediction process separately for each coding unit to calculate a prediction block.
  • intra-frame prediction usually uses the respective angle mode and non-angle mode to predict the current block to obtain the predicted block, and then selects the optimal prediction of the current block according to the rate-distortion cost information calculated from the predicted block and the original block. mode, and then perform intra prediction based on the prediction mode.
  • the current intra-frame angle prediction does not use all reference pixels, and it is easy to ignore the correlation between some pixels and the current block.
  • the intra-frame prediction filtering technology improves the pixel prediction accuracy through point-to-point filtering, which can effectively enhance the spatial correlation. Therefore, the intra-frame prediction filtering technology is adopted to improve the pixel prediction accuracy through point-to-point filtering, which can effectively enhance the spatial correlation, thereby improving the intra-frame prediction accuracy.
  • the embodiments of the present application provide an intra-frame prediction method, an encoder, a decoder, and a storage medium, which can smooth the prediction process of processed image blocks and improve encoding and decoding efficiency.
  • an embodiment of the present application provides an intra-frame prediction method, which is applied to an encoder, and the method includes:
  • intra-frame prediction filtering Perform intra-frame prediction filtering and intra-frame prediction smoothing filtering on the initial prediction block respectively, to obtain a first-type prediction value and a second-type prediction value; wherein, the intra-frame prediction smoothing filtering adopts at least two adjacent reference values A process of filtering the current block for multiple adjacent reference pixels in each adjacent reference pixel set in the pixel set;
  • the first-type predicted value and the second-type predicted value respectively, perform rate-distortion cost calculation with the original pixel value of the current block, and determine the current prediction mode corresponding to the optimal rate-distortion cost;
  • the index information of the current prediction mode and the filter identifier are written into the code stream, where the filter identifier represents an identifier corresponding to intra-frame prediction filtering and/or intra-frame prediction smoothing filtering.
  • an embodiment of the present application provides an intra-frame prediction method, which is applied to a decoder, and the method includes:
  • the filtering flag indicates that the intra-frame prediction filtering permission flag is valid and the intra-frame prediction smoothing filtering permission flag is valid, at least obtain the current prediction mode and the intra-frame prediction filtering use flag from the code stream;
  • intra-frame prediction filtering use flag When the intra-frame prediction filtering use flag is invalid and a valid intra-frame prediction smoothing filter use flag is obtained from the code stream, based on the current prediction mode, perform intra-frame prediction smoothing filtering processing on the current block to obtain the first a predicted value.
  • an encoder including:
  • the first determining part is configured to traverse the intra prediction mode, and determine the initial prediction value of the initial prediction block corresponding to the current block;
  • the first filtering part is configured to perform intra-frame prediction filtering and intra-frame prediction smoothing filtering respectively on the initial prediction block to obtain a first-type prediction value and a second-type prediction value; wherein, the intra-frame prediction smoothing filtering process A process of filtering the current block for using a plurality of adjacent reference pixels in each adjacent reference pixel set in at least two adjacent reference pixel sets;
  • the first determining part is further configured to use the initial prediction value, the first type prediction value and the second type prediction value to calculate the rate-distortion cost with the original pixel value of the current block, respectively, and determine the corresponding optimal rate-distortion cost.
  • a first prediction part configured to use the current prediction mode to perform intra prediction on the current block
  • the writing part is configured to write the index information of the current prediction mode and the filtering identification into the code stream, wherein the filtering identification represents the identification corresponding to intra-frame prediction filtering and/or intra-frame prediction smoothing filtering.
  • an embodiment of the present application provides a decoder, including:
  • the parsing part is configured to obtain the code stream, parse the header information of the code stream, and obtain the filtering identifier; and when the filtering identifier is an indication that the intra-frame prediction filtering permission flag is valid, and the intra-frame prediction smoothing filtering permission flag is valid, Obtain at least the current prediction mode and intra-frame prediction filtering use identifier from the code stream;
  • the second prediction part is configured to, based on the current prediction mode, frame the current block when the intra-frame prediction filtering use flag is invalid and a valid intra-frame prediction smoothing filter use flag is obtained from the code stream
  • the intra-prediction smoothing filtering process is performed to obtain the first predicted value.
  • an encoder including:
  • the first memory stores a computer program executable on a first processor, which when executed by the first processor implements the intra prediction method of the encoder.
  • an embodiment of the present application also provides a decoder, including:
  • the second memory stores a computer program executable on a second processor, which when executed by the second processor implements the intra prediction method of the decoder.
  • an embodiment of the present application provides a storage medium, including:
  • a computer program is stored thereon, and when the computer program is executed by the first processor, the intra-frame prediction method of the encoder is realized; or, when the computer program is executed by the second processor, the intra-frame prediction method of the decoder is realized. method of prediction.
  • Embodiments of the present application provide an intra-frame prediction method, an encoder, a decoder, and a storage medium.
  • intra-frame prediction filtering in the encoder, it is possible to use at least two adjacent reference pixel sets in each adjacent reference pixel set. The process of filtering the current block of a neighboring reference pixel.
  • Intra-frame prediction smoothing filtering can refer to multiple adjacent reference pixels in one adjacent reference pixel set, so such prediction will be smoother, and the prediction for blurred images will be more accurate, thereby improving coding efficiency. .
  • 1A-1C are exemplary distribution diagrams of components in different color formats provided by the embodiments of the present application.
  • FIG. 2 is a schematic diagram of the division of an exemplary coding unit provided by an embodiment of the present application.
  • FIG. 3 is a schematic diagram of an exemplary prediction mode provided by an embodiment of the present application.
  • 4A-4C are schematic diagrams of three filtering situations of exemplary intra-frame prediction filtering provided by an embodiment of the present application.
  • FIG. 5 is a schematic diagram of the composition and structure of an exemplary network architecture for video encoding and decoding provided by an embodiment of the present application
  • FIG. 6 is a structural diagram of an exemplary video coding system provided by an embodiment of the present application.
  • FIG. 7 is a structural diagram of an exemplary video decoding system provided by an embodiment of the present application.
  • FIG. 8 is a flowchart 1 of an intra-frame prediction method provided by an embodiment of the present application.
  • FIGS. 9A-9C are schematic diagrams of exemplary target reference pixels provided by embodiments of the present application.
  • FIG. 10 is a second flowchart of an intra-frame prediction method provided by an embodiment of the present application.
  • FIG. 11 is a flowchart of an intra-frame prediction method further provided by an embodiment of the present application.
  • FIG. 12 is a schematic structural diagram 1 of an encoder according to an embodiment of the present application.
  • FIG. 13 is a second schematic structural diagram of an encoder according to an embodiment of the application.
  • FIG. 14 is a schematic structural diagram 1 of a decoder according to an embodiment of the present application.
  • FIG. 15 is a second schematic structural diagram of a decoder according to an embodiment of the present application.
  • a first image component, a second image component and a third image component are generally used to represent the current block (Coding Block, CB); wherein, the three image components are a luminance component and a blue chrominance component respectively. and a red chrominance component, specifically, the luminance component is usually represented by the symbol Y, the blue chrominance component is usually represented by the symbol Cb or U, and the red chrominance component is usually represented by the symbol Cr or V; in this way, the video image can use the YCbCr format Representation can also be represented in YUV format.
  • YCbCr YUV
  • YUV ratio is generally measured as 4:2:0, 4:2:2 or 4:4:4
  • Y represents brightness ( Luma)
  • Cb(U) represents blue chromaticity
  • Cr(V) represents red chromaticity
  • U and V represent chromaticity (Chroma) to describe color and saturation.
  • 1A to 1C show the distribution diagrams of each component in different color formats, wherein white is the Y component, and black and gray are the UV components.
  • 4:2:0 means that every 4 pixels has 4 luminance components and 2 chrominance components (YYYYCbCr)
  • 4:2:2 means that every 4 A pixel has 4 luminance components, 4 chrominance components (YYYYCbCrCbCr)
  • 4:4:4 represents a full pixel display (YYYYCbCrCbCrCbCrCbCr).
  • LCUs Large Coding Units
  • CU Coding Unit
  • Prediction Unit PU
  • the hybrid coding framework may include modules such as prediction, transform (Transform), quantization (Quantization), entropy coding (EntropyCoding), and loop filtering (In Loop Filter); wherein, the prediction module may include intraPrediction (intraPrediction) And inter prediction (interPrediction), inter prediction can include motion estimation (motion estimation) and motion compensation (motion compensation). Since there is a strong correlation between adjacent pixels in a frame of a video image, the use of intra-frame prediction in the video coding and decoding technology can eliminate the spatial redundancy between adjacent pixels.
  • Inter-frame prediction can refer to the image information of different frames, and use motion estimation to search for the motion vector information that best matches the current divided block to eliminate temporal redundancy; transformation converts the predicted image block to the frequency domain, redistributes the energy, and combines quantization The information that is not sensitive to the human eye can be removed to eliminate visual redundancy; entropy coding can eliminate character redundancy according to the current context model and the probability information of the binary code stream.
  • the encoder first reads the image information, and divides the image into several coding tree units (Coding Tree Units, CTUs), and a coding tree unit can be further divided into several coding units ( Coding Unit, CU), these coding units can be rectangular blocks or square blocks, and the specific relationship can be referred to as shown in FIG. 2 .
  • CTUs Coding Tree Units
  • CU Coding Unit
  • the current coding unit cannot refer to the information of different frame images, and can only use the adjacent coding units of the same frame image as reference information for prediction, that is, according to most of the current left-to-right, top-to-bottom
  • the current coding unit can refer to the upper left coding unit, the upper coding unit and the left coding unit are used as reference information to predict the current coding unit, and the current coding unit is used as the reference information of the next coding unit. image for prediction.
  • the input digital video is in color format, that is, the input source of the current mainstream digital video encoder is YUV 4:2:0 format, that is, every 4 pixels of the image consists of 4 Y components and 2 UV components.
  • the Y component and UV component will be encoded separately, and the encoding tools and technologies used are slightly different.
  • the decoding end will also decode according to different formats.
  • the current block is mainly predicted with reference to the image information of adjacent blocks of the current frame, the residual information is obtained by calculating the residual between the predicted block and the original image block, and then the process of transformation and quantization is performed.
  • the residual information is transmitted to the decoding end.
  • the decoding end receives and parses the code stream, the residual information is obtained through inverse transformation and inverse quantization, and the reconstructed image block is obtained by superimposing the residual information on the predicted image block predicted by the decoding end.
  • intra prediction usually uses the respective angle mode and non-angle mode to predict the current coding block to obtain the prediction block, and selects the optimal prediction mode for the current coding unit according to the rate-distortion information calculated from the prediction block and the original block. , and then transmit the prediction mode to the decoding end via the code stream.
  • the decoding end parses the prediction mode, predicts the prediction image of the current decoding block, and superimposes the residual pixels transmitted by the code stream to obtain the reconstructed image.
  • the H.264/AVC standard has only 8 angle prediction modes and 1 non-angle prediction mode; H.265/HEVC is extended to 33 angle prediction modes and 2 non-angle prediction modes; and the latest general video coding standard H .266/VVC adopts 67 prediction modes, among which 2 non-angle prediction modes are reserved, and the angle mode is expanded from 33 kinds of H.265 to 65 kinds.
  • the intra-frame prediction will be more accurate, and it will be more in line with the needs of the current society for the development of high-definition and ultra-high-definition video.
  • the domestic digital audio and video coding standard 3 (Audio Video coding Standard, AVS3) also continues to expand the angle mode and non-angle mode.
  • AVS3 Analog Video coding Standard
  • IPF Intra Prediction Filter
  • the intra prediction filtering technology points out that not all reference pixels are used in the current intra-frame angle prediction, and it is easy to ignore some pixels and
  • the correlation between the current coding units, the intra-frame prediction filtering technology improves the pixel prediction accuracy through point-to-point filtering, which can effectively enhance the spatial correlation, thereby improving the intra-frame prediction accuracy.
  • the IPF technology takes the prediction mode from upper right to lower left in AVS3 as an example, as shown in FIG. 3 .
  • the direction of the prediction mode is from upper right to lower left, and the generated prediction value of the current coding unit mainly uses the reference pixels of the adjacent blocks in the upper MRB row, that is, the prediction pixels of the current coding unit do not refer to the reconstructed pixels of the adjacent blocks on the left.
  • the current coding unit and the left reconstruction block are in a spatially adjacent relationship. If only the upper MRB pixels are referenced instead of the left URB pixels, the spatial correlation is likely to be lost, resulting in poor prediction effect.
  • the intra-frame prediction filter technology is a filtering method applied to all prediction modes of intra-frame prediction to improve the accuracy of intra-frame prediction.
  • IPF is mainly realized through the following processes:
  • the intra-frame prediction filtering technology judges the current prediction mode of the coding unit and divides it into a horizontal angle prediction mode, a vertical angle prediction mode and a non-angle prediction mode;
  • IPF technology uses different filters to filter input pixels
  • the IPF technology uses different filter coefficients to filter the input pixels
  • the input pixels of the IPF technology are the predicted pixels obtained in each prediction mode, and the output pixels are the final predicted pixels after IPF filtering.
  • the IPF technology has an allowable flag bit ipf_enable_flag (intra-frame prediction filtering permission flag), which is an exemplary, binary variable.
  • a value of '1' indicates that intra-frame prediction filtering can be used; a value of '0' indicates that intra-frame prediction filtering should not be used.
  • the IPF technology also uses the flag ipf_flag (intra-frame prediction filtering use flag), which is an exemplary, binary variable.
  • a value of '1' indicates that intra-frame prediction filtering should be used; a value of '0' indicates that intra-frame prediction filtering should not be used.
  • Predictive filtering if the flag bit ipf_flag does not exist in the code stream, the default value is 0.
  • the IPF technology classifies prediction modes 0, 1 and 2 as non-angle prediction modes, and uses the first three-tap filter to filter the prediction pixels; Mode 50 is classified as a vertical angle-like prediction mode, and the predicted pixels are filtered using the first two-tap filter; The second two-tap filter filters the predicted pixels. in,
  • P'(x, y) is the final prediction value of the pixel located at the position (x, y) of the current chrominance prediction block
  • f(x) and f(y) are the horizontal filtering of the reconstructed pixels with reference to the adjacent blocks on the left, respectively.
  • the coefficient and the vertical filter coefficient of the reconstructed pixel of the adjacent block on the reference upper side, P(-1,y) and P(x,-1) are the reconstructed pixels on the left side of row y and the reconstructed pixels on the upper side of column x, respectively
  • P(x,y) is the original predicted pixel value in the current chrominance component prediction block.
  • the values of x and y do not exceed the range of the width and height of the current coding unit.
  • the values of the above-mentioned horizontal filter coefficients and vertical filter coefficients are related to the size of the current coding unit and the distances from the predicted pixels in the current prediction block to the left reconstructed pixels and the top reconstructed pixels.
  • the above-mentioned horizontal filter coefficients may be divided into different filter coefficient groups according to the size of the current coding unit.
  • Table 1 presents the filter coefficients of the intra-frame prediction filtering technique.
  • FIGS. 4A to 4C show schematic diagrams of three filtering situations of intra-frame prediction filtering, respectively only referring to the upper reference pixels to filter the prediction value in the current coding unit (that is, the upper two-tap filtering in FIG. 4A ); Only the left reference pixel is referenced to filter the measurements in the current coding unit pre-filter (ie, the left two-tap filtering of FIG. 4B ); and both the upper and left reference pixels are referenced to the prediction in the current coding unit block values are filtered (ie, three-tap filtering on the upper and left sides of Figure 4C).
  • the IPF technology acts on the intra-frame prediction part of the video coding hybrid framework, and is specifically applied to all prediction modes of intra-frame prediction, acting on both the encoding end and the decoding end.
  • FIG. 5 is a schematic diagram of the composition and structure of the network architecture for video encoding and decoding according to an embodiment of the present application, as shown in FIG. 5 .
  • the network architecture includes one or more electronic devices 12 to 1N and a communication network 01 , wherein the electronic devices 12 to 1N can perform video interaction through the communication network 01 .
  • the electronic device can be various types of devices with video codec functions, for example, the electronic device can include mobile phones, tablet computers, personal computers, personal digital assistants, navigators, digital phones, video phones, Televisions, sensing devices, servers, etc., are not limited in the embodiments of the present application.
  • the intra-frame prediction apparatus in the embodiment of the present application may be the above-mentioned electronic device.
  • the electronic device in the embodiments of the present application has a video encoding and decoding function, and generally includes a video encoder (ie, encoder) and a video decoder (ie, decoder).
  • a video encoder ie, encoder
  • a video decoder ie, decoder
  • the video coding system 11 includes:
  • CTU Coding Tree Unit
  • inter prediction unit 115 is used to perform inter prediction encoding of the received video reconstruction block relative to one or more blocks in one or more reference frames to provide temporal Prediction information; where, motion estimation is the process of generating a motion vector that can estimate the motion of the video reconstruction block, and then motion compensation is performed based on the motion vector determined by motion estimation; After the mode, the inter prediction unit 115 is also used to provide the selected inter prediction data to the encoding unit 119, and also send the calculated motion vector data to the encoding unit 119; in addition, the inverse quantization unit 116 and the inverse The transform unit 117 is used for the reconstruction of the video reconstruction block, reconstructing the residual block in the pixel domain, the reconstructed residual block is passed through the loop filtering unit 118 to remove the blocking artifacts, and then, the reconstructed residual block is A predictive block added to the frame of the decoded image buffer unit 110 to generate a reconstructed video reconstruction block; the encoding unit 119 is used to encode
  • the decoded image buffer unit 110 is used for storing reconstructed video reconstruction blocks for prediction reference. As the video image encoding proceeds, new reconstructed video reconstruction blocks are continuously generated, and these reconstructed video reconstruction blocks are all stored in the decoded image buffer unit 110 .
  • FIG. 7 is a schematic structural diagram of a video encoding system according to an embodiment of the present application.
  • the video encoding system 12 includes:
  • the intra-frame prediction first traverses all prediction modes, calculates the predicted pixels under each intra-frame prediction mode, and calculates the rate-distortion cost according to the original pixels;
  • this process opens the IPF technology, first calculates the predicted pixels under each intra-frame prediction mode, and obtains the prediction block of the current coding unit;
  • the IPF current coding unit identification position is no, indicating that the IPF technology is not used, and is also transmitted to the decoding end through the code stream; if the rate-distortion in e) is smaller, then the prediction mode index code recorded in e) is used as the current code.
  • the optimal prediction mode of the unit is transmitted to the decoding end through the code stream, and the current IPF coding unit identification position is true, indicating that the IPF technology is used, and is also transmitted to the decoding end
  • the predicted value is superimposed with the residual information after operations such as transformation and quantization to obtain the reconstructed block of the current coding unit as the reference information of the subsequent coding unit.
  • the decoding end obtains the code stream and parses to obtain the digital video sequence information, and parses to obtain the IPF permission flag of the current video sequence, the coding mode of the current decoding unit is the intra-frame prediction coding mode, and the IPF usage flag of the current decoding unit.
  • the reconstructed block of the current decoding unit is obtained by superimposing the restored residual information of the prediction block, and output after post-processing.
  • the intra-frame prediction filtering technology can effectively improve the coding efficiency of intra-frame prediction, greatly enhance the spatial correlation of intra-frame prediction, and solve the problem of ignoring unused reference pixels in the process of intra-frame prediction. Impact.
  • the intra-frame prediction filtering technology and the current intra-frame prediction mode cannot solve similar problems well, and pixel-by-pixel filtering based on reference pixels can improve the correlation between the prediction block and the reference block.
  • it cannot solve the smoothing problem inside the prediction block. That is to say, a prediction block calculated according to a single prediction mode usually shows a better prediction effect in an image with a clearer texture, and therefore the residual error will be smaller and smaller, and the coding efficiency will be improved.
  • over-sharpened prediction may lead to increased and larger residual errors, poor prediction effect and reduced coding efficiency.
  • the present technology is implemented on the prediction block when performing intra-frame prediction, and for some image blocks that need smoothing processing, the present application proposes an intra-frame prediction method based on the smoothing-processing intra-frame prediction filtering enhancement technology.
  • An intra-frame prediction method provided in this embodiment of the present application mainly acts on the intra-frame prediction unit 114 of the video coding system 11 and the intra-frame prediction unit 123 of the video decoding system 12;
  • the encoding system 11 can obtain a better prediction effect through the intra-frame prediction method provided in the embodiment of the present application, and correspondingly, at the decoding end, the video decoding and restoration quality can also be improved.
  • An embodiment of the present application provides an intra-frame prediction method, which is applied to a video encoding device, that is, an encoder.
  • the functions implemented by the method can be implemented by the processor in the video encoding device calling program codes, and of course the program codes can be stored in a computer storage medium.
  • the video encoding device includes at least a processor and a storage medium.
  • the current coding unit and the current decoding unit are both represented by the current block.
  • FIG. 8 is a schematic flowchart of an implementation of an intra-frame prediction method according to an embodiment of the present application. As shown in FIG. 8 , the method includes:
  • an intra-frame prediction smoothing filtering technology (Intra Prediction Smooth, IPS) is proposed, and on the basis of IPF, the IPS technology can be superimposed to realize intra-frame prediction. That is, when both IPF and IPS are allowed, the intra-frame prediction process on the encoder side is performed.
  • IPS Intra Prediction Smooth
  • a video image may be divided into multiple image blocks, and each image block to be encoded currently may be called an encoding block, wherein each encoding block may include a first image component, a second image component, and a first image component. Three image components; and the current block is the coding block in the video image currently to be predicted for the first image component, the second image component or the third image component.
  • the current block performs the first image component prediction, and the first image component is a luminance component, that is, the image component to be predicted is a luminance component, then the current block may also be called a luminance block; or, it is assumed that the current block performs the second image component prediction prediction, and the second image component is a chrominance component, that is, the image component to be predicted is a chrominance component, then the current block may also be called a chrominance block.
  • the encoder determines the to-be-predicted image component of the current block; based on the parameters of the current block, various prediction modes are used to predict and encode the to-be-predicted image component to obtain an initial predicted value, and then calculate various predictions based on the initial predicted value.
  • the rate-distortion cost corresponding to each prediction mode in the mode select the minimum rate-distortion cost (ie, the first minimum rate-distortion cost) from the multiple calculated rate-distortion costs, and determine the minimum rate-distortion cost as the optimal rate-distortion cost cost, the prediction mode corresponding to the optimal rate-distortion cost is determined as the prediction mode of the current block.
  • the encoder after determining that the encoder traverses all intra-frame prediction modes, determines the initial prediction value of the initial prediction block corresponding to the current block, and then uses the IPF and IPS technologies to filter the current block, the first block is obtained.
  • the first-class predicted value and the second-class predicted value, the initial predicted value, the first-class predicted value and the second-class predicted value are used to calculate the rate-distortion cost with the original pixel value of the current block respectively, and the three corresponding minimum values are determined.
  • the prediction mode corresponding to changing the optimal rate-distortion cost is the current prediction mode.
  • intra-frame prediction is performed on the current block to complete the prediction of the image component to be predicted.
  • intra prediction smoothing filtering is a process of filtering the current block by using multiple adjacent reference pixels in each adjacent reference pixel set in at least two adjacent reference pixel sets.
  • Intra-frame prediction filtering is a process of filtering the current block by using one adjacent reference pixel in each adjacent reference pixel set in at least two adjacent reference pixel sets.
  • Intra-frame prediction smoothing filtering can refer to multiple adjacent reference pixels in one adjacent reference pixel set, so such prediction will be smoother, and the prediction for blurred images will be more accurate, thereby improving coding efficiency. .
  • the encoder adopts IPF and IPS technology, and after filtering the current block, the process of obtaining the first type of prediction value and the second type of prediction value: one is to obtain the initial prediction block corresponding to the current block.
  • the initial prediction block is processed by IPF and IPS respectively; the other is that the encoder traverses all intra prediction modes again, and only calculates the initial prediction block in each intra prediction mode when IPF is turned on. , and then perform IPF filtering on the initial prediction block to determine the first type of prediction value.
  • the encoder traverses all intra-frame prediction modes again, and only calculates the initial prediction block in each intra-frame prediction mode when IPS is turned on, and then performs IPS filtering on the initial prediction block to determine the second type of prediction value.
  • the application examples are not limited.
  • the encoder performs intra-frame prediction filtering and intra-frame prediction smoothing filtering on the initial prediction block respectively, and the process of obtaining the first type prediction value and the second type prediction value may be: The left adjacent reference pixel point and the upper adjacent reference pixel point of each pixel in the prediction block, perform intra-frame prediction filtering on each pixel of the initial prediction block to obtain the first type of prediction value; and the encoder is based on the initial prediction value. According to the preset horizontal distance and the preset vertical distance corresponding to each pixel in the prediction block, the initial prediction block is subjected to intra-frame prediction smoothing filtering processing to obtain the second type of prediction value.
  • the encoder performs IPF filtering on the prediction block of the current block, selects the corresponding filter according to the currently traversed prediction mode, and selects the corresponding filter coefficient group according to the size of the current block.
  • Table 1 after the filter coefficient is determined, the prediction block is filtered based on the filter coefficient to obtain the rate-distortion cost of each prediction mode, and the smallest one is the second minimum rate-distortion cost.
  • the encoder may determine, based on the preset horizontal distance and the preset vertical distance corresponding to each pixel in the initial prediction block, at least two phases corresponding to each pixel. At least two adjacent reference block ranges of the adjacent reference pixel set; from the at least two adjacent reference pixel sets, correspondingly determine at least two target reference pixel sets belonging to the at least two adjacent reference block ranges; wherein, each The target reference pixel sets include at least one target reference pixel; based on the at least two target reference pixel sets, intra-frame prediction smoothing filtering is performed on each pixel to obtain a second-type prediction value.
  • the process of the encoder correspondingly determining at least two target reference pixel sets belonging to the at least two adjacent reference pixel sets from the at least two adjacent reference pixel sets may be:
  • Manner 1 The encoder determines, from at least two adjacent reference pixel sets, pixels at two boundary pixel positions of each adjacent reference block range as a target reference pixel set, thereby determining at least two target reference pixel sets.
  • Mode 2 The encoder determines, from at least two adjacent reference pixel sets, M pixels at any pixel position within the range of each adjacent reference block as a target reference pixel set, thereby determining at least two target reference pixel sets , where M is greater than 0 and does not exceed the maximum number of pixels in the corresponding adjacent reference pixel set.
  • the at least two adjacent reference pixel sets include: a left adjacent reference pixel set and an upper adjacent reference pixel set;
  • the at least two adjacent reference block ranges include: a left adjacent reference pixel set A pixel range and an upper adjacent reference pixel range;
  • the at least two target reference pixel sets include: a left target adjacent reference pixel set and an upper target adjacent reference pixel set.
  • the at least two adjacent reference pixel sets include: when the left adjacent reference pixel set and the upper adjacent reference pixel set, the encoder determines the preset horizontal distance and the preset vertical distance corresponding to the current pixel to The current pixel of the current block is used as the reference, and the preset horizontal distance is moved left and right to obtain the first horizontal pixel position and the second horizontal pixel position; and the preset vertical distance is moved up and down respectively to obtain the first vertical pixel position and the second vertical pixel position. pixel location. Based on the first horizontal pixel position and the second horizontal pixel position, it is mapped to the upper adjacent reference pixel set, and the first boundary limit and the second boundary limit of the pixel position are determined.
  • the pixel position between is the upper adjacent reference pixel range corresponding to the current pixel corresponding to the upper adjacent reference pixel set. Based on the first vertical pixel position and the second vertical pixel position, it is mapped to the left adjacent reference pixel set. , the third boundary limit and the fourth boundary limit of the pixel position are determined, and the pixel position between the third boundary limit and the fourth boundary limit is the left adjacent reference pixel set corresponding to the current pixel corresponding to the left adjacent pixel set Reference pixel range, so that at least two adjacent reference block ranges are obtained. For each adjacent reference pixel set, the encoder may select at least one adjacent reference pixel (ie, the target reference pixel set) from the reference pixels included in the corresponding adjacent reference pixel range for performing intra prediction, thereby At least two target reference pixel sets are selected.
  • the encoder may select at least one adjacent reference pixel (ie, the target reference pixel set) from the reference pixels included in the corresponding adjacent reference pixel range for performing intra prediction, thereby At least two target reference pixel sets are
  • the encoder does this for each pixel of the current block, so as to complete the processing of the current block, and for the current pixel, the encoder selects a target reference pixel set for an adjacent reference pixel set
  • the number of reference pixels can be one or more, which is determined based on the position of the current pixel in the current block.
  • the encoder can select the reference pixel value corresponding to the boundary position of each adjacent reference pixel range to form the target reference pixel set, or can select any M reference pixels from each adjacent reference pixel range to form the target reference pixel.
  • the corresponding determined number of M reference pixels may be different or may be the same, which is not limited in this embodiment of the present application.
  • the target reference pixel set corresponding to the upper adjacent reference pixel set may be UPNum
  • the target reference pixel set corresponding to the left adjacent reference pixel set may be LFNum
  • UPNum and LFNum may be the same or different.
  • the preset horizontal distance is less than or equal to the width of the initial prediction block; the preset vertical distance is less than or equal to the height of the initial prediction block.
  • the preset horizontal distance and the preset vertical distance corresponding to the first type of pixels in the position of n rows and m columns of the initial prediction block are different from those of other types except n rows and m columns of the initial prediction block.
  • the preset horizontal distance and the preset vertical distance corresponding to the pixels are different; n and m are natural numbers; the preset horizontal distance and the preset vertical distance corresponding to each pixel in the first type of pixels are at least partially the same, or both are different; other The preset horizontal distance and the preset vertical distance corresponding to each pixel in the class of pixels are at least partially the same, or both are different.
  • the preset horizontal distance and the preset vertical distance are equal, and the specific values can be set in the encoder and decoder according to the actual situation, or can be preset in the encoder and decoder according to the size of the current prediction block.
  • the embodiments of the present application are not limited.
  • the n rows and m columns of the initial prediction block may be the first n rows and the first m columns, or may be in other forms, which are not limited in the embodiments of the present application.
  • the above-mentioned IPS technology may have different filtering processes for the first n rows and the first m columns of the initial prediction block.
  • the first n rows and the first m columns may use the reference pixels corresponding to the first positional relationship
  • other non-first n rows and m columns may use the reference pixels corresponding to the first positional relationship.
  • the reference pixels corresponding to the second positional relationship may be used for the filtering of the predicted pixels of the row and the first m columns.
  • the first three rows and the first three columns use the same HD and VD, which is 1, while the first three rows and the first three columns use the same HD and VD, which is 2.
  • n and m do not exceed the size of the current block, and n and m may be the same or different, which are not limited in this embodiment of the present application.
  • the encoder may determine filter coefficients according to the principle mechanism of IPF; based on at least two target reference pixel sets and filter coefficients, perform intra-frame prediction smoothing filtering on each pixel to obtain the second type Predictive value.
  • the filter coefficients corresponding to the first type of pixels in the position of n rows and m columns of the initial prediction block are different from the filter coefficients corresponding to other types of pixels except n rows and m columns of the initial prediction block; n and m are natural numbers;
  • the filter coefficients corresponding to each pixel in one type of pixels are at least partially the same, or all different; the filter coefficients corresponding to each pixel in other types of pixels are at least partially the same, or all different.
  • the IPS technology may have multiple different filter coefficients, for example, the coefficients used by the current predicted pixel in the first n rows and the first m columns are different from the filter coefficients used in the non-first n rows and non-first m columns, etc. , and the filter coefficients corresponding to each pixel in the first n rows and the first m columns may also be the same, partly the same, or all different; the filter coefficients corresponding to each pixel in the non-first n rows and non-first m columns It can also be the same, partially the same, or all different.
  • the encoder may determine a weight coefficient; based on at least two target reference pixel sets and the weight coefficient, weighted summation is performed on each pixel to complete intra-frame prediction After smoothing and filtering, the second type of predicted value is obtained.
  • the IPS technology can have multiple different weight coefficients, such as the coefficients used by the current predicted pixel in the first n rows and the first m columns are different from the weight coefficients used in the non-first n rows and non-first m columns, etc., and,
  • the weight coefficients corresponding to each pixel in the first n rows and the first m columns can also be the same, partially the same, or different; the weight coefficients corresponding to each pixel in the non-first n rows and non-first m columns are also Can be the same, partially the same, or all different.
  • the encoder uses the position of the current predicted pixel as a reference, and uses the horizontal distance HD and vertical distance VD as offsets to obtain UPNum reconstruction reference pixels adjacent to the upper side and LFNum reconstruction reference pixels on the left side, and obtain The reference pixel and the current predicted pixel are filtered, or the weighted summation can be used to obtain the accurate predicted pixel as the final predicted pixel of the current position in the predicted block.
  • a is the current pixel
  • b and c are the adjacent reference pixel ranges
  • the encoder obtains 4 reference pixels and predicted pixels.
  • a 5-tap filter or weighting is formed to obtain the accurately corrected predicted pixel value.
  • the coefficients of the first two rows and the first two columns of the four reference pixels can be set to be 10.
  • the weighting coefficient of the predicted pixels is 216, and the sum needs to be divided by 256 or shifted to the right by 8 bits, and the VD and HD of the first two rows and columns can be set to 1; set 4 reference pixels not the first two rows
  • the weighting coefficients of the first two columns and the non-corrected ones are both 13, and the uncorrected predicted pixel coefficient is 204, and the sum needs to be divided by 256 or shifted to the right by 8 bits.
  • a schematic diagram of 4 target reference pixels with a distance of 1 a is the current pixel, b and c are the adjacent reference pixel ranges, and the encoder obtains 4 reference pixels and predicted pixels to form a 5-tap Filter or weight to obtain the accurately corrected predicted pixel value.
  • the weighting coefficient of the 4 reference pixels can be set to be 7, and the weighting coefficient of the imprecisely corrected predicted pixel is 228. The sum of Divide by 256 or shift right by 8 bits.
  • a is the current pixel
  • b and c are the adjacent reference pixel ranges
  • the encoder obtains two reference pixels and predicted pixels to form a 3-tap Filter or weight to obtain the accurately corrected predicted pixel value.
  • the weighting coefficient of the two reference pixels can be set to be 14, and the weighting coefficient of the uncorrected predicted pixel is 228. The sum of Divide by 256 or shift right by 8 bits.
  • the predicted pixels obtained by the encoder performing the IPS technology may be different from other cases, for example, the second type of prediction of the predicted pixels used by the IPS is obtained by using 4-tap different filter coefficients or different filter coefficients of other taps. value etc.
  • the encoder uses IPS technology to filter the current block
  • the reconstructed reference pixels need to be acquired, that is, the coordinates of the reference pixels in the target reference pixel set exceed the upper adjacent reference pixel set or the left side
  • the size range of the adjacent reference pixel set can also be filled with the adjacent reference pixel set of the current block, and then the target reference pixel set is selected.
  • the encoder uses the leftmost reference pixel in the upper adjacent reference pixel set to pad to the left until the upper phase is determined. until the first boundary of the range of adjacent reference blocks, and/or, using the rightmost reference pixel in the set of adjacent reference pixels on the upper side to fill to the right, until the second boundary of the range of adjacent reference blocks on the upper side is determined.
  • the preset vertical distance is less than or equal to the height of the initial prediction block
  • the lowermost reference pixel in the set of adjacent reference pixels on the left side is used to fill down until the third boundary limit of the range of the adjacent reference block on the left side is determined. and/or, the uppermost reference pixel in the set of adjacent reference pixels on the left side is used to fill upwards until the fourth boundary limit of the range of the adjacent reference block on the left side is determined.
  • the encoder fills the leftmost reconstruction reference pixel with the uppermost adjacent reconstructed pixel to the left until the required range is satisfied, and uses the rightmost reconstruction reference pixel of the upper adjacent reconstruction pixel to fill to the right until the requirement is met.
  • the requested range use the reconstructed reference pixel on the uppermost side of the adjacent reconstructed pixels on the left to fill up until the required range is satisfied, and use the adjacent reconstructed pixels on the left to fill down until the required range is satisfied.
  • the above filling process can also be replaced by a conditional judgment method, and the pixels beyond the range can directly request the pixels at the most boundary, etc., which are not limited in this embodiment of the present application.
  • the encoder performs IPS filtering on the prediction block of the current block, selects the corresponding filter according to the currently traversed prediction mode, determines the filter coefficient, and filters the prediction block based on the filter coefficient to obtain each filter coefficient.
  • the rate-distortion cost of a prediction mode, and the smallest one is selected to be the third smallest rate-distortion cost.
  • the encoder after the encoder obtains the predicted value of the first type and the predicted value of the second type, the encoder adopts the initial predicted value, the predicted value of the first type and the predicted value of the second type, which are respectively the same as the current predicted value.
  • the original pixel value of the block is subjected to rate-distortion cost calculation, and the first minimum rate-distortion cost corresponding to the initial prediction value, the second minimum rate-distortion cost corresponding to the first type of prediction value, and the first minimum rate-distortion cost corresponding to the second type of prediction value are determined.
  • Three minimum rate-distortion costs determine the minimum optimal rate-distortion cost from the first minimum rate-distortion cost, the second minimum rate-distortion cost and the third minimum rate-distortion cost; take the prediction mode corresponding to the obtained optimal rate-distortion cost as Current forecast mode.
  • the encoder can use the current prediction mode to perform intra-frame prediction on the current block, and write the index information of the current prediction mode and the filter flag into the code stream, where the filter flag represents intra-frame prediction filtering and/or intra-frame prediction Predictive smoothing filter corresponding flag.
  • the filtering flags include: an intra-frame prediction filtering permission flag, an intra-frame prediction filtering use flag, an intra-frame prediction smoothing filtering permission flag, and an intra-frame prediction smoothing filtering use flag.
  • the encoder determines that the intra-frame prediction filtering permission flag is valid, and the intra-frame prediction smoothing filtering permission flag is valid, and sets the intra-frame prediction filtering permission flag to be valid. Allows markers and intra-frame prediction smoothing filtering to allow markers to be written into the stream;
  • the transmission of the intra-frame prediction filtering use flag and the intra-frame prediction smoothing filter use flag follows one of the following transmission modes to the decoder:
  • the optimal rate-distortion cost is the second minimum rate-distortion cost, determine that the intra-frame prediction filtering use flag is valid, and write the intra-frame prediction filtering use flag into the code stream;
  • the optimal rate-distortion cost is the third minimum rate-distortion cost
  • the optimal rate-distortion cost is the first minimum rate-distortion cost
  • a binary number may be used to indicate whether it is valid, for example, 1 indicates valid, 0 indicates invalid, or 0 indicates valid, 1 indicates invalid, etc.
  • the specific setting method and setting value are not limited in this embodiment of the present application.
  • the encoder implements the process of filtering the current block by using one adjacent reference pixel in each adjacent reference pixel set in at least two adjacent reference pixel sets by using intra-frame prediction filtering.
  • Intra-frame prediction smoothing filtering can refer to multiple adjacent reference pixels in one adjacent reference pixel set, so such prediction will be smoother, and the prediction for blurred images will be more accurate, thereby improving coding efficiency. .
  • the filtering identifiers of the s are the same identifier, or use one-to-one corresponding identifiers.
  • the IPS technology is applicable to each encoding channel, wherein the luminance component and the chrominance component respectively use independent identification bits to indicate whether the IPS technology is used or not.
  • the encoder when the encoder uses one filter flag for three coding channels, it can save bits and data amount of transmitted flag bits, and reduce the computational complexity.
  • the encoder may first determine the current block to see if it meets the conditions for IPS processing.
  • the intra-frame prediction method proposed in this application is adopted. in,
  • the initial prediction block is divided to obtain prediction sub-blocks that meet the prediction area threshold range; the intra-frame prediction processing flow is implemented for the prediction sub-block.
  • the range of the predicted area threshold can be no less than 4 ⁇ 4 or 8 ⁇ 8, and no more than 64 ⁇ 64 or 32 ⁇ 32, and the detailed value can be set according to the actual situation. If the area is too small, IPS processing can not be performed, but if the area is too large, it can be divided to meet the conditions for using IPS before processing.
  • the encoder limits the use range of the IPS, and does not use the IPS technology for the unit with a small prediction block area, so as to reduce the transmission identification bits and reduce the computational complexity.
  • the encoder after determining the initial prediction value of the initial prediction block corresponding to the current block, the encoder adopts the current prediction mode, and before performing intra prediction on the current block, the encoder may Type, the prediction mode set is determined, and the prediction mode set is the prediction mode that enables the intra-frame prediction smoothing filtering technology; for the initial prediction value, the initial sub-prediction value predicted by the prediction mode set is used to perform intra-frame prediction smoothing filtering.
  • the second type of prediction value; the initial prediction value and the second type of prediction value are used to calculate the rate-distortion cost with the original pixel value of the current block respectively, and the current prediction mode corresponding to the optimal rate-distortion cost is determined.
  • the encoder limits the scope of use of the IPS, filters the prediction modes of the current block, and only uses the IPS technology for part of the prediction modes, so as to reduce transmission flag bits and reduce computational complexity.
  • the manner in which the encoder obtains the predicted value of the first type and the predicted value of the second type may further include:
  • the encoder After traversing the intra-frame prediction mode and determining the initial prediction value of the initial prediction block corresponding to the current block, the encoder uses the initial prediction value, the first-type prediction value and the second-type prediction value to compare with the original pixel value of the current block respectively. Calculate the rate-distortion cost. Before determining the current prediction mode corresponding to the optimal rate-distortion cost, perform intra-frame prediction filtering on the initial prediction block to obtain the first type of prediction value; traverse the intra-frame prediction mode and use intra-frame prediction smoothing filtering. Perform prediction estimation on the current block to obtain the second type of prediction value.
  • the encoder can traverse the intra prediction mode again, and based on the preset horizontal distance and preset vertical distance corresponding to each pixel in the current block, perform intra prediction smoothing filtering processing on the current block to obtain the second type of prediction value; without filtering the initial predicted value.
  • the implementation principle of the filter processing performed on the current block by the encoder is the same as the principle of filtering the initial prediction block in the foregoing embodiment, and details are not described herein again.
  • the encoder obtains coding information, including intra-frame prediction filtering (IPF) allowable identification bits and the allowable identification bits of this technical solution (IPS), etc., and after obtaining the image information, the image is divided into several CTUs, and further divided into several CTUs.
  • IPF intra-frame prediction filtering
  • IPS technical solution
  • the IPF allowable flag bit and the IPS allowable flag bit are both '1', perform all the following steps; if the IPF allowable flag bit is '1' and the IPS allowable flag bit is '0', Then only a), b), c), d), e), and h) are executed; if the IPF allowable flag is '0' and the IPS allowable flag is '1', only a), b), f), g), and h); if both the IPF allow flag and the IPS allow flag are '0', only a), b) and h) are performed. If the above-mentioned IPF or IPS allows the flag bit to be '0', then the coding end defaults to the use flag bit of the corresponding technology of all coding blocks to be '0':
  • the current coding unit traverses all intra-frame prediction modes, calculates the prediction block under each prediction mode, and calculates the rate-distortion cost information of the current prediction mode according to the original pixel block;
  • this process uses IPF technology, first calculates the predicted pixels under each intra-frame prediction mode, and obtains the prediction block of the current coding unit;
  • this process uses IPS technology, calculates the predicted pixels under each intra-frame prediction mode, and calculates the predicted pixel process using IPS technology, which can be added for several tap interpolation filters.
  • IPS technology calculates the predicted pixels under each intra-frame prediction mode, and calculates the predicted pixel process using IPS technology, which can be added for several tap interpolation filters.
  • Several reference pixels are used as new interpolation filters to predict the predicted pixels of the process.
  • the above-mentioned several tap interpolation filters can be 8 tap filters, and the above-mentioned several reference pixels can be 4.
  • the above-mentioned several reference pixels are obtained.
  • the process is as follows: taking the position of each predicted pixel as a reference, and taking the horizontal distance HD and vertical distance VD as offsets, obtain UPNum reconstruction reference pixels adjacent to the upper side and LFNum reconstruction reference pixels on the left side.
  • the above HD and VD can be 0, 1, 2, 3, ..., N, etc., and N does not exceed the width or height of the current block.
  • the above UPNum and LFNum can be 0, 1, 2, 3, ..., M, etc., and M does not exceed the number of adjacent reconstruction reference pixels on the upper side of the current block and the number of adjacent reconstruction reference pixels on the left side;
  • the IPF allows the identification bit to be '0' and the IPS allows the identification bit to be '0', then transmit the prediction mode index recorded in b) to the decoding end through the code stream; if the IPF allows the identification bit to be '1' and If the IPS allows the flag bit to be '0', compare the minimum cost value recorded in b) with the minimum cost value recorded in e), and if the rate-distortion cost in b) is smaller, compare the prediction recorded in b)
  • the mode index coding is used as the optimal prediction mode of the current coding unit and is transmitted to the decoding end through the code stream, and the IPF current coding unit identification position is '0', indicating that the IPF technology is not used, and is also transmitted to the decoding end through the code stream; if e) in The rate distortion is smaller, then the prediction mode index code recorded in e) is transmitted to the decoding end as the optimal prediction mode of the current coding unit through the code stream, and the IPF current
  • the code stream is transmitted to the decoding end, and the IPF usage flag position of the current coding unit is '0' and the IPS usage flag position is '1', indicating that the IPF technology is not used but the IPS technology is used, and is also transmitted to the decoding end through the code stream.
  • the encoder superimposes the predicted block and the inversely transformed and inversely quantized residuals to obtain a reconstructed coding unit block, which is used as a prediction reference block for the next coding block.
  • An embodiment of the present application provides an intra-frame prediction method, as shown in FIG. 10 , including:
  • S202 Perform intra-frame prediction smoothing filtering processing on the initial prediction block respectively, to obtain a second type of prediction value; wherein, the intra-frame prediction smoothing filtering is performed by using the multi-dimensional data in each adjacent reference pixel set in at least two adjacent reference pixel sets. The process of filtering the current block with adjacent reference pixels.
  • S205 Write the index information of the current prediction mode and the filter flag into the code stream, where the filter flag represents the flag corresponding to the intra-frame prediction smoothing filter.
  • the encoder can only determine whether to perform IPS, and it is not dependent on the implementation of the IPF technology, so that the intra-frame prediction smoothing filtering permission flag is valid.
  • the encoder may only perform IPS, so that when the IPS technology is enabled, the encoder performs intra-frame prediction smoothing filtering on the initial prediction blocks respectively to obtain the second type of prediction value, and the encoder only needs to use the initial prediction block.
  • the predicted value and the second type of predicted value are respectively calculated with the original pixel value of the current block to calculate the rate-distortion cost, and the current prediction mode corresponding to the optimal rate-distortion cost can be determined.
  • the detailed filtering processing method is consistent with the implementation of the foregoing embodiment, and is not repeated here.
  • the filtering flag includes: an intra-frame prediction smoothing filtering permission flag and an intra-frame prediction smoothing filtering use flag.
  • the encoder writes the index information of the current prediction mode into the code stream; determines that the intra-frame prediction smoothing filtering permission flag is valid, and writes the intra-frame prediction smoothing filtering permission flag into the code stream.
  • Intra-frame prediction smoothing filtering uses one of the following transmission modes:
  • the optimal rate-distortion cost is the third minimum rate-distortion cost, it is determined that the intra-frame prediction smoothing filtering use flag is valid, and the intra-frame prediction smoothing filtering use flag is written into the code stream.
  • the optimal rate-distortion cost is the first minimum rate-distortion cost, it is determined that the intra-frame prediction smoothing filtering use flag is invalid, and the intra-frame prediction smoothing filtering use flag is written into the code stream.
  • the encoder obtains coding information, including intra-frame prediction filtering (IPF) allowable identification bits and the allowable identification bits of this technical solution (IPS), etc., and after obtaining the image information, the image is divided into several CTUs, and further divided into several CTUs.
  • IPF intra-frame prediction filtering
  • IPS technical solution
  • the IPF allowable flag bit and the IPS allowable flag bit are both '1', perform all the following steps; if the IPF allowable flag bit is '1' and the IPS allowable flag bit is '0', Then only execute a), b), c), d), e), and i); if the IPF allow flag bit is '0' and the IPS allow flag bit is '1', then only execute a), b), f), g), h), and i); if both the IPF allow flag and the IPS allow flag are '0', only execute a), b) and i). If the above-mentioned IPF or IPS allows the flag bit to be '0', then the coding end defaults to the use flag bit of the corresponding technology of all coding blocks to be '0':
  • the current coding unit traverses all intra-frame prediction modes, calculates the prediction block under each prediction mode, and calculates the rate-distortion cost information of the current prediction mode according to the original pixel block;
  • this process uses IPF technology, first calculates the predicted pixels under each intra-frame prediction mode, and obtains the prediction block of the current coding unit;
  • this process uses IPS technology, first calculates the predicted pixels in each intra-frame prediction mode, and obtains the prediction block of the current coding unit.
  • the interpolation filter for calculating intra-frame predicted pixels in this process may be different from steps a) and c), and a stronger filter can be used to obtain smoother predicted pixels.
  • the specific implementation can be an 8-tap interpolation filter for the angle prediction mode, Exemplarily, the filter coefficients may be shown in the following 4 groups, and each group may include 32 types of 8-tap filter coefficients.
  • the prediction block of the current coding unit is accurately corrected for the predicted pixels, and the accurate correction process can be based on the position where the current predicted pixels are located, and the horizontal distance HD and the vertical distance VD are used as offsets to obtain the adjacent UPNum on the upper side.
  • Reconstruction reference pixels and LFNum reconstruction reference pixels on the left filter the obtained reference pixels and the current prediction pixel, or perform weighted summation to obtain the accurate prediction pixel as the final prediction pixel at the current position in the prediction block.
  • the above HD and VD can be 0, 1, 2, 3, ..., N, etc., and N does not exceed the width or height of the current block.
  • the above UPNum and LFNum can be 0, 1, 2, 3, ..., M, etc., and M does not exceed the number of adjacent reconstruction reference pixels on the upper side of the current block and the number of adjacent reconstruction reference pixels on the left side;
  • the IPF allows the identification bit to be '0' and the IPS allows the identification bit to be '0', then transmit the prediction mode index recorded in b) to the decoding end through the code stream; if the IPF allows the identification bit to be '1' and If the IPS allows the flag bit to be '0', compare the minimum cost value recorded in b) with the minimum cost value recorded in e), and if the rate-distortion cost in b) is smaller, compare the prediction recorded in b)
  • the mode index coding is used as the optimal prediction mode of the current coding unit and is transmitted to the decoding end through the code stream, and the IPF current coding unit identification position is '0', indicating that the IPF technology is not used, and is also transmitted to the decoding end through the code stream; if e) in The rate distortion is smaller, then the prediction mode index code recorded in e) is transmitted to the decoding end as the optimal prediction mode of the current coding unit through the code stream, and the IPF current
  • the stream is transmitted to the decoding end, and the IPF use identification position of the current coding unit is '0' and the IPS use identification position is '1', indicating that the IPF technology is not used but the IPS technology is used, and is also transmitted to the decoding end through the code stream.
  • the encoder superimposes the prediction block and the inversely transformed and inversely quantized residuals to obtain a reconstructed coding unit block, which is used as a prediction reference block for the next coding unit.
  • the encoder adopts intra-frame prediction smoothing filtering technology, which strengthens the spatial correlation of intra-frame prediction, and multiple reference pixels in a neighboring reference pixel set are smoothed and filtered, avoiding the possibility of over-sharp prediction. It will cause the residual error to increase and become larger, the prediction effect is not good, and the coding efficiency is improved.
  • An embodiment of the present application provides an intra-frame prediction method, which is applied to a video decoding device, that is, a decoder.
  • the functions implemented by the method can be implemented by the processor in the video decoding device calling program codes, and of course the program codes can be stored in a computer storage medium.
  • the video encoding device includes at least a processor and a storage medium.
  • FIG. 11 is a schematic flowchart of an implementation of an intra-frame prediction method according to an embodiment of the present application. As shown in FIG. 11 , the method includes:
  • the decoder can parse from the header information of the code stream to determine whether the IPF and IPS processing is performed. That is, the intra-frame prediction filtering permission flag in the filtering flags, and/or the intra-frame prediction smoothing filtering permission flag can be parsed out.
  • the filtering flags are indicative of the Intra prediction filtering permission flag being valid, and the intra prediction smoothing filtering permission flag being valid.
  • the decoder can obtain at least the current prediction mode and the intra-frame prediction filtering use flag from the code stream.
  • the intra-frame prediction filtering use flag is invalid, and the valid intra-frame prediction smoothing filter use flag is obtained from the code stream
  • identifying that is, when the current block adopts the current prediction mode determined by the IPS technology, based on the current prediction mode, perform intra-frame prediction smoothing filtering on the current block to obtain the first prediction value, and then based on the first prediction value.
  • the predicted value goes through the subsequent decoding process.
  • the current prediction mode is obtained from the code stream; based on the current prediction mode, intra prediction filtering is performed on the current block to obtain the second prediction value.
  • the decoder performs intra-frame prediction smoothing filtering processing on the current block based on the preset horizontal distance and the preset vertical distance corresponding to each pixel in the current block to obtain the first predicted value.
  • At least two adjacent sets of at least two adjacent reference pixels corresponding to each pixel are determined based on a preset horizontal distance and a preset vertical distance corresponding to each pixel in the current block reference block range;
  • each target reference pixel set includes at least one target reference pixel
  • intra-frame prediction smoothing filtering is performed on each pixel to obtain a first predicted value.
  • the decoder correspondingly determines, from the at least two adjacent reference pixel sets, at least two target reference pixel sets belonging to the range of the at least two adjacent reference blocks, including the following manners :
  • Manner 1 From the at least two adjacent reference pixel sets, determine the pixels at the two boundary pixel positions of each adjacent reference block range as a target reference pixel set, so as to determine the at least two target reference pixels gather.
  • Mode 2 From the at least two adjacent reference pixel sets, determine M pixels at any pixel position within the range of each adjacent reference block as a target reference pixel set, so as to determine the at least two target reference pixel sets. A set of pixels, where M is greater than 0 and does not exceed the maximum number of pixels in the corresponding adjacent reference pixel set.
  • the at least two adjacent reference pixel sets include: a left adjacent reference pixel set and an upper adjacent reference pixel set;
  • the at least two adjacent reference block ranges include: a left adjacent reference pixel range and an upper adjacent reference pixel range;
  • the at least two target reference pixel sets include: a left target adjacent reference pixel set and an upper target adjacent reference pixel set.
  • the preset horizontal distance is less than or equal to the width of the initial prediction block; the preset vertical distance is less than or equal to the height of the initial prediction block;
  • the preset horizontal distance and the preset vertical distance corresponding to each pixel in the first type of pixels are at least partially the same, or both are different;
  • the preset horizontal distance and the preset vertical distance corresponding to each pixel in the other types of pixels are at least partially the same, or both are different.
  • performing an intra-frame prediction smoothing filtering process on each pixel based on at least two target reference pixel sets to obtain a second type of prediction value including: a decoder determining a filter coefficient; and based on at least two The target refers to the set of pixels and filter coefficients, and performs intra-frame prediction smoothing filtering processing on each pixel to obtain the second type of prediction value.
  • the filter coefficients corresponding to the first type of pixels at the positions of n rows and m columns in the initial prediction block are different from the filter coefficients corresponding to other types of pixels except for n rows and m columns in the initial prediction block; n and m are natural numbers;
  • the filter coefficients corresponding to each pixel in the first type of pixels are at least partially the same, or all different;
  • the filter coefficients corresponding to each pixel in the other types of pixels are at least partially the same, or all different.
  • the leftmost reference pixel in the upper adjacent reference pixel set is used to pad to the left until the upper adjacent reference is determined until the first boundary of the block range, and/or, using the rightmost reference pixel in the upper adjacent reference pixel set to fill to the right, until the second boundary of the upper adjacent reference block range is determined;
  • the lowermost reference pixel in the set of adjacent reference pixels on the left side is used to fill down until the third boundary limit of the range of the adjacent reference block on the left side is determined, And/or, the uppermost reference pixel in the set of adjacent reference pixels on the left side is used to fill upwards until the fourth boundary limit of the range of the adjacent reference block on the left side is determined.
  • the decoder obtains the code stream, and parses the code stream to obtain the IPF permission flag and the IPS permission flag of the current video sequence.
  • the allowable flag bit of the IPF and the allowable flag bit of the IPS are both '1', perform all the following steps; if the allowable flag bit of the IPF is '1' and the allowable flag bit of the IPS is ' 0', only perform steps a), b), c), d) and g); if the allowable flag bit of IPF is '0' and the allowable flag bit of IPS is '1', then only a), b ), e), f) and g) steps; if the allowable identification bit of IPF and the allowable identification bit of IPS are both '0', then only steps a), b) and g) are performed. If the above-mentioned IPF or IPS allows the flag bit to be '0', the decoding end defaults to the use flag bit of the corresponding technology of all decoding blocks to be '0':
  • step f obtain the use identification bit of IPF, if the use identification bit of IPF is '1', then skip the remaining process of this step and skip step f); if the use identification bit of IPF is '0', then parse and obtain IPS usage flag. If the use flag bit of IPS is '0', then no additional operation is performed on the current prediction block, and step f) is skipped; if the use flag bit of IPS is '1', then f) is executed;
  • the prediction block of the current coding unit is accurately corrected for the predicted pixels.
  • the accurate correction process can be based on the position where the current predicted pixels are located, and the horizontal distance HD and the vertical distance VD are used as offsets to obtain the adjacent UPNum on the upper side.
  • Reconstruction reference pixels and LFNum reconstruction reference pixels on the left filter the obtained reference pixels and the current prediction pixel, or perform weighted summation to obtain the accurate prediction pixel as the final prediction pixel at the current position in the prediction block.
  • the above HD and VD can be 0, 1, 2, 3, ..., N, etc., and N does not exceed the width or height of the current block.
  • the above UPNum and LFNum can be 0, 1, 2, 3, ..., M, etc., and M does not exceed the number of adjacent reconstruction reference pixels on the upper side of the current block and the number of adjacent reconstruction reference pixels on the left side;
  • step g) superimposing the residual information restored in step a) of the prediction block to obtain the reconstructed block of the current decoding unit, which is output after post-processing.
  • the implementation on the decoder side may also adopt the following implementation manners, which are not limited in this embodiment of the present application. As follows:
  • the decoder obtains the code stream, and parses the code stream to obtain the IPF allowable flag bit and the IPS allowable flag bit of the current video sequence.
  • the allowable flag bit of the IPF and the allowable flag bit of the IPS are both '1', perform all the following steps; if the allowable flag bit of the IPF is '1' and the allowable flag bit of the IPS is ' 0', then only execute steps a), b), and d); if the allowable identification bit of IPF is '0' and the allowable identification bit of IPS is '1', then only steps a), c) and d) are executed ; If the allowable identification bit of IPF and the allowable identification bit of IPS are both '0', then only steps a) and d) are performed. If the above-mentioned IPF or IPS allows the flag bit to be '0', the decoding end defaults to the use flag bit of the corresponding technology of all decoding blocks to be '0':
  • the process can use an interpolation filter that is different from the generation of predicted pixels in steps d) and e), and can use an 8-tap interpolation filter plus use the location of each predicted pixel.
  • interpolation filters for generating predicted pixels, and the specific process of obtaining the above-mentioned several reconstructed reference pixels is: taking the horizontal distance HD and the vertical distance VD as offsets, obtaining the adjacent UPNum on the upper side. Reconstructed reference pixels and left LFNum reconstructed reference pixels.
  • the above HD and VD can be 0, 1, 2, 3, ..., N, etc., and N does not exceed the width or height of the current block.
  • the above UPNum and LFNum can be 0, 1, 2, 3, ..., M, etc., and M does not exceed the number of adjacent reconstruction reference pixels on the upper side of the current block and the number of adjacent reconstruction reference pixels on the left side;
  • step g) superimposing the residual information restored in step a) of the prediction block to obtain the reconstructed block of the current decoding unit, which is output after post-processing.
  • the decoder implements the process of filtering the current block by using one adjacent reference pixel in each adjacent reference pixel set in at least two adjacent reference pixel sets by using intra-frame prediction filtering.
  • intra-frame prediction smoothing filtering can refer to multiple adjacent reference pixels in one adjacent reference pixel set, so such prediction will be smoother and the prediction for blurred images will be more accurate, thereby improving decoding efficiency.
  • an encoder 1 including:
  • the first determination part 10 is configured to traverse the intra prediction mode, and determine the initial prediction value of the initial prediction block corresponding to the current block;
  • the first filtering part 11 is configured to perform intra-frame prediction filtering and intra-frame prediction smoothing filtering respectively on the initial prediction block to obtain a first-type prediction value and a second-type prediction value; wherein, the intra-frame prediction smoothing Filtering is a process of filtering the current block by using a plurality of adjacent reference pixels in each adjacent reference pixel set in at least two adjacent reference pixel sets;
  • the first determining part 10 is further configured to use the initial prediction value, the first type prediction value and the second type prediction value to calculate the rate-distortion cost with the original pixel value of the current block, respectively, to determine the optimal rate-distortion cost. the corresponding current prediction mode;
  • a first prediction part 12 configured to use the current prediction mode to perform intra-frame prediction on the current block
  • the writing part 13 is configured to write the index information of the current prediction mode and the filter flag into the code stream, wherein the filter flag represents the flag corresponding to intra-frame prediction filtering and/or intra-frame prediction smoothing filtering.
  • the first filtering part 11 is further configured to, based on the adjacent reference pixels on the left and the adjacent reference pixels on the upper side of each pixel in the initial prediction block, filter Perform intra-frame prediction filtering on each pixel of the initial prediction block to obtain the first type of prediction value; and based on the preset horizontal distance and preset vertical distance corresponding to each pixel in the initial prediction block, determine the The initial prediction block is subjected to intra-frame prediction smoothing filtering to obtain the second type of prediction value.
  • the first filtering part 11 is further configured to, based on the preset horizontal distance and the preset vertical distance corresponding to each pixel in the initial prediction block, determine the at least two adjacent reference block ranges of at least two adjacent reference pixel sets corresponding to each pixel; from the at least two adjacent reference pixel sets, correspondingly determined to belong to the at least two adjacent reference pixel sets At least two target reference pixel sets within the block range; wherein, each target reference pixel set includes at least one target reference pixel; based on the at least two target reference pixel sets, intra prediction smoothing is performed on each pixel filtering to obtain the predicted value of the second type.
  • the first filtering part 11 is further configured to determine, from the at least two adjacent reference pixel sets, pixels at two boundary pixel positions of each adjacent reference block range is one target reference pixel set, thereby determining the at least two target reference pixel sets.
  • the first filtering part 11 is further configured to determine, from the at least two adjacent reference pixel sets, M arbitrary pixel positions within the range of each adjacent reference block The pixel is one target reference pixel set, so that the at least two target reference pixel sets are determined, wherein M is greater than 0 and does not exceed the maximum number of pixels of the corresponding adjacent reference pixel sets.
  • the at least two adjacent reference pixel sets include: a left adjacent reference pixel set and an upper adjacent reference pixel set;
  • the at least two adjacent reference block ranges include: a left adjacent reference pixel range and an upper adjacent reference pixel range;
  • the at least two target reference pixel sets include: a left target adjacent reference pixel set and an upper target adjacent reference pixel set.
  • the preset horizontal distance is less than or equal to the width of the initial prediction block; the preset vertical distance is less than or equal to the height of the initial prediction block;
  • the preset horizontal distance and the preset vertical distance corresponding to the first type of pixels in the position of n rows and m columns of the initial prediction block are the same as those corresponding to other types of pixels except n rows and m columns of the initial prediction block.
  • the preset horizontal distance is different from the preset vertical distance; n and m are natural numbers;
  • the preset horizontal distance corresponding to each pixel in the first type of pixels is at least partially the same as the preset vertical distance, or both are different;
  • the preset horizontal distance and the preset vertical distance corresponding to each of the other types of pixels are at least partially the same, or both are different.
  • the first filtering part 11 is further configured to determine a filter coefficient; based on the at least two target reference pixel sets and the filter coefficient, perform intra-frame filtering on each pixel A prediction smoothing filtering process is performed to obtain the second type of prediction value.
  • the filter coefficients corresponding to the first type of pixels in the positions of n rows and m columns of the initial prediction block are corresponding to other types of pixels except for n rows and m columns of the initial prediction block.
  • Different filter coefficients; n and m are natural numbers;
  • the filter coefficients corresponding to each pixel in the first type of pixels are at least partially the same, or all different;
  • the filter coefficients corresponding to each pixel in the other types of pixels are at least partially the same, or all different.
  • the first determining part 10 is further configured to use the initial prediction value, the first type prediction value and the second type prediction value to perform rate-distortion costing on the original pixel values of the current block, respectively.
  • the filtering flags include: intra-frame prediction filtering permission flags, intra-frame prediction filtering use flags, intra-frame prediction smoothing filtering permission flags, and intra-frame prediction smoothing filtering use flags;
  • the writing part 13 is further configured to write the index information of the current prediction mode into the code stream; determine that the intra-frame prediction filtering permission flag is valid, and the intra-frame prediction smoothing filtering permission flag is valid, and writing the intra-frame prediction filtering permission flag and the intra-frame prediction smoothing filtering permission flag into the code stream;
  • the optimal rate-distortion cost is the second minimum rate-distortion cost, determining that the intra-frame prediction filtering use flag is valid, and writing the intra-frame prediction filtering use flag into the code stream;
  • the optimal rate-distortion cost is the third minimum rate-distortion cost
  • the prediction filtering uses the flag and the intra-frame prediction smoothing filtering uses the flag to write the code stream;
  • the optimal rate-distortion cost is the first minimum rate-distortion cost
  • the predictive filtering uses the flag and the intra prediction smoothing filtering uses the flag to write the code stream.
  • the first filtering part 11 is further configured to traverse the intra-frame prediction mode, and after determining the initial prediction value of the initial prediction block corresponding to the current block, use the current block Prediction mode, before performing intra-frame prediction on the current block, perform intra-frame prediction smoothing filtering processing on the initial prediction block respectively, to obtain the second type of prediction value; wherein, the intra-frame prediction smoothing filter adopts at least two A process of filtering the current block for multiple adjacent reference pixels in each adjacent reference pixel set in the adjacent reference pixel set;
  • the first determining part 10 is further configured to use the initial predicted value and the second type of predicted value to perform rate-distortion cost calculation with the original pixel value of the current block, respectively, to determine the rate-distortion cost corresponding to the optimal rate-distortion cost. the current prediction mode.
  • the first filtering part 11 is further configured to traverse the intra-frame prediction mode, and after determining the initial prediction value of the initial prediction block corresponding to the current block, use the current block prediction mode, before performing intra prediction on the current block, determine a prediction mode set according to the type of intra prediction mode, and the prediction mode set is a prediction mode enabling intra prediction smoothing filtering technology; for the initial prediction In the value, the initial sub-prediction value predicted by the prediction mode set is used to perform intra-frame prediction smoothing filtering processing to obtain the second type of prediction value;
  • the first determining part 10 is further configured to use the initial predicted value and the second type of predicted value to perform rate-distortion cost calculation with the original pixel value of the current block, respectively, to determine the rate-distortion cost corresponding to the optimal rate-distortion cost. the current prediction mode.
  • the filtering flags include: an intra-frame prediction smoothing filtering permission flag and an intra-frame prediction smoothing filtering use flag;
  • the writing part 13 is further configured to write the index information of the current prediction mode into the code stream; determine that the intra-frame prediction smoothing filtering permission flag is valid, and write the intra-frame prediction smoothing filtering permission flag incoming code stream;
  • the optimal rate-distortion cost is the third minimum rate-distortion cost, it is determined that the intra-frame prediction smoothing filter usage flag is valid, and the intra-frame prediction smoothing filter usage flag is written into the code stream.
  • the optimal rate-distortion cost is the first minimum rate-distortion cost, it is determined that the intra-frame prediction smoothing filter usage flag is invalid, and the intra-frame prediction smoothing filter usage flag is written into the code stream.
  • the encoding type of the initial prediction value corresponds to multiple encoding channels
  • the filtering identifiers corresponding to each encoding channel use the same identifier, or use one-to-one corresponding identifiers.
  • the first filtering part 11 is further configured to allow the use of intra-frame prediction smoothing to perform filtering when the size of the initial prediction block satisfies the prediction area threshold range; When the size is greater than the upper limit of the prediction area threshold range, the initial prediction block is divided to obtain prediction sub-blocks satisfying the prediction area threshold range; the intra prediction processing flow is implemented for the prediction sub-block.
  • the first filtering part 11 is further configured to traverse the intra-frame prediction mode, and after determining the initial prediction value of the initial prediction block corresponding to the current block, use the initial prediction value , the first type of prediction value and the second type of prediction value, respectively carry out rate-distortion cost calculation with the original pixel value of the current block, and before determining the current prediction mode corresponding to the optimal rate-distortion cost, the initial prediction block is respectively framed
  • the intra-prediction filtering is performed to obtain the first type of prediction value; the intra-frame prediction mode is traversed, and the intra-frame prediction smoothing filtering process is used to predict and estimate the current block to obtain the second type of prediction value.
  • the first filtering part 11 is further configured to traverse the intra prediction mode, and based on the preset horizontal distance and the preset vertical distance corresponding to each pixel in the current block, The current block is subjected to intra-frame prediction smoothing filtering to obtain the second type of prediction value.
  • the first determining part 10 is further configured to, when the preset horizontal distance is greater than the width of the initial prediction block, use the leftmost reference pixel set in the upper adjacent reference pixel set
  • the reference pixels on the upper side are filled to the left until the first boundary limit of the range of the upper adjacent reference block is determined, and/or, the rightmost reference pixel in the upper adjacent reference pixel set is used to fill to the right , until the second boundary limit of the upper adjacent reference block range is determined;
  • the lowermost reference pixel in the set of adjacent reference pixels on the left side is used to fill down, until the range of the adjacent reference block on the left side is determined. Until the third boundary limit, and/or, use the uppermost reference pixel in the left adjacent reference pixel set to fill upwards until the fourth boundary limit of the left adjacent reference block range is determined.
  • the encoder implements the process of filtering the current block by using one adjacent reference pixel in each adjacent reference pixel set in at least two adjacent reference pixel sets by using intra-frame prediction filtering.
  • Intra-frame prediction smoothing filtering can refer to multiple adjacent reference pixels in one adjacent reference pixel set, so such prediction will be smoother, and the prediction for blurred images will be more accurate, thereby improving coding efficiency. .
  • an encoder including:
  • the first memory 15 stores a computer program that can be run on the first processor 14 , and the first processor 14 executes the program for an intra-frame prediction method on the encoder side.
  • an embodiment of the present application provides a decoder 2, including:
  • the parsing part 20 is configured to obtain the code stream, parse the header information of the code stream, and obtain the filtering identification; and when the filtering identification indicates that the intra-frame prediction filtering permission flag is valid, and the intra-frame prediction smoothing filtering permission flag is valid , at least obtain the current prediction mode and intra prediction filtering use identifier from the code stream;
  • the second prediction part 21 is configured to, based on the current prediction mode, perform a prediction on the current block when the intra-frame prediction filtering use flag is invalid and a valid intra-frame prediction smoothing filter use flag is obtained from the code stream.
  • the intra-frame prediction is smoothed and filtered to obtain the first predicted value.
  • the parsing part 20 is further configured to at least acquire the current prediction mode and the intra-frame prediction filtering use flag from the code stream, when the intra-frame prediction filtering use flag is valid When , obtain the current prediction mode from the code stream;
  • the second prediction part 21 is further configured to perform intra-frame prediction filtering processing on the current block based on the current prediction mode to obtain a second prediction value.
  • the second prediction part 21 is further configured to frame the current block based on a preset horizontal distance and a preset vertical distance corresponding to each pixel in the current block
  • the intra-prediction smoothing filtering process is performed to obtain the first predicted value.
  • the second prediction part 21 is further configured to, based on the preset horizontal distance and the preset vertical distance corresponding to each pixel in the current block, determine a at least two adjacent reference block ranges of at least two adjacent reference pixel sets corresponding to each pixel;
  • each target reference pixel set includes at least one target reference pixel
  • intra-frame prediction smoothing filtering is performed on each pixel to obtain the first predicted value.
  • the second prediction part 21 is further configured to determine, from the at least two adjacent reference pixel sets, pixels at two boundary pixel positions of each adjacent reference block range is one target reference pixel set, thereby determining the at least two target reference pixel sets.
  • the second prediction part 21 is further configured to determine, from the at least two adjacent reference pixel sets, M arbitrary pixel positions within the range of each adjacent reference block The pixel is one target reference pixel set, so that the at least two target reference pixel sets are determined, wherein M is greater than 0 and does not exceed the maximum number of pixels of the corresponding adjacent reference pixel sets.
  • the at least two adjacent reference pixel sets include: a left adjacent reference pixel set and an upper adjacent reference pixel set;
  • the at least two adjacent reference block ranges include: a left adjacent reference pixel range and an upper adjacent reference pixel range;
  • the at least two target reference pixel sets include: a left target adjacent reference pixel set and an upper target adjacent reference pixel set.
  • the preset horizontal distance is less than or equal to the width of the initial prediction block; the preset vertical distance is less than or equal to the height of the initial prediction block;
  • the preset horizontal distance and the preset vertical distance corresponding to the first type of pixels in the position of n rows and m columns of the initial prediction block are the same as those corresponding to other types of pixels except n rows and m columns of the initial prediction block.
  • the preset horizontal distance is different from the preset vertical distance; n and m are natural numbers;
  • the preset horizontal distance corresponding to each pixel in the first type of pixels is at least partially the same as the preset vertical distance, or both are different;
  • the preset horizontal distance and the preset vertical distance corresponding to each of the other types of pixels are at least partially the same, or both are different.
  • the second prediction part 21 is further configured to determine filter coefficients; based on the at least two target reference pixel sets and the filter coefficients, perform intra-frame performance on each pixel A prediction smoothing filtering process is performed to obtain the second type of prediction value.
  • the filter coefficients corresponding to the first type of pixels in the positions of n rows and m columns of the initial prediction block are corresponding to other types of pixels except for n rows and m columns of the initial prediction block.
  • Different filter coefficients; n and m are natural numbers;
  • the filter coefficients corresponding to each pixel in the first type of pixels are at least partially the same, or all different;
  • the filter coefficients corresponding to each pixel in the other types of pixels are at least partially the same, or all different.
  • the leftmost reference pixel in the set of adjacent reference pixels on the upper side is used to pad to the left, until it is determined that the upper until the first boundary of the range of the adjacent reference blocks on the side, and/or, use the rightmost reference pixel in the set of adjacent reference pixels on the upper side to fill to the right, until the range of the adjacent reference blocks on the upper side is determined. until the second boundary;
  • the lowermost reference pixel in the set of adjacent reference pixels on the left side is used to fill down, until the range of the adjacent reference block on the left side is determined. Until the third boundary limit, and/or, use the uppermost reference pixel in the left adjacent reference pixel set to fill upwards until the fourth boundary limit of the left adjacent reference block range is determined.
  • the decoder implements the process of filtering the current block by using one adjacent reference pixel in each adjacent reference pixel set in at least two adjacent reference pixel sets by using intra-frame prediction filtering.
  • intra-frame prediction smoothing filtering can refer to multiple adjacent reference pixels in a set of adjacent reference pixels, so such prediction will be smoother, and the prediction for blurred images will be more accurate, thereby improving decoding efficiency.
  • an embodiment of the present application further provides a decoder, including:
  • the second memory 22 stores a computer program that can be executed on the second processor 23, and the second processor 23 implements the intra-frame prediction method on the decoder side when executing the program.
  • an embodiment of the present application provides a storage medium on which a computer program is stored.
  • the computer program is executed by the first processor, an intra-frame prediction method of the encoder is implemented; or, the computer program is processed by the second processor.
  • the decoder executes, implements the decoder's intra prediction method.
  • Intra-frame prediction smoothing filtering can refer to multiple adjacent reference pixels in one adjacent reference pixel set, so such prediction will be smoother, and the prediction for blurred images will be more accurate, thereby improving coding efficiency.

Landscapes

  • Engineering & Computer Science (AREA)
  • Multimedia (AREA)
  • Signal Processing (AREA)
  • Compression Or Coding Systems Of Tv Signals (AREA)

Abstract

Des modes de réalisation de la présente demande utilisent un procédé de prédiction intra, un codeur, un décodeur et un support de stockage, le procédé comprenant les étapes consistant à : traverser de modes de prédiction intra pour déterminer une valeur de prédiction initiale d'un bloc de prédiction initiale correspondant à un bloc actuel ; réaliser un filtrage de prédiction intra et un filtrage lisse de prédiction intra respectivement sur le bloc de prédiction initiale pour obtenir une valeur de prédiction de premier type et une valeur de prédiction de second type, le filtrage lisse de prédiction intra étant un processus de filtrage réalisé sur le bloc actuel à l'aide de multiples pixels adjacents de référence dans chacun des ensembles d'au moins deux ensembles de pixels adjacents de référence ; utiliser la valeur de prédiction initiale, la valeur de prédiction de premier type et la valeur de prédiction de second type pour calculer un coût de distorsion de débit d'une valeur de pixel d'origine du bloc actuel et déterminer un mode de prédiction actuelle correspondant à un coût de distorsion de débit optimal ; réaliser une prédiction intra sur le bloc actuel à l'aide du mode de prédiction actuelle ; et écrire des informations d'indice du mode de prédiction actuelle et un identifiant de filtrage dans un flux de codes, l'identifiant de filtrage représentant un identifiant correspondant à un filtrage de prédiction intra et/ou à un filtrage lisse de prédiction intra.
PCT/CN2020/121668 2020-10-16 2020-10-16 Procédé de prédiction intra, codeur, décodeur et support de stockage WO2022077490A1 (fr)

Priority Applications (4)

Application Number Priority Date Filing Date Title
CN202310996326.3A CN116980596A (zh) 2020-10-16 2020-10-16 一种帧内预测方法、编码器、解码器及存储介质
MX2023000279A MX2023000279A (es) 2020-10-16 2020-10-16 Metodo de intra prediccion, codificador, decodificador y medio de almacenamiento.
CN202080106028.7A CN116325727A (zh) 2020-10-16 2020-10-16 一种帧内预测方法、编码器、解码器及存储介质
PCT/CN2020/121668 WO2022077490A1 (fr) 2020-10-16 2020-10-16 Procédé de prédiction intra, codeur, décodeur et support de stockage

Applications Claiming Priority (1)

Application Number Priority Date Filing Date Title
PCT/CN2020/121668 WO2022077490A1 (fr) 2020-10-16 2020-10-16 Procédé de prédiction intra, codeur, décodeur et support de stockage

Publications (1)

Publication Number Publication Date
WO2022077490A1 true WO2022077490A1 (fr) 2022-04-21

Family

ID=81208894

Family Applications (1)

Application Number Title Priority Date Filing Date
PCT/CN2020/121668 WO2022077490A1 (fr) 2020-10-16 2020-10-16 Procédé de prédiction intra, codeur, décodeur et support de stockage

Country Status (3)

Country Link
CN (2) CN116325727A (fr)
MX (1) MX2023000279A (fr)
WO (1) WO2022077490A1 (fr)

Cited By (1)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN114598880A (zh) * 2022-05-07 2022-06-07 深圳传音控股股份有限公司 图像处理方法、智能终端及存储介质

Citations (5)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
WO2016091122A1 (fr) * 2014-12-11 2016-06-16 Mediatek Inc. Procédé et appareil de prédiction intra en codage vidéo
CN107148778A (zh) * 2014-10-31 2017-09-08 联发科技股份有限公司 用于视频编码的改进的定向帧内预测方法
CN109076237A (zh) * 2015-11-18 2018-12-21 联发科技股份有限公司 在视频和图像压缩中使用帧内预测滤波器的帧内预测模式的方法和装置
CN110166773A (zh) * 2019-06-25 2019-08-23 浙江大华技术股份有限公司 帧内预测方法、视频编码方法、视频处理装置、存储介质
US20200007895A1 (en) * 2018-07-02 2020-01-02 Qualcomm Incorporated Combining mode dependent intra smoothing (mdis) with intra interpolation filter switching

Patent Citations (5)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN107148778A (zh) * 2014-10-31 2017-09-08 联发科技股份有限公司 用于视频编码的改进的定向帧内预测方法
WO2016091122A1 (fr) * 2014-12-11 2016-06-16 Mediatek Inc. Procédé et appareil de prédiction intra en codage vidéo
CN109076237A (zh) * 2015-11-18 2018-12-21 联发科技股份有限公司 在视频和图像压缩中使用帧内预测滤波器的帧内预测模式的方法和装置
US20200007895A1 (en) * 2018-07-02 2020-01-02 Qualcomm Incorporated Combining mode dependent intra smoothing (mdis) with intra interpolation filter switching
CN110166773A (zh) * 2019-06-25 2019-08-23 浙江大华技术股份有限公司 帧内预测方法、视频编码方法、视频处理装置、存储介质

Non-Patent Citations (2)

* Cited by examiner, † Cited by third party
Title
A. FILIPPOV (HUAWEI), V. RUFITSKIY (HUAWEI), E. ALSHINA (HUAWEI): "AHG9: On high-level syntax for smoothing intra prediction tools", 131. MPEG MEETING; 20200629 - 20200703; ONLINE; (MOTION PICTURE EXPERT GROUP OR ISO/IEC JTC1/SC29/WG11), 29 May 2020 (2020-05-29), pages 1 - 4, XP030285436 *
A. K. RAMASUBRAMONIAN (QUALCOMM), G. VAN DER AUWERA (QUALCOMM), M. KARCZEWICZ (QUALCOMM): "Non-CE3: Simplification of reference sample smoothing in intra prediction", 15. JVET MEETING; 20190703 - 20190712; GOTHENBURG; (THE JOINT VIDEO EXPLORATION TEAM OF ISO/IEC JTC1/SC29/WG11 AND ITU-T SG.16 ), 26 June 2019 (2019-06-26), Gothenburg SE, pages 1 - 3, XP030220156 *

Cited By (1)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN114598880A (zh) * 2022-05-07 2022-06-07 深圳传音控股股份有限公司 图像处理方法、智能终端及存储介质

Also Published As

Publication number Publication date
MX2023000279A (es) 2023-02-09
CN116325727A (zh) 2023-06-23
CN116980596A (zh) 2023-10-31

Similar Documents

Publication Publication Date Title
JP7425241B2 (ja) 双方向オプティカルフローに基づく映像符号化及び復号化
US11523111B2 (en) Encoder, a decoder and corresponding methods using an adaptive loop filter
KR102123958B1 (ko) 코딩 프로세스에서의 실시간 비디오 노이즈를 감소시키는 방법, 단말기, 및 컴퓨터로 판독할 수 있는 비휘발성 저장 매체
JP5937212B2 (ja) デプス符号化
US20220159292A1 (en) Filtering method for removing blocking artifact and apparatus
EP4107945A1 (fr) Prédiction intra basée sur un réseau neuronal pour codage ou décodage vidéo
CN114287133A (zh) 帧内模式中的预测样点滤波的加权因子
CN114223200B (zh) 位置相关帧内预测样点滤波
CN115336270A (zh) 用于视频滤波的方法和装置
JP7481332B2 (ja) デコードのための予測方法及びその装置、並びにコンピュータ記憶媒体
CN114071161B (zh) 图像编码方法、图像解码方法及相关装置
WO2022077490A1 (fr) Procédé de prédiction intra, codeur, décodeur et support de stockage
KR20210070368A (ko) 비디오 이미지 요소 예측 방법 및 장치, 컴퓨터 저장 매체
CN113489974B (zh) 帧内预测方法、视频/图像编解码方法及相关装置
WO2022022622A1 (fr) Procédé de codage d'image, procédé de décodage d'image et appareil associé
US20210014511A1 (en) Image processing apparatus and method
WO2020258052A1 (fr) Procédé et dispositif de prédiction de composante d'image et support de stockage informatique
WO2020214564A1 (fr) Procédé et appareil de codage et de décodage vidéo à flux optique reposant sur une compensation de mouvement à limites lissées
JP4719108B2 (ja) 映像符号化方法、映像符号化装置、映像符号化プログラムおよびそのプログラムを記録したコンピュータ読み取り可能な記録媒体
CN115398893B (zh) 用于在视频编解码中滤波的方法和用于视频解码的设备
US10992942B2 (en) Coding method, decoding method, and coding device
WO2020258053A1 (fr) Procédé et appareil de prédiction de composante d'image et support de stockage informatique
JP2023030585A (ja) 画像符号化装置及び方法及びプログラム
WO2022197326A1 (fr) Procédé et appareil de codage vidéo
TW202406338A (zh) 用於視訊編解碼的具有色度分類器的自適應環路濾波器的方法和裝置

Legal Events

Date Code Title Description
121 Ep: the epo has been informed by wipo that ep was designated in this application

Ref document number: 20957268

Country of ref document: EP

Kind code of ref document: A1

NENP Non-entry into the national phase

Ref country code: DE

122 Ep: pct application non-entry in european phase

Ref document number: 20957268

Country of ref document: EP

Kind code of ref document: A1