WO2022266971A1 - Procédé de codage, procédé de décodage, codeur, décodeur et support de stockage informatique - Google Patents

Procédé de codage, procédé de décodage, codeur, décodeur et support de stockage informatique Download PDF

Info

Publication number
WO2022266971A1
WO2022266971A1 PCT/CN2021/102213 CN2021102213W WO2022266971A1 WO 2022266971 A1 WO2022266971 A1 WO 2022266971A1 CN 2021102213 W CN2021102213 W CN 2021102213W WO 2022266971 A1 WO2022266971 A1 WO 2022266971A1
Authority
WO
WIPO (PCT)
Prior art keywords
block
value
current block
prediction
identification information
Prior art date
Application number
PCT/CN2021/102213
Other languages
English (en)
Chinese (zh)
Inventor
唐桐
Original Assignee
Oppo广东移动通信有限公司
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Application filed by Oppo广东移动通信有限公司 filed Critical Oppo广东移动通信有限公司
Priority to PCT/CN2021/102213 priority Critical patent/WO2022266971A1/fr
Priority to CN202180098944.5A priority patent/CN117413515A/zh
Publication of WO2022266971A1 publication Critical patent/WO2022266971A1/fr

Links

Images

Classifications

    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04NPICTORIAL COMMUNICATION, e.g. TELEVISION
    • H04N19/00Methods or arrangements for coding, decoding, compressing or decompressing digital video signals
    • H04N19/10Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using adaptive coding
    • H04N19/102Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using adaptive coding characterised by the element, parameter or selection affected or controlled by the adaptive coding
    • H04N19/103Selection of coding mode or of prediction mode
    • H04N19/11Selection of coding mode or of prediction mode among a plurality of spatial predictive coding modes
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04NPICTORIAL COMMUNICATION, e.g. TELEVISION
    • H04N19/00Methods or arrangements for coding, decoding, compressing or decompressing digital video signals
    • H04N19/10Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using adaptive coding
    • H04N19/169Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using adaptive coding characterised by the coding unit, i.e. the structural portion or semantic portion of the video signal being the object or the subject of the adaptive coding
    • H04N19/17Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using adaptive coding characterised by the coding unit, i.e. the structural portion or semantic portion of the video signal being the object or the subject of the adaptive coding the unit being an image region, e.g. an object
    • H04N19/176Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using adaptive coding characterised by the coding unit, i.e. the structural portion or semantic portion of the video signal being the object or the subject of the adaptive coding the unit being an image region, e.g. an object the region being a block, e.g. a macroblock
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04NPICTORIAL COMMUNICATION, e.g. TELEVISION
    • H04N19/00Methods or arrangements for coding, decoding, compressing or decompressing digital video signals
    • H04N19/50Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using predictive coding
    • H04N19/593Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using predictive coding involving spatial prediction techniques

Definitions

  • the embodiments of the present application relate to the technical field of video coding and decoding, and in particular, to a coding and decoding method, an encoder, a decoder, and a computer storage medium.
  • H.265/High Efficiency Video Coding (HEVC) can no longer meet the needs of the rapid development of video applications.
  • JVET Joint Video Exploration Team
  • VVC Very Video Coding
  • VTM VVC Test Model
  • MIP matrix weighted Intra Prediction
  • Embodiments of the present application provide an encoding and decoding method, an encoder, a decoder, and a computer storage medium, which can reduce encoding complexity and further improve encoding and decoding efficiency.
  • the embodiment of the present application proposes an encoding method, which is applied to an encoder, and the method includes:
  • bypass identification information Based on the block type, determine the value of bypass identification information corresponding to the current block, where the bypass identification information is used to identify whether the current block adopts matrix weighted intra prediction MIP coding;
  • the current block is encoded according to the target block division manner.
  • the embodiment of the present application proposes a code stream, the code stream includes one or more of the value of the bypass identification information, the division method of the target block, the target prediction mode of the node sub-block, and the residual value coded bits.
  • the embodiment of the present application proposes a decoding method, which is applied to a decoder, and the method includes:
  • a reconstruction value for the current block is determined based on the predicted value and the residual value.
  • the embodiment of the present application proposes an encoder, the encoder includes a first determination unit and an encoding unit; wherein,
  • the first determining unit is configured to determine a block type of the current block; based on the block type, determine a value of bypass identification information corresponding to the current block, and the bypass identification information is used to identify the current block Whether to use matrix-weighted intra-frame prediction MIP coding; determine the target block division method based on the value of the bypass identification information; and encode the current block according to the target block division method.
  • the encoding unit is configured to encode the current block according to the target block division manner.
  • the embodiment of the present application provides an encoder, the encoder includes a first memory and a first processor; wherein,
  • the first memory is used to store a computer program capable of running on the first processor
  • the first processor is configured to execute the method as described in any one of the above when running the computer program.
  • the embodiment of the present application proposes a decoder, the decoder includes an analysis unit and a second determination unit; wherein,
  • the parsing unit is configured to parse the code stream, and determine the target block division method of the current block; analyze the code stream based on the target block division method, and determine the predicted value of the current block; based on the target block division method, analyze code stream, determining the residual value of the current block;
  • the second determination unit is configured to determine a reconstruction value of the current block based on the prediction value and the residual value.
  • the embodiment of the present application provides a decoder, where the decoder includes a second memory and a second processor; wherein,
  • the second memory is used to store a computer program capable of running on the second processor
  • the second processor is configured to execute the above method when running the computer program.
  • the embodiment of the present application provides a computer storage medium, wherein the computer storage medium stores a computer program, and when the computer program is executed, the method described in any one of the foregoing is implemented.
  • the embodiment of the present application provides a codec system, wherein the codec system consists of an encoder and a decoder.
  • the embodiment of the present application provides a codec method, an encoder, a decoder, and a computer storage medium.
  • the block type of the current block is determined; based on the block type, the value of the bypass identification information corresponding to the current block is determined.
  • the bypass identification information is used to identify whether the current block adopts matrix weighted intra prediction MIP encoding; based on the value of the bypass identification information, determine the target block division method; according to the target block division method, the The current block is encoded.
  • analyze the code stream to determine the target block division method of the current block On the decoder side, analyze the code stream to determine the target block division method of the current block; analyze the code stream based on the target block division method to determine the predicted value of the current block; analyze the code stream based on the target block division method to determine the residual of the current block Difference; based on the predicted value and the residual value, the reconstruction value of the current block is determined.
  • the block type of the current block can be determined based on the color parameters of the current block, and whether to use MIP for encoding prediction can be determined based on the block type of the current block, and MIP can be skipped for screen content blocks with a small number of brightness sampling values and a wide range of brightness sampling
  • the encoding process significantly reduces encoding complexity and encoding time, thereby improving encoding and decoding efficiency.
  • Fig. 1 is a schematic flow diagram of a MIP method provided by the related art
  • FIG. 2A is a schematic block diagram of an encoder provided in an embodiment of the present application.
  • FIG. 2B is a schematic block diagram of a decoder provided in an embodiment of the present application.
  • FIG. 3 is a schematic flowchart of an encoding method provided in an embodiment of the present application.
  • FIG. 4 is a schematic flowchart of an exemplary determination of MIP bypass identification information provided by an embodiment of the present application
  • Fig. 5 is a schematic flow chart of another decoding method provided by the embodiment of the present application.
  • FIG. 6 is a schematic diagram of the composition and structure of an encoder provided in an embodiment of the present application.
  • FIG. 7 is a schematic diagram of a specific hardware structure of an encoder provided in an embodiment of the present application.
  • FIG. 8 is a schematic structural diagram of a decoder provided in an embodiment of the present application.
  • FIG. 9 is a schematic diagram of a specific hardware structure of a decoder provided in an embodiment of the present application.
  • references to “some embodiments” describe a subset of all possible embodiments, but it is understood that “some embodiments” may be the same subset or a different subset of all possible embodiments, and Can be combined with each other without conflict.
  • first ⁇ second ⁇ third involved in the embodiment of the present application is only used to distinguish similar objects, and does not represent a specific ordering of objects. Understandably, “first ⁇ second ⁇ The specific order or sequence of "third” can be interchanged where allowed, so that the embodiments of the application described herein can be implemented in an order other than that illustrated or described herein.
  • JVET Joint Video Experts Team
  • VVC Versatile Video Coding
  • VVC's reference software testing platform VVC Test Model, VTM
  • the embodiment of the present application provides a kind of intra-frame prediction technology based on linear affine change, which can also be called MIP technology.
  • Figure 1 shows a specific flowchart of the MIP technology. prediction, inverse transform and inverse quantization, loop filtering and other operations, then the current block can also be called a luma block, if the application is to perform operations such as chrominance component prediction, inverse transform and inverse quantization, and loop filtering on the current block , then the current block can also be called a chroma block.
  • the luminance block will be taken as an example for description below.
  • MIP For predicting a luma block of width W and height H, MIP takes H reconstructed pixels on the left side of the luma block and W reconstructed pixels above the luma block as input, and then averages, matrix-vector multiplication and linear interpolation Get the final predicted pixel. It is roughly divided into three steps:
  • the input long reference boundary pixels bdry top and bdry left are converted into short reference boundary pixels bdry top red and bdry left red by calculating the neighborhood average to reduce the amount of calculation and model parameter storage in the prediction process Space, and then connect bdry top red and bdry left red to form the final vector bdry red with a length of 4 or 8.
  • a and b are pre-trained and stored in sets S0, S1, and S2.
  • S0 is used for 4x4 blocks, including mipMatrix4x4 (corresponding to A) and mipBias4x4 (corresponding to b) in S0
  • S1 corresponds to 4x8, 8x4, 8x8 blocks, including mipMatrix8x8 (corresponding to A) and mipBias8x8 (corresponding to b) in S1
  • mipMatrix16x16 (corresponding to A) and mipBias16x16 (corresponding to B) are included in S2.
  • the predicted pixel value is is the prediction result of using MIP to predict the luma block. It should be noted that if the difference between the horizontal direction and the vertical direction needs to be performed, the interpolation order is fixed.
  • the encoder 10 includes a transform and quantization unit 101, an intra frame estimation unit 102, an intra frame prediction unit 103, a motion compensation unit 104, a motion estimation unit 105, an inverse transform and inverse quantization unit 106, a filter control unit
  • CABAC Context-based Adaptive Binary Arithmatic Coding
  • a video coding block can be obtained by dividing the coding tree unit (Coding Tree Unit, CTU), and then the residual pixel information obtained after intra-frame or inter-frame prediction is paired by the transformation and quantization unit 101
  • the video coding block is transformed, including transforming the residual information from the pixel domain to the transform domain, and quantizing the obtained transform coefficients to further reduce the bit rate;
  • the intra frame estimation unit 102 and the intra frame prediction unit 103 are used for Intra-frame prediction is performed on the video coding block; specifically, the intra-frame estimation unit 102 and the intra-frame prediction unit 103 are used to determine the intra-frame prediction mode to be used to code the video coding block;
  • the motion compensation unit 104 and the motion estimation unit 105 is used to perform inter-frame predictive encoding of the received video coding block relative to one or more blocks in one or more reference frames to provide temporal prediction information;
  • the motion estimation performed by the motion estimation unit 105 is used to generate motion vectors process, the motion vector can estimate the motion of the video
  • the context content can be based on adjacent coding blocks, and can be used to encode the information indicating the determined intra-frame prediction mode, and output the code stream of the video signal; and the decoded image buffer unit 110 is used to store the reconstructed video coding block for forecast reference. As the video image encoding progresses, new reconstructed video encoding blocks will be continuously generated, and these reconstructed video encoding blocks will be stored in the decoded image buffer unit 110 .
  • the decoder 20 includes a decoding unit 201, an inverse transform and inverse quantization unit 202, an intra prediction unit 203, a motion compensation unit 204, a filtering unit 205, and a decoded image buffer unit 206, etc., wherein the decoding unit 201 Header information decoding and CABAC decoding can be realized, and the filtering unit 205 can realize DBF filtering/SAO filtering/ALF filtering.
  • the code stream of the video signal is output; the code stream is input into the video decoding system 20, and first passes through the decoding unit 201 to obtain the decoded transform coefficient; for the transform coefficient, pass
  • the inverse transform and inverse quantization unit 202 performs processing to generate a residual block in the pixel domain; the intra prediction unit 203 is operable to generate residual blocks based on the determined intra prediction mode and data from previously decoded blocks of the current frame or picture.
  • the motion compensation unit 204 determines the prediction information for the video decoding block by parsing motion vectors and other associated syntax elements, and uses the prediction information to generate the predictive properties of the video decoding block being decoded block; a decoded video block is formed by summing the residual block from the inverse transform and inverse quantization unit 202 with the corresponding predictive block produced by the intra prediction unit 203 or the motion compensation unit 204; the decoded video signal Video quality can be improved by filtering unit 205 in order to remove block artifacts; the decoded video blocks are then stored in the decoded picture buffer unit 206, which stores reference pictures for subsequent intra prediction or motion compensation , and is also used for the output of the video signal, that is, the restored original video signal is obtained.
  • the encoding and decoding method in the embodiment of the present application can be applied to a video encoding and decoding chip, and the encoding performance can be significantly improved by using the MIP mode.
  • it can be applied to the intra/inter prediction part shown in Figure 2A (indicated by a black bold box, specifically including the intra estimation unit 102 and the intra prediction unit 103), and can also be applied to the intra prediction unit 103 shown in Figure 2B.
  • the shown intra/inter prediction part (indicated by a black bold box, specifically includes the intra prediction unit 203).
  • the codec method in the embodiment of the present application can be applied to a video encoding system (referred to as “encoder” for short), or a video decoding system (referred to as “decoder” for short), and can even be used at the same time It is applied to a video encoding system and a video decoding system, but it is not limited here.
  • the "current block” specifically refers to the block currently to be encoded in the video image (also referred to as “coding block” for short); when the embodiment of the present application When applied to the decoder 20, the “current block” specifically refers to a block currently to be decoded in a video image (may also be referred to simply as a "decoded block”).
  • an implementation method of an encoder using MIP is as follows: first, an input image is divided into multiple non-overlapping CTU blocks. Then, each CTU is processed sequentially according to the raster scanning order, and the CTU is divided into several CUs according to various block division methods.
  • the various block division methods for the CTU can include quadtree division, vertical binary tree division, and horizontal binary tree division.
  • determining the optimal block division method from a variety of block division methods includes the following steps, for the i-th block division method Spilt[i]: 1 use linear interpolation-based prediction
  • the method of intra-frame prediction select the optimal intra-frame prediction mode bestRegIntraMode[i] and prediction cost bestRegIntraCost[i], 2
  • the MIP method uses the MIP method to perform frame
  • For intra prediction select the optimal MIP prediction mode bestMipIntraMode[i] and prediction cost bestMipIntraCost[i]; 3 compare bestRegIntraCost[i] and bestMipIntraCost[i], and select the optimal intra prediction mode corresponding to Spilt[i] bestIntraMode[ i] and prediction cost bestIntraCost[i]; 4 Use other methods, such as inter-frame prediction for prediction, and select the optimal prediction mode best
  • the residual block is predicted according to the optimal block division method, and the residual block is transformed, quantized, and entropy encoded, and the prediction information such as the block division mode is encoded, and the output code stream is waiting for transmission.
  • an implementation method of a decoder using MIP is as follows: first, the input code stream is subjected to entropy decoding, inverse quantization, and inverse transformation to obtain a residual block; then, the image is reconstructed according to the residual block , the reconstruction process mainly includes the following three steps: 1 Determine the partition tree of the current CTU according to the block partition mode; 2 Process each CU of the partition tree in turn according to the raster scan order, and use the prediction mode bestMode[opt] of each CU to calculate the prediction value Pred; 3 superimpose the residual value and predicted value of the current CU to obtain the reconstructed CU. Finally, the reconstructed image is sent to the Deblocking Filter (DBF)/Sample Adaptive Offset (SAO) filter/Adaptive Loop Filter (ALF), and the filtered image Send it to the cache area and wait for the video to play.
  • DPF Deblocking Filter
  • SAO Sample Adaptive Offset
  • ALF Adaptive Loop Filter
  • the biggest advantage of MIP technology is that it can predict textured image blocks more accurately.
  • the prediction model of MIP is trained based on massive data, and the prediction model is more accurate than conventional intra-frame prediction technology.
  • the linear interpolation calculation model is more complex and more accurate; however, for flat and simple image blocks, the conventional intra prediction technology can be used to predict more accurately, and then the MIP technology can obtain a small gain, but it will be certain degree increases the coding complexity.
  • many screen content videos are rich in flat areas, such as large areas of solid-color background areas, etc. For such areas, direct use of conventional intra-frame prediction techniques can already complete accurate prediction tasks.
  • screen content videos generally contain a large number of repetitive patterns and a limited number of colors.
  • the embodiment of the present application provides an encoding method.
  • the block type of the current block is determined based on the color parameters of the current block; based on the block type, the bypass identifier corresponding to the current block is determined, and the bypass identifier is used to identify whether to use
  • the matrix-weighted intra-frame prediction MIP encodes the current block; based on the MIP bypass identification information, the target block division method is determined; and the current block is encoded according to the target block division method.
  • the block type of the current block can be determined based on the color parameters of the current block, and whether to use MIP for encoding prediction can be determined based on the block type of the current block, and MIP can be skipped for screen content blocks with a small number of brightness sampling values and a wide range of brightness sampling
  • the encoding process significantly reduces encoding complexity and encoding time, thereby improving encoding and decoding efficiency.
  • FIG. 3 shows a schematic flowchart of an encoding method provided in an embodiment of the present application.
  • the method may include:
  • the encoding method in the embodiment of the present application is applied to an encoder.
  • the video image can be divided into multiple image blocks, and each image block to be encoded can be called a coding block, and the current block here specifically refers to the coding block to be currently coded, which It may be a CTU or even a CU, which is not limited in this embodiment of the present application.
  • the video image may be divided according to a preset size to obtain N coding blocks, where N is an integer greater than zero, and the N blocks do not overlap each other.
  • the preset size refers to a preset block size value.
  • the preset size can be any one of 8, 16, 32, 64, etc., or any one of 8 ⁇ 8, 16 ⁇ 16, 32 ⁇ 32, 64 ⁇ 64, etc.
  • the application examples are not specifically limited.
  • the preset size may be 8 ⁇ 8. In this case, the video image is divided into N 8 ⁇ 8 coding blocks.
  • the embodiment of the present application mainly provides a MIP adaptive bypass technology based on content analysis to analyze image content.
  • MIP predictive coding is skipped; otherwise, MIP predictive coding is used.
  • the encoding complexity can be reduced while ensuring the encoding performance.
  • the judgment of whether to use MIP predictive coding can be made based on a frame of video image, or the judgment of whether to use MIP predictive coding can be made based on CUs in one or more regions of a frame of video image. Specifically, it can be based on The selection is made according to the actual situation, and the embodiment of the present application does not make a specific limitation.
  • the color parameter of the current block in the video image is acquired; the block type of the current block is determined based on the color parameter of the current block, wherein the color parameter includes the number of brightness sampling values and the range of brightness sampling, specifically, the determination of the color parameter in the video image
  • the process of the color parameter of the current block includes: for the current block of the video image, obtaining the maximum brightness component sampling value, the minimum brightness component sampling value and the number of brightness sampling values of the current block; according to the maximum brightness component sampling value of the current block and the minimum luminance component sampling value to determine the luminance sampling range of the current block.
  • the number of luminance sample values of the current block is determined by the number of luminance component sample values of the current block, and the number of luminance sample values of the current block is determined by counting the numbers of all different luminance component sample values in the current block.
  • the current block can be a luma block or a chroma block, that is, in this application, the luma component sampling value of the luma block can be obtained, and then the block type of the luma block can be determined based on the luma component sampling value of the luma block, or Obtain the luminance component sampling value of the chroma block, and then determine the block type of the chroma block based on the luminance component sampling value of the chroma block, which can be selected according to the actual situation, and is not specifically limited in the embodiment of the present application.
  • the maximum luminance component sampling value of the current block is 2 to the 8th power, that is, 255, and the minimum luminance component sampling value is 0. Then, the range of the number of brightness sample values of the current block is between 0 and 256.
  • the difference between the maximum luminance component sampling value and the minimum luminance component sampling value of the current block is calculated, and the difference is based on the luminance sampling range of the current block. Specifically, as shown in formula (2),
  • R is the brightness sampling range
  • Y max and Y min are the maximum brightness component sampling value and the minimum brightness component sampling value of the current block respectively.
  • the content parameters of the current block are determined according to the number of luminance sampling values of the current block and the luminance sampling range of the current block , then compare the content parameter of the current block with the first preset threshold, if the content parameter of the current block is smaller than the first preset threshold, then determine that the block type of the current block is the first type block; if the content parameter of the current block is greater than or equal to the first preset threshold, then it is determined that the block type of the current block is the second type block.
  • the first type of block is a natural content block whose number of luminance sampling values is not less than the preset number threshold and whose luminance sampling range is not greater than the preset range threshold;
  • the second type of block is that the number of luminance sampling values is less than the preset number threshold and the luminance sampling range Blocks of screen content that are larger than a preset range threshold. That is to say, the number of luminance sampling values in the first type block is larger and the luminance sampling range is smaller, and the number of luminance sampling values in the second type block is smaller and the luminance sampling range is wider.
  • N is the number of luminance sample values
  • D is a content parameter.
  • N is the number of luminance sample values
  • D is a content parameter.
  • the content parameter is used to represent the size of the brightness sampling range and the number of brightness sampling values of the current block. That is to say, when the number of luma sample values in the current block is small and the range of luma sample values is large, the content parameter is larger; when the number of luma sample values in the current block is large and the range of luma sample values is small, the content parameter The smaller the parameter.
  • the first preset threshold can be set according to the actual situation, and the embodiment of the present application does not make specific limitations. In practical applications, the value of the first preset threshold can be 3, 4, 5, etc. .
  • block type determination is performed on at least one coding block in the video image in sequence, wherein at least one coding block may be a coding block corresponding to one or more slices, or may be a coding block corresponding to one or more tiles, or It may be a coding block corresponding to CUs of one or more pre-set regions, or a coding block corresponding to a frame of video image, which may be selected according to actual conditions, and is not specifically limited in this embodiment of the present application.
  • the video image is divided into N slices, wherein each slice includes multiple CTUs, and the type of the slice can be determined according to the block types of all coding blocks in the slice; or the video image is horizontally or vertically Divided into N rectangular areas, the rectangular area is called a tile, each tile includes an integer number of CTUs, and the type of the tile is determined according to the block type of all coding blocks in the tile; or according to all codes in the CU of one or more regions
  • the block type of the block determines the type of CU in one or more regions; or determines the type of a frame of video image according to the block types of all coding blocks in a frame of video image.
  • the specific ones can be selected according to the actual situation, and are not specifically limited in this embodiment of the present application.
  • multi-scale block type fusion may be performed on at least one coding block based on the block type corresponding to at least one coding block; and based on the fused block Type, to calculate the proportion of screen content blocks.
  • one frame of video images can be fused, and the proportion of screen content blocks can be calculated.
  • four 8 ⁇ 8 blocks are sequentially combined into 16 ⁇ 16 blocks, and determine the 4 block types in each 16 ⁇ 16 block, if more than 3 block types among the 4 block types belong to the same block type, the entire 16 ⁇ 16 block type will be fused into one block type, and the block type is the same block type mentioned above; if two block types among the four block types are the same block type, block type fusion is not performed on the 16 ⁇ 16 block type.
  • bypass identification information Based on the block type, determine a value of bypass identification information corresponding to the current block, where the bypass identification information is used to identify whether the current block adopts matrix weighted intra prediction MIP encoding.
  • bypass identification information is newly added to indicate whether to use MIP coding for the current block.
  • it is determined whether to MIP is used for intra-frame prediction, and then the specific value of the bypass flag is determined. Then, MIP predictive coding can be performed or skipped based on the value of the bypass flag information, which can be used for screens with a small number of luminance sampling values and a wide range of luminance sampling. Content blocks skip the MIP encoding process, significantly reducing encoding complexity.
  • the block type corresponding to at least one coding block in the video image can be sequentially determined, and if the number of the second type blocks in the block type corresponding to at least one coding block is greater than the second preset threshold, then determine the bypass flag
  • the value of the information is a first value; if the number of blocks of the second type in the block type corresponding to at least one coding block is not greater than a second preset threshold, then it is determined that the value of the bypass identification information is a second value.
  • at least one coding block may include the current block.
  • the first value and the second value are different, and the first value and the second value may be in the form of parameters or numbers.
  • the bypass identification information of the video image may be a flag, which is not limited here.
  • the bypass identification information is a flag
  • the first value can be set to 1, and the second value can be set to 0; in another specific example, the first The value can also be set to true, and the second value can also be set to false; even in yet another specific example, the first value can also be set to 0, and the second value can also be set to 1; or, the first value can also be set to Set to false, the second value can also be set to true.
  • the first value and the second value in the embodiment of the present application are not limited in any way.
  • the value of the bypass identification information can be determined based on the block types of the first 10 image blocks in a frame of video image, specifically, if there are at least 8 image block block types in the first 10 image blocks is a screen content block, then the value of the bypass identification information is the second value, representing an image block that adopts a MIP encoded video image; if there are no block types of 8 image blocks in the first 10 image blocks that are screen content blocks, then The value of the bypass identification information is the first value, which indicates that an image block of the MIP-encoded video image is skipped.
  • the screen content block ratio is compared with the second preset threshold, and if the screen content block ratio is If the proportion of the content block is not less than the third preset threshold, the value of the bypass identification information is determined to be the first value; if the proportion of the screen content block is less than the third preset threshold, then the value of the bypass identification information is determined to be the first value. binary value.
  • the specific value of the second preset threshold can be selected according to the actual situation, and the embodiment of the present application does not make specific limitations. In practical applications, the second preset threshold can be set to 0.1, 0.2 or 0.3, etc. .
  • the MIP encoding of the current block is skipped, and the MIP encoding process can be skipped for screen content blocks with a small number of luminance sampling values and a wide luminance sampling range, significantly The encoding complexity is greatly reduced; if the value of the bypass identification information is the second value, the current block is encoded using MIP.
  • the level of bypass identification information can be one of the following: image level, that is, calculate a bypass identification information for a frame of video image; slice level (slice), that is, for a frame of video image A slice calculates a bypass identification information as a MIP predictive coding constraint of a slice; slice level (tile), that is, calculates a bypass identification information for a tile in a frame of video image as a tile MIP predictive coding constraint.
  • image level that is, calculate a bypass identification information for a frame of video image
  • slice level (slice) that is, for a frame of video image
  • a slice calculates a bypass identification information as a MIP predictive coding constraint of a slice
  • slice level (tile) that is, calculates a bypass identification information for a tile in a frame of video image as a tile MIP predictive coding constraint.
  • bypass identification information can be determined for a frame of video image, and the specific implementation process is shown in Figure 4,
  • 2 includes:
  • step 2 If it is judged that the video sequence has not been encoded, go to step 2.
  • the image block is determined to be a natural content block; correspondingly, for 2.5, if D is less than T2, the image block is determined to be For the screen content block.
  • the target block division method is determined based on the value of the bypass identification information. Specifically, the video image is divided into multiple non-overlapping CTU blocks, Afterwards, each CTU is processed sequentially according to the raster scanning order, and the CTU is divided into several CUs according to various block division methods. At this time, it is necessary to determine the prediction mode and prediction cost corresponding to each block division method.
  • the target block division mode is determined in the division mode, and the value of the bypass identification information is used to determine the prediction mode and prediction cost corresponding to the various block division modes.
  • the encoder after determining the target block division method based on the value of the bypass identification information, in order for the decoder to obtain the target block division method and the prediction mode corresponding to the target block division method, the encoder needs to The division method and the prediction mode corresponding to the target block division method are encoded, and then written into the code stream to be transmitted from the encoder to the decoder. Furthermore, it is also possible to perform prediction based on the division method of the target block to obtain the residual value of the current block; perform transformation, quantization, and entropy coding on the residual value, and then write the coded bits into the code stream.
  • the processing sequence is to determine the target prediction mode of each node sub-block in turn; and determine the predicted value of the node sub-block according to the target prediction mode; determine the residual value of the node sub-block according to the original value and the predicted value of the node sub-block; finally , encode the target prediction mode and residual value of the node sub-block, and write the coded bits into the code stream.
  • the embodiment of the present application can also provide a code stream, which includes one or more of the value of the bypass identification information, the division method of the target block, the target prediction mode of the node sub-block, and the residual value coding bits.
  • the value of the bypass identification information may also be encoded, and the encoded bits are written into the code stream.
  • an implementation method of an encoder based on content analysis-based MIP adaptive bypass technology is as follows: first, an input image is divided into multiple non-overlapping CTU blocks. Then, process each CTU sequentially according to the raster scanning order, divide the CTU into several CUs according to various block division methods, and determine the optimal block division method from the various block division methods includes the following steps, for the i-th block division method Spilt[i]: 1 Perform intra-frame prediction based on linear interpolation prediction method, select the optimal intra-frame prediction mode bestRegIntraMode[i] and prediction cost bestRegIntraCost[i], 2 After judging that the block size of the current CU satisfies the corresponding The block size is limited and the value of the bypass identification information is the second value, that is, the current CU is coded using MIP, then intra-frame prediction is performed using the MIP method, and the optimal MIP prediction mode bestMipIntraMode[i] and prediction cost bestMi
  • the residual block is predicted according to the optimal block division method, and the residual block is transformed, quantized, and entropy encoded, and the prediction information such as the block division mode is encoded, and the output code stream is waiting for transmission.
  • the test sequence Contains three types of screen sequences: TGM: Text and graphics with motion; M: mixed content; A: animation.
  • TGM Text and graphics with motion
  • M mixed content
  • A animation.
  • TGM sequences, 1 M sequence the average changes of BD-rate on Y, Cb, and Cr components are -0.182%, -0.568%, -0.034%, respectively, while the encoding time The average change was -4.3%.
  • the data shows that this technology can save a certain amount of encoding time without reducing the encoding performance.
  • the technology proposed in the embodiments of the present application can reduce coding complexity without reducing coding performance.
  • this technology directly bypasses MIP predictive coding for typical screen content sequences (that is, a small number of luminance sampling values and a wide range of luminance sampling), because these typical screen content sequences directly use IBC, PLT and traditional intra prediction methods Better coding efficiency can be obtained, so although the MIP prediction is bypassed, there is still no loss of coding performance, and the coding time is reduced by 4.3%.
  • the technology can significantly reduce the coding complexity while maintaining the coding performance substantially equivalent to that of the original technology.
  • An embodiment of the present application provides an encoding method, which is applied to an encoder. Determine the color parameters of the current block in the video image; obtain the block type of the current block based on the color parameters of the current block; determine the value of the bypass identification information corresponding to the current block based on the block type, and the value of the bypass identification information is used for identification Whether to use matrix-weighted intra-frame prediction MIP to encode the current block; determine the division method of the target block based on the value of the bypass identification information; encode the current block according to the division method of the target block, so that it can be based on the color parameter of the current block Determine the block type of the current block, and determine whether to use MIP for encoding prediction based on the block type of the current block. It can skip the MIP encoding process for screen content blocks with a small number of luminance sampling values and a wide luminance sampling range, which significantly reduces the encoding complexity. and reduce the encoding time, thereby improving the encoding and decoding efficiency.
  • FIG. 5 shows a schematic flowchart of a decoding method provided in an embodiment of the present application. As shown in Figure 5, the method may include:
  • the decoding method in the embodiment of the present application is applied to a decoder.
  • the video image can be divided into multiple image blocks, wherein each image block to be decoded can be called a decoding block, and the current block here specifically refers to the decoding block currently to be decoded ; After the decoding is complete, you can wait for the video to play.
  • the target block division method of the current block is determined, and then the target block division mode is written into the code stream. Therefore, the target block can be obtained by parsing the code stream at the decoding end
  • the division method is used for determining the prediction value of the current block and the residual value of the current block during the decoding process.
  • judging whether the video image is decoded using MIP may be represented by a bypass flag.
  • the method may also include:
  • bypass identification information If the value of the bypass identification information is the first value, skip MIP decoding of the current block; or, if the value of the bypass identification information is the second value, use MIP to decode the current block.
  • the first value and the second value are different, and the first value and the second value may be in the form of parameters or numbers.
  • the value of the bypass identification information may be a flag (flag), which is not limited here.
  • the first value can be set to 0, and the second value can be set to 1 ;
  • the first value can be set to false, and the second value can be set to true; even in yet another specific example, the first value can also be set to 1, and the second value can also be set to 0; or, the first value can also be set to true, and the second value can also be set to false.
  • the first value and the second value in the embodiment of the present application are not limited in any way.
  • the decoder parses the code stream to obtain the value of the bypass identification information, if the value of the MIP bypass identification information is 0, then it is judged that the encoder The end skips the MIP predictive coding process, which can save coding speed and significantly reduce coding complexity. At this time, the decoder side skips the process of MIP decoding the current block. Otherwise, if the value of the MIP bypass identification information is 1, it is determined that the encoder uses MIP for predictive coding, and at this time, the decoder also uses MIP to decode the current block.
  • the division tree of the current block is determined, wherein the division tree includes one or more node sub-blocks obtained by dividing the current block. Then, according to the preset node sub-block processing sequence, the code stream of each node sub-block of the partition tree is sequentially analyzed, and the target prediction mode of each node sub-block is determined; the predicted value of each node sub-block is determined according to the prediction mode.
  • the parsing the code stream based on the target block division mode, and determining the residual value of the current block may include:
  • the code stream of each node sub-block of the partition tree is sequentially analyzed according to the preset processing order of the node sub-blocks, and the residual value of each node sub-block is determined.
  • the preset node sub-block processing sequence may be a preset scanning sequence. That is to say, the embodiment of the present application can sequentially analyze the code stream of each node sub-block of the partition tree according to the preset scanning order, and then determine the residual value of each node sub-block.
  • the determining the reconstruction value of the current block based on the prediction value and the residual value may include: performing an addition calculation on the prediction value and the residual value to determine the reconstruction value of the current block.
  • the predicted value of the current block can also be obtained by parsing the code stream; and the residual value of the current block can also be obtained by parsing the code stream; in this way, by analyzing the code stream and The residual value is added to determine the reconstruction value of the current block.
  • the implementation method of a decoder based on content analysis-based MIP adaptive bypass technology is as follows: first, the input code stream is subjected to entropy decoding, inverse quantization, and inverse transformation, which can be The residual value is obtained; then, the image is reconstructed according to the residual block, and the reconstruction process here mainly includes the following three steps: 1 Determine the partition tree of the current CTU according to the partition method of the block target block. 2 Process each CU of the partition tree sequentially according to the raster scan order, and use the prediction mode of each CU to calculate the predicted value. 3 The residual value and predicted value of the current CU are superimposed to obtain the reconstructed CU.
  • the reconstructed image is sent to the Deblocking Filter (DBF)/Sample Adaptive Offset (SAO)/Adaptive Loop Filter (ALF) filter, and the filtered image is sent to the cache area and wait for the video to play.
  • DPF Deblocking Filter
  • SAO Sample Adaptive Offset
  • ALF Adaptive Loop Filter
  • This embodiment provides a decoding method, which is applied to a decoder. Analyze the code stream to determine the target block division method of the current block; analyze the code stream based on the target block division method to determine the predicted value of the current block; analyze the code stream based on the target block division method to determine the residual value of the current block; based on the prediction value and residual value, determine the reconstruction value for the current block.
  • the block type of the current block can be determined based on the color parameters of the current block, and whether to use MIP for encoding prediction can be determined based on the block type of the current block, and MIP can be skipped for screen content blocks with a small number of brightness sampling values and a wide range of brightness sampling
  • the encoding process significantly reduces encoding complexity and encoding time, thereby improving encoding and decoding efficiency.
  • FIG. 6 shows a schematic structural diagram of an encoder 1 provided in the embodiment of the present application.
  • the encoder 11 may include: a first determining unit 10, and an encoding unit 11; wherein,
  • the first determining unit 10 is configured to determine the block type of the current block; based on the block type, determine the value of the bypass identification information corresponding to the current block, and the bypass identification information is used to identify whether to use Matrix weighted intra prediction MIP encoding the current block; based on the value of the bypass identification information, determine the target block division method.
  • the coding unit 11 is configured to code the current block according to the target block division method.
  • the first determining unit 10 is configured to determine a color parameter of a current block in a video image; and determine a block type of the current block based on the color parameter of the current block.
  • the color parameters include the number of brightness samples and the range of brightness samples; the encoder 1 further includes:
  • the first determining unit 10 is further configured to determine the maximum luminance component sampling value of the current block, the minimum luminance component sampling value, and the number of luminance sampling values of the current block; according to the maximum luminance component sampling value of the current block and the minimum luminance component sampling value to determine the luminance sampling range of the current block.
  • the first determining unit 10 is further configured to determine the content parameters of the current block according to the number of luminance sampling values of the current block and the luminance sampling range of the current block; if the current If the content parameter of the block is less than the first preset threshold, it is determined that the block type of the current block is the first type block; if the content parameter of the current block is greater than or equal to the first preset threshold, then it is determined that the current block The block type of the block is a second type block.
  • the first determination unit 10 is further configured to determine the block type corresponding to at least one coding block contained in the video image; if the second type of the block type corresponding to the at least one coding block If the number of blocks is greater than the second preset threshold, it is determined that the value of the bypass identification information corresponding to the current block is the first value; if the number of the second type of blocks in the block type corresponding to the at least one coding block is less than or equal to The second preset threshold determines that the value of the bypass identification information corresponding to the current block is a second value.
  • the encoder further includes: a fusion unit and a calculation unit;
  • the first determining unit 10 is further configured to determine a block type corresponding to at least one coding block in the video image
  • the fusion unit is configured to fuse the at least one coding block based on the block type corresponding to the at least one coding block to obtain a fused coding block;
  • the calculation unit is configured to calculate the screen content block proportion based on the block type of the fused coding block
  • the first determination unit 10 is further configured to determine that the value of the bypass identification information is the first value if the proportion of the screen content block is greater than or equal to a second preset threshold; if the screen content block If the proportion is smaller than the second preset threshold, it is determined that the value of the bypass identification information is a second value.
  • MIP encoding of the current block is skipped. If the value of the bypass identification information is the second value, MIP is used to encode the current block.
  • the level of the bypass identification information includes at least one of the following: picture level, slice level, slice level, coding tree unit CTU level, coding unit CU level.
  • the first determining unit 10 is further configured to determine prediction modes and prediction costs corresponding to multiple division methods based on the value of the bypass identification information; from the multiple block division methods Determine the target block division method with the smallest prediction cost and the target prediction mode of each node sub-block; wherein, each node sub-block is obtained by dividing the current block based on the target block division method .
  • the first determination unit 10 is further configured to use a linear interpolation prediction method to perform intra-frame prediction to obtain the first prediction mode and the first prediction cost; if determined based on the value of the bypass identification information If MIP is used for intra-frame prediction, MIP is used for intra-frame prediction to obtain the second prediction mode and the second prediction cost; based on the comparison result of the first prediction cost and the second prediction cost, the first Determining the intra prediction cost and the intra prediction mode corresponding to the intra prediction cost from the prediction cost and the second prediction cost; if it is determined based on the value of the bypass identification information that MIP is skipped for intra prediction, Then determine the first prediction cost as the intra prediction cost, and determine the first prediction mode as the intra prediction mode; use an inter prediction method to perform inter prediction, and obtain the inter prediction mode and frame Inter-prediction cost; based on the comparison result of the intra-frame prediction cost and the inter-frame prediction cost, determine the prediction mode and prediction corresponding to multiple block division methods from the intra-frame prediction cost and the inter-frame prediction
  • the first determination unit 10 is further configured to divide the current block according to the target block division method to obtain a division tree of the current block, wherein the division tree includes the One or more node sub-blocks obtained by dividing the current block; determine the target prediction mode of each of the node sub-blocks; determine the predicted value of the node sub-block according to the target prediction mode; according to the node sub-block the original value of the block and the predicted value, and determine the residual value of the node sub-block;
  • the encoding unit 11 is further configured to encode the target prediction mode and the residual value of the node sub-block, and write the encoded bits into the code stream.
  • the encoding unit 11 is further configured to encode the target block division manner, and write encoded bits into a code stream.
  • a "unit" may be a part of a circuit, a part of a processor, a part of a program or software, etc., of course it may also be a module, or it may be non-modular.
  • each component in this embodiment may be integrated into one processing unit, each unit may exist separately physically, or two or more units may be integrated into one unit.
  • the above-mentioned integrated units can be implemented in the form of hardware or in the form of software function modules.
  • the integrated unit is implemented in the form of a software function module and is not sold or used as an independent product, it can be stored in a computer-readable storage medium.
  • the technical solution of this embodiment is essentially or It is said that the part that contributes to the prior art or the whole or part of the technical solution can be embodied in the form of a software product, the computer software product is stored in a storage medium, and includes several instructions to make a computer device (which can It is a personal computer, a server, or a network device, etc.) or a processor (processor) that executes all or part of the steps of the method described in this embodiment.
  • the aforementioned storage medium includes: U disk, mobile hard disk, read only memory (Read Only Memory, ROM), random access memory (Random Access Memory, RAM), magnetic disk or optical disk and other various media that can store program codes.
  • the embodiment of the present application provides a computer storage medium, which is applied to the encoder 1, and the computer storage medium stores a computer program, and when the computer program is executed by the first processor, it realizes any one of the above-mentioned embodiments. Methods.
  • FIG. 7 shows a schematic diagram of a specific hardware structure of the encoder provided by the embodiment of the present application.
  • it may include: a first communication interface 12 , a first memory 13 and a first processor 14 ; each component is coupled together through a first bus system 15 .
  • the first bus system 15 is used to realize connection and communication between these components.
  • the first bus system 15 also includes a power bus, a control bus and a status signal bus.
  • the various buses are labeled as the first bus system 15 in FIG. 7 . in,
  • the first communication interface 12 is used for receiving and sending signals during the process of sending and receiving information with other external network elements;
  • the first memory 13 is used to store computer programs that can run on the first processor 14;
  • the first processor 14 is configured to, when running the computer program, execute:
  • bypass identification information Based on the block type, determine the value of bypass identification information corresponding to the current block, where the bypass identification information is used to identify whether to use matrix weighted intra prediction MIP to encode the current block;
  • the current block is encoded according to the target block division manner.
  • the first memory 13 in the embodiment of the present application may be a volatile memory or a nonvolatile memory, or may include both volatile and nonvolatile memories.
  • the non-volatile memory can be read-only memory (Read-Only Memory, ROM), programmable read-only memory (Programmable ROM, PROM), erasable programmable read-only memory (Erasable PROM, EPROM), electronically programmable Erase Programmable Read-Only Memory (Electrically EPROM, EEPROM) or Flash.
  • the volatile memory can be Random Access Memory (RAM), which acts as external cache memory.
  • RAM Static Random Access Memory
  • DRAM Dynamic Random Access Memory
  • SRAM Dynamic Random Access Memory
  • Synchronous Dynamic Random Access Memory Synchronous Dynamic Random Access Memory
  • SDRAM double data rate synchronous dynamic random access memory
  • Double Data Rate SDRAM DDRSDRAM
  • enhanced SDRAM ESDRAM
  • Synchlink DRAM SLDRAM
  • Direct Memory Bus Random Access Memory Direct Rambus RAM, DRRAM
  • the first memory 13 of the systems and methods described herein is intended to include, but not be limited to, these and any other suitable type of memory.
  • the first processor 14 may be an integrated circuit chip with signal processing capability. In the implementation process, each step of the above method may be completed by an integrated logic circuit of hardware in the first processor 1003 or an instruction in the form of software.
  • the above-mentioned first processor 14 can be a general-purpose processor, a digital signal processor (Digital Signal Processor, DSP), an application specific integrated circuit (Application Specific Integrated Circuit, ASIC), a ready-made programmable gate array (Field Programmable Gate Array, FPGA) Or other programmable logic devices, discrete gate or transistor logic devices, discrete hardware components.
  • DSP Digital Signal Processor
  • ASIC Application Specific Integrated Circuit
  • FPGA Field Programmable Gate Array
  • Various methods, steps, and logic block diagrams disclosed in the embodiments of the present application may be implemented or executed.
  • a general-purpose processor may be a microprocessor, or the processor may be any conventional processor, or the like.
  • the steps of the method disclosed in connection with the embodiments of the present application may be directly implemented by a hardware decoding processor, or implemented by a combination of hardware and software modules in the decoding processor.
  • the software module can be located in a mature storage medium in the field such as random access memory, flash memory, read-only memory, programmable read-only memory or electrically erasable programmable memory, register.
  • the storage medium is located in the first memory 13, and the first processor 14 reads the information in the first memory 13, and completes the steps of the above method in combination with its hardware.
  • the embodiments described in this application may be implemented by hardware, software, firmware, middleware, microcode or a combination thereof.
  • the processing unit can be implemented in one or more application specific integrated circuits (Application Specific Integrated Circuits, ASIC), digital signal processor (Digital Signal Processing, DSP), digital signal processing device (DSP Device, DSPD), programmable Logic device (Programmable Logic Device, PLD), Field-Programmable Gate Array (Field-Programmable Gate Array, FPGA), general-purpose processor, controller, microcontroller, microprocessor, other devices for performing the functions described in this application electronic unit or its combination.
  • the techniques described herein can be implemented through modules (eg, procedures, functions, and so on) that perform the functions described herein.
  • Software codes can be stored in memory and executed by a processor. Memory can be implemented within the processor or external to the processor.
  • the first processor 14 is further configured to execute the method described in any one of the foregoing embodiments when running the computer program.
  • This embodiment provides an encoder, which may include a first determining unit and an encoding unit.
  • the block type of the current block can be determined based on the color parameters of the current block, and whether to use MIP for encoding prediction can be determined based on the block type of the current block, and MIP can be skipped for screen content blocks with a small number of brightness sampling values and a wide range of brightness sampling
  • the encoding process significantly reduces encoding complexity and encoding time, thereby improving encoding and decoding efficiency.
  • FIG. 8 shows a schematic diagram of the composition and structure of a decoder 2 provided by the embodiment of the present application.
  • the decoder 2 may include: an analysis unit 20 and a second determination unit 21; wherein,
  • the parsing unit 20 is configured to parse the code stream to determine the target block division method of the current block; analyze the code stream based on the target block division method, and determine the predicted value of the current block; based on the target block division method, Analyzing the code stream to determine the residual value of the current block;
  • the second determination unit 21 is configured to determine a reconstruction value of the current block based on the predicted value and the residual value.
  • the parsing unit 20 is further configured to parse the code stream to obtain the value of the bypass identification information; if the value of the bypass identification information is the first value, skip the MIP decoding the current block; or, if the value of the bypass identification information is a second value, decoding the current block by using MIP.
  • the second determining unit 21 is further configured to determine the partition tree of the current block according to the partition method of the target block, wherein the partition tree includes partitioning the current block to obtain One or more node subblocks of .
  • the parsing unit 20 is further configured to parse the code stream of each node sub-block of the partition tree, and determine the target prediction mode of each node sub-block;
  • the second determination unit 21 is further configured to determine the prediction value of each node sub-block according to the target prediction mode.
  • the second determining unit 21 is further configured to sequentially analyze the code stream of each node sub-block of the partition tree according to the preset node sub-block processing order, and determine the code stream of each node sub-block The residual value of the block.
  • a "unit” may be a part of a circuit, a part of a processor, a part of a program or software, etc., of course it may also be a module, or it may be non-modular.
  • each component in this embodiment may be integrated into one processing unit, each unit may exist separately physically, or two or more units may be integrated into one unit.
  • the above-mentioned integrated units can be implemented in the form of hardware or in the form of software function modules.
  • the integrated units are implemented in the form of software function modules and are not sold or used as independent products, they can be stored in a computer-readable storage medium.
  • this embodiment provides a computer storage medium, which is applied to the decoder 2.
  • the computer storage medium stores a computer program, and when the computer program is executed by the second processor, any one of the preceding embodiments is implemented. the method described.
  • FIG. 9 shows a schematic diagram of a specific hardware structure of the decoder 2 provided by the embodiment of the present application.
  • it may include: a second communication interface 22 , a second memory 23 and a second processor 24 ; each component is coupled together through a second bus system 25 .
  • the second bus system 25 is used to realize connection and communication between these components.
  • the second bus system 25 also includes a power bus, a control bus and a status signal bus.
  • the various buses are labeled as the second bus system 25 in FIG. 9 . in,
  • the second communication interface 22 is used for receiving and sending signals during the process of sending and receiving information with other external network elements;
  • the second memory 23 is used to store computer programs that can run on the second processor 24;
  • the second processor 24 is configured to, when running the computer program, execute:
  • a reconstruction value for the current block is determined based on the predicted value and the residual value.
  • the second processor 24 is further configured to execute the method described in any one of the foregoing embodiments when running the computer program.
  • the hardware function of the second memory 23 is similar to that of the first memory 12
  • the hardware function of the second processor 24 is similar to that of the first processor 14 ; details will not be described here.
  • This embodiment provides a decoder, which may include an analyzing unit and a second determining unit.
  • the block type of the current block can be determined based on the color parameters of the current block, and whether to use MIP for encoding prediction can be determined based on the block type of the current block, and MIP can be skipped for screen content blocks with a small number of brightness sampling values and a wide range of brightness sampling
  • the encoding process significantly reduces encoding complexity and encoding time, thereby improving encoding and decoding efficiency.
  • the embodiment of the present application also provides a codec system, the codec system is composed of the encoder shown in Figure 7 and the codec shown in Figure 9 Decoder composition shown.
  • the embodiment of the present application provides a codec method, an encoder, a decoder, and a computer storage medium.
  • the color parameters of the current block in the video image are determined; and the block type of the current block is determined based on the color parameters of the current block. ;Based on the block type, determine the value of the bypass identification information corresponding to the current block, and the bypass identification information is used to identify whether to use matrix weighted intra prediction MIP to encode the current block; based on the value of the bypass identification information, determine the target block division Mode; encode the current block according to the division mode of the target block.
  • the decoder side analyze the code stream, determine the target block division method of the current block, the prediction mode of the current block, and the residual value of the current block; determine the prediction value of the current block based on the division method and prediction mode of the target block; determine the prediction value based on the prediction value and the residual value to determine the reconstruction value of the current block.
  • the block type of the current block can be determined based on the color parameters of the current block, and whether to use MIP for encoding prediction can be determined based on the block type of the current block, and MIP can be skipped for screen content blocks with a small number of brightness sampling values and a wide range of brightness sampling
  • the encoding process significantly reduces encoding complexity and encoding time, thereby improving encoding and decoding efficiency.

Landscapes

  • Engineering & Computer Science (AREA)
  • Multimedia (AREA)
  • Signal Processing (AREA)
  • Compression Or Coding Systems Of Tv Signals (AREA)

Abstract

Sont divulgués dans les modes de réalisation de la présente demande un procédé de codage, un procédé de décodage, un codeur, un décodeur et un support de stockage informatique. Le procédé de codage comprend les étapes suivantes : déterminer un type de bloc du bloc courant ; sur la base du type de bloc, déterminer une valeur d'informations d'identification de dérivation correspondant au bloc courant, les informations d'identification de dérivation étant utilisées pour identifier s'il faut coder le bloc courant en utilisant une prédiction intra pondérée par matrice (MIP) ; déterminer un mode de division de bloc cible sur la base de la valeur des informations d'identification de dérivation ; et coder le bloc courant selon le mode de division de bloc cible. De cette manière, il est possible d'utiliser un type de bloc du bloc courant pour déterminer s'il faut effectuer un codage et une prédiction à l'aide d'une MIP, de telle sorte qu'un processus de codage par MIP d'un bloc de contenu d'écran qui a un petit nombre de valeurs d'échantillonnage de luminosité et une large plage d'échantillonnage de luminosité peut être sauté, ce qui réduit significativement la complexité du codage et qui réduit le temps de codage ; de plus, l'efficacité de codage et de décodage peut être améliorée.
PCT/CN2021/102213 2021-06-24 2021-06-24 Procédé de codage, procédé de décodage, codeur, décodeur et support de stockage informatique WO2022266971A1 (fr)

Priority Applications (2)

Application Number Priority Date Filing Date Title
PCT/CN2021/102213 WO2022266971A1 (fr) 2021-06-24 2021-06-24 Procédé de codage, procédé de décodage, codeur, décodeur et support de stockage informatique
CN202180098944.5A CN117413515A (zh) 2021-06-24 2021-06-24 编解码方法、编码器、解码器以及计算机存储介质

Applications Claiming Priority (1)

Application Number Priority Date Filing Date Title
PCT/CN2021/102213 WO2022266971A1 (fr) 2021-06-24 2021-06-24 Procédé de codage, procédé de décodage, codeur, décodeur et support de stockage informatique

Publications (1)

Publication Number Publication Date
WO2022266971A1 true WO2022266971A1 (fr) 2022-12-29

Family

ID=84544013

Family Applications (1)

Application Number Title Priority Date Filing Date
PCT/CN2021/102213 WO2022266971A1 (fr) 2021-06-24 2021-06-24 Procédé de codage, procédé de décodage, codeur, décodeur et support de stockage informatique

Country Status (2)

Country Link
CN (1) CN117413515A (fr)
WO (1) WO2022266971A1 (fr)

Cited By (1)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
WO2024149398A1 (fr) * 2023-01-13 2024-07-18 Douyin Vision Co., Ltd. Procédé, appareil et support de traitement vidéo

Citations (5)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
KR20190122617A (ko) * 2009-05-12 2019-10-30 엘지전자 주식회사 비디오 신호 처리 방법 및 장치
CN111541896A (zh) * 2020-04-27 2020-08-14 中南大学 一种基于vvc的帧内预测模式的优化方法及系统
CN111698504A (zh) * 2019-03-11 2020-09-22 杭州海康威视数字技术股份有限公司 编码方法、解码方法及装置
CN112118448A (zh) * 2019-06-21 2020-12-22 杭州海康威视数字技术股份有限公司 一种编解码方法、装置及存储介质
CN112689146A (zh) * 2020-12-18 2021-04-20 重庆邮电大学 一种基于启发学习的vvc帧内预测快速模式选择方法

Patent Citations (5)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
KR20190122617A (ko) * 2009-05-12 2019-10-30 엘지전자 주식회사 비디오 신호 처리 방법 및 장치
CN111698504A (zh) * 2019-03-11 2020-09-22 杭州海康威视数字技术股份有限公司 编码方法、解码方法及装置
CN112118448A (zh) * 2019-06-21 2020-12-22 杭州海康威视数字技术股份有限公司 一种编解码方法、装置及存储介质
CN111541896A (zh) * 2020-04-27 2020-08-14 中南大学 一种基于vvc的帧内预测模式的优化方法及系统
CN112689146A (zh) * 2020-12-18 2021-04-20 重庆邮电大学 一种基于启发学习的vvc帧内预测快速模式选择方法

Cited By (1)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
WO2024149398A1 (fr) * 2023-01-13 2024-07-18 Douyin Vision Co., Ltd. Procédé, appareil et support de traitement vidéo

Also Published As

Publication number Publication date
CN117413515A (zh) 2024-01-16

Similar Documents

Publication Publication Date Title
JP7277616B2 (ja) ビデオ・データを処理する方法、装置及び記憶媒体
CN113711597B (zh) 上下文建模和多重变换矩阵的选择
WO2022104498A1 (fr) Procédé de prédiction intra-trame, codeur, décodeur et support de stockage informatique
CN111327904B (zh) 图像重建方法和装置
WO2021004152A1 (fr) Procédé de prédiction de composante d'image, codeur, décodeur et support de mémoire
CN104205843A (zh) Hevc中无损编码中的改进帧内预测
WO2021139572A1 (fr) Procédé de codage, procédé de décodage, codeur, décodeur, et support de stockage
WO2022067805A1 (fr) Procédé de prédiction d'image, codeur, décodeur et support de stockage informatique
WO2022116824A1 (fr) Procédé de décodage vidéo, procédé de codage vidéo, dispositifs associés et support d'enregistrement
CN115988223A (zh) 帧内预测模式的确定、图像编码以及图像解码方法
WO2022266971A1 (fr) Procédé de codage, procédé de décodage, codeur, décodeur et support de stockage informatique
WO2020224476A1 (fr) Procédé, appareil et dispositif de division d'image
WO2020147514A1 (fr) Codeur vidéo, décodeur vidéo et procédé correspondant
WO2022227082A1 (fr) Procédés de division de bloc, codeurs, décodeurs et support de stockage informatique
CN118158441A (zh) 图像划分方法及装置
CN116567232A (zh) 图像块的划分方法、视频编码方法、装置以及设备
WO2020143684A1 (fr) Procédé, dispositif, appareil et système de prédiction d'image et support d'enregistrement
CN116982262A (zh) 视频编码中依赖性量化的状态转换
WO2021134303A1 (fr) Procédé de transformation, codeur, décodeur et support de stockage
WO2024007120A1 (fr) Procédé de codage et de décodage, codeur, décodeur et support de stockage
WO2023193254A1 (fr) Procédé de décodage, procédé de codage, décodeur et codeur
WO2023141970A1 (fr) Procédé de décodage, procédé de codage, décodeur, codeur et système de codage et de décodage
WO2024077553A1 (fr) Procédé et appareil de codage vidéo, procédé et appareil de décodage vidéo, dispositif, système, et support de stockage
WO2024108391A1 (fr) Procédé et appareil de codage vidéo, procédé et appareil de décodage vidéo, et dispositifs, système et support d'enregistrement
WO2023193260A1 (fr) Procédé de codage/décodage, flux de codes, codeur, décodeur et support de stockage

Legal Events

Date Code Title Description
121 Ep: the epo has been informed by wipo that ep was designated in this application

Ref document number: 21946466

Country of ref document: EP

Kind code of ref document: A1

WWE Wipo information: entry into national phase

Ref document number: 202180098944.5

Country of ref document: CN

NENP Non-entry into the national phase

Ref country code: DE

122 Ep: pct application non-entry in european phase

Ref document number: 21946466

Country of ref document: EP

Kind code of ref document: A1