WO2020132908A1 - 解码预测方法、装置及计算机存储介质 - Google Patents

解码预测方法、装置及计算机存储介质 Download PDF

Info

Publication number
WO2020132908A1
WO2020132908A1 PCT/CN2018/123657 CN2018123657W WO2020132908A1 WO 2020132908 A1 WO2020132908 A1 WO 2020132908A1 CN 2018123657 W CN2018123657 W CN 2018123657W WO 2020132908 A1 WO2020132908 A1 WO 2020132908A1
Authority
WO
WIPO (PCT)
Prior art keywords
reference pixel
point
midpoint
edge
block
Prior art date
Application number
PCT/CN2018/123657
Other languages
English (en)
French (fr)
Inventor
霍俊彦
马彦卓
万帅
杨付正
李新伟
Original Assignee
Oppo广东移动通信有限公司
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Priority to CN202310449638.2A priority Critical patent/CN116456081A/zh
Priority to PCT/CN2018/123657 priority patent/WO2020132908A1/zh
Application filed by Oppo广东移动通信有限公司 filed Critical Oppo广东移动通信有限公司
Priority to SG11202102435VA priority patent/SG11202102435VA/en
Priority to BR112021006547A priority patent/BR112021006547A2/pt
Priority to KR1020217023381A priority patent/KR20210106547A/ko
Priority to MX2021003940A priority patent/MX2021003940A/es
Priority to JP2021518718A priority patent/JP7481332B2/ja
Priority to CA3222255A priority patent/CA3222255A1/en
Priority to CN202310451318.0A priority patent/CN116456082A/zh
Priority to CN202110246654.2A priority patent/CN113068025B/zh
Priority to CN202310454159.XA priority patent/CN116614622A/zh
Priority to CN202310455271.5A priority patent/CN116456084A/zh
Priority to CN202110270601.4A priority patent/CN113068026B/zh
Priority to CN202310452465.XA priority patent/CN116456083A/zh
Priority to CA3113787A priority patent/CA3113787C/en
Priority to EP18945329.3A priority patent/EP3829167A4/en
Priority to AU2018454766A priority patent/AU2018454766A1/en
Priority to CN201880095849.8A priority patent/CN112470468A/zh
Publication of WO2020132908A1 publication Critical patent/WO2020132908A1/zh
Priority to PH12021550515A priority patent/PH12021550515A1/en
Priority to IL281399A priority patent/IL281399A/en
Priority to ZA2021/01726A priority patent/ZA202101726B/en
Priority to US17/234,639 priority patent/US11297313B2/en
Priority to US17/677,802 priority patent/US11683477B2/en
Priority to US17/677,810 priority patent/US11785208B2/en
Priority to US17/677,807 priority patent/US11683478B2/en
Priority to US17/677,813 priority patent/US11677936B2/en
Priority to US18/328,705 priority patent/US20230308643A1/en

Links

Images

Classifications

    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04NPICTORIAL COMMUNICATION, e.g. TELEVISION
    • H04N19/00Methods or arrangements for coding, decoding, compressing or decompressing digital video signals
    • H04N19/10Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using adaptive coding
    • H04N19/102Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using adaptive coding characterised by the element, parameter or selection affected or controlled by the adaptive coding
    • H04N19/103Selection of coding mode or of prediction mode
    • H04N19/105Selection of the reference unit for prediction within a chosen coding or prediction mode, e.g. adaptive choice of position and number of pixels used for prediction
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04NPICTORIAL COMMUNICATION, e.g. TELEVISION
    • H04N19/00Methods or arrangements for coding, decoding, compressing or decompressing digital video signals
    • H04N19/10Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using adaptive coding
    • H04N19/102Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using adaptive coding characterised by the element, parameter or selection affected or controlled by the adaptive coding
    • H04N19/132Sampling, masking or truncation of coding units, e.g. adaptive resampling, frame skipping, frame interpolation or high-frequency transform coefficient masking
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04NPICTORIAL COMMUNICATION, e.g. TELEVISION
    • H04N19/00Methods or arrangements for coding, decoding, compressing or decompressing digital video signals
    • H04N19/10Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using adaptive coding
    • H04N19/134Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using adaptive coding characterised by the element, parameter or criterion affecting or controlling the adaptive coding
    • H04N19/146Data rate or code amount at the encoder output
    • H04N19/149Data rate or code amount at the encoder output by estimating the code amount by means of a model, e.g. mathematical model or statistical model
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04NPICTORIAL COMMUNICATION, e.g. TELEVISION
    • H04N19/00Methods or arrangements for coding, decoding, compressing or decompressing digital video signals
    • H04N19/10Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using adaptive coding
    • H04N19/169Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using adaptive coding characterised by the coding unit, i.e. the structural portion or semantic portion of the video signal being the object or the subject of the adaptive coding
    • H04N19/17Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using adaptive coding characterised by the coding unit, i.e. the structural portion or semantic portion of the video signal being the object or the subject of the adaptive coding the unit being an image region, e.g. an object
    • H04N19/176Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using adaptive coding characterised by the coding unit, i.e. the structural portion or semantic portion of the video signal being the object or the subject of the adaptive coding the unit being an image region, e.g. an object the region being a block, e.g. a macroblock
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04NPICTORIAL COMMUNICATION, e.g. TELEVISION
    • H04N19/00Methods or arrangements for coding, decoding, compressing or decompressing digital video signals
    • H04N19/10Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using adaptive coding
    • H04N19/169Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using adaptive coding characterised by the coding unit, i.e. the structural portion or semantic portion of the video signal being the object or the subject of the adaptive coding
    • H04N19/182Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using adaptive coding characterised by the coding unit, i.e. the structural portion or semantic portion of the video signal being the object or the subject of the adaptive coding the unit being a pixel
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04NPICTORIAL COMMUNICATION, e.g. TELEVISION
    • H04N19/00Methods or arrangements for coding, decoding, compressing or decompressing digital video signals
    • H04N19/10Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using adaptive coding
    • H04N19/169Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using adaptive coding characterised by the coding unit, i.e. the structural portion or semantic portion of the video signal being the object or the subject of the adaptive coding
    • H04N19/186Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using adaptive coding characterised by the coding unit, i.e. the structural portion or semantic portion of the video signal being the object or the subject of the adaptive coding the unit being a colour or a chrominance component
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04NPICTORIAL COMMUNICATION, e.g. TELEVISION
    • H04N19/00Methods or arrangements for coding, decoding, compressing or decompressing digital video signals
    • H04N19/44Decoders specially adapted therefor, e.g. video decoders which are asymmetric with respect to the encoder

Definitions

  • the embodiments of the present application relate to the technical field of video encoding and decoding, and in particular, to a decoding prediction method, device, and computer storage medium.
  • H.265/High Efficiency Video Coding (HEVC) is the latest international video compression standard.
  • the compression performance of H.265/HEVC is higher than the previous generation video coding standard H.264/Advanced Video Coding (Advanced Video Coding).
  • AVC Advanced Video Coding
  • AVC increased by about 50%, but still can not meet the rapid development of video applications, especially ultra-high definition, virtual reality (Virtual Reality, VR) and other new video applications.
  • JEM Joint Exploration Test Model
  • VVC Versatile Video Coding
  • VTM Versatile Video Coding
  • a codec prediction method based on a linear model has been integrated, through which the chrominance component of the current block to be decoded is predicted from the linear model.
  • the subset of neighboring reference pixels constructed by neighboring reference pixels is unreasonable, which results in a high search complexity and reduces the decoding prediction performance of video images.
  • the embodiments of the present application are expected to provide a decoding prediction method, device, and computer storage medium. Since the adjacent reference pixels in the adjacent reference pixel subset not only consider the importance and dispersion, but also the adjacent reference pixel subset The number of pixels is small, which reduces the search complexity, improves the prediction performance of video image decoding, and thus reduces the bit rate.
  • an embodiment of the present application provides a decoding prediction method.
  • the method includes:
  • a reference pixel corresponding to the position of the reference pixel to be selected is selected from the first set of adjacent reference pixels, and the selected reference pixels are combined to form a neighboring reference pixel set;
  • an embodiment of the present application provides a decoding prediction device, the decoding prediction device includes an acquisition unit, a determination unit, a selection unit, and a decoding unit,
  • the obtaining unit is configured to obtain reference pixel points adjacent to at least one side of the block to be decoded to obtain a first set of adjacent reference pixels;
  • the determining unit is configured to determine a reference point from the at least one edge, and determine the position of the reference pixel point to be selected corresponding to the at least one edge according to a preset number of samples;
  • the selecting unit is configured to select a reference pixel corresponding to the position of the reference pixel to be selected from the first set of adjacent reference pixels based on the position of the reference pixel to be selected, and to select the selected reference pixel Points form a subset of adjacent reference pixels;
  • the decoding unit is configured to perform predictive decoding on the block to be decoded based on the adjacent reference pixel subset.
  • an embodiment of the present application provides a decoding prediction device, the decoding prediction device includes: a memory and a processor;
  • the memory is used to store a computer program that can run on the processor
  • the processor is configured to execute the steps of the method according to the first aspect when running the computer program.
  • an embodiment of the present application provides a computer storage medium that stores a decoding prediction program, and when the decoding prediction program is executed by at least one processor, the steps of the method according to the first aspect are implemented.
  • Embodiments of the present application provide a decoding prediction method, device, and computer storage medium.
  • reference pixel points adjacent to at least one side of a block to be decoded are obtained to obtain a first set of adjacent reference pixels; and then from the at least one side Determine a reference point, and determine the position of the reference pixel to be selected corresponding to the at least one edge according to the preset number of samples; and then select from the first neighboring reference pixel set based on the position of the reference pixel to be selected Reference pixel points corresponding to the positions of the reference pixel points to be selected, the selected reference pixel points form a subset of adjacent reference pixels; finally, based on the subset of adjacent reference pixels, predictive decoding of the block to be decoded ; Because the selection of adjacent reference pixels in the adjacent reference pixel subset takes into account both importance and dispersion, the model parameters constructed based on the adjacent reference pixel subset are more accurate, which can improve the prediction performance of video image decoding. ; And the number of pixels in the adjacent reference pixel
  • FIGS. 1A to 1C are schematic structural diagrams of video image sampling formats in related technical solutions, respectively;
  • FIGS. 2A and 2B are schematic diagrams of sampling the adjacent reference values of the first image component and the adjacent reference values of the second image component of the decoding block in the related technical solution, respectively;
  • FIG. 3 is a schematic structural diagram of a prediction model constructed by a decoding block based on a maximum value and a minimum value in a related technical solution;
  • 4A and 4B are schematic structural diagrams of selecting a neighboring reference pixel for a square decoding block and a non-square decoding block provided by the related technical solutions;
  • 5A and 5B are a schematic diagram of a conventional technical solution provided by the related technical solutions and a structure of selecting adjacent reference pixels in the L0138 proposal;
  • FIG. 6 is a schematic block diagram of a composition of a video encoding system provided by an embodiment of the present application.
  • FIG. 7 is a schematic block diagram of a composition of a video decoding system provided by an embodiment of the present application.
  • FIG. 8 is a schematic flowchart of a decoding prediction method provided by an embodiment of the present application.
  • FIG. 9 is a schematic structural diagram of selecting a subset of adjacent reference pixels corresponding to an upper side of a block to be decoded according to an embodiment of the present application.
  • FIG. 10 is a schematic structural diagram of selecting a subset of adjacent reference pixels corresponding to the left side of a block to be decoded according to an embodiment of the present application
  • FIG. 11 is a schematic structural diagram of selecting another subset of adjacent reference pixels corresponding to the upper side of a block to be decoded according to an embodiment of the present application;
  • FIG. 12 is a schematic structural diagram of selecting another subset of adjacent reference pixels corresponding to the upper side of a block to be decoded according to an embodiment of the present application
  • FIG. 13 is a schematic structural diagram of selecting another subset of adjacent reference pixels corresponding to the upper side of a block to be decoded according to an embodiment of the present application
  • FIG. 14 is a schematic structural diagram of selecting another subset of adjacent reference pixels corresponding to the upper side of a block to be decoded according to an embodiment of the present application;
  • FIG. 15 is a schematic structural diagram of a decoding prediction device according to an embodiment of the present application.
  • 16 is a schematic diagram of a specific hardware structure of a decoding prediction device provided by an embodiment of the present application.
  • the first image component, the second image component, and the third image component are generally used to characterize the decoded block; where the three image components are a luminance component, a blue chroma component, and a red chroma component, respectively Specifically, the luminance component is generally represented by the symbol Y, the blue chrominance component is generally represented by the symbol Cb, and the red chrominance component is generally represented by the symbol Cr.
  • the first image component may be the luminance component Y
  • the second image component may be the blue chrominance component Cb
  • the third image component may be the red chrominance component Cr
  • the commonly used sampling format is the YCbCr format.
  • the YCbCr format includes the following types, as shown in FIGS. 1A to 1C, respectively, where the cross (X) in the figure indicates the sampling point of the first image component, and the circle ( ⁇ ) indicates the second Image component or third image component sampling point.
  • YCbCr format includes:
  • 4:4:4 format as shown in Figure 1A, it means that the second image component or the third image component is not downsampled; it takes 4 first image components for every 4 consecutive pixels on each scan line Sampling samples, 4 sampling samples of the second image component and 4 sampling samples of the third image component;
  • 4:2:2 format as shown in Figure 1B, it means that the first image component is sampled horizontally with respect to the second image component or the third image component, without vertical downsampling; it is on each scan line Take 4 sample samples of the first image component, 2 sample samples of the second image component and 2 sample samples of the third image component every 4 consecutive pixel points;
  • 4:2:0 format as shown in Figure 1C, it means that the first image component is 2:1 horizontal downsampling and 2:1 vertical downsampling with respect to the second image component or the third image component; Every 2 consecutive pixels on the horizontal scanning line and the vertical scanning line take 2 sample samples of the first image component, 1 sample sample of the second image component and 1 sample sample of the third image component.
  • the video image adopts the YCbCr 4:2:0 format
  • the first image component of the video image is a 2N ⁇ 2N size decoding block
  • the corresponding second image component or third image component is N ⁇ N size Of the decoded block, where N is the side length of the decoded block.
  • the following will take the 4:2:0 format as an example for description, but the technical solutions of the embodiments of the present application are also applicable to other sampling formats.
  • CCLM implements the prediction between the first image component to the second image component, the first image component to the third image component, and the second image component and the third image component.
  • the prediction from the component to the second image component is described as an example, but the technical solutions of the embodiments of the present application can also be applied to the prediction of other image components.
  • the prediction mode of CCLM is used in the VTM, in which case the first image component and the second image component are of the same decoding block, the second The image component is predicted based on the reconstruction value of the first image component of the same decoding block, for example, using the prediction model shown in equation (1):
  • i, j represent the position coordinates of the pixels in the decoded block
  • i represents the horizontal direction
  • j represents the vertical direction
  • Pred C [i,j] represents the pixel coordinates in the decoded block with position coordinates [i,j]
  • the predicted value of the second image component, Rec L [i,j] represents the reconstructed value of the first image component corresponding to the pixel with the position coordinate of [i,j] (downsampled) in the same decoding block
  • ⁇ and ⁇ are the above Predict the model parameters of the model.
  • model parameters ⁇ and ⁇ which can be the regression construction method evaluated by the least square method, the construction method of the maximum value and the minimum value, or even other construction methods. This is not specifically limited. The following will take the construction method of regression based on least squares and the construction method of maximum and minimum values as examples.
  • L(n) represents the adjacent reference value of the first image component corresponding to the down-sampled left side and the upper side
  • C(n) represents the phase of the second image component corresponding to the left side and the upper side
  • N is the side length of the decoded block of the second image component
  • n 1, 2, ..., 2N.
  • 2A and 2B which respectively show schematic diagrams of sampling the adjacent reference value of the first image component and the adjacent reference value of the second image component of the decoding block in the related technical solution; wherein, in FIG. 2A, the bold The larger box is used to highlight the first image component decoding block 21, and the gray solid circle is used to indicate the adjacent reference value L(n) of the first image component decoding block 21; in FIG.
  • FIG. 2A shows a 2N ⁇ 2N size first image component decoding block 21.
  • a 2N ⁇ 2N size first image component corresponds to the size of a second image component N ⁇ N, as shown at 22 in FIG. 2B; that is, FIGS. 2A and 2B are schematic diagrams of decoding blocks obtained by sampling the first image component and the second image component for the same decoding block, respectively.
  • L max and L min represent the maximum and minimum values obtained by searching the adjacent reference values of the first image component corresponding to the left side and the upper side after downsampling
  • C max and C min represent L max and L min The adjacent reference value of the second image component corresponding to the reference pixel at the corresponding position.
  • FIG. 4A it shows a schematic structural diagram of a square decoding block for selecting adjacent reference pixels provided by the related technical solution; as shown in FIG. 4A, the decoding block is a square decoding block, and the left side of the decoding block and All adjacent pixels corresponding to the upper side can be used as reference pixels; for the first image component, downsampling is required first, so that the downsampled first image component has the same resolution as the second image component.
  • gray solid circles are used to represent adjacent reference pixel points selected by the square decoding block.
  • FIG. 4B it shows a schematic structural diagram of a non-square decoding block provided by a related technical solution to select neighboring reference pixels; as shown in FIG. 4B, the decoding block is a non-square decoding block, and the width and width of the decoding block The heights are not equal.
  • the first image component needs to be down-sampled first, so that the down-sampled first image component and the second image component have the same resolution; on the other hand, the adjacent pixels corresponding to the long side of the decoded block The points need to be further down-sampled so that the number of adjacent reference pixel points obtained on the long side is equal to the number of adjacent reference pixel points on the short side.
  • gray solid circles are used to represent adjacent reference pixel points selected by the non-square decoding block.
  • the pixels in a row adjacent to the upper side and the adjacent pixels in a column adjacent to the left side can be used as adjacent Reference pixel
  • the adjacent pixel corresponding to the effective side is used as the adjacent reference pixel
  • the model parameter ⁇ is set to 0, and the model parameter ⁇ is set to the middle of the second image component Value 512, that is, the predicted value of the second image component corresponding to all pixels in the decoded block is 512.
  • N is defined as the length of the short side of the decoded block of the second image component
  • N is defined as the length of the short side of the decoded block of the second image component
  • adjacent pixels corresponding to the left and upper sides of the decoded block there are a total of 2N adjacent Reference pixels can be used in CCLM.
  • the following description is based on the case where adjacent pixel points corresponding to the left and upper sides of the decoded block can be used.
  • the required computational complexity is proportional to the number of adjacent reference pixels used for CCLM.
  • the L0138 proposal of the 12th meeting of JVET proposes a technical solution for reducing the number of adjacent reference pixel points for CCLM based on the size of the second image component block corresponding to the block to be decoded.
  • Table 1 shows a relationship table between the size of the second image component block corresponding to the block to be decoded and the number of adjacent reference pixels provided by the related technical solution.
  • N 1 is the number of adjacent reference pixels used for CCLM in the traditional technical solution
  • N 2 is the number of adjacent reference pixels used for CCLM after the minus point provided by the L0138 proposal.
  • the point reduction method in the L0138 proposal is obtained by downsampling adjacent reference pixel points after acquiring them in the traditional technical solution.
  • VTM For non-square decoded blocks, a long-side downsampling scheme has been provided in VTM (as shown in FIG. 4B). And as in the L0138 proposal described above, this proposal proposes a further downsampling set of downsampling.
  • the selection of neighboring reference pixels in the downsampling process will have an impact on decoding prediction performance.
  • the subset of neighboring reference pixels obtained by the default downsampling scheme in VTM the corresponding decoding prediction performance is Not ideal, so the L0138 proposal proposes a selection scheme that can use another subset of adjacent reference pixel points during downsampling.
  • the pixels of the adjacent row corresponding to the upper side of the decoded block are selected from the leftmost edge.
  • FIG. 5A it shows a schematic structural diagram of selecting a conventional reference pixel provided by a related technical solution; as shown in FIG. 5A, for the pixels of the adjacent row corresponding to the long side of the decoding block The pixels on the left edge are taken for sampling; see FIG. 5B, which shows a schematic diagram of the structure of the L0138 proposal to select adjacent reference pixels provided by the related technical solutions; as shown in FIG. 5B, for the long side of the decoded block The pixel points of the corresponding adjacent row are sampled from the pixel point on the rightmost edge; the sampling interval here is the same as the traditional technical solution, and no further details are provided in the embodiments of the present application.
  • the value range of all adjacent reference point pixels can be more comprehensively covered in the calculation of CCLM model parameters, but due to the lack of consideration of the importance of each adjacent reference pixel point, it is limited Under the premise of the number, it still cannot represent the overall characteristics of the long side to the maximum extent.
  • the larger the length of the decoded block the greater the number of adjacent reference pixels used for CCLM, resulting in a greater number of searches for determining the maximum and minimum values. Therefore, the The computational complexity is still relatively large.
  • the correlation between the searched pixel and the current decoded block is very low, it can be confirmed that the pixel is a bad point. In this way, if the two pixels obtained by the search contain dead pixels, a large model deviation will be caused to the prediction model. Therefore, although the method of constructing model parameters based on the maximum and minimum values is of low complexity, its robustness is not good, which degrades the decoding prediction performance.
  • embodiments of the present application provide a decoding prediction method, by using the midpoint of at least one edge of the block to be decoded as a reference point, and determining the number of samples to be selected according to a preset number of samples Reference pixel positions, to put adjacent reference pixel points selected by taking into account importance and dispersion into adjacent reference pixel subsets; based on the adjacent reference pixel subsets, predictively decode the block to be decoded ; In this way, because the model parameters constructed based on the adjacent reference pixel subset are more accurate, the decoding prediction performance can be improved; and the number of pixels in the adjacent reference pixel subset is less, which also reduces the complexity of the search Degrees, which in turn reduces the bit rate.
  • the embodiments of the present application will be described in detail below with reference to the drawings.
  • the video encoding system 600 includes transform and quantization 601, intra estimation 602, and intra prediction 603. , Motion compensation 604, motion estimation 605, inverse transform and inverse quantization 606, filter control analysis 607, deblocking filtering and sample adaptive indentation (Sample Adaptive ffset, SAO) filtering 608, header information encoding and context-based adaptation Binary arithmetic coding (Context-based Adaptive Binary Arithmatic Coding, CABAC) 609 and decoded image buffer 610 and other components; for the original input video signal, a video coding block can be obtained by dividing the coding tree block (Coding Tree Unit, CTU), Then, the residual pixel information obtained after intra-frame or inter-frame prediction is transformed and quantized 601 to transform the video coding block, including transforming the residual information from the pixel domain to the transform domain, and quantizing the resulting
  • CTU Coding Tree Unit
  • the reconstructed residual block is subjected to filter control analysis 607 and deblocking filtering and SAO filtering 608 Remove blockiness artifacts, and then add the reconstructed residual block to a predictive block in the frame of the decoded image buffer 610 to generate a reconstructed video coding block; header encoding and CABAC 609 are used for encoding Various encoding parameters and quantized transform coefficients.
  • the context content can be based on adjacent encoding blocks, which can be used to encode information indicating the determined intra-prediction mode and output the code stream of the video signal;
  • the decoded image buffer 610 is used to store the reconstructed video coding block for prediction reference. As the video image encoding proceeds, new reconstructed video encoding blocks will be continuously generated, and these reconstructed video encoding blocks will be stored in the decoded image buffer 610.
  • the video decoding system 700 includes header information decoding and CABAC decoding 701, inverse transform and inverse quantization 702 , Intra prediction 703, motion compensation 704, de-blocking filtering and SAO filtering 705, and decoded image buffer 706 and other components; after the input video signal undergoes the encoding process of FIG.
  • the code stream of the video signal is output; the code stream is input to the video In the decoding system 700, firstly through header information decoding and CABAC decoding 701, it is used to obtain the decoded transform coefficients; the transform coefficients are processed by inverse transform and inverse quantization 702 to generate residual blocks in the pixel domain; intra Prediction 703 can be used to generate prediction data for the current video decoding block based on the determined intra prediction mode and data from the previous decoded block of the current frame or picture; motion compensation 704 is determined by parsing the motion vector and other associated syntax elements Prediction information for the video decoding block, and use the prediction information to generate the predictive block of the video decoding block being decoded; by the residual block from the inverse transform and inverse quantization 702 and by intra prediction 703 or motion compensation The corresponding predictive blocks generated by 704 are summed to form a decoded video block; the decoded video signal is deblocked and SAO filtered 705 to remove blockiness artifacts, which can improve the video quality; then the
  • the embodiment of the present application is mainly applied to the intra prediction part 603 shown in FIG. 6 and the intra prediction part 703 shown in FIG. 7; that is to say, the embodiment of the present application can be applied to both an encoding system and an application Decoding system, but this embodiment of the present application does not specifically limit this.
  • FIG. 8 shows a schematic flowchart of a decoding prediction method provided by an embodiment of the present application.
  • the method may include:
  • S802 Determine a reference point from the at least one edge, and determine the position of the reference pixel point to be selected corresponding to the at least one edge according to the preset number of samples;
  • S804 Perform predictive decoding on the block to be decoded based on the adjacent reference pixel subset.
  • the block to be decoded is currently a decoding block to be subjected to second image component prediction or third image component prediction.
  • At least one side of the block to be decoded may refer to the upper side of the block to be decoded, or may refer to the left side of the block to be decoded, or may even refer to the upper side and the left side of the block to be decoded. No specific restrictions.
  • the reference point may be the midpoint of the at least one side, or the position of the first reference pixel to the left of the midpoint of the at least one side, or may be the position of the at least one side.
  • the position of the first reference pixel point to the right of the midpoint may even be the position of other reference pixel points of the at least one side, which is not specifically limited in this embodiment of the present application.
  • the decoding prediction method of the embodiments of the present application can also be applied to the encoding system.
  • the encoding prediction performance of the video image can be improved, and the encoding compression can be improved. Efficiency to save coding rate.
  • the construction of the adjacent reference pixel subset in the decoding system will be described as an example.
  • reference pixel points adjacent to at least one side of the block to be decoded are obtained to obtain a first set of adjacent reference pixels; then a reference point is determined from the at least one side, and sampled according to a preset number Determine the position of the reference pixel to be selected corresponding to the at least one edge; and then select the position corresponding to the position of the reference pixel to be selected from the first neighboring reference pixel set based on the position of the reference pixel to be selected Reference pixel points, and the selected reference pixel points form a subset of neighboring reference pixels; finally, based on the subset of neighboring reference pixels, predictively decode the block to be decoded; All reference pixel points adjacent to the upper or left side of the block to be decoded are all involved in the decoding prediction search operation, but on the premise of taking into account the importance and dispersion, by selecting the adjacent reference pixels in the appropriate position to A subset of adjacent reference pixels is formed, so that the number of pixels in the adjacent reference pixel sub
  • the determining a reference point from the at least one edge includes:
  • the midpoint of the at least one side is determined based on the length of the at least one side, and the midpoint of the at least one side is used as the reference point.
  • the determining a reference point from the at least one edge includes:
  • the position of the first reference pixel point with the middle position to the right is taken as the at least one The reference point of one side, or the position of the first reference pixel point with the middle position to the left as the reference point of the at least one side.
  • the importance of the reference pixel adjacent to at least one side of the block to be decoded is related to its corresponding position, in order to make the reference pixel in the adjacent reference pixel subset represent the characteristics of the entire adjacent side, It is necessary to select the reference pixel located at the center of the edge as much as possible, so as to eliminate the points of relatively low importance (such as the reference pixels on both sides of the edge).
  • the position of the first reference pixel point whose middle position is right or left may be used as the reference point of the side.
  • FIG. 9 shows a schematic structural diagram of selecting a subset of adjacent reference pixels corresponding to the upper side of a block to be decoded provided by an embodiment of the present application.
  • the midpoint of the side is selected as the center (the dotted line position shown in FIG. 9 ), and the center is used as the reference point for reference Pixel selection.
  • the midpoint is between 7 and 8, that is, the midpoint can be selected from 7 or 8; in FIG. 9, 8 is used as the reference point as
  • the preset number of samples is 4, you can determine the position of the reference pixels to be selected (as shown in the example of gray dots in Figure 9) as 2, 6, 10, and 14; according to these reference pixel positions, you can choose Corresponding reference pixels form a subset of adjacent reference pixels.
  • the determining a reference point from the at least one edge includes:
  • the position of the first reference pixel point at the lower side of the intermediate position is used as the at least one
  • the reference point of one side, or the position of the first reference pixel above the middle position is used as the reference point of the at least one side.
  • the importance of the reference pixel adjacent to at least one side of the block to be decoded is related to its corresponding position, in order to make the reference pixel in the adjacent reference pixel subset represent the characteristics of the entire adjacent side, It is necessary to select the reference pixel located at the center of the edge as much as possible, so as to eliminate the points of relatively low importance (such as the reference pixels on both sides of the edge).
  • the position of the first reference pixel point whose middle position is lower or upper may be used as the reference point of the side.
  • FIG. 10 shows a schematic structural diagram of selecting a subset of adjacent reference pixels corresponding to the left side of a block to be decoded provided by an embodiment of the present application.
  • the midpoint of the side is selected as the center (the dotted line position shown in FIG. 10), and the center is used as the reference point for reference Pixel selection.
  • the midpoint is between 3 and 4, that is, the midpoint can be selected from 3 or 4; in FIG. 10, 4 is selected as the reference point
  • the preset number of samples is 2, you can determine the position of the reference pixel to be selected (as shown in the example of gray dots in Figure 10) as 2 and 6; according to the position of these reference pixels, you can select the corresponding reference pixel Points to form a subset of adjacent reference pixels.
  • the middle position of the left side or the upper side of the block to be decoded is between two points.
  • the first pixel point at the middle position to the right is used as the midpoint of the side; however, the embodiment of the present application may also use the first pixel point at the center position to the left as the midpoint of the side ,
  • FIG. 11 An example of the structure shown in FIG. 11.
  • the first pixel point at the left of the middle position (3 in FIG. 11) is taken as the midpoint of the side. Since the preset number of samples is 2, then the reference pixel point to be selected can be determined
  • the positions (such as the gray dot examples in FIG.
  • the reference pixel points corresponding to these reference pixel point positions can also be selected to form a subset of adjacent reference pixels. Therefore, in the embodiment of the present application, for the upper side of the block to be decoded, either the first pixel point with the middle position to the right can be used as the midpoint of the side, or the first pixel with the middle position to the left The point is used as the midpoint of the side, which is not specifically limited in this embodiment of the present application; in addition, for the left side of the block to be decoded, the first pixel point in the middle position can be used as the midpoint of the side, or The first pixel at the upper middle position is taken as the midpoint of the side, which is not specifically limited in the embodiment of the present application.
  • the upper side of the block to be decoded will be described as an example below, but the prediction method of the embodiment of the present application is also applicable to the left side of the block to be decoded, or even other sides of the block to be decoded.
  • represents the sampling interval
  • length represents the number of reference pixels in a row adjacent to the upper side of the block to be decoded, or the number of reference pixels in a column adjacent to the left side of the block to be decoded
  • N 2 represents the pending
  • shift means Select the starting point of the reference pixel.
  • the position of the reference pixel to be selected can be determined, such as 2, 6, 10, and 14, and then the corresponding reference pixel can be selected to form a subset of adjacent reference pixels.
  • the number of preset samples corresponding to the left side and the number of preset samples corresponding to the upper side may be the same or different, and the embodiments of the present application are not specifically limited.
  • the embodiment of the present application also provides a solution for determining the preset number of samples, as shown in Table 2.
  • N 2 ′ in Table 2 can also be used to replace N 2 , which is brought into the above formula (4) and formula (5) for calculation, so that the adjacent reference pixel subset formed More accurate. Thereby improving decoding prediction performance.
  • N represents the length of the short side of the block to be decoded
  • N 1 represents the number of adjacent reference pixels provided in the traditional technical solution
  • N 2 is the number of adjacent reference pixels provided in the L0138 proposal
  • the reference pixel subset contains 4 reference pixels; when the length of the short side of the block to be decoded is greater than 4, the adjacent reference pixel The subset can contain 8 reference pixels.
  • the determining a reference point from the at least one edge, and determining the position of the reference pixel point to be selected corresponding to the at least one edge according to the preset number of samples includes:
  • the position of the reference pixel to be selected corresponding to the at least one edge is determined according to the first sampling interval.
  • using the midpoint of the at least one edge as a reference point, and determining the position of the reference pixel to be selected corresponding to the at least one edge according to the first sampling interval includes:
  • the midpoint value is a non-integer
  • round the reference pixel position on the left side of the midpoint value to a small value and determine the rounded reference pixel position as the reference pixel position to be selected; for the The position of the reference pixel point to the right of the midpoint value is rounded up, and the rounded reference pixel point position is determined as the position of the reference pixel point to be selected.
  • using the midpoint of the at least one edge as a reference point, and determining the position of the reference pixel to be selected corresponding to the at least one edge according to the first sampling interval includes:
  • the midpoint value is a non-integer
  • round the reference pixel position to the left of the midpoint value to a large value and determine the rounded reference pixel position as the reference pixel position to be selected; for the The position of the reference pixel point to the right of the midpoint value is rounded down, and the rounded reference pixel point position is determined as the position of the reference pixel point to be selected.
  • the first sampling interval corresponding to the side can be calculated.
  • the middle position of the left side or the upper side of the block to be decoded is between two points, which is calculated at this time
  • the midpoint value of is a non-integer
  • the calculated reference pixel position is also a non-integer; however, if the length of the left or upper side of the block to be decoded is not an integer multiple of 2, then the left or upper side of the block to be decoded
  • the middle position of the side will not be between two points.
  • the calculated midpoint value is an integer
  • the calculated reference pixel position is also an integer; that is, the calculated midpoint value can be an integer, It may also be a non-integer; correspondingly, the calculated reference pixel position may also be an integer or a non-integer; the embodiments of the present application are not specifically limited.
  • the calculated reference pixel position is also an integer.
  • the calculated reference pixel position can be directly used as the reference pixel position to be selected; when calculated When the midpoint value of is a non-integer, correspondingly, the calculated position of the reference pixel is also a non-integer.
  • the position of the reference pixel to be selected can be determined by rounding up or down.
  • the first sampling interval is 4. Since the length of the upper side is 8, that is to say, the middle position of the upper side is between two points.
  • the reference point position can be calculated to be 1.5 and 5.5; at this time, the reference pixel position on the left of the midpoint value can be rounded down to obtain one of the reference pixel positions to be selected, and the middle The position of the reference pixel on the right of the point value is rounded to a large value, and then another reference pixel position to be selected is 6; this method can also be called an outward rounding scheme, as shown in FIG. 12; this can be Reference pixel points corresponding to positions 1 and 6 constitute a subset of adjacent reference pixels.
  • the reference pixel position on the left side of the midpoint value can also be rounded up to obtain one of the reference pixel positions to be selected is 2, and the middle The position of the reference pixel on the right of the point value is rounded down, you can get another position of the reference pixel to be selected as 5; this method can also be called an inward rounding scheme, as shown in Figure 13; this can be The reference pixel points corresponding to positions 2 and 5 constitute a subset of adjacent reference pixels.
  • the method further includes:
  • the position of the reference pixel to be selected corresponding to the at least one edge is determined according to the second sampling interval.
  • the method further includes:
  • the first sampling interval may be fine-tuned, such as adding or subtracting 1 to the first sampling interval to obtain the second sampling interval.
  • the adjusted second sampling interval may be 3 or 5.
  • the adjustment of the first sampling interval may be adjusted by a small range (for example, plus 1 or minus 1), but the specific setting of the adjustment range is not specifically limited in the embodiment of the present application.
  • uniform sampling can be performed according to the first sampling interval or the second sampling interval; or even sampling can be performed according to the first sampling interval and the second sampling interval.
  • the positions of the reference pixel points to be selected determined afterwards may be symmetrically distributed on both sides of the reference point, or may be distributed asymmetrically on both sides of the reference point; the embodiments of the present application are not specifically limited.
  • the determining a reference point from the at least one edge, and determining the position of the reference pixel point to be selected corresponding to the at least one edge according to the preset number of samples includes:
  • the position of consecutive reference pixels near the reference point is determined as the position of the reference pixel to be selected according to the preset number of samples, and the reference point is located The middle position of the position of the reference pixel to be selected.
  • the adjacent reference pixel points are related to the reference pixel point at the intermediate position in comparison with the reconstruction value of the first image component of the block to be decoded, then a preset number of consecutive samples near the intermediate position can be used for reference
  • the pixel position is used as the reference pixel position to be selected.
  • This method can be referred to as a continuous point-taking scheme for the intermediate position, as shown in FIG. 14; in this way, the reference pixel points corresponding to the 2, 3, and 4 positions can be formed into adjacent reference pixels. Subset.
  • the neighboring reference pixel subset in this embodiment constitutes the neighboring reference
  • Table 3 The number of pixels and corresponding positions of reference pixels to be selected are shown in Table 3.
  • the reference pixel positions of a preset number of consecutive samples near the middle position are used as reference pixel positions to be selected to form a subset of adjacent reference pixels; prediction is performed according to the subset of adjacent reference pixels
  • Decoding based on the L0138 proposal, can still save the bit rate and increase the decoding gain, thereby improving the decoding prediction performance.
  • the block to be decoded includes a square decoded block or a non-square decoded block.
  • the method when the block to be decoded is a non-square decoded block, the method further includes:
  • Delete the reference pixel corresponding to the end position of the long side perform an initial offset on the long side of the deleted reference pixel according to a preset offset, and take the offset reference pixel position as the starting point, according to the first
  • the three sampling intervals sample the long side of the deleted reference pixel to determine the position of the reference pixel to be selected corresponding to the long side.
  • embodiments of the present application can be applied to square decoding blocks or non-square decoding blocks, and the embodiments of the present application are not specifically limited.
  • the block to be decoded is a non-square decoding block
  • the position of the reference pixel is used as the starting point, and then the long side of the deleted reference pixel is sampled to determine the position of the reference pixel to be selected corresponding to the long side.
  • the preset offset may be one-half of the third sampling interval or other values, which is not specifically limited in the embodiment of the present application.
  • reference pixel sampling is performed on the long side of the non-square decoding block as shown in FIG. 4B.
  • the third sampling interval can be determined as 4; the long side is from the first reference pixel position on the left as the starting point, and then selected according to the third sampling interval
  • the same number of reference pixels as the adjacent reference pixels on the short side are generated.
  • the long side of the non-square decoding block is initially offset first, so that the sampled reference pixel point Capable of covering the entire long edge.
  • the preset offset is one-half of the third sampling interval, that is, the preset offset is 2, that is to say, the sampling in this embodiment starts with 2 as the starting point.
  • the subset of pixels can better cover the characteristics of the entire long side.
  • the determining a reference point from the at least one edge, and determining the position of the reference pixel point to be selected corresponding to the at least one edge according to the preset number of samples includes:
  • the position of the reference pixel point to be selected is determined according to a preset number of samples.
  • the embodiments of the present application may directly select reference pixels based on the midpoint of the left or upper side of the block to be decoded to form a subset of adjacent reference pixels; the embodiments of the present application may also First delete the start and end positions corresponding to the left/upper side of the block to be decoded (for example, for the upper side, the start position is the left edge position and the end position is the right edge position; for the left side In terms of edges, the starting position is the upper edge position, and the end position is the lower edge position), corresponding to the reference pixel points, the reference pixel points in the middle part are retained, and then the selected reference pixel points are formed into adjacent Reference pixel subset.
  • the preset number of samples can be any value; generally speaking, the preset number of samples is less than N 1 shown in Table 2; however, the embodiments of the present application are not limited to the preset number of samples The number is equal to N 2 or N 2 '. Since the number of preset samples is smaller than N 1 shown in Table 2, the search complexity of reference pixels can be reduced to improve decoding prediction performance. Specifically, by changing the ⁇ and shift values in the above equations (4) and (5), the position of the reference pixel to be selected can be determined with different starting points and sampling intervals, and then the corresponding Of reference pixels to form a subset of adjacent reference pixels.
  • the predictive decoding of the block to be decoded based on the adjacent reference pixel subset includes:
  • the prediction model is used to characterize the prediction relationship between the first image component and the second image component corresponding to each pixel in the block to be decoded;
  • the model parameters ⁇ and ⁇ can be constructed; in this way, the prediction model can be established according to the above formula (1); according to the prediction model, the block to be decoded can be determined Perform predictive decoding. Because the adjacent reference pixel points included in the adjacent reference pixel subset take into account both importance and dispersion, the constructed model parameters are more accurate, thereby better improving decoding prediction performance, and thus reducing the bit rate.
  • the above embodiment provides a decoding prediction method, by obtaining reference pixel points adjacent to at least one side of the block to be decoded to obtain a first set of adjacent reference pixels; determining a reference point from the at least one side, according to the Set the number of samples to determine the position of the reference pixel to be selected corresponding to the at least one edge; based on the position of the reference pixel to be selected, select the position of the reference pixel to be selected from the first neighboring reference pixel set Corresponding reference pixel points, the selected reference pixel points form a neighboring reference pixel subset; based on the neighboring reference pixel subset, predictive decoding of the block to be decoded; due to the phase of the neighboring reference pixel subset Adjacent reference pixels not only take into account the importance and dispersion, but also have a smaller number of pixels in the adjacent reference pixel subset, thereby reducing the complexity of searching, improving the prediction performance of video image decoding, and thus reducing the bit rate.
  • FIG. 15 shows a schematic structural diagram of a decoding prediction apparatus 150 provided by an embodiment of the present application.
  • the decoding prediction apparatus 150 may include: an obtaining unit 1501, a determining unit 1502, an selecting unit 1503, and a decoding unit 1504, wherein,
  • the obtaining unit 1501 is configured to obtain reference pixel points adjacent to at least one side of the block to be decoded to obtain a first set of adjacent reference pixels;
  • the determining unit 1502 is configured to determine a reference point from the at least one edge, and determine the position of the reference pixel point to be selected corresponding to the at least one edge according to a preset number of samples;
  • the selecting unit 1503 is configured to select a reference pixel corresponding to the position of the reference pixel to be selected from the first set of adjacent reference pixels based on the position of the reference pixel to be selected, and to select the selected reference Pixels form a subset of adjacent reference pixels;
  • the decoding unit 1504 is configured to perform predictive decoding on the block to be decoded based on the adjacent reference pixel subset.
  • the determining unit 1502 is specifically configured to determine the intermediate position based on the length of the at least one side, and if the midpoint of the at least one side is located at an intermediate position between two reference pixel points The position of the first reference pixel point to the right is used as the reference point of the at least one side, or the position of the first reference pixel point to the left of the middle position is used as the reference point of the at least one side.
  • the determining unit 1502 is specifically configured to determine the intermediate position based on the length of the at least one side, and if the midpoint of the at least one side is located at an intermediate position between two reference pixel points The position of the first reference pixel point below is used as the reference point of the at least one side, or the position of the first reference pixel point above the middle position is used as the reference point of the at least one side.
  • the decoding prediction device 150 further includes a calculation unit 1505 configured to calculate a first sampling interval based on the preset number of samples and the length of the at least one side;
  • the determining unit 1502 is specifically configured to determine the position of the reference pixel point to be selected corresponding to the at least one edge according to the first sampling interval using the midpoint of the at least one edge as a reference point.
  • the calculation unit 1505 is further configured to calculate a midpoint value of the at least one side based on the length of the at least one side; and calculate a reference based on the midpoint value and the first sampling interval Pixel position;
  • the determining unit 1502 is specifically configured to, when the midpoint value is a non-integer, round the reference pixel position on the left side of the midpoint value to a small value, and determine the rounded reference pixel position as a pending Select the position of the reference pixel; round the position of the reference pixel to the right of the midpoint value, and determine the position of the rounded reference pixel as the position of the reference pixel to be selected.
  • the calculation unit 1505 is further configured to calculate a midpoint value of the at least one side based on the length of the at least one side; and calculate a reference based on the midpoint value and the first sampling interval Pixel position;
  • the determining unit 1502 is specifically configured to, when the midpoint value is a non-integer value, round the position of the reference pixel on the left side of the midpoint value to a large value, and determine the rounded reference pixel position as a pending Select the position of the reference pixel; round the position of the reference pixel to the right of the midpoint value to a small value, and determine the position of the rounded reference pixel as the position of the reference pixel to be selected.
  • the decoding prediction device 150 further includes an adjustment unit 1506 configured to adjust the first sampling interval to obtain a second sampling interval;
  • the determining unit 1502 is further configured to determine the position of the reference pixel point to be selected corresponding to the at least one edge according to the second sampling interval using the midpoint of the at least one edge as a reference point.
  • the determining unit 1502 is further configured to determine the position of the reference pixel to be selected corresponding to the left side of the reference point according to the first sampling interval using the midpoint of the at least one side as the reference point, The position of the reference pixel point to be selected corresponding to the right side of the reference point is determined according to the second sampling interval.
  • the determining unit 1502 is further configured to use the midpoint of the at least one side as a reference point, and determine the position of consecutive reference pixels near the reference point according to the preset number of samples It is the position of the reference pixel to be selected, and the reference point is located in the middle of the position of the reference pixel to be selected.
  • the block to be decoded includes a square decoded block or a non-square decoded block.
  • the obtaining unit 1501 is further configured to delete reference pixel points corresponding to the start position and the end position of the at least one edge, respectively, to obtain a second set of adjacent reference pixels;
  • the determining unit 1502 is further configured to determine the position of the reference pixel point to be selected according to a preset number of samples based on the second neighboring reference pixel set based on the midpoint of the at least one side as a reference point.
  • the decoding prediction device 150 further includes a building unit 1507 configured to determine model parameters based on the neighboring reference pixel subsets; and build a prediction model according to the model parameters; wherein, The prediction model is used to characterize the prediction relationship between the first image component and the second image component corresponding to each pixel in the block to be decoded;
  • the decoding unit 1504 is specifically configured to perform predictive decoding on the block to be decoded based on the prediction model.
  • the “unit” may be a part of a circuit, a part of a processor, a part of a program or software, etc. Of course, it may also be a module or non-modular. Moreover, each component in this embodiment may be integrated into one processing unit, or each unit may exist alone physically, or two or more units may be integrated into one unit. The above integrated unit may be implemented in the form of hardware or software function modules.
  • the integrated unit is implemented in the form of a software function module and is not sold or used as an independent product, it may be stored in a computer-readable storage medium.
  • the technical solution of this embodiment essentially or It is said that part of the contribution to the existing technology or all or part of the technical solution can be embodied in the form of a software product.
  • the computer software product is stored in a storage medium and includes several instructions to make a computer device (may It is a personal computer, a server, or a network device, etc.) or a processor (processor) that performs all or part of the steps of the method described in this embodiment.
  • the foregoing storage media include various media that can store program codes, such as a USB flash drive, a mobile hard disk, a read-only memory (Read Only Memory, ROM), a random access memory (Random Access Memory, RAM), a magnetic disk, or an optical disk.
  • program codes such as a USB flash drive, a mobile hard disk, a read-only memory (Read Only Memory, ROM), a random access memory (Random Access Memory, RAM), a magnetic disk, or an optical disk.
  • this embodiment provides a computer storage medium that stores a decoding prediction program, and when the decoding prediction program is executed by at least one processor, the steps of the method described in the foregoing technical solution shown in FIG. 8 are implemented.
  • FIG. 16 shows a specific hardware structure example of the decoding prediction device 150 provided by the embodiment of the present application, which may include: a network interface 1601, a memory 1602, and a processor 1603 ;
  • the various components are coupled together via the bus system 1604.
  • the bus system 1604 is used to implement connection and communication between these components.
  • the bus system 1604 also includes a power bus, a control bus, and a status signal bus.
  • various buses are marked as the bus system 1604 in FIG. 16.
  • the network interface 1601 is used to receive and send signals in the process of sending and receiving information with other external network elements;
  • the memory 1602 is used to store a computer program that can run on the processor 1603;
  • the processor 1603 is configured to execute:
  • a reference pixel corresponding to the position of the reference pixel to be selected is selected from the first set of adjacent reference pixels, and the selected reference pixels are combined to form a neighboring reference pixel set;
  • the memory 1602 in the embodiment of the present application may be a volatile memory or a nonvolatile memory, or may include both volatile and nonvolatile memory.
  • the non-volatile memory may be read-only memory (Read-Only Memory, ROM), programmable read-only memory (Programmable ROM, PROM), erasable programmable read-only memory (Erasable PROM, EPROM), electronically Erasable programmable read only memory (Electrically, EPROM, EEPROM) or flash memory.
  • the volatile memory may be a random access memory (Random Access Memory, RAM), which is used as an external cache.
  • RAM static random access memory
  • DRAM dynamic random access memory
  • DRAM synchronous dynamic random access memory
  • SDRAM double data rate synchronous dynamic random access memory
  • Double Data Rate SDRAM DDRSDRAM
  • enhanced SDRAM ESDRAM
  • Synchlink DRAM SLDRAM
  • Direct Rambus RAM Direct Rambus RAM
  • the processor 1603 may be an integrated circuit chip with signal processing capabilities. In the implementation process, each step of the above method may be completed by an integrated logic circuit of hardware in the processor 1603 or an instruction in the form of software.
  • the aforementioned processor 1603 may be a general-purpose processor, a digital signal processor (Digital Signal Processor, DSP), an application specific integrated circuit (Application Specific Integrated Circuit, ASIC), an existing programmable gate array (Field Programmable Gate Array, FPGA) or other Programmable logic devices, discrete gates or transistor logic devices, discrete hardware components.
  • DSP Digital Signal Processor
  • ASIC Application Specific Integrated Circuit
  • FPGA Field Programmable Gate Array
  • the methods, steps, and logical block diagrams disclosed in the embodiments of the present application may be implemented or executed.
  • the general-purpose processor may be a microprocessor or the processor may be any conventional processor or the like.
  • the steps of the method disclosed in conjunction with the embodiments of the present application may be directly embodied and executed by a hardware decoding processor, or may be executed and completed by a combination of hardware and software modules in the decoding processor.
  • the software module may be located in a mature storage medium in the art, such as random access memory, flash memory, read-only memory, programmable read-only memory, or electrically erasable programmable memory, and registers.
  • the storage medium is located in the memory 1602, and the processor 1603 reads the information in the memory 1602, and completes the steps of the above method in combination with its hardware.
  • the embodiments described herein may be implemented in hardware, software, firmware, middleware, microcode, or a combination thereof.
  • the processing unit can be implemented in one or more application specific integrated circuits (Application Specific Integrated Circuits, ASIC), digital signal processor (Digital Signal Processing, DSP), digital signal processing device (DSP Device, DSPD), programmable Logic device (Programmable Logic Device, PLD), field-programmable gate array (Field-Programmable Gate Array, FPGA), general-purpose processor, controller, microcontroller, microprocessor, others used to perform the functions described in this application Electronic unit or its combination.
  • ASIC Application Specific Integrated Circuits
  • DSP Digital Signal Processing
  • DSP Device digital signal processing device
  • DSPD digital signal processing device
  • PLD programmable Logic Device
  • FPGA field-programmable gate array
  • controller microcontroller, microprocessor, others used to perform the functions described in this application Electronic unit or its combination.
  • the techniques described herein may be implemented through modules (eg, procedures, functions, etc.) that perform the functions described herein.
  • the software codes can be stored in memory and executed by the processor.
  • the memory may be implemented in the processor or external to the processor.
  • the processor 1603 is further configured to execute the steps of the method in the foregoing technical solution shown in FIG. 8 when running the computer program.
  • reference pixel points adjacent to at least one side of the block to be decoded are obtained to obtain a first set of adjacent reference pixels; then a reference point is determined from the at least one side, according to a preset number of samples Determining the position of the reference pixel point to be selected corresponding to the at least one edge; and then selecting a reference corresponding to the position of the reference pixel point to be selected from the first neighboring reference pixel set based on the position of the reference pixel point to be selected Pixels, the selected reference pixels are used to form a subset of neighboring reference pixels; finally, based on the subset of neighboring reference pixels, predictive decoding of the block to be decoded; because of the neighboring reference in the subset of neighboring reference pixels
  • the selection of pixels takes into account both importance and dispersion.
  • the model parameters constructed based on the adjacent reference pixel subset are more accurate, which can improve the prediction performance of video image decoding; and the number of pixels in the adjacent reference pixel subset The number is small, which also reduces the complexity of the search, improves the prediction performance of video image decoding, and thus reduces the bit rate.

Landscapes

  • Engineering & Computer Science (AREA)
  • Multimedia (AREA)
  • Signal Processing (AREA)
  • Physics & Mathematics (AREA)
  • Algebra (AREA)
  • General Physics & Mathematics (AREA)
  • Mathematical Analysis (AREA)
  • Mathematical Optimization (AREA)
  • Pure & Applied Mathematics (AREA)
  • Compression Or Coding Systems Of Tv Signals (AREA)
  • Signal Processing For Digital Recording And Reproducing (AREA)
  • Compression Of Band Width Or Redundancy In Fax (AREA)

Abstract

本申请实施例公开了一种解码预测方法、装置及计算机存储介质,该方法包括:获取与待解码块至少一个边相邻的参考像素点,得到第一相邻参考像素集合;从所述至少一个边中确定一基准点,按照预设采样个数确定所述至少一个边对应的待选择参考像素点位置;基于所述待选择参考像素点位置,从所述第一相邻参考像素集合中选取与所述待选择参考像素点位置对应的参考像素点,将选取得到的参考像素点组成相邻参考像素子集;基于所述相邻参考像素子集,对所述待解码块进行预测解码。

Description

解码预测方法、装置及计算机存储介质 技术领域
本申请实施例涉及视频编解码的技术领域,尤其涉及一种解码预测方法、装置及计算机存储介质。
背景技术
随着人们对视频显示质量要求的提高,高清和超高清视频等新视频应用形式应运而生。H.265/高效率视频编码(High Efficiency Video Coding,HEVC)是目前最新的国际视频压缩标准,H.265/HEVC的压缩性能比前一代视频编码标准H.264/先进视频编码(Advanced Video Coding,AVC)提高约50%,但仍然满足不了视频应用迅速发展的需求,尤其是超高清、虚拟现实(Virtual Reality,VR)等新视频应用。
ITU-T的视频编码专家组和ISO/IEC的运动图像专家组于2015年成立联合视频研究组(Joint Video Exploration Team,JVET)着手制定下一代的视频编码标准。联合探索测试模型(Joint Exploration Test Model,JEM)为通用的参考软件平台,不同编码工具基于此平台验证。2018年4月,JVET正式命名下一代视频编码标准为多功能视频编码(Versatile Video Coding,VVC),其相应的测试模型为VTM。在JEM和VTM参考软件中,已经集成了一种基于线性模型的编解码预测方法,通过该线性模型由当前待解码块的亮度分量预测其色度分量。然而,在构建线性模型时,由相邻参考像素所构造的相邻参考像素子集不太合理,导致搜索复杂度较高,且降低了视频图像的解码预测性能。
发明内容
有鉴于此,本申请实施例期望提供一种解码预测方法、装置及计算机存储介质,由于相邻参考像素子集中的相邻参考像素不仅兼顾了重要性和分散性,而且相邻参考像素子集中的像素个数较少,从而降低了搜索的复杂度,提升了视频图像解码的预测性能,进而降低了比特率。
本申请实施例的技术方案可以如下实现:
第一方面,本申请实施例提供了一种解码预测方法,所述方法包括:
获取与待解码块至少一个边相邻的参考像素点,得到第一相邻参考像素集合;
从所述至少一个边中确定一基准点,按照预设采样个数确定所述至少一个边对应的待选择参考像素点位置;
基于所述待选择参考像素点位置,从所述第一相邻参考像素集合中选取与所述待选择参考像素点位置对应的参考像素点,将选取得到的参考像素点组成相邻参考像素子集;
基于所述相邻参考像素子集,对所述待解码块进行预测解码。
第二方面,本申请实施例提供了一种解码预测装置,所述解码预测装置包括获取单元、确定单元、选取单元和解码单元,
所述获取单元,配置为获取与待解码块至少一个边相邻的参考像素点,得到第一相邻参考像素集合;
所述确定单元,配置为从所述至少一个边中确定一基准点,按照预设采样个数确定所述至少一个边对应的待选择参考像素点位置;
所述选取单元,配置为基于所述待选择参考像素点位置,从所述第一相邻参考像素集 合中选取与所述待选择参考像素点位置对应的参考像素点,将选取得到的参考像素点组成相邻参考像素子集;
所述解码单元,配置为基于所述相邻参考像素子集,对所述待解码块进行预测解码。
第三方面,本申请实施例提供了一种解码预测装置,所述解码预测装置包括:存储器和处理器;
所述存储器,用于存储能够在所述处理器上运行的计算机程序;
所述处理器,用于在运行所述计算机程序时,执行如第一方面所述方法的步骤。
第四方面,本申请实施例提供了一种计算机存储介质,所述计算机存储介质存储有解码预测程序,所述解码预测程序被至少一个处理器执行时实现如第一方面所述方法的步骤。
本申请实施例提供了一种解码预测方法、装置及计算机存储介质,首先获取与待解码块至少一个边相邻的参考像素点,得到第一相邻参考像素集合;然后从所述至少一个边中确定一基准点,按照预设采样个数确定所述至少一个边对应的待选择参考像素点位置;再基于所述待选择参考像素点位置,从所述第一相邻参考像素集合中选取与所述待选择参考像素点位置对应的参考像素点,将选取得到的参考像素点组成相邻参考像素子集;最后基于所述相邻参考像素子集,对所述待解码块进行预测解码;由于该相邻参考像素子集中相邻参考像素点的选取兼顾了重要性和分散性,基于该相邻参考像素子集所构造的模型参数较为准确,可以使得视频图像解码的预测性能得到提升;而且该相邻参考像素子集中的像素个数较少,从而还降低了搜索的复杂度,提升了视频图像解码的预测性能,进而降低了比特率。
附图说明
图1A至图1C分别为相关技术方案中视频图像采样格式的结构示意图;
图2A和图2B分别为相关技术方案中解码块的第一图像分量相邻参考值和第二图像分量相邻参考值的采样示意图;
图3为相关技术方案中解码块基于最大值和最小值构造预测模型的结构示意图;
图4A和图4B分别为相关技术方案提供的一种正方形解码块和非正方形解码块选取相邻参考像素点的结构示意图;
图5A和图5B分别为相关技术方案提供的一种传统技术方案和L0138提案选取相邻参考像素点的结构示意图;
图6为本申请实施例提供的一种视频编码系统的组成框图示意图;
图7为本申请实施例提供的一种视频解码系统的组成框图示意图;
图8为本申请实施例提供的一种解码预测方法的流程示意图;
图9为本申请实施例提供的一种待解码块上侧边对应的相邻参考像素子集选取的结构示意图;
图10为本申请实施例提供的一种待解码块左侧边对应的相邻参考像素子集选取的结构示意图;
图11为本申请实施例提供的另一种待解码块上侧边对应的相邻参考像素子集选取的结构示意图;
图12为本申请实施例提供的又一种待解码块上侧边对应的相邻参考像素子集选取的结构示意图;
图13为本申请实施例提供的再一种待解码块上侧边对应的相邻参考像素子集选取的结构示意图;
图14为本申请实施例提供的再一种待解码块上侧边对应的相邻参考像素子集选取的结构示意图;
图15为本申请实施例提供的一种解码预测装置的组成结构示意图;
图16为本申请实施例提供的一种解码预测装置的具体硬件结构示意图。
具体实施方式
为了能够更加详尽地了解本申请实施例的特点与技术内容,下面结合附图对本申请实施例的实现进行详细阐述,所附附图仅供参考说明之用,并非用来限定本申请实施例。
在视频图像中,一般采用第一图像分量、第二图像分量和第三图像分量来表征解码块;其中,这三个图像分量分别为一个亮度分量、一个蓝色色度分量和一个红色色度分量,具体地,亮度分量通常使用符号Y表示,蓝色色度分量通常使用符号Cb表示,红色色度分量通常使用符号Cr表示。
在本申请实施例中,第一图像分量可以为亮度分量Y,第二图像分量可以为蓝色色度分量Cb,第三图像分量可以为红色色度分量Cr,但是本申请实施例对此不作具体限定。目前常用的采样格式为YCbCr格式,YCbCr格式包括以下几种,分别如图1A至图1C所示,其中,图中的叉(X)表示第一图像分量采样点,圈(○)表示第二图像分量或第三图像分量采样点。YCbCr格式包括:
4:4:4格式:如图1A所示,表示第二图像分量或第三图像分量没有下采样;它是在每个扫描行上每4个连续的像素点取4个第一图像分量的采样样本、4个第二图像分量的采样样本和4个第三图像分量的采样样本;
4:2:2格式:如图1B所示,表示第一图像分量相对于第二图像分量或第三图像分量进行2:1的水平采样,没有竖直下采样;它是在每个扫描行上每4个连续的像素点取4个第一图像分量的采样样本、2个第二图像分量的采样样本和2个第三图像分量的采样样本;
4:2:0格式:如图1C所示,表示第一图像分量相对于第二图像分量或第三图像分量进行2:1的水平下采样和2:1的竖直下采样;它是在水平扫描行和垂直扫描行上每2个连续的像素点取2个第一图像分量的采样样本、1个第二图像分量的采样样本和1个第三图像分量的采样样本。
在视频图像采用YCbCr为4:2:0格式的情况下,若视频图像的第一图像分量为2N×2N大小的解码块,则对应的第二图像分量或第三图像分量为N×N大小的解码块,其中N为解码块的边长。在本申请实施例中,下述将以4:2:0格式为例进行描述,但是本申请实施例的技术方案同样适用于其他采样格式。
在下一代视频编码标准H.266中,为了进一步提升了编解码性能,针对分量间预测(Cross-component Prediction,CCP)进行了扩展改进,提出了分量间线性模型预测(Cross-component Linear Model Prediction,CCLM)。在H.266中,CCLM实现了第一图像分量到第二图像分量、第一图像分量到第三图像分量以及第二图像分量与第三图像分量之间的预测,下述将以第一图像分量到第二图像分量的预测为例进行描述,但是本申请实施例的技术方案同样也可以适用于其他图像分量的预测。
可以理解地,为了减少第一图像分量与第二图像分量之间的冗余,在VTM中使用CCLM的预测模式,此时第一图像分量和第二图像分量为同一解码块的,该第二图像分量是基于同一解码块的第一图像分量重建值进行预测的,例如使用如式(1)所示的预测模型:
Pred C[i,j]=α·Rec L[i,j]+β                      (1)
其中,i,j表示解码块中像素点的位置坐标,i表示水平方向,j表示竖直方向,Pred C[i,j]表示解码块中位置坐标为[i,j]的像素点对应的第二图像分量预测值,Rec L[i,j]表示同一解码块中(经过下采样的)位置坐标为[i,j]的像素点对应的第一图像分量重建值,α和β是上述预测模型的模型参数。
模型参数α和β的构造有多种方式,可以是以最小二乘法为评价的回归构造方式,也可以是最大值与最小值的构造方式,甚至还可以是其他构造方式,本申请实施例对此不作具体限定。下面将分别以基于最小二乘法的回归构造方式和最大值与最小值的构造方式为例进行描述。
在VVC中,最初接受了一种基于最小二乘法的回归构造模型参数的推导方法,具体地,可以通过最小化解码块周围的第一图像分量相邻参考值和第二图像分量相邻参考值的回归误差进行推导得到,如下述式(2)所示的α和β:
Figure PCTCN2018123657-appb-000001
其中,L(n)表示经过下采样的左侧边和上侧边所对应的第一图像分量相邻参考值,C(n)表示左侧边和上侧边所对应的第二图像分量相邻参考值,N为第二图像分量解码块的边长,n=1,2,...,2N。参见图2A和图2B,其分别示出了相关技术方案中解码块的第一图像分量相邻参考值和第二图像分量相邻参考值的采样示意图;其中,在图2A中,加粗的较大方框用于突出指示第一图像分量解码块21,而灰色实心圆圈用于指示第一图像分量解码块21的相邻参考值L(n);在图2B中,加粗的较大方框用于突出指示第二图像分量解码块22,而灰色实心圆圈用于指示第二图像分量解码块22的相邻参考值C(n)。图2A示出了2N×2N大小的第一图像分量解码块21,对于4:2:0格式的视频图像来说,一个2N×2N大小的第一图像分量相对应的第二图像分量的大小为N×N,如图2B中的22所示;也就是说,图2A和图2B是针对同一解码块分别进行第一图像分量采样和第二图像分量采样所得到的解码块示意图。
在VVC中,近期接受了一种简化版模型参数的推导方法,具体地,可以通过搜索最大的第一图像分量相邻参考值和最小的第一图像分量相邻参考值,根据“两点确定一线”原则来推导出模型参数,如下述式(3)所示的α和β:
Figure PCTCN2018123657-appb-000002
其中,L max和L min表示经过下采样的左侧边和上侧边所对应的第一图像分量相邻参考值中搜索得到的最大值和最小值,C max和C min表示L max和L min对应位置的参考像素点所对应的第二图像分量相邻参考值。参见图3,其示出了相关技术方案中解码块基于最大值和最小值构造预测模型的结构示意图;其中,横坐标表示解码块的第一图像分量相邻参考值,纵坐标表示解码块的第二图像分量相邻参考值,根据L max和L min以及C max和C min,通过式(3)可以计算得到模型参数α和β,所构建的预测模型为C=α·L+β;这里的L表示解码块中其中一个像素点所对应的第一图像分量重建值,C表示解码块中该像素点所对应的第二图像分量预测值。
对于CCLM中相邻参考像素点集合的构造,现有技术中存在多种情况,下面将分别进行描述。
(a)从解码块的形状进行区分
参见图4A,其示出了相关技术方案提供的一种正方形解码块选取相邻参考像素点的结构示意图;如图4A所示,该解码块为正方形解码块,该解码块的左侧边和上侧边所对应的所有相邻像素点都可以作为参考像素点;针对第一图像分量首先需要进行下采样,这样经过下采样的第一图像分量与第二图像分量具有相同的分辨率。在图4A中,灰色实心圆 圈用于表示正方形解码块所选取的相邻参考像素点。
参见图4B,其示出了相关技术方案提供的一种非正方形解码块选取相邻参考像素点的结构示意图;如图4B所示,该解码块为非正方形解码块,该解码块的宽和高不相等。一方面,针对第一图像分量首先需要进行下采样,这样经过下采样的第一图像分量与第二图像分量具有相同的分辨率;另一方面,该解码块的长边所对应的相邻像素点需要进一步进行下采样,使得长边得到的相邻参考像素点个数与短边的相邻参考像素点个数相等。在图4B中,灰色实心圆圈用于表示非正方形解码块所选取的相邻参考像素点。
(b)从解码块左侧边或者上侧边所对应的相邻像素点的存在性进行区分
当解码块左侧边和上侧边所对应的相邻像素点都可以使用时,与上侧边相邻一行的像素点和与左侧边相邻一列的相邻像素点都可以作为相邻参考像素点;
当解码块左侧边或者上侧边所对应的相邻像素点仅一边可以使用时,将有效边所对应的相邻像素点作为相邻参考像素点;
当解码块左侧边和上侧边所对应的相邻像素点都不可以使用时,此时无相邻参考像素点,模型参数α设置为0,模型参数β设置为第二图像分量的中间值512,即解码块内所有像素点所对应的第二图像分量预测值均为512。
需要说明的是,假定N定义为第二图像分量解码块短边的长度,那么当解码块左侧边和上侧边所对应的相邻像素点都可以使用时,则总共有2N个相邻参考像素点可以用于CCLM中。另外,除特殊说明外,下文中均是以解码块左侧边和上侧边所对应的相邻像素点都可以使用的情况进行描述。
(c)相邻参考像素子集(减点)的方案
在计算CCLM的模型参数过程中,所需要的运算复杂度与用于CCLM的相邻参考像素点的个数成正比。这样,为了降低运算复杂度,JVET第12次会议的L0138提案提出了一种基于待解码块对应的第二图像分量块的大小减少用于CCLM的相邻参考像素点个数的技术方案。参见表1,其示出了相关技术方案提供的一种待解码块对应的第二图像分量块的大小与相邻参考像素点个数之间的关系表。在表1中,N 1为传统技术方案中用于CCLM的相邻参考像素点个数,N 2为L0138提案提供的减点后用于CCLM的相邻参考像素点个数。具体的,L0138提案中的减点方法是在传统技术方案中获取到相邻参考像素点之后对其进行下采样而得到的。
表1
第二图像分量块的大小 N 1 N 2
2×2 4 2
2×n/n×2(n>2) 4 4
4×4/4×n/n×4(n>4) 8 8
8×8/8×n/n×8(n>8) 16 8
16×16/16×n/n×16(n>16) 32 8
32×32 64 8
(d)相邻参考像素点的下采样方案
对于非正方形解码块,VTM中已经提供了针对长边的下采样方案(如图4B所示)。而如上面所述的L0138提案,该提案中提出了进一步减点的下采样集合。
在L0138提案中,下采样过程中相邻参考像素点的选取会对解码预测性能有所影响,按照VTM中默认的下采样方案所得到的相邻参考像素点子集,其对应的解码预测性能并不理想,因而L0138提案提出了可以在下采样时采用另外的相邻参考像素点子集的选取方案。具体地,VTM中所使用的传统技术方案中,对非正方形解码块的长边进行下采样时,对于解码块上侧边对应的相邻一行的像素点是从最左边缘开始选点,对于左侧边对应的相邻一列的像素点是从最上边缘开始选点;而在L0138提案中提出了从与传统技术方案的相 反边缘开始选点的方案,具体可以参考图5A和图5B所示的选点结构示例。
参见图5A,其示出了相关技术方案提供的一种传统技术方案选取相邻参考像素点的结构示意图;如图5A所示,对于该解码块长边对应的相邻一行的像素点从最左边缘的像素点开始取点采样的;参见图5B,其示出了相关技术方案提供的一种L0138提案选取相邻参考像素点的结构示意图;如图5B所示,对于该解码块长边对应的相邻一行的像素点从最右边缘的像素点开始取点采样的;这里的采样间隔与传统技术方案相同,本申请实施例不再赘述。利用L0138提案的取点方案,可以在CCLM的模型参数计算中比较全面地覆盖到所有相邻参考点像素的取值范围,但是由于没有考虑到每个相邻参考像素点的重要性,在有限个数的前提下,仍然不能最大限度代表长边的整体特性。
在相关技术方案中,一方面从运算复杂度考虑,在VTM中虽然接受了从2N个点中搜索最大的第一图像分量相邻参考值和最小的第一图像分量相邻参考值,根据“两点确定一线”原则来推导出模型参数(参见式(3))。该方案仅利用了两个相邻参考像素点,相比于基于最小二乘法回归构造模型参数的运算复杂度大幅度减少;但是该方案仍然有相当大的复杂度,主要是因为需要在相邻参考像素点集合内确定出最大值和最小值,从而需要进行4N次的比较运算操作来确定出最大值和最小值。而且解码块的长度越大,其用于CCLM的相邻参考像素点个数就越大,导致用于确定出最大值和最小值的搜索次数也就越大,所以,该方案所带来的运算复杂度仍然比较大。另一方面从预测准确度考虑,由于在搜索最大值和最小值的过程中,若搜索得到的像素点与当前解码块的相关性很低,可以确认出该像素点为坏点。这样,如果搜索得到的这两个像素点中包含有坏点,就会对预测模型造成较大的模型偏差。因此,基于最大值和最小值构造模型参数的方法虽然复杂度较低,但其鲁棒性不好,使得解码预测性能有所下降。
基于此,当前最新的L0138提案中提出了相邻参考像素子集的概念,虽然针对上述不足有所改进,不仅降低了相邻参考像素点个数,使得运算复杂度进一步降低;而且还通过选择合适位置的像素点(如图5B所示),解码预测性能略有提高,但是该方案仍然有提升空间。
为了更好地提升编解码预测的性能,本申请实施例提供了一种解码预测方法,通过以所述待解码块至少一个边的中点为基准点,按照预设采样个数确定出待选择参考像素点位置,以将兼顾重要性和分散性所选取的相邻参考像素点放入到相邻参考像素子集中;基于所述相邻参考像素子集,对所述待解码块进行预测解码;这样,由于基于该相邻参考像素子集所构造的模型参数较为准确,可以使得解码预测性能得到提升;而且该相邻参考像素子集中的像素个数较少,从而还降低了搜索的复杂度,进而降低了比特率。下面将结合附图对本申请实施例进行详细描述。
参见图6,其示出了本申请实施例提供的一种视频编码系统的组成框图示例;如图6所示,该视频编码系统600包括变换与量化601、帧内估计602、帧内预测603、运动补偿604、运动估计605、反变换与反量化606、滤波器控制分析607、去方块滤波及样本自适应缩进(Sample Adaptive 0ffset,SAO)滤波608、头信息编码及基于上下文的自适应二进制算术编码(Context-based Adaptive Binary Arithmatic Coding,CABAC)609和解码图像缓存610等部件;针对输入的原始视频信号,通过编码树块(Coding Tree Unit,CTU)的划分可以得到一个视频编码块,然后对经过帧内或帧间预测后得到的残差像素信息通过变换与量化601对该视频编码块进行变换,包括将残差信息从像素域变换到变换域,并对所得的变换系数进行量化,用以进一步减少比特率;帧内估计602和帧内预测603是用于对该视频编码块进行帧内预测;明确地说,帧内估计602和帧内预测603用于确定待用以编码该视频编码块的帧内预测模式;运动补偿604和运动估计605用于执行所接收的视频编码块相对于一或多个参考帧中的一或多个块的帧间预测编码以提供时间预测信息;由运动估计605执行的运动估计为产生运动向量的过程,所述运动向量可以估计该视频编码块的 运动,然后由运动补偿604基于由运动估计605所确定的运动向量执行运动补偿;在确定帧内预测模式之后,帧内预测603还用于将所选择的帧内预测数据提供到头信息编码及CABAC 609,而且运动估计605将所计算确定的运动向量数据也发送到头信息编码及CABAC 609;此外,反变换与反量化606是用于该视频编码块的重构建,在像素域中重构建残差块,该重构建残差块通过滤波器控制分析607和去方块滤波及SAO滤波608去除方块效应伪影,然后将该重构残差块添加到解码图像缓存610的帧中的一个预测性块,用以产生经重构建的视频编码块;头信息编码及CABAC 609是用于编码各种编码参数及量化后的变换系数,在基于CABAC的编码算法中,上下文内容可基于相邻编码块,可用于编码指示所确定的帧内预测模式的信息,输出该视频信号的码流;而解码图像缓存610是用于存放重构建的视频编码块,用于预测参考。随着视频图像编码的进行,会不断生成新的重构建的视频编码块,这些重构建的视频编码块都会被存放在解码图像缓存610中。
参见图7,其示出了本申请实施例提供的一种视频解码系统的组成框图示例;如图7所示,该视频解码系统700包括头信息解码及CABAC解码701、反变换与反量化702、帧内预测703、运动补偿704、去方块滤波及SAO滤波705和解码图像缓存706等部件;输入的视频信号经过图6的编码处理之后,输出该视频信号的码流;该码流输入视频解码系统700中,首先经过头信息解码及CABAC解码701,用于得到解码后的变换系数;针对该变换系数通过反变换与反量化702进行处理,以便在像素域中产生残差块;帧内预测703可用于基于所确定的帧内预测模式和来自当前帧或图片的先前经解码块的数据而产生当前视频解码块的预测数据;运动补偿704是通过剖析运动向量和其他关联语法元素来确定用于视频解码块的预测信息,并使用该预测信息以产生正被解码的视频解码块的预测性块;通过对来自反变换与反量化702的残差块与由帧内预测703或运动补偿704产生的对应预测性块进行求和,而形成解码的视频块;该解码的视频信号通过去方块滤波及SAO滤波705以便去除方块效应伪影,可以改善视频质量;然后将经解码的视频块存储于解码图像缓存706中,解码图像缓存706存储用于后续帧内预测或运动补偿的参考图像,同时也用于视频信号的输出,即得到了所恢复的原始视频信号。
本申请实施例主要应用在如图6所示的帧内预测603部分和如图7所示的帧内预测703部分;也就是说,本申请实施例既可以应用于编码系统,也可以应用于解码系统,但是本申请实施例对此不作具体限定。
基于上述图6或者图7的应用场景示例,参见图8,其示出了本申请实施例提供的一种解码预测方法的流程示意图,该方法可以包括:
S801:获取与待解码块至少一个边相邻的参考像素点,得到第一相邻参考像素集合;
S802:从所述至少一个边中确定一基准点,按照预设采样个数确定所述至少一个边对应的待选择参考像素点位置;
S803:基于所述待选择参考像素点位置,从所述第一相邻参考像素集合中选取与所述待选择参考像素点位置对应的参考像素点,将选取得到的参考像素点组成相邻参考像素子集;
S804:基于所述相邻参考像素子集,对所述待解码块进行预测解码。
需要说明的是,待解码块为当前待进行第二图像分量预测或者第三图像分量预测的解码块。待解码块至少一个边可以是指待解码块的上侧边,也可以是指待解码块的左侧边,甚至还可以是指待解码块的上侧边和左侧边,本申请实施例不作具体限定。
还需要说明的是,基准点可以是所述至少一个边的中点,也可以是所述至少一个边的中点偏左的第一个参考像素点位置,还可以是所述至少一个边的中点偏右的第一参考像素点位置,甚至还可以是所述至少一个边的其他参考像素点位置,本申请实施例不作具体限定。
在本申请实施例中,本申请实施例的解码预测方法同样可以适用于编码系统中,通过 在编码系统中构造相邻参考像素子集,可以提升视频图像的编码预测性能,并且提高了编码压缩效率,以节省编码码率。下面将仅以解码系统中构造相邻参考像素子集为例进行描述。
在本申请实施例中,首先获取与待解码块至少一个边相邻的参考像素点,得到第一相邻参考像素集合;然后从所述至少一个边中确定一基准点,按照预设采样个数确定所述至少一个边对应的待选择参考像素点位置;再基于所述待选择参考像素点位置,从所述第一相邻参考像素集合中选取与所述待选择参考像素点位置对应的参考像素点,将选取得到的参考像素点组成相邻参考像素子集;最后基于所述相邻参考像素子集,对所述待解码块进行预测解码;由于本申请实施例中不再是将与待解码块的上侧边或者左侧边相邻的所有参考像素点全部参与解码预测的搜索运算,而是在兼顾重要性和分散性的前提下,通过选取合适位置的相邻参考像素来组成相邻参考像素子集,这样该相邻参考像素子集中的像素个数较少,不仅能够降低搜索的复杂度,而且还可以提升解码预测性能,进而降低了比特率。
在一些实施例中,所述从所述至少一个边中确定一基准点,包括:
基于所述至少一个边的长度,确定所述至少一个边的中点,将所述至少一个边的中点作为所述基准点。
在一些实施例中,当所述基准点为所述至少一个边的中点时,所述从所述至少一个边中确定一基准点,包括:
基于所述至少一个边的长度,若所述至少一个边的中点位于两个参考像素点之间的中间位置,则将所述中间位置偏右的第一个参考像素点位置作为所述至少一个边的基准点,或者将所述中间位置偏左的第一个参考像素点位置作为所述至少一个边的基准点。
需要说明的是,考虑到与待解码块至少一个边相邻的参考像素点的重要性与其对应的位置有关,为了使得相邻参考像素子集中的参考像素点能够代表整个相邻边的特性,需要尽可能选取处于该边中心位置的参考像素点,以将重要性比较低的点(比如该边的两侧边缘的参考像素点)进行剔除。在本申请实施例中,如果以待解码块上侧边为例进行说明,那么可以将中间位置偏右或者偏左的第一个参考像素点位置作为该边的基准点。
示例性地,参见图9,其示出了本申请实施例提供的一种待解码块上侧边对应的相邻参考像素子集选取的结构示意图。如图9所示,对于待解码块的上侧边上分布的所有参考像素点,选取该边的中点作为中心(如图9中所示的虚线位置),以该中心作为基准点进行参考像素点的选取。假定该解码块上侧边的边长为16,预设采样个数为4,此时可以得到采样间隔Δ为16/4=4。这样,由于上侧边的长度为16,可以确定出中点在7和8之间,也就是说,中点可以选取7,也可以选取8;在图9中,以选取8作为基准点为例,由于预设采样个数为4,那么可以确定出待选择参考像素点位置(如图9中灰色点示例)为2、6、10和14;根据这些参考像素点位置可以选取出与之对应的参考像素点,以此组成相邻参考像素子集。
在一些实施例中,当所述基准点为所述至少一个边的中点时,所述从所述至少一个边中确定一基准点,包括:
基于所述至少一个边的长度,若所述至少一个边的中点位于两个参考像素点之间的中间位置,则将所述中间位置偏下的第一个参考像素点位置作为所述至少一个边的基准点,或者将所述中间位置偏上的第一个参考像素点位置作为所述至少一个边的基准点。
需要说明的是,考虑到与待解码块至少一个边相邻的参考像素点的重要性与其对应的位置有关,为了使得相邻参考像素子集中的参考像素点能够代表整个相邻边的特性,需要尽可能选取处于该边中心位置的参考像素点,以将重要性比较低的点(比如该边的两侧边缘的参考像素点)进行剔除。在本申请实施例中,如果以待解码块左侧边为例进行说明,那么可以将中间位置偏下或者偏上的第一个参考像素点位置作为该边的基准点。
示例性地,参见图10,其示出了本申请实施例提供的一种待解码块左侧边对应的相邻 参考像素子集选取的结构示意图。如图10所示,对于待解码块的左侧边上分布的所有参考像素点,选取该边的中点作为中心(如图10中所示的虚线位置),以该中心作为基准点进行参考像素点的选取。假定该解码块上侧边的边长为8,预设采样个数为2,此时可以得到采样间隔Δ为8/2=4。这样,由于左侧边的长度为8,可以确定出中点在3和4之间,也就是说,中点可以选取3,也可以选取4;在图10中,以选取4作为基准点为例,由于预设采样个数为2,那么可以确定出待选择参考像素点位置(如图10中灰色点示例)为2和6;根据这些参考像素点位置可以选取出与之对应的参考像素点,以此组成相邻参考像素子集。
在实际应用中,由于待解码块左侧边或者上侧边的边长都是2的整数倍,那么待解码块左侧边或者上侧边的中间位置均是处于两个点之间。在图9的示例中,是以中间位置偏右的第一个像素点作为该边的中点;但是本申请实施例也可以将中间位置偏左的第一个像素点作为该边的中点,如图11所示的结构示例。在图11中,以将中间位置偏左的第一个像素点(如图11中的3)作为该边的中点,由于预设采样个数为2,那么可以确定出待选择参考像素点位置(如图11中灰色点示例)为1和5,根据这些参考像素点位置也可以选取出与之对应的参考像素点,以此组成相邻参考像素子集。因此,在本申请实施例中,针对待解码块的上侧边,既可以将中间位置偏右的第一个像素点作为该边的中点,也可以将中间位置偏左的第一个像素点作为该边的中点,本申请实施例不作具体限定;另外,针对待解码块的左侧边,既可以将中间位置偏下的第一个像素点作为该边的中点,也可以将中间位置偏上的第一个像素点作为该边的中点,本申请实施例也不作具体限定。
除特别说明外,下文中将以待解码块的上侧边为例进行说明,但是本申请实施例的预测方法同样适用于待解码块的左侧边,甚至是待解码块的其他边。
可以理解地,如果不考虑与待解码块左侧边或者上侧边相邻的参考像素点的存在性,那么还可以根据式(4)和式(5)进行相邻参考像素子集的构造,
△=length/(N 2/2)                              (4)
shift=△/2                                   (5)
其中,Δ表示采样间隔,length表示与待解码块上侧边相邻的一行参考像素点的个数、或者与待解码块左侧边相邻的一列参考像素点的个数,N 2表示待解码块所期望的构成相邻参考像素子集中相邻参考像素点个数(一般来说,左侧边和上侧边各自为二分之一,但是本申请实施例不作具体限定),shift表示选取参考像素点的起始点位置。这里,当待解码块左侧边或者上侧边的中间位置均是处于两个点之间时,此时如果以中间位置偏右的第一个像素点作为该边的中点,那么起始点位置shift=△/2;此时如果以中间位置偏左的第一个像素点作为该边的中点,那么起始点位置shift=△/2-1。
示例性地,以图9所示的上侧边为例,length等于16,N 2等于8,假定左侧边和上侧边各自为二分之一,即上侧边的预设采样个数为4,那么根据式(4)和式(5)分别计算得到△=length/(N 2/2)=4,shift=△/2=2,即以2为起始点位置、4为采样间隔,首先可以确定出待选择参考像素点位置,比如为2、6、10和14,进而可以选取到对应的参考像素点,以组成相邻参考像素子集。这里,需要注意的是,左侧边对应的预设采样个数和上侧边对应的预设采样个数,两者的取值可以相同,也可以不同,本申请实施例不作具体限定。
另外,本申请实施例还提供了一种确定预设采样个数的方案,如表2所示。这样,本申请实施例还可以用表2中的N 2‘替代N 2,将其带入上述的式(4)和式(5)中进行计算,以使得所组成的相邻参考像素子集更为准确。从而提高了解码预测性能。
参见表2,其示出了本申请实施例提供的一种待解码块对应的第二图像分量块的大小与相邻参考像素点个数之间的关系表。在表2中,N表示待解码块短边的长度,N 1表示传统技术方案中提供的相邻参考像素点个数,N 2为L0138提案中提供的相邻参考像素点个数, N 2‘表示本本申请实施例中提供的相邻参考像素点个数。根据表2可以看出,当待解码块短边的长度小于或等于4时,相邻参考像素子集中包含4个参考像素点;当待解码块短边的长度大于4时,相邻参考像素子集中可以包含8个参考像素点。
表2
第二图像分量块的大小 短边的长度 N 1 N 2 N 2
2×n/n×2(n≥2) 2 4 2 4
4×n/n×4(n≥4) 4 8 4 4
8×n/n×8(n≥8) 8 16 8 8
16×n/n×16(n≥16) 16 32 8 8
32×32 32 64 8 8
在一些实施例中,所述从所述至少一个边中确定一基准点,按照预设采样个数确定所述至少一个边对应的待选择参考像素点位置,包括:
基于所述预设采样个数和所述至少一个边的长度,计算得到第一采样间隔;
以所述至少一个边的中点为基准点,按照所述第一采样间隔确定所述至少一个边对应的待选择参考像素点位置。
可选地,所述以所述至少一个边的中点为基准点,按照所述第一采样间隔确定所述至少一个边对应的待选择参考像素点位置,包括:
基于所述至少一个边的长度,计算所述至少一个边的中点值;
根据所述中点值和第一采样间隔,计算得到参考像素点位置;
当所述中点值为非整数时,对所述中点值左侧的参考像素点位置向小取整,将取整后的参考像素点位置确定为待选择参考像素点位置;对所述中点值右侧的参考像素点位置向大取整,将取整后的参考像素点位置确定为待选择参考像素点位置。
可选地,所述以所述至少一个边的中点为基准点,按照所述第一采样间隔确定所述至少一个边对应的待选择参考像素点位置,包括:
基于所述至少一个边的长度,计算所述至少一个边的中点值;
根据所述中点值和第一采样间隔,计算得到参考像素点位置;
当所述中点值为非整数时,对所述中点值左侧的参考像素点位置向大取整,将取整后的参考像素点位置确定为待选择参考像素点位置;对所述中点值右侧的参考像素点位置向小取整,将取整后的参考像素点位置确定为待选择参考像素点位置。
需要说明的是,根据预设采样个数和待解码块一个边的长度,可以计算得到该边对应的第一采样间隔。另外,由于待解码块左侧边或者上侧边的边长都是2的整数倍,那么待解码块左侧边或者上侧边的中间位置均是处于两个点之间,此时计算得到的中点值为非整数,计算得到的参考像素点位置也为非整数;然而如果待解码块左侧边或者上侧边的边长不是2的整数倍,那么待解码块左侧边或者上侧边的中间位置不会处于两个点之间,此时计算得到的中点值为整数,计算得到的参考像素点位置也为整数;也就是说,计算得到的中点值可以为整数,也可以为非整数;对应地,计算得到的参考像素点位置也可以为整数,也可以为非整数;本申请实施例不作具体限定。
这样,当计算得到的中点值为整数时,对应地,计算得到的参考像素点位置也为整数,此时可以直接将计算得到的参考像素点位置作为待选择参考像素点位置;当计算得到的中点值为非整数时,对应地,计算得到的参考像素点位置也为非整数,此时可以通过向大取整或者向小取整来确定出待选择参考像素点位置。
举例来说,以图11所示的上侧边为例,假定预设采样个数为2个,上侧边的长度为8,那么可以确定出第一采样间隔为4。由于上侧边的长度为8,也就是说,上侧边的中间位置处于两个点之间。由于参考像素点排列是以0开始,结束为7,通过计算可以得到上侧边的中点实际位置为3.5;由于第一采样间隔为4,在中点的左右两侧分别移位4/2个点, 则可以计算得到参考点位置为1.5和5.5;此时可以将中点值左侧的参考像素点位置向小取整,能够得到其中一个待选择参考像素点位置为1,而将中点值右侧的参考像素点位置向大取整,则可以得到另一个待选择参考像素点位置为6;该方法也可以称之为向外取整方案,如图12所示;这样可以将1和6位置所对应的参考像素点组成相邻参考像素子集。另外,当计算得到参考点位置为1.5和5.5时,此时还可以将中点值左侧的参考像素点位置向大取整,能够得到其中一个待选择参考像素点位置为2,而将中点值右侧的参考像素点位置向小取整,则可以得到另一个待选择参考像素点位置为5;该方法也可以称之为向内取整方案,如图13所示;这样可以将2和5位置所对应的参考像素点组成相邻参考像素子集。
在一些实施例中,在所述计算得到第一采样间隔之后,所述方法还包括:
对所述第一采样间隔进行调整,得到第二采样间隔;
以所述至少一个边的中点为基准点,按照所述第二采样间隔确定所述至少一个边对应的待选择参考像素点位置。
在一些实施例中,在所述得到第二采样间隔之后,所述方法还包括:
以所述至少一个边的中点为基准点,按照所述第一采样间隔确定所述基准点左侧对应的待选择参考像素点位置,按照所述第二采样间隔确定所述基准点右侧对应的待选择参考像素点位置。
需要说明的是,当计算得到第一采样间隔之后,还可以对第一采样间隔进行微调,比如对第一采样间隔进行加1或者减1的操作,以此得到第二采样间隔。例如,第一采样间隔为4,那么调整后的第二采样间隔可以为3或者5。在本申请实施例中,针对第一采样间隔的调整,可以进行小幅度(比如,加1或者减1)的调整,但是针对调整幅度的具体设定,本申请实施例不作具体限定。
另外,当确定出待解码块至少一条边的基准点之后,可以按照第一采样间隔或者第二采样间隔进行均匀采样;也可以按照第一采样间隔和第二采样间隔进行分均匀采样,而且采样后所确定的待选择参考像素点位置可以对称分布在基准点两侧,也可以非对称分布在基准点两侧;本申请实施例不作具体限定。
在一些实施例中,所述从所述至少一个边中确定一基准点,按照预设采样个数确定所述至少一个边对应的待选择参考像素点位置,包括:
以所述至少一个边的中点为基准点,根据所述预设采样个数将处于所述基准点附近的连续的参考像素点位置确定为待选择参考像素点位置,且所述基准点位于所述待选择参考像素点位置的中间位置。
需要说明的是,由于相邻参考像素点中与待解码块的第一图像分量重建值比较相关的是处于中间位置的参考像素点,那么可以将中间位置附近连续的预设采样个数的参考像素点位置作为待选择参考像素点位置,该方法可以称之为中间位置连续取点方案,如图14所示;这样可以将2、3和4位置所对应的参考像素点组成相邻参考像素子集。
可以理解地,假定与待解码块上侧边或者左侧边相邻的一行/一列上的参考像素点位置从0开始编号,那么本实施例中所组成的相邻参考像素子集中相邻参考像素点个数以及对应的待选择参考像素点位置如表3所示。
表3
Figure PCTCN2018123657-appb-000003
在本实施例中,将中间位置附近连续的预设采样个数的参考像素点位置作为待选择参考像素点位置,以此组成相邻参考像素子集;根据该相邻参考像素子集进行预测解码,在L0138提案的基础上仍然能够节省比特率,提高了解码增益,从而使得解码预测性能得到提升。
在一些实施例中,所述待解码块包括正方形解码块或者非正方形解码块。
进一步地,在一些实施例中,当所述待解码块为非正方形解码块时,所述方法还包括:
基于所述待解码块至少一个边的长度,确定所述待解码块的长边以及所述长边对应的第三采样间隔;
删除所述长边的末端位置对应的参考像素点,对删除参考像素点的长边按照预设偏移量进行初始偏移,以偏移后的参考像素点位置作为起始点,按照所述第三采样间隔对删除参考像素点的长边进行采样,以确定所述长边对应的待选择参考像素点位置。
需要说明的是,本申请实施例既可以应用于正方形解码块,也可以应用于非正方形解码块,本申请实施例不作具体限定。
还需要说明的是,当待解码块为非正方形解码块时,待解码块左侧边或者上侧边中必然存在一个长边和一个短边;根据长边与短边之间的比值,可以得到长边对应的第三采样间隔。在对长边进行采样之前,可以先删除所述长边的末端位置对应的参考像素点,然后再对删除参考像素点的长边按照预设偏移量进行初始偏移,以偏移后的参考像素点位置作为起始点,然后再对删除参考像素点的长边进行采样,以确定该长边对应的待选择参考像素点位置。在本申请实施例中,预设偏移量可以是第三采样间隔的二分之一,也可以是其他值,本申请实施例不作具体限定。
示例性地,在相关技术方案中,针对非正方形解码块的长边进行参考像素点采样如图4B所示。在图4B中,根据长边与短边之间的比值,可以确定出第三采样间隔为4;长边是从左侧第一个参考像素点位置作为起始点,然后按照第三采样间隔选取出与短边的相邻参考像素点个数相同的参考像素点。由于这样会使得所采样的参考像素点偏左,不能覆盖整个长边的特性;因此,在本实施例中,针对非正方形解码块的长边首先进行初始偏移,使得所采样的参考像素点能够覆盖整个长边的特性。比如预设偏移量为第三采样间隔的二分之一,即预设偏移量为2,也就是说,本实施例进行采样是以2为起始点的,这样所组成的相邻参考像素子集能够更好地覆盖整个长边的特性。
在一些实施例中,所述从所述至少一个边中确定一基准点,按照预设采样个数确定所述至少一个边对应的待选择参考像素点位置,包括:
删掉所述至少一个边的起始位置和末端位置分别对应的参考像素点,得到第二相邻参考像素集合;
以所述至少一个边的中点为基准点,基于所述第二相邻参考像素集合,按照预设采样个数确定待选择参考像素点位置。
需要说明的是,本申请实施例可以直接基于待解码块左侧边或者上侧边的中点为基准点进行参考像素点的选取,以组成相邻参考像素子集;本申请实施例还可以先删除待解码块左侧边/上侧边对应的起始位置和末端位置(比如,针对上侧边来说,起始位置为左侧边缘位置,末端位置为右侧边缘位置;针对左侧边来说,起始位置为上侧边缘位置,末端位置为下侧边缘位置)分别对应的参考像素点,保留了中间部分的参考像素点,然后再将所选取的参考像素点来组成相邻参考像素子集。
还需要说明的是,预设采样个数可以任意取值;一般来说,预设采样个数小于表2中所示的N 1即可;但是本申请实施例并不局限于预设采样个数等于N 2或N 2‘。由于预设采样个数小于表2中所示的N 1,这样就可以降低参考像素点的搜索复杂度,以提升解码预测性能。具体来说,可以通过改变上述式(4)和式(5)中的Δ和shift值,也就可以实现以不同的起始点和采样间隔来确定待选择参考像素点位置,进而可以选取到对应的参考像素点, 以组成相邻参考像素子集。
在一些实施例中,所述基于所述相邻参考像素子集,对所述待解码块进行预测解码,包括:
基于所述相邻参考像素子集,确定模型参数;
根据所述模型参数,建立预测模型;其中,所述预测模型用于表征所述待解码块中每个像素点对应的第一图像分量与第二图像分量之间的预测关系;
基于所述预测模型,对所述待解码块进行预测解码。
需要说明的是,当得到相邻参考像素子集之后,就可以构造出模型参数α和β;这样,根据上述式(1)就可以建立预测模型;根据预测模型,就可以对该待解码块进行预测解码。由于相邻参考像素子集中所包括的相邻参考像素点兼顾了重要性和分散性,使得所构造的模型参数更为准确,从而更好地提升了解码预测性能,进而降低了比特率。
上述实施例提供了一种解码预测方法,通过获取与待解码块至少一个边相邻的参考像素点,得到第一相邻参考像素集合;从所述至少一个边中确定一基准点,按照预设采样个数确定所述至少一个边对应的待选择参考像素点位置;基于所述待选择参考像素点位置,从所述第一相邻参考像素集合中选取与所述待选择参考像素点位置对应的参考像素点,将选取得到的参考像素点组成相邻参考像素子集;基于所述相邻参考像素子集,对所述待解码块进行预测解码;由于相邻参考像素子集中的相邻参考像素不仅兼顾了重要性和分散性,而且相邻参考像素子集中的像素个数较少,从而降低了搜索的复杂度,提升了视频图像解码的预测性能,进而降低了比特率。
基于前述图8所示技术方案相同的发明构思,参见图15,其示出了本申请实施例提供的一种解码预测装置150的组成结构示意图。该解码预测装置150可以包括:获取单元1501、确定单元1502、选取单元1503和解码单元1504,其中,
所述获取单元1501,配置为获取与待解码块至少一个边相邻的参考像素点,得到第一相邻参考像素集合;
所述确定单元1502,配置为从所述至少一个边中确定一基准点,按照预设采样个数确定所述至少一个边对应的待选择参考像素点位置;
所述选取单元1503,配置为基于所述待选择参考像素点位置,从所述第一相邻参考像素集合中选取与所述待选择参考像素点位置对应的参考像素点,将选取得到的参考像素点组成相邻参考像素子集;
所述解码单元1504,配置为基于所述相邻参考像素子集,对所述待解码块进行预测解码。
在上述方案中,所述确定单元1502,具体配置为基于所述至少一个边的长度,若所述至少一个边的中点位于两个参考像素点之间的中间位置,则将所述中间位置偏右的第一个参考像素点位置作为所述至少一个边的基准点,或者将所述中间位置偏左的第一个参考像素点位置作为所述至少一个边的基准点。
在上述方案中,所述确定单元1502,具体配置为基于所述至少一个边的长度,若所述至少一个边的中点位于两个参考像素点之间的中间位置,则将所述中间位置偏下的第一个参考像素点位置作为所述至少一个边的基准点,或者将所述中间位置偏上的第一个参考像素点位置作为所述至少一个边的基准点。
在上述方案中,参见图15,所述解码预测装置150还包括计算单元1505,配置为基于所述预设采样个数和所述至少一个边的长度,计算得到第一采样间隔;
所述确定单元1502,具体配置为以所述至少一个边的中点为基准点,按照所述第一采样间隔确定所述至少一个边对应的待选择参考像素点位置。
在上述方案中,所述计算单元1505,还配置为基于所述至少一个边的长度,计算所述至少一个边的中点值;以及根据所述中点值和第一采样间隔,计算得到参考像素点位置;
所述确定单元1502,具体配置为当所述中点值为非整数时,对所述中点值左侧的参考像素点位置向小取整,将取整后的参考像素点位置确定为待选择参考像素点位置;对所述中点值右侧的参考像素点位置向大取整,将取整后的参考像素点位置确定为待选择参考像素点位置。
在上述方案中,所述计算单元1505,还配置为基于所述至少一个边的长度,计算所述至少一个边的中点值;以及根据所述中点值和第一采样间隔,计算得到参考像素点位置;
所述确定单元1502,具体配置为当所述中点值为非整数时,对所述中点值左侧的参考像素点位置向大取整,将取整后的参考像素点位置确定为待选择参考像素点位置;对所述中点值右侧的参考像素点位置向小取整,将取整后的参考像素点位置确定为待选择参考像素点位置。
在上述方案中,参见图15,所述解码预测装置150还包括调整单元1506,配置为对所述第一采样间隔进行调整,得到第二采样间隔;
所述确定单元1502,还配置为以所述至少一个边的中点为基准点,按照所述第二采样间隔确定所述至少一个边对应的待选择参考像素点位置。
在上述方案中,所述确定单元1502,还配置为以所述至少一个边的中点为基准点,按照所述第一采样间隔确定所述基准点左侧对应的待选择参考像素点位置,按照所述第二采样间隔确定所述基准点右侧对应的待选择参考像素点位置。
在上述方案中,所述确定单元1502,还配置为以所述至少一个边的中点为基准点,根据所述预设采样个数将处于所述基准点附近的连续的参考像素点位置确定为待选择参考像素点位置,且所述基准点位于所述待选择参考像素点位置的中间位置。
在上述方案中,所述待解码块包括正方形解码块或者非正方形解码块。
在上述方案中,所述获取单元1501,还配置为删掉所述至少一个边的起始位置和末端位置分别对应的参考像素点,得到第二相邻参考像素集合;
所述确定单元1502,还配置为以所述至少一个边的中点为基准点,基于所述第二相邻参考像素集合,按照预设采样个数确定待选择参考像素点位置。
在上述方案中,参见图15,所述解码预测装置150还包括建立单元1507,配置为基于所述相邻参考像素子集,确定模型参数;以及根据所述模型参数,建立预测模型;其中,所述预测模型用于表征所述待解码块中每个像素点对应的第一图像分量与第二图像分量之间的预测关系;
所述解码单元1504,具体配置为基于所述预测模型,对所述待解码块进行预测解码。
可以理解地,在本实施例中,“单元”可以是部分电路、部分处理器、部分程序或软件等等,当然也可以是模块,还可以是非模块化的。而且在本实施例中的各组成部分可以集成在一个处理单元中,也可以是各个单元单独物理存在,也可以两个或两个以上单元集成在一个单元中。上述集成的单元既可以采用硬件的形式实现,也可以采用软件功能模块的形式实现。
所述集成的单元如果以软件功能模块的形式实现并非作为独立的产品进行销售或使用时,可以存储在一个计算机可读取存储介质中,基于这样的理解,本实施例的技术方案本质上或者说对现有技术做出贡献的部分或者该技术方案的全部或部分可以以软件产品的形式体现出来,该计算机软件产品存储在一个存储介质中,包括若干指令用以使得一台计算机设备(可以是个人计算机,服务器,或者网络设备等)或processor(处理器)执行本实施例所述方法的全部或部分步骤。而前述的存储介质包括:U盘、移动硬盘、只读存储器(Read Only Memory,ROM)、随机存取存储器(Random Access Memory,RAM)、磁碟或者光盘等各种可以存储程序代码的介质。
因此,本实施例提供了一种计算机存储介质,该计算机存储介质存储有解码预测程序,所述解码预测程序被至少一个处理器执行时实现前述图8所示技术方案中所述方法的步骤。
基于上述解码预测装置150的组成以及计算机存储介质,参见图16,其示出了本申请实施例提供的解码预测装置150的具体硬件结构示例,可以包括:网络接口1601、存储器1602和处理器1603;各个组件通过总线系统1604耦合在一起。可理解,总线系统1604用于实现这些组件之间的连接通信。总线系统1604除包括数据总线之外,还包括电源总线、控制总线和状态信号总线。但是为了清楚说明起见,在图16中将各种总线都标为总线系统1604。其中,网络接口1601,用于在与其他外部网元之间进行收发信息过程中,信号的接收和发送;
存储器1602,用于存储能够在处理器1603上运行的计算机程序;
处理器1603,用于在运行所述计算机程序时,执行:
获取与待解码块至少一个边相邻的参考像素点,得到第一相邻参考像素集合;
从所述至少一个边中确定一基准点,按照预设采样个数确定所述至少一个边对应的待选择参考像素点位置;
基于所述待选择参考像素点位置,从所述第一相邻参考像素集合中选取与所述待选择参考像素点位置对应的参考像素点,将选取得到的参考像素点组成相邻参考像素子集;
基于所述相邻参考像素子集,对所述待解码块进行预测解码。
可以理解,本申请实施例中的存储器1602可以是易失性存储器或非易失性存储器,或可包括易失性和非易失性存储器两者。其中,非易失性存储器可以是只读存储器(Read-Only Memory,ROM)、可编程只读存储器(Programmable ROM,PROM)、可擦除可编程只读存储器(Erasable PROM,EPROM)、电可擦除可编程只读存储器(Electrically EPROM,EEPROM)或闪存。易失性存储器可以是随机存取存储器(Random Access Memory,RAM),其用作外部高速缓存。通过示例性但不是限制性说明,许多形式的RAM可用,例如静态随机存取存储器(Static RAM,SRAM)、动态随机存取存储器(Dynamic RAM,DRAM)、同步动态随机存取存储器(Synchronous DRAM,SDRAM)、双倍数据速率同步动态随机存取存储器(Double Data Rate SDRAM,DDRSDRAM)、增强型同步动态随机存取存储器(Enhanced SDRAM,ESDRAM)、同步连接动态随机存取存储器(Synchlink DRAM,SLDRAM)和直接内存总线随机存取存储器(Direct Rambus RAM,DRRAM)。本文描述的系统和方法的存储器1602旨在包括但不限于这些和任意其它适合类型的存储器。
而处理器1603可能是一种集成电路芯片,具有信号的处理能力。在实现过程中,上述方法的各步骤可以通过处理器1603中的硬件的集成逻辑电路或者软件形式的指令完成。上述的处理器1603可以是通用处理器、数字信号处理器(Digital Signal Processor,DSP)、专用集成电路(Application Specific Integrated Circuit,ASIC)、现成可编程门阵列(Field Programmable Gate Array,FPGA)或者其他可编程逻辑器件、分立门或者晶体管逻辑器件、分立硬件组件。可以实现或者执行本申请实施例中的公开的各方法、步骤及逻辑框图。通用处理器可以是微处理器或者该处理器也可以是任何常规的处理器等。结合本申请实施例所公开的方法的步骤可以直接体现为硬件译码处理器执行完成,或者用译码处理器中的硬件及软件模块组合执行完成。软件模块可以位于随机存储器,闪存、只读存储器,可编程只读存储器或者电可擦写可编程存储器、寄存器等本领域成熟的存储介质中。该存储介质位于存储器1602,处理器1603读取存储器1602中的信息,结合其硬件完成上述方法的步骤。
可以理解的是,本文描述的这些实施例可以用硬件、软件、固件、中间件、微码或其组合来实现。对于硬件实现,处理单元可以实现在一个或多个专用集成电路(Application Specific Integrated Circuits,ASIC)、数字信号处理器(Digital Signal Processing,DSP)、数字信号处理设备(DSP Device,DSPD)、可编程逻辑设备(Programmable Logic Device,PLD)、现场可编程门阵列(Field-Programmable Gate Array,FPGA)、通用处理器、控制器、微控制器、微处理器、用于执行本申请所述功能的其它电子单元或其组合中。
对于软件实现,可通过执行本文所述功能的模块(例如过程、函数等)来实现本文所述的技术。软件代码可存储在存储器中并通过处理器执行。存储器可以在处理器中或在处理器外部实现。
可选地,作为另一个实施例,处理器1603还配置为在运行所述计算机程序时,执行前述图8所示技术方案中所述方法的步骤。
需要说明的是:本申请实施例所记载的技术方案之间,在不冲突的情况下,可以任意组合。
以上所述,仅为本申请的具体实施方式,但本申请的保护范围并不局限于此,任何熟悉本技术领域的技术人员在本申请揭露的技术范围内,可轻易想到变化或替换,都应涵盖在本申请的保护范围之内。因此,本申请的保护范围应以所述权利要求的保护范围为准。
工业实用性
本申请实施例中,首先获取与待解码块至少一个边相邻的参考像素点,得到第一相邻参考像素集合;然后从所述至少一个边中确定一基准点,按照预设采样个数确定所述至少一个边对应的待选择参考像素点位置;再基于所述待选择参考像素点位置,从所述第一相邻参考像素集合中选取与所述待选择参考像素点位置对应的参考像素点,将选取得到的参考像素点组成相邻参考像素子集;最后基于所述相邻参考像素子集,对所述待解码块进行预测解码;由于该相邻参考像素子集中相邻参考像素点的选取兼顾了重要性和分散性,基于该相邻参考像素子集所构造的模型参数较为准确,可以使得视频图像解码的预测性能得到提升;而且该相邻参考像素子集中的像素个数较少,从而还降低了搜索的复杂度,提升了视频图像解码的预测性能,进而降低了比特率。

Claims (26)

  1. 一种解码预测方法,其中,所述方法包括:
    获取与待解码块至少一个边相邻的参考像素点,得到第一相邻参考像素集合;
    从所述至少一个边中确定一基准点,按照预设采样个数确定所述至少一个边对应的待选择参考像素点位置;
    基于所述待选择参考像素点位置,从所述第一相邻参考像素集合中选取与所述待选择参考像素点位置对应的参考像素点,将选取得到的参考像素点组成相邻参考像素子集;
    基于所述相邻参考像素子集,对所述待解码块进行预测解码。
  2. 根据权利要求1所述的方法,其中,当所述基准点为所述至少一个边的中点时,所述从所述至少一个边中确定一基准点,包括:
    基于所述至少一个边的长度,若所述至少一个边的中点位于两个参考像素点之间的中间位置,则将所述中间位置偏右的第一个参考像素点位置作为所述至少一个边的基准点,或者将所述中间位置偏左的第一个参考像素点位置作为所述至少一个边的基准点。
  3. 根据权利要求1所述的方法,其中,当所述基准点为所述至少一个边的中点时,所述从所述至少一个边中确定一基准点,包括:
    基于所述至少一个边的长度,若所述至少一个边的中点位于两个参考像素点之间的中间位置,则将所述中间位置偏下的第一个参考像素点位置作为所述至少一个边的基准点,或者将所述中间位置偏上的第一个参考像素点位置作为所述至少一个边的基准点。
  4. 根据权利要求1所述的方法,其中,所述从所述至少一个边中确定一基准点,按照预设采样个数确定所述至少一个边对应的待选择参考像素点位置,包括:
    基于所述预设采样个数和所述至少一个边的长度,计算得到第一采样间隔;
    以所述至少一个边的中点为基准点,按照所述第一采样间隔确定所述至少一个边对应的待选择参考像素点位置。
  5. 根据权利要求4所述的方法,其中,所述以所述至少一个边的中点为基准点,按照所述第一采样间隔确定所述至少一个边对应的待选择参考像素点位置,包括:
    基于所述至少一个边的长度,计算所述至少一个边的中点值;
    根据所述中点值和第一采样间隔,计算得到参考像素点位置;
    当所述中点值为非整数时,对所述中点值左侧的参考像素点位置向小取整,将取整后的参考像素点位置确定为待选择参考像素点位置;对所述中点值右侧的参考像素点位置向大取整,将取整后的参考像素点位置确定为待选择参考像素点位置。
  6. 根据权利要求4所述的方法,其中,所述以所述至少一个边的中点为基准点,按照所述第一采样间隔确定所述至少一个边对应的待选择参考像素点位置,包括:
    基于所述至少一个边的长度,计算所述至少一个边的中点值;
    根据所述中点值和第一采样间隔,计算得到参考像素点位置;
    当所述中点值为非整数时,对所述中点值左侧的参考像素点位置向大取整,将取整后的参考像素点位置确定为待选择参考像素点位置;对所述中点值右侧的参考像素点位置向小取整,将取整后的参考像素点位置确定为待选择参考像素点位置。
  7. 根据权利要求4所述的方法,其中,在所述计算得到第一采样间隔之后,所述方法还包括:
    对所述第一采样间隔进行调整,得到第二采样间隔;
    以所述至少一个边的中点为基准点,按照所述第二采样间隔确定所述至少一个边对应的待选择参考像素点位置。
  8. 根据权利要求7所述的方法,其中,在所述得到第二采样间隔之后,所述方法还包括:
    以所述至少一个边的中点为基准点,按照所述第一采样间隔确定所述基准点左侧对应的待选择参考像素点位置,按照所述第二采样间隔确定所述基准点右侧对应的待选择参考像素点位置。
  9. 根据权利要求1所述的方法,其中,所述从所述至少一个边中确定一基准点,按照预设采样个数确定所述至少一个边对应的待选择参考像素点位置,包括:
    以所述至少一个边的中点为基准点,根据所述预设采样个数将处于所述基准点附近的连续的参考像素点位置确定为待选择参考像素点位置,且所述基准点位于所述待选择参考像素点位置的中间位置。
  10. 根据权利要求1至9任一项所述的方法,其中,所述待解码块包括正方形解码块或者非正方形解码块。
  11. 根据权利要求1所述的方法,其中,所述从所述至少一个边中确定一基准点,按照预设采样个数确定所述至少一个边对应的待选择参考像素点位置,包括:
    删掉所述至少一个边的起始位置和末端位置分别对应的参考像素点,得到第二相邻参考像素集合;
    以所述至少一个边的中点为基准点,基于所述第二相邻参考像素集合,按照预设采样个数确定待选择参考像素点位置。
  12. 根据权利要求1至11任一项所述的方法,其中,所述基于所述相邻参考像素子集,对所述待解码块进行预测解码,包括:
    基于所述相邻参考像素子集,确定模型参数;
    根据所述模型参数,建立预测模型;其中,所述预测模型用于表征所述待解码块中每个像素点对应的第一图像分量与第二图像分量之间的预测关系;
    基于所述预测模型,对所述待解码块进行预测解码。
  13. 一种解码预测装置,其中,所述解码预测装置包括获取单元、确定单元、选取单元和解码单元,
    所述获取单元,配置为获取与待解码块至少一个边相邻的参考像素点,得到第一相邻参考像素集合;
    所述确定单元,配置为从所述至少一个边中确定一基准点,按照预设采样个数确定所述至少一个边对应的待选择参考像素点位置;
    所述选取单元,配置为基于所述待选择参考像素点位置,从所述第一相邻参考像素集合中选取与所述待选择参考像素点位置对应的参考像素点,将选取得到的参考像素点组成相邻参考像素子集;
    所述解码单元,配置为基于所述相邻参考像素子集,对所述待解码块进行预测解码。
  14. 根据权利要求13所述的解码预测装置,其中,所述确定单元,具体配置为基于所述至少一个边的长度,若所述至少一个边的中点位于两个参考像素点之间的中间位置,则将所述中间位置偏右的第一个参考像素点位置作为所述至少一个边的基准点,或者将所述中间位置偏左的第一个参考像素点位置作为所述至少一个边的基准点。
  15. 根据权利要求13所述的解码预测装置,其中,所述确定单元,具体配置为基于所述至少一个边的长度,若所述至少一个边的中点位于两个参考像素点之间的中间位置,则将所述中间位置偏下的第一个参考像素点位置作为所述至少一个边的基准点,或者将所述中间位置偏上的第一个参考像素点位置作为所述至少一个边的基准点。
  16. 根据权利要求13所述的解码预测装置,其中,所述解码预测装置还包括计算单元,配置为基于所述预设采样个数和所述至少一个边的长度,计算得到第一采样间隔;
    所述确定单元,具体配置为以所述至少一个边的中点为基准点,按照所述第一采样间隔确定所述至少一个边对应的待选择参考像素点位置。
  17. 根据权利要求16所述的解码预测装置,其中,所述计算单元,还配置为基于所述至少一个边的长度,计算所述至少一个边的中点值;以及根据所述中点值和第一采样间隔,计算得到参考像素点位置;所述确定单元,具体配置为当所述中点值为非整数时,对所述中点值左侧的参考像素点位置向小取整,将取整后的参考像素点位置确定为待选择参考像素点位置;对所述中点值右侧的参考像素点位置向大取整,将取整后的参考像素点位置确定为待选择参考像素点位置。
  18. 根据权利要求16所述的解码预测装置,其中,所述计算单元,还配置为基于所述至少一个边的长度,计算所述至少一个边的中点值;以及根据所述中点值和第一采样间隔,计算得到参考像素点位置;所述确定单元,具体配置为当所述中点值为非整数时,对所述中点值左侧的参考像素点位置向大取整,将取整后的参考像素点位置确定为待选择参考像素点位置;对所述中点值右侧的参考像素点位置向小取整,将取整后的参考像素点位置确定为待选择参考像素点位置。
  19. 根据权利要求16所述的解码预测装置,其中,所述解码预测装置还包括调整单元,配置为对所述第一采样间隔进行调整,得到第二采样间隔;
    所述确定单元,还配置为以所述至少一个边的中点为基准点,按照所述第二采样间隔确定所述至少一个边对应的待选择参考像素点位置。
  20. 根据权利要求19所述的解码预测装置,其中,所述确定单元,还配置为以所述至少一个边的中点为基准点,按照所述第一采样间隔确定所述基准点左侧对应的待选择参考像素点位置,按照所述第二采样间隔确定所述基准点右侧对应的待选择参考像素点位置。
  21. 根据权利要求13所述的解码预测装置,其中,所述确定单元,还配置为以所述至少一个边的中点为基准点,根据所述预设采样个数将处于所述基准点附近的连续的参考像素点位置确定为待选择参考像素点位置,且所述基准点位于所述待选择参考像素点位置的中间位置。
  22. 根据权利要求13至21任一项所述的解码预测装置,其中,所述待解码块包括正方形解码块或者非正方形解码块。
  23. 根据权利要求13所述的解码预测装置,其中,所述获取单元,还配置为删掉所述至少一个边的起始位置和末端位置分别对应的参考像素点,得到第二相邻参考像素集合;
    所述确定单元,还配置为以所述至少一个边的中点为基准点,基于所述第二相邻参考像素集合,按照预设采样个数确定待选择参考像素点位置。
  24. 根据权利要求13至23任一项所述的解码预测装置,其中,所述解码预测装置还包括建立单元,配置为基于所述相邻参考像素子集,确定模型参数;以及根据所述模型参数,建立预测模型;其中,所述预测模型用于表征所述待解码块中每个像素点对应的第一图像分量与第二图像分量之间的预测关系;
    所述解码单元,具体配置为基于所述预测模型,对所述待解码块进行预测解码。
  25. 一种解码预测装置,其中,所述解码预测装置包括:存储器和处理器;
    所述存储器,用于存储能够在所述处理器上运行的计算机程序;
    所述处理器,用于在运行所述计算机程序时,执行如权利要求1至12任一项所述方法的步骤。
  26. 一种计算机存储介质,其中,所述计算机存储介质存储有解码预测程序,所述解码预测程序被至少一个处理器执行时实现如权利要求1至12任一项所述方法的步骤。
PCT/CN2018/123657 2018-12-25 2018-12-25 解码预测方法、装置及计算机存储介质 WO2020132908A1 (zh)

Priority Applications (27)

Application Number Priority Date Filing Date Title
CA3113787A CA3113787C (en) 2018-12-25 2018-12-25 Decoding prediction method, apparatus and computer storage medium which use reference samples adjacent to at least one side of decoding block
CN202310452465.XA CN116456083A (zh) 2018-12-25 2018-12-25 解码预测方法、装置及计算机存储介质
SG11202102435VA SG11202102435VA (en) 2018-12-25 2018-12-25 Decoding prediction method and apparatus, and computer storage medium
PCT/CN2018/123657 WO2020132908A1 (zh) 2018-12-25 2018-12-25 解码预测方法、装置及计算机存储介质
KR1020217023381A KR20210106547A (ko) 2018-12-25 2018-12-25 디코딩 예측 방법, 장치 및 컴퓨터 저장 매체
MX2021003940A MX2021003940A (es) 2018-12-25 2018-12-25 Metodo y aparato de prediccion de decodificacion y medio de almacenamiento computacional.
JP2021518718A JP7481332B2 (ja) 2018-12-25 2018-12-25 デコードのための予測方法及びその装置、並びにコンピュータ記憶媒体
CA3222255A CA3222255A1 (en) 2018-12-25 2018-12-25 Decoding prediction method and apparatus, and computer storage medium
CN202310451318.0A CN116456082A (zh) 2018-12-25 2018-12-25 编码预测方法、装置及计算机存储介质
CN202110246654.2A CN113068025B (zh) 2018-12-25 2018-12-25 解码预测方法、装置及计算机存储介质
EP18945329.3A EP3829167A4 (en) 2018-12-25 2018-12-25 DECODING PREDICTION METHOD AND APPARATUS, AND COMPUTER STORAGE MEDIUM
CN202310455271.5A CN116456084A (zh) 2018-12-25 2018-12-25 解码预测方法、装置及计算机存储介质
CN202110270601.4A CN113068026B (zh) 2018-12-25 2018-12-25 编码预测方法、装置及计算机存储介质
CN202310449638.2A CN116456081A (zh) 2018-12-25 2018-12-25 解码预测方法、装置及计算机存储介质
BR112021006547A BR112021006547A2 (pt) 2018-12-25 2018-12-25 método de previsão para decodificação, dispositivo de previsão para decodificação e mídia de armazenamento de computador que armazena um programa de previsão para decodificação
CN202310454159.XA CN116614622A (zh) 2018-12-25 2018-12-25 解码预测方法、装置及计算机存储介质
AU2018454766A AU2018454766A1 (en) 2018-12-25 2018-12-25 Decoding prediction method and apparatus, and computer storage medium
CN201880095849.8A CN112470468A (zh) 2018-12-25 2018-12-25 解码预测方法、装置及计算机存储介质
PH12021550515A PH12021550515A1 (en) 2018-12-25 2021-03-10 Decoding prediction method and apparatus, and computer storage medium
IL281399A IL281399A (en) 2018-12-25 2021-03-10 Method and apparatus for predicting decoding and means for computer storage
ZA2021/01726A ZA202101726B (en) 2018-12-25 2021-03-15 Decoding prediction method and apparatus, and computer storage medium
US17/234,639 US11297313B2 (en) 2018-12-25 2021-04-19 Prediction method for decoding and apparatus, and computer storage medium
US17/677,802 US11683477B2 (en) 2018-12-25 2022-02-22 Prediction method for decoding and apparatus, and computer storage medium
US17/677,810 US11785208B2 (en) 2018-12-25 2022-02-22 Prediction method for decoding and apparatus, and computer storage medium
US17/677,807 US11683478B2 (en) 2018-12-25 2022-02-22 Prediction method for decoding and apparatus, and computer storage medium
US17/677,813 US11677936B2 (en) 2018-12-25 2022-02-22 Prediction method for decoding and apparatus, and computer storage medium
US18/328,705 US20230308643A1 (en) 2018-12-25 2023-06-02 Prediction method for decoding and apparatus, and computer storage medium

Applications Claiming Priority (1)

Application Number Priority Date Filing Date Title
PCT/CN2018/123657 WO2020132908A1 (zh) 2018-12-25 2018-12-25 解码预测方法、装置及计算机存储介质

Related Child Applications (1)

Application Number Title Priority Date Filing Date
US17/234,639 Continuation US11297313B2 (en) 2018-12-25 2021-04-19 Prediction method for decoding and apparatus, and computer storage medium

Publications (1)

Publication Number Publication Date
WO2020132908A1 true WO2020132908A1 (zh) 2020-07-02

Family

ID=71126759

Family Applications (1)

Application Number Title Priority Date Filing Date
PCT/CN2018/123657 WO2020132908A1 (zh) 2018-12-25 2018-12-25 解码预测方法、装置及计算机存储介质

Country Status (14)

Country Link
US (6) US11297313B2 (zh)
EP (1) EP3829167A4 (zh)
JP (1) JP7481332B2 (zh)
KR (1) KR20210106547A (zh)
CN (8) CN116456082A (zh)
AU (1) AU2018454766A1 (zh)
BR (1) BR112021006547A2 (zh)
CA (2) CA3222255A1 (zh)
IL (1) IL281399A (zh)
MX (1) MX2021003940A (zh)
PH (1) PH12021550515A1 (zh)
SG (1) SG11202102435VA (zh)
WO (1) WO2020132908A1 (zh)
ZA (1) ZA202101726B (zh)

Cited By (1)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
JP7476426B2 (ja) 2021-06-30 2024-04-30 テンセント・アメリカ・エルエルシー ビデオコーディングのための方法、装置、およびコンピュータプログラム

Families Citing this family (2)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
KR20210108397A (ko) 2019-01-02 2021-09-02 광동 오포 모바일 텔레커뮤니케이션즈 코포레이션 리미티드 예측 디코딩 방법, 장치 및 컴퓨터 저장 매체
CN116737607B (zh) * 2023-08-16 2023-11-21 之江实验室 样本数据缓存方法、系统、计算机设备和存储介质

Citations (4)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN102611885A (zh) * 2011-01-20 2012-07-25 华为技术有限公司 一种编解码方法和装置
CN104125463A (zh) * 2013-04-26 2014-10-29 华为技术有限公司 一种图像预测编码方法及图像编码器
CN106688238A (zh) * 2013-10-17 2017-05-17 华为技术有限公司 改进后的深度图帧内编码的参考像素点选择和滤波
US20170318305A1 (en) * 2014-10-31 2017-11-02 Nec Corporation Prediction-image generation device and prediction-image generation method

Family Cites Families (59)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US7813432B2 (en) * 2004-12-30 2010-10-12 Intel Corporation Offset buffer for intra-prediction of digital video
TWI401972B (zh) 2009-06-23 2013-07-11 Acer Inc 時間性錯誤隱藏方法
US9736486B2 (en) * 2010-02-08 2017-08-15 Nokia Technologies Oy Apparatus, a method and a computer program for video coding
CN102215396A (zh) * 2010-04-09 2011-10-12 华为技术有限公司 一种视频编解码方法和系统
KR101772459B1 (ko) * 2010-05-17 2017-08-30 엘지전자 주식회사 신규한 인트라 예측 모드
WO2011149265A2 (en) * 2010-05-25 2011-12-01 Lg Electronics Inc. New planar prediction mode
CN107682702B (zh) 2010-09-30 2020-08-11 太阳专利托管公司 图像解码方法、图像解码装置
US20120163457A1 (en) 2010-12-28 2012-06-28 Viktor Wahadaniah Moving picture decoding method, moving picture coding method, moving picture decoding apparatus, moving picture coding apparatus, and moving picture coding and decoding apparatus
US10034018B2 (en) 2011-09-23 2018-07-24 Velos Media, Llc Decoded picture buffer management
US9280835B2 (en) 2011-10-18 2016-03-08 Nokia Technologies Oy Method for coding and an apparatus based on a DC prediction value
GB2498550B (en) 2012-01-19 2016-02-24 Canon Kk Method and device for processing components of an image for encoding or decoding
CN103220506B (zh) 2012-01-19 2015-11-25 华为技术有限公司 一种编解码方法和设备
CN102547296B (zh) 2012-02-27 2015-04-01 开曼群岛威睿电通股份有限公司 移动估计加速电路、移动估计方法及环路滤波加速电路
GB2501518A (en) * 2012-04-27 2013-10-30 Canon Kk Encoding and Decoding an Image Comprising Blocks of Pixels Including Selecting the Prediction Mode of Prediction Units
US9667942B2 (en) * 2012-11-20 2017-05-30 Qualcomm Incorporated Adaptive luminance compensation in three dimensional video coding
JP6005572B2 (ja) * 2013-03-28 2016-10-12 Kddi株式会社 動画像符号化装置、動画像復号装置、動画像符号化方法、動画像復号方法、およびプログラム
CN110225356B (zh) * 2013-04-08 2024-02-13 Ge视频压缩有限责任公司 多视图解码器
CN105474639B (zh) * 2013-07-10 2018-08-07 凯迪迪爱通信技术有限公司 视频编码装置、视频解码装置、视频系统、视频编码方法、视频解码方法以及程序
US10129542B2 (en) 2013-10-17 2018-11-13 Futurewei Technologies, Inc. Reference pixel selection and filtering for intra coding of depth map
WO2015100515A1 (en) * 2013-12-30 2015-07-09 Qualcomm Incorporated Simplification of segment-wise dc coding of large prediction blocks in 3d video coding
KR102425577B1 (ko) * 2014-06-18 2022-07-26 삼성전자주식회사 휘도차를 보상하기 위한 인터 레이어 비디오 부호화 방법 및 그 장치, 비디오 복호화 방법 및 그 장치
WO2016182317A1 (ko) * 2015-05-12 2016-11-17 삼성전자 주식회사 인트라 예측을 수행하는 영상 복호화 방법 및 그 장치 및 인트라 예측을 수행하는 영상 부호화 방법 및 그 장치
US10887597B2 (en) * 2015-06-09 2021-01-05 Qualcomm Incorporated Systems and methods of determining illumination compensation parameters for video coding
KR20180019092A (ko) * 2015-06-16 2018-02-23 엘지전자 주식회사 영상 코딩 시스템에서 조도 보상에 기반한 블록 예측 방법 및 장치
US20160373742A1 (en) * 2015-06-18 2016-12-22 Qualcomm Incorporated Intra prediction and intra mode coding
KR102460912B1 (ko) * 2015-07-08 2022-10-28 인터디지털 매디슨 페턴트 홀딩스 에스에이에스 교차 평면 필터링을 이용한 향상된 크로마 코딩
US10200719B2 (en) * 2015-11-25 2019-02-05 Qualcomm Incorporated Modification of transform coefficients for non-square transform units in video coding
US10390021B2 (en) * 2016-03-18 2019-08-20 Mediatek Inc. Method and apparatus of video coding
EP3453174A1 (en) * 2016-05-06 2019-03-13 VID SCALE, Inc. Method and system for decoder-side intra mode derivation for block-based video coding
US10397569B2 (en) * 2016-06-03 2019-08-27 Mediatek Inc. Method and apparatus for template-based intra prediction in image and video coding
US10368107B2 (en) * 2016-08-15 2019-07-30 Qualcomm Incorporated Intra video coding using a decoupled tree structure
US10326986B2 (en) * 2016-08-15 2019-06-18 Qualcomm Incorporated Intra video coding using a decoupled tree structure
WO2018062921A1 (ko) * 2016-09-30 2018-04-05 엘지전자 주식회사 영상 코딩 시스템에서 블록 분할 및 인트라 예측 방법 및 장치
US10880570B2 (en) * 2016-10-05 2020-12-29 Qualcomm Incorporated Systems and methods of adaptively determining template size for illumination compensation
US10880546B2 (en) * 2016-10-11 2020-12-29 Lg Electronics Inc. Method and apparatus for deriving intra prediction mode for chroma component
US10477240B2 (en) * 2016-12-19 2019-11-12 Qualcomm Incorporated Linear model prediction mode with sample accessing for video coding
EP4072134A1 (en) * 2016-12-27 2022-10-12 Samsung Electronics Co., Ltd. Video affine mode encoding method and device therefor, and decoding method and device therefor
US11277635B2 (en) * 2017-03-17 2022-03-15 Vid Scale, Inc. Predictive coding for 360-degree video based on geometry padding
WO2018236028A1 (ko) * 2017-06-21 2018-12-27 엘지전자(주) 인트라 예측 모드 기반 영상 처리 방법 및 이를 위한 장치
US11190799B2 (en) * 2017-06-21 2021-11-30 Lg Electronics Inc. Intra-prediction mode-based image processing method and apparatus therefor
WO2019009546A1 (ko) * 2017-07-04 2019-01-10 엘지전자 주식회사 인터 예측 기반의 영상 처리 방법 및 이를 위한 장치
US20200145649A1 (en) * 2017-07-10 2020-05-07 Lg Electronics Inc. Method and apparatus for reducing noise in frequency-domain in image coding system
GB2567249A (en) * 2017-10-09 2019-04-10 Canon Kk New sample sets and new down-sampling schemes for linear component sample prediction
GB2567861A (en) 2017-10-27 2019-05-01 Sony Corp Image data encoding and decoding
US11722694B2 (en) 2018-01-26 2023-08-08 Interdigital Vc Holdings, Inc. Method and apparatus for video encoding and decoding based on a linear model responsive to neighboring samples
US10523948B2 (en) * 2018-02-05 2019-12-31 Tencent America LLC Method and apparatus for video coding
US10462483B1 (en) * 2018-04-26 2019-10-29 Tencent America LLC Method and apparatus for video coding
US10609402B2 (en) * 2018-05-02 2020-03-31 Tencent America LLC Method and apparatus for prediction and transform for small blocks
AU2019292471A1 (en) 2018-06-30 2021-01-28 Guangdong Oppo Mobile Telecommunications Corp., Ltd. Merge mode-based inter-prediction method and apparatus
US11128872B2 (en) * 2018-07-16 2021-09-21 Qualcomm Incorporated Position dependent intra prediction combination with wide angle intra prediction
US10491893B1 (en) * 2018-08-29 2019-11-26 Tencent America LLC Method and apparatus for multi-line intra prediction
KR20200028856A (ko) 2018-09-07 2020-03-17 김기백 인트라 예측을 이용한 영상 부호화/복호화 방법 및 장치
IL307586A (en) 2018-09-21 2023-12-01 Guangdong Oppo Mobile Telecommunications Corp Ltd A method for encoding/decoding image signals and a device therefor
IL304953A (en) 2018-09-21 2023-10-01 Guangdong Oppo Mobile Telecommunications Corp Ltd A method of encoding/decoding image signals and a device therefor
KR20230169407A (ko) * 2018-10-05 2023-12-15 후아웨이 테크놀러지 컴퍼니 리미티드 인트라 예측 방법 및 디바이스
IL281823B2 (en) 2018-10-12 2024-01-01 Guangdong Oppo Mobile Telecommunications Corp Ltd A method for encoding/decoding an image signal and a device for it
US11197005B2 (en) * 2018-11-08 2021-12-07 Qualcomm Incorporated Cross-component prediction for video coding
US11178396B2 (en) * 2018-11-14 2021-11-16 Tencent America LLC Constrained intra prediction and unified most probable mode list generation
CA3085391C (en) * 2018-11-23 2023-10-24 Lg Electronics Inc. Method for decoding image on basis of cclm prediction in image coding system, and device therefor

Patent Citations (4)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN102611885A (zh) * 2011-01-20 2012-07-25 华为技术有限公司 一种编解码方法和装置
CN104125463A (zh) * 2013-04-26 2014-10-29 华为技术有限公司 一种图像预测编码方法及图像编码器
CN106688238A (zh) * 2013-10-17 2017-05-17 华为技术有限公司 改进后的深度图帧内编码的参考像素点选择和滤波
US20170318305A1 (en) * 2014-10-31 2017-11-02 Nec Corporation Prediction-image generation device and prediction-image generation method

Cited By (1)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
JP7476426B2 (ja) 2021-06-30 2024-04-30 テンセント・アメリカ・エルエルシー ビデオコーディングのための方法、装置、およびコンピュータプログラム

Also Published As

Publication number Publication date
CN113068025B (zh) 2023-05-12
CA3113787A1 (en) 2020-07-02
ZA202101726B (en) 2022-07-27
CN116614622A (zh) 2023-08-18
EP3829167A1 (en) 2021-06-02
US20220182609A1 (en) 2022-06-09
CN113068026A (zh) 2021-07-02
CN112470468A (zh) 2021-03-09
US11785208B2 (en) 2023-10-10
US20230308643A1 (en) 2023-09-28
US11683477B2 (en) 2023-06-20
EP3829167A4 (en) 2021-08-11
BR112021006547A2 (pt) 2021-07-06
JP7481332B2 (ja) 2024-05-10
CA3113787C (en) 2024-02-13
MX2021003940A (es) 2021-05-27
CA3222255A1 (en) 2020-07-02
US20220182608A1 (en) 2022-06-09
CN113068026B (zh) 2023-05-05
US20220182610A1 (en) 2022-06-09
KR20210106547A (ko) 2021-08-30
US11297313B2 (en) 2022-04-05
CN116456083A (zh) 2023-07-18
CN116456084A (zh) 2023-07-18
US11683478B2 (en) 2023-06-20
AU2018454766A1 (en) 2021-04-08
CN116456082A (zh) 2023-07-18
IL281399A (en) 2021-04-29
US20220182611A1 (en) 2022-06-09
SG11202102435VA (en) 2021-04-29
JP2022520295A (ja) 2022-03-30
US20210243431A1 (en) 2021-08-05
US11677936B2 (en) 2023-06-13
PH12021550515A1 (en) 2022-02-21
CN116456081A (zh) 2023-07-18
CN113068025A (zh) 2021-07-02

Similar Documents

Publication Publication Date Title
TWI705694B (zh) 片級內部區塊複製及其他視訊寫碼改善
WO2020132908A1 (zh) 解码预测方法、装置及计算机存储介质
JP2023113908A (ja) 予測デコーディング方法、装置及びコンピュータ記憶媒体
AU2019357929A1 (en) Video image component prediction method and apparatus, and computer storage medium
WO2020056767A1 (zh) 视频图像分量的预测方法、装置及计算机存储介质
WO2022140905A1 (zh) 预测方法、编码器、解码器以及存储介质
CN115988202A (zh) 一种用于帧内预测的设备和方法
RU2781240C1 (ru) Способ и устройство предсказания для декодирования и компьютерный носитель данных
CN118266217A (zh) 用于视频处理的方法、设备和介质

Legal Events

Date Code Title Description
121 Ep: the epo has been informed by wipo that ep was designated in this application

Ref document number: 18945329

Country of ref document: EP

Kind code of ref document: A1

ENP Entry into the national phase

Ref document number: 2018945329

Country of ref document: EP

Effective date: 20210226

ENP Entry into the national phase

Ref document number: 3113787

Country of ref document: CA

ENP Entry into the national phase

Ref document number: 2021518718

Country of ref document: JP

Kind code of ref document: A

ENP Entry into the national phase

Ref document number: 2018454766

Country of ref document: AU

Date of ref document: 20181225

Kind code of ref document: A

REG Reference to national code

Ref country code: BR

Ref legal event code: B01A

Ref document number: 112021006547

Country of ref document: BR

NENP Non-entry into the national phase

Ref country code: DE

ENP Entry into the national phase

Ref document number: 112021006547

Country of ref document: BR

Kind code of ref document: A2

Effective date: 20210406

ENP Entry into the national phase

Ref document number: 20217023381

Country of ref document: KR

Kind code of ref document: A