US20210006818A1 - Picture prediction method and related apparatus - Google Patents
Picture prediction method and related apparatus Download PDFInfo
- Publication number
- US20210006818A1 US20210006818A1 US17/025,901 US202017025901A US2021006818A1 US 20210006818 A1 US20210006818 A1 US 20210006818A1 US 202017025901 A US202017025901 A US 202017025901A US 2021006818 A1 US2021006818 A1 US 2021006818A1
- Authority
- US
- United States
- Prior art keywords
- picture block
- pixel
- current picture
- motion vector
- block
- Prior art date
- Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
- Abandoned
Links
Images
Classifications
-
- H—ELECTRICITY
- H04—ELECTRIC COMMUNICATION TECHNIQUE
- H04N—PICTORIAL COMMUNICATION, e.g. TELEVISION
- H04N19/00—Methods or arrangements for coding, decoding, compressing or decompressing digital video signals
- H04N19/10—Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using adaptive coding
- H04N19/169—Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using adaptive coding characterised by the coding unit, i.e. the structural portion or semantic portion of the video signal being the object or the subject of the adaptive coding
- H04N19/17—Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using adaptive coding characterised by the coding unit, i.e. the structural portion or semantic portion of the video signal being the object or the subject of the adaptive coding the unit being an image region, e.g. an object
- H04N19/172—Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using adaptive coding characterised by the coding unit, i.e. the structural portion or semantic portion of the video signal being the object or the subject of the adaptive coding the unit being an image region, e.g. an object the region being a picture, frame or field
-
- H—ELECTRICITY
- H04—ELECTRIC COMMUNICATION TECHNIQUE
- H04N—PICTORIAL COMMUNICATION, e.g. TELEVISION
- H04N19/00—Methods or arrangements for coding, decoding, compressing or decompressing digital video signals
- H04N19/10—Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using adaptive coding
- H04N19/169—Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using adaptive coding characterised by the coding unit, i.e. the structural portion or semantic portion of the video signal being the object or the subject of the adaptive coding
- H04N19/17—Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using adaptive coding characterised by the coding unit, i.e. the structural portion or semantic portion of the video signal being the object or the subject of the adaptive coding the unit being an image region, e.g. an object
- H04N19/176—Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using adaptive coding characterised by the coding unit, i.e. the structural portion or semantic portion of the video signal being the object or the subject of the adaptive coding the unit being an image region, e.g. an object the region being a block, e.g. a macroblock
-
- H—ELECTRICITY
- H04—ELECTRIC COMMUNICATION TECHNIQUE
- H04N—PICTORIAL COMMUNICATION, e.g. TELEVISION
- H04N19/00—Methods or arrangements for coding, decoding, compressing or decompressing digital video signals
- H04N19/10—Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using adaptive coding
- H04N19/169—Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using adaptive coding characterised by the coding unit, i.e. the structural portion or semantic portion of the video signal being the object or the subject of the adaptive coding
- H04N19/182—Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using adaptive coding characterised by the coding unit, i.e. the structural portion or semantic portion of the video signal being the object or the subject of the adaptive coding the unit being a pixel
-
- H—ELECTRICITY
- H04—ELECTRIC COMMUNICATION TECHNIQUE
- H04N—PICTORIAL COMMUNICATION, e.g. TELEVISION
- H04N19/00—Methods or arrangements for coding, decoding, compressing or decompressing digital video signals
- H04N19/50—Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using predictive coding
- H04N19/503—Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using predictive coding involving temporal prediction
- H04N19/51—Motion estimation or motion compensation
-
- H—ELECTRICITY
- H04—ELECTRIC COMMUNICATION TECHNIQUE
- H04N—PICTORIAL COMMUNICATION, e.g. TELEVISION
- H04N19/00—Methods or arrangements for coding, decoding, compressing or decompressing digital video signals
- H04N19/50—Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using predictive coding
- H04N19/503—Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using predictive coding involving temporal prediction
- H04N19/51—Motion estimation or motion compensation
- H04N19/513—Processing of motion vectors
-
- H—ELECTRICITY
- H04—ELECTRIC COMMUNICATION TECHNIQUE
- H04N—PICTORIAL COMMUNICATION, e.g. TELEVISION
- H04N19/00—Methods or arrangements for coding, decoding, compressing or decompressing digital video signals
- H04N19/50—Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using predictive coding
- H04N19/503—Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using predictive coding involving temporal prediction
- H04N19/51—Motion estimation or motion compensation
- H04N19/513—Processing of motion vectors
- H04N19/517—Processing of motion vectors by encoding
- H04N19/52—Processing of motion vectors by encoding by predictive encoding
-
- H—ELECTRICITY
- H04—ELECTRIC COMMUNICATION TECHNIQUE
- H04N—PICTORIAL COMMUNICATION, e.g. TELEVISION
- H04N19/00—Methods or arrangements for coding, decoding, compressing or decompressing digital video signals
- H04N19/50—Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using predictive coding
- H04N19/503—Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using predictive coding involving temporal prediction
- H04N19/51—Motion estimation or motion compensation
- H04N19/537—Motion estimation other than block-based
- H04N19/54—Motion estimation other than block-based using feature points or meshes
-
- H—ELECTRICITY
- H04—ELECTRIC COMMUNICATION TECHNIQUE
- H04N—PICTORIAL COMMUNICATION, e.g. TELEVISION
- H04N19/00—Methods or arrangements for coding, decoding, compressing or decompressing digital video signals
- H04N19/50—Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using predictive coding
- H04N19/593—Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using predictive coding involving spatial prediction techniques
-
- H—ELECTRICITY
- H04—ELECTRIC COMMUNICATION TECHNIQUE
- H04N—PICTORIAL COMMUNICATION, e.g. TELEVISION
- H04N19/00—Methods or arrangements for coding, decoding, compressing or decompressing digital video signals
- H04N19/70—Methods or arrangements for coding, decoding, compressing or decompressing digital video signals characterised by syntax aspects related to video coding, e.g. related to compression standards
Definitions
- the subject matter and the claimed application were made by or on the behalf of Huawei Technologies Co., Ltd., of Shenzhen, Guangdong province, P.R. China and University of Science and Technology of China, of Baohe District, Hefei City, Anhui province, China, under a joint research agreement titled “Research and Development of Next Generation Video Coding Standards and Technologies.”
- the joint research agreement was in effect on or before the claimed application was made, and the claimed application was made as a result of activities undertaken within the scope of the joint research agreement.
- the present application relates to the field of picture processing technologies, and specifically, to a picture prediction method and a related apparatus.
- HEVC High Efficiency Video Coding
- a basic principle of video compression coding is to use correlation between a space domain, a time domain, and a code word to remove redundancy as much as possible.
- a prevalent practice is to use a block-based hybrid video coding framework to implement video compression coding by performing steps of prediction (including intra-frame prediction and inter-frame prediction), transform, quantization, entropy coding, and the like.
- This coding framework shows a high viability, and therefore, HEVC still uses this block-based hybrid video coding framework.
- motion estimation or motion compensation is a key technology that affects coding/decoding efficiency.
- all conventional motion estimation or motion compensation algorithms are block motion compensation algorithms that are established based on a translational motion model.
- motion in the real world is diversified, and irregular motion such as scaling up/down, rotation, or parabolic motion is ubiquitous.
- an irregular motion model such as an affine motion model, a rotational motion model, or a scaling motion model
- computational complexity of conventional picture prediction performed based on a non-translational motion model is generally quite high.
- Embodiments of the present application provide a picture prediction method and a related apparatus, so as to reduce calculation complexity of picture prediction based on a non-translational motion model.
- a first aspect of the present application provides a picture prediction method, including:
- K is an integer greater than 1
- the K pixel samples include a first vertex angle pixel sample in the current picture block, a motion vector predictor of the first vertex angle pixel sample is obtained based on a motion vector of a preset first spatially adjacent picture block of the current picture block, and the first spatially adjacent picture block is spatially adjacent to the first vertex angle pixel sample; and performing, based on a non-translational motion model and the motion vector predictors of the K pixel samples, pixel value prediction on the current picture block.
- the K pixel samples include at least two pixel samples in an upper left pixel sample, an upper right pixel sample, or a lower left pixel sample in the current picture block;
- the upper left pixel sample in the current picture block is an upper left vertex of the current picture block or a pixel block that is in the current picture block and that includes an upper left vertex of the current picture block
- the lower left pixel sample in the current picture block is a lower left vertex of the current picture block or a pixel block that is in the current picture block and that includes a lower left vertex of the current picture block
- the upper right pixel sample in the current picture block is an upper right vertex of the current picture block or a pixel block that is in the current picture block and that includes an upper right vertex of the current picture block.
- the first spatially adjacent picture block is a picture block X1, a picture block X2, or a picture block X3;
- a motion vector predictor of the upper left pixel sample in the current picture block is obtained based on a motion vector of the picture block X1, the picture block X1 is a spatially adjacent picture block of the current picture block, and the picture block X1 is spatially adjacent to the upper left pixel sample in the current picture block;
- a motion vector predictor of the upper right pixel sample in the current picture block is obtained based on a motion vector of the picture block X2, the picture block X2 is a spatially adjacent picture block of the current picture block, and the picture block X2 is spatially adjacent to the upper right pixel sample in the current picture block;
- a motion vector predictor of the lower left pixel sample in the current picture block is obtained based on a motion vector of the picture block X3, the picture block X3 is a spatially adjacent picture block of the current picture block, and the picture block X3 is spatially adjacent to the lower left pixel sample in the current picture block.
- the first vertex angle pixel sample is the upper left pixel sample in the current picture block
- the first spatially adjacent picture block is the picture block X1
- the picture block X1 is a spatially adjacent picture block on the upper left of the current picture block.
- the picture block X2 is a spatially adjacent picture block above the current picture block
- an absolute value of a difference between a motion vector absolute value of the picture block X2 and a motion vector absolute value of the picture block X1 is greater than or equal to an absolute value of a difference between a motion vector absolute value of a picture block X5 and the motion vector absolute value of the picture block X1, and the picture block X5 is any one of remaining at least some spatially adjacent picture blocks, other than the picture block X2, that are of the current picture block and that are spatially adjacent to the upper right vertex.
- the picture block X3 is a spatially adjacent picture block on the left of the current picture block;
- an absolute value of a difference between a motion vector absolute value of the picture block X3 and the motion vector absolute value of the picture block X1 is greater than or equal to an absolute value of a difference between a motion vector absolute value of a picture block X6 and the motion vector absolute value of the picture block X1, and the picture block X6 is any one of remaining at least some spatially adjacent picture blocks, other than the picture block X3, that are of the current picture block and that are spatially adjacent to the lower left vertex.
- the performing, based on a non-translational motion model and the motion vector predictors of the K pixel samples, pixel value prediction on the current picture block includes:
- a pixel sample i is any pixel sample except the first vertex angle pixel sample in the K pixel samples, and if a prediction direction corresponding to a motion vector predictor of the pixel sample i is different from a prediction direction corresponding to the motion vector predictor of the first vertex angle pixel sample, a reference frame index corresponding to the motion vector predictor of the pixel sample i is 0.
- the reference frame Y1 is a reference frame corresponding to the motion vector predictor of the first vertex angle pixel sample
- that the motion vector predictor of the pixel sample i is scaled to the reference frame Y1 includes: if reference frame indexes corresponding to the motion vector predictors of the first vertex angle pixel sample and the pixel sample i are different, the motion vector predictor of the pixel sample i is scaled to the reference frame Y1.
- the performing motion estimation on the K pixel samples by using motion vector predictors of K2 pixel samples and motion vector predictors that are of the K1 pixel samples and that are scaled to the reference frame Y1, to obtain motion vectors of the K pixel samples includes:
- the non-translational motion model is any one of the following models: an affine transformation model, a parabolic motion model, a rotary motion model, a perspective motion model, a shearing motion model, or a scalable motion model.
- the picture prediction method is applied to a video coding process or the picture prediction method is applied to a video decoding process.
- a second aspect of the present application provides a picture prediction apparatus, including:
- a determining unit configured to determine motion vector predictors of K pixel samples in a current picture block, where K is an integer greater than 1, the K pixel samples include a first vertex angle pixel sample in the current picture block, a motion vector predictor of the first vertex angle pixel sample is obtained based on a motion vector of a preset first spatially adjacent picture block of the current picture block, and the first spatially adjacent picture block is spatially adjacent to the first vertex angle pixel sample;
- a predicting unit configured to perform, based on a non-translational motion model and the motion vector predictors of the K pixel samples, pixel value prediction on the current picture block.
- the K pixel samples include at least two pixel samples in an upper left pixel sample, an upper right pixel sample, or a lower left pixel sample in the current picture block;
- the upper left pixel sample in the current picture block is an upper left vertex of the current picture block or a pixel block that is in the current picture block and that includes an upper left vertex of the current picture block
- the lower left pixel sample in the current picture block is a lower left vertex of the current picture block or a pixel block that is in the current picture block and that includes a lower left vertex of the current picture block
- the upper right pixel sample in the current picture block is an upper right vertex of the current picture block or a pixel block that is in the current picture block and that includes an upper right vertex of the current picture block.
- the first spatially adjacent picture block is a picture block X1, a picture block X2, or a picture block X3;
- a motion vector predictor of the upper left pixel sample in the current picture block is obtained based on a motion vector of the picture block X1, the picture block X1 is a spatially adjacent picture block of the current picture block, and the picture block X1 is spatially adjacent to the upper left pixel sample in the current picture block;
- a motion vector predictor of the upper right pixel sample in the current picture block is obtained based on a motion vector of the picture block X2, the picture block X2 is a spatially adjacent picture block of the current picture block, and the picture block X2 is spatially adjacent to the upper right pixel sample in the current picture block;
- a motion vector predictor of the lower left pixel sample in the current picture block is obtained based on a motion vector of the picture block X3, the picture block X3 is a spatially adjacent picture block of the current picture block, and the picture block X3 is spatially adjacent to the lower left pixel sample in the current picture block.
- the first vertex angle pixel sample is the upper left pixel sample in the current picture block
- the first spatially adjacent picture block is the picture block X1
- the picture block X1 is a spatially adjacent picture block on the upper left of the current picture block.
- the picture block X2 is a spatially adjacent picture block above the current picture block
- an absolute value of a difference between a motion vector absolute value of the picture block X2 and a motion vector absolute value of the picture block X1 is greater than or equal to an absolute value of a difference between a motion vector absolute value of a picture block X5 and the motion vector absolute value of the picture block X1, and the picture block X5 is any one of remaining at least some spatially adjacent picture blocks, other than the picture block X2, that are of the current picture block and that are spatially adjacent to the upper right vertex.
- the picture block X3 is a spatially adjacent picture block on the left of the current picture block;
- an absolute value of a difference between a motion vector absolute value of the picture block X3 and the motion vector absolute value of the picture block X1 is greater than or equal to an absolute value of a difference between a motion vector absolute value of a picture block X6 and the motion vector absolute value of the picture block X1, and the picture block X6 is any one of remaining at least some spatially adjacent picture blocks, other than the picture block X3, that are of the current picture block and that are spatially adjacent to the lower left vertex.
- the predicting unit is specifically configured to: when reference frames corresponding to motion vector predictors of K1 pixel samples in the K pixel samples are not a reference frame Y1, scale the motion vector predictors of the K1 pixel samples to the reference frame Y1, perform motion estimation on the K pixel samples by using motion vector predictors of K2 pixel samples and motion vector predictors that are of the K1 pixel samples and that are scaled to the reference frame Y1, to obtain motion vectors of the K pixel samples, and perform pixel value prediction on the current picture block by using the non-translational motion model and the motion vectors of the K pixel samples, where the K2 pixel samples are remaining pixel samples other than the K1 pixel samples in the K pixel samples, and K1 and K2 are positive integers.
- a pixel sample i is any pixel sample except the first vertex angle pixel sample in the K pixel samples, and if a prediction direction corresponding to a motion vector predictor of the pixel sample i is different from a prediction direction corresponding to the motion vector predictor of the first vertex angle pixel sample, a reference frame index corresponding to the motion vector predictor of the pixel sample i is 0.
- the reference frame Y1 is a reference frame corresponding to the motion vector predictor of the first vertex angle pixel sample
- the predicting unit is specifically configured to: if reference frame indexes corresponding to the motion vector predictors of the first vertex angle pixel sample and the pixel sample i are different, scale the motion vector predictor of the pixel sample i to the reference frame Y1.
- the non-translational motion model is any one of the following models: an affine transformation model, a parabolic motion model, a rotary motion model, a perspective motion model, a shearing motion model, or a scalable motion model.
- the picture prediction apparatus is applied to a video coding apparatus or the picture prediction apparatus is applied to a video decoding apparatus.
- determined motion vector predictors of K pixel samples are referenced during pixel value prediction performed, based on a non-translational motion model, on a current picture block.
- the K pixel samples include a first vertex angle pixel sample in the current picture block.
- the motion vector predictors of the K pixel samples are directly used for prediction, a motion vector predictor of the first vertex angle pixel sample in the K pixel samples is obtained based on a motion vector of a preset first spatially adjacent picture block of the current picture block, and there is a single optional case for the motion vector predictors of the K pixel samples.
- FIG. 1 - a is a schematic diagram of a prediction unit division manner corresponding to intra-frame prediction according to an embodiment of the present application
- FIG. 1 - b is a schematic diagram of several prediction unit division manners corresponding to inter-frame prediction according to an embodiment of the present application;
- FIG. 1 - c is a schematic flowchart of a picture prediction method according to an embodiment of the present application
- FIG. 2 is a schematic diagram of possible adjacent reference blocks of a current picture block according to an embodiment of the present application
- FIG. 3 - a is a schematic flowchart of another picture prediction method according to an embodiment of the present application.
- FIG. 3 - b is another schematic diagram of possible adjacent reference blocks of a current picture block according to an embodiment of the present application.
- FIG. 3 - c is a schematic diagram of a reference frame of a current picture block according to an embodiment of the present application.
- FIG. 3 - d is a schematic diagram of a current picture block according to an embodiment of the present application.
- FIG. 3 - e is a schematic diagram of affine transformation of a picture block according to an embodiment of the present application.
- FIG. 4 - a is a schematic flowchart of another picture prediction method according to an embodiment of the present application.
- FIG. 4 - b is another schematic diagram of adjacent reference blocks of a current picture block according to an embodiment of the present application.
- FIG. 5 is a schematic flowchart of another picture prediction method according to an embodiment of the present application.
- FIG. 6 is a schematic flowchart of another picture prediction method according to an embodiment of the present application.
- FIG. 7 is a schematic flowchart of a video decoding method according to an embodiment of the present application.
- FIG. 8 is a schematic diagram of a picture prediction apparatus according to an embodiment of the present application.
- FIG. 9 is a schematic diagram of another picture prediction apparatus according to an embodiment of the present application.
- FIG. 10 is a schematic diagram of another picture prediction apparatus according to an embodiment of the present application.
- the embodiments of the present application provide a picture prediction method and a related apparatus, so as to reduce calculation complexity of picture prediction based on a non-translational motion model.
- a video sequence includes a series of pictures, the pictures are further divided into slices, and the slices are further divided into blocks.
- Video coding is to perform coding processing from left to right and from top to bottom row by row starting from an upper left corner position of a picture by using a block as a unit.
- a concept of a block is further extended.
- MB macroblock
- coding unit In the HEVC standard, basic concepts such as a coding unit (CU), a prediction unit (PU), and a transform unit (TU) are used, and multiple units are classified according to functions, and a completely new tree-based structure is used for description.
- the coding unit may be divided into smaller coding units according to a quadtree, and the smaller coding unit may be further divided, to form a quadtree structure.
- the prediction unit and the transform unit also have similar tree structures. Regardless of whether a unit is a coding unit, a prediction unit, or a transform unit, the unit belongs to the concept of a block in essence.
- the coding unit is similar to a macroblock MB or a coding block, and is a basic unit for partitioning and coding an picture.
- the prediction unit may correspond to a prediction block, and is a basic unit for predictive coding.
- the coding unit is further divided into multiple prediction units according to a division mode.
- the transform unit may correspond to a transform block, and is a basic unit for transforming a prediction residual.
- a size of the coding unit may include four levels: 64 ⁇ 64, 32 ⁇ 32, 16 ⁇ 16, and 8 ⁇ 8. Coding units at each level may be divided into prediction units of different sizes according to intra-frame prediction and inter-frame prediction.
- FIG. 1 - a and FIG. 1 - b FIG. 1 - a shows an example of a prediction unit division manner corresponding to intra-frame prediction.
- FIG. 1 - b shows examples of several prediction unit division manners corresponding to inter-frame prediction.
- a skip mode and a direct mode become effective tools for improving coding efficiency.
- Blocks of the two coding modes used when a bit rate is low can occupy more than a half of an entire coding sequence.
- the skip mode When the skip mode is used, a motion vector of a current picture block can be derived by using nearby motion vectors only by adding a skip mode flag to a bit stream, and a value of a reference block is directly copied according to the motion vector as a reconstructed value of the current picture block.
- an encoder may derive the motion vector of the current picture block by using the adjacent motion vectors, and directly copy the value of the reference block according to the motion vector as a predictor of the current picture block, and perform predictive coding on the current picture block by using the predictor in an encoder.
- some new coding tools are introduced to further improve video coding efficiency.
- a merge coding mode and an advanced motion vector prediction (AMVP) mode are two important inter-frame prediction tools.
- a candidate motion information set is constructed by using motion information (including a prediction direction, a motion vector, and a reference frame index) of an adjacent coded block of a current coding block, candidate motion information that enables coding efficiency to be the highest may be selected as motion information of the current coding block by means of comparison, a predictor of the current coding block is found in a reference frame, predictive coding is performed on the current coding block, and an index value indicating an adjacent coded block whose motion information is selected is written into a bitstream.
- motion information including a prediction direction, a motion vector, and a reference frame index
- a motion vector of an adjacent coded block is used as a motion vector predictor of a current coding block, a motion vector that enables coding efficiency to be the highest may be selected to predict a motion vector of the current coding block, and an index value that indicates a selected adjacent motion vector may be written into a video bitstream.
- the picture prediction method provided in the embodiments of the present application is executed by a video coding apparatus or a video decoding apparatus.
- the video coding apparatus or the video decoding apparatus may be any apparatus that needs to output or store a video, for example, a laptop computer, a tablet computer, a personal computer, a mobile phone, or a video server.
- the picture prediction method includes: determining motion vector predictors (MVP) of K pixel samples in a current picture block, where K is an integer greater than 1, the K pixel samples include a first vertex angle pixel sample in the current picture block, a motion vector predictor of the first vertex angle pixel sample is obtained based on a motion vector of a preset first spatially adjacent picture block of the current picture block, and the first spatially adjacent picture block is spatially adjacent to the first vertex angle pixel sample; and performing, based on a non-translational motion model and the motion vector predictors of the K pixel samples, pixel value prediction on the current picture block.
- MVP motion vector predictors
- FIG. 1 - c is a schematic flowchart of a picture prediction method according to an embodiment of the present application. As shown in an example of FIG. 1 - c , the picture prediction method provided in this embodiment of the present application may include the following steps.
- K is an integer greater than 1
- the K pixel samples include a first vertex angle pixel sample in the current picture block, and a motion vector predictor of the first vertex angle pixel sample is obtained based on a motion vector of a preset first spatially adjacent picture block of the current picture block.
- the first spatially adjacent picture block is spatially adjacent to the first vertex angle pixel sample.
- the motion vector predictor of the first vertex angle pixel sample is obtained based on the motion vector of the preset first spatially adjacent picture block of the current picture block.
- the motion vector predictor of the first vertex angle pixel sample may be equal to the motion vector of the preset first spatially adjacent picture block of the current picture block (that is, the motion vector of the preset first spatially adjacent picture block of the current picture block is used as the motion vector predictor of the first vertex angle pixel sample); or a motion vector predictor obtained by performing preset transformation on the motion vector of the preset first spatially adjacent picture block of the current picture block may be used as the motion vector predictor of the first vertex angle pixel sample.
- the non-translational motion model is any one of the following models: an affine transformation model, a parabolic motion model, a rotary motion model, a perspective motion model, a shearing motion model, or a scalable motion model. It may be understood that the non-translational motion model is not limited to the foregoing specific examples.
- determined motion vector predictors of K pixel samples are referenced during pixel value prediction performed, based on a non-translational motion model, on a current picture block.
- the K pixel samples include a first vertex angle pixel sample in the current picture block.
- the motion vector predictors of the K pixel samples are directly used for prediction, a motion vector predictor of the first vertex angle pixel sample in the K pixel samples is obtained based on a motion vector of a preset first spatially adjacent picture block of the current picture block, and there is a single optional case for the motion vector predictors of the K pixel samples.
- the first spatially adjacent picture block is spatially adjacent to the first vertex angle pixel sample, and the first spatially adjacent picture block of the current picture block is preset (that is, it may be considered that the first spatially adjacent picture block of the current picture block is agreed or set in advance). In this way, this is helpful in quickly determining the motion vector predictors of the K pixel samples. In addition, because redundancy is removed by making full use of a spatial correlation between picture blocks, this is helpful in further improving coding efficiency.
- the K pixel samples include at least two pixel samples in an upper left pixel sample, an upper right pixel sample, a lower left pixel sample, or a lower right pixel sample in the current picture block.
- a size of the pixel block is, for example, 2 ⁇ 2, 1 ⁇ 2, 4 ⁇ 2, 4 ⁇ 4, or another size.
- the upper left pixel sample in the current picture block is an upper left vertex of the current picture block or a pixel block that is in the current picture block and that includes an upper left vertex of the current picture block.
- the lower left pixel sample in the current picture block is a lower left vertex of the current picture block or a pixel block that is in the current picture block and that includes a lower left vertex of the current picture block.
- the upper right pixel sample in the current picture block is an upper right vertex of the current picture block or a pixel block that is in the current picture block and that includes an upper right vertex of the current picture block.
- the lower right pixel sample in the current picture block is a lower right vertex of the current picture block or a pixel block that is in the current picture block and that includes a lower right vertex of the current picture block.
- the K pixel samples may further include a center pixel sample in the current picture block.
- the center pixel sample in the current picture block may be a center pixel in the current picture block or a pixel block that is in the current picture block and that includes a center pixel in the current picture block.
- the first spatially adjacent picture block may be a picture block X1, a picture block X2, a picture block X3, or a picture block X4.
- a motion vector predictor of the upper left pixel sample in the current picture block may be obtained based on a motion vector of the picture block X1, the picture block X1 is a spatially adjacent picture block of the current picture block, and the picture block X1 is spatially adjacent to the upper left pixel sample in the current picture block.
- a motion vector predictor of the upper right pixel sample in the current picture block may be obtained based on a motion vector of the picture block X2, the picture block X2 is a spatially adjacent picture block of the current picture block, and the picture block X2 is spatially adjacent to the upper right pixel sample in the current picture block.
- a motion vector predictor of the lower left pixel sample in the current picture block may be obtained based on a motion vector of the picture block X3, the picture block X3 is a spatially adjacent picture block of the current picture block, and the picture block X3 is spatially adjacent to the lower left pixel sample in the current picture block.
- a motion vector predictor of the lower right pixel sample in the current picture block may be obtained based on a motion vector of the picture block X4, the picture block X4 is a spatially adjacent picture block of the current picture block, and the picture block X4 is spatially adjacent to the lower right pixel sample in the current picture block.
- the first vertex angle pixel sample may be the upper left pixel sample in the current picture block
- the first spatially adjacent picture block may be the picture block X1.
- the picture block X1 is a spatially adjacent picture block on the upper left of the current picture block, or the picture block X1 is a spatially adjacent picture block on the left of the current picture block, or the picture block X1 is a spatially adjacent picture block above the current picture block.
- the picture block X2 is a spatially adjacent picture block above the current picture block, or the picture block X2 is a spatially adjacent picture block on the upper right of the current picture block, or the picture block X2 is a spatially adjacent picture block on the right of the current picture block.
- an absolute value of a difference between a motion vector absolute value of the picture block X2 and a motion vector absolute value of the picture block X1 is greater than or equal to an absolute value of a difference between a motion vector absolute value of a picture block X5 and the motion vector absolute value of the picture block X1, and the picture block X5 is any one of remaining at least some spatially adjacent picture blocks, other than the picture block X2, that are of the current picture block and that are spatially adjacent to the upper right vertex. That is, a location of the picture block X2 may be determined based on a specific policy, or a location of the picture block X2 may be directly agreed.
- the picture block X3 is a spatially adjacent picture block on the left of the current picture block, or the picture block X3 is a spatially adjacent picture block on the lower left of the current picture block, or the picture block X3 is a spatially adjacent picture block below the current picture block.
- an absolute value of a difference between a motion vector absolute value of the picture block X3 and the motion vector absolute value of the picture block X1 is greater than or equal to an absolute value of a difference between a motion vector absolute value of a picture block X6 and the motion vector absolute value of the picture block X1, and the picture block X6 is any one of remaining at least some spatially adjacent picture blocks, other than the picture block X3, that are of the current picture block and that are spatially adjacent to the lower left vertex. That is, a location of the picture block X3 may be determined based on a specific policy, or a location of the picture block X3 may be directly agreed.
- the picture block X4 is a spatially adjacent picture block on the right of the current picture block, or the picture block X4 is a spatially adjacent picture block on the lower right of the current picture block, or the picture block X4 is a spatially adjacent picture block below the current picture block.
- an absolute value of a difference between a motion vector absolute value of the picture block X4 and the motion vector absolute value of the picture block X1 is greater than or equal to an absolute value of a difference between a motion vector absolute value of a picture block X7 and the motion vector absolute value of the picture block X1, and the picture block X7 is any one of remaining at least some spatially adjacent picture blocks, other than the picture block X4, that are of the current picture block and that are spatially adjacent to the lower left vertex.
- FIG. 2 shows examples of some possible locations of the picture block X1, the picture block X2, the picture block X3, and the picture block X4.
- the performing, based on a non-translational motion model and the motion vector predictors of the K pixel samples, pixel value prediction on the current picture block may specifically include: if reference frames corresponding to the motion vector predictors of the K pixel samples are a reference frame Y1, performing motion estimation on the K pixel samples by using the motion vector predictors of the K pixel samples, to obtain motion vectors of the K pixel samples; and performing pixel value prediction on the current picture block by using the non-translational motion model and the motion vectors of the K pixel samples.
- the performing, based on a non-translational motion model and the motion vector predictors of the K pixel samples, pixel value prediction on the current picture block may specifically include: when reference frames corresponding to motion vector predictors of K1 pixel samples in the K pixel samples are not a reference frame Y1, scaling the motion vector predictors of the K1 pixel samples to the reference frame Y1; performing motion estimation on the K pixel samples by using motion vector predictors of K2 pixel samples and motion vector predictors that are of the K1 pixel samples and that are scaled to the reference frame Y1, to obtain motion vectors of the K pixel samples; and performing pixel value prediction on the current picture block by using the non-translational motion model and the motion vectors of the K pixel samples, where the K2 pixel samples are remaining pixel samples other than the K1 pixel samples in the K pixel samples, and K1 and K2 are positive integers.
- the reference frame Y1 may be a reference frame corresponding to the motion vector predictor of the first vertex angle pixel sample, or the reference frame Y1 may be a reference frame corresponding to a motion vector predictor of another pixel sample in the K pixel samples.
- a pixel sample i is any pixel sample except the first vertex angle pixel sample in the K pixel samples, and if a prediction direction corresponding to a motion vector predictor of the pixel sample i is different from a prediction direction corresponding to the motion vector predictor of the first vertex angle pixel sample, a reference frame index corresponding to the motion vector predictor of the pixel sample i is 0.
- that the motion vector predictor of the pixel sample i is scaled to the reference frame Y1 may include: if reference frame indexes corresponding to the motion vector predictors of the first vertex angle pixel sample and the pixel sample i are different, the motion vector predictor of the pixel sample i is scaled to the reference frame Y1.
- the performing motion estimation on the K pixel samples by using motion vector predictors of K2 pixel samples and motion vector predictors that are of the K1 pixel samples and that are scaled to the reference frame Y1, to obtain motion vectors of the K pixel samples may include: performing motion estimation on the K pixel samples based on an iterative search algorithm (or a motion estimation algorithm of another type) by using the motion vector predictors of the K2 pixel samples and the motion vector predictors that are of the K1 pixel samples and that are scaled to the reference frame Y1, to obtain the motion vectors of the K pixel samples.
- a step of performing motion estimation on the K pixel samples may be omitted.
- a normalization step may even be omitted. That is, the following step may be omitted: when the reference frames corresponding to the motion vector predictors of the K1 pixel samples in the K pixel samples are not the reference frame Y1, the motion vector predictors of the K1 pixel samples are scaled to the reference frame Y1.
- any one of the foregoing picture prediction methods provided in this embodiment may be applied to a video coding process or a video decoding process.
- FIG. 3 - a is a schematic flowchart of another picture prediction method according to another embodiment of the present application.
- the another picture prediction method provided in the another embodiment of the present application may include the following steps.
- K 3.
- the K pixel samples include an upper left vertex, an upper right vertex, and a lower left vertex of the current picture block.
- a motion vector predictor of the upper left vertex of the current picture block is equal to a motion vector of an adjacent picture block A on the upper left of the current picture block.
- a motion vector predictor of the upper right vertex of the current picture block is equal to a motion vector of an adjacent picture block B above the current picture block, or a motion vector predictor of the upper right vertex of the current picture block may be equal to a motion vector of an adjacent picture block C on the upper right of the current picture block.
- a motion vector predictor of the lower left vertex of the current picture block may be equal to a motion vector of an adjacent picture block E on the lower left of the current picture block, or a motion vector predictor of the lower left vertex of the current picture block may be equal to a motion vector of an adjacent picture block D on the left of the current picture block.
- the motion vector predictors of the upper right vertex and the lower left vertex of the current picture block may be determined based on the following formulas:
- mv A represents the motion vector (MV) of the picture block A.
- mvp 1 represents an MVP of the upper right vertex of the current picture block
- mv B represents the MV of the picture block B
- mv C represents the MV of the picture block C
- mvp 2 represents the MVP of the upper right vertex of the current picture block
- mv D represents the MV of the picture block D
- mv E and represents the MV of the picture block E.
- a picture block that has a larger difference with a motion vector absolute value of the block A is selected from the spatially adjacent picture block B and picture block C.
- a picture block that has a larger difference with the motion vector absolute value of the block A is selected from the spatially adjacent picture block D and picture block E.
- the performing motion estimation on the K pixel samples by using the motion vector predictors of the K pixel samples, to obtain motion vectors of the K pixel samples includes: scaling the motion vector predictors of the K1 pixel samples to the reference frame Y1 (that is, performing normalization processing on the motion vector predictors of the K1 pixel samples); and performing motion estimation on the K pixel samples by using motion vector predictors of K2 pixel samples and motion vector predictors that are of the K1 pixel samples and that are scaled to the reference frame Y1, to obtain the motion vectors of the K pixel samples.
- a reference frame index of a motion vector predictor of a pixel sample i may be first revised.
- the pixel sample i is any pixel sample (for example, the lower left vertex or the upper right vertex), except the upper left vertex of the current picture block, in the K pixel samples.
- a prediction direction corresponding to the motion vector predictor of the pixel sample i is different from a prediction direction corresponding to the motion vector predictor of the upper left vertex (for example, the prediction direction corresponding to the motion vector predictor of the pixel sample i is forward prediction, and the prediction direction corresponding to the motion vector predictor of the upper left vertex is backward prediction; or the prediction direction corresponding to the motion vector predictor of the pixel sample i is backward prediction, and the prediction direction corresponding to the motion vector predictor of the upper left vertex is forward prediction), a reference frame index corresponding to the motion vector predictor of the pixel sample i is 0.
- a method for scaling a motion vector predictor to a reference frame may be as follows: It is assumed that a current coding frame index is CurPoc, and a reference frame of the picture block A (that is, a reference frame corresponding to the MVP of the upper left vertex) is a destination reference frame.
- a destination reference frame index is DesPoc
- a reference frame corresponding to the MVP of the upper right vertex or the lower left vertex is a source reference frame
- a source reference frame index is SrcPoc
- the motion vector predictor of the upper right vertex or the lower left vertex is MVP
- a motion vector predictor that is of the upper right vertex or the lower left vertex and that is obtained after scaling is MVs.
- MVs may be obtained by means of calculation by using the following formulas:
- MV S sign ⁇ ⁇ ( ScaleFactor ⁇ MVP ) ⁇ abs ⁇ ⁇ ( ⁇ MVP )
- ScaleFactor C ⁇ u ⁇ r ⁇ P ⁇ o ⁇ c - D ⁇ e ⁇ s ⁇ P ⁇ o ⁇ c C ⁇ u ⁇ r ⁇ P ⁇ o ⁇ c - S ⁇ r ⁇ c ⁇ P ⁇ o ⁇ c
- reference frames corresponding to the motion vector predictors of the K pixel samples are the reference frame Y1
- no normalization processing is performed on the motion vector predictors of the K1 pixel samples (because the motion vector predictors of the K1 pixel samples have been in a normalized state)
- motion estimation may be performed on the K pixel samples directly by using the motion vector predictors of the K pixel samples, to obtain the motion vectors of the K pixel samples.
- motion estimation is performed on the K pixel samples to obtain the motion vectors of the K pixel samples may include: motion estimation is performed on the K pixel samples based on an iterative search algorithm, to obtain the motion vectors of the K pixel samples.
- that motion estimation is performed on the K pixel samples based on each iteration in the iterative search algorithm, to obtain the motion vectors of the K pixel samples may include the following steps.
- a size of a current picture block X is S1 ⁇ S2, and coordinates of each pixel p in the current picture block X are (x, y).
- ⁇ ′ 1,x (p) and ⁇ ′ 1,y (p) respectively represent derivatives of each pixel p in the current picture block X in a direction x and a direction y.
- the derivatives are calculated by using a Sobel operator.
- Another operator for example, a Gaussian operation
- Templates of the Sobel operator are shown in examples of formula (6) and formula (7).
- sobel x [ - 1 0 1 - 2 0 2 - 1 0 1 ] formula ⁇ ⁇ ( 6 )
- sobel y [ 1 2 1 0 0 0 - 1 - 2 - 1 ] formula ⁇ ⁇ ( 7 )
- An adjacent pixel of a pixel located on the edge of the current picture block X may not exist, and ⁇ ′ 1,x (p) and ⁇ ′ 1,y (p) may be replaced with corresponding values of the adjacent pixel of the pixel located on the edge of the current picture block X.
- org[x][y] and pred[x][y] respectively represent an original pixel value and a predicated pixel value of the pixel that is in the current picture block X and whose coordinates are (x, y).
- ⁇ ⁇ H xkl [ H x , 11 H x , 12 H x , 13 H x , 21 H x , 22 H x , 23 H x , 31 H x , 32 H x , 33 ]
- ⁇ ⁇ H ykl [ H y , 11 H y , 12 H y , 13 H y , 21 H y , 22 H y , 23 H y , 31 H y , 32 H y , 33 ]
- ⁇ ⁇ g xk [ g x
- a motion vector variation ⁇ x of each of the three pixel samples in the current picture block in the direction x and a motion vector variation ⁇ y of each pixel sample in the direction y in this iteration may be obtained by means of motion estimation.
- One or more iterations are performed based on an iterative manner in the foregoing steps S1 to S4, so that the motion vectors of the K pixel samples in the current picture block can be obtained by means of motion estimation.
- subtraction may be further performed between the motion vectors that are of the K pixel samples and that are obtained by means of motion estimation and the corresponding MVPs, to obtain motion vector differences (MVD) of the K pixel samples, and the MVDs of the K pixel samples may be written into a video bitstream.
- MVPs motion vector differences
- a syntactic element affine_no_MVD_flag may be used to indicate whether to write an MVD corresponding to a pixel sample into the video bitstream. For example, if the MVDs corresponding to the foregoing K pixel samples are all 0, affine_no_MVD_flag may be set to true, and no MVD is written into the video bitstream; or if the MVDs corresponding to the foregoing K pixel samples are not all 0, affine_no_MVD_flag may be set to false, and the MVDs corresponding to the K pixel samples are written into the video bitstream.
- motion vectors (vx 0 , vy 0 ), (vx 1 , vy 1 ), and (vx 2 , vy 2 ) of three vertices whose coordinates are (0, 0) (S 1 , 0), and (0, S 2 ) are calculated by performing step 301 and step 302 .
- affine transformation formula (12) A schematic diagram of affine transformation may be shown in FIG. 3 - e.
- a motion vector of the pixel whose coordinates are (x, y) is calculated as follows:
- motion information of each pixel in the current picture block may be used for motion vector prediction for another picture block
- one piece of motion information may be stored for each 4 ⁇ 4 block, and a motion vector of each 4 ⁇ 4 block is stored as motion information of a pixel on an upper left corner or any other pixel in the 4 ⁇ 4 block.
- the affine transformation model is mainly used as an example of a non-translational motion model for description, and a mechanism for calculating, based on another non-translational motion model and the motion vectors of the K pixel samples, the motion vector of each pixel in the current picture block has a similar principle although formulas different from those used in this embodiment may be used in the mechanism. Examples are not listed one by one herein.
- a residual of the current picture block may be obtained by using an original pixel value of the current picture block and a predicted pixel value that is of the current picture block and that is obtained by means of pixel value prediction, and the residual of the current picture block may be written into the video bitstream.
- determined motion vector predictors of K pixel samples are referenced during pixel value prediction performed, based on a non-translational motion model, on a current picture block.
- the K pixel samples include a first vertex angle pixel sample in the current picture block.
- the motion vector predictors of the K pixel samples are directly used for prediction, a motion vector predictor of the first vertex angle pixel sample in the K pixel samples is obtained based on a motion vector of a preset first spatially adjacent picture block of the current picture block, and there is a single optional case for the motion vector predictors of the K pixel samples.
- prediction directions of the MVPs of the K pixel samples in the current picture block are normalized, reference frame indexes are processed, and motion vectors corresponding to the MVPs of the K pixel samples are scaled and mapped to a same reference frame according to reference frame indexes.
- a motion vector of each pixel in the current picture block may be calculated according to optimal motion vectors of the K pixel samples by using an affine transformation formula.
- FIG. 4 - a is a schematic flowchart of another picture prediction method according to another embodiment of the present application.
- the another picture prediction method provided in the another embodiment of the present application may include the following steps.
- K 3.
- the K pixel samples include an upper left vertex, an upper right vertex, and a lower left vertex of the current picture block.
- a motion vector predictor of the upper left vertex of the current picture block is equal to a motion vector of an adjacent picture block A on the upper left of the current picture block.
- a motion vector predictor of the upper right vertex of the current picture block is equal to a motion vector of an adjacent picture block B above the current picture block.
- a motion vector predictor of the lower left vertex of the current picture block may be equal to a motion vector of an adjacent picture block D on the left of the current picture block.
- the performing motion estimation on the K pixel samples by using the motion vector predictors of the K pixel samples, to obtain motion vectors of the K pixel samples includes: scaling the motion vector predictors of the K1 pixel samples to the reference frame Y1 (that is, performing normalization processing on the motion vector predictors of the K1 pixel samples); and performing motion estimation on the K pixel samples by using motion vector predictors of K2 pixel samples and motion vector predictors that are of the K1 pixel samples and that are scaled to the reference frame Y1, to obtain the motion vectors of the K pixel samples.
- a reference frame index of a motion vector predictor of a pixel sample i may be first revised.
- the pixel sample i is any pixel sample (for example, the lower left vertex or the upper right vertex), except the upper left vertex of the current picture block, in the K pixel samples.
- a prediction direction corresponding to the motion vector predictor of the pixel sample i is different from a prediction direction corresponding to the motion vector predictor of the upper left vertex (for example, the prediction direction corresponding to the motion vector predictor of the pixel sample i is forward prediction, and the prediction direction corresponding to the motion vector predictor of the upper left vertex is backward prediction; or the prediction direction corresponding to the motion vector predictor of the pixel sample i is backward prediction, and the prediction direction corresponding to the motion vector predictor of the upper left vertex is forward prediction), a reference frame index corresponding to the motion vector predictor of the pixel sample i is 0.
- a method for scaling a motion vector predictor to a reference frame may be as follows: It is assumed that a current coding frame index is CurPoc, and a reference frame of the picture block A (that is, a reference frame corresponding to the MVP of the upper left vertex) is a destination reference frame.
- a destination reference frame index is DesPoc
- a reference frame corresponding to the MVP of the upper right vertex or the lower left vertex is a source reference frame
- a source reference frame index is SrcPoc
- the motion vector predictor of the upper right vertex or the lower left vertex is MVP
- a motion vector predictor that is of the upper right vertex or the lower left vertex and that is obtained after scaling is MVs.
- MVs may be obtained by means of calculation by using the following formulas:
- MV S sign ⁇ ( ScaleFactor ⁇ MVP ) ⁇ abs ⁇ ( ScaleFactor ⁇ MVP )
- ScaleFactor CurPoc - DesPoc CurPoc - SrcPoc
- reference frames corresponding to the motion vector predictors of the K pixel samples are the reference frame Y1
- no normalization processing is performed on the motion vector predictors of the K1 pixel samples (because the motion vector predictors of the K1 pixel samples have been in a normalized state)
- motion estimation may be performed on the K pixel samples directly by using the motion vector predictors of the K pixel samples, to obtain the motion vectors of the K pixel samples.
- motion estimation is performed on the K pixel samples to obtain the motion vectors of the K pixel samples may include: motion estimation is performed on the K pixel samples based on an iterative search algorithm, to obtain the motion vectors of the K pixel samples.
- an iterative search algorithm For a specific iterative search algorithm, refer to related descriptions in the foregoing embodiment. Details are not described herein again.
- subtraction may be further performed between the motion vectors that are of the K pixel samples and that are obtained by means of motion estimation and the corresponding MVPs, to obtain MVDs of the K pixel samples, and the MVDs of the K pixel samples may be written into a video bitstream.
- a syntactic element affine_no_MVD_flag may be used to indicate whether to write an MVD corresponding to a pixel sample into the video bitstream. For example, if the MVDs corresponding to the foregoing K pixel samples are all 0, affine_no_MVD_flag may be set to true, and no MVD is written into the video bitstream; or if the MVDs corresponding to the foregoing K pixel samples are not all 0, affine_no_MVD_flag may be set to false, and the MVDs corresponding to the K pixel samples are written into the video bitstream.
- motion information of each pixel in the current picture block may be used for motion vector prediction for another picture block
- one piece of motion information may be stored for each 4 ⁇ 4 (or 8 ⁇ 8) block, and a motion vector of each 4 ⁇ 4 (or 8 ⁇ 8) block is stored as motion information of a pixel on an upper left corner or any other pixel in the 4'4 (or 8 ⁇ 8) block.
- the motion information may include a motion vector, a prediction direction corresponding to the motion vector, and a reference frame index corresponding to the motion vector.
- the affine transformation model is mainly used as an example of a non-translational motion model for description, and a mechanism for calculating, based on another non-translational motion model and the motion vectors of the K pixel samples, the motion vector of each pixel in the current picture block has a similar principle although formulas different from those used in this embodiment may be used in the mechanism. Examples are not listed one by one herein.
- a residual of the current picture block may be obtained by using an original pixel value of the current picture block and a predicted pixel value that is of the current picture block and that is obtained by means of pixel value prediction, and the residual of the current picture block may be written into the video bitstream.
- determined motion vector predictors of K pixel samples are referenced during pixel value prediction performed, based on a non-translational motion model, on a current picture block.
- the K pixel samples include a first vertex angle pixel sample in the current picture block.
- the motion vector predictors of the K pixel samples are directly used for prediction, a motion vector predictor of the first vertex angle pixel sample in the K pixel samples is obtained based on a motion vector of a preset first spatially adjacent picture block of the current picture block, and there is a single optional case for the motion vector predictors of the K pixel samples.
- prediction directions of the MVPs of the K pixel samples in the current picture block are normalized, reference frame indexes are processed, and motion vectors corresponding to the MVPs of the K pixel samples are scaled and mapped to a same reference frame according to reference frame indexes.
- a motion vector of each pixel in the current picture block may be calculated according to optimal motion vectors of the K pixel samples by using an affine transformation formula.
- FIG. 5 is a schematic flowchart of another picture prediction method according to another embodiment of the present application.
- the another picture prediction method provided in the another embodiment of the present application may include the following steps.
- K 3.
- the K pixel samples include an upper left vertex, an upper right vertex, and a lower left vertex of the current picture block.
- motion information of each pixel in the current picture block may be used for motion vector prediction for another picture block
- one piece of motion information may be stored for each 4 ⁇ 4 (or 8 ⁇ 8) block, and a motion vector of each 4 ⁇ 4 (or 8 ⁇ 8) block is stored as motion information of a pixel on an upper left corner or any other pixel in the 4 ⁇ 4 (or 8 ⁇ 8) block.
- the motion information may include a motion vector, a prediction direction corresponding to the motion vector, and a reference frame index corresponding to the motion vector.
- the affine transformation model is mainly used as an example of a non-translational motion model for description, and a mechanism for calculating, based on another non-translational motion model and the motion vectors of the K pixel samples, the motion vector of each pixel in the current picture block has a similar principle although formulas different from those used in this embodiment may be used in the mechanism. Examples are not listed one by one herein.
- a residual of the current picture block may be obtained by using an original pixel value of the current picture block and a predicted pixel value that is of the current picture block and that is obtained by means of pixel value prediction, and the residual of the current picture block may be written into a video bitstream.
- determined motion vector predictors of K pixel samples are referenced during pixel value prediction performed, based on a non-translational motion model, on a current picture block.
- the K pixel samples include a first vertex angle pixel sample in the current picture block.
- the motion vector predictors of the K pixel samples are directly used for prediction, a motion vector predictor of the first vertex angle pixel sample in the K pixel samples is obtained based on a motion vector of a preset first spatially adjacent picture block of the current picture block, and there is a single optional case for the motion vector predictors of the K pixel samples.
- prediction directions of the MVPs of the K pixel samples in the current picture block are normalized, reference frame indexes are processed, and motion vectors corresponding to the MVPs of the K pixel samples are scaled and mapped to a same reference frame according to reference frame indexes.
- the foregoing solution is helpful in resolving a problem that inter-frame prediction accuracy is low in a current video coding standard when a video includes content of a non-translational feature, for example, lens zooming, object deformation, or moving object rotation.
- FIG. 6 is a schematic flowchart of another picture prediction method according to another embodiment of the present application.
- the another picture prediction method provided in the another embodiment of the present application may include the following steps.
- K 3.
- the K pixel samples include an upper left vertex, an upper right vertex, and a lower left vertex of the current picture block.
- the performing motion estimation on the K pixel samples by using the motion vector predictors of the K pixel samples, to obtain motion vectors of the K pixel samples includes: scaling the motion vector predictors of the K1 pixel samples to the reference frame Y1 (that is, performing normalization processing on the motion vector predictors of the K1 pixel samples); and performing motion estimation on the K pixel samples by using motion vector predictors of K2 pixel samples and motion vector predictors that are of the K1 pixel samples and that are scaled to the reference frame Y1, to obtain the motion vectors of the K pixel samples.
- a reference frame index of a motion vector predictor of a pixel sample i may be first revised.
- the pixel sample i is any pixel sample (for example, the lower left vertex or the upper right vertex), except the upper left vertex of the current picture block, in the K pixel samples.
- a prediction direction corresponding to the motion vector predictor of the pixel sample i is different from a prediction direction corresponding to the motion vector predictor of the upper left vertex (for example, the prediction direction corresponding to the motion vector predictor of the pixel sample i is forward prediction, and the prediction direction corresponding to the motion vector predictor of the upper left vertex is backward prediction; or the prediction direction corresponding to the motion vector predictor of the pixel sample i is backward prediction, and the prediction direction corresponding to the motion vector predictor of the upper left vertex is forward prediction), a reference frame index corresponding to the motion vector predictor of the pixel sample i is 0.
- reference frames corresponding to the motion vector predictors of the K pixel samples are the reference frame Y1
- no normalization processing is performed on the motion vector predictors of the K1 pixel samples (because the motion vector predictors of the K1 pixel samples have been in a normalized state)
- motion estimation may be performed on the K pixel samples directly by using the motion vector predictors of the K pixel samples, to obtain the motion vectors of the K pixel samples.
- motion estimation is performed on the K pixel samples to obtain the motion vectors of the K pixel samples may include: motion estimation is performed on the K pixel samples based on an iterative search algorithm, to obtain the motion vectors of the K pixel samples.
- an iterative search algorithm For a specific iterative search algorithm, refer to related descriptions in the foregoing embodiment. Details are not described herein again.
- subtraction may be further performed between the motion vectors that are of the K pixel samples and that are obtained by means of motion estimation and the corresponding MVPs, to obtain MVDs of the K pixel samples, and the MVDs of the K pixel samples may be written into a video bitstream.
- a syntactic element affine_no_MVD_flag may be used to indicate whether to write an MVD corresponding to a pixel sample into the video bitstream. For example, if the MVDs corresponding to the foregoing K pixel samples are all 0, affine_no_MVD_flag may be set to true, and no MVD is written into the video bitstream; or if the MVDs corresponding to the foregoing K pixel samples are not all 0, affine_no_MVD_flag may be set to false, and the MVDs corresponding to the K pixel samples are written into the video bitstream.
- motion information of each pixel in the current picture block may be used for motion vector prediction for another picture block
- one piece of motion information may be stored for each 4 ⁇ 4 (or 8 ⁇ 8) block, and a motion vector of each 4 ⁇ 4 (or 8 ⁇ 8) block is stored as motion information of a pixel on an upper left corner or any other pixel in the 4 ⁇ 4 (or 8 ⁇ 8) block.
- the motion information may include a motion vector, a prediction direction corresponding to the motion vector, and a reference frame index corresponding to the motion vector.
- the affine transformation model is mainly used as an example of a non-translational motion model for description, and a mechanism for calculating, based on another non-translational motion model and the motion vectors of the K pixel samples, the motion vector of each pixel in the current picture block has a similar principle although formulas different from those used in this embodiment may be used in the mechanism. Examples are not listed one by one herein.
- step 604 Calculate a rate-distortion cost of using an affine transformation-based inter-frame prediction mode (step 601 to step 603 ) for the current prediction unit, and compare the rate-distortion cost with a calculated rate-distortion cost of using an HEVC inter-frame prediction mode.
- an optimal inter-frame prediction mode of the current picture block is set to the affine transformation-based inter-frame prediction mode.
- a syntactic element affine MC flag may be set to true to indicate that the optimal inter-frame prediction mode of the current picture block is the affine transformation-based inter-frame prediction mode.
- a syntactic element affine_MC_flag is set to false to indicate that an optimal inter-frame prediction mode of the current picture block is the HEVC inter-frame prediction mode.
- a predicted pixel value of each pixel may be obtained by using an HEVC interpolation filter; or if the motion vector of the pixel is greater than 1 ⁇ 4 pixel precision, predicted pixel values of four 1 ⁇ 4 pixels at the shortest distance from the pixel are first obtained by using an HEVC interpolation filter, and then, bilinear interpolation is performed to obtain a predicted pixel value of the pixel.
- a predicted pixel value is obtained by using the HEVC interpolation filter.
- a residual of the current picture block may be obtained by using an original pixel value of the current picture block and a predicted pixel value that is of the current picture block and that is obtained by means of pixel value prediction, and the residual of the current picture block is written into the video bitstream.
- the residual of the current prediction unit is written into the video bitstream after undergoing transformation, quantization, and entropy encoding.
- determined motion vector predictors of K pixel samples are referenced during pixel value prediction performed, based on a non-translational motion model, on a current picture block.
- the K pixel samples include a first vertex angle pixel sample in the current picture block.
- the motion vector predictors of the K pixel samples are directly used for prediction, a motion vector predictor of the first vertex angle pixel sample in the K pixel samples is obtained based on a motion vector of a preset first spatially adjacent picture block of the current picture block, and there is a single optional case for the motion vector predictors of the K pixel samples.
- prediction directions of the MVPs of the K pixel samples in the current picture block are normalized, reference frame indexes are processed, and motion vectors corresponding to the MVPs of the K pixel samples are scaled and mapped to a same reference frame according to reference frame indexes.
- a motion vector of each pixel in the current picture block may be calculated according to optimal motion vectors of the K pixel samples by using an affine transformation formula.
- FIG. 7 is a schematic flowchart of a video decoding method according to an embodiment of the present application. As shown in an example of FIG. 7 , the video decoding method provided in this embodiment of the present application may include the following steps.
- a decoder decodes a syntactic element affine_MC_flag in a video bitstream, to determine whether an affine transformation-based inter-frame prediction mode is used for a current picture block.
- the decoder decodes a syntactic element affine_inter_dir_flag in the video bitstream, to determine an inter-frame prediction direction used for the current picture block.
- the decoder determines MVPs of K pixel samples in the current picture block according to the syntactic element affine_inter_dir_flag.
- the K pixel samples in the current picture block may be K pixel samples in a current picture block that are described in any one of the foregoing embodiments.
- the decoder determines, by decoding a syntactic element affine_no_ MVD_flag that is in the video bitstream and is corresponding to the current picture block, whether MVDs corresponding to the K pixel samples in the current picture block are 0. If affine_no_MVD_flag is true, it indicates that the MVDs corresponding to the K pixel samples in the current picture block are all 0, and motion vectors corresponding to the obtained MVPs of the K pixel samples may be used as motion vectors of the K pixel samples.
- affine_no_MVD_flag is false, it indicates that the MVDs corresponding to the K pixel samples are not all 0, the MVDs corresponding to the K pixel samples are obtained by means of separate decoding, and an MVD corresponding to each pixel sample and a corresponding MVP are added to obtain a motion vector of the pixel sample.
- the decoder decodes a residual coefficient of the current picture block, and performs dequantization and inverse transformation to obtain a residual pixel in the current picture block.
- the decoder calculates, based on an affine transformation model and motion vectors of the K pixel samples, a motion vector of each pixel in the current picture block, and performs, based on the calculated motion vector of each pixel in the current picture block, pixel value prediction on the current picture block, to obtain a predicted pixel value of the current picture block.
- the affine transformation model is mainly used as an example of a non-translational motion model for description, and a mechanism for calculating, based on another non-translational motion model and the motion vectors of the K pixel samples, the motion vector of each pixel in the current picture block has a similar principle although formulas different from those used in this embodiment may be used in the mechanism. Examples are not listed one by one herein.
- the decoder reconstructs the current picture block according to the residual pixel in the current picture block and the predicted pixel value of the current picture block.
- motion information of each pixel in the current picture block may be used for motion vector prediction for another picture block
- one piece of motion information may be stored for each 4 ⁇ 4 (or 8 ⁇ 8) block, and a motion vector of each 4 ⁇ 4 (or 8 ⁇ 8) block is stored as motion information of a pixel on an upper left corner or any other pixel in the 4 ⁇ 4 (or 8 ⁇ 8) block.
- the motion information may include a motion vector, a prediction direction corresponding to the motion vector, and a reference frame index corresponding to the motion vector.
- determined motion vector predictors of K pixel samples are referenced during pixel value prediction performed, based on a non-translational motion model, on a current picture block.
- the K pixel samples include a first vertex angle pixel sample in the current picture block.
- the motion vector predictors of the K pixel samples are directly used for prediction, a motion vector predictor of the first vertex angle pixel sample in the K pixel samples is obtained based on a motion vector of a preset first spatially adjacent picture block of the current picture block, and there is a single optional case for the motion vector predictors of the K pixel samples.
- prediction directions of the MVPs of the K pixel samples in the current picture block are normalized, reference frame indexes are processed, and motion vectors corresponding to the MVPs of the K pixel samples are scaled and mapped to a same reference frame according to reference frame indexes.
- a motion vector of each pixel in the current picture block may be calculated according to optimal motion vectors of the K pixel samples by using an affine transformation formula.
- An emulation test is performed on some solutions, provided in the embodiments of the present application, on an encoder HM11.0 of the latest video coding standard HEVC. For example, a test is performed in two configurations: a low delay (LDB) and random access (RA), and performance of the solutions in the embodiments of the present application is compared with performance of the HM11.0 solution.
- LLB low delay
- RA random access
- Six test sequences whose affine features are relatively apparent are selected for testing. The test sequences are listed in Table 1, and emulation results are shown in Table 2.
- a performance counter Bjontegaard delta rate may be obtained by means of calculation by using results of four points whose quantization parameters (QPs) are 22, 27, 32, and 37, and peak signal to noise ratios (PSNR) of three components Y, U, and V are respectively calculated.
- QPs quantization parameters
- PSNR peak signal to noise ratios
- An affine transformation model is a motion model that meets the following model:
- a, b, c, d, e, and f are constant parameters
- (x, y) are original coordinates of a pixel
- (x′, y′) are coordinates obtained after affine transformation is performed on the pixel
- Vx is a motion vector in an x-axis direction
- Vy is a motion vector in a y-axis direction.
- a rotary motion model may be shown as follows:
- [ x ′ y ′ ] [ cos ⁇ ⁇ ⁇ sin ⁇ ⁇ ⁇ a - sin ⁇ ⁇ ⁇ cos ⁇ ⁇ ⁇ b ] ⁇ [ x y 1 ] .
- ⁇ is a rotation angle
- a and b are parameters.
- a perspective motion model may be shown as follows:
- a shearing motion model may be shown as follows:
- a scalable motion model may be shown as follows:
- V x ⁇ X ⁇ X + ⁇ X
- V y ⁇ X ⁇ y + ⁇ y .
- ⁇ x , ⁇ y , ⁇ x , and ⁇ y are constant parameters.
- a quadratic motion model may be shown as follows:
- an embodiment of the present application provides a picture prediction apparatus 800 , and the picture prediction apparatus 800 may include:
- a determining unit 810 configured to determine motion vector predictors of K pixel samples in a current picture block, where K is an integer greater than 1, the K pixel samples include a first vertex angle pixel sample in the current picture block, a motion vector predictor of the first vertex angle pixel sample is obtained based on a motion vector of a preset first spatially adjacent picture block of the current picture block, and the first spatially adjacent picture block is spatially adjacent to the first vertex angle pixel sample; and
- a predicting unit 820 configured to perform, based on a non-translational motion model and the motion vector predictors of the K pixel samples, pixel value prediction on the current picture block.
- the motion vector predictor of the first vertex angle pixel sample is obtained based on the motion vector of the preset first spatially adjacent picture block of the current picture block.
- the motion vector predictor of the first vertex angle pixel sample may be equal to the motion vector of the preset first spatially adjacent picture block of the current picture block (that is, the motion vector of the preset first spatially adjacent picture block of the current picture block is used as the motion vector predictor of the first vertex angle pixel sample); or a motion vector predictor obtained by performing preset transformation on the motion vector of the preset first spatially adjacent picture block of the current picture block may be used as the motion vector predictor of the first vertex angle pixel sample.
- the non-translational motion model is any one of the following models: an affine transformation model, a parabolic motion model, a rotary motion model, a perspective motion model, a shearing motion model, or a scalable motion model. It may be understood that the non-translational motion model is not limited to the foregoing specific examples.
- the K pixel samples include at least two pixel samples in an upper left pixel sample, an upper right pixel sample, a lower left pixel sample, or a lower right pixel sample in the current picture block.
- a size of the pixel block is, for example, 2 ⁇ 2, 1 ⁇ 2, 4 ⁇ 2, 4 ⁇ 4, or another size.
- the upper left pixel sample in the current picture block is an upper left vertex of the current picture block or a pixel block that is in the current picture block and that includes an upper left vertex of the current picture block.
- the lower left pixel sample in the current picture block is a lower left vertex of the current picture block or a pixel block that is in the current picture block and that includes a lower left vertex of the current picture block.
- the upper right pixel sample in the current picture block is an upper right vertex of the current picture block or a pixel block that is in the current picture block and that includes an upper right vertex of the current picture block.
- the lower right pixel sample in the current picture block is a lower right vertex of the current picture block or a pixel block that is in the current picture block and that includes a lower right vertex of the current picture block.
- the K pixel samples may further include a center pixel sample in the current picture block.
- the center pixel sample in the current picture block may be a center pixel in the current picture block or a pixel block that is in the current picture block and that includes a center pixel in the current picture block.
- the first spatially adjacent picture block may be a picture block X1, a picture block X2, a picture block X3, or a picture block X4.
- a motion vector predictor of the upper left pixel sample in the current picture block may be obtained based on a motion vector of the picture block X1, the picture block X1 is a spatially adjacent picture block of the current picture block, and the picture block X1 is spatially adjacent to the upper left pixel sample in the current picture block.
- a motion vector predictor of the upper right pixel sample in the current picture block may be obtained based on a motion vector of the picture block X2, the picture block X2 is a spatially adjacent picture block of the current picture block, and the picture block X2 is spatially adjacent to the upper right pixel sample in the current picture block.
- a motion vector predictor of the lower left pixel sample in the current picture block may be obtained based on a motion vector of the picture block X3, the picture block X3 is a spatially adjacent picture block of the current picture block, and the picture block X3 is spatially adjacent to the lower left pixel sample in the current picture block.
- a motion vector predictor of the lower right pixel sample in the current picture block may be obtained based on a motion vector of the picture block X4, the picture block X4 is a spatially adjacent picture block of the current picture block, and the picture block X4 is spatially adjacent to the lower right pixel sample in the current picture block.
- the first vertex angle pixel sample may be the upper left pixel sample in the current picture block
- the first spatially adjacent picture block may be the picture block X1.
- the picture block X1 is a spatially adjacent picture block on the upper left of the current picture block, or the picture block X1 is a spatially adjacent picture block on the left of the current picture block, or the picture block X1 is a spatially adjacent picture block above the current picture block.
- the picture block X2 is a spatially adjacent picture block above the current picture block, or the picture block X2 is a spatially adjacent picture block on the upper right of the current picture block, or the picture block X2 is a spatially adjacent picture block on the right of the current picture block.
- an absolute value of a difference between a motion vector absolute value of the picture block X2 and a motion vector absolute value of the picture block X1 is greater than or equal to an absolute value of a difference between a motion vector absolute value of a picture block X5 and the motion vector absolute value of the picture block X1, and the picture block X5 is any one of remaining at least some spatially adjacent picture blocks, other than the picture block X2, that are of the current picture block and that are spatially adjacent to the upper right vertex. That is, a location of the picture block X2 may be determined based on a specific policy, or a location of the picture block X2 may be directly agreed.
- the picture block X3 is a spatially adjacent picture block on the left of the current picture block, or the picture block X3 is a spatially adjacent picture block on the lower left of the current picture block, or the picture block X3 is a spatially adjacent picture block below the current picture block.
- an absolute value of a difference between a motion vector absolute value of the picture block X3 and the motion vector absolute value of the picture block X1 is greater than or equal to an absolute value of a difference between a motion vector absolute value of a picture block X6 and the motion vector absolute value of the picture block X1, and the picture block X6 is any one of remaining at least some spatially adjacent picture blocks, other than the picture block X3, that are of the current picture block and that are spatially adjacent to the lower left vertex. That is, a location of the picture block X3 may be determined based on a specific policy, or a location of the picture block X3 may be directly agreed.
- the picture block X4 is a spatially adjacent picture block on the right of the current picture block, or the picture block X4 is a spatially adjacent picture block on the lower right of the current picture block, or the picture block X4 is a spatially adjacent picture block below the current picture block.
- an absolute value of a difference between a motion vector absolute value of the picture block X4 and the motion vector absolute value of the picture block X1 is greater than or equal to an absolute value of a difference between a motion vector absolute value of a picture block X7 and the motion vector absolute value of the picture block X1, and the picture block X7 is any one of remaining at least some spatially adjacent picture blocks, other than the picture block X4, that are of the current picture block and that are spatially adjacent to the lower left vertex.
- FIG. 2 shows examples of some possible locations of the picture block X1, the picture block X2, the picture block X3, and the picture block X4.
- the predicting unit 820 may be configured to: if reference frames corresponding to the motion vector predictors of the K pixel samples are a reference frame Y1, perform motion estimation on the K pixel samples by using the motion vector predictors of the K pixel samples, to obtain motion vectors of the K pixel samples; and perform pixel value prediction on the current picture block by using the non-translational motion model and the motion vectors of the K pixel samples.
- the predicting unit 820 may be specifically configured to: when reference frames corresponding to motion vector predictors of K1 pixel samples in the K pixel samples are not a reference frame Y1, scale the motion vector predictors of the K1 pixel samples to the reference frame Y1, perform motion estimation on the K pixel samples by using motion vector predictors of K2 pixel samples and motion vector predictors that are of the K1 pixel samples and that are scaled to the reference frame Y1, to obtain motion vectors of the K pixel samples, and perform pixel value prediction on the current picture block by using the non-translational motion model and the motion vectors of the K pixel samples, where the K2 pixel samples are remaining pixel samples other than the K1 pixel samples in the K pixel samples, and K1 and K2 are positive integers.
- the reference frame Y1 may be a reference frame corresponding to the motion vector predictor of the first vertex angle pixel sample, or the reference frame Y1 may be a reference frame corresponding to a motion vector predictor of another pixel sample in the K pixel samples.
- a pixel sample i is any pixel sample except the first vertex angle pixel sample in the K pixel samples, and if a prediction direction corresponding to a motion vector predictor of the pixel sample i is different from a prediction direction corresponding to the motion vector predictor of the first vertex angle pixel sample, a reference frame index corresponding to the motion vector predictor of the pixel sample i is 0.
- the predicting unit 820 may scale the motion vector predictor of the pixel sample i to the reference frame Y1.
- the predicting unit 820 may be specifically configured to perform motion estimation on the K pixel samples based on an iterative search algorithm (or a motion estimation algorithm of another type) by using the motion vector predictors of the K2 pixel samples and the motion vector predictors that are of the K1 pixel samples and that are scaled to the reference frame Y1, to obtain the motion vectors of the K pixel samples.
- the performing motion estimation on the K pixel samples by using motion vector predictors of K2 pixel samples and motion vector predictors that are of the K1 pixel samples and that are scaled to the reference frame Y1, to obtain motion vectors of the K pixel samples includes: performing motion estimation on the K pixel samples based on an iterative search algorithm by using the motion vector predictors of the K2 pixel samples and the motion vector predictors that are of the K1 pixel samples and that are scaled to the reference frame Y1, to obtain the motion vectors of the K pixel samples.
- the picture prediction apparatus 800 is applied to a video coding apparatus or the picture prediction apparatus 800 is applied to a video decoding apparatus.
- the picture prediction apparatus 800 may be any apparatus that needs to output and play a video, for example, a laptop computer, a tablet computer, a personal computer, or a mobile phone.
- determined motion vector predictors of K pixel samples are referenced during pixel value prediction performed, based on a non-translational motion model, on a current picture block.
- the K pixel samples include a first vertex angle pixel sample in the current picture block.
- the motion vector predictors of the K pixel samples are directly used for prediction, a motion vector predictor of the first vertex angle pixel sample in the K pixel samples is obtained based on a motion vector of a preset first spatially adjacent picture block of the current picture block, and there is a single optional case for the motion vector predictors of the K pixel samples.
- the first spatially adjacent picture block is spatially adjacent to the first vertex angle pixel sample, and the first spatially adjacent picture block of the current picture block is preset (that is, it may be considered that the first spatially adjacent picture block of the current picture block is agreed or set in advance). In this way, this is helpful in quickly determining the motion vector predictors of the K pixel samples. In addition, because redundancy is removed by making full use of a spatial correlation between picture blocks, this is helpful in further improving coding efficiency.
- FIG. 9 is a schematic diagram of a picture prediction apparatus 900 according to an embodiment of the present application.
- the picture prediction apparatus 900 may include at least one bus 901 , at least one processor 902 connected to the bus 901 , and at least one memory 903 connected to the bus 901 .
- the processor 902 invokes, by using the bus 901 , code stored in the memory 903 , so as to determine motion vector predictors of K pixel samples in a current picture block, where K is an integer greater than 1, the K pixel samples include a first vertex angle pixel sample in the current picture block, a motion vector predictor of the first vertex angle pixel sample is obtained based on a motion vector of a preset first spatially adjacent picture block of the current picture block, and the first spatially adjacent picture block is spatially adjacent to the first vertex angle pixel sample; and perform, based on a non-translational motion model and the motion vector predictors of the K pixel samples, pixel value prediction on the current picture block.
- the motion vector predictor of the first vertex angle pixel sample is obtained based on the motion vector of the preset first spatially adjacent picture block of the current picture block.
- the motion vector predictor of the first vertex angle pixel sample may be equal to the motion vector of the preset first spatially adjacent picture block of the current picture block (that is, the motion vector of the preset first spatially adjacent picture block of the current picture block is used as the motion vector predictor of the first vertex angle pixel sample); or a motion vector predictor obtained by performing preset transformation on the motion vector of the preset first spatially adjacent picture block of the current picture block may be used as the motion vector predictor of the first vertex angle pixel sample.
- the non-translational motion model is any one of the following models: an affine transformation model, a parabolic motion model, a rotary motion model, a perspective motion model, a shearing motion model, or a scalable motion model. It may be understood that the non-translational motion model is not limited to the foregoing specific examples.
- the K pixel samples include at least two pixel samples in an upper left pixel sample, an upper right pixel sample, a lower left pixel sample, or a lower right pixel sample in the current picture block.
- a size of the pixel block is, for example, 2 ⁇ 2, 1 ⁇ 2, 4 ⁇ 2, 4 ⁇ 4, or another size.
- the upper left pixel sample in the current picture block is an upper left vertex of the current picture block or a pixel block that is in the current picture block and that includes an upper left vertex of the current picture block.
- the lower left pixel sample in the current picture block is a lower left vertex of the current picture block or a pixel block that is in the current picture block and that includes a lower left vertex of the current picture block.
- the upper right pixel sample in the current picture block is an upper right vertex of the current picture block or a pixel block that is in the current picture block and that includes an upper right vertex of the current picture block.
- the lower right pixel sample in the current picture block is a lower right vertex of the current picture block or a pixel block that is in the current picture block and that includes a lower right vertex of the current picture block.
- the K pixel samples may further include a center pixel sample in the current picture block.
- the center pixel sample in the current picture block may be a center pixel in the current picture block or a pixel block that is in the current picture block and that includes a center pixel in the current picture block.
- the first spatially adjacent picture block may be a picture block X1, a picture block X2, a picture block X3, or a picture block X4.
- a motion vector predictor of the upper left pixel sample in the current picture block may be obtained based on a motion vector of the picture block X1, the picture block X1 is a spatially adjacent picture block of the current picture block, and the picture block X1 is spatially adjacent to the upper left pixel sample in the current picture block.
- a motion vector predictor of the upper right pixel sample in the current picture block may be obtained based on a motion vector of the picture block X2, the picture block X2 is a spatially adjacent picture block of the current picture block, and the picture block X2 is spatially adjacent to the upper right pixel sample in the current picture block.
- a motion vector predictor of the lower left pixel sample in the current picture block may be obtained based on a motion vector of the picture block X3, the picture block X3 is a spatially adjacent picture block of the current picture block, and the picture block X3 is spatially adjacent to the lower left pixel sample in the current picture block.
- a motion vector predictor of the lower right pixel sample in the current picture block may be obtained based on a motion vector of the picture block X4, the picture block X4 is a spatially adjacent picture block of the current picture block, and the picture block X4 is spatially adjacent to the lower right pixel sample in the current picture block.
- the first vertex angle pixel sample may be the upper left pixel sample in the current picture block
- the first spatially adjacent picture block may be the picture block X1.
- the picture block X1 is a spatially adjacent picture block on the upper left of the current picture block, or the picture block X1 is a spatially adjacent picture block on the left of the current picture block, or the picture block X1 is a spatially adjacent picture block above the current picture block.
- the picture block X2 is a spatially adjacent picture block above the current picture block, or the picture block X2 is a spatially adjacent picture block on the upper right of the current picture block, or the picture block X2 is a spatially adjacent picture block on the right of the current picture block.
- an absolute value of a difference between a motion vector absolute value of the picture block X2 and a motion vector absolute value of the picture block X1 is greater than or equal to an absolute value of a difference between a motion vector absolute value of a picture block X5 and the motion vector absolute value of the picture block X1, and the picture block X5 is any one of remaining at least some spatially adjacent picture blocks, other than the picture block X2, that are of the current picture block and that are spatially adjacent to the upper right vertex. That is, a location of the picture block X2 may be determined based on a specific policy, or a location of the picture block X2 may be directly agreed.
- the picture block X3 is a spatially adjacent picture block on the left of the current picture block, or the picture block X3 is a spatially adjacent picture block on the lower left of the current picture block, or the picture block X3 is a spatially adjacent picture block below the current picture block.
- an absolute value of a difference between a motion vector absolute value of the picture block X3 and the motion vector absolute value of the picture block X1 is greater than or equal to an absolute value of a difference between a motion vector absolute value of a picture block X6 and the motion vector absolute value of the picture block X1, and the picture block X6 is any one of remaining at least some spatially adjacent picture blocks, other than the picture block X3, that are of the current picture block and that are spatially adjacent to the lower left vertex. That is, a location of the picture block X3 may be determined based on a specific policy, or a location of the picture block X3 may be directly agreed.
- the picture block X4 is a spatially adjacent picture block on the right of the current picture block, or the picture block X4 is a spatially adjacent picture block on the lower right of the current picture block, or the picture block X4 is a spatially adjacent picture block below the current picture block.
- an absolute value of a difference between a motion vector absolute value of the picture block X4 and the motion vector absolute value of the picture block X1 is greater than or equal to an absolute value of a difference between a motion vector absolute value of a picture block X7 and the motion vector absolute value of the picture block X1, and the picture block X7 is any one of remaining at least some spatially adjacent picture blocks, other than the picture block X4, that are of the current picture block and that are spatially adjacent to the lower left vertex.
- the processor 902 may be configured to: if reference frames corresponding to the motion vector predictors of the K pixel samples are a reference frame Y1, perform motion estimation on the K pixel samples by using the motion vector predictors of the K pixel samples, to obtain motion vectors of the K pixel samples; and perform pixel value prediction on the current picture block by using the non-translational motion model and the motion vectors of the K pixel samples.
- the processor 902 may be configured to: when reference frames corresponding to motion vector predictors of K1 pixel samples in the K pixel samples are not a reference frame Y1, scale the motion vector predictors of the K1 pixel samples to the reference frame Y1, perform motion estimation on the K pixel samples by using motion vector predictors of K2 pixel samples and motion vector predictors that are of the K1 pixel samples and that are scaled to the reference frame Y1, to obtain motion vectors of the K pixel samples, and perform pixel value prediction on the current picture block by using the non-translational motion model and the motion vectors of the K pixel samples, where the K2 pixel samples are remaining pixel samples other than the K1 pixel samples in the K pixel samples, and K1 and K2 are positive integers.
- the reference frame Y1 may be a reference frame corresponding to the motion vector predictor of the first vertex angle pixel sample, or the reference frame Y1 may be a reference frame corresponding to a motion vector predictor of another pixel sample in the K pixel samples.
- a pixel sample i is any pixel sample except the first vertex angle pixel sample in the K pixel samples, and if a prediction direction corresponding to a motion vector predictor of the pixel sample i is different from a prediction direction corresponding to the motion vector predictor of the first vertex angle pixel sample, a reference frame index corresponding to the motion vector predictor of the pixel sample i is 0.
- the processor 902 may be configured to scale the motion vector predictor of the pixel sample i to the reference frame Y1.
- the processor 902 may be configured to perform motion estimation on the K pixel samples based on an iterative search algorithm (or a motion estimation algorithm of another type) by using the motion vector predictors of the K2 pixel samples and the motion vector predictors that are of the K1 pixel samples and that are scaled to the reference frame Y1, to obtain the motion vectors of the K pixel samples.
- the picture prediction apparatus 900 may be any apparatus that needs to output and play a video, for example, a laptop computer, a tablet computer, a personal computer, or a mobile phone.
- determined motion vector predictors of K pixel samples are referenced during pixel value prediction performed, based on a non-translational motion model, on a current picture block.
- the K pixel samples include a first vertex angle pixel sample in the current picture block.
- the motion vector predictors of the K pixel samples are directly used for prediction, a motion vector predictor of the first vertex angle pixel sample in the K pixel samples is obtained based on a motion vector of a preset first spatially adjacent picture block of the current picture block, and there is a single optional case for the motion vector predictors of the K pixel samples.
- the first spatially adjacent picture block is spatially adjacent to the first vertex angle pixel sample, and the first spatially adjacent picture block of the current picture block is preset (that is, it may be considered that the first spatially adjacent picture block of the current picture block is agreed or set in advance). In this way, this is helpful in quickly determining the motion vector predictors of the K pixel samples. In addition, because redundancy is removed by making full use of a spatial correlation between picture blocks, this is helpful in further improving coding efficiency.
- FIG. 10 is a structural block diagram of a picture prediction apparatus 1000 according to another embodiment of the present application.
- the picture prediction apparatus 1000 may include at least one processor 1001 , at least one network interface 1004 or other user interface 1003 , a memory 1005 , and at least one communications bus 1002 .
- the communications bus 1002 is configured to implement a connection and communication between these components.
- the picture prediction apparatus 1000 may optionally include the user interface 1003 that includes a display (for example, a touchscreen, a liquid crystal display (LCD), a holographic imaging device, a cathode ray tube (CRT), or a Projector), a click device (for example, a mouse, a trackball, a touchpad, or a touchscreen), a camera and/or a pickup apparatus, and the like.
- a display for example, a touchscreen, a liquid crystal display (LCD), a holographic imaging device, a cathode ray tube (CRT), or a Projector
- a click device for example, a mouse, a trackball, a touchpad, or a touchscreen
- a camera and/or a pickup apparatus and the like.
- the memory 1005 may include a read-only memory and a random access memory, and provide an instruction and data for the processor 1001 .
- a part of the memory 1005 may further include a nonvolatile random access memory (NVRAM).
- NVRAM nonvolatile random access memory
- the memory 1005 stores the following element: an executable module or a data structure, or a subset thereof, or an extension set thereof:
- an operating system 10051 including various system programs, and used to implement various basic services and process hardware-based tasks;
- an application program module 10052 including various application programs, and used to implement various application services.
- the processor 1001 is configured to: determine motion vector predictors of K pixel samples in a current picture block, where K is an integer greater than 1, the K pixel samples include a first vertex angle pixel sample in the current picture block, a motion vector predictor of the first vertex angle pixel sample is obtained based on a motion vector of a preset first spatially adjacent picture block of the current picture block, and the first spatially adjacent picture block is spatially adjacent to the first vertex angle pixel sample; and perform, based on a non-translational motion model and the motion vector predictors of the K pixel samples, pixel value prediction on the current picture block.
- the motion vector predictor of the first vertex angle pixel sample is obtained based on the motion vector of the preset first spatially adjacent picture block of the current picture block.
- the motion vector predictor of the first vertex angle pixel sample may be equal to the motion vector of the preset first spatially adjacent picture block of the current picture block (that is, the motion vector of the preset first spatially adjacent picture block of the current picture block is used as the motion vector predictor of the first vertex angle pixel sample); or a motion vector predictor obtained by performing preset transformation on the motion vector of the preset first spatially adjacent picture block of the current picture block may be used as the motion vector predictor of the first vertex angle pixel sample.
- the non-translational motion model is any one of the following models: an affine transformation model, a parabolic motion model, a rotary motion model, a perspective motion model, a shearing motion model, or a scalable motion model. It may be understood that the non-translational motion model is not limited to the foregoing specific examples.
- the K pixel samples include at least two pixel samples in an upper left pixel sample, an upper right pixel sample, a lower left pixel sample, or a lower right pixel sample in the current picture block.
- a size of the pixel block is, for example, 2 ⁇ 2, 1 ⁇ 2, 4 ⁇ 2, 4 ⁇ 4, or another size.
- the upper left pixel sample in the current picture block is an upper left vertex of the current picture block or a pixel block that is in the current picture block and that includes an upper left vertex of the current picture block.
- the lower left pixel sample in the current picture block is a lower left vertex of the current picture block or a pixel block that is in the current picture block and that includes a lower left vertex of the current picture block.
- the upper right pixel sample in the current picture block is an upper right vertex of the current picture block or a pixel block that is in the current picture block and that includes an upper right vertex of the current picture block.
- the lower right pixel sample in the current picture block is a lower right vertex of the current picture block or a pixel block that is in the current picture block and that includes a lower right vertex of the current picture block.
- the K pixel samples may further include a center pixel sample in the current picture block.
- the center pixel sample in the current picture block may be a center pixel in the current picture block or a pixel block that is in the current picture block and that includes a center pixel in the current picture block.
- the first spatially adjacent picture block may be a picture block X1, a picture block X2, a picture block X3, or a picture block X4.
- a motion vector predictor of the upper left pixel sample in the current picture block may be obtained based on a motion vector of the picture block X1, the picture block X1 is a spatially adjacent picture block of the current picture block, and the picture block X1 is spatially adjacent to the upper left pixel sample in the current picture block.
- a motion vector predictor of the upper right pixel sample in the current picture block may be obtained based on a motion vector of the picture block X2, the picture block X2 is a spatially adjacent picture block of the current picture block, and the picture block X2 is spatially adjacent to the upper right pixel sample in the current picture block.
- a motion vector predictor of the lower left pixel sample in the current picture block may be obtained based on a motion vector of the picture block X3, the picture block X3 is a spatially adjacent picture block of the current picture block, and the picture block X3 is spatially adjacent to the lower left pixel sample in the current picture block.
- a motion vector predictor of the lower right pixel sample in the current picture block may be obtained based on a motion vector of the picture block X4, the picture block X4 is a spatially adjacent picture block of the current picture block, and the picture block X4 is spatially adjacent to the lower right pixel sample in the current picture block.
- the first vertex angle pixel sample may be the upper left pixel sample in the current picture block
- the first spatially adjacent picture block may be the picture block X1.
- the picture block X1 is a spatially adjacent picture block on the upper left of the current picture block, or the picture block X1 is a spatially adjacent picture block on the left of the current picture block, or the picture block X1 is a spatially adjacent picture block above the current picture block.
- the picture block X2 is a spatially adjacent picture block above the current picture block, or the picture block X2 is a spatially adjacent picture block on the upper right of the current picture block, or the picture block X2 is a spatially adjacent picture block on the right of the current picture block.
- an absolute value of a difference between a motion vector absolute value of the picture block X2 and a motion vector absolute value of the picture block X1 is greater than or equal to an absolute value of a difference between a motion vector absolute value of a picture block X5 and the motion vector absolute value of the picture block X1, and the picture block X5 is any one of remaining at least some spatially adjacent picture blocks, other than the picture block X2, that are of the current picture block and that are spatially adjacent to the upper right vertex. That is, a location of the picture block X2 may be determined based on a specific policy, or a location of the picture block X2 may be directly agreed.
- the picture block X3 is a spatially adjacent picture block on the left of the current picture block, or the picture block X3 is a spatially adjacent picture block on the lower left of the current picture block, or the picture block X3 is a spatially adjacent picture block below the current picture block.
- an absolute value of a difference between a motion vector absolute value of the picture block X3 and the motion vector absolute value of the picture block X1 is greater than or equal to an absolute value of a difference between a motion vector absolute value of a picture block X6 and the motion vector absolute value of the picture block X1, and the picture block X6 is any one of remaining at least some spatially adjacent picture blocks, other than the picture block X3, that are of the current picture block and that are spatially adjacent to the lower left vertex. That is, a location of the picture block X3 may be determined based on a specific policy, or a location of the picture block X3 may be directly agreed.
- the picture block X4 is a spatially adjacent picture block on the right of the current picture block, or the picture block X4 is a spatially adjacent picture block on the lower right of the current picture block, or the picture block X4 is a spatially adjacent picture block below the current picture block.
- an absolute value of a difference between a motion vector absolute value of the picture block X4 and the motion vector absolute value of the picture block X1 is greater than or equal to an absolute value of a difference between a motion vector absolute value of a picture block X7 and the motion vector absolute value of the picture block X1, and the picture block X7 is any one of remaining at least some spatially adjacent picture blocks, other than the picture block X4, that are of the current picture block and that are spatially adjacent to the lower left vertex.
- the processor 1001 may be configured to: if reference frames corresponding to the motion vector predictors of the K pixel samples are a reference frame Y1, perform motion estimation on the K pixel samples by using the motion vector predictors of the K pixel samples, to obtain motion vectors of the K pixel samples; and perform pixel value prediction on the current picture block by using the non-translational motion model and the motion vectors of the K pixel samples.
- the processor 1001 may be configured to: when reference frames corresponding to motion vector predictors of K1 pixel samples in the K pixel samples are not a reference frame Y1, scale the motion vector predictors of the K1 pixel samples to the reference frame Y1, perform motion estimation on the K pixel samples by using motion vector predictors of K2 pixel samples and motion vector predictors that are of the K1 pixel samples and that are scaled to the reference frame Y1, to obtain motion vectors of the K pixel samples, and perform pixel value prediction on the current picture block by using the non-translational motion model and the motion vectors of the K pixel samples, where the K2 pixel samples are remaining pixel samples other than the K1 pixel samples in the K pixel samples, and K1 and K2 are positive integers.
- the reference frame Y1 may be a reference frame corresponding to the motion vector predictor of the first vertex angle pixel sample, or the reference frame Y1 may be a reference frame corresponding to a motion vector predictor of another pixel sample in the K pixel samples.
- a pixel sample i is any pixel sample except the first vertex angle pixel sample in the K pixel samples, and if a prediction direction corresponding to a motion vector predictor of the pixel sample i is different from a prediction direction corresponding to the motion vector predictor of the first vertex angle pixel sample, a reference frame index corresponding to the motion vector predictor of the pixel sample i is 0.
- the processor 1001 may be configured to scale the motion vector predictor of the pixel sample i to the reference frame Y1.
- the processor 1001 may be configured to perform motion estimation on the K pixel samples based on an iterative search algorithm (or a motion estimation algorithm of another type) by using the motion vector predictors of the K2 pixel samples and the motion vector predictors that are of the K1 pixel samples and that are scaled to the reference frame Y1, to obtain the motion vectors of the K pixel samples.
- an iterative search algorithm or a motion estimation algorithm of another type
- the picture prediction apparatus 1000 may be any apparatus that needs to output and play a video, for example, a laptop computer, a tablet computer, a personal computer, or a mobile phone.
- determined motion vector predictors of K pixel samples are referenced during pixel value prediction performed, based on a non-translational motion model, on a current picture block.
- the K pixel samples include a first vertex angle pixel sample in the current picture block.
- the motion vector predictors of the K pixel samples are directly used for prediction, a motion vector predictor of the first vertex angle pixel sample in the K pixel samples is obtained based on a motion vector of a preset first spatially adjacent picture block of the current picture block, and there is a single optional case for the motion vector predictors of the K pixel samples.
- the first spatially adjacent picture block is spatially adjacent to the first vertex angle pixel sample, and the first spatially adjacent picture block of the current picture block is preset (that is, it may be considered that the first spatially adjacent picture block of the current picture block is agreed or set in advance). In this way, this is helpful in quickly determining the motion vector predictors of the K pixel samples. In addition, because redundancy is removed by making full use of a spatial correlation between picture blocks, this is helpful in further improving coding efficiency.
- the disclosed apparatus may be implemented in other manners.
- the described apparatus embodiment is merely an example.
- the unit division is merely logical function division and may be other division in actual implementation.
- a plurality of units or components may be combined or integrated into another system, or some features may be ignored or not performed.
- the displayed or discussed mutual couplings or direct couplings or communication connections may be implemented through some interfaces.
- the indirect couplings or communication connections between the apparatuses or units may be implemented in electronic or other forms.
- the units described as separate parts may or may not be physically separate, and parts displayed as units may or may not be physical units, may be located in one position, or may be distributed on a plurality of network units. Some or all of the units may be selected according to actual needs to achieve the objectives of the solutions in the embodiments.
- functional units in the embodiments of the present application may be integrated into one processing unit, or each of the units may exist alone physically, or two or more units are integrated into one unit.
- the integrated unit may be implemented in a form of hardware, or may be implemented in a form of a software functional unit.
- the integrated unit When the foregoing integrated unit is implemented in the form of a software functional unit and sold or used as an independent product, the integrated unit may be stored in a computer-readable storage medium.
- the software product is stored in a storage medium and includes instructions for instructing a computer device (which may be a personal computer, a server, or a network device) to perform all or some of the steps of the methods described in the embodiments of the present application.
- the foregoing storage medium includes: any medium that can store program code, such as a Universal Serial Bus (USB) flash drive, a read-only memory (ROM), a random access memory (RAM), a removable hard disk, a magnetic disk, or an optical disc.
- USB Universal Serial Bus
- ROM read-only memory
- RAM random access memory
- removable hard disk a magnetic disk, or an optical disc.
Landscapes
- Engineering & Computer Science (AREA)
- Multimedia (AREA)
- Signal Processing (AREA)
- Compression Or Coding Systems Of Tv Signals (AREA)
Abstract
A picture prediction method and a related apparatus are disclosed. The picture prediction method includes: determining motion vector predictors of K pixel samples in a current picture block, where K is an integer greater than 1, the K pixel samples include a first vertex angle pixel sample in the current picture block, a motion vector predictor of the first vertex angle pixel sample is obtained based on a motion vector of a preset first spatially adjacent picture block of the current picture block, and the first spatially adjacent picture block is spatially adjacent to the first vertex angle pixel sample; and performing, based on a non-translational motion model and the motion vector predictors of the K pixel samples, pixel value prediction on the current picture block. Solutions in the embodiments of the present application are helpful in reducing calculation complexity of picture prediction based on a non-translational motion model.
Description
- This application is a continuation of U.S. patent application Ser. No. 15/461,274, filed on Mar. 16, 2017, which is a continuation of International Application No. PCT/CN2015/076659, filed on Apr. 15, 2015. The International Application claims priority to Chinese Patent Application No. 201410526608.8, filed on Sep. 30, 2014. All of the aforementioned patent applications are hereby incorporated by reference in their entireties.
- The subject matter and the claimed application were made by or on the behalf of Huawei Technologies Co., Ltd., of Shenzhen, Guangdong Province, P.R. China and University of Science and Technology of China, of Baohe District, Hefei City, Anhui Province, China, under a joint research agreement titled “Research and Development of Next Generation Video Coding Standards and Technologies.” The joint research agreement was in effect on or before the claimed application was made, and the claimed application was made as a result of activities undertaken within the scope of the joint research agreement.
- The present application relates to the field of picture processing technologies, and specifically, to a picture prediction method and a related apparatus.
- With development of photoelectric acquisition technologies and continuous increase of requirements for high-definition digital videos, an amount of video data is increasingly large. Due to limited heterogeneous transmission bandwidths and diversified video applications, higher requirements are continuously imposed on video coding efficiency. A task of developing a High Efficiency Video Coding (HEVC) standard is initiated according to the requirements.
- A basic principle of video compression coding is to use correlation between a space domain, a time domain, and a code word to remove redundancy as much as possible. Currently, a prevalent practice is to use a block-based hybrid video coding framework to implement video compression coding by performing steps of prediction (including intra-frame prediction and inter-frame prediction), transform, quantization, entropy coding, and the like. This coding framework shows a high viability, and therefore, HEVC still uses this block-based hybrid video coding framework.
- In various video coding/decoding solutions, motion estimation or motion compensation is a key technology that affects coding/decoding efficiency. In various conventional video coding/decoding solutions, it is assumed that motion of an object always meets a translational motion, and that motion of every part of the entire object is the same. Basically, all conventional motion estimation or motion compensation algorithms are block motion compensation algorithms that are established based on a translational motion model. However, motion in the real world is diversified, and irregular motion such as scaling up/down, rotation, or parabolic motion is ubiquitous. Since the ninth decade of the last century, video coding experts have realized universality of irregular motion, and wished to introduce an irregular motion model (a non-translational motion model such as an affine motion model, a rotational motion model, or a scaling motion model) to improve video coding efficiency. However, computational complexity of conventional picture prediction performed based on a non-translational motion model is generally quite high.
- Embodiments of the present application provide a picture prediction method and a related apparatus, so as to reduce calculation complexity of picture prediction based on a non-translational motion model.
- A first aspect of the present application provides a picture prediction method, including:
- determining motion vector predictors of K pixel samples in a current picture block, where K is an integer greater than 1, the K pixel samples include a first vertex angle pixel sample in the current picture block, a motion vector predictor of the first vertex angle pixel sample is obtained based on a motion vector of a preset first spatially adjacent picture block of the current picture block, and the first spatially adjacent picture block is spatially adjacent to the first vertex angle pixel sample; and performing, based on a non-translational motion model and the motion vector predictors of the K pixel samples, pixel value prediction on the current picture block.
- With reference to the first aspect, in a first possible implementation manner of the first aspect, the K pixel samples include at least two pixel samples in an upper left pixel sample, an upper right pixel sample, or a lower left pixel sample in the current picture block; and
- the upper left pixel sample in the current picture block is an upper left vertex of the current picture block or a pixel block that is in the current picture block and that includes an upper left vertex of the current picture block, the lower left pixel sample in the current picture block is a lower left vertex of the current picture block or a pixel block that is in the current picture block and that includes a lower left vertex of the current picture block, and the upper right pixel sample in the current picture block is an upper right vertex of the current picture block or a pixel block that is in the current picture block and that includes an upper right vertex of the current picture block.
- With reference to the first possible implementation manner of the first aspect, in a second possible implementation manner of the first aspect, the first spatially adjacent picture block is a picture block X1, a picture block X2, or a picture block X3;
- a motion vector predictor of the upper left pixel sample in the current picture block is obtained based on a motion vector of the picture block X1, the picture block X1 is a spatially adjacent picture block of the current picture block, and the picture block X1 is spatially adjacent to the upper left pixel sample in the current picture block;
- a motion vector predictor of the upper right pixel sample in the current picture block is obtained based on a motion vector of the picture block X2, the picture block X2 is a spatially adjacent picture block of the current picture block, and the picture block X2 is spatially adjacent to the upper right pixel sample in the current picture block; and
- a motion vector predictor of the lower left pixel sample in the current picture block is obtained based on a motion vector of the picture block X3, the picture block X3 is a spatially adjacent picture block of the current picture block, and the picture block X3 is spatially adjacent to the lower left pixel sample in the current picture block.
- With reference to the second possible implementation manner of the first aspect, in a third possible implementation manner of the first aspect,
- the first vertex angle pixel sample is the upper left pixel sample in the current picture block, the first spatially adjacent picture block is the picture block X1, and the picture block X1 is a spatially adjacent picture block on the upper left of the current picture block.
- With reference to the third possible implementation manner of the first aspect, in a fourth possible implementation manner of the first aspect,
- the picture block X2 is a spatially adjacent picture block above the current picture block; or
- an absolute value of a difference between a motion vector absolute value of the picture block X2 and a motion vector absolute value of the picture block X1 is greater than or equal to an absolute value of a difference between a motion vector absolute value of a picture block X5 and the motion vector absolute value of the picture block X1, and the picture block X5 is any one of remaining at least some spatially adjacent picture blocks, other than the picture block X2, that are of the current picture block and that are spatially adjacent to the upper right vertex.
- With reference to the third possible implementation manner of the first aspect or the fourth possible implementation manner of the first aspect, in a fifth possible implementation manner of the first aspect,
- the picture block X3 is a spatially adjacent picture block on the left of the current picture block; or
- an absolute value of a difference between a motion vector absolute value of the picture block X3 and the motion vector absolute value of the picture block X1 is greater than or equal to an absolute value of a difference between a motion vector absolute value of a picture block X6 and the motion vector absolute value of the picture block X1, and the picture block X6 is any one of remaining at least some spatially adjacent picture blocks, other than the picture block X3, that are of the current picture block and that are spatially adjacent to the lower left vertex.
- With reference to any one of the first aspect, or the first to the fifth possible implementation manners of the first aspect, in a sixth possible implementation manner of the first aspect, the performing, based on a non-translational motion model and the motion vector predictors of the K pixel samples, pixel value prediction on the current picture block includes:
- when reference frames corresponding to motion vector predictors of K1 pixel samples in the K pixel samples are not a reference frame Y1, scaling the motion vector predictors of the K1 pixel samples to the reference frame Y1, performing motion estimation on the K pixel samples by using motion vector predictors of K2 pixel samples and motion vector predictors that are of the K1 pixel samples and that are scaled to the reference frame Y1, to obtain motion vectors of the K pixel samples, and performing pixel value prediction on the current picture block by using the non-translational motion model and the motion vectors of the K pixel samples, where the K2 pixel samples are remaining pixel samples other than the K1 pixel samples in the K pixel samples, and K1 and K2 are positive integers.
- With reference to the sixth possible implementation manner of the first aspect, in a seventh possible implementation manner of the first aspect, a pixel sample i is any pixel sample except the first vertex angle pixel sample in the K pixel samples, and if a prediction direction corresponding to a motion vector predictor of the pixel sample i is different from a prediction direction corresponding to the motion vector predictor of the first vertex angle pixel sample, a reference frame index corresponding to the motion vector predictor of the pixel sample i is 0.
- With reference to the seventh possible implementation manner of the first aspect, in an eighth possible implementation manner of the first aspect, the reference frame Y1 is a reference frame corresponding to the motion vector predictor of the first vertex angle pixel sample, and that the motion vector predictor of the pixel sample i is scaled to the reference frame Y1 includes: if reference frame indexes corresponding to the motion vector predictors of the first vertex angle pixel sample and the pixel sample i are different, the motion vector predictor of the pixel sample i is scaled to the reference frame Y1.
- With reference to the sixth possible implementation manner of the first aspect or the seventh possible implementation manner of the first aspect or the eighth possible implementation manner of the first aspect, in a ninth possible implementation manner of the first aspect, the performing motion estimation on the K pixel samples by using motion vector predictors of K2 pixel samples and motion vector predictors that are of the K1 pixel samples and that are scaled to the reference frame Y1, to obtain motion vectors of the K pixel samples includes:
- performing motion estimation on the K pixel samples based on an iterative search algorithm by using the motion vector predictors of the K2 pixel samples and the motion vector predictors that are of the K1 pixel samples and that are scaled to the reference frame Y1, to obtain the motion vectors of the K pixel samples.
- With reference to any one of the first aspect, or the first to the ninth possible implementation manners of the first aspect, in a tenth possible implementation manner of the first aspect, the non-translational motion model is any one of the following models: an affine transformation model, a parabolic motion model, a rotary motion model, a perspective motion model, a shearing motion model, or a scalable motion model.
- With reference to any one of the first aspect, or the first to the tenth possible implementation manners of the first aspect, in an eleventh possible implementation manner of the first aspect,
- the picture prediction method is applied to a video coding process or the picture prediction method is applied to a video decoding process.
- A second aspect of the present application provides a picture prediction apparatus, including:
- a determining unit, configured to determine motion vector predictors of K pixel samples in a current picture block, where K is an integer greater than 1, the K pixel samples include a first vertex angle pixel sample in the current picture block, a motion vector predictor of the first vertex angle pixel sample is obtained based on a motion vector of a preset first spatially adjacent picture block of the current picture block, and the first spatially adjacent picture block is spatially adjacent to the first vertex angle pixel sample; and
- a predicting unit, configured to perform, based on a non-translational motion model and the motion vector predictors of the K pixel samples, pixel value prediction on the current picture block.
- With reference to the second aspect, in a first possible implementation manner of the second aspect, the K pixel samples include at least two pixel samples in an upper left pixel sample, an upper right pixel sample, or a lower left pixel sample in the current picture block; and
- the upper left pixel sample in the current picture block is an upper left vertex of the current picture block or a pixel block that is in the current picture block and that includes an upper left vertex of the current picture block, the lower left pixel sample in the current picture block is a lower left vertex of the current picture block or a pixel block that is in the current picture block and that includes a lower left vertex of the current picture block, and the upper right pixel sample in the current picture block is an upper right vertex of the current picture block or a pixel block that is in the current picture block and that includes an upper right vertex of the current picture block.
- With reference to the first possible implementation manner of the second aspect, in a second possible implementation manner of the second aspect, the first spatially adjacent picture block is a picture block X1, a picture block X2, or a picture block X3;
- a motion vector predictor of the upper left pixel sample in the current picture block is obtained based on a motion vector of the picture block X1, the picture block X1 is a spatially adjacent picture block of the current picture block, and the picture block X1 is spatially adjacent to the upper left pixel sample in the current picture block;
- a motion vector predictor of the upper right pixel sample in the current picture block is obtained based on a motion vector of the picture block X2, the picture block X2 is a spatially adjacent picture block of the current picture block, and the picture block X2 is spatially adjacent to the upper right pixel sample in the current picture block; and
- a motion vector predictor of the lower left pixel sample in the current picture block is obtained based on a motion vector of the picture block X3, the picture block X3 is a spatially adjacent picture block of the current picture block, and the picture block X3 is spatially adjacent to the lower left pixel sample in the current picture block.
- With reference to the second possible implementation manner of the second aspect, in a third possible implementation manner of the second aspect,
- the first vertex angle pixel sample is the upper left pixel sample in the current picture block, the first spatially adjacent picture block is the picture block X1, and the picture block X1 is a spatially adjacent picture block on the upper left of the current picture block.
- With reference to the third possible implementation manner of the second aspect, in a fourth possible implementation manner of the second aspect,
- the picture block X2 is a spatially adjacent picture block above the current picture block; or
- an absolute value of a difference between a motion vector absolute value of the picture block X2 and a motion vector absolute value of the picture block X1 is greater than or equal to an absolute value of a difference between a motion vector absolute value of a picture block X5 and the motion vector absolute value of the picture block X1, and the picture block X5 is any one of remaining at least some spatially adjacent picture blocks, other than the picture block X2, that are of the current picture block and that are spatially adjacent to the upper right vertex.
- With reference to the third possible implementation manner of the second aspect or the fourth possible implementation manner of the second aspect, in a fifth possible implementation manner of the second aspect,
- the picture block X3 is a spatially adjacent picture block on the left of the current picture block; or
- an absolute value of a difference between a motion vector absolute value of the picture block X3 and the motion vector absolute value of the picture block X1 is greater than or equal to an absolute value of a difference between a motion vector absolute value of a picture block X6 and the motion vector absolute value of the picture block X1, and the picture block X6 is any one of remaining at least some spatially adjacent picture blocks, other than the picture block X3, that are of the current picture block and that are spatially adjacent to the lower left vertex.
- With reference to any one of the second aspect, or the first to the fifth possible implementation manners of the second aspect, in a sixth possible implementation manner of the second aspect, the predicting unit is specifically configured to: when reference frames corresponding to motion vector predictors of K1 pixel samples in the K pixel samples are not a reference frame Y1, scale the motion vector predictors of the K1 pixel samples to the reference frame Y1, perform motion estimation on the K pixel samples by using motion vector predictors of K2 pixel samples and motion vector predictors that are of the K1 pixel samples and that are scaled to the reference frame Y1, to obtain motion vectors of the K pixel samples, and perform pixel value prediction on the current picture block by using the non-translational motion model and the motion vectors of the K pixel samples, where the K2 pixel samples are remaining pixel samples other than the K1 pixel samples in the K pixel samples, and K1 and K2 are positive integers.
- With reference to the sixth possible implementation manner of the second aspect, in a seventh possible implementation manner of the second aspect, a pixel sample i is any pixel sample except the first vertex angle pixel sample in the K pixel samples, and if a prediction direction corresponding to a motion vector predictor of the pixel sample i is different from a prediction direction corresponding to the motion vector predictor of the first vertex angle pixel sample, a reference frame index corresponding to the motion vector predictor of the pixel sample i is 0.
- With reference to the seventh possible implementation manner of the second aspect, in an eighth possible implementation manner of the second aspect, the reference frame Y1 is a reference frame corresponding to the motion vector predictor of the first vertex angle pixel sample, and the predicting unit is specifically configured to: if reference frame indexes corresponding to the motion vector predictors of the first vertex angle pixel sample and the pixel sample i are different, scale the motion vector predictor of the pixel sample i to the reference frame Y1.
- With reference to any one of the second aspect, or the first to the eighth possible implementation manners of the second aspect, in a ninth possible implementation manner of the second aspect, the non-translational motion model is any one of the following models: an affine transformation model, a parabolic motion model, a rotary motion model, a perspective motion model, a shearing motion model, or a scalable motion model.
- With reference to any one of the second aspect, or the first to the ninth possible implementation manners of the second aspect, in a tenth possible implementation manner of the second aspect, the picture prediction apparatus is applied to a video coding apparatus or the picture prediction apparatus is applied to a video decoding apparatus.
- It may be learned that in the solutions in the embodiments of the present application, determined motion vector predictors of K pixel samples are referenced during pixel value prediction performed, based on a non-translational motion model, on a current picture block. The K pixel samples include a first vertex angle pixel sample in the current picture block. The motion vector predictors of the K pixel samples are directly used for prediction, a motion vector predictor of the first vertex angle pixel sample in the K pixel samples is obtained based on a motion vector of a preset first spatially adjacent picture block of the current picture block, and there is a single optional case for the motion vector predictors of the K pixel samples. Therefore, a mechanism that is used in a conventional technology and in which motion vector predictors of K pixel samples are selected from multiple optional motion vector predictor sets of the K pixel samples by means of massive calculation is abandoned. This is helpful in avoiding transferring selection information of motion information predictors in a bitstream, is helpful in improving coding efficiency, and is also helpful in significantly reducing calculation complexity of picture prediction based on a non-translational motion model.
-
FIG. 1 -a is a schematic diagram of a prediction unit division manner corresponding to intra-frame prediction according to an embodiment of the present application; -
FIG. 1 -b is a schematic diagram of several prediction unit division manners corresponding to inter-frame prediction according to an embodiment of the present application; -
FIG. 1 -c is a schematic flowchart of a picture prediction method according to an embodiment of the present application; -
FIG. 2 is a schematic diagram of possible adjacent reference blocks of a current picture block according to an embodiment of the present application; -
FIG. 3 -a is a schematic flowchart of another picture prediction method according to an embodiment of the present application; -
FIG. 3 -b is another schematic diagram of possible adjacent reference blocks of a current picture block according to an embodiment of the present application; -
FIG. 3 -c is a schematic diagram of a reference frame of a current picture block according to an embodiment of the present application; -
FIG. 3 -d is a schematic diagram of a current picture block according to an embodiment of the present application; -
FIG. 3 -e is a schematic diagram of affine transformation of a picture block according to an embodiment of the present application; -
FIG. 4 -a is a schematic flowchart of another picture prediction method according to an embodiment of the present application; -
FIG. 4 -b is another schematic diagram of adjacent reference blocks of a current picture block according to an embodiment of the present application; -
FIG. 5 is a schematic flowchart of another picture prediction method according to an embodiment of the present application; -
FIG. 6 is a schematic flowchart of another picture prediction method according to an embodiment of the present application; -
FIG. 7 is a schematic flowchart of a video decoding method according to an embodiment of the present application; -
FIG. 8 is a schematic diagram of a picture prediction apparatus according to an embodiment of the present application; -
FIG. 9 is a schematic diagram of another picture prediction apparatus according to an embodiment of the present application; and -
FIG. 10 is a schematic diagram of another picture prediction apparatus according to an embodiment of the present application. - The embodiments of the present application provide a picture prediction method and a related apparatus, so as to reduce calculation complexity of picture prediction based on a non-translational motion model.
- To make persons skilled in the art better understand the technical solutions in the present application, the following clearly describes the technical solutions in the embodiments of the present application with reference to the accompanying drawings in the embodiments of the present application. The described embodiments are merely some rather than all of the embodiments of the present application. All other embodiments obtained by persons of ordinary skill in the art based on the embodiments of the present application without creative efforts shall fall within the protection scope of the present application.
- The following provides detailed descriptions separately.
- In the specification, claims, and accompanying drawings of the present application, terms “first”, “second”, “third”, “fourth”, and the like are intended to distinguish between different objects but do not indicate a particular order. In addition, terms “include”, “have”, and any other variant thereof are intended to cover a non-exclusive inclusion. For example, a process, a method, a system, a product, or a device that includes a series of steps or units is not limited to the listed steps or units, but optionally further includes an unlisted step or unit, or optionally further includes another inherent step or unit of the process, the method, the product, or the device.
- The following first describes some concepts that may be involved in the embodiments of the present application.
- In most coding frameworks, a video sequence includes a series of pictures, the pictures are further divided into slices, and the slices are further divided into blocks. Video coding is to perform coding processing from left to right and from top to bottom row by row starting from an upper left corner position of a picture by using a block as a unit. In some new video coding standards, a concept of a block is further extended. There is a macroblock (MB) in the H.264 standard, and the macroblock may be further divided into multiple prediction blocks that may be used for predictive coding. In the HEVC standard, basic concepts such as a coding unit (CU), a prediction unit (PU), and a transform unit (TU) are used, and multiple units are classified according to functions, and a completely new tree-based structure is used for description. For example, the coding unit may be divided into smaller coding units according to a quadtree, and the smaller coding unit may be further divided, to form a quadtree structure. The prediction unit and the transform unit also have similar tree structures. Regardless of whether a unit is a coding unit, a prediction unit, or a transform unit, the unit belongs to the concept of a block in essence. The coding unit is similar to a macroblock MB or a coding block, and is a basic unit for partitioning and coding an picture. The prediction unit may correspond to a prediction block, and is a basic unit for predictive coding. The coding unit is further divided into multiple prediction units according to a division mode. The transform unit may correspond to a transform block, and is a basic unit for transforming a prediction residual.
- In the HEVC standard, a size of the coding unit may include four levels: 64×64, 32×32, 16×16, and 8×8. Coding units at each level may be divided into prediction units of different sizes according to intra-frame prediction and inter-frame prediction. For example, as shown in
FIG. 1 -a andFIG. 1 -b,FIG. 1 -a shows an example of a prediction unit division manner corresponding to intra-frame prediction.FIG. 1 -b shows examples of several prediction unit division manners corresponding to inter-frame prediction. - During development and evolution of video coding technologies, video coding experts figure out various methods to use temporal and spatial correlation between adjacent coding/decoding blocks to try to improve coding efficiency. In the H264/Advanced Video Coding (AVC) standard, a skip mode and a direct mode become effective tools for improving coding efficiency. Blocks of the two coding modes used when a bit rate is low can occupy more than a half of an entire coding sequence. When the skip mode is used, a motion vector of a current picture block can be derived by using nearby motion vectors only by adding a skip mode flag to a bit stream, and a value of a reference block is directly copied according to the motion vector as a reconstructed value of the current picture block. In addition, when the direct mode is used, an encoder may derive the motion vector of the current picture block by using the adjacent motion vectors, and directly copy the value of the reference block according to the motion vector as a predictor of the current picture block, and perform predictive coding on the current picture block by using the predictor in an encoder. In the evolved HEVC standard, some new coding tools are introduced to further improve video coding efficiency. A merge coding mode and an advanced motion vector prediction (AMVP) mode are two important inter-frame prediction tools. During merge coding, a candidate motion information set is constructed by using motion information (including a prediction direction, a motion vector, and a reference frame index) of an adjacent coded block of a current coding block, candidate motion information that enables coding efficiency to be the highest may be selected as motion information of the current coding block by means of comparison, a predictor of the current coding block is found in a reference frame, predictive coding is performed on the current coding block, and an index value indicating an adjacent coded block whose motion information is selected is written into a bitstream. When the advanced motion vector prediction mode is used, a motion vector of an adjacent coded block is used as a motion vector predictor of a current coding block, a motion vector that enables coding efficiency to be the highest may be selected to predict a motion vector of the current coding block, and an index value that indicates a selected adjacent motion vector may be written into a video bitstream.
- The technical solutions in the embodiments of the present application are further discussed in the following.
- The following first describes a picture prediction method provided in the embodiments of the present application. The picture prediction method provided in the embodiments of the present application is executed by a video coding apparatus or a video decoding apparatus. The video coding apparatus or the video decoding apparatus may be any apparatus that needs to output or store a video, for example, a laptop computer, a tablet computer, a personal computer, a mobile phone, or a video server.
- In an embodiment of a picture prediction method in the present application, the picture prediction method includes: determining motion vector predictors (MVP) of K pixel samples in a current picture block, where K is an integer greater than 1, the K pixel samples include a first vertex angle pixel sample in the current picture block, a motion vector predictor of the first vertex angle pixel sample is obtained based on a motion vector of a preset first spatially adjacent picture block of the current picture block, and the first spatially adjacent picture block is spatially adjacent to the first vertex angle pixel sample; and performing, based on a non-translational motion model and the motion vector predictors of the K pixel samples, pixel value prediction on the current picture block.
- Referring to
FIG. 1 -c,FIG. 1 -c is a schematic flowchart of a picture prediction method according to an embodiment of the present application. As shown in an example ofFIG. 1 -c, the picture prediction method provided in this embodiment of the present application may include the following steps. - 101. Determine motion vector predictors of K pixel samples in a current picture block.
- K is an integer greater than 1, the K pixel samples include a first vertex angle pixel sample in the current picture block, and a motion vector predictor of the first vertex angle pixel sample is obtained based on a motion vector of a preset first spatially adjacent picture block of the current picture block. The first spatially adjacent picture block is spatially adjacent to the first vertex angle pixel sample.
- Optionally, in some possible implementation manners of the present application, the motion vector predictor of the first vertex angle pixel sample is obtained based on the motion vector of the preset first spatially adjacent picture block of the current picture block. Specifically, the motion vector predictor of the first vertex angle pixel sample may be equal to the motion vector of the preset first spatially adjacent picture block of the current picture block (that is, the motion vector of the preset first spatially adjacent picture block of the current picture block is used as the motion vector predictor of the first vertex angle pixel sample); or a motion vector predictor obtained by performing preset transformation on the motion vector of the preset first spatially adjacent picture block of the current picture block may be used as the motion vector predictor of the first vertex angle pixel sample.
- 102. Perform, based on a non-translational motion model and the motion vector predictors of the K pixel samples, pixel value prediction on the current picture block.
- Optionally, in some possible implementation manners of the present application, the non-translational motion model is any one of the following models: an affine transformation model, a parabolic motion model, a rotary motion model, a perspective motion model, a shearing motion model, or a scalable motion model. It may be understood that the non-translational motion model is not limited to the foregoing specific examples.
- It may be learned that in the technical solution in this embodiment, determined motion vector predictors of K pixel samples are referenced during pixel value prediction performed, based on a non-translational motion model, on a current picture block. The K pixel samples include a first vertex angle pixel sample in the current picture block. The motion vector predictors of the K pixel samples are directly used for prediction, a motion vector predictor of the first vertex angle pixel sample in the K pixel samples is obtained based on a motion vector of a preset first spatially adjacent picture block of the current picture block, and there is a single optional case for the motion vector predictors of the K pixel samples. Therefore, a mechanism that is used in a conventional technology and in which motion vector predictors of K pixel samples are selected from multiple optional motion vector predictor sets of the K pixel samples by means of massive calculation is abandoned. This is helpful in avoiding transferring selection information of motion information predictors in a bitstream, is helpful in improving coding efficiency, and is also helpful in significantly reducing calculation complexity of picture prediction based on a non-translational motion model.
- Further, the first spatially adjacent picture block is spatially adjacent to the first vertex angle pixel sample, and the first spatially adjacent picture block of the current picture block is preset (that is, it may be considered that the first spatially adjacent picture block of the current picture block is agreed or set in advance). In this way, this is helpful in quickly determining the motion vector predictors of the K pixel samples. In addition, because redundancy is removed by making full use of a spatial correlation between picture blocks, this is helpful in further improving coding efficiency.
- There may be various manners of selecting the K pixel samples.
- Optionally, in some possible implementation manners of the present application, the K pixel samples include at least two pixel samples in an upper left pixel sample, an upper right pixel sample, a lower left pixel sample, or a lower right pixel sample in the current picture block.
- If the pixel sample is a pixel block, a size of the pixel block is, for example, 2×2, 1×2, 4×2, 4×4, or another size.
- Optionally, in some possible implementation manners of the present application, the upper left pixel sample in the current picture block is an upper left vertex of the current picture block or a pixel block that is in the current picture block and that includes an upper left vertex of the current picture block. The lower left pixel sample in the current picture block is a lower left vertex of the current picture block or a pixel block that is in the current picture block and that includes a lower left vertex of the current picture block. The upper right pixel sample in the current picture block is an upper right vertex of the current picture block or a pixel block that is in the current picture block and that includes an upper right vertex of the current picture block. The lower right pixel sample in the current picture block is a lower right vertex of the current picture block or a pixel block that is in the current picture block and that includes a lower right vertex of the current picture block.
- Optionally, in some possible implementation manners of the present application, the K pixel samples may further include a center pixel sample in the current picture block. The center pixel sample in the current picture block may be a center pixel in the current picture block or a pixel block that is in the current picture block and that includes a center pixel in the current picture block.
- Optionally, in some possible implementation manners of the present application, the first spatially adjacent picture block may be a picture block X1, a picture block X2, a picture block X3, or a picture block X4.
- A motion vector predictor of the upper left pixel sample in the current picture block may be obtained based on a motion vector of the picture block X1, the picture block X1 is a spatially adjacent picture block of the current picture block, and the picture block X1 is spatially adjacent to the upper left pixel sample in the current picture block.
- A motion vector predictor of the upper right pixel sample in the current picture block may be obtained based on a motion vector of the picture block X2, the picture block X2 is a spatially adjacent picture block of the current picture block, and the picture block X2 is spatially adjacent to the upper right pixel sample in the current picture block.
- A motion vector predictor of the lower left pixel sample in the current picture block may be obtained based on a motion vector of the picture block X3, the picture block X3 is a spatially adjacent picture block of the current picture block, and the picture block X3 is spatially adjacent to the lower left pixel sample in the current picture block.
- A motion vector predictor of the lower right pixel sample in the current picture block may be obtained based on a motion vector of the picture block X4, the picture block X4 is a spatially adjacent picture block of the current picture block, and the picture block X4 is spatially adjacent to the lower right pixel sample in the current picture block.
- Optionally, in some possible implementation manners of the present application, the first vertex angle pixel sample may be the upper left pixel sample in the current picture block, and the first spatially adjacent picture block may be the picture block X1. The picture block X1 is a spatially adjacent picture block on the upper left of the current picture block, or the picture block X1 is a spatially adjacent picture block on the left of the current picture block, or the picture block X1 is a spatially adjacent picture block above the current picture block.
- Optionally, in some possible implementation manners of the present application, the picture block X2 is a spatially adjacent picture block above the current picture block, or the picture block X2 is a spatially adjacent picture block on the upper right of the current picture block, or the picture block X2 is a spatially adjacent picture block on the right of the current picture block. Still optionally, an absolute value of a difference between a motion vector absolute value of the picture block X2 and a motion vector absolute value of the picture block X1 is greater than or equal to an absolute value of a difference between a motion vector absolute value of a picture block X5 and the motion vector absolute value of the picture block X1, and the picture block X5 is any one of remaining at least some spatially adjacent picture blocks, other than the picture block X2, that are of the current picture block and that are spatially adjacent to the upper right vertex. That is, a location of the picture block X2 may be determined based on a specific policy, or a location of the picture block X2 may be directly agreed.
- Optionally, in some possible implementation manners of the present application, the picture block X3 is a spatially adjacent picture block on the left of the current picture block, or the picture block X3 is a spatially adjacent picture block on the lower left of the current picture block, or the picture block X3 is a spatially adjacent picture block below the current picture block. Still optionally, an absolute value of a difference between a motion vector absolute value of the picture block X3 and the motion vector absolute value of the picture block X1 is greater than or equal to an absolute value of a difference between a motion vector absolute value of a picture block X6 and the motion vector absolute value of the picture block X1, and the picture block X6 is any one of remaining at least some spatially adjacent picture blocks, other than the picture block X3, that are of the current picture block and that are spatially adjacent to the lower left vertex. That is, a location of the picture block X3 may be determined based on a specific policy, or a location of the picture block X3 may be directly agreed.
- Optionally, in some possible implementation manners of the present application, the picture block X4 is a spatially adjacent picture block on the right of the current picture block, or the picture block X4 is a spatially adjacent picture block on the lower right of the current picture block, or the picture block X4 is a spatially adjacent picture block below the current picture block. Still optionally, an absolute value of a difference between a motion vector absolute value of the picture block X4 and the motion vector absolute value of the picture block X1 is greater than or equal to an absolute value of a difference between a motion vector absolute value of a picture block X7 and the motion vector absolute value of the picture block X1, and the picture block X7 is any one of remaining at least some spatially adjacent picture blocks, other than the picture block X4, that are of the current picture block and that are spatially adjacent to the lower left vertex.
- Referring to
FIG. 2 ,FIG. 2 shows examples of some possible locations of the picture block X1, the picture block X2, the picture block X3, and the picture block X4. - Optionally, in some possible implementation manners of the present application, the performing, based on a non-translational motion model and the motion vector predictors of the K pixel samples, pixel value prediction on the current picture block may specifically include: if reference frames corresponding to the motion vector predictors of the K pixel samples are a reference frame Y1, performing motion estimation on the K pixel samples by using the motion vector predictors of the K pixel samples, to obtain motion vectors of the K pixel samples; and performing pixel value prediction on the current picture block by using the non-translational motion model and the motion vectors of the K pixel samples.
- Optionally, in some possible implementation manners of the present application, the performing, based on a non-translational motion model and the motion vector predictors of the K pixel samples, pixel value prediction on the current picture block may specifically include: when reference frames corresponding to motion vector predictors of K1 pixel samples in the K pixel samples are not a reference frame Y1, scaling the motion vector predictors of the K1 pixel samples to the reference frame Y1; performing motion estimation on the K pixel samples by using motion vector predictors of K2 pixel samples and motion vector predictors that are of the K1 pixel samples and that are scaled to the reference frame Y1, to obtain motion vectors of the K pixel samples; and performing pixel value prediction on the current picture block by using the non-translational motion model and the motion vectors of the K pixel samples, where the K2 pixel samples are remaining pixel samples other than the K1 pixel samples in the K pixel samples, and K1 and K2 are positive integers.
- Optionally, in some possible implementation manners of the present application, the reference frame Y1 may be a reference frame corresponding to the motion vector predictor of the first vertex angle pixel sample, or the reference frame Y1 may be a reference frame corresponding to a motion vector predictor of another pixel sample in the K pixel samples.
- Optionally, in some possible implementation manners of the present application, a pixel sample i is any pixel sample except the first vertex angle pixel sample in the K pixel samples, and if a prediction direction corresponding to a motion vector predictor of the pixel sample i is different from a prediction direction corresponding to the motion vector predictor of the first vertex angle pixel sample, a reference frame index corresponding to the motion vector predictor of the pixel sample i is 0.
- Optionally, in some possible implementation manners of the present application, that the motion vector predictor of the pixel sample i is scaled to the reference frame Y1 may include: if reference frame indexes corresponding to the motion vector predictors of the first vertex angle pixel sample and the pixel sample i are different, the motion vector predictor of the pixel sample i is scaled to the reference frame Y1.
- Optionally, in some possible implementation manners of the present application, the performing motion estimation on the K pixel samples by using motion vector predictors of K2 pixel samples and motion vector predictors that are of the K1 pixel samples and that are scaled to the reference frame Y1, to obtain motion vectors of the K pixel samples may include: performing motion estimation on the K pixel samples based on an iterative search algorithm (or a motion estimation algorithm of another type) by using the motion vector predictors of the K2 pixel samples and the motion vector predictors that are of the K1 pixel samples and that are scaled to the reference frame Y1, to obtain the motion vectors of the K pixel samples.
- Certainly, in some other possible application scenarios, a step of performing motion estimation on the K pixel samples may be omitted. A normalization step may even be omitted. That is, the following step may be omitted: when the reference frames corresponding to the motion vector predictors of the K1 pixel samples in the K pixel samples are not the reference frame Y1, the motion vector predictors of the K1 pixel samples are scaled to the reference frame Y1.
- It may be understood that any one of the foregoing picture prediction methods provided in this embodiment may be applied to a video coding process or a video decoding process.
- To help better understand and implement the foregoing solutions in the embodiments of the present application, the following provides further descriptions with reference to a more specific application scenario.
- Referring to
FIG. 3 -a,FIG. 3 -a is a schematic flowchart of another picture prediction method according to another embodiment of the present application. As shown in an example ofFIG. 3 -a, the another picture prediction method provided in the another embodiment of the present application may include the following steps. - 301. Determine motion vector predictors of K pixel samples in a current picture block.
- In an example of this embodiment, K=3.
- In this embodiment, the K pixel samples include an upper left vertex, an upper right vertex, and a lower left vertex of the current picture block.
- As shown in an example of
FIG. 3 -b, a motion vector predictor of the upper left vertex of the current picture block is equal to a motion vector of an adjacent picture block A on the upper left of the current picture block. A motion vector predictor of the upper right vertex of the current picture block is equal to a motion vector of an adjacent picture block B above the current picture block, or a motion vector predictor of the upper right vertex of the current picture block may be equal to a motion vector of an adjacent picture block C on the upper right of the current picture block. A motion vector predictor of the lower left vertex of the current picture block may be equal to a motion vector of an adjacent picture block E on the lower left of the current picture block, or a motion vector predictor of the lower left vertex of the current picture block may be equal to a motion vector of an adjacent picture block D on the left of the current picture block. - For example, the motion vector predictors of the upper right vertex and the lower left vertex of the current picture block may be determined based on the following formulas:
-
- In the foregoing formula (1) and formula (2),
mvA represents the motion vector (MV) of the picture block A. - In the foregoing formula (1),
mvp1 represents an MVP of the upper right vertex of the current picture block,mvB represents the MV of the picture block B, andmvC represents the MV of the picture block C. In the foregoing formula (2),mvp2 represents the MVP of the upper right vertex of the current picture block,mvD represents the MV of the picture block D,mvE and represents the MV of the picture block E. - That is, for the MVP of the upper right vertex, a picture block that has a larger difference with a motion vector absolute value of the block A is selected from the spatially adjacent picture block B and picture block C. For the MVP of the lower left vertex, a picture block that has a larger difference with the motion vector absolute value of the block A is selected from the spatially adjacent picture block D and picture block E.
- 302. Perform motion estimation on the K pixel samples by using the motion vector predictors of the K pixel samples, to obtain motion vectors of the K pixel samples.
- When reference frames corresponding to motion vector predictors of K1 pixel samples in the K pixel samples are not a reference frame Y1, the performing motion estimation on the K pixel samples by using the motion vector predictors of the K pixel samples, to obtain motion vectors of the K pixel samples includes: scaling the motion vector predictors of the K1 pixel samples to the reference frame Y1 (that is, performing normalization processing on the motion vector predictors of the K1 pixel samples); and performing motion estimation on the K pixel samples by using motion vector predictors of K2 pixel samples and motion vector predictors that are of the K1 pixel samples and that are scaled to the reference frame Y1, to obtain the motion vectors of the K pixel samples.
- Before normalization processing is performed on the motion vector predictors of the K1 pixel samples, a reference frame index of a motion vector predictor of a pixel sample i may be first revised. The pixel sample i is any pixel sample (for example, the lower left vertex or the upper right vertex), except the upper left vertex of the current picture block, in the K pixel samples. If a prediction direction corresponding to the motion vector predictor of the pixel sample i is different from a prediction direction corresponding to the motion vector predictor of the upper left vertex (for example, the prediction direction corresponding to the motion vector predictor of the pixel sample i is forward prediction, and the prediction direction corresponding to the motion vector predictor of the upper left vertex is backward prediction; or the prediction direction corresponding to the motion vector predictor of the pixel sample i is backward prediction, and the prediction direction corresponding to the motion vector predictor of the upper left vertex is forward prediction), a reference frame index corresponding to the motion vector predictor of the pixel sample i is 0.
- Optionally, in some possible implementation manners of the present application, a method for scaling a motion vector predictor to a reference frame may be as follows: It is assumed that a current coding frame index is CurPoc, and a reference frame of the picture block A (that is, a reference frame corresponding to the MVP of the upper left vertex) is a destination reference frame. A destination reference frame index is DesPoc, a reference frame corresponding to the MVP of the upper right vertex or the lower left vertex is a source reference frame, a source reference frame index is SrcPoc, the motion vector predictor of the upper right vertex or the lower left vertex is MVP, and a motion vector predictor that is of the upper right vertex or the lower left vertex and that is obtained after scaling is MVs. As shown in
FIG. 3 -c, MVs may be obtained by means of calculation by using the following formulas: -
- Particularly, if reference frames corresponding to the motion vector predictors of the K pixel samples are the reference frame Y1, no normalization processing is performed on the motion vector predictors of the K1 pixel samples (because the motion vector predictors of the K1 pixel samples have been in a normalized state), and motion estimation may be performed on the K pixel samples directly by using the motion vector predictors of the K pixel samples, to obtain the motion vectors of the K pixel samples.
- Optionally, that motion estimation is performed on the K pixel samples to obtain the motion vectors of the K pixel samples may include: motion estimation is performed on the K pixel samples based on an iterative search algorithm, to obtain the motion vectors of the K pixel samples.
- Optionally, in some possible implementation manners of the present application, that motion estimation is performed on the K pixel samples based on each iteration in the iterative search algorithm, to obtain the motion vectors of the K pixel samples may include the following steps.
- S1. Calculate φk(p) of each pixel according to formulas (3), (4), and (5).
- In φk(p), k=1, 2, 3.
- As shown in
FIG. 3 -d, it is assumed that a size of a current picture block X is S1×S2, and coordinates of each pixel p in the current picture block X are (x, y). -
- S2. Calculate Ψ′1,x(p) and Ψ′1,y(p).
- Ψ′1,x(p) and Ψ′1,y(p) respectively represent derivatives of each pixel p in the current picture block X in a direction x and a direction y. In an example of this embodiment, the derivatives are calculated by using a Sobel operator. Another operator (for example, a Gaussian operation) may be used. Templates of the Sobel operator are shown in examples of formula (6) and formula (7).
-
- Assuming that coordinates of a current pixel p are (x, y), calculation formulas of Ψ′1,x(p) and Ψ′1,y(p) are shown in examples of formula (8) and formula (9), where pred[x] [y] represents a predicted pixel value of a pixel that is in the current picture block X and whose coordinates are (x, y).
-
- An adjacent pixel of a pixel located on the edge of the current picture block X may not exist, and Ψ′1,x(p) and Ψ′1,y(p) may be replaced with corresponding values of the adjacent pixel of the pixel located on the edge of the current picture block X.
- S3. Calculate a prediction error e(p).
- It is assumed that the coordinates of the current pixel p are (x, y).
-
e(p)=org[x][y]−pred[x][y] formula (10) - where org[x][y] and pred[x][y] respectively represent an original pixel value and a predicated pixel value of the pixel that is in the current picture block X and whose coordinates are (x, y).
- S4. Assuming that δx1, δx2, δx3, δy1, δy2, and δy3 are respectively motion vector variations of three vertices of a current picture block X in a direction x and a direction y, obtain δx1, δx2, δx3, δy1, δy2, and δy3 by means of calculation by solving a linear equation set (11).
-
- Based on the foregoing steps S1 to S4, a motion vector variation δx of each of the three pixel samples in the current picture block in the direction x and a motion vector variation δy of each pixel sample in the direction y in this iteration may be obtained by means of motion estimation.
- One or more iterations are performed based on an iterative manner in the foregoing steps S1 to S4, so that the motion vectors of the K pixel samples in the current picture block can be obtained by means of motion estimation.
- Optionally, in some possible implementation manners of the present application, subtraction may be further performed between the motion vectors that are of the K pixel samples and that are obtained by means of motion estimation and the corresponding MVPs, to obtain motion vector differences (MVD) of the K pixel samples, and the MVDs of the K pixel samples may be written into a video bitstream.
- Further, a syntactic element affine_no_MVD_flag may be used to indicate whether to write an MVD corresponding to a pixel sample into the video bitstream. For example, if the MVDs corresponding to the foregoing K pixel samples are all 0, affine_no_MVD_flag may be set to true, and no MVD is written into the video bitstream; or if the MVDs corresponding to the foregoing K pixel samples are not all 0, affine_no_MVD_flag may be set to false, and the MVDs corresponding to the K pixel samples are written into the video bitstream.
- 303. Calculate, based on an affine transformation model and the motion vectors of the K pixel samples, a motion vector of each pixel in the current picture block.
- It is assumed that the size of the current prediction block X is S1×S2. Motion vectors (vx0, vy0), (vx1, vy1), and (vx2, vy2) of three vertices whose coordinates are (0, 0) (S1, 0), and (0, S2) are calculated by performing
step 301 andstep 302. - The coordinates and the motion vectors of the three vertices are substituted into an affine transformation formula (12), so that the motion vector of each pixel in the current picture block can be calculated. A schematic diagram of affine transformation may be shown in
FIG. 3 -e. - A motion vector of the pixel whose coordinates are (x, y) is calculated as follows:
-
- Further, because motion information of each pixel in the current picture block may be used for motion vector prediction for another picture block, one piece of motion information may be stored for each 4×4 block, and a motion vector of each 4×4 block is stored as motion information of a pixel on an upper left corner or any other pixel in the 4×4 block.
- It should be noted that in this embodiment, the affine transformation model is mainly used as an example of a non-translational motion model for description, and a mechanism for calculating, based on another non-translational motion model and the motion vectors of the K pixel samples, the motion vector of each pixel in the current picture block has a similar principle although formulas different from those used in this embodiment may be used in the mechanism. Examples are not listed one by one herein.
- 304. Perform, based on the calculated motion vector of each pixel in the current picture block, pixel value prediction on the current picture block.
- Further, during video coding, a residual of the current picture block may be obtained by using an original pixel value of the current picture block and a predicted pixel value that is of the current picture block and that is obtained by means of pixel value prediction, and the residual of the current picture block may be written into the video bitstream.
- It may be learned that in the technical solution in this embodiment, determined motion vector predictors of K pixel samples are referenced during pixel value prediction performed, based on a non-translational motion model, on a current picture block. The K pixel samples include a first vertex angle pixel sample in the current picture block. The motion vector predictors of the K pixel samples are directly used for prediction, a motion vector predictor of the first vertex angle pixel sample in the K pixel samples is obtained based on a motion vector of a preset first spatially adjacent picture block of the current picture block, and there is a single optional case for the motion vector predictors of the K pixel samples. Therefore, a mechanism that is used in a conventional technology and in which motion vector predictors of K pixel samples are selected from multiple optional motion vector predictor sets of the K pixel samples by means of massive calculation is abandoned. This is helpful in avoiding transferring selection information of motion information predictors in a bitstream, is helpful in improving coding efficiency, and is also helpful in significantly reducing calculation complexity of picture prediction based on a non-translational motion model.
- Further, prediction directions of the MVPs of the K pixel samples in the current picture block are normalized, reference frame indexes are processed, and motion vectors corresponding to the MVPs of the K pixel samples are scaled and mapped to a same reference frame according to reference frame indexes. Further, a motion vector of each pixel in the current picture block may be calculated according to optimal motion vectors of the K pixel samples by using an affine transformation formula. The foregoing solution is helpful in resolving a problem that inter-frame prediction accuracy is low in a current video coding standard when a video includes content of a non-translational feature, for example, lens zooming, object deformation, or moving object rotation. In addition, a disadvantage that an affine transformation technology proposed in the conventional technology is hardly applied to multi-direction prediction and multiple-reference-frames-based prediction is overcome, and this is helpful in making full use of a temporary correlation between video sequences to remove temporary redundancy, thereby improving coding efficiency.
- Referring to
FIG. 4 -a,FIG. 4 -a is a schematic flowchart of another picture prediction method according to another embodiment of the present application. As shown in an example ofFIG. 4 -a, the another picture prediction method provided in the another embodiment of the present application may include the following steps. - 401. Determine motion vector predictors of K pixel samples in a current picture block.
- In an example of this embodiment, K=3.
- In this embodiment, the K pixel samples include an upper left vertex, an upper right vertex, and a lower left vertex of the current picture block.
- As shown in an example of
FIG. 4 -b, a motion vector predictor of the upper left vertex of the current picture block is equal to a motion vector of an adjacent picture block A on the upper left of the current picture block. A motion vector predictor of the upper right vertex of the current picture block is equal to a motion vector of an adjacent picture block B above the current picture block. A motion vector predictor of the lower left vertex of the current picture block may be equal to a motion vector of an adjacent picture block D on the left of the current picture block. - 402. Perform motion estimation on the K pixel samples by using the motion vector predictors of the K pixel samples, to obtain motion vectors of the K pixel samples.
- When reference frames corresponding to motion vector predictors of K1 pixel samples in the K pixel samples are not a reference frame Y1, the performing motion estimation on the K pixel samples by using the motion vector predictors of the K pixel samples, to obtain motion vectors of the K pixel samples includes: scaling the motion vector predictors of the K1 pixel samples to the reference frame Y1 (that is, performing normalization processing on the motion vector predictors of the K1 pixel samples); and performing motion estimation on the K pixel samples by using motion vector predictors of K2 pixel samples and motion vector predictors that are of the K1 pixel samples and that are scaled to the reference frame Y1, to obtain the motion vectors of the K pixel samples.
- Before normalization processing is performed on the motion vector predictors of the K1 pixel samples, a reference frame index of a motion vector predictor of a pixel sample i may be first revised. The pixel sample i is any pixel sample (for example, the lower left vertex or the upper right vertex), except the upper left vertex of the current picture block, in the K pixel samples. If a prediction direction corresponding to the motion vector predictor of the pixel sample i is different from a prediction direction corresponding to the motion vector predictor of the upper left vertex (for example, the prediction direction corresponding to the motion vector predictor of the pixel sample i is forward prediction, and the prediction direction corresponding to the motion vector predictor of the upper left vertex is backward prediction; or the prediction direction corresponding to the motion vector predictor of the pixel sample i is backward prediction, and the prediction direction corresponding to the motion vector predictor of the upper left vertex is forward prediction), a reference frame index corresponding to the motion vector predictor of the pixel sample i is 0.
- Optionally, in some possible implementation manners of the present application, a method for scaling a motion vector predictor to a reference frame may be as follows: It is assumed that a current coding frame index is CurPoc, and a reference frame of the picture block A (that is, a reference frame corresponding to the MVP of the upper left vertex) is a destination reference frame. A destination reference frame index is DesPoc, a reference frame corresponding to the MVP of the upper right vertex or the lower left vertex is a source reference frame, a source reference frame index is SrcPoc, the motion vector predictor of the upper right vertex or the lower left vertex is MVP, and a motion vector predictor that is of the upper right vertex or the lower left vertex and that is obtained after scaling is MVs. MVs may be obtained by means of calculation by using the following formulas:
-
- Particularly, if reference frames corresponding to the motion vector predictors of the K pixel samples are the reference frame Y1, no normalization processing is performed on the motion vector predictors of the K1 pixel samples (because the motion vector predictors of the K1 pixel samples have been in a normalized state), and motion estimation may be performed on the K pixel samples directly by using the motion vector predictors of the K pixel samples, to obtain the motion vectors of the K pixel samples.
- Optionally, that motion estimation is performed on the K pixel samples to obtain the motion vectors of the K pixel samples may include: motion estimation is performed on the K pixel samples based on an iterative search algorithm, to obtain the motion vectors of the K pixel samples. For a specific iterative search algorithm, refer to related descriptions in the foregoing embodiment. Details are not described herein again.
- Optionally, in some possible implementation manners of the present application, subtraction may be further performed between the motion vectors that are of the K pixel samples and that are obtained by means of motion estimation and the corresponding MVPs, to obtain MVDs of the K pixel samples, and the MVDs of the K pixel samples may be written into a video bitstream.
- Further, a syntactic element affine_no_MVD_flag may be used to indicate whether to write an MVD corresponding to a pixel sample into the video bitstream. For example, if the MVDs corresponding to the foregoing K pixel samples are all 0, affine_no_MVD_flag may be set to true, and no MVD is written into the video bitstream; or if the MVDs corresponding to the foregoing K pixel samples are not all 0, affine_no_MVD_flag may be set to false, and the MVDs corresponding to the K pixel samples are written into the video bitstream.
- 403. Calculate, based on an affine transformation model and the motion vectors of the K pixel samples, a motion vector of each pixel in the current picture block.
- Further, because motion information of each pixel in the current picture block may be used for motion vector prediction for another picture block, one piece of motion information may be stored for each 4×4 (or 8×8) block, and a motion vector of each 4×4 (or 8×8) block is stored as motion information of a pixel on an upper left corner or any other pixel in the 4'4 (or 8×8) block.
- The motion information may include a motion vector, a prediction direction corresponding to the motion vector, and a reference frame index corresponding to the motion vector.
- It should be noted that in this embodiment, the affine transformation model is mainly used as an example of a non-translational motion model for description, and a mechanism for calculating, based on another non-translational motion model and the motion vectors of the K pixel samples, the motion vector of each pixel in the current picture block has a similar principle although formulas different from those used in this embodiment may be used in the mechanism. Examples are not listed one by one herein.
- 404. Perform, based on the calculated motion vector of each pixel in the current picture block, pixel value prediction on the current picture block.
- Further, during video coding, a residual of the current picture block may be obtained by using an original pixel value of the current picture block and a predicted pixel value that is of the current picture block and that is obtained by means of pixel value prediction, and the residual of the current picture block may be written into the video bitstream.
- It may be learned that in the technical solution in this embodiment, determined motion vector predictors of K pixel samples are referenced during pixel value prediction performed, based on a non-translational motion model, on a current picture block. The K pixel samples include a first vertex angle pixel sample in the current picture block. The motion vector predictors of the K pixel samples are directly used for prediction, a motion vector predictor of the first vertex angle pixel sample in the K pixel samples is obtained based on a motion vector of a preset first spatially adjacent picture block of the current picture block, and there is a single optional case for the motion vector predictors of the K pixel samples. Therefore, a mechanism that is used in a conventional technology and in which motion vector predictors of K pixel samples are selected from multiple optional motion vector predictor sets of the K pixel samples by means of massive calculation is abandoned. This is helpful in avoiding transferring selection information of motion information predictors in a bitstream, is helpful in improving coding efficiency, and is also helpful in significantly reducing calculation complexity of picture prediction based on a non-translational motion model.
- Further, prediction directions of the MVPs of the K pixel samples in the current picture block are normalized, reference frame indexes are processed, and motion vectors corresponding to the MVPs of the K pixel samples are scaled and mapped to a same reference frame according to reference frame indexes. Further, a motion vector of each pixel in the current picture block may be calculated according to optimal motion vectors of the K pixel samples by using an affine transformation formula. The foregoing solution is helpful in resolving a problem that inter-frame prediction accuracy is low in a current video coding standard when a video includes content of a non-translational feature, for example, lens zooming, object deformation, or moving object rotation. In addition, a disadvantage that an affine transformation technology proposed in the conventional technology is hardly applied to multi-direction prediction and multiple-reference-frames-based prediction is overcome, and this is helpful in making full use of a temporary correlation between video sequences to remove temporary redundancy, thereby improving coding efficiency.
- Referring to
FIG. 5 ,FIG. 5 is a schematic flowchart of another picture prediction method according to another embodiment of the present application. As shown in an example ofFIG. 5 , the another picture prediction method provided in the another embodiment of the present application may include the following steps. - 501. Determine motion vector predictors of K pixel samples in a current picture block.
- In an example of this embodiment, K=3.
- In this embodiment, the K pixel samples include an upper left vertex, an upper right vertex, and a lower left vertex of the current picture block.
- For a specific manner of determining the motion vector predictors of the K pixel samples in the current picture block in
step 501, refer to any manner in the foregoing embodiment. - 502. Use the motion vector predictors of the foregoing K pixel samples as motion vectors of the foregoing K pixel samples, and calculate, based on an affine transformation model and the motion vectors of the K pixel samples, a motion vector of each pixel in the current picture block.
- Further, because motion information of each pixel in the current picture block may be used for motion vector prediction for another picture block, one piece of motion information may be stored for each 4×4 (or 8×8) block, and a motion vector of each 4×4 (or 8×8) block is stored as motion information of a pixel on an upper left corner or any other pixel in the 4×4 (or 8×8) block.
- The motion information may include a motion vector, a prediction direction corresponding to the motion vector, and a reference frame index corresponding to the motion vector.
- It should be noted that in this embodiment, the affine transformation model is mainly used as an example of a non-translational motion model for description, and a mechanism for calculating, based on another non-translational motion model and the motion vectors of the K pixel samples, the motion vector of each pixel in the current picture block has a similar principle although formulas different from those used in this embodiment may be used in the mechanism. Examples are not listed one by one herein.
- 503. Perform, based on the calculated motion vector of each pixel in the current picture block, pixel value prediction on the current picture block.
- Further, during video coding, a residual of the current picture block may be obtained by using an original pixel value of the current picture block and a predicted pixel value that is of the current picture block and that is obtained by means of pixel value prediction, and the residual of the current picture block may be written into a video bitstream.
- It may be learned that in the technical solution in this embodiment, determined motion vector predictors of K pixel samples are referenced during pixel value prediction performed, based on a non-translational motion model, on a current picture block. The K pixel samples include a first vertex angle pixel sample in the current picture block. The motion vector predictors of the K pixel samples are directly used for prediction, a motion vector predictor of the first vertex angle pixel sample in the K pixel samples is obtained based on a motion vector of a preset first spatially adjacent picture block of the current picture block, and there is a single optional case for the motion vector predictors of the K pixel samples. Therefore, a mechanism that is used in a conventional technology and in which motion vector predictors of K pixel samples are selected from multiple optional motion vector predictor sets of the K pixel samples by means of massive calculation is abandoned. This is helpful in avoiding transferring selection information of motion information predictors in a bitstream, is helpful in improving coding efficiency, and is also helpful in significantly reducing calculation complexity of picture prediction based on a non-translational motion model.
- Further, prediction directions of the MVPs of the K pixel samples in the current picture block are normalized, reference frame indexes are processed, and motion vectors corresponding to the MVPs of the K pixel samples are scaled and mapped to a same reference frame according to reference frame indexes. The foregoing solution is helpful in resolving a problem that inter-frame prediction accuracy is low in a current video coding standard when a video includes content of a non-translational feature, for example, lens zooming, object deformation, or moving object rotation. In addition, a disadvantage that an affine transformation technology proposed in the conventional technology is hardly applied to multi-direction prediction and multiple-reference-frames-based prediction is overcome, and this is helpful in making full use of a temporary correlation between video sequences to remove temporary redundancy, thereby improving coding efficiency.
- Referring to
FIG. 6 ,FIG. 6 is a schematic flowchart of another picture prediction method according to another embodiment of the present application. As shown in an example ofFIG. 6 , the another picture prediction method provided in the another embodiment of the present application may include the following steps. - 601. Determine motion vector predictors of K pixel samples in a current picture block.
- In an example of this embodiment, K=3.
- In this embodiment, the K pixel samples include an upper left vertex, an upper right vertex, and a lower left vertex of the current picture block.
- For a specific manner of determining the motion vector predictors of the K pixel samples in the current picture block in
step 601, refer to any manner in the foregoing embodiment. - 602. Perform motion estimation on the K pixel samples by using the motion vector predictors of the K pixel samples, to obtain motion vectors of the K pixel samples.
- When reference frames corresponding to motion vector predictors of K1 pixel samples in the K pixel samples are not a reference frame Y1, the performing motion estimation on the K pixel samples by using the motion vector predictors of the K pixel samples, to obtain motion vectors of the K pixel samples includes: scaling the motion vector predictors of the K1 pixel samples to the reference frame Y1 (that is, performing normalization processing on the motion vector predictors of the K1 pixel samples); and performing motion estimation on the K pixel samples by using motion vector predictors of K2 pixel samples and motion vector predictors that are of the K1 pixel samples and that are scaled to the reference frame Y1, to obtain the motion vectors of the K pixel samples.
- Before normalization processing is performed on the motion vector predictors of the K1 pixel samples, a reference frame index of a motion vector predictor of a pixel sample i may be first revised. The pixel sample i is any pixel sample (for example, the lower left vertex or the upper right vertex), except the upper left vertex of the current picture block, in the K pixel samples. If a prediction direction corresponding to the motion vector predictor of the pixel sample i is different from a prediction direction corresponding to the motion vector predictor of the upper left vertex (for example, the prediction direction corresponding to the motion vector predictor of the pixel sample i is forward prediction, and the prediction direction corresponding to the motion vector predictor of the upper left vertex is backward prediction; or the prediction direction corresponding to the motion vector predictor of the pixel sample i is backward prediction, and the prediction direction corresponding to the motion vector predictor of the upper left vertex is forward prediction), a reference frame index corresponding to the motion vector predictor of the pixel sample i is 0.
- Particularly, if reference frames corresponding to the motion vector predictors of the K pixel samples are the reference frame Y1, no normalization processing is performed on the motion vector predictors of the K1 pixel samples (because the motion vector predictors of the K1 pixel samples have been in a normalized state), and motion estimation may be performed on the K pixel samples directly by using the motion vector predictors of the K pixel samples, to obtain the motion vectors of the K pixel samples.
- Optionally, that motion estimation is performed on the K pixel samples to obtain the motion vectors of the K pixel samples may include: motion estimation is performed on the K pixel samples based on an iterative search algorithm, to obtain the motion vectors of the K pixel samples. For a specific iterative search algorithm, refer to related descriptions in the foregoing embodiment. Details are not described herein again.
- Optionally, in some possible implementation manners of the present application, subtraction may be further performed between the motion vectors that are of the K pixel samples and that are obtained by means of motion estimation and the corresponding MVPs, to obtain MVDs of the K pixel samples, and the MVDs of the K pixel samples may be written into a video bitstream.
- Further, a syntactic element affine_no_MVD_flag may be used to indicate whether to write an MVD corresponding to a pixel sample into the video bitstream. For example, if the MVDs corresponding to the foregoing K pixel samples are all 0, affine_no_MVD_flag may be set to true, and no MVD is written into the video bitstream; or if the MVDs corresponding to the foregoing K pixel samples are not all 0, affine_no_MVD_flag may be set to false, and the MVDs corresponding to the K pixel samples are written into the video bitstream.
- 603. Calculate, based on an affine transformation model and the motion vectors of the K pixel samples, a motion vector of each pixel in the current picture block.
- Further, because motion information of each pixel in the current picture block may be used for motion vector prediction for another picture block, one piece of motion information may be stored for each 4×4 (or 8×8) block, and a motion vector of each 4×4 (or 8×8) block is stored as motion information of a pixel on an upper left corner or any other pixel in the 4×4 (or 8×8) block.
- The motion information may include a motion vector, a prediction direction corresponding to the motion vector, and a reference frame index corresponding to the motion vector.
- It should be noted that in this embodiment, the affine transformation model is mainly used as an example of a non-translational motion model for description, and a mechanism for calculating, based on another non-translational motion model and the motion vectors of the K pixel samples, the motion vector of each pixel in the current picture block has a similar principle although formulas different from those used in this embodiment may be used in the mechanism. Examples are not listed one by one herein.
- 604. Calculate a rate-distortion cost of using an affine transformation-based inter-frame prediction mode (step 601 to step 603) for the current prediction unit, and compare the rate-distortion cost with a calculated rate-distortion cost of using an HEVC inter-frame prediction mode.
- If the rate-distortion cost of using the affine transformation-based inter-frame prediction mode is less than the rate-distortion cost of using the HEVC inter-frame prediction mode, an optimal inter-frame prediction mode of the current picture block is set to the affine transformation-based inter-frame prediction mode. Specifically, a syntactic element affine MC flag may be set to true to indicate that the optimal inter-frame prediction mode of the current picture block is the affine transformation-based inter-frame prediction mode. If the rate-distortion cost of using the affine transformation-based inter-frame prediction mode is not less than the rate-distortion cost of using the HEVC inter-frame prediction mode, a syntactic element affine_MC_flag is set to false to indicate that an optimal inter-frame prediction mode of the current picture block is the HEVC inter-frame prediction mode.
- 605. Perform, based on the calculated motion vector of each pixel in the current picture block, pixel value prediction on the current picture block.
- For example, for a luminance component, if the motion vector of the pixel is ¼ pixel precision, a predicted pixel value of each pixel may be obtained by using an HEVC interpolation filter; or if the motion vector of the pixel is greater than ¼ pixel precision, predicted pixel values of four ¼ pixels at the shortest distance from the pixel are first obtained by using an HEVC interpolation filter, and then, bilinear interpolation is performed to obtain a predicted pixel value of the pixel. For a chrominance component, a predicted pixel value is obtained by using the HEVC interpolation filter.
- Further, during video coding, a residual of the current picture block may be obtained by using an original pixel value of the current picture block and a predicted pixel value that is of the current picture block and that is obtained by means of pixel value prediction, and the residual of the current picture block is written into the video bitstream. Specifically, the residual of the current prediction unit is written into the video bitstream after undergoing transformation, quantization, and entropy encoding.
- It may be learned that in the technical solution in this embodiment, determined motion vector predictors of K pixel samples are referenced during pixel value prediction performed, based on a non-translational motion model, on a current picture block. The K pixel samples include a first vertex angle pixel sample in the current picture block. The motion vector predictors of the K pixel samples are directly used for prediction, a motion vector predictor of the first vertex angle pixel sample in the K pixel samples is obtained based on a motion vector of a preset first spatially adjacent picture block of the current picture block, and there is a single optional case for the motion vector predictors of the K pixel samples. Therefore, a mechanism that is used in a conventional technology and in which motion vector predictors of K pixel samples are selected from multiple optional motion vector predictor sets of the K pixel samples by means of massive calculation is abandoned. This is helpful in avoiding transferring selection information of motion information predictors in a bitstream, is helpful in improving coding efficiency, and is also helpful in significantly reducing calculation complexity of picture prediction based on a non-translational motion model.
- Further, prediction directions of the MVPs of the K pixel samples in the current picture block are normalized, reference frame indexes are processed, and motion vectors corresponding to the MVPs of the K pixel samples are scaled and mapped to a same reference frame according to reference frame indexes. Further, a motion vector of each pixel in the current picture block may be calculated according to optimal motion vectors of the K pixel samples by using an affine transformation formula. The foregoing solution is helpful in resolving a problem that inter-frame prediction accuracy is low in a current video coding standard when a video includes content of a non-translational feature, for example, lens zooming, object deformation, or moving object rotation. In addition, a disadvantage that an affine transformation technology proposed in the conventional technology is hardly applied to multi-direction prediction and multiple-reference-frames-based prediction is overcome, and this is helpful in making full use of a temporary correlation between video sequences to remove temporary redundancy, thereby improving coding efficiency.
- Referring to
FIG. 7 ,FIG. 7 is a schematic flowchart of a video decoding method according to an embodiment of the present application. As shown in an example ofFIG. 7 , the video decoding method provided in this embodiment of the present application may include the following steps. - 701. A decoder decodes a syntactic element affine_MC_flag in a video bitstream, to determine whether an affine transformation-based inter-frame prediction mode is used for a current picture block.
- 702. If the affine transformation-based inter-frame prediction mode is used for the current picture block, the decoder decodes a syntactic element affine_inter_dir_flag in the video bitstream, to determine an inter-frame prediction direction used for the current picture block.
- 703. The decoder determines MVPs of K pixel samples in the current picture block according to the syntactic element affine_inter_dir_flag.
- The K pixel samples in the current picture block may be K pixel samples in a current picture block that are described in any one of the foregoing embodiments.
- 704. The decoder determines, by decoding a syntactic element affine_no_ MVD_flag that is in the video bitstream and is corresponding to the current picture block, whether MVDs corresponding to the K pixel samples in the current picture block are 0. If affine_no_MVD_flag is true, it indicates that the MVDs corresponding to the K pixel samples in the current picture block are all 0, and motion vectors corresponding to the obtained MVPs of the K pixel samples may be used as motion vectors of the K pixel samples. If affine_no_MVD_flag is false, it indicates that the MVDs corresponding to the K pixel samples are not all 0, the MVDs corresponding to the K pixel samples are obtained by means of separate decoding, and an MVD corresponding to each pixel sample and a corresponding MVP are added to obtain a motion vector of the pixel sample.
- 705. The decoder decodes a residual coefficient of the current picture block, and performs dequantization and inverse transformation to obtain a residual pixel in the current picture block.
- 706. The decoder calculates, based on an affine transformation model and motion vectors of the K pixel samples, a motion vector of each pixel in the current picture block, and performs, based on the calculated motion vector of each pixel in the current picture block, pixel value prediction on the current picture block, to obtain a predicted pixel value of the current picture block.
- It should be noted that in this embodiment, the affine transformation model is mainly used as an example of a non-translational motion model for description, and a mechanism for calculating, based on another non-translational motion model and the motion vectors of the K pixel samples, the motion vector of each pixel in the current picture block has a similar principle although formulas different from those used in this embodiment may be used in the mechanism. Examples are not listed one by one herein.
- 707. The decoder reconstructs the current picture block according to the residual pixel in the current picture block and the predicted pixel value of the current picture block.
- Further, because motion information of each pixel in the current picture block may be used for motion vector prediction for another picture block, one piece of motion information may be stored for each 4×4 (or 8×8) block, and a motion vector of each 4×4 (or 8×8) block is stored as motion information of a pixel on an upper left corner or any other pixel in the 4×4 (or 8×8) block.
- The motion information may include a motion vector, a prediction direction corresponding to the motion vector, and a reference frame index corresponding to the motion vector.
- It may be learned that in the technical solution in this embodiment, determined motion vector predictors of K pixel samples are referenced during pixel value prediction performed, based on a non-translational motion model, on a current picture block. The K pixel samples include a first vertex angle pixel sample in the current picture block. The motion vector predictors of the K pixel samples are directly used for prediction, a motion vector predictor of the first vertex angle pixel sample in the K pixel samples is obtained based on a motion vector of a preset first spatially adjacent picture block of the current picture block, and there is a single optional case for the motion vector predictors of the K pixel samples. Therefore, a mechanism that is used in a conventional technology and in which motion vector predictors of K pixel samples are selected from multiple optional motion vector predictor sets of the K pixel samples by means of massive calculation is abandoned. This is helpful in avoiding transferring selection information of motion information predictors in a bitstream, is helpful in improving coding efficiency, and is also helpful in significantly reducing calculation complexity of picture prediction based on a non-translational motion model.
- Further, prediction directions of the MVPs of the K pixel samples in the current picture block are normalized, reference frame indexes are processed, and motion vectors corresponding to the MVPs of the K pixel samples are scaled and mapped to a same reference frame according to reference frame indexes. Further, a motion vector of each pixel in the current picture block may be calculated according to optimal motion vectors of the K pixel samples by using an affine transformation formula. The foregoing solution is helpful in resolving a problem that inter-frame prediction accuracy is low in a current video coding standard when a video includes content of a non-translational feature, for example, lens zooming, object deformation, or moving object rotation. In addition, a disadvantage that an affine transformation technology proposed in the conventional technology is hardly applied to multi-direction prediction and multiple-reference-frames-based prediction is overcome, and this is helpful in making full use of a temporary correlation between video sequences to remove temporary redundancy, thereby improving coding efficiency.
- An emulation test is performed on some solutions, provided in the embodiments of the present application, on an encoder HM11.0 of the latest video coding standard HEVC. For example, a test is performed in two configurations: a low delay (LDB) and random access (RA), and performance of the solutions in the embodiments of the present application is compared with performance of the HM11.0 solution. Six test sequences whose affine features are relatively apparent are selected for testing. The test sequences are listed in Table 1, and emulation results are shown in Table 2.
-
TABLE 1 Sequence name Resolution Frame rate Sequence length Intotree 1920 × 1080 50 3 seconds Cactus 1920 × 1080 50 3 seconds Tractor 1920 × 1080 25 3 seconds City 1280 × 720 60 5 seconds Crew 1280 × 720 60 5 seconds Slide Show 1280 × 720 20 5 seconds -
TABLE 2 LDB RA Sequence BD-Rate name Y U V Y U V Intotree −1.1% 0.1% −0.1% −1.6% −1.1% −1.7% Cactus −3.8% −2.6% −2.3% −4.7% −3.7% −3.3% Tractor −13.2% −11.3% −10.6% −17.8% −14.6% −14.5% City −1.1% −0.6% −0.1% −1.5% −1.4% −1.7% Crew −0.6% −0.2% −0.6% −0.7% −0.8% −1.0% SlideShow −7.4% −8.7% −8.6% −4.0% −4.2% −4.2% Average −4.5% −3.9% −3.7% −5.0% −4.3% −4.4% value - A performance counter Bjontegaard delta rate (BD-rate) may be obtained by means of calculation by using results of four points whose quantization parameters (QPs) are 22, 27, 32, and 37, and peak signal to noise ratios (PSNR) of three components Y, U, and V are respectively calculated.
- It may be learned that for the solutions in the embodiments of the present application, in the LDB configuration, an average BD-rate (Y) performance increases by 4.5%, and in the random access configuration, an average BD-rate (Y) performance increases by 5.0%, and even by 17.8%. That is, when PSNR quality is the same, compared with the HM11.0 solution, the solutions in the embodiments can reduce a bit rate of a maximum of 17.8%.
- The following describes some other non-translational motion models.
- An affine transformation model is a motion model that meets the following model:
-
- where a, b, c, d, e, and f are constant parameters, (x, y) are original coordinates of a pixel, (x′, y′) are coordinates obtained after affine transformation is performed on the pixel, Vx is a motion vector in an x-axis direction, and Vy is a motion vector in a y-axis direction.
- A rotary motion model may be shown as follows:
-
- where α is a rotation angle, and a and b are parameters.
- A perspective motion model may be shown as follows:
-
- where
-
- is a perspective matrix.
- A shearing motion model may be shown as follows:
-
- where
-
- is a shear matrix, and others are constant parameters.
- A scalable motion model may be shown as follows:
-
- where ρx, ρy, θx, and θy and are constant parameters.
- A quadratic motion model may be shown as follows:
-
- where u and v represent original coordinates, x and y represent coordinates obtained after transformation, and others are constant parameters.
- It may be understood that the foregoing examples are only examples of some possible representation forms of various motion models, and certainly, the various motion models may have other forms.
- The following further provides related apparatuses for implementing the foregoing solutions.
- Referring to
FIG. 8 , an embodiment of the present application provides apicture prediction apparatus 800, and thepicture prediction apparatus 800 may include: - a determining
unit 810, configured to determine motion vector predictors of K pixel samples in a current picture block, where K is an integer greater than 1, the K pixel samples include a first vertex angle pixel sample in the current picture block, a motion vector predictor of the first vertex angle pixel sample is obtained based on a motion vector of a preset first spatially adjacent picture block of the current picture block, and the first spatially adjacent picture block is spatially adjacent to the first vertex angle pixel sample; and - a predicting
unit 820, configured to perform, based on a non-translational motion model and the motion vector predictors of the K pixel samples, pixel value prediction on the current picture block. - Optionally, in some possible implementation manners of the present application, the motion vector predictor of the first vertex angle pixel sample is obtained based on the motion vector of the preset first spatially adjacent picture block of the current picture block. Specifically, the motion vector predictor of the first vertex angle pixel sample may be equal to the motion vector of the preset first spatially adjacent picture block of the current picture block (that is, the motion vector of the preset first spatially adjacent picture block of the current picture block is used as the motion vector predictor of the first vertex angle pixel sample); or a motion vector predictor obtained by performing preset transformation on the motion vector of the preset first spatially adjacent picture block of the current picture block may be used as the motion vector predictor of the first vertex angle pixel sample.
- Optionally, in some possible implementation manners of the present application, the non-translational motion model is any one of the following models: an affine transformation model, a parabolic motion model, a rotary motion model, a perspective motion model, a shearing motion model, or a scalable motion model. It may be understood that the non-translational motion model is not limited to the foregoing specific examples.
- There may be various manners of selecting the K pixel samples.
- Optionally, in some possible implementation manners of the present application, the K pixel samples include at least two pixel samples in an upper left pixel sample, an upper right pixel sample, a lower left pixel sample, or a lower right pixel sample in the current picture block.
- If the pixel sample is a pixel block, a size of the pixel block is, for example, 2×2, 1×2, 4×2, 4×4, or another size.
- Optionally, in some possible implementation manners of the present application, the upper left pixel sample in the current picture block is an upper left vertex of the current picture block or a pixel block that is in the current picture block and that includes an upper left vertex of the current picture block. The lower left pixel sample in the current picture block is a lower left vertex of the current picture block or a pixel block that is in the current picture block and that includes a lower left vertex of the current picture block. The upper right pixel sample in the current picture block is an upper right vertex of the current picture block or a pixel block that is in the current picture block and that includes an upper right vertex of the current picture block. The lower right pixel sample in the current picture block is a lower right vertex of the current picture block or a pixel block that is in the current picture block and that includes a lower right vertex of the current picture block.
- Optionally, in some possible implementation manners of the present application, the K pixel samples may further include a center pixel sample in the current picture block. The center pixel sample in the current picture block may be a center pixel in the current picture block or a pixel block that is in the current picture block and that includes a center pixel in the current picture block.
- Optionally, in some possible implementation manners of the present application, the first spatially adjacent picture block may be a picture block X1, a picture block X2, a picture block X3, or a picture block X4.
- A motion vector predictor of the upper left pixel sample in the current picture block may be obtained based on a motion vector of the picture block X1, the picture block X1 is a spatially adjacent picture block of the current picture block, and the picture block X1 is spatially adjacent to the upper left pixel sample in the current picture block.
- A motion vector predictor of the upper right pixel sample in the current picture block may be obtained based on a motion vector of the picture block X2, the picture block X2 is a spatially adjacent picture block of the current picture block, and the picture block X2 is spatially adjacent to the upper right pixel sample in the current picture block.
- A motion vector predictor of the lower left pixel sample in the current picture block may be obtained based on a motion vector of the picture block X3, the picture block X3 is a spatially adjacent picture block of the current picture block, and the picture block X3 is spatially adjacent to the lower left pixel sample in the current picture block.
- A motion vector predictor of the lower right pixel sample in the current picture block may be obtained based on a motion vector of the picture block X4, the picture block X4 is a spatially adjacent picture block of the current picture block, and the picture block X4 is spatially adjacent to the lower right pixel sample in the current picture block.
- Optionally, in some possible implementation manners of the present application, the first vertex angle pixel sample may be the upper left pixel sample in the current picture block, and the first spatially adjacent picture block may be the picture block X1. The picture block X1 is a spatially adjacent picture block on the upper left of the current picture block, or the picture block X1 is a spatially adjacent picture block on the left of the current picture block, or the picture block X1 is a spatially adjacent picture block above the current picture block.
- Optionally, in some possible implementation manners of the present application, the picture block X2 is a spatially adjacent picture block above the current picture block, or the picture block X2 is a spatially adjacent picture block on the upper right of the current picture block, or the picture block X2 is a spatially adjacent picture block on the right of the current picture block. Still optionally, an absolute value of a difference between a motion vector absolute value of the picture block X2 and a motion vector absolute value of the picture block X1 is greater than or equal to an absolute value of a difference between a motion vector absolute value of a picture block X5 and the motion vector absolute value of the picture block X1, and the picture block X5 is any one of remaining at least some spatially adjacent picture blocks, other than the picture block X2, that are of the current picture block and that are spatially adjacent to the upper right vertex. That is, a location of the picture block X2 may be determined based on a specific policy, or a location of the picture block X2 may be directly agreed.
- Optionally, in some possible implementation manners of the present application, the picture block X3 is a spatially adjacent picture block on the left of the current picture block, or the picture block X3 is a spatially adjacent picture block on the lower left of the current picture block, or the picture block X3 is a spatially adjacent picture block below the current picture block. Still optionally, an absolute value of a difference between a motion vector absolute value of the picture block X3 and the motion vector absolute value of the picture block X1 is greater than or equal to an absolute value of a difference between a motion vector absolute value of a picture block X6 and the motion vector absolute value of the picture block X1, and the picture block X6 is any one of remaining at least some spatially adjacent picture blocks, other than the picture block X3, that are of the current picture block and that are spatially adjacent to the lower left vertex. That is, a location of the picture block X3 may be determined based on a specific policy, or a location of the picture block X3 may be directly agreed.
- Optionally, in some possible implementation manners of the present application, the picture block X4 is a spatially adjacent picture block on the right of the current picture block, or the picture block X4 is a spatially adjacent picture block on the lower right of the current picture block, or the picture block X4 is a spatially adjacent picture block below the current picture block. Still optionally, an absolute value of a difference between a motion vector absolute value of the picture block X4 and the motion vector absolute value of the picture block X1 is greater than or equal to an absolute value of a difference between a motion vector absolute value of a picture block X7 and the motion vector absolute value of the picture block X1, and the picture block X7 is any one of remaining at least some spatially adjacent picture blocks, other than the picture block X4, that are of the current picture block and that are spatially adjacent to the lower left vertex.
- Referring to
FIG. 2 ,FIG. 2 shows examples of some possible locations of the picture block X1, the picture block X2, the picture block X3, and the picture block X4. - Optionally, in some possible implementation manners of the present application, the predicting
unit 820 may be configured to: if reference frames corresponding to the motion vector predictors of the K pixel samples are a reference frame Y1, perform motion estimation on the K pixel samples by using the motion vector predictors of the K pixel samples, to obtain motion vectors of the K pixel samples; and perform pixel value prediction on the current picture block by using the non-translational motion model and the motion vectors of the K pixel samples. - Optionally, the predicting
unit 820 may be specifically configured to: when reference frames corresponding to motion vector predictors of K1 pixel samples in the K pixel samples are not a reference frame Y1, scale the motion vector predictors of the K1 pixel samples to the reference frame Y1, perform motion estimation on the K pixel samples by using motion vector predictors of K2 pixel samples and motion vector predictors that are of the K1 pixel samples and that are scaled to the reference frame Y1, to obtain motion vectors of the K pixel samples, and perform pixel value prediction on the current picture block by using the non-translational motion model and the motion vectors of the K pixel samples, where the K2 pixel samples are remaining pixel samples other than the K1 pixel samples in the K pixel samples, and K1 and K2 are positive integers. - Optionally, in some possible implementation manners of the present application, the reference frame Y1 may be a reference frame corresponding to the motion vector predictor of the first vertex angle pixel sample, or the reference frame Y1 may be a reference frame corresponding to a motion vector predictor of another pixel sample in the K pixel samples.
- Optionally, in some possible implementation manners of the present application, a pixel sample i is any pixel sample except the first vertex angle pixel sample in the K pixel samples, and if a prediction direction corresponding to a motion vector predictor of the pixel sample i is different from a prediction direction corresponding to the motion vector predictor of the first vertex angle pixel sample, a reference frame index corresponding to the motion vector predictor of the pixel sample i is 0.
- Optionally, in some possible implementation manners of the present application, if reference frame indexes corresponding to the motion vector predictors of the first vertex angle pixel sample and the pixel sample i are different, the predicting
unit 820 may scale the motion vector predictor of the pixel sample i to the reference frame Y1. - Optionally, in some possible implementation manners of the present application, the predicting
unit 820 may be specifically configured to perform motion estimation on the K pixel samples based on an iterative search algorithm (or a motion estimation algorithm of another type) by using the motion vector predictors of the K2 pixel samples and the motion vector predictors that are of the K1 pixel samples and that are scaled to the reference frame Y1, to obtain the motion vectors of the K pixel samples. - Optionally, in some possible implementation manners of the present application, the performing motion estimation on the K pixel samples by using motion vector predictors of K2 pixel samples and motion vector predictors that are of the K1 pixel samples and that are scaled to the reference frame Y1, to obtain motion vectors of the K pixel samples includes: performing motion estimation on the K pixel samples based on an iterative search algorithm by using the motion vector predictors of the K2 pixel samples and the motion vector predictors that are of the K1 pixel samples and that are scaled to the reference frame Y1, to obtain the motion vectors of the K pixel samples.
- The
picture prediction apparatus 800 is applied to a video coding apparatus or thepicture prediction apparatus 800 is applied to a video decoding apparatus. - It may be understood that functions of all functional modules in the
picture prediction apparatus 800 in this embodiment may be specifically implemented according to the methods in the foregoing method embodiments. For a specific implementation process, refer to related descriptions in the foregoing method embodiments, and details are not described herein again. Thepicture prediction apparatus 800 may be any apparatus that needs to output and play a video, for example, a laptop computer, a tablet computer, a personal computer, or a mobile phone. - It may be learned that in the technical solution in this embodiment, determined motion vector predictors of K pixel samples are referenced during pixel value prediction performed, based on a non-translational motion model, on a current picture block. The K pixel samples include a first vertex angle pixel sample in the current picture block. The motion vector predictors of the K pixel samples are directly used for prediction, a motion vector predictor of the first vertex angle pixel sample in the K pixel samples is obtained based on a motion vector of a preset first spatially adjacent picture block of the current picture block, and there is a single optional case for the motion vector predictors of the K pixel samples. Therefore, a mechanism that is used in a conventional technology and in which motion vector predictors of K pixel samples are selected from multiple optional motion vector predictor sets of the K pixel samples by means of massive calculation is abandoned. This is helpful in avoiding transferring selection information of motion information predictors in a bitstream, is helpful in improving coding efficiency, and is also helpful in significantly reducing calculation complexity of picture prediction based on a non-translational motion model.
- Further, the first spatially adjacent picture block is spatially adjacent to the first vertex angle pixel sample, and the first spatially adjacent picture block of the current picture block is preset (that is, it may be considered that the first spatially adjacent picture block of the current picture block is agreed or set in advance). In this way, this is helpful in quickly determining the motion vector predictors of the K pixel samples. In addition, because redundancy is removed by making full use of a spatial correlation between picture blocks, this is helpful in further improving coding efficiency.
- Referring to
FIG. 9 ,FIG. 9 is a schematic diagram of apicture prediction apparatus 900 according to an embodiment of the present application. Thepicture prediction apparatus 900 may include at least onebus 901, at least oneprocessor 902 connected to thebus 901, and at least onememory 903 connected to thebus 901. - The
processor 902 invokes, by using thebus 901, code stored in thememory 903, so as to determine motion vector predictors of K pixel samples in a current picture block, where K is an integer greater than 1, the K pixel samples include a first vertex angle pixel sample in the current picture block, a motion vector predictor of the first vertex angle pixel sample is obtained based on a motion vector of a preset first spatially adjacent picture block of the current picture block, and the first spatially adjacent picture block is spatially adjacent to the first vertex angle pixel sample; and perform, based on a non-translational motion model and the motion vector predictors of the K pixel samples, pixel value prediction on the current picture block. - Optionally, in some possible implementation manners of the present application, the motion vector predictor of the first vertex angle pixel sample is obtained based on the motion vector of the preset first spatially adjacent picture block of the current picture block. Specifically, the motion vector predictor of the first vertex angle pixel sample may be equal to the motion vector of the preset first spatially adjacent picture block of the current picture block (that is, the motion vector of the preset first spatially adjacent picture block of the current picture block is used as the motion vector predictor of the first vertex angle pixel sample); or a motion vector predictor obtained by performing preset transformation on the motion vector of the preset first spatially adjacent picture block of the current picture block may be used as the motion vector predictor of the first vertex angle pixel sample.
- Optionally, in some possible implementation manners of the present application, the non-translational motion model is any one of the following models: an affine transformation model, a parabolic motion model, a rotary motion model, a perspective motion model, a shearing motion model, or a scalable motion model. It may be understood that the non-translational motion model is not limited to the foregoing specific examples.
- There may be various manners of selecting the K pixel samples.
- Optionally, in some possible implementation manners of the present application, the K pixel samples include at least two pixel samples in an upper left pixel sample, an upper right pixel sample, a lower left pixel sample, or a lower right pixel sample in the current picture block.
- If the pixel sample is a pixel block, a size of the pixel block is, for example, 2×2, 1×2, 4×2, 4×4, or another size.
- Optionally, in some possible implementation manners of the present application, the upper left pixel sample in the current picture block is an upper left vertex of the current picture block or a pixel block that is in the current picture block and that includes an upper left vertex of the current picture block. The lower left pixel sample in the current picture block is a lower left vertex of the current picture block or a pixel block that is in the current picture block and that includes a lower left vertex of the current picture block. The upper right pixel sample in the current picture block is an upper right vertex of the current picture block or a pixel block that is in the current picture block and that includes an upper right vertex of the current picture block. The lower right pixel sample in the current picture block is a lower right vertex of the current picture block or a pixel block that is in the current picture block and that includes a lower right vertex of the current picture block.
- Optionally, in some possible implementation manners of the present application, the K pixel samples may further include a center pixel sample in the current picture block. The center pixel sample in the current picture block may be a center pixel in the current picture block or a pixel block that is in the current picture block and that includes a center pixel in the current picture block.
- Optionally, in some possible implementation manners of the present application, the first spatially adjacent picture block may be a picture block X1, a picture block X2, a picture block X3, or a picture block X4.
- A motion vector predictor of the upper left pixel sample in the current picture block may be obtained based on a motion vector of the picture block X1, the picture block X1 is a spatially adjacent picture block of the current picture block, and the picture block X1 is spatially adjacent to the upper left pixel sample in the current picture block.
- A motion vector predictor of the upper right pixel sample in the current picture block may be obtained based on a motion vector of the picture block X2, the picture block X2 is a spatially adjacent picture block of the current picture block, and the picture block X2 is spatially adjacent to the upper right pixel sample in the current picture block.
- A motion vector predictor of the lower left pixel sample in the current picture block may be obtained based on a motion vector of the picture block X3, the picture block X3 is a spatially adjacent picture block of the current picture block, and the picture block X3 is spatially adjacent to the lower left pixel sample in the current picture block.
- A motion vector predictor of the lower right pixel sample in the current picture block may be obtained based on a motion vector of the picture block X4, the picture block X4 is a spatially adjacent picture block of the current picture block, and the picture block X4 is spatially adjacent to the lower right pixel sample in the current picture block.
- Optionally, in some possible implementation manners of the present application, the first vertex angle pixel sample may be the upper left pixel sample in the current picture block, and the first spatially adjacent picture block may be the picture block X1. The picture block X1 is a spatially adjacent picture block on the upper left of the current picture block, or the picture block X1 is a spatially adjacent picture block on the left of the current picture block, or the picture block X1 is a spatially adjacent picture block above the current picture block.
- Optionally, in some possible implementation manners of the present application, the picture block X2 is a spatially adjacent picture block above the current picture block, or the picture block X2 is a spatially adjacent picture block on the upper right of the current picture block, or the picture block X2 is a spatially adjacent picture block on the right of the current picture block. Still optionally, an absolute value of a difference between a motion vector absolute value of the picture block X2 and a motion vector absolute value of the picture block X1 is greater than or equal to an absolute value of a difference between a motion vector absolute value of a picture block X5 and the motion vector absolute value of the picture block X1, and the picture block X5 is any one of remaining at least some spatially adjacent picture blocks, other than the picture block X2, that are of the current picture block and that are spatially adjacent to the upper right vertex. That is, a location of the picture block X2 may be determined based on a specific policy, or a location of the picture block X2 may be directly agreed.
- Optionally, in some possible implementation manners of the present application, the picture block X3 is a spatially adjacent picture block on the left of the current picture block, or the picture block X3 is a spatially adjacent picture block on the lower left of the current picture block, or the picture block X3 is a spatially adjacent picture block below the current picture block. Still optionally, an absolute value of a difference between a motion vector absolute value of the picture block X3 and the motion vector absolute value of the picture block X1 is greater than or equal to an absolute value of a difference between a motion vector absolute value of a picture block X6 and the motion vector absolute value of the picture block X1, and the picture block X6 is any one of remaining at least some spatially adjacent picture blocks, other than the picture block X3, that are of the current picture block and that are spatially adjacent to the lower left vertex. That is, a location of the picture block X3 may be determined based on a specific policy, or a location of the picture block X3 may be directly agreed.
- Optionally, in some possible implementation manners of the present application, the picture block X4 is a spatially adjacent picture block on the right of the current picture block, or the picture block X4 is a spatially adjacent picture block on the lower right of the current picture block, or the picture block X4 is a spatially adjacent picture block below the current picture block. Still optionally, an absolute value of a difference between a motion vector absolute value of the picture block X4 and the motion vector absolute value of the picture block X1 is greater than or equal to an absolute value of a difference between a motion vector absolute value of a picture block X7 and the motion vector absolute value of the picture block X1, and the picture block X7 is any one of remaining at least some spatially adjacent picture blocks, other than the picture block X4, that are of the current picture block and that are spatially adjacent to the lower left vertex.
- Optionally, in some possible implementation manners of the present application, the
processor 902 may be configured to: if reference frames corresponding to the motion vector predictors of the K pixel samples are a reference frame Y1, perform motion estimation on the K pixel samples by using the motion vector predictors of the K pixel samples, to obtain motion vectors of the K pixel samples; and perform pixel value prediction on the current picture block by using the non-translational motion model and the motion vectors of the K pixel samples. - Optionally, in some possible implementation manners of the present application, the
processor 902 may be configured to: when reference frames corresponding to motion vector predictors of K1 pixel samples in the K pixel samples are not a reference frame Y1, scale the motion vector predictors of the K1 pixel samples to the reference frame Y1, perform motion estimation on the K pixel samples by using motion vector predictors of K2 pixel samples and motion vector predictors that are of the K1 pixel samples and that are scaled to the reference frame Y1, to obtain motion vectors of the K pixel samples, and perform pixel value prediction on the current picture block by using the non-translational motion model and the motion vectors of the K pixel samples, where the K2 pixel samples are remaining pixel samples other than the K1 pixel samples in the K pixel samples, and K1 and K2 are positive integers. - Optionally, in some possible implementation manners of the present application, the reference frame Y1 may be a reference frame corresponding to the motion vector predictor of the first vertex angle pixel sample, or the reference frame Y1 may be a reference frame corresponding to a motion vector predictor of another pixel sample in the K pixel samples.
- Optionally, in some possible implementation manners of the present application, a pixel sample i is any pixel sample except the first vertex angle pixel sample in the K pixel samples, and if a prediction direction corresponding to a motion vector predictor of the pixel sample i is different from a prediction direction corresponding to the motion vector predictor of the first vertex angle pixel sample, a reference frame index corresponding to the motion vector predictor of the pixel sample i is 0.
- Optionally, in some possible implementation manners of the present application, if reference frame indexes corresponding to the motion vector predictors of the first vertex angle pixel sample and the pixel sample i are different, the
processor 902 may be configured to scale the motion vector predictor of the pixel sample i to the reference frame Y1. - Optionally, in some possible implementation manners of the present application, the
processor 902 may be configured to perform motion estimation on the K pixel samples based on an iterative search algorithm (or a motion estimation algorithm of another type) by using the motion vector predictors of the K2 pixel samples and the motion vector predictors that are of the K1 pixel samples and that are scaled to the reference frame Y1, to obtain the motion vectors of the K pixel samples. - It may be understood that functions of all functional modules in the
picture prediction apparatus 900 in this embodiment may be specifically implemented according to the methods in the foregoing method embodiments. For a specific implementation process, refer to related descriptions in the foregoing method embodiments, and details are not described herein again. Thepicture prediction apparatus 900 may be any apparatus that needs to output and play a video, for example, a laptop computer, a tablet computer, a personal computer, or a mobile phone. - It may be learned that in the technical solution in this embodiment, determined motion vector predictors of K pixel samples are referenced during pixel value prediction performed, based on a non-translational motion model, on a current picture block. The K pixel samples include a first vertex angle pixel sample in the current picture block. The motion vector predictors of the K pixel samples are directly used for prediction, a motion vector predictor of the first vertex angle pixel sample in the K pixel samples is obtained based on a motion vector of a preset first spatially adjacent picture block of the current picture block, and there is a single optional case for the motion vector predictors of the K pixel samples. Therefore, a mechanism that is used in a conventional technology and in which motion vector predictors of K pixel samples are selected from multiple optional motion vector predictor sets of the K pixel samples by means of massive calculation is abandoned. This is helpful in avoiding transferring selection information of motion information predictors in a bitstream, is helpful in improving coding efficiency, and is also helpful in significantly reducing calculation complexity of picture prediction based on a non-translational motion model.
- Further, the first spatially adjacent picture block is spatially adjacent to the first vertex angle pixel sample, and the first spatially adjacent picture block of the current picture block is preset (that is, it may be considered that the first spatially adjacent picture block of the current picture block is agreed or set in advance). In this way, this is helpful in quickly determining the motion vector predictors of the K pixel samples. In addition, because redundancy is removed by making full use of a spatial correlation between picture blocks, this is helpful in further improving coding efficiency.
- Referring to
FIG. 10 ,FIG. 10 is a structural block diagram of apicture prediction apparatus 1000 according to another embodiment of the present application. Thepicture prediction apparatus 1000 may include at least oneprocessor 1001, at least onenetwork interface 1004 orother user interface 1003, amemory 1005, and at least onecommunications bus 1002. Thecommunications bus 1002 is configured to implement a connection and communication between these components. Thepicture prediction apparatus 1000 may optionally include theuser interface 1003 that includes a display (for example, a touchscreen, a liquid crystal display (LCD), a holographic imaging device, a cathode ray tube (CRT), or a Projector), a click device (for example, a mouse, a trackball, a touchpad, or a touchscreen), a camera and/or a pickup apparatus, and the like. - The
memory 1005 may include a read-only memory and a random access memory, and provide an instruction and data for theprocessor 1001. A part of thememory 1005 may further include a nonvolatile random access memory (NVRAM). - In some implementation manners, the
memory 1005 stores the following element: an executable module or a data structure, or a subset thereof, or an extension set thereof: - an
operating system 10051, including various system programs, and used to implement various basic services and process hardware-based tasks; and - an
application program module 10052, including various application programs, and used to implement various application services. - In this embodiment of the present application, by invoking a program or an instruction stored in the
memory 1005, theprocessor 1001 is configured to: determine motion vector predictors of K pixel samples in a current picture block, where K is an integer greater than 1, the K pixel samples include a first vertex angle pixel sample in the current picture block, a motion vector predictor of the first vertex angle pixel sample is obtained based on a motion vector of a preset first spatially adjacent picture block of the current picture block, and the first spatially adjacent picture block is spatially adjacent to the first vertex angle pixel sample; and perform, based on a non-translational motion model and the motion vector predictors of the K pixel samples, pixel value prediction on the current picture block. - Optionally, in some possible implementation manners of the present application, the motion vector predictor of the first vertex angle pixel sample is obtained based on the motion vector of the preset first spatially adjacent picture block of the current picture block. Specifically, the motion vector predictor of the first vertex angle pixel sample may be equal to the motion vector of the preset first spatially adjacent picture block of the current picture block (that is, the motion vector of the preset first spatially adjacent picture block of the current picture block is used as the motion vector predictor of the first vertex angle pixel sample); or a motion vector predictor obtained by performing preset transformation on the motion vector of the preset first spatially adjacent picture block of the current picture block may be used as the motion vector predictor of the first vertex angle pixel sample.
- Optionally, in some possible implementation manners of the present application, the non-translational motion model is any one of the following models: an affine transformation model, a parabolic motion model, a rotary motion model, a perspective motion model, a shearing motion model, or a scalable motion model. It may be understood that the non-translational motion model is not limited to the foregoing specific examples.
- There may be various manners of selecting the K pixel samples.
- Optionally, in some possible implementation manners of the present application, the K pixel samples include at least two pixel samples in an upper left pixel sample, an upper right pixel sample, a lower left pixel sample, or a lower right pixel sample in the current picture block.
- If the pixel sample is a pixel block, a size of the pixel block is, for example, 2×2, 1×2, 4×2, 4×4, or another size.
- Optionally, in some possible implementation manners of the present application, the upper left pixel sample in the current picture block is an upper left vertex of the current picture block or a pixel block that is in the current picture block and that includes an upper left vertex of the current picture block. The lower left pixel sample in the current picture block is a lower left vertex of the current picture block or a pixel block that is in the current picture block and that includes a lower left vertex of the current picture block. The upper right pixel sample in the current picture block is an upper right vertex of the current picture block or a pixel block that is in the current picture block and that includes an upper right vertex of the current picture block. The lower right pixel sample in the current picture block is a lower right vertex of the current picture block or a pixel block that is in the current picture block and that includes a lower right vertex of the current picture block.
- Optionally, in some possible implementation manners of the present application, the K pixel samples may further include a center pixel sample in the current picture block. The center pixel sample in the current picture block may be a center pixel in the current picture block or a pixel block that is in the current picture block and that includes a center pixel in the current picture block.
- Optionally, in some possible implementation manners of the present application, the first spatially adjacent picture block may be a picture block X1, a picture block X2, a picture block X3, or a picture block X4.
- A motion vector predictor of the upper left pixel sample in the current picture block may be obtained based on a motion vector of the picture block X1, the picture block X1 is a spatially adjacent picture block of the current picture block, and the picture block X1 is spatially adjacent to the upper left pixel sample in the current picture block.
- A motion vector predictor of the upper right pixel sample in the current picture block may be obtained based on a motion vector of the picture block X2, the picture block X2 is a spatially adjacent picture block of the current picture block, and the picture block X2 is spatially adjacent to the upper right pixel sample in the current picture block.
- A motion vector predictor of the lower left pixel sample in the current picture block may be obtained based on a motion vector of the picture block X3, the picture block X3 is a spatially adjacent picture block of the current picture block, and the picture block X3 is spatially adjacent to the lower left pixel sample in the current picture block.
- A motion vector predictor of the lower right pixel sample in the current picture block may be obtained based on a motion vector of the picture block X4, the picture block X4 is a spatially adjacent picture block of the current picture block, and the picture block X4 is spatially adjacent to the lower right pixel sample in the current picture block.
- Optionally, in some possible implementation manners of the present application, the first vertex angle pixel sample may be the upper left pixel sample in the current picture block, and the first spatially adjacent picture block may be the picture block X1. The picture block X1 is a spatially adjacent picture block on the upper left of the current picture block, or the picture block X1 is a spatially adjacent picture block on the left of the current picture block, or the picture block X1 is a spatially adjacent picture block above the current picture block.
- Optionally, in some possible implementation manners of the present application, the picture block X2 is a spatially adjacent picture block above the current picture block, or the picture block X2 is a spatially adjacent picture block on the upper right of the current picture block, or the picture block X2 is a spatially adjacent picture block on the right of the current picture block. Still optionally, an absolute value of a difference between a motion vector absolute value of the picture block X2 and a motion vector absolute value of the picture block X1 is greater than or equal to an absolute value of a difference between a motion vector absolute value of a picture block X5 and the motion vector absolute value of the picture block X1, and the picture block X5 is any one of remaining at least some spatially adjacent picture blocks, other than the picture block X2, that are of the current picture block and that are spatially adjacent to the upper right vertex. That is, a location of the picture block X2 may be determined based on a specific policy, or a location of the picture block X2 may be directly agreed.
- Optionally, in some possible implementation manners of the present application, the picture block X3 is a spatially adjacent picture block on the left of the current picture block, or the picture block X3 is a spatially adjacent picture block on the lower left of the current picture block, or the picture block X3 is a spatially adjacent picture block below the current picture block. Still optionally, an absolute value of a difference between a motion vector absolute value of the picture block X3 and the motion vector absolute value of the picture block X1 is greater than or equal to an absolute value of a difference between a motion vector absolute value of a picture block X6 and the motion vector absolute value of the picture block X1, and the picture block X6 is any one of remaining at least some spatially adjacent picture blocks, other than the picture block X3, that are of the current picture block and that are spatially adjacent to the lower left vertex. That is, a location of the picture block X3 may be determined based on a specific policy, or a location of the picture block X3 may be directly agreed.
- Optionally, in some possible implementation manners of the present application, the picture block X4 is a spatially adjacent picture block on the right of the current picture block, or the picture block X4 is a spatially adjacent picture block on the lower right of the current picture block, or the picture block X4 is a spatially adjacent picture block below the current picture block. Still optionally, an absolute value of a difference between a motion vector absolute value of the picture block X4 and the motion vector absolute value of the picture block X1 is greater than or equal to an absolute value of a difference between a motion vector absolute value of a picture block X7 and the motion vector absolute value of the picture block X1, and the picture block X7 is any one of remaining at least some spatially adjacent picture blocks, other than the picture block X4, that are of the current picture block and that are spatially adjacent to the lower left vertex.
- Optionally, in some possible implementation manners of the present application, the
processor 1001 may be configured to: if reference frames corresponding to the motion vector predictors of the K pixel samples are a reference frame Y1, perform motion estimation on the K pixel samples by using the motion vector predictors of the K pixel samples, to obtain motion vectors of the K pixel samples; and perform pixel value prediction on the current picture block by using the non-translational motion model and the motion vectors of the K pixel samples. - Optionally, in some possible implementation manners of the present application, the
processor 1001 may be configured to: when reference frames corresponding to motion vector predictors of K1 pixel samples in the K pixel samples are not a reference frame Y1, scale the motion vector predictors of the K1 pixel samples to the reference frame Y1, perform motion estimation on the K pixel samples by using motion vector predictors of K2 pixel samples and motion vector predictors that are of the K1 pixel samples and that are scaled to the reference frame Y1, to obtain motion vectors of the K pixel samples, and perform pixel value prediction on the current picture block by using the non-translational motion model and the motion vectors of the K pixel samples, where the K2 pixel samples are remaining pixel samples other than the K1 pixel samples in the K pixel samples, and K1 and K2 are positive integers. - Optionally, in some possible implementation manners of the present application, the reference frame Y1 may be a reference frame corresponding to the motion vector predictor of the first vertex angle pixel sample, or the reference frame Y1 may be a reference frame corresponding to a motion vector predictor of another pixel sample in the K pixel samples.
- Optionally, in some possible implementation manners of the present application, a pixel sample i is any pixel sample except the first vertex angle pixel sample in the K pixel samples, and if a prediction direction corresponding to a motion vector predictor of the pixel sample i is different from a prediction direction corresponding to the motion vector predictor of the first vertex angle pixel sample, a reference frame index corresponding to the motion vector predictor of the pixel sample i is 0.
- Optionally, in some possible implementation manners of the present application, if reference frame indexes corresponding to the motion vector predictors of the first vertex angle pixel sample and the pixel sample i are different, the
processor 1001 may be configured to scale the motion vector predictor of the pixel sample i to the reference frame Y1. - Optionally, in some possible implementation manners of the present application, the
processor 1001 may be configured to perform motion estimation on the K pixel samples based on an iterative search algorithm (or a motion estimation algorithm of another type) by using the motion vector predictors of the K2 pixel samples and the motion vector predictors that are of the K1 pixel samples and that are scaled to the reference frame Y1, to obtain the motion vectors of the K pixel samples. - It may be understood that functions of all functional modules in the
picture prediction apparatus 1000 in this embodiment may be specifically implemented according to the methods in the foregoing method embodiments. For a specific implementation process, refer to related descriptions in the foregoing method embodiments, and details are not described herein again. Thepicture prediction apparatus 1000 may be any apparatus that needs to output and play a video, for example, a laptop computer, a tablet computer, a personal computer, or a mobile phone. - It may be learned that in the technical solution in this embodiment, determined motion vector predictors of K pixel samples are referenced during pixel value prediction performed, based on a non-translational motion model, on a current picture block. The K pixel samples include a first vertex angle pixel sample in the current picture block. The motion vector predictors of the K pixel samples are directly used for prediction, a motion vector predictor of the first vertex angle pixel sample in the K pixel samples is obtained based on a motion vector of a preset first spatially adjacent picture block of the current picture block, and there is a single optional case for the motion vector predictors of the K pixel samples. Therefore, a mechanism that is used in a conventional technology and in which motion vector predictors of K pixel samples are selected from multiple optional motion vector predictor sets of the K pixel samples by means of massive calculation is abandoned. This is helpful in avoiding transferring selection information of motion information predictors in a bitstream, is helpful in improving coding efficiency, and is also helpful in significantly reducing calculation complexity of picture prediction based on a non-translational motion model.
- Further, the first spatially adjacent picture block is spatially adjacent to the first vertex angle pixel sample, and the first spatially adjacent picture block of the current picture block is preset (that is, it may be considered that the first spatially adjacent picture block of the current picture block is agreed or set in advance). In this way, this is helpful in quickly determining the motion vector predictors of the K pixel samples. In addition, because redundancy is removed by making full use of a spatial correlation between picture blocks, this is helpful in further improving coding efficiency.
- It should be noted that, for ease of description, the foregoing method embodiments are expressed as a series of actions. However, persons skilled in the art should appreciate that the present application is not limited to the described action sequence, because according to the present application, some steps may be performed in other sequences or performed simultaneously. In addition, persons skilled in the art should also appreciate that all the embodiments described in the specification are examples of embodiments, and the related actions and modules are not necessarily mandatory to the present application.
- In the foregoing embodiments, all the embodiments have respective focuses of description. For a part that is not described in detail in an embodiment, refer to related descriptions in other embodiments.
- In the embodiments provided in this application, it should be understood that the disclosed apparatus may be implemented in other manners. For example, the described apparatus embodiment is merely an example. For example, the unit division is merely logical function division and may be other division in actual implementation. For example, a plurality of units or components may be combined or integrated into another system, or some features may be ignored or not performed. In addition, the displayed or discussed mutual couplings or direct couplings or communication connections may be implemented through some interfaces. The indirect couplings or communication connections between the apparatuses or units may be implemented in electronic or other forms.
- The units described as separate parts may or may not be physically separate, and parts displayed as units may or may not be physical units, may be located in one position, or may be distributed on a plurality of network units. Some or all of the units may be selected according to actual needs to achieve the objectives of the solutions in the embodiments.
- In addition, functional units in the embodiments of the present application may be integrated into one processing unit, or each of the units may exist alone physically, or two or more units are integrated into one unit. The integrated unit may be implemented in a form of hardware, or may be implemented in a form of a software functional unit.
- When the foregoing integrated unit is implemented in the form of a software functional unit and sold or used as an independent product, the integrated unit may be stored in a computer-readable storage medium. Based on such an understanding, the technical solutions in the present application essentially, or the part contributing to the prior art, or all or some of the technical solutions may be implemented in the form of a software product. The software product is stored in a storage medium and includes instructions for instructing a computer device (which may be a personal computer, a server, or a network device) to perform all or some of the steps of the methods described in the embodiments of the present application. The foregoing storage medium includes: any medium that can store program code, such as a Universal Serial Bus (USB) flash drive, a read-only memory (ROM), a random access memory (RAM), a removable hard disk, a magnetic disk, or an optical disc.
- The foregoing embodiments are merely intended for describing the technical solutions in the present application, but not for limiting the present application. Although the present application is described in detail with reference to the foregoing embodiments, persons of ordinary skill in the art should understand that they may still make modifications to the technical solutions described in the foregoing embodiments or make equivalent replacements to some technical features thereof, without departing from the spirit and scope of the technical solutions in the embodiments of the present application. Additionally, statements made herein characterizing the invention refer to an embodiment of the invention and not necessarily all embodiments.
Claims (22)
1. A picture prediction method for a video coding process, the method comprising:
determining, by an encoding apparatus, motion vector predictors (MVPs) of K pixel samples in a current picture block, wherein K is an integer greater than 1;
obtaining, by the encoding apparatus, respective motion vectors of the K pixel samples via motion estimation;
obtaining, by the encoding apparatus, motion vector differences (MVDs) of the K pixel samples based on the respective motion vectors of the K pixel samples and corresponding MVPs;
based on the MVDs of the K pixel samples all being zero, writing, by the encoding apparatus, a flag into a bitstream, wherein the value of the flag indicates that the MVDs of the K pixel samples all being zero; and
calculating, based on a non-translational motion model , a motion vector of each pixel or each sub-block in the current picture block; and
performing, based on the calculated motion vector of each pixel or each sub-block in the current picture block, pixel value prediction on the current picture block.
wherein the non-translational motion model is represented by vx=a×x+b×y+e and vy=c×x+d×y+f, where (x,y) corresponds to the coordinates of a respective pixel or sub-block in the current picture block, and a, b, e, c, d and f are parameters of the non-translational motion model and are derived based on the respective motion vectors of the K pixel samples of the current picture block and a width and a height of the current picture bloc and wherein (vx, vy) represents a respective motion vector of a respective pixel or sub-block calculated based on the non-translational motion model.
2. The method according to claim 1 , wherein:
the K pixel samples comprise at least two pixel samples in an upper left pixel sample, an upper right pixel sample, or a lower left pixel sample in the current picture block; and
the upper left pixel sample in the current picture block is an upper left vertex of the current picture block or a pixel block that is in the current picture block and that comprises an upper left vertex of the current picture block,
the lower left pixel sample in the current picture block is a lower left vertex of the current picture block or a pixel block that is in the current picture block and that comprises a lower left vertex of the current picture block, and
the upper right pixel sample in the current picture block is an upper right vertex of the current picture block or a pixel block that is in the current picture block and that comprises an upper right vertex of the current picture block.
3. The method according to claim 1 , wherein a respective motion vector of a respective pixel whose coordinates are (x, y) is calculated as follows:
wherein the size of the current picture block is S1×S2, and wherein the respective motion vectors of the K pixel samples comprises motion vectors (vx0, vy0), (vx1, vy1), and (vx2, vy2) of three vertices whose coordinates are (0, 0) (S1, 0) and (0, S2) where S1 is the width of the current picture block and S2 is the height of the current picture block, and wherein (vx, vy) represents a respective motion vector of a respective pixel or a sub-block calculated based on the non-translational motion model.
4. The method according to claim 1 , wherein the flag indicates:
whether or not any MVD of a pixel sample of the K pixel samples is written into the bitstream; or
whether or not all MVDs of the K pixel samples in the current picture block are 0.
5. The method according to claim 1 , wherein obtaining the MVDs of the K pixel samples comprises:
obtaining the MVDs of the K pixel samples by performing subtraction between the respective motion vectors of the K pixel samples and the corresponding MVPs.
6. The method according to claim 1 , wherein the non-translational motion model is an affine transformation model, a parabolic motion model, a rotary motion model, a perspective motion model, a shearing motion model, or a scalable motion model.
7. An encoding apparatus, comprising:
a non-transitory memory having processor-executable instructions stored thereon; and
a processor, coupled to the non-transitory memory, configured to execute the processor-executable instructions to facilitate:
determining motion vector predictors (MVPs) of K pixel samples in a current picture block, wherein K is an integer greater than 1;
obtaining respective motion vectors of the K pixel samples via motion estimation;
obtaining motion vector differences (MVDs) of the K pixel samples based on the respective motion vectors of the K pixel samples and corresponding MVPs;
based on the MVDs of the K pixel samples all being zero, writing a flag into a bitstream, wherein the value of the flag indicates that the MVDs of the K pixel samples all being zero;
calculating, based on a non-translational motion model , a motion vector of each pixel or each sub-block in the current picture block; and
performing, based on the calculated motion vector of each pixel or each sub-block in the current picture block, pixel value prediction on the current picture block.
wherein the non-translational motion model is represented by vx=a×x+b×y+e and vy=c×x+d×y+f, where (x,y) corresponds to the coordinates of a respective pixel or sub-block in the current picture block, and a, b, e, c, d and f are parameters of the non-translational motion model and are derived based on the respective motion vectors of the K pixel samples of the current picture block and a width and a height of the current picture bloc and wherein (vx, vy) represents a respective motion vector of a respective pixel or sub-block calculated based on the non-translational motion model.
8. The apparatus according to claim 7 , wherein:
the K pixel samples comprise at least two pixel samples in an upper left pixel sample, an upper right pixel sample, or a lower left pixel sample in the current picture block; and
the upper left pixel sample in the current picture block is an upper left vertex of the current picture block or a pixel block that is in the current picture block and that comprises an upper left vertex of the current picture block,
the lower left pixel sample in the current picture block is a lower left vertex of the current picture block or a pixel block that is in the current picture block and that comprises a lower left vertex of the current picture block, and
the upper right pixel sample in the current picture block is an upper right vertex of the current picture block or a pixel block that is in the current picture block and that comprises an upper right vertex of the current picture block.
9. The apparatus according to claim 7 , wherein a respective motion vector of a respective pixel whose coordinates are (x, y) is calculated as follows:
wherein the size of the current picture block is S1×S2, and wherein the respective motion vectors of the K pixel samples comprises motion vectors (vx0, vy0), (vx1, vy1), and (vx2, vy2) of three vertices whose coordinates are (0, 0) (S1, 0) and (0, S2) where S1 is the width of the current picture block and S2 is the height of the current picture block, and wherein (vx, vy) represents a respective motion vector of a respective pixel or a sub-block calculated based on the non-translational motion model.
10. The apparatus according to claim 7 , wherein the flag indicates:
whether or not any MVD of a pixel sample of the K pixel samples is written into the bitstream; or
whether or not all MVDs of the K pixel samples in the current picture block are 0.
11. The apparatus according to claim 7 , wherein obtaining the MVDs of the K pixel samples comprises obtaining the MVDs of the K pixel samples by performing subtraction between the respective motion vectors of the K pixel samples and the corresponding MVPs.
12. The apparatus according to claim 7 , wherein the non-translational motion model is an affine transformation model, a parabolic motion model, a rotary motion model, a perspective motion model, a shearing motion model, or a scalable motion model.
13. A picture prediction method for a video decoding process, comprising:
determining, by a decoding apparatus, motion vector predictors (MVPs) of K pixel samples in a current picture block, wherein K is an integer greater than 1;
decoding a flag of the current picture block from a bitstream;
based on the value of the flag indicating that the MVDs of the K pixel samples all being zero, determining the MVDs of the K pixel samples are all 0; and obtaining respective motion vectors of the K pixel samples based on the MVDs of the K pixel samples and corresponding MVPs of the K pixel samples;
calculating, based on a non-translational motion model, a motion vector of each pixel or each sub-block in the current picture block; and
performing, based on the calculated motion vector of each pixel or each sub-block in the current picture block, pixel value prediction on the current picture block.
wherein the non-translational motion model is represented by vx=a×x+b×y+e and vy=c×x+d×y+f, where (x,y) corresponds to the coordinates of a respective pixel or sub-block in the current picture block, and a, b, e, c, d and f are parameters of the non-translational motion model and are derived based on the respective motion vectors of the K pixel samples of the current picture block and a width and a height of the current picture bloc and wherein (vx, vy) represents a respective motion vector of a respective pixel or sub-block calculated based on the non-translational motion model.
14. The method according to claim 13 , wherein:
the K pixel samples comprise at least two pixel samples in an upper left pixel sample, an upper right pixel sample, or a lower left pixel sample in the current picture block; and
the upper left pixel sample in the current picture block is an upper left vertex of the current picture block or a pixel block that is in the current picture block and that comprises an upper left vertex of the current picture block,
the lower left pixel sample in the current picture block is a lower left vertex of the current picture block or a pixel block that is in the current picture block and that comprises a lower left vertex of the current picture block, and
the upper right pixel sample in the current picture block is an upper right vertex of the current picture block or a pixel block that is in the current picture block and that comprises an upper right vertex of the current picture block.
15. The method according to claim 13 , wherein a respective motion vector of a respective pixel whose coordinates are (x, y) is calculated as follows:
wherein the size of the current picture block is S1×S2, and wherein the respective motion vectors of the K pixel samples comprises motion vectors (vx0, vy0), (vx1, vy1), and (vx2, vy2) of three vertices whose coordinates are (0, 0) (S1, 0) and (0, S2) where S1 is the width of the current picture block and S2 is the height of the current picture block, and wherein (vx, vy) represents a respective motion vector of a respective pixel or a sub-block calculated based on the non-translational motion model.
16. The method according to claim 13 , wherein the flag indicates:
whether or not any MVD of a pixel sample of the K pixel samples is written into the bitstream; or
whether or not all MVDs of the K pixel samples in the current picture block are 0.
17. The method according to claim 13 , wherein the non-translational motion model is an affine transformation model, a parabolic motion model, a rotary motion model, a perspective motion model, a shearing motion model, or a scalable motion model.
18. A decoding apparatus, comprising:
a non-transitory memory having processor-executable instructions stored thereon; and
a processor, coupled to the non-transitory memory, configured to execute the processor-executable instructions to facilitate:
determining motion vector predictors (MVPs) of K pixel samples in a current picture block, wherein K is an integer greater than 1;
decoding a flag of the current picture block from a bitstream;
based on the value of the flag indicating that the MVDs of the K pixel samples all being zero, determining the MVDs of the K pixel samples are all 0; and obtaining respective motion vectors of the K pixel samples based on the MVDs of the K pixel samples and corresponding MVPs of the K pixel samples;
calculating, based on a non-translational motion model, a motion vector of each pixel or each sub-block in the current picture block; and
performing, based on the calculated motion vector of each pixel or each sub-block in the current picture block, pixel value prediction on the current picture block.
wherein the non-translational motion model is represented by vx=a×x+b×y+e and vy=c×x+d×y+f, where (x,y) corresponds to the coordinates of a respective pixel or sub-block in the current picture block, and a, b, e, c, d and f are parameters of the non-translational motion model and are derived based on the respective motion vectors of the K pixel samples of the current picture block and a width and a height of the current picture bloc and wherein (vx, vy) represents a respective motion vector of a respective pixel or sub-block calculated based on the non-translational motion model.
19. The apparatus according to claim 18 , wherein:
the K pixel samples comprise at least two pixel samples in an upper left pixel sample, an upper right pixel sample, or a lower left pixel sample in the current picture block; and
the upper left pixel sample in the current picture block is an upper left vertex of the current picture block or a pixel block that is in the current picture block and that comprises an upper left vertex of the current picture block,
the lower left pixel sample in the current picture block is a lower left vertex of the current picture block or a pixel block that is in the current picture block and that comprises a lower left vertex of the current picture block, and
the upper right pixel sample in the current picture block is an upper right vertex of the current picture block or a pixel block that is in the current picture block and that comprises an upper right vertex of the current picture block.
20. The apparatus according to claim 18 , wherein a respective motion vector of a respective pixel whose coordinates are (x, y) is calculated as follows:
wherein the size of the current picture block is S1×S2, and wherein the respective motion vectors of the K pixel samples comprises motion vectors (vx0, vy0), (vx1, vy1), and (vx2, vy2) of three vertices whose coordinates are (0, 0) (S1, 0) and (0, S2) where S1 is the width of the current picture block and S2 is the height of the current picture block, and wherein (vx, vy) represents a respective motion vector of a respective pixel or a sub-block calculated based on the non-translational motion model.
21. The apparatus according to claim 18 , wherein the flag indicates:
whether or not any MVD of a pixel sample of the K pixel samples is written into the bitstream; or
whether or not all MVDs of the K pixel samples in the current picture block are 0.
22. The apparatus according to claim 18 , wherein the non-translational motion model is an affine transformation model, a parabolic motion model, a rotary motion model, a perspective motion model, a shearing motion model, or a scalable motion model.
Priority Applications (1)
Application Number | Priority Date | Filing Date | Title |
---|---|---|---|
US17/025,901 US20210006818A1 (en) | 2014-09-30 | 2020-09-18 | Picture prediction method and related apparatus |
Applications Claiming Priority (5)
Application Number | Priority Date | Filing Date | Title |
---|---|---|---|
CN201410526608.8 | 2014-09-30 | ||
CN201410526608.8A CN104539966B (en) | 2014-09-30 | 2014-09-30 | Image prediction method and relevant apparatus |
PCT/CN2015/076659 WO2016050051A1 (en) | 2014-09-30 | 2015-04-15 | Image prediction method and relevant device |
US15/461,274 US10827194B2 (en) | 2014-09-30 | 2017-03-16 | Picture prediction method and related apparatus |
US17/025,901 US20210006818A1 (en) | 2014-09-30 | 2020-09-18 | Picture prediction method and related apparatus |
Related Parent Applications (1)
Application Number | Title | Priority Date | Filing Date |
---|---|---|---|
US15/461,274 Continuation US10827194B2 (en) | 2014-09-30 | 2017-03-16 | Picture prediction method and related apparatus |
Publications (1)
Publication Number | Publication Date |
---|---|
US20210006818A1 true US20210006818A1 (en) | 2021-01-07 |
Family
ID=52855412
Family Applications (2)
Application Number | Title | Priority Date | Filing Date |
---|---|---|---|
US15/461,274 Active 2035-09-15 US10827194B2 (en) | 2014-09-30 | 2017-03-16 | Picture prediction method and related apparatus |
US17/025,901 Abandoned US20210006818A1 (en) | 2014-09-30 | 2020-09-18 | Picture prediction method and related apparatus |
Family Applications Before (1)
Application Number | Title | Priority Date | Filing Date |
---|---|---|---|
US15/461,274 Active 2035-09-15 US10827194B2 (en) | 2014-09-30 | 2017-03-16 | Picture prediction method and related apparatus |
Country Status (7)
Country | Link |
---|---|
US (2) | US10827194B2 (en) |
EP (3) | EP3836546A1 (en) |
JP (3) | JP6546277B2 (en) |
KR (3) | KR102060184B1 (en) |
CN (4) | CN112087630B (en) |
BR (2) | BR122020019754B1 (en) |
WO (1) | WO2016050051A1 (en) |
Families Citing this family (36)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
CN107534770B (en) | 2015-03-10 | 2019-11-05 | 华为技术有限公司 | Image prediction method and relevant device |
CN109005407B (en) * | 2015-05-15 | 2023-09-01 | 华为技术有限公司 | Video image encoding and decoding method, encoding device and decoding device |
CN106331722B (en) | 2015-07-03 | 2019-04-26 | 华为技术有限公司 | Image prediction method and relevant device |
WO2017022973A1 (en) * | 2015-08-04 | 2017-02-09 | 엘지전자 주식회사 | Inter prediction method and device in video coding system |
CN108965869B (en) | 2015-08-29 | 2023-09-12 | 华为技术有限公司 | Image prediction method and device |
CN108965871B (en) * | 2015-09-29 | 2023-11-10 | 华为技术有限公司 | Image prediction method and device |
WO2017118411A1 (en) * | 2016-01-07 | 2017-07-13 | Mediatek Inc. | Method and apparatus for affine inter prediction for video coding system |
CN107027040B9 (en) | 2016-01-29 | 2020-08-28 | 华为技术有限公司 | Filtering method and device for removing blocking effect |
CN111526361B (en) * | 2016-02-06 | 2022-05-13 | 华为技术有限公司 | Image coding and decoding method and device |
ES2737841B1 (en) * | 2016-02-25 | 2021-07-27 | Kt Corp | Method and apparatus for processing video signals |
WO2017156705A1 (en) * | 2016-03-15 | 2017-09-21 | Mediatek Inc. | Affine prediction for video coding |
US10638152B2 (en) | 2016-03-15 | 2020-04-28 | Mediatek Inc. | Method and apparatus of video coding with affine motion compensation |
US10659801B2 (en) | 2016-03-24 | 2020-05-19 | Lg Electronics Inc. | Method and apparatus for inter prediction in video coding system |
US10958931B2 (en) * | 2016-05-11 | 2021-03-23 | Lg Electronics Inc. | Inter prediction method and apparatus in video coding system |
US10560712B2 (en) | 2016-05-16 | 2020-02-11 | Qualcomm Incorporated | Affine motion prediction for video coding |
WO2017201678A1 (en) * | 2016-05-24 | 2017-11-30 | 华为技术有限公司 | Image prediction method and related device |
US10778999B2 (en) | 2016-09-30 | 2020-09-15 | Qualcomm Incorporated | Frame rate up-conversion coding mode with affine motion model |
CN108271023B (en) | 2017-01-04 | 2021-11-19 | 华为技术有限公司 | Image prediction method and related device |
US10701390B2 (en) * | 2017-03-14 | 2020-06-30 | Qualcomm Incorporated | Affine motion information derivation |
CN107071496B (en) * | 2017-05-14 | 2020-01-17 | 北京工业大学 | Fast depth selection method for H.265/HEVC inter-frame coding unit |
CN116708780A (en) | 2017-08-11 | 2023-09-05 | 华为技术有限公司 | Method, device and equipment for encoding and decoding video image |
CN107454429B (en) * | 2017-08-21 | 2019-12-24 | 武汉兴图新科电子股份有限公司 | Video error concealment method based on motion vector extrapolation and image gradient weighting |
CN109587479B (en) * | 2017-09-29 | 2023-11-10 | 华为技术有限公司 | Inter-frame prediction method and device for video image and coder-decoder |
US11877001B2 (en) * | 2017-10-10 | 2024-01-16 | Qualcomm Incorporated | Affine prediction in video coding |
CN109922336B (en) | 2017-12-12 | 2023-07-18 | 华为技术有限公司 | Inter-frame prediction method and device for video data |
US20190208211A1 (en) * | 2018-01-04 | 2019-07-04 | Qualcomm Incorporated | Generated affine motion vectors |
CN114513671B (en) * | 2018-04-02 | 2024-04-09 | 华为技术有限公司 | Video encoding and decoding method and device |
WO2020007093A1 (en) * | 2018-07-02 | 2020-01-09 | 华为技术有限公司 | Image prediction method and apparatus |
SG11202013202YA (en) * | 2018-07-02 | 2021-01-28 | Huawei Tech Co Ltd | Motion vector prediction method and related apparatus |
CN108960160B (en) * | 2018-07-10 | 2021-03-09 | 深圳地平线机器人科技有限公司 | Method and device for predicting structured state quantity based on unstructured prediction model |
MX2021000351A (en) | 2018-07-11 | 2021-05-12 | Huawei Tech Co Ltd | Video encoder, video decoder, and corresponding method. |
WO2020024275A1 (en) * | 2018-08-03 | 2020-02-06 | 华为技术有限公司 | Inter-frame prediction method and device |
CN110971899B (en) * | 2018-10-01 | 2021-06-01 | 华为技术有限公司 | Method for determining motion information, and inter-frame prediction method and device |
JP7189234B2 (en) * | 2018-12-21 | 2022-12-13 | パナソニック インテレクチュアル プロパティ コーポレーション オブ アメリカ | Encoding device and decoding device |
CN109672886B (en) * | 2019-01-11 | 2023-07-04 | 京东方科技集团股份有限公司 | Image frame prediction method and device and head display equipment |
US11350108B2 (en) * | 2019-03-18 | 2022-05-31 | Tencent America LLC | Affine inter prediction refinement with optical flow |
Family Cites Families (25)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
US5923786A (en) * | 1995-07-17 | 1999-07-13 | Sony Corporation | Method and device for encoding and decoding moving images |
US5652616A (en) * | 1996-08-06 | 1997-07-29 | General Instrument Corporation Of Delaware | Optimal disparity estimation for stereoscopic video coding |
US7206346B2 (en) * | 1997-06-25 | 2007-04-17 | Nippon Telegraph And Telephone Corporation | Motion vector predictive encoding method, motion vector decoding method, predictive encoding apparatus and decoding apparatus, and storage media storing motion vector predictive encoding and decoding programs |
US7016539B1 (en) * | 1998-07-13 | 2006-03-21 | Cognex Corporation | Method for fast, robust, multi-dimensional pattern recognition |
CN1194544C (en) * | 2003-04-25 | 2005-03-23 | 北京工业大学 | Video encoding method based on prediction time and space domain conerent movement vectors |
US20100215101A1 (en) * | 2007-04-09 | 2010-08-26 | Yong Joon Jeon | Method and an apparatus for processing a video signal |
CN101350920A (en) * | 2007-07-17 | 2009-01-21 | 北京华辰广正科技发展有限公司 | Method for estimating global motion facing to panorama video |
CN101198064A (en) * | 2007-12-10 | 2008-06-11 | 武汉大学 | Movement vector prediction method in resolution demixing technology |
CN101340583B (en) * | 2008-08-08 | 2010-07-28 | 浙江大学 | Moving object detection apparatus based on H. 264 encoding domain |
WO2010093430A1 (en) * | 2009-02-11 | 2010-08-19 | Packetvideo Corp. | System and method for frame interpolation for a compressed video bitstream |
US8363721B2 (en) * | 2009-03-26 | 2013-01-29 | Cisco Technology, Inc. | Reference picture prediction for video coding |
CN101771878B (en) * | 2010-01-14 | 2011-05-25 | 广西大学 | Self-adaptively selecting global motion estimation method for panoramic video coding |
KR102166520B1 (en) * | 2010-04-13 | 2020-10-16 | 지이 비디오 컴프레션, 엘엘씨 | Sample region merging |
WO2011128259A1 (en) * | 2010-04-13 | 2011-10-20 | Fraunhofer-Gesellschaft zur Förderung der angewandten Forschung e.V. | A video decoder and a video encoder using motion-compensated prediction |
CN103039075B (en) * | 2010-05-21 | 2015-11-25 | Jvc建伍株式会社 | Picture coding device, method for encoding images and picture decoding apparatus, picture decoding method |
JP5686018B2 (en) * | 2010-05-21 | 2015-03-18 | 株式会社Jvcケンウッド | Image coding apparatus, image coding method, and image coding program |
WO2012006299A1 (en) * | 2010-07-08 | 2012-01-12 | Dolby Laboratories Licensing Corporation | Systems and methods for multi-layered image and video delivery using reference processing signals |
KR102034004B1 (en) * | 2010-10-08 | 2019-10-18 | 지이 비디오 컴프레션, 엘엘씨 | Picture coding supporting block partitioning and block merging |
JP5716437B2 (en) | 2011-02-08 | 2015-05-13 | 株式会社Jvcケンウッド | Image coding apparatus, image coding method, and image coding program |
KR102232113B1 (en) * | 2011-03-21 | 2021-03-25 | 엘지전자 주식회사 | Method for selecting motion vector predictor and device using same |
US20130170553A1 (en) * | 2012-01-04 | 2013-07-04 | Qualcomm Incorporated | Coding motion vector difference |
CN103475883B (en) * | 2013-09-26 | 2016-07-06 | 北京航空航天大学 | A kind of HEVC estimation divided based on moving region shifts to an earlier date terminating method |
JP6614472B2 (en) * | 2013-09-30 | 2019-12-04 | サン パテント トラスト | Image encoding method, image decoding method, image encoding device, and image decoding device |
US10531116B2 (en) * | 2014-01-09 | 2020-01-07 | Qualcomm Incorporated | Adaptive motion vector resolution signaling for video coding |
CN103905826A (en) * | 2014-04-10 | 2014-07-02 | 北京工业大学 | Self-adaptation global motion estimation method |
-
2014
- 2014-09-30 CN CN202010800185.XA patent/CN112087630B/en active Active
- 2014-09-30 CN CN201410526608.8A patent/CN104539966B/en active Active
- 2014-09-30 CN CN202010800184.5A patent/CN112087629B/en active Active
- 2014-09-30 CN CN201711022009.2A patent/CN107734335B/en active Active
-
2015
- 2015-04-15 EP EP20195878.2A patent/EP3836546A1/en active Pending
- 2015-04-15 WO PCT/CN2015/076659 patent/WO2016050051A1/en active Application Filing
- 2015-04-15 JP JP2017518250A patent/JP6546277B2/en active Active
- 2015-04-15 BR BR122020019754-4A patent/BR122020019754B1/en active IP Right Grant
- 2015-04-15 EP EP15846328.1A patent/EP3197165B1/en active Active
- 2015-04-15 KR KR1020197003118A patent/KR102060184B1/en active IP Right Grant
- 2015-04-15 KR KR1020197037916A patent/KR102160564B1/en active IP Right Grant
- 2015-04-15 EP EP20195877.4A patent/EP3836547A1/en active Pending
- 2015-04-15 KR KR1020177007259A patent/KR101945894B1/en active IP Right Grant
- 2015-04-15 BR BR112017006018-3A patent/BR112017006018B1/en active IP Right Grant
-
2017
- 2017-03-16 US US15/461,274 patent/US10827194B2/en active Active
-
2019
- 2019-06-20 JP JP2019114759A patent/JP7123863B2/en active Active
-
2020
- 2020-09-18 US US17/025,901 patent/US20210006818A1/en not_active Abandoned
-
2021
- 2021-07-26 JP JP2021121469A patent/JP7252282B2/en active Active
Also Published As
Publication number | Publication date |
---|---|
CN104539966B (en) | 2017-12-22 |
JP7252282B2 (en) | 2023-04-04 |
JP7123863B2 (en) | 2022-08-23 |
US20170188041A1 (en) | 2017-06-29 |
KR20190014590A (en) | 2019-02-12 |
KR20170044686A (en) | 2017-04-25 |
CN107734335B (en) | 2020-11-06 |
CN104539966A (en) | 2015-04-22 |
US10827194B2 (en) | 2020-11-03 |
EP3836547A1 (en) | 2021-06-16 |
EP3197165A4 (en) | 2017-08-02 |
JP2019198092A (en) | 2019-11-14 |
CN112087629A (en) | 2020-12-15 |
BR112017006018B1 (en) | 2023-12-19 |
CN112087630A (en) | 2020-12-15 |
EP3836546A1 (en) | 2021-06-16 |
BR122020019754B1 (en) | 2023-12-12 |
JP6546277B2 (en) | 2019-07-17 |
KR102060184B1 (en) | 2019-12-27 |
KR101945894B1 (en) | 2019-02-08 |
KR20190143496A (en) | 2019-12-30 |
EP3197165B1 (en) | 2020-09-16 |
JP2017535162A (en) | 2017-11-24 |
WO2016050051A1 (en) | 2016-04-07 |
BR112017006018A2 (en) | 2018-06-26 |
CN112087629B (en) | 2021-08-20 |
KR102160564B1 (en) | 2020-09-28 |
CN107734335A (en) | 2018-02-23 |
EP3197165A1 (en) | 2017-07-26 |
CN112087630B (en) | 2022-04-08 |
JP2021182752A (en) | 2021-11-25 |
Similar Documents
Publication | Publication Date | Title |
---|---|---|
US20210006818A1 (en) | Picture prediction method and related apparatus | |
US11968386B2 (en) | Picture prediction method and related apparatus | |
US10659803B2 (en) | Picture prediction method and related apparatus | |
US11831902B2 (en) | Picture prediction method and picture prediction apparatus |
Legal Events
Date | Code | Title | Description |
---|---|---|---|
STPP | Information on status: patent application and granting procedure in general |
Free format text: APPLICATION DISPATCHED FROM PREEXAM, NOT YET DOCKETED |
|
STPP | Information on status: patent application and granting procedure in general |
Free format text: DOCKETED NEW CASE - READY FOR EXAMINATION |
|
STPP | Information on status: patent application and granting procedure in general |
Free format text: NON FINAL ACTION MAILED |
|
STPP | Information on status: patent application and granting procedure in general |
Free format text: RESPONSE TO NON-FINAL OFFICE ACTION ENTERED AND FORWARDED TO EXAMINER |
|
STPP | Information on status: patent application and granting procedure in general |
Free format text: FINAL REJECTION MAILED |
|
STPP | Information on status: patent application and granting procedure in general |
Free format text: ADVISORY ACTION MAILED |
|
STCB | Information on status: application discontinuation |
Free format text: ABANDONED -- FAILURE TO RESPOND TO AN OFFICE ACTION |