US20210321112A1 - Coding prediction method and apparatus, and computer storage medium - Google Patents
Coding prediction method and apparatus, and computer storage medium Download PDFInfo
- Publication number
- US20210321112A1 US20210321112A1 US17/357,621 US202117357621A US2021321112A1 US 20210321112 A1 US20210321112 A1 US 20210321112A1 US 202117357621 A US202117357621 A US 202117357621A US 2021321112 A1 US2021321112 A1 US 2021321112A1
- Authority
- US
- United States
- Prior art keywords
- coding
- block
- coding block
- prediction
- motion estimation
- Prior art date
- Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
- Granted
Links
Images
Classifications
-
- H—ELECTRICITY
- H04—ELECTRIC COMMUNICATION TECHNIQUE
- H04N—PICTORIAL COMMUNICATION, e.g. TELEVISION
- H04N19/00—Methods or arrangements for coding, decoding, compressing or decompressing digital video signals
- H04N19/10—Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using adaptive coding
- H04N19/134—Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using adaptive coding characterised by the element, parameter or criterion affecting or controlling the adaptive coding
- H04N19/136—Incoming video signal characteristics or properties
- H04N19/137—Motion inside a coding unit, e.g. average field, frame or block difference
-
- H—ELECTRICITY
- H04—ELECTRIC COMMUNICATION TECHNIQUE
- H04N—PICTORIAL COMMUNICATION, e.g. TELEVISION
- H04N19/00—Methods or arrangements for coding, decoding, compressing or decompressing digital video signals
- H04N19/10—Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using adaptive coding
- H04N19/169—Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using adaptive coding characterised by the coding unit, i.e. the structural portion or semantic portion of the video signal being the object or the subject of the adaptive coding
- H04N19/17—Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using adaptive coding characterised by the coding unit, i.e. the structural portion or semantic portion of the video signal being the object or the subject of the adaptive coding the unit being an image region, e.g. an object
- H04N19/176—Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using adaptive coding characterised by the coding unit, i.e. the structural portion or semantic portion of the video signal being the object or the subject of the adaptive coding the unit being an image region, e.g. an object the region being a block, e.g. a macroblock
-
- H—ELECTRICITY
- H04—ELECTRIC COMMUNICATION TECHNIQUE
- H04N—PICTORIAL COMMUNICATION, e.g. TELEVISION
- H04N19/00—Methods or arrangements for coding, decoding, compressing or decompressing digital video signals
- H04N19/10—Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using adaptive coding
- H04N19/102—Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using adaptive coding characterised by the element, parameter or selection affected or controlled by the adaptive coding
- H04N19/103—Selection of coding mode or of prediction mode
- H04N19/105—Selection of the reference unit for prediction within a chosen coding or prediction mode, e.g. adaptive choice of position and number of pixels used for prediction
-
- H—ELECTRICITY
- H04—ELECTRIC COMMUNICATION TECHNIQUE
- H04N—PICTORIAL COMMUNICATION, e.g. TELEVISION
- H04N19/00—Methods or arrangements for coding, decoding, compressing or decompressing digital video signals
- H04N19/10—Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using adaptive coding
- H04N19/134—Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using adaptive coding characterised by the element, parameter or criterion affecting or controlling the adaptive coding
- H04N19/146—Data rate or code amount at the encoder output
- H04N19/147—Data rate or code amount at the encoder output according to rate distortion criteria
-
- H—ELECTRICITY
- H04—ELECTRIC COMMUNICATION TECHNIQUE
- H04N—PICTORIAL COMMUNICATION, e.g. TELEVISION
- H04N19/00—Methods or arrangements for coding, decoding, compressing or decompressing digital video signals
- H04N19/10—Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using adaptive coding
- H04N19/134—Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using adaptive coding characterised by the element, parameter or criterion affecting or controlling the adaptive coding
- H04N19/157—Assigned coding mode, i.e. the coding mode being predefined or preselected to be further used for selection of another element or parameter
- H04N19/159—Prediction type, e.g. intra-frame, inter-frame or bidirectional frame prediction
-
- H—ELECTRICITY
- H04—ELECTRIC COMMUNICATION TECHNIQUE
- H04N—PICTORIAL COMMUNICATION, e.g. TELEVISION
- H04N19/00—Methods or arrangements for coding, decoding, compressing or decompressing digital video signals
- H04N19/50—Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using predictive coding
- H04N19/503—Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using predictive coding involving temporal prediction
- H04N19/51—Motion estimation or motion compensation
- H04N19/513—Processing of motion vectors
- H04N19/517—Processing of motion vectors by encoding
- H04N19/52—Processing of motion vectors by encoding by predictive encoding
-
- H—ELECTRICITY
- H04—ELECTRIC COMMUNICATION TECHNIQUE
- H04N—PICTORIAL COMMUNICATION, e.g. TELEVISION
- H04N19/00—Methods or arrangements for coding, decoding, compressing or decompressing digital video signals
- H04N19/50—Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using predictive coding
- H04N19/503—Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using predictive coding involving temporal prediction
- H04N19/51—Motion estimation or motion compensation
- H04N19/537—Motion estimation other than block-based
- H04N19/54—Motion estimation other than block-based using feature points or meshes
-
- H—ELECTRICITY
- H04—ELECTRIC COMMUNICATION TECHNIQUE
- H04N—PICTORIAL COMMUNICATION, e.g. TELEVISION
- H04N19/00—Methods or arrangements for coding, decoding, compressing or decompressing digital video signals
- H04N19/50—Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using predictive coding
- H04N19/503—Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using predictive coding involving temporal prediction
- H04N19/51—Motion estimation or motion compensation
- H04N19/567—Motion estimation based on rate distortion criteria
-
- H—ELECTRICITY
- H04—ELECTRIC COMMUNICATION TECHNIQUE
- H04N—PICTORIAL COMMUNICATION, e.g. TELEVISION
- H04N19/00—Methods or arrangements for coding, decoding, compressing or decompressing digital video signals
- H04N19/50—Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using predictive coding
- H04N19/593—Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using predictive coding involving spatial prediction techniques
Definitions
- High Efficiency Video Coding HEVC
- SCC Screen Content Coding
- IBC Infra Block Copy
- Embodiments of the disclosure relate to the technical field of video coding and decoding, and particularly to a coding prediction method and apparatus as well as a decoding method and a decoder.
- the embodiments of the disclosure provide a coding prediction method, which may include the following operations.
- Motion Vector (MV) Predictors For a coding block of an intra prediction type, Motion Vector (MV) Predictors (MVPs) of at least two control points of the coding block are determined.
- MV Motion Vector
- MVPs Motion Vector Predictors
- Affine motion model-based motion estimation is performed on the coding block based on the MVPs of the at least two control points to obtain a first coding parameter of the coding block, the first coding parameter indicating a group of coding parameters corresponding to a minimum Rate-Distortion Cost (RDcost) and obtained by performing performing motion estimation on the coding block in a non-translation motion manner.
- RDcost Rate-Distortion Cost
- Prediction coding is performed on the coding block based on the first coding parameter.
- the embodiments of the disclosure provide a coding prediction apparatus, which may include a determination unit, a motion estimation unit and a prediction unit.
- the determination unit may be configured to determine MVPs of at least two control points of a coding block of an intra prediction type.
- the motion estimation unit may be configured to perform affine motion model-based motion estimation on the coding block based on the MVPs of the at least two control points to obtain a first coding parameter of the coding block, the first coding parameter indicating a group of coding parameters corresponding to a minimum RDcost and obtained by performing motion estimation on the coding block in a non-translation motion manner.
- the prediction unit may be configured to perform prediction coding on the coding block based on the first coding parameter.
- the embodiments of the disclosure provide a coding prediction apparatus, which may include a memory and a processor.
- the memory may be configured to store a computer program capable of running in the processor.
- the processor may be configured to run the computer program to implement the operations of any method as described in the first aspect.
- the embodiments of the disclosure provide a decoding method, which includes: creating a motion vector (MV) candidate list, receiving an index of a motion vector predictor (MVP index) from an encoder, obtaining a block vector (BV) based on the MVP index, and decoding a coding block subjected to prediction coding based on the BV to obtain a prediction pixel value of the coding block.
- MV motion vector
- MVP index motion vector predictor
- BV block vector
- the embodiments of the disclosure provide a decoder including a processor and memory storing a computer program executable by the processor.
- the processor is configured to run the computer program to implement operations of: creating a motion vector (MV) candidate list, receiving an index of a motion vector predictor (MVP index) from an encoder, obtaining a block vector (BV) based on the MVP index, and decoding a coding block subjected to prediction coding based on the BV to obtain a prediction pixel value of the coding block.
- MVP index motion vector predictor
- BV block vector
- FIG. 1 is a structure diagram of an IBC coding mode according to a related technical solution.
- FIG. 2 is a flowchart of creating an MV candidate list according to a related technical solution.
- FIG. 3 is a flowchart of a Non-Merge mode under an IBC coding mode according to a related technical solution.
- FIG. 4 is a structure diagram of a neighbouring block configuration in a Merge mode under an IBC coding mode according to a related technical solution.
- FIG. 5 is a flowchart of a Merge mode under an IBC coding mode according to a related technical solution.
- FIG. 6 is a flowchart of a coding prediction method according to an embodiment of the disclosure.
- FIG. 7A and FIG. 7B are structure diagrams of a non-translation motion of a coding block according to an embodiment of the disclosure.
- FIG. 8A is a schematic diagram of MVs of at least two control points of a coding block according to an embodiment of the disclosure.
- FIG. 8B is a schematic diagram of an MV sample of each sub block in a coding block according to an embodiment of the disclosure.
- FIG. 9A is a structure diagram of creating a candidate list based on five neighbouring locks according to an embodiment of the disclosure.
- FIG. 9B is a structure diagram of deducing MV at a corresponding position of a coding block based on a neighbouring block (A) according to an embodiment of the disclosure.
- FIG. 10 is a detailed flowchart of a coding prediction method according to an embodiment of the disclosure.
- FIG. 11 is a composition structure diagram of a coding prediction apparatus according to an embodiment of the disclosure.
- FIG. 12 is a composition structure diagram of another coding prediction apparatus according to an embodiment of the disclosure.
- FIG. 13 is a specific hardware structure diagram of a coding prediction apparatus according to an embodiment of the disclosure.
- the Joint Video Exploration Team is a working team set up by the Video Coding Experts Group (VCEG) of the International Telecommunication Union Telecommunication Standardization Sector (ITU-T) and the Moving Picture Experts Group (MPEG) of the International Organization for Standardization (ISO)/International Electrotechnical Commission (IEC) in October, 2015.
- the mission of the working team is to make a next-generation video coding standard.
- a Joint Exploration Test Model (JEM) is a universal reference software platform, and verification of different coding tools is implemented based on this platform.
- the JVET formally named the next-generation video coding standard after Versatile Video Coding (VVC) in April, 2018, and a corresponding test model is a View Transformation Model (VTM).
- the JVET adds more novel tools, for example, a Quadtree with nested multi-type tree (MT) (consisting of structures of a Quad Tree (QT), a Binary Tree (BT) and a Trigeminal Tree (TT)), affine motion compensation, Sub-Block Based Temporary Motion Vector Prediction (SbTMVP) and Adaptive Motion Vector Resolution (AMVR).
- MT Quadtree with nested multi-type tree
- QT Quad Tree
- BT Binary Tree
- TT Trigeminal Tree
- affine motion compensation Sub-Block Based Temporary Motion Vector Prediction
- SBTMVP Sub-Block Based Temporary Motion Vector Prediction
- AMVR Adaptive Motion Vector Resolution
- a basic principle of video coding compression is to remove a redundancy by use of a correlation among a space domain, a time domain and a codeword.
- a popular coding manner is to implement video coding compression based on a block-based hybrid video coding picturework through prediction (including intra prediction and inter prediction), transformation, quantization, entropy coding and the like. Such a coding picturework is high in vitality, and the block-based hybrid video coding picturework is still adopted for HEVC.
- an SCC standard is extended based on HEVC, and standardization of the standard has been completed basically in 2016. Coding technologies such as IBC, a Palette Mode (PLT), Adaptive Color Transform (ACT) and AMVR are added to the SCC standard to improve the coding efficiency.
- IBC Intra prediction is performed by SCC
- PHT Palette Mode
- ACT Adaptive Color Transform
- AMVR AMVR
- IBC coding mode may also be included in addition to Conventional intra Prediction (CIP).
- CIP Conventional intra Prediction
- the IBC coding mode is a method like motion compensation, and a reference block matched with a current coding block is found from a current picture and is represented by a BV.
- the IBC coding mode will be introduced below in combination with FIG. 1 in detail.
- the slashed regions are search regions (i.e., coded regions in a current picture)
- the black shaded blocks are respectively a current CU (a present coding block) and a best block predictor matched with the current CU, and a distance between the current CU and the matched best block predictor is called a BV.
- an IBC technology is proposed for SCC.
- a basic idea thereof is like conventional inter motion estimation, specifically a coded region of a current picture is searched for a reference block matched with a block to be encoded to obtain a distance between the two blocks. The distance is called a BC. Then a prediction residual is acquired based on the BV. The block to be encoded is finally encoded.
- the IBC coding mode also called a Coding Picture Reference (CPR) mode
- CPR Coding Picture Reference
- the IBC coding mode may further be divided into a first sub coding mode (for example, an IBC Non-Merge mode) and a second sub coding mode (for example, an IBC Merge mode), and the two coding modes are applicable to coding blocks smaller than or equal to 16 ⁇ 16.
- the IBC coding mode will be introduced below through the two coding modes in detail.
- the first sub coding mode takes the IBC Non-Merge mode as an example.
- the coding mode includes two search manners: hash-based search and pattern search.
- FIG. 2 A creation process of the MV candidate list is shown in FIG. 2 .
- FIG. 2 a flowchart of creating an MV candidate list according to a related technical solution is shown. The flow may include the following operations.
- two candidate MVs are selected from the space-domain candidate list.
- a candidate MV is selected from the time-domain candidate list.
- a candidate MV (0, 0) is added to form an initial candidate MV list.
- the MV candidate list is formed based on the first two MVs that are reserved.
- two MVPs may be obtained based on the MV candidate list obtained in S 208 . After the two MVPs are obtained, search may be started.
- a search manner adopted first is hash-based search, and this search manner is for accelerating search of the whole picture.
- Hash key matching between a current coding block (Curblock) and a reference block (Refblock) is implemented by taking a 4 ⁇ 4 block as a basic unit.
- selection of the reference block may be extended to all available size ranges of the whole picture.
- FIG. 3 a flowchart of a Non-Merge mode under an IBC coding mode according to a related technical solution is shown. The flow may include the following operations.
- a reference block candidate list is created based on hash-based search, a hash key matching relationship being formed between each reference block in the reference block candidate list and a coding block.
- the reference block candidate list is traversed, and a corresponding BV between the coding block and each reference block is calculated.
- an RDcost corresponding to each BV is calculated by taking each MVP as a search starting point based on the two MVPs.
- a first parameter corresponding to a minimum RDcost is returned and reserved, the first parameter at least including a first BV and a first MVP index.
- the operation that “the reference block candidate list is created” is specifically implemented as follows. At first, a mapping table between a coding block and a hash key is created by taking a 4 ⁇ 4 coding black as a basic unit in a whole current picture based on a position of the coding block. Then, hash key matching is performed on the coding block and a reference block. Only when a hash key of the coding block is equal to a hash key of the reference block, it is determined that the reference block may be matched with the coding block. The reference block candidate list may be created based on matched reference blocks.
- the RDcost is calculated based on the BV and the MVP. That is, an RDcost has a corresponding relationship with a BV and an MVP.
- the reference block candidate list may be traversed, and there may be a BV between the coding block and each reference block in the reference block candidate list. Then, the RDcost corresponding to each BV may be calculated by taking each BVP as the search starting point based on the two MVPs, and in such a manner, an RDcost set may be obtained. The minimum RDcost may be selected from the RDcost set. Finally, a coding parameter corresponding to the minimum RDcost may be reserved as the first parameter.
- S 303 when traversing is not ended, S 303 is re-executed to continue executing the flow.
- S 306 is executed. Since the first parameter corresponding to the minimum RDcost is reserved, after S 307 , motion compensation may be performed on the coding block based on the first parameter to determine a prediction residual of the coding block, thereby performing prediction coding on the coding block.
- the first parameter may not only include the first BV and the first MVP index but also include a first result (cost). No specific limits are made thereto in the embodiments of the disclosure.
- a pattern search may further be carried out.
- a current picture is taken as a P picture, an inter prediction type is adopted, and the current picture is placed at the end of REF_PLC_LIST_0 as a reference picture for inter prediction.
- a search process may include the following operations.
- a BV candidate list of the coding block is created based on a neighbouring block of the coding block.
- a BV array of the coding block and a corresponding cost array are created based on the BV candidate list.
- S 312 the BV candidate list is traversed, a Sum of Absolute Differences (SAD) cost corresponding to each BV is calculated, and the BV array and the cost array are updated to a SAD cost ascending sequence.
- SAD Sum of Absolute Differences
- an RDcost corresponding to the second BV is calculated by taking each MVP as a search starting point, and a second parameter corresponding to a minimum RDcost is returned and reserved, the second parameter at least including a second BV and a second MVP index.
- the operation “the BV candidate list of the coding block is created” is specifically implemented as follows. At first, neighbouring blocks (Neiblocks) in an AMVP space-domain candidate list corresponding to the current coding block are accessed respectively. When an IBC coding mode is adopted for a neighbouring block, a reference block (NeiRefblock) of the neighbouring, block may be reused as a reference of the coding block. Then, a BV between the coding block and the NeiRefblock is calculated and added to the BV candidate list of the coding block, thereby creating the BV candidate list of the coding block.
- the BV array and the corresponding cost array may be created based on the BV candidate list and the two arrays are arranged in a SAD cost ascending order.
- the BV candidate list may be traversed, and the sequences of the two arrays may be updated in real time based on the SAD cost corresponding to each calculated BV.
- the RDcost corresponding to the returned second BV may be calculated by each MVP as a search starting point, and the second parameter corresponding to the minimum RDcost may be stored.
- motion compensation may also be performed on the coding block based on the second parameter to determine the prediction residual of the coding block, thereby performing prediction coding on the coding block.
- the second parameter may not only include the second BV and the second MVP index but also include the second cost. No specific limits are made thereto in the embodiments of the disclosure.
- Motion compensation may be performed on the coding block based on the best results obtained by the two search manners to determine the prediction residual, and then prediction coding may be performed on the coding block.
- the encoder On an encoder side, the encoder is only required to transmit a corresponding MVP index and a Motion Vector Difference (MVD) between a BV and an MVP.
- the decoder On a decoder side, the decoder may create a MV candidate list at first in the same manner and may obtain the BV based on the transmitted MVP index and MVD, thereby performing decoding processing on the coding block subjected to prediction coding to obtain a prediction pixel value of the coding block. Then, the decoder may add the prediction pixel value and the prediction residual transmitted by the encoder to obtain a reconstructed pixel value corresponding to each pixel in the coding block.
- MVP Motion Vector Difference
- the second sub coding mode takes the IBC Merge mode as an example.
- a current picture is taken as a P picture, an inter prediction type is adopted, and the current picture is taken as a reference picture for inter prediction.
- FIG. 4 a structure diagram of a neighbouring block configuration in a Merge mode under an IBC coding mode according to a related technical solution is shown.
- an MV candidate list of a coding block may be created according to priority A1 ⁇ B1 ⁇ B0 ⁇ A0 ⁇ B2 as shown in FIG. 4 .
- FIG. 5 a flowchart of a Merge mode under an IBC coding mode according to a related technical solution is shown. The flow may include the following operations.
- the MV candidate list is traversed, an MVP in the MV candidate list is directly determined as a BV of the coding block, and an RDcost corresponding to each BV is calculated.
- a third parameter corresponding to a minimum RDcost in the calculated RDcosts is reserved, the third parameter at least including a third MVP index.
- the operation that “the space-domain MV candidate list corresponding to the coding block is created” is specifically as follows. In combination with a configuration structure shown in FIG. 4 , at most four space-domain candidate MVs are selected according to the sequence A1 ⁇ B1 ⁇ B0 ⁇ A0 ⁇ B2 and placed in the MV candidate list, so that the MV candidate list corresponding to the coding block may be created.
- the encoder is only required to transmit an MVP index.
- the decoder may create an MV candidate list at first in the same manner and may directly obtain a BV based on the transmitted MVP index, thereby performing decoding processing on the coding block subjected to prediction coding.
- the reference block matched with the coding block is still obtained based on a translation motion model, and consequently, a higher coding efficiency may be achieved in the IBC coding mode only for blocks repeatedly appearing in an SC.
- a higher coding efficiency may be achieved in the IBC coding mode only for blocks repeatedly appearing in an SC.
- complex motion conditions such as scaling, rotation and transformation, and thus there is a further optimization space for encoding.
- the embodiments of the disclosure provide a coding prediction method.
- An affine motion model is added based on the IBC coding mode, and a translation motion model is replaced with the affine motion model, so that the number of coding bits may further be reduced, and the coding rate is improved.
- the coding prediction method provided in the embodiments of the disclosure may be applied not only to an intra prediction type of an encoder but also to an intra prediction type of a decoder. That is, the embodiments of the disclosure may be applied not only to a coding system but also to a decoding system. No specific limits are made in the embodiments of the disclosure.
- the method may include the following operations.
- MVPs of at least two control points of the coding block are determined.
- affine motion model-based motion estimation is performed on the coding block based on the predicted motion vector values of the at least two control points to obtain a first coding parameter of the coding block, the first coding parameter indicating a group of coding parameters corresponding to a minimum rate distortion cost and obtained by performing motion estimation on the coding block in a non-translation motion manner.
- prediction coding is performed on the coding block based on the first coding parameter.
- multiple groups of coding parameters may be obtained by “performing affine motion model-based motion estimation on the coding block based on the MVPs of the at least two control points”.
- RDcost calculation may be performed for different coding parameters to obtain RDcosts corresponding to the different coding parameters.
- the minimum RDcost may be selected from these RDcosts, and the group of coding parameters corresponding to the minimum RDcost are taken as the first coding parameters.
- the coding block is a coding block to be encoded in a current picture, and the intra prediction type is set. Accordingly, for the coding block of the intra prediction type, the MVPs of at least two control points of the coding block are determined at first; then affine motion model-based motion estimation is performed on the coding block based on the MVPs of the at least two control points to obtain the first coding parameter corresponding to the coding block, the first coding parameter indicating the group of coding parameters corresponding to the minimum RDcost and obtained by performing non-translation motion manner-based. motion estimation on the coding block, and finally, prediction coding is performed on the coding block based on the first coding parameter.
- affine motion model-based motion estimation With addition of affine motion model-based motion estimation, the existence of complex non-translation motions, such as scaling, rotation and transformation, of a coding block of the intra prediction type in a screen picture may be eliminated, the number of coding bits is further reduced, and the coding rate is further improved.
- an affine motion model is a novel technology for an inter prediction type.
- inter prediction is based on motion compensation of a coding block, and a translation motion model is usually adopted. That is, there is made such a hypothesis that MVs of all pixels in a coding block are the same and are usually replaced with a BV of a top-left vertex of the coding block.
- the coding block may also have non-translation motions such as a scaling motion, a rotational rotation and a transformation motion.
- FIG. 7A and FIG. 7B show a rotational motion of a coding block and a scaling motion of the coding block respectively.
- ⁇ is a scaling coefficient, and a value of ⁇ is specifically set according to a practical condition. No specific limits are made in the embodiment of the disclosure.
- FIG. 8A a schematic diagram of MVs of at least two control points of a coding block according to an embodiment of the disclosure is shown.
- two control points i.e., MVs of top-left and top-right vertexes
- MVs of top-left and top-right vertexes are adopted for the coding block, namely ⁇ right arrow over (v) ⁇ 0 and ⁇ right arrow over (v) ⁇ 1 shown in FIG.
- the affine motion model is also called a four-parameter affine model.
- the affine motion model is also called a six-parameter affine model.
- an MV of each 4 ⁇ 4 sub block in the coding block may be deduced through formula (1):
- an MV of each 4 ⁇ 4 sub block in the coding block may be deduced through formula (2):
- w and h represent a width and height of the coding block respectively
- FIG. 8B a schematic diagram of an MV sample of each sub block in a coding block according to an embodiment of the disclosure is shown.
- coding prediction of each sub block may be completed by a motion compensation difference filter in combination with the MVs deduced through the formula (1) or the formula (2). Therefore, complex motion conditions may be described better by affine motion compensation.
- the affine motion model is applied to the coding block based on the intra prediction type, so that the number of the coding bits may further be reduced, and the coding rate is further improved.
- a coding mode corresponding to the coding block at least includes an IBC coding mode and an IBCAffine coding mode. That is, the IBC coding mode may be selected for prediction coding of the coding block, or the IBCAffine coding mode may be selected for prediction coding.
- the IBC coding mode may be selected for prediction coding of the coding block, or the IBCAffine coding mode may be selected for prediction coding.
- no specific limits are made in the embodiment of the disclosure.
- the method may further include the following operations.
- a best BV of the coding block is acquired.
- the IBCAffine coding mode is selected for the coding block.
- the IBC coding mode is selected for the coding block.
- the operation that the best BV of the coding block is acquired may include the following operations.
- IBC search is performed on the coding block, at least one reference block matched with a hash key of the coding block is selected, and a first candidate block list of the coding block is created based on the at least one reference block.
- the first candidate block list is traversed, and a BV between the coding block and each reference block in the first candidate block list is calculated.
- RDcosts corresponding to different BVs are calculated based on the BVs respectively, and the BV corresponding to the minimum RDcost in the RDcosts is determined as the best BV of the coding block.
- the preset multiple is a preset judgment value configured to determine whether to adopt the IBCAffine coding mode for the coding block or not.
- the preset multiple may be set to be 1.05 times.
- the preset multiple may be specifically limited according to a practical condition. No specific limits are made in the embodiment of the disclosure.
- a flag value of the best coding mode may be set to be True, thereby indicating the coding mode adopted for the coding block in a coding system and further performing prediction coding based on the coding mode.
- the operation that affine motion model-based motion estimation is performed on the coding block based on the MVPs of the at least two control points to obtain the first coding parameter of the coding block may include the following operation.
- Affine motion model-based motion estimation is performed on the coding block based on the MVPs of the at least two control points through the IBCAffine coding mode to obtain the first coding parameter of the coding block, the first coding parameter indicating a group of coding parameters corresponding to a minimum RDcost and obtained by performing affine motion model-based motion estimation on the coding block.
- the method may farther include the following operations.
- Translation motion model-based motion estimation is performed on the coding block based on the IBC coding mode to obtain a second coding parameter of the coding block, the second coding parameter indicating a group of coding parameters corresponding to a minimum RDcost and obtained by performing translation motion model-based motion estimation on the coding block.
- Prediction coding is performed on the coding block based on the second coding parameter.
- the method may further include the following operations.
- the best BV is directly determined as a third coding parameter of the coding block.
- Prediction coding is performed on the coding block based on the third coding parameter.
- a flag value of the best coding mode may be set to be True, thereby indicating the coding mode adopted for the coding block.
- the IBC coding mode may be divided into a first sub coding mode (for example, an IBC Non-Merge mode) and a second sub coding mode for example, an IBC Merge mode), and the IBCAffine coding mode may further be divided into a third sub coding mode (for example, an IBCAffine Non-Merge mode) and a fourth sub coding mode (for example, an IBCAffine Merge mode).
- the first sub coding mode and the second sub coding mode have been specifically described above, the third sub coding mode and the fourth sub coding mode will be specifically described below respectively.
- an IBC search (the search process is similar to the search process of the first sub coding mode) may be first performed on a coding block and a first candidate block list of the coding block may be created, a hash key matching relationship being formed between each reference block in the first candidate block list and the coding block.
- the first candidate block list may be traversed in combination with the two preset MVPs obtained in FIG. 2 , and the RDcost corresponding to each BV may be calculated by taking the two preset MVPs as search starting points respectively.
- the BV corresponding to the minimum RDcost in the calculated RDcosts is determined as the best BV of the coding block.
- the best BV When the best BV is not equal to 0, it indicates that hash keys at any positions in the coding block are equal and IBCAffine coding mode-based motion estimation is not performed. In such case, the best BV may be directly determined as the third coding parameter corresponding to the coding block, and prediction coding may be performed on the coding block based on the third parameter.
- IBCAffine coding mode-based motion estimation is required to be performed on the coding block (namely affine motion model-based motion estimation is performed on the coding block), and an obtained best result (for example, a result corresponding to the minimum RDcost) is called the first motion estimation result, represented by IBCAffineCost.
- IBC coding mode-based motion estimation is also required to be performed on the coding block (translation motion model-based motion estimation is performed on the coding block), and an obtained best result (for example, a result corresponding to the minimum RDcost) is called the second motion estimation result, represented by IBCCost.
- IBCCost the second motion estimation result
- the preset multiple is 1.05 times.
- the IBCAffineCost the IBCAffine coding mode is selected for the coding block, and the first coding parameter corresponding to the minimum RDcost (for example, a first BV and a corresponding first MVP index).
- the IBC coding mode is selected for the coding block, and the second coding parameter corresponding to the minimum RDcost (for example, a second BV and a corresponding second MVP index) may be obtained. Therefore, motion compensation may be performed according to the obtained first coding parameter or second coding parameter to obtain a prediction residual of the coding block to implement prediction coding of the coding block.
- the second coding parameter corresponding to the minimum RDcost for example, a second BV and a corresponding second MVP index
- the operation that affine motion model-based motion estimation is performed on the coding block based on the MVPs of the at least two control points to obtain the first coding parameter of the coding block may include the following operations.
- a prediction value of at least one pixel of the coding block in a reference picture is calculated based on the MVPs of the at least two control points.
- An iterative operation is executed on a matching error of the at least one pixel of the coding block between an original picture and the reference picture and on a gradient matrix of the prediction value, and an MV is updated based on the iterative operation.
- the first coding parameter of the coding block is obtained based on the updated MV.
- the third sub coding mode it is also needed to take the current picture as a P picture, set a prediction mode to be an inter prediction mode and place the current picture at the end of REF_PIC_LIST_0 as a reference picture for inter prediction.
- two candidate MVPs of the coding block are acquired by inter Affine AMVP (AffineAMVP), each candidate MVP including at least two control points.
- affine motion search is performed by taking each candidate MVP as a search starting point.
- a prediction value s i of a pixel at a corresponding, position in a reference picture is represented through a formula (3):
- ⁇ x i and ⁇ y i represent MVs of the coding block and are not changed linearly but determined by four parameters (a, b, ⁇ 0 and ⁇ 1 ) in the following formula (4):
- a and b represent translation components of a pixel coordinate of the coding block in the reference picture after rotational transformation respectively
- ⁇ 0 and ⁇ 1 represent parameters in a rotation matrix after scaling transformation is performed on the coding block based on the coefficient ⁇
- ⁇ T represents transpose of ⁇ from a row vector to a column vector.
- G xi and G yi represent gradient values and are obtained by Sobel gradient calculation.
- a prediction error of any pixel (x i , y i ) in the coding block may be calculated through the following formula (6):
- Part 0 represents a prediction error of any pixel in the coding block
- part 1 represents a matching error of any pixel (x i , y i ) in the coding block between an original picture and a reference picture
- part 2 represents a gradient matrix of a prediction value of a pixel at a corresponding position in the reference picture.
- the iterative operation may be executed based on the gradient of the prediction value and the matching error, and the iterative operation process may update the MVs.
- the prediction error is minimum, and the obtained MVs are updated MVs that are finally required.
- the preset count threshold is a preset iteration count required by determining that the prediction error is minimum.
- the preset count threshold may be 5; and for the six-parameter affine model, the preset count threshold may be 4.
- the preset count threshold may be specifically set according to a practical condition. No specific limits are made in the embodiment of the disclosure.
- the method may further include the following operations.
- a second candidate block list of the coding block is created, a reference block in the second candidate block list being spatially neighbouring to the coding block and the reference block being encoded based on the IBCAffine coding mode.
- the second candidate block list may be traversed, and MVs of control points at corresponding, positions of the coding block may be calculated based on MVs of at least two control points of each reference block in the second candidate block list respectively.
- a fourth coding parameter corresponding to the coding block may be acquired from the MVs, the fourth coding parameter indicating a group of coding parameters corresponding to the minimum RDcost among the obtained MVs of the coding block.
- Prediction coding may be performed on the coding block based on the fourth coding parameter.
- FIG. 9A a structure diagram of creating a candidate list based on five neighbouring blocks according to an embodiment of the disclosure is shown.
- the five neighbouring blocks A, B, C, D and E shown in FIG. 9A are sequentially accessed according to a sequence A ⁇ B ⁇ C ⁇ D ⁇ E.
- the IBCAffine coding mode is adopted for a neighbouring block and the reference picture is the last picture (i.e., the current picture) in the sequence REF_PIC_LIST_0, the neighbouring block is placed in the second candidate block list as a reference block.
- the second candidate block list is traversed to make a selection based on the number of the control points.
- the formula (1) is selected to deduce the MVs the control points at the corresponding positions of the current coding block; and when the number of the control points is 3, the formula (2) is selected to deduce the MVs of the control points at the corresponding positions of the current coding block, specifically referring to the example shown in FIG. 9B .
- the result corresponding to the minimum prediction residual i.e., the minimum RDcost
- the encoder is required to transmit MVP indexes in the second candidate block list and prediction residuals to a decoder.
- the decoder may create a candidate block list the same as that of the encoder and calculate an MV of the current coding block through the formula (1) or the formula (2) based on the MVP indexes transmitted in a bit stream, thereby performing decoding processing on the coding block subjected to prediction coding.
- the embodiment provides the coding prediction method.
- the MVPs of at least two control points of the coding block are determined at first; then affine motion model-based motion estimation is performed on the coding block based on the MVPs of the at least two control points to obtain the first coding parameter of the coding block, the first coding parameter indicating, the group of coding parameters corresponding to the minimum RDcost and obtained by performing non-translation motion manner-based motion estimation on the coding block; and finally, prediction coding is performed on the coding block based on the first coding parameter.
- affine motion model-based motion estimation With addition of affine motion model-based motion estimation, the existence of non-translation motions, such as scaling, rotation and transformation, of a coding block of the intra prediction type in a screen picture may be eliminated, the number of coding bits is further reduced, and the coding rate is further improved.
- FIG. 10 a detailed flow of a coding prediction method according to an embodiment of the disclosure is shown.
- the detailed flow may include the following operations.
- IBC search is performed on a coding block, and a first candidate block list of the coding block is created.
- the first candidate block list is traversed, and a BV between the coding block and each reference block in the first candidate block list is calculated.
- RDcosts corresponding to different BVs are calculated based on the BVs respectively, and a BV corresponding to a minimum RDcost in the RDcosts is determined as a best BV.
- the coding prediction method provided in the embodiment may be applied not only to the intra prediction type of an encoder but also to an intra prediction type of a decoder. That is, the embodiment of the disclosure may be applied not only to a coding system and but also to a decoding system. However, no specific limits are made in the embodiment of the disclosure.
- a composition of a coding prediction apparatus 110 is shown, which may include a determination unit 1101 , a motion estimation unit 1102 and a prediction unit 1103 .
- the determination unit 1101 is configured to, for a coding block of an intra prediction type, determine MVPs of at least two control points of the coding block.
- the motion estimation unit 1102 is configured to perform affine motion model-based motion estimation on the coding block based on the MVPs of the at least two control points to obtain a first coding parameter of the coding block, the first coding parameter indicating a group of coding parameters corresponding to a minimum RDcost and obtained by performing motion estimation on the coding block in a non-translation motion manner.
- the prediction unit 1103 is configured to perform prediction coding on the coding block based on the first coding parameter.
- the coding prediction apparatus 110 may further include an acquisition unit 1104 and a judgment unit 1105 .
- the acquisition unit 1104 is configured to acquire a best BV of the coding block.
- the judgment unit 1105 is configured to, when the best BV is equal to 0, calculate a first motion estimation result of the coding block based on the IBCAffine coding mode and a second motion estimation result of the coding block based on the IBC coding mode respectively; when the first motion estimation result is not greater than a preset multiple of the second motion estimation result, select the IBCAffine coding mode for the coding block; and when the first motion estimation result is greater than the preset multiple of the second motion estimation result, select the IBC coding mode for the coding block.
- the acquisition unit 1104 is specifically configured to, perform IBC search on the coding block, select at least one reference block matched with a hash key of the coding block, create a first candidate block list of the coding block based on the at least one reference block, traverse the first candidate block list, calculate a BV between the coding block and each reference block in the tint candidate block list, calculate RDcosts corresponding to different BVs based on the BVs respectively and determine the BV corresponding to the minimum RDcost in the RDcosts as the best BV of the coding block.
- the motion estimation unit 1102 is specifically configured to perform affine motion model-based motion estimation on the coding block based on the MVPs of the at least two control points through the IBCAffine coding mode to obtain the first coding parameter of the coding block, the first coding parameter indicating the group of coding parameters corresponding to a minimum RDcost and obtained by performing the affine motion model-based motion estimation on the coding block.
- the motion estimation unit 1102 is further configured to perform translation motion model-based motion estimation on the coding block based on the IBC coding mode to obtain a second coding parameter of the coding block, the second coding parameter indicating a group of coding parameters corresponding to a minimum RDcost and obtained by performing translation motion model-based motion estimation on the coding block.
- the prediction unit 1103 is further configured to perform prediction coding on the coding block based on the second coding parameter.
- the motion estimation unit 1102 is further configured to, when the best BV is not equal to 0, directly determine the best BV as a third coding parameter corresponding to the coding block.
- the prediction unit 1103 is further configured to perform prediction coding on the coding block based on the third coding parameter.
- the motion estimation unit 1102 is specifically configured to calculate a prediction value of at least one pixel of the coding block in a reference picture based on the MVPs of the at least two control points; execute an iterative operation on a matching error of the at least one pixel of the coding block between an original picture and the reference picture and on a gradient matrix of the prediction value; update an MV based on the iterative operation; when a count of the iterative operation meets a preset count threshold, obtain an updated MV; and obtain the first coding parameter of the coding block based on the updated MV.
- the motion estimation unit 1102 is further configured to, for the coding block of the intra prediction type, create a second candidate block list of the coding block, a reference block in the second candidate block list being spatially neighbouring to the coding block and the reference block being encoded based on the IBCAffine coding mode; traverse the second candidate block list and calculate MVs of control points at corresponding positions of the coding block based on MVs of at least two control points of each reference block in the second candidate block list respectively; and acquire a fourth coding parameter corresponding to the coding block from the MVs, the fourth coding parameter indicating a group of coding parameters corresponding to the minimum RDcost among the obtained MVs of the coding block.
- the prediction unit 1103 is further configured to perform prediction coding on the coding block based on the fourth coding parameter.
- unit may be part of a circuit, part of a processor, part of a program or software and the like, of course, may also be modular and may also be non-modular.
- each component in the embodiment may be integrated into a processing unit, each unit may also exist independently, and two or more than two units may also be integrated into a unit.
- the integrated unit may be implemented in a hardware form and may also be implemented in form of software function module.
- the integrated unit When implemented in form of software function module and sold or used not as an independent product, the integrated unit may be stored in a computer-readable storage medium.
- the technical solution of the embodiment substantially or parts making contributions to the conventional art or all or part of the technical solution may be embodied in form of software product, and the computer software product is stored in a storage medium, including a plurality of instructions configured to enable a computer device (which may be a personal computer, a server, a network device or the like) or a processor to execute all or part of the operations of the method in the embodiment.
- the storage medium includes: various media capable of storing program codes such as a U disk, a mobile hard disk, a Read Only Memory (ROM), a Random Access Memory (RAM), a magnetic disk or an optical disk.
- the embodiment provides a computer storage medium, which stores a coding prediction program.
- the coding prediction program is executed by at least one processor to implement the operations of the method in the technical solution shown in FIG. 6 .
- FIG. 13 a specific hardware structure example of the coding prediction apparatus 110 according to an embodiment of the disclosure is shown, which may include a network interface 1301 , a memory 1302 and a processor 1303 . Each component is coupled together through a bus system 1304 . It can be understood that the bus system 1304 is configured to implement connection communication between these components.
- the bus system 1304 includes a data bus and may further include a power bus, a control bus and a state signal bus. However, for clear description, various buses in FIG. 13 are marked as the bus system 1304 .
- the network interface 1301 is configured to receive and send a signal in a process of receiving and sending information with another external network element.
- the memory 1302 is configured to store a computer program capable of running in the processor 1303 .
- the processor 1303 is configured to run the computer program to execute the to liming operations.
- MVPs of at least two control points of the coding block are determined.
- Affine motion model-based motion estimation is performed on the coding block based on the MVPs of the at least two control points to obtain a first coding parameter of the coding block, the first coding parameter indicating a group of coding parameters corresponding to a minimum RDcost and obtained by performing motion estimation on the coding block in a non-translation motion manner.
- Prediction coding is performed. on the coding block based on the first coding parameter.
- the memory 1302 in the embodiment of the disclosure may be a volatile memory or a nonvolatile memory, or may include both the volatile and nonvolatile memories.
- the nonvolatile memory may be a ROM, a Programmable ROM (PROM), an Erasable PROM (EPROM), an Electrically EPROM (EEPROM) or a flash memory.
- the volatile memory may be a RAM, and is used as an external high-speed cache.
- RAMs in various forms may be adopted, such as a Static RAM (SRAM), a Dynamic RAM (DRAM), a Synchronous DRAM (SDRAM), a Double Data Rate SDRAM (DDRSDRAM), an Enhanced SDRAM (ESDRAM), a Synchlink DRAM (SLDRAM) and a Direct Rambus RAM (DRRAM).
- SRAM Static RAM
- DRAM Dynamic RAM
- SDRAM Synchronous DRAM
- DDRSDRAM Double Data Rate SDRAM
- ESDRAM Enhanced SDRAM
- SLDRAM Synchlink DRAM
- DRRAM Direct Rambus RAM
- the processor 1303 may be an integrated circuit chip with a signal processing capability. In an implementation process, each operation of the method may be completed by an integrated logic circuit of hardware in the processor 1303 or an instruction in a software form.
- the processor 1303 may be a universal processor, a Digital Signal Processor (DSP), an Application Specific Integrated Circuit (ASIC), a Field Programmable Gate Array (FPGA) or another Programmable Logic Device (PLD), discrete gate or transistor logical device and discrete hardware component.
- DSP Digital Signal Processor
- ASIC Application Specific Integrated Circuit
- FPGA Field Programmable Gate Array
- PLD Programmable Logic Device
- Each method, operation and logical block diagram disclosed in the embodiments of the disclosure may be implemented or executed.
- the universal processor may be a microprocessor or the processor may also be any conventional processor and the like.
- the operations of the method disclosed in combination with the embodiments of the disclosure may be directly embodied to be executed and completed by a hardware decoding processor or executed and completed by a combination of hardware and software modules in the decoding processor.
- the software module may be located in a mature storage medium in this field such as a RAM, a flash memory, a ROM, a PROM or EEPROM and a register.
- the storage medium is located in the memory 1302 .
- the processor 1303 reads information in the memory 1302 and completes the operations of the method in combination with hardware.
- the processing unit may be implemented in one or more ASICs, DSPs, DSP Devices (DSPDs), PLDs, FPGAs, universal processors, controllers, microcontrollers, microprocessors, other electronic units configured to execute the functions in the disclosure or combinations thereof.
- ASICs ASICs, DSPs, DSP Devices (DSPDs), PLDs, FPGAs, universal processors, controllers, microcontrollers, microprocessors, other electronic units configured to execute the functions in the disclosure or combinations thereof.
- the technology of the disclosure may be implemented through the modules (for example, processes and functions) executing the functions in the disclosure.
- a software code may be stored in the memory and executed by the processor.
- the memory may be implemented in the processor or outside the processor.
- the processor 1303 is further configured to run the computer program to implement the operations of the method in the technical solution shown in FIG. 6 .
- sequence numbers of the embodiments of the disclosure are adopted not to represent superiority-inferiority of the embodiments but only for description.
- MVPs of at least two control points of the coding block are determined at first; then affine motion model-based motion estimation is performed on the coding block based on the MVPs of the at least two control points to obtain a first coding parameter of the coding block, the first coding parameter indicating a group of coding parameters corresponding to a RDcost and obtained by performing non-translation motion manner-based motion estimation on the coding block; and finally, prediction coding is performed on the coding block based on the first coding parameter.
- affine motion model-based motion estimation With addition of affine motion model-based motion estimation, The existence of non-translation motions, such as scaling, rotation and transformation, of a coding block of the ultra prediction type in a screen picture may be eliminated, the number of coding bits is further reduced, and the coding rate is further improved.
Landscapes
- Engineering & Computer Science (AREA)
- Multimedia (AREA)
- Signal Processing (AREA)
- Compression Or Coding Systems Of Tv Signals (AREA)
Abstract
Description
- This application is a continuation application of International Patent Application No. PCT/CN2018/124504 filed on Dec. 27, 2018, the disclosure of which is incorporated herein by reference in its entirety.
- With the rapid development of sciences and technologies, users have higher requirements on video watching and processing, and particularly, have more and more demands on Screen Content Videos (SCVs). High Efficiency Video Coding (HEVC) is for adaptation to videos with high resolutions and high picture rates, but the efficiency of typical Screen Content Coding (SCC) is not improved greatly. Therefore, an extended coding standard, i.e., SCC, has been proposed based on HEVC with regard to high contrast, limited colour data, more duplicated regions and other characteristics of screen content.
- In most coding standards, adaptive inter/intra prediction is used based on blocks. For example, in the SCC standard, a basic block unit for video coding is called a Coding Unit (CU), and the same coding parameter is shared for pixels in the CU to improve the coding efficiency. The latest proposed coding and decoding technology is affine motion estimation and compensation capable of effectively tracking more complex motions, for example, rotation, scaling and transformation of a moving object, and is mainly applied to SCC of an inter prediction type. For an intra prediction type, an Infra Block Copy (IBC) coding mode is generally adopted in existing solutions. In the IBC coding mode, a Two-Dimensional (2D) translation motion model is considered only, and thus there is a further optimization space for coding.
- Embodiments of the disclosure relate to the technical field of video coding and decoding, and particularly to a coding prediction method and apparatus as well as a decoding method and a decoder.
- According to a first aspect, the embodiments of the disclosure provide a coding prediction method, which may include the following operations.
- For a coding block of an intra prediction type, Motion Vector (MV) Predictors (MVPs) of at least two control points of the coding block are determined.
- Affine motion model-based motion estimation is performed on the coding block based on the MVPs of the at least two control points to obtain a first coding parameter of the coding block, the first coding parameter indicating a group of coding parameters corresponding to a minimum Rate-Distortion Cost (RDcost) and obtained by performing performing motion estimation on the coding block in a non-translation motion manner.
- Prediction coding is performed on the coding block based on the first coding parameter.
- According to a second aspect, the embodiments of the disclosure provide a coding prediction apparatus, which may include a determination unit, a motion estimation unit and a prediction unit.
- The determination unit may be configured to determine MVPs of at least two control points of a coding block of an intra prediction type.
- The motion estimation unit may be configured to perform affine motion model-based motion estimation on the coding block based on the MVPs of the at least two control points to obtain a first coding parameter of the coding block, the first coding parameter indicating a group of coding parameters corresponding to a minimum RDcost and obtained by performing motion estimation on the coding block in a non-translation motion manner.
- The prediction unit may be configured to perform prediction coding on the coding block based on the first coding parameter.
- According to a third aspect, the embodiments of the disclosure provide a coding prediction apparatus, which may include a memory and a processor.
- The memory may be configured to store a computer program capable of running in the processor.
- The processor may be configured to run the computer program to implement the operations of any method as described in the first aspect.
- According to a fourth aspect, the embodiments of the disclosure provide a decoding method, which includes: creating a motion vector (MV) candidate list, receiving an index of a motion vector predictor (MVP index) from an encoder, obtaining a block vector (BV) based on the MVP index, and decoding a coding block subjected to prediction coding based on the BV to obtain a prediction pixel value of the coding block.
- According to a fifth aspect, the embodiments of the disclosure provide a decoder including a processor and memory storing a computer program executable by the processor. The processor is configured to run the computer program to implement operations of: creating a motion vector (MV) candidate list, receiving an index of a motion vector predictor (MVP index) from an encoder, obtaining a block vector (BV) based on the MVP index, and decoding a coding block subjected to prediction coding based on the BV to obtain a prediction pixel value of the coding block.
-
FIG. 1 is a structure diagram of an IBC coding mode according to a related technical solution. -
FIG. 2 is a flowchart of creating an MV candidate list according to a related technical solution. -
FIG. 3 is a flowchart of a Non-Merge mode under an IBC coding mode according to a related technical solution. -
FIG. 4 is a structure diagram of a neighbouring block configuration in a Merge mode under an IBC coding mode according to a related technical solution. -
FIG. 5 is a flowchart of a Merge mode under an IBC coding mode according to a related technical solution. -
FIG. 6 is a flowchart of a coding prediction method according to an embodiment of the disclosure. -
FIG. 7A andFIG. 7B are structure diagrams of a non-translation motion of a coding block according to an embodiment of the disclosure. -
FIG. 8A is a schematic diagram of MVs of at least two control points of a coding block according to an embodiment of the disclosure. -
FIG. 8B is a schematic diagram of an MV sample of each sub block in a coding block according to an embodiment of the disclosure. -
FIG. 9A is a structure diagram of creating a candidate list based on five neighbouring locks according to an embodiment of the disclosure. -
FIG. 9B is a structure diagram of deducing MV at a corresponding position of a coding block based on a neighbouring block (A) according to an embodiment of the disclosure. -
FIG. 10 is a detailed flowchart of a coding prediction method according to an embodiment of the disclosure. -
FIG. 11 is a composition structure diagram of a coding prediction apparatus according to an embodiment of the disclosure. -
FIG. 12 is a composition structure diagram of another coding prediction apparatus according to an embodiment of the disclosure. -
FIG. 13 is a specific hardware structure diagram of a coding prediction apparatus according to an embodiment of the disclosure. - In order to make the characteristics and technical contents of the embodiments of the disclosure understood in more detail, implementation of the embodiments of the disclosure will be described below in combination with the drawings in detail. The appended drawings are only adopted for description as references and not intended to limit the embodiments of the disclosure.
- The Joint Video Exploration Team (JVET) is a working team set up by the Video Coding Experts Group (VCEG) of the International Telecommunication Union Telecommunication Standardization Sector (ITU-T) and the Moving Picture Experts Group (MPEG) of the International Organization for Standardization (ISO)/International Electrotechnical Commission (IEC) in October, 2015. The mission of the working team is to make a next-generation video coding standard. A Joint Exploration Test Model (JEM) is a universal reference software platform, and verification of different coding tools is implemented based on this platform. The JVET formally named the next-generation video coding standard after Versatile Video Coding (VVC) in April, 2018, and a corresponding test model is a View Transformation Model (VTM). In July, 2018, the JEVT hammered out algorithm descriptions and coding methods of VVC working draft 2 and VTM2 in the 11th meeting. Based on an original HEVC Test Model (HM), the JVET adds more novel tools, for example, a Quadtree with nested multi-type tree (MT) (consisting of structures of a Quad Tree (QT), a Binary Tree (BT) and a Trigeminal Tree (TT)), affine motion compensation, Sub-Block Based Temporary Motion Vector Prediction (SbTMVP) and Adaptive Motion Vector Resolution (AMVR).
- A basic principle of video coding compression is to remove a redundancy by use of a correlation among a space domain, a time domain and a codeword. A popular coding manner is to implement video coding compression based on a block-based hybrid video coding picturework through prediction (including intra prediction and inter prediction), transformation, quantization, entropy coding and the like. Such a coding picturework is high in vitality, and the block-based hybrid video coding picturework is still adopted for HEVC.
- Based on a characteristic of a screen content, an SCC standard is extended based on HEVC, and standardization of the standard has been completed basically in 2016. Coding technologies such as IBC, a Palette Mode (PLT), Adaptive Color Transform (ACT) and AMVR are added to the SCC standard to improve the coding efficiency. When intra prediction is performed by SCC, an IBC coding mode may also be included in addition to Conventional intra Prediction (CIP). The IBC coding mode is a method like motion compensation, and a reference block matched with a current coding block is found from a current picture and is represented by a BV. The IBC coding mode will be introduced below in combination with
FIG. 1 in detail. - Referring to
FIG. 1 , a structure diagram of an IBC coding mode according to a related technical solution is shown. As shown inFIG. 1 , the slashed regions are search regions (i.e., coded regions in a current picture), the black shaded blocks are respectively a current CU (a present coding block) and a best block predictor matched with the current CU, and a distance between the current CU and the matched best block predictor is called a BV. In an existing extended HEVC standard, i.e., an HEVC-SCC standard, an IBC technology is proposed for SCC. A basic idea thereof is like conventional inter motion estimation, specifically a coded region of a current picture is searched for a reference block matched with a block to be encoded to obtain a distance between the two blocks. The distance is called a BC. Then a prediction residual is acquired based on the BV. The block to be encoded is finally encoded. - In a latest solution for a VVC reference model BMS2.1, the IBC coding mode, also called a Coding Picture Reference (CPR) mode, is mainly adopted for SCC. Here, the IBC coding mode may further be divided into a first sub coding mode (for example, an IBC Non-Merge mode) and a second sub coding mode (for example, an IBC Merge mode), and the two coding modes are applicable to coding blocks smaller than or equal to 16×16. The IBC coding mode will be introduced below through the two coding modes in detail.
- The first sub coding mode takes the IBC Non-Merge mode as an example. The coding mode includes two search manners: hash-based search and pattern search.
- Before search, an MV candidate list is needed to be created based on an inter Advanced Motion Vector Prediction (AMVP) technology. A creation process of the MV candidate list is shown in
FIG. 2 . Referring toFIG. 2 , a flowchart of creating an MV candidate list according to a related technical solution is shown. The flow may include the following operations. - In S201, a space-domain candidate list is acquired.
- In S202, a time-domain candidate list is acquired.
- In S203, two candidate MVs are selected from the space-domain candidate list.
- In S204, a candidate MV is selected from the time-domain candidate list.
- In S205, same candidate MVs in the selected candidate MVs are merged.
- In S206, a candidate MV (0, 0) is added to form an initial candidate MV list.
- In S207, the first two MVs in the initial candidate MV list are reserved.
- In S208, the MV candidate list is formed based on the first two MVs that are reserved.
- It is to be noted that two MVPs may be obtained based on the MV candidate list obtained in S208. After the two MVPs are obtained, search may be started.
- A search manner adopted first is hash-based search, and this search manner is for accelerating search of the whole picture. Hash key matching between a current coding block (Curblock) and a reference block (Refblock) is implemented by taking a 4×4 block as a basic unit. Here, selection of the reference block may be extended to all available size ranges of the whole picture. Referring to
FIG. 3 , a flowchart of a Non-Merge mode under an IBC coding mode according to a related technical solution is shown. The flow may include the following operations. - In S301, two MVPs are acquired based on an MV candidate list.
- In S302, a reference block candidate list is created based on hash-based search, a hash key matching relationship being formed between each reference block in the reference block candidate list and a coding block.
- In S303, the reference block candidate list is traversed, and a corresponding BV between the coding block and each reference block is calculated.
- In S304, an RDcost corresponding to each BV is calculated by taking each MVP as a search starting point based on the two MVPs.
- In S305, it is determined whether traversing of the reference block candidate list is ended or not.
- In S306, after traversing is ended, a first parameter corresponding to a minimum RDcost is returned and reserved, the first parameter at least including a first BV and a first MVP index.
- In S307, it is determined whether the first BV is equal to 0 or not.
- In S308, when the first BV is not equal to 0, prediction coding is performed on the coding block after motion compensation is performed on the coding block.
- It is to be noted that the operation that “the reference block candidate list is created” is specifically implemented as follows. At first, a mapping table between a coding block and a hash key is created by taking a 4×4 coding black as a basic unit in a whole current picture based on a position of the coding block. Then, hash key matching is performed on the coding block and a reference block. Only when a hash key of the coding block is equal to a hash key of the reference block, it is determined that the reference block may be matched with the coding block. The reference block candidate list may be created based on matched reference blocks.
- It is also to be noted that the RDcost is calculated based on the BV and the MVP. That is, an RDcost has a corresponding relationship with a BV and an MVP. The reference block candidate list may be traversed, and there may be a BV between the coding block and each reference block in the reference block candidate list. Then, the RDcost corresponding to each BV may be calculated by taking each BVP as the search starting point based on the two MVPs, and in such a manner, an RDcost set may be obtained. The minimum RDcost may be selected from the RDcost set. Finally, a coding parameter corresponding to the minimum RDcost may be reserved as the first parameter.
- In the embodiments of the disclosure, for S305, when traversing is not ended, S303 is re-executed to continue executing the flow. When traversing is ended, S306 is executed. Since the first parameter corresponding to the minimum RDcost is reserved, after S307, motion compensation may be performed on the coding block based on the first parameter to determine a prediction residual of the coding block, thereby performing prediction coding on the coding block. In the embodiments of the disclosure, the first parameter may not only include the first BV and the first MVP index but also include a first result (cost). No specific limits are made thereto in the embodiments of the disclosure.
- Furthermore, when the first BV corresponding to the minimum RDcost after S306 is equal to 0 or no reference block matched with the coding block is found by hash-based search in S302, a pattern search may further be carried out. In such a search manner, a current picture is taken as a P picture, an inter prediction type is adopted, and the current picture is placed at the end of REF_PLC_LIST_0 as a reference picture for inter prediction. Still referring to
FIG. 3 , a search process may include the following operations. - In S309, when the first BV is equal to 0, pattern-search-based motion estimation is implemented.
- In S310, a BV candidate list of the coding block is created based on a neighbouring block of the coding block.
- In S311, a BV array of the coding block and a corresponding cost array are created based on the BV candidate list.
- In S312, the BV candidate list is traversed, a Sum of Absolute Differences (SAD) cost corresponding to each BV is calculated, and the BV array and the cost array are updated to a SAD cost ascending sequence.
- In S313, it is determined whether traversing of the BV candidate list is ended or not.
- In S314, after traversing is ended, a second BV and a second cost both corresponding to a minimum SAD cost are selected and returned.
- In S315, it is determined whether the second BV is equal to 0 or not.
- In S316, when the second BV is not equal to 0, an RDcost corresponding to the second BV is calculated by taking each MVP as a search starting point, and a second parameter corresponding to a minimum RDcost is returned and reserved, the second parameter at least including a second BV and a second MVP index.
- In S317, when the second BV is equal to 0, the flow is ended.
- It is to be noted that the operation “the BV candidate list of the coding block is created” is specifically implemented as follows. At first, neighbouring blocks (Neiblocks) in an AMVP space-domain candidate list corresponding to the current coding block are accessed respectively. When an IBC coding mode is adopted for a neighbouring block, a reference block (NeiRefblock) of the neighbouring, block may be reused as a reference of the coding block. Then, a BV between the coding block and the NeiRefblock is calculated and added to the BV candidate list of the coding block, thereby creating the BV candidate list of the coding block.
- It is also to be noted that the BV array and the corresponding cost array may be created based on the BV candidate list and the two arrays are arranged in a SAD cost ascending order. The BV candidate list may be traversed, and the sequences of the two arrays may be updated in real time based on the SAD cost corresponding to each calculated BV. After traversing is ended, only 0th elements in the two arrays, i.e., the second cost and second BV corresponding to the minimum SAD cost, are returned. Next, the RDcost corresponding to the returned second BV may be calculated by each MVP as a search starting point, and the second parameter corresponding to the minimum RDcost may be stored. Then, after S316, motion compensation may also be performed on the coding block based on the second parameter to determine the prediction residual of the coding block, thereby performing prediction coding on the coding block. In the embodiments of the disclosure, the second parameter may not only include the second BV and the second MVP index but also include the second cost. No specific limits are made thereto in the embodiments of the disclosure.
- Motion compensation may be performed on the coding block based on the best results obtained by the two search manners to determine the prediction residual, and then prediction coding may be performed on the coding block. On an encoder side, the encoder is only required to transmit a corresponding MVP index and a Motion Vector Difference (MVD) between a BV and an MVP. On a decoder side, the decoder may create a MV candidate list at first in the same manner and may obtain the BV based on the transmitted MVP index and MVD, thereby performing decoding processing on the coding block subjected to prediction coding to obtain a prediction pixel value of the coding block. Then, the decoder may add the prediction pixel value and the prediction residual transmitted by the encoder to obtain a reconstructed pixel value corresponding to each pixel in the coding block.
- The second sub coding mode takes the IBC Merge mode as an example. In the coding mode, a current picture is taken as a P picture, an inter prediction type is adopted, and the current picture is taken as a reference picture for inter prediction. Referring to
FIG. 4 , a structure diagram of a neighbouring block configuration in a Merge mode under an IBC coding mode according to a related technical solution is shown. As shown inFIG. 4 , an MV candidate list of a coding block may be created according to priority A1→B1→B0→A0→B2 as shown inFIG. 4 . Referring toFIG. 5 , a flowchart of a Merge mode under an IBC coding mode according to a related technical solution is shown. The flow may include the following operations. - In S501, a space-domain MV candidate list corresponding to a coding block is created.
- In S502, the MV candidate list is traversed, an MVP in the MV candidate list is directly determined as a BV of the coding block, and an RDcost corresponding to each BV is calculated.
- In S503, after traversing is ended, a third parameter corresponding to a minimum RDcost in the calculated RDcosts is reserved, the third parameter at least including a third MVP index.
- In S504, after motion compensation is performed on the coding block, prediction coding is performed on the coding block based on the third parameter.
- It is to be noted that the operation that “the space-domain MV candidate list corresponding to the coding block is created” is specifically as follows. In combination with a configuration structure shown in
FIG. 4 , at most four space-domain candidate MVs are selected according to the sequence A1→B1→B0→A0→B2 and placed in the MV candidate list, so that the MV candidate list corresponding to the coding block may be created. - In the second sub coding mode, since a BV and an MVP are the same, no MVD in the first sub coding mode exists. On an encoder side, the encoder is only required to transmit an MVP index. On the decoder side, the decoder may create an MV candidate list at first in the same manner and may directly obtain a BV based on the transmitted MVP index, thereby performing decoding processing on the coding block subjected to prediction coding.
- In the IBC coding mode, the reference block matched with the coding block is still obtained based on a translation motion model, and consequently, a higher coding efficiency may be achieved in the IBC coding mode only for blocks repeatedly appearing in an SC. However, in an SCC scenario, there may also be such complex motion conditions such as scaling, rotation and transformation, and thus there is a further optimization space for encoding.
- The embodiments of the disclosure provide a coding prediction method. An affine motion model is added based on the IBC coding mode, and a translation motion model is replaced with the affine motion model, so that the number of coding bits may further be reduced, and the coding rate is improved. Here, the coding prediction method provided in the embodiments of the disclosure may be applied not only to an intra prediction type of an encoder but also to an intra prediction type of a decoder. That is, the embodiments of the disclosure may be applied not only to a coding system but also to a decoding system. No specific limits are made in the embodiments of the disclosure.
- The embodiments of the disclosure will be described below in combination with the drawings in detail by taking the intra prediction type of an encoder as an example.
- Referring to
FIG. 6 , a flow example of a coding prediction method according to an embodiment of the disclosure is shown. The method may include the following operations. - In S601, for a coding block of an intra prediction type, MVPs of at least two control points of the coding block are determined.
- In S602, affine motion model-based motion estimation is performed on the coding block based on the predicted motion vector values of the at least two control points to obtain a first coding parameter of the coding block, the first coding parameter indicating a group of coding parameters corresponding to a minimum rate distortion cost and obtained by performing motion estimation on the coding block in a non-translation motion manner.
- In S603, prediction coding is performed on the coding block based on the first coding parameter.
- It is to be noted that multiple groups of coding parameters may be obtained by “performing affine motion model-based motion estimation on the coding block based on the MVPs of the at least two control points”. RDcost calculation may be performed for different coding parameters to obtain RDcosts corresponding to the different coding parameters. The minimum RDcost may be selected from these RDcosts, and the group of coding parameters corresponding to the minimum RDcost are taken as the first coding parameters.
- In the embodiment of the disclosure, the coding block is a coding block to be encoded in a current picture, and the intra prediction type is set. Accordingly, for the coding block of the intra prediction type, the MVPs of at least two control points of the coding block are determined at first; then affine motion model-based motion estimation is performed on the coding block based on the MVPs of the at least two control points to obtain the first coding parameter corresponding to the coding block, the first coding parameter indicating the group of coding parameters corresponding to the minimum RDcost and obtained by performing non-translation motion manner-based. motion estimation on the coding block, and finally, prediction coding is performed on the coding block based on the first coding parameter. With addition of affine motion model-based motion estimation, the existence of complex non-translation motions, such as scaling, rotation and transformation, of a coding block of the intra prediction type in a screen picture may be eliminated, the number of coding bits is further reduced, and the coding rate is further improved.
- It is to be noted that an affine motion model is a novel technology for an inter prediction type. In HEVC, inter prediction is based on motion compensation of a coding block, and a translation motion model is usually adopted. That is, there is made such a hypothesis that MVs of all pixels in a coding block are the same and are usually replaced with a BV of a top-left vertex of the coding block. However, during a practical application, the coding block may also have non-translation motions such as a scaling motion, a rotational rotation and a transformation motion.
FIG. 7A andFIG. 7B show a rotational motion of a coding block and a scaling motion of the coding block respectively. ρ is a scaling coefficient, and a value of ρ is specifically set according to a practical condition. No specific limits are made in the embodiment of the disclosure. - Here, motion compensation is implemented based on the affine motion model through MVs of control points. The control points are represented in two modes, for example, two control points and three control points. A reference block (or called a mapped block) corresponding to the coding block may be found from a reference picture by use of the MVs of the control points. Referring to
FIG. 8A , a schematic diagram of MVs of at least two control points of a coding block according to an embodiment of the disclosure is shown. When two control points, i.e., MVs of top-left and top-right vertexes, are adopted for the coding block, namely {right arrow over (v)}0 and {right arrow over (v)}1 shown inFIG. 8A , the affine motion model is also called a four-parameter affine model. When three control points, i.e., MVs of top-left, top-right and bottom-left vertexes, are adopted for the coding block, namely {right arrow over (v)}0, {right arrow over (v)}1 and {right arrow over (v)}2 shown toFIG. 8A , the affine motion model is also called a six-parameter affine model. - When the four-parameter affine model is adopted, an MV of each 4×4 sub block in the coding block may be deduced through formula (1):
-
- When the six-parameter affine model is adopted, an MV of each 4×4 sub block in the coding block may be deduced through formula (2):
-
- w and h represent a width and height of the coding block respectively, the MV of the top-left vertex is {right arrow over (v)}0=(v0x, v0y), the MV of the top-right vertex is {right arrow over (v)}1=(v1x, v1y), the MV at the bottom-left vertex is {right arrow over (v)}1=(v2x, v2y), and an MV corresponding to each pixel (x, y) in the coding block is {right arrow over (v)}=(vx, vy).
- Referring to
FIG. 8B , a schematic diagram of an MV sample of each sub block in a coding block according to an embodiment of the disclosure is shown. For an MV sample of each sub block shown inFIG. 8B , coding prediction of each sub block may be completed by a motion compensation difference filter in combination with the MVs deduced through the formula (1) or the formula (2). Therefore, complex motion conditions may be described better by affine motion compensation. In the embodiment of the disclosure, the affine motion model is applied to the coding block based on the intra prediction type, so that the number of the coding bits may further be reduced, and the coding rate is further improved. - It can be understood that a coding mode corresponding to the coding block at least includes an IBC coding mode and an IBCAffine coding mode. That is, the IBC coding mode may be selected for prediction coding of the coding block, or the IBCAffine coding mode may be selected for prediction coding. However, no specific limits are made in the embodiment of the disclosure.
- In some embodiments, before the operation that the MVPs of the at least two control points of the coding block are determined for the coding block of the intra prediction type, the method may further include the following operations.
- A best BV of the coding block is acquired.
- When the best BV is equal to 0, a first motion estimation result of the coding block based on the IBCAffine coding mode and a second motion estimation result of the coding block based on the IBC coding mode are calculated respectively.
- When the first motion estimation result is not greater than a preset multiple of the second motion estimation result, the IBCAffine coding mode is selected for the coding block.
- When the first motion estimation result is greater than the preset multiple of the second motion estimation result, the IBC coding mode is selected for the coding block.
- Furthermore, in some embodiments, the operation that the best BV of the coding block is acquired may include the following operations.
- IBC search is performed on the coding block, at least one reference block matched with a hash key of the coding block is selected, and a first candidate block list of the coding block is created based on the at least one reference block.
- The first candidate block list is traversed, and a BV between the coding block and each reference block in the first candidate block list is calculated.
- RDcosts corresponding to different BVs are calculated based on the BVs respectively, and the BV corresponding to the minimum RDcost in the RDcosts is determined as the best BV of the coding block.
- It is to be noted that the preset multiple is a preset judgment value configured to determine whether to adopt the IBCAffine coding mode for the coding block or not. In the embodiment of the disclosure, the preset multiple may be set to be 1.05 times. However, in a practical application, the preset multiple may be specifically limited according to a practical condition. No specific limits are made in the embodiment of the disclosure.
- It is also to be noted that, for the two coding modes (the IBC coding mode and the IBCAffine coding mode) for the coding block, when the best coding mode (i.e., the coding mode presently selected for the coding block) is determined for the coding block, a flag value of the best coding mode may be set to be True, thereby indicating the coding mode adopted for the coding block in a coding system and further performing prediction coding based on the coding mode.
- Furthermore, in some embodiments, when the IBCAffine coding mode is selected for the coding block, the operation that affine motion model-based motion estimation is performed on the coding block based on the MVPs of the at least two control points to obtain the first coding parameter of the coding block may include the following operation.
- Affine motion model-based motion estimation is performed on the coding block based on the MVPs of the at least two control points through the IBCAffine coding mode to obtain the first coding parameter of the coding block, the first coding parameter indicating a group of coding parameters corresponding to a minimum RDcost and obtained by performing affine motion model-based motion estimation on the coding block.
- Furthermore, in some embodiments, when the IBC coding mode is selected for the coding block, the method may farther include the following operations.
- Translation motion model-based motion estimation is performed on the coding block based on the IBC coding mode to obtain a second coding parameter of the coding block, the second coding parameter indicating a group of coding parameters corresponding to a minimum RDcost and obtained by performing translation motion model-based motion estimation on the coding block.
- Prediction coding is performed on the coding block based on the second coding parameter.
- Furthermore, in some embodiments, after the operation that the best BV of the coding block is acquired, the method may further include the following operations.
- When the best BV is not equal to 0, the best BV is directly determined as a third coding parameter of the coding block.
- Prediction coding is performed on the coding block based on the third coding parameter.
- It is to be noted that, for the two coding modes, when the best coding mode (i.e., the coding mode presently selected for the coding block) is determined for the coding block, a flag value of the best coding mode may be set to be True, thereby indicating the coding mode adopted for the coding block.
- It is also to be noted that the IBC coding mode may be divided into a first sub coding mode (for example, an IBC Non-Merge mode) and a second sub coding mode for example, an IBC Merge mode), and the IBCAffine coding mode may further be divided into a third sub coding mode (for example, an IBCAffine Non-Merge mode) and a fourth sub coding mode (for example, an IBCAffine Merge mode). The first sub coding mode and the second sub coding mode have been specifically described above, the third sub coding mode and the fourth sub coding mode will be specifically described below respectively.
- In the third sub coding mode, an IBC search (the search process is similar to the search process of the first sub coding mode) may be first performed on a coding block and a first candidate block list of the coding block may be created, a hash key matching relationship being formed between each reference block in the first candidate block list and the coding block. The first candidate block list may be traversed in combination with the two preset MVPs obtained in
FIG. 2 , and the RDcost corresponding to each BV may be calculated by taking the two preset MVPs as search starting points respectively. The BV corresponding to the minimum RDcost in the calculated RDcosts is determined as the best BV of the coding block. When the best BV is not equal to 0, it indicates that hash keys at any positions in the coding block are equal and IBCAffine coding mode-based motion estimation is not performed. In such case, the best BV may be directly determined as the third coding parameter corresponding to the coding block, and prediction coding may be performed on the coding block based on the third parameter. When the best BV is equal to 0, on one hand, IBCAffine coding mode-based motion estimation is required to be performed on the coding block (namely affine motion model-based motion estimation is performed on the coding block), and an obtained best result (for example, a result corresponding to the minimum RDcost) is called the first motion estimation result, represented by IBCAffineCost. On the other hand, IBC coding mode-based motion estimation is also required to be performed on the coding block (translation motion model-based motion estimation is performed on the coding block), and an obtained best result (for example, a result corresponding to the minimum RDcost) is called the second motion estimation result, represented by IBCCost. Here, there is made such a hypothesis that the preset multiple is 1.05 times. In case of IBCAffineCost≤1.05×IBCCost, the IBCAffine coding mode is selected for the coding block, and the first coding parameter corresponding to the minimum RDcost (for example, a first BV and a corresponding first MVP index). In case of IBCAffineCost>1.05×IBCCost, the IBC coding mode is selected for the coding block, and the second coding parameter corresponding to the minimum RDcost (for example, a second BV and a corresponding second MVP index) may be obtained. Therefore, motion compensation may be performed according to the obtained first coding parameter or second coding parameter to obtain a prediction residual of the coding block to implement prediction coding of the coding block. - In some embodiments, the operation that affine motion model-based motion estimation is performed on the coding block based on the MVPs of the at least two control points to obtain the first coding parameter of the coding block may include the following operations.
- A prediction value of at least one pixel of the coding block in a reference picture is calculated based on the MVPs of the at least two control points.
- An iterative operation is executed on a matching error of the at least one pixel of the coding block between an original picture and the reference picture and on a gradient matrix of the prediction value, and an MV is updated based on the iterative operation.
- When a count of the iterative operation meets a preset count threshold, an updated MV is obtained.
- The first coding parameter of the coding block is obtained based on the updated MV.
- It is to be noted that, in the third sub coding mode, it is also needed to take the current picture as a P picture, set a prediction mode to be an inter prediction mode and place the current picture at the end of REF_PIC_LIST_0 as a reference picture for inter prediction. At first, two candidate MVPs of the coding block are acquired by inter Affine AMVP (AffineAMVP), each candidate MVP including at least two control points. Then, affine motion search is performed by taking each candidate MVP as a search starting point.
- Specifically, for any pixel (xi, yi) in the current coding block, a prediction value si of a pixel at a corresponding, position in a reference picture is represented through a formula (3):
-
- Δxi and Δyi represent MVs of the coding block and are not changed linearly but determined by four parameters (a, b, ω0 and ω1) in the following formula (4):
-
- a and b represent translation components of a pixel coordinate of the coding block in the reference picture after rotational transformation respectively, ω0 and ω1 represent parameters in a rotation matrix after scaling transformation is performed on the coding block based on the coefficient ρ, and ĀT represents transpose of Ā from a row vector to a column vector.
- First-order taylor expansion is performed on the formula (3) to obtain formula (5), as shown below:
-
- Gxi and Gyi represent gradient values and are obtained by Sobel gradient calculation. According to the formula (5), a prediction error of any pixel (xi, yi) in the coding block may be calculated through the following formula (6):
-
-
Part 0 represents a prediction error of any pixel in the coding block, part 1 represents a matching error of any pixel (xi, yi) in the coding block between an original picture and a reference picture, and part 2 represents a gradient matrix of a prediction value of a pixel at a corresponding position in the reference picture. - Accordingly, the iterative operation may be executed based on the gradient of the prediction value and the matching error, and the iterative operation process may update the MVs. When the count of the iterative operation meets the preset count threshold, the prediction error is minimum, and the obtained MVs are updated MVs that are finally required.
- It is also to be noted that the preset count threshold is a preset iteration count required by determining that the prediction error is minimum. In the embodiment of the disclosure, for the four-parameter affine model, the preset count threshold may be 5; and for the six-parameter affine model, the preset count threshold may be 4. During the practical application, the preset count threshold may be specifically set according to a practical condition. No specific limits are made in the embodiment of the disclosure.
- In some embodiments, for the coding block of the intra prediction type, the method may further include the following operations.
- A second candidate block list of the coding block is created, a reference block in the second candidate block list being spatially neighbouring to the coding block and the reference block being encoded based on the IBCAffine coding mode.
- The second candidate block list may be traversed, and MVs of control points at corresponding, positions of the coding block may be calculated based on MVs of at least two control points of each reference block in the second candidate block list respectively.
- A fourth coding parameter corresponding to the coding block may be acquired from the MVs, the fourth coding parameter indicating a group of coding parameters corresponding to the minimum RDcost among the obtained MVs of the coding block.
- Prediction coding may be performed on the coding block based on the fourth coding parameter.
- It is to be noted that, in the fourth sub coding mode, it is needed to create the second candidate block list of the coding block at first. Referring to
FIG. 9A , a structure diagram of creating a candidate list based on five neighbouring blocks according to an embodiment of the disclosure is shown. The five neighbouring blocks A, B, C, D and E shown inFIG. 9A are sequentially accessed according to a sequence A→B→C→D→E. When the IBCAffine coding mode is adopted for a neighbouring block and the reference picture is the last picture (i.e., the current picture) in the sequence REF_PIC_LIST_0, the neighbouring block is placed in the second candidate block list as a reference block. Then, the second candidate block list is traversed to make a selection based on the number of the control points. When the number of the control points is 2, the formula (1) is selected to deduce the MVs the control points at the corresponding positions of the current coding block; and when the number of the control points is 3, the formula (2) is selected to deduce the MVs of the control points at the corresponding positions of the current coding block, specifically referring to the example shown inFIG. 9B . The result corresponding to the minimum prediction residual (i.e., the minimum RDcost) may be selected from the obtained MVs and returned, the returned result being the fourth coding parameter, and prediction coding may be performed on the coding block based on the fourth coding parameter. That is, on an encoder side, the encoder is required to transmit MVP indexes in the second candidate block list and prediction residuals to a decoder. On the decoder side, the decoder may create a candidate block list the same as that of the encoder and calculate an MV of the current coding block through the formula (1) or the formula (2) based on the MVP indexes transmitted in a bit stream, thereby performing decoding processing on the coding block subjected to prediction coding. - The embodiment provides the coding prediction method. For the coding block of the intra prediction type, the MVPs of at least two control points of the coding block are determined at first; then affine motion model-based motion estimation is performed on the coding block based on the MVPs of the at least two control points to obtain the first coding parameter of the coding block, the first coding parameter indicating, the group of coding parameters corresponding to the minimum RDcost and obtained by performing non-translation motion manner-based motion estimation on the coding block; and finally, prediction coding is performed on the coding block based on the first coding parameter. With addition of affine motion model-based motion estimation, the existence of non-translation motions, such as scaling, rotation and transformation, of a coding block of the intra prediction type in a screen picture may be eliminated, the number of coding bits is further reduced, and the coding rate is further improved.
- Based on the same inventive concept of the abovementioned embodiment, referring to
FIG. 10 , a detailed flow of a coding prediction method according to an embodiment of the disclosure is shown. The detailed flow may include the following operations. - In S1001, IBC search is performed on a coding block, and a first candidate block list of the coding block is created.
- In S1002, the first candidate block list is traversed, and a BV between the coding block and each reference block in the first candidate block list is calculated.
- In S1003, RDcosts corresponding to different BVs are calculated based on the BVs respectively, and a BV corresponding to a minimum RDcost in the RDcosts is determined as a best BV.
- In S1004, it is determined whether the best BV is equal to 0 or not.
- In S1005, when the best BV is equal to 0, affine motion model-based motion estimation is performed on the coding block based on an IBCAffine coding mode, and a first motion estimation result IBCAffineCost and a first coding parameter both corresponding to the minimum RDcost are obtained by calculation.
- In S1006, when the best BV is equal to 0, translation motion model-based motion estimation is performed on the coding block based on an IBC coding mode, and a second motion estimation result IBCCost and a second coding parameter both corresponding to the minimum RDcost are obtained by calculation.
- In S1007, it is determined whether IBCAffineCost≤1.05×IBCCost is true or not.
- In S1008, when IBCAffineCost≤1.05×IBCCost is true, the IBCAffine coding mode is selected for the coding block, and the first coding parameter of the coding block is obtained.
- In S1009, when IBCAffineCost≤1.05×IBCCost is not true, the IBC coding mode is selected for the coding block, and the second coding parameter of the coding block is obtained.
- In S1010, when the best BV is not equal to 0, the best BV is directly determined as a third coding parameter of the coding block.
- In S1011, after motion compensation is performed an the coding block, prediction coding is performed on the coding block.
- It is to be noted that, when the IBCAffine coding mode is not adopted for the coding block, processing may be performed directly according to the flowchart shown in
FIG. 3 and elaborations are omitted herein. In addition, after S1004, it is determined whether the best BV is equal to 0 or not. When the best BV is equal to 0, S1005 and S1006 are executed. When the best BV is not equal to 0, S1010 is executed. S1011 may be executed after S1008, or S1009 or S1010. - It is also to be noted that the coding prediction method provided in the embodiment may be applied not only to the intra prediction type of an encoder but also to an intra prediction type of a decoder. That is, the embodiment of the disclosure may be applied not only to a coding system and but also to a decoding system. However, no specific limits are made in the embodiment of the disclosure.
- Through the embodiment, specific implementation of the abovementioned embodiment is elaborated. It can be seen that, through the technical solution of the embodiment, with addition of affine motion model-based motion estimation, the existence of non-translation motions, such as scaling, rotation and transformation, of a coding block of the intra prediction type in a screen picture may be eliminated, the number of coding bits is further reduced, and the coding rate is further improved.
- Based on the same inventive concept of the abovementioned embodiments, referring to
FIG. 11 , a composition of acoding prediction apparatus 110 according to an embodiment of the disclosure is shown, which may include adetermination unit 1101, amotion estimation unit 1102 and aprediction unit 1103. - The
determination unit 1101 is configured to, for a coding block of an intra prediction type, determine MVPs of at least two control points of the coding block. - The
motion estimation unit 1102 is configured to perform affine motion model-based motion estimation on the coding block based on the MVPs of the at least two control points to obtain a first coding parameter of the coding block, the first coding parameter indicating a group of coding parameters corresponding to a minimum RDcost and obtained by performing motion estimation on the coding block in a non-translation motion manner. - The
prediction unit 1103 is configured to perform prediction coding on the coding block based on the first coding parameter. - In the solution, referring to
FIG. 12 , thecoding prediction apparatus 110 may further include anacquisition unit 1104 and ajudgment unit 1105. - The
acquisition unit 1104 is configured to acquire a best BV of the coding block. - The
judgment unit 1105 is configured to, when the best BV is equal to 0, calculate a first motion estimation result of the coding block based on the IBCAffine coding mode and a second motion estimation result of the coding block based on the IBC coding mode respectively; when the first motion estimation result is not greater than a preset multiple of the second motion estimation result, select the IBCAffine coding mode for the coding block; and when the first motion estimation result is greater than the preset multiple of the second motion estimation result, select the IBC coding mode for the coding block. - In the solution, the
acquisition unit 1104 is specifically configured to, perform IBC search on the coding block, select at least one reference block matched with a hash key of the coding block, create a first candidate block list of the coding block based on the at least one reference block, traverse the first candidate block list, calculate a BV between the coding block and each reference block in the tint candidate block list, calculate RDcosts corresponding to different BVs based on the BVs respectively and determine the BV corresponding to the minimum RDcost in the RDcosts as the best BV of the coding block. - In the solution, the
motion estimation unit 1102 is specifically configured to perform affine motion model-based motion estimation on the coding block based on the MVPs of the at least two control points through the IBCAffine coding mode to obtain the first coding parameter of the coding block, the first coding parameter indicating the group of coding parameters corresponding to a minimum RDcost and obtained by performing the affine motion model-based motion estimation on the coding block. - In the solution, the
motion estimation unit 1102 is further configured to perform translation motion model-based motion estimation on the coding block based on the IBC coding mode to obtain a second coding parameter of the coding block, the second coding parameter indicating a group of coding parameters corresponding to a minimum RDcost and obtained by performing translation motion model-based motion estimation on the coding block. - The
prediction unit 1103 is further configured to perform prediction coding on the coding block based on the second coding parameter. - In the solution, the
motion estimation unit 1102 is further configured to, when the best BV is not equal to 0, directly determine the best BV as a third coding parameter corresponding to the coding block. - The
prediction unit 1103 is further configured to perform prediction coding on the coding block based on the third coding parameter. - In the solution, the
motion estimation unit 1102 is specifically configured to calculate a prediction value of at least one pixel of the coding block in a reference picture based on the MVPs of the at least two control points; execute an iterative operation on a matching error of the at least one pixel of the coding block between an original picture and the reference picture and on a gradient matrix of the prediction value; update an MV based on the iterative operation; when a count of the iterative operation meets a preset count threshold, obtain an updated MV; and obtain the first coding parameter of the coding block based on the updated MV. - In the solution, the
motion estimation unit 1102 is further configured to, for the coding block of the intra prediction type, create a second candidate block list of the coding block, a reference block in the second candidate block list being spatially neighbouring to the coding block and the reference block being encoded based on the IBCAffine coding mode; traverse the second candidate block list and calculate MVs of control points at corresponding positions of the coding block based on MVs of at least two control points of each reference block in the second candidate block list respectively; and acquire a fourth coding parameter corresponding to the coding block from the MVs, the fourth coding parameter indicating a group of coding parameters corresponding to the minimum RDcost among the obtained MVs of the coding block. - The
prediction unit 1103 is further configured to perform prediction coding on the coding block based on the fourth coding parameter. - It can be understood that, in the embodiment, “unit” may be part of a circuit, part of a processor, part of a program or software and the like, of course, may also be modular and may also be non-modular. In addition, each component in the embodiment may be integrated into a processing unit, each unit may also exist independently, and two or more than two units may also be integrated into a unit. The integrated unit may be implemented in a hardware form and may also be implemented in form of software function module.
- When implemented in form of software function module and sold or used not as an independent product, the integrated unit may be stored in a computer-readable storage medium. Based on such an understanding, the technical solution of the embodiment substantially or parts making contributions to the conventional art or all or part of the technical solution may be embodied in form of software product, and the computer software product is stored in a storage medium, including a plurality of instructions configured to enable a computer device (which may be a personal computer, a server, a network device or the like) or a processor to execute all or part of the operations of the method in the embodiment. The storage medium includes: various media capable of storing program codes such as a U disk, a mobile hard disk, a Read Only Memory (ROM), a Random Access Memory (RAM), a magnetic disk or an optical disk.
- Therefore, the embodiment provides a computer storage medium, which stores a coding prediction program. The coding prediction program is executed by at least one processor to implement the operations of the method in the technical solution shown in
FIG. 6 . - Based on the composition of the
coding prediction apparatus 110 and the computer storage medium, referring toFIG. 13 , a specific hardware structure example of thecoding prediction apparatus 110 according to an embodiment of the disclosure is shown, which may include anetwork interface 1301, amemory 1302 and aprocessor 1303. Each component is coupled together through a bus system 1304. It can be understood that the bus system 1304 is configured to implement connection communication between these components. The bus system 1304 includes a data bus and may further include a power bus, a control bus and a state signal bus. However, for clear description, various buses inFIG. 13 are marked as the bus system 1304. Thenetwork interface 1301 is configured to receive and send a signal in a process of receiving and sending information with another external network element. - The
memory 1302 is configured to store a computer program capable of running in theprocessor 1303. - The
processor 1303 is configured to run the computer program to execute the to liming operations. - For a coding block of an intra prediction type, MVPs of at least two control points of the coding block are determined.
- Affine motion model-based motion estimation is performed on the coding block based on the MVPs of the at least two control points to obtain a first coding parameter of the coding block, the first coding parameter indicating a group of coding parameters corresponding to a minimum RDcost and obtained by performing motion estimation on the coding block in a non-translation motion manner.
- Prediction coding is performed. on the coding block based on the first coding parameter.
- It can be understood that the
memory 1302 in the embodiment of the disclosure may be a volatile memory or a nonvolatile memory, or may include both the volatile and nonvolatile memories. The nonvolatile memory may be a ROM, a Programmable ROM (PROM), an Erasable PROM (EPROM), an Electrically EPROM (EEPROM) or a flash memory. The volatile memory may be a RAM, and is used as an external high-speed cache. It is exemplarily but unlimitedly described that RAMs in various forms may be adopted, such as a Static RAM (SRAM), a Dynamic RAM (DRAM), a Synchronous DRAM (SDRAM), a Double Data Rate SDRAM (DDRSDRAM), an Enhanced SDRAM (ESDRAM), a Synchlink DRAM (SLDRAM) and a Direct Rambus RAM (DRRAM). It is to be noted that thememory 1302 of a system and method described herein is intended to include, but not limited to, memories of these and any other proper types. - The
processor 1303 may be an integrated circuit chip with a signal processing capability. In an implementation process, each operation of the method may be completed by an integrated logic circuit of hardware in theprocessor 1303 or an instruction in a software form. Theprocessor 1303 may be a universal processor, a Digital Signal Processor (DSP), an Application Specific Integrated Circuit (ASIC), a Field Programmable Gate Array (FPGA) or another Programmable Logic Device (PLD), discrete gate or transistor logical device and discrete hardware component. Each method, operation and logical block diagram disclosed in the embodiments of the disclosure may be implemented or executed. The universal processor may be a microprocessor or the processor may also be any conventional processor and the like. The operations of the method disclosed in combination with the embodiments of the disclosure may be directly embodied to be executed and completed by a hardware decoding processor or executed and completed by a combination of hardware and software modules in the decoding processor. The software module may be located in a mature storage medium in this field such as a RAM, a flash memory, a ROM, a PROM or EEPROM and a register. The storage medium is located in thememory 1302. Theprocessor 1303 reads information in thememory 1302 and completes the operations of the method in combination with hardware. - It can be understood that these embodiments described herein may be implemented by hardware, software, firmware, middleware, a microcode or a combination thereof. In case of implementation with the hardware, the processing unit may be implemented in one or more ASICs, DSPs, DSP Devices (DSPDs), PLDs, FPGAs, universal processors, controllers, microcontrollers, microprocessors, other electronic units configured to execute the functions in the disclosure or combinations thereof.
- In case of implementation with the software, the technology of the disclosure may be implemented through the modules (for example, processes and functions) executing the functions in the disclosure. A software code may be stored in the memory and executed by the processor. The memory may be implemented in the processor or outside the processor.
- Optionally, as another embodiment, the
processor 1303 is further configured to run the computer program to implement the operations of the method in the technical solution shown inFIG. 6 . - It is to be noted that terms “include” and “contain” or any other variant thereof is intended to cover nonexclusive inclusions herein, so that a process, method, object or device including a series of elements not only includes those elements but also includes other elements which are not clearly listed or may further include elements intrinsic to the process, the method, the object or the device. Under the condition of no more limitations, an element defined by the statement “including a/an . . . ” does not exclude existence of the same other elements in a process, method, object or device including the element.
- The sequence numbers of the embodiments of the disclosure are adopted not to represent superiority-inferiority of the embodiments but only for description.
- From the above descriptions about the implementation modes, those skilled in the art may clearly know that the method of the abovementioned embodiments may be implemented in a manner of combining software and a necessary universal hardware platform, and of course, may also be implemented through hardware, but the former is a preferred implementation mode under many circumstances. Based on such an understanding, the technical solutions of the disclosure substantially or parts making contributions to the conventional art may be embodied in form of software product, and the computer software product is stored in a storage medium (for example, a ROM/RAM, a magnetic disk and an optical disk), including a plurality of instructions configured to enable a computer (which may be a personal computer, a server, a network device or the like) to execute the method in each embodiment of the disclosure.
- The embodiments of the disclosure are described above in combination with the drawings, but the disclosure is not limited to the abovementioned specific implementation modes. The abovementioned specific implementation modes are not restrictive but only schematic, those of ordinary skill in the art may be inspired by the disclosure to implement many forms without departing from the purpose of the disclosure and the scope of protection of the claims, and all these shall fall within the scope of protection of the disclosure.
- In the embodiments of the disclosure, for a coding block of an intra prediction type, MVPs of at least two control points of the coding block are determined at first; then affine motion model-based motion estimation is performed on the coding block based on the MVPs of the at least two control points to obtain a first coding parameter of the coding block, the first coding parameter indicating a group of coding parameters corresponding to a RDcost and obtained by performing non-translation motion manner-based motion estimation on the coding block; and finally, prediction coding is performed on the coding block based on the first coding parameter. With addition of affine motion model-based motion estimation, The existence of non-translation motions, such as scaling, rotation and transformation, of a coding block of the ultra prediction type in a screen picture may be eliminated, the number of coding bits is further reduced, and the coding rate is further improved.
Claims (20)
Applications Claiming Priority (1)
Application Number | Priority Date | Filing Date | Title |
---|---|---|---|
PCT/CN2018/124504 WO2020133115A1 (en) | 2018-12-27 | 2018-12-27 | Coding prediction method and apparatus, and computer storage medium |
Related Parent Applications (1)
Application Number | Title | Priority Date | Filing Date |
---|---|---|---|
PCT/CN2018/124504 Continuation WO2020133115A1 (en) | 2018-12-27 | 2018-12-27 | Coding prediction method and apparatus, and computer storage medium |
Publications (2)
Publication Number | Publication Date |
---|---|
US20210321112A1 true US20210321112A1 (en) | 2021-10-14 |
US11632553B2 US11632553B2 (en) | 2023-04-18 |
Family
ID=71126695
Family Applications (1)
Application Number | Title | Priority Date | Filing Date |
---|---|---|---|
US17/357,621 Active US11632553B2 (en) | 2018-12-27 | 2021-06-24 | Coding prediction method and apparatus, and computer storage medium |
Country Status (4)
Country | Link |
---|---|
US (1) | US11632553B2 (en) |
EP (1) | EP3902257A4 (en) |
CN (1) | CN113287309A (en) |
WO (1) | WO2020133115A1 (en) |
Cited By (4)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
US20220094968A1 (en) * | 2020-09-24 | 2022-03-24 | Kalyan Goswami | Affine Intra Block Copy Refinement |
US20220248048A1 (en) * | 2019-12-27 | 2022-08-04 | Zhejiang Dahua Technology Co., Ltd. | Radio frequency assembly and electronic device |
CN116156174A (en) * | 2023-02-23 | 2023-05-23 | 格兰菲智能科技有限公司 | Data encoding processing method, device, computer equipment and storage medium |
WO2023171988A1 (en) * | 2022-03-11 | 2023-09-14 | 현대자동차주식회사 | Image encoding/decoding method and apparatus, and recording medium storing bitstream |
Families Citing this family (3)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
CN112218092B (en) * | 2020-10-17 | 2022-09-06 | 浙江大华技术股份有限公司 | Encoding method, apparatus and storage medium for string encoding technique |
CN116170594B (en) * | 2023-04-19 | 2023-07-14 | 中国科学技术大学 | Coding method and device based on rate distortion cost prediction |
CN116760986B (en) * | 2023-08-23 | 2023-11-14 | 腾讯科技(深圳)有限公司 | Candidate motion vector generation method, candidate motion vector generation device, computer equipment and storage medium |
Citations (4)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
US20190028731A1 (en) * | 2016-01-07 | 2019-01-24 | Mediatek Inc. | Method and apparatus for affine inter prediction for video coding system |
US20190215534A1 (en) * | 2016-09-12 | 2019-07-11 | Sony Corporation | Image processing apparatus and image processing method |
US11012715B2 (en) * | 2018-02-08 | 2021-05-18 | Qualcomm Incorporated | Intra block copy for video coding |
US20210258570A1 (en) * | 2018-10-29 | 2021-08-19 | Huawei Technologies Co., Ltd. | Video picture prediction method and apparatus |
Family Cites Families (3)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
JP6614472B2 (en) | 2013-09-30 | 2019-12-04 | サン パテント トラスト | Image encoding method, image decoding method, image encoding device, and image decoding device |
CN105163116B (en) | 2015-08-29 | 2018-07-31 | 华为技术有限公司 | The method and apparatus of image prediction |
US10805630B2 (en) * | 2017-04-28 | 2020-10-13 | Qualcomm Incorporated | Gradient based matching for motion search and derivation |
-
2018
- 2018-12-27 EP EP18945171.9A patent/EP3902257A4/en not_active Withdrawn
- 2018-12-27 CN CN201880100324.9A patent/CN113287309A/en active Pending
- 2018-12-27 WO PCT/CN2018/124504 patent/WO2020133115A1/en unknown
-
2021
- 2021-06-24 US US17/357,621 patent/US11632553B2/en active Active
Patent Citations (4)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
US20190028731A1 (en) * | 2016-01-07 | 2019-01-24 | Mediatek Inc. | Method and apparatus for affine inter prediction for video coding system |
US20190215534A1 (en) * | 2016-09-12 | 2019-07-11 | Sony Corporation | Image processing apparatus and image processing method |
US11012715B2 (en) * | 2018-02-08 | 2021-05-18 | Qualcomm Incorporated | Intra block copy for video coding |
US20210258570A1 (en) * | 2018-10-29 | 2021-08-19 | Huawei Technologies Co., Ltd. | Video picture prediction method and apparatus |
Non-Patent Citations (1)
Title |
---|
Huanbang Chen et al., (hereinafter Chen) "Description of SDR, HDR and 360o video coding proposal by Huawey, GoPro,HiSilicon, and Samsung" JVET-J0025_v2, San Diego, US, 10-20 Apr. 2018. (Year: 2018) * |
Cited By (5)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
US20220248048A1 (en) * | 2019-12-27 | 2022-08-04 | Zhejiang Dahua Technology Co., Ltd. | Radio frequency assembly and electronic device |
US20220094968A1 (en) * | 2020-09-24 | 2022-03-24 | Kalyan Goswami | Affine Intra Block Copy Refinement |
US11818384B2 (en) * | 2020-09-24 | 2023-11-14 | Ofinno, Llc | Affine intra block copy refinement |
WO2023171988A1 (en) * | 2022-03-11 | 2023-09-14 | 현대자동차주식회사 | Image encoding/decoding method and apparatus, and recording medium storing bitstream |
CN116156174A (en) * | 2023-02-23 | 2023-05-23 | 格兰菲智能科技有限公司 | Data encoding processing method, device, computer equipment and storage medium |
Also Published As
Publication number | Publication date |
---|---|
EP3902257A4 (en) | 2022-01-05 |
CN113287309A (en) | 2021-08-20 |
WO2020133115A1 (en) | 2020-07-02 |
EP3902257A1 (en) | 2021-10-27 |
US11632553B2 (en) | 2023-04-18 |
Similar Documents
Publication | Publication Date | Title |
---|---|---|
US20210321112A1 (en) | Coding prediction method and apparatus, and computer storage medium | |
US11917159B2 (en) | Prediction method, encoder, decoder and computer storage medium | |
CN109964484B (en) | Method and apparatus for motion vector symbol prediction in video coding | |
US20190116376A1 (en) | Motion vector predictors using affine motion model in video coding | |
US8675735B2 (en) | Predicted reference information generating method, video encoding and decoding methods, apparatuses therefor, programs therefor, and storage media which store the programs | |
US20180220149A1 (en) | Inter prediction method and device in video coding system | |
CN110944195A (en) | Modification of motion vectors with adaptive motion vector resolution | |
CN118590651A (en) | Multiple hypotheses for sub-block prediction block | |
US11895320B2 (en) | History-based motion vector prediction | |
CN114449288A (en) | Method and apparatus for pattern-based motion vector derivation for video coding | |
CN116708814A (en) | Video encoding and decoding method and apparatus performed by video encoder and decoder | |
CN113273187A (en) | Affine-based Merge with Motion Vector Difference (MVD) | |
JP2024038060A (en) | Method for refining accuracy of sub-pixel of error surface base for decoder side motion vector refinement | |
CN113302936B (en) | Control method for Merge with MVD | |
WO2019242686A1 (en) | Method and apparatus of motion vector buffer management for video coding system | |
CN113826394A (en) | Improvement of adaptive motion vector difference resolution in intra block copy mode | |
CN111466116B (en) | Method and device for affine interframe prediction of video coding and decoding system | |
US12052425B2 (en) | Prediction value determination method, encoder and computer storage medium | |
CN117321995A (en) | Method, apparatus and medium for video processing | |
CN113348667B (en) | Resolution method of distance index under Merge with MVD | |
CN114097219A (en) | Storage of motion information in history-based motion vector prediction tables | |
US20230388484A1 (en) | Method and apparatus for asymmetric blending of predictions of partitioned pictures | |
CN117597921A (en) | Method, apparatus and medium for video processing | |
RU2020138539A (en) | INTERACTION BETWEEN LUT AND AMVP |
Legal Events
Date | Code | Title | Description |
---|---|---|---|
AS | Assignment |
Owner name: GUANGDONG OPPO MOBILE TELECOMMUNICATIONS CORP., LTD., CHINA Free format text: ASSIGNMENT OF ASSIGNORS INTEREST;ASSIGNORS:LIANG, FAN;HAN, HAIYANG;CAO, SIQI;SIGNING DATES FROM 20210615 TO 20210618;REEL/FRAME:056661/0376 |
|
FEPP | Fee payment procedure |
Free format text: ENTITY STATUS SET TO UNDISCOUNTED (ORIGINAL EVENT CODE: BIG.); ENTITY STATUS OF PATENT OWNER: LARGE ENTITY |
|
STPP | Information on status: patent application and granting procedure in general |
Free format text: DOCKETED NEW CASE - READY FOR EXAMINATION |
|
STPP | Information on status: patent application and granting procedure in general |
Free format text: NON FINAL ACTION MAILED |
|
STPP | Information on status: patent application and granting procedure in general |
Free format text: RESPONSE TO NON-FINAL OFFICE ACTION ENTERED AND FORWARDED TO EXAMINER |
|
STPP | Information on status: patent application and granting procedure in general |
Free format text: NOTICE OF ALLOWANCE MAILED -- APPLICATION RECEIVED IN OFFICE OF PUBLICATIONS |
|
STCF | Information on status: patent grant |
Free format text: PATENTED CASE |