WO2020003281A1 - Traitement de flux binaire vidéo à l'aide d'un mode de fusion étendu et d'informations de mouvement signalées d'un bloc - Google Patents

Traitement de flux binaire vidéo à l'aide d'un mode de fusion étendu et d'informations de mouvement signalées d'un bloc Download PDF

Info

Publication number
WO2020003281A1
WO2020003281A1 PCT/IB2019/055590 IB2019055590W WO2020003281A1 WO 2020003281 A1 WO2020003281 A1 WO 2020003281A1 IB 2019055590 W IB2019055590 W IB 2019055590W WO 2020003281 A1 WO2020003281 A1 WO 2020003281A1
Authority
WO
WIPO (PCT)
Prior art keywords
candidates
candidate
list
mvd
precision
Prior art date
Application number
PCT/IB2019/055590
Other languages
English (en)
Inventor
Hongbin Liu
Li Zhang
Kai Zhang
Yue Wang
Original Assignee
Beijing Bytedance Network Technology Co., Ltd.
Bytedance Inc.
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Application filed by Beijing Bytedance Network Technology Co., Ltd., Bytedance Inc. filed Critical Beijing Bytedance Network Technology Co., Ltd.
Publication of WO2020003281A1 publication Critical patent/WO2020003281A1/fr

Links

Classifications

    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04NPICTORIAL COMMUNICATION, e.g. TELEVISION
    • H04N19/00Methods or arrangements for coding, decoding, compressing or decompressing digital video signals
    • H04N19/10Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using adaptive coding
    • H04N19/169Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using adaptive coding characterised by the coding unit, i.e. the structural portion or semantic portion of the video signal being the object or the subject of the adaptive coding
    • H04N19/17Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using adaptive coding characterised by the coding unit, i.e. the structural portion or semantic portion of the video signal being the object or the subject of the adaptive coding the unit being an image region, e.g. an object
    • H04N19/176Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using adaptive coding characterised by the coding unit, i.e. the structural portion or semantic portion of the video signal being the object or the subject of the adaptive coding the unit being an image region, e.g. an object the region being a block, e.g. a macroblock
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04NPICTORIAL COMMUNICATION, e.g. TELEVISION
    • H04N19/00Methods or arrangements for coding, decoding, compressing or decompressing digital video signals
    • H04N19/10Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using adaptive coding
    • H04N19/102Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using adaptive coding characterised by the element, parameter or selection affected or controlled by the adaptive coding
    • H04N19/103Selection of coding mode or of prediction mode
    • H04N19/105Selection of the reference unit for prediction within a chosen coding or prediction mode, e.g. adaptive choice of position and number of pixels used for prediction
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04NPICTORIAL COMMUNICATION, e.g. TELEVISION
    • H04N19/00Methods or arrangements for coding, decoding, compressing or decompressing digital video signals
    • H04N19/10Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using adaptive coding
    • H04N19/134Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using adaptive coding characterised by the element, parameter or criterion affecting or controlling the adaptive coding
    • H04N19/157Assigned coding mode, i.e. the coding mode being predefined or preselected to be further used for selection of another element or parameter
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04NPICTORIAL COMMUNICATION, e.g. TELEVISION
    • H04N19/00Methods or arrangements for coding, decoding, compressing or decompressing digital video signals
    • H04N19/10Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using adaptive coding
    • H04N19/169Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using adaptive coding characterised by the coding unit, i.e. the structural portion or semantic portion of the video signal being the object or the subject of the adaptive coding
    • H04N19/184Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using adaptive coding characterised by the coding unit, i.e. the structural portion or semantic portion of the video signal being the object or the subject of the adaptive coding the unit being bits, e.g. of the compressed video stream
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04NPICTORIAL COMMUNICATION, e.g. TELEVISION
    • H04N19/00Methods or arrangements for coding, decoding, compressing or decompressing digital video signals
    • H04N19/50Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using predictive coding
    • H04N19/503Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using predictive coding involving temporal prediction
    • H04N19/51Motion estimation or motion compensation
    • H04N19/513Processing of motion vectors
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04NPICTORIAL COMMUNICATION, e.g. TELEVISION
    • H04N19/00Methods or arrangements for coding, decoding, compressing or decompressing digital video signals
    • H04N19/50Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using predictive coding
    • H04N19/503Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using predictive coding involving temporal prediction
    • H04N19/51Motion estimation or motion compensation
    • H04N19/513Processing of motion vectors
    • H04N19/517Processing of motion vectors by encoding
    • H04N19/52Processing of motion vectors by encoding by predictive encoding
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04NPICTORIAL COMMUNICATION, e.g. TELEVISION
    • H04N19/00Methods or arrangements for coding, decoding, compressing or decompressing digital video signals
    • H04N19/50Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using predictive coding
    • H04N19/503Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using predictive coding involving temporal prediction
    • H04N19/51Motion estimation or motion compensation
    • H04N19/513Processing of motion vectors
    • H04N19/521Processing of motion vectors for estimating the reliability of the determined motion vectors or motion vector field, e.g. for smoothing the motion vector field or for correcting motion vectors
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04NPICTORIAL COMMUNICATION, e.g. TELEVISION
    • H04N19/00Methods or arrangements for coding, decoding, compressing or decompressing digital video signals
    • H04N19/70Methods or arrangements for coding, decoding, compressing or decompressing digital video signals characterised by syntax aspects related to video coding, e.g. related to compression standards
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04NPICTORIAL COMMUNICATION, e.g. TELEVISION
    • H04N19/00Methods or arrangements for coding, decoding, compressing or decompressing digital video signals
    • H04N19/50Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using predictive coding
    • H04N19/503Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using predictive coding involving temporal prediction
    • H04N19/51Motion estimation or motion compensation
    • H04N19/523Motion estimation or motion compensation with sub-pixel accuracy

Definitions

  • FIG. 7 shows an example of combined bi-predictive merge candidate.
  • FIG. 9 shows an example illustration of motion vector scaling for spatial motion vector candidate.
  • FIG. 10 shows an example of neighboring samples used for deriving IC parameters.
  • FIG. 13 shows an example of MVP for AF_INTER.
  • FIGS. 14A and 14B show examples of candidates for AF_MERGE.
  • FIG. 21 shows examples of non-adjacent merge candidates.
  • the present document provides various techniques that can be used by a decoder of video bitstreams to improve the quality of decompressed or decoded digital video. Furthermore, a video encoder may also implement these techniques during the process of encoding in order to reconstruct decoded frames used for further encoding.
  • a merge mode is specified whereby the motion parameters for the current PU are obtained from neighboring PUs, including spatial and temporal candidates.
  • the merge mode can be applied to any inter-predicted PU, not only for skip mode.
  • the alternative to merge mode is the explicit transmission of motion parameters, where motion vector (to be more precise, motion vector difference compared to a motion vector predictor), corresponding reference picture index for each reference picture list and reference picture list usage are signaled explicitly per each PU.
  • Such a mode is named Advanced motion vector prediction (AMVP) in this document.
  • the PU is produced from two blocks of samples. This is referred to as‘bi-prediction’. Bi-prediction is available for B-slices only.
  • Step 1.2 Redundancy check for spatial candidates
  • the position for the temporal candidate is selected between candidates Co and Ci, as depicted in FIG. 6. If PU at position Co is not available, is intra coded, or is outside of the current CTU row, position Ci is used. Otherwise, position Co is used in the derivation of the temporal merge candidate.
  • merge candidates Besides spatial and temporal merge candidates, there are two additional types of merge candidates: combined bi-predictive merge candidate and zero merge candidate.
  • Combined bi- predictive merge candidates are generated by utilizing spatial and temporal merge candidates.
  • Combined bi-predictive merge candidate is used for B-Slice only.
  • the combined bi-predictive candidates are generated by combining the first reference picture list motion parameters of an initial candidate with the second reference picture list motion parameters of another. If these two tuples provide different motion hypotheses, they will form a new bi-predictive candidate. As an example, FIG.
  • motion estimation can be performed in parallel whereby the motion vectors for all prediction units inside a given region are derived
  • HEVC defines the motion estimation region (MER) whose size is signalled in the picture parameter set using the
  • FIG. 8 summarizes derivation process for motion vector prediction candidate.
  • one motion vector candidate is selected from two candidates, which are derived based on two different co-located positions. After the first list of spatio-temporal candidates is made, duplicated motion vector candidates in the list are removed. If the number of potential candidates is larger than two, motion vector candidates whose reference picture index within the associated reference picture list is larger than 1 are removed from the list. If the number of spatio-temporal motion vector candidates is smaller than two, additional zero motion vector candidates is added to the list.
  • the quarter luma sample MV resolution is used for the CU.
  • the MVPs in the AMVP candidate list for the CU are rounded to the corresponding precision.
  • motion vector accuracy is one-quarter pel (one -quarter luma sample and one- eighth chroma sample for 4:2:0 video).
  • JEM the accuracy for the internal motion vector storage and the merge candidate increases to 1/16 pel.
  • the higher motion vector accuracy (1/16 pel) is used in motion compensation inter prediction for the CU coded with skip/merge mode.
  • LIC Local Illumination Compensation
  • CU inter-mode coded coding unit
  • a least square error method is employed to derive the parameters a and b by using the neighbouring samples of the current CU and their corresponding reference samples. More specifically, as illustrated in FIG. 10, the subsampled (2:1 subsampling) neighbouring samples of the CU and the corresponding samples (identified by motion information of the current CU or sub-CU) in the reference picture are used. The IC parameters are derived and applied for each prediction direction separately.
  • LIC is disabled for the entire picture when there is no obvious illumination change between a current picture and its reference pictures. To identify this situation, histograms of a current picture and every reference picture of the current picture are calculated at the encoder. If the histogram difference between the current picture and every reference picture of the current picture is smaller than a given threshold, LIC is disabled for the current picture; otherwise, LIC is enabled for the current picture.
  • Equation 2 M and N should be adjusted downward if necessary to make it a divisor of w and h, respectively.
  • an index indicating the position of the CPMVP in the candidate list is signalled in the bitstream.
  • a FRUC flag is signalled for a CU when its merge flag is tme.
  • a merge index is signalled and the regular merge mode is used.
  • an additional FRUC mode flag is signalled to indicate which method (bilateral matching or template matching) is to be used to derive motion information for the block.
  • the decision on whether using FRUC merge mode for a CU is based on RD cost selection as done for normal merge candidate. That is the two matching modes (bilateral matching and template matching) are both checked for a CU by using RD cost selection. The one leading to the minimal cost is further compared to other CU modes. If a FRUC matching mode is the most efficient one, FRUC flag is set to tme for the CU and the related matching mode is used.
  • Motion derivation process in FRUC merge mode has two steps.
  • a CU-level motion search is first performed, then followed by a Sub-CU level motion refinement.
  • an initial motion vector is derived for the whole CU based on bilateral matching or template matching.
  • a list of MV candidates is generated and the candidate which leads to the minimum matching cost is selected as the starting point for further CU level refinement.
  • a local search based on bilateral matching or template matching around the starting point is performed and the MV results in the minimum matching cost is taken as the MV for the whole CU.
  • the motion information is further refined at sub-CU level with the derived CU motion vectors as the starting points.
  • the following derivation process is performed for a W c H CU motion information derivation.
  • H CU is derived.
  • the CU is further split into M c M sub-CUs.
  • the value of M is calculated as in (3), D is a predefined splitting depth which is set to 3 by default in the JEM.
  • the MV for each sub-CU is derived.
  • the bilateral matching is used to derive motion information of the current CU by finding the closest match between two blocks along the motion trajectory of the current CU in two different reference pictures.
  • the motion vectors MV0 and MV1 pointing to the two reference blocks shall be proportional to the temporal distances, i.e., TD0 and TD1, between the current picture and the two reference pictures.
  • the bilateral matching becomes mirror based bi-directional MV.
  • each valid MV of a merge candidate is used as an input to generate a MV pair with the assumption of bilateral matching.
  • one valid MV of a merge candidate is (MVa, refa) at reference list A.
  • the reference picture refb of its paired bilateral MV is found in the other reference list B so that refa and refb are temporally at different sides of the current picture. If such a refb is not available in reference list B, refb is determined as a reference which is different from refa and its temporal distance to the current picture is the minimal one in list B.
  • MVb is derived by scaling MVa based on the temporal distance between the current picture and refa, refb.
  • MVs from the interpolated MV field are also added to the CU level candidate list. More specifically, the interpolated MVs at the position (0, 0), (W/2, 0), (0, H/2) and (W/2, H/2) of the current CU are added.
  • the MV candidate set at sub-CU level consists of:
  • ATMVP and STMVP candidates are limited to the four first ones.
  • interpolated motion field is generated for the whole picture based on unilateral ME. Then the motion field may be used later as CU level or sub-CU level MV candidates.
  • the matching cost is a bit different at different steps.
  • the matching cost is the sum of absolute difference (SAD) of bilateral matching or template matching.
  • SAD absolute difference
  • the matching cost C of bilateral matching at sub-CU level search is calculated as follows:
  • MV and MV S indicate the current MV and the starting MV, respectively.
  • SAD is still used as the matching cost of template matching at sub-CU level search.
  • MV is derived by using luma samples only. The derived motion will be used for both luma and chroma for MC inter prediction. After MV is decided, final MC is performed using 8-taps interpolation filter for luma and 4-taps interpolation filter for chroma.
  • MV refinement is a pattern based MV search with the criterion of bilateral matching cost or template matching cost.
  • two search patterns are supported - an unrestricted center-biased diamond search (UCBDS) and an adaptive cross search for MV refinement at the CU level and sub-CU level, respectively.
  • UMBDS center-biased diamond search
  • the MV is directly searched at quarter luma sample MV accuracy, and this is followed by one-eighth luma sample MV refinement.
  • the search range of MV refinement for the CU and sub-CU step are set equal to 8 luma samples.
  • the encoder can choose among uni-prediction from listO, uni-prediction from listl or bi-prediction for a CU. The selection is based on a template matching cost as follows:
  • costBi ⁇ factor * min (costO, costl )
  • costO is the SAD of listO template matching
  • costl is the SAD of listl template matching
  • costBi is the SAD of bi -prediction template matching.
  • the value of factor is equal to 1.25, which means that the selection process is biased toward bi-prediction.
  • the inter prediction direction selection is only applied to the CU-level template matching process.
  • bi-prediction operation for the prediction of one block region, two prediction blocks, formed using a motion vector (MV) of listO and a MV of listl, respectively, are combined to form a single prediction signal.
  • MV motion vector
  • DMVR decoder-side motion vector refinement
  • the two motion vectors of the bi-prediction are further refined by a bilateral template matching process.
  • the bilateral template matching applied in the decoder to perform a distortion-based search between a bilateral template and the reconstmction samples in the reference pictures in order to obtain a refined MV without transmission of additional motion information.
  • a bilateral template is generated as the weighted combination (i.e. average) of the two prediction blocks, from the initial MV0 of listO and MV1 of listl, respectively, as shown in FIG. 18.
  • the template matching operation consists of calculating cost measures between the generated template and the sample region (around the initial prediction block) in the reference picture. For each of the two reference pictures, the MV that yields the minimum template cost is considered as the updated MV of that list to replace the original one.
  • nine MV candidates are searched for each list. The nine MV candidates include the original MV and 8 surrounding MVs with one luma sample offset to the original MV in either the horizontal or vertical direction, or both.
  • the two new MVs i.e., MVO' and MV1' as shown in FIG. 18, are used for generating the final bi -prediction results.
  • a sum of absolute differences (SAD) is used as the cost measure.
  • SAD sum of absolute differences
  • Tencent proposes to derive additional spatial merge candidates from positions in an outer reference area which has an offset of (-96, -96) to the current block.
  • each candidate B (i, j) or C (i, j) has an offset of 16 in the vertical direction compared to its previous B or C candidates.
  • Each candidate A (i, j) or D (i, j) has an offset of 16 in the horizontal direction compared to its previous A or D candidates.
  • Each E (i, j) has an offset of 16 in both horizontal direction and vertical direction compared to its previous E candidates. The candidates are checked from inside to the outside.
  • FIG. 22 shows an example of a UMVE search process
  • FIG. 23 shows an example of UMVE search points.
  • UMVE provides a new motion vector expression with simplified signaling.
  • the expression method includes starting point, motion magnitude, and motion direction.
  • Base candidate index defines the starting point.
  • Base candidate index indicates the best candidate among candidates in the list as follows.
  • Distance index is motion magnitude information.
  • Distance index indicates the pre defined distance from the starting point information. Pre-defined distance is as follows:
  • motion information of a merge candidate is inherited by current block, including motion vector, reference pictures, prediction direction, LIC flag etc. Only a merge index is signaled, which is efficient in many cases. However, the inherited motion information, especially motion vector maybe not good enough.
  • the MVD can only has non-zero component in either horizontal direction or vertical direction but not both direction. Meanwhile, it also signals the MVD information, i.e., the distance index or motion magnitude information.
  • motion information such as prediction direction, reference indices/pictures, motion vectors, LIC flag, affine flag, Intra Block Copy (IBC) flag, MVD precision, MVD values
  • IBC Intra Block Copy
  • an EMM list is constmcted, and an index is signaled to indicate the first part of motion information of which candidate is inherited by the current block (e.g., PU/CU). Meanwhile, additional information (i.e., 2nd part of the motion information) like MVD is further signaled.
  • the motion information candidate list is constmcted by inserting motion information of spatial neighboring blocks, temporal neighboring blocks or non-adjacent blocks.
  • the prediction direction is not inherited and is explicitly signaled. In this case, it is proposed to constmct two or multiple motion information candidate lists.
  • the MVD precision is set to the highest supported precision (e.g., 1 ⁇ 4) by default.
  • the MVD precision is set to the most frequent MVD precision appeared in the merge candidate list.
  • the MVD precision is set to an arbitrary valid MVD precision.
  • v. Rounding operations may be applied to achieve this.
  • the MVD precision is set to the highest supported precision (e.g., 1 ⁇ 4) by default.
  • the MVD precision is set to the most frequent MVD precision appeared in the merge candidate list.
  • the MVD precision is set to an arbitrary valid MVD precision.
  • EMM mode can work together with DMVR or template matching. In this case, some of or all candidates are further refined by DMVR or template matching.
  • FRUC is not further refined.
  • FIG. 24 is a flowchart for an example method 2400 of processing a video bitstream.
  • the method 2400 includes constructing (2402) a list of EMM candidates; determining (2404), based on a first set of bits in a bitstream representation of a current block, the motion information inherited by the current block from the list; determining (2406), based on a second set of bits in the bitstream representation of a current block, the signaled motion information of the current block; and performing (2408), based on the list of EMM candidates and the signaled motion information, a conversion between the current block and the bitstream representation, wherein a motion vector difference (MVD) precision of the list of EMM candidates is based on at least one candidate inserted into the list of EMM candidates.
  • MMD motion vector difference
  • a method of video bitstream processing comprising: constmcting a list of extended merge mode (EMM) candidates; determining, based on a first set of bits in a bitstream representation of a current block, the motion information inherited by the current block from the list; determining, based on a second set of bits in the bitstream representation of a current block, the signaled motion information of the current block; and performing, based on the list of EMM candidates and the signaled motion information, a conversion between the current block and the bitstream representation, wherein a motion vector difference (MVD) precision of the list of EMM candidates is based on at least one candidate inserted into the list of EMM candidates.
  • EMM extended merge mode
  • the characteristic includes the size of the current block being greater than a threshold.
  • An apparatus in a video system comprising a processor and a non-transitory memory with instmctions thereon, wherein the instructions upon execution by the processor, cause the processor to implement the method in any one of examples 1 to 18.
  • JEM-7 https://jvet.hhi.fraunhofer.de/svn/svn_HMJEMSoftware/tags/ HM-l6.6- JEM-7.0.
  • FIG. 25 is a block diagram of a video processing apparatus 2500.
  • the apparatus 2500 may be used to implement one or more of the methods described herein.
  • the apparatus 2500 may be embodied in a smartphone, tablet, computer, Internet of Things (IoT) receiver, and so on.
  • the apparatus 2500 may include one or more processors 2502, one or more memories 2504 and video processing hardware 2506.
  • the processor(s) 2502 may be configured to implement one or more methods (including, but not limited to, method 2400) described in the present document.
  • the memory (memories) 2504 may be used for storing data and code used for implementing the methods and techniques described herein.
  • the video processing hardware 2506 may be used to implement, in hardware circuitry, some techniques described in the present document.
  • the video coding methods may be implemented using an apparatus that is implemented on a hardware platform as described with respect to FIG. 25.
  • the disclosed and other solutions, examples, embodiments, modules and the functional operations described in this document can be implemented in digital electronic circuitry, or in computer software, firmware, or hardware, including the structures disclosed in this document and their structural equivalents, or in combinations of one or more of them.
  • the disclosed and other embodiments can be implemented as one or more computer program products, i.e., one or more modules of computer program instmctions encoded on a computer readable medium for execution by, or to control the operation of, data processing apparatus.
  • the computer readable medium can be a machine -readable storage device, a machine-readable storage substrate, a memory device, a composition of matter effecting a machine-readable propagated signal, or a combination of one or more them.
  • the term“data processing apparatus” encompasses all apparatus, devices, and machines for processing data, including by way of example a
  • the apparatus can include, in addition to hardware, code that creates an execution environment for the computer program in question, e.g., code that constitutes processor firmware, a protocol stack, a database management system, an operating system, or a combination of one or more of them.
  • a propagated signal is an artificially generated signal, e.g., a machine-generated electrical, optical, or electromagnetic signal, that is generated to encode information for transmission to suitable receiver apparatus.
  • the processes and logic flows described in this document can be performed by one or more programmable processors executing one or more computer programs to perform functions by operating on input data and generating output.
  • the processes and logic flows can also be performed by, and apparatus can also be implemented as, special purpose logic circuitry, e.g., an FPGA (field programmable gate array) or an ASIC (application specific integrated circuit).
  • processors suitable for the execution of a computer program include, by way of example, both general and special purpose microprocessors, and any one or more processors of any kind of digital computer.
  • a processor will receive instructions and data from a read only memory or a random-access memory or both.
  • the essential elements of a computer are a processor for performing instmctions and one or more memory devices for storing instructions and data.
  • a computer will also include, or be operatively coupled to receive data from or transfer data to, or both, one or more mass storage devices for storing data, e.g., magnetic, magneto optical disks, or optical disks.
  • mass storage devices for storing data, e.g., magnetic, magneto optical disks, or optical disks.
  • a computer need not have such devices.

Landscapes

  • Engineering & Computer Science (AREA)
  • Multimedia (AREA)
  • Signal Processing (AREA)
  • Compression Or Coding Systems Of Tv Signals (AREA)

Abstract

L'invention concerne des procédés, des dispositifs et des systèmes d'utilisation d'un mode de fusion étendu (EMM) en codage vidéo. Un exemple de procédé de traitement vidéo consiste à construire une liste de modes de fusion étendus (EMM) candidats ; à déterminer, sur la base d'un premier ensemble de bits dans une représentation par flux binaire d'un bloc courant, les informations de mouvement héritées par le bloc courant à partir de la liste ; à déterminer, sur la base d'un second ensemble de bits dans la représentation par flux binaire d'un bloc courant, les informations de mouvement signalées du bloc courant ; et à effectuer, sur la base de la liste d'EMM candidats et des informations de mouvement signalées, une conversion entre le bloc courant et la représentation par flux binaire, une précision de différence de vecteurs de mouvement (MVD) de la liste d'EMM candidats étant basée sur au moins un candidat inséré dans la liste d'EMM candidats.
PCT/IB2019/055590 2018-06-29 2019-07-01 Traitement de flux binaire vidéo à l'aide d'un mode de fusion étendu et d'informations de mouvement signalées d'un bloc WO2020003281A1 (fr)

Applications Claiming Priority (2)

Application Number Priority Date Filing Date Title
CN2018093646 2018-06-29
CNPCT/CN2018/093646 2018-06-29

Publications (1)

Publication Number Publication Date
WO2020003281A1 true WO2020003281A1 (fr) 2020-01-02

Family

ID=67253944

Family Applications (3)

Application Number Title Priority Date Filing Date
PCT/IB2019/055590 WO2020003281A1 (fr) 2018-06-29 2019-07-01 Traitement de flux binaire vidéo à l'aide d'un mode de fusion étendu et d'informations de mouvement signalées d'un bloc
PCT/IB2019/055583 WO2020003276A1 (fr) 2018-06-29 2019-07-01 Signalisation de mode emm
PCT/IB2019/055579 WO2020003273A1 (fr) 2018-06-29 2019-07-01 Mode de fusion étendue

Family Applications After (2)

Application Number Title Priority Date Filing Date
PCT/IB2019/055583 WO2020003276A1 (fr) 2018-06-29 2019-07-01 Signalisation de mode emm
PCT/IB2019/055579 WO2020003273A1 (fr) 2018-06-29 2019-07-01 Mode de fusion étendue

Country Status (3)

Country Link
CN (3) CN110662055B (fr)
TW (3) TWI736923B (fr)
WO (3) WO2020003281A1 (fr)

Cited By (3)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
KR20210046750A (ko) * 2018-08-28 2021-04-28 에프쥐 이노베이션 컴퍼니 리미티드 비디오 데이터를 코딩하기 위한 디바이스 및 방법
US20210235094A1 (en) * 2018-07-13 2021-07-29 Tencent America LLC Method and apparatus for video coding
WO2022214097A1 (fr) * 2021-04-09 2022-10-13 Beijing Bytedance Network Technology Co., Ltd. Procédé, dispositif et support de traitement vidéo

Families Citing this family (6)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
KR20220042125A (ko) 2019-08-10 2022-04-04 베이징 바이트댄스 네트워크 테크놀로지 컴퍼니, 리미티드 서브픽처 디코딩에서의 버퍼 관리
JP7322290B2 (ja) 2019-10-02 2023-08-07 北京字節跳動網絡技術有限公司 ビデオビットストリームにおけるサブピクチャシグナリングのための構文
CN117676135A (zh) 2019-10-18 2024-03-08 北京字节跳动网络技术有限公司 子图片与环路滤波之间的相互影响
EP4074038A4 (fr) * 2020-01-12 2023-01-25 Beijing Bytedance Network Technology Co., Ltd. Contraintes de codage et de décodage vidéo
WO2021192892A1 (fr) * 2020-03-27 2021-09-30 株式会社コナミデジタルエンタテインメント Système de distribution de vidéo, procédé de commande de distribution de vidéo et programme informatique
WO2022214077A1 (fr) * 2021-04-10 2022-10-13 Beijing Bytedance Network Technology Co., Ltd. Amélioration de mouvement gpm

Citations (2)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US20180098087A1 (en) * 2016-09-30 2018-04-05 Qualcomm Incorporated Frame rate up-conversion coding mode
EP3468194A1 (fr) * 2017-10-05 2019-04-10 Thomson Licensing Inférence et prédiction de mode découplé

Family Cites Families (11)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
WO2013009104A2 (fr) * 2011-07-12 2013-01-17 한국전자통신연구원 Procédé de prédiction inter et appareil pour celui-ci
CN107197272B (zh) * 2011-08-29 2019-12-06 苗太平洋控股有限公司 以合并模式对图像编码的方法
US9357214B2 (en) * 2012-12-07 2016-05-31 Qualcomm Incorporated Advanced merge/skip mode and advanced motion vector prediction (AMVP) mode for 3D video
KR101854003B1 (ko) * 2013-07-02 2018-06-14 경희대학교 산학협력단 복수의 레이어를 포함하는 영상의 부호화 및 복호화 방법
CN103561263B (zh) * 2013-11-06 2016-08-24 北京牡丹电子集团有限责任公司数字电视技术中心 基于运动矢量约束和加权运动矢量的运动补偿预测方法
CN106031170B (zh) * 2014-04-01 2019-04-26 联发科技股份有限公司 运动信息编码的方法
US11330284B2 (en) * 2015-03-27 2022-05-10 Qualcomm Incorporated Deriving motion information for sub-blocks in video coding
US10812791B2 (en) * 2016-09-16 2020-10-20 Qualcomm Incorporated Offset vector identification of temporal motion vector predictor
US10750190B2 (en) * 2016-10-11 2020-08-18 Lg Electronics Inc. Video decoding method and device in video coding system
CN107396106A (zh) * 2017-06-26 2017-11-24 深圳市亿联智能有限公司 一种基于h.265编码标准的视频加密算法
CN107396102B (zh) * 2017-08-30 2019-10-08 中南大学 一种基于Merge技术运动矢量的帧间模式快速选择方法及装置

Patent Citations (2)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US20180098087A1 (en) * 2016-09-30 2018-04-05 Qualcomm Incorporated Frame rate up-conversion coding mode
EP3468194A1 (fr) * 2017-10-05 2019-04-10 Thomson Licensing Inférence et prédiction de mode découplé

Non-Patent Citations (4)

* Cited by examiner, † Cited by third party
Title
"Algorithm description of Joint Exploration Test Model 2 (JEM2)", 114. MPEG MEETING;22-2-2016 - 26-2-2016; SAN DIEGO; (MOTION PICTURE EXPERT GROUP OR ISO/IEC JTC1/SC29/WG11),, no. N16066, 4 April 2016 (2016-04-04), XP030022739 *
A. ALSHINE. ALSHINA: "Description of SDR, HDR and 360° video coding technology proposal by Samsung, Huawei, GoPro, and HiSilicon - mobile application scenario", JVET-J0024, April 2018 (2018-04-01)
C. ROSEWARNEB. BROSSM. NACCARIK. SHARMANG. SULLIVAN: "High Efficiency Video Coding (HEVC) Test Model 16 (HM 16) Improved Encoder Description Update 7", JCTVC-Y1002, October 2016 (2016-10-01)
J. CHENE. ALSHINAG. J. SULLIVANJ.-R. OHMJ. BOYCE: "Algorithm description of Joint Exploration Test Model 7 (JEM7", JVET-G1001, August 2017 (2017-08-01)

Cited By (6)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US20210235094A1 (en) * 2018-07-13 2021-07-29 Tencent America LLC Method and apparatus for video coding
US11611756B2 (en) * 2018-07-13 2023-03-21 Tencent America LLC Method and apparatus for video coding
KR20210046750A (ko) * 2018-08-28 2021-04-28 에프쥐 이노베이션 컴퍼니 리미티드 비디오 데이터를 코딩하기 위한 디바이스 및 방법
EP3844950A4 (fr) * 2018-08-28 2022-05-25 FG Innovation Company Limited Dispositif et procédé de codage de données vidéo
KR102585855B1 (ko) 2018-08-28 2023-10-11 에프쥐 이노베이션 컴퍼니 리미티드 비디오 데이터를 코딩하기 위한 디바이스 및 방법
WO2022214097A1 (fr) * 2021-04-09 2022-10-13 Beijing Bytedance Network Technology Co., Ltd. Procédé, dispositif et support de traitement vidéo

Also Published As

Publication number Publication date
TW202002651A (zh) 2020-01-01
TW202017370A (zh) 2020-05-01
CN110662041A (zh) 2020-01-07
WO2020003276A1 (fr) 2020-01-02
CN110662046B (zh) 2022-03-25
TW202002650A (zh) 2020-01-01
TWI722467B (zh) 2021-03-21
TWI736923B (zh) 2021-08-21
CN110662055A (zh) 2020-01-07
WO2020003273A1 (fr) 2020-01-02
CN110662046A (zh) 2020-01-07
TWI731362B (zh) 2021-06-21
CN110662041B (zh) 2022-07-29
CN110662055B (zh) 2022-07-05

Similar Documents

Publication Publication Date Title
US11477458B2 (en) Mode dependent motion vector difference precision set
US11284068B2 (en) Indication method of maximum number of candidates
US11876932B2 (en) Size selective application of decoder side refining tools
US11956465B2 (en) Difference calculation based on partial position
WO2020108650A1 (fr) Procédé d'indication de drapeau dans un mode de copie à l'intérieur des blocs
WO2019234673A1 (fr) Dmvr de chrominance
CN110662041B (zh) 视频比特流处理的方法和装置,存储视频比特流的方法和非暂时性计算机可读记录介质
WO2020177683A1 (fr) Activation de bio sur la base des informations dans l'en-tête d'image
WO2020151765A1 (fr) Interpolation pour bi-prédiction avec pondération au niveau des unités de codage (cu)
WO2020003262A1 (fr) Mode de bi-prédiction symétrique de codage vidéo
US11863771B2 (en) Updating of history based motion vector prediction tables
WO2020182187A1 (fr) Poids adaptatif dans une prédiction à hypothèses multiples dans un codage vidéo

Legal Events

Date Code Title Description
121 Ep: the epo has been informed by wipo that ep was designated in this application

Ref document number: 19740072

Country of ref document: EP

Kind code of ref document: A1

NENP Non-entry into the national phase

Ref country code: DE

32PN Ep: public notification in the ep bulletin as address of the adressee cannot be established

Free format text: NOTING OF LOSS OF RIGHTS PURSUANT TO RULE 112(1) EPC (EPO FORM 1205A DATED 26.04.2021)

122 Ep: pct application non-entry in european phase

Ref document number: 19740072

Country of ref document: EP

Kind code of ref document: A1