WO2023055153A1 - Procédé, dispositif et support d'enregistrement pour le codage/décodage d'image - Google Patents

Procédé, dispositif et support d'enregistrement pour le codage/décodage d'image Download PDF

Info

Publication number
WO2023055153A1
WO2023055153A1 PCT/KR2022/014710 KR2022014710W WO2023055153A1 WO 2023055153 A1 WO2023055153 A1 WO 2023055153A1 KR 2022014710 W KR2022014710 W KR 2022014710W WO 2023055153 A1 WO2023055153 A1 WO 2023055153A1
Authority
WO
WIPO (PCT)
Prior art keywords
block
information
prediction
mode
neural network
Prior art date
Application number
PCT/KR2022/014710
Other languages
English (en)
Korean (ko)
Inventor
권형진
김동현
김연희
김종호
도지훈
이주영
임웅
정세윤
최진수
이태진
강현구
김동욱
정승원
Original Assignee
한국전자통신연구원
고려대학교 산학협력단
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Application filed by 한국전자통신연구원, 고려대학교 산학협력단 filed Critical 한국전자통신연구원
Priority claimed from KR1020220124695A external-priority patent/KR20230046269A/ko
Publication of WO2023055153A1 publication Critical patent/WO2023055153A1/fr

Links

Images

Classifications

    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06TIMAGE DATA PROCESSING OR GENERATION, IN GENERAL
    • G06T9/00Image coding
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04NPICTORIAL COMMUNICATION, e.g. TELEVISION
    • H04N19/00Methods or arrangements for coding, decoding, compressing or decompressing digital video signals
    • H04N19/10Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using adaptive coding
    • H04N19/102Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using adaptive coding characterised by the element, parameter or selection affected or controlled by the adaptive coding
    • H04N19/119Adaptive subdivision aspects, e.g. subdivision of a picture into rectangular or non-rectangular coding blocks
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04NPICTORIAL COMMUNICATION, e.g. TELEVISION
    • H04N19/00Methods or arrangements for coding, decoding, compressing or decompressing digital video signals
    • H04N19/10Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using adaptive coding
    • H04N19/102Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using adaptive coding characterised by the element, parameter or selection affected or controlled by the adaptive coding
    • H04N19/129Scanning of coding units, e.g. zig-zag scan of transform coefficients or flexible macroblock ordering [FMO]
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04NPICTORIAL COMMUNICATION, e.g. TELEVISION
    • H04N19/00Methods or arrangements for coding, decoding, compressing or decompressing digital video signals
    • H04N19/10Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using adaptive coding
    • H04N19/102Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using adaptive coding characterised by the element, parameter or selection affected or controlled by the adaptive coding
    • H04N19/132Sampling, masking or truncation of coding units, e.g. adaptive resampling, frame skipping, frame interpolation or high-frequency transform coefficient masking
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04NPICTORIAL COMMUNICATION, e.g. TELEVISION
    • H04N19/00Methods or arrangements for coding, decoding, compressing or decompressing digital video signals
    • H04N19/50Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using predictive coding
    • H04N19/593Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using predictive coding involving spatial prediction techniques
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04NPICTORIAL COMMUNICATION, e.g. TELEVISION
    • H04N19/00Methods or arrangements for coding, decoding, compressing or decompressing digital video signals
    • H04N19/80Details of filtering operations specially adapted for video compression, e.g. for pixel interpolation
    • H04N19/82Details of filtering operations specially adapted for video compression, e.g. for pixel interpolation involving filtering within a prediction loop

Definitions

  • the present invention relates to a method, apparatus and recording medium for video encoding/decoding.
  • the present invention is the benefit of the filing date of Korean Patent Application No. 10-2021-0129157 filed on September 29, 2021, the benefit of the filing date of Korean Patent Application No. 10-2022-0107154 filed on August 25, 2022, and the Claims the benefit of the filing date of Korean Patent Application No. 10-2022-0124695 filed on the 29th of the month, all of which are incorporated herein.
  • HD high definition
  • Ultra High Definition (UHD) TV which has four times higher resolution than FHD TV, as well as High Definition TV (HDTV) and Full HD (FHD) TV. has increased, and according to this increase in interest, an image encoding/decoding technology for an image having a higher resolution and quality is required.
  • UHD Ultra High Definition
  • HDTV High Definition TV
  • FHD Full HD
  • video compression technology there are various technologies such as inter prediction technology, intra prediction technology, transformation and quantization technology, and entropy encoding technology.
  • the inter-prediction technique is a technique of predicting a value of a pixel included in a current picture using pictures before and/or after the current picture.
  • Intra-prediction technology is a technology of predicting a value of a pixel included in a current picture by using information about pixels in the current picture.
  • the transformation and quantization technique is a technique for compressing the energy of the residual image.
  • the entropy coding technique assigns short codes to values with a high frequency of occurrence and assigns long codes to values with a low frequency of occurrence.
  • data for video can be effectively compressed, transmitted, and stored.
  • An embodiment may provide a method, apparatus, and recording medium for providing learning in an in-loop filter and predictive deep neural network.
  • selecting a deep neural network generating coded deep neural network information by encoding deep neural network information used to construct the deep neural network; and generating a bitstream including the encoded deep neural network information.
  • the image encoding method may further include performing learning in the deep neural network.
  • the step of performing learning in the deep neural network may include: performing learning in a teacher network; conducting first learning in the student network; and performing second learning in the student network.
  • Learning in the student network may be performed using knowledge distillation using the teacher network.
  • Learning in an in-loop filter based on the deep neural network may be performed.
  • An input of the in-loop filter may be an image in which compression degradation occurs.
  • An output of the in-loop filter may be an image from which compression degradation has been removed.
  • Learning in the in-loop filter may be performed to minimize an error between the output of the in-loop filter and the original image.
  • the deep neural network may be a predictive deep neural network that generates a predicted image similar to an original image through prediction.
  • receiving a bitstream including coded deep neural network information generating deep neural network information used to construct a deep neural network by decoding the encoded deep neural network information; and selecting the deep neural network.
  • the image decoding method may further include performing learning in the deep neural network.
  • the step of performing learning in the deep neural network may include: performing learning in a teacher network; conducting first learning in the student network; and performing second learning in the student network.
  • Learning in the student network may be performed using knowledge distillation using the teacher network.
  • Learning in an in-loop filter based on the deep neural network may be performed.
  • An input of the in-loop filter may be an image in which compression degradation occurs.
  • An output of the in-loop filter may be an image from which compression degradation has been removed.
  • Learning in the in-loop filter may be performed to minimize an error between the output of the in-loop filter and the original image.
  • the deep neural network may be a predictive deep neural network that generates a predicted image similar to an original image through prediction.
  • bitstream in another aspect, includes encoded deep neural network information, and decoding the encoded deep neural network information is performed to A computer readable recording medium in which deep neural network information used to construct a deep neural network is generated and the deep neural network is selected is provided.
  • Learning in the deep neural network may be performed.
  • a first learning in the student network may be performed.
  • Second learning in the student network may be performed.
  • Learning in the student network may be performed using knowledge distillation using the teacher network.
  • Learning in an in-loop filter based on the deep neural network may be performed.
  • An input of the in-loop filter may be an image in which compression degradation occurs.
  • An output of the in-loop filter may be an image from which compression degradation has been removed.
  • Learning in the in-loop filter may be performed to minimize an error between the output of the in-loop filter and the original image.
  • a method, apparatus, and recording medium for providing learning in an in-loop filter and predictive deep neural network are provided.
  • FIG. 1 is a block diagram showing a configuration according to an embodiment of an encoding device to which the present invention is applied.
  • FIG. 2 is a block diagram showing a configuration according to an embodiment of a decoding device to which the present invention is applied.
  • FIG. 3 is a diagram schematically illustrating a division structure of an image when encoding and decoding an image.
  • FIG. 4 is a diagram illustrating a form of a prediction unit that a coding unit may include.
  • FIG. 5 is a diagram illustrating a form of a transform unit that may be included in a coding unit.
  • FIG. 6 shows division of a block according to an example.
  • FIG. 7 is a diagram for explaining an embodiment of an intra prediction process.
  • FIG. 8 is a diagram for explaining reference samples used in an intra prediction process.
  • FIG. 9 is a diagram for explaining an embodiment of an inter prediction process.
  • 11 illustrates an order of adding motion information of spatial candidates to a merge list according to an example.
  • FIG 13 illustrates diagonal scanning according to an example.
  • 16 is a structural diagram of an encoding device according to an embodiment.
  • 17 is a structural diagram of a decryption device according to an embodiment.
  • FIG. 18 is a flowchart of an encoding method according to an embodiment.
  • 19 is a flowchart of a decoding method according to an embodiment.
  • 20 is a flowchart of learning in a deep neural network according to an example.
  • 21 illustrates learning in a teacher network according to an example.
  • 22 shows learnings in an intra-prediction deep neural network and an inter-prediction deep neural network according to an embodiment.
  • FIG. 23 illustrates a first learning in a student network according to one embodiment.
  • FIG. 24 illustrates a first learning in a student network according to one embodiment.
  • 25 illustrates second learning in a student network according to one embodiment.
  • first and second may be used to describe various components, but the components should not be limited by the terms. These terms are only used for the purpose of distinguishing one component from another. For example, a first element may be termed a second element, and similarly, a second element may be termed a first element, without departing from the scope of the present invention.
  • the term "and/or" may include any combination of a plurality of related listed items or any of a plurality of related listed items.
  • each component is listed and included as each component for convenience of description, and at least two of each component are combined to form one component, or one component is divided into a plurality of components to perform functions. It can be performed, and integrated embodiments and separate embodiments of each of these components are also included in the scope of the present invention as long as they do not depart from the essence of the present invention.
  • the term “at least one” may mean one or more numbers such as 1, 2, 3, and 4. In embodiments, the term “a plurality of” may mean one of two or more numbers, such as 2, 3, and 4.
  • Some of the components of the embodiments are not essential components that perform essential functions in the present invention, but may be optional components for improving performance.
  • Embodiments may be implemented by including only essential components in implementing the essence of the embodiments, excluding components used for performance improvement.
  • a structure including only essential components excluding optional components used for performance improvement is also included in the scope of the embodiments.
  • an image may mean one picture constituting a video, and may also indicate the video itself.
  • "encoding and/or decoding an image” may mean “encoding and/or decoding a video”, and may mean “encoding and/or decoding one of images constituting a video”.
  • video and “motion picture(s)” may be used interchangeably and may be used interchangeably.
  • the target image may be an encoding target image that is an encoding target and/or a decoding target image that is a decoding target.
  • the target image may be an input image input to an encoding device or an input image input to a decoding device.
  • the target image may be a current image that is a target of current encoding and/or decoding.
  • the terms “target image” and “current image” may be used interchangeably and may be used interchangeably.
  • image image
  • picture image
  • frame image
  • screen image
  • the target block may be an encoding target block that is an encoding target and/or a decoding target block that is a decoding target.
  • the target block may be a current block that is a target of current encoding and/or decoding.
  • the terms “target block” and “current block” may be used interchangeably and may be used interchangeably.
  • the current block may refer to a coding target block that is an encoding target during encoding and/or a decoding target block that is a decoding target during decoding.
  • the current block may be at least one of a coding block, a prediction block, a residual block, and a transform block.
  • block and “unit” may be used interchangeably and may be used interchangeably. Or “block” may represent a specific unit.
  • region and “segment” may be used interchangeably.
  • each of the specified information, data, flag, index and element, attribute, etc. may have a value.
  • the value "0" of information, data, flags, indexes, elements, and attributes may represent false, logical false, or a first predefined value. That is to say, the value "0", false, logic false and the first predefined value may be used interchangeably.
  • the value "1" of information, data, flags, indexes, elements, and attributes may represent true, logical true, or a second predefined value. In other words, the value "1", true, logically true, and the second predefined value may be used interchangeably.
  • i When a variable such as i or j is used to indicate a row, column, or index, the value of i may be an integer greater than or equal to 0, or may be an integer greater than or equal to 1. That is to say, row, column, index, etc. may be counted from 0, in embodiments, may be counted from 1.
  • the term “one or more” or the term “at least one” may mean the term “plurality”. “One or more” or “at least one” may be used interchangeably with “plural”.
  • Encoder An encoder may mean a device that performs encoding. In other words, an encoder may mean an encoding device.
  • a decoder may refer to a device that performs decoding.
  • a decryptor may mean a decryption device.
  • a unit may represent a unit of encoding and/or decoding of an image.
  • the terms “unit” and “block” may be used interchangeably and may be used interchangeably.
  • a unit may be an MxN array of samples. M and N may each be a positive integer.
  • a unit may refer to an array of samples in a two-dimensional form.
  • a unit may be a region created by dividing one image.
  • a unit may be a specified area within one image.
  • One image may be divided into a plurality of units.
  • the unit may refer to the divided parts when one image is divided into subdivided parts and encoding or decoding of the divided parts is performed.
  • a predefined process for a unit may be performed according to a unit type.
  • the type of unit is a macro unit, a coding unit (CU), a prediction unit (PU), a residual unit, and a transform unit (TU), etc. can be classified as Or, depending on the function, the unit is a block, macroblock, coding tree unit, coding tree block, coding unit, coding block, or prediction unit. It may mean a prediction unit, a prediction block, a residual unit, a residual block, a transform unit, a transform block, and the like.
  • the target unit may be at least one of a CU, a PU, a residual unit, and a TU to be encoded and/or decoded.
  • a unit may refer to information including a luma component block, a chroma component block corresponding to the luma component block, and a syntax element for each block, in order to be referred to as a block.
  • Units may vary in size and shape. Also, units can have various sizes and shapes. In particular, the shape of the unit may include not only a square but also a two-dimensional geometric figure such as a rectangle, a trapezoid, a triangle, and a pentagon.
  • the unit information may include at least one or more of a unit type, a size of a unit, a depth of a unit, a coding order of a unit, a decoding order of a unit, and the like.
  • the unit type may indicate one of CU, PU, residual unit, and TU.
  • One unit can be further divided into sub-units having a smaller size than the unit.
  • Depth may mean the degree of division of a unit. Also, the depth of a unit may indicate a level at which the unit exists when the unit(s) are expressed as a tree structure.
  • Unit division information may include depth about the depth of the unit. Depth may indicate the number and/or extent to which a unit is divided.
  • the root node has the shallowest depth and the leaf node has the deepest depth.
  • the root node may be the highest node.
  • a leaf node may be the lowest node.
  • One unit may be hierarchically divided into a plurality of sub-units while having depth information based on a tree structure.
  • a unit and a sub-unit generated by division of the unit may correspond to a node and a child node of the node, respectively.
  • Each divided sub-unit may have a depth. Since depth represents the number and/or degree of division of a unit, division information of a sub-unit may include information about the size of the sub-unit.
  • the highest node may correspond to the first non-split unit.
  • the highest node may be referred to as a root node.
  • the highest node may have the smallest depth value. In this case, the highest node may have a depth of level 0.
  • a node with a depth of level 1 may represent a unit created as the original unit is split once.
  • a node with a depth of level 2 may represent a unit created as the original unit is split twice.
  • a node with a depth of level n may represent a unit created as the initial unit is split n times.
  • a leaf node may be the lowest node and may be a node that cannot be further divided.
  • the depth of a leaf node may be the maximum level.
  • the predefined value of the maximum level may be 3.
  • -QT depth may indicate the depth for quad division.
  • BT depth may indicate a depth for binary partitioning.
  • the TT depth may indicate a depth for ternary division.
  • a sample may be a base unit constituting a block.
  • a sample may be represented as values from 0 to 2 Bd -1 according to a bit depth (Bd).
  • a sample can be a pixel or a pixel value.
  • pixel In the following, the terms “pixel”, “pixel” and “sample” may be used in the same meaning and may be used interchangeably.
  • a CTU may consist of one luma component (Y) coding tree block and two chroma component (Cb, Cr) coding tree blocks related to the luma component coding tree block. there is.
  • the CTU may mean including syntax elements for the above blocks and each block of the above blocks.
  • Each coding tree unit is a quad tree (QT), binary tree (BT), and ternary tree (TT) to construct sub units such as a coding unit, a prediction unit, and a transform unit. It can be segmented using one or more segmentation schemes.
  • a quad tree may mean a quarternary tree.
  • each coding tree unit may be split using a MultiType Tree (MTT) using one or more splitting schemes.
  • MTT MultiType Tree
  • - CTU may be used as a term to refer to a pixel block, which is a processing unit in the process of decoding and encoding an image, as in segmentation of an input image.
  • a coding tree block may be used as a term to refer to any one of a Y coding tree block, a Cb coding tree block, and a Cr coding tree block.
  • a neighboring block may mean a block adjacent to a target block.
  • a neighboring block may mean a reconstructed neighboring block.
  • neighboring block and “adjacent block” may be used in the same meaning and may be used interchangeably.
  • a neighboring block may mean a reconstructed neighboring block.
  • a spatial neighbor block may be a block that is spatially adjacent to the target block.
  • Neighboring blocks may include spatial neighboring blocks.
  • the target block and spatial neighboring blocks may be included in the target picture.
  • a spatial neighboring block may mean a block whose boundary meets the target block or a block located within a predetermined distance from the target block.
  • a spatial neighboring block may mean a block adjacent to a vertex of a target block.
  • the block adjacent to the vertex of the target block may be a block vertically adjacent to a neighboring block horizontally adjacent to the target block or a block horizontally adjacent to a neighboring block vertically adjacent to the target block.
  • Temporal neighbor block may be a block that is temporally adjacent to the target block. Neighboring blocks may include temporal neighboring blocks.
  • a temporal neighboring block may include a co-located block (col block).
  • a collocated block may be a block in an already reconstructed co-located picture (col picture).
  • a position of a collocated block in a collocated picture may correspond to a position of a target block in a target picture.
  • the position of the collocated block in the collocated picture may be the same as the position of the target block in the target picture.
  • a collocated picture may be a picture included in a reference picture list.
  • a temporal neighboring block may be a block temporally adjacent to a spatial neighboring block of a target block.
  • the prediction mode may be information indicating a mode used for intra prediction or a mode used for inter prediction.
  • a prediction unit may mean a base unit for prediction such as inter prediction, intra prediction, inter compensation, intra compensation, and motion compensation.
  • One prediction unit may be divided into a plurality of smaller-sized partitions or sub-prediction units.
  • a plurality of partitions may also be a basis unit in performing prediction or compensation.
  • a partition generated by division of a prediction unit may also be a prediction unit.
  • Prediction unit partition A prediction unit partition may mean a form in which a prediction unit is divided.
  • a reconstructed neighboring unit may be a unit that has already been decoded and reconstructed in a neighbor of a target unit.
  • the reconstructed neighbor unit may be a spatial neighbor unit or a temporal neighbor unit to the target unit.
  • the reconstructed spatial neighboring unit may be a unit in the target picture and already reconstructed through encoding and/or decoding.
  • the reconstructed temporal neighbor unit may be a unit in the reference picture and a unit that has already been reconstructed through encoding and/or decoding.
  • a position of the reconstructed temporal neighboring unit within the reference image may be the same as a position within the target unit's target picture or may correspond to a position within the target unit's target picture.
  • the reconstructed temporal neighboring unit may be a neighboring block of a corresponding block in the reference picture.
  • the position of the corresponding block in the reference image may correspond to the position of the target block in the target image.
  • the correspondence of the positions of the blocks may mean that the positions of the blocks are the same, and may mean that one block is included in another block, and one block occupies a specified position of the other block. can mean doing
  • a picture can be divided into one or more sub-pictures.
  • a sub-picture may consist of one or more tile rows and one or more tile columns.
  • a sub-picture may be an area having a square shape or a rectangular (ie, non-square) shape within a picture.
  • a sub-picture may include one or more CTUs. .
  • a sub-picture may be a rectangular area of one or more slices within one picture.
  • One sub-picture may include one or more tiles, one or more bricks, and/or one or more slices.
  • a tile can be a square or rectangular (ie, non-square) area within a picture.
  • a tile may contain one or more CTUs.
  • a tile can be divided into one or more bricks.
  • a brick may mean one or more CTU rows within a tile.
  • Each brick may contain one or more CTU rows.
  • a tile that is not divided into two or more can also mean a brick.
  • a slice can include one or more tiles within a picture. Or, a slice may include one or more bricks within a tile.
  • each sub-picture boundary may always be a slice boundary.
  • each vertical sub-picture boundary may always be a vertical tile boundary.
  • a parameter set may correspond to header information among structures in a bitstream.
  • - Parameter sets include a Video Parameter Set (VPS), a Sequence Parameter Set (SPS), a Picture Parameter Set (PPS), an Adaptation Parameter Set (APS), and a decoding parameter It may include at least one of a set (Decoding Parameter Set; DPS), and the like.
  • VPS Video Parameter Set
  • SPS Sequence Parameter Set
  • PPS Picture Parameter Set
  • APS Adaptation Parameter Set
  • decoding parameter It may include at least one of a set (Decoding Parameter Set; DPS), and the like.
  • Information signaled through a parameter set may be applied to pictures referring to the parameter set.
  • information in the VPS may be applied to pictures referring to the VPS.
  • Information in the SPS may be applied to pictures referring to the SPS.
  • Information in the PPS may be applied to pictures referring to the PPS.
  • a parameter set may refer to an upper parameter set.
  • PPS may refer to SPS.
  • SPS may refer to VPS.
  • the parameter set may include tile group, slice header information, and tile header information.
  • a tile group may refer to a group including a plurality of tiles. Also, the meaning of a tile group may be the same as that of a slice.
  • Rate-distortion optimization The encoding device uses a combination of the size of the coding unit, the prediction mode, the size of the prediction unit, motion information, and the size of the conversion unit to provide high encoding efficiency. Distortion optimization can be used.
  • the rate-distortion optimization method may calculate a rate-distortion cost of each combination in order to select an optimal combination among the above combinations.
  • the rate-distortion cost can be calculated using the formula “D+ ⁇ *R”.
  • a combination that minimizes the rate-distortion cost according to the equation “D+ ⁇ *R” can be selected as an optimal combination in the rate-distortion optimization method.
  • D may be the mean square error of difference values between the original transform coefficients and the reconstructed transform coefficients within the transform unit.
  • R can represent rate.
  • R may represent a bit rate using related context information.
  • R may include not only coding parameter information such as a prediction mode, motion information, and a coded block flag, but also bits generated by encoding transform coefficients.
  • the encoding device may perform processes such as inter prediction, intra prediction, transformation, quantization, entropy encoding, inverse quantization, and/or inverse transformation to calculate accurate D and R. These processes can greatly increase complexity in an encoding device.
  • Bitstream may mean a string of bits including coded image information.
  • Parsing may mean determining a value of a syntax element by entropy decoding a bitstream. Alternatively, parsing may mean entropy decoding itself.
  • Symbol may mean at least one of a syntax element, a coding parameter, and a transform coefficient of a coding target unit and/or a decoding target unit. Also, a symbol may mean an object of entropy encoding or a result of entropy decoding.
  • a reference picture may refer to an image that a unit refers to for inter prediction or motion compensation.
  • the reference picture may be an image including a reference unit referred to by a target unit for inter prediction or motion compensation.
  • reference picture and “reference image” may be used interchangeably and may be used interchangeably.
  • the reference picture list may be a list including one or more reference pictures used for inter prediction or motion compensation.
  • the types of reference picture lists are List Combined (LC), List 0 (L0), List 1 (L1), List 2 (List 2; L2), and List 3 (List 3; L3). ), etc. may be present.
  • One or more reference picture lists may be used for inter prediction.
  • the inter prediction indicator may indicate the direction of inter prediction for a target unit. Inter prediction can be one of uni-prediction and bi-prediction, etc. Alternatively, the inter prediction indicator may indicate the number of reference pictures used when generating a prediction unit of a target unit. Alternatively, the inter prediction indicator may indicate the number of prediction blocks used for inter prediction or motion compensation of the target unit.
  • the prediction list utilization flag may indicate whether a prediction unit is generated using at least one reference picture in a specific reference picture list.
  • An inter prediction indicator can be derived using the prediction list utilization flag.
  • the prediction list utilization flag can be derived using the inter prediction indicator. For example, when the prediction list utilization flag indicates a first value of 0, it may indicate that a prediction block is not generated using a reference picture in the reference picture list for the target unit. When the prediction list utilization flag indicates the second value of 1, it may indicate that a prediction unit is generated using the reference picture list for the target unit.
  • the reference picture index may be an index indicating a specific reference picture in a reference picture list.
  • POC Picture Order Count
  • Motion Vector A motion vector may be a two-dimensional vector used in inter prediction or motion compensation.
  • a motion vector may mean an offset between a target image and a reference image.
  • MV can be expressed as (mv x , mv y ).
  • mv x may represent a horizontal component
  • mv y may represent a vertical component.
  • the search range may be a two-dimensional area where MVs are searched during inter prediction.
  • the size of the search area may be MxN.
  • M and N may each be a positive integer.
  • Motion vector candidate may mean a block as a prediction candidate or a motion vector of a block as a prediction candidate when predicting a motion vector.
  • a motion vector candidate may be included in a motion vector candidate list.
  • Motion vector candidate list may refer to a list constructed using one or more motion vector candidates.
  • Motion vector candidate index may mean an indicator indicating a motion vector candidate in the motion vector candidate list.
  • the motion vector candidate index may be an index of a motion vector predictor.
  • Motion information includes not only motion vectors, reference picture indices and inter prediction indicators, but also reference picture list information, reference pictures, motion vector candidates, motion vector candidate indices, merge candidates and merge indices, etc. It may mean information including at least one of
  • a merge candidate list may refer to a list constructed using one or more merge candidates.
  • a merge candidate is a spatial merge candidate, a temporal merge candidate, a combined merge candidate, a combined bi-prediction merge candidate, a candidate based on history, a candidate based on an average of two candidates, and zero It may mean a merge candidate and the like.
  • a merge candidate may include an inter prediction indicator, and may include motion information such as a reference picture index for each list, a motion vector, a prediction list utilization flag, and an inter prediction indicator.
  • a merge index may be an indicator pointing to a merge candidate in a merge candidate list.
  • the merge index may indicate a reconstructed unit that derives a merge candidate from among reconstructed units spatially adjacent to the target unit and reconstructed units temporally adjacent to the target unit.
  • the merge index may indicate at least one piece of motion information of a merge candidate.
  • a transform unit may be a basic unit in residual signal encoding and/or residual signal decoding, such as transform, inverse transform, quantization, inverse quantization, transform coefficient encoding, and transform coefficient decoding.
  • One transform unit may be divided into a plurality of sub-transform units having smaller sizes.
  • the transformation may include one or more of a first-order transformation and a second-order transformation
  • the inverse transformation may include one or more of a first-order inverse transformation and a second-order inverse transformation.
  • Scaling may refer to a process of multiplying a transform coefficient level by a factor.
  • Scaling may be referred to as dequantization.
  • a quantization parameter may mean a value used when generating a transform coefficient level for a transform coefficient in quantization.
  • the quantization parameter may refer to a value used when generating a transform coefficient by scaling a transform coefficient level in inverse quantization.
  • the quantization parameter may be a value mapped to a quantization step size.
  • the delta quantization parameter may refer to a difference value between a predicted quantization parameter and a quantization parameter of a target unit.
  • a scan may refer to a method of arranging the order of coefficients within a unit, block or matrix. For example, arranging a 2D array into a 1D array may be referred to as a scan. Alternatively, arranging a one-dimensional array into a two-dimensional array may also be referred to as scan or inverse scan.
  • a transform coefficient may be a coefficient value generated by performing transformation in an encoding device.
  • the transform coefficient may be a coefficient value generated by performing at least one of entropy decoding and inverse quantization in the decoding apparatus.
  • a quantized level generated by applying quantization to a transform coefficient or a residual signal or a quantized transform coefficient level may also be included in the meaning of a transform coefficient.
  • a quantized level may refer to a value generated by performing quantization on a transform coefficient or a residual signal in an encoding device.
  • the quantized level may mean a value to be subjected to inverse quantization when the decoding apparatus performs inverse quantization.
  • a quantized transform coefficient level which is a result of transform and quantization, may also be included in the meaning of the quantized level.
  • Non-zero transform coefficient may mean a transform coefficient having a non-zero value or a transform coefficient level having a non-zero value.
  • the non-zero transform coefficient may refer to a transform coefficient whose value is not 0 or a transform coefficient level whose value is not 0.
  • a quantization matrix may mean a matrix used in a quantization process or an inverse quantization process to improve subjective or objective picture quality of an image.
  • a quantization matrix may also be referred to as a scaling list.
  • Quantization matrix coefficient A quantization matrix coefficient may mean each element in a quantization matrix. Quantization matrix coefficients may also be referred to as matrix coefficients.
  • the default matrix may be a quantization matrix predefined in an encoding device and a decoding device.
  • Non-default matrix may be a quantization matrix that is not predefined in the encoding device and the decoding device.
  • the non-default matrix may refer to a quantization matrix signaled from an encoding device to a decoding device by a user.
  • MPM may indicate an intra prediction mode that is highly likely to be used for intra prediction of a target block.
  • the encoding device and the decoding device may determine one or more MPMs based on a coding parameter related to the target block and an attribute of an object related to the target block.
  • the encoding device and the decoding device may determine one or more MPMs based on the intra prediction mode of the reference block.
  • Reference blocks may be plural.
  • the plurality of reference blocks may include a spatial neighboring block adjacent to the left side of the target block and a spatial neighboring block adjacent to the top of the target block. In other words, one or more different MPMs may be determined depending on which intra prediction modes are used for reference blocks.
  • One or more MPMs may be determined in the same way in the encoding device and the decoding device.
  • the encoding device and the decoding device may share an MPM list including one or more identical MPMs.
  • An MPM list can be a list containing one or more MPMs. The number of one or more MPMs in the MPM list may be predefined.
  • the MPM indicator may indicate an MPM used for intra prediction of a target block among one or more MPMs in the MPM list.
  • the MPM indicator may be an index to an MPM list.
  • the MPM list is determined in the same way in the encoding device and the decoding device, the MPM list itself may not need to be transmitted from the encoding device to the decoding device.
  • the MPM indicator may be signaled from the encoding device to the decoding device. As the MPM indicator is signaled, the decoding apparatus may determine an MPM to be used for intra prediction of the target block among MPMs in the MPM list.
  • the MPM use indicator may indicate whether an MPM use mode is to be used for prediction of a target block.
  • the MPM use mode may be a mode for determining an MPM to be used for intra prediction of a target block by using an MPM list.
  • the MPM use indicator may be signaled from the encoding device to the decoding device.
  • Signaling may indicate that information is transmitted from an encoding device to a decoding device.
  • signaling may mean that an encoding device includes information in a bitstream or a recording medium.
  • Information signaled by the encoding device may be used by the decoding device.
  • the encoding device may generate encoded information by performing encoding on signaled information.
  • Encoded information may be transmitted from an encoding device to a decoding device.
  • the decoding apparatus may obtain information by decoding the transmitted encoded information.
  • encoding may be entropy encoding
  • decoding may be entropy decoding.
  • Selective signaling of information may mean that an encoding device selectively includes information in a bitstream or a recording medium (according to specific conditions). Selective signaling of information may mean that a decoding apparatus selectively extracts information from a bitstream (according to a specific condition).
  • Omission of signaling Signaling of information may be omitted. Omission of signaling of information about information may mean that an encoding device does not include information in a bitstream or a recording medium (according to a specific condition). Omission of signaling for information may mean that the decoding apparatus does not extract information from the bitstream (according to a specific condition).
  • Variables, coding parameters and constants, etc. can have values that can be computed.
  • a statistical value may be a value generated by an operation on the values of these specified objects.
  • the statistical value is an average value, a weighted average value, a weighted sum, a minimum value, a maximum value, and a mode for values such as a specified variable, a specified coding parameter, and a specified constant. It can be one or more of a value, a median value, and an interpolated value.
  • FIG. 1 is a block diagram showing a configuration according to an embodiment of an encoding device to which the present invention is applied.
  • the encoding device 100 may be an encoder, a video encoding device, or an image encoding device.
  • a video may include one or more images.
  • the encoding apparatus 100 may sequentially encode one or more images of a video.
  • an encoding apparatus 100 includes an inter prediction unit 110, an intra prediction unit 120, a switch 115, a subtractor 125, a transform unit 130, a quantization unit 140, and entropy encoding. It may include a unit 150, an inverse quantization unit 160, an inverse transform unit 170, an adder 175, a filter unit 180, and a reference picture buffer 190.
  • the encoding apparatus 100 may perform encoding on a target image using an intra mode and/or an inter mode.
  • the prediction mode for the target block may be one of an intra mode and an inter mode.
  • intra mode intra prediction mode
  • in-picture mode in-picture prediction mode
  • inter mode inter prediction mode
  • inter-screen mode inter-prediction mode
  • video may refer to only a part of an image or may refer to a block.
  • processing of “image” may indicate sequential processing of a plurality of blocks.
  • the encoding device 100 may generate a bitstream including encoded information through encoding of a target image, and output and store the generated bitstream.
  • the generated bitstream may be stored in a computer readable recording medium and may be streamed through a wired and/or wireless transmission medium.
  • the switch 115 can be switched to intra.
  • the switch 115 can be switched to inter.
  • the encoding apparatus 100 may generate a prediction block for the target block. Also, after the prediction block is generated, the encoding apparatus 100 may encode a residual block for the target block by using the target block and the residual of the prediction block.
  • the intra predictor 120 may use a pixel of a block already encoded and/or decoded, which is adjacent to the target block, as a reference sample.
  • the intra predictor 120 may perform spatial prediction on the target block using the reference sample, and generate prediction samples for the target block through the spatial prediction.
  • a prediction sample may mean a sample within a prediction block.
  • the inter prediction unit 110 may include a motion estimation unit and a motion compensation unit.
  • the motion prediction unit may search for a region that best matches the target block from the reference image in the motion prediction process, and derive motion vectors for the target block and the searched region using the searched region. can do. At this time, the motion prediction unit may use a search area as a search target area.
  • the reference picture may be stored in the reference picture buffer 190, and when encoding and/or decoding of the reference picture is processed, the encoded and/or decoded reference picture may be stored in the reference picture buffer 190.
  • the reference picture buffer 190 may be a decoded picture buffer (DPB).
  • DPB decoded picture buffer
  • the motion compensator may generate a prediction block for the target block by performing motion compensation using a motion vector.
  • the motion vector may be a 2D vector used for inter prediction.
  • the motion vector may indicate an offset between the target image and the reference image.
  • the motion estimation unit and the motion compensation unit may generate a prediction block by applying an interpolation filter to a partial region in a reference image when a motion vector has a non-integer value.
  • methods of motion prediction and motion compensation of PUs included in the CU based on the CU include skip mode, merge mode, and advanced motion vector prediction (Advanced Motion Vector Prediction (AMVP) mode and current picture reference mode may be determined, and inter prediction or motion compensation may be performed according to each mode.
  • AMVP Advanced Motion Vector Prediction
  • the subtractor 125 may generate a residual block that is a difference between the target block and the prediction block.
  • a residual block may also be referred to as a residual signal.
  • the residual signal may mean a difference between the original signal and the prediction signal.
  • the residual signal may be a signal generated by transforming, quantizing, or transforming and quantizing the difference between the original signal and the predicted signal.
  • a residual block may be a residual signal for a block unit.
  • the transform unit 130 may generate transform coefficients by performing transform on the residual block, and may output the generated transform coefficients.
  • the transform coefficient may be a coefficient value generated by performing transform on the residual block.
  • the conversion unit 130 may use one of a plurality of predefined conversion methods in performing the conversion.
  • a plurality of predefined transform methods may include discrete cosine transform (DCT), discrete sine transform (DST), and Karhunen-Loeve transform (KLT) based transform. there is.
  • DCT discrete cosine transform
  • DST discrete sine transform
  • KLT Karhunen-Loeve transform
  • a transform method used for transforming the residual block may be determined according to at least one of coding parameters for the target block and/or neighboring blocks. For example, the transform method may be determined based on at least one of an inter prediction mode for a PU, an intra prediction mode for a PU, a TU size, and a TU shape. Alternatively, transformation information indicating a transformation method may be signaled from the encoding apparatus 100 to the decoding apparatus 200.
  • the transform unit 130 may skip transforming the residual block.
  • a quantized transform coefficient level or a quantized level may be generated by applying quantization to the transform coefficients.
  • a quantized transform coefficient level and a quantized level may also be referred to as a transform coefficient.
  • the quantization unit 140 may generate a quantized transform coefficient level (ie, a quantized level or a quantized coefficient) by quantizing a transform coefficient according to a quantization parameter.
  • the quantization unit 140 may output the generated quantized transform coefficient level.
  • the quantization unit 140 may quantize the transform coefficient using a quantization matrix.
  • the entropy encoding unit 150 may generate a bitstream by performing entropy encoding according to a probability distribution based on values calculated by the quantization unit 140 and/or coding parameter values calculated in the encoding process. .
  • the entropy encoding unit 150 may output the generated bitstream.
  • the entropy encoding unit 150 may perform entropy encoding on information about pixels of an image and information for decoding an image.
  • information for decoding an image may include a syntax element and the like.
  • entropy encoding When entropy encoding is applied, a small number of bits may be allocated to a symbol having a high probability of occurrence, and a large number of bits may be allocated to a symbol having a low probability of occurrence. As symbols are represented through such allocation, the size of bitstrings for symbols that are encoding targets can be reduced. Therefore, compression performance of image encoding can be improved through entropy encoding.
  • the entropy encoding unit 150 uses exponential golomb, context-adaptive variable length coding (CAVLC), and context-adaptive binary arithmetic coding (Context-Adaptive Binary Coding) for entropy encoding.
  • a coding method such as Arithmetic Coding (CABAC) may be used.
  • CABAC Arithmetic Coding
  • the entropy encoding unit 150 may perform entropy encoding using a variable length coding/code (VLC) table.
  • VLC variable length coding/code
  • the entropy encoding unit 150 may derive a binarization method for a target symbol.
  • the entropy encoding unit 150 may derive a probability model of the target symbol/bin.
  • the entropy encoding unit 150 may perform arithmetic encoding using the derived binarization method, probability model, and context model.
  • the entropy encoding unit 150 may change coefficients in the form of a 2-dimensional block into a form of a 1-dimensional vector through a transform coefficient scanning method in order to encode the quantized transform coefficient level.
  • a coding parameter may be information required for encoding and/or decoding.
  • the coding parameter may include information that is encoded in the encoding device 100 and transmitted from the encoding device 100 to the decoding device, and may include information that can be derived in an encoding or decoding process. For example, as information transmitted to the decoding device, there is a syntax element.
  • a coding parameter may include information derived from an encoding process or a decoding process, as well as information (or flags and indexes, etc.) encoded in an encoding device and signaled from an encoding device to a decoding device, such as a syntax element. there is. Also, the coding parameter may include information required for encoding or decoding an image.
  • Information indicating whether or not to use information indicating whether to use additional (secondary) transform, primary transform selection information (or primary transform index), secondary transform selection information (or secondary transform index), residual Information indicating the presence or absence of a signal, a coded block pattern, a coded block flag, a quantization parameter, a residual quantization parameter, a quantization matrix, information about an intra-loop filter, and an intra-loop filter Information indicating whether to apply, coefficients of intra-loop filter, filter tap of intra-loop, shape/form of intra-loop filter, information indicating whether to apply deblocking filter, information indicating whether to apply deblocking filter, coefficient, filter tap of deblocking filter, strength of deblocking filter, shape/shape of deblocking filter, information indicating whether to apply adaptive sample offset, adaptive sample offset value, adaptive sample offset category, adaptive sample Offset type, information indicating whether the adaptive in-loop filter is applied, coefficients of the adaptive in-loop filter, filter tap of the adaptive in-loop filter, shape/shape
  • information related to the aforementioned coding parameters may also be included in the coding parameters.
  • Information used to calculate and/or derive the aforementioned coding parameters may also be included in the coding parameters.
  • Information calculated or derived using the aforementioned coding parameters may also be included in the coding parameters.
  • the primary transform selection information may indicate a primary transform applied to the target block.
  • the secondary transform selection information may indicate a secondary transform applied to the target block.
  • the residual signal may represent a difference between the original signal and the predicted signal.
  • the residual signal may be a signal generated by transforming a difference between the original signal and the predicted signal.
  • the residual signal may be a signal generated by transforming and quantizing a difference between the original signal and the predicted signal.
  • a residual block may be a residual signal for a block.
  • signaling information may mean that the encoding device 100 includes entropy-encoded information generated by performing entropy encoding on a flag or index in a bitstream, , in the decoding apparatus 200, may mean obtaining information by performing entropy decoding on entropy-encoded information extracted from a bitstream.
  • the information may include flags and indexes.
  • a signal may refer to signaled information.
  • information on images and blocks may be referred to as signals.
  • the terms “information” and “signal” may be used interchangeably and may be used interchangeably.
  • the specific signal may be a signal representing a specific block.
  • An original signal may be a signal representing a target block.
  • a prediction signal may be a signal representing a prediction block.
  • a residual signal may be a signal representing a residual block.
  • a bitstream may include information according to a specified syntax.
  • the encoding device 100 may generate a bitstream including information according to a specified syntax.
  • the encoding device 200 may obtain information from a bitstream according to a specified syntax.
  • the encoded target image may be used as a reference image for other image(s) to be processed later. Accordingly, the encoding apparatus 100 may reconstruct or decode the encoded target image and store the reconstructed or decoded image in the reference picture buffer 190 as a reference image. Inverse quantization and inverse transformation may be performed on the encoded target image for decoding.
  • the quantized level may be inversely quantized in the inverse quantization unit 160 and inversely transformed in the inverse transformation unit 170 .
  • the inverse quantization unit 160 may generate inverse quantized coefficients by performing inverse quantization on the quantized level.
  • the inverse transform unit 170 may generate inverse quantized and inverse transformed coefficients by performing an inverse transform on the inverse quantized coefficients.
  • the inverse quantized and inverse transformed coefficients may be combined with the prediction block through the adder 175.
  • a reconstructed block may be generated by adding the inverse quantized and inverse transformed coefficients and the prediction block.
  • the inverse quantized and/or inverse transformed coefficient may mean a coefficient on which at least one of dequantization and inverse-transformation has been performed, and may mean a reconstructed residual block.
  • the reconstructed block may mean a recovered block or a decoded block.
  • the reconstructed block may pass through the filter unit 180 .
  • the filter unit 180 may include at least one of a deblocking filter, a sample adaptive offset (SAO), an adaptive loop filter (ALF), and a non-local filter (NLF). One or more may be applied to a reconstructed sample, reconstructed block or reconstructed picture.
  • the filter unit 180 may also be referred to as an in-loop filter.
  • the deblocking filter may remove block distortion generated at a boundary between blocks in a reconstructed picture.
  • it may be determined whether to apply the deblocking filter to the target block based on pixel(s) included in several columns or rows included in the block.
  • the applied filter may vary according to the required strength of deblocking filtering. In other words, among different filters, a filter determined according to the strength of deblocking filtering may be applied to the target block.
  • a deblocking filter is applied to the target block, a long-tap filter, a strong filter, a weak filter, and a Gaussian filter are selected according to the strength of the deblocking filtering required. ) may be applied to the target block.
  • horizontal filtering and vertical filtering may be processed in parallel.
  • the SAO may add an appropriate offset to a pixel value of a pixel to compensate for a coding error.
  • the SAO may perform correction using an offset on a difference between an original image and an image to which deblocking is applied in units of pixels for an image to which deblocking is applied.
  • offset correction on an image a method of dividing pixels included in the image into a certain number of areas, determining an area to be offset from among the divided areas, and applying the offset to the determined area will be used.
  • a method of applying an offset in consideration of edge information of each pixel of the image may be used.
  • ALF may perform filtering based on a value obtained by comparing the reconstructed image and the original image. After dividing the pixels included in the image into predetermined groups, a filter to be applied to each divided group may be determined, and filtering may be performed differentially for each group. Information related to whether to apply the adaptive loop filter may be signaled for each CU. This information may be signaled for the luma signal. The shape and filter coefficients of ALF to be applied to each block may be different for each block. Alternatively, a fixed type of ALF may be applied to the block regardless of the characteristics of the block.
  • the non-local filter may perform filtering based on reconstructed blocks similar to the target block.
  • a region similar to the target block may be selected from the reconstructed image, and filtering of the target block may be performed using statistical properties of the selected similar region.
  • Information related to whether to apply the non-local filter may be signaled to the CU. Also, shapes and filter coefficients of non-local filters to be applied to blocks may be different according to blocks.
  • a reconstructed block or a reconstructed image that has passed through the filter unit 180 may be stored in the reference picture buffer 190 as a reference picture.
  • a reconstructed block that has passed through the filter unit 180 may be part of a reference picture.
  • the reference picture may be a reconstructed picture composed of reconstructed blocks that have passed through the filter unit 180.
  • the stored reference picture can then be used for inter prediction or motion compensation.
  • FIG. 2 is a block diagram showing a configuration according to an embodiment of a decoding device to which the present invention is applied.
  • the decoding device 200 may be a decoder, a video decoding device, or an image decoding device.
  • the decoding apparatus 200 includes an entropy decoding unit 210, an inverse quantization unit 220, an inverse transform unit 230, an intra prediction unit 240, an inter prediction unit 250, and a switch 245. , an adder 255, a filter unit 260, and a reference picture buffer 270.
  • the decoding device 200 may receive the bitstream output from the encoding device 100.
  • the decoding apparatus 200 may receive a bitstream stored in a computer readable recording medium or may receive a bitstream streamed through a wired/wireless transmission medium.
  • the decoding apparatus 200 may perform intra mode and/or inter mode decoding on a bitstream. Also, the decoding apparatus 200 may generate a reconstructed image or a decoded image through decoding, and output the generated reconstructed image or decoded image.
  • conversion to an intra mode or an inter mode according to a prediction mode used for decoding may be performed by the switch 245 .
  • the prediction mode used for decoding is an intra mode
  • the switch 245 may be switched to intra mode.
  • the prediction mode used for decoding is the inter mode
  • the switch 245 may be switched to inter mode.
  • the decoding apparatus 200 may obtain a reconstructed residual block by decoding the input bitstream and generate a prediction block. When the reconstructed residual block and the prediction block are obtained, the decoding apparatus 200 may generate a reconstructed block to be decoded by summing the reconstructed residual block and the prediction block.
  • the entropy decoding unit 210 may generate symbols by performing entropy decoding on the bitstream based on the probability distribution of the bitstream.
  • the generated symbols may include symbols in the form of quantized transform coefficient levels (ie, quantized levels or quantized coefficients).
  • the entropy decoding method may be similar to the above-described entropy encoding method.
  • the entropy decoding method may be a reverse process of the above-described entropy encoding method.
  • the entropy decoding unit 210 may change a coefficient in the form of a 1-dimensional vector into a form of a 2-dimensional block through a transform coefficient scanning method in order to decode the quantized transform coefficient level.
  • the coefficients may be changed into a 2D block form by scanning the coefficients of a block using an upper-right diagonal scan.
  • which scan to be used among the upper-right diagonal scan, vertical scan, and horizontal scan may be determined according to the size of the block and/or the intra prediction mode.
  • the quantized coefficient may be inversely quantized in the inverse quantization unit 220 .
  • the inverse quantization unit 220 may generate inverse quantized coefficients by performing inverse quantization on the quantized coefficients.
  • the inverse quantized coefficient may be inversely transformed in the inverse transform unit 230 .
  • the inverse transform unit 230 may generate a reconstructed residual block by performing an inverse transform on the inverse quantized coefficient.
  • a reconstructed residual block may be generated.
  • the inverse quantization unit 220 may apply a quantization matrix to the quantized coefficients in generating the reconstructed residual block.
  • the intra predictor 240 may generate a prediction block by performing spatial prediction on the target block using pixel values of previously decoded blocks adjacent to the target block.
  • the inter prediction unit 250 may include a motion compensation unit. Alternatively, the inter prediction unit 250 may be referred to as a motion compensation unit.
  • the motion compensator may generate a prediction block by performing motion compensation on the target block using a motion vector and a reference image stored in the reference picture buffer 270 .
  • the motion compensator may apply an interpolation filter to a partial region in the reference image and generate a prediction block using the reference image to which the interpolation filter is applied.
  • the motion compensation unit may determine which mode among skip mode, merge mode, AMVP mode, and current picture reference mode is a motion compensation method used for a PU included in a CU based on the CU to perform motion compensation, and the determined mode Accordingly, motion compensation may be performed.
  • the reconstructed residual block and the prediction block may be added through an adder 255.
  • Adder 255 may produce a reconstructed block by adding the reconstructed residual block and the prediction block.
  • the reconstructed block may pass through the filter unit 260 .
  • the filter unit 260 may apply at least one of a deblocking filter, an SAO filter, an ALF filter, and a non-local filter to a reconstructed block or a reconstructed image.
  • a reconstructed image may be a picture including a reconstructed block.
  • the filter unit 260 may output a reconstructed image.
  • the reconstructed block and/or the reconstructed image that has passed through the filter unit 260 may be stored as a reference picture in the reference picture buffer 270 .
  • a reconstructed block that has passed through the filter unit 260 may be part of a reference picture.
  • the reference picture may be a reconstructed image composed of reconstructed blocks that have passed through the filter unit 260 .
  • the stored reference picture can then be used for inter prediction and/or motion compensation.
  • FIG. 3 is a diagram schematically illustrating a division structure of an image when encoding and decoding an image.
  • 3 may schematically show an example in which one unit is divided into a plurality of sub-units.
  • a coding unit may be used in encoding and decoding.
  • a unit may be a term that collectively refers to 1) a block including image samples and 2) a syntax element.
  • “division of a unit” may mean “division of a block corresponding to a unit”.
  • a CU may be used as a base unit for image encoding and/or decoding.
  • the CU may be used as a unit to which one selected mode of intra mode and inter mode is applied in video encoding and/or decoding.
  • it may be determined which mode among the intra mode and the inter mode is to be applied to each CU.
  • a CU may be a base unit in encoding and/or decoding of prediction, transform, quantization, inverse transform, inverse quantization, and transform coefficients.
  • an image 300 may be sequentially divided into units of largest coding units (LCUs). For each LCU, a partition structure may be determined.
  • LCU may be used as the same meaning as Coding Tree Unit (CTU).
  • CTU Coding Tree Unit
  • Division of a unit may mean division of a block corresponding to the unit.
  • the block division information may include depth information about the depth of a unit. Depth information may indicate the number and/or degree of division of a unit.
  • One unit may be hierarchically divided into a plurality of sub-units with depth information based on a tree structure.
  • Each divided sub-unit may have depth information.
  • Depth information may be information indicating the size of a CU. Depth information may be stored for each CU.
  • Each CU may have depth information.
  • CUs generated by splitting may have a depth increased by 1 from the depth of the split CU.
  • the division structure may refer to a distribution of CUs in the LCU 310 to efficiently encode an image. This distribution may be determined according to whether one CU is to be divided into a plurality of CUs.
  • the number of divided CUs may be a positive integer greater than or equal to 2, including 2, 4, 8 and 16, and the like.
  • the horizontal and vertical sizes of the CU generated by division may be smaller than the horizontal and vertical sizes of the CU before division, depending on the number of CUs generated by division.
  • the horizontal size and vertical size of the CU generated by division may be half of the horizontal size and half of the vertical size of the CU before division.
  • a divided CU may be recursively divided into a plurality of CUs in the same way.
  • at least one of the horizontal size and the vertical size of the divided CU may be reduced compared to at least one of the horizontal and vertical sizes of the CU before division.
  • the division of the CU may be made recursively to a predefined depth or to a predefined size.
  • the depth of CU may have a value of 0 to 3.
  • the size of the CU may range from 64x64 to 8x8 depending on the depth of the CU.
  • the depth of the LCU 310 may be 0, and the depth of the smallest coding unit (SCU) may be a predefined maximum depth.
  • the LCU may be a CU having the largest coding unit size as described above, and the SCU may be a CU having the smallest coding unit size.
  • the division may start from the LCU 310, and the depth of the CU may increase by 1 whenever the horizontal size and/or the vertical size of the CU are reduced by the division.
  • a CU that is not split may have a size of 2Nx2N.
  • a CU of 2Nx2N size may be divided into 4 CUs having a size of NxN.
  • the size of N can be halved every time the depth increases by 1.
  • an LCU having a depth of 0 may be 64x64 pixels or a 64x64 block. 0 may be the minimum depth.
  • a SCU with a depth of 3 can be 8x8 pixels or an 8x8 block. 3 may be the maximum depth.
  • the CU of the 64x64 block, which is the LCU may be expressed as a depth of 0.
  • a CU of a 32x32 block can be represented with a depth of 1.
  • a CU of a 16x16 block can be represented with a depth of 2.
  • a CU of an 8x8 block, which is an SCU can be expressed as a depth of 3.
  • Information on whether the CU is split may be expressed through split information of the CU.
  • the division information may be 1 bit of information. All CUs except for the SCU may include partition information.
  • a value of partition information of a CU that is not split may be a first value
  • a value of partition information of a CU that is split may be a second value.
  • the split information indicates whether the CU splits, the first value may be 0 and the second value may be 1.
  • each CU of the 4 CUs generated by the division are half of the horizontal size and half of the vertical size of the CU before the division, respectively.
  • the sizes of the 4 CUs may be 16x16.
  • the CU is divided in a quad-tree form. In other words, it can be seen that quad-tree partitioning is applied to the CU.
  • the horizontal size or vertical size of each CU of the two CUs generated by the split is half of the horizontal size or half of the vertical size of the CU before the split, respectively.
  • the sizes of the two divided CUs may be 16x32.
  • the sizes of the two divided CUs may be 32x16.
  • the three divided CUs may be generated by dividing the horizontal or vertical size of the CU before being divided at a ratio of 1:2:1.
  • the 3 divided CUs may have sizes of 16x8, 16x16 and 16x8, respectively, from the top.
  • the three divided CUs may have sizes of 8x32, 16x32, and 8x32 from the left, respectively.
  • Quad-tree partitioning and binary-tree partitioning are applied to the LCU 310 of FIG. 3 .
  • a 64x64 Coding Tree Unit may be divided into a plurality of smaller CUs by a recursive quad-tree structure.
  • One CU can be divided into 4 CUs with equal sizes.
  • CUs can be partitioned recursively, and each CU can have a structure of a quad tree.
  • the CTU 320 of FIG. 3 is an example of a CTU to which quad tree partitioning, binary tree partitioning, and ternary tree partitioning are all applied.
  • At least one of quad tree partitioning, binary tree partitioning, and ternary tree partitioning may be applied to the CTU. Partitions can be applied based on a specified priority order.
  • quad tree partitioning may be preferentially applied to CTUs.
  • a CU that cannot be further divided into a quad tree may correspond to a leaf node of a quad tree.
  • a CU corresponding to a leaf node of a quad tree may be a root node of a binary tree and/or a ternary tree. That is, a CU corresponding to a leaf node of a quad tree may be split in the form of a binary tree or a ternary tree, or may not be split any further.
  • quad tree splitting is not applied again to a CU generated by applying binary tree splitting or ternary tree splitting to a CU corresponding to a leaf node of a quad tree, so that block splitting and/or signaling of block splitting information is performed. can be done effectively.
  • Quad splitting information having a first value may indicate that the CU is split in the form of a quad tree.
  • Quad splitting information having a second value may indicate that the CU is not split in the form of a quad tree.
  • the quad division information may be a flag having a specified length (eg, 1 bit).
  • a CU corresponding to a leaf node of a quad tree may be partitioned in a binary tree form or a ternary tree form.
  • a CU generated by binary tree splitting or ternary tree splitting may be split again into a binary tree shape or a ternary tree shape, or may not be split any further.
  • Partitioning in the case where there is no priority between binary tree partitioning and ternary tree partitioning may be referred to as multi-type tree partitioning. That is, a CU corresponding to a leaf node of a quad tree may become a root node of a multi-type tree.
  • Splitting of a CU corresponding to each node of the multi-type tree may be signaled using at least one of information indicating whether or not the multi-type tree is split, splitting direction information, and splitting tree information. For splitting of a CU corresponding to each node of the multi-type tree, information indicating whether to split or not, splitting direction information, and splitting tree information may be signaled sequentially.
  • information indicating whether a multi-type tree having a first value (eg, “1”) is split may indicate that the corresponding CU is split into a multi-type tree.
  • Information indicating whether the multi-type tree having a second value (eg, “0”) is split may indicate that the corresponding CU is not split into a multi-type tree.
  • the corresponding CU may further include splitting direction information.
  • the splitting direction information may indicate a splitting direction of multi-type tree splitting.
  • the division direction information having a first value (eg, “1”) may indicate that the corresponding CU is divided in the vertical direction.
  • the division direction information having a second value (eg, “0”) may indicate that the corresponding CU is divided in the horizontal direction.
  • the corresponding CU may further include split tree information.
  • Split tree information may indicate a tree used for multi-type tree split.
  • split tree information having a first value may indicate that the corresponding CU is split in the form of a binary tree.
  • Split tree information having a second value (eg, “0”) may indicate that the corresponding CU is split in the form of a ternary tree.
  • each of the aforementioned information indicating whether to split or not, split tree information, and split direction information may be a flag having a specified length (eg, 1 bit).
  • At least one of the above-described quad splitting information, information indicating whether to split a multi-type tree, splitting direction information, and splitting tree information may be entropy-encoded and/or entropy-decoded.
  • information of neighboring CUs adjacent to the target CU may be used.
  • context information for entropy encoding and/or entropy decoding of information of a target CU may be derived based on information of a neighboring CU.
  • the information of the neighboring CU may include at least one of 1) quad split information, 2) information indicating whether the multi-type tree is split, 3) split direction information, and 4) split tree information of the neighboring CU.
  • binary tree splitting may be preferentially performed among binary tree splitting and ternary tree splitting. That is, binary tree splitting is applied first, and a CU corresponding to a leaf node of the binary tree may be set as a root node of the ternary tree. In this case, quad tree splitting and binary tree splitting may not be performed on a CU corresponding to a node of a ternary tree.
  • a CU that is not further split by quad tree splitting, binary tree splitting, and/or ternary tree splitting may become a unit of encoding, prediction, and/or transformation. That is, for prediction and/or transformation, the CU may not be further split. Accordingly, a partitioning structure and partitioning information for dividing a CU into prediction units and/or transform units may not exist in the bitstream.
  • this CU may be recursively split until the size of the CU is less than or equal to the size of the maximum transform block.
  • the CU may be divided into four 32x32 blocks for transform.
  • the CU may be divided into two 32x32 blocks for transform.
  • whether to split the CU may be determined by comparing the horizontal size (and/or vertical size) of the CU and the horizontal size (and/or vertical size) of the largest transform block. For example, when the horizontal size of the CU is greater than the horizontal size of the largest transform block, the CU may be vertically divided into two parts. In addition, when the vertical size of the CU is greater than the vertical size of the largest transform block, the CU may be divided into two horizontally.
  • Information on the maximum size and/or minimum size of a CU and information on the maximum size and/or minimum size of a transform block may be signaled or determined at a higher level for a CU.
  • the upper level may be a sequence level, a picture level, a tile level, a tile group level, and a slice level.
  • the minimum size of a CU may be determined to be 4x4.
  • the maximum size of a transform block may be determined to be 64x64.
  • the minimum size of a transform block may be determined to be 4x4.
  • Information about the minimum size of a CU corresponding to a leaf node of a quad tree ie quad tree minimum size
  • the maximum depth of a path from the root node of a multi-type tree to a leaf node ie multi-type tree maximum size. depth
  • the higher level may be a sequence level, a picture level, a slice level, a tile group level, and a tile level.
  • Information on the minimum size of the quad tree and/or information on the maximum depth of the multi-type tree may be separately signaled or determined for each intra-slice and inter-slice.
  • Difference information about the size of the CTU and the maximum size of the transform block may be signaled or determined at a higher level for the CU.
  • the higher level may be a sequence level, a picture level, a slice level, a tile group level, and a tile level.
  • Information about the maximum size of the CU corresponding to each node of the binary tree ie, the maximum size of the binary tree
  • the maximum size of the CU corresponding to each node of the ternary tree ie, the maximum size of the ternary tree
  • the maximum size of a ternary tree may be 128x128.
  • the minimum size of a CU corresponding to each node of a binary tree say, the minimum size of a binary tree
  • the minimum size of a CU corresponding to each node of a ternary tree say, the minimum size of a ternary tree
  • the maximum size of the binary tree and/or the maximum size of the ternary tree may be signaled or determined at the slice level.
  • the minimum size of the binary tree and/or the minimum size of the ternary tree may be signaled or determined at the slice level.
  • quad splitting information information indicating whether to split a multi-type tree, splitting tree information, and/or splitting direction information may or may not exist in the bitstream.
  • the CU may not include quad splitting information, and the quad splitting information for the CU may be inferred as a second value.
  • the size (horizontal size and vertical size) of a CU corresponding to a node of a multi-type tree is larger than the binary tree maximum size (horizontal size and vertical size) and/or the ternary tree maximum size (horizontal size and vertical size) If larger, the CU may not be partitioned into binary tree form and/or ternary tree form. According to this determination method, information indicating whether to split the multi-type tree may not be signaled and may be inferred as a second value.
  • the size (horizontal size and vertical size) of a CU corresponding to a node of a multi-type tree is equal to the minimum size (horizontal size and vertical size) of a binary tree, or the size of a CU (horizontal size and vertical size) is equal to the size of a ternary tree.
  • the CU may not be split into a binary tree shape and/or a ternary tree shape.
  • information indicating whether to split the multi-type tree may not be signaled and may be inferred as a second value. This is because when the CU is divided into a binary tree shape and/or a ternary tree shape, a CU smaller than the minimum size of the binary tree and/or the minimum size of the ternary tree is generated.
  • binary tree splitting or ternary tree splitting may be limited based on the size of the virtual pipeline data unit (ie, pipeline buffer size). For example, binary tree splitting or ternary tree splitting may be limited when a CU is split into sub-CUs not suitable for a pipeline buffer size by binary tree splitting or ternary tree splitting.
  • the pipeline buffer size may be equal to the size of the largest transform block (eg, 64X64).
  • the pipeline buffer size is 64X64, the following partitions can be limited.
  • N and/or M is 128) CUs
  • the CU may not be split into a binary tree form and/or a ternary tree form. According to this determination method, information indicating whether to split the multi-type tree may not be signaled and may be inferred as a second value.
  • a multi-type tree Information indicating whether to divide may be signaled. Otherwise, the CU may not be partitioned into binary tree form and/or ternary tree form. According to this determination method, information indicating whether to split the multi-type tree may not be signaled and may be inferred as a second value.
  • splitting direction information may be signaled. Otherwise, the division direction information may not be signaled and may be inferred as a value indicating a direction in which the CU may be divided.
  • split tree information may be signaled. Otherwise, the split tree information may not be signaled and may be inferred as a value indicating a tree applicable to split of a CU.
  • FIG. 4 is a diagram illustrating a form of a prediction unit that a coding unit may include.
  • a CU that is not further divided may be divided into one or more prediction units (PUs).
  • PUs prediction units
  • a PU may be a basic unit for prediction.
  • a PU may be coded and decoded in any one of skip mode, inter mode, and intra mode.
  • a PU may be divided into various types according to each mode.
  • the target block described above with reference to FIG. 1 and the target block described above with reference to FIG. 2 may be PUs.
  • a CU may not be divided into PUs.
  • the size of a CU and the size of a PU may be the same.
  • skip mode there may not be a split within a CU.
  • a 2Nx2N mode 410 in which sizes of PU and CU are the same may be supported without division.
  • inter mode 8 divided types can be supported within the CU.
  • 2Nx2N mode 410, 2NxN mode 415, Nx2N mode 420, NxN mode 425, 2NxnU mode 430, 2NxnD mode 435, nLx2N mode 440 and nRx2N Mode 445 may be supported.
  • 2Nx2N mode 410 and NxN mode 425 may be supported.
  • a PU having a size of 2Nx2N may be encoded.
  • a PU having a size of 2Nx2N may mean a PU having the same size as that of a CU.
  • a 2Nx2N PU may have a size of 64x64, 32x32, 16x16 or 8x8.
  • a PU having a size of NxN may be encoded.
  • the size of a PU when the size of a PU is 8x8, 4 divided PUs can be coded.
  • the size of the divided PU may be 4x4.
  • the PU When a PU is coded using an intra mode, the PU may be coded using one intra prediction mode among a plurality of intra prediction modes.
  • HEVC High Efficiency Video Coding
  • HEVC High Efficiency Video Coding
  • Which mode among the 2Nx2N mode 410 and the NxN mode 425 the PU will be encoded may be determined by a rate-distortion cost.
  • the encoding apparatus 100 may perform an encoding operation on a PU having a size of 2Nx2N.
  • the encoding operation may be encoding a PU in each of a plurality of intra prediction modes usable by the encoding apparatus 100 .
  • An optimal intra prediction mode for a PU having a size of 2Nx2N may be derived through an encoding operation.
  • An optimal intra prediction mode may be an intra prediction mode that generates a minimum rate-distortion cost for encoding a 2Nx2N PU among a plurality of intra prediction modes usable by the encoding apparatus 100 .
  • the encoding apparatus 100 may sequentially perform an encoding operation on each PU of NxN-divided PUs.
  • the encoding operation may be encoding a PU in each of a plurality of intra prediction modes usable by the encoding apparatus 100 .
  • An optimal intra prediction mode for an NxN-sized PU may be derived through an encoding operation.
  • An optimal intra-prediction mode may be an intra-prediction mode that generates the lowest rate-distortion cost for encoding an NxN-sized PU among a plurality of intra-prediction modes usable by the encoding apparatus 100.
  • the encoding apparatus 100 may determine which of the 2Nx2N-sized PU and the NxN-sized PUs to encode based on the comparison of the rate-distortion cost of the 2Nx2N-sized PU and the rate-distortion costs of the NxN-sized PUs.
  • One CU may be divided into one or more PUs, and a PU may also be divided into a plurality of PUs.
  • the horizontal size and vertical size of each of the 4 PUs generated by the division are half of the horizontal size and half of the vertical size of the PU before the division, respectively.
  • the sizes of the 4 PUs may be 16x16.
  • the horizontal size or vertical size of each PU of the two PUs generated by the split is half of the horizontal size or half of the vertical size of the PU before splitting, respectively.
  • the sizes of the 2 PUs may be 16x32.
  • the sizes of the two divided PUs may be 32x16.
  • FIG. 5 is a diagram illustrating a form of a transform unit that may be included in a coding unit.
  • a transform unit may be a basic unit used for transformation, quantization, inverse transformation, inverse quantization, entropy encoding, and entropy decoding processes within the CU.
  • a TU may have a square shape or a rectangular shape.
  • the shape of the TU may be determined depending on the size and/or shape of the CU.
  • a CU that is no longer split into CUs may be split into one or more TUs.
  • the division structure of the TU may be a quad-tree structure.
  • one CU 510 may be divided one or more times according to a quad-tree structure.
  • one CU 510 may be composed of TUs of various sizes.
  • one CU When one CU is split two or more times, the CU can be considered to be split recursively. Through splitting, one CU may be composed of TUs having various sizes.
  • one CU may be divided into one or more TUs based on the number of vertical and/or horizontal lines dividing the CU.
  • a CU may be divided into symmetric TUs or asymmetric TUs.
  • information on the size and/or shape of a TU may be signaled from the encoding device 100 to the decoding device 200.
  • the size and/or shape of the TU may be derived from information on the size and/or shape of the CU.
  • a CU may not be divided into TUs.
  • the size of a CU and the size of a TU may be the same.
  • One CU may be divided into one or more TUs, and a TU may also be divided into a plurality of TUs.
  • each TU of the 4 TUs generated by the division are half of the horizontal size and half of the vertical size of the TU before the division, respectively.
  • the sizes of the 4 divided TUs may be 16x16.
  • the TU is divided in the form of a quad-tree.
  • the horizontal size or vertical size of each TU of the two TUs generated by the split is half of the horizontal size or half of the vertical size of the TU before the split, respectively.
  • the sizes of the two split TUs may be 16x32.
  • the sizes of the two divided TUs may be 32x16.
  • CUs may be divided in other ways than shown in FIG. 5 .
  • one CU may be split into 3 CUs.
  • the horizontal size or vertical size of the three divided CUs may be 1/4, 1/2, and 1/4 of the horizontal or vertical size of the CU before division, respectively.
  • the sizes of the 3 divided CUs may be 8x32, 16x32 and 8x32, respectively.
  • the CU is split in the form of a ternary tree.
  • One of the exemplified quad tree-type partitioning, binary tree-type partitioning, and ternary tree-type partitioning may be applied for partitioning of the CU, and a plurality of partitioning schemes may be combined together and used for partitioning of the CU. .
  • a case in which a plurality of partitioning methods are combined and used may be referred to as a composite tree type of partitioning.
  • FIG. 6 shows division of a block according to an example.
  • a target block may be divided as shown in FIG. 6 .
  • the target block may be a CU.
  • an indicator indicating division information may be signaled from the encoding apparatus 100 to the decoding apparatus 200.
  • the division information may be information indicating how the target block is divided.
  • Splitting information includes a split flag (hereinafter referred to as “split_flag”), a quad-binary flag (hereinafter referred to as “QB_flag”), a quad-tree flag (hereinafter referred to as “quadtree_flag”), and a binary tree flag (hereinafter referred to as “binarytree_flag”). ”) and a binary type flag (hereinafter referred to as "Btype_flag”).
  • split_flag may be a flag indicating whether a block is split. For example, a value of 1 of split_flag may indicate that a block is split. A value of 0 of split_flag may indicate that the block is not split.
  • QB_flag may be a flag indicating in which form a block is divided into a quad tree form and a binary tree form. For example, a value of 0 of QB_flag may indicate that a block is divided in a quad tree form. A value of 1 of QB_flag may indicate that a block is divided in the form of a binary tree. Alternatively, a value of 0 of QB_flag may indicate that the block is divided in the form of a binary tree. A value of 1 for QB_flag may indicate that a block is divided in a quad tree form.
  • quadtree_flag may be a flag indicating whether a block is split into a quad tree. For example, a value of 1 of quadtree_flag may indicate that a block is split into a quad tree. A value of 0 of quadtree_flag may indicate that a block is not split in the form of a quad tree.
  • binarytree_flag may be a flag indicating whether a block is split in the form of a binary tree. For example, a value of 1 in binarytree_flag may indicate that a block is split into a binary tree. A value of 0 in binarytree_flag may indicate that a block is not split in the form of a binary tree.
  • Btype_flag may be a flag indicating whether the block is split vertically or horizontally when the block is split in the form of a binary tree. For example, a value of 0 in Btype_flag may indicate that a block is divided in a horizontal direction. A value of 1 of Btype_flag may indicate that a block is divided in a vertical direction. Alternatively, a value of 0 of Btype_flag may indicate that the block is divided in the vertical direction. A value of 1 of Btype_flag may indicate that the block is divided in the horizontal direction.
  • partition information for the block of FIG. 6 can be derived by signaling at least one of quadtree_flag, binarytree_flag, and Btype_flag as shown in Table 1 below.
  • the split information for the block of FIG. 6 can be derived by signaling at least one of split_flag, QB_flag, and Btype_flag as shown in Table 2 below.
  • the partitioning method may be limited to a quad tree, or a binary tree, depending on the size and/or shape of the block.
  • split_flag may be a flag indicating whether to split into a quad tree type or a flag indicating whether to split into a binary tree type.
  • the size and shape of the block may be derived according to the depth information of the block, and the depth information may be signaled from the encoding apparatus 100 to the decoding apparatus 200.
  • the specified range may be defined by at least one of a maximum block size and a minimum block size in which quad-tree partitioning is possible.
  • Information indicating the maximum block size and/or minimum block size in which only quad tree-type division is possible may be signaled from the encoding apparatus 100 to the decoding apparatus 200 through a bitstream. Also, this information may be signaled for at least one unit of video, sequence, picture, parameter, tile group, and slice (or segment).
  • the maximum block size and/or the minimum block size may be fixed sizes predefined in the encoding apparatus 100 and the decoding apparatus 200. For example, when the size of a block is larger than 64x64 and smaller than 256x256, only quad-tree partitioning may be possible. In this case, split_flag may be a flag indicating whether to split into a quad tree.
  • the divided block may be at least one of CU and TU.
  • split_flag may be a flag indicating whether to split into a quad tree.
  • the specified range may be defined by at least one of a maximum block size and a minimum block size in which binary tree or ternary tree partitioning is possible.
  • Information representing a maximum block size and/or a minimum block size in which only binary tree-type division or ternary tree-type division is possible may be signaled from the encoding apparatus 100 to the decoding apparatus 200 through a bitstream.
  • this information may be signaled for at least one unit of sequence, picture, and slice (or segment).
  • the maximum block size and/or the minimum block size may be fixed sizes predefined in the encoding apparatus 100 and the decoding apparatus 200. For example, when the size of a block is greater than 8x8 and less than 16x16, only binary tree partitioning may be possible. In this case, split_flag may be a flag indicating whether to split into a binary tree type or a ternary tree type.
  • quad tree partitioning may be equally applied to binary tree and/or ternary tree partitioning.
  • the division of a block may be limited by the previous division.
  • each divided block may be additionally divided only into the specified tree shape.
  • the specified tree type may be at least one of a binary tree type, a ternary tree type, and a quad tree type.
  • the aforementioned indicator may not be signaled.
  • FIG. 7 is a diagram for explaining an embodiment of an intra prediction process.
  • Arrows extending from the center to the periphery of the graph of FIG. 7 may represent prediction directions of directional intra prediction modes. Also, numbers displayed adjacent to arrows may represent examples of mode values assigned to an intra prediction mode or a prediction direction of the intra prediction mode.
  • the number 0 may indicate a planar mode, which is a non-directional intra prediction mode.
  • Number 1 may indicate a DC mode, which is a non-directional intra prediction mode.
  • Intra coding and/or decoding may be performed using reference samples of neighboring blocks of the target block.
  • a neighboring block may be a reconstructed neighboring block.
  • a reference sample may mean a neighboring sample.
  • intra encoding and/or decoding may be performed using a value of a reference sample included in a reconstructed neighboring block or a coding parameter.
  • the encoding apparatus 100 and/or the decoding apparatus 200 may generate a prediction block by performing intra prediction on the target block based on sample information in the target image.
  • the encoding apparatus 100 and/or the decoding apparatus 200 may generate a prediction block for a target block by performing intra prediction based on sample information in a target image.
  • the encoding device 100 and/or the decoding device 200 may perform directional prediction and/or non-directional prediction based on at least one reconstructed reference sample.
  • a prediction block may refer to a block generated as a result of performing intra prediction.
  • a prediction block may correspond to at least one of CU, PU, and TU.
  • a unit of a prediction block may be the size of at least one of CU, PU, and TU.
  • the prediction block may have a square shape with a size of 2Nx2N or NxN.
  • the size of NxN may include 4x4, 8x8, 16x16, 32x32, and 64x64.
  • the prediction block may be a square block having a size of 2x2, 4x4, 8x8, 16x16, 32x32, or 64x64, or may be a rectangular block having a size of 2x8, 4x8, 2x16, 4x16, or 8x16. there is.
  • Intra prediction may be performed according to an intra prediction mode for a target block.
  • the number of intra prediction modes that the target block may have may be a predefined fixed value or may be a value determined differently according to properties of the prediction block.
  • the properties of the prediction block may include the size of the prediction block and the type of the prediction block.
  • the properties of the prediction block may indicate coding parameters for the prediction block.
  • the number of intra prediction modes may be fixed to N regardless of the size of a prediction block.
  • the number of intra prediction modes may be 3, 5, 9, 17, 34, 35, 36, 65, 67, or 95.
  • the intra prediction mode may be a non-directional mode or a directional mode.
  • the intra prediction mode may include two non-directional modes and 65 directional modes, corresponding to numbers 0 to 66 shown in FIG. 7 .
  • the intra prediction mode may include two non-directional modes and 93 directional modes, corresponding to numbers -14 to 80 shown in FIG. 7 .
  • the two non-directional modes may include DC mode and Planar mode.
  • the directional mode may be a prediction mode having a specific direction or a specific angle.
  • the directional mode may also be referred to as an argular mode.
  • the intra prediction mode may be expressed as at least one of a mode number, a mode value, a mode angle, and a mode direction. That is to say, the terms “(mode) number of an intra prediction mode”, “(mode) value of an intra prediction mode”, “(mode) angle of an intra prediction mode” and “(mode) direction of an intra prediction mode) have the same meaning. and can be used interchangeably.
  • the number of intra prediction modes may be M.
  • M may be 1 or more.
  • the number of intra prediction modes may be M including the number of non-directional modes and the number of directional modes.
  • the number of intra prediction modes may be fixed to M regardless of block sizes and/or color components.
  • the number of intra prediction modes may be fixed to either 35 or 67 regardless of the size of a block.
  • the number of intra prediction modes may be different according to the shape, size, and/or color component type of a block.
  • directional prediction modes indicated by dotted lines may be applied only to prediction for non-square blocks.
  • the number of intra prediction modes may increase. Alternatively, as the block size increases, the number of intra prediction modes may decrease. When the size of a block is 4x4 or 8x8, the number of intra prediction modes may be 67. When the size of a block is 16x16, the number of intra prediction modes may be 35. When the size of a block is 32x32, the number of intra prediction modes may be 19. When the size of a block is 64x64, the number of intra prediction modes may be 7.
  • the number of intra prediction modes may be different depending on whether the color component is a luma signal or a chroma signal.
  • the number of intra prediction modes of the luma component block may be greater than the number of intra prediction modes of the chroma component block.
  • prediction may be performed in a vertical direction based on a pixel value of a reference sample.
  • prediction may be performed in a horizontal direction based on a pixel value of a reference sample.
  • the encoding apparatus 100 and the decoding apparatus 200 may perform intra prediction on a target unit using a reference sample according to an angle corresponding to the directional mode.
  • An intra prediction mode positioned to the right of the vertical mode may be named a vertical-right mode.
  • An intra-prediction mode positioned below the horizontal mode may be named a horizontal-below mode.
  • intra prediction modes having a mode value of one of 51, 52, 53, 54, 55, 56, 57, 58, 59, 60, 61, 62, 63, 64, 65, and 66 are vertical It can be the right modes.
  • Intra prediction modes having a mode value of 2, 3, 4, 5, 6, 7, 8, 9, 10, 11, 12, 13, 14, 15, 16, and 17 may be horizontal bottom modes.
  • the non-directional mode may include a DC mode and a planar mode.
  • the mode value of the DC mode may be 1.
  • the mode value of the planner mode may be 0.
  • the directional mode may include an angular mode. Modes other than the DC mode and the planner mode among the plurality of intra prediction modes may be directional modes.
  • a prediction block may be generated based on an average of pixel values of a plurality of reference samples. For example, a pixel value of a prediction block may be determined based on an average of pixel values of a plurality of reference samples.
  • the number of intra prediction modes and the mode value of each intra prediction mode described above may be merely illustrative.
  • the number of the aforementioned intra prediction modes and the mode value of each intra prediction mode may be differently defined according to embodiments, implementations, and/or needs.
  • a step of checking whether samples included in the reconstructed neighboring block can be used as reference samples of the target block may be performed. If there is a sample that cannot be used as a reference sample of the target block among the samples of the neighboring block, a value generated by copying and/or interpolation using at least one sample value among the samples included in the reconstructed neighboring block This can be replaced with the sample value of a sample that cannot be used as a reference sample. If a value generated by copying and/or interpolation is replaced with a sample value of a sample, the sample may be used as a reference sample of the target block.
  • a filter may be applied to at least one of a reference sample or a prediction sample based on at least one of an intra prediction mode and a size of a target block.
  • the type of filter applied to at least one of the reference sample and the prediction sample may differ according to at least one of an intra prediction mode of the target block, a size of the target block, and a shape of the target block.
  • the type of filter may be classified according to one or more of a length of a filter tap, a value of a filter coefficient, and a filter strength.
  • the length of the filter tap may mean the number of filter taps. Also, the number of filter taps may mean the length of a filter.
  • the intra prediction mode is the planner mode
  • the upper reference sample of the target sample in generating the prediction block of the target block, the upper reference sample of the target sample, the left reference sample of the target sample, and the upper right reference sample of the target block according to the location of the prediction target sample in the prediction block.
  • a sample value of the prediction target sample may be generated using a weight-sum of the lower left reference sample of the target block.
  • an average value of upper reference samples and left reference samples of the target block may be used to generate a prediction block of the target block.
  • filtering using values of reference samples may be performed on specified rows or specified columns within the target block.
  • the specified rows may be one or more top rows adjacent to the reference sample.
  • the specified columns may be one or more left columns adjacent to the reference sample.
  • a prediction block may be generated using an upper reference sample, a left reference sample, an upper right reference sample, and/or a lower left reference sample of the target block.
  • Interpolation in units of real numbers may be performed to generate the prediction samples described above.
  • the intra prediction mode of the target block may be predicted from the intra prediction modes of neighboring blocks of the target block, and information used for prediction may be entropy encoded/decoded.
  • the intra prediction modes of the target block and the neighboring block are the same, it may be signaled that the intra prediction modes of the target block and the neighboring block are the same using a predefined flag.
  • an indicator indicating an intra prediction mode identical to that of a target block among intra prediction modes of a plurality of neighboring blocks may be signaled.
  • information of the intra prediction mode of the target block may be encoded and/or decoded using entropy encoding and/or decoding.
  • FIG. 8 is a diagram for explaining reference samples used in an intra prediction process.
  • the reconstructed reference samples used for intra prediction of the target block include below-left reference samples, left reference samples, above-left corner reference samples, and above reference samples. s and above-right reference samples, etc.
  • the left reference samples may refer to reconstructed reference pixels adjacent to the left side of the target block.
  • Top reference samples may refer to reconstructed reference pixels adjacent to the top of the target block.
  • the top left corner reference sample may refer to a reconstructed reference pixel located at the top left corner of the target block.
  • the lower left reference samples may refer to a reference sample located at the lower end of the left sample line among samples located on the same line as the left sample line composed of the left reference samples.
  • the upper right reference samples may refer to reference samples located to the right of the upper pixel line among samples located on the same line as the upper sample line composed of the upper reference samples.
  • each of the lower left reference samples, the left reference samples, the upper reference samples, and the upper right reference samples may be N.
  • a prediction block may be generated through intra prediction of the target block.
  • Generation of the prediction block may include determining values of pixels of the prediction block.
  • the size of the target block and the prediction block may be the same.
  • Reference samples used for intra prediction of the target block may vary according to the intra prediction mode of the target block.
  • a direction of an intra prediction mode may represent a dependency relationship between reference samples and pixels of a prediction block.
  • a value of a specified reference sample may be used as a value of one or more specified pixels of a prediction block.
  • the specified reference sample and the specified one or more pixels of the prediction block may be samples and pixels specified as a straight line in the direction of the intra prediction mode.
  • the value of the specified reference sample may be copied to a value of a pixel located in a direction opposite to the direction of the intra prediction mode.
  • the value of a pixel of the prediction block may be a value of a reference sample located in the direction of the intra prediction mode based on the position of the pixel.
  • intra prediction mode of the target block when the intra prediction mode of the target block is a vertical mode, upper reference samples may be used for intra prediction.
  • a value of a pixel of a prediction block may be a value of a reference sample positioned vertically above the position of the pixel. Accordingly, the top reference samples that are top adjacent to the target block may be used for intra prediction. Also, values of pixels in one row of the prediction block may be the same as values of upper reference samples.
  • left reference samples may be used for intra prediction.
  • a value of a pixel of a prediction block may be a value of a reference sample located horizontally to the left of the pixel. Accordingly, left reference samples adjacent to the left of the target block may be used for intra prediction. Also, values of pixels of one column of the prediction block may be the same as values of left reference samples.
  • the mode value of the intra prediction mode of the target block is 34
  • at least some of the left reference samples, the upper left corner reference sample, and the upper reference samples may be used for intra prediction.
  • the mode value of the intra prediction mode is 34
  • the value of the pixel of the prediction block may be the value of the reference sample located diagonally at the top left of the pixel.
  • At least some of the upper right reference samples may be used for intra prediction.
  • At least some of lower left reference samples may be used for intra prediction.
  • the upper left corner reference sample may be used for intra prediction.
  • the number of reference samples used to determine the pixel value of one pixel of the prediction block may be one or two or more.
  • a pixel value of a pixel of a prediction block may be determined according to the position of the pixel and the position of the reference sample indicated by the direction of the intra prediction mode.
  • the position of the reference sample indicated by the position of the pixel and the direction of the intra prediction mode is an integer position
  • a value of one reference sample indicated by the integer position may be used to determine a pixel value of a pixel of the prediction block.
  • an interpolated reference sample may be generated based on two reference samples closest to the position of the reference sample. there is.
  • the value of the interpolated reference sample may be used to determine a pixel value of a pixel of the predictive block.
  • an interpolated value will be generated based on the values of the two samples.
  • a prediction block generated by prediction may not be the same as the original target block.
  • a prediction error which is a difference between a target block and a prediction block, may exist, and a prediction error may also exist between pixels of the target block and pixels of the prediction block.
  • a larger prediction error may occur as the distance between a pixel of a prediction block and a reference sample increases.
  • a discontinuity may occur between a prediction block generated by such a prediction error and a neighboring block.
  • Filtering of prediction blocks may be used to reduce prediction errors. Filtering may be adaptively applying a filter to a region considered to have a large prediction error among prediction blocks. For example, a region considered to have a large prediction error may be a boundary of a prediction block. Also, depending on the intra-prediction mode, a region considered to have a large prediction error among prediction blocks may be different, and filter characteristics may be different.
  • reference lines 0 to 3 may be used for intra prediction of a target block.
  • Each reference line in FIG. 8 may represent a reference sample line including one or more reference samples. The smaller the reference line number, the closer the reference sample line may be to the target block.
  • Samples of segment A and segment F may be obtained through padding using the nearest samples of segment B and segment E, respectively, instead of being obtained from reconstructed neighboring blocks.
  • Index information indicating a reference sample line to be used for intra prediction of a target block may be signaled.
  • the index information may indicate a reference sample line used for intra prediction of a target block among a plurality of reference sample lines.
  • index information may have a value of 0 to 3.
  • reference sample line 0 When the upper boundary of the target block is the boundary of the CTU, only reference sample line 0 may be available. Therefore, in this case, index information may not be signaled. When a reference sample line other than reference sample line 0 is used, filtering of a prediction block described later may not be performed.
  • a prediction block for a target block of a second color component may be generated based on a corresponding reconstructed block of a first color component.
  • the first color component may be a luma component
  • the second color component may be a chroma component
  • parameters of a linear model between a first color component and a second color component may be derived based on a template.
  • the template may include the top reference sample and/or the left reference sample of the target block, and the top reference sample and/or left reference sample of the reconstructed block of the first color component corresponding to these reference samples. there is.
  • the parameters of the linear model are: 1) the value of the sample of the first color component having the largest value among the samples in the template, 2) the value of the sample of the second color component corresponding to the sample of this first color component, 3) the value of the sample of the first color component having the minimum value among the samples in the template and 4) the value of the sample of the second color component corresponding to the sample of the first color component.
  • a prediction block for the target block may be generated by applying the corresponding reconstructed block to the linear model.
  • sub-sampling may be performed on the neighboring samples of the reconstructed block of the first color component and the corresponding reconstructed block. For example, when one sample of the second color component corresponds to four samples of the first color component, one corresponding sample may be calculated by subsampling the four samples of the first color component. there is. When subsampling is performed, derivation of parameters of the linear model and intra prediction between color components may be performed based on the subsampled corresponding samples.
  • Whether intra prediction between color components is performed and/or a template range may be signaled as an intra prediction mode.
  • a target block may be divided into 2 or 4 sub-blocks in a horizontal direction and/or a vertical direction.
  • Divided sub-blocks may be sequentially reconstructed. That is, as intra prediction is performed on a sub-block, a sub-prediction block for the sub-block may be generated. In addition, as inverse quantization and/or inverse transformation are performed on the sub-block, a sub-residual block for the sub-block may be generated. A reconstructed sub-block may be generated by adding the sub-prediction block to the sub-residual block. The reconstructed subblock may be used as a reference sample for intra prediction of a subblock of a later order.
  • a sub-block may be a block including a specified number (eg, 16) or more samples.
  • the target block may be divided into two sub-blocks.
  • the target block is a 4x4 block, the target block cannot be divided into sub blocks. If the target block has other sizes, the target block may be divided into 4 sub-blocks.
  • Such sub-block-based intra prediction may be limited to being performed only when the reference sample line 0 is used.
  • filtering of a prediction block described later may not be performed.
  • a final prediction block may be generated by performing filtering on a prediction block generated by intra prediction.
  • Filtering may be performed by applying specific weights to the filtering target sample, the left reference sample, the top reference sample, and/or the top left reference sample.
  • a weight and/or a reference sample (or a range of reference samples or a location of the reference sample) used for filtering may be determined based on at least one of a block size, an intra prediction mode, and a position of a sample to be filtered within a prediction block. there is.
  • filtering may be performed only for a specified intra prediction mode (eg, DC mode, planar mode, vertical mode, horizontal mode, diagonal mode, and/or adjacent diagonal mode).
  • a specified intra prediction mode eg, DC mode, planar mode, vertical mode, horizontal mode, diagonal mode, and/or adjacent diagonal mode.
  • the adjacent diagonal mode may be a mode having a number obtained by adding k to the number of the diagonal mode, or a mode having a number obtained by subtracting k from the number of the diagonal mode.
  • the number of adjacent diagonal modes may be the sum of the number of diagonal modes and k, and may be the difference between the number of diagonal modes and k.
  • k may be a positive integer of 8 or less.
  • the intra prediction mode of the target block may be derived using intra prediction modes of neighboring blocks existing around the target block, and the derived intra prediction mode may be entropy-encoded and/or entropy-decoded.
  • the intra-prediction mode of the target block and the intra-prediction mode of the neighboring block are the same, information that the intra-prediction mode of the target block and the intra-prediction mode of the neighboring block are the same may be signaled using specified flag information. .
  • indicator information about a neighboring block having the same intra prediction mode as that of the target block among intra prediction modes of a plurality of neighboring blocks may be signaled.
  • entropy encoding and/or entropy decoding based on the intra prediction mode of the neighboring block is performed to obtain information about the intra prediction mode of the target block.
  • Entropy encoding and/or entropy decoding may be performed.
  • FIG. 9 is a diagram for explaining an embodiment of an inter prediction process.
  • the rectangle shown in FIG. 9 may represent an image (or picture). Also, arrows in FIG. 9 may indicate prediction directions. An arrow pointing from a first picture to a second picture may indicate that the second picture refers to the first picture. That is, an image may be encoded and/or decoded according to a prediction direction.
  • Each image may be classified into an intra picture (I picture), a uni-prediction picture (P picture), and a bi-prediction picture (B picture) according to an encoding type.
  • I picture intra picture
  • P picture uni-prediction picture
  • B picture bi-prediction picture
  • Each picture may be coded and/or decoded according to the coding type of each picture.
  • the target image may be encoded using data within the image itself without inter prediction referring to another image.
  • an I picture can be coded only with intra prediction.
  • the target image When the target image is a P picture, the target image may be coded through inter prediction using only reference pictures existing in one direction.
  • the unidirectional direction may be a forward direction or a reverse direction.
  • the target video When the target video is a B picture, the target video may be coded through inter prediction using reference pictures existing in both directions or inter prediction using reference pictures existing in one of forward and backward directions.
  • both directions may be forward and reverse directions.
  • P-pictures and B-pictures that are coded and/or decoded using reference pictures may be regarded as pictures in which inter prediction is used.
  • Inter prediction or motion compensation may be performed using a reference image and motion information.
  • the encoding apparatus 100 may perform inter prediction and/or motion compensation on a target block.
  • the decoding apparatus 200 may perform inter prediction and/or motion compensation corresponding to inter prediction and/or motion compensation performed in the encoding apparatus 100 on a target block.
  • Motion information on the target block may be derived during inter prediction by each of the encoding apparatus 100 and the decoding apparatus 200 .
  • Motion information may be derived using motion information of a reconstructed neighboring block, motion information of a collocated block, and/or motion information of a block adjacent to the collocated block.
  • the encoding apparatus 100 or the decoding apparatus 200 performs prediction and/or motion compensation by using motion information of a spatial candidate and/or a temporal candidate as motion information of a target block.
  • a target block may mean a PU and/or a PU partition.
  • a spatial candidate may be a reconstructed block that is spatially adjacent to the target block.
  • a temporal candidate may be a reconstructed block corresponding to a target block in an already reconstructed collocated picture (col picture).
  • the encoding apparatus 100 and the decoding apparatus 200 may improve encoding efficiency and decoding efficiency by using motion information of spatial candidates and/or temporal candidates.
  • Motion information of spatial candidates may be referred to as spatial motion information.
  • Motion information of temporal candidates may be referred to as temporal motion information.
  • motion information of a spatial candidate may be motion information of a PU including the spatial candidate.
  • Motion information of the temporal candidate may be motion information of a PU including the temporal candidate.
  • Motion information of the candidate block may be motion information of a PU including the candidate block.
  • Inter prediction may be performed using a reference picture.
  • a reference picture may be at least one of a previous picture of the target picture or a subsequent picture of the target picture.
  • a reference picture may refer to an image used for prediction of a target block.
  • a region within a reference picture may be specified by using a reference picture index (or refIdx) indicating a reference picture and a motion vector to be described later.
  • a specified region within a reference picture may represent a reference block.
  • Inter prediction may select a reference picture and may select a reference block corresponding to a target block within the reference picture. In addition, inter prediction may generate a prediction block for a target block using the selected reference block.
  • Motion information may be derived during inter prediction by each of the encoding apparatus 100 and the decoding apparatus 200 .
  • a spatial candidate may be a block that 1) exists in the target picture, 2) has already been reconstructed through encoding and/or decoding, and 3) is located adjacent to or at a corner of the target block.
  • the block located at the corner of the target block may be a block vertically adjacent to a neighboring block horizontally adjacent to the target block or a block horizontally adjacent to a neighboring block vertically adjacent to the target block.
  • a block located at a corner of a target block may have the same meaning as "a block adjacent to a corner of a target block”.
  • a "block located at a corner of a target block” may be included in a "block adjacent to the target block”.
  • spatial candidates include a reconstructed block located to the left of the target block, a reconstructed block located on top of the target block, a reconstructed block located at the lower left corner of the target block, and a reconstructed block located at the upper right corner of the target block. It can be a reconstructed block or a reconstructed block located in the upper left corner of the target block.
  • Each of the encoding apparatus 100 and the decoding apparatus 200 may identify a block existing at a position spatially corresponding to a target block within a col picture.
  • a position of a target block in a target picture and a position of an identified block in a collocated picture may correspond to each other.
  • Each of the encoding apparatus 100 and the decoding apparatus 200 may determine a col block existing at a predefined relative position with respect to the identified block as a temporal candidate.
  • the predefined relative position may be a position inside and/or outside the identified block.
  • the collocated block may include a first collocated block and a second collocated block.
  • the first collocated block may be a block located at the coordinates (xP + nPSW, yP + nPSH).
  • the second collocated block may be a block located at coordinates (xP + (nPSW >> 1), yP + (nPSH >> 1).
  • the second call block may be selectively used when the first call block is unavailable.
  • the motion vector of the target block may be determined based on the motion vector of the collocated block.
  • Each of the encoding apparatus 100 and the decoding apparatus 200 may scale a motion vector of a collocated block.
  • a scaled motion vector of the collocated block may be used as a motion vector of the target block.
  • a motion vector of motion information of a temporal candidate stored in the list may be a scaled motion vector.
  • a ratio of the motion vector of the target block and the motion vector of the collocated block may be equal to the ratio of the first temporal distance to the second temporal distance.
  • the first temporal distance may be the distance between the reference picture of the target block and the target picture.
  • the second temporal distance may be a distance between a reference picture of a collocated block and a collocated picture.
  • a method of deriving motion information may change according to an inter prediction mode of a target block.
  • inter prediction modes applied for inter prediction an Advanced Motion Vector Predictor (AMVP) mode, a merge mode and a skip mode, a merge mode with motion vector difference
  • AMVP Advanced Motion Vector Predictor
  • Merge mode may also be referred to as motion merge mode. Below, each of the modes is described in detail.
  • the encoding apparatus 100 may search for similar blocks in the neighborhood of the target block.
  • the encoding apparatus 100 may obtain a prediction block by performing prediction on a target block using motion information of a similar block found.
  • the encoding apparatus 100 may encode a residual block that is a difference between a target block and a prediction block.
  • each of the encoding apparatus 100 and the decoding apparatus 200 may generate a predictive motion vector candidate list using a motion vector of a spatial candidate, a motion vector of a temporal candidate, and a zero vector.
  • the predicted motion vector candidate list may include one or more predicted motion vector candidates. At least one of the motion vector of the spatial candidate, the motion vector of the temporal candidate, and the zero vector may be determined and used as the predicted motion vector candidate.
  • predicted motion vector (candidate) and “motion vector (candidate)” may be used interchangeably and may be used interchangeably.
  • predicted motion vector candidate and “AMVP candidate” may be used interchangeably and may be used interchangeably.
  • prediction motion vector candidate list and “AMVP candidate list” may be used in the same meaning and may be used interchangeably.
  • Spatial candidates may include reconstructed spatial neighboring blocks.
  • the motion vector of the reconstructed neighboring block may be referred to as a spatial prediction motion vector candidate.
  • Temporal candidates may include a collocated block and a block adjacent to the collocated block.
  • a motion vector of a collocated block or a motion vector of a block adjacent to the collocated block may be referred to as a temporal prediction motion vector candidate.
  • the zero vector may be a (0, 0) motion vector.
  • the predictive motion vector candidate may be a motion vector predictor for motion vector prediction. Also, in the encoding apparatus 100, the predicted motion vector candidate may be an initial motion vector search position.
  • the encoding apparatus 100 may determine a motion vector to be used for encoding a target block within a search range by using the predictive motion vector candidate list. Also, the encoding apparatus 100 may determine a predictive motion vector candidate to be used as a predictive motion vector of the target block from among predictive motion vector candidates in the predictive motion vector candidate list.
  • a motion vector to be used for encoding a target block may be a motion vector that can be encoded with minimal cost.
  • the encoding apparatus 100 may determine whether to use the AMVP mode in encoding the target block.
  • the encoding apparatus 100 may generate a bitstream including inter prediction information required for inter prediction.
  • the decoding apparatus 200 may perform inter prediction on a target block using inter prediction information of a bitstream.
  • the inter-prediction information includes 1) mode information indicating whether the AMVP mode is used, 2) predictive motion vector index, 3) motion vector difference (MVD), 4) reference direction, and 5) reference picture index. can do.
  • predicted motion vector index and “AMVP index” may be used interchangeably and may be used interchangeably.
  • the inter prediction information may include a residual signal.
  • the decoding apparatus 200 may obtain a predicted motion vector index, a motion vector difference, a reference direction, and a reference picture index from a bitstream through entropy decoding.
  • the predictive motion vector index may indicate a predictive motion vector candidate used for prediction of a target block among predictive motion vector candidates included in the predictive motion vector candidate list.
  • the decoding apparatus 200 may derive a predicted motion vector candidate using the predicted motion vector candidate list, and may determine motion information of a target block based on the derived motion vector predicted candidate.
  • the decoding apparatus 200 may determine a motion vector candidate for a target block from among motion vector predictor candidates included in a predictor motion vector candidate list by using the predictor motion vector index.
  • the decoding apparatus 200 may select a predicted motion vector candidate indicated by a predicted motion vector index as a predicted motion vector of a target block from among motion vector predicted candidates included in the predicted motion vector candidate list.
  • the encoding apparatus 100 may generate an entropy-encoded predicted motion vector index by applying entropy encoding to the predicted motion vector index, and may generate a bitstream including the entropy-encoded predicted motion vector index.
  • the entropy-encoded predicted motion vector index may be signaled from the encoding apparatus 100 to the decoding apparatus 200 through a bitstream.
  • the decoding apparatus 200 may extract an entropy-encoded predictive motion vector index from a bitstream and obtain the predictive motion vector index by applying entropy decoding to the entropy-encoded predictive motion vector index.
  • a motion vector actually used for inter prediction of a target block may not coincide with a predicted motion vector.
  • MVD may be used to represent a difference between a motion vector actually used for inter prediction of a target block and a predicted motion vector.
  • the encoding apparatus 100 may derive a predicted motion vector similar to a motion vector actually used for inter prediction of a target block in order to use an MVD having a size as small as possible.
  • MVD may be a difference between a motion vector of a target block and a predicted motion vector.
  • the encoding apparatus 100 may calculate the MVD and generate an entropy-encoded MVD by applying entropy encoding to the MVD.
  • the encoding apparatus 100 may generate a bitstream including an entropy-encoded MDV.
  • the MVD may be transmitted from the encoding device 100 to the decoding device 200 through a bitstream.
  • the decoding apparatus 200 may extract the entropy-encoded MVD from the bitstream and obtain the MVD by applying entropy decoding to the entropy-encoded MVD.
  • the decoding apparatus 200 may derive the motion vector of the target block by adding the MVD and the predicted motion vector.
  • the motion vector of the target block derived by the decoding apparatus 200 may be the sum of the MVD and the motion vector candidate.
  • the encoding apparatus 100 may generate entropy-encoded MVD resolution information by applying entropy encoding to the calculated MVD resolution information, and may generate a bitstream including the entropy-encoded MVD resolution information.
  • the decoding apparatus 200 may extract entropy-encoded MVD resolution information from a bitstream, and obtain MVD resolution information by applying entropy decoding to the entropy-encoded MVD resolution information.
  • the decoding apparatus 200 may adjust the resolution of the MVD using the MVD resolution information.
  • the encoding device 100 may calculate the MVD based on the affine model.
  • the decoding apparatus 200 may derive an affine control motion vector of the target block through the sum of the MVD and the affine control motion vector candidate, and may derive a motion vector for a subblock using the affine control motion vector. there is.
  • a reference direction may indicate a reference picture list used for prediction of a target block.
  • the reference direction may indicate one of a reference picture list L0 and a reference picture list L1.
  • the reference direction indicates only a reference picture list used for prediction of the target block, and may not indicate that directions of reference pictures are limited to a forward direction or a backward direction.
  • each of the reference picture list L0 and the reference picture list L1 may include forward and/or backward pictures.
  • That the reference direction is uni-direction may mean that one reference picture list is used.
  • Bi-direction of the reference direction may mean that two reference picture lists are used.
  • the reference direction can indicate one of: that only the reference picture list L0 is used, that only the reference picture list L1 is used, and two reference picture lists.
  • the reference picture index may indicate a reference picture used for prediction of a target block among reference pictures of a reference picture list.
  • the encoding apparatus 100 may generate an entropy-coded reference picture index by applying entropy encoding to the reference picture index, and may generate a bitstream including the entropy-coded reference picture index.
  • the entropy-encoded reference picture index may be signaled from the encoding device 100 to the decoding device 200 through a bitstream.
  • the decoding apparatus 200 may extract an entropy-encoded reference picture index from a bitstream and obtain the reference picture index by applying entropy decoding to the entropy-encoded reference picture index.
  • two reference picture lists are used for prediction of a target block.
  • One reference picture index and one motion vector may be used for each reference picture list.
  • two prediction blocks may be specified for the target block. For example, a (final) prediction block of the target block may be generated through an average or a weighted-sum of two prediction blocks of the target block.
  • the motion vector of the target block may be derived by the predicted motion vector index, MVD, reference direction, and reference picture index.
  • the decoding apparatus 200 may generate a prediction block for the target block based on the derived motion vector and the reference picture index.
  • the prediction block may be a reference block indicated by a derived motion vector in a reference picture indicated by a reference picture index.
  • the amount of bits transmitted from the encoding apparatus 100 to the decoding apparatus 200 can be reduced and encoding efficiency can be improved.
  • Motion information of a neighboring block reconstructed with respect to the target block may be used.
  • the encoding apparatus 100 may not separately encode the motion information of the target block itself.
  • Motion information of the target block is not encoded, and other information capable of inducing motion information of the target block through motion information of the reconstructed neighboring block may be encoded instead.
  • other information is encoded instead, the amount of bits transmitted to the decoding apparatus 200 can be reduced and encoding efficiency can be improved.
  • the encoding apparatus 100 and the decoding apparatus 200 may use an identifier and/or an index indicating which unit's motion information among reconstructed neighboring units is used as the motion information of the target unit.
  • Merge may refer to merging of motions of a plurality of blocks. Merge may mean applying motion information of one block to another block as well.
  • merge mode may refer to a mode in which motion information of a target block is derived from motion information of neighboring blocks.
  • the encoding apparatus 100 may perform prediction of motion information of a target block by using motion information of a spatial candidate and/or motion information of a temporal candidate.
  • the spatial candidate may include a reconstructed spatial neighboring block spatially adjacent to the target block.
  • a spatial neighboring block may include a left neighboring block and an upper neighboring block.
  • Temporal candidates may include call blocks.
  • spatial candidate and “spatial merge candidate” may be used interchangeably and may be used interchangeably.
  • the terms “temporal candidate” and “temporal merge candidate” may be used interchangeably and may be used interchangeably.
  • the encoding apparatus 100 may obtain a prediction block through prediction.
  • the encoding apparatus 100 may encode a residual block that is a difference between a target block and a prediction block.
  • each of the encoding apparatus 100 and the decoding apparatus 200 may generate a merge candidate list using motion information of spatial candidates and/or motion information of temporal candidates.
  • Motion information may include 1) a motion vector, 2) a reference picture index, and 3) a reference direction.
  • the reference direction can be unidirectional or bidirectional.
  • the reference direction may mean an inter prediction indicator.
  • the merge candidate list may include merge candidates.
  • a merge candidate may be motion information.
  • the merge candidate list may be a list in which motion information is stored.
  • Merge candidates may be motion information such as temporal candidates and/or spatial candidates.
  • the merge candidate list may include motion information such as a temporal candidate and/or a spatial candidate.
  • the merge candidate list may include a new merge candidate generated by a combination of merge candidates already existing in the merge candidate list.
  • the merge candidate list may include new motion information generated by combining motion information already existing in the merge candidate list.
  • the merge candidate list may include a history-based merge candidate.
  • a history-based merge candidate may be motion information of a block encoded and/or decoded prior to a target block.
  • the merge candidate list may include a merge candidate based on an average of two merge candidates.
  • Merge candidates may be specified modes for deriving inter prediction information.
  • a merge candidate may be information indicating a specific mode for deriving inter prediction information.
  • Inter prediction information of a target block may be derived according to a specified mode indicated by a merge candidate.
  • the specified mode may include a process of deriving a series of inter prediction information.
  • This specified mode may be an inter prediction information derivation mode or a motion information derivation mode.
  • Inter prediction information of a target block may be derived according to a mode indicated by a merge candidate selected by a merge index among merge candidates in the merge candidate list.
  • the motion information derivation modes in the merge candidate list may be at least one of 1) a motion information derivation mode in units of sub-blocks and 2) an affine motion information derivation mode.
  • the merge candidate list may include motion information of a zero vector.
  • a zero vector may be referred to as a zero merge candidate.
  • the motion information in the merge candidate list is: 1) motion information of a spatial candidate, 2) motion information of a temporal candidate, 3) motion information generated by a combination of motion information already existing in the merge candidate list, 4) zero vector may be at least one of
  • Motion information may include 1) a motion vector, 2) a reference picture index, and 3) a reference direction.
  • a reference direction may be referred to as an inter prediction indicator.
  • the reference direction can be unidirectional or bidirectional.
  • a unidirectional reference direction may represent L0 prediction or L1 prediction.
  • the merge candidate list may be generated before prediction by merge mode is performed.
  • the number of merge candidates in the merge candidate list may be predefined.
  • the encoding device 100 and the decoding device 200 may add merge candidates to the merge candidate list according to a predefined method and a predefined order so that the merge candidate list includes a predefined number of merge candidates.
  • the merge candidate list of the encoding device 100 and the merge candidate list of the decoding device 200 may be the same through a predefined method and a predefined ranking.
  • Merge may be applied in units of CUs or units of PUs.
  • the encoding apparatus 100 may transmit a bitstream including predefined information to the decoding apparatus 200.
  • the predefined information includes 1) information indicating whether merging is to be performed for each block partition, 2) which block to merge with among blocks that are spatial candidates and/or temporal candidates for the target block. It may contain information on whether
  • the encoding apparatus 100 may determine a merge candidate to be used for encoding a target block. For example, the encoding apparatus 100 may perform predictions on a target block using merge candidates of the merge candidate list and generate residual blocks for the merge candidates. The encoding apparatus 100 may use a merge candidate that requires the least cost in encoding the prediction and residual blocks for encoding the target block.
  • the encoding apparatus 100 may determine whether to use merge mode in encoding a target block.
  • the encoding apparatus 100 may generate a bitstream including inter prediction information required for inter prediction.
  • the encoding apparatus 100 may generate entropy-encoded inter-prediction information by performing entropy encoding on the inter-prediction information, and may transmit a bitstream including the entropy-encoded inter-prediction information to the decoding apparatus 200.
  • Entropy-encoded inter prediction information may be signaled from the encoding apparatus 100 to the decoding apparatus 200 through a bitstream.
  • the decoding apparatus 200 may extract entropy-encoded inter prediction information from a bitstream and obtain inter prediction information by performing entropy decoding on the entropy-encoded inter prediction information.
  • the decoding apparatus 200 may perform inter prediction on a target block using inter prediction information of a bitstream.
  • the inter prediction information may include 1) mode information indicating whether merge mode is used, 2) merge index, and 3) correction information.
  • the inter prediction information may include a residual signal.
  • the decoding apparatus 200 may obtain the merge index from the bitstream only when the mode information indicates that the merge mode is used.
  • Mode information may be a merge flag.
  • a unit of mode information may be a block.
  • Information about a block may include mode information, and the mode information may indicate whether merge mode is applied to the block.
  • the merge index may indicate a merge candidate used for prediction of a target block among merge candidates included in the merge candidate list.
  • the merge index may indicate which block among neighboring blocks that are spatially or temporally adjacent to the target block is merged with.
  • the encoding apparatus 100 may select a merge candidate having the highest encoding performance among merge candidates included in the merge candidate list, and set a merge index value to indicate the selected merge candidate.
  • the correction information may be information used for motion vector correction.
  • the encoding device 100 may generate correction information.
  • the decoding apparatus 200 may correct the motion vector of the merge candidate selected by the merge index based on the correction information.
  • the correction information may include at least one of information indicating whether correction is performed, correction direction information, and correction size information.
  • a prediction mode for correcting a motion vector based on signaled correction information may be referred to as a merge mode with motion vector difference.
  • the decoding apparatus 200 may perform prediction on a target block by using a merge candidate indicated by a merge index among merge candidates included in the merge candidate list.
  • the motion vector of the target block may be specified by the motion vector of the merge candidate indicated by the merge index, the reference picture index, and the reference direction.
  • the skip mode may be a mode in which motion information of a spatial candidate or motion information of a temporal candidate is applied to a target block as it is. Also, the skip mode may be a mode not using a residual signal. That is to say, when skip mode is used, the reconstructed block may be the same as the predicted block.
  • a difference between merge mode and skip mode may be transmission or use of a residual signal.
  • skip mode may be similar to merge mode except that no residual signal is transmitted or used.
  • the encoding apparatus 100 transmits information indicating which block's motion information among spatial or temporal candidate blocks is used as the motion information of the target block to the decoding apparatus 200 through a bitstream.
  • the encoding apparatus 100 may generate entropy-encoded information by performing entropy encoding on such information, and may signal the entropy-encoded information to the decoding apparatus 200 through a bitstream.
  • the decoding apparatus 200 may extract entropy-encoded information from a bitstream and obtain information by performing entropy decoding on the entropy-encoded information.
  • the encoding device 100 may not transmit other syntax element information such as MVD to the decoding device 200.
  • the encoding apparatus 100 may not signal a syntax element related to at least one of the MVD, the coded block flag, and the transform coefficient level to the decoding apparatus 200.
  • Skip mode can also use a merge candidate list.
  • the merge candidate list can be used in both merge mode and skip mode.
  • the merge candidate list may be referred to as a "skip candidate list” or a "merge/skip candidate list”.
  • skip mode may use a separate candidate list different from merge mode.
  • the merge candidate list and the merge candidate may be replaced with the skip candidate list and the skip candidate, respectively.
  • the merge candidate list may be generated before prediction by skip mode is performed.
  • the encoding apparatus 100 may determine a merge candidate to be used for encoding a target block. For example, the encoding apparatus 100 may perform predictions on a target block using merge candidates of a merge candidate list. The encoding apparatus 100 may use a merge candidate requiring a minimum prediction cost for encoding a target block.
  • the encoding apparatus 100 may determine whether to use a skip mode in encoding a target block.
  • the encoding apparatus 100 may generate a bitstream including inter prediction information required for inter prediction.
  • the decoding apparatus 200 may perform inter prediction on a target block using inter prediction information of a bitstream.
  • the inter-prediction information may include 1) mode information indicating whether a skip mode is used and 2) a skip index.
  • the skip index may be the same as the aforementioned merge index.
  • the target block When the skip mode is used, the target block may be coded without a residual signal. Inter prediction information may not include a residual signal. Alternatively, the bitstream may not include a residual signal.
  • the decoding apparatus 200 may obtain the skip index from the bitstream only when the mode information indicates that the skip mode is used. As mentioned above, the merge index and skip index may be the same. The decoding apparatus 200 may obtain the skip index from the bitstream only when the mode information indicates that merge mode or skip mode is used.
  • the skip index may indicate a merge candidate used for prediction of a target block among merge candidates included in the merge candidate list.
  • the decoding apparatus 200 may perform prediction on a target block by using a merge candidate indicated by a skip index among merge candidates included in the merge candidate list.
  • a motion vector of a target block may be specified by a motion vector of a merge candidate indicated by a skip index, a reference picture index, and a reference direction.
  • the current picture reference mode may mean a prediction mode using a pre-reconstructed region in a target picture to which a target block belongs.
  • a motion vector may be used to specify the pre-reconstructed area. Whether the target block is encoded in the current picture reference mode can be determined using the reference picture index of the target block.
  • a flag or index indicating whether the target block is a block encoded in the current picture reference mode may be signaled from the encoding device 100 to the decoding device 200. Alternatively, whether the target block is a block coded in the current picture reference mode may be inferred through a reference picture index of the target block.
  • the target picture may exist at a fixed position or an arbitrary position in the reference picture list for the target block.
  • the fixed position may be a position where the value of the reference picture index is 0 or the last position.
  • a separate reference picture index indicating such an arbitrary position may be signaled from the encoding apparatus 100 to the decoding apparatus 200.
  • a sub-block merge mode may mean a mode for deriving motion information for a sub-block of a CU.
  • motion information of a call sub-block of a target sub-block in a reference image ie, a sub-block based temporal merge candidate
  • an affine control point motion vector A subblock merge candidate list may be generated using an affine control point motion vector merge candidate.
  • divided target blocks may be generated by dividing the target block in a diagonal direction.
  • motion information of each divided target block may be derived, and prediction samples for each divided target block may be derived using the derived motion information.
  • a prediction sample of the target block may be derived through a weighted sum of prediction samples of the divided target blocks.
  • the inter-intra combined prediction mode may be a mode in which a prediction sample of a target block is derived by using a weighted sum of prediction samples generated by inter prediction and prediction samples generated by intra prediction.
  • the decoding apparatus 200 may perform self-correction on the derived motion information. For example, the decoding apparatus 200 may search a specific area based on a reference block indicated by the derived motion information to search for motion information having a minimum sum of absolute differences (SAD). and the searched motion information may be derived as corrected motion information.
  • SAD minimum sum of absolute differences
  • the decoding apparatus 200 may perform compensation for prediction samples derived through inter prediction using an optical flow.
  • motion information to be used for prediction of a target block among motion information in the list may be specified through an index of the list.
  • the encoding apparatus 100 may signal only the index of an element that causes the least cost in inter prediction of a target block among elements in the list.
  • the encoding device 100 may encode the index and signal the encoded index.
  • the aforementioned lists may have to be derived in the same manner based on the same data in the encoding apparatus 100 and the decoding apparatus 200.
  • the same data may include a reconstructed picture and a reconstructed block.
  • the order of the elements within the list may need to be constant.
  • a large block in the middle may represent a target block.
  • Five small blocks may represent spatial candidates.
  • Coordinates of the target block may be (xP, yP), and the size of the target block may be (nPSW, nPSH).
  • the spatial candidate A 0 may be a block adjacent to the lower left corner of the target block.
  • a 0 may be a block occupying pixels of coordinates (xP - 1, yP + nPSH).
  • Spatial candidate A 1 may be a block adjacent to the left of the target block.
  • a 1 may be the lowest block among blocks adjacent to the left side of the target block.
  • a 1 may be a block adjacent to the top of A 0 .
  • a 1 may be a block occupying pixels of coordinates (xP - 1, yP + nPSH - 1).
  • Spatial candidate B 0 may be a block adjacent to the upper right corner of the target block.
  • B 0 may be a block occupying pixels of coordinates (xP + nPSW, yP - 1).
  • Spatial candidate B 1 may be a block adjacent to the top of the target block.
  • B 1 may be the rightmost block among blocks adjacent to the top of the target block.
  • B 1 may be a block adjacent to the left side of B 0 .
  • B 1 may be a block occupying pixels of coordinates (xP + nPSW - 1, yP - 1).
  • Spatial candidate B 2 may be a block adjacent to the upper left corner of the target block.
  • B 2 may be a block occupying pixels of coordinates (xP - 1, yP - 1).
  • a candidate block may include a spatial candidate and a temporal candidate.
  • the above determination may be made by sequentially applying steps 1) to 4) below.
  • Step 1) If the PU including the candidate block is outside the boundary of the picture, the availability of the candidate block may be set to false. "Availability is set to false” may mean the same as “availability is set to unavailability”.
  • Step 2 If the PU including the candidate block is outside the boundary of the slice, the availability of the candidate block may be set to false. If the target block and the candidate block are located in different slices, the availability of the candidate block may be set to false.
  • Step 3 If the PU including the candidate block is outside the boundary of the tile, the availability of the candidate block may be set to false. If the target block and the candidate block are located in different tiles, the availability of the candidate block may be set to false.
  • Step 4 If the prediction mode of the PU including the candidate block is an intra prediction mode, availability of the candidate block may be set to false. If the PU containing the candidate block does not use inter prediction, the availability of the candidate block may be set to false.
  • 11 illustrates an order of adding motion information of spatial candidates to a merge list according to an example.
  • the order of A 1 , B 1 , B 0 , A 0 and B 2 may be used. That is, in the order of A 1 , B 1 , B 0 , A 0 and B 2 , motion information of available spatial candidates may be added to the merge list.
  • the maximum number of merge candidates in the merge list may be set.
  • the set maximum number is indicated by N.
  • the set number may be transmitted from the encoding device 100 to the decoding device 200.
  • a slice header of a slice may include N.
  • the maximum number of merge candidates of the merge list for the target block of the slice may be set by the slice header.
  • the value of N may be 5 by default.
  • Motion information (ie, merge candidates) may be added to the merge list in the order of steps 1) to 4) below.
  • Step 1) Available spatial candidates among spatial candidates may be added to the merge list.
  • Motion information of available spatial candidates may be added to the merge list in the order shown in FIG. 11 . In this case, if motion information of an available spatial candidate overlaps with other motion information already existing in the merge list, the motion information may not be added to the merge list. Checking whether it overlaps with other motion information existing in the list can be abbreviated as "redundancy check”.
  • a maximum of N pieces of motion information may be added.
  • Step 2 If the number of pieces of motion information in the merge list is smaller than N and a temporal candidate is available, the motion information of the temporal candidate may be added to the merge list. In this case, if motion information of an available temporal candidate overlaps with other motion information already existing in the merge list, the motion information may not be added to the merge list.
  • Step 3 If the number of pieces of motion information in the merge list is smaller than N and the type of the target slice is "B”, the combined motion information generated by combined bi-prediction is added to the merge list.
  • the target slice may be a slice including the target block.
  • the combined motion information may be a combination of L0 motion information and L1 motion information.
  • the L0 motion information may be motion information referring only to the reference picture list L0.
  • the L1 motion information may be motion information referring only to the reference picture list L1.
  • L0 motion information there may be one or more L0 motion information. Also, within the merge list, there may be one or more L1 motion information.
  • the combined motion information may be one or more.
  • which L0 motion information and which L1 motion information among one or more pieces of L0 motion information and one or more pieces of L1 motion information are to be used may be predefined.
  • One or more pieces of combined motion information may be generated in a predefined order by combined bidirectional prediction using pairs of different pieces of motion information in a merge list.
  • One of the pairs of different motion information may be L0 motion information and the other may be L1 motion information.
  • the combined motion information added with the highest priority may be a combination of L0 motion information having a merge index of 0 and L1 motion information having a merge index of 1. If the motion information with a merge index of 0 is not L0 motion information or the motion information with a merge index of 1 is not L1 motion information, the combined motion information may not be generated and added.
  • Motion information added next may be a combination of L0 motion information having a merge index of 1 and L1 motion information having a merge index of 0. The following specific combinations may follow other combinations in the field of encoding/decoding of video.
  • the combined motion information may not be added to the merge list.
  • the zero vector motion information may be motion information in which a motion vector is a zero vector.
  • One or more zero vector motion information may be provided.
  • Reference picture indexes of one or more zero vector motion information may be different from each other.
  • the value of the reference picture index of the first zero vector motion information may be 0.
  • the value of the reference picture index of the second zero vector motion information may be 1.
  • the number of zero vector motion information may be equal to the number of reference pictures in the reference picture list.
  • the reference direction of the zero vector motion information may be bidirectional. Both motion vectors may be zero vectors.
  • the number of zero vector motion information may be the smaller of the number of reference pictures in the reference picture list L0 and the number of reference pictures in the reference picture list L1.
  • a unidirectional reference direction may be used for a reference picture index applicable to only one reference picture list.
  • the encoding apparatus 100 and/or the decoding apparatus 200 may sequentially add zero vector motion information to the merge list while changing the reference picture index.
  • the zero vector motion information may not be added to the merge list.
  • steps 1) to 4) described above is merely exemplary, and the order between steps may be interchanged. Also, some of the steps may be omitted according to predefined conditions.
  • the maximum number of motion vector predictor candidates in the predictor motion vector candidate list may be predefined.
  • the predefined maximum number is denoted by N.
  • the predefined maximum number may be 2.
  • Motion information (ie, predicted motion vector candidates) may be added to the predicted motion vector candidate list in the order of steps 1) to 3) below.
  • Step 1) Among the spatial candidates, available spatial candidates may be added to the predicted motion vector candidate list.
  • Spatial candidates may include a first spatial candidate and a second spatial candidate.
  • the first spatial candidate may be one of A 0 , A 1 , scaled A 0 and scaled A 1 .
  • the second spatial candidate may be one of B 0 , B 1 , B 2 , scaled B 0 , scaled B 1 , and scaled B 2 .
  • Motion information of available spatial candidates may be added to the predicted motion vector candidate list in the order of the first spatial candidate and the second spatial candidate.
  • the motion information may not be added to the predictor motion vector candidate list.
  • the value of N is 2, if the motion information of the second spatial candidate is identical to the motion information of the first spatial candidate, the motion information of the second spatial candidate may not be added to the predicted motion vector candidate list.
  • a maximum of N pieces of motion information may be added.
  • Step 2 If the number of pieces of motion information in the predicted motion vector candidate list is smaller than N and a temporal candidate is available, the motion information of the temporal candidate may be added to the predicted motion vector candidate list. In this case, when motion information of an available temporal candidate overlaps with other motion information already existing in the motion vector predictor candidate list, the motion information may not be added to the predictor motion vector candidate list.
  • Step 3 If the number of pieces of motion information in the predicted motion vector candidate list is smaller than N, zero vector motion information may be added to the predicted motion vector candidate list.
  • One or more zero vector motion information may be provided.
  • Reference picture indexes of one or more zero vector motion information may be different from each other.
  • the encoding apparatus 100 and/or the decoding apparatus 200 may sequentially add zero vector motion information to the predictive motion vector candidate list while changing the reference picture index.
  • the zero vector motion information may not be added to the predicted motion vector candidate list.
  • steps 1) to 3) described above is merely exemplary, and the order between steps may be interchanged. Also, some of the steps may be omitted according to predefined conditions.
  • a quantized level may be generated by performing a transform and/or quantization process on the residual signal.
  • a residual signal may be generated as a difference between an original block and a prediction block.
  • the prediction block may be a block generated by intra prediction or inter prediction.
  • the residual signal may be converted into the frequency domain through a transform process that is part of the quantization process.
  • Transformation kernels used for transformation may include various DCT kernels such as Discrete Cosine Transform (DCT) type 2 (DCT-II) and Discrete Sine Transform (DST) kernels. .
  • DCT Discrete Cosine Transform
  • DCT-II Discrete Cosine Transform
  • DST Discrete Sine Transform
  • transform kernels may perform a separable transform or a 2D (2D) non-separable transform on the residual signal.
  • the separable transform may be a transform that performs a one-dimensional (1D) transform on the residual signal in each of a horizontal direction and a vertical direction.
  • DCT type and DST type adaptively used for 1D conversion may include DCT-V, DCT-VIII, DST-I and DST-VII in addition to DCT-II as shown in Table 3 and Table 4 below, respectively. there is.
  • a transform set may be used in deriving the DCT type or DST type to be used for transformation.
  • Each transform set may include a plurality of transform candidates.
  • Each transformation candidate may be a DCT type or a DST type.
  • Table 5 below shows an example of a transform set applied in the horizontal direction and a transform set applied in the vertical direction according to the intra prediction mode.
  • transform sets applied in the horizontal and vertical directions may be predefined according to the intra prediction mode of the target block.
  • the encoding apparatus 100 may perform transform and inverse transform on the residual signal using a transform included in a transform set corresponding to the intra prediction mode of the target block.
  • the decoding apparatus 200 may perform an inverse transform on the residual signal using a transform included in a transform set corresponding to the intra prediction mode of the target block.
  • the set of transforms applied to the residual signal may be determined as exemplified in Tables 3, 4, and 5, and may not be signaled. Transformation indication information may be signaled from the encoding device 100 to the decoding device 200.
  • the transform indication information may be information indicating which transform candidate among a plurality of transform candidates included in a transform set applied to the residual signal is used.
  • transform sets each having three transforms may be configured according to the intra prediction mode.
  • An optimal transform method can be selected among all 9 multiple transform methods resulting from a combination of three transforms in the horizontal direction and three transforms in the vertical direction. Encoding efficiency can be improved by encoding and/or decoding the residual signal using such an optimal conversion method.
  • information on which transform among transforms belonging to the transform set is used for at least one of the vertical transform and the horizontal transform may be entropy encoded and/or decoded.
  • a truncated unary binarization may be used to encode and/or decode this information.
  • a method using various transforms may be applied to a residual signal generated by intra prediction or inter prediction.
  • Transformation may include at least one of a primary transformation and a secondary transformation.
  • a transform coefficient may be generated by performing a primary transform on the residual signal, and a secondary transform coefficient may be generated by performing a secondary transform on the transform coefficient.
  • a primary transformation may be named primary.
  • the primary transform may be referred to as an adaptive multiple transform (AMT).
  • AMT may mean that different transforms are applied to each of the 1D directions (ie, the vertical direction and the horizontal direction).
  • the secondary transform may be a transform for improving the energy concentration of the transform coefficient generated by the primary transform.
  • the second-order transformation may be a separable transformation or a non-separable transformation.
  • the non-separable transform may be a non-separable secondary transform (NSST).
  • the primary transformation may be performed using at least one of a plurality of predefined transformation methods.
  • a plurality of predefined transform methods include a Discrete Cosine Transform (DCT), a Discrete Sine Transform (DST), and a Karhunen-Loeve Transform (KLT) based transform.
  • DCT Discrete Cosine Transform
  • DST Discrete Sine Transform
  • KLT Karhunen-Loeve Transform
  • the primary transform may be a transform having various transform types according to a kernel function defining DCT or DST.
  • the transform type includes: 1) prediction mode of the target block (eg, one of intra prediction and inter prediction), 2) size of the target block, 3) shape of the target block, 4) intra prediction mode of the target block , 5) components of the target block (eg, one of a luma component and a chroma component) and 6) a partition type applied to the target block (eg, Quad Tree (QT), Binary Tree (BT)). ) and one of a ternary tree (TT)).
  • prediction mode of the target block eg, one of intra prediction and inter prediction
  • the target block eg, one of intra prediction and inter prediction
  • components of the target block eg, one of a luma component and a chroma component
  • TT ternary tree
  • the primary transform includes transforms such as DCT-2, DCT-5, DCT-7, DST-7, DST-1, DST-8 and DCT-8 according to the transform kernel presented in Table 6 below. can do.
  • Table 6 various transform types and transform kernel functions for Multiple Transform Selection (MTS) are illustrated.
  • MTS may mean that a combination of one or more DCT and/or DST transform kernels is selected for horizontal and/or vertical transform of the residual signal.
  • i and j may be integer values greater than or equal to 0 and less than or equal to N-1.
  • a secondary transform may be performed on transform coefficients generated by performing the primary transform.
  • a set of transforms can be defined in the second-order transform.
  • Methods for deriving and/or determining a set of transforms, such as those described above, may be applied to first-order as well as second-order transforms.
  • a primary transform and a secondary transform can be determined for a specified object.
  • the first transform and the second transform may be applied to one or more signal components of a luma component and a chroma component.
  • Whether to apply the primary transform and/or the secondary transform may be determined according to at least one of coding parameters for a target block and/or a neighboring block.
  • whether to apply the first transform and/or the second transform may be determined by the size and/or shape of the target block.
  • transformation information indicating a transformation method used for a target may be derived by using specified information.
  • the transformation information may include an index of a transformation to be used for primary transformation and/or secondary transformation.
  • the transform information may indicate that the primary transform and/or the secondary transform are not used.
  • the transform method(s) applied to the primary transform and/or the secondary transform indicated by transform information may be applied to the target block and/or neighboring blocks. It may be determined according to at least one of the coding parameters for
  • transformation information indicating a transformation method for a specified target may be signaled from the encoding apparatus 100 to the decoding apparatus 200.
  • an index indicating the primary transform, whether or not a secondary transform is used, and an index indicating the secondary transform may be derived as transformation information in the decoding apparatus 200 there is.
  • conversion information indicating whether or not the primary transform is used for one CU, an index indicating the primary transform, whether or not a secondary transform is used, and an index indicating the secondary transform may be signaled.
  • a quantized transform coefficient (ie, a quantized level) may be generated by performing quantization on a residual signal or a result generated by performing the primary transform and/or the secondary transform.
  • FIG 13 illustrates diagonal scanning according to an example.
  • the quantized transform coefficients may be scanned according to at least one of (up-right) diagonal scanning, vertical scanning, and horizontal scanning according to at least one of an intra prediction mode, a block size, and a block shape.
  • a block may be a transform unit.
  • Each scanning can start at a specified start point and end at a specified end point.
  • quantized transform coefficients may be changed into a one-dimensional vector form by scanning coefficients of a block using the diagonal scanning of FIG. 13 .
  • the horizontal scanning of FIG. 14 or the vertical scanning of FIG. 15 may be used instead of diagonal scanning according to the block size and/or intra prediction mode.
  • Vertical scanning may be scanning two-dimensional block form coefficients in a column direction.
  • Horizontal scanning may be scanning two-dimensional block form coefficients in a row direction.
  • which of diagonal scanning, vertical scanning, and horizontal scanning is to be used may be determined according to the size of the block and/or the inter-prediction mode.
  • quantized transform coefficients may be scanned along a diagonal, horizontal, or vertical direction.
  • Quantized transform coefficients may be expressed in block form.
  • a block may include a plurality of sub-blocks. Each sub-block may be defined according to a minimum block size or a minimum block shape.
  • a scanning order according to a type or direction of scanning may be applied to subblocks first. Also, a scanning order according to a scanning direction may be applied to quantized transform coefficients in a sub-block.
  • transform coefficients quantized by primary transform, secondary transform, and quantization of the residual signal of the target block are can be created Thereafter, one of three scanning orders may be applied to the four 4x4 subblocks, and quantized transform coefficients may be scanned according to the scanning order for each 4x4 subblock.
  • the encoding apparatus 100 may generate an entropy-encoded quantized transform coefficient by performing entropy encoding on the scanned quantized transform coefficients, and may generate a bitstream including the entropy-coded quantized transform coefficients. .
  • the decoding apparatus 200 may extract entropy-encoded quantized transform coefficients from a bitstream and generate quantized transform coefficients by performing entropy decoding on the entropy-coded quantized transform coefficients.
  • Quantized transform coefficients may be arranged in a 2D block form through inverse scanning. At this time, as a reverse scanning method, at least one of (upper right) diagonal scan, vertical scan, and horizontal scan may be performed.
  • inverse quantization may be performed on quantized transform coefficients.
  • the second-order inverse transform may be performed on the result generated by performing the inverse quantization.
  • the first-order inverse transform may be performed on the result generated by performing the second-order inverse transform.
  • a reconstructed residual signal may be generated by performing a first-order inverse transform on a result generated by performing a second-order inverse transform.
  • inverse mapping of a dynamic range may be performed before in-loop filtering.
  • the dynamic range may be divided into 16 equal pieces, and a mapping function for each piece may be signaled.
  • the mapping function may be signaled at the slice level or tile group level.
  • An inverse mapping function for performing inverse mapping may be derived based on the mapping function.
  • In-loop filtering, reference picture storage, and motion compensation may be performed in the inversely mapped region.
  • a prediction block generated through inter prediction may be converted into a mapped region by mapping using a mapping function, and the converted prediction block may be used to generate a reconstructed block.
  • a prediction block generated by intra prediction can be used to generate a reconstructed block without mapping and/or inverse mapping.
  • the residual block may be converted into an inversely mapped region by scaling the chroma component of the mapped region.
  • Whether scaling is available may be signaled at the slice level or tile group level.
  • scaling can be applied only when mapping for luma components is available, and partitioning of luma components and partitioning of chroma components follow the same tree structure.
  • Scaling may be performed based on an average of values of samples of a luma prediction block corresponding to the chroma prediction block.
  • the luma prediction block may mean a mapped luma prediction block.
  • a value required for scaling may be derived by referring to a look-up table using an index of a piece to which an average of values of samples of the luma prediction block belongs.
  • the residual block may be converted into an inversely mapped region. Then, for the chroma component block, reconstruction, intra prediction, inter prediction, in-loop filtering, and reference picture storage may be performed in the inversely mapped region.
  • mapping and/or inverse mapping of the luma component and chroma component is available may be signaled through a sequence parameter set.
  • a prediction block of the target block may be generated based on the block vector.
  • a block vector may indicate displacement between a target block and a reference block.
  • a reference block may be a block within a target image.
  • a prediction mode in which a prediction block is generated by referring to a target image may be referred to as an intra block copy (IBC) mode.
  • IBC intra block copy
  • IBC mode can be applied to CUs of a specified size.
  • IBC mode can be applied to MxN CUs.
  • M and N may be 64 or less.
  • the IBC mode may include a skip mode, a merge mode, and an AMVP mode.
  • a merge candidate list may be constructed, and a merge index may be signaled so that one merge candidate among merge candidates of the merge candidate list may be specified.
  • a block vector of a specified merge candidate may be used as a block vector of a target block.
  • a differential block vector may be signaled.
  • the predicted block vector may be derived from the left neighboring block and the top neighboring block of the target block.
  • an index of which neighboring block is to be used may be signaled.
  • the prediction block of the IBC mode may be included in the target CTU or the left CTU, and may be limited to a block within a pre-constructed region.
  • the value of the block vector may be limited so that the prediction block of the target block is located within a specified region.
  • the specified area may be an area of three 64x64 blocks that are encoded and/or decoded prior to the 64x64 block including the target block. As the value of the block vector is limited in this way, memory consumption and device complexity according to the implementation of the IBC mode can be reduced.
  • 16 is a structural diagram of an encoding device according to an embodiment.
  • the encoding device 1600 may correspond to the aforementioned encoding device 100.
  • the encoding device 1600 includes a processing unit 1610, a memory 1630, a user interface (UI) input device 1650, a UI output device 1660, and storage that communicate with each other through a bus 1690. (1640). Also, the encoding device 1600 may further include a communication unit 1620 connected to the network 1699.
  • the processor 1610 may be a semiconductor device that executes processing instructions stored in a central processing unit (CPU), memory 1630, or storage 1640.
  • the processing unit 1610 may be at least one hardware processor.
  • the processing unit 1610 may generate and process signals, data, or information input to the encoding device 1600, output from the encoding device 1600, or used inside the encoding device 1600, signals, Inspection, comparison, and judgment related to data or information can be performed. In other words, in an embodiment, data or information generation and processing, and data or information related inspection, comparison, and judgment may be performed by the processing unit 1610 .
  • the processing unit 1610 includes an inter prediction unit 110, an intra prediction unit 120, a switch 115, a subtractor 125, a transform unit 130, a quantization unit 140, an entropy encoding unit 150, and an inverse quantization unit. It may include a unit 160, an inverse transform unit 170, an adder 175, a filter unit 180, and a reference picture buffer 190.
  • Program modules may be included in the encoding device 1600 in the form of an operating system, application program modules, and other program modules.
  • Program modules may be physically stored on various known storage devices. Also, at least some of these program modules may be stored in a remote storage device capable of communicating with the encoding device 1600 .
  • Program modules include routines, subroutines, programs, objects, components, and data that perform functions or operations according to an embodiment or implement abstract data types according to an embodiment.
  • a data structure, etc. may be encompassed, but is not limited thereto.
  • Program modules may include instructions or codes executed by at least one processor of the encoding device 1600 .
  • the processing unit 1610 includes an inter prediction unit 110, an intra prediction unit 120, a switch 115, a subtractor 125, a transform unit 130, a quantization unit 140, an entropy encoding unit 150, and an inverse quantization unit. Instructions or codes of the unit 160, the inverse transform unit 170, the adder 175, the filter unit 180, and the reference picture buffer 190 may be executed.
  • Storage may represent memory 1630 and/or storage 1640 .
  • Memory 1630 and storage 1640 may be various forms of volatile or non-volatile storage media.
  • the memory 1630 may include at least one of a ROM 1631 and a RAM 1632 .
  • the storage unit may store data or information used for the operation of the encoding device 1600 .
  • data or information of the encoding device 1600 may be stored in a storage unit.
  • the storage unit may store pictures, blocks, lists, motion information, inter prediction information, and bitstreams.
  • the encoding device 1600 may be implemented in a computer system including a recording medium that can be read by a computer.
  • the recording medium may store at least one module required for the encoding device 1600 to operate.
  • the memory 1630 may store at least one module, and the at least one module may be configured to be executed by the processing unit 1610 .
  • a function related to communication of data or information of the encoding device 1600 may be performed through the communication unit 1620 .
  • the communication unit 1620 may transmit a bitstream to a decoding apparatus 1700 to be described later.
  • 17 is a structural diagram of a decryption device according to an embodiment.
  • the decoding device 1700 may correspond to the decoding device 200 described above.
  • the decryption device 1700 includes a processing unit 1710, a memory 1730, a user interface (UI) input device 1750, a UI output device 1760, and storage that communicate with each other through a bus 1790. (1740). Also, the decoding apparatus 1700 may further include a communication unit 1720 connected to the network 1799.
  • UI user interface
  • the processor 1710 may be a semiconductor device that executes processing instructions stored in a central processing unit (CPU), memory 1730, or storage 1740.
  • the processing unit 1710 may be at least one hardware processor.
  • the processing unit 1710 may generate and process signals, data, or information input to the decoding device 1700, output from the decoding device 1700, or used inside the decoding device 1700, and signals, Inspection, comparison, and judgment related to data or information can be performed. In other words, in an embodiment, data or information generation and processing, and data or information related inspection, comparison, and judgment may be performed by the processing unit 1710 .
  • the processing unit 1710 includes an entropy decoding unit 210, an inverse quantization unit 220, an inverse transform unit 230, an intra prediction unit 240, an inter prediction unit 250, a switch 245, an adder 255, a filter A unit 260 and a reference picture buffer 270 may be included.
  • An entropy decoding unit 210, an inverse quantization unit 220, an inverse transform unit 230, an intra prediction unit 240, an inter prediction unit 250, a switch 245, an adder 255, a filter unit 260, and At least some of the reference picture buffers 270 may be program modules, and may communicate with an external device or system.
  • Program modules may be included in the decryption device 1700 in the form of an operating system, application program modules, and other program modules.
  • Program modules may be physically stored on various known storage devices. Also, at least some of these program modules may be stored in a remote storage device capable of communicating with the decryption device 1700 .
  • Program modules include routines, subroutines, programs, objects, components, and data that perform functions or operations according to an embodiment or implement abstract data types according to an embodiment.
  • a data structure, etc. may be encompassed, but is not limited thereto.
  • the program modules may include instructions or codes executed by at least one processor of the decoding device 1700 .
  • the processing unit 1710 includes an entropy decoding unit 210, an inverse quantization unit 220, an inverse transform unit 230, an intra prediction unit 240, an inter prediction unit 250, a switch 245, an adder 255, a filter Instructions or codes of unit 260 and reference picture buffer 270 may be executed.
  • Storage may represent memory 1730 and/or storage 1740 .
  • Memory 1730 and storage 1740 may be various forms of volatile or non-volatile storage media.
  • the memory 1730 may include at least one of a ROM 1731 and a RAM 1732 .
  • the storage unit may store data or information used for the operation of the decoding device 1700 .
  • data or information of the decoding device 1700 may be stored in the storage unit.
  • the storage unit may store pictures, blocks, lists, motion information, inter prediction information, and bitstreams.
  • the decryption device 1700 may be implemented in a computer system including a recording medium that can be read by a computer.
  • the recording medium may store at least one module required for the decoding apparatus 1700 to operate.
  • the memory 1730 may store at least one module, and the at least one module may be configured to be executed by the processing unit 1710 .
  • a function related to communication of data or information of the decoding device 1700 may be performed through the communication unit 1720 .
  • the communication unit 1720 may receive a bitstream from the encoding device 1600.
  • the processing unit may represent the processing unit 1610 of the encoding device 1600 and/or the processing unit 1710 of the decoding device 1700.
  • the processing unit may represent switch 115 and/or switch 245.
  • the processing unit may represent the inter prediction unit 110, the subtractor 125, and the adder 175, and may represent the inter prediction unit 250 and the adder 255.
  • the processing unit may represent the intra prediction unit 120, the subtractor 125, and the adder 175, and may represent the intra prediction unit 240 and the adder 255.
  • the processing unit may represent the transform unit 130 and the inverse transform unit 170, and may represent the inverse transform unit 230.
  • the processing unit may represent the quantization unit 140 and the inverse quantization unit 160, and may represent the inverse quantization unit 220.
  • the processing unit may represent the entropy encoding unit 150 and/or the entropy decoding unit 210.
  • the processing unit may represent the filter unit 180 and/or the filter unit 260.
  • the processing unit may indicate the reference picture buffer 190 and/or the reference picture buffer 270 .
  • Video compression techniques based on deep neural networks can show superior performance than conventional video compression techniques.
  • video compression techniques based on deep neural networks may have the disadvantage of requiring a very large execution time compared to conventional video compression techniques due to the enormous amount of computation of deep neural networks.
  • a method, apparatus, and recording medium for efficiently lightening an in-loop filter and a predictive deep neural network may be provided in order to improve encoding efficiency of an image and reduce a time required for encoding.
  • a method, apparatus, and recording medium for providing learning in a lightweight in-loop filter and lightweight predictive deep neural network through a knowledge distillation technique may be provided. there is.
  • an in-loop filter for improving the efficiency and speed of encoding an image and a learning method in a predictive deep neural network may be provided.
  • Deep neural network and “neural network” may be used interchangeably. That is to say, “deep” is an optional expression and may be removed or omitted in embodiments.
  • neural network and “network” may be used interchangeably. That is to say, the term “neural network” and the term “network” may have the same meaning.
  • the terms “learning” and “training” may be used interchangeably. That is to say, the term “learning” and the term “training” can have the same meaning.
  • FIG. 18 is a flowchart of an encoding method according to an embodiment.
  • a deep neural network may include one or more of an in-loop filter and a predictive deep neural network.
  • the deep neural network may represent one or more of an in-loop filter based on the deep neural network and a predictive deep neural network for prediction.
  • the deep neural network may be replaced with an in-loop filter and/or predictive deep neural network.
  • step 1810 the processing unit 1610 of the encoding device 1600 may perform learning in the deep neural network.
  • processing unit 1610 may select a deep neural network.
  • the processing unit 1610 may perform encoding on deep neural network information.
  • the processing unit 1610 may generate encoded deep neural network information by encoding the deep neural network information.
  • the processing unit 1610 may generate a bitstream including deep neural network information or encoded deep neural network information.
  • the processing unit 1610 may store the bitstream in the storage unit 1630 .
  • the communication unit 1620 may transmit the bitstream to the decoding device 1700.
  • Deep neural network information or encoded deep neural network information may be signaled from the encoding device 1600 to the decoding device 1700 through a bitstream.
  • 19 is a flowchart of a decoding method according to an embodiment.
  • a deep neural network may include one or more of an in-loop filter and a predictive deep neural network.
  • the deep neural network may represent one or more of an in-loop filter based on the deep neural network and a predictive deep neural network for prediction.
  • the deep neural network may be replaced with an in-loop filter and/or predictive deep neural network.
  • the communication unit 1730 may receive a bitstream from the encoding device 1600.
  • step 1920 the processing unit 1710 of the decoding device 1700 may perform learning in the deep neural network.
  • Steps 1910 and 1920 may be performed concurrently. Alternatively, step 1920 may be performed before step 1910 .
  • the storage unit 1720 may store a bitstream.
  • the processing unit 1710 may obtain a bitstream from the storage unit 1720 or the communication unit 1730.
  • the bitstream may include deep neural network information or encoded deep neural network information.
  • the processor 1610 may decode the encoded deep neural network information.
  • the processing unit 1610 may generate deep neural network information by encoding the encoded deep neural network information.
  • processing unit 1610 may select a deep neural network.
  • 20 is a flowchart of learning in a deep neural network according to an example.
  • Step 1810 may include steps 2010, 2020 and 2030 below.
  • Step 1920 may include steps 2010, 2020 and 2030 below.
  • learning in a deep neural network may refer to learning in an in-loop filter and/or learning in a predictive deep neural network.
  • the processing unit may perform learning on a teacher network.
  • the processing unit may perform a first learning in the student network.
  • the first learning may refer to first stage learning.
  • the processing unit may perform a second learning in the student network.
  • the second learning may refer to second stage learning.
  • the size of the input of the in-loop filter may be equal to the size of the reconstructed image or reconstructed block (before in-loop filtering is applied).
  • Learning in the filter described in the embodiments may be learning under a specific condition or learning for a specific target.
  • the specific condition may be a condition in which the value of a coding parameter is a specific value.
  • a specific condition may be that a value of a quantization parameter is a specific value.
  • the specific value could be 22, 27, 32, 37 or 42.
  • a specific object may be an image or a component of a block.
  • a specific object may be a luma component and/or a chroma component.
  • a specific target may be a slice type.
  • slice types may include Intra A , Intra B , and Inter .
  • low-latency B and Random Access can utilize intra-slice filters trained using different quantization parameters.
  • different quantization parameters may be ⁇ 19, 24, 29, 34, 39 ⁇ .
  • the in-loop filter may restore a reconstructed image or a reconstructed block for each component of an input image.
  • 21 illustrates learning in a teacher network according to an example.
  • in-loop filter learning based on a deep neural network may be performed to improve the quality of an image in which compression deterioration has occurred.
  • learning in an in-loop filter based on a deep neural network may be performed as shown in FIG. 21 .
  • An in-loop filter may include a filtering deep neural network.
  • a filtering deep neural network may include a plurality of modules.
  • An input of the in-loop filter may include an image in which compression degradation occurs and coding information of the image.
  • An output of the in-loop filter may be an image from which compression degradation has been removed.
  • elimination of compression degradation may not imply complete elimination of compression degradation.
  • an image from which compression degradation is removed may be an image from which compression degradation is partially removed.
  • an image from which compression degradation has been removed may be an image generated by applying a specific process to an image in which compression degradation occurs.
  • an image from which compression degradation has been removed may be an image closer to an original image than an image in which compression degradation has occurred.
  • an image from which compression degradation has been removed may simply mean an output of an in-loop filter, and the in-loop filter may perform processing to remove or reduce compression degradation in the image.
  • an image may refer to a target image including a target block.
  • the coding information may include the aforementioned coding parameters.
  • the coding information may be information related to the target image or target block described in the embodiments.
  • the input of the in-loop filter may be an image in which compression degradation occurs.
  • the output of the in-loop filter may be an image from which compression degradation has been removed.
  • learning in an in-loop filter based on a deep neural network may be performed so that an error between an output of the in-loop filter and an original image is minimized.
  • an original image may refer to an original block in an original image corresponding to a target block.
  • learnings in a plurality of in-loop filters may be performed according to various degrees of compression degradation.
  • an in-loop filter based on a deep neural network to be trained may be an image reconstruction deep neural network.
  • the image reconstruction deep neural network may include a plurality of convolutional neural network layers or fully-connected neural network layers.
  • a plurality of neural network layers may be grouped into one module.
  • 22 shows learnings in an intra-prediction deep neural network and an inter-prediction deep neural network according to an embodiment.
  • a deep neural network can be a predictive deep neural network.
  • the predictive deep neural network may generate a prediction image (or prediction block) similar to the original image (or original block) through prediction.
  • step 2010 learning in the predictive deep neural network may be performed to generate a prediction image (or prediction block) similar to the original image (or original block) through intra prediction and/or inter prediction.
  • block may be replaced with the term “signal”.
  • prediction block may be replaced with the term “prediction signal”.
  • learning in a predictive deep neural network can be configured as shown in FIG. 22 .
  • the predictive deep neural network may include an intra-predictive deep neural network and/or an inter-predictive deep neural network.
  • the predictive deep neural network may be an intra-predictive deep neural network or an inter-predictive deep neural network.
  • An inter-prediction deep neural network may include a plurality of modules.
  • Inputs to the intra-prediction deep neural network may include neighboring samples of the target block and coding information of an image.
  • the surrounding samples may be plural.
  • the output of the intra-prediction deep neural network may be an improved prediction block of the target block.
  • Inputs to the inter-prediction deep neural network may include a prediction block of a target block and coding information of an image.
  • a prediction block of the target block may be input to a first module among a plurality of modules.
  • Coding information of an image may be input to each module of a plurality of modules.
  • Coding information of an image may be input to one or more modules among a plurality of modules.
  • the output of the inter-prediction deep neural network may be an improved prediction block of the target block.
  • an improved prediction block of the target block may be generated using the output of the intra-prediction deep neural network and the output of the inter-prediction deep neural network.
  • an input of the intra prediction deep neural network may include at least one of a neighboring sample of the target block and a prediction block of the target block.
  • the neighboring samples may be samples adjacent to the target block.
  • the peripheral sample may be the reference sample described above in the embodiments.
  • Neighboring samples may be samples that have already been encoded/decoded.
  • learning in the predictive deep neural network may be performed to minimize an error between the output of the predictive deep neural network and the original image.
  • an input of an inter-prediction deep neural network may be a prediction block of a current block.
  • a prediction block may be a block in a reference picture.
  • a prediction block may be determined through compensation using motion that minimizes rate-distortion loss. It may be a prediction block in .
  • a reference picture may be a frame that has already been encoded/decoded.
  • learning in a plurality of predictive deep neural networks may be performed according to various degrees of compression degradation.
  • a predictive deep neural network may include multiple convolutional neural network layers or fully-connected neural network layers.
  • a plurality of neural network layers may be grouped into one module.
  • the coding information of the image may be used for learning in the teacher network.
  • coding information of an image used in learning in a teacher network may include a coding parameter.
  • the coding information of an image includes 1) information on block division of the image, 2) information on the encoding mode of the target block, 3) information on the encoding mode of neighboring blocks, and 4) information on quantization of the target block. and 5) information on the number of times the in-loop filter is performed on the target block.
  • coding information of an image may be input to a teacher network.
  • the coding information of the image may be input to the teacher network together with the input image of the teacher network or neighboring samples.
  • coding information of an image may be input to a specific module of the teacher network.
  • coding information of the image may be input to a specific module together with an output of a previous module.
  • a previous module can be a module linked to a specific module.
  • the previous module may be a module located before a specific module.
  • coding information of an image may be converted according to a specific position corresponding to an output of a specific module of the teacher network, and the converted coding information may be input to the specific position.
  • encoding information of an image may be input to directly change model parameters within a module of a teacher network.
  • the trained deep neural network described in the embodiments can be utilized as a teacher network in knowledge distillation.
  • Additional information may be input to the above-described filtering deep neural network and prediction deep neural network.
  • the additional information may include a prediction image (or prediction block), a slice quantization parameter, a base quantization parameter, and a slice type. Also, the additional information may include coding parameters.
  • the student network may be constructed with a deep neural network structure similar to that of the teacher network.
  • the pre-trained teacher network may be a deep neural network with greater complexity compared to the student network on which learning is to be performed. Meanwhile, the complexity of the teacher network and the complexity of the student network may be the same. The difference between the complexity of the teacher network and the complexity of the student network may not be large.
  • the number of modules in the teacher network and the number of modules in the student network may be the same.
  • the structure of the modules of the teacher network and the structure of the modules of the student network may be the same or similar to each other.
  • the size of the feature map of the teacher network and the size of the feature map of the student network may be the same.
  • the size of the activity vector of the teacher network and the size of the activity vector of the student network may be the same.
  • Learning in a student network can be performed using knowledge distillation using a pre-trained teacher network.
  • information of the teacher network may be transferred to the student network, and learning may be performed in the student network using the transferred information.
  • the teacher network can be pre-trained.
  • the student network can be trained using Attention Transfer (AT) loss.
  • AT Attention Transfer
  • Intermediate features of the teacher network and the student network can be averaged along the channel axis to obtain one-channel attention maps.
  • a student network can be trained to mimic the attention map obtained from the teacher network.
  • features can be mapped to a map of states.
  • the degree of compression degradation of the input of the teacher network and the degree of compression degradation of the input of the student network may be the same.
  • the size of neighboring samples input to the teacher network and the size of neighboring samples input to the student network may be the same.
  • learning in an in-loop filter based on a deep neural network may be performed until knowledge distillation errors converge through knowledge distillation.
  • learning when learning using knowledge distillation is performed, additional learning may not be performed in the learned teacher network, and learning may be performed only in the student network.
  • learning may refer to learning within a student network.
  • learning when learning using knowledge distillation is performed, learning may be performed in both the teacher network and the student network together.
  • learning may refer to learning within a teacher network and learning within a student network.
  • learning may be performed such that values of corresponding model parameters in the deep neural network of the teacher network and the deep neural network of the student network are equal.
  • learning may be performed such that an error between the result of the teacher network and the result of the student network is minimized.
  • learning on student networks can be performed using intermediate feature maps or activation vectors.
  • the above intermediate feature map or activation vector may be an output of a specific module among a plurality of modules of the teacher network.
  • a first module in the teacher network and a second module in the student network may correspond to each other. Correspondence between the first module in the teacher network and the second module in the student network may be established. Learning may be performed so that outputs from modules corresponding to each other in the teacher network and the student network have characteristics similar to each other.
  • the outputs may include intermediate feature maps or activity vectors.
  • learning may be performed on the teacher network and the student network such that it is difficult for a separate discriminator deep neural network to distinguish between outputs from the teacher network and outputs from the student network.
  • the output may include a feature map, an active vector, or an output image.
  • the output image may be an image from which compression deterioration has been removed.
  • feature maps or active vectors may be extracted from modules at the same position among the entire structure of the teacher network and the entire structure of the student network. Learning may be performed using extracted feature maps or extracted learning vectors.
  • results from all modules of the teacher network and the student network may be used for learning.
  • training may be performed such that the complexity of the student network is minimized.
  • learning can be performed such that the complexity of the student network and the knowledge distillation error are minimized at the same time.
  • the information of the teacher network may include one or more of model parameters, intermediate feature maps, active vectors, and filtered images.
  • Intermediate feature maps and activation vectors can be functions of specific training images.
  • the learning image may be an input image input to the teacher network.
  • the filtered image may be the (final) output from the teacher network.
  • information of a plurality of teacher networks may be used for learning in the teacher network and the student network.
  • information of a specific teacher network among a plurality of teacher networks may be used for learning.
  • the specific teacher network may be a teacher network that generates an output having the smallest error with respect to the original video among a plurality of teacher networks.
  • the information of the teacher network may include coding information of a learning image.
  • Information of the teacher network including coding information of the learning image may be used for learning.
  • Learning using knowledge distillation in embodiments may include one or more of the approaches described above.
  • Equation 1 the calculation process of the attention module can be expressed as Equation 1 below.
  • F_out F_in * f ( Rec , Pred , BS , QP ) + F_in
  • F_in may be an input of an attention module.
  • F_out may be the output of the attention module.
  • Rec, Pred, BS, and QP may represent reconstructed images, predicted images, boundary strengths, and quantization parameters, respectively.
  • the quantization parameter may be a sequence-level input quantization parameter.
  • f may consist of two convolutions. The activation function can be applied after the first convolutional layer. The purpose of f may be to generate an attention map from external information, and then to perform recalibration on the feature map F_in .
  • Inputs to the neural network may include reconstructed images, predicted images, boundary strengths, and quantization parameters. (Split can additionally be taken as an input within intra-slice models.)
  • Whether or not filtering is to be applied may be determined for each slice or each block.
  • a conditional parameter may be additionally determined from the candidate list.
  • the candidate list may include three candidates derived from the quantization parameter.
  • q may represent a sequence level quantization parameter.
  • the candidate list may include conditional parameters ⁇ q, q-5, q-10 ⁇ . This selection process may be based on the rate-distortion cost of the encoding device 1600.
  • the deep neural network information may include information indicating whether to use the selection process and indexes for conditional eye parameters.
  • the index may indicate a conditional parameter used for filtering among conditional parameters.
  • the granularity of the decision of the filter and the choice of parameters may vary based on the resolution and quantization parameters. For example, given higher resolution and larger quantization parameters, decisions and selections can be performed over larger domains.
  • the candidate parameter list for the temporal layer may be different.
  • the third candidate q - 10 can be replaced by q + 5.
  • parameter selection can be disabled while on/off control is still maintained.
  • a scaling factor may be signaled for each color component in a picture header.
  • the difference between the input image and the image filtered by the neural network may be scaled by a scaling factor before the difference is added to the input image. This scaling may be termed residual scaling.
  • the input image used in residual scaling may be the output of a deblocking filter.
  • Equation 2 may represent a process of residual scaling.
  • R NN can be the output of a neural network filter.
  • R DB can be the output of the deblocking filter.
  • w may be a weight
  • R Refine can be the result of residual scaling.
  • FIG. 23 illustrates a first learning in a student network according to one embodiment.
  • step 2020 learning in a learning network using a teacher network may be performed for learning in an in-loop filter using knowledge distillation.
  • first learning in a student network using knowledge distillation may be performed according to the configuration shown in FIG. 23 .
  • feature maps that are results of a module of a teacher network and a module of a student network can be extracted.
  • Each feature map of the extracted feature maps may be converted into an attention map.
  • Attention maps may be created using the extracted feature maps.
  • an average may be derived for each axis of the feature map, and an attention map may be constructed using the derived average.
  • a sum may be derived for each axis of the feature map, and an attention map may be constructed using the derived sum.
  • the first feature map may be the result of a module of a teacher network.
  • the second feature map may be the result of a module of the student network.
  • the attention map may be constructed by transforming the feature map using a transform module having learnable parameters.
  • the first attention map may be an attention map obtained from a result of a module of the teacher network.
  • the second attention map may be an attention map obtained from the result of the module of the student network. Learning may be performed to minimize an error between the first attentional map and the second attentional map.
  • the learning loss function may be configured as in Equation 3 below.
  • L step1 may represent a loss function of learning.
  • the attention map may represent the attention map derived from the feature map output from the ith module of the teacher network.
  • the attention map may represent the attention map derived from the feature map output from the i-th module of the student network.
  • FIG. 24 illustrates a first learning in a student network according to one embodiment.
  • step 2020 for learning in a predictive deep neural network using knowledge distillation, learning in a learning network using a teacher network may be performed.
  • first learning in a student network using knowledge distillation may be performed according to the configuration shown in FIG. 24 .
  • activation vectors that are results of a module of a teacher network and a module of a student network can be extracted.
  • the first activation vector may be the result of a module of the teacher network.
  • the second activation vector may be the result of a module of the student network.
  • a first activation vector can be extracted from the module of the teacher network.
  • a second activation vector may be extracted from a module of the student network.
  • a transformation module for equalizing the dimensions of the first activation vector and the dimensions of the second activation vector may be used.
  • a plurality of transform modules may be added to the predictive deep neural network.
  • the transform module may change the dimension of the second active vector to be the same as the dimension of the first active vector by performing transform on the second active vector.
  • the transform module may generate a second active vector having a changed dimension by performing a transform on the second active vector. Parameters of the transformation module may be determined by learning.
  • learning may be performed such that an error between activation vectors from a module of a teacher network and a module of a student network is minimized.
  • the active vectors may be a first active vector and a second active vector.
  • the activity vectors can be a first activity vector and a second activity vector with altered dimensions.
  • the learning loss function may be configured as in Equation 4 below.
  • L step1 may represent a loss function of learning.
  • the teacher network may represent the first activation vector output from the i-th module of the teacher network.
  • the student network may be a second activation vector having a modified dimension generated by performing transformation on a second activation vector output from the i-th module of the student network.
  • 25 illustrates second learning in a student network according to one embodiment.
  • step 2030 second learning may proceed with respect to the student network that has completed the first learning.
  • 1) the original image or surrounding samples and 2) the output of the teacher network may be used.
  • fine-tune may be performed on the student network to which the first learning is applied.
  • the output of the teacher network and the original image can be used as a supervision signal for fine-tuning.
  • Both directors can be used to measure the L1 loss.
  • Each of the two directors can be named imitation loss (to the output of the teacher network) and reconstruction loss (to the original video) respectively.
  • Equation 5 The purpose of the second learning can be expressed as Equation 5 below.
  • L Finetune can be an L1 loss.
  • L Reconstuction may be the reconstruction loss.
  • L imitation may be imitation loss.
  • can be (empirically) set to 1.
  • the second learning of the student network may be performed according to the configuration shown in FIG. 25 .
  • training in an in-loop filter and predictive deep neural network may be performed such that the error between the output of the student network and the output of the teacher network is minimized.
  • the output of the student network may be a resulting image output from the student network.
  • the output of the teacher network may be a result image output from the teacher network.
  • training in an in-loop filter and predictive deep neural network may be performed so that an error between the output of the student network and the original image is minimized.
  • the first error may be an error between the output of the student network and the output of the teacher network.
  • the second error may be an error between the output of the student network and the original image.
  • weights of the first error and the second error in learning may be adjusted by a Lagrange multiplier.
  • the value of the Lagrange multiplier may be set to 1.
  • learning in the in-loop filter and prediction deep neural network may be performed until the first error and the second error converge.
  • the learning loss function may be configured as in Equation 6 below.
  • L step2 may represent a loss function of learning.
  • the first loss function of the first learning in the student network and the second loss function of the second learning in the student network can be summed as one third loss function.
  • Learning in the in-loop filter and predictive deep neural network may be performed using the third loss function.
  • a deep neural network may be selected. Selecting a deep neural network may mean performing filtering on a target image using the selected deep neural network.
  • the target image may be a reconstructed image.
  • the target image may mean a target block or a reconstructed block.
  • filtering of the target image may be performed using a filter based on the learned deep neural network.
  • the deep neural network filter may be a trained deep neural network described in the embodiments.
  • the deep neural network filter may be a filter using a trained deep neural network described in the embodiments.
  • a filter may refer to a deep neural network filter.
  • the N filters may be N deep neural network filters.
  • the N deep neural network filters may each perform image reconstructions of different degrees. That is to say, the N deep neural network filters may differ from each other in the degree of image reconstruction.
  • the N deep neural network filters may be student networks trained using knowledge distillation.
  • a filter is applied to a specific target may be determined.
  • a specific object may be a unit including a target block.
  • a specific object is a (reconstructed) picture, (reconstructed) picture, (reconstructed) slice, (reconstructed) Coding Tree Unit (CTU), or (reconstructed) may be a block.
  • the size of the (reconstructed) block may be HxW.
  • N, H and W may be positive integers.
  • one filter out of N filters may be selected.
  • a filter candidate set may be configured according to the compression degree and compression method for a specific object (or image).
  • a filter candidate set may be some filters among N filters. In the filter selection described in the embodiments, one filter may be selected from among the filters of the filter candidate set.
  • a filter candidate set may be M deep neural network filters. M may be an integer less than or equal to N. The M deep neural network filters of the filter candidate set may each perform image reconstructions of different degrees.

Landscapes

  • Engineering & Computer Science (AREA)
  • Multimedia (AREA)
  • Signal Processing (AREA)
  • Physics & Mathematics (AREA)
  • General Physics & Mathematics (AREA)
  • Theoretical Computer Science (AREA)
  • Compression Or Coding Systems Of Tv Signals (AREA)

Abstract

L'invention concerne un procédé, un dispositif et un support d'enregistrement pour le codage/décodage d'image. Des modes de réalisation concernent un procédé, un dispositif et un support d'enregistrement pour alléger efficacement un filtre en boucle et un réseau neuronal profond prédictif afin d'améliorer une efficacité de codage d'une image et de raccourcir le temps nécessaire au codage. Des modes de réalisation concernent l'apprentissage dans un filtre en boucle léger et un réseau neuronal profond prédictif léger par distillation de connaissance. Selon des modes de réalisation, un procédé d'apprentissage dans un filtre en boucle et un réseau neuronal profond prédictif est fourni afin d'améliorer l'efficacité et la vitesse de codage d'image et de raccourcir le temps nécessaire au codage.
PCT/KR2022/014710 2021-09-29 2022-09-29 Procédé, dispositif et support d'enregistrement pour le codage/décodage d'image WO2023055153A1 (fr)

Applications Claiming Priority (4)

Application Number Priority Date Filing Date Title
KR20210129157 2021-09-29
KR10-2021-0129157 2021-09-29
KR10-2022-0124695 2022-09-29
KR1020220124695A KR20230046269A (ko) 2021-09-29 2022-09-29 영상 부호화/복호화를 위한 방법, 장치 및 기록 매체

Publications (1)

Publication Number Publication Date
WO2023055153A1 true WO2023055153A1 (fr) 2023-04-06

Family

ID=85783226

Family Applications (1)

Application Number Title Priority Date Filing Date
PCT/KR2022/014710 WO2023055153A1 (fr) 2021-09-29 2022-09-29 Procédé, dispositif et support d'enregistrement pour le codage/décodage d'image

Country Status (1)

Country Link
WO (1) WO2023055153A1 (fr)

Citations (5)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
WO2017222140A1 (fr) * 2016-06-24 2017-12-28 한국과학기술원 Procédés et dispositifs de codage et de décodage comprenant un filtre en boucle à base de cnn
KR20200045128A (ko) * 2018-10-22 2020-05-04 삼성전자주식회사 모델 학습 방법 및 장치, 및 데이터 인식 방법
US20200145661A1 (en) * 2017-07-06 2020-05-07 Samsung Electronics Co., Ltd. Method for encoding/decoding image, and device therefor
KR20200109904A (ko) * 2019-03-15 2020-09-23 (주)인시그널 Dnn 기반 이미지 또는 비디오 코딩을 위한 시스템 및 방법
KR20210071886A (ko) * 2020-06-09 2021-06-16 베이징 바이두 넷컴 사이언스 앤 테크놀로지 코., 엘티디. 모델의 증류 방법, 장치, 전자기기 및 저장매체

Patent Citations (5)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
WO2017222140A1 (fr) * 2016-06-24 2017-12-28 한국과학기술원 Procédés et dispositifs de codage et de décodage comprenant un filtre en boucle à base de cnn
US20200145661A1 (en) * 2017-07-06 2020-05-07 Samsung Electronics Co., Ltd. Method for encoding/decoding image, and device therefor
KR20200045128A (ko) * 2018-10-22 2020-05-04 삼성전자주식회사 모델 학습 방법 및 장치, 및 데이터 인식 방법
KR20200109904A (ko) * 2019-03-15 2020-09-23 (주)인시그널 Dnn 기반 이미지 또는 비디오 코딩을 위한 시스템 및 방법
KR20210071886A (ko) * 2020-06-09 2021-06-16 베이징 바이두 넷컴 사이언스 앤 테크놀로지 코., 엘티디. 모델의 증류 방법, 장치, 전자기기 및 저장매체

Similar Documents

Publication Publication Date Title
WO2019112394A1 (fr) Procédé et appareil de codage et décodage utilisant un partage d'informations sélectif entre des canaux
WO2019190224A1 (fr) Dispositif et procédé de codage/décodage d'image, et support d'enregistrement mémorisant un flux binaire
WO2018026148A1 (fr) Procédé et dispositif de codage/décodage d'images, et support d'enregistrement stockant un flux binaire
WO2019107927A1 (fr) Procédé et appareil de prédiction intra bidirectionnelle
WO2021015581A1 (fr) Procédé, appareil et support d'enregistrement pour coder/décoder une image à l'aide d'un partitionnement géométrique
WO2020184977A1 (fr) Procédé et dispositif de codage/décodage sur la base d'une copie intra-bloc et support de stockage de flux binaire
WO2020256528A1 (fr) Procédé et dispositif de codage/décodage vidéo, et support d'informations de flux binaire
WO2021054805A1 (fr) Procédé et dispositif de codage/décodage d'image, et support d'enregistrement stockant un flux binaire
WO2020050600A1 (fr) Procédé et dispositif de codage/décodage vidéo, et support d'enregistrement pour stockage de flux binaire
WO2020256422A1 (fr) Procédé et dispositif de codage/décodage d'informations à prédiction inter
WO2018174618A1 (fr) Procédé et dispositif de prédiction à l'aide d'un bloc de référence
WO2020256522A1 (fr) Procédé et appareil de codage d'image et de décodage d'image à l'aide d'une segmentation de zone
WO2021112651A1 (fr) Procédé et dispositif permettant de coder/décoder une image au moyen d'un mode palette, et support d'enregistrement
WO2021112652A1 (fr) Procédé, appareil et support d'enregistrement pour codage/décodage d'image différentielle basée sur une zone
WO2022019613A1 (fr) Procédé, appareil et support d'enregistrement pour coder/décoder une image à l'aide d'un partitionnement géométrique
WO2019074273A1 (fr) Procédé et dispositif utilisant des informations d'inter-prédiction
WO2020256377A1 (fr) Procédé et dispositif de codage/décodage d'images basés sur une division en sous-blocs
WO2020130710A1 (fr) Dispositif et procédé de codage/décodage d'image, et support d'enregistrement sur lequel un flux binaire est stocké
WO2019117645A1 (fr) Procédé et dispositif de codage et de décodage d'image utilisant un réseau de prédiction
WO2020256495A1 (fr) Procédé, dispositif et support d'enregistrement pour coder/décoder une image à l'aide d'une représentation de référence
WO2020256324A1 (fr) Procédé et appareil de codage/décodage vidéo et support d'enregistrement contenant en mémoire un flux binaire
WO2020209671A1 (fr) Procédé et dispositif de signalisation d'un signal lié à un mode de prédiction pour prédiction intra
WO2019147067A1 (fr) Méthode et appareil de codage et de décodage d'image à l'aide d'informations de mouvement temporelles
WO2021029720A1 (fr) Procédé, appareil et support d'enregistrement permettant de coder/décoder une image en faisant appel au partitionnement
WO2020256405A1 (fr) Procédé de codage/décodage sur la base d'une copie intra-bloc, dispositif et support de stockage de flux binaire

Legal Events

Date Code Title Description
121 Ep: the epo has been informed by wipo that ep was designated in this application

Ref document number: 22876925

Country of ref document: EP

Kind code of ref document: A1