WO2015120823A1 - Procédé et dispositif de compression d'image utilisant un espace de stockage de pixel de référence en de multiples formes - Google Patents

Procédé et dispositif de compression d'image utilisant un espace de stockage de pixel de référence en de multiples formes Download PDF

Info

Publication number
WO2015120823A1
WO2015120823A1 PCT/CN2015/073204 CN2015073204W WO2015120823A1 WO 2015120823 A1 WO2015120823 A1 WO 2015120823A1 CN 2015073204 W CN2015073204 W CN 2015073204W WO 2015120823 A1 WO2015120823 A1 WO 2015120823A1
Authority
WO
WIPO (PCT)
Prior art keywords
decoding
reference pixel
matching
module
pixel sample
Prior art date
Application number
PCT/CN2015/073204
Other languages
English (en)
Chinese (zh)
Inventor
林涛
Original Assignee
同济大学
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Application filed by 同济大学 filed Critical 同济大学
Publication of WO2015120823A1 publication Critical patent/WO2015120823A1/fr

Links

Images

Classifications

    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04NPICTORIAL COMMUNICATION, e.g. TELEVISION
    • H04N19/00Methods or arrangements for coding, decoding, compressing or decompressing digital video signals
    • H04N19/50Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using predictive coding
    • H04N19/503Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using predictive coding involving temporal prediction
    • H04N19/51Motion estimation or motion compensation
    • H04N19/573Motion compensation with multiple frame prediction using two or more reference frames in a given prediction direction
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04NPICTORIAL COMMUNICATION, e.g. TELEVISION
    • H04N19/00Methods or arrangements for coding, decoding, compressing or decompressing digital video signals
    • H04N19/10Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using adaptive coding
    • H04N19/102Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using adaptive coding characterised by the element, parameter or selection affected or controlled by the adaptive coding
    • H04N19/103Selection of coding mode or of prediction mode
    • H04N19/105Selection of the reference unit for prediction within a chosen coding or prediction mode, e.g. adaptive choice of position and number of pixels used for prediction
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04NPICTORIAL COMMUNICATION, e.g. TELEVISION
    • H04N19/00Methods or arrangements for coding, decoding, compressing or decompressing digital video signals
    • H04N19/10Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using adaptive coding
    • H04N19/134Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using adaptive coding characterised by the element, parameter or criterion affecting or controlling the adaptive coding
    • H04N19/157Assigned coding mode, i.e. the coding mode being predefined or preselected to be further used for selection of another element or parameter
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04NPICTORIAL COMMUNICATION, e.g. TELEVISION
    • H04N19/00Methods or arrangements for coding, decoding, compressing or decompressing digital video signals
    • H04N19/10Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using adaptive coding
    • H04N19/169Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using adaptive coding characterised by the coding unit, i.e. the structural portion or semantic portion of the video signal being the object or the subject of the adaptive coding
    • H04N19/17Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using adaptive coding characterised by the coding unit, i.e. the structural portion or semantic portion of the video signal being the object or the subject of the adaptive coding the unit being an image region, e.g. an object
    • H04N19/176Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using adaptive coding characterised by the coding unit, i.e. the structural portion or semantic portion of the video signal being the object or the subject of the adaptive coding the unit being an image region, e.g. an object the region being a block, e.g. a macroblock
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04NPICTORIAL COMMUNICATION, e.g. TELEVISION
    • H04N19/00Methods or arrangements for coding, decoding, compressing or decompressing digital video signals
    • H04N19/42Methods or arrangements for coding, decoding, compressing or decompressing digital video signals characterised by implementation details or hardware specially adapted for video compression or decompression, e.g. dedicated software implementation
    • H04N19/43Hardware specially adapted for motion estimation or compensation
    • H04N19/433Hardware specially adapted for motion estimation or compensation characterised by techniques for memory access
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04NPICTORIAL COMMUNICATION, e.g. TELEVISION
    • H04N19/00Methods or arrangements for coding, decoding, compressing or decompressing digital video signals
    • H04N19/50Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using predictive coding
    • H04N19/503Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using predictive coding involving temporal prediction

Definitions

  • the present invention relates to a digital video compression encoding and decoding system, and more particularly to a method and apparatus for encoding and decoding composite images and video containing computer screen images.
  • a digital video signal of an image is a sequence of images.
  • a frame of image is usually a rectangular area composed of several pixels, and a digital video signal is a sequence of video images composed of tens of frames to thousands of frames of images, sometimes simply referred to as a video sequence or sequence.
  • Encoding a digital video signal encodes a frame by frame image.
  • coding one frame of image in almost all international standards for video image encoding such as MPEG-1/2/4, H.264/AVC and the latest international video compression standard HEVC (High Efficiency Video Coding)
  • a frame image is divided into sub-images of a plurality of blocks of MxM pixels, it is called a "Coding Unit (CU)", and the sub-images are coded one by one with CU as a basic coding unit.
  • the size of the commonly used M is 8, 16, 32, 64. Therefore, encoding a video image sequence is to sequentially encode each coding unit of each frame.
  • decoding each coding unit of each frame is sequentially decoded in the same order, and finally the entire video image sequence is reconstructed.
  • each CU in one frame of image can be different, some are 8x8, some are 64x64, and so on.
  • LCUs Large Coding Units
  • NxN pixels the same size and having NxN pixels
  • each LCU is further divided into trees.
  • CTU Coding Tree Unit
  • One of the LCUs is composed of three 32x32 pixel CUs and four 16x16 pixel CUs, so that seven tree-structured CUs form a CTU.
  • the other LCU consists of two 32x32 pixel CUs, three 16x16 pixel CUs, and 20 8x8 pixel CUs.
  • Such 25 CUs in a tree structure constitute another CTU.
  • Encoding one frame of image is to sequentially encode one CU in one CTU. At any one time, the CU being coded is referred to as the current coded CU.
  • Decoding one frame of image also sequentially decodes one CU in one CTU in the same order.
  • the CU being decoded is referred to as the currently decoded CU.
  • the current encoded CU or the currently decoded CU is typically the current CU.
  • one CU is usually further divided into smaller sub-areas.
  • the sub-regions include, but are not limited to, a prediction unit (PU), a transform unit (TU), an asymmetric partition (AMP) region, a macroblock, a block, a microblock, and a strip (width or height is one pixel or one pixel component) Area), a variable-sized rectangular area, a variable-sized pixel string (segment) or a pixel component string (segment) or a pixel index string (segment).
  • Encoding (and corresponding decoding) a CU is to encode one sub-region (and corresponding decoding).
  • a sub-area is called a coding sub-area
  • a sub-area is called a decoding sub-area.
  • the coding sub-area and the decoding sub-area are collectively referred to as a codec sub-area.
  • the sub-regions particularly prediction units, transform units, asymmetrically divided regions, macroblocks, blocks, microblocks, strips
  • the coding sub-region and the decoding sub-region are often referred to as coding blocks and decoding blocks, respectively, in many cases, collectively referred to as codec blocks.
  • a color pixel consists of three components.
  • the two most commonly used pixel color formats are the GBR color format consisting of a green component, a blue component, and a red component, and a generic term called YUV consisting of a luma component and two chroma components.
  • Color formats such as YCbCr color format. Therefore, when encoding a sub-area, a sub-area can be divided into three component planes (G plane, B plane, R plane or Y plane, U plane, V plane), and the three component planes are respectively coded;
  • the three components of one pixel can be bundled into one 3-tuple, and the sub-regions composed of these 3-tuples are encoded as a whole.
  • the arrangement of the former pixel and its components is called the planar format of the image (and its sub-regions), and the arrangement of the latter pixel and its components is called the overlay format of the image (and its sub-regions). (packed format).
  • a planar format is arranged. First arrange all WxH G components of a frame image (or a CU) with a width of W pixels of H pixels, then arrange all WxH B components, and finally arrange all WxH R components:
  • a stacking format is arranged by first arranging the G components of the first pixel, then arranging the B component and the R component, then arranging the G component, the B component, and the R component of the second pixel, and so on, and finally arranging.
  • palette index representation Another common representation of pixels and their components is the palette index form.
  • the value of a component of a pixel is represented by the index of the palette.
  • the palette space stores the value or approximate value of the three components of the pixel that needs to be represented.
  • the address of the palette is called the index of the pixel (called the palette color) stored in this address.
  • An index can represent one component of a pixel, and an index can also represent three components of a pixel.
  • the palette can be one or more. In the case of multiple palettes, a complete index is actually composed of the palette number and the index of the numbered palette.
  • the index representation of a pixel and its components is an index to represent this pixel.
  • the index representation of a pixel is also referred to in the prior art as an indexed color or a pseudo color representation of a pixel, or is often referred to directly as an indexed pixel or a pseudo pixel (pseudo pixel). ) or pixel index or cable lead. Indexes are sometimes referred to as indices. Representing a pixel in its indexed representation is also referred to as indexing or exponentialization.
  • the palette index representation of a pixel and its components is a color-based aggregation and sorting format of the pixel and its components.
  • CMYK color format CMYK color format
  • grayscale color format CMYK color format
  • the index representation of a pixel is also a color format of a pixel, called an index color format or a pseudo color format.
  • the YUV color format can be subdivided into several seed formats according to whether the chroma component is downsampled: a YUV 4:4:4 pixel color format consisting of 1 Y component, 1 U component, and 1 V component.
  • the left and right adjacent pixels are composed of two Y components, one U component, and one V component in a YUV 4:2:2 pixel color format; four pixels arranged in a left and right adjacent position by 2x2 spatial position are composed of four pixels.
  • YUV4: 2:0 pixel color format consisting of Y component, 1 U component, and 1 V component.
  • a component is generally represented by a number of 8 to 16 bits.
  • the YUV4:2:2 pixel color format and the YUV4:2:0 pixel color format are all downsampled for the YUV4:4:4 pixel color format.
  • a pixel component is also referred to as a pixel sample or simply as a sample.
  • a sample can be an 8-bit number, ie one sample occupies one byte.
  • a sample can also be a 10-bit number or a 12-bit number or a 14-bit number or a 16-bit number.
  • reconstructed pixels are generated, which are further divided into different degrees of partially reconstructed pixels generated in the encoding or decoding process and fully reconstructed pixels generated after the encoding or decoding process is completed. . If the fully reconstructed pixel samples have equal values to the original input pixel samples prior to encoding, the encoding and decoding process that is experienced is referred to as lossless encoding and decoding. If the fully reconstructed pixel samples have unequal values from the original input pixel samples prior to encoding, the encoding and decoding process that is experienced is referred to as lossy encoding and decoding.
  • the generated reconstructed pixel samples are usually saved as historical data and used as reference pixel samples for subsequent sub-region encoding or decoding.
  • the storage space in which the reconstructed pixel history data is saved is referred to as a reference pixel sample storage space.
  • the reference pixel sample storage space is limited and only a portion of the historical data can be saved.
  • the historical data in the reference pixel sample storage space may also include reconstructed pixel samples of the reconstructed sub-region of the current CU.
  • a notable feature of computer screen images is that there are often many similar or even identical pixel patterns within the same frame of image.
  • Chinese or foreign text that often appears in computer screen images is composed of a few basic strokes, and many similar or identical strokes can be found in the same frame image.
  • Menus, icons, etc. which are common in computer screen images, also have many similar or identical patterns.
  • the intra prediction method used in the existing image and video compression technology only refers to adjacent pixel samples, and cannot improve the compression efficiency by using the similarity or the similarity in one frame image.
  • the intra motion compensation method in the prior art is also called intra block copy mode, and uses several fixed size (8x8, 16x16, 32x32, 64x64 pixels) blocks for intraframe.
  • Intra block matching coding does not achieve a finer match with a variety of different sizes and shapes.
  • the microblock matching method, the fine division matching method, the string matching method, and the palette matching method also called the palette index matching method or the index matching method.
  • the palette matching method also called the palette index matching method or the index matching method.
  • the historical data of the encoded (or decoded) reconstructed pixel samples must be used as the reference pixel samples.
  • the larger the reference pixel sample storage space the better the efficiency of prediction and matching.
  • the reference pixel sample storage space can generally accommodate historical data of a few frames or even tens of pixels (such as palette pixels) and dozens of frames.
  • the historical data of the reconstructed pixel is represented in a single form and placed in a single reference pixel sample storage space module. Take a frame of 640x480 pixels of YUV4:4:4 image and each pixel sample occupies one byte.
  • the most common form of historical data is in a storage space module with a linear (1D) address.
  • the linear placement order is a planar format and each plane is divided into equally sized blocks, each block having WxH pixel samples (ie, one component of the pixel), first arranged in blocks in a certain linear order.
  • Each prediction method or method of matching has its specific and most effective form of historical data.
  • Historical data and reference pixel sample storage space will greatly reduce coding efficiency. Therefore, multiple forms of historical data and reference pixel sample storage space must be used to improve prediction or matching performance and reduce the bandwidth required to read and write historical data.
  • the main technical feature of the present invention is that when encoding or decoding a plurality of prediction and matching modes in combination with a current coding unit, various forms of historical data and reference pixel sample storage space are used accordingly.
  • pixel sample storage space usually consists of four elements:
  • pixels and their component arrangement formats such as stack format and plane format
  • Pixel color format such as GBR color format, RGB color format, YCbCr color format, YUV color format, CMYK color format, YIQ color format, HSV color format, HSL color format, index color format, etc., different color formats Usually can be converted to each other;
  • Pixel aggregation and sorting format For example, pixels are aggregated and ordered according to their position coordinates: the historical data of a pixel whose shape is a 2-dimensional area or its components is aggregated in a 2-dimensional array space, or aggregated in a 1-dimensional array space. In the case of 1 dimension, the linear placement of the historical data in a storage space having a linear (1 dimensional) address is also determined; a typical linear placement is to first divide the 2-dimensional region into equal sizes.
  • each block having WxH pixels (in the case of a stack format) or one component thereof (in the case of a planar format), and then placing the blocks in a determined linear order (eg row order or column order or depth)
  • the 4-point tree order for D is placed in front of the linear address storage space, and the pixels inside each block or one of its components are scanned according to a certain linearity (for example: line scan or column scan or zigzag scan or Z-scan or bow scan or 4-point tree scan with depth D is arranged in 1D data and placed in linear address storage space; for example, pixels are clustered and sorted according to their colors: color palette color, texture color
  • the pixels in the palette are sorted according to one of their characteristics to form a reference pixel sample storage space with a linear address; two examples of the characteristics: 1) the palette color appears in a CU Sequence, 2) the frequency of occurrence of colors in a color palette in the CU;
  • Downsampling format of chroma components such as YUV4:4:4 format or YUV4:2:2 format or YUV4:2:0 format.
  • the 4 elements determine the form of the reference pixel sample storage space. One difference between these four elements is the form of a different reference pixel sample storage space.
  • the most basic characteristic feature is to determine A (2 ⁇ A ⁇ 5) kinds of predictive coding modes and matching coding modes, and determine B (2 ⁇ B ⁇ A ⁇ 5) of historical data.
  • A (2 ⁇ A ⁇ 5) kinds of predictive coding modes and matching coding modes
  • B (2 ⁇ B ⁇ A ⁇ 5) of historical data.
  • Different forms of reference pixel sample storage space, and class A predictive coding mode and matching coding mode are classified into class B, and respectively correspond to B reference pixel sample storage spaces.
  • each predictive coding mode or matching coding mode performs predictive coding or matching coding according to the corresponding reference pixel sample storage space according to its classification, and selects A prediction according to a predetermined evaluation criterion.
  • One of the coding mode and the matching coding mode encodes the current sub-area. Regardless of which prediction encoding mode and matching encoding method are selected, the current sub-region is encoded, and the generated new historical data is converted into the main reference pixel sample storage space and placed in the main reference pixel sample storage space. It may also be converted to other reference pixel sample storage spaces (which may be in another level of partial reconstruction) and placed in other reference pixel sample storage spaces. Historical data in different reference pixel sample storage spaces needs to be synchronized.
  • the most basic characteristic feature is to determine A (2 ⁇ A ⁇ 5) kinds of predictive decoding mode and copy decoding mode, and determine B (2 ⁇ B ⁇ A ⁇ 5) of historical data.
  • A (2 ⁇ A ⁇ 5) kinds of predictive decoding mode and copy decoding mode
  • B (2 ⁇ B ⁇ A ⁇ 5) of historical data.
  • Different types of reference pixel sample storage spaces, and class A prediction decoding mode and matching decoding mode are classified into class B, and respectively correspond to B reference pixel sample storage spaces.
  • B different forms of reference pixel sample storage space may have different spatial sizes to store different amounts of historical data, wherein the one storing the most historical data is called the primary reference pixel sample storage space, and the other reference pixel samples are
  • the historical data in the storage space is a subset of the historical data of the primary reference pixel sample storage space, but has different representations and may be in different degrees of partial reconstruction (such as having been color clustered or matched) The residual or the matching residual has not been added).
  • each of the prediction decoding mode or the copy decoding mode performs prediction decoding or copy decoding according to the corresponding reference pixel sample storage space according to the classification.
  • the information read from the video stream data or the information read from the video stream data plus the characteristics of the current decoding sub-region and the neighboring sub-regions one of the A prediction decoding modes and the copy decoding mode is selected. Decode the current subregion. Regardless of which prediction decoding mode and copy decoding mode are selected to decode the current sub-area, the generated new historical data is converted into the main reference pixel sample storage space and placed in the main reference pixel sample storage space. It may also be converted to other reference pixel sample storage spaces (which may be in another level of partial reconstruction) and placed in other reference pixel sample storage spaces. Historical data in different reference pixel sample storage spaces needs to be synchronized.
  • an image encoding method or apparatus comprising at least one of the steps or modules for performing the following functions and operations:
  • the video code stream includes at least decoding corresponding decoding in the decoding method and the device.
  • an image decoding method or apparatus comprising at least one of the steps or modules for performing the following functions and operations:
  • the prediction and matching coding module is configured to be a plurality of coding processing units using different coding modes, wherein the coding processing unit is configured to perform data using one of intra prediction coding, inter prediction coding, or a plurality of different matching coding modes. deal with;
  • An encoding storage module configured to store a plurality of storage units of different forms of reconstructed pixel sample data
  • An encoding data control module configured to control the prediction and matching encoding module to read data of a corresponding one or more storage units in the code storage module that is preset; and to control the prediction according to the read data
  • the matching coding module hands over the data to be encoded to the corresponding coding processing unit; and controls the prediction and matching coding mode
  • the block writes output data of the encoding processing unit to one or more storage units of the code storage module that are preset.
  • the prediction and copy decoding module is configured to be a plurality of decoding processing units using different decoding modes, wherein the decoding processing unit is configured to perform data using one of intra prediction decoding, inter prediction decoding, or a plurality of different copy decoding modes. deal with;
  • Decoding the storage module configured to store a plurality of storage units of different forms of reconstructed pixel sample data
  • a decoding data control module configured to control the prediction and copy decoding module to read data of a corresponding one or more storage units in the preset decoding storage module; and to control the prediction according to the read data
  • the copy decoding module hands over the data to be decoded to the corresponding decoding processing unit; and controls the prediction and copy decoding module to write the output data of the decoding processing unit to one or more storage units in the preset decoding storage module.
  • Predictive coding mode module matching coding mode 1 module, matching coding mode 2 module, ..., matching coding mode A-1 module: A is a positive integer satisfying 2 ⁇ A ⁇ 5; Using input prediction (intra prediction or inter prediction or both) coding mode, matching coding mode 1, matching coding mode 2, ..., matching coding mode A-1, an input coding unit of the input video image Performing a prediction or matching coding operation on a current coding sub-region; the output of the prediction coding mode module is a prediction mode, an inter-predicted motion vector, and a prediction residual, that is, an original input pixel sample and prediction of the input video image a difference between the pixel samples; the output of each of the matching coding mode modules is a matching mode, a matching position, an unmatched sample, and a matching residual; the matching position is used to represent the currently encoded pixel in the current coding sub-region a variable of a position of the matched reference pixel sample in the reference pixel sample storage space, the reference image matching the
  • the unmatched samples are found in the reference pixel sample storage space according to a predetermined matching criterion, if a matching matching criterion is determined Very loose, allowing arbitrarily large matching errors, so that a match can always be found, then there is no unmatched sample as an output in this matching coding mode; the matching residual is the original input pixel sample and the matching pixel The difference between the samples, if a matching matching criterion is an absolutely accurate lossless matching, the matching residual is zero, that is, the matching encoding method has no matching residual as an output, if a matching If the matching criterion determined by the coding mode is an approximate lossy match, the matching residual may not be zero, and another case of lossy matching is to first perform sample quantization, color quantization or based on the original input pixel sample.
  • Pre-processing of pixel clustering of colors and then performing matching encoding operations.
  • sample quantization, color quantization, or color-based pixel clustering is lossy, even if the matching encoding operation itself is lossless,
  • the matching residual ie, the difference between the original input pixel sample and the matched pixel sample
  • the matching residual may also be non-zero;
  • the reference pixel sample storage space stores the reconstructed pixel sample history data generated in the encoding process in a unique mutually different form, and is used as a reference pixel sample required for subsequent encoding and reconstruction of various operations;
  • Different forms of reference pixel sample storage space may have different spatial sizes to store different amounts of historical data, wherein the one storing the most historical data is called the primary reference pixel sample storage space, and the other reference pixel sample storage spaces.
  • the historical data in the data is a subset of the historical data of the primary reference pixel sample storage space, but has different representations and may be in different degrees of partial reconstruction (such as having been color clustered or added matching residuals) Or no matching residuals are added; corresponding to the B reference pixel sample storage spaces, the predictive coding mode module, the matching coding mode 1 module, and the matching code Mode 2 module, ..., matching coding mode A-1 module
  • a coding mode is also divided into B class correspondingly, corresponding to B reference pixel sample storage space; in the current coding sub-area
  • each predictive coding mode or matching coding mode uses the corresponding reference pixel sample storage space for predictive coding or matching coding according to its classification, and selects A predictive coding mode and matching coding mode according to predetermined evaluation criteria.
  • the form conversion module of the pixel sample storage space converts one form of the form of the reference pixel sample storage space into another form; the reconstructed pixel generated and output by the module 3) passes through the module. Converted into the form of the main reference pixel sample storage space and placed in the module 2) the main reference pixel sample storage space, or may be converted into other reference pixel sample storage space by the module (may be at another level Partial reconstruction stage) and put into the other reference pixel sample storage space of module 2), used as reference pixel samples needed for subsequent encoding and reconstruction of various operations; this module makes different reference pixel samples Historical data in the storage space is synchronized with each other.
  • An embodiment of the decoding apparatus of the present invention is composed of the following modules:
  • Code stream data parsing and partial decoding module performing entropy decoding on the video stream of the currently decoded CU including the prediction mode, the motion vector, the matching mode, the matching position, the unmatched sample compressed data, and all other syntax element compressed data.
  • parsing and performing partial decoding such as transform decoding, prediction and compensation, ie inverse operation of residual operation, DPCM decoding, first-order and high-order differential decoding, mapping Parameters such as prediction mode, motion vector, matching mode, matching position, and unmatched samples obtained by decoding, run-length decoding, and index decoding are sent to each prediction decoding mode module and matching decoding mode module;
  • Other syntax elements such as prediction residuals and matching residual entropy decoded output data (ie, the result of entropy decoding) are sent to the remaining common techniques of decoding and reconstruction modules; in particular, based on parsing from video stream data.
  • the information is based on the information parsed from the video stream data plus the current decoding sub-region and the neighboring sub-region. Selecting a corresponding prediction decoding mode or matching decoding mode, and sending several parameters and variables corresponding to the prediction mode, the motion vector, the matching mode, the matching position, the unmatched sample, and the like to a corresponding prediction decoding mode module or matching Decoding mode module, starting the corresponding predictive decoding mode module or matching decoding mode module to decode the current decoding sub-region;
  • Predictive decoding mode module Predictive decoding mode module, matching decoding mode 1 module, matching decoding mode 2 module, ..., matching decoding mode A-1 module: A is a positive integer satisfying 2 ⁇ A ⁇ 5; the A decoding mode modules respectively Using the prediction (intra prediction or inter prediction or both) decoding mode, matching decoding mode 1, matching decoding mode 2, ..., matching decoding mode A-1, the A decoding mode corresponds to the corresponding A current decoding The region performs prediction or matching decoding operation; the input of the prediction decoding mode module is a motion vector of a prediction mode and an inter prediction; the input of the matching decoding mode module is a matching mode, a matching position, and possibly an unmatched sample.
  • the matching position is used to indicate from the reference
  • the matching reference pixel samples are copied and pasted to the position of the currently decoded pixel sample (referred to as the matched pixel sample) of the current decoded sub-region, the matched reference pixel
  • the sample values are referred to as matching pixel samples (obviously, the matched pixel samples are replicas of matching pixel samples, both of which are numerically equal), and the positions of the matching pixel samples are not necessarily in the reference
  • One area connected together in the pixel sample storage space may also be several areas separated in the reference pixel sample storage space; the unmatched samples are directly parsed and decoded from the video stream data.
  • the unmatched sample is typically not present in the reference pixel sample storage space;
  • the A decoding mode module The output is a predicted pixel sample or the matched pixel sample (which is numerically equal to the matched pixel sample) plus the unmatched sample (present in some matching decoding modes); Pixel sample values are matched and the match is not possible to sample all together constitute the complete data matching the decoded output of the current decoding CU;
  • the reference pixel sample storage space stores the reconstructed pixel sample history data generated in the decoding process in a unique mutually different form, and is used as a reference pixel sample required for subsequent decoding and reconstruction of various operations;
  • Different forms of reference pixel sample storage space may have different spatial sizes to store different amounts of historical data, wherein the one storing the most historical data is called the primary reference pixel sample storage space, and the other reference pixel sample storage spaces.
  • the historical data in the data is a subset of the historical data of the primary reference pixel sample storage space, but has different representations and may be in different stages of partial reconstruction (such as matching residuals or no matching residuals)
  • the A decoding modes of the decoding mode A-1 module are also divided into B classes correspondingly, corresponding to the B reference pixel sample storage spaces; when decoding the current decoding sub-region, each prediction decoding The mode or the matching decoding mode performs prediction decoding or matching decoding according to the corresponding reference pixel sample storage space according to its classification, and each reference pixel sample storage space is all decoding modes corresponding to the decoding mode (possibly There are a variety of) providing reference pixel samples;
  • the remaining common techniques of decoding and reconstructing modules performing the remaining common techniques on the current decoding sub-region, such as inverse transform, inverse quantization, compensation corresponding to prediction residuals and matching residuals (ie, taking Inverse operation of residual operation), prediction and compensation (ie inverse operation of residual operation), DPCM, first-order and high-order difference, mapping, run, index, deblocking filtering, sample adaptive compensation (Sample Adaptive Offset), decoding and reconstruction operations;
  • the input to this module is all other syntax elements of module 1) output such as entropy decoded output data for prediction residuals and matching residuals and modules 2)
  • the output of the block is the predicted pixel sample or the matched pixel sample plus the unmatched sample that may exist;
  • the output of the module is a reconstructed pixel (including fully reconstructed pixels and varying degrees of partial reconstruction) Pixels); the reconstructed pixels are converted into a form of a primary reference pixel sample storage space and placed in the module 3) the primary reference pixel sample storage space, and
  • Form conversion module of pixel sample storage space convert one form of the form of B reference pixel sample storage space into another form; module 4) reconstructed pixels generated and output must pass through this module Converted into the form of the main reference pixel sample storage space and placed in the module 3) the main reference pixel sample storage space, may also be converted into other reference pixel sample storage space by the module (may be in another degree Partial reconstruction stage) and put into the other reference pixel sample storage space of module 3), used as reference pixel samples needed for subsequent decoding and reconstruction of various operations; this module makes different reference pixel samples Historical data in the storage space is synchronized with each other.
  • the positive integer A is equal to 4, and the four coding mode modules are a prediction coding mode module, a block matching coding mode module, a string matching coding mode module, and a palette matching coding mode module; the positive integer B is equal to 2,
  • the two reference pixel sample storage spaces are reference pixel storage spaces in the form of a planar format and a two-dimensional array space (referred to as a planar two-dimensional reference space) and a reference to a form of a stacked format and a one-dimensional array space.
  • a pixel sample storage space (referred to as a stacked 1-dimensional reference space);
  • the primary reference pixel sample storage space is the planar 2-dimensional reference space, and historical data in the stacked 1-dimensional reference space are all A subset of the historical data of a planar 2-dimensional reference space, but with different representations and possibly at different degrees of partial reconstruction (eg if a matching residual has been added or a matching residual has not been added); the 4 encodings
  • the method is also divided into two categories correspondingly, and the prediction coding mode and the block matching coding mode are both the first type coding mode, corresponding to the planar two-dimensional reference space, and the string matching coding mode and the color palette.
  • Encoding with a second coding method are based, and the The stacked packet corresponds to a 1-dimensional reference space; when encoding the current coding sub-region, the predictive coding mode uses the planar 2-dimensional reference space for predictive coding, and the block-matching coding mode also uses the planar 2-dimensional reference space.
  • Block matching coding, the string matching coding method uses the overlapping packet 1D reference space for string matching coding, and the palette matching coding mode also uses the overlapping packet 1D reference space for palette matching coding;
  • the dimension reference space provides reference pixel samples for predictive coding and block matching encoding, and the stacked 1-dimensional reference space provides reference pixel samples for string matching encoding and palette matching encoding;
  • a block-matching coding scheme is a block of a certain size (such as 64x64 samples, 32x32 samples, 16x16 samples, 8x8 samples, 8x4 samples, 4x8 samples, 4x4 samples, etc.) (referred to as a matched block, which is Matching encoding is performed in units of a 2-dimensional coordinate, the matching pixel samples form a matching block in the planar 2-dimensional reference space, and the position within one frame of the image is also available.
  • a two-dimensional coordinate representation and thus in the block matching coding mode, the matching position can be represented by a difference between a 2-dimensional coordinate of the matching block and a 2-dimensional coordinate of the matched block, which is called a displacement vector;
  • the string matching coding method arranges the pixel samples of the current coding sub-region into a 1-dimensional pixel sample string in a stack format according to a predetermined linear scan format, and uses a variable-length pixel sample string (referred to as being matched).
  • the string whose position can be represented by a 2D coordinate or a linear address, is matched and encoded.
  • the matched pixel sample forms a matching string in the 1D-dimensional reference space of the stack, and the position is also available.
  • a 2-dimensional coordinate can also be represented by a linear address. Therefore, in the string matching coding mode, the matching position can use the difference between the 2-dimensional coordinates of the matching string and the 2-dimensional coordinates of the matched string, and the linear address of the matching string can also be used.
  • the difference between the linear addresses of the matched strings is generally referred to as the displacement vector. Since the length of the matching string (equal to the length of the matched string) is variable, another variable called the matching length and the displacement vector are needed. Together, that is, (displacement vector, matching length), the matching position is completely represented; the result of performing string matching encoding on the current coding CU is M (M ⁇ 1) matching strings and N (N) ⁇ 0) unmatched pixel samples, output M pairs (displacement vector, matching length) and N unmatched pixel samples;
  • the palette matching encoding uses only a part of the pixels of the stacked 1-dimensional reference space (generally containing a number of pixel samples in the current CU) as reference pixels, and thus a predetermined method from the stacked 1-dimensional reference space Pick out and update a set of K pixels (usually 4 ⁇ K ⁇ 64). These K pixels form a palette. Each pixel in the palette is represented by an index. Using the pixels of the palette as reference pixels, and the matching position of the matching pixel samples is an index of the matching pixel samples in a palette, all of the matching pixels of the current encoding CU The index of the sample value constitutes an index array as an index map;
  • One of the functions of the remaining common techniques of coding and reconstruction modules is to perform various commonly used transformations and predictions on parameters and variables such as matching patterns, displacement vectors, matching lengths, index mappings, unmatched samples, and the like. And find the residual, DPCM, first-order and high-order difference, mapping, run, index coding;
  • the form conversion module of the pixel sample storage space performs mutual conversion between the following two expression forms of the pixel: 1) a format of a planar format and a 2-dimensional array space, 2) a form of a stacked packet format and a 1-dimensional array space;
  • the palette matching encoding mode module and the color palette are optional and may be omitted, so that in the encoding device, the positive integer A is equal to 3, and the three encoding mode modules are predictions respectively.
  • the string matching coding mode module is optional and may be omitted. Therefore, in the coding apparatus, the positive integer A is equal to 3, and the three coding mode modules are respectively a prediction coding mode module and a block matching coding mode. Module, palette matching encoding mode module.
  • the positive integer A is equal to 4, and the four decoding mode modules are a prediction decoding mode module, a block matching decoding mode module, a string matching decoding mode module, and a palette matching decoding mode module; the positive integer B is equal to 2,
  • the two reference pixel sample storage spaces are reference pixel storage spaces in the form of a planar format and a two-dimensional array space (referred to as a planar two-dimensional reference space) and a reference to a form of a stacked format and a one-dimensional array space.
  • a pixel sample storage space (referred to as a stacked 1-dimensional reference space);
  • the primary reference pixel sample storage space is the planar 2-dimensional reference space, and historical data in the stacked 1-dimensional reference space are all A subset of the historical data of the planar 2-dimensional reference space, but with different representations and possibly at different degrees of partial reconstruction (eg if a matching residual has been added or a matching residual has not been added);
  • the 4 decodings The mode is also divided into two categories correspondingly, and the prediction decoding mode and the block matching decoding mode are both the first type decoding mode, corresponding to the planar 2-dimensional reference space, and the string matching decoding mode and the color palette.
  • the decoding mode is a type 2 decoding mode corresponding to the stacked packet 1-dimensional reference space; when decoding the current decoding CU, the prediction decoding mode uses the planar 2-dimensional reference space for prediction decoding, and the block
  • the matching decoding method also uses the planar 2-dimensional reference space for block matching decoding, and the string matching decoding method uses the stacked packet 1-dimensional reference space for string matching decoding, and the palette matching decoding method also uses the stacked packet 1D.
  • the reference space performs palette matching decoding; the planar 2-dimensional reference space provides reference pixel samples for predictive decoding and block matching decoding, and the stacked 1-dimensional reference space provides for string matching decoding and palette matching decoding. Reference pixel sample;
  • a block-matching decoding method is a block of a certain size (such as 64x64 samples, 32x32 samples, 16x16 samples, 8x8 samples, 8x4 samples, 4x8 samples, 4x4 samples, etc.) (referred to as a matched block, which is Matching decoding is performed in units of a 2-dimensional coordinate, which forms a matching block in the planar 2-dimensional reference space, and the position within one frame of the image is also available.
  • a two-dimensional coordinate representation and thus in the block matching decoding mode, the matching position can be represented by a difference between a 2-dimensional coordinate of the matching block and a 2-dimensional coordinate of the matched block, which is called a displacement vector;
  • the string matching decoding method arranges the pixel samples of the currently decoded CU into a 1-dimensional pixel sample string in a stack format according to a predetermined linear scan format, and uses a variable length pixel sample string (referred to as a matched string). , the position of which can be represented by a 2D coordinate or a linear address.
  • the matching pixel sample forms a matching string in the 1st dimension reference space of the stack, and the position is also available.
  • the 2-dimensional coordinates can also be represented by a linear address. Therefore, in the string matching decoding mode, the matching position can use both the difference between the 2-dimensional coordinates of the matching string and the 2-dimensional coordinates of the matched string, and the linear address of the matching string can also be used.
  • the difference between the linear addresses of the matching strings is generally referred to as the displacement vector.
  • the length of the matching string is variable, another variable called the matching length is required along with the displacement vector. , ie (displacement vector, matching length), to completely represent the matching position; the input when performing string matching decoding on the current decoding CU is parsed from the video stream data And the matching pattern obtained by decoding, M (M ⁇ 1) pairs (displacement vector, matching length) and N (N ⁇ 0) unmatched pixel samples;
  • Palette matching decoding uses only a portion of the pixels of the stacked 1-dimensional reference space (typically containing a number of pixel samples within the current CU) as reference pixels, thus following a predetermined method from the stacked 1-dimensional reference space Pick out and update a set of K pixels (usually 4 ⁇ K ⁇ 64). These K pixels form a palette. Each pixel in the palette is represented by an index. The palette matches the decoding method.
  • the matching position of the matching pixel samples is an index of the matching pixel samples in a palette
  • all of the matching pixels of the currently decoded CU The index of the sample form an index array as an index map; the index map is parsed and decoded from the video stream data, and is one of the input of the palette matching decoding;
  • One of the functions of the code stream data parsing and partial decoding module is to represent parameters such as a matching pattern, a displacement vector, a matching length, an index map, an unmatched sample, and the like obtained by entropy decoding from the video code stream.
  • the partial decoded data of the syntax element performs various commonly used transform decoding, prediction and compensation, that is, inverse operation of residual operation, DPCM decoding, first-order and high-order differential decoding, mapping decoding, run-length decoding, and index decoding.
  • the operation obtains the original matching mode, the displacement vector, the matching length, the index mapping, the unmatched sample and other parameters and variables, and is used as a block matching decoding mode module, a string matching decoding mode module, and a palette matching decoding mode module input;
  • the form conversion module of the pixel sample storage space performs mutual conversion between the following two expression forms of the pixel: 1) a format of a planar format and a 2-dimensional array space, 2) a form of a stacked packet format and a 1-dimensional array space;
  • the palette matching decoding mode module and the color palette are optional and may be omitted, so that in the decoding device, the positive integer A is equal to 3, and the three decoding mode modules are predictions respectively.
  • the string matching decoding mode module is optional and may be omitted, so that in the decoding device, the positive integer A is equal to 3, and the three decoding mode modules are respectively a prediction decoding mode module, and a block matching decoding mode. Module, palette matching decoding mode module.
  • the coding unit in the video stream ie the CU part, consists of syntax elements loaded with the following information:
  • CU header prediction and matching mode identification code, prediction mode or matching mode, motion vector 1 or matching position 1, unmatched pixel sample 1, motion vector 2 or matching position 2, unmatched pixel sample 2, ......... ..., more motion vectors or matching positions, more unmatched pixel samples, prediction residuals or matching residuals, other coding results;
  • the order in which all other of the syntax elements are placed in the code stream is not unique, and any predetermined predetermined order may be employed; any one of the syntax elements may also be split into several parts. The parts may be placed centrally in the same place in the code stream, or may be placed in different places in the code stream; any number of syntax elements may also be combined into one syntax element; in addition to the CU header syntax elements and prediction and matching The mode identifier code syntax element, and other syntax elements may not exist in the compressed code stream data of a certain CU;
  • the prediction and matching mode identification code may take the following code values and have the following semantics:
  • the current encoding or decoding CU (referred to as the current CU) adopts predictive coding or decoding 1
  • the current encoding or decoding CU (referred to as the current CU) adopts matching encoding or decoding mode 1 2
  • the current encoding or decoding CU (referred to as the current CU) adopts matching encoding or decoding mode 2 ... ...2015........ A-1
  • the current encoding or decoding CU (referred to as the current CU for short) adopts matching encoding or decoding mode A-1
  • the value ranges from 1 to B, and corresponds to the B reference pixel sample storage spaces, respectively.
  • the code value of the current CU prediction and matching mode identification code and the A-ary array ClassOfCoding[A] can be followed.
  • the following judgment conditions determine which reference pixel sample storage space the current CU uses for encoding or decoding:
  • the reference pixel sample storage space j is used.
  • Embodiment 1 of a video code stream including a prediction and matching mode identification code and other coding results, and a class 2 coding or decoding mode
  • the coding unit in the video stream ie the CU part, consists of syntax elements loaded with the following information:
  • CU header prediction and matching mode identification code, prediction mode or matching mode, motion vector 1 or displacement vector 1 or (displacement vector 1, matching length 1) or index map 1, unmatched pixel sample 1, motion vector 2 or displacement Vector 2 or (displacement vector 2, matching length 2) or index map 2, unmatched pixel samples 2, .........., more motion vectors or displacement vectors or (displacement vectors, matching lengths) or index mapping, more Multiple unmatched pixel samples, prediction residuals or matching residuals, other coding results;
  • the order in which all other of the syntax elements are placed in the code stream is not unique, and any predetermined predetermined order may be employed; any one of the syntax elements may also be split into several parts. The parts may be placed centrally in the same place in the code stream, or may be placed in different places in the code stream; any number of syntax elements may also be combined into one syntax element; in addition to the CU header syntax elements and prediction and matching The mode identifier code syntax element, and other syntax elements may not exist in the video code stream data of a certain CU;
  • the prediction and matching mode identification code may take the following code values and have the following semantics:
  • the current encoding or decoding CU (referred to as the current CU) adopts predictive coding or decoding 1
  • the current encoding or decoding CU (referred to as the current CU) adopts block matching encoding or decoding mode.
  • the current encoding or decoding CU uses string matching encoding or decoding.
  • the current encoding or decoding CU (referred to as the current CU) adopts palette matching encoding or decoding mode.
  • the division of the four coding modes or the four decoding modes is defined by a predetermined 4-element array ClassOfCoding[4], and each ClassOfCoding[i] (where i satisfies 0 ⁇ i ⁇ 4) Yes:
  • the values 1 and 2 correspond to the reference pixel sample storage space 1 (ie, the planar 2-dimensional reference space) and the reference pixel sample storage space 2 (the stacked 1-dimensional reference space), respectively, thereby, from the current CU
  • the code value of the prediction and matching mode identification code and the 4-ary array ClassOfCoding[4] can determine which reference pixel sample storage space the current CU uses for encoding or decoding according to the following judgment conditions:
  • the current CU is encoded or decoded by using a planar 2-dimensional reference space
  • the current CU is encoded or decoded by using the planar 2-dimensional reference space;
  • the current CU is encoded or decoded by using the overlapping 1D reference space;
  • the current CU is encoded or decoded using the superimposed 1-dimensional reference space.
  • Embodiment 2 of a video code stream including a prediction and matching mode identification code and other coding results, and a class 2 coding or decoding mode
  • the coding unit in the video stream ie the CU part, consists of syntax elements loaded with the following information:
  • CU header prediction and matching mode identification code, prediction mode or matching mode, motion vector 1 or displacement vector 1 or (displacement vector 1, matching length 1), unmatched pixel sample 1, motion vector 2 or displacement vector 2 or ( Displacement vector 2, matching length 2), unmatched pixel samples 2, .........., more motion vectors or displacement vectors Quantity or (displacement vector, matching length), more unmatched pixel samples, prediction residuals or matching residuals, other coding results;
  • the order in which all other of the syntax elements are placed in the code stream is not unique, and any predetermined predetermined order may be employed; any one of the syntax elements may also be split into several parts. The parts may be placed centrally in the same place in the code stream, or may be placed in different places in the code stream; any number of syntax elements may also be combined into one syntax element; in addition to the CU header syntax elements and prediction and matching The mode identifier code syntax element, and other syntax elements may not exist in the video code stream data of a certain CU;
  • the prediction and matching mode identification code may take the following code values and have the following semantics:
  • the current encoding or decoding CU (referred to as the current CU) adopts predictive coding or decoding 1
  • the current encoding or decoding CU (referred to as the current CU) adopts block matching encoding or decoding mode.
  • the current encoding or decoding CU (referred to as the current CU) uses string matching encoding or decoding.
  • the division of the three coding modes or the three decoding modes is defined by a predetermined 3-element array ClassOfCoding[3], and each ClassOfCoding[i] (where i satisfies 0 ⁇ i ⁇ 3) Yes:
  • the values 1 and 2 correspond to the reference pixel sample storage space 1 (ie, the planar 2-dimensional reference space) and the reference pixel sample storage space 2 (the stacked 1-dimensional reference space), respectively, thereby, from the current CU
  • the code value of the prediction and matching mode identification code and the 3-ary array ClassOfCoding[3] can determine which reference pixel sample storage space the current CU uses for encoding or decoding according to the following judgment conditions:
  • the current CU is encoded or decoded by using a planar 2-dimensional reference space
  • the current CU is encoded or decoded by using the planar 2-dimensional reference space;
  • the current CU is encoded or decoded using the superimposed 1-dimensional reference space.
  • the coding unit in the video stream ie the CU part, consists of syntax elements loaded with the following information:
  • CU header prediction and matching mode identification code, prediction mode or matching mode, motion vector 1 or displacement vector 1 or index map 1, unmatched pixel sample 1, motion vector 2 or displacement vector 2 or index map 2, unmatched pixel Sample 2, ........., more motion vectors or displacement vectors or index maps, more unmatched pixel samples, prediction residuals or matching residuals, other coding results;
  • the order in which all other of the syntax elements are placed in the code stream is not unique, and any predetermined predetermined order may be employed; any one of the syntax elements may also be split into several parts. The parts may be placed centrally in the same place in the code stream, or may be placed in different places in the code stream; any number of syntax elements may also be combined into one syntax element; in addition to the CU header syntax elements and prediction and matching The mode identifier code syntax element, and other syntax elements may not exist in the video code stream data of a certain CU;
  • the prediction and matching mode identification code may take the following code values and have the following semantics:
  • the current encoding or decoding CU (referred to as the current CU) adopts predictive coding or decoding 1
  • the current encoding or decoding CU (referred to as the current CU) adopts block matching encoding or decoding mode.
  • the current encoding or decoding CU (referred to as the current CU) adopts palette matching encoding or decoding mode.
  • the division of the three coding modes or the three decoding modes is defined by a predetermined 3-element array ClassOfCoding[3], and each ClassOfCoding[i] (where i satisfies 0 ⁇ i ⁇ 3) Yes:
  • the values 1 and 2 correspond to the reference pixel sample storage space 1 (ie, the planar 2-dimensional reference space) and the reference pixel sample storage space 2 (the stacked 1-dimensional reference space), respectively, thereby, from the current CU
  • the code value of the prediction and matching mode identification code and the 3-ary array ClassOfCoding[3] can determine which reference pixel sample storage space the current CU uses for encoding or decoding according to the following judgment conditions:
  • the current CU is encoded or decoded by using a planar 2-dimensional reference space
  • the current CU is encoded or decoded by using the planar 2-dimensional reference space;
  • the current CU is encoded or decoded using the superimposed 1-dimensional reference space.
  • the reference pixel sample storage space in the form of the planar format and the 2-dimensional array space is composed of pixel samples of two frames of images placed in the following order;
  • the Y component of the first row of pixels of the L-1 frame image is placed in the scanning order from left to right.
  • the Y component of the second row of pixels of the L-1 frame image is placed in the scanning order from left to right.
  • the Y component of the last row of pixels of the L-1 frame image is placed in the scanning order from left to right.
  • the U component of the first row of pixels of the L-1 frame image is placed in the scanning order from left to right.
  • the U component of the second row of pixels of the L-1 frame image is placed in the scanning order from left to right.
  • the U component of the last row of pixels of the L-1 frame image is placed in the scanning order from left to right.
  • the V component of the first row of pixels of the L-1 frame image is placed in the scanning order from left to right.
  • the V component of the second row of pixels of the L-1 frame image is placed in the scanning order from left to right.
  • the V component of the last row of pixels of the L-1 frame image is placed in the scanning order from left to right.
  • the Y component of the second row of pixels of the Lth frame image is placed in the scanning order from left to right.
  • the U component of the first row of pixels of the Lth frame image is placed in the scanning order from left to right.
  • the U component of the second row of pixels of the Lth frame image is placed in the scanning order from left to right.
  • the V component of the first row of pixels of the Lth frame image is placed in the scanning order from left to right.
  • the V component of the second row of pixels of the Lth frame image is placed in the scanning order from left to right.
  • the V component of the last row of pixels of the Lth frame image is placed in the scanning order from left to right.
  • the above Lth frame image is the frame image in the current encoding or decoding
  • the L-1 frame image is the frame image before the Lth frame image (in the encoding or decoding order).
  • the reference pixel sample storage space in the form of the stack format and the 1-dimensional array space is composed of pixel samples of 1 frame image, and the 1 frame image is first divided into 64 ⁇ 64 pixels (overlapping format, 64 ⁇ 64 ⁇ 3 pixels)
  • the block of the sample) gives each block a sequence number according to the raster scan order of the block, that is, the sequence number of the block in the upper left corner is 1, the block number on the right side of the block numbered 1 is 2, and the block number 2 is The sequence number of the block on the right is 3, and so on. If the block with the sequence number X is the rightmost block and there is no block on the right side, the sequence of the leftmost block directly below the block with the sequence number X is X+1.
  • the block on the right side of the block with the sequence number X+1 has the sequence number X+2, the block number on the right side of the block with the sequence number X+2 is X+3, and so on, all the blocks of the image of one frame are There is a sequence number; then, all pixels are placed and arranged in the following order to form a 1-dimensional array space:
  • the first column of the block of sequence number 1 is arranged in pixels from top to bottom in one column, and arranged in Y, U, and V in one pixel.
  • the second column of the block of sequence number 1 is arranged in pixels from top to bottom in one column, and arranged in Y, U, and V in one pixel.
  • the 64th column of the block of sequence number 1 is arranged in pixels from top to bottom in one column, and arranged in Y, U, and V in one pixel.
  • the first column of the block of the number 2 is arranged in pixels from top to bottom in one column, and arranged in Y, U, and V in one pixel.
  • the second column of the block with the number 2 is arranged in pixels from top to bottom in one column, and arranged in Y, U, and V in one pixel.
  • the 64th column of the block with the number 2 is arranged in pixels from top to bottom in one column, and arranged in Y, U, and V in one pixel.
  • the first column of pixels of the largest block the columns are arranged in pixels from top to bottom, and the pixels are arranged in Y, U, and V.
  • the second column of pixels with the largest serial number arranged in pixels from top to bottom in one column, and arranged in Y, U, and V in one pixel.
  • the 64th column pixel of the largest block of the largest number is arranged in pixels from top to bottom in one column, and is arranged in Y, U, and V in one pixel.
  • the blocks are arranged in a row order and the pixels in the block are arranged in a column scan format.
  • Change the row order in this embodiment to another order such as column order or 4-point tree order with depth D
  • change the column scan format to other scan formats such as line scan or zigzag scan or Z scan or depth D. More examples can be obtained by 4-tree scanning and other formats.
  • the size of the block is 64 x 64 pixels. More embodiments can be obtained by changing the size of the block, such as 32x32 or 16x16 or 8x8 or 8x4 or 4x8 or 4x4 pixels.
  • the reference pixel sample storage space in the form of the stack format and the 1-dimensional array space is composed of pixel samples of 1 frame image, and the 1 frame image is first divided into 64 ⁇ 64 pixels (overlapping format, 64 ⁇ 64 ⁇ 3 pixels)
  • the block of the sample (abbreviated as 64x64 block) gives each block a sequence number according to the raster scan order of the block, that is, the sequence number of the block in the upper left corner is 1, and the block number on the right side of the block numbered 1 is 2. On the right side of the block numbered 2 The serial number of the block is 3, and so on.
  • the block with the sequence number X is the rightmost block and there is no block on the right side
  • the sequence of the leftmost block directly below the block with the sequence number X is X+1
  • the serial number is The block of the right side of the block of X+1 has the serial number X+2
  • the block of the right side of the block of the sequence number X+2 has the serial number of X+3, and so on, all the blocks of the image of one frame have A sequence number; then, divide a 64x64 block into sub-blocks of different depths according to the 4-point tree method and give each sub-block a sequence number; the division method is as follows:
  • the division with depth 1 is to divide 64x64 blocks into 32x32 sub-blocks with up to the left and right and 4 depths, and the numbers of the sub-blocks with depth 1 being upper left, upper right, lower left, and lower right are called the depth 1 serial number. 1, 2, 3, 4, as shown in Figure 5;
  • the subdivision with depth of 2 is divided into four sub-blocks with a depth of 1 and divided into four 16x16 sub-blocks with a depth of 2, and the sub-blocks with a depth of 1 are divided into upper left, upper right, lower left, and lower right.
  • the sub-blocks of four sub-blocks of depth 2 (called depth 2 numbers) are 1, 2, 3, 4, and the sub-blocks of depth 1 and number 2 are divided into upper left, upper right, lower left, and lower right.
  • the sub-blocks of 2 are called 5, 6, 7, 8, and the sub-blocks of depth 1 and 3 are divided into upper sub-tops, upper right, lower left, and lower right sub-blocks of depth 2
  • the serial number (called the depth 2 serial number) is 9, 10, 11, 12, and the sub-blocks with the depth 1 and the number 4 are divided into the upper left, upper right, lower left, and lower right sub-blocks with a depth of 2 (called The depth 2 serial number is 13, 14, 15, 16 respectively, so that there are 16 sub-blocks with a depth of 2, as shown in FIG. 6;
  • the subdivision with a depth of 3 is divided into 16 sub-blocks of depth 2 and divided into 8 x 8 sub-blocks of 3 depths and 3 depths, and the sub-blocks whose depth 2 is k (1 ⁇ k ⁇ 16) are divided into upper left.
  • the upper, lower left, and lower right sub-blocks of depth 3 are called 4x(k-1)+1, 4x(k-1)+2, 4x(k-1). +3,4x(k-1)+4, so that there are 64 sub-blocks with a depth of 3, as shown in Figure 7;
  • the partition with a depth of 4 is divided into 64 sub-blocks with a depth of 3 and divided into 4 x 4 sub-blocks with 4 depths up and down, and a sub-block whose depth 3 is k (1 ⁇ k ⁇ 64).
  • the upper, lower left, and lower right sub-blocks of depth 4 (called depth 4 numbers) are 4x(k-1)+1, 4x(k-1)+2, 4x(k-1), respectively. +3,4x(k-1)+4, so that there are 256 sub-blocks with a depth of 4, as shown in Figure 8;
  • the first column of the sub-block of depth D is number 1, and the pixel values in one column are arranged in a predetermined order.
  • the second column of the sub-block of depth D is number 1, and the pixel values in one column are arranged in a predetermined order.
  • the last column of the sub-block with the depth D number is 1, and the pixel values in the 1 column are arranged in a predetermined order.
  • the first column of the sub-block of depth D is number 2, and the pixel values in one column are arranged in a predetermined order.
  • the second column of the sub-block of depth D is number 2, and the pixel values in one column are arranged in a predetermined order.
  • the last column of the sub-block whose depth D is number 2, and the pixel values in the 1 column are arranged in a predetermined order.
  • the first column of pixels of the sub-block with the largest depth D number, and the pixel values in one column are arranged in a predetermined order
  • the second column of the sub-block with the largest depth D number, the pixel values in one column are arranged in a predetermined order
  • the last column pixel of the sub-block with the largest depth D number, and the pixel values in one column are arranged in a predetermined order.
  • the size of the block is 64 x 64 pixels. More embodiments can be obtained by changing the size of the block, such as 32x32 or 16x16 or 8x8 or 8x4 or 4x8 or 4x4 pixels.
  • the analyzing and evaluating the characteristics of the codec sub-region and/or the neighboring region in the codec method or device to select and determine a codec mode of the codec sub-region and a corresponding reference pixel sample storage space thereof includes at least one of the following or a combination thereof:
  • the decoding method or device adopts at least the following two decoding modes and two corresponding reference pixel samples respectively Value storage space:
  • the corresponding reference pixel sample storage space is a reference pixel sample storage space 1 composed of reconstructed pixel samples of a part of the current decoding CU but adjacent to the CU;
  • the corresponding reference pixel sample storage space is the reference pixel sample storage space 2 composed of part or all of the reconstructed index of the current decoding CU;
  • the value of the reference pixel is copied from the reference pixel sample storage space 1 by using the pixel string copying method. a predicted value or a reconstructed value of the currently decoded pixel; if the reference pixel of the currently decoded pixel is located inside the currently decoded CU, the value of the reference index is copied from the reference pixel sample storage space 2 as the current decoded pixel by using an index copy manner Index and get the palette color corresponding to the index from the palette as the predicted or reconstructed value of the currently decoded pixel.
  • the reference pixel sample storage space 1 in the implementation or variant 11 adopts a pixel color format of a non-index color format; the reference pixel sample storage space 2 in the implementation or variant 11 adopts a pixel color format of an index color format.
  • the decoding method or apparatus adopts at least the following two decoding modes and two corresponding reference pixel sample storage spaces:
  • the corresponding reference pixel sample storage space is a reference pixel sample storage space 1 composed of reconstructed pixel samples of a part of the current decoding CU but adjacent to the CU;
  • the corresponding reference pixel sample storage space is a palette associated with the currently decoded CU, that is, part or all of the pixels of the currently decoded CU may be represented by an index of the palette;
  • the value of the reference pixel is copied from the reference pixel sample storage space 1 by the pixel string copy method.
  • the index copy mode is adopted, the reference index is first decoded, and then the target is obtained from the palette.
  • the palette color corresponding to the index is used as a predicted value or a reconstructed value of the currently decoded pixel.
  • the reference pixel sample storage space 1 in the implementation or variant example 13 adopts the pixel color format of the non-index color format; the reference pixel sample storage space composed of the palette in the implementation or variant example 13 adopts the pixel color of the index color format. format.
  • FIG. 1 is a block diagram showing the composition of an encoding apparatus of the present invention.
  • FIG. 2 is a block diagram showing the composition of a decoding apparatus of the present invention.
  • 3 is an embodiment of four coding modes and two different forms of reference pixel sample storage spaces
  • 4 is an embodiment of four decoding modes and two different forms of reference pixel sample storage spaces.
  • Figure 5 is a sub-block of depth 1 and a depth 1 number of each sub-block
  • Figure 6 is a sub-block of depth 2 and a depth 2 number of each sub-block
  • Figure 7 is a sub-block of depth 3 and a depth 3 number of each sub-block
  • Figure 8 is a sub-block of depth 4 and a depth 4 number of each sub-block

Landscapes

  • Engineering & Computer Science (AREA)
  • Multimedia (AREA)
  • Signal Processing (AREA)
  • Compression Or Coding Systems Of Tv Signals (AREA)

Abstract

L'invention concerne un procédé et un dispositif de compression d'image. Lors de l'exécution d'un codage (ou d'un décodage) prédictif et de mise en correspondance sur une unité de codage (CU), représenter dans au moins deux formats les données historiques utilisées en tant que valeurs d'échantillon de pixel de référence, et stocker dans au moins deux espaces de stockage de valeur d'échantillon de pixel de référence, respectivement. Un espace de stockage de valeur d'échantillon de pixel de référence est l'espace de stockage de valeur d'échantillon de pixel de référence principal comprenant la plupart des données historiques ; les données historiques dans l'autre espace de stockage de valeur d'échantillon de pixel de référence est un sous-ensemble des données historiques dans l'espace de stockage de valeur d'échantillon de pixel de référence principal, mais présente différentes formes. Tous les modes de codage (ou de décodage) prédictif et de mise en correspondance sont classés de manière correspondante en au moins deux types, et utilisent respectivement des espaces de stockage de valeur d'échantillon de pixel de référence différents pour effectuer un codage (ou un décodage) prédictif et de mise en correspondance respectif. Les différents espaces de stockage de valeur d'échantillon de pixel de référence stockent réellement les mêmes données historiques uniquement en différentes formes de représentation. Par conséquent, les données historiques dans les différents espaces de stockage de valeur d'échantillon de pixel de référence doivent rester synchronisées.
PCT/CN2015/073204 2014-02-16 2015-02-16 Procédé et dispositif de compression d'image utilisant un espace de stockage de pixel de référence en de multiples formes WO2015120823A1 (fr)

Applications Claiming Priority (2)

Application Number Priority Date Filing Date Title
CN201410051372 2014-02-16
CN201410051372.7 2014-02-16

Publications (1)

Publication Number Publication Date
WO2015120823A1 true WO2015120823A1 (fr) 2015-08-20

Family

ID=53799612

Family Applications (1)

Application Number Title Priority Date Filing Date
PCT/CN2015/073204 WO2015120823A1 (fr) 2014-02-16 2015-02-16 Procédé et dispositif de compression d'image utilisant un espace de stockage de pixel de référence en de multiples formes

Country Status (2)

Country Link
CN (1) CN104853211A (fr)
WO (1) WO2015120823A1 (fr)

Cited By (9)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN107483942A (zh) * 2016-06-08 2017-12-15 同济大学 视频数据压缩码流的解码、视频数据的编码方法及装置
CN107770553A (zh) * 2016-08-21 2018-03-06 上海天荷电子信息有限公司 采用多类匹配参数及参数存储地址的数据压缩方法和装置
CN109995573A (zh) * 2018-01-02 2019-07-09 中国移动通信有限公司研究院 确定用于预测计算资源的样本空间的方法、计算资源分配方法及装置
CN110662029A (zh) * 2018-06-30 2020-01-07 上海天荷电子信息有限公司 动态和自适应地限制取值范围的数据压缩方法和装置
CN110944178A (zh) * 2018-09-22 2020-03-31 上海天荷电子信息有限公司 使用多个以编码参数关联的候选集的数据压缩方法和装置
CN111034202A (zh) * 2017-07-05 2020-04-17 奥兰治 图像编码和解码方法、编码和解码设备及对应计算机程序
CN111866512A (zh) * 2020-07-29 2020-10-30 腾讯科技(深圳)有限公司 视频解码方法、视频编码方法、装置、设备及存储介质
CN112383781A (zh) * 2013-08-16 2021-02-19 上海天荷电子信息有限公司 参考块的位置决定其重构阶段的块匹配编解码方法和装置
CN115119046A (zh) * 2022-06-02 2022-09-27 绍兴市北大信息技术科创中心 一种参考像素集的图像编解码方法、装置及系统

Families Citing this family (13)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN107770543B (zh) * 2016-08-21 2023-11-10 上海天荷电子信息有限公司 多类匹配参数中按顺序递增截断值的数据压缩方法和装置
CN107770540B (zh) * 2016-08-21 2024-01-09 上海天荷电子信息有限公司 融合多种具有不同参考关系的基元的数据压缩方法和装置
CN114501020A (zh) * 2016-08-21 2022-05-13 上海天荷电子信息有限公司 使用最近运动矢量和邻近运动矢量的图像压缩方法和装置
CN107770544B (zh) * 2016-08-21 2023-12-08 上海天荷电子信息有限公司 采用多类匹配参数及前参数缓冲区的数据压缩方法和装置
CN108235007B (zh) * 2016-12-12 2023-06-27 上海天荷电子信息有限公司 各模式使用不同精度同种编码参数的数据压缩方法和装置
CN108184113B (zh) * 2017-12-05 2021-12-03 上海大学 一种基于图像间参考的图像压缩编码方法和系统
WO2020207421A1 (fr) * 2019-04-09 2020-10-15 Beijing Bytedance Network Technology Co., Ltd. Construction d'entrée de codage en mode palette
CN114175662B (zh) 2019-07-20 2023-11-24 北京字节跳动网络技术有限公司 调色板模式使用指示的条件相关编解码
CN112532989B (zh) * 2020-06-02 2022-02-25 腾讯科技(深圳)有限公司 视频编码方法、视频解码方法及相关设备
CN112532988B (zh) * 2020-06-02 2022-05-06 腾讯科技(深圳)有限公司 视频编码方法、视频解码方法及相关设备
CN113395515B (zh) * 2021-04-08 2022-06-14 同济大学 对分量下采样格式数据进行点预测的编码解码方法及装置
CN113808225B (zh) * 2021-09-27 2023-09-19 东华理工大学南昌校区 一种图像的无损编码方法
CN116996687A (zh) * 2023-07-21 2023-11-03 深圳市战音科技有限公司 一种图像拆分编码方法、装置及电子设备

Citations (5)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
JP2009212790A (ja) * 2008-03-04 2009-09-17 Mitsubishi Electric Corp 予測モード選択装置および画像符号化装置
US20110110424A1 (en) * 2009-11-12 2011-05-12 Hitachi Kokusai Electric Inc. Video Encoder and Data Processing Method
CN102497546A (zh) * 2011-11-28 2012-06-13 同济大学 一种屏幕图像压缩装置及压缩方法
CN102761738A (zh) * 2011-12-05 2012-10-31 上海天荷电子信息有限公司 基于混合色度采样率的图像压缩方法和装置
CN104244007A (zh) * 2013-06-13 2014-12-24 上海天荷电子信息有限公司 基于任意形状匹配的图像压缩方法和装置

Family Cites Families (1)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
JP2012142865A (ja) * 2011-01-05 2012-07-26 Sony Corp 画像処理装置及び画像処理方法

Patent Citations (5)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
JP2009212790A (ja) * 2008-03-04 2009-09-17 Mitsubishi Electric Corp 予測モード選択装置および画像符号化装置
US20110110424A1 (en) * 2009-11-12 2011-05-12 Hitachi Kokusai Electric Inc. Video Encoder and Data Processing Method
CN102497546A (zh) * 2011-11-28 2012-06-13 同济大学 一种屏幕图像压缩装置及压缩方法
CN102761738A (zh) * 2011-12-05 2012-10-31 上海天荷电子信息有限公司 基于混合色度采样率的图像压缩方法和装置
CN104244007A (zh) * 2013-06-13 2014-12-24 上海天荷电子信息有限公司 基于任意形状匹配的图像压缩方法和装置

Cited By (16)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN112383781B (zh) * 2013-08-16 2023-05-02 上海天荷电子信息有限公司 参考块的位置决定其重构阶段的块匹配编解码方法和装置
CN112383781A (zh) * 2013-08-16 2021-02-19 上海天荷电子信息有限公司 参考块的位置决定其重构阶段的块匹配编解码方法和装置
CN107483942A (zh) * 2016-06-08 2017-12-15 同济大学 视频数据压缩码流的解码、视频数据的编码方法及装置
CN107483942B (zh) * 2016-06-08 2023-07-14 同济大学 视频数据压缩码流的解码、视频数据的编码方法及装置
CN107770553A (zh) * 2016-08-21 2018-03-06 上海天荷电子信息有限公司 采用多类匹配参数及参数存储地址的数据压缩方法和装置
CN107770553B (zh) * 2016-08-21 2023-06-27 上海天荷电子信息有限公司 采用多类匹配参数及参数存储地址的数据压缩方法和装置
CN111034202B (zh) * 2017-07-05 2023-08-22 奥兰治 图像编码和解码方法、编码和解码设备及对应计算机程序
CN111034202A (zh) * 2017-07-05 2020-04-17 奥兰治 图像编码和解码方法、编码和解码设备及对应计算机程序
CN109995573B (zh) * 2018-01-02 2022-04-01 中国移动通信有限公司研究院 确定用于预测计算资源的样本空间的方法、计算资源分配方法及装置
CN109995573A (zh) * 2018-01-02 2019-07-09 中国移动通信有限公司研究院 确定用于预测计算资源的样本空间的方法、计算资源分配方法及装置
CN110662029A (zh) * 2018-06-30 2020-01-07 上海天荷电子信息有限公司 动态和自适应地限制取值范围的数据压缩方法和装置
CN110944178B (zh) * 2018-09-22 2023-03-07 上海天荷电子信息有限公司 数据压缩的编码方法及编码装置、解码方法及解码装置
CN110944178A (zh) * 2018-09-22 2020-03-31 上海天荷电子信息有限公司 使用多个以编码参数关联的候选集的数据压缩方法和装置
CN111866512A (zh) * 2020-07-29 2020-10-30 腾讯科技(深圳)有限公司 视频解码方法、视频编码方法、装置、设备及存储介质
CN115119046A (zh) * 2022-06-02 2022-09-27 绍兴市北大信息技术科创中心 一种参考像素集的图像编解码方法、装置及系统
CN115119046B (zh) * 2022-06-02 2024-04-16 绍兴市北大信息技术科创中心 一种参考像素集的图像编解码方法、装置及系统

Also Published As

Publication number Publication date
CN104853211A (zh) 2015-08-19

Similar Documents

Publication Publication Date Title
WO2015120823A1 (fr) Procédé et dispositif de compression d'image utilisant un espace de stockage de pixel de référence en de multiples formes
CN104378644B (zh) 定宽度变长度像素样值串匹配增强的图像压缩方法和装置
WO2015120818A1 (fr) Méthodes et dispositifs de codage et de décodage d'image
US11451835B2 (en) Method and device for entropy coding and decoding using mapping of matching relationship parameters
CN105491376B (zh) 图像编码、解码方法及装置
EP3876537A1 (fr) Procédé de codage d'image, procédé de décodage d'image, dispositif de codage et dispositif de décodage
EP3310059B1 (fr) Procédé d'encodage et de décodage d'image, dispositif de traitement d'image, et support de stockage informatique
CN113784124B (zh) 使用多种形状子块进行精细划分的块匹配编码和解码方法
US11653019B2 (en) Image coding and decoding methods, image processing device and computer storage medium
JP6794364B2 (ja) 画像符号化方法及び装置並びに画像復号方法及び装置
CN104811731A (zh) 多层次子块匹配图像压缩方法
CN105992003B (zh) 依据排序或频度对调色板颜色编号的图像压缩方法和装置
WO2016124154A1 (fr) Procédé et appareil de codage d'image ainsi que procédé et appareil de décodage d'image
EP3893191A1 (fr) Procédé et appareil de traitement d'images
WO2016119667A1 (fr) Procédé et appareil de traitement d'images

Legal Events

Date Code Title Description
121 Ep: the epo has been informed by wipo that ep was designated in this application

Ref document number: 15748501

Country of ref document: EP

Kind code of ref document: A1

NENP Non-entry into the national phase

Ref country code: DE

32PN Ep: public notification in the ep bulletin as address of the adressee cannot be established

Free format text: NOTING OF LOSS OF RIGHTS PURSUANT TO RULE 112(1) EPC (EPO FORM 1205 DATED 26/01/2017)

122 Ep: pct application non-entry in european phase

Ref document number: 15748501

Country of ref document: EP

Kind code of ref document: A1