WO2016202189A1 - Procédés de codage et de décodage d'image, dispositif de traitement d'image, et support de stockage informatique - Google Patents

Procédés de codage et de décodage d'image, dispositif de traitement d'image, et support de stockage informatique Download PDF

Info

Publication number
WO2016202189A1
WO2016202189A1 PCT/CN2016/085015 CN2016085015W WO2016202189A1 WO 2016202189 A1 WO2016202189 A1 WO 2016202189A1 CN 2016085015 W CN2016085015 W CN 2016085015W WO 2016202189 A1 WO2016202189 A1 WO 2016202189A1
Authority
WO
WIPO (PCT)
Prior art keywords
dynamic
copy
limit
motion vector
image
Prior art date
Application number
PCT/CN2016/085015
Other languages
English (en)
Chinese (zh)
Inventor
林涛
李明
吴平
尚国强
吴钊
Original Assignee
同济大学
中兴通讯股份有限公司
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Priority claimed from CN201610364479.6A external-priority patent/CN106254878B/zh
Application filed by 同济大学, 中兴通讯股份有限公司 filed Critical 同济大学
Priority to US15/736,006 priority Critical patent/US11159818B2/en
Priority to EP16810930.4A priority patent/EP3310056A4/fr
Priority to EP21178252.9A priority patent/EP3896976A1/fr
Publication of WO2016202189A1 publication Critical patent/WO2016202189A1/fr
Priority to US17/482,832 priority patent/US11653019B2/en

Links

Images

Classifications

    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04NPICTORIAL COMMUNICATION, e.g. TELEVISION
    • H04N19/00Methods or arrangements for coding, decoding, compressing or decompressing digital video signals
    • H04N19/10Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using adaptive coding
    • H04N19/102Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using adaptive coding characterised by the element, parameter or selection affected or controlled by the adaptive coding
    • H04N19/103Selection of coding mode or of prediction mode
    • H04N19/105Selection of the reference unit for prediction within a chosen coding or prediction mode, e.g. adaptive choice of position and number of pixels used for prediction
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04NPICTORIAL COMMUNICATION, e.g. TELEVISION
    • H04N19/00Methods or arrangements for coding, decoding, compressing or decompressing digital video signals
    • H04N19/10Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using adaptive coding
    • H04N19/102Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using adaptive coding characterised by the element, parameter or selection affected or controlled by the adaptive coding
    • H04N19/103Selection of coding mode or of prediction mode
    • H04N19/11Selection of coding mode or of prediction mode among a plurality of spatial predictive coding modes
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04NPICTORIAL COMMUNICATION, e.g. TELEVISION
    • H04N19/00Methods or arrangements for coding, decoding, compressing or decompressing digital video signals
    • H04N19/10Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using adaptive coding
    • H04N19/102Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using adaptive coding characterised by the element, parameter or selection affected or controlled by the adaptive coding
    • H04N19/13Adaptive entropy coding, e.g. adaptive variable length coding [AVLC] or context adaptive binary arithmetic coding [CABAC]
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04NPICTORIAL COMMUNICATION, e.g. TELEVISION
    • H04N19/00Methods or arrangements for coding, decoding, compressing or decompressing digital video signals
    • H04N19/10Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using adaptive coding
    • H04N19/134Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using adaptive coding characterised by the element, parameter or criterion affecting or controlling the adaptive coding
    • H04N19/136Incoming video signal characteristics or properties
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04NPICTORIAL COMMUNICATION, e.g. TELEVISION
    • H04N19/00Methods or arrangements for coding, decoding, compressing or decompressing digital video signals
    • H04N19/10Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using adaptive coding
    • H04N19/169Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using adaptive coding characterised by the coding unit, i.e. the structural portion or semantic portion of the video signal being the object or the subject of the adaptive coding
    • H04N19/17Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using adaptive coding characterised by the coding unit, i.e. the structural portion or semantic portion of the video signal being the object or the subject of the adaptive coding the unit being an image region, e.g. an object
    • H04N19/176Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using adaptive coding characterised by the coding unit, i.e. the structural portion or semantic portion of the video signal being the object or the subject of the adaptive coding the unit being an image region, e.g. an object the region being a block, e.g. a macroblock
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04NPICTORIAL COMMUNICATION, e.g. TELEVISION
    • H04N19/00Methods or arrangements for coding, decoding, compressing or decompressing digital video signals
    • H04N19/46Embedding additional information in the video signal during the compression process
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04NPICTORIAL COMMUNICATION, e.g. TELEVISION
    • H04N19/00Methods or arrangements for coding, decoding, compressing or decompressing digital video signals
    • H04N19/50Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using predictive coding
    • H04N19/593Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using predictive coding involving spatial prediction techniques

Definitions

  • the present invention relates to digital video encoding and decoding technologies, and in particular, to an image encoding and decoding method, an image processing device, and a computer storage medium.
  • HEVC High Efficiency Video Coding
  • a notable feature of computer screen images is that there are often many similar or even identical pixel patterns (pi ⁇ el pattern) within the same frame image.
  • pixel patterns pi ⁇ el pattern
  • Chinese or foreign text that often appears in computer screen images is composed of a few basic strokes, and many similar or identical strokes can be found in the same frame image.
  • Menus, icons, etc., which are common in computer screen images, also have many similar or identical patterns. Therefore, the existing image and video compression technologies usually adopt various copying methods, including at least the following copying methods:
  • Intra block copying means intraframe block matching or intraframe motion compensation or block matching or block copying.
  • Intra-frame micro-block copying means intra-frame micro-block matching or micro-block matching or micro-block copying.
  • Intra-frame lines are copied, that is, intra-frame matching or strip matching or strip copying.
  • Intra-frame string replication means intra-frame string matching or string matching or string copying or pixel string copying.
  • Palette index string copy is a palette or index string copy.
  • index-pixel string fusion copy mode also called palette-pixel string fusion copy mode or merged pixel string copy palette mode or fusion palette Pixel string copy mode or palette with pixel string copy mode.
  • the most common type of reference pixel sample set is a reference area adjacent to or near the current pixel sample segment.
  • the reference area is finite, so the range of values of the copy position, ie the motion vector, is also limited.
  • the range of the value of the motion vector is not limited, thereby consuming unnecessary bits, which greatly reduces the compression efficiency of the copy mode.
  • the multi-pixel segment copy mode with multiple motion vectors in the codec block such as the microblock copy mode, the strip copy mode, the pixel string copy mode, the palette and the pixel string copy mode, and the compression efficiency.
  • an embodiment of the present invention provides an image encoding and decoding method, an image processing device, and a computer storage medium.
  • the coded bits of the reference area size information and/or the reference area position and the size information and/or the value range information of the motion vector are written into one or more of the video code streams:
  • Video parameter set Video parameter set, sequence parameter set, image parameter set, slice header, coding tree unit (CTU) header, coding unit (CU) header, codec block header.
  • CTU coding tree unit
  • CU coding unit
  • the copy parameter is a motion vector.
  • the dynamic limit of the copy parameter is one of a dynamic upper limit, a dynamic lower limit, a dynamic left limit, and a dynamic right limit of the motion vector, or multiple or all.
  • the method further includes:
  • a truncated k-th order exponential Columbus code binarization and entropy coding having a cutoff value corresponding to the dynamic limit is performed on the copy parameter, where k ⁇ 0.
  • the copy parameters are decoded according to the dynamic limit.
  • the parsing the video code stream to obtain the dynamic limit of the copy parameter of the decoded sample segment includes:
  • Parsing the video code stream and acquiring at least one of the following information: reference area size information, reference area position and size information, and value range information of the motion vector;
  • a dynamic limit of the copy parameter of the decoded sample segment is determined based on the acquired information.
  • the method before decoding the replication parameter according to the dynamic limit, the method further includes:
  • the following one or more corresponding code streams in the video code stream are parsed, and at least one of the following information is obtained: reference area size information, reference area position and size information, and value range information of the motion vector.
  • Video parameter set sequence parameter set, image parameter set, strip header, CTU header, CU header, codec block header.
  • the copy parameter is a motion vector.
  • the dynamic limit of the copy parameter is one of a dynamic upper limit, a dynamic lower limit, a dynamic left limit, and a dynamic right limit of the motion vector, or multiple or all.
  • the method further includes:
  • a truncated k-th order exponential Columbus code binarization and entropy decoding having a truncation value corresponding to the dynamic limit is performed on the copy parameter, where k ⁇ 0.
  • a video code stream including reference area position and size information and copy parameter information is generated.
  • the coded bits of the reference area size information and/or the reference area position and the size information and/or the value range information of the motion vector are written into one or more of the video code streams:
  • Video parameter set sequence parameter set, image parameter set, strip header, CTU header, CU header, codec block header.
  • the copy parameter is a motion vector.
  • the dynamic limit of the copy parameter is one, a plurality, or all of a dynamic upper limit, a dynamic lower limit, a dynamic left limit, and a dynamic right limit of the motion vector.
  • the dynamic position of the reference area is represented by one or more area motion vectors.
  • the reference region when the motion vector of the coded sample segment is encoded according to the dynamic position and size of the reference region, the reference region is composed of K sub-reference regions having the same shape and size as the codec block. Where K ⁇ 1;
  • the dynamic positions of the K sub-reference regions are respectively represented by corresponding K sub-reference region motion vectors, and the motion vector of the current codec segment is decomposed into a main motion vector and a slave motion vector, wherein the master mobile The vector is equal to the sub-reference region motion vector of the sub-reference region in which the reference sample segment is located.
  • the copy parameters of the current decoded sample segment are decoded according to the dynamic position and size of the reference region.
  • the following one or more corresponding code streams in the video code stream are parsed, and at least one of the following information is obtained: reference area size information, reference area position and size information, and value range information of the motion vector. ,include:
  • Video parameter set sequence parameter set, image parameter set, strip header, CTU header, CU header, codec block header.
  • the copy parameter is a motion vector.
  • the dynamic limit of the copy parameter is one of a dynamic upper limit, a dynamic lower limit, a dynamic left limit, and a dynamic right limit of the motion vector, or multiple or all.
  • the dynamic position of the reference area is represented by one or more area motion vectors.
  • the reference region when the motion vector of the current decoded sample segment is decoded according to the dynamic position and size of the reference region, the reference region is composed of K sub-reference regions having the same shape and size as the current codec block; , K ⁇ 1;
  • the dynamic positions of the K sub-reference regions are respectively represented by corresponding K sub-reference region motion vectors, and the motion vector of the current codec segment is decomposed into a main motion vector and a slave motion vector, wherein the master mobile The vector is equal to the sub-reference area motion vector of the sub-reference area in which the current reference sample segment is located.
  • a first acquiring unit configured to acquire a dynamic limit of a copy parameter of the current coded sample segment according to a size of the reference area
  • a coding unit configured to encode the copy parameter according to the dynamic limit
  • a generating unit configured to generate a video code stream including reference area size information and copy parameter information.
  • An parsing unit configured to parse the video code stream to obtain a dynamic limit of a copy parameter of the decoded sample segment
  • a decoding unit configured to decode the copy parameter according to the dynamic limit.
  • a coding unit configured to encode a copy parameter of the coded sample segment according to a dynamic position and size of the reference region
  • a generating unit configured to generate a video code stream including reference area location and size information and copy parameter information.
  • the parsing unit is configured to parse the video code stream, and at least obtain the reference area location and size information and the copy parameter information;
  • the decoding unit is configured to decode the copy parameter of the current decoded sample segment according to the dynamic position and size of the reference region.
  • the computer storage medium provided by the embodiment of the present invention stores a computer program for performing the above image encoding and/or image decoding method.
  • encoding the image includes: acquiring a dynamic limit of the copy parameter of the current coded sample segment according to the size of the reference region; encoding the copy parameter according to the dynamic limit; generating the reference region Video stream of size information and copy parameter information.
  • decoding the image comprises: parsing the video code stream, obtaining a dynamic limit of the copy parameter of the decoded sample segment; and decoding the copy parameter according to the dynamic limit.
  • encoding the image includes: encoding a copy parameter of the encoded sample segment according to a dynamic position and size of the reference region; and generating a video code stream including the reference region position and size information and the copy parameter information.
  • the decoding of the image comprises: parsing the video code stream, acquiring at least the reference area position and size information and the copy parameter information; and decoding the copy parameter of the current decoded sample segment according to the dynamic position and size of the reference area.
  • the range of the value of the motion vector is limited, the number of bits is saved, and the compression efficiency of the copy mode is improved.
  • FIG. 1 is a schematic flowchart 1 of an image encoding method according to an embodiment of the present invention.
  • FIG. 2 is a schematic flowchart 1 of an image decoding method according to an embodiment of the present invention.
  • FIG. 3 is a second schematic flowchart of an image encoding method according to an embodiment of the present invention.
  • FIG. 4 is a second schematic flowchart of an image decoding method according to an embodiment of the present invention.
  • FIG. 5 is a schematic flowchart 3 and a flow chart 4 of an image encoding method according to an embodiment of the present invention
  • FIG. 6 is a diagram showing an example of a size of a reference area, a position of a reference area, and a dynamic limit of a motion vector in an embodiment of the present invention
  • FIG. 7 is a schematic flowchart 3 and a flow diagram 4 of an image decoding method according to an embodiment of the present invention.
  • FIG. 8 is a first schematic structural diagram of an image processing apparatus according to an embodiment of the present invention.
  • FIG. 9 is a second schematic structural diagram of an image processing apparatus according to an embodiment of the present invention.
  • FIG. 10 is a third schematic structural diagram of an image processing apparatus according to an embodiment of the present invention.
  • FIG. 11 is a fourth structural diagram of an image processing apparatus according to an embodiment of the present invention.
  • the natural form of a digital video signal of an image is a sequence of images.
  • a frame of image is usually a rectangular area composed of several pixels, and a digital video signal is a sequence of video images composed of tens of frames to thousands of frames of images, sometimes simply referred to as a video sequence or sequence.
  • Encoding a digital video signal encodes a frame by frame image. At any one time, the image of the frame being encoded is referred to as the current encoded image.
  • decoding a compressed video stream (also referred to as a bit stream) of a digital video signal is decoding a stream of one frame by one compressed image.
  • the image of the frame being decoded is referred to as the current decoded image.
  • the current encoded image or the currently decoded image is collectively referred to as the current image.
  • one frame of image is divided into blocks M.
  • the sub-image of the M pixel is called a coding block (from the perspective of decoding, that is, a decoding block, collectively referred to as a codec block) or a coding unit (CU, Coding Unit), and the sub-image is performed one by one with the CU as a basic coding unit.
  • the size of the commonly used M is 4, 8, 16, 32, 64. Therefore, encoding the video image sequence is to encode one CU for each coding unit of each frame image, that is, CU.
  • the CU being coded is referred to as the current coded CU.
  • decoding the code stream of the video image sequence is also decoding one CU for each CU of each frame image, and finally reconstructing the entire video image sequence.
  • the CU being decoded is referred to as the currently decoded CU.
  • the current coding CU or the current decoding CU is collectively referred to as the current CU.
  • each CU in one frame of image can be different, some are 8 ⁇ 8, some are 64 ⁇ 64, etc. .
  • LCUs Lorgest Coding Units
  • each LCU is further divided into a tree structure. Multiple sizes are not necessarily The same CU. Therefore, the LCU is also called a Coding Tree Unit (CTU).
  • CTU Coding Tree Unit
  • One of the LCUs is composed of three 32 ⁇ 32 pixel CUs and four 16 ⁇ 16 pixel CUs, so that seven CUs in a tree structure constitute one CTU.
  • the other LCU is composed of two 32 ⁇ 32 pixel CUs, three 16 ⁇ 16 pixel CUs, and 20 8 ⁇ 8 pixel CUs.
  • Such 25 CUs in a tree structure constitute another CTU.
  • Encoding one frame of image is to sequentially encode one CU in one CTU.
  • LCU is synonymous with CTU.
  • a CU whose size is equal to the CTU is called a CU with a depth of zero.
  • a CU obtained by dividing the CU up and down and left and right by a depth of 0 is called a CU having a depth of 1.
  • a CU with a depth of 1 and a CU divided into four equal parts is called a CU with a depth of 2.
  • a CU obtained by dividing the CU of the depth of 2 into four equal parts is called a CU having a depth of 3.
  • the CTU being coded is referred to as the current coded CTU.
  • the CTU being decoded is referred to as the current decoded CTU.
  • the current coding CTU or the current decoding CTU is collectively referred to as the current CTU.
  • the CU can also be further divided into sub-areas.
  • Sub-regions include, but are not limited to, prediction unit (PU), transform unit (TU), asymmetric partition (AMP) regions.
  • PU prediction unit
  • TU transform unit
  • AMP asymmetric partition
  • a color pixel usually consists of three components.
  • the two most commonly used pixel color formats are a GBR color format consisting of a green component, a blue component, and a red component, and a luminance (luma) component and two chroma components.
  • YUV color format The color format commonly known as YUV actually includes multiple color formats, such as the YCbCr color format.
  • a CU when encoding a CU, a CU can be divided into three component planes (G plane, B plane, R plane or Y plane, U plane, V plane), and the three component planes are respectively coded;
  • the three component bundles of one pixel are combined into one 3-tuple, and the CUs composed of these 3-tuples are encoded as a whole.
  • the arrangement of the former pixel and its components is called the planar format of the image (and its CU), and the arrangement of the latter pixel and its components is called the stacked format of the image (and its CU). Format).
  • Pixel Both the GBR color format and the YUV color format are pixel 3-component representation formats.
  • the palette index representation format In addition to the 3-component representation format of a pixel, another common representation format for a pixel is the palette index representation format.
  • the value of a pixel can also be represented by the index of the palette.
  • the palette space stores the value or approximate value of the three components of the color of the pixel to be represented, and the address of the palette is referred to as the index of the color of the pixel stored in this address.
  • An index can represent a component of a pixel's color, and an index can also represent three components of a pixel's color.
  • the palette can be one or more. In the case of multiple palettes, a complete index is actually composed of the palette number (which one of the multiple palettes) and the index of the palette of the sequence number.
  • the index representation format of a pixel is to represent this pixel with an index. If the pixels in an image region (such as a coded block or a decoded block) cannot all be represented by the palette color (ie, at least one pixel in the image region, the value of no three components is equal or approximately equal to the pixel)
  • the palette color and its index there is usually a special index called escape color in the palette, which is used to represent pixels that cannot be represented by normal palette colors. Therefore, if the index of one pixel is an index of the escape color, the pixel needs to express its color with another dedicated 3 components.
  • the normal color and escape color in the palette are called palette colors, but the escape color is a virtual color. There is no physical space in the palette to store this color.
  • the index of the escape color is usually the last index of the palette.
  • the index representation format of a pixel is also referred to as the pixel's index color (inde ⁇ ed color) or pseudo color representation format, or is often referred to directly as an index pixel (inde ⁇ ed pi ⁇ el) or pseudo pixel (pseudo) Pi ⁇ el) or pixel index or index. Indexes are sometimes referred to as indices.
  • the representation of a pixel in its index representation format is also referred to as indexing or indexing.
  • CMYK representation formats Other commonly used pixel representation formats include CMYK representation formats and grayscale representation formats.
  • the YUV color format can be subdivided into several seed formats according to whether the chroma component is downsampled: 1 pixel consists of 1 Y component, 1 U component, and 1 V component. YUV4:4:4 Pixel color format; YUV4:2:2 pixel color format consisting of 2 Y components, 1 U component, and 1 V component; 2 left and right adjacent 4 ⁇ 2 spatial positions
  • the pixels are composed of 4 Y components, 1 U component, and 1 V component in a YUV 4:2:0 pixel color format.
  • a component is generally represented by a number of 8 to 16 bits.
  • the YUV4:2:2 pixel color format and the YUV4:2:0 pixel color format are all downsampled for the YUV4:4:4 pixel color format.
  • a pixel component is also referred to as a pixel sample (pi x el sample) or simply as a sample.
  • the most basic element when encoding or decoding can be one pixel, one pixel component, or one pixel index (ie, index pixel).
  • a pixel or a pixel component or an index pixel, which is the most basic element of encoding or decoding, is collectively referred to as a pixel sample, sometimes referred to as a pixel value, or simply as a sample.
  • pixel sample value In the embodiment of the present invention, "pixel sample value”, “pixel value”, “sample value”, “index pixel”, and “pixel index” are synonyms, and depending on the context, whether “pixel” or “one pixel” is clearly indicated The “component” still means “index pixel” or both. If it is not clear from the context, then it means either of the three.
  • a coded block or a decoded block is an area consisting of a number of pixel values.
  • the shape of the codec block may be a rectangle, a square, a parallelogram, a trapezoid, a polygon, a circle, an ellipse, and the like.
  • a rectangle also includes a rectangle whose width or height is one pixel value that degenerates into a line (ie, a line segment or a line shape).
  • each codec block may have a different shape and size.
  • some or all of the codec blocks may overlap each other, or all codec blocks may not overlap each other.
  • a codec block may be composed of "pixels”, or may be composed of “components of pixels”, or may be composed of "index pixels”, or may be composed of a mixture of the three, or any of the three. Mixed composition.
  • a codec block refers to an area in which encoding or decoding is performed in a frame image, including but not limited to at least one of the following: a maximum coding unit LCU, A coding tree unit CTU, a coding unit CU, a sub-region of a CU, a prediction unit PU, and a transform unit TU.
  • a notable feature of computer screen images is that there are often many similar or even identical pixel patterns (pi ⁇ el pattern) within the same frame image.
  • pixel patterns pi ⁇ el pattern
  • Chinese or foreign text that often appears in computer screen images is composed of a few basic strokes, and many similar or identical strokes can be found in the same frame image.
  • Menus, icons, etc., which are common in computer screen images, also have many similar or identical patterns. Therefore, the copying method adopted by image and video compression technology includes at least the following copying methods:
  • Intra block copying means intraframe block matching or intraframe motion compensation or block matching or block copying.
  • the basic operation of block copy coding or decoding is to copy a current code block or a current decoding block (referred to as the current block) from the reconstructed reference pixel sample set by the same size as the current block (the same number of pixel samples).
  • the reference block is referenced and the value of the reference block is assigned to the current block.
  • the copy parameter of the block copy mode includes the displacement vector of the current block, indicating the relative position between the reference block and the current block.
  • a current block has a displacement vector.
  • Intra-frame micro-block copying means intra-frame micro-block matching or micro-block matching or micro-block copying.
  • a current block such as 8 ⁇ 8 pixel samples
  • several microblocks such as microblocks of 4 ⁇ 2 pixel samples or microblocks of 8 ⁇ 2 pixel samples or 2 ⁇ 4 pixels
  • the microblock of the sample or the microblock of the 2 ⁇ 8 pixel sample) the basic operation of the microblock copy encoding or decoding is to encode the microblock or the decoded microblock (referred to as the current microblock) for each of the current block, from A reference microblock is copied within the reconstructed reference pixel sample set, and the value of the reference microblock is assigned to the current microblock.
  • the copy parameter of the microblock copy mode includes a displacement vector of the current microblock, indicating a relative position between the reference microblock and the current microblock.
  • a current microblock has a displacement vector. How many displacement vectors are there in how many microblocks a current block is divided into.
  • Intra-frame lines are copied, that is, intra-frame matching or strip matching or strip copying.
  • a strip is a microblock having a height of 1 or a width of 1, such as a microblock of 4x1 or 8x1 or 1x4 or 1x8 pixel samples.
  • the basic operation of strip copy encoding or decoding is for each encoding strip or decoding strip in the current block. (referred to as the current bar), copy a reference bar from the reconstructed reference pixel sample set, and assign the value of the reference bar to the current bar.
  • strip copying is a special case of microblock copying.
  • the copy parameter of the strip copy mode includes the displacement vector of the current strip, indicating the relative position between the reference strip and the current strip.
  • a current bar has a displacement vector. How many displacement vectors are there in how many bars a current block is divided into.
  • Intra-frame string replication means intra-frame string matching or string matching or string copying or pixel string copying.
  • a current coded block or a current decoded block (referred to as the current block) is divided into several variable-length pixel sample strings.
  • the string here refers to arranging the pixel samples in a two-dimensional area of an arbitrary shape into a string whose length is much larger than the width (for example, a string having a width of 1 pixel sample and a length of 37 pixel samples or a width of 2)
  • a string of pixel samples of length 111 pixels typically but not limited to a length that is an independent encoding or decoding parameter and a width that is a predetermined or derived parameter from other encoding or decoding parameters).
  • the basic operation of string copy encoding or decoding is to copy each reference string from the reconstructed reference pixel sample set to each encoded string or decoded string (referred to as the current string) in the current block, and the value of the reference string Assign to the current string.
  • the copy parameters of the string copy mode include the displacement vector and the copy length of the current string, that is, the copy size, which respectively represent the relative position between the reference string and the current string and the length of the current string, that is, the number of pixel samples.
  • the length of the current string is also the length of the reference string.
  • a current string has a displacement vector and a copy length. How many displacement vectors and how many copy lengths a current block is divided into.
  • the displacement vector is also called the copy position, and its representation forms are: 2D coordinates, linear address, distance, pointer, index, number, mode number, and so on.
  • Palette index string copy is a palette or index string copy.
  • first construct or acquire a palette and then use some or all of the pixels of the current encoding block or the current decoding block (referred to as the current block) with the palette in the palette.
  • the index of the color is expressed, and then the index is encoded and decoded, including but not limited to: dividing the index of a current block into several index strings of variable length, that is, performing index string copy encoding and decoding.
  • Index string copy The basic operation of encoding or decoding is to encode a string or index decoding string (referred to as the current index string) for each index in the current block, and copy a reference index string from the indexed reconstructed reference pixel sample set, and The index value of the reference index string is assigned to the current index string.
  • the copy parameter of the index string copy mode includes the displacement vector and the copy length of the current index string, that is, the copy size, respectively indicating the relative position between the reference index string and the current index string and the length of the current index string, that is, the number of corresponding pixel samples.
  • the length of the current index string is also the length of the reference index string.
  • a current index string has a displacement vector and a copy length.
  • the displacement vector is also called the copy position, and its representation forms are: 2D coordinates, linear address, distance, pointer, index, number, mode number, and so on.
  • the index string is also called a palette color string or a palette pixel string or a palette string
  • the index string copy mode is also called a palette copy mode or a palette mode.
  • strings are also called runs or strokes. Therefore, the index string is also called an index run or index run or simply a run or trip.
  • the palette color of the current block is derived from the pixel color and/or palette color candidate set of the current block, and the palette color candidate set is composed of a portion of the palette color of the codec block that has completed the codec.
  • index-pixel string fusion copy mode also called palette-pixel string fusion copy mode or merged pixel string copy palette mode or fusion palette Pixel string copy mode or palette with pixel string copy mode.
  • index-pixel string fusion copy mode also called palette-pixel string fusion copy mode or merged pixel string copy palette mode or fusion palette Pixel string copy mode or palette with pixel string copy mode.
  • Different fusion schemes may differ in the following aspects: 1) the number of types of strings, 2) the range of values of parameters, 3) the range of values of one or several replicated parameters, and 4) the location of the displacement vector Range of values, 5) current index or current pixel position, 6) position of the current sample segment, 7) position of the reference index or reference pixel, 8) position of the reference sample segment; 9) copy shape.
  • a pixel sample segment referred to as a sample segment.
  • the basic constituent elements of a sample segment are pixel or pixel components or pixel indices.
  • a sample segment has a copy parameter that represents the relationship between the current pixel sample segment and the reference pixel sample segment. Therefore, a sample segment is the smallest unit of a copy operation with the same replication relationship.
  • a copy parameter includes a plurality of copy parameter components, and the copy parameter components include at least: displacement vector horizontal component, displacement vector vertical component, 1 dimensional displacement vector, linear address, relative linear address, index, palette linear address, relative index, toning Board relative linear address, copy length, copy width, copy height, rectangle width, rectangle length, unmatched pixels (also known as no reference pixels, ie non-replicated pixels that are not copied from elsewhere).
  • pixel samples or indexes need to be arranged in a certain order.
  • the arrangement is also called the scanning method.
  • the scanning method can be divided into the following types according to its path shape:
  • A) Horizontal Z-scan mode is also called horizontal raster scan mode.
  • a pixel sample or index of a coded block or a decoded block (collectively referred to as a codec block) is arranged line by line and arranged in the same direction (all left to right or all right to left) in all rows. Rows and rows can be arranged from top to bottom or from bottom to top.
  • Vertical Z-scan mode is also called vertical raster scan mode.
  • a coding block or a decoding block Pixel samples or indexes, called codec blocks, are arranged in columns and columns, and are arranged in the same direction (all from top to bottom or all from bottom to top) in all columns. Columns and columns can be arranged from left to right or from right to left.
  • a pixel sample or index of a coded block or a decoded block (collectively referred to as a codec block), arranged line by line, arranged in one direction (eg, from left to right) in an odd line and in another (opposite) direction in an even line (eg, from right to left). Rows and rows can be arranged from top to bottom or from bottom to top.
  • a pixel sample or index of a coded block or a decoded block (collectively referred to as a codec block), arranged in columns and columns, arranged in one direction (eg, from top to bottom) in an odd column and in another in an even column (opposite ) Directions (eg, from bottom to top). Columns and columns can be arranged from left to right or from right to left.
  • the basic copy shape has the following two types:
  • the current string and the reference string are each a one-dimensional sample string formed in the order of a predetermined scanning manner in each codec block, and have equal lengths, but the two-dimensional regions formed by the two strings are not necessarily the same. Two-dimensional shape.
  • the current strings are arranged in the order of the predetermined scanning mode within the current codec block.
  • the reference string remains exactly the same two-dimensional shape as the current string, with equal length.
  • Each of the above basic replicated shapes can be subdivided into a plurality of replicated shapes according to a specific scanning manner, such as a vertical arcuate one-dimensional linear replica shape, a horizontal Z-shaped two-dimensional conformal replica shape, and the like.
  • copying is an operation of reconstruction and decoding, and the corresponding encoding operation is “matching”. Therefore, various copying methods such as block matching mode, microblock copying mode, line copying mode, pixel string copying mode, and index string copying mode are also called block matching mode, microblock matching mode, line matching mode, and pixel string matching mode. , index string matching method, and so on.
  • reference pixel sample set is a reference area adjacent or near the current pixel sample segment.
  • the reference area is finite, so the range of values of the copy position, ie the motion vector, is also limited.
  • the range of the value of the motion vector is not limited, thereby consuming unnecessary bits, which greatly reduces the compression efficiency of the copy mode.
  • the multi-pixel segment copy mode with multiple motion vectors in one codec block such as the microblock copy mode, the strip copy mode, the pixel string copy mode, the palette and the pixel string copy mode, and the compression efficiency.
  • an embodiment of the present invention provides an image encoding and decoding method and an image processing device. Before encoding and decoding the motion vector, first limit the value range of the motion vector.
  • the size of the reference area of one or more codec blocks is fixed. In the embodiment of the present invention, the size of the reference area of one or more codec blocks is bounded. In the embodiment of the present invention, the size of the reference area of the one or more codec blocks is adaptively restricted, that is, the limitation is by other properties of the codec block (eg, the location of the codec block, or The codec block is determined similarly to the pixels of other regions in the image, etc.). In the embodiment of the present invention, the position of the reference area of one or more codec blocks is fixed.
  • the location of the reference area of the one or more codec blocks is adaptively restricted, that is, the limitation is by other properties of the codec block (eg, the location of the codec block, or The codec block is determined similarly to the pixels of other regions in the image, etc.).
  • the value range of the motion vector of one or more current pixel sample segments is fixed. In the embodiment of the present invention, the range of values of the motion vectors of one or more current pixel sample segments is bounded.
  • the value range of the motion vector of one or more current pixel sample segments is adaptively limited, that is, the limit is determined by other properties of the current pixel sample segment (eg, the current The position of the pixel sample segment, or the position of the current pixel sample segment, or the similarity of the current pixel sample segment to pixels of other regions in the image, etc.).
  • the size and location of the reference area of one codec block are restricted, and the size and location of the reference area of different codec blocks may be limited. It is different and is adaptive and dynamic.
  • the range of values of the motion vector of a current pixel sample segment is limited.
  • the limits of the range of motion vectors of different current pixel sample segments may be different, adaptive and dynamically changing.
  • the reference area is an adaptive and dynamic limited area
  • the value range of the motion vector is an adaptive and dynamic limited value range.
  • the motion vector is subjected to self-limiting coding and decoding, that is, adaptive and dynamic limited coding and decoding.
  • the self-limiting codec includes at least a binarization method and an entropy codec having a truncation value corresponding to the restriction.
  • the self-limiting codec includes at least a truncated k-th order exponential Columbus code binarization method and an entropy codec having a truncation value corresponding to the restriction, where k ⁇ 0.
  • the truncated k-th order exponential Columbus code is a k-th order exponential Columbus code whose prefix and/or its suffix has a cutoff value.
  • the self-limiting codec includes at least decomposing a set of several motion vectors into one main motion vector and several slave motion vectors.
  • the adaptive and dynamic limited area is K (K ⁇ 1) sub-reference areas having the same size and shape as the current codec block, and the self-limiting codec includes at least the current codec.
  • the motion vector of the block is divided into corresponding K groups, and each motion vector of each group is decomposed into a common main motion vector and each slave motion vector.
  • one or several of a video parameter set and/or a sequence parameter set and/or an image parameter set and/or a slice header and/or a CTU header and/or a CU header and/or a codec block header of a video code stream The reference area is specified and restricted with one or several syntax elements.
  • a video parameter set and/or a sequence parameter set and/or an image parameter set and/or a slice header and/or a CTU header and/or a CU header and/or a codec block header of a video code stream Or a number of places with one or several syntax elements to specify and limit the range of values of the motion vector.
  • FIG. 1 is a schematic flowchart 1 of an image encoding method according to an embodiment of the present invention. As shown in FIG. 1 , the image encoding method includes the following steps:
  • Step 101 Obtain a dynamic limit of a copy parameter of the current coded sample segment according to the size of the reference area.
  • Step 102 Encode the copy parameter according to the dynamic limit, and generate a video code stream including reference area size information and copy parameter information.
  • the coded bits of the reference area size information and/or the reference area position and the size information and/or the value range information of the motion vector are written into one or more of the video code streams.
  • Video parameter set sequence parameter set, image parameter set, strip header, CTU header, CU header, codec block header.
  • the video parameter set usually one or several syntax elements of the video parameter set that exist directly or implicitly;
  • Sequence parameter set usually one or several syntax elements that are directly or implicitly derived from a sequence parameter set
  • Image parameter set usually one or several syntax elements of the image parameter set that exist directly or implicitly;
  • Strip head A grammatical element that is usually one or more of the leading or implicit derivation of a strip
  • CTU header A syntax element that is usually one or several directly or implicitly derived from a CTU header
  • CU header usually one or several syntax elements that are directly or implicitly derived from the CU header;
  • Codec block header Usually one or several syntax elements that are directly or implicitly derived from the codec block header.
  • the copy parameter is a motion vector.
  • the dynamic limit of the copy parameter is one of a dynamic upper limit, a dynamic lower limit, a dynamic left limit, and a dynamic right limit of the motion vector, or multiple or all.
  • a truncated k-th order exponential Columbus code binarization and entropy coding having a cutoff value corresponding to the dynamic limit is performed on the copy parameter, where k ⁇ 0.
  • FIG. 2 is a schematic flowchart 1 of an image decoding method according to an embodiment of the present invention. As shown in FIG. 2, the image decoding method includes the following steps:
  • Step 201 Parse the video code stream to obtain a dynamic limit of the copy parameter of the decoded sample segment.
  • Step 202 Decode the replication parameter according to the dynamic limit.
  • the parsing the video code stream to obtain the dynamic limit of the copy parameter of the decoded sample segment includes:
  • Parsing the video code stream and acquiring at least one of the following information: reference area size information, reference area position and size information, and value range information of the motion vector;
  • a dynamic limit of the copy parameter of the decoded sample segment is determined based on the acquired information.
  • the method before decoding the replication parameter according to the dynamic limit, the method further includes:
  • the following one or more corresponding code streams in the video code stream are parsed, and at least one of the following information is obtained: reference area size information, reference area position and size information, and value of the motion vector.
  • Range information including:
  • Video parameter set sequence parameter set, image parameter set, strip header, CTU header, CU header, codec block header.
  • the copy parameter is a motion vector.
  • the dynamic limit of the copy parameter is one of a dynamic upper limit, a dynamic lower limit, a dynamic left limit, and a dynamic right limit of the motion vector, or multiple or all.
  • a truncated k-th order exponential Columbus code binarization and entropy decoding having a cutoff value corresponding to the dynamic limit is performed on the copy parameter, Where k ⁇ 0.
  • FIG. 3 is a schematic flowchart 2 of an image encoding method according to an embodiment of the present invention. As shown in FIG. 3, the image encoding method includes the following steps:
  • Step 301 Encode the copy parameter of the coded sample segment according to the dynamic position and size of the reference region.
  • Step 302 Generate a video code stream including reference area location and size information and copy parameter information.
  • the reference bit size information and/or the reference region position and the coded bits of the size information and/or the value range information of the motion vector are written in the following part of the video code stream or Multiple places:
  • Video parameter set sequence parameter set, image parameter set, strip header, CTU header, CU header, codec block header.
  • the copy parameter is a motion vector.
  • the dynamic limit of the copy parameter is one, a plurality, or all of a dynamic upper limit, a dynamic lower limit, a dynamic left limit, and a dynamic right limit of the motion vector.
  • the dynamic position of the reference area is represented by one or more area motion vectors.
  • the reference region when the motion vector of the coded sample segment is encoded according to the dynamic position and size of the reference region, the reference region is composed of K children of the same shape and size as the codec block.
  • Reference area composition wherein, K ⁇ 1;
  • the dynamic positions of the K sub-reference regions are respectively represented by corresponding K sub-reference region motion vectors, and the motion vector of the current codec segment is decomposed into a main motion vector and a slave motion vector, wherein the master mobile The vector is equal to the sub-reference region motion vector of the sub-reference region in which the reference sample segment is located.
  • FIG. 4 is a second schematic flowchart of an image decoding method according to an embodiment of the present invention, as shown in FIG. 4,
  • the image decoding method includes the following steps:
  • Step 401 Parse the video code stream, and obtain at least reference area location and size information and copy parameter information.
  • Step 402 Decode the copy parameter of the current decoded sample segment according to the dynamic position and size of the reference region.
  • the following one or more corresponding code streams in the video code stream are parsed, and at least one of the following information is obtained: reference area size information, reference area position and size information, and motion vector acquisition.
  • Value range information including:
  • Video parameter set sequence parameter set, image parameter set, strip header, CTU header, CU header, codec block header.
  • the copy parameter is a motion vector.
  • the dynamic limit of the copy parameter is one of a dynamic upper limit, a dynamic lower limit, a dynamic left limit, and a dynamic right limit of the motion vector, or multiple or all.
  • the dynamic position of the reference area is represented by one or more area motion vectors.
  • the reference region when the motion vector of the current decoded sample segment is decoded according to the dynamic position and size of the reference region, the reference region is composed of K sub-reference regions having the same shape and size as the current codec block. Composition; wherein, K ⁇ 1;
  • the dynamic positions of the K sub-reference regions are respectively represented by corresponding K sub-reference region motion vectors, and the motion vector of the current codec segment is decomposed into a main motion vector and a slave motion vector, wherein the master mobile The vector is equal to the sub-reference area motion vector of the sub-reference area in which the current reference sample segment is located.
  • FIG. 5( a ) is a third schematic flowchart of an image encoding method according to an embodiment of the present invention. As shown in the figure, in the process:
  • the video code stream of the area size information and the copy parameter information the dynamic limit of the copy parameter of the current coded sample segment is obtained from the size of the reference area, and the copy parameter is encoded according to the dynamic limit.
  • FIG. 5(b) is a schematic flowchart diagram 4 of an image encoding method according to an embodiment of the present invention. As shown in the figure, in the process:
  • Coding with a copy mode including at least the following functions and operations and generating a video code stream containing at least reference area position and size information and copy parameter information: encoding the copy parameters of the current coded sample segment according to the dynamic position and size of the reference area .
  • the copy parameter is a motion vector.
  • Figure 6 is an example.
  • the reference area is a current CTU and a plurality of adjacent CTUs, such that the reference area size is a current CTU and a size of a plurality of adjacent CTUs.
  • Fig. 6(a) is an example.
  • the dynamic limit of the copy parameter is a dynamic upper limit, a dynamic lower limit, a dynamic left limit, and a dynamic right limit of the motion vector.
  • Fig. 6(a) is an example.
  • the copy parameters are binarized and entropy encoded with a cutoff value corresponding to the dynamic limit based on the dynamic limit.
  • the truncated k-th order exponential Columbus code binarization and entropy coding having a truncation value corresponding to the dynamic limit is performed on the copy parameter according to the dynamic limit, wherein k ⁇ 0.
  • the truncated k-th order exponent Columbus code is a k-th order exponential Columbus code whose prefix and/or its suffix has a cutoff value.
  • the dynamic position of the reference area is represented by one or several area motion vectors.
  • the main motion vector in Fig. 6(b) is an example.
  • the dynamic position of the reference area is represented by one or several area motion vectors, and the size of the reference area is the size of one or several current coding blocks.
  • Fig. 6(b) is an example.
  • the reference area consists of one or several regions of the same shape and size as the current coded block, and the dynamic position of the reference region is represented by a corresponding one or several region motion vectors.
  • Fig. 6(b) is an example.
  • the motion vector of the current coded sample segment when the motion vector of the current coded sample segment is encoded according to the dynamic position and size of the reference region, the motion vector of a group of the current coded sample segments is decomposed into a primary motion vector and a plurality of slaves.
  • Moving vector Fig. 6(b) is an example.
  • the reference region is K (K ⁇ 1) sub-references having the same shape and size as the current coded block.
  • the regional components, and the dynamic positions of the K sub-reference regions are respectively represented by corresponding K (K ⁇ 1) sub-reference region motion vectors, and the motion vector of the current coded sample segment is decomposed into a main motion vector and a slave A motion vector, the primary motion vector being equal to a sub-reference region motion vector of a sub-reference region in which the current reference sample segment is located.
  • Fig. 6(b) is an example.
  • the reference region size information is present in a video parameter set and/or a sequence parameter set and/or an image parameter set and/or a slice header and/or a CTU header and/or a CU header and/or of a video bitstream. Or code one or more places in the block header.
  • the reference region location and size information is present in a video parameter set and/or a sequence parameter set and/or an image parameter set and/or a slice header and/or a CTU header and/or a CU header of the video stream. And/or encode one or more places in the block header.
  • FIG. 7( a ) is a third schematic flowchart of an image decoding method according to an embodiment of the present invention. As shown in the figure, in the process:
  • Parsing the video code stream acquiring at least the reference area size information and the copy parameter information, and decoding by using a copy mode that includes at least the following functions and operations: obtaining a dynamic limit of the copy parameter of the current decoded sample segment from the size of the reference area, according to the The dynamic boundary decodes the copy parameters.
  • FIG. 7(b) is a schematic flowchart diagram 4 of an image decoding method according to an embodiment of the present invention, as shown in the figure. In the process:
  • Parsing the video code stream obtaining at least the reference area position and size information and the copy parameter information, and decoding by using a copy mode including at least the following functions and operations: decoding the copy parameters of the current decoded sample segment according to the dynamic position and size of the reference area .
  • the copy parameter is a motion vector.
  • Figure 6 is an example.
  • the reference area is a current CTU and a plurality of adjacent CTUs, such that the reference area size is a current CTU and a size of a plurality of adjacent CTUs.
  • Fig. 6(a) is an example.
  • the dynamic limit of the copy parameter is a dynamic upper limit, a dynamic lower limit, a dynamic left limit, and a dynamic right limit of the motion vector.
  • Fig. 6(a) is an example.
  • the copying parameter is binarized and entropy decoded with a cutoff value corresponding to the dynamic limit according to the dynamic limit.
  • the truncated k-th order exponential Columbus code binarization and entropy decoding having a truncation value corresponding to the dynamic limit is performed on the copy parameter according to the dynamic limit, wherein k ⁇ 0.
  • the truncated k-th order exponent Columbus code is a k-th order exponential Columbus code whose prefix and/or its suffix has a cutoff value.
  • the dynamic position of the reference area is represented by one or several area motion vectors.
  • the main motion vector in Fig. 6(b) is an example.
  • the dynamic position of the reference region is represented by one or several region motion vectors, and the size of the reference region is the size of one or several current decoded blocks.
  • Fig. 6(b) is an example.
  • the reference area is composed of one or several regions of the same shape and size as the current decoded block, and the dynamic position of the reference region is represented by a corresponding one or several region motion vectors.
  • Fig. 6(b) is an example.
  • the current decoded sample segment is based on the dynamic position and size of the reference region.
  • a set of motion vectors of a plurality of currently decoded sample segments are decomposed into a main motion vector and a plurality of slave motion vectors.
  • Fig. 6(b) is an example.
  • the reference region is composed of K (K ⁇ 1) sub-references of the same shape and size as the current decoded block.
  • the regional components, and the dynamic positions of the K sub-reference regions are respectively represented by corresponding K (K ⁇ 1) sub-reference region motion vectors, and the motion vector of the current decoded sample segment is decomposed into a main motion vector and a slave A motion vector, the primary motion vector being equal to a sub-reference region motion vector of a sub-reference region in which the current reference sample segment is located.
  • Fig. 6(b) is an example.
  • the reference region size information is present in a video parameter set and/or a sequence parameter set and/or an image parameter set and/or a slice header and/or a CTU header and/or a CU header and/or of a video bitstream. Or code one or more places in the block header.
  • the reference region location and size information is present in a video parameter set and/or a sequence parameter set and/or an image parameter set and/or a slice header and/or a CTU header and/or a CU header of the video stream. And/or encode one or more places in the block header.
  • the invention is applicable to the encoding and decoding of images in a stacked format.
  • the invention is equally applicable to the encoding and decoding of component plane format images.
  • FIG. 8 is a first schematic structural diagram of an image processing apparatus according to an embodiment of the present invention.
  • the image processing apparatus can encode an image. As shown in FIG. 8, the image processing apparatus includes:
  • the first obtaining unit 81 is configured to acquire, according to the size of the reference area, a dynamic limit of the copy parameter of the current coded sample segment;
  • the encoding unit 82 is configured to encode the copy parameter according to the dynamic limit
  • the generating unit 83 is configured to generate a video code stream including reference area size information and copy parameter information.
  • each unit in the image processing device may be implemented by a central processing unit (CPU) or a microprocessor (Micro Processor Unit, MPU) located in the image processing device. Or a digital signal processor (DSP), or a Field Programmable Gate Array (FPGA).
  • CPU central processing unit
  • MPU Micro Processor Unit
  • DSP digital signal processor
  • FPGA Field Programmable Gate Array
  • FIG. 9 is a second schematic structural diagram of an image processing apparatus according to an embodiment of the present invention.
  • the image processing apparatus can decode an image. As shown in FIG. 9, the image processing apparatus includes:
  • the parsing unit 91 is configured to parse the video code stream to obtain a dynamic limit of the copy parameter of the decoded sample segment;
  • the decoding unit 92 is configured to decode the copy parameter according to the dynamic limit.
  • each unit in the image processing device may be implemented by a central processing unit (CPU) or a microprocessor (Micro Processor Unit, MPU) located in the image processing device. Or a digital signal processor (DSP), or a Field Programmable Gate Array (FPGA).
  • CPU central processing unit
  • MPU Micro Processor Unit
  • DSP digital signal processor
  • FPGA Field Programmable Gate Array
  • FIG. 10 is a third structural diagram of an image processing apparatus according to an embodiment of the present invention.
  • the image processing apparatus can encode an image. As shown in FIG. 10, the image processing apparatus includes:
  • the encoding unit 1001 is configured to encode the copy parameter of the encoded sample segment according to the dynamic position and size of the reference region;
  • the generating unit 1002 is configured to generate a video code stream including reference area position and size information and copy parameter information.
  • each unit in the image processing device may be implemented by a central processing unit (CPU) or a microprocessor (Micro Processor Unit, MPU) located in the image processing device. Or a digital signal processor (DSP), or a Field Programmable Gate Array (FPGA).
  • CPU central processing unit
  • MPU Micro Processor Unit
  • DSP digital signal processor
  • FPGA Field Programmable Gate Array
  • FIG. 11 is a schematic structural diagram of an image processing apparatus according to an embodiment of the present invention.
  • the image processing apparatus can decode an image. As shown in FIG. 11, the image processing apparatus includes:
  • the parsing unit 1101 is configured to parse the video code stream, and obtain at least reference area location and size information and copy parameter information;
  • the decoding unit 1102 is configured to decode the copy parameters of the current decoded sample segment according to the dynamic position and size of the reference region.
  • each unit in the image processing device may be implemented by a central processing unit (CPU) or a microprocessor (Micro Processor Unit, MPU) located in the image processing device. Or a digital signal processor (DSP), or a Field Programmable Gate Array (FPGA).
  • CPU central processing unit
  • MPU Micro Processor Unit
  • DSP digital signal processor
  • FPGA Field Programmable Gate Array
  • embodiments of the present invention can be provided as a method, system, or computer program product. Accordingly, the present invention can take the form of a hardware embodiment, a software embodiment, or a combination of software and hardware. Moreover, the invention can take the form of a computer program product embodied on one or more computer-usable storage media (including but not limited to disk storage and optical storage, etc.) including computer usable program code.
  • Embodiments of the Invention may also be stored in a computer readable storage medium if it is implemented in the form of a software function module and sold or used as a standalone product. Based on such understanding, the technical solution of the embodiments of the present invention may be embodied in the form of a software product in essence or in the form of a software product stored in a storage medium, including a plurality of instructions.
  • a computer device (which may be a personal computer, server, or network device, etc.) is caused to perform all or part of the methods described in various embodiments of the present invention.
  • the foregoing storage medium includes various media that can store program codes, such as a USB flash drive, a mobile hard disk, a read only memory (ROM), a magnetic disk, or an optical disk.
  • program codes such as a USB flash drive, a mobile hard disk, a read only memory (ROM), a magnetic disk, or an optical disk.
  • an embodiment of the present invention further provides a computer storage medium, wherein a computer program for executing an image encoding method and/or an image decoding method according to an embodiment of the present invention is stored.
  • the computer program instructions can also be stored in a computer readable memory that can direct a computer or other programmable data processing device to operate in a particular manner, such that the instructions stored in the computer readable memory produce an article of manufacture comprising the instruction device.
  • the apparatus implements the functions specified in one or more blocks of a flow or a flow and/or block diagram of the flowchart.
  • These computer program instructions can also be loaded onto a computer or other programmable data processing device such that a series of operational steps are performed on a computer or other programmable device to produce computer-implemented processing for execution on a computer or other programmable device. Instructions are provided for implementation The steps of a function specified in a block or blocks of a flow or a flow and/or a block diagram of a flow chart.
  • the dynamic boundary of the copy parameter of the current coded sample segment is obtained according to the size of the reference region; the copy parameter is encoded according to the dynamic limit; and the reference region size information and the copy parameter information are generated.
  • Video stream Parsing the video code stream, obtaining a dynamic limit of the copy parameter of the decoded sample segment; decoding the copy parameter according to the dynamic limit.
  • the copy parameter of the coded sample segment is encoded according to the dynamic position and size of the reference region; and the video code stream including the reference region position and size information and the copy parameter information is generated.
  • Parsing the video code stream obtaining at least the reference area position and size information and the copy parameter information; and decoding the copy parameters of the current decoded sample segment according to the dynamic position and size of the reference area.

Landscapes

  • Engineering & Computer Science (AREA)
  • Multimedia (AREA)
  • Signal Processing (AREA)
  • Compression Or Coding Systems Of Tv Signals (AREA)

Abstract

L'invention concerne des procédés de codage et de décodage d'image, un dispositif de traitement d'image, et un support de stockage informatique. Le procédé de codage d'image comprend les étapes consistant à : obtenir une plage dynamique d'un paramètre de copie d'un segment de valeur d'échantillonnage de codage actuel, d'après une taille d'une zone de référence ; et coder le paramètre de copie d'après la plage dynamique pour générer un flux binaire vidéo contenant des informations concernant la taille de la zone de référence et des informations concernant le paramètre de copie. Le procédé de décodage comprend les étapes consistant à : analyser un flux binaire vidéo pour obtenir une plage dynamique d'un paramètre de copie d'un segment de valeur d'échantillonnage de décodage ; et décoder le paramètre de copie d'après la plage dynamique. En variante, le procédé de codage d'image comprend les étapes consistant à : coder un paramètre de copie d'un segment de valeur d'échantillonnage de codage d'après une position dynamique et la taille d'une zone de référence ; et générer un flux binaire vidéo contenant des informations concernant la position et la taille de la zone de référence et des informations concernant le paramètre de copie. En variante, le procédé de codage d'image comprend les étapes consistant à : coder un paramètre de copie d'un segment de valeur d'échantillonnage de codage actuel d'après une position dynamique et la taille d'une zone de référence ; et générer un flux binaire vidéo contenant des informations concernant la position et la taille de la zone de référence et des informations concernant le paramètre de copie.
PCT/CN2016/085015 2015-06-14 2016-06-06 Procédés de codage et de décodage d'image, dispositif de traitement d'image, et support de stockage informatique WO2016202189A1 (fr)

Priority Applications (4)

Application Number Priority Date Filing Date Title
US15/736,006 US11159818B2 (en) 2015-06-14 2016-06-06 Image coding and decoding methods, image processing device and computer storage medium
EP16810930.4A EP3310056A4 (fr) 2015-06-14 2016-06-06 Procédés de codage et de décodage d'image, dispositif de traitement d'image, et support de stockage informatique
EP21178252.9A EP3896976A1 (fr) 2015-06-14 2016-06-06 Procédés de codage et de décodage d'images, dispositif de traitement d'images, et support de stockage informatique
US17/482,832 US11653019B2 (en) 2015-06-14 2021-09-23 Image coding and decoding methods, image processing device and computer storage medium

Applications Claiming Priority (4)

Application Number Priority Date Filing Date Title
CN201510322825.X 2015-06-14
CN201510322825 2015-06-14
CN201610364479.6 2016-05-27
CN201610364479.6A CN106254878B (zh) 2015-06-14 2016-05-27 一种图像编码及解码方法、图像处理设备

Related Child Applications (2)

Application Number Title Priority Date Filing Date
US15/736,006 A-371-Of-International US11159818B2 (en) 2015-06-14 2016-06-06 Image coding and decoding methods, image processing device and computer storage medium
US17/482,832 Continuation US11653019B2 (en) 2015-06-14 2021-09-23 Image coding and decoding methods, image processing device and computer storage medium

Publications (1)

Publication Number Publication Date
WO2016202189A1 true WO2016202189A1 (fr) 2016-12-22

Family

ID=57544935

Family Applications (1)

Application Number Title Priority Date Filing Date
PCT/CN2016/085015 WO2016202189A1 (fr) 2015-06-14 2016-06-06 Procédés de codage et de décodage d'image, dispositif de traitement d'image, et support de stockage informatique

Country Status (1)

Country Link
WO (1) WO2016202189A1 (fr)

Cited By (4)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN110662029A (zh) * 2018-06-30 2020-01-07 上海天荷电子信息有限公司 动态和自适应地限制取值范围的数据压缩方法和装置
CN112532990A (zh) * 2020-11-23 2021-03-19 绍兴文理学院 一种串长度参数编码、解码方法和装置
CN112601086A (zh) * 2020-12-07 2021-04-02 绍兴文理学院 一种串长度参数混合编码、解码方法及装置
CN113452995A (zh) * 2021-03-31 2021-09-28 同济大学 当前串与参考串有不同扫描方向的数据编解码方法及装置

Citations (9)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US20040057515A1 (en) * 2002-09-20 2004-03-25 Shinichiro Koto Video encoding method and video decoding method
CN1703097A (zh) * 2004-05-26 2005-11-30 松下电器产业株式会社 编码运动矢量的设备和方法及其程序和介质
CN101094398A (zh) * 2006-06-20 2007-12-26 中兴通讯股份有限公司 一种基于时间分辨率变换的视频转换处理方法
CN101309405A (zh) * 2007-05-14 2008-11-19 华为技术有限公司 参考数据载入的方法及装置
CN101416524A (zh) * 2006-04-07 2009-04-22 微软公司 运动估计搜索范围和扩展的运动矢量范围的动态选择
US20120263231A1 (en) * 2011-04-18 2012-10-18 Minhua Zhou Temporal Motion Data Candidate Derivation in Video Coding
CN102939757A (zh) * 2010-06-04 2013-02-20 索尼公司 图像处理装置和方法
WO2013184949A1 (fr) * 2012-06-08 2013-12-12 Qualcomm Incorporated Sous-échantillonnage d'informations en mode prédiction dans un codage de couche d'amélioration
CN103647972A (zh) * 2008-12-03 2014-03-19 株式会社日立制作所 运动图像解码方法和运动图像编码方法

Patent Citations (9)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US20040057515A1 (en) * 2002-09-20 2004-03-25 Shinichiro Koto Video encoding method and video decoding method
CN1703097A (zh) * 2004-05-26 2005-11-30 松下电器产业株式会社 编码运动矢量的设备和方法及其程序和介质
CN101416524A (zh) * 2006-04-07 2009-04-22 微软公司 运动估计搜索范围和扩展的运动矢量范围的动态选择
CN101094398A (zh) * 2006-06-20 2007-12-26 中兴通讯股份有限公司 一种基于时间分辨率变换的视频转换处理方法
CN101309405A (zh) * 2007-05-14 2008-11-19 华为技术有限公司 参考数据载入的方法及装置
CN103647972A (zh) * 2008-12-03 2014-03-19 株式会社日立制作所 运动图像解码方法和运动图像编码方法
CN102939757A (zh) * 2010-06-04 2013-02-20 索尼公司 图像处理装置和方法
US20120263231A1 (en) * 2011-04-18 2012-10-18 Minhua Zhou Temporal Motion Data Candidate Derivation in Video Coding
WO2013184949A1 (fr) * 2012-06-08 2013-12-12 Qualcomm Incorporated Sous-échantillonnage d'informations en mode prédiction dans un codage de couche d'amélioration

Non-Patent Citations (2)

* Cited by examiner, † Cited by third party
Title
CHEN, CHUNCHI ET AL.: "Screen Content Coding Using Non-Square Intra Block Copy for HEVC", PROCEEDINGS OF THE 2014 IEEE INTERNATIONAL CONFERENCE ON MU1TIMEDIA AND EXPO, 18 July 2014 (2014-07-18), XP032639288, ISSN: 1945-7871 *
See also references of EP3310056A4 *

Cited By (5)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN110662029A (zh) * 2018-06-30 2020-01-07 上海天荷电子信息有限公司 动态和自适应地限制取值范围的数据压缩方法和装置
CN112532990A (zh) * 2020-11-23 2021-03-19 绍兴文理学院 一种串长度参数编码、解码方法和装置
CN112532990B (zh) * 2020-11-23 2024-03-08 绍兴文理学院 一种串长度参数编码、解码方法和装置
CN112601086A (zh) * 2020-12-07 2021-04-02 绍兴文理学院 一种串长度参数混合编码、解码方法及装置
CN113452995A (zh) * 2021-03-31 2021-09-28 同济大学 当前串与参考串有不同扫描方向的数据编解码方法及装置

Similar Documents

Publication Publication Date Title
CN112383780B (zh) 点匹配参考集和索引来回扫描串匹配的编解码方法和装置
US10341651B2 (en) Image coding method, decoding method, coding device and decoding device
CN111800640B (zh) 方向交替变换来回扫描串匹配的图像编码解码方法和装置
CN105491376B (zh) 图像编码、解码方法及装置
US10575022B2 (en) Image encoding and decoding method, image processing device and computer storage medium
CN106254865B (zh) 一种图像编码及解码方法、图像处理设备
CN105847842B (zh) 图像编码方法及装置、图像解码方法及装置
US10178405B2 (en) Enhanced coding and decoding using intra block copy mode
US11653019B2 (en) Image coding and decoding methods, image processing device and computer storage medium
CN104754362B (zh) 使用精细划分块匹配的图像压缩方法
WO2016202189A1 (fr) Procédés de codage et de décodage d'image, dispositif de traitement d'image, et support de stockage informatique
KR102532391B1 (ko) 영상 부호화 방법과 장치 및 영상 복호화 방법과 장치
CN105992003B (zh) 依据排序或频度对调色板颜色编号的图像压缩方法和装置
CN106303534B (zh) 多种索引串与像素串融合复制方式的图像压缩方法和装置
WO2016197893A1 (fr) Procédé de codage et de décodage d'image, dispositif de traitement d'image et support de stockage informatique
CN106303535B (zh) 参考像素取自不同程度重构像素的图像压缩方法和装置
WO2016119746A1 (fr) Procédé et appareil de codage d'image, et procédé et appareil de décodage d'image
WO2016197898A1 (fr) Procédé d'encodage et de décodage d'image, dispositif de traitement d'image, et support de stockage informatique
EP3253055B1 (fr) Procédé et appareil de codage/décodage d'image
WO2016124154A1 (fr) Procédé et appareil de codage d'image ainsi que procédé et appareil de décodage d'image
CN105828079B (zh) 图像处理方法及装置
WO2016119666A1 (fr) Procédé et appareil de codage/décodage d'image

Legal Events

Date Code Title Description
121 Ep: the epo has been informed by wipo that ep was designated in this application

Ref document number: 16810930

Country of ref document: EP

Kind code of ref document: A1

WWE Wipo information: entry into national phase

Ref document number: 15736006

Country of ref document: US

NENP Non-entry into the national phase

Ref country code: DE

WWE Wipo information: entry into national phase

Ref document number: 2016810930

Country of ref document: EP