WO2014048372A1 - Procédé et dispositif de traitement d'image, codeur et décodeur - Google Patents

Procédé et dispositif de traitement d'image, codeur et décodeur Download PDF

Info

Publication number
WO2014048372A1
WO2014048372A1 PCT/CN2013/084504 CN2013084504W WO2014048372A1 WO 2014048372 A1 WO2014048372 A1 WO 2014048372A1 CN 2013084504 W CN2013084504 W CN 2013084504W WO 2014048372 A1 WO2014048372 A1 WO 2014048372A1
Authority
WO
WIPO (PCT)
Prior art keywords
block
sub
target image
image
target
Prior art date
Application number
PCT/CN2013/084504
Other languages
English (en)
Chinese (zh)
Inventor
杨海涛
李礼
李斌
李厚强
Original Assignee
华为技术有限公司
中国科学技术大学
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Application filed by 华为技术有限公司, 中国科学技术大学 filed Critical 华为技术有限公司
Publication of WO2014048372A1 publication Critical patent/WO2014048372A1/fr

Links

Classifications

    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04NPICTORIAL COMMUNICATION, e.g. TELEVISION
    • H04N19/00Methods or arrangements for coding, decoding, compressing or decompressing digital video signals
    • H04N19/50Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using predictive coding
    • H04N19/59Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using predictive coding involving spatial sub-sampling or interpolation, e.g. alteration of picture size or resolution
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04NPICTORIAL COMMUNICATION, e.g. TELEVISION
    • H04N19/00Methods or arrangements for coding, decoding, compressing or decompressing digital video signals
    • H04N19/30Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using hierarchical techniques, e.g. scalability
    • H04N19/33Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using hierarchical techniques, e.g. scalability in the spatial domain
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04NPICTORIAL COMMUNICATION, e.g. TELEVISION
    • H04N19/00Methods or arrangements for coding, decoding, compressing or decompressing digital video signals
    • H04N19/46Embedding additional information in the video signal during the compression process
    • H04N19/463Embedding additional information in the video signal during the compression process by compressing encoding parameters before transmission
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04NPICTORIAL COMMUNICATION, e.g. TELEVISION
    • H04N19/00Methods or arrangements for coding, decoding, compressing or decompressing digital video signals
    • H04N19/50Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using predictive coding
    • H04N19/503Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using predictive coding involving temporal prediction
    • H04N19/51Motion estimation or motion compensation
    • H04N19/513Processing of motion vectors
    • H04N19/517Processing of motion vectors by encoding
    • H04N19/52Processing of motion vectors by encoding by predictive encoding
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04NPICTORIAL COMMUNICATION, e.g. TELEVISION
    • H04N19/00Methods or arrangements for coding, decoding, compressing or decompressing digital video signals
    • H04N19/10Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using adaptive coding
    • H04N19/169Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using adaptive coding characterised by the coding unit, i.e. the structural portion or semantic portion of the video signal being the object or the subject of the adaptive coding
    • H04N19/187Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using adaptive coding characterised by the coding unit, i.e. the structural portion or semantic portion of the video signal being the object or the subject of the adaptive coding the unit being a scalable video layer
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04NPICTORIAL COMMUNICATION, e.g. TELEVISION
    • H04N19/00Methods or arrangements for coding, decoding, compressing or decompressing digital video signals
    • H04N19/50Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using predictive coding
    • H04N19/503Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using predictive coding involving temporal prediction
    • H04N19/51Motion estimation or motion compensation

Definitions

  • the present invention relates to the field of video processing and, more particularly, to a method, apparatus, encoder and decoder for image processing. Background technique
  • Scalable coding is also known as layered coding.
  • SVC scalable video coding
  • the image can be subjected to down-resolution processing to obtain a low-resolution image, and the original image is referred to as a high-resolution image as a contrast, and the encoder separately determines the low-quality (eg, low-resolution) image.
  • the high quality (for example, high resolution) image is encoded to obtain high quality image coding information and low quality image coding information.
  • an image is divided into a plurality of image layers based on quality parameters including spatial resolution, temporal resolution, or signal-to-noise ratio intensity.
  • the goal of SVC is to make high-quality image layers make full use of low-quality image layer information, improve the efficiency of inter-layer prediction, and make it more efficient when encoding high-quality images.
  • the motion information of the image block of the appropriately scaled low-quality layer image is directly used as the motion information of the corresponding image block in the high-quality layer image, but, for example, there are one or more sub-blocks in the low-quality layer image (corresponding to The sub-block) is encoded in the intra mode (ie, the motion information of the corresponding sub-block is empty), and the sub-block in the high-quality layer image block cannot acquire the motion information from the corresponding sub-block in the low-quality layer image.
  • the motion information of the sub-block is constructed according to a given method.
  • the motion information derived from these derivations is not accurate, which affects the coding performance of the sub-block and further affects the coding efficiency of the entire high-quality layer image.
  • Embodiments of the present invention provide a method and apparatus for image processing, which can improve coding performance of a sub-block that cannot obtain motion information from a corresponding sub-block included in a base layer image in a target image block of an enhancement layer image.
  • a method for image processing comprising: when determining that motion information of a first base layer image sub-block corresponding to a first target image sub-block of a target image block is empty, according to Determining a second target image, a size of the target image block, a size of each target image sub-block included in the target image block, and second indication information indicating a position of the first target image sub-block in the target image block Determining, according to the motion information of the second target image sub-block, determining first reference information for encoding the first target image sub-block, where the first base layer image sub-block is located in the base layer image An image block, the target image block being located in the enhancement layer image, the base layer image corresponding to the enhancement layer image, and a spatial location of the first base image block sub-block in the base layer image and the first target An image sub-block corresponds to a spatial position in the enhancement layer image; encoding the target image block to generate a target code stream and included in the target code
  • the size according to the target image block, the size of the target image sub-block included in the target image block, and the second position indicating the position of the first target image sub-block in the target image block Instructing information, determining the second target image sub-block, comprising: determining the second target image sub-block according to any one of the following formulas,
  • Idx 2 ((l - ldx x %N / (N / 2)) x 2 + ⁇ ldx x %N / (N / 4) %2)) x N / 4 ;
  • Idx 2 Idx N x N + ((l - ldx x %N / (N / 2)) x 2 + (l - ldx x %N / (N / 4) %2)) x N / 4 ;
  • Representing third indication information indicating a position of the second target image sub-block in the target image block, / indicating the second indication information, N being according to a size of the target image block and a size of the target image sub-block definite.
  • the determining, according to the motion information of the second target image sub-block, determining, for encoding the first target image sub-block The first reference information includes: if the motion information of the second target image sub-block is empty, determining that the first reference information is zero motion information.
  • the encoding the target image block includes: according to the first reference information, The first target image sub-block performs motion compensation processing.
  • the target image block according to the reference information Performing encoding includes: performing deblocking filtering processing on pixels located near a boundary between the target image sub-blocks.
  • the method For encoding the target image block, the method includes: entropy encoding the first indication information, so that the first indication information is in the target code stream, and the skip mode flag bit or the merged MERGE mode flag bit information adjacent.
  • the encoding the target image block according to the reference information includes: determining, according to the reference image block located in the preset position in the enhancement layer image, whether the context is encoded by using reference information; according to the context, The first indication information is entropy encoded.
  • a method for image processing comprising: acquiring first indication information from a target code stream; and a first base layer corresponding to a first target image sub-block of the target image block
  • the motion information of the image sub-block is empty, based on the first indication information, according to the size of the target image block, the size of each target image sub-block included in the target image block, and the sub-block for indicating the first target image Determining, by the second indication information of the position in the target image block, the second target image sub-block; determining, according to the motion information of the second target image sub-block, the sub-block for the first target image Decoding the first reference information, wherein the first base layer image sub-block is an image block located in the base layer image, the target image block being located in the enhancement layer image, the base layer image corresponding to the enhancement layer image And the spatial position of the first basic image block sub-block in the base layer image corresponds to a spatial position of the first target image sub-block in the enhancement layer image;
  • the size according to the target image block, the size of the target image sub-block included in the target image block, and the second position indicating the position of the first target image sub-block in the target image block Instructing information, determining the second target image sub-block, comprising: determining the second target image sub-block according to any one of the following formulas,
  • the determining, according to the motion information of the second target image sub-block, determining, for encoding the first target image sub-block The first reference information includes: if the motion information of the second target image sub-block is empty, determining that the first reference information is zero motion information.
  • the decoding the target code stream includes: according to the first reference information, The first target image sub-block performs motion compensation processing.
  • the target image block according to the reference information Performing decoding includes: performing deblocking filtering processing on pixels located near a boundary between the target image sub-blocks.
  • the secondary target Obtaining the first indication information in the code stream includes: acquiring, from the target code stream, the first indication information, where the first indication information is in the target code stream and is merged with the skip mode flag bit
  • the obtaining the first indication information from the target code stream includes: determining, according to whether the reference image block located in the preset position in the enhancement layer image is decoded by using reference information, determining a context according to the context; Decoding to determine the first indication information.
  • an apparatus for image processing comprising: an obtaining unit, configured to determine motion information of a first base layer image sub-block corresponding to a first target image sub-block of a target image block When empty, according to the size of the target image block, the size of each target image sub-block included in the target image block, and the second indication information indicating the position of the first target image sub-block in the target image block, Determining a second target image sub-block; configured to determine, according to motion information of the second target image sub-block, first reference information used to encode the first target image sub-block, where the first base layer image sub- The block is an image block located in the base layer image, the target image block is located in the enhancement layer image, the base layer image corresponds to the enhancement layer image, and the first basic image block sub-block is in the base layer image a spatial position corresponding to a spatial position of the first target image sub-block in the enhancement layer image; a coding unit that encodes the target image block to generate a mesh
  • the acquiring unit is specifically configured to determine the second target image sub-block according to any one of the following formulas,
  • W3 ⁇ 4 / ⁇ /NxN + ((l-/ x 1 N/(N/2))x2 + (l-/ ⁇ N/(N/4) 2))xN/4; where The third indication information of the position of the second target image sub-block in the target image block, / represents the second indication information, and N is determined according to the size of the target image block and the size of the target image sub-block.
  • the acquiring unit is specifically configured to: if the motion information of the second target image sub-block is empty, determine the first reference information Zero motion information.
  • the coding unit is specifically configured to use the first target image according to the first reference information.
  • the sub-block performs motion compensation processing.
  • the coding unit is further configured to perform deblocking filtering processing on pixels located near a boundary between the target image sub-blocks.
  • the coding unit Specifically, the first indication information is entropy encoded such that the first indication information is adjacent to the skip mode flag bit or the merged MERGE mode flag bit information in the target code stream.
  • the coding unit is specifically configured to: according to whether the reference image block located in the preset position in the enhancement layer image is encoded by using reference information, determine a context; and perform entropy coding on the first indication information according to the context. .
  • a fourth aspect provides an apparatus for image processing, the apparatus comprising: a decoding unit, configured to acquire first indication information from a target code stream; and an acquiring unit, configured to: when the first target image is associated with the target image block And when the motion information of the first base layer image sub-block corresponding to the block is empty, based on the first indication information acquired by the decoding unit, according to the size of the target image block, each target image sub-block included in the target image block.
  • a second indication information indicating a position of the first target image sub-block in the target image block determining a second target image sub-block; and determining, according to the motion information of the second target image sub-block,
  • the first reference information for decoding the first target image sub-block wherein the first base layer image sub-block is an image block located in the base layer image, the target image block being located in the enhancement layer image, the basic a layer image corresponding to the enhancement layer image, and a spatial position of the first basic image block sub-block in the base layer image is in the first target image sub-block
  • the spatial position in the enhancement layer image corresponds; the decoding unit is further configured to decode the target code stream to obtain the target image block.
  • the acquiring unit is specifically configured to determine the second target image sub-block according to any one of the following formulas,
  • the acquiring unit is specifically configured to determine the first reference information if the motion information of the second
  • the decoding unit is specifically configured to: according to the first reference information, the first target image The sub-block performs motion compensation processing.
  • the decoding unit is further configured to locate the target image The pixels near the boundary between the sub-blocks are subjected to deblocking filtering processing.
  • the decoding unit Specifically, the first indication information is obtained from the target code stream, where the first indication information is adjacent to the skip mode flag bit or the merged MERGE mode flag bit information in the target code stream.
  • the decoding unit is specifically configured to: according to whether the reference image block located in the preset position in the enhancement layer image is decoded by using reference information, determine a context, and perform entropy decoding according to the context to determine the first Instructions.
  • an encoder for image processing comprising: a bus; a processor coupled to the bus; a memory coupled to the bus; wherein the processor calls the memory through the bus a program stored in the method for determining, when the motion information of the first base layer image sub-block corresponding to the first target image sub-block of the target image block is empty, according to the size of the target image block, the target image block Decoding a size of each target image sub-block and second indication information indicating a position of the first target image sub-block in the target image block, determining a second target image sub-block; for using the second target image
  • the motion information of the sub-block determines first reference information for encoding the first target image sub-block, wherein the first base layer image sub-block is an image block located in the base layer image, where the target image block is located In the enhancement layer image, the base layer image corresponds to the enhancement layer image, and the spatial position of the first basic image block sub-block in the base layer image and the The first target image sub
  • Idx 2 + idx, %N/(N/2))x2 + (l- Idx l %N /(N / 4)%2)) ⁇ N /4 ; Idx 2 + ((l - ldx x %N/(N/2))x2 + (ldx x %N / (N / 4) %2)) xN/4; W3 ⁇ 4 / ⁇ /NxN + ((l-/ x 1 N/(N/2))x2 + (l-/ ⁇ N/(N/4) 2))xN/4; wherein, the representation is used to indicate that the second target image sub-block is in the target image block
  • the third indication information of the location, / represents the second indication information, and N is determined according to the size of the target image block and the size of the target image sub-block.
  • the processor is specifically configured to: if the motion information of the second target image sub-block is empty, determine the first reference information Zero motion information.
  • the processor is specifically configured to: according to the first reference information, the first target image The sub-block performs motion compensation processing.
  • the processor is specifically configured to locate the target image The pixels near the boundary between the sub-blocks are subjected to deblocking filtering processing.
  • the processor Specifically, the first indication information is entropy encoded such that the first indication information is adjacent to the skip mode flag bit or the merged MERGE mode flag bit information in the target code stream.
  • the processor is specifically configured to: according to whether the reference image block located in the preset position in the enhancement layer image is encoded by using reference information, determine a context; and perform entropy coding on the first indication information according to the context. .
  • a decoder for image processing comprising: a bus; a processor coupled to the bus; a memory coupled to the bus; wherein the processor calls the memory through the bus a program stored in the target code stream for obtaining first indication information; for when motion information of the first base layer image sub-block corresponding to the first target image sub-block of the target image block is empty And based on the first indication information, according to the size of the target image block, the target Determining, by the image block, a size of each target image sub-block and a second indication information indicating a position of the first target image sub-block in the target image block, determining a second target image sub-block; Moving information of the target image sub-block, determining first reference information for decoding the first target image sub-block, wherein the first base layer image sub-block is an image block located in the base layer image, the target image The block is located in the enhancement layer image, the base layer image corresponds to the enhancement layer image, and the spatial position of the first
  • the processor is specifically configured to determine the second target image sub-block according to any one of the following formulas,
  • Idx 2 + ⁇ idx, %N/(N/2))x2 + (l- Idx, %N /(N / 4)%2)) ⁇ N /4 ; Idx 2 + ((l - ldx x %N/(N/2))x2 + (ldx x %N / (N / 4) %2)) xN/4; W3 ⁇ 4 / ⁇ /NxN + ((l-/ 1 %N/(N/2))x2 + (l-/ ⁇ %N/(N/4)%2))xN/4; wherein, the representation is used to indicate that the second target image sub-block is in the target image block
  • the third indication information of the position in the middle indicates that the second indication information is determined according to the size of the target image block and the size of the target image sub-block.
  • the processor is specifically configured to determine the first reference information if the motion information of the second target image sub-block is empty Zero motion information.
  • the processor is specifically configured to: according to the first reference information, the first target image The sub-block performs motion compensation processing.
  • the processor is specifically configured to locate the target image The pixels near the boundary between the sub-blocks are subjected to deblocking filtering processing.
  • the processor Specifically, the first indication information is obtained from the target code stream, where the first indication information is adjacent to the skip mode flag bit or the merged MERGE mode flag bit information in the target code stream.
  • the processor is specifically configured to determine, according to whether the reference image block located in the preset position in the enhancement layer image is used by using reference information, to determine a context, and perform entropy decoding according to the context to determine The first indication information.
  • a method for image processing comprising: when determining that motion information of a first base layer image sub-block corresponding to a first target image sub-block of a target image block is empty, according to a reconstructed pixel of the first base layer image sub-block, determining second reference information for encoding the first target image sub-block, wherein the first base layer image sub-block is an image block located in the base layer image And the target image block is located in the enhancement layer image, the base layer image corresponds to the enhancement layer image, and the spatial position of the first basic image block sub-block in the base layer image and the first target image sub-block Corresponding to a spatial position in the enhancement layer image; encoding the target image block to generate a target code stream and fourth indication information included in the target code stream.
  • the encoding the target image block comprises: performing deblocking filtering processing on pixels located near a boundary between the target image sub-blocks.
  • the encoding the target image block includes: entropy encoding the fourth indication information, so that the fourth indication information Adjacent to the skip mode flag bit or the merged MERGE mode flag bit information in the target bitstream.
  • the encoding the target image block includes: performing, according to whether the reference image block located at the preset position in the enhancement layer image uses reference information Encoding, determining a context; according to the context, the fourth indication information is entropy encoded.
  • a method for image processing comprising: acquiring fourth indication information from a target code stream; determining a first basic corresponding to a first target image sub-block of the target image block When the motion information of the layer image sub-block is empty, based on the fourth indication information, determining second reference information for encoding the first target image sub-block according to the reconstructed pixel of the first base layer image sub-block,
  • the first base layer image sub-block is an image block located in the base layer image
  • the target image block is located in the enhancement layer image
  • the base layer image corresponds to the enhancement layer image
  • the first basic image block is The spatial position of the sub-block in the base layer image corresponds to the spatial position of the first target image sub-block in the enhancement layer image
  • the target code stream is decoded to obtain the target image block.
  • the decoding of the target code stream includes: The pixels near the boundary between the target image sub-blocks perform deblocking filtering processing.
  • the obtaining, by the target code stream, the fourth indication information including: obtaining the fourth indication information from the target code stream,
  • the fourth indication information is adjacent to the skip mode flag bit or the merged MERGE mode flag bit information in the target code stream.
  • the obtaining, by the target code stream, the fourth indication information including: according to the enhancement layer Whether the reference image block located in the preset position in the image is decoded using the reference information to determine a context; according to the context, entropy decoding is performed to determine the fourth indication information.
  • an apparatus for image processing comprising: an acquisition unit, configured to determine motion information of a first base layer image sub-block corresponding to a first target image sub-block of a target image block When the space is empty, determining, according to the reconstructed pixels of the first base layer image sub-block, second reference information for encoding the first target image sub-block, where the first base layer image sub-block is located at the base layer An image block in the image, the target image block being located in the enhancement layer image, the base layer image corresponding to the enhancement layer image, and a spatial position of the first basic image block sub-block in the base layer image and the first A target image sub-block corresponds to a spatial position in the enhancement layer image; a coding unit is configured to encode the target image block to generate a target code stream and fourth indication information included in the target code stream.
  • the coding unit is specifically configured to perform deblocking filtering processing on pixels located near a boundary between the target image sub-blocks.
  • the coding unit is specifically configured to entropy encode the fourth indication information, so that the fourth indication information is in the target code
  • the stream is adjacent to the skip mode flag bit or the merged MERGE mode flag bit information.
  • the coding unit is specifically configured to use the reference image located in the preset position in the enhancement layer image. Whether the block is encoded using reference information, determining a context; and for entropy encoding the fourth indication information according to the context.
  • an apparatus for image processing comprising: a decoding unit, configured to acquire fourth indication information from a target code stream; and an acquiring unit, configured to determine a first target image with a target image block
  • a decoding unit configured to acquire fourth indication information from a target code stream
  • an acquiring unit configured to determine a first target image with a target image block
  • the motion information of the first base layer image sub-block corresponding to the sub-block is empty, based on the fourth indication information acquired by the decoding unit, according to the reconstructed pixel of the first base layer image sub-block, a second reference information for encoding the first target image sub-block, wherein the first base layer image sub-block is an image block located in the base layer image, the target image block being located in the enhancement layer image, a base layer image corresponding to the enhancement layer image, and a spatial position of the first base image block sub-block in the base layer image corresponds to a spatial position of the first target image sub-block in the enhancement layer image;
  • the coding unit is further configured
  • the decoding unit is specifically configured to perform deblocking filtering processing on pixels located near a boundary between the target image sub-blocks.
  • the decoding unit is specifically configured to obtain fourth indication information from the target code stream, where the fourth indication information is The target code stream is adjacent to the skip mode flag bit or the merged MERGE mode flag bit information.
  • the decoding unit is specifically configured to use the reference image located in the preset position in the enhancement layer image. Whether the block is decoded using reference information, determining a context; and performing entropy decoding according to the context to determine the fourth indication information.
  • an encoder for image processing comprising: a bus; a processor coupled to the bus; a memory coupled to the bus; wherein the processor calls the bus through the bus a program stored in the memory for reconstructing the motion of the first base layer image sub-block according to the first base layer image sub-block corresponding to the first target image sub-block of the target image block a second reference information for encoding the first target image sub-block, wherein the first base layer image sub-block is an image block located in a base layer image, the target image block being located in the enhancement layer image Corresponding to the enhancement layer image, and the spatial position of the first basic image block sub-block in the base layer image and the spatial position of the first target image sub-block in the enhancement layer image Corresponding; encoding the target image block to generate a target code stream and fourth indication information included in the target code stream.
  • the processor specifically performs deblocking filtering on pixels located near a boundary between the target image sub-blocks.
  • the processor is specifically configured to entropy encode the fourth indication information, so that the fourth indication information is at the target The code stream is adjacent to the skip mode flag bit or the merged MERGE mode flag bit information.
  • the processor is specifically configured to use the reference located in the preset position in the enhancement layer image. Whether the image block is encoded using the reference information, determining a context; and for entropy encoding the fourth indication information according to the context.
  • a decoder for image processing comprising: a bus; a processor connected to the bus; a memory connected to the bus; wherein the processor calls the bus through the bus a program stored in the memory for acquiring fourth indication information from the target code stream; for determining that motion information of the first base layer image sub-block corresponding to the first target image sub-block of the target image block is Empty space, based on the fourth indication information, determining second reference information for encoding the first target image sub-block according to the reconstructed pixel of the first base layer image sub-block, wherein the first base layer image The sub-block is an image block located in the base layer image, the target image block is located in the enhancement layer image, the base layer image corresponds to the enhancement layer image, and the first basic image block sub-block is in the base layer image a spatial position corresponding to a spatial position of the first target image sub-block in the enhancement layer image; configured to decode the target code stream to obtain the target image
  • the processor is specifically configured to perform deblocking filtering processing on pixels located near a boundary between the target image sub-blocks.
  • the processor is specifically configured to obtain fourth indication information from the target code stream, where the fourth indication information is The target bitstream is adjacent to the skip mode flag bit or the merged MERGE mode flag bit information.
  • the processor is specifically configured to determine, according to whether the reference image block located at the preset position in the enhancement layer image is decoded by using reference information, to determine a context according to the context, to determine the first Four instructions.
  • a method and apparatus for image processing according to an embodiment of the present invention for a first target image sub-block in which a motion information cannot be obtained from a corresponding sub-block included in a base layer image in a target image block of an enhancement layer image, according to the Position of a target image sub-block determines a second target image sub-block, and according to motion information of the second target image sub-block or first base layer image sub-block corresponding to the first target image sub-block in spatial position Reconstructing a pixel, determining reference information for the first target image sub-block, and encoding according to the reference information, can improve coding performance of the first target image sub-block.
  • 1 is a schematic flow chart of a method for image processing according to an embodiment of the present invention.
  • 2 is a schematic diagram of sub-block partitioning and sub-block indexing according to an embodiment of the invention.
  • FIG. 3 is a schematic flow chart of a method for image processing in accordance with another embodiment of the present invention.
  • 4 is a schematic block diagram of an apparatus for image processing in accordance with an embodiment of the present invention.
  • FIG. 5 is a schematic block diagram of an apparatus for image processing according to another embodiment of the present invention.
  • 6 is a schematic block diagram of an encoder for image processing in accordance with an embodiment of the present invention.
  • FIG. 7 is a schematic block diagram of a decoder for image processing according to another embodiment of the present invention.
  • 8 is a schematic flowchart of a method for image processing according to still another embodiment of the present invention.
  • FIG. 9 is a schematic flowchart of a method for image processing according to still another embodiment of the present invention.
  • 11 is a schematic block diagram of an apparatus for image processing according to still another embodiment of the present invention.
  • Figure 12 is a schematic block diagram of an encoder for image processing in accordance with still another embodiment of the present invention.
  • Figure 13 is a schematic block diagram of a decoder for image processing in accordance with still another embodiment of the present invention. detailed description
  • Fig. 1 shows a schematic flow chart of a method 100 for image processing according to an embodiment of the present invention, which is described from the perspective of an encoding end. As shown in FIG. 1, the method 100 includes:
  • S120 Determine, according to the motion information of the second target image sub-block, the first target image. a first reference information encoded by a sub-block, wherein the first base layer image sub-block is an image block located in a base layer image, the target image block being located in an enhancement layer image, the base layer image and the enhancement layer Corresponding to the image, and a spatial position of the first basic image block sub-block in the base layer image corresponds to a spatial position of the first target image sub-block in the enhancement layer image;
  • the image when the image is hierarchically encoded, for example, in spatial scalable coding, the image may be subjected to resolution processing to obtain a low-resolution image, and the original image is referred to as a high-resolution image as a contrast, respectively.
  • the low resolution image and the high resolution image are encoded.
  • a high quality image to be encoded is referred to herein as an enhancement layer image
  • a corresponding low quality image to be encoded e.g., the low resolution image
  • a base layer image a corresponding low quality image to be encoded
  • the target image is an image processed by a layered coding technique
  • the basic layer refers to a quality in layered coding (including frame rate, spatial resolution, temporal resolution, signal to noise ratio intensity or quality level).
  • the enhancement layer refers to the layer with higher quality (including frame rate, spatial resolution, temporal resolution, signal-to-noise ratio intensity or quality level) in the layered coding.
  • the corresponding base layer may be any layer lower in quality than the enhancement layer, for example, if currently There are five layers, and the coding quality is sequentially improved (that is, the first layer has the lowest quality and the fifth layer has the highest quality).
  • the enhancement layer is the fourth layer
  • the base layer may be the first layer or the second layer. It is the third layer or the fourth layer.
  • the corresponding enhancement layer can be any layer of lower quality than the base layer.
  • the enhancement layer image is the image in the currently processed enhancement layer
  • the base layer image is the image in the base layer at the same time as the enhancement layer image.
  • the quality of the base layer image is lower than the quality of the enhancement layer image.
  • the target image block is the image block being processed in the enhancement layer image.
  • the base layer image block is an image block in the base layer image that has a corresponding relationship with the target image block in spatial position.
  • the correspondence between the image block in the base layer and the image block in the enhancement layer may be calculated according to the resolution proportional relationship between the base layer image and the enhancement layer image. For example, in a system including the X direction and the y direction, if the enhancement layer image is resolved in the X direction and the y direction The rate is 2 times of the image of the layer, and the image block with the pixel coordinates of the upper left corner of the enhancement layer is (2x, 2y) and the size is (2m) X (2n), and the corresponding block in the base layer image may be the upper left.
  • the sub-blocks described later refer to sub-blocks of the target image block (image blocks in the enhancement layer), and the corresponding sub-blocks described later refer to the image blocks corresponding to the sub-blocks in the base layer.
  • the motion information may include one or more of a prediction direction, a reference image index, or a motion vector, where the prediction direction may be divided into one-way and two-way prediction, and the one-way prediction may be further divided into forward prediction.
  • forward prediction refers to the use of a forward reference picture list, ie, a reference picture in list 0 to generate a prediction signal
  • backward prediction refers to the use of a backward reference picture list, ie, a reference picture in list 1 to generate a prediction.
  • Bidirectional prediction refers to the use of reference pictures in list 0 and list 1 to generate prediction signals; for unidirectional prediction, a reference picture index is required to indicate the reference picture selected in list 0 or listl.
  • each motion vector includes a horizontal direction component X and a vertical direction component y, which can be recorded as (X, y), for unidirectional prediction,
  • a motion vector indicates the displacement of the prediction signal in the selected list 0 or list 1 reference image, for bidirectional prediction,
  • listO list 1 reference picture and the reference picture respectively indicate the front and the rear to the prediction signal to the prediction signal in the selected displacement.
  • the target image block may be configured by at least two sub-blocks (ie, target image sub-blocks), wherein the size of the sub-block may be determined according to a preset value.
  • the sub-block size is 4x4 as an example. For example, if the size of the target image block is 16x 16, it can be determined that the target image block includes 16 sub-blocks (size 4x4). Therefore, in the embodiment of the present invention, a corresponding sub-block of each sub-block in the target image block (belonging to the corresponding image block) may be determined, and motion information of the corresponding sub-block is determined.
  • the coordinates of the corresponding position of the pixel in the base layer image may be determined according to the coordinates of a certain pixel point in the sub-block (denoted as: u ⁇ E x , E y ) n ) : u(B x , B y y ), and the image block in the base layer containing the corresponding position coordinates is used as the corresponding child, and can be calculated according to the following formulas 1 and 2: ( ⁇ , ⁇ ):
  • S is an accuracy control factor (for example, in the embodiment of the present invention, it can be set to 16), and D r can be calculated according to the following formula 5, and D y is calculated according to the following formula 6
  • ScaledBaseHeight where ⁇ represents the width of the base layer image, BiweH gfe represents the height of the base layer image, ScaledBaseWidth represents the width of the enhancement layer image, and ScaledBaseHeight represents the height of the enhancement layer image.
  • the corresponding sub-block can be determined, and in the case where the corresponding sub-block includes motion information, the prediction direction and the reference image index in the motion information can be directly used as the sub-block (first target image sub-block) Prediction direction and reference image index.
  • the motion vector of the corresponding sub-block can be according to the following Equations 7 to 10 Scaling is performed, and the scaled motion vector is taken as the motion vector (EMV x , EMV ⁇ ) of the sub-block (first target image sub-block).
  • EMV x (BMV x x ScaledBaseWidth + R BW ) / BaseWidth ( 7 )
  • EMV y (BMV y x ScaledBaseHeight + R BH ) I BaseHeight ( 8 )
  • R BH sgn(BMV y ) * BaseHeight 12 ( 10)
  • the motion information of the corresponding sub-block can be directly used as the motion information of the sub-block.
  • motion information of the corresponding sub-block may be
  • the method further includes: And determining, according to the encoding mode of the base layer image, whether the first base layer image sub-block corresponding to the first target image sub-block includes motion information.
  • whether the corresponding sub-block includes motion information may be determined according to an encoding mode of the base layer image (corresponding image block). For example, if the base layer image uses the intra prediction coding mode, it may be determined that the corresponding sub-block does not contain motion information (i.e., the motion information of the first basic layer image sub-block is empty).
  • the corresponding sub-block when the corresponding sub-block includes motion information, the corresponding sub-block may be determined according to the process as described above and its motion information may be acquired. When the corresponding sub-block does not include motion information, the above process may be skipped.
  • the second target image sub-block may be determined by the following method 1, and S120, according to the The motion information of the two target image sub-blocks acquires the first reference information.
  • the target image sub-block may be filled with motion information so that the filled motion information can be used as reference information.
  • the index allocation method of the sub-block may be the same as the prior art.
  • the description thereof is omitted, and the division and indexing of the sub-blocks are shown in FIG.
  • the processing level may be determined according to the size of the target image block and the size of the sub-block, and the processing layer is recursively layer by layer according to the processing level.
  • each processing unit (referred to as the first processing unit) in the lowest level hierarchy (referred to as the first layer) may be defined to include four sub-blocks, the previous layer of the first layer.
  • Each processing unit (referred to as the second layer) (denoted as the second layer) includes four first processing units, and so on, in order to avoid redundancy, the recursive description is omitted.
  • two levels may be included.
  • sub-block 0 ⁇ sub-block 3 constitute the first processing unit
  • sub-block 4 ⁇ sub The block 7 constitutes the first processing unit 1
  • the sub-block 8 ⁇ sub-block 11 constitutes the first processing unit 2
  • the sub-block 12 ⁇ sub-block 15 constitutes the first processing unit 3.
  • the first processing unit 0 to the first processing unit 3 constitute the second processing unit 0. It should be understood that the above-described hierarchical division method is merely illustrative, and the present invention is not limited thereto.
  • the index number of the sub-block may be used for each first processing unit. (for example, from small to large) sequentially determining whether the motion information of each sub-block is empty, and if the motion information of the sub-block is empty, it may be based on the sub-block adjacent to the first processing unit (second target image sub-).
  • the motion information of an example of the block determines its motion information. For example, if the motion information of the sub-block with index 0 (ie, sub-block 0, belonging to the first processing unit 0) is empty, motion information of other sub-blocks belonging to the same processing unit (first processing unit 0) may be acquired. And use the motion information as the motion information of the sub-block 0.
  • the obtaining order may be, for example, first obtaining motion information of a sub-block with an index of 1 (sub-block 1, that is, an example of a second target image sub-block, adjacent to sub-block 0 in the horizontal direction), if sub-block 1 If the motion information is empty, the motion information of the sub-block whose index is 2 (sub-block 2, that is, another example of the second target image sub-block, adjacent to the sub-block 0 in the vertical direction) may be acquired again. If the motion information of the sub-block 2 is empty, then the sub-block with the index of 3 (sub-block 3, that is, another example of the second target image sub-block, adjacent to the sub-block 0 in the diagonal direction) may be acquired. Sports information.
  • the motion information can be filled by the above method.
  • the method for filling the motion information of the sub-blocks whose motion information is empty is only an exemplary description, and the present invention is not limited thereto.
  • the vertical direction may also be obtained first.
  • the motion information of the sub-block (here, the adjacent sub-block), and the motion information of the specified sub-block (here, the adjacent sub-block) in the horizontal direction is acquired, and the specified sub-block in the diagonal direction is acquired (here , for adjacent sub-blocks) motion information. That is, the order of acquisition can be arbitrarily changed.
  • the motion information filled for the sub-block can be directly used. That is, the motion information of the second target image sub-block may refer to the motion information of the base layer corresponding sub-block of the second target image sub-block, or may be the method of filling the motion information according to the embodiment of the present invention. Image sub-block-filled motion information (from other enhancement layer sub-blocks).
  • the first processing unit may be in the same processing unit as the sub-block 0 from the first layer (for example, the first processing unit 0)
  • Other sub-blocks eg, sub-block 1 - sub-block 3 acquire motion information.
  • the sub-block for example, sub-block 0
  • the specified sub-block in the first layer for example, sub-block 1 ⁇ sub-block 3
  • the motion information is empty, you can get Taking motion information of a predetermined sub-block (another example of the second target image sub-block) in the predetermined first processing unit (for example, the first processing unit 1 to the first processing unit 3) in the second processing unit, and taking the motion
  • the information is the motion information of the sub-block (for example, sub-block 0).
  • first processing unit 0 if the corresponding sub-blocks of all the sub-blocks in one first processing unit (for example, the first processing unit 0) are all empty, other first processing units in the second processing unit may be acquired (for example, the first processing).
  • the motion information of the predetermined sub-block in the first to third processing units 3 for convenience of explanation, the sub-block in the upper left corner of each first processing unit is described as an example, and the motion information is used as the first Motion information of each sub-block within a processing unit (first processing unit 0).
  • the order of obtaining may be, for example, first obtaining a sub-block of the upper left corner of the first processing unit (the first processing unit 1 adjacent to the first processing unit 0 in the horizontal direction) having an index of 1 (sub-block 4, ie
  • the motion information of the second target image sub-block is empty. If the motion information of the sub-block 4 is empty, the motion information of the other sub-blocks in the first processing unit 1 is also considered to be empty, so that the index can be re-acquired.
  • a sub-block of the upper left corner of the first processing unit (first processing unit 2 adjacent to the first processing unit 0 in the vertical direction) of 2 (subblock 8, that is, another example of the second target image sub-block)
  • the motion information of the sub-block 8 is considered to be empty if the motion information of the sub-block 8 is empty, so that the motion information of the other sub-blocks in the first processing unit 2 is also empty, so that the first processing unit with the index of 3 can be reacquired (first The processing unit 3, which is adjacent to the first processing unit 0 in the diagonal direction, has motion information of the sub-block in the upper left corner (sub-block 12, that is, another example of the second target image sub-block).
  • the motion information can be filled by the above method.
  • the method for filling the motion information of the sub-blocks whose motion information is empty is only an exemplary description, and the present invention is not limited thereto.
  • the vertical direction may also be obtained first. Obtaining motion information of a predetermined sub-block of the first processing unit (here, adjacent to the first processing unit), and acquiring a predetermined sub-block of a predetermined first processing unit (here, an adjacent first processing unit) in the horizontal direction The motion information further acquires motion information of a prescribed sub-block of the first processing unit (here, the adjacent first processing unit) in the diagonal direction.
  • predetermined sub-block is not limited to the sub-block in the upper left corner of the first processing unit, and may be a sub-block at an arbitrary position in the same first processing unit.
  • the target image block enumerated above is 16 x 16 and the size of the sub-block is 4 x 4, the target image block includes only two layers, and the above recursive process ends.
  • the size of the target image block is larger, for example, 32 x 32, and the size of the sub-block is 4 x 4, then the target image block includes three layers, and the recursive operation can be continued in the same manner as described above.
  • Image block All sub-blocks get motion information.
  • the second target image sub-block includes:
  • Idx 2 + ⁇ idx, %N/(N/2))x2 + (l- Idx, %N /(N / 4)%2)) ⁇ N /4 ; Idx 2 + ((l - ldx x %N/(N/2))x2 + ⁇ ldx x %N / (N / 4) %2)) xN/4; W3 ⁇ 4 / ⁇ /NxN + ((l-/ x 1 N/(N/2))x2 + (l-/ ⁇ N/(N/4) 2))xN/4; wherein, the representation is used to indicate that the second target image sub-block is in the target image block
  • the third indication information of the location, / ⁇ indicates the second indication information, and N is determined according to the size of the target image block and the size of the target image sub-block.
  • the third index information indicating the position of the second target image sub-block in the target image block is indicated, / ⁇ indicates the second index information, % indicates a modulo operation or a remainder operation, and N indicates the target image The number of sub-blocks included in the block.
  • the second target image sub-block within the currently processed hierarchy may be determined according to the index of the sub-block currently being processed, where ⁇ corresponds to the layer currently being processed, and ⁇ is according to the target
  • the size of the image block and the size of the sub-block are determined. For example, if the size of the target image block is 16x 16, and the size of the sub-block is 4x4, the target image block includes two layers as described above, and when processing the first layer, The number of sub-blocks included in each processing unit (first processing unit) in the layer, here is 4. When processing the second layer, ⁇ is the number of sub-blocks included in each processing unit (second processing unit) in the layer, here 16.
  • determining the first reference information of the first target image sub-block according to the motion information of the second target image sub-block includes:
  • the motion information of the second target image sub-block is empty, it is determined that the first reference information is zero motion information.
  • zero motion information is used as the motion information of the sub-block.
  • zero motion information can be constructed in the following manner. For example, in predictive coded image frames (P frames), the predicted side of zero motion information For unidirectional prediction, the reference image index is 0, motion vector (0, 0). In the bidirectionally predictive coded image frame (B frame), the prediction direction of the zero motion information is bidirectional prediction, both reference image indices are 0, and both motion vectors are (0, 0).
  • the foregoing method of using the zero motion information as the motion information of the sub-block may be performed after processing the last layer, or may be performed on any other layer.
  • the treatment is carried out, and the present invention is not particularly limited.
  • the method for obtaining the motion information of the sub-blocks listed above is only an exemplary description of the present invention, and the present invention is not limited thereto.
  • the coding mode of the layer image determines whether the corresponding sub-block contains motion information.
  • the base layer image uses the intra prediction coding mode, it may be determined that the corresponding sub-block does not contain motion information (i.e., the motion information of the first base layer image sub-block is empty). If it is determined that only one of the sub-blocks of the target image block (specifically, its corresponding sub-block) has motion information, the motion information of the sub-block may be used as motion information of the other sub-blocks.
  • the first reference information of the first target image sub-block can be determined.
  • the first reference information of the first target image block (the corresponding sub-block in the base layer does not include motion information) is acquired by the above method 1.
  • the encoding the target image block includes:
  • the first target image sub-block may be encoded (specifically, motion compensation processing) based on reference information (specifically, motion information) of the first target image sub-block.
  • independent motion compensation processing may be performed on the first target image sub-block according to the motion information acquired or filled for the first target image sub-block as described above.
  • the motion information of the third target image block (the corresponding sub-block in the base layer includes motion information) may be acquired, and the method for acquiring the motion information of the third target image block may be the same as the prior art. The description thereof is omitted here.
  • the target image block after performing motion compensation processing on all the sub-blocks of the target image block, after acquiring the prediction signal of the target image block, the target image block can be predictively encoded, thereby calculating the rate distortion cost.
  • an indication identifier (first indication information) can be determined to indicate the decoding end, and the first target image block is obtained by the above method 1 and method 2.
  • the encoding, according to the reference information, the first target image sub-block comprising: entropy encoding the first indication information, so that the first indication information is in the target code stream and the skip mode
  • the flag bit or the merged MERGE mode flag bit information is adjacent.
  • the first indication information may be configured at a position adjacent to the skip mode flag bit information.
  • the first indication information may be placed as the first information of the target image block in the target code stream before the skip mode flag bit information, and the first indication information may be placed as the target after the skip mode flag bit information.
  • the first indication information can be placed adjacent to the MERGE mode flag bit information.
  • the first indication information may be placed before the MERGE mode flag bit information, and the first indication information may be placed after the MERGE mode flag bit information.
  • the MERGE may be the same mode as the prior art, and the determination method and the configuration position may be the same as the prior art.
  • the description thereof is omitted.
  • the first indication information may be a binary flag bit. Therefore, when the first indication information is entropy encoded, it is not necessary to perform binarization processing on the first indication information.
  • the context used for entropy encoding the first indication information of the binarization may be selected, and optionally, encoding the first target image sub-block according to the reference information, including: according to the enhancement Whether the reference image block located in the preset position in the layer image is encoded using the reference information to determine the context;
  • the first indication information is entropy encoded.
  • the context can contain three contexts of 0, 1, and 2. This embodiment determines which context to use based on whether the left and upper image blocks use the base layer mode. In this embodiment, for example, depending on whether the image blocks on the left and the top of the target image block use their respective first indication information, which context is used, for example, if the image blocks on the left and above the target image block are not Using the first indication information, then selecting a context model with an index of 0. If the image block on the left and top of the target image block has a first indication information, then select a context model with an index of 1 if the target image block is left and above. The image block uses the first indication information, then the context with the index of 2 is selected. Table 1 0 1 2 Intra coded slice is undefined (or unused) CNU CNU
  • the first indication information can be binary arithmetic coded according to the context selected as described above, and the used context model is updated.
  • the process may be the same as the prior art, and the description thereof is omitted here to avoid redundancy.
  • each of the first target image sub-blocks may be encoded according to the obtained reference information, thereby completing encoding of the target image block, and generating the information of the target image block and the entropy-encoded information.
  • the first indication information is added to the code stream (target stream).
  • the encoded target image including the base layer image and the enhancement layer image
  • the processing procedure may be the same as the prior art, where , the description is omitted.
  • the target image information may be acquired from the code stream, and the target image (specifically, the target image block) is determined, and the obtained first indication information (entropy-encoded information) is entropy solved or similar. The description thereof is omitted here.
  • a binary symbol string (bin string) representing the first indication information may be parsed from the code stream according to the selected context.
  • the binary arithmetic decoding corresponds to binary arithmetic coding at the encoding end.
  • the decoding end needs to obtain the first reference information of the first target image sub-block by using the same method as the encoding end, and the decoding end needs to use The first reference information of the first target image sub-block is acquired in the same manner as the encoding end.
  • the method of indicating the first indication information listed above is merely illustrative, and the present invention is not limited thereto. It should be noted that, in the embodiment of the present invention, when the reference information of the first target image sub-block is acquired by using the method 1, the reference information (specifically, the motion information) of the first target image sub-block may be used.
  • the first target image sub-block is decoded (specifically, motion compensation processing). Specifically, independent motion compensation processing may be performed on the first target image sub-block according to the motion information filled in the first target image sub-block as described above.
  • the encoding the target image block includes:
  • Deblocking filtering processing is performed on pixels located near the boundary between the target image sub-blocks. Specifically, it is also possible to perform filtering processing on pixels in the vicinity of the boundary between the sub-blocks of the target image block.
  • a method for image processing according to an embodiment of the present invention for a first target image sub-block in which a motion information cannot be obtained from a corresponding sub-block included in a base layer image in a target image block of an enhancement layer image, according to the first target image Positioning the sub-block determines a second target image sub-block, and determining first reference information for the first target image sub-block according to the motion information of the second target image sub-block, and encoding according to the first reference information, The coding performance of the first target image sub-block can be improved.
  • Fig. 3 shows a schematic flow chart of a method 200 for image processing according to an embodiment of the present invention, which is described from the perspective of a decoding end. As shown in FIG. 2, the method 200 includes:
  • S230 Determine first reference information used to decode the first target image sub-block according to the motion information of the second target image sub-block, where the first base layer image sub-block is located in the base layer image.
  • the sub-block corresponds to a spatial position in the enhancement layer image;
  • the decoding end may acquire the target image information from the code stream, and determine the target image (specifically, the target image block), and acquire the first indication information (entropy-encoded information).
  • the obtaining the first indication information from the target code stream includes:
  • the first indication information is obtained from the target code stream, where the first indication information is adjacent to the skip mode flag bit or the merged MERGE mode flag bit information in the target code stream.
  • the first indication information may be configured at a position adjacent to the skip mode flag bit information.
  • the first indication information may be placed as the first information of the target image block in the target code stream before the skip mode flag bit information, and the first indication information may be placed as the target after the skip mode flag bit information.
  • the first indication information can be placed adjacent to the MERGE mode flag bit information.
  • the first indication information may be placed before the MERGE mode flag bit information, and the first indication information may be placed after the MERGE mode flag bit information.
  • the MERGE may be the same as the prior art, and the determination method and the configuration position may be the same as those in the prior art. Here, in order to avoid redundancy, the description thereof is omitted.
  • the decoding end may perform entropy decoding on the obtained first indication information.
  • the first indication information may be a binary flag bit. Therefore, when entropy decoding the first indication information, it is not necessary to binarize the first indication information.
  • the upper and lower texts used for entropy encoding the first indication information of the binarization may be selected, and optionally, the first indication information is obtained from the target code stream, including:
  • entropy decoding is performed to determine the first indication information.
  • the context can contain three contexts of 0, 1, and 2.
  • This embodiment determines which context to use based on whether the left and upper image blocks use the base layer mode. In this embodiment, for example, depending on whether the image blocks on the left and the top of the target image block use their respective first indication information, which context is used, for example, if the image blocks on the left and above the target image block are not Using the first indication information, then selecting a context model with an index of 0. If the image block on the left and top of the target image block has a first indication information, then select a context model with an index of 1 if the target image block is left and above. The image block uses the first indication information, then the context with the index of 2 is selected.
  • the first indication information can be binary arithmetically decoded according to the context selected as above, and the used context model is updated.
  • the process may be the same as the prior art, and the description thereof is omitted here to avoid redundancy.
  • the decoding end needs to obtain the first reference information of the first target image sub-block by using the same method as the encoding end, and the decoding end needs to use The first reference information of the first target image sub-block is acquired in the same manner as the encoding end.
  • the decoding end may determine, according to the first indication information, whether the first reference information of the first target image sub-block needs to be acquired, and the following is a description of the case where the encoding end needs to acquire the first reference information of the first target image sub-block. .
  • the image can be processed to obtain a low resolution image, and the original image is referred to as a high resolution image as a contrast, and the encoder separately determines the low resolution.
  • the rate image and the high resolution image are encoded.
  • a high quality image to be encoded is referred to herein as an enhancement layer image
  • a corresponding low quality image to be encoded is referred to as a base layer image.
  • the target image is an image processed by a layered coding technique
  • the basic layer refers to a quality in layered coding (including frame rate, spatial resolution, temporal resolution, signal to noise ratio intensity or quality level).
  • the enhancement layer refers to the layer with higher quality (including frame rate, spatial resolution, temporal resolution, signal-to-noise ratio intensity or quality level) in the layered coding.
  • the corresponding base layer may be any layer lower in quality than the enhancement layer, for example, if currently There are five layers, and the coding quality is sequentially improved (that is, the first layer has the lowest quality and the fifth layer has the highest quality).
  • the enhancement layer is the fourth layer
  • the base layer may be the first layer or the second layer. It is the third layer or the fourth layer.
  • the corresponding enhancement layer can be any layer of lower quality than the base layer.
  • the enhancement layer image is the image in the currently processed enhancement layer
  • the base layer image is the image in the base layer at the same time as the enhancement layer image.
  • the quality of the base layer image is lower than the quality of the enhancement layer image.
  • the target image block is the image block being processed in the enhancement layer image.
  • the base layer image block is an image block in the base layer image that has a corresponding relationship with the target image block in spatial position.
  • the correspondence between the image block in the base layer and the image block in the enhancement layer can be calculated according to the resolution proportional relationship between the base layer image and the enhancement layer image. For example, in a system including an X direction and a y direction, if the resolution of the enhancement layer image in the X direction and the y direction is twice that of the original layer image, the pixel coordinates of the upper left corner in the enhancement layer are (2x, 2y). And an image block of size (2m ) X ( 2n ), the corresponding block in the base layer image may be an image block whose pixel coordinates are (X , y ) and whose size is mxn in the upper left corner.
  • the sub-blocks described later refer to sub-blocks of the target image block (image blocks in the enhancement layer), and the corresponding sub-blocks described later refer to the image blocks corresponding to the sub-blocks in the base layer.
  • the motion information may include one or more of a prediction direction, a reference image index, or a motion vector, where the prediction direction may be divided into one-way and two-way prediction, and the one-way prediction may be further divided into forward prediction.
  • forward prediction refers to the use of a forward reference picture list, ie, a reference picture in list 0 to generate a prediction signal
  • backward prediction refers to using a backward reference picture list, ie, a reference picture in list 1 to generate a prediction.
  • Bidirectional prediction refers to the use of reference pictures in list 0 and list 1 to generate prediction signals; for unidirectional prediction, a reference picture index is required to indicate the reference picture selected in list 0 or list l.
  • each motion vector includes a horizontal direction component X and a vertical direction component y, which can be written as (X, y), for one-way prediction, Requires a motion vector to indicate the displacement of the predicted signal in the selected list 0 or list 1 reference image, for bidirectional prediction
  • Two motion vectors are required to indicate the displacement of the forward prediction signal and the backward prediction signal in the selected list O reference image and the list 1 reference image, respectively.
  • the target image block may be configured by at least two sub-blocks (ie, target image sub-blocks), wherein the size of the sub-block may be determined according to a preset value.
  • the sub-block size is 4 x 4 as an example. For example, if the size of the target image block is 16 16, it can be determined that the target image block includes 16 sub-blocks (size 4 x 4). Therefore, in the embodiment of the present invention, a corresponding sub-block (a first base layer image sub-block) in each of the target image blocks (the first target image sub-block) in the base layer may be determined, and the Corresponding to the motion information of the sub-block.
  • a first base layer image sub-block corresponding to the first target image sub-block may be determined, and in a case where the first base layer image sub-block includes motion information, prediction in the motion information may be directly used
  • the direction and reference image index are used as the prediction direction of the sub-block (first target image sub-block) and the reference image index.
  • the motion vector ( ⁇ , ⁇ ) of the first base layer image sub-block may be scaled according to the following Equations 7 to 10, and the scaled motion vector is used as the sub-block (first target image sub-block) Motion vector (EMV x , EMV y ).
  • the motion information of the corresponding sub-block can be directly used as the motion information of the sub-block.
  • motion information of the corresponding sub-block may be
  • the method further includes:
  • the first base layer image sub-block corresponding to the first target image sub-block includes motion information.
  • whether the corresponding sub-block includes motion information may be determined according to an encoding mode of the base layer image (corresponding image block). For example, if the base layer image uses the intra prediction coding mode, it may be determined that the corresponding sub-block does not contain motion information (i.e., the motion information of the first basic layer image sub-block is empty).
  • the corresponding sub-block when the corresponding sub-block includes motion information, the corresponding sub-block may be determined according to the above process and its motion information is obtained. When the corresponding sub-block does not include motion information, the above procedure may be skipped.
  • the second target image sub-block may be determined by the following method 2, and S230, according to the The motion information of the two target image sub-blocks acquires the first reference information.
  • the target image sub-block may be filled with motion information, so that the filled motion information can be used For reference information.
  • the index allocation method of the sub-block may be the same as the prior art.
  • the description thereof is omitted, and the division and indexing of the sub-blocks are shown in FIG.
  • the processing level may be determined according to the size of the target image block and the size of the sub-block, and the processing layer is recursively layer by layer according to the processing level.
  • each processing unit (referred to as the first processing unit) in the lowest level hierarchy (referred to as the first layer) may be defined to include four sub-blocks, the previous layer of the first layer.
  • Each processing unit (referred to as the second layer) (denoted as the second layer) includes four first processing units, and so on, in order to avoid redundancy, the recursive description is omitted.
  • two levels may be included.
  • sub-block 0 ⁇ sub-block 3 constitute the first processing unit 0
  • the block 7 constitutes the first processing unit 1
  • the sub-block 8 ⁇ sub-block 11 constitutes the first processing unit 2
  • the sub-block 12 ⁇ sub-block 15 constitutes the first processing unit 3.
  • the first processing unit 0 to the first processing unit 3 constitute the second processing unit 0. It should be understood that the above-described hierarchical division method is merely illustrative, and the present invention is not limited thereto.
  • the motion information of each sub-block is empty according to the index number of the sub-block (for example, from small to large) may be determined, if the motion information of the sub-block is empty. Then, motion information of the sub-block (an example of the second target image sub-block) adjacent to the first processing unit may be determined based on the motion information. For example, if the motion information of the sub-block with index 0 (ie, sub-block 0, belonging to the first processing unit 0) is empty, motion information of other sub-blocks belonging to the same processing unit (first processing unit 0) may be acquired. And use the motion information as the motion information of the sub-block 0.
  • the obtaining order may be, for example, first obtaining motion information of a sub-block with an index of 1 (sub-block 1, that is, an example of a second target image sub-block, adjacent to sub-block 0 in the horizontal direction), if sub-block 1 If the motion information is empty, the motion information of the sub-block whose index is 2 (sub-block 2, that is, another example of the second target image sub-block, adjacent to the sub-block 0 in the vertical direction) may be acquired again. If the motion information of the sub-block 2 is empty, then the sub-block with the index of 3 (sub-block 3, that is, another example of the second target image sub-block, adjacent to the sub-block 0 in the diagonal direction) may be acquired. Sports information.
  • the motion information can be filled by the above method.
  • the method for filling the motion information of the sub-blocks whose motion information is empty is only an exemplary description, and the present invention is not limited thereto.
  • the motion information of the specified sub-block here, the adjacent sub-block. That is, the order of acquisition can be arbitrarily changed.
  • the motion information filled for the sub-block can be directly used. That is, the motion information of the second target image sub-block may refer to the motion information of the base layer corresponding sub-block of the second target image sub-block, or may be the method of filling the motion information according to the embodiment of the present invention. Image sub-block-filled motion information (from other enhancement layer sub-blocks).
  • the first processing unit may be in the same processing unit as the sub-block 0 from the first layer (for example, the first processing unit 0)
  • Other sub-blocks eg, sub-block 1 - sub-block 3 acquire motion information.
  • the sub-block for example, sub-block 0
  • the specified sub-block in the first layer for example, sub-block 1 ⁇ sub-block 3
  • the predetermined sub-block in the predetermined first processing unit for example, the first processing unit 1 to the first processing unit 3 in the second processing unit may be acquired (another example of the second target image sub-block)
  • the motion information is used as the motion information of the sub-block (for example, sub-block 0).
  • first processing unit 0 if the corresponding sub-blocks of all the sub-blocks in one first processing unit (for example, the first processing unit 0) are all empty, other first processing units in the second processing unit may be acquired (for example, the first processing).
  • the motion information of the predetermined sub-block in the first to third processing units 3 for convenience of explanation, the sub-block in the upper left corner of each first processing unit is described as an example, and the motion information is used as the first Motion information of each sub-block within a processing unit (first processing unit 0).
  • the order of obtaining may be, for example, first obtaining a sub-block of the upper left corner of the first processing unit (the first processing unit 1 adjacent to the first processing unit 0 in the horizontal direction) having an index of 1 (sub-block 4, ie
  • the motion information of the second target image sub-block is empty. If the motion information of the sub-block 4 is empty, the motion information of the other sub-blocks in the first processing unit 1 is also considered to be empty, so that the index can be re-acquired.
  • a sub-block of the upper left corner of the first processing unit (first processing unit 2 adjacent to the first processing unit 0 in the vertical direction) of 2 (subblock 8, that is, another example of the second target image sub-block) Motion information If the motion information of the sub-block 8 is empty, it can be considered that the motion information of other sub-blocks in the first processing unit 2 is also empty, so that the first processing unit with the index of 3 can be re-acquired (the first processing unit 3, The motion information of the sub-block (sub-block 12, that is, another example of the second target image sub-block) of the upper left corner adjacent to the first processing unit 0 in the diagonal direction. Similarly, for the first processing unit whose motion information is empty, the motion information can be filled by the above method.
  • the method for filling the motion information of the sub-blocks whose motion information is empty is only an exemplary description, and the present invention is not limited thereto.
  • the vertical direction may also be obtained first.
  • the motion information further acquires motion information of a prescribed sub-block of the first processing unit (here, the adjacent first processing unit) in the diagonal direction. That is, the order of acquisition can be arbitrarily changed.
  • the above-mentioned "predetermined sub-block” is not limited to the sub-block in the upper left corner of the first processing unit, and may be a sub-block at an arbitrary position in the same first processing unit.
  • the target image block since the size of the target image block enumerated above is 16x 16, and the size of the sub-block is 4x 4, the target image block includes only two layers, and the above recursion process ends. However, the size of the target image block is larger, for example, 32x 32, and the size of the sub-block is 4x4, then the target image block includes three layers, and the recursive operation can be continued in the same manner as described above, as the target image block All sub-blocks get motion information.
  • the second target image sub-block includes:
  • Idx 2 + ⁇ idx, %N/(N/2))x2 + (l- Idx, %N /(N / 4)%2)) ⁇ N /4 ; Idx 2 + ((l - ldx x %N/(N/2))x2 + (ldx x %N / (N / 4) %2)) xN/4; W3 ⁇ 4 / ⁇ /NxN + ((l-/ 1 %N/(N/2))x2 + (l-/ ⁇ %N/(N/4)%2))xN/4; wherein, the representation is used to indicate that the second target image sub-block is in the target image block
  • the third indication information of the position in the /, the second indication information, N is determined according to the size of the target image block and the size of the target image sub-block.
  • indicating third index information indicating a position of the second target image sub-block in the target image block indicates the second index information, % indicates a modulo operation or a remainder operation, and N indicates that the target image block includes The number of sub-blocks.
  • the second target image sub-block within the currently processed hierarchy may be determined according to the index of the sub-block currently being processed, where N corresponds to the layer currently being processed, and N is according to the target
  • the size of the image block and the size of the sub-block are determined. For example, if the size of the target image block is 16 x 16 and the size of the sub-block is 4 x 4, the target image block includes two layers as described above, and the first layer is processed.
  • N is the number of sub-blocks included in each processing unit (first processing unit) in the layer, here is 4.
  • N is the number of sub-blocks included in each processing unit (second processing unit) in the layer, here 16.
  • determining the first reference information of the first target image sub-block according to the motion information of the second target image sub-block includes:
  • the motion information of the second target image sub-block is empty, it is determined that the first reference information is zero motion information.
  • zero motion information is used as the motion information of the sub-block.
  • zero motion information can be constructed in the following manner. For example, in a predictive coded picture frame (P frame), the prediction direction of the zero motion information is unidirectional prediction, and the reference picture index is 0, motion vector (0, 0). In the bidirectionally predictive coded picture frame (B frame), the prediction direction of the zero motion information is bidirectional prediction, both reference picture indices are 0, and both motion vectors are (0, 0).
  • the foregoing method of using the zero motion information as the motion information of the sub-block may be performed after processing the last layer, or may be performed on any other layer.
  • the treatment is carried out, and the present invention is not particularly limited.
  • the method for obtaining the motion information of the sub-blocks listed above is only an exemplary description of the present invention, and the present invention is not limited thereto.
  • the coding mode of the layer image determines whether the corresponding sub-block contains motion information.
  • the base layer image uses the intra prediction coding mode, it may be determined that the corresponding sub-block does not contain motion information (i.e., the motion information of the first base layer image sub-block is empty). If it is determined that only one of the sub-blocks of the target image block (specifically, its corresponding sub-block) has motion information, the motion information of the sub-block may be used as motion information of the other sub-blocks.
  • the first reference information of the first target image sub-block can be determined.
  • the first target sub-block may be decoded according to the obtained first reference information, for example, according to the first reference information (specifically, motion information) of the first target image sub-block, the first The target image sub-block is decoded (specifically, motion compensation processing).
  • the first target image sub-block may be subjected to motion compensation processing according to the motion information filled for the first target image sub-block as described above.
  • the motion information can be obtained by the same method as the prior art, and The decoding is performed, and the process can be the same as in the prior art.
  • the description thereof will be omitted.
  • decoding the target image block according to the reference information comprises: performing deblocking filtering processing on pixels located near a boundary between the target image sub-blocks.
  • a method for image processing according to an embodiment of the present invention for a first target image sub-block in which a motion information cannot be obtained from a corresponding sub-block included in a base layer image in a target image block of an enhancement layer image, according to the first target image Positioning the sub-block determines a second target image sub-block, and determining reference information for the first target image sub-block according to the motion information of the second target image sub-block, and encoding according to the reference information, the first The coding performance of the target image sub-block.
  • FIGS. 1 to 3 a method for image processing according to an embodiment of the present invention is described in detail with reference to FIGS. 1 to 3.
  • a device for image processing according to an embodiment of the present invention will be described in detail with reference to FIGS. 4 to 5. .
  • FIG. 4 shows a schematic block diagram of an apparatus 300 for image processing in accordance with an embodiment of the present invention.
  • the apparatus 300 includes:
  • the acquiring unit 310 is configured to, when determining that the motion information of the first base layer image sub-block corresponding to the first target image sub-block of the target image block is empty, according to the size of the target image block, the target image block includes Determining a second target image sub-block, the size of each target image sub-block and second indication information indicating a position of the first target image sub-block in the target image block;
  • the target image block being located in the enhancement layer image, the base layer image corresponding to the enhancement layer image, and a spatial position of the first basic image block sub-block in the base layer image and the first target image
  • the sub-block corresponds to a spatial position in the enhancement layer image
  • the encoding unit 320 encodes the target image block to generate a target code stream and includes the The first indication information in the target code stream.
  • the obtaining unit 310 is specifically configured to determine the second target image sub-block according to any one of the following formulas.
  • Idx 2 / 2) ⁇ 2 + (1 - Idx l %N /(N/4) %2)) xN/4; Idx 2 (N/2))x2 + (ldx x %N / (N / 4) %2)) xN/4;
  • W3 ⁇ 4 / ⁇ /NxN + ((l-/ x 1 N/(N/2))x2 + (l-/ ⁇ N/(N/4) 2))xN/4; where The third indication information of the position of the second target image sub-block in the target image block, / represents the second indication information, and N is determined according to the size of the target image block and the size of the target image sub-block.
  • the acquiring unit 310 is specifically configured to determine that the first reference information is zero motion information if the motion information of the second target image sub-block is empty.
  • the encoding unit 320 is specifically configured to perform motion compensation processing on the first target image sub-block according to the first reference information.
  • the encoding unit 320 is further configured to perform deblocking filtering processing on pixels located near a boundary between the target image sub-blocks.
  • the encoding unit 320 is specifically configured to entropy encode the first indication information, so that the first indication information is adjacent to the skip mode flag bit or the merged MERGE mode flag bit information in the target code stream.
  • the encoding unit 320 is specifically configured to determine, according to whether the reference image block located in the preset position in the enhancement layer image is encoded by using reference information, to determine a context;
  • the apparatus 300 for image processing according to an embodiment of the present invention may correspond to an encoding end in the method of the embodiment of the present invention, and each unit in the image processing apparatus 300, that is, a module and the above other operations and/or functions respectively In order to implement the corresponding process of the method 100 in FIG. 1, for the sake of cleaning, no further details are provided herein.
  • An apparatus for image processing for a first target image sub-block in which a motion information cannot be obtained from a corresponding sub-block included in a base layer image in a target image block of an enhancement layer image, by according to the first target image Positioning the sub-block determines a second target image sub-block, and determining reference information for the first target image sub-block according to the motion information of the second target image sub-block, and encoding according to the reference information, thereby improving the The coding performance of a target image sub-block.
  • FIG. 5 shows a schematic block of an apparatus 400 for image processing in accordance with an embodiment of the present invention.
  • the apparatus 400 includes:
  • the decoding unit 410 is configured to obtain first indication information from the target code stream.
  • the obtaining unit 420 configured to: when the motion information of the first base layer image sub-block corresponding to the first target image sub-block of the target image block is empty, based on the first indication information acquired by the decoding unit, according to the target Determining a second target image sub-block, the size of the image block, the size of each target image sub-block included in the target image block, and second indication information indicating a position of the first target image sub-block in the target image block ;
  • the target image block being located in the enhancement layer image, the base layer image corresponding to the enhancement layer image, and a spatial position of the first basic image block sub-block in the base layer image and the first target image
  • the sub-block corresponds to a spatial position in the enhancement layer image
  • the decoding unit 410 is further configured to decode the target code stream to obtain the target image block.
  • the acquiring unit 420 is specifically configured to determine the second target image sub-block according to any one of the following formulas.
  • the acquiring unit 420 is specifically configured to determine that the first reference information is zero motion information if the motion information of the second target image sub-block is empty.
  • the decoding unit 410 is specifically configured to perform motion compensation processing on the first target image sub-block according to the first reference information.
  • the decoding unit 410 is further configured to perform deblocking filtering processing on pixels located near a boundary between the target image sub-blocks.
  • the decoding unit 410 is specifically configured to obtain, from the target code stream, the first indication information, where the first indication information is in the target code stream, and the skip mode flag bit or the merged MERGE mode flag bit information is adjacent.
  • the decoding unit 410 is specifically configured to be located at a preset location according to the enhancement layer image. Whether the reference image block is decoded using the reference information to determine the context;
  • the apparatus 400 for image processing may correspond to a decoding end in the method of the embodiment of the present invention, and the units in the apparatus 400 for image processing, that is, modules and other operations and/or functions described above In order to implement the corresponding process of the method 200 in FIG. 3, respectively, for the sake of cleaning, no further details are provided herein.
  • An apparatus for image processing for a first target image sub-block in which a motion information cannot be obtained from a corresponding sub-block included in a base layer image in a target image block of an enhancement layer image, by according to the first target image Positioning the sub-block determines a second target image sub-block, and determining reference information for the first target image sub-block according to the motion information of the second target image sub-block, and encoding according to the reference information, thereby improving the The coding performance of a target image sub-block.
  • Fig. 6 shows a schematic block diagram of an encoder 500 for image processing in accordance with an embodiment of the present invention.
  • the encoder 500 can include:
  • processor 520 connected to the bus
  • the processor 520 calls, by the bus 510, a program stored in the memory 530, for determining that the motion information of the first base layer image sub-block corresponding to the first target image sub-block of the target image block is Space-time, determining according to the size of the target image block, the size of each target image sub-block included in the target image block, and the second indication information indicating the position of the first target image sub-block in the target image block.
  • a second target image sub-block
  • the target image block being located in the enhancement layer image, the base layer image corresponding to the enhancement layer image, and a spatial position of the first basic image block sub-block in the base layer image and the first target image
  • the sub-block corresponds to a spatial position in the enhancement layer image
  • the processor 520 is specifically configured to determine the second target image sub-block according to any one of the following formulas,
  • the processor 520 is specifically configured to determine that the first reference information is zero motion information if the motion information of the second target image sub-block is empty.
  • the processor 520 is specifically configured to perform motion compensation processing on the first target image sub-block according to the first reference information.
  • the processor 520 is specifically configured to perform deblocking filtering processing on pixels located near a boundary between the target image sub-blocks.
  • the processor 520 is specifically configured to entropy encode the first indication information, so that the first indication information is adjacent to the skip mode flag bit or the merged MERGE mode flag bit information in the target code stream.
  • the processor 520 is specifically configured to determine, according to whether the reference image block located in the preset position in the enhancement layer image is encoded by using reference information, to determine a context;
  • the encoder 500 for image processing may correspond to an encoding end in the method of the embodiment of the present invention, and each unit in the encoder 500 for image processing, that is, a module and the above-described other operations and/or For the purpose of implementing the corresponding process of the method 100 in FIG. 1 , the functions are not described here.
  • An encoder for image processing for a first target image sub-block in which a motion information cannot be obtained from a corresponding sub-block included in a base layer image in a target image block of an enhancement layer image, according to the first target Positioning the image sub-block determines a second target image sub-block, and determining reference information for the first target image sub-block according to the motion information of the second target image sub-block, and encoding according to the reference information, The coding performance of the first target image sub-block.
  • FIG. 7 shows a schematic block diagram of a decoder 600 for image processing in accordance with an embodiment of the present invention.
  • the decoder 600 can include: Bus 610;
  • processor 620 connected to the bus
  • the processor 620 by using the bus 610, invokes a program stored in the memory 630 to obtain first indication information from the target code stream.
  • the target image block Determining a second target image sub-block by including a size of each target image sub-block and second indication information indicating a position of the first target image sub-block in the target image block;
  • the target image block being located in the enhancement layer image, the base layer image corresponding to the enhancement layer image, and a spatial position of the first basic image block sub-block in the base layer image and the first target image
  • the sub-block corresponds to a spatial position in the enhancement layer image
  • the processor 620 is specifically configured to determine the second target image sub-block according to any one of the following formulas.
  • Idx 2 + ⁇ idx, %N/(N/2))x2 + (l- Idx, %N /(N / 4)%2)) ⁇ N /4 ; Idx 2 + ((l - ldx x %N/(N/2))x2 + ⁇ ldx x %N / (N / 4) %2)) xN/4; W3 ⁇ 4 / ⁇ /NxN + ((l-/ x 1 N/(N/2))x2 + (l-/ ⁇ N/(N/4) 2))xN/4; wherein, the representation is used to indicate that the second target image sub-block is in the target image block
  • the third indication information of the location, / represents the second indication information, and N is determined according to the size of the target image block and the size of the target image sub-block.
  • the processor 620 is specifically configured to determine that the first reference information is zero motion information if the motion information of the second target image sub-block is empty.
  • the processor 620 is specifically configured to perform motion compensation processing on the first target image sub-block according to the first reference information.
  • the processor 620 is specifically configured to perform deblocking filtering processing on pixels located near a boundary between the target image sub-blocks.
  • the processor 620 is specifically configured to obtain, from the target code stream, the first indication information, where the first indication information is in the target code stream, and the skip mode flag bit or the merged MERGE mode.
  • the flag information is adjacent.
  • the processor 620 is specifically configured to determine, according to whether the reference image block located in the preset position in the enhancement layer image is decoded by using reference information, to determine a context;
  • the decoder 600 for image processing may correspond to a decoding end in the method of the embodiment of the present invention, and each unit in the decoder 600 for image processing, that is, a module and the other operations described above and/or For the purpose of implementing the corresponding process of the method 200 in FIG. 5, the functions are not described here.
  • a decoder for image processing for a first target image sub-block in which a motion information cannot be obtained from a corresponding sub-block included in a base layer image in a target image block of an enhancement layer image, according to the first target Positioning the image sub-block determines a second target image sub-block, and determining reference information for the first target image sub-block according to the motion information of the second target image sub-block, and encoding according to the reference information, The coding performance of the first target image sub-block.
  • Fig. 8 shows a schematic flow chart of a method 700 for image processing according to an embodiment of the present invention, which is described from the perspective of an encoding end. As shown in FIG. 8, the method 700 includes:
  • the motion information of the first base layer image sub-block corresponding to the first target image sub-block of the target image block is empty, determining, according to the reconstructed pixel of the first base layer image sub-block, a second reference information encoded by the first target image sub-block, wherein the first base layer image sub-block is an image block located in the base layer image, the target image block is located in the enhancement layer image, and the base layer image is The enhancement layer image corresponds to, and a spatial position of the first basic image block sub-block in the base layer image corresponds to a spatial position of the first target image sub-block in the enhancement layer image;
  • S720 Encode the target image block to generate a target code stream and fourth indication information included in the target code stream.
  • the image when the image is hierarchically encoded, for example, in spatial scalable coding, the image may be subjected to resolution processing to obtain a low-resolution image, and the original image is referred to as a high-resolution image as a contrast, respectively.
  • the low resolution image and the high resolution image are encoded.
  • a high-quality image to be encoded is referred to herein as an enhancement layer image
  • a corresponding low-quality image to be encoded (for example, the low-resolution image) is referred to as a base layer image.
  • the target image is an image processed by a layered coding technique
  • the basic layer refers to a quality in layered coding (including frame rate, spatial resolution, temporal resolution, and signal to noise).
  • the enhancement layer refers to a layer with higher quality (including frame rate, spatial resolution, temporal resolution, signal-to-noise ratio intensity, or quality level) in the layered coding. .
  • the corresponding base layer may be any layer lower in quality than the enhancement layer, for example, if currently There are five layers, and the coding quality is sequentially improved (that is, the first layer has the lowest quality and the fifth layer has the highest quality).
  • the enhancement layer is the fourth layer
  • the base layer may be the first layer or the second layer. It is the third layer or the fourth layer.
  • the corresponding enhancement layer can be any layer of lower quality than the base layer.
  • the enhancement layer image is the image in the currently processed enhancement layer
  • the base layer image is the image in the base layer at the same time as the enhancement layer image.
  • the quality of the base layer image is lower than the quality of the enhancement layer image.
  • the target image block is the image block being processed in the enhancement layer image.
  • the base layer image block is an image block in the base layer image that has a corresponding relationship with the target image block in spatial position.
  • the correspondence between the image block in the base layer and the image block in the enhancement layer can be calculated according to the resolution proportional relationship between the base layer image and the enhancement layer image. For example, in a system including an X direction and a y direction, if the resolution of the enhancement layer image in the X direction and the y direction is twice that of the original layer image, the pixel coordinates of the upper left corner in the enhancement layer are (2x, 2y). And an image block of size (2m ) X ( 2n ), the corresponding block in the base layer image may be an image block whose pixel coordinates are (X , y ) and whose size is mxn in the upper left corner.
  • the sub-blocks described later refer to sub-blocks of the target image block (image blocks in the enhancement layer), and the corresponding sub-blocks described later refer to the image blocks corresponding to the sub-blocks in the base layer.
  • the motion information may include one or more of a prediction direction, a reference image index, or a motion vector, where the prediction direction may be divided into one-way and two-way prediction, and the one-way prediction may be further divided into forward prediction.
  • forward prediction refers to the use of a forward reference picture list, ie, a reference picture in list 0 to generate a prediction signal
  • backward prediction refers to using a backward reference picture list, ie, a reference picture in list 1 to generate a prediction.
  • Bidirectional prediction refers to the use of reference pictures in list 0 and list 1 to generate prediction signals; for unidirectional prediction, a reference picture index is required to indicate the reference picture selected in list 0 or list l.
  • each motion vector includes a horizontal direction
  • the component x and the vertical component y can be written as (X, y).
  • a motion vector is required to indicate the displacement of the prediction signal in the selected list 0 or list 1 reference image.
  • Two motion vectors indicating the displacement of the forward prediction signal and the backward prediction signal in the selected listO reference image and the list 1 reference image, respectively.
  • the target image block may be configured by at least two sub-blocks (ie, target image sub-blocks), wherein the size of the sub-block may be determined according to a preset value.
  • the sub-block size is 4x4 as an example. For example, if the size of the target image block is 16x 16, it can be determined that the target image block includes 16 sub-blocks (size 4x4). Therefore, in the embodiment of the present invention, a corresponding sub-block of each sub-block in the target image block (belonging to the corresponding image block) may be determined, and motion information of the corresponding sub-block is determined.
  • the coordinates of the corresponding position of the pixel in the base layer image may be determined according to the coordinates of a certain pixel point in the sub-block (denoted as: u ⁇ E x , E y ) n ) u (B x , B y y ), and the image block in the base layer containing the corresponding position coordinates is used as the corresponding sub-block.
  • the motion vector ( ⁇ , ⁇ of the first base layer image sub-block is scaled, and the scaled motion vector is used as the motion vector (EMV x , EMV y ) of the sub-block (first target image sub-block).
  • the motion information of the corresponding sub-block can be directly used as the motion information of the sub-block.
  • motion information of the corresponding sub-block may be
  • the method further includes:
  • the first base layer image sub-block corresponding to the first target image sub-block includes motion information.
  • whether the corresponding sub-block includes motion information may be determined according to an encoding mode of the base layer image (corresponding image block). For example, if the base layer image uses the intra prediction coding mode, it may be determined that the corresponding sub-block does not contain motion information (i.e., the motion information of the first basic layer image sub-block is empty).
  • the correspondence may be determined as described above.
  • the sub-block acquires its motion information, and when the corresponding sub-block does not include motion information, the above process can be skipped.
  • the second reference information can be obtained by the following method 3.
  • the reconstructed pixel of the sub-block corresponding to the first target image sub-block in the base layer image may be acquired, and the reconstructed pixel is subjected to upsampling processing, and the prediction signal of the first target image sub-block is generated as a reference. information.
  • the encoding process is performed on the target image block according to the reference information, including: performing motion compensation processing on the first target image sub-block according to the reference information.
  • the first target image sub-block may be encoded (specifically, predictive coding processing) based on reference information of the first target image sub-block (specifically, the prediction signal). Specifically, the reconstructed pixel is appropriately upsampled to serve as a prediction signal for the current sub-block. After the prediction signal of the current block is obtained, the current block can be predictively coded to calculate the rate distortion cost.
  • the encoding processing the target image block according to the reference information comprises: performing deblocking filtering processing on pixels located near a boundary between the target image sub-blocks.
  • second reference information of the first target image block (the corresponding sub-block in the base layer does not include motion information) is acquired by the above method 3
  • the third target image block is acquired (the corresponding sub-block in the base layer includes The motion information of the motion information)
  • the method of acquiring the motion information of the third target image block may be the same as the prior art, and the description thereof is omitted here, so that the distortion rate cost of the target image block can be calculated, if the distortion rate is costed
  • an indication identifier (fourth indication information) may be determined to indicate the decoding end, and the reference information of the first target image block is obtained by the foregoing method 1 and method 2, and the third target image block is acquired (corresponding in the basic layer)
  • the sub-block includes motion information of the motion information).
  • the encoding process is performed on the first target image sub-block according to the reference information, including:
  • the first indication information is subjected to entropy encoding processing such that the first indication information is adjacent to the skip mode flag bit or the merged MERGE mode flag bit information in the target code stream.
  • the first indication information in the target code stream, may be configured at a position adjacent to the skip mode flag bit information.
  • the first indication information may be placed as the first information of the target image block in the target code stream before the skip mode flag bit information, and the first indication information may be placed as the target after the skip mode flag bit information.
  • the above skip mode may be the same as the prior art, and the determination method and the configuration position may be the same as the prior art.
  • the description thereof is omitted.
  • the first indication information can be placed adjacent to the MERGE mode flag bit information.
  • the first indication information may be placed before the MERGE mode flag bit information, and the first indication information may be placed after the MERGE mode flag bit information.
  • the MERGE may be the same mode as the prior art, and the determination method and the configuration position may be the same as the prior art.
  • the description thereof is omitted.
  • the first indication information may be a binary flag bit. Therefore, when the first indication information is entropy encoded, it is not necessary to perform binarization processing on the first indication information.
  • the context used for performing the entropy coding process on the binarized first indication information may be selected.
  • the first target image sub-block is encoded according to the reference information, including:
  • the first indication information is subjected to entropy coding processing.
  • the context can contain three contexts of 0, 1, and 2.
  • This embodiment determines which context to use based on whether the left and upper image blocks use the base layer mode. In this embodiment, for example, depending on whether the image blocks on the left and the top of the target image block use their respective first indication information, which context is used, for example, if the image blocks on the left and above the target image block are not Using the first indication information, then selecting a context model with an index of 0. If the image block on the left and top of the target image block has a first indication information, then select a context model with an index of 1 if the target image block is left and above. The image block uses the first indication information, then the context with the index of 2 is selected.
  • the first indication information can be binary arithmetic coded according to the context selected as described above, and the used context model is updated.
  • the process may be the same as the prior art, and the description thereof is omitted here to avoid redundancy.
  • each first target image sub-block may be encoded according to the obtained reference information, thereby completing encoding processing on the target image block, and generating the generated target image.
  • the information of the block and the first indication information after the entropy coding process described above are added to the code stream (target code stream).
  • target code stream information of the target image (including the base layer image and the enhancement layer image) after the encoding process may be included, and the process may be the same as the prior art, here, in order to avoid For the details, the description is omitted.
  • the decoding end needs to obtain the second reference information of the first target image sub-block by using the same method as the encoding end, and the decoding end needs to use
  • the second reference information of the first target image sub-block is acquired in the same manner as the encoding end.
  • the decoding end may determine, according to the fourth indication information, whether the second reference information of the first target image sub-block needs to be acquired, and the following is a description of the case where the encoding end needs to acquire the second reference information of the first target image sub-block. .
  • the target image information may be acquired from the code stream, and the target image (specifically, the target image block) is determined, and the acquired fourth indication information (the information after the entropy coding process) is subjected to entropy decoding processing.
  • the process of context selection and context update is the same as or similar to the processing of the above encoding end, and the description thereof is omitted here.
  • a binary symbol string (bin string) representing the fourth indication information may be parsed from the code stream according to the selected context.
  • the binary arithmetic decoding process corresponds to the binary arithmetic coding process at the encoding end.
  • the sub-block corresponding to the first target image sub-block in the base layer image may be acquired.
  • the first target image sub-block may be subjected to encoding processing (specifically, predictive encoding processing) based on reference information of the first target image sub-block (specifically, the prediction signal).
  • the reconstructed pixel is upsampled as a prediction signal of the current sub-block.
  • the current block may be subjected to motion compensation processing, and the additionally decoded residual signal may also be superimposed to obtain the reconstructed signal.
  • a method for image processing according to an embodiment of the present invention for a first target image sub-block in which a motion information cannot be obtained from a corresponding sub-block included in a base layer image in a target image block of an enhancement layer image, Determining for the second target image sub-block according to the position of the first target image sub-block, and determining, according to the reconstructed pixel of the first base layer image sub-block corresponding to the first target image sub-block in spatial position
  • the reference information of the first target image sub-block, and performing encoding processing according to the reference information can improve coding performance of the first target image sub-block.
  • Figure 9 shows a schematic flow diagram of a method 800 for image processing in accordance with an embodiment of the present invention as described from the perspective of a decoding end.
  • the method 800 includes:
  • the decoding end may acquire the target image information from the code stream, determine the target image (specifically, the target image block), and acquire the first indication information (the information after entropy coding processing).
  • the obtaining the first indication information from the target code stream includes:
  • the first indication information is obtained from the target code stream, where the first indication information is adjacent to the skip mode flag bit or the merged MERGE mode flag bit information in the target code stream.
  • the first indication information may be configured at a position adjacent to the skip mode flag bit information.
  • the first indication information may be placed as the first information of the target image block in the target code stream before the skip mode flag bit information, and the first indication information may be placed as the target after the skip mode flag bit information.
  • the first indication information can be placed adjacent to the MERGE mode flag bit information.
  • the first indication information may be placed before the MERGE mode flag bit information, and the first indication information may be placed after the MERGE mode flag bit information.
  • the MERGE can be with The same mode of the prior art, the determination method and the configuration position thereof may be the same as the prior art, and the description thereof is omitted here to avoid redundancy.
  • the decoding end may perform entropy decoding processing on the obtained first indication information.
  • the first indication information may be a binary flag bit. Therefore, when entropy decoding the first indication information, it is not necessary to binarize the first indication information.
  • the context used for performing the entropy coding process on the first indication information of the binarization may be selected.
  • the obtaining the first indication information from the target code stream includes:
  • entropy decoding processing is performed to determine the first indication information.
  • the context can contain three contexts of 0, 1, and 2.
  • This embodiment determines which context to use based on whether the left and upper image blocks use the base layer mode. In this embodiment, for example, depending on whether the image blocks on the left and the top of the target image block use their respective first indication information, which context is used, for example, if the image blocks on the left and above the target image block are not Using the first indication information, then selecting a context model with an index of 0. If the image block on the left and top of the target image block has a first indication information, then select a context model with an index of 1 if the target image block is left and above. The image block uses the first indication information, then the context with the index of 2 is selected.
  • the first indication information can be binary arithmetically decoded according to the context selected as above, and the used context model is updated.
  • the process may be the same as the prior art, and the description thereof is omitted here to avoid redundancy.
  • the decoding end needs to obtain the first reference information of the first target image sub-block by using the same method as the encoding end, and the decoding end needs to use The first reference information of the first target image sub-block is acquired in the same manner as the encoding end.
  • the decoding end may determine, according to the first indication information, whether the first reference information of the first target image sub-block needs to be acquired, and the following is a description of the case where the encoding end needs to acquire the first reference information of the first target image sub-block. .
  • the image is hierarchically encoded, for example, in spatial scalable coding
  • the image may be subjected to resolution processing to obtain a low-resolution image, and the original image is referred to as a high-resolution image as a comparison, and the encoder separately encodes the low-resolution image and the high-resolution image.
  • a high quality image to be encoded is referred to herein as an enhancement layer image
  • a corresponding low quality image to be encoded (eg, the low resolution image) is referred to as a base layer image.
  • the target image is an image processed by a layered coding technique
  • the basic layer refers to a quality in layered coding (including frame rate, spatial resolution, temporal resolution, signal to noise ratio intensity or quality level).
  • the enhancement layer refers to the layer with higher quality (including frame rate, spatial resolution, temporal resolution, signal-to-noise ratio intensity or quality level) in the layered coding.
  • the corresponding base layer may be any layer lower in quality than the enhancement layer, for example, if currently There are five layers, and the coding quality is sequentially improved (that is, the first layer has the lowest quality and the fifth layer has the highest quality).
  • the enhancement layer is the fourth layer
  • the base layer may be the first layer or the second layer. It is the third layer or the fourth layer.
  • the corresponding enhancement layer can be any layer of lower quality than the base layer.
  • the enhancement layer image is the image in the currently processed enhancement layer
  • the base layer image is the image in the base layer at the same time as the enhancement layer image.
  • the quality of the base layer image is lower than the quality of the enhancement layer image.
  • the target image block is the image block being processed in the enhancement layer image.
  • the base layer image block is an image block in the base layer image that has a corresponding relationship with the target image block in spatial position.
  • the correspondence between the image block in the base layer and the image block in the enhancement layer can be calculated according to the resolution proportional relationship between the base layer image and the enhancement layer image. For example, in a system including an X direction and a y direction, if the resolution of the enhancement layer image in the X direction and the y direction is twice that of the original layer image, the pixel coordinates of the upper left corner in the enhancement layer are (2x, 2y). And an image block of size (2m ) X ( 2n ), the corresponding block in the base layer image may be an image block whose pixel coordinates are (X , y ) and whose size is mxn in the upper left corner.
  • the sub-blocks described later refer to sub-blocks of the target image block (image blocks in the enhancement layer), and the corresponding sub-blocks described later refer to the image blocks corresponding to the sub-blocks in the base layer.
  • the motion information may include one or more of a prediction direction, a reference image index, or a motion vector, where the prediction direction may be divided into one-way and two-way prediction, and the one-way prediction Can be divided into forward prediction and backward prediction, forward prediction refers to the use of forward reference image list, ie list
  • the reference image in ( list ) 0 produces a prediction signal
  • the backward prediction refers to the use of a backward reference image list, that is, a reference image in list 1 to generate a prediction signal
  • the bidirectional prediction refers to generating prediction using the reference images in list 0 and list 1 at the same time.
  • Signal; for unidirectional prediction a reference image index is required to indicate the reference image selected in list 0 or list l.
  • two reference image indices are required, indicating the reference selected in list 0 and list 1, respectively.
  • Image each motion vector includes a horizontal direction component X and a vertical direction component y, which can be written as (X, y).
  • a motion vector is required to indicate the prediction signal in the selected list 0 or list 1 reference image.
  • two motion vectors are required to indicate the displacement of the forward prediction signal and the backward prediction signal in the selected list O reference image and the list 1 reference image, respectively.
  • the target image block may be configured by at least two sub-blocks (ie, target image sub-blocks), wherein the size of the sub-block may be determined according to a preset value.
  • the sub-block size is 4 x 4 as an example. For example, if the size of the target image block is 16 x 16 , it can be determined that the target image block includes 16 sub-blocks (the size is 4 x 4 ). Therefore, in the embodiment of the present invention, a corresponding sub-block (a first base layer image sub-block) in each of the target image blocks (the first target image sub-block) in the base layer may be determined, and the Corresponding to the motion information of the sub-block.
  • the pixel may be determined in the base layer image according to the coordinates of a certain pixel point in the sub-block (the first target image sub-block) (referred to as: "(E x , E y ) " )
  • the coordinates of the corresponding position (denoted as "" (and the image block in the base layer containing the corresponding position coordinates is taken as the corresponding sub-block (the first base layer image sub-block).
  • the motion vectors (EMV x , EMV y ) of the first target image sub-block are calculated according to the following Equations 1 to 10.
  • the motion information of the corresponding sub-block can be directly used as the motion information of the sub-block.
  • motion information of the corresponding sub-block may be
  • the method further includes:
  • whether the corresponding sub-block includes motion information may be determined according to an encoding mode of the base layer image (corresponding image block). For example, if the base layer image uses the intra prediction coding mode, it may be determined that the corresponding sub-block does not contain motion information (i.e., the motion information of the first basic layer image sub-block is empty).
  • the corresponding sub-block when the corresponding sub-block includes the motion information, the corresponding sub-block may be determined according to the process as described above and the motion information thereof may be obtained.
  • the method may be obtained by the following method 4 Second reference information of a target image sub-block.
  • the reconstructed pixel of the sub-block corresponding to the first target image sub-block in the base layer image may be acquired, and the reconstructed pixel is subjected to upsampling processing, and the prediction signal of the first target image sub-block is generated as a reference. information.
  • the decoding process of the target image block according to the reference information includes: performing motion compensation processing on the first target image sub-block according to the reference information.
  • the first target sub-block may be decoded according to the obtained reference information. If the reference information is acquired by the method 3, the reference information of the first target image sub-block may be used (specifically, Motion information), decoding processing (specifically, motion compensation processing) on the first target image sub-block. Specifically, motion compensation processing may be performed on the first target image sub-block according to the motion information filled for the first target image sub-block as described above.
  • Motion information specifically, Motion information
  • decoding processing specifically, motion compensation processing
  • the reconstructed pixel of the sub-block corresponding to the first target image sub-block in the base layer image may be acquired, and the reconstructed pixel is subjected to upsampling processing to generate the first
  • the prediction signal of the target image sub-block is used as reference information.
  • the first target image sub-block may be subjected to encoding processing (specifically, predictive encoding processing) based on reference information of the first target image sub-block (specifically, the prediction signal).
  • the reconstructed pixel is upsampled as a prediction signal of the current sub-block.
  • the current block may be subjected to motion compensation processing, and the additionally decoded residual signal may be superimposed to obtain the reconstructed signal.
  • the motion information can be obtained by the same method as the prior art, and The decoding process is performed, and the process can be the same as in the prior art.
  • the decoding process on the target image block according to the reference information includes: performing deblocking filtering processing on pixels located near a boundary between the target image sub-blocks.
  • a method for image processing according to an embodiment of the present invention for a first target image sub-block in which a motion information cannot be obtained from a corresponding sub-block included in a base layer image in a target image block of an enhancement layer image, according to the first target image Positioning the sub-block determines a second target image sub-block, and determining, according to the reconstructed pixel of the first base layer image sub-block corresponding to the first target image sub-block in spatial position, determining the sub-block for the first target image Referring to the information and performing encoding processing according to the reference information, the coding performance of the first target image sub-block can be improved.
  • FIGS. 8 to 9 a method for image processing according to an embodiment of the present invention is described in detail with reference to FIGS. 8 to 9.
  • a device for image processing according to an embodiment of the present invention will be described in detail with reference to FIGS. 10 to 11. .
  • FIG. 10 shows a schematic block diagram of an apparatus 900 for image processing in accordance with an embodiment of the present invention.
  • the apparatus 900 includes:
  • the obtaining unit 910 is configured to determine, according to the reconstructed pixels of the first base layer image sub-block, when determining that motion information of the first base layer image sub-block corresponding to the first target image sub-block of the target image block is empty Second reference information for encoding the first target image sub-block, wherein the first base layer image sub-block is an image block located in a base layer image, the target image block being located in an enhancement layer image, a base layer image corresponding to the enhancement layer image, and a spatial position of the first base image block sub-block in the base layer image corresponds to a spatial position of the first target image sub-block in the enhancement layer image;
  • the encoding unit 920 is configured to encode the target image block to generate a target code stream and fourth indication information included in the target code stream.
  • the coding unit 920 is specifically configured to perform deblocking filtering processing on pixels located near a boundary between the target image sub-blocks.
  • the encoding unit 920 is specifically configured to entropy encode the fourth indication information, so that the fourth indication information is adjacent to the skip mode flag bit or the merged MERGE mode flag bit information in the target code stream.
  • the coding unit 920 is specifically configured to determine, according to whether the reference image block located in the preset position in the enhancement layer image is encoded by using reference information, to determine a context;
  • the apparatus 900 for image processing according to an embodiment of the present invention may correspond to an embodiment of the present invention.
  • the modules in the image processing device 900, and the other operations and/or functions described above, respectively, in order to implement the corresponding processes of the method 700 in FIG. 8, for the sake of cleaning no further description is provided herein. .
  • An apparatus for image processing according to an embodiment of the present invention for a first target image sub-block in which a motion information cannot be obtained from a corresponding sub-block included in a base layer image in a target image block of an enhancement layer image, by according to the first target image Positioning the sub-block determines a second target image sub-block, and determining, according to the reconstructed pixel of the first base layer image sub-block corresponding to the first target image sub-block in spatial position, determining the sub-block for the first target image Referring to the information and performing encoding processing according to the reference information, the coding performance of the first target image sub-block can be improved.
  • FIG 11 shows a schematic block diagram of an apparatus 1000 for image processing in accordance with an embodiment of the present invention.
  • the device 1000 includes:
  • the decoding unit 1010 is configured to obtain fourth indication information from the target code stream.
  • the obtaining unit 1020 is configured to, when determining that the motion information of the first base layer image sub-block corresponding to the first target image sub-block of the target image block is empty, based on the fourth indication information acquired by the decoding unit, according to the first a reconstructed pixel of a base layer image sub-block, determining second reference information for encoding the first target image sub-block, wherein the first base layer image sub-block is an image block located in the base layer image, The target image block is located in the enhancement layer image, the base layer image corresponding to the enhancement layer image, and the spatial position of the first basic image block sub-block in the base layer image and the first target image sub-block are in the Corresponding to the spatial position in the enhancement layer image;
  • the encoding unit 1010 is further configured to decode the target code stream to obtain the target image block.
  • the decoding unit 1010 is specifically configured to perform deblocking filtering processing on pixels located near a boundary between the target image sub-blocks.
  • the decoding unit 1010 is specifically configured to obtain fourth indication information from the target code stream, where the fourth indication information is in the target code stream, and the skip mode flag bit or the merged MERGE mode flag bit information is adjacent.
  • the decoding unit 1010 is specifically configured to determine, according to whether the reference image block located in the preset position in the enhancement layer image is decoded by using reference information, to determine a context;
  • the apparatus 1000 for image processing may correspond to a decoding end in the method of the embodiment of the present invention, and the units in the apparatus 1000 for image processing are modules and other operations and/or functions described above.
  • Clean, no longer repeat here In order to implement the corresponding process of the method 800 in FIG. 9, respectively, Clean, no longer repeat here.
  • An apparatus for image processing according to an embodiment of the present invention for a first target image sub-block in which a motion information cannot be obtained from a corresponding sub-block included in a base layer image in a target image block of an enhancement layer image, by according to the first target image Positioning the sub-block determines a second target image sub-block, and determining, according to the reconstructed pixel of the first base layer image sub-block corresponding to the first target image sub-block in spatial position, determining the sub-block for the first target image Referring to the information and performing encoding processing according to the reference information, the coding performance of the first target image sub-block can be improved.
  • Figure 12 shows a schematic block diagram of an encoder 1100 for image processing in accordance with an embodiment of the present invention.
  • the encoder 1100 can include:
  • processor 1120 connected to the bus
  • the processor 1120 calls the program stored in the memory 1130 through the bus 1110, for determining that the motion information of the first base layer image sub-block corresponding to the first target image sub-block of the target image block is Empty space, determining, according to the reconstructed pixels of the first base layer image sub-block, second reference information for encoding the first target image sub-block, wherein the first base layer image sub-block is located in the base layer image An image block in which the target image block is located in the enhancement layer image, the base layer image corresponding to the enhancement layer image, and a spatial position of the first basic image block sub-block in the base layer image and the first The spatial position of the target image sub-block corresponds in the enhancement layer image;
  • the target image block is encoded to generate a target code stream and fourth indication information included in the target code stream.
  • the processor 1120 is specifically configured to perform deblocking filtering processing on pixels located near a boundary between the target image sub-blocks.
  • the processor 1120 is specifically configured to entropy encode the fourth indication information, so that the fourth indication information is adjacent to the skip mode flag bit or the merged MERGE mode flag bit information in the target code stream.
  • the processor 1120 is specifically configured to be located at a preset location according to the enhancement layer image. Whether the reference image block is encoded using reference information to determine a context;
  • the encoder 1100 for image processing may correspond to an encoding end in the method of the embodiment of the present invention, and each unit in the encoder 1100 for image processing, that is, a module and the above-described other operations and/or For the purpose of implementing the corresponding process of the method 700 in FIG. 8 , the functions are not described herein.
  • An encoder for image processing for a first target image sub-block in which a motion information cannot be obtained from a corresponding sub-block included in a base layer image in a target image block of an enhancement layer image, according to the first target Positioning the image sub-block determines a second target image sub-block, and determining, according to the reconstructed pixel of the first base layer image sub-block corresponding to the first target image sub-block in spatial position, determining the first target image sub-block
  • the reference information, and encoding processing according to the reference information can improve the coding performance of the first target image sub-block.
  • Figure 13 shows a schematic block diagram of a decoder 1200 for image processing in accordance with an embodiment of the present invention.
  • the decoder 1200 can include:
  • processor 1220 connected to the bus
  • the processor 1220 calls the program stored in the memory 1230 through the bus 1210 to obtain fourth indication information from the target code stream.
  • the first base layer image sub-block is an image block located in a base layer image, the target image block being located in the enhancement layer image Corresponding to the enhancement layer image, and the spatial position of the first basic image block sub-block in the base layer image and the spatial position of the first target image sub-block in the enhancement layer image Corresponding
  • the processor 1220 is specifically configured to perform deblocking filtering processing on pixels located near a boundary between the target image sub-blocks.
  • the processor 1220 is specifically configured to obtain fourth indication information from the target code stream, where the fourth indication information is in the target code stream and the skip mode flag bit or the merged MERGE mode The type flag information is adjacent.
  • the processor 1220 is specifically configured to determine, according to whether the reference image block located in the preset position in the enhancement layer image is decoded by using reference information, to determine a context;
  • the decoder 1200 for image processing may correspond to a decoding end in the method of the embodiment of the present invention, and each unit in the decoder 1200 for image processing is a module and the other operations described above and/or For the purpose of implementing the corresponding process of the method 800 in FIG. 9, the functions are not described here.
  • a decoder for image processing for a first target image sub-block in which a motion information cannot be obtained from a corresponding sub-block included in a base layer image in a target image block of an enhancement layer image, according to the first target Positioning the image sub-block determines a second target image sub-block, and determining, according to the reconstructed pixel of the first base layer image sub-block corresponding to the first target image sub-block in spatial position, determining the first target image sub-block
  • the reference information, and encoding processing according to the reference information can improve the coding performance of the first target image sub-block.
  • the decoding end and the decoding end are required to obtain the reference information, that is, if the encoding end uses the method 1 to obtain the reference information (the first reference information) ), the decoder will use method 2 (corresponding to method 1) to obtain reference information (first reference information).
  • the decoding end processing method may be determined according to the described encoding end processing method, or the encoding end processing method may be determined according to the described decoding end processing method.
  • the size of the sequence numbers of the above processes does not mean the order of execution, and the order of execution of each process should be determined by its function and internal logic, and should not be taken to the embodiments of the present invention.
  • the implementation process constitutes any limitation.
  • the disclosed systems, devices, and methods may be implemented in other ways.
  • the device embodiments described above are merely illustrative.
  • the division of the unit is only a logical function division.
  • there may be another division manner for example, multiple units or components may be combined or Can be integrated into another system, or some features can be ignored, or not executed.
  • the coupling or direct coupling or communication connection shown or discussed may be an indirect coupling or communication connection through some interface, device or unit, and may be electrical, mechanical or otherwise.
  • the units described as separate components may or may not be physically separate, and the components displayed as units may or may not be physical units, that is, may be located in one place, or may be distributed to multiple network units. Some or all of the units may be selected according to actual needs to achieve the objectives of the solution of the embodiment.
  • each functional unit in each embodiment of the present invention may be integrated into one processing unit, or each unit may exist physically separately, or two or more units may be integrated into one unit.
  • the functions, if implemented in the form of software functional units and sold or used as separate products, may be stored in a computer readable storage medium.
  • the technical solution of the present invention which is essential to the prior art or part of the technical solution, may be embodied in the form of a software product stored in a storage medium, including
  • the instructions are used to cause a computer device (which may be a personal computer, server, or network device, etc.) to perform all or part of the steps of the methods described in various embodiments of the present invention.
  • the foregoing storage medium includes: a U disk, a mobile hard disk, a read-only memory (ROM), a random access memory (RAM), a magnetic disk or an optical disk, and the like, which can store program codes. .

Landscapes

  • Engineering & Computer Science (AREA)
  • Multimedia (AREA)
  • Signal Processing (AREA)
  • Compression Or Coding Systems Of Tv Signals (AREA)

Abstract

L'invention concerne un procédé de traitement d'image. Selon ledit procédé, s'il est déterminé qu'il n'y a pas d'information de mouvement concernant un premier sous-bloc d'image de couche de base correspondant à un premier sous-bloc d'image cible d'un bloc d'image, déterminer un second sous-bloc d'image cible selon la taille du bloc d'image cible, la taille de chaque sous-bloc d'image cible compris dans le bloc d'image cible et une seconde information d'indication pour indiquer la position du premier sous-bloc d'image cible dans le bloc d'image cible; déterminer une première information de référence pour coder le premier sous-bloc d'image cible selon l'information de mouvement concernant le second sous-bloc d'image cible, le premier sous-bloc d'image de couche de base étant un bloc d'image situé dans l'image de couche de base, le bloc d'image cible étant situé dans une image de couche d'enrichissement et l'image de la couche de base correspondant à l'image de la couche d'enrichissement; et coder le bloc de l'image cible de manière à générer un flux de codes cibles et une première information d'indication contenue dans le flux de codes cibles.
PCT/CN2013/084504 2012-09-29 2013-09-27 Procédé et dispositif de traitement d'image, codeur et décodeur WO2014048372A1 (fr)

Applications Claiming Priority (2)

Application Number Priority Date Filing Date Title
CN201210375019.5 2012-09-29
CN201210375019.5A CN103716629B (zh) 2012-09-29 2012-09-29 用于图像处理的方法、装置、编码器和解码器

Publications (1)

Publication Number Publication Date
WO2014048372A1 true WO2014048372A1 (fr) 2014-04-03

Family

ID=50387015

Family Applications (1)

Application Number Title Priority Date Filing Date
PCT/CN2013/084504 WO2014048372A1 (fr) 2012-09-29 2013-09-27 Procédé et dispositif de traitement d'image, codeur et décodeur

Country Status (2)

Country Link
CN (1) CN103716629B (fr)
WO (1) WO2014048372A1 (fr)

Cited By (2)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
RU2793825C1 (ru) * 2019-08-26 2023-04-06 Хуавэй Текнолоджиз Ко., Лтд. Способ и устройство для хранения информации о движении
US11706417B2 (en) 2018-09-10 2023-07-18 Huawei Technologies Co., Ltd. Video decoding method and video decoder

Families Citing this family (5)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
KR20170026334A (ko) * 2014-07-06 2017-03-08 엘지전자 주식회사 비디오 신호의 처리 방법 및 이를 위한 장치
CN109996075B (zh) * 2017-12-29 2022-07-12 华为技术有限公司 一种图像解码方法及解码器
CN112040243B (zh) 2018-06-04 2021-06-29 华为技术有限公司 获取运动矢量的方法和装置
CN112598572B (zh) * 2019-10-01 2022-04-15 浙江大学 一种筛选子块图像与处理单元的方法及装置
CN114339262B (zh) * 2020-09-30 2023-02-14 华为技术有限公司 熵编/解码方法及装置

Citations (3)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN101156454A (zh) * 2005-02-07 2008-04-02 三星电子株式会社 用来压缩多层运动向量的方法与装置
CN101755458A (zh) * 2006-07-11 2010-06-23 诺基亚公司 可缩放视频编码
US20120201301A1 (en) * 2006-01-11 2012-08-09 Qualcomm Incorporated Video coding with fine granularity spatial scalability

Family Cites Families (2)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
KR100664929B1 (ko) * 2004-10-21 2007-01-04 삼성전자주식회사 다 계층 기반의 비디오 코더에서 모션 벡터를 효율적으로압축하는 방법 및 장치
CN101198064A (zh) * 2007-12-10 2008-06-11 武汉大学 一种分辨率分层技术中的运动矢量预测方法

Patent Citations (3)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN101156454A (zh) * 2005-02-07 2008-04-02 三星电子株式会社 用来压缩多层运动向量的方法与装置
US20120201301A1 (en) * 2006-01-11 2012-08-09 Qualcomm Incorporated Video coding with fine granularity spatial scalability
CN101755458A (zh) * 2006-07-11 2010-06-23 诺基亚公司 可缩放视频编码

Cited By (2)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US11706417B2 (en) 2018-09-10 2023-07-18 Huawei Technologies Co., Ltd. Video decoding method and video decoder
RU2793825C1 (ru) * 2019-08-26 2023-04-06 Хуавэй Текнолоджиз Ко., Лтд. Способ и устройство для хранения информации о движении

Also Published As

Publication number Publication date
CN103716629A (zh) 2014-04-09
CN103716629B (zh) 2017-02-22

Similar Documents

Publication Publication Date Title
CN112889269B (zh) 视频解码方法及装置
CN113545042B (zh) 视频编解码的方法和装置
CN113261298B (zh) 视频编解码的方法和装置
JP6701409B2 (ja) 符号化装置、復号装置、符号化方法、復号方法、及びプログラム
TWI532369B (zh) 用於基於高效率視訊寫碼之可擴縮寫碼視訊資訊之裝置及方法
KR101658324B1 (ko) 비디오 코딩을 위한 방법 및 장치
CN112055968B (zh) 视频解码方法、装置、存储介质以及电子设备
JP6352390B2 (ja) 高レベルシンタックス専用shvcにおける一般化残差予測ならびにそのシグナリングおよび管理
CN112005551B (zh) 一种视频图像预测方法及装置
WO2014048372A1 (fr) Procédé et dispositif de traitement d'image, codeur et décodeur
JP6573879B2 (ja) ビデオ情報のスケーラブルコーディングのためのデバイスおよび方法
JP6495268B2 (ja) パラメータセット内のビューidビット深度のシグナリング
TWI535273B (zh) 經組態以寫碼視訊資訊之設備及視訊寫碼裝置、編碼及解碼視訊資訊之方法與非暫態電腦可讀媒體
TWI504239B (zh) 針對基於高效率視訊寫碼之可縮放寫碼之運動場升頻取樣
CN113491113B (zh) 视频编解码的方法和装置
JP6367309B2 (ja) 映像情報のスケーラブルコーディングのためのデバイス及び方法
WO2020052534A1 (fr) Procédé de décodage vidéo et décodeur vidéo
WO2020042630A1 (fr) Procédé et appareil de prédiction d'images vidéo
CN111953996B (zh) 视频解码的方法和装置
CN113498607B (zh) 视频编码方法、解码方法、装置和可读介质
WO2014048378A1 (fr) Procédé et dispositif de traitement d'image, codeur et décodeur
CN111726622B (zh) 视频编解码的方法、装置及介质
TWI543590B (zh) 用於可縮放視訊寫碼之空間運動向量縮放
CN118042127A (zh) 视频编码、解码方法及装置
CN113545039B (zh) 视频编解码的方法和装置

Legal Events

Date Code Title Description
121 Ep: the epo has been informed by wipo that ep was designated in this application

Ref document number: 13842955

Country of ref document: EP

Kind code of ref document: A1

NENP Non-entry into the national phase

Ref country code: DE

122 Ep: pct application non-entry in european phase

Ref document number: 13842955

Country of ref document: EP

Kind code of ref document: A1