CN103716629A - Image processing method, device, coder and decoder - Google Patents

Image processing method, device, coder and decoder Download PDF

Info

Publication number
CN103716629A
CN103716629A CN201210375019.5A CN201210375019A CN103716629A CN 103716629 A CN103716629 A CN 103716629A CN 201210375019 A CN201210375019 A CN 201210375019A CN 103716629 A CN103716629 A CN 103716629A
Authority
CN
China
Prior art keywords
image
block
sub
target image
idx
Prior art date
Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
Granted
Application number
CN201210375019.5A
Other languages
Chinese (zh)
Other versions
CN103716629B (en
Inventor
杨海涛
李礼
李斌
李厚强
Current Assignee (The listed assignees may be inaccurate. Google has not performed a legal analysis and makes no representation or warranty as to the accuracy of the list.)
University of Science and Technology of China USTC
Huawei Technologies Co Ltd
Original Assignee
University of Science and Technology of China USTC
Huawei Technologies Co Ltd
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Application filed by University of Science and Technology of China USTC, Huawei Technologies Co Ltd filed Critical University of Science and Technology of China USTC
Priority to CN201210375019.5A priority Critical patent/CN103716629B/en
Priority to PCT/CN2013/084504 priority patent/WO2014048372A1/en
Publication of CN103716629A publication Critical patent/CN103716629A/en
Application granted granted Critical
Publication of CN103716629B publication Critical patent/CN103716629B/en
Active legal-status Critical Current
Anticipated expiration legal-status Critical

Links

Images

Classifications

    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04NPICTORIAL COMMUNICATION, e.g. TELEVISION
    • H04N19/00Methods or arrangements for coding, decoding, compressing or decompressing digital video signals
    • H04N19/50Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using predictive coding
    • H04N19/59Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using predictive coding involving spatial sub-sampling or interpolation, e.g. alteration of picture size or resolution
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04NPICTORIAL COMMUNICATION, e.g. TELEVISION
    • H04N19/00Methods or arrangements for coding, decoding, compressing or decompressing digital video signals
    • H04N19/30Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using hierarchical techniques, e.g. scalability
    • H04N19/33Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using hierarchical techniques, e.g. scalability in the spatial domain
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04NPICTORIAL COMMUNICATION, e.g. TELEVISION
    • H04N19/00Methods or arrangements for coding, decoding, compressing or decompressing digital video signals
    • H04N19/46Embedding additional information in the video signal during the compression process
    • H04N19/463Embedding additional information in the video signal during the compression process by compressing encoding parameters before transmission
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04NPICTORIAL COMMUNICATION, e.g. TELEVISION
    • H04N19/00Methods or arrangements for coding, decoding, compressing or decompressing digital video signals
    • H04N19/50Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using predictive coding
    • H04N19/503Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using predictive coding involving temporal prediction
    • H04N19/51Motion estimation or motion compensation
    • H04N19/513Processing of motion vectors
    • H04N19/517Processing of motion vectors by encoding
    • H04N19/52Processing of motion vectors by encoding by predictive encoding
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04NPICTORIAL COMMUNICATION, e.g. TELEVISION
    • H04N19/00Methods or arrangements for coding, decoding, compressing or decompressing digital video signals
    • H04N19/10Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using adaptive coding
    • H04N19/169Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using adaptive coding characterised by the coding unit, i.e. the structural portion or semantic portion of the video signal being the object or the subject of the adaptive coding
    • H04N19/187Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using adaptive coding characterised by the coding unit, i.e. the structural portion or semantic portion of the video signal being the object or the subject of the adaptive coding the unit being a scalable video layer
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04NPICTORIAL COMMUNICATION, e.g. TELEVISION
    • H04N19/00Methods or arrangements for coding, decoding, compressing or decompressing digital video signals
    • H04N19/50Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using predictive coding
    • H04N19/503Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using predictive coding involving temporal prediction
    • H04N19/51Motion estimation or motion compensation

Landscapes

  • Engineering & Computer Science (AREA)
  • Multimedia (AREA)
  • Signal Processing (AREA)
  • Compression Or Coding Systems Of Tv Signals (AREA)

Abstract

Disclosed is an image processing method which includes: determining a second target image sub-block according to the size of a target image block, the size of each target image sub-block included in the target image block and second indication information used for indicating the position of a first target image sub-block in the target image block when it is determined that monition information of a first basic-layer image sub-block corresponding to the first target image sub-block of the target image block is vacant; determining first reference information used for coding the first target image sub-block according to the motion information of the second target image sub-block, wherein the first-layer image sub-block is an image block in a basic-layer image and the target image block is in a reinforced-layer image and the basic-layer image is corresponding to the reinforced-layer image; and coding the target image block so as to generate a target code stream and first indication information included in the target code stream.

Description

Method, device code device and the decoder for image, processed
Technical field
The present invention relates to field of video processing, and more specifically, relate to a kind of method, device code device and decoder of processing for image.
Background technology
Along with the fast development of the Internet and becoming increasingly abundant of people's material spirit culture, application demand for video in the Internet is especially more and more for the application demand of HD video, and the data volume of HD video is very large, wanting HD video can transmit in band-limited the Internet, and the problem that must first solve is exactly HD video compressed encoding problem.
In network environment (such as the Internet), because the network bandwidth is limited, terminal equipment and user's demand is all different, so for certain specific application, the code stream of first compression is not satisfactory and effective, for some specific user or equipment, or even nonsensical.The effective method addressing this problem utilizes scalable video (SVC, scalable video coding) technology exactly.Scalable coding is called again hierarchical coding.In this SVC technology, according to the mass parameter that comprises spatial resolution, temporal resolution or signal to noise ratio intensity etc., image is carried out to hierarchical coding.For example, when spatial scalable is encoded, image can be carried out to resolution decreasing processing and obtain low-resolution image, as a comparison original image is called to high-definition picture, encoder respectively to this low quality (for example, low resolution) image and this high-quality (for example, high-resolution) image is encoded, and obtains high quality graphic coded message and low-quality image coded message.In this SVC technology, the mass parameter according to comprising spatial resolution, temporal resolution or signal to noise ratio intensity etc., is divided into a plurality of image layer by an image.The target of SVC is exactly to allow high quality graphic layer utilize fully the information of low-quality image layer as far as possible, improves the efficiency of inter-layer prediction, can efficiency when making the high image of coding quality higher.
In order to improve inter-layer prediction efficiency, in the prior art, if there is at least one image block to use inter-frame forecast mode in low quality layer image in all image blocks corresponding with the image block of high quality layer image, directly use the movable information of image block of the low quality layer image after suitable convergent-divergent as the movable information of respective image piece in high quality layer image, but, for example, in low quality layer image, there are one or more sub-blocks (corresponding sub-block) to adopt frame mode (, the movable information of this correspondence sub-block is empty) in the situation of coding, sub-block in high quality layer image block cannot be obtained movable information in the corresponding sub-block from low quality layer image.In this situation, can obtain according to given method construct the movable information of this sub-block.Yet the movable information that these derivations obtain is inaccurate, thereby can affect the coding efficiency to this sub-block, and further affect the code efficiency of whole high quality layer image.
Therefore, be desirable to provide a kind of method, can improve the coding efficiency that obtains the sub-block of movable information for corresponding sub-block that can not be from low quality layer image in the target image piece of high quality layer image.
Summary of the invention
The embodiment of the present invention provides a kind of method and apparatus of processing for image, can improve the coding efficiency that obtains the sub-block of movable information for the corresponding sub-block that can not comprise from basic tomographic image in the target image piece of enhancement layer image.
First aspect, a kind of method of processing for image is provided, the method comprises: when determining that the movable information of the first basic tomographic image sub-block corresponding with the first object image subblock of target image piece is empty, the size of each the target image sub-block comprising according to the size of this target image piece, this target image piece and be used to indicate the second indication information of the position of first object image subblock in this target image piece, determines the second target image sub-block; According to the movable information of this second target image sub-block, be identified for the first reference information that this first object image subblock is encoded, wherein, this first basic tomographic image sub-block is the image block that is arranged in basic tomographic image, this target image piece is arranged in enhancement layer image, the basic tomographic image being somebody's turn to do is corresponding with this enhancement layer image, and this first primary image piece sub-block locus in this basic tomographic image is corresponding with this first object image subblock locus in this enhancement layer image; This target image piece is encoded, to generate target code stream and to be contained in the first indication information in this target code stream.
In a kind of possible execution mode, the size of the target image sub-block that this comprises according to the size of target image piece, this target image piece and be used to indicate the second indication information of the position of first object image subblock in this target image piece, determine the second target image sub-block, comprise: according to following arbitrary formula, determine this second target image sub-block
Idx 2=Idx 1/N×N+((Idx 1%N/(N/2))×2+(1-Idx 1%N/(N/4)%2))×N/4;
Idx 2=Idx 1/N×N+((1-Idx 1%N/(N/2))×2+(Idx 1%N/(N/4)%2))×N/4;
Idx 2=Idx 1/N×N+((1-Idx 1%N/(N/2))×2+(1-Idx 1%N/(N/4)%2))×N/4;
Wherein, Idx 2represent to be used to indicate the 3rd indication information of this position of the second target image sub-block in this target image piece, Idx 1represent this second indication information, N determines according to the size of the size of this target image piece and this target image sub-block.
In conjunction with first aspect and the possible execution mode of the first, in the possible execution mode of the second, this is according to the movable information of this second target image sub-block, be identified for the first reference information that this first object image subblock is encoded, comprise: if the movable information of this second target image sub-block for empty, is determined this first reference information, be zero movable information.
In conjunction with first aspect, execution mode and the possible execution mode of the second that the first is possible, in the third possible execution mode, this is encoded and comprises this target image piece: according to this first reference information, this first object image subblock is carried out to motion compensation process.
In conjunction with first aspect, possible execution mode and the third the possible execution mode of execution mode, the second that the first is possible, in the 4th kind of possible execution mode, this is according to this reference information, this target image piece is encoded, comprising: the pixel to the boundary vicinity between this target image sub-block is carried out block-eliminating effect filtering processing.
In conjunction with first aspect, possible execution mode, the third possible execution mode and the 4th kind of possible execution mode of execution mode, the second that the first is possible, in the 5th kind of possible execution mode, this is according to this reference information, this target image piece is encoded, comprise: this first indication information is carried out to entropy coding, so that this first indication information is adjacent with skip mode flag bit or fusion MERGE mode flags position information in this target code stream.
In conjunction with first aspect, possible execution mode, the third possible execution mode, the 4th kind of possible execution mode and the 5th kind of possible execution mode of execution mode, the second that the first is possible, in the 6th kind of possible execution mode, this is according to this reference information, this target image piece is encoded, comprise: according to the reference image block that is positioned at predeterminated position in this enhancement layer image, whether use reference information to encode, determine context; According to this context, this first indication information is carried out to entropy coding.
Second aspect, provides a kind of method of processing for image, and the method comprises: from target code stream, obtain the first indication information; When the movable information of the first corresponding basic tomographic image sub-block of the first object image subblock with target image piece is sky, based on this first indication information, the size of each the target image sub-block comprising according to the size of this target image piece, this target image piece and be used to indicate the second indication information of the position of first object image subblock in this target image piece, determines the second target image sub-block; According to the movable information of this second target image sub-block, be identified for the first reference information that this first object image subblock is decoded, wherein, this first basic tomographic image sub-block is the image block that is arranged in basic tomographic image, this target image piece is arranged in enhancement layer image, the basic tomographic image being somebody's turn to do is corresponding with this enhancement layer image, and this first primary image piece sub-block locus in this basic tomographic image is corresponding with this first object image subblock locus in this enhancement layer image; This object code stream is decoded, to obtain this target image piece.
In a kind of possible execution mode, the size of the target image sub-block that this comprises according to the size of target image piece, this target image piece and be used to indicate the second indication information of the position of first object image subblock in this target image piece, determine the second target image sub-block, comprise: according to following arbitrary formula, determine this second target image sub-block
Idx 2=Idx 1/N×N+((Idx 1%N/(N/2))×2+(1-Idx 1%N/(N/4)%2))×N/4;
Idx 2=Idx 1/N×N+((1-Idx 1%N/(N/2))×2+(Idx 1%N/(N/4)%2))×N/4;
Idx 2=Idx 1/N×N+((1-Idx 1%N/(N/2))×2+(1-Idx 1%N/(N/4)%2))×N/4;
Wherein, Idx 2represent to be used to indicate the 3rd indication information of this position of the second target image sub-block in this target image piece, Idx 1represent this second indication information, N determines according to the size of the size of this target image piece and this target image sub-block.
In conjunction with second aspect and the possible execution mode of the first, in the possible execution mode of the second, this is according to the movable information of this second target image sub-block, be identified for the first reference information that this first object image subblock is encoded, comprise: if the movable information of this second target image sub-block for empty, is determined this first reference information, be zero movable information.
In conjunction with second aspect, execution mode and the possible execution mode of the second that the first is possible, in the third possible execution mode, this is decoded and comprises this object code stream: according to this first reference information, this first object image subblock is carried out to motion compensation process.
In conjunction with second aspect, possible execution mode and the third the possible execution mode of execution mode, the second that the first is possible, in the 4th kind of possible execution mode, this is according to this reference information, this target image piece is decoded, comprising: the pixel to the boundary vicinity between this target image sub-block is carried out block-eliminating effect filtering processing.
In conjunction with second aspect, possible execution mode, the third possible execution mode and the 4th kind of possible execution mode of execution mode, the second that the first is possible, in the 5th kind of possible execution mode, should be from target code stream, obtain the first indication information, comprise: from target code stream, obtain the first indication information, wherein, this first indication information is adjacent with skip mode flag bit or fusion MERGE mode flags position information in this target code stream.
In conjunction with second aspect, possible execution mode, the third possible execution mode, the 4th kind of possible execution mode and the 5th kind of possible execution mode of execution mode, the second that the first is possible, in the 6th kind of possible execution mode, should be from target code stream, obtain the first indication information, comprise: according to the reference image block that is positioned at predeterminated position in this enhancement layer image, whether use reference information to decode, determine context; According to this context, carry out entropy decoding, to determine this first indication information.
The third aspect, a kind of device of processing for image is provided, this device comprises: acquiring unit, for when determining that the movable information of the first basic tomographic image sub-block corresponding with the first object image subblock of target image piece is empty, the size of each the target image sub-block comprising according to the size of this target image piece, this target image piece and be used to indicate the second indication information of the position of first object image subblock in this target image piece, determines the second target image sub-block; Be used for according to the movable information of this second target image sub-block, be identified for the first reference information that this first object image subblock is encoded, wherein, this first basic tomographic image sub-block is the image block that is arranged in basic tomographic image, this target image piece is arranged in enhancement layer image, the basic tomographic image being somebody's turn to do is corresponding with this enhancement layer image, and this first primary image piece sub-block locus in this basic tomographic image is corresponding with this first object image subblock locus in this enhancement layer image; Coding unit, encodes to this target image piece, to generate target code stream and to be contained in the first indication information in this target code stream.
In a kind of possible execution mode, this acquiring unit, specifically for according to following arbitrary formula, is determined this second target image sub-block,
Idx 2=Idx 1/N×N+((Idx 1%N/(N/2))×2+(1-Idx 1%N/(N/4)%2))×N/4;
Idx 2=Idx 1/N×N+((1-Idx 1%N/(N/2))×2+(Idx 1%N/(N/4)%2))×N/4;
Idx 2=Idx 1/N×N+((1-Idx 1%N/(N/2))×2+(1-Idx 1%N/(N/4)%2))×N/4;
Wherein, Idx 2represent to be used to indicate the 3rd indication information of this position of the second target image sub-block in this target image piece, Idx 1represent this second indication information, N determines according to the size of the size of this target image piece and this target image sub-block.
In conjunction with the third aspect and the possible execution mode of the first, in the possible execution mode of the second, if this acquiring unit is empty specifically for the movable information of this second target image sub-block, determine that this first reference information is zero movable information.
In conjunction with the third aspect, execution mode and the possible execution mode of the second that the first is possible, in the third possible execution mode, this coding unit, specifically for according to this first reference information, carries out motion compensation process to this first object image subblock.
In conjunction with the third aspect, possible execution mode and the third the possible execution mode of execution mode, the second that the first is possible, in the 4th kind of possible execution mode, this coding unit also carries out block-eliminating effect filtering processing for the pixel of the boundary vicinity between this target image sub-block.
In conjunction with the third aspect, possible execution mode, the third possible execution mode and the 4th kind of possible execution mode of execution mode, the second that the first is possible, in the 5th kind of possible execution mode, this coding unit is specifically for this first indication information is carried out to entropy coding, so that this first indication information is adjacent with skip mode flag bit or fusion MERGE mode flags position information in this target code stream.
In conjunction with the third aspect, possible execution mode, the third possible execution mode, the 4th kind of possible execution mode and the 5th kind of possible execution mode of execution mode, the second that the first is possible, in the 6th kind of possible execution mode, this coding unit, specifically for whether using reference information to encode according to the reference image block that is positioned at predeterminated position in this enhancement layer image, is determined context; For according to this context, this first indication information is carried out to entropy coding.
Fourth aspect, the device that provides a kind of image to process, this device comprises: decoding unit, for from target code stream, obtains the first indication information; Acquiring unit, while being empty for the movable information when the first corresponding basic tomographic image sub-block of the first object image subblock with target image piece, this first indication information obtaining based on this decoding unit, the size of each the target image sub-block comprising according to the size of this target image piece, this target image piece and be used to indicate the second indication information of the position of first object image subblock in this target image piece, determines the second target image sub-block; Be used for according to the movable information of this second target image sub-block, be identified for the first reference information that this first object image subblock is decoded, wherein, this first basic tomographic image sub-block is the image block that is arranged in basic tomographic image, this target image piece is arranged in enhancement layer image, the basic tomographic image being somebody's turn to do is corresponding with this enhancement layer image, and this first primary image piece sub-block locus in this basic tomographic image is corresponding with this first object image subblock locus in this enhancement layer image; This decoding unit is also for this object code stream is decoded, to obtain this target image piece.
In a kind of possible execution mode, this acquiring unit, specifically for according to following arbitrary formula, is determined this second target image sub-block,
Idx 2=Idx 1/N×N+((Idx 1%N/(N/2))×2+(1-Idx 1%N/(N/4)%2))×N/4;
Idx 2=Idx 1/N×N+((1-Idx 1%N/(N/2))×2+(Idx 1%N/(N/4)%2))×N/4;
Idx 2=Idx 1/N×N+((1-Idx 1%N/(N/2))×2+(1-Idx 1%N/(N/4)%2))×N/4;
Wherein, Idx 2represent to be used to indicate the 3rd indication information of this position of the second target image sub-block in this target image piece, Idx 1represent this second indication information, N determines according to the size of the size of this target image piece and this target image sub-block.
In conjunction with fourth aspect and the possible execution mode of the first, in the possible execution mode of the second, if this acquiring unit is empty specifically for the movable information of this second target image sub-block, determine that this first reference information is zero movable information.
In conjunction with fourth aspect, execution mode and the possible execution mode of the second that the first is possible, in the third possible execution mode, this decoding unit, specifically for according to this first reference information, carries out motion compensation process to this first object image subblock.
In conjunction with fourth aspect, possible execution mode and the third the possible execution mode of execution mode, the second that the first is possible, in the 4th kind of possible execution mode, this decoding unit also carries out block-eliminating effect filtering processing for the pixel of the boundary vicinity between this target image sub-block.
In conjunction with fourth aspect, possible execution mode, the third possible execution mode and the 4th kind of possible execution mode of execution mode, the second that the first is possible, in the 5th kind of possible execution mode, this decoding unit is specifically for from target code stream, obtain the first indication information, wherein, this first indication information is adjacent with skip mode flag bit or fusion MERGE mode flags position information in this target code stream.
In conjunction with fourth aspect, possible execution mode, the third possible execution mode, the 4th kind of possible execution mode and the 5th kind of possible execution mode of execution mode, the second that the first is possible, in the 6th kind of possible execution mode, this decoding unit, specifically for whether using reference information to decode according to the reference image block that is positioned at predeterminated position in this enhancement layer image, is determined context; For according to this context, carry out entropy decoding, to determine this first indication information.
The 5th aspect, provides a kind of encoder of processing for image, and this encoder comprises: bus; The processor being connected with this bus; The memory being connected with this bus; Wherein, this processor is by this bus, call the program of storing in this memory, for when determining that the movable information of the first basic tomographic image sub-block corresponding with the first object image subblock of target image piece is empty, the size of each the target image sub-block comprising according to the size of this target image piece, this target image piece and be used to indicate the second indication information of the position of first object image subblock in this target image piece, determines the second target image sub-block; Be used for according to the movable information of this second target image sub-block, be identified for the first reference information that this first object image subblock is encoded, wherein, this first basic tomographic image sub-block is the image block that is arranged in basic tomographic image, this target image piece is arranged in enhancement layer image, the basic tomographic image being somebody's turn to do is corresponding with this enhancement layer image, and this first primary image piece sub-block locus in this basic tomographic image is corresponding with this first object image subblock locus in this enhancement layer image; For this target image piece is encoded, to generate target code stream and to be contained in the first indication information in this target code stream.
In a kind of possible execution mode, this processor, specifically for according to following arbitrary formula, is determined this second target image sub-block,
Idx 2=Idx 1/N×N+((Idx 1%N/(N/2))×2+(1-Idx 1%N/(N/4)%2))×N/4;
Idx 2=Idx 1/N×N+((1-Idx 1%N/(N/2))×2+(Idx 1%N/(N/4)%2))×N/4;
Idx 2=Idx 1/N×N+((1-Idx 1%N/(N/2))×2+(1-Idx 1%N/(N/4)%2))×N/4;
Wherein, Idx 2represent to be used to indicate the 3rd indication information of this position of the second target image sub-block in this target image piece, Idx 1represent this second indication information, N determines according to the size of the size of this target image piece and this target image sub-block.
In conjunction with the 5th aspect and the possible execution mode of the first, in the possible execution mode of the second, if this processor is empty specifically for the movable information of this second target image sub-block, determine that this first reference information is zero movable information.
In conjunction with the 5th aspect, execution mode and the possible execution mode of the second that the first is possible, in the third possible execution mode, this processor, specifically for according to this first reference information, carries out motion compensation process to this first object image subblock.
In conjunction with the 5th aspect, possible execution mode and the third the possible execution mode of execution mode, the second that the first is possible, in the 4th kind of possible execution mode, this processor carries out block-eliminating effect filtering processing specifically for the pixel of the boundary vicinity between this target image sub-block.
In conjunction with the 5th aspect, possible execution mode, the third possible execution mode and the 4th kind of possible execution mode of execution mode, the second that the first is possible, in the 5th kind of possible execution mode, this processor is specifically for this first indication information is carried out to entropy coding, so that this first indication information is adjacent with skip mode flag bit or fusion MERGE mode flags position information in this target code stream.
In conjunction with the 5th aspect, possible execution mode, the third possible execution mode, the 4th kind of possible execution mode and the 5th kind of possible execution mode of execution mode, the second that the first is possible, in the 6th kind of possible execution mode, this processor, specifically for whether using reference information to encode according to the reference image block that is positioned at predeterminated position in this enhancement layer image, is determined context; For according to this context, this first indication information is carried out to entropy coding.
The 6th aspect, provides a kind of decoder of processing for image, and this decoder comprises: bus; The processor being connected with this bus; The memory being connected with this bus; Wherein, this processor, by this bus, calls the program of storing in this memory, for from target code stream, obtains the first indication information; While being empty for the movable information when the first corresponding basic tomographic image sub-block of the first object image subblock with target image piece, based on this first indication information, the size of each the target image sub-block comprising according to the size of this target image piece, this target image piece and be used to indicate the second indication information of the position of first object image subblock in this target image piece, determines the second target image sub-block; Be used for according to the movable information of this second target image sub-block, be identified for the first reference information that this first object image subblock is decoded, wherein, this first basic tomographic image sub-block is the image block that is arranged in basic tomographic image, this target image piece is arranged in enhancement layer image, the basic tomographic image being somebody's turn to do is corresponding with this enhancement layer image, and this first primary image piece sub-block locus in this basic tomographic image is corresponding with this first object image subblock locus in this enhancement layer image; For this object code stream is decoded, to obtain this target image piece.
In a kind of possible execution mode, this processor, specifically for according to following arbitrary formula, is determined this second target image sub-block,
Idx 2=Idx 1/N×N+((Idx 1%N/(N/2))×2+(1-Idx 1%N/(N/4)%2))×N/4;
Idx 2=Idx 1/N×N+((1-Idx 1%N/(N/2))×2+(Idx 1%N/(N/4)%2))×N/4;
Idx 2=Idx 1/N×N+((1-Idx 1%N/(N/2))×2+(1-Idx 1%N/(N/4)%2))×N/4;
Wherein, Idx 2represent to be used to indicate the 3rd indication information of this position of the second target image sub-block in this target image piece, Idx 1represent this second indication information, N determines according to the size of the size of this target image piece and this target image sub-block.
In conjunction with the 6th aspect and the possible execution mode of the first, in the possible execution mode of the second, if this processor is empty specifically for the movable information of this second target image sub-block, determine that this first reference information is zero movable information.
In conjunction with the 6th aspect, execution mode and the possible execution mode of the second that the first is possible, in the third possible execution mode, this processor, specifically for according to this first reference information, carries out motion compensation process to this first object image subblock.
In conjunction with the 6th aspect, possible execution mode and the third the possible execution mode of execution mode, the second that the first is possible, in the 4th kind of possible execution mode, this processor carries out block-eliminating effect filtering processing specifically for the pixel of the boundary vicinity between this target image sub-block.
In conjunction with the 6th aspect, possible execution mode, the third possible execution mode and the 4th kind of possible execution mode of execution mode, the second that the first is possible, in the 5th kind of possible execution mode, this processor is specifically for from target code stream, obtain the first indication information, wherein, this first indication information is adjacent with skip mode flag bit or fusion MERGE mode flags position information in this target code stream.
In conjunction with the 6th aspect, possible execution mode, the third possible execution mode, the 4th kind of possible execution mode and the 5th kind of possible execution mode of execution mode, the second that the first is possible, in the 6th kind of possible execution mode, this processor, specifically for whether using reference information to decode according to the reference image block that is positioned at predeterminated position in this enhancement layer image, is determined context; For according to this context, carry out entropy decoding, to determine this first indication information.
The 7th aspect, a kind of method of processing for image is provided, the method comprises: when determining that the movable information of the first basic tomographic image sub-block corresponding with the first object image subblock of target image piece is empty, according to the reconstruction pixel of this first basic tomographic image sub-block, be identified for the second reference information that this first object image subblock is encoded, wherein, this first basic tomographic image sub-block is the image block that is arranged in basic tomographic image, this target image piece is arranged in enhancement layer image, the basic tomographic image being somebody's turn to do is corresponding with this enhancement layer image, and this first primary image piece sub-block locus in this basic tomographic image is corresponding with this first object image subblock locus in this enhancement layer image, this target image piece is encoded, to generate target code stream and to be contained in the 4th indication information in this target code stream.
In a kind of possible execution mode, this is encoded and comprises this target image piece: the pixel to the boundary vicinity between this target image sub-block is carried out block-eliminating effect filtering processing.
In conjunction with the 7th aspect and the possible execution mode of the first, in the possible execution mode of the second, this is encoded and comprises this target image piece: the 4th indication information is carried out to entropy coding, so that the 4th indication information is adjacent with skip mode flag bit or fusion MERGE mode flags position information in this target code stream.
In conjunction with the 7th aspect, execution mode and the possible execution mode of the second that the first is possible, whether this is encoded and comprises this target image piece: according to the reference image block that is positioned at predeterminated position in this enhancement layer image, use reference information to encode, determine context; According to this context, the 4th indication information is carried out to entropy coding.
Eight aspect, provides a kind of method of processing for image, and the method comprises: from target code stream, obtain the 4th indication information, when determining that the movable information of the first basic tomographic image sub-block corresponding with the first object image subblock of target image piece is empty, based on the 4th indication information, according to the reconstruction pixel of this first basic tomographic image sub-block, be identified for the second reference information that this first object image subblock is encoded, wherein, this first basic tomographic image sub-block is the image block that is arranged in basic tomographic image, this target image piece is arranged in enhancement layer image, the basic tomographic image being somebody's turn to do is corresponding with this enhancement layer image, and this first primary image piece sub-block locus in this basic tomographic image is corresponding with this first object image subblock locus in this enhancement layer image, this object code stream is decoded, to obtain this target image piece.
In a kind of possible execution mode, this is decoded and comprises this object code stream: the pixel to the boundary vicinity between this target image sub-block is carried out block-eliminating effect filtering processing.
In conjunction with eight aspect and the possible execution mode of the first, in the possible execution mode of the second, should be from target code stream, obtain the 4th indication information, comprise: from target code stream, obtain the 4th indication information, wherein, the 4th indication information is adjacent with skip mode flag bit or fusion MERGE mode flags position information in this target code stream.
In conjunction with eight aspect, execution mode and the possible execution mode of the second that the first is possible, in the third possible execution mode, should be from target code stream, obtain the 4th indication information, comprise: according to the reference image block that is positioned at predeterminated position in this enhancement layer image, whether use reference information to decode, determine context; According to this context, carry out entropy decoding, to determine the 4th indication information.
The 9th aspect, a kind of device of processing for image is provided, this device comprises: acquiring unit, for when determining that the movable information of the first basic tomographic image sub-block corresponding with the first object image subblock of target image piece is empty, according to the reconstruction pixel of this first basic tomographic image sub-block, be identified for the second reference information that this first object image subblock is encoded, wherein, this first basic tomographic image sub-block is the image block that is arranged in basic tomographic image, this target image piece is arranged in enhancement layer image, the basic tomographic image being somebody's turn to do is corresponding with this enhancement layer image, and this first primary image piece sub-block locus in this basic tomographic image is corresponding with this first object image subblock locus in this enhancement layer image, coding unit, for this target image piece is encoded, to generate target code stream and to be contained in the 4th indication information in this target code stream.
In a kind of possible execution mode, this coding unit carries out block-eliminating effect filtering processing specifically for the pixel of the boundary vicinity between this target image sub-block.
In conjunction with the 9th aspect and the possible execution mode of the first, in the possible execution mode of the second, this coding unit is specifically for the 4th indication information is carried out to entropy coding, so that the 4th indication information is adjacent with skip mode flag bit or fusion MERGE mode flags position information in this target code stream.
In conjunction with the 9th aspect, execution mode and the possible execution mode of the second that the first is possible, in the third possible execution mode, this coding unit, specifically for whether using reference information to encode according to the reference image block that is positioned at predeterminated position in this enhancement layer image, is determined context; For according to this context, the 4th indication information is carried out to entropy coding.
The tenth aspect, the device that provides a kind of image to process, this device comprises: decoding unit, for from target code stream, obtains the 4th indication information, acquiring unit, for when determining that the movable information of the first basic tomographic image sub-block corresponding with the first object image subblock of target image piece is empty, the 4th indication information obtaining based on this decoding unit, according to the reconstruction pixel of this first basic tomographic image sub-block, be identified for the second reference information that this first object image subblock is encoded, wherein, this first basic tomographic image sub-block is the image block that is arranged in basic tomographic image, this target image piece is arranged in enhancement layer image, the basic tomographic image being somebody's turn to do is corresponding with this enhancement layer image, and this first primary image piece sub-block locus in this basic tomographic image is corresponding with this first object image subblock locus in this enhancement layer image, this coding unit is also for this object code stream is decoded, to obtain this target image piece.
In a kind of possible execution mode, this decoding unit carries out block-eliminating effect filtering processing specifically for the pixel of the boundary vicinity between this target image sub-block.
In conjunction with the tenth aspect and the possible execution mode of the first, in the possible execution mode of the second, this decoding unit is specifically for from target code stream, obtain the 4th indication information, wherein, the 4th indication information is adjacent with skip mode flag bit or fusion MERGE mode flags position information in this target code stream.
In conjunction with the tenth aspect, execution mode and the possible execution mode of the second that the first is possible, in the third possible execution mode, this decoding unit, specifically for whether using reference information to decode according to the reference image block that is positioned at predeterminated position in this enhancement layer image, is determined context; For according to this context, carry out entropy decoding, to determine the 4th indication information.
The tenth on the one hand, and a kind of encoder of processing for image is provided, and this encoder comprises: bus, the processor being connected with this bus, the memory being connected with this bus, wherein, this processor is by this bus, call the program of storing in this memory, for when determining that the movable information of the first basic tomographic image sub-block corresponding with the first object image subblock of target image piece is empty, according to the reconstruction pixel of this first basic tomographic image sub-block, be identified for the second reference information that this first object image subblock is encoded, wherein, this first basic tomographic image sub-block is the image block that is arranged in basic tomographic image, this target image piece is arranged in enhancement layer image, the basic tomographic image being somebody's turn to do is corresponding with this enhancement layer image, and this first primary image piece sub-block locus in this basic tomographic image is corresponding with this first object image subblock locus in this enhancement layer image, this target image piece is encoded, to generate target code stream and to be contained in the 4th indication information in this target code stream.
In a kind of possible execution mode, this processor is specifically used the pixel of the boundary vicinity between this target image sub-block is carried out to block-eliminating effect filtering processing.
In conjunction with the tenth one side and the possible execution mode of the first, in the possible execution mode of the second, this processor is specifically for the 4th indication information is carried out to entropy coding, so that the 4th indication information is adjacent with skip mode flag bit or fusion MERGE mode flags position information in this target code stream.
In conjunction with the tenth on the one hand, execution mode and the possible execution mode of the second that the first is possible, in the third possible execution mode, this processor, specifically for whether using reference information to encode according to the reference image block that is positioned at predeterminated position in this enhancement layer image, is determined context; For according to this context, the 4th indication information is carried out to entropy coding.
The 12 aspect, provides a kind of decoder of processing for image, and this decoder comprises: bus, the processor being connected with this bus, the memory being connected with this bus, wherein, this processor, by this bus, calls the program of storing in this memory, for from target code stream, obtains the 4th indication information, for when determining that the movable information of the first basic tomographic image sub-block corresponding with the first object image subblock of target image piece is empty, based on the 4th indication information, according to the reconstruction pixel of this first basic tomographic image sub-block, be identified for the second reference information that this first object image subblock is encoded, wherein, this first basic tomographic image sub-block is the image block that is arranged in basic tomographic image, this target image piece is arranged in enhancement layer image, the basic tomographic image being somebody's turn to do is corresponding with this enhancement layer image, and this first primary image piece sub-block locus in this basic tomographic image is corresponding with this first object image subblock locus in this enhancement layer image, for this object code stream is decoded, to obtain this target image piece.
In a kind of possible execution mode, this processor carries out block-eliminating effect filtering processing specifically for the pixel of the boundary vicinity between this target image sub-block.
In conjunction with the 12 aspect and the possible execution mode of the first, in the possible execution mode of the second, this processor is specifically for from target code stream, obtain the 4th indication information, wherein, the 4th indication information is adjacent with skip mode flag bit or fusion MERGE mode flags position information in this target code stream.
In conjunction with the 12 aspect, this processor, specifically for whether using reference information to decode according to the reference image block that is positioned at predeterminated position in this enhancement layer image, determine context; For according to this context, carry out entropy decoding, to determine the 4th indication information.
According to the method and apparatus of processing for image of the embodiment of the present invention, for the corresponding sub-block that can not comprise from basic tomographic image in the target image piece of enhancement layer image, obtain the first object image subblock of movable information, by according to the location positioning second target image sub-block of described first object image subblock, and according to the movable information of this second target image sub-block or on locus the reconstruction pixel of the first basic tomographic image sub-block corresponding with this first object image subblock, determine the reference information for this first object image subblock, and encode according to this reference information, can improve the coding efficiency of this first object image subblock.
Accompanying drawing explanation
In order to be illustrated more clearly in the technical scheme of the embodiment of the present invention, to the accompanying drawing of required use in the embodiment of the present invention be briefly described below, apparently, below described accompanying drawing be only some embodiments of the present invention, for those of ordinary skills, do not paying under the prerequisite of creative work, can also obtain according to these accompanying drawings other accompanying drawing.
Fig. 1 is the indicative flowchart of the method processed for image according to an embodiment of the invention.
Fig. 2 is the schematic diagram of partition and sub-block index according to an embodiment of the invention.
Fig. 3 is the meaning flow chart of the method processed for image according to another embodiment of the present invention.
Fig. 4 is the schematic block diagram of the device processed for image according to an embodiment of the invention.
Fig. 5 is the schematic block diagram of the device processed for image according to another embodiment of the present invention.
Fig. 6 is the schematic block diagram of the encoder processed for image according to an embodiment of the invention.
Fig. 7 is the schematic block diagram of the decoder processed for image according to another embodiment of the present invention.
Fig. 8 is the indicative flowchart of the method for processing for image according to yet another embodiment of the invention
Fig. 9 is the indicative flowchart of the method for processing for image according to yet another embodiment of the invention
Figure 10 is the schematic block diagram of the device of processing for image according to yet another embodiment of the invention.
Figure 11 is the schematic block diagram of the device of processing for image according to yet another embodiment of the invention.
Figure 12 is the schematic block diagram of the encoder of processing for image according to yet another embodiment of the invention.
Figure 13 is the schematic block diagram of the decoder of processing for image according to yet another embodiment of the invention.
Embodiment
Below in conjunction with the accompanying drawing in the embodiment of the present invention, the technical scheme in the embodiment of the present invention is clearly and completely described, obviously, described embodiment is the present invention's part embodiment, rather than whole embodiment.Embodiment based in the present invention, those of ordinary skills, not making the every other embodiment obtaining under creative work prerequisite, belong to the scope of protection of the invention.
Fig. 1 show from coding side angle, describe according to the indicative flowchart of the method 100 of processing for image of the embodiment of the present invention.As shown in Figure 1, the method 100 comprises:
S110, when determining that the movable information of the first basic tomographic image sub-block corresponding with the first object image subblock of target image piece is empty, the size of each the target image sub-block comprising according to the size of this target image piece, this target image piece and be used to indicate the second indication information of the position of first object image subblock in this target image piece, determines the second target image sub-block;
S120, according to the movable information of this second target image sub-block, be identified for the first reference information that this first object image subblock is encoded, wherein, this first basic tomographic image sub-block is the image block that is arranged in basic tomographic image, this target image piece is arranged in enhancement layer image, the basic tomographic image being somebody's turn to do is corresponding with this enhancement layer image, and this first primary image piece sub-block locus in this basic tomographic image is corresponding with this first object image subblock locus in this enhancement layer image;
S130, encodes to this target image piece, to generate target code stream and to be contained in the first indication information in this target code stream.
Specifically, image is being carried out to hierarchical coding, for example, when spatial scalable is encoded, image can be carried out resolution processes to obtain low-resolution image, as a comparison original image is called to high-definition picture, encoder is encoded to this low-resolution image and this high-definition picture respectively.For convenience of describing, the high image to be encoded of quality is called to enhancement layer image herein, the image low-quality to be encoded of correspondence (for example described low-resolution image) is called to basic tomographic image.
In embodiments of the present invention, target image is the image that uses hierarchical coding technology to process, basic layer refers to the lower layer of quality (comprising the parameters such as frame rate, spatial resolution, temporal resolution, signal to noise ratio intensity or credit rating) in hierarchical coding, and enhancement layer refers to the higher layer of quality (comprising the parameters such as frame rate, spatial resolution, temporal resolution, signal to noise ratio intensity or credit rating) in hierarchical coding.It should be noted that, in embodiments of the present invention, in embodiments of the present invention, for a given enhancement layer, basic layer corresponding thereto can be quality lower than arbitrary layer of this enhancement layer, for example, if five layers of current existence, (coding quality improves successively, ground floor quality is minimum, and layer 5 quality is the highest), if enhancement layer is the 4th layer, basic layer can be ground floor, can be also the second layer, also can be the 3rd layer, also can be the 4th layer.In like manner, for a given basic layer, enhancement layer corresponding thereto can be the arbitrary layer of quality lower than this basic layer.
Enhancement layer image is the image in the enhancement layer of pre-treatment, basic tomographic image be in basic layer with the image of enhancement layer image at synchronization.
In sum, in embodiments of the present invention, the quality of this basic tomographic image is lower than the quality of this enhancement layer image.
The image block of target image piece for processing in this enhancement layer image.
Basic tomographic image piece is in basic tomographic image, to have on locus the image block of corresponding relation with this target image piece.
In embodiments of the present invention, the image block in basic layer and the corresponding relation of the image block in enhancement layer can calculate according to the resolution proportionate relationship between basic tomographic image and enhancement layer image.For example, within comprising the system of x direction and y direction, if enhancement layer image is respectively 2 times of basic tomographic image in the resolution of x direction and y direction, the pixel coordinate for the upper left corner in enhancement layer is (2x, 2y) and size be the image block of (2m) * (2n), corresponding blocks in its basic tomographic image can be that the pixel coordinate in the upper left corner is that (x, y) and size are the image block of m * n.
In embodiments of the present invention, aftermentioned sub-block refers to the sub-block (image block in enhancement layer) of target image piece, and the corresponding sub-block of aftermentioned refers to the corresponding image block in basic layer of this sub-block.
In embodiments of the present invention, movable information can comprise prediction direction, one or more in reference picture index or motion vector, wherein, prediction direction can be divided into unidirectional and bi-directional predicted, single directional prediction can be divided into forward prediction and back forecast again, forward prediction refers to use forward direction reference picture list, it is the reference picture generation prediction signal in list (list) 0, back forecast refers to use backward reference picture list, it is the reference picture generation prediction signal in list1, bi-directional predicted finger is used the reference picture in list0 and list1 to produce prediction signal simultaneously, for single directional prediction, need a reference picture index indication selected reference picture in list0 or list1, for bi-directional predicted, need two reference picture index, indicate respectively selected reference picture in list0 and list1, each motion vector comprises horizontal direction component x and vertical direction component y, can be denoted as (x, y), for single directional prediction, need the displacement of a motion vector indication predicting signal in selected list0 or list1 reference picture, for bi-directional predicted, need two motion vectors, indicate respectively the displacement in selected list0 reference picture and list1 reference picture of forward prediction signal and back forecast signal.
In embodiments of the present invention, target image piece can be regarded as and consists of at least two sub-blocks (that is, target image sub-block), and wherein, the size of this sub-block can determine according to preset value, for convenience of explanation, below, the sub-block size of take describes as example as 4 * 4.For example, if the size of target image piece is 16 * 16, can determine that this target image piece comprises 16 sub-blocks (size is 4 * 4).Thereby, in embodiments of the present invention, can determine each sub-block in this target image piece corresponding sub-block (belonging to this correspondence image piece) in basic layer, and determine the movable information of this correspondence sub-block.
In embodiments of the present invention, can according to the coordinate of a certain pixel in sub-block, (be denoted as: " (E x, E y) "), determine that the coordinate of this pixel correspondence position in basic tomographic image (is denoted as: " (B x, B y) "), and using the image block in the basic layer that comprises affiliated corresponding position coordinates as described corresponding sub-block.In embodiments of the present invention, can be according to calculating (B with following formula 1 and formula 2 x, B y):
B x = Round ( E x × D x + R x 2 s - 4 ) - - - ( 1 )
B y = Round ( F y × D y + R y 2 s - 4 ) - - - ( 2 )
Wherein, Round() represent to block the operation of fractional part, R xand R yrepresent side-play amount, can be according to calculating R with following formula 3 x, according to calculating R with following formula 4 y
R x=2 s-5 (3)
R y=2 s-5 (4)
Wherein, S is precision controlling elements (for example, in the embodiment of the present invention, can be set to 16), can be according to calculating D with following formula 5 x, according to calculating D with following formula 6 y
D x = Round ( 2 s * BaseWidth ScaledBaseWidth ) - - - ( 5 )
D y = Round ( 2 s * BaseHeight ScaledBaseHeight ) - - - ( 6 )
Wherein, BaseWidth represents the wide of basic tomographic image, and BaseHeight represents the height of basic tomographic image, and ScaledBaseWidth represents the wide of enhancement layer image, and ScaledBaseHeight represents the height of enhancement layer image.
Thereby, can determine this correspondence sub-block, and, in the situation that this correspondence sub-block comprises movable information, can directly use prediction direction in this movable information and reference picture index as prediction direction and the reference picture index of this sub-block (first object image subblock).Can according to following formula 7 to formula 10, the motion vector (BMV to this correspondence sub-block x, BMV y) carry out convergent-divergent, and using the motion vector after convergent-divergent as this sub-block the motion vector (EMV of (first object image subblock) x, EMV y).
EMV x=(BMV x×ScaledBaseWidth+R BW)/BaseWidth (7)
EMV y=(BMV y×ScaledBaseHeight+R BH)/BaseHeight (8)
R BW=sgn(BMV x)*BaseWidth/2 (9)
R BH=sgn(BMV y)*BaseHeight/2 (10)
Wherein, sgn (x) is-symbol function, can obtain the symbol of x.
Here, it should be noted that, if basic tomographic image is identical with the resolution of enhancement layer image, without carrying out above-mentioned zoom operations, and can directly use the movable information of corresponding sub-block as the movable information of sub-block.
By above method, can determine that the corresponding sub-block that can comprise from basic tomographic image in target image piece obtains the sub-block of movable information.
Thereby, for the sub-block of the corresponding sub-block acquisition movable information that can comprise from basic tomographic image, can be by the movable information of its corresponding sub-block
Alternatively, in embodiments of the present invention, the method also comprises:
According to the coding mode of described basic tomographic image, determine whether the first basic tomographic image sub-block corresponding with first object image subblock comprises movable information.
Specifically, in embodiments of the present invention, can determine whether this correspondence sub-block includes movable information according to the coding mode of described basic tomographic image (correspondence image piece).For example, if basic tomographic image is used intraframe predictive coding pattern, can determine that this correspondence sub-block does not comprise movable information (that is, the movable information of the first basic tomographic image sub-block is empty).
Thereby, when this correspondence sub-block comprises movable information, can determine this correspondence sub-block and obtain its movable information by process as mentioned above, when this correspondence sub-block does not comprise movable information, can skip above flow process.
For the corresponding sub-block that can not comprise from basic tomographic image, obtain the sub-block (that is, first object image subblock) of movable information, S110, can be by the following method 1, determine the second target image sub-block, and S120, movable information according to this second target image sub-block, obtains the first reference information.
Method 1
Can be that this target image sub-block fills movable information, thereby can using the movable information of this filling as with reference to information.
Below, describe in detail in the embodiment of the present invention, fill the method for movable information.
Specifically, without loss of generality, for example, if the size of target image piece is 16 * 16, the size of sub-block is 4 * 4, and in embodiments of the present invention, the index assignment method of sub-block can be same as the prior art, here, the description thereof will be omitted, shown in Fig. 2 division and the index of sub-block.
In embodiments of the present invention, can determine and process level according to the size of the size of target image piece and sub-block, and process by processing level Layer by layer recurrence.For example, in embodiments of the present invention, the level that can stipulate the bottom (is denoted as, ground floor) each processing unit in (being denoted as the first processing unit) comprises four sub-blocks, and each processing unit of the last layer of ground floor time (being denoted as the second layer) (is denoted as, the second processing unit) comprise four the first processing units, by that analogy, for fear of repeating, omit recursion explanation.Thereby, as example and non-limiting, in the target image piece shown in Fig. 2, can comprise two levels, in ground floor, sub-block 0 ~ sub-block 3 has formed the first processing unit 0, and sub-block 4 ~ sub-block 7 has formed the first processing unit 1, sub-block 8 ~ sub-block 11 has formed the first processing unit 2, and sub-block 12 ~ sub-block 15 has formed the first processing unit 3.In the second layer, the first processing unit 0 ~ the first processing unit 3 has formed the second processing unit 0.Should be understood that the above hierarchical division method of enumerating is only exemplary illustration, the present invention is not limited thereto.
In embodiments of the present invention, for each the first processing unit, can be according to the call number of sub-block (for example, whether the movable information of from small to large) judging successively each sub-block is as empty, if the movable information of this sub-block is empty, the movable information of sub-block (example of the second target image sub-block) that can be based on being adjacent in this first processing unit is determined its movable information.For example, if the movable information of the sub-block that index is 0 (that is, sub-block 0, belongs to the first processing unit 0) is for empty, can obtain the movable information that belongs to other sub-blocks in same processing unit (the first processing unit 0), and the movable information using this movable information as this sub-block 0.Obtaining order can be, for example, first obtain index and be 1 sub-block (sub-block 1, , one example of the second target image sub-block, adjacent with sub-block 0 in the horizontal direction) movable information, if the movable information of sub-block 1 is empty, can obtain index is 2(sub-block 2 again, , another example of the second target image sub-block, in the vertical direction is adjacent with sub-block 0) the movable information of sub-block, if the movable information of sub-block 2 is empty, can obtain index is 3(sub-block 3 again, , an example again of the second target image sub-block, adjacent with sub-block 0 in diagonal) the movable information of sub-block.In like manner, for each movable information, be empty sub-block, all can to its movable information, fill by above method.Should understand, the movable information fill method that is empty sub-block to movable information of more than enumerating is only exemplary illustration, the present invention is not limited thereto, for example, and for the above-mentioned order of obtaining, also can first obtain regulation sub-block on vertical direction (here, for adjacent sub-blocks) movable information, then obtain the movable information of the regulation sub-block (being adjacent sub-blocks here) in horizontal direction, obtain again the movable information of the regulation sub-block (being adjacent sub-blocks here) in diagonal.That is, this obtains sequentially and can change arbitrarily.
Thereby, through the above-mentioned processing of carrying out in ground floor, each sub-block in each first processing unit is carried out after above-mentioned processing, as long as have the movable information of at least one sub-block in four sub-blocks in this first processing unit not for empty, can be just that empty sub-block is filled (obtaining in other words) movable information for all movable informations in this first processing unit.
It should be noted that, for filled the sub-block of movable information according to said method, while needing to use the movable information of this sub-block, can directly use the movable information of filling into this sub-block in subsequent treatment.; the movable information of the second target image sub-block can refer to the movable information of the corresponding sub-block of basic layer of this second target image sub-block, can be also (from other enhancement layer sub-blocks) movable information of filling for this second target image sub-block according to the method for the filling movable information of the embodiment of the present invention.
Therefore, for example, when the corresponding sub-block of definite some sub-blocks (, sub-block 0) does not comprise movable information, can be from ground floor with this sub-block 0 in same the first processing unit (for example, the first processing unit 0) other sub-blocks (for example, sub-block 1 ~ sub-block 3) are obtained movable information.When with this sub-block (for example, sub-block 0) in same the first processing unit (for example, the first processing unit 0) other sub-blocks (regulation sub-block in ground floor, for example, sub-block 1 ~ sub-block 3) movable information is empty, (for example can obtain regulation the first processing unit in the second processing unit, the first processing unit 1 ~ the first processing unit 3) movable information of regulation sub-block (another example of the second target image sub-block) in, and using this movable information as this sub-block the movable information of (for example, sub-block 0).
; for example, if first processing unit (; the corresponding sub-block of all sub-blocks the first processing unit 0) is sky; (for example can obtain other first processing units in the second processing unit; the first processing unit 1 ~ the first processing unit 3) the regulation sub-block in (for convenience of explanation; in the present invention, take the sub-block in each the first processing unit upper left corner describes as example) movable information, and using this movable information as this first processing unit the movable information of each sub-block in (the first processing unit 0).Obtaining order can be, for example, first obtain index and be the first processing unit (the first processing unit 1 of 1, it is adjacent with the first processing unit 0 in the horizontal direction) the sub-block (sub-block 4 in the upper left corner, , one example of the second target image sub-block) movable information, if the movable information of sub-block 4 is empty, can think that the movable information of these first processing unit, 1 interior other sub-blocks is also sky, thereby can obtain again index and be the first processing unit (the first processing unit 2 of 2, its in the vertical direction is adjacent with the first processing unit 0) the sub-block (sub-block 8 in the upper left corner, , another example of the second target image sub-block) movable information, if the movable information of sub-block 8 is empty, can think that the movable information of these first processing unit, 2 interior other sub-blocks is also sky, thereby can obtain again index and be the first processing unit (the first processing unit 3 of 3, it is adjacent with the first processing unit 0 in diagonal) the sub-block (sub-block 12 in the upper left corner, , an example again of the second target image sub-block) movable information.In like manner, for each movable information, be the first empty processing unit, all can to its movable information, fill by above method.Should understand, the movable information fill method that is empty sub-block to movable information of more than enumerating is only exemplary illustration, the present invention is not limited thereto, for example, for the above-mentioned order of obtaining, also can first obtain regulation the first processing unit on vertical direction (here, for adjacent the first processing unit) the movable information of regulation sub-block, obtain again regulation the first processing unit in horizontal direction (here, for adjacent the first processing unit) the movable information of regulation sub-block, obtain again regulation the first processing unit in diagonal (here, for adjacent the first processing unit) the movable information of regulation sub-block.That is, this obtains sequentially and can change arbitrarily.And, as above-mentioned " regulation sub-block ", be not limited to the sub-block in the first processing unit upper left corner, can be also the sub-block of optional position in same the first processing unit.
It should be noted that, because the size of the above-mentioned target image piece of enumerating is 16 * 16, the size of sub-block is 4 * 4, so this target image piece only comprises two-layerly, and above-mentioned recursive procedure finishes.But the size of target image piece is for larger, for example, 32 * 32, and the size of sub-block is 4 * 4, this target image piece comprises three layers, can proceed recursive operation according to method same as described above, for all sub-blocks when target image piece are obtained movable information.
Alternatively, the size of the target image sub-block that this comprises according to the size of target image piece, this target image piece and be used to indicate the second indication information of the position of first object image subblock in this target image piece, determines the second target image sub-block, comprising:
According to following arbitrary formula, determine this second target image sub-block,
Idx 2=Idx 1/N×N+((Idx 1%N/(N/2))×2+(1-Idx 1%N/(N/4)%2))×N/4;
Idx 2=Idx 1/N×N+((1-Idx 1%N/(N/2))×2+(Idx 1%N/(N/4)%2))×N/4;
Idx 2=Idx 1/N×N+((1-Idx 1%N/(N/2))×2+(1-Idx 1%N/(N/4)%2))×N/4;
Wherein, Idx 2represent to be used to indicate the 3rd indication information of this position of the second target image sub-block in this target image piece, Idx 1represent this second indication information, N determines according to the size of the size of this target image piece and this target image sub-block.
Wherein, Idx 2represent to be used to indicate the 3rd index information of this position of the second target image sub-block in this target image piece, Idx 1represent this second index information, % represents modular arithmetic or remainder operation, and N represents the quantity of the sub-block that this target image piece comprises.
Specifically, can be according to the index of the current sub-block of processing according to above-mentioned each formula, determine the second target image sub-block in the level of pre-treatment, wherein N is corresponding with the current level of processing, and, N determines according to the size of the size of target image piece and sub-block, for example, if the size of target image piece is 16 * 16, the size of sub-block is 4 * 4, as mentioned above this target image piece comprise two-layer, when processing ground floor, N, for the sub-block quantity that each processing unit (the first processing unit) in this layer comprises, is 4 here.When processing the second layer, N, for the sub-block quantity that in this layer, each processing unit (the second processing unit) comprises, is 16 here.
Above, during upper left corner sub-block that to have enumerated when above-mentioned " regulation sub-block " be described processing unit, the formula using.But the present invention does not limit therewith, can also, according to the position in " regulation sub-block " described processing unit, above-mentioned formula be changed.
Alternatively, this determines the first reference information of this first object image subblock according to the movable information of this second target image sub-block, comprising:
If the movable information of this second target image sub-block is empty, determine that this first reference information is zero movable information.
Specifically, if after above-mentioned processing, can not fill movable information for this sub-block, use zero movable information as the movable information of this sub-block.In embodiments of the present invention, can construct in accordance with the following methods zero movable information.For example, in predictive-coded picture frame (P frame), the prediction direction of zero movable information is single directional prediction, and reference picture index is 0, motion vector (0,0).In bidirectionally predictive coded picture frame (B frame), the prediction direction of zero movable information is bi-directional predicted, and two reference picture index are 0, two motion vector and are (0,0).
It should be noted that, when target image piece comprises a plurality of processing levels, above-mentioned use zero movable information, as the method for the movable information of this sub-block, can carry out after last level is processed, also can be to other, arbitrary level carry out after processing, the present invention is also not particularly limited.Should understand, the method of the movable information that obtains sub-block of more than enumerating is only exemplary illustration of the present invention, the present invention is not limited thereto, for example, in the present invention, be in example, can also as mentioned above, according to the coding mode of described basic tomographic image (correspondence image piece), determine whether this correspondence sub-block includes movable information.For example, if basic tomographic image is used intraframe predictive coding pattern, can determine that this correspondence sub-block does not comprise movable information (that is, the movable information of the first basic tomographic image sub-block is empty).If determine in all sub-blocks of target image piece and only have a sub-block (specifically, being its corresponding sub-block) to there is movable information, movable information that can be using the movable information of this sub-block as other sub-blocks.
Thus, by said method, can determine the first reference information of first object image subblock.
In S130, by said method 1, obtaining first object image block first reference information of (in basic layer, corresponding sub-block does not comprise movable information).
Alternatively, this is encoded to this target image piece, comprising:
According to this first reference information, this first object image subblock is carried out to motion compensation process.
Specifically, can also be according to the reference information of first object image subblock (specifically movable information), to this first object image subblock encode (being specifically, motion compensation process).Specifically, can, according to the movable information that obtains or fill for this first object image subblock as mentioned above, to this first object image subblock, carry out independently motion compensation process.
In embodiments of the present invention, can also obtain the 3rd target image piece movable information of (in basic layer, corresponding sub-block comprises movable information), the method for obtaining the movable information of the 3rd target image piece can be same as the prior art, and the description thereof will be omitted here.
Thereby, all sub-blocks of target image piece are being carried out after motion compensation process, obtain after the prediction signal of target image piece, can carry out predictive coding to target image piece, thereby calculate rate distortion costs.Can calculate after the distortion rate cost of target image piece, if this distortion rate Least-cost, can determine a sign (the first indication information), with indication decoding end, by said method 1 and method 2, obtain the reference information of first object image block, and obtain the 3rd target image piece movable information of (in basic layer, corresponding sub-block comprises movable information).And this first indication information is carried out to entropy coding.
Alternatively, this is encoded to this first object image subblock according to this reference information, comprising:
This first indication information is carried out to entropy coding, so that this first indication information is adjacent with skip mode flag bit or fusion MERGE mode flags position information in this target code stream.
Specifically, in embodiments of the present invention, in target code stream, the first indication information can be configured in to the position adjacent with skip mode flag bit information.Specifically,
For example, the first indication information can be placed in to skip mode flag bit information first information at target code stream as target image piece before, the first indication information can also be placed in to skip mode flag bit information second information at target code stream as target image piece afterwards.Above-mentioned skip mode can be pattern same as the prior art, and it determines that method and allocation position can be same as the prior art, and here, for fear of repeating, the description thereof will be omitted.
Again for example, the first indication information can be placed in to the position adjacent with MERGE mode flags position information.Concrete, before the first indication information can being placed in to MERGE mode flags position information, after the first indication information can also being placed in to MERGE mode flags position information.Described MERGE can be pattern same as the prior art, and it determines that method and allocation position can be same as the prior art, and here, for fear of repeating, the description thereof will be omitted.
In embodiments of the present invention, this first indication information can be a binary flags position.Therefore, when the first indication information being carried out to entropy coding, without the first indication information is carried out to binary conversion treatment.
Thereafter, can select for the first indication information of binaryzation being carried out to the context that entropy when coding used, alternatively, this is encoded to this first object image subblock according to this reference information, comprising:
According to the reference image block that is positioned at predeterminated position in this enhancement layer image, whether use reference information to encode, determine context;
According to this context, this first indication information is carried out to entropy coding.
Specifically.As shown in the following Table 1, this context can comprise 0,1,2 three context.Whether the present embodiment is determined and is used which context with basic layer model according to the image block of the left side and top.In the present embodiment, for example, can whether use its first indication information separately according to the image block of the target image piece left side and top, determine and use any context, for example, if the image block of the target image piece left side and top is not used the first indication information, selecting so index is 0 context model, if the image block of the target image piece left side and top has one to use the first indication information, selecting so index is 1 context model, if the image block of the target image piece left side and top is all used the first indication information, selecting so index is 2 context.
Table 1
Figure BDA00002218336900231
Figure BDA00002218336900241
, can according to the context as mentioned above selected to this first indication information carry out binary arithmetic coding, and upgrade the context model of using thereafter.In embodiments of the present invention, this process can be same as the prior art, and for fear of repeating, the description thereof will be omitted here.
In embodiments of the present invention, can be according to obtained reference information, each first object image subblock is encoded, and then complete the coding to target image piece, and the first indication information after the information of the target image piece of generation and above-mentioned entropy coding is added to code stream (target code stream).
Here, it should be noted that, in this target code stream, can comprise target image (comprising basic tomographic image and the enhancement layer image) information after coding, and this processing procedure can be same as the prior art, here, for fear of repeating, the description thereof will be omitted.
In decoding end, can from code stream, obtain target image information, and definite target image (specifically, target image piece), the first indication information obtaining (information after entropy coding) is carried out to entropy decoding, in this processing procedure, context is selected with the process of updating context and the processing of above-mentioned coding side same or similar, and the description thereof will be omitted here.
Can from code stream, resolve the binary character word string (bin string) that obtains representing the first indication information according to the context of selecting.Wherein, this binary arithmetic decoding is corresponding with the binary arithmetic coding of coding side.
In embodiments of the present invention, can stipulate, for example, when the first indication information is 1, decoding end need to be used the method identical with coding side to obtain the first reference information of first object image subblock, and decoding end need to be used the method identical with coding side to obtain the first reference information of first object image subblock.
The indicating means that should be understood that above the first indication information of enumerating is only exemplary illustration, and the present invention is not limited thereto.
It should be noted that, in embodiments of the present invention, when using method 1 is obtained the reference information of first object image subblock, can be according to the reference information of first object image subblock (specifically movable information), to this first object image subblock decode (being specifically, motion compensation process).Specifically, can, according to being the movable information that this first object image subblock is filled as mentioned above, to this first object image subblock, carry out independently motion compensation process.
Alternatively, this is encoded and comprises this target image piece:
Pixel to the boundary vicinity between this target image sub-block is carried out block-eliminating effect filtering processing.
Specifically, can also carry out filtering processing to the boundary vicinity pixel between each sub-block of target image piece.
According to the method for processing for image of the embodiment of the present invention, for the corresponding sub-block that can not comprise from basic tomographic image in the target image piece of enhancement layer image, obtain the first object image subblock of movable information, by according to the location positioning second target image sub-block of this first object image subblock, and according to the movable information of this second target image sub-block, determine the first reference information for this first object image subblock, and encode according to this first reference information, can improve the coding efficiency of this first object image subblock.
Fig. 3 show from decoding end angle, describe according to the indicative flowchart of the method 200 of processing for image of the embodiment of the present invention.As shown in Figure 2, the method 200 comprises:
S210, from target code stream, obtains the first indication information;
S220, when the movable information of the first corresponding basic tomographic image sub-block of the first object image subblock with target image piece is sky, based on this first indication information, the size of each the target image sub-block comprising according to the size of this target image piece, this target image piece and be used to indicate the second indication information of the position of first object image subblock in this target image piece, determines the second target image sub-block;
S230, according to the movable information of this second target image sub-block, be identified for the first reference information that this first object image subblock is decoded, wherein, this first basic tomographic image sub-block is the image block that is arranged in basic tomographic image, this target image piece is arranged in enhancement layer image, the basic tomographic image being somebody's turn to do is corresponding with this enhancement layer image, and this first primary image piece sub-block locus in this basic tomographic image is corresponding with this first object image subblock locus in this enhancement layer image;
S230, decodes to this object code stream, to obtain this target image piece.
Specifically, at S210, decoding end can be obtained target image information from code stream, and definite target image (specifically, being target image piece), and the first indication information obtaining (information after entropy coding).
Alternatively, should from target code stream, obtain the first indication information, comprise:
From target code stream, obtain the first indication information, wherein, this first indication information is adjacent with skip mode flag bit or fusion MERGE mode flags position information in this target code stream.
Specifically, in embodiments of the present invention, in target code stream, the first indication information can be configured in to the position adjacent with skip mode flag bit information.Specifically,
For example, the first indication information can be placed in to skip mode flag bit information first information at target code stream as target image piece before, the first indication information can also be placed in to skip mode flag bit information second information at target code stream as target image piece afterwards.Above-mentioned skip mode can be pattern same as the prior art, and it determines that method and allocation position can be same as the prior art, and here, for fear of repeating, the description thereof will be omitted.
Again for example, the first indication information can be placed in to the position adjacent with MERGE mode flags position information.Concrete, before the first indication information can being placed in to MERGE mode flags position information, after the first indication information can also being placed in to MERGE mode flags position information.This MERGE can be pattern same as the prior art, and it determines that method and allocation position can be same as the prior art, and here, for fear of repeating, the description thereof will be omitted.
Thereafter, decoding end can be carried out entropy decoding to the first indication information obtaining.
In embodiments of the present invention, this first indication information can be a binary flags position.Therefore, when the first indication information being carried out to entropy decoding, without making the first indication information binaryzation.
, can select for first indication information of binaryzation carried out to context that entropy when coding used, alternatively, should from target code stream obtain first indication information, comprise thereafter:
According to the reference image block that is positioned at predeterminated position in this enhancement layer image, whether use reference information to decode, determine context;
According to this context, carry out entropy decoding, to determine this first indication information.
Specifically.As shown in Table 1 above, this context can comprise 0,1,2 three context.Whether the present embodiment is determined and is used which context with basic layer model according to the image block of the left side and top.In the present embodiment, for example, can whether use its first indication information separately according to the image block of the target image piece left side and top, determine and use any context, for example, if the image block of the target image piece left side and top is not used the first indication information, selecting so index is 0 context model, if the image block of the target image piece left side and top has one to use the first indication information, selecting so index is 1 context model, if the image block of the target image piece left side and top is all used the first indication information, selecting so index is 2 context.
, can according to the context of as above this selection to this first indication information carry out binary arithmetic decoding, and upgrade the context model of using thereafter.In embodiments of the present invention, this process can be same as the prior art, and for fear of repeating, the description thereof will be omitted here.
In embodiments of the present invention, can stipulate, for example, when the first indication information is 1, decoding end need to be used the method identical with coding side to obtain the first reference information of first object image subblock, and decoding end need to be used the method identical with coding side to obtain the first reference information of first object image subblock.
The indicating means that should be understood that above the first indication information of enumerating is only exemplary illustration, and the present invention is not limited thereto.
Therefore, decoding end can determine whether to obtain the first reference information of first object image subblock according to this first indication information, below, the situation that need to obtain the first reference information of first object image subblock to coding side describes.
Image is being carried out to hierarchical coding, for example, when spatial scalable is encoded, image can be carried out resolution processes to obtain low-resolution image, as a comparison original image is called to high-definition picture, encoder is encoded to this low-resolution image and this high-definition picture respectively.For convenience of describing, the high image to be encoded of quality is called to enhancement layer image herein, the image low-quality to be encoded of correspondence (for example this low-resolution image) is called to basic tomographic image.
In embodiments of the present invention, target image is the image that uses hierarchical coding technology to process, basic layer refers to the lower layer of quality (comprising the parameters such as frame rate, spatial resolution, temporal resolution, signal to noise ratio intensity or credit rating) in hierarchical coding, and enhancement layer refers to the higher layer of quality (comprising the parameters such as frame rate, spatial resolution, temporal resolution, signal to noise ratio intensity or credit rating) in hierarchical coding.It should be noted that, in embodiments of the present invention, in embodiments of the present invention, for a given enhancement layer, basic layer corresponding thereto can be quality lower than arbitrary layer of this enhancement layer, for example, if five layers of current existence, (coding quality improves successively, ground floor quality is minimum, and layer 5 quality is the highest), if enhancement layer is the 4th layer, basic layer can be ground floor, can be also the second layer, also can be the 3rd layer, also can be the 4th layer.In like manner, for a given basic layer, enhancement layer corresponding thereto can be the arbitrary layer of quality lower than this basic layer.
Enhancement layer image is the image in the enhancement layer of pre-treatment, basic tomographic image be in basic layer with the image of enhancement layer image at synchronization.
To sum up be somebody's turn to do, in embodiments of the present invention, the quality of this basic tomographic image is lower than the quality of this enhancement layer image.
The image block of target image piece for processing in this enhancement layer image.
Basic tomographic image piece is in basic tomographic image, to have on locus the image block of corresponding relation with this target image piece.
In embodiments of the present invention, the image block in basic layer and the corresponding relation of the image block in enhancement layer can calculate according to the resolution proportionate relationship between basic tomographic image and enhancement layer image.For example, within comprising the system of x direction and y direction, if enhancement layer image is respectively 2 times of basic tomographic image in the resolution of x direction and y direction, the pixel coordinate for the upper left corner in enhancement layer is (2x, 2y) and size be the image block of (2m) * (2n), corresponding blocks in its basic tomographic image can be that the pixel coordinate in the upper left corner is that (x, y) and size are the image block of m * n.
In embodiments of the present invention, aftermentioned sub-block refers to the sub-block (image block in enhancement layer) of target image piece, and the corresponding sub-block of aftermentioned refers to the corresponding image block in basic layer of this sub-block.
In embodiments of the present invention, movable information can comprise prediction direction, one or more in reference picture index or motion vector, wherein, prediction direction can be divided into unidirectional and bi-directional predicted, single directional prediction can be divided into forward prediction and back forecast again, forward prediction refers to use forward direction reference picture list, it is the reference picture generation prediction signal in list (list) 0, back forecast refers to use backward reference picture list, it is the reference picture generation prediction signal in list1, bi-directional predicted finger is used the reference picture in list0 and list1 to produce prediction signal simultaneously, for single directional prediction, need a reference picture index indication selected reference picture in list0 or list1, for bi-directional predicted, need two reference picture index, indicate respectively selected reference picture in list0 and list1, each motion vector comprises horizontal direction component x and vertical direction component y, can be denoted as (x, y), for single directional prediction, need the displacement of a motion vector indication predicting signal in selected list0 or list1 reference picture, for bi-directional predicted, need two motion vectors, indicate respectively the displacement in selected list0 reference picture and list1 reference picture of forward prediction signal and back forecast signal.
In embodiments of the present invention, target image piece can be regarded as and consists of at least two sub-blocks (that is, target image sub-block), and wherein, the size of this sub-block can determine according to preset value, for convenience of explanation, below, the sub-block size of take describes as example as 4 * 4.For example, if the size of target image piece is 16 * 16, can determine that this target image piece comprises 16 sub-blocks (size is 4 * 4).Thereby, in embodiments of the present invention, can determine each sub-block (first object image subblock) in this target image piece corresponding sub-block (the first basic tomographic image sub-block) in basic layer, and the movable information of definite this correspondence sub-block.
In embodiments of the present invention, can according to the coordinate of a certain pixel in sub-block (first object image subblock), (be denoted as: " (E x, E y) "), determine that the coordinate of this pixel correspondence position in basic tomographic image (is denoted as: " (B x, B y) "), and using the image block in the basic layer that comprises affiliated corresponding position coordinates as described corresponding sub-block (the first basic tomographic image sub-block).In embodiments of the present invention, can calculate (B according to following formula 1 to formula 10 x, B y):
Thereby, can determine the first basic tomographic image sub-block corresponding with this first object image subblock, and, in the situation that this first basic tomographic image sub-block comprises movable information, can directly use prediction direction in this movable information and reference picture index as prediction direction and the reference picture index of this sub-block (first object image subblock).Can according to following formula 7 to formula 10, the motion vector (BMV to this first basic tomographic image sub-block x, BMV y) carry out convergent-divergent, and using the motion vector after convergent-divergent as this sub-block the motion vector (EMV of (first object image subblock) x, EMV y).
Here, it should be noted that, if basic tomographic image is identical with the resolution of enhancement layer image, without carrying out above-mentioned zoom operations, and can directly use the movable information of corresponding sub-block as the movable information of sub-block.
By above method, can determine that the corresponding sub-block that can comprise from basic tomographic image in target image piece obtains the sub-block of movable information.
Thereby, for the sub-block of the corresponding sub-block acquisition movable information that can comprise from basic tomographic image, can be by the movable information of its corresponding sub-block
Alternatively, in embodiments of the present invention, the method also comprises:
According to the coding mode of described basic tomographic image, determine whether the first basic tomographic image sub-block corresponding with first object image subblock comprises movable information.
Specifically, in embodiments of the present invention, can determine whether this correspondence sub-block includes movable information according to the coding mode of described basic tomographic image (correspondence image piece).For example, if basic tomographic image is used intraframe predictive coding pattern, can determine that this correspondence sub-block does not comprise movable information (that is, the movable information of the first basic tomographic image sub-block is empty).
Thereby, when this correspondence sub-block comprises movable information, can determine this correspondence sub-block and obtain its movable information by as above this process, when this correspondence sub-block does not comprise movable information, can skip above flow process.
For the corresponding sub-block that can not comprise from basic tomographic image, obtain the sub-block (that is, first object image subblock) of movable information, S220, can be by the following method 2, determine the second target image sub-block, and S230, movable information according to this second target image sub-block, obtains the first reference information.
Method 2
Can be that this target image sub-block fills movable information, thereby can using the movable information of this filling as with reference to information.
Below, describe in detail in the embodiment of the present invention, fill the method for movable information.
Specifically, without loss of generality, for example, if the size of target image piece is 16 * 16, the size of sub-block is 4 * 4, and in embodiments of the present invention, the index assignment method of sub-block can be same as the prior art, here, the description thereof will be omitted, shown in Fig. 2 division and the index of sub-block.
In embodiments of the present invention, can determine and process level according to the size of the size of target image piece and sub-block, and process by processing level Layer by layer recurrence.For example, in embodiments of the present invention, the level that can stipulate the bottom (is denoted as, ground floor) each processing unit in (being denoted as the first processing unit) comprises four sub-blocks, and each processing unit of the last layer of ground floor time (being denoted as the second layer) (is denoted as, the second processing unit) comprise four the first processing units, by that analogy, for fear of repeating, omit recursion explanation.Thereby, as example and non-limiting, in the target image piece shown in Fig. 2, can comprise two levels, in ground floor, sub-block 0 ~ sub-block 3 has formed the first processing unit 0, and sub-block 4 ~ sub-block 7 has formed the first processing unit 1, sub-block 8 ~ sub-block 11 has formed the first processing unit 2, and sub-block 12 ~ sub-block 15 has formed the first processing unit 3.In the second layer, the first processing unit 0 ~ the first processing unit 3 has formed the second processing unit 0.Should be understood that the above hierarchical division method of enumerating is only exemplary illustration, the present invention is not limited thereto.
In embodiments of the present invention, for each the first processing unit, can be according to the call number of sub-block (for example, whether the movable information of from small to large) judging successively each sub-block is as empty, if the movable information of this sub-block is empty, the movable information of sub-block (example of the second target image sub-block) that can be based on being adjacent in this first processing unit is determined its movable information.For example, if the movable information of the sub-block that index is 0 (that is, sub-block 0, belongs to the first processing unit 0) is for empty, can obtain the movable information that belongs to other sub-blocks in same processing unit (the first processing unit 0), and the movable information using this movable information as this sub-block 0.Obtaining order can be, for example, first obtain index and be 1 sub-block (sub-block 1, , one example of the second target image sub-block, adjacent with sub-block 0 in the horizontal direction) movable information, if the movable information of sub-block 1 is empty, can obtain index is 2(sub-block 2 again, , another example of the second target image sub-block, in the vertical direction is adjacent with sub-block 0) the movable information of sub-block, if the movable information of sub-block 2 is empty, can obtain index is 3(sub-block 3 again, , an example again of the second target image sub-block, adjacent with sub-block 0 in diagonal) the movable information of sub-block.In like manner, for each movable information, be empty sub-block, all can to its movable information, fill by above method.Should understand, the movable information fill method that is empty sub-block to movable information of more than enumerating is only exemplary illustration, the present invention is not limited thereto, for example, and for the above-mentioned order of obtaining, also can first obtain regulation sub-block on vertical direction (here, for adjacent sub-blocks) movable information, then obtain the movable information of the regulation sub-block (being adjacent sub-blocks here) in horizontal direction, obtain again the movable information of the regulation sub-block (being adjacent sub-blocks here) in diagonal.That is, this obtains sequentially and can change arbitrarily.
Thereby, through the above-mentioned processing of carrying out in ground floor, each sub-block in each first processing unit is carried out after above-mentioned processing, as long as have the movable information of at least one sub-block in four sub-blocks in this first processing unit not for empty, can be just that empty sub-block is filled (obtaining in other words) movable information for all movable informations in this first processing unit.
It should be noted that, for filled the sub-block of movable information according to said method, while needing to use the movable information of this sub-block, can directly use the movable information of filling into this sub-block in subsequent treatment.; the movable information of the second target image sub-block can refer to the movable information of the corresponding sub-block of basic layer of this second target image sub-block, can be also (from other enhancement layer sub-blocks) movable information of filling for this second target image sub-block according to the method for the filling movable information of the embodiment of the present invention.
Therefore, for example, when the corresponding sub-block of definite some sub-blocks (, sub-block 0) does not comprise movable information, can be from ground floor with this sub-block 0 in same the first processing unit (for example, the first processing unit 0) other sub-blocks (for example, sub-block 1 ~ sub-block 3) are obtained movable information.When with this sub-block (for example, sub-block 0) in same the first processing unit (for example, the first processing unit 0) other sub-blocks (regulation sub-block in ground floor, for example, sub-block 1 ~ sub-block 3) movable information is empty, (for example can obtain regulation the first processing unit in the second processing unit, the first processing unit 1 ~ the first processing unit 3) movable information of regulation sub-block (another example of the second target image sub-block) in, and using this movable information as this sub-block the movable information of (for example, sub-block 0).
; for example, if first processing unit (; the corresponding sub-block of all sub-blocks the first processing unit 0) is sky; (for example can obtain other first processing units in the second processing unit; the first processing unit 1 ~ the first processing unit 3) the regulation sub-block in (for convenience of explanation; in the present invention, take the sub-block in each the first processing unit upper left corner describes as example) movable information, and using this movable information as this first processing unit the movable information of each sub-block in (the first processing unit 0).Obtaining order can be, for example, first obtain index and be the first processing unit (the first processing unit 1 of 1, it is adjacent with the first processing unit 0 in the horizontal direction) the sub-block (sub-block 4 in the upper left corner, , one example of the second target image sub-block) movable information, if the movable information of sub-block 4 is empty, can think that the movable information of these first processing unit, 1 interior other sub-blocks is also sky, thereby can obtain again index and be the first processing unit (the first processing unit 2 of 2, its in the vertical direction is adjacent with the first processing unit 0) the sub-block (sub-block 8 in the upper left corner, , another example of the second target image sub-block) movable information, if the movable information of sub-block 8 is empty, can think that the movable information of these first processing unit, 2 interior other sub-blocks is also sky, thereby can obtain again index and be the first processing unit (the first processing unit 3 of 3, it is adjacent with the first processing unit 0 in diagonal) the sub-block (sub-block 12 in the upper left corner, , an example again of the second target image sub-block) movable information.In like manner, for each movable information, be the first empty processing unit, all can to its movable information, fill by above method.Should understand, the movable information fill method that is empty sub-block to movable information of more than enumerating is only exemplary illustration, the present invention is not limited thereto, for example, for the above-mentioned order of obtaining, also can first obtain regulation the first processing unit on vertical direction (here, for adjacent the first processing unit) the movable information of regulation sub-block, obtain again regulation the first processing unit in horizontal direction (here, for adjacent the first processing unit) the movable information of regulation sub-block, obtain again regulation the first processing unit in diagonal (here, for adjacent the first processing unit) the movable information of regulation sub-block.That is, this obtains sequentially and can change arbitrarily.And, as above-mentioned " regulation sub-block ", be not limited to the sub-block in the first processing unit upper left corner, can be also the sub-block of optional position in same the first processing unit.
It should be noted that, because the size of the above-mentioned target image piece of enumerating is 16 * 16, the size of sub-block is 4 * 4, so this target image piece only comprises two-layerly, and above-mentioned recursive procedure finishes.But the size of target image piece is for larger, for example, 32 * 32, and the size of sub-block is 4 * 4, this target image piece comprises three layers, can proceed recursive operation according to method same as described above, for all sub-blocks when target image piece are obtained movable information.
Alternatively, the size of the target image sub-block that this comprises according to the size of target image piece, this target image piece and be used to indicate the second indication information of the position of first object image subblock in this target image piece, determines the second target image sub-block, comprising:
According to following arbitrary formula, determine this second target image sub-block,
Idx 2=Idx 1/N×N+((Idx 1%N/(N/2))×2+(1-Idx 1%N/(N/4)%2))×N/4;
Idx 2=Idx 1/N×N+((1-Idx 1%N/(N/2))×2+(Idx 1%N/(N/4)%2))×N/4;
Idx 2=Idx 1/N×N+((1-Idx 1%N/(N/2))×2+(1-Idx 1%N/(N/4)%2))×N/4;
Wherein, Idx 2represent to be used to indicate the 3rd indication information of this position of the second target image sub-block in this target image piece, Idx 1represent this second indication information, N determines according to the size of the size of this target image piece and this target image sub-block.
Wherein, Idx 2represent to be used to indicate the 3rd index information of this position of the second target image sub-block in this target image piece, Idx 1represent this second index information, % represents modular arithmetic or remainder operation, and N represents the quantity of the sub-block that this target image piece comprises.
Specifically, can be according to the index of the current sub-block of processing according to above-mentioned each formula, determine the second target image sub-block in the level of pre-treatment, wherein N is corresponding with the current level of processing, and, N determines according to the size of the size of target image piece and sub-block, for example, if the size of target image piece is 16 * 16, the size of sub-block is 4 * 4, as mentioned above this target image piece comprise two-layer, when processing ground floor, N, for the sub-block quantity that each processing unit (the first processing unit) in this layer comprises, is 4 here.When processing the second layer, N, for the sub-block quantity that in this layer, each processing unit (the second processing unit) comprises, is 16 here.
Above, during upper left corner sub-block that to have enumerated when above-mentioned " regulation sub-block " be described processing unit, the formula using.But the present invention does not limit therewith, can also, according to the position in " regulation sub-block " described processing unit, above-mentioned formula be changed.
Alternatively, this determines the first reference information of this first object image subblock according to the movable information of this second target image sub-block, comprising:
If the movable information of this second target image sub-block is empty, determine that this first reference information is zero movable information.
Specifically, if after above-mentioned processing, can not fill movable information for this sub-block, use zero movable information as the movable information of this sub-block.In embodiments of the present invention, can construct in accordance with the following methods zero movable information.For example, in predictive-coded picture frame (P frame), the prediction direction of zero movable information is single directional prediction, and reference picture index is 0, motion vector (0,0).In bidirectionally predictive coded picture frame (B frame), the prediction direction of zero movable information is bi-directional predicted, and two reference picture index are 0, two motion vector and are (0,0).
It should be noted that, when target image piece comprises a plurality of processing levels, above-mentioned use zero movable information, as the method for the movable information of this sub-block, can carry out after last level is processed, also can be to other, arbitrary level carry out after processing, the present invention is also not particularly limited.Should understand, the method of the movable information that obtains sub-block of more than enumerating is only exemplary illustration of the present invention, the present invention is not limited thereto, for example, in the present invention, be in example, can also as mentioned above, according to the coding mode of described basic tomographic image (correspondence image piece), determine whether this correspondence sub-block includes movable information.For example, if basic tomographic image is used intraframe predictive coding pattern, can determine that this correspondence sub-block does not comprise movable information (that is, the movable information of the first basic tomographic image sub-block is empty).If determine in all sub-blocks of target image piece and only have a sub-block (specifically, being its corresponding sub-block) to there is movable information, movable information that can be using the movable information of this sub-block as other sub-blocks.
Thus, by said method, can determine the first reference information of first object image subblock.
At S240, can be according to the first reference information obtaining, this first object sub-block is decoded, for example, can be according to the first reference information of first object image subblock (specifically movable information), to this first object image subblock decode (being specifically, motion compensation process).Specifically, can, according to being the movable information that this first object image subblock is filled as mentioned above, to this first object image subblock, carry out motion compensation process.
In embodiments of the present invention, for the 3rd target image sub-block in target image piece (the corresponding sub-block in basic layer comprises movable information), can be by method same as the prior art, obtain its movable information, and it is decoded, this process can be same as the prior art, here, for fear of repeating, the description thereof will be omitted.
Alternatively, this is decoded to this target image piece according to this reference information, comprising:
Pixel to the boundary vicinity between this target image sub-block is carried out block-eliminating effect filtering processing.
According to the method for processing for image of the embodiment of the present invention, for the corresponding sub-block that can not comprise from basic tomographic image in the target image piece of enhancement layer image, obtain the first object image subblock of movable information, by according to the location positioning second target image sub-block of this first object image subblock, and determine the reference information for this first object image subblock according to the movable information of this second target image sub-block, and encode according to this reference information, can improve the coding efficiency of this first object image subblock.
Above, in conjunction with Fig. 1 to Fig. 3, describe the method for processing for image according to the embodiment of the present invention in detail, below, in connection with Fig. 4 to Fig. 5, describe the device of processing for image according to the embodiment of the present invention in detail.
Fig. 4 shows according to the schematic block diagram of the device 300 of processing for image of the embodiment of the present invention.As shown in Figure 4, this device 300 comprises:
Acquiring unit 310, for when determining that the movable information of the first basic tomographic image sub-block corresponding with the first object image subblock of target image piece is empty, the size of each the target image sub-block comprising according to the size of this target image piece, this target image piece and be used to indicate the second indication information of the position of first object image subblock in this target image piece, determines the second target image sub-block;
Be used for according to the movable information of this second target image sub-block, be identified for the first reference information that this first object image subblock is encoded, wherein, this first basic tomographic image sub-block is the image block that is arranged in basic tomographic image, this target image piece is arranged in enhancement layer image, the basic tomographic image being somebody's turn to do is corresponding with this enhancement layer image, and this first primary image piece sub-block locus in this basic tomographic image is corresponding with this first object image subblock locus in this enhancement layer image;
Coding unit 320, encodes to this target image piece, to generate target code stream and to be contained in the first indication information in this target code stream.
Alternatively, this acquiring unit 310, specifically for according to following arbitrary formula, is determined this second target image sub-block,
Idx 2=Idx 1/N×N+((Idx 1%N/(N/2))×2+(1-Idx 1%N/(N/4)%2))×N/4;
Idx 2=Idx 1/N×N+((1-Idx 1%N/(N/2))×2+(Idx 1%N/(N/4)%2))×N/4;
Idx 2=Idx 1/N×N+((1-Idx 1%N/(N/2))×2+(1-Idx 1%N/(N/4)%2))×N/4;
Wherein, Idx 2represent to be used to indicate the 3rd indication information of this position of the second target image sub-block in this target image piece, Idx 1represent this second indication information, N determines according to the size of the size of this target image piece and this target image sub-block.
Alternatively, if this acquiring unit 310 is empty specifically for the movable information of this second target image sub-block, determine that this first reference information is zero movable information.
Alternatively, this coding unit 320, specifically for according to this first reference information, carries out motion compensation process to this first object image subblock.
Alternatively, this coding unit 320 also carries out block-eliminating effect filtering processing for the pixel of the boundary vicinity between this target image sub-block.
Alternatively, this coding unit 320 is specifically for this first indication information is carried out to entropy coding, so that this first indication information is adjacent with skip mode flag bit or fusion MERGE mode flags position information in this target code stream.
Alternatively, this coding unit 320, specifically for whether using reference information to encode according to the reference image block that is positioned at predeterminated position in this enhancement layer image, is determined context;
For according to this context, this first indication information is carried out to entropy coding.
Can be corresponding to the coding side of the method for the embodiment of the present invention according to the device 300 of processing for image of the embodiment of the present invention, and, each unit in the device 300 that this image is processed is that module and above-mentioned other operations and/or function are respectively in order to realize the corresponding flow process of the method 100 in Fig. 1, for simplicity, do not repeat them here.
According to the device of processing for image of the embodiment of the present invention, for the corresponding sub-block that can not comprise from basic tomographic image in the target image piece of enhancement layer image, obtain the first object image subblock of movable information, by according to the location positioning second target image sub-block of this first object image subblock, and according to the movable information of this second target image sub-block, determine the reference information for this first object image subblock, and encode according to this reference information, can improve the coding efficiency of this first object image subblock.
Fig. 5 shows according to the schematic block diagram of the device 400 of processing for image of the embodiment of the present invention.As shown in Figure 5, this device 400 comprises:
Decoding unit 410, for from target code stream, obtains the first indication information;
Acquiring unit 420, while being empty for the movable information when the first corresponding basic tomographic image sub-block of the first object image subblock with target image piece, this first indication information obtaining based on this decoding unit, the size of each the target image sub-block comprising according to the size of this target image piece, this target image piece and be used to indicate the second indication information of the position of first object image subblock in this target image piece, determines the second target image sub-block;
Be used for according to the movable information of this second target image sub-block, be identified for the first reference information that this first object image subblock is decoded, wherein, this first basic tomographic image sub-block is the image block that is arranged in basic tomographic image, this target image piece is arranged in enhancement layer image, the basic tomographic image being somebody's turn to do is corresponding with this enhancement layer image, and this first primary image piece sub-block locus in this basic tomographic image is corresponding with this first object image subblock locus in this enhancement layer image;
This decoding unit 410 is also for this object code stream is decoded, to obtain this target image piece.
Alternatively, this acquiring unit 420, specifically for according to following arbitrary formula, is determined this second target image sub-block,
Idx 2=Idx 1/N×N+((Idx 1%N/(N/2))×2+(1-Idx 1%N/(N/4)%2))×N/4;
Idx 2=Idx 1/N×N+((1-Idx 1%N/(N/2))×2+(Idx 1%N/(N/4)%2))×N/4;
Idx 2=Idx 1/N×N+((1-Idx 1%N/(N/2))×2+(1-Idx 1%N/(N/4)%2))×N/4;
Wherein, Idx 2represent to be used to indicate the 3rd indication information of this position of the second target image sub-block in this target image piece, Idx 1represent this second indication information, N determines according to the size of the size of this target image piece and this target image sub-block.
Alternatively, if this acquiring unit 420 is empty specifically for the movable information of this second target image sub-block, determine that this first reference information is zero movable information.
Alternatively, this decoding unit 410, specifically for according to this first reference information, carries out motion compensation process to this first object image subblock.
Alternatively, this decoding unit 410 also carries out block-eliminating effect filtering processing for the pixel of the boundary vicinity between this target image sub-block.
Alternatively, this decoding unit 410, specifically for from target code stream, obtains the first indication information, and wherein, this first indication information is adjacent with skip mode flag bit or fusion MERGE mode flags position information in this target code stream.
Alternatively, this decoding unit 410, specifically for whether using reference information to decode according to the reference image block that is positioned at predeterminated position in this enhancement layer image, is determined context;
For according to this context, carry out entropy decoding, to determine this first indication information.
Can be corresponding to the method decoding end of the embodiment of the present invention according to the device 400 of processing for image of the embodiment of the present invention, and, each unit of the device 400 that should process for image is that module and above-mentioned other operations and/or function are respectively in order to realize the corresponding flow process of the method 200 in Fig. 3, for simplicity, do not repeat them here.
According to the device of processing for image of the embodiment of the present invention, for the corresponding sub-block that can not comprise from basic tomographic image in the target image piece of enhancement layer image, obtain the first object image subblock of movable information, by according to the location positioning second target image sub-block of this first object image subblock, and according to the movable information of this second target image sub-block, determine the reference information for this first object image subblock, and encode according to this reference information, can improve the coding efficiency of this first object image subblock.
Above, in conjunction with Fig. 1 to Fig. 5, describe the method and apparatus of processing for image according to the embodiment of the present invention in detail, below in conjunction with Fig. 6 and Fig. 7, describe the encoder of processing for image according to the embodiment of the present invention in detail.
Fig. 6 shows according to the schematic block diagram of the encoder 500 of processing for image of the embodiment of the present invention.As shown in Figure 6, this encoder 500 can comprise:
Bus 510;
The processor 520 being connected with this bus;
The memory 530 being connected with this bus;
Wherein, this processor 520 is by this bus 510, call the program of storage in this memory 530, for when determining that the movable information of the first basic tomographic image sub-block corresponding with the first object image subblock of target image piece is empty, the size of each the target image sub-block comprising according to the size of this target image piece, this target image piece and be used to indicate the second indication information of the position of first object image subblock in this target image piece, determines the second target image sub-block;
Be used for according to the movable information of this second target image sub-block, be identified for the first reference information that this first object image subblock is encoded, wherein, this first basic tomographic image sub-block is the image block that is arranged in basic tomographic image, this target image piece is arranged in enhancement layer image, the basic tomographic image being somebody's turn to do is corresponding with this enhancement layer image, and this first primary image piece sub-block locus in this basic tomographic image is corresponding with this first object image subblock locus in this enhancement layer image;
For this target image piece is encoded, to generate target code stream and to be contained in the first indication information in this target code stream.
Alternatively, this processor 520, specifically for according to following arbitrary formula, is determined this second target image sub-block,
Idx 2=Idx 1/N×N+((Idx 1%N/(N/2))×2+(1-Idx 1%N/(N/4)%2))×N/4;
Idx 2=Idx 1/N×N+((1-Idx 1%N/(N/2))×2+(Idx 1%N/(N/4)%2))×N/4;
Idx 2=Idx 1/N×N+((1-Idx 1%N/(N/2))×2+(1-Idx 1%N/(N/4)%2))×N/4;
Wherein, Idx 2represent to be used to indicate the 3rd indication information of this position of the second target image sub-block in this target image piece, Idx 1represent this second indication information, N determines according to the size of the size of this target image piece and this target image sub-block.
Alternatively, if this processor 520 is empty specifically for the movable information of this second target image sub-block, determine that this first reference information is zero movable information.
Alternatively, this processor 520, specifically for according to this first reference information, carries out motion compensation process to this first object image subblock.
Alternatively, this processor 520 carries out block-eliminating effect filtering processing specifically for the pixel of the boundary vicinity between this target image sub-block.
Alternatively, this processor 520 is specifically for this first indication information is carried out to entropy coding, so that this first indication information is adjacent with skip mode flag bit or fusion MERGE mode flags position information in this target code stream.
Alternatively, this processor 520, specifically for whether using reference information to encode according to the reference image block that is positioned at predeterminated position in this enhancement layer image, is determined context;
For according to this context, this first indication information is carried out to entropy coding.
Can be corresponding to the method coding side of the embodiment of the present invention according to the encoder 500 of processing for image of the embodiment of the present invention, and, each unit of the encoder 500 that should process for image is that module and above-mentioned other operations and/or function are respectively in order to realize the corresponding flow process of the method 100 in Fig. 1, for simplicity, do not repeat them here.
According to the encoder of processing for image of the embodiment of the present invention, for the corresponding sub-block that can not comprise from basic tomographic image in the target image piece of enhancement layer image, obtain the first object image subblock of movable information, by according to the location positioning second target image sub-block of this first object image subblock, and according to the movable information of this second target image sub-block, determine the reference information for this first object image subblock, and encode according to this reference information, can improve the coding efficiency of this first object image subblock.
Fig. 7 shows according to the schematic block diagram of the decoder 600 of processing for image of the embodiment of the present invention.As shown in Figure 7, this decoder 600 can comprise:
Bus 610;
The processor 620 being connected with this bus;
The memory 630 being connected with this bus;
Wherein, this processor 620, by this bus 610, calls the program of storage in this memory 630, for from target code stream, obtains the first indication information;
While being empty for the movable information when the first corresponding basic tomographic image sub-block of the first object image subblock with target image piece, based on this first indication information, the size of each the target image sub-block comprising according to the size of this target image piece, this target image piece and be used to indicate the second indication information of the position of first object image subblock in this target image piece, determines the second target image sub-block;
Be used for according to the movable information of this second target image sub-block, be identified for the first reference information that this first object image subblock is decoded, wherein, this first basic tomographic image sub-block is the image block that is arranged in basic tomographic image, this target image piece is arranged in enhancement layer image, the basic tomographic image being somebody's turn to do is corresponding with this enhancement layer image, and this first primary image piece sub-block locus in this basic tomographic image is corresponding with this first object image subblock locus in this enhancement layer image;
For this object code stream is decoded, to obtain this target image piece.
Alternatively, this processor 620, specifically for according to following arbitrary formula, is determined this second target image sub-block,
Idx 2=Idx 1/N×N+((Idx 1%N/(N/2))×2+(1-Idx 1%N/(N/4)%2))×N/4;
Idx 2=Idx1/N×N+((1-Idx 1%N/(N/2))×2+(Idx 1%N/(N/4)%2))×N/4;
Idx 2=Idx 1/N×N+((1-Idx 1%N/(N/2))×2+(1-Idx 1%N/(N/4)%2))×N/4;
Wherein, Idx 2represent to be used to indicate the 3rd indication information of this position of the second target image sub-block in this target image piece, Idx 1represent this second indication information, N determines according to the size of the size of this target image piece and this target image sub-block.
Alternatively, if this processor 620 is empty specifically for the movable information of this second target image sub-block, determine that this first reference information is zero movable information.
Alternatively, this processor 620, specifically for according to this first reference information, carries out motion compensation process to this first object image subblock.
Alternatively, this processor 620 carries out block-eliminating effect filtering processing specifically for the pixel of the boundary vicinity between this target image sub-block.
Alternatively, this processor 620, specifically for from target code stream, obtains the first indication information, and wherein, this first indication information is adjacent with skip mode flag bit or fusion MERGE mode flags position information in this target code stream.
Alternatively, this processor 620, specifically for whether using reference information to decode according to the reference image block that is positioned at predeterminated position in this enhancement layer image, is determined context;
For according to this context, carry out entropy decoding, to determine this first indication information.
Can be corresponding to the method decoding end of the embodiment of the present invention according to the decoder 600 of processing for image of the embodiment of the present invention, and, each unit of the decoder 600 that should process for image is that module and above-mentioned other operations and/or function are respectively in order to realize the corresponding flow process of the method 200 in Fig. 5, for simplicity, do not repeat them here.
According to the decoder of processing for image of the embodiment of the present invention, for the corresponding sub-block that can not comprise from basic tomographic image in the target image piece of enhancement layer image, obtain the first object image subblock of movable information, by according to the location positioning second target image sub-block of this first object image subblock, and according to the movable information of this second target image sub-block, determine the reference information for this first object image subblock, and encode according to this reference information, can improve the coding efficiency of this first object image subblock.
Fig. 8 show from coding side angle, describe according to the indicative flowchart of the method 700 of processing for image of the embodiment of the present invention.As shown in Figure 8, the method 700 comprises:
S710, when determining that the movable information of the first basic tomographic image sub-block corresponding with the first object image subblock of target image piece is empty, according to the reconstruction pixel of this first basic tomographic image sub-block, be identified for the second reference information that this first object image subblock is encoded, wherein, this first basic tomographic image sub-block is the image block that is arranged in basic tomographic image, this target image piece is arranged in enhancement layer image, the basic tomographic image being somebody's turn to do is corresponding with this enhancement layer image, and this first primary image piece sub-block locus in this basic tomographic image is corresponding with this first object image subblock locus in this enhancement layer image,
S720, encodes to this target image piece, to generate target code stream and to be contained in the 4th indication information in this target code stream.
Specifically, image is being carried out to hierarchical coding, for example, when spatial scalable is encoded, image can be carried out resolution processes to obtain low-resolution image, as a comparison original image is called to high-definition picture, encoder is respectively to the processing of encoding of this low-resolution image and this high-definition picture.For convenience of describing, the high image to be encoded of quality is called to enhancement layer image herein, the image low-quality to be encoded of correspondence (for example described low-resolution image) is called to basic tomographic image.
In embodiments of the present invention, target image is the image that uses hierarchical coding technology to process, basic layer refers to the lower layer of quality (comprising the parameters such as frame rate, spatial resolution, temporal resolution, signal to noise ratio intensity or credit rating) in hierarchical coding, and enhancement layer refers to the higher layer of quality (comprising the parameters such as frame rate, spatial resolution, temporal resolution, signal to noise ratio intensity or credit rating) in hierarchical coding.It should be noted that, in embodiments of the present invention, in embodiments of the present invention, for a given enhancement layer, basic layer corresponding thereto can be quality lower than arbitrary layer of this enhancement layer, for example, if five layers of current existence, (coding quality improves successively, ground floor quality is minimum, and layer 5 quality is the highest), if enhancement layer is the 4th layer, basic layer can be ground floor, can be also the second layer, also can be the 3rd layer, also can be the 4th layer.In like manner, for a given basic layer, enhancement layer corresponding thereto can be the arbitrary layer of quality lower than this basic layer.
Enhancement layer image is the image in the enhancement layer of pre-treatment, basic tomographic image be in basic layer with the image of enhancement layer image at synchronization.
In sum, in embodiments of the present invention, the quality of this basic tomographic image is lower than the quality of this enhancement layer image.
The image block of target image piece for processing in this enhancement layer image.
Basic tomographic image piece is in basic tomographic image, to have on locus the image block of corresponding relation with this target image piece.
In embodiments of the present invention, the image block in basic layer and the corresponding relation of the image block in enhancement layer can calculate according to the resolution proportionate relationship between basic tomographic image and enhancement layer image.For example, within comprising the system of x direction and y direction, if enhancement layer image is respectively 2 times of basic tomographic image in the resolution of x direction and y direction, the pixel coordinate for the upper left corner in enhancement layer is (2x, 2y) and size be the image block of (2m) * (2n), corresponding blocks in its basic tomographic image can be that the pixel coordinate in the upper left corner is that (x, y) and size are the image block of m * n.
In embodiments of the present invention, aftermentioned sub-block refers to the sub-block (image block in enhancement layer) of target image piece, and the corresponding sub-block of aftermentioned refers to the corresponding image block in basic layer of this sub-block.
In embodiments of the present invention, movable information can comprise prediction direction, one or more in reference picture index or motion vector, wherein, prediction direction can be divided into unidirectional and bi-directional predicted, single directional prediction can be divided into forward prediction and back forecast again, forward prediction refers to use forward direction reference picture list, it is the reference picture generation prediction signal in list (list) 0, back forecast refers to use backward reference picture list, it is the reference picture generation prediction signal in list1, bi-directional predicted finger is used the reference picture in list 0 and list 1 to produce prediction signal simultaneously, for single directional prediction, need a reference picture index indication selected reference picture in list0 or list1, for bi-directional predicted, need two reference picture index, indicate respectively selected reference picture in list 0 and list1, each motion vector comprises horizontal direction component x and vertical direction component y, can be denoted as (x, y), for single directional prediction, need the displacement of a motion vector indication predicting signal in selected list0 or list1 reference picture, for bi-directional predicted, need two motion vectors, indicate respectively the displacement in selected list0 reference picture and list1 reference picture of forward prediction signal and back forecast signal.
In embodiments of the present invention, target image piece can be regarded as and consists of at least two sub-blocks (that is, target image sub-block), and wherein, the size of this sub-block can determine according to preset value, for convenience of explanation, below, the sub-block size of take describes as example as 4 * 4.For example, if the size of target image piece is 16 * 16, can determine that this target image piece comprises 16 sub-blocks (size is 4 * 4).Thereby, in embodiments of the present invention, can determine each sub-block in this target image piece corresponding sub-block (belonging to this correspondence image piece) in basic layer, and determine the movable information of this correspondence sub-block.
In embodiments of the present invention, can according to the coordinate of a certain pixel in sub-block, (be denoted as: " (E x, E y) "), determine that the coordinate of this pixel correspondence position in basic tomographic image (is denoted as: " (B x, B y) "), and using the image block in the basic layer that comprises affiliated corresponding position coordinates as described corresponding sub-block.In embodiments of the present invention, can according to above formula 1 to formula 10, the motion vector (BMV to this first basic tomographic image sub-block x, BMV y) carry out convergent-divergent, and using the motion vector after convergent-divergent as this sub-block the motion vector (EMV of (first object image subblock) x, EMV y).
Here, it should be noted that, if basic tomographic image is identical with the resolution of enhancement layer image, without carrying out above-mentioned zoom operations, and can directly use the movable information of corresponding sub-block as the movable information of sub-block.
By above method, can determine that the corresponding sub-block that can comprise from basic tomographic image in target image piece obtains the sub-block of movable information.
Thereby, for the sub-block of the corresponding sub-block acquisition movable information that can comprise from basic tomographic image, can be by the movable information of its corresponding sub-block
Alternatively, in embodiments of the present invention, the method also comprises:
According to the coding mode of described basic tomographic image, determine whether the first basic tomographic image sub-block corresponding with first object image subblock comprises movable information.
Specifically, in embodiments of the present invention, can determine whether this correspondence sub-block includes movable information according to the coding mode of described basic tomographic image (correspondence image piece).For example, if basic tomographic image is used intraframe predictive coding pattern, can determine that this correspondence sub-block does not comprise movable information (that is, the movable information of the first basic tomographic image sub-block is empty).
Thereby, when this correspondence sub-block comprises movable information, can determine this correspondence sub-block and obtain its movable information by process as mentioned above, when this correspondence sub-block does not comprise movable information, can skip above flow process.
For the corresponding sub-block that can not comprise from basic tomographic image, obtain the sub-block (that is, first object image subblock) of movable information, can be by the following method 3, obtain its second reference information.
Method 3
Specifically, can obtain the reconstruction pixel of sub-block corresponding with this first object image subblock in basic tomographic image, and this reconstruction pixel is carried out to up-sampling processing, the prediction signal that generates this first object image subblock is as with reference to information.
Alternatively, this is according to this reference information, to the processing of encoding of this target image piece, comprising:
According to this reference information, this first object image subblock is carried out to motion compensation process.
Specifically, can be according to the reference information of first object image subblock (specifically, being this prediction signal), to the processing (being specifically, that predictive coding is processed) of encoding of this first object image subblock.Specifically, rebuild pixel and carry out after suitable up-sampling the prediction signal as current sub-block.After getting the prediction signal of current block, just can carry out predictive coding to current block, thereby calculate rate distortion costs.
Alternatively, this is according to this reference information, to the processing of encoding of this target image piece, comprising:
Pixel to the boundary vicinity between this target image sub-block is carried out block-eliminating effect filtering processing.
In S230, by said method 3, obtaining first object image block second reference information of (in basic layer, corresponding sub-block does not comprise movable information), and obtain the 3rd target image piece movable information of (in basic layer, corresponding sub-block comprises movable information), the method of obtaining the movable information of the 3rd target image piece can be same as the prior art, here the description thereof will be omitted, thereby, can calculate after the distortion rate cost of target image piece, if this distortion rate Least-cost, can determine a sign (the 4th indication information), with indication decoding end, by said method 1 and method 2, obtain the reference information of first object image block, and obtain the 3rd target image piece movable information of (in basic layer, corresponding sub-block comprises movable information).And the 4th indication information is carried out to entropy coding and process.
Alternatively, this is according to this reference information, to the processing of encoding of this first object image subblock, comprising:
This first indication information is carried out to entropy coding and process, so that this first indication information is adjacent with skip mode flag bit or fusion MERGE mode flags position information in this target code stream.
Specifically, in embodiments of the present invention, in target code stream, the first indication information can be configured in to the position adjacent with skip mode flag bit information.Specifically,
For example, the first indication information can be placed in to skip mode flag bit information first information at target code stream as target image piece before, the first indication information can also be placed in to skip mode flag bit information second information at target code stream as target image piece afterwards.Above-mentioned skip mode can be pattern same as the prior art, and it determines that method and allocation position can be same as the prior art, and here, for fear of repeating, the description thereof will be omitted.
Again for example, the first indication information can be placed in to the position adjacent with MERGE mode flags position information.Concrete, before the first indication information can being placed in to MERGE mode flags position information, after the first indication information can also being placed in to MERGE mode flags position information.Described MERGE can be pattern same as the prior art, and it determines that method and allocation position can be same as the prior art, and here, for fear of repeating, the description thereof will be omitted.
In embodiments of the present invention, this first indication information can be a binary flags position.Therefore, when the first indication information being carried out to entropy coding, without the first indication information is carried out to binary conversion treatment.
Thereafter, can select the context that uses when entropy coding is processed for the first indication information of binaryzation is carried out, alternatively, this is according to this reference information, to the processing of encoding of this first object image subblock, comprising:
According to the reference image block that is positioned at predeterminated position in this enhancement layer image, whether use the reference information processing of encoding, determine context;
According to this context, this first indication information is carried out to entropy coding and process.
Specifically.As shown in Table 1 above, this context can comprise 0,1,2 three context.Whether the present embodiment is determined and is used which context with basic layer model according to the image block of the left side and top.In the present embodiment, for example, can whether use its first indication information separately according to the image block of the target image piece left side and top, determine and use any context, for example, if the image block of the target image piece left side and top is not used the first indication information, selecting so index is 0 context model, if the image block of the target image piece left side and top has one to use the first indication information, selecting so index is 1 context model, if the image block of the target image piece left side and top is all used the first indication information, selecting so index is 2 context.
, can according to the context as mentioned above selected to this first indication information carry out binary arithmetic coding, and upgrade the context model of using thereafter.In embodiments of the present invention, this process can be same as the prior art, and for fear of repeating, the description thereof will be omitted here.
In embodiments of the present invention, can be according to obtained reference information, to the processing of encoding of each first object image subblock, and then complete the coding processing to target image piece, and the first indication information after the information of the target image piece of generation and the processing of above-mentioned entropy coding is added to code stream (target code stream).
Here, it should be noted that, in this target code stream, can comprise target image (comprising basic tomographic image and the enhancement layer image) information after coding is processed, and this processing procedure can be same as the prior art, here, for fear of repeating, the description thereof will be omitted.
In embodiments of the present invention, can stipulate, for example, when the 4th indication information is 1, decoding end need to be used the method identical with coding side to obtain the second reference information of first object image subblock, and decoding end need to be used the method identical with coding side to obtain the second reference information of first object image subblock.
The indicating means that should be understood that above the 4th indication information of enumerating is only exemplary illustration, and the present invention is not limited thereto.
Therefore, decoding end can determine whether to obtain the second reference information of first object image subblock according to the 4th indication information, below, the situation that need to obtain the second reference information of first object image subblock to coding side describes.
In decoding end, can from code stream, obtain target image information, and definite target image (specifically, target image piece), the 4th indication information obtaining (information after entropy coding is processed) is carried out to entropy decoding to be processed, in this processing procedure, context is selected with the process of updating context and the processing of above-mentioned coding side same or similar, and the description thereof will be omitted here.
Can from code stream, resolve the binary character word string (bin string) that obtains representing the 4th indication information according to the context of selecting.Wherein, this binary arithmetic decoding is processed and is processed corresponding with the binary arithmetic coding of coding side.
It should be noted that, in embodiments of the present invention, when using method 3 is obtained the second reference information of first object image subblock, can obtain the reconstruction pixel of sub-block corresponding with this first object image subblock in basic tomographic image, and this reconstruction pixel is carried out to up-sampling processing, the prediction signal that generates this first object image subblock is as with reference to information.Can be according to the reference information of first object image subblock (specifically, being this prediction signal), to the processing (being specifically, that predictive coding is processed) of encoding of this first object image subblock.Specifically, rebuild after pixel up-sampling the prediction signal as current sub-block.After getting the prediction signal of current block, just can carry out motion compensation process to current block, and the residual signals that decoding obtains in addition that can also superpose, to obtain reconstruction signal.
And, can also carry out filtering processing to the boundary vicinity pixel between each sub-block of target image piece.
According to the method for processing for image of the embodiment of the present invention, for the corresponding sub-block that can not comprise from basic tomographic image in the target image piece of enhancement layer image, obtain the first object image subblock of movable information, by according to the location positioning second target image sub-block of this first object image subblock, and according to the reconstruction pixel of the first basic tomographic image sub-block corresponding with this first object image subblock on locus, determine the reference information for this first object image subblock, and according to the processing of encoding of this reference information, can improve the coding efficiency of this first object image subblock.
Fig. 9 show from decoding end angle, describe according to the indicative flowchart of the method 800 of processing for image of the embodiment of the present invention.As shown in Figure 9, the method 800 comprises:
S210, from target code stream, obtains the 4th indication information;
S220, when determining that the movable information of the first basic tomographic image sub-block corresponding with the first object image subblock of target image piece is empty, based on the 4th indication information, according to the reconstruction pixel of this first basic tomographic image sub-block, be identified for the second reference information that this first object image subblock is encoded, wherein, this first basic tomographic image sub-block is the image block that is arranged in basic tomographic image, this target image piece is arranged in enhancement layer image, the basic tomographic image being somebody's turn to do is corresponding with this enhancement layer image, and this first primary image piece sub-block locus in this basic tomographic image is corresponding with this first object image subblock locus in this enhancement layer image,
S230, decodes to this object code stream, to obtain this target image piece.
Specifically, at S210, decoding end can be obtained target image information from code stream, and definite target image (specifically, being target image piece), and the first indication information obtaining (information after entropy coding is processed).
Alternatively, should from target code stream, obtain the first indication information, comprise:
From target code stream, obtain the first indication information, wherein, this first indication information is adjacent with skip mode flag bit or fusion MERGE mode flags position information in this target code stream.
Specifically, in embodiments of the present invention, in target code stream, the first indication information can be configured in to the position adjacent with skip mode flag bit information.Specifically,
For example, the first indication information can be placed in to skip mode flag bit information first information at target code stream as target image piece before, the first indication information can also be placed in to skip mode flag bit information second information at target code stream as target image piece afterwards.Above-mentioned skip mode can be pattern same as the prior art, and it determines that method and allocation position can be same as the prior art, and here, for fear of repeating, the description thereof will be omitted.
Again for example, the first indication information can be placed in to the position adjacent with MERGE mode flags position information.Concrete, before the first indication information can being placed in to MERGE mode flags position information, after the first indication information can also being placed in to MERGE mode flags position information.This MERGE can be pattern same as the prior art, and it determines that method and allocation position can be same as the prior art, and here, for fear of repeating, the description thereof will be omitted.
Thereafter, decoding end can be carried out entropy decoding processing to the first indication information obtaining.
In embodiments of the present invention, this first indication information can be a binary flags position.Therefore, when the first indication information being carried out to entropy decoding, without making the first indication information binaryzation.
, can select the context that for first indication information of binaryzation carried out when entropy coding process use, alternatively, should from target code stream, obtain the first indication information, comprise thereafter:
According to the reference image block that is positioned at predeterminated position in this enhancement layer image, whether use the reference information processing of decoding, determine context;
According to this context, carry out entropy decoding and process, to determine this first indication information.
Specifically.As shown in Table 1 above, this context can comprise 0,1,2 three context.Whether the present embodiment is determined and is used which context with basic layer model according to the image block of the left side and top.In the present embodiment, for example, can whether use its first indication information separately according to the image block of the target image piece left side and top, determine and use any context, for example, if the image block of the target image piece left side and top is not used the first indication information, selecting so index is 0 context model, if the image block of the target image piece left side and top has one to use the first indication information, selecting so index is 1 context model, if the image block of the target image piece left side and top is all used the first indication information, selecting so index is 2 context.
, can according to the context of as above this selection to this first indication information carry out binary arithmetic decoding, and upgrade the context model of using thereafter.In embodiments of the present invention, this process can be same as the prior art, and for fear of repeating, the description thereof will be omitted here.
In embodiments of the present invention, can stipulate, for example, when the first indication information is 1, decoding end need to be used the method identical with coding side to obtain the first reference information of first object image subblock, and decoding end need to be used the method identical with coding side to obtain the first reference information of first object image subblock.
The indicating means that should be understood that above the first indication information of enumerating is only exemplary illustration, and the present invention is not limited thereto.
Therefore, decoding end can determine whether to obtain the first reference information of first object image subblock according to this first indication information, below, the situation that need to obtain the first reference information of first object image subblock to coding side describes.
In embodiments of the present invention, image is being carried out to hierarchical coding, for example, when spatial scalable is encoded, image can be carried out resolution processes to obtain low-resolution image, as a comparison original image is called to high-definition picture, encoder is respectively to the processing of encoding of this low-resolution image and this high-definition picture.For convenience of describing, the high image to be encoded of quality is called to enhancement layer image herein, the image low-quality to be encoded of correspondence (for example described low-resolution image) is called to basic tomographic image.
In embodiments of the present invention, target image is the image that uses hierarchical coding technology to process, basic layer refers to the lower layer of quality (comprising the parameters such as frame rate, spatial resolution, temporal resolution, signal to noise ratio intensity or credit rating) in hierarchical coding, and enhancement layer refers to the higher layer of quality (comprising the parameters such as frame rate, spatial resolution, temporal resolution, signal to noise ratio intensity or credit rating) in hierarchical coding.It should be noted that, in embodiments of the present invention, in embodiments of the present invention, for a given enhancement layer, basic layer corresponding thereto can be quality lower than arbitrary layer of this enhancement layer, for example, if five layers of current existence, (coding quality improves successively, ground floor quality is minimum, and layer 5 quality is the highest), if enhancement layer is the 4th layer, basic layer can be ground floor, can be also the second layer, also can be the 3rd layer, also can be the 4th layer.In like manner, for a given basic layer, enhancement layer corresponding thereto can be the arbitrary layer of quality lower than this basic layer.
Enhancement layer image is the image in the enhancement layer of pre-treatment, basic tomographic image be in basic layer with the image of enhancement layer image at synchronization.
In sum, in embodiments of the present invention, the quality of this basic tomographic image is lower than the quality of this enhancement layer image.
The image block of target image piece for processing in this enhancement layer image.
Basic tomographic image piece is in basic tomographic image, to have on locus the image block of corresponding relation with this target image piece.
In embodiments of the present invention, the image block in basic layer and the corresponding relation of the image block in enhancement layer can calculate according to the resolution proportionate relationship between basic tomographic image and enhancement layer image.For example, within comprising the system of x direction and y direction, if enhancement layer image is respectively 2 times of basic tomographic image in the resolution of x direction and y direction, the pixel coordinate for the upper left corner in enhancement layer is (2x, 2y) and size be the image block of (2m) * (2n), corresponding blocks in its basic tomographic image can be that the pixel coordinate in the upper left corner is that (x, y) and size are the image block of m * n.
In embodiments of the present invention, aftermentioned sub-block refers to the sub-block (image block in enhancement layer) of target image piece, and the corresponding sub-block of aftermentioned refers to the corresponding image block in basic layer of this sub-block.
In embodiments of the present invention, movable information can comprise prediction direction, one or more in reference picture index or motion vector, wherein, prediction direction can be divided into unidirectional and bi-directional predicted, single directional prediction can be divided into forward prediction and back forecast again, forward prediction refers to use forward direction reference picture list, it is the reference picture generation prediction signal in list (list) 0, back forecast refers to use backward reference picture list, it is the reference picture generation prediction signal in list1, bi-directional predicted finger is used the reference picture in list0 and list1 to produce prediction signal simultaneously, for single directional prediction, need a reference picture index indication selected reference picture in list0 or list1, for bi-directional predicted, need two reference picture index, indicate respectively selected reference picture in list0 and list1, each motion vector comprises horizontal direction component x and vertical direction component y, can be denoted as (x, y), for single directional prediction, need the displacement of a motion vector indication predicting signal in selected list0 or list1 reference picture, for bi-directional predicted, need two motion vectors, indicate respectively the displacement in selected list0 reference picture and list1 reference picture of forward prediction signal and back forecast signal.
In embodiments of the present invention, target image piece can be regarded as and consists of at least two sub-blocks (that is, target image sub-block), and wherein, the size of this sub-block can determine according to preset value, for convenience of explanation, below, the sub-block size of take describes as example as 4 * 4.For example, if the size of target image piece is 16 * 16, can determine that this target image piece comprises 16 sub-blocks (size is 4 * 4).Thereby, in embodiments of the present invention, can determine each sub-block (first object image subblock) in this target image piece corresponding sub-block (the first basic tomographic image sub-block) in basic layer, and the movable information of definite this correspondence sub-block.
In embodiments of the present invention, can according to the coordinate of a certain pixel in sub-block (first object image subblock), (be denoted as: " (E x, E y) "), determine that the coordinate of this pixel correspondence position in basic tomographic image (is denoted as: " (B x, B y) "), and using the image block in the basic layer that comprises affiliated corresponding position coordinates as described corresponding sub-block (the first basic tomographic image sub-block).In embodiments of the present invention, can calculate according to following formula 1 to formula 10 motion vector (the EMV of first object image subblock x, EMV y).
Here, it should be noted that, if basic tomographic image is identical with the resolution of enhancement layer image, without carrying out above-mentioned zoom operations, and can directly use the movable information of corresponding sub-block as the movable information of sub-block.
By above method, can determine that the corresponding sub-block that can comprise from basic tomographic image in target image piece obtains the sub-block of movable information.
Thereby, for the sub-block of the corresponding sub-block acquisition movable information that can comprise from basic tomographic image, can be by the movable information of its corresponding sub-block
Alternatively, in embodiments of the present invention, the method also comprises:
According to the coding mode of described basic tomographic image, determine whether the first basic tomographic image sub-block corresponding with first object image subblock comprises movable information.
Specifically, in embodiments of the present invention, can determine whether this correspondence sub-block includes movable information according to the coding mode of described basic tomographic image (correspondence image piece).For example, if basic tomographic image is used intraframe predictive coding pattern, can determine that this correspondence sub-block does not comprise movable information (that is, the movable information of the first basic tomographic image sub-block is empty).
Thereby, when this correspondence sub-block comprises movable information, can determine this correspondence sub-block and obtain its movable information by process as mentioned above, when this correspondence sub-block does not comprise movable information, can, by with below method 4, obtain the second reference information of this first object image subblock.
Method 4
Specifically, can obtain the reconstruction pixel of sub-block corresponding with this first object image subblock in basic tomographic image, and this reconstruction pixel is carried out to up-sampling processing, the prediction signal that generates this first object image subblock is as with reference to information.
Alternatively, this is according to this reference information, to the processing of decoding of this target image piece, comprising:
According to this reference information, this first object image subblock is carried out to motion compensation process.
At S230, can be according to the reference information obtaining, to the processing of decoding of this first object sub-block, at this reference information, be obtain by method 3 in the situation that, can be according to the reference information of first object image subblock (specifically movable information), to the processing (specifically, being motion compensation process) of decoding of this first object image subblock.Specifically, can, according to being the movable information that this first object image subblock is filled as mentioned above, to this first object image subblock, carry out motion compensation process.
At this reference information, be obtain by method 4 in the situation that, can obtain the reconstruction pixel of sub-block corresponding with this first object image subblock in basic tomographic image, and this reconstruction pixel is carried out to up-sampling processing, the prediction signal that generates this first object image subblock is as with reference to information.Can be according to the reference information of first object image subblock (specifically, being this prediction signal), to the processing (being specifically, that predictive coding is processed) of encoding of this first object image subblock.Specifically, rebuild after pixel up-sampling the prediction signal as current sub-block.After getting the prediction signal of current block, just can carry out motion compensation process to current block, and the residual signals that decoding obtains in addition that can also superpose, to obtain reconstruction signal.
In embodiments of the present invention, for the 3rd target image sub-block in target image piece (the corresponding sub-block in basic layer comprises movable information), can be by method same as the prior art, obtain its movable information, and to its processing of decoding, this process can be same as the prior art, here, for fear of repeating, the description thereof will be omitted.
Alternatively, this is according to this reference information, to the processing of decoding of this target image piece, comprising:
Pixel to the boundary vicinity between this target image sub-block is carried out block-eliminating effect filtering processing.
According to the method for processing for image of the embodiment of the present invention, for the corresponding sub-block that can not comprise from basic tomographic image in the target image piece of enhancement layer image, obtain the first object image subblock of movable information, by according to the location positioning second target image sub-block of this first object image subblock, and according to the reconstruction pixel of the first basic tomographic image sub-block corresponding with this first object image subblock on locus, determine the reference information for this first object image subblock, and according to the processing of encoding of this reference information, can improve the coding efficiency of this first object image subblock.
Above, in conjunction with Fig. 8 to Fig. 9, describe the method for processing for image according to the embodiment of the present invention in detail, below, in connection with Figure 10 to Figure 11, describe the device of processing for image according to the embodiment of the present invention in detail.
Figure 10 shows according to the schematic block diagram of the device 900 of processing for image of the embodiment of the present invention.As shown in figure 10, this device 900 comprises:
Acquiring unit 910, for when determining that the movable information of the first basic tomographic image sub-block corresponding with the first object image subblock of target image piece is empty, according to the reconstruction pixel of this first basic tomographic image sub-block, be identified for the second reference information that this first object image subblock is encoded, wherein, this first basic tomographic image sub-block is the image block that is arranged in basic tomographic image, this target image piece is arranged in enhancement layer image, the basic tomographic image being somebody's turn to do is corresponding with this enhancement layer image, and this first primary image piece sub-block locus in this basic tomographic image is corresponding with this first object image subblock locus in this enhancement layer image,
Coding unit 920, for this target image piece is encoded, to generate target code stream and to be contained in the 4th indication information in this target code stream.
Alternatively, this coding unit 920 carries out block-eliminating effect filtering processing specifically for the pixel of the boundary vicinity between this target image sub-block.
Alternatively, this coding unit 920 is specifically for the 4th indication information is carried out to entropy coding, so that the 4th indication information is adjacent with skip mode flag bit or fusion MERGE mode flags position information in this target code stream.
Alternatively, this coding unit 920, specifically for whether using reference information to encode according to the reference image block that is positioned at predeterminated position in this enhancement layer image, is determined context;
For according to this context, the 4th indication information is carried out to entropy coding.
Can be corresponding to the coding side of the method for the embodiment of the present invention according to the device 900 of processing for image of the embodiment of the present invention, and, each unit in the device 900 that this image is processed is that module and above-mentioned other operations and/or function are respectively in order to realize the corresponding flow process of the method 700 in Fig. 8, for simplicity, do not repeat them here.
According to the device of processing for image of the embodiment of the present invention, for the corresponding sub-block that can not comprise from basic tomographic image in the target image piece of enhancement layer image, obtain the first object image subblock of movable information, by according to the location positioning second target image sub-block of this first object image subblock, and according to the reconstruction pixel of the first basic tomographic image sub-block corresponding with this first object image subblock on locus, determine the reference information for this first object image subblock, and according to the processing of encoding of this reference information, can improve the coding efficiency of this first object image subblock.
Figure 11 shows according to the schematic block diagram of the device 1000 of processing for image of the embodiment of the present invention.As shown in figure 11, this device 1000 comprises:
Decoding unit 1010, for from target code stream, obtains the 4th indication information;
Acquiring unit 1020, for when determining that the movable information of the first basic tomographic image sub-block corresponding with the first object image subblock of target image piece is empty, the 4th indication information obtaining based on this decoding unit, according to the reconstruction pixel of this first basic tomographic image sub-block, be identified for the second reference information that this first object image subblock is encoded, wherein, this first basic tomographic image sub-block is the image block that is arranged in basic tomographic image, this target image piece is arranged in enhancement layer image, the basic tomographic image being somebody's turn to do is corresponding with this enhancement layer image, and this first primary image piece sub-block locus in this basic tomographic image is corresponding with this first object image subblock locus in this enhancement layer image,
This coding unit 1010 is also for this object code stream is decoded, to obtain this target image piece.
Alternatively, this decoding unit 1010 carries out block-eliminating effect filtering processing specifically for the pixel of the boundary vicinity between this target image sub-block.
Alternatively, this decoding unit 1010, specifically for from target code stream, obtains the 4th indication information, and wherein, the 4th indication information is adjacent with skip mode flag bit or fusion MERGE mode flags position information in this target code stream.
Alternatively, this decoding unit 1010, specifically for whether using reference information to decode according to the reference image block that is positioned at predeterminated position in this enhancement layer image, is determined context;
For according to this context, carry out entropy decoding, to determine the 4th indication information.
Can be corresponding to the method decoding end of the embodiment of the present invention according to the device 1000 of processing for image of the embodiment of the present invention, and, each unit of the device 1000 that should process for image is that module and above-mentioned other operations and/or function are respectively in order to realize the corresponding flow process of the method 800 in Fig. 9, for simplicity, do not repeat them here.
According to the device of processing for image of the embodiment of the present invention, for the corresponding sub-block that can not comprise from basic tomographic image in the target image piece of enhancement layer image, obtain the first object image subblock of movable information, by according to the location positioning second target image sub-block of this first object image subblock, and according to the reconstruction pixel of the first basic tomographic image sub-block corresponding with this first object image subblock on locus, determine the reference information for this first object image subblock, and according to the processing of encoding of this reference information, can improve the coding efficiency of this first object image subblock.
Above, in conjunction with Fig. 8 to Figure 11, describe the method and apparatus of processing for image according to the embodiment of the present invention in detail, below in conjunction with Figure 12 and Figure 13, describe the encoder of processing for image according to the embodiment of the present invention in detail.
Figure 12 shows according to the schematic block diagram of the encoder 1100 of processing for image of the embodiment of the present invention.As shown in figure 12, this encoder 1100 can comprise:
Bus 1110;
The processor 1120 being connected with this bus;
The memory 1130 being connected with this bus;
Wherein, this processor 1120 is by this bus 1110, call the program of storage in this memory 1130, for when determining that the movable information of the first basic tomographic image sub-block corresponding with the first object image subblock of target image piece is empty, according to the reconstruction pixel of this first basic tomographic image sub-block, be identified for the second reference information that this first object image subblock is encoded, wherein, this first basic tomographic image sub-block is the image block that is arranged in basic tomographic image, this target image piece is arranged in enhancement layer image, the basic tomographic image being somebody's turn to do is corresponding with this enhancement layer image, and this first primary image piece sub-block locus in this basic tomographic image is corresponding with this first object image subblock locus in this enhancement layer image,
This target image piece is encoded, to generate target code stream and to be contained in the 4th indication information in this target code stream.
Alternatively, this processor 1120 carries out block-eliminating effect filtering processing specifically for the pixel of the boundary vicinity between this target image sub-block.
Alternatively, this processor 1120 is specifically for the 4th indication information is carried out to entropy coding, so that the 4th indication information is adjacent with skip mode flag bit or fusion MERGE mode flags position information in this target code stream.
Alternatively, this processor 1120, specifically for whether using reference information to encode according to the reference image block that is positioned at predeterminated position in this enhancement layer image, is determined context;
For according to this context, the 4th indication information is carried out to entropy coding.
Can be corresponding to the method coding side of the embodiment of the present invention according to the encoder 1100 of processing for image of the embodiment of the present invention, and, each unit of the encoder 1100 that should process for image is that module and above-mentioned other operations and/or function are respectively in order to realize the corresponding flow process of the method 700 in Fig. 8, for simplicity, do not repeat them here.
According to the encoder of processing for image of the embodiment of the present invention, for the corresponding sub-block that can not comprise from basic tomographic image in the target image piece of enhancement layer image, obtain the first object image subblock of movable information, by according to the location positioning second target image sub-block of this first object image subblock, and according to the reconstruction pixel of the first basic tomographic image sub-block corresponding with this first object image subblock on locus, determine the reference information for this first object image subblock, and according to the processing of encoding of this reference information, can improve the coding efficiency of this first object image subblock.
Figure 13 shows according to the schematic block diagram of the decoder 1200 of processing for image of the embodiment of the present invention.As shown in Figure 7, this decoder 1200 can comprise:
Bus 1210;
The processor 1220 being connected with this bus;
The memory 1230 being connected with this bus;
Wherein, this processor 1220, by this bus 1210, calls the program of storage in this memory 1230, for from target code stream, obtains the 4th indication information;
For when determining that the movable information of the first basic tomographic image sub-block corresponding with the first object image subblock of target image piece is empty, based on the 4th indication information, according to the reconstruction pixel of this first basic tomographic image sub-block, be identified for the second reference information that this first object image subblock is encoded, wherein, this first basic tomographic image sub-block is the image block that is arranged in basic tomographic image, this target image piece is arranged in enhancement layer image, the basic tomographic image being somebody's turn to do is corresponding with this enhancement layer image, and this first primary image piece sub-block locus in this basic tomographic image is corresponding with this first object image subblock locus in this enhancement layer image,
For this object code stream is decoded, to obtain this target image piece.
Alternatively, this processor 1220 carries out block-eliminating effect filtering processing specifically for the pixel of the boundary vicinity between this target image sub-block.
Alternatively, this processor 1220, specifically for from target code stream, obtains the 4th indication information, and wherein, the 4th indication information is adjacent with skip mode flag bit or fusion MERGE mode flags position information in this target code stream.
Alternatively, this processor 1220, specifically for whether using reference information to decode according to the reference image block that is positioned at predeterminated position in this enhancement layer image, is determined context;
For according to this context, carry out entropy decoding, to determine the 4th indication information.
Can be corresponding to the method decoding end of the embodiment of the present invention according to the decoder 1200 of processing for image of the embodiment of the present invention, and, each unit of the decoder 1200 that should process for image is that module and above-mentioned other operations and/or function are respectively in order to realize the corresponding flow process of the method 800 in Fig. 9, for simplicity, do not repeat them here.
According to the decoder of processing for image of the embodiment of the present invention, for the corresponding sub-block that can not comprise from basic tomographic image in the target image piece of enhancement layer image, obtain the first object image subblock of movable information, by according to the location positioning second target image sub-block of this first object image subblock, and according to the reconstruction pixel of the first basic tomographic image sub-block corresponding with this first object image subblock on locus, determine the reference information for this first object image subblock, and according to the processing of encoding of this reference information, can improve the coding efficiency of this first object image subblock.
It should be noted that, for the reference information that coding side is obtained with decoding end consistent, so the method for obtaining reference information that requires coding side to use with decoding end is consistent,, if coding side using method 1 is obtained reference information (the first reference information), it is corresponding with method 1 that decoding end is wanted using method 2() obtain reference information (the first reference information).In other words, can be according to described coding side processing method correspondence definite decoding end processing method really, or the definite coding side processing method corresponding according to described decoding end processing method.
Should be understood that term "and/or" herein, is only a kind of incidence relation of describing affiliated partner, and expression can exist three kinds of relations, and for example, A and/or B, can represent: individualism A exists A and B, these three kinds of situations of individualism B simultaneously.In addition, character "/", generally represents that forward-backward correlation is to liking a kind of relation of "or" herein.
Should understand, in various embodiment of the present invention, the size of the sequence number of above-mentioned each process does not also mean that the priority of execution sequence, and the execution sequence of each process should determine with its function and internal logic, and should not form any restriction to the implementation process of the embodiment of the present invention.
Those of ordinary skills can recognize, unit and the algorithm steps of each example of describing in conjunction with embodiment disclosed herein, can realize with the combination of electronic hardware or computer software and electronic hardware.These functions are carried out with hardware or software mode actually, depend on application-specific and the design constraint of technical scheme.Professional and technical personnel can specifically should be used for realizing described function with distinct methods to each, but this realization should not thought and exceeds scope of the present invention.
Those skilled in the art can be well understood to, and for convenience and simplicity of description, the specific works process of the system of foregoing description, device and unit, can, with reference to the corresponding process in preceding method embodiment, not repeat them here.
In the several embodiment that provide in the application, should be understood that disclosed system, apparatus and method can realize by another way.For example, device embodiment described above is only schematic, for example, the division of described unit, be only that a kind of logic function is divided, during actual realization, can have other dividing mode, for example a plurality of unit or assembly can in conjunction with or can be integrated into another system, or some features can ignore, or do not carry out.Another point, shown or discussed coupling each other or direct-coupling or communication connection can be by some interfaces, indirect coupling or the communication connection of device or unit can be electrically, machinery or other form.
The described unit as separating component explanation can or can not be also physically to separate, and the parts that show as unit can be or can not be also physical locations, can be positioned at a place, or also can be distributed in a plurality of network element.Can select according to the actual needs some or all of unit wherein to realize the object of the present embodiment scheme.
In addition, each functional unit in each embodiment of the present invention can be integrated in a processing unit, can be also that the independent physics of unit exists, and also can be integrated in a unit two or more unit.
If described function usings that the form of SFU software functional unit realizes and during as production marketing independently or use, can be stored in a computer read/write memory medium.Understanding based on such, the part that technical scheme of the present invention contributes to prior art in essence in other words or the part of this technical scheme can embody with the form of software product, this computer software product is stored in a storage medium, comprise that some instructions are with so that a computer equipment (can be personal computer, server, or the network equipment etc.) carry out all or part of step of method described in each embodiment of the present invention.And aforesaid storage medium comprises: various media that can be program code stored such as USB flash disk, portable hard drive, read-only memory (ROM, Read-Only Memory), random access memory (RAM, Random Access Memory), magnetic disc or CDs.
The above; be only the specific embodiment of the present invention, but protection scope of the present invention is not limited to this, is anyly familiar with those skilled in the art in the technical scope that the present invention discloses; can expect easily changing or replacing, within all should being encompassed in protection scope of the present invention.Therefore, protection scope of the present invention should be as the criterion with the protection range of described claim.

Claims (66)

1. a method of processing for image, is characterized in that, described method comprises:
When determining that the movable information of the first basic tomographic image sub-block corresponding with the first object image subblock of target image piece is empty, the size of each the target image sub-block comprising according to the size of described target image piece, described target image piece and be used to indicate the second indication information of the position of first object image subblock in described target image piece, determines the second target image sub-block;
According to the movable information of described the second target image sub-block, be identified for the first reference information that described first object image subblock is encoded, wherein, the described first basic tomographic image sub-block is the image block that is arranged in basic tomographic image, described target image piece is arranged in enhancement layer image, described basic tomographic image is corresponding with described enhancement layer image, and described the first locus of primary image piece sub-block in described basic tomographic image is corresponding with the locus of described first object image subblock in described enhancement layer image;
Described target image piece is encoded, to generate target code stream and to be contained in the first indication information in described target code stream.
2. method according to claim 1, it is characterized in that: the size of the described target image sub-block comprising according to the size of target image piece, described target image piece and be used to indicate the second indication information of the position of first object image subblock in described target image piece, determine the second target image sub-block, comprising:
According to following arbitrary formula, determine described the second target image sub-block,
Idx 2=Idx 1/N×N+((Idx 1%N/(N/2))×2+(1-Idx 1%N/(N/4)%2))×N/4;
Idx 2=Idx 1/N×N+((1-Idx 1%N/(N/2))×2+(Idx 1%N/(N/4)%2))×N/4;
Idx 2=Idx 1/N×N+((1-Idx 1%N/(N/2))×2+(1-Idx 1%N/(N/4)%2))×N/4;
Wherein, Idx 2represent to be used to indicate the 3rd indication information of the position of described the second target image sub-block in described target image piece, Idx 1represent described the second indication information, N determines according to the size of the size of described target image piece and described target image sub-block.
3. method according to claim 1 and 2, is characterized in that, described according to the movable information of described the second target image sub-block, is identified for the first reference information that described first object image subblock is encoded, comprising:
If the movable information of described the second target image sub-block is empty, determine that described the first reference information is zero movable information.
4. it is characterized in that according to the method in any one of claims 1 to 3: described described target image piece is encoded and comprised:
According to described the first reference information, described first object image subblock is carried out to motion compensation process.
5. according to the method described in any one in claim 1 to 4, it is characterized in that: described described target image piece is encoded and comprised:
Pixel to the boundary vicinity between described target image sub-block is carried out block-eliminating effect filtering processing.
6. according to the method described in any one in claim 1 to 5, it is characterized in that: described described target image piece is encoded and comprised:
Described the first indication information is carried out to entropy coding, so that described the first indication information is adjacent with skip mode flag bit or fusion MERGE mode flags position information in described target code stream.
7. according to the method described in any one in claim 1 to 6, it is characterized in that: described described target image piece is encoded and comprised:
According to the reference image block that is positioned at predeterminated position in described enhancement layer image, whether use reference information to encode, determine context;
According to described context, described the first indication information is carried out to entropy coding.
8. a method of processing for image, is characterized in that, described method comprises:
From target code stream, obtain the first indication information;
When the movable information of the first corresponding basic tomographic image sub-block of the first object image subblock with target image piece is sky, based on described the first indication information, the size of each the target image sub-block comprising according to the size of described target image piece, described target image piece and be used to indicate the second indication information of the position of first object image subblock in described target image piece, determines the second target image sub-block;
According to the movable information of described the second target image sub-block, be identified for the first reference information that described first object image subblock is decoded, wherein, the described first basic tomographic image sub-block is the image block that is arranged in basic tomographic image, described target image piece is arranged in enhancement layer image, described basic tomographic image is corresponding with described enhancement layer image, and described the first locus of primary image piece sub-block in described basic tomographic image is corresponding with the locus of described first object image subblock in described enhancement layer image;
Described object code stream is decoded, to obtain described target image piece.
9. method according to claim 8, it is characterized in that: the size of the described target image sub-block comprising according to the size of target image piece, described target image piece and be used to indicate the second indication information of the position of first object image subblock in described target image piece, determine the second target image sub-block, comprising:
According to following arbitrary formula, determine described the second target image sub-block,
Idx 2=Idx 1/N×N+((Idx 1%N/(N/2))×2+(1-Idx 1%N/(N/4)%2))×N/4;
Idx 2=Idx 1/N×N+((1-Idx 1%N/(N/2))×2+(Idx 1%N/(N/4)%2))×N/4;
Idx 2=Idx 1/N×N+((1-Idx 1%N/(N/2))×2+(1-Idx 1%N/(N/4)%2))×N/4;
Wherein, Idx 2represent to be used to indicate the 3rd indication information of the position of described the second target image sub-block in described target image piece, Idx 1represent described the second indication information, N determines according to the size of the size of described target image piece and described target image sub-block.
10. method according to claim 8 or claim 9, is characterized in that, described according to the movable information of described the second target image sub-block, is identified for the first reference information that described first object image subblock is encoded, comprising:
If the movable information of described the second target image sub-block is empty, determine that described the first reference information is zero movable information.
Method in 11. according to Claim 8 to 10 described in any one, is characterized in that: described described object code stream is decoded and comprised:
According to described the first reference information, described first object image subblock is carried out to motion compensation process.
Method in 12. according to Claim 8 to 11 described in any one, is characterized in that: described described object code stream is decoded and comprised:
Pixel to the boundary vicinity between described target image sub-block is carried out block-eliminating effect filtering processing.
Method in 13. according to Claim 8 to 12 described in any one, is characterized in that: described from target code stream, obtain the first indication information, and comprising:
From target code stream, obtain the first indication information, wherein, described the first indication information is adjacent with skip mode flag bit or fusion MERGE mode flags position information in described target code stream.
Method in 14. according to Claim 8 to 13 described in any one, is characterized in that: described from target code stream, obtain the first indication information, and comprising:
According to the reference image block that is positioned at predeterminated position in described enhancement layer image, whether use reference information to decode, determine context;
According to described context, carry out entropy decoding, to determine described the first indication information.
15. 1 kinds of devices of processing for image, is characterized in that, described device comprises:
Acquiring unit, for when determining that the movable information of the first basic tomographic image sub-block corresponding with the first object image subblock of target image piece is empty, the size of each the target image sub-block comprising according to the size of described target image piece, described target image piece and be used to indicate the second indication information of the position of first object image subblock in described target image piece, determines the second target image sub-block;
Be used for according to the movable information of described the second target image sub-block, be identified for the first reference information that described first object image subblock is encoded, wherein, the described first basic tomographic image sub-block is the image block that is arranged in basic tomographic image, described target image piece is arranged in enhancement layer image, described basic tomographic image is corresponding with described enhancement layer image, and described the first locus of primary image piece sub-block in described basic tomographic image is corresponding with the locus of described first object image subblock in described enhancement layer image;
Coding unit, encodes to described target image piece, to generate target code stream and to be contained in the first indication information in described target code stream.
16. devices according to claim 15, is characterized in that: described acquiring unit, specifically for according to following arbitrary formula, is determined described the second target image sub-block,
Idx 2=Idx 1/N×N+((Idx 1%N/(N/2))×2+(1-Idx 1%N/(N/4)%2))×N/4;
Idx 2=Idx 1/N×N+((1-Idx 1%N/(N/2))×2+(Idx 1%N/(N/4)%2))×N/4;
Idx 2=Idx 1/N×N+((1-Idx 1%N/(N/2))×2+(1-Idx 1%N/(N/4)%2))×N/4;
Wherein, Idx 2represent to be used to indicate the 3rd indication information of the position of described the second target image sub-block in described target image piece, Idx 1represent described the second indication information, N determines according to the size of the size of described target image piece and described target image sub-block.
17. according to the device described in claim 15 or 16, it is characterized in that, if described acquiring unit is empty specifically for the movable information of described the second target image sub-block, determines that described the first reference information is zero movable information.
18. according to claim 15 to the device described in any one in 17, it is characterized in that: described coding unit, specifically for according to described the first reference information, carries out motion compensation process to described first object image subblock.
19. according to claim 15 to the device described in any one in 18, it is characterized in that: described coding unit carries out block-eliminating effect filtering processing specifically for the pixel of the boundary vicinity between described target image sub-block.
20. according to claim 15 to the device described in any one in 19, it is characterized in that: described coding unit is specifically for described the first indication information is carried out to entropy coding, so that described the first indication information is adjacent with skip mode flag bit or fusion MERGE mode flags position information in described target code stream.
21. according to claim 15 to the device described in any one in 20, it is characterized in that: described coding unit, specifically for whether using reference information to encode according to the reference image block that is positioned at predeterminated position in described enhancement layer image, is determined context;
For according to described context, described the first indication information is carried out to entropy coding.
22. 1 kinds of devices of processing for image, is characterized in that, described device comprises:
Decoding unit, for from target code stream, obtains the first indication information;
Acquiring unit, while being empty for the movable information when the first corresponding basic tomographic image sub-block of the first object image subblock with target image piece, described the first indication information obtaining based on described decoding unit, the size of each the target image sub-block comprising according to the size of described target image piece, described target image piece and be used to indicate the second indication information of the position of first object image subblock in described target image piece, determines the second target image sub-block;
Be used for according to the movable information of described the second target image sub-block, be identified for the first reference information that described first object image subblock is decoded, wherein, the described first basic tomographic image sub-block is the image block that is arranged in basic tomographic image, described target image piece is arranged in enhancement layer image, described basic tomographic image is corresponding with described enhancement layer image, and described the first locus of primary image piece sub-block in described basic tomographic image is corresponding with the locus of described first object image subblock in described enhancement layer image;
Described decoding unit is also for described object code stream is decoded, to obtain described target image piece.
23. devices according to claim 22, is characterized in that: described acquiring unit, specifically for according to following arbitrary formula, is determined described the second target image sub-block,
Idx 2=Idx 1/N×N+((Idx 1%N/(N/2))×2+(1-Idx 1%N/(N/4)%2))×N/4;
Idx 2=Idx 1/N×N+((1-Idx 1%N/(N/2))×2+(Idx 1%N/(N/4)%2))×N/4;
Idx 2=Idx 1/N×N+((1-Idx 1%N/(N/2))×2+(1-Idx 1%N/(N/4)%2))×N/4;
Wherein, Idx 2represent to be used to indicate the 3rd indication information of the position of described the second target image sub-block in described target image piece, Idx 1represent described the second indication information, N determines according to the size of the size of described target image piece and described target image sub-block.
24. according to the device described in claim 22 or 23, it is characterized in that, if described acquiring unit is empty specifically for the movable information of described the second target image sub-block, determines that described the first reference information is zero movable information.
25. according to the device described in any one in claim 22 to 24, it is characterized in that: described decoding unit, specifically for according to described the first reference information, carries out motion compensation process to described first object image subblock.
26. according to the device described in any one in claim 22 to 25, it is characterized in that: described decoding unit carries out block-eliminating effect filtering processing specifically for the pixel of the boundary vicinity between described target image sub-block.
27. according to the device described in any one in claim 22 to 26, it is characterized in that: described decoding unit is specifically for from target code stream, obtain the first indication information, wherein, described the first indication information is adjacent with skip mode flag bit or fusion MERGE mode flags position information in described target code stream.
28. according to the device described in any one in claim 22 to 27, it is characterized in that: described decoding unit, specifically for whether using reference information to decode according to the reference image block that is positioned at predeterminated position in described enhancement layer image, is determined context;
For according to described context, carry out entropy decoding, to determine described the first indication information.
29. 1 kinds of encoders of processing for image, is characterized in that, described encoder comprises:
Bus;
The processor being connected with described bus;
The memory being connected with described bus;
Wherein, described processor is by described bus, call the program of storing in described memory, for when determining that the movable information of the first basic tomographic image sub-block corresponding with the first object image subblock of target image piece is empty, the size of each the target image sub-block comprising according to the size of described target image piece, described target image piece and be used to indicate the second indication information of the position of first object image subblock in described target image piece, determines the second target image sub-block;
Be used for according to the movable information of described the second target image sub-block, be identified for the first reference information that described first object image subblock is encoded, wherein, the described first basic tomographic image sub-block is the image block that is arranged in basic tomographic image, described target image piece is arranged in enhancement layer image, described basic tomographic image is corresponding with described enhancement layer image, and described the first locus of primary image piece sub-block in described basic tomographic image is corresponding with the locus of described first object image subblock in described enhancement layer image;
For described target image piece is encoded, to generate target code stream and to be contained in the first indication information in described target code stream
30. encoders according to claim 29, is characterized in that: described processor, specifically for according to following arbitrary formula, is determined described the second target image sub-block,
Idx 2=Idx 1/N×N+((Idx 1%N/(N/2))×2+(1-Idx 1%N/(N/4)%2))×N/4;
Idx 2=Idx 1/N×N+((1-Idx 1%N/(N/2))×2+(Idx 1%N/(N/4)%2))×N/4;
Idx 2=Idx 1/N×N+((1-Idx 1%N/(N/2))×2+(1-Idx 1%N/(N/4)%2))×N/4;
Wherein, Idx 2represent to be used to indicate the 3rd indication information of the position of described the second target image sub-block in described target image piece, Idx 1represent described the second indication information, N determines according to the size of the size of described target image piece and described target image sub-block.
31. according to the encoder described in claim 29 or 30, it is characterized in that, if described processor is empty specifically for the movable information of described the second target image sub-block, determines that described the first reference information is zero movable information.
32. according to the encoder described in any one in claim 29 to 31, it is characterized in that: described processor, specifically for according to described reference information, carries out motion compensation process to described first object image subblock.
33. according to the encoder described in any one in claim 29 to 32, it is characterized in that: described processor carries out block-eliminating effect filtering processing specifically for the pixel of the boundary vicinity between described target image sub-block.
34. according to the encoder described in any one in claim 29 to 33, it is characterized in that: described processor is specifically for described the first indication information is carried out to entropy coding, so that described the first indication information is adjacent with skip mode flag bit or fusion MERGE mode flags position information in described target code stream.
35. according to the encoder described in any one in claim 29 to 34, it is characterized in that: described processor, specifically for whether using reference information to encode according to the reference image block that is positioned at predeterminated position in described enhancement layer image, is determined context;
For according to described context, described the first indication information is carried out to entropy coding.
36. 1 kinds of decoders of processing for image, is characterized in that, described decoder comprises:
Bus;
The processor being connected with described bus;
The memory being connected with described bus;
Wherein, described processor, by described bus, calls the program of storing in described memory, for from target code stream, obtains the first indication information;
While being empty for the movable information when the first corresponding basic tomographic image sub-block of the first object image subblock with target image piece, based on described the first indication information, the size of each the target image sub-block comprising according to the size of described target image piece, described target image piece and be used to indicate the second indication information of the position of first object image subblock in described target image piece, determines the second target image sub-block;
Be used for according to the movable information of described the second target image sub-block, be identified for the first reference information that described first object image subblock is decoded, wherein, the described first basic tomographic image sub-block is the image block that is arranged in basic tomographic image, described target image piece is arranged in enhancement layer image, described basic tomographic image is corresponding with described enhancement layer image, and described the first locus of primary image piece sub-block in described basic tomographic image is corresponding with the locus of described first object image subblock in described enhancement layer image;
For described object code stream is decoded, to obtain described target image piece.
37. decoders according to claim 36, is characterized in that: described processor, specifically for according to following arbitrary formula, is determined described the second target image sub-block,
Idx 2=Idx 1/N×N+((Idx 1%N/(N/2))×2+(1-Idx 1%N/(N/4)%2))×N/4;
Idx 2=Idx 1/N×N+((1-Idx 1%N/(N/2))×2+(Idx 1%N/(N/4)%2))×N/4;
Idx 2=Idx 1/N×N+((1-Idx 1%N/(N/2))×2+(1-Idx 1%N/(N/4)%2))×N/4;
Wherein, Idx 2represent to be used to indicate the 3rd indication information of the position of described the second target image sub-block in described target image piece, Idx 1represent described the second indication information, N determines according to the size of the size of described target image piece and described target image sub-block.
38. according to the decoder described in claim 36 or 37, it is characterized in that, if described processor is empty specifically for the movable information of described the second target image sub-block, determines that described the first reference information is zero movable information.
39. according to the decoder described in any one in claim 36 to 38, it is characterized in that: described processor, specifically for according to described the first reference information, carries out motion compensation process to described first object image subblock.
40. according to the decoder described in any one in claim 36 to 39, it is characterized in that: described processor carries out block-eliminating effect filtering processing specifically for the pixel of the boundary vicinity between described target image sub-block.
41. according to the decoder described in any one in claim 36 to 40, it is characterized in that: described processor is specifically for from target code stream, obtain the first indication information, wherein, described the first indication information is adjacent with skip mode flag bit or fusion MERGE mode flags position information in described target code stream.
42. according to the decoder described in any one in claim 36 to 41, it is characterized in that: described processor, specifically for whether using reference information to decode according to the reference image block that is positioned at predeterminated position in described enhancement layer image, is determined context;
For according to described context, carry out entropy decoding, to determine described the first indication information.
43. 1 kinds of methods of processing for image, is characterized in that, described method comprises:
When determining that the movable information of the first basic tomographic image sub-block corresponding with the first object image subblock of target image piece is empty, according to the reconstruction pixel of the described first basic tomographic image sub-block, be identified for the second reference information that described first object image subblock is encoded, wherein, the described first basic tomographic image sub-block is the image block that is arranged in basic tomographic image, described target image piece is arranged in enhancement layer image, described basic tomographic image is corresponding with described enhancement layer image, and described the first locus of primary image piece sub-block in described basic tomographic image is corresponding with the locus of described first object image subblock in described enhancement layer image,
Described target image piece is encoded, to generate target code stream and to be contained in the 4th indication information in described target code stream.
44. according to the method described in claim 43, it is characterized in that: described described target image piece is encoded and comprised:
Pixel to the boundary vicinity between described target image sub-block is carried out block-eliminating effect filtering processing.
45. according to the method described in claim 43 or 44, it is characterized in that: described described target image piece is encoded and comprised:
Described the 4th indication information is carried out to entropy coding, so that described the 4th indication information is adjacent with skip mode flag bit or fusion MERGE mode flags position information in described target code stream.
46. according to the method described in any one in claim 43 to 45, it is characterized in that: described described target image piece is encoded and comprised:
According to the reference image block that is positioned at predeterminated position in described enhancement layer image, whether use reference information to encode, determine context;
According to described context, described the 4th indication information is carried out to entropy coding.
47. 1 kinds of methods of processing for image, is characterized in that, described method comprises:
From target code stream, obtain the 4th indication information;
When determining that the movable information of the first basic tomographic image sub-block corresponding with the first object image subblock of target image piece is empty, based on described the 4th indication information, according to the reconstruction pixel of the described first basic tomographic image sub-block, be identified for the second reference information that described first object image subblock is encoded, wherein, the described first basic tomographic image sub-block is the image block that is arranged in basic tomographic image, described target image piece is arranged in enhancement layer image, described basic tomographic image is corresponding with described enhancement layer image, and described the first locus of primary image piece sub-block in described basic tomographic image is corresponding with the locus of described first object image subblock in described enhancement layer image,
Described object code stream is decoded, to obtain described target image piece.
48. according to the method described in claim 47, it is characterized in that: described described object code stream is decoded and comprised:
Pixel to the boundary vicinity between described target image sub-block is carried out block-eliminating effect filtering processing.
49. according to the method described in claim 47 or 48, it is characterized in that: described from target code stream, obtain the 4th indication information, and comprising:
From target code stream, obtain the 4th indication information, wherein, described the 4th indication information is adjacent with skip mode flag bit or fusion MERGE mode flags position information in described target code stream.
50. according to the method described in any one in claim 47 to 48, it is characterized in that: described from target code stream, obtain the 4th indication information, and comprising:
According to the reference image block that is positioned at predeterminated position in described enhancement layer image, whether use reference information to decode, determine context;
According to described context, carry out entropy decoding, to determine described the 4th indication information.
51. 1 kinds of devices of processing for image, is characterized in that, described device comprises:
Acquiring unit, for when determining that the movable information of the first basic tomographic image sub-block corresponding with the first object image subblock of target image piece is empty, according to the reconstruction pixel of the described first basic tomographic image sub-block, be identified for the second reference information that described first object image subblock is encoded, wherein, the described first basic tomographic image sub-block is the image block that is arranged in basic tomographic image, described target image piece is arranged in enhancement layer image, described basic tomographic image is corresponding with described enhancement layer image, and described the first locus of primary image piece sub-block in described basic tomographic image is corresponding with the locus of described first object image subblock in described enhancement layer image,
Coding unit, for described target image piece is encoded, to generate target code stream and to be contained in the 4th indication information in described target code stream.
52. according to the device described in claim 51, it is characterized in that: described coding unit carries out block-eliminating effect filtering processing specifically for the pixel of the boundary vicinity between described target image sub-block.
53. according to the device described in claim 51 or 52, it is characterized in that: described coding unit is specifically for described the 4th indication information is carried out to entropy coding, so that described the 4th indication information is adjacent with skip mode flag bit or fusion MERGE mode flags position information in described target code stream.
54. according to the device described in any one in claim 51 to 53, it is characterized in that: described coding unit, specifically for whether using reference information to encode according to the reference image block that is positioned at predeterminated position in described enhancement layer image, is determined context;
For according to described context, described the 4th indication information is carried out to entropy coding.
55. 1 kinds of devices of processing for image, is characterized in that, described device comprises:
Decoding unit, for from target code stream, obtains the 4th indication information;
Acquiring unit, for when determining that the movable information of the first basic tomographic image sub-block corresponding with the first object image subblock of target image piece is empty, the 4th indication information obtaining based on described decoding unit, according to the reconstruction pixel of the described first basic tomographic image sub-block, be identified for the second reference information that described first object image subblock is encoded, wherein, the described first basic tomographic image sub-block is the image block that is arranged in basic tomographic image, described target image piece is arranged in enhancement layer image, described basic tomographic image is corresponding with described enhancement layer image, and described the first locus of primary image piece sub-block in described basic tomographic image is corresponding with the locus of described first object image subblock in described enhancement layer image,
Described coding unit is also for described object code stream is decoded, to obtain described target image piece.
56. according to the device described in claim 55, it is characterized in that: described decoding unit carries out block-eliminating effect filtering processing specifically for the pixel of the boundary vicinity between described target image sub-block.
57. according to the device described in claim 55 or 56, it is characterized in that: described decoding unit is specifically for from target code stream, obtain the 4th indication information, wherein, described the 4th indication information is adjacent with skip mode flag bit or fusion MERGE mode flags position information in described target code stream.
58. according to the device described in any one in claim 55 to 57, it is characterized in that: described decoding unit, specifically for whether using reference information to decode according to the reference image block that is positioned at predeterminated position in described enhancement layer image, is determined context;
For according to described context, carry out entropy decoding, to determine described the 4th indication information.
59. 1 kinds of encoders of processing for image, is characterized in that, described encoder comprises:
Bus;
The processor being connected with described bus;
The memory being connected with described bus;
Wherein, described processor is by described bus, call the program of storing in described memory, for when determining that the movable information of the first basic tomographic image sub-block corresponding with the first object image subblock of target image piece is empty, according to the reconstruction pixel of the described first basic tomographic image sub-block, be identified for the second reference information that described first object image subblock is encoded, wherein, the described first basic tomographic image sub-block is the image block that is arranged in basic tomographic image, described target image piece is arranged in enhancement layer image, described basic tomographic image is corresponding with described enhancement layer image, and described the first locus of primary image piece sub-block in described basic tomographic image is corresponding with the locus of described first object image subblock in described enhancement layer image,
Described target image piece is encoded, to generate target code stream and to be contained in the 4th indication information in described target code stream.
60. according to the encoder described in claim 59, it is characterized in that: described processor is specifically used the pixel of the boundary vicinity between described target image sub-block is carried out to block-eliminating effect filtering processing.
61. according to the encoder described in claim 59 or 60, it is characterized in that, described processor is specifically for described the 4th indication information is carried out to entropy coding, so that described the 4th indication information is adjacent with skip mode flag bit or fusion MERGE mode flags position information in described target code stream.
62. according to the encoder described in any one in claim 59 to 61, it is characterized in that: described processor, specifically for whether using reference information to encode according to the reference image block that is positioned at predeterminated position in described enhancement layer image, is determined context;
For according to described context, described the 4th indication information is carried out to entropy coding.
63. 1 kinds of decoders of processing for image, is characterized in that, described decoder comprises:
Bus;
The processor being connected with described bus;
The memory being connected with described bus;
Wherein, described processor, by described bus, calls the program of storing in described memory, for from target code stream, obtains the 4th indication information;
For when determining that the movable information of the first basic tomographic image sub-block corresponding with the first object image subblock of target image piece is empty, based on described the 4th indication information, according to the reconstruction pixel of the described first basic tomographic image sub-block, be identified for the second reference information that described first object image subblock is encoded, wherein, the described first basic tomographic image sub-block is the image block that is arranged in basic tomographic image, described target image piece is arranged in enhancement layer image, described basic tomographic image is corresponding with described enhancement layer image, and described the first locus of primary image piece sub-block in described basic tomographic image is corresponding with the locus of described first object image subblock in described enhancement layer image,
For described object code stream is decoded, to obtain described target image piece.
64. according to the decoder described in claim 63, it is characterized in that: described processor carries out block-eliminating effect filtering processing specifically for the pixel of the boundary vicinity between described target image sub-block.
65. according to the decoder described in claim 63 or 64, it is characterized in that, described processor, specifically for from target code stream, obtains the 4th indication information, wherein, described the 4th indication information is adjacent with skip mode flag bit or fusion MERGE mode flags position information in described target code stream.
66. according to the decoder described in any one in claim 63 to 65, it is characterized in that: described processor, specifically for whether using reference information to decode according to the reference image block that is positioned at predeterminated position in described enhancement layer image, is determined context;
For according to described context, carry out entropy decoding, to determine described the 4th indication information.
CN201210375019.5A 2012-09-29 2012-09-29 Image processing method, device, coder and decoder Active CN103716629B (en)

Priority Applications (2)

Application Number Priority Date Filing Date Title
CN201210375019.5A CN103716629B (en) 2012-09-29 2012-09-29 Image processing method, device, coder and decoder
PCT/CN2013/084504 WO2014048372A1 (en) 2012-09-29 2013-09-27 Method and device for image processing, coder and decoder

Applications Claiming Priority (1)

Application Number Priority Date Filing Date Title
CN201210375019.5A CN103716629B (en) 2012-09-29 2012-09-29 Image processing method, device, coder and decoder

Publications (2)

Publication Number Publication Date
CN103716629A true CN103716629A (en) 2014-04-09
CN103716629B CN103716629B (en) 2017-02-22

Family

ID=50387015

Family Applications (1)

Application Number Title Priority Date Filing Date
CN201210375019.5A Active CN103716629B (en) 2012-09-29 2012-09-29 Image processing method, device, coder and decoder

Country Status (2)

Country Link
CN (1) CN103716629B (en)
WO (1) WO2014048372A1 (en)

Cited By (6)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN106664416A (en) * 2014-07-06 2017-05-10 Lg电子株式会社 Method for processing video signal, and apparatus therefor
CN109996075A (en) * 2017-12-29 2019-07-09 华为技术有限公司 A kind of picture decoding method and decoder
CN110891180A (en) * 2018-09-10 2020-03-17 华为技术有限公司 Video decoding method and video decoder
CN112040243A (en) * 2018-06-04 2020-12-04 华为技术有限公司 Method and device for obtaining motion vector
CN112598572A (en) * 2019-10-01 2021-04-02 浙江大学 Method and device for screening subblock images and processing units
WO2022068716A1 (en) * 2020-09-30 2022-04-07 华为技术有限公司 Entropy encoding/decoding method and device

Citations (4)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN1764280A (en) * 2004-10-21 2006-04-26 三星电子株式会社 Method and apparatus based on multilayer effective compressing motion vector in video encoder
CN101198064A (en) * 2007-12-10 2008-06-11 武汉大学 Movement vector prediction method in resolution demixing technology
CN101755458A (en) * 2006-07-11 2010-06-23 诺基亚公司 Scalable video coding
US20120201301A1 (en) * 2006-01-11 2012-08-09 Qualcomm Incorporated Video coding with fine granularity spatial scalability

Family Cites Families (1)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
KR100704626B1 (en) * 2005-02-07 2007-04-09 삼성전자주식회사 Method and apparatus for compressing multi-layered motion vectors

Patent Citations (4)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN1764280A (en) * 2004-10-21 2006-04-26 三星电子株式会社 Method and apparatus based on multilayer effective compressing motion vector in video encoder
US20120201301A1 (en) * 2006-01-11 2012-08-09 Qualcomm Incorporated Video coding with fine granularity spatial scalability
CN101755458A (en) * 2006-07-11 2010-06-23 诺基亚公司 Scalable video coding
CN101198064A (en) * 2007-12-10 2008-06-11 武汉大学 Movement vector prediction method in resolution demixing technology

Cited By (18)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN106664416B (en) * 2014-07-06 2019-11-05 Lg电子株式会社 Handle the method and device thereof of vision signal
US10567755B2 (en) 2014-07-06 2020-02-18 Lg Electronics Inc. Method for processing video signal, and apparatus therefor
CN106664416A (en) * 2014-07-06 2017-05-10 Lg电子株式会社 Method for processing video signal, and apparatus therefor
CN109996075A (en) * 2017-12-29 2019-07-09 华为技术有限公司 A kind of picture decoding method and decoder
CN109996075B (en) * 2017-12-29 2022-07-12 华为技术有限公司 Image decoding method and decoder
CN112040243B (en) * 2018-06-04 2021-06-29 华为技术有限公司 Method and device for obtaining motion vector
US11812050B2 (en) 2018-06-04 2023-11-07 Huawei Technologies Co., Ltd. Motion vector obtaining method and apparatus
US11463721B2 (en) 2018-06-04 2022-10-04 Huawei Technologies Co., Ltd. Motion vector obtaining method and apparatus
CN112040243A (en) * 2018-06-04 2020-12-04 华为技术有限公司 Method and device for obtaining motion vector
CN112703735A (en) * 2018-09-10 2021-04-23 华为技术有限公司 Video decoding method and video decoder
CN112703735B (en) * 2018-09-10 2022-03-08 华为技术有限公司 Video encoding/decoding method, related apparatus and computer-readable storage medium
CN112040247A (en) * 2018-09-10 2020-12-04 华为技术有限公司 Video decoding method and video decoder
US11706417B2 (en) 2018-09-10 2023-07-18 Huawei Technologies Co., Ltd. Video decoding method and video decoder
CN110891180A (en) * 2018-09-10 2020-03-17 华为技术有限公司 Video decoding method and video decoder
CN110891180B (en) * 2018-09-10 2023-11-17 华为技术有限公司 Video decoding method and video decoder
CN112598572B (en) * 2019-10-01 2022-04-15 浙江大学 Method and device for screening subblock images and processing units
CN112598572A (en) * 2019-10-01 2021-04-02 浙江大学 Method and device for screening subblock images and processing units
WO2022068716A1 (en) * 2020-09-30 2022-04-07 华为技术有限公司 Entropy encoding/decoding method and device

Also Published As

Publication number Publication date
CN103716629B (en) 2017-02-22
WO2014048372A1 (en) 2014-04-03

Similar Documents

Publication Publication Date Title
CN104796709B (en) The method and apparatus that the coding unit of image boundary is encoded and decodes
CN104967852B (en) The method and apparatus that image is coded and decoded by intra prediction
CN104581161B (en) With the use of method and the equipment of large-scale conversion cell encoding and decoded picture
CN106060547B (en) The method and apparatus of decoding high resolution image
CN103716629A (en) Image processing method, device, coder and decoder
CN102098519B (en) Video encoding method and decoding method as well as encoding and decoding device
CN106464889A (en) Inter-layer video decoding method and apparatus therefor performing sub-block-based prediction, and inter-layer video encoding method and apparatus therefor performing sub-block-based prediction
CN104718758A (en) Transform-domain intra prediction
CN103079067B (en) Motion vector predictor list builder method and video coding-decoding method and device
CN104255032A (en) Inter-layer prediction method and encoding device and decoding device using same
CN103716631A (en) Image processing method, device, coder and decoder
CN106105208A (en) Scalable video/coding/decoding method and equipment
CN103596003B (en) Interframe predication quick mode selecting method for high-performance video coding
CN102474620A (en) Spatial prediction method and apparatus in layered video coding
CN102685505B (en) Intra-frame prediction method and prediction device
CN102497545B (en) Content adaptive and art directable scalable video coding
CN101584220A (en) Method and system for encoding a video signal, encoded video signal, method and system for decoding a video signal
KR101845622B1 (en) Adaptive rdpcm method for video coding, video encoding method based on adaptive rdpcm and video decoding method based on adaptive rdpcm
CN102215392A (en) Intra-frame predicting method or device for estimating pixel value
CN107005705A (en) The method and apparatus for being encoded or being decoded to multi-layer image using inter-layer prediction
CN103716622B (en) Image processing method and device
CN103248885B (en) Intra-frame image prediction decoding method and Video Codec
CN105049854A (en) Methods for encoding and decoding a picture and corresponding devices
CN103716627B (en) layered coding and decoding method and device
CN109302615B (en) Self-adaptive template prediction method for video compression

Legal Events

Date Code Title Description
C06 Publication
PB01 Publication
C10 Entry into substantive examination
SE01 Entry into force of request for substantive examination
C14 Grant of patent or utility model
GR01 Patent grant