CN101385352A - Inter-layer prediction method for video signal - Google Patents

Inter-layer prediction method for video signal Download PDF

Info

Publication number
CN101385352A
CN101385352A CNA200780005672XA CN200780005672A CN101385352A CN 101385352 A CN101385352 A CN 101385352A CN A200780005672X A CNA200780005672X A CN A200780005672XA CN 200780005672 A CN200780005672 A CN 200780005672A CN 101385352 A CN101385352 A CN 101385352A
Authority
CN
China
Prior art keywords
picture
macro block
layer
basic unit
inter
Prior art date
Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
Granted
Application number
CNA200780005672XA
Other languages
Chinese (zh)
Other versions
CN101385352B (en
Inventor
朴胜煜
全柄文
朴志皓
Current Assignee (The listed assignees may be inaccurate. Google has not performed a legal analysis and makes no representation or warranty as to the accuracy of the list.)
LG Electronics Inc
Original Assignee
LG Electronics Inc
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Priority claimed from KR1020060111893A external-priority patent/KR20070075257A/en
Priority claimed from KR1020070001587A external-priority patent/KR20070075293A/en
Priority claimed from KR1020070001582A external-priority patent/KR20070095180A/en
Application filed by LG Electronics Inc filed Critical LG Electronics Inc
Priority to CN201210585882.3A priority Critical patent/CN103096078B/en
Priority claimed from PCT/KR2007/000147 external-priority patent/WO2007081139A1/en
Publication of CN101385352A publication Critical patent/CN101385352A/en
Application granted granted Critical
Publication of CN101385352B publication Critical patent/CN101385352B/en
Active legal-status Critical Current
Anticipated expiration legal-status Critical

Links

Images

Classifications

    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04NPICTORIAL COMMUNICATION, e.g. TELEVISION
    • H04N19/00Methods or arrangements for coding, decoding, compressing or decompressing digital video signals
    • H04N19/50Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using predictive coding
    • H04N19/503Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using predictive coding involving temporal prediction
    • H04N19/51Motion estimation or motion compensation
    • H04N19/53Multi-resolution motion estimation; Hierarchical motion estimation
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04NPICTORIAL COMMUNICATION, e.g. TELEVISION
    • H04N19/00Methods or arrangements for coding, decoding, compressing or decompressing digital video signals
    • H04N19/10Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using adaptive coding
    • H04N19/102Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using adaptive coding characterised by the element, parameter or selection affected or controlled by the adaptive coding
    • H04N19/132Sampling, masking or truncation of coding units, e.g. adaptive resampling, frame skipping, frame interpolation or high-frequency transform coefficient masking
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04NPICTORIAL COMMUNICATION, e.g. TELEVISION
    • H04N19/00Methods or arrangements for coding, decoding, compressing or decompressing digital video signals
    • H04N19/10Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using adaptive coding
    • H04N19/134Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using adaptive coding characterised by the element, parameter or criterion affecting or controlling the adaptive coding
    • H04N19/157Assigned coding mode, i.e. the coding mode being predefined or preselected to be further used for selection of another element or parameter
    • H04N19/16Assigned coding mode, i.e. the coding mode being predefined or preselected to be further used for selection of another element or parameter for a given display mode, e.g. for interlaced or progressive display mode
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04NPICTORIAL COMMUNICATION, e.g. TELEVISION
    • H04N19/00Methods or arrangements for coding, decoding, compressing or decompressing digital video signals
    • H04N19/10Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using adaptive coding
    • H04N19/169Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using adaptive coding characterised by the coding unit, i.e. the structural portion or semantic portion of the video signal being the object or the subject of the adaptive coding
    • H04N19/17Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using adaptive coding characterised by the coding unit, i.e. the structural portion or semantic portion of the video signal being the object or the subject of the adaptive coding the unit being an image region, e.g. an object
    • H04N19/176Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using adaptive coding characterised by the coding unit, i.e. the structural portion or semantic portion of the video signal being the object or the subject of the adaptive coding the unit being an image region, e.g. an object the region being a block, e.g. a macroblock
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04NPICTORIAL COMMUNICATION, e.g. TELEVISION
    • H04N19/00Methods or arrangements for coding, decoding, compressing or decompressing digital video signals
    • H04N19/30Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using hierarchical techniques, e.g. scalability
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04NPICTORIAL COMMUNICATION, e.g. TELEVISION
    • H04N19/00Methods or arrangements for coding, decoding, compressing or decompressing digital video signals
    • H04N19/85Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using pre-processing or post-processing specially adapted for video compression
    • H04N19/88Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using pre-processing or post-processing specially adapted for video compression involving rearrangement of data among different coding units, e.g. shuffling, interleaving, scrambling or permutation of pixel data or permutation of transform coefficient data among different blocks

Abstract

The present invention relates to a method for conducting interlayer motion prediction in encoding or decoding of video signal. The present method identifies the base layer and the types on the current layer or the types of the macro blocks comprised in the image; if the image on the current layer or the types of the macro blocks comprised in the image is identified as intra mode, and the image on the base layer or the types of the macro blocks comprised in the image is identified as non-interlaced, the motion information of the macro block on the base layer is copied to form blocks on a virtual layer; the motion information copied from the blocks on the virtual layer is used for the interlayer motion prediction of the macro blocks of the image on the current layer.

Description

The inter-layer prediction method that is used for vision signal
1. technical field
The present invention relates to be used for when encoding/decoding video signal, carrying out the method for inter-layer prediction.
2. background technology
Liftable level Video Codec (SVC) encodes video into the picture sequence with high image quality, guarantee that simultaneously the part (specifically, from entire frame sequence discontinuous the partial frame sequence selected) of encoded picture sequence can be decoded and be used for showing this video with low image quality.
Although can show the low image quality video by the part that receives and handle according to the picture sequence of liftable level scheme coding, if but still the existence bit rate reduce then problem that picture quality significantly descends.A solution of this problem provides the auxiliary picture sequence of low bit rate---picture sequence that for example has small screen size and/or low frame per second---as the one deck at least in the hierarchical structure.
When hypothesis provided two sequences, auxiliary (descend) picture sequence was called as basic unit, and lead (on) picture sequence is called as enhancing or reinforced layer.The vision signal of basic unit and enhancement layer has redundancy, and is two-layer because identical video signal source is encoded into.In order to improve the encoding-decoding efficiency of enhancement layer, the vision signal of enhancement layer uses the information through encoding and decoding (movable information or texture information) of basic unit to come encoding and decoding.
Although can as shown in Figure 1a single video source 1 be encoded into a plurality of layer with different transmissibilitys, a plurality of video source 2b that also can will comprise shown in Fig. 1 b under the different scanning pattern of identical content 2a are encoded into corresponding each layer.Equally, in this case, the encoder on coding upper strata can improve the encoding and decoding gain by the encoded information and executing inter-layer prediction that utilizes lower floor, because two source 2b provide identical content 2a.
Therefore, need provide a kind of inter-layer prediction method of when different source codes being become corresponding each layer, the scan pattern of vision signal being included in consideration.When coding interlaced video, it can be encoded into idol field and strange field, and the strange and even macro block that also can be encoded in the frame is right.Correspondingly, also must consider to be used for the picture type of encoding and decoding interlaced video signal for inter-layer prediction.
Generally speaking, enhancement layer provides resolution to be higher than the picture of basic unit's resolution.Correspondingly, if all layer picture has different resolution when different source codes being become corresponding each layer, then also need to carry out interpolation and improve screen resolution (that is picture size).Because the image of the basic unit's picture that uses in inter-layer prediction for the encoding and decoding for prediction approaches the image of enhancement layer picture more, the encoding and decoding rate is just high more, so need provide a kind of scan pattern of the vision signal with all layer to include the interpolating method of consideration in.
3. summary of the invention
The purpose of this invention is to provide a kind of method that in two-layer, has inter-layer prediction under the situation that one deck has the interlaced video signal component at least.
Another object of the present invention provides a kind ofly carries out method to all layer inter-layer motion prediction of picture with different spatial resolutions (liftable level) according to picture type.
Another purpose of the present invention provides the method for a kind of execution to all layer inter-layer texture prediction of picture with different spaces and/or temporal resolution (liftable level).
A kind of inter-layer motion prediction method according to the present invention comprises: the motion related information of internal schema macro block is arranged to the motion related information of inter mode macro block, and this internal schema is two vertical macro blocks that adjoin of basic unit with the inter mode macro block; Right movable information is used for inter-layer motion prediction to obtain vertically to adjoin macro block based on these two macro blocks that vertically adjoin then.
Another kind of inter-layer motion prediction method according to the present invention comprises: the inter mode piece that will be arranged to have 0 motion related information as the internal schema macro block of one of two internal schemas of vertically adjoining of basic unit and inter mode macro block; Right movable information is used for inter-layer motion prediction to obtain vertically to adjoin macro block based on these two macro blocks that vertically adjoin then.
Another kind of inter-layer motion prediction method according to the present invention comprises: from the derive movable information of single macro block of the right movable information that vertically adjoins the frame macro block of basic unit; And with the movable information derived information of forecasting as the right movable information separately of the movable information of the field macro block in anterior layer or the field macro block in anterior layer.
Another kind of inter-layer motion prediction method according to the present invention comprises from the movable information of single macro block of basic unit or the movable information that vertically adjoins single right macro block of a macro block that is selected from basic unit two macro blocks movable information separately of deriving; And with the movable information separately derived as when the frame macro block of anterior layer information of forecasting to separately movable information.
A kind of inter-layer motion prediction method of all layer that is used for having the picture of different resolution according to the present invention comprises: by using the Forecasting Methodology of conversion framing macro block the picture of lower floor to be transformed into the frame picture of equal resolution according to the type of picture and the type selecting ground of picture macro block; Rise this frame picture of sampling so that it has the resolution identical with the resolution on upper strata; Use then and be applicable to the type of the frame macro block in this frame picture and the inter-layer prediction method of the macro block (mb) type in the picture of upper strata through rising sampling.
Another kind of all layer the inter-layer motion prediction method that is used for having the picture of different resolution according to the present invention comprises: identify lower floor and upper strata picture type and/or be included in the type of the macro block of these pictures; Lower layer pictures is used the virtual screen that has the aspect ratio identical with structure from single the right method of macroblock prediction frame macro block according to the result who identifies with the aspect ratio of upper strata picture; Rise this virtual screen of sampling; Utilize this virtual screen to the upper layer application inter-layer motion prediction then through rising sampling.
Another kind of all layer the inter-layer motion prediction method that is used for having the picture of different resolution according to the present invention comprises: identify lower floor and upper strata picture type and/or be included in the type of the macro block of these pictures; Lower layer pictures is used the virtual screen that has the aspect ratio identical with structure from single the right method of macroblock prediction frame macro block according to the result who identifies with the aspect ratio of upper strata picture; And utilize the virtual screen that is constructed that the picture on upper strata is used inter-layer motion prediction.
Another kind of all layer the inter-layer motion prediction method that is used to have the picture of different resolution according to the present invention comprises: the type that identifies lower floor and upper strata picture; If the type of lower layer pictures is that the type of field and upper strata picture is line by line, the movable information that then copies the piece in the lower layer pictures is with the constructing virtual picture; Rise this virtual screen of sampling; And between this virtual screen and upper strata picture through rising sampling application of frame macro block-macroblock motion prediction method.
Another kind of all layer the inter-layer motion prediction method that is used to have the picture of different resolution according to the present invention comprises: the type that identifies lower floor and upper strata picture; If the type of lower layer pictures is and the type of upper strata picture is line by line, the movable information of piece that then copies lower floor is with the constructing virtual picture; And use this virtual screen to come the upper strata picture is used inter-layer motion prediction.
In an embodiment of the present invention, in inter-layer motion prediction, sequentially predict partition mode, reference key and motion vector.
In another embodiment of the present invention, sequentially prediction reference index, motion vector and partition mode.
In another embodiment of the present invention, the field macro block right movable information that be used for the virtual basic unit of inter-layer motion prediction is to derive from the right movable information of the frame macro block of basic unit.
In another embodiment of the present invention, the movable information of field macro block of an occasionally strange picture that be used for the virtual basic unit of inter-layer motion prediction is to derive from the right movable information of the frame macro block of basic unit.
In another embodiment of the present invention, select macro block, and the right movable information of frame macro block that will be used for the virtual basic unit of inter-layer motion prediction is to derive from the movable information of selected macro block from the field macro block centering of basic unit.
In another embodiment of the present invention, the right movable information of frame macro block that be used for the virtual basic unit of inter-layer motion prediction is to derive from the movable information of the field macro block of the occasionally strange picture of basic unit.
In another embodiment of the present invention, the information of the field macro block in the occasionally strange picture of basic unit is copied with other constructing virtual field macro block, and the right movable information of frame macro block that will be used for the virtual basic unit of inter-layer motion prediction is to derive from the right movable information of field macro block that constructs in this way.
A kind of inter-layer texture prediction method according to the present invention comprises: by basic unit vertically to adjoin the frame macro block right to structure macro block; And with the field macro block that constructed to separately texture information as when the field of anterior layer macro block to separately texture prediction information.
Another kind of inter-layer texture prediction method according to the present invention comprises: vertically adjoin the frame macro block to constructing single macro block by basic unit; And with the texture information of single the macro block that constructed as texture prediction information when the field of anterior layer macro block.
Another kind of inter-layer texture prediction method according to the present invention comprises: by single macro block of basic unit or vertically to adjoin a macro block right to structure frame macro block; And with the frame macro block that constructed to separately texture information as when the frame macro block of anterior layer to separately texture prediction information.
Another kind of inter-layer texture prediction method according to the present invention comprise by basic unit vertically adjoin a macro block to structure N to the frame macro block, wherein N is the integer greater than 1; And with the N that constructed to frame macro block texture information separately as when the anterior layer meta in the N of different time position to frame macro block texture prediction information separately.
Another kind of inter-layer texture prediction method according to the present invention comprises: each frame of lower floor is divided into a plurality of pictures has the temporal resolution identical with the upper strata to allow lower floor; Rise each isolated picture of institute of sampling in vertical direction to expand each isolated picture of institute in vertical direction; The inter-layer texture prediction that then each field picture through rising sampling is used for each frame on upper strata.
Another kind of inter-layer texture prediction method according to the present invention comprises: each picture that rises sampling lower floor in vertical direction is to expand each picture in vertical direction; And the inter-layer texture prediction that each field picture through rising sampling is used for each frame on upper strata.
Another kind of inter-layer texture prediction method according to the present invention comprises: each frame on upper strata is divided into a plurality of pictures; The picture of down-sampled lower floor is to dwindle the picture of lower floor in vertical direction; The inter-layer texture prediction that will be used for isolated the picture on upper strata then through down-sampled picture.
A kind of method of inter-layer prediction encoded video signal of utilizing according to the present invention comprises: determine to be to use the 2N piece texture information separately of constructing with the selected row of selecting of order layout then by the row of the 2N piece in the arbitrariness picture of alternately selecting basic unit in inter-layer texture prediction, also be to use by interpolation and be selected from the 2N piece texture information separately that a piece of the 2N piece of basic unit is constructed; And will indicate this information of determining to bring in the information encoded.
A kind of inter-layer prediction that utilizes according to the present invention comes the method for decoded video signal to comprise: check whether specific indication information is included in the signal that receives; And determine in inter-layer texture prediction, to be to use the 2N piece texture information separately of constructing by the selected row of selecting of order layout then by the row of the 2N piece in the arbitrariness picture of alternately selecting basic unit based on the result who is checked, also be to use by interpolation and be selected from the 2N piece texture information separately that a piece of the 2N piece of basic unit is constructed.
In an embodiment of the present invention, each frame of upper strata or lower floor is divided into two field pictures.
In an embodiment of the present invention, if specific indication information is not included in the received signal, then this situation is considered as with receive comprise the signal that is set as 0 indication information and determined its separately texture information will to be used for the situation of piece of inter-layer prediction identical.
A kind of method that the vision signal of basic unit is used for inter-layer texture prediction according to the present invention comprises: the interlaced video signal of basic unit is divided into the strange field component of even summation; On vertical and/or horizontal direction, the strange field component of even summation is amplified separately; To be used for inter-layer texture prediction through the strange field component group combination of the even summation that amplifies then.
The another kind of method that the vision signal of basic unit is used for inter-layer texture prediction according to the present invention comprises: the progressive video signal of basic unit is divided into idol row group and strange row group; On vertical and/or horizontal direction, the very capable group of even summation is amplified separately; The very row group combination and be used for inter-layer texture prediction of even summation that will be through amplifying.
The another kind of method that the vision signal of basic unit is used for inter-layer texture prediction according to the present invention comprises: amplify the interlaced video signal of basic unit so that it has the resolution identical with the progressive video signal on upper strata on vertical and/or horizontal direction; And the inter-layer texture prediction of carrying out the vision signal on upper strata based on vision signal through amplifying.
The another kind of method that the vision signal of basic unit is used for inter-layer texture prediction according to the present invention comprises: amplify the progressive video signal of basic unit so that it has the resolution identical with the interlaced video signal on upper strata on vertical and/or horizontal direction; And the inter-layer texture prediction of carrying out the vision signal on upper strata based on vision signal through amplifying.
In one embodiment of the invention, vision signal is separated and amplification is to carry out in macro block rank (or i.e. basis at macro block on).
In another embodiment of the present invention, vision signal is separated and is amplified and carry out on the picture rank.
In another embodiment of the present invention, if it is used the picture format difference of two layers of inter-layer texture prediction,, then carry out vision signal and separate and amplify if promptly one deck comprises line by line picture and another layer comprises interlaced picture.
In another embodiment of the present invention, if all be interlacing, then carry out vision signal and separate and amplify its two-layer picture of using inter-layer texture prediction.
4. accompanying drawing summary
Fig. 1 a and 1b illustrate and single video frequency source coding are become a plurality of layers method;
Fig. 2 a and 2b schematically illustrate the configuration of application according to the video signal coding apparatus of inter-layer prediction method of the present invention;
Fig. 2 c and 2d illustrate the type of the picture sequence that is used for coding interlaced vision signal;
Fig. 3 a and 3b schematically show the process of wherein constructing basic unit's picture for inter-layer texture prediction according to an embodiment of the invention and carrying out de-blocking filter;
Fig. 4 a to 4f schematically shows the movable information process of utilizing the movable information of frame macro block to derive of field macro block of the virtual basic unit of the inter-layer motion prediction that wherein will be used for MBAFF frame midfield macro block according to an embodiment of the invention;
Fig. 4 g schematically shows according to an embodiment of the invention the program that the right texture information of macro block wherein is used to the right texture prediction of field macro block in the MBAFF frame;
Fig. 4 h illustrates according to embodiments of the invention the frame macro block being transformed into the right method of a macro block;
Fig. 5 a and 5b illustrate reference key and movable information derivation program according to another embodiment of the invention;
Fig. 6 a to 6c schematically shows the derive program of movable information of the field macro block in the virtual basic unit of the movable information that wherein utilizes the frame macro block according to an embodiment of the invention;
Fig. 6 d schematically shows according to an embodiment of the invention the program that the right texture information of frame macro block wherein is used to the texture prediction of the field macro block in the picture;
Fig. 7 a and 7b illustrate reference key and movable information derivation program according to another embodiment of the invention;
The movable information that Fig. 8 a to 8c schematically shows the field macroblock frame macro block of the virtual basic unit that wherein will be used for inter-layer motion prediction according to an embodiment of the invention is a program of utilizing the movable information of the field macro block of MBAFF frame to derive;
Fig. 8 d schematically shows according to an embodiment of the invention the program that the right texture information of field macro block in the MBAFF frame wherein is used to the right texture prediction of frame macro block;
Fig. 8 e illustrates according to embodiments of the invention the field macro block the right method of conversion framing macro block;
The program that when Fig. 8 f and 8g schematically show according to an embodiment of the invention macro block centering then and there only a macro block is inter mode the right texture information of field macro block in the MBAFF frame is used for the right inter-layer prediction of frame macro block;
Fig. 8 h schematically shows according to an embodiment of the invention the program that the right texture information of field macro block in the MBAFF frame wherein is used to many texture predictions to the frame macro block;
Fig. 9 a and 9b illustrate reference key and movable information derivation program according to another embodiment of the invention;
The movable information that Figure 10 a to 10c schematically shows the frame macro block of the virtual basic unit that will be used for inter-layer motion prediction according to an embodiment of the invention therein is a program of utilizing the movable information of the field macro block of a picture to derive;
The texture information that Figure 10 d schematically shows the field macro block in its midfield picture according to an embodiment of the invention is used to the program of the right texture prediction of frame macro block;
Figure 11 illustrates reference key and movable information derivation program according to another embodiment of the invention;
The movable information that Figure 12 a and 12b schematically show the frame macro block of the virtual basic unit that wherein will be used for inter-layer motion prediction according to another embodiment of the invention is a program of utilizing the movable information of the field macro block of a picture to derive;
Figure 13 a to 13d schematically shows the program that the movable information of movable information utilization field macro block of the field macro block of the virtual basic unit that will be used for inter-layer motion prediction is according to an embodiment of the invention derived according to the type of picture respectively;
Figure 14 a to 14k respectively according to the type of picture illustrate according to various embodiments of the present invention in all layer the spatial resolution method of motion prediction between execution level simultaneously not;
Figure 15 a and 15b schematically show according to an embodiment of the invention enhancement layer be line by line and the picture that will have the basic unit of different spatial resolutions when basic unit is interlacing is used for the program of inter-layer texture prediction;
Figure 16 a and 16b schematically show according to an embodiment of the invention wherein for the picture with basic unit is used for inter-layer texture prediction macro block with picture to being divided into the program that macro block and isolated macro block are exaggerated;
Figure 17 a and 17b schematically show according to an embodiment of the invention enhancement layer be interlacing and basic unit is the picture that will have the basic unit of different spatial resolutions in line by line is used for the program of inter-layer texture prediction;
Figure 18 schematically shows according to an embodiment of the invention at enhancement layer and the picture that will have the basic unit of different spatial resolutions when basic unit all is interlacing is used for the program of inter-layer prediction;
It is the program that progressive frame sequence and two-layer picture type and temporal resolution are not predicted between application layer simultaneously that Figure 19 a illustrates according to an embodiment of the invention at enhancement layer;
It is progressive frame sequence and the two-layer program of using inter-layer prediction when having different picture types with identical resolution that Figure 19 b illustrates according to an embodiment of the invention at enhancement layer;
It is the program that progressive frame sequence and two-layer picture type and temporal resolution are not predicted between application layer simultaneously that Figure 20 illustrates according to an embodiment of the invention in basic unit; And
It is progressive frame sequence and the two-layer program of using inter-layer prediction when having different picture types with identical resolution that Figure 21 illustrates according to an embodiment of the invention in basic unit.
5. embodiments of the present invention
With reference now to accompanying drawing, describes embodiments of the invention in detail.
Fig. 2 a schematically shows the building block of application according to the video signal coding apparatus of inter-layer prediction method of the present invention.Although the device of Fig. 2 a is realized as and incoming video signal is encoded into two-layer, the principle of the following description of the present invention also is applicable in vision signal and is encoded into three layers or even interlayer process when more multi-layered.
Enhancement layer (EL) encoder 20 places of inter-layer prediction method according to the present invention in the device of Fig. 2 a carry out.Encoded information (movable information and texture information) receives at basic unit (EL) encoder 21 places.Based on the information and executing inter-layer texture prediction or the motion prediction that are received.If needed, the information that received of decoding and then based on the information and executing prediction that decodes.Certainly, in the present invention, shown in Fig. 2 b, incoming video signal can be to use the video source 3 of the basic unit that has been encoded to come encoding and decoding.Inter-layer prediction method as described below is suitable equally in this case.
In the situation of Fig. 2 a, can have BS encoder 21 coding interlaced vision signals therein or therein the encoded video source 3 of Fig. 2 b by two kinds of methods of encoding and decoding.Particularly, in one of these two kinds of methods, shown in Fig. 3 a, interlaced video signal is encoded into a sequence simply on the basis by the field, and in another approach, shown in Fig. 3 b, each frame that comes tectonic sequence is encoded into frame sequence with frame by macro block with two (even summation is strange) field.The last macro block of macro block centering in Bian Ma the frame is called as " top macro block " in this way, and macro block is called as " end macro block " down.If the top macro block is made of idol (or strange) field picture component, then end macro block is made of strange (or idol) field picture component.Gou Zao frame is called as macro block adaptive frame/field (MBAFF) frame in this way.The MBAFF frame can comprise that not only each self-contained macro block strange and an idol macro block is right, can comprise that also the macro block of each self-contained two frame macro block is right.
Correspondingly, when the macro block in the picture had the interlaced picture component, it may be the macro block in the field, and also may be the macro block in the frame.Each macro block with interlaced picture component is called as a macro block, and each macro block that has line by line (scanning) picture content is called the frame macro block.
Therefore, need by determine will the macro block of EL encoder 20 places coding and will inter-layer prediction at macro block in the base layer macro block type separately used be that a frame macro block (mb) type or a macro block (mb) type are determined inter-layer prediction method.If macro block is a macro block, then need by determine it be in the field or the MBAFF frame in a field macro block determine inter-layer prediction method.
To at each situation this method be described respectively.Before describing, suppose that the resolution when anterior layer equals the resolution of basic unit.That is, suppose that SpatialScalabilityType () is 0.Description when the resolution of anterior layer is higher than basic unit's resolution will provide after a while.In following description and accompanying drawing, term " top " and " idol " (or strange) are used interchangeably, and term " end " and " very " (or by chance) are used interchangeably.
In order to utilize basic unit to come inter-layer prediction, at first need the basic unit of decoding with coding or decoding enhancement layer.Therefore, it is as follows at first to describe base layer decoder.
In decoding during basic unit, the basic unit's movable information such as partition mode, reference key and motion vector of not only decoding, the texture of the basic unit of also decoding.
When the texture of basic unit is decoded when being used for inter-layer texture prediction, be not that all images sample data of basic unit is all decoded, this is in order to reduce the load of decoder.The image pattern data of internal schema macro block are decoded comes out, and the inter mode macro block be only residual error data---to be the error information between the image pattern data---decoded comes out and need not adjoin picture and carry out motion compensation.
In addition, the basic unit's texture decoder that is used for inter-layer texture prediction is not by on the basis of macro block but carrying out by on the basis of picture, goes up the basic unit picture consistent with the enhancement layer picture with the structure time.Basic unit's picture is to construct by the image pattern data that reconstruct from the internal schema macro block with from the residual error data that the inter mode macro block decodes as mentioned above.
Internal schema such as DCT and quantification or inter mode motion compensation and conversion are carried out on the image block basis, are for example carrying out on 16 x, 16 macroblock basis or on 4 x, 4 sub-piece bases.This causes the branch blocking artefacts at block boundary place to make picture distortion.Use de-blocking filter and reduce these minutes blocking artefacts.Deblocking filter makes the edge of image block smoothly to improve the quality of frame of video.
Whether using de-blocking filter reduces piecemeal distortion and depends on that image block is in the intensity of boundary and the gradient of border surrounding pixel.The dynamics of deblocking filter or the degree pixel value before by the image block partition mode of quantization parameter, internal schema, inter mode, indicator collet size etc., motion vector, de-blocking filter etc. is determined.
Deblocking filter in the inter-layer prediction is the internal schema macro block that is applied in basic unit's picture on basis of texture prediction of basic internal schema (intraBL or the interlayer internal schema) macro block as enhancement layer.
In the time will shown in Fig. 2 c, being encoded into picture sequence entirely according to inter-layer prediction method coding two-layer, this two-layer frame format that is counted as entirely, thus make and can easily derive the coding/decoding process that comprises de-blocking filter from encoding-decoding process at frame format.
Now will be at the picture format of the basic unit situation different with the picture format of enhancement layer---be that basic unit is frame (or promptly line by line) form and the basic unit situation for (or being interlacing) form, basic unit is the situation of frame format for field form basic unit, although or the enhancement layer shown in Fig. 2 c and 2d and the both of basic unit are for the field form but one of enhancement layer and basic unit are encoded into a picture sequence that another is encoded into the situation of MBAFF frame---the method for carrying out de-blocking filter according to embodiments of the invention is described.
Fig. 3 a and 3b schematically show and construct basic unit's picture is used for the de-blocking filter of inter-layer texture prediction with execution process according to an embodiment of the invention therein.
It is the embodiment for a form of frame format and basic unit that Fig. 3 a illustrates enhancement layer wherein, and basic unit is the embodiment of frame format for field form basic unit and Fig. 3 b illustrates wherein.
In these embodiments, for inter-layer texture prediction, the inter mode macro block of basic unit and the texture of internal schema macro block are decoded, comprise basic unit's picture of image pattern data and residual error data with structure, and recently rise the picture of sampling being constructed according to the resolution of the resolution (or being screen size) of basic unit and enhancement layer to reduce after dividing blocking artefacts at the picture that deblocking filter is applied to constructed.
First method among Fig. 3 a and the 3b (method 1) is that wherein basic unit is divided into two field pictures to carry out the method for de-blocking filter.In the method, when utilizing when creating enhancement layer with the basic unit of different pictures form coding, basic unit's picture is divided into an idol row picture and a strange row picture, and these two field pictures are deblocked (that is, be used to deblock filtering) and rise sampling.Then these two picture splicings are become single picture, and carry out inter-layer texture prediction based on this single picture.
This first method comprises following three steps.
At separating step (field, the end (or the idol field) picture that step 1), basic unit's picture are divided into field, top (or the strange) picture that comprises idol and go and comprise strange row.Basic unit's picture is to comprise the residual error data (inter mode data) that reconstructs from the data flow of basic unit by motion compensation and the video pictures of image pattern data (internal schema data).
In the step (step 2) of deblocking, separated picture deblocked by deblocking filter in separating step.Here, can use conventional deblocking filter as this deblocking filter.
When the resolution of the resolution of enhancement layer and basic unit not simultaneously, the field picture through deblocking recently rises sampling according to the resolution of the resolution of enhancement layer and basic unit.
The splicing step (step 3), field, top picture through rising sampling and field, end picture through rising sampling in an alternating manner by interlacing scan to be spliced into single picture.Afterwards, carry out the texture prediction of enhancement layer based on this single picture.
In second method in Fig. 3 a and 3b (method 2), when utilizing the basic unit of encoding to create enhancement layer with the different pictures form, basic unit's picture is not divided into two field pictures but it is directly deblocked to it and rises sampling, and carry out inter-layer texture prediction based on the picture of gained as a result.
In this second method, and to not be divided into top and bottom field picture by the corresponding basic unit of the enhancement layer picture picture that inter-layer texture prediction is encoded but deblocked immediately, rise sampling then.Afterwards, carry out the texture prediction of enhancement layer based on this picture through rising sampling.
The deblocking filter that is applied to basic unit's picture of constructing for inter-layer motion prediction only is applied to comprising the zone of the image pattern data that decode from the internal schema macro block, and is not applied to comprising the zone of residual error data.
Be encoded into a form in the basic unit among Fig. 3 a---be that basic unit is encoded into a picture sequence or is encoded into shown in Fig. 2 d under the situation of MBAFF frame shown in Fig. 2 c, in order to use second method, the row that need carry out interlacing scan top and bottom field picture alternately with it is combined into single picture (under the situation of Fig. 2 c) or alternately the row of the right top and bottom macro block of interlaced field macro block it is combined into the process of single picture (under the situation of Fig. 2 d).This process will be described in detail with reference to figure 8d and 8e.To be to comprise the residual error data (inter mode data) that reconstructs by motion compensation and the field picture or the macro block of image pattern data (internal schema data) by interleaved top and bottom field picture or top and bottom macro block.
In addition, (basic unit) in the MBAFF frame shown in Fig. 2 d right top and bottom macro block of macro block is different patterns and selects the internal schema piece to be used under the situation of the right inter-layer texture prediction of the macro block of enhancement layer (under the situation of Fig. 8 g that describes after a while) from these macro blocks, shown in Fig. 2 d be encoded in the field right basic unit of macro block in the MBAFF frame any frame (picture) in time with the inconsistent situation of enhancement layer picture under (under the situation of Fig. 8 h that describes after a while), or be under the situation of base layer prediction of the field macro block shown in Fig. 2 c (under the situation of Figure 10 d that describes after a while) with picture at texture with the right enhancement layer of macro block, one that chooses in the macro block is risen and is sampled into interim macro block to (" 841 " among Fig. 8 g and " 851 " among Fig. 8 h and " 852 ") or two interim macro blocks (" 1021 " among Figure 10 d), and deblocking filter is applied to internal schema macro block in these macro blocks.
The inter-layer texture prediction of describing in following various embodiment is based on carrying out through the basic unit's picture that deblocks of describing among the embodiment of Fig. 3 a and 3b.
Now will be at according to the macro block (mb) type in the anterior layer of wanting encoding and decoding with will be used for describing inter-layer prediction method respectively when each situation that the macro block (mb) type of the basic unit of the inter-layer prediction of the macro block of anterior layer is classified.In this description, hypothesis equals the spatial resolution of basic unit when the spatial resolution of anterior layer as described above.
I. frame MB-〉the situation of field MB in the MBAFF frame
In this case, the macro block in anterior layer (EL) is encoded into the field macro block in the MBAFF frame, and will be used for macro block when the basic unit of the inter-layer prediction of the macro block of the anterior layer framing macro block that is encoded.In last macro block in the basic unit and the following macro block in the macro block of included vision signal composition and a pair of coordination in anterior layer included vision signal composition be identical.It is right that upper and lower (top and bottom) macro block will be called as macro block, and term " to " will be used to describe a pair of piece that vertically adjoins in the following description.At first, the description inter-layer motion prediction is as follows.
EL encoder 20 uses by the macro block of basic unit is become macro-block partition mode that single macro block (by being compressed to half size in vertical direction) the obtains partition mode as current macro to 410 merger.Fig. 4 a illustrates the specific example of this process.As shown, at first, the respective macroblock of basic unit is become single macro block (S41) to 410 merger, and the partition mode of the macro block that obtains by merger is copied into another macro block to construct macro block to 411 (S42).Afterwards, with this macro block that macro block 411 partition mode separately is applied to virtual basic unit to 412 (S43).
Yet, when corresponding macro block is merged into single macro block to 410, may be created on unallowed zoning in the partition mode.In order to prevent this situation, EL encoder 20 is determined partition mode according to following rule.
1) two 8 x of the top and bottom of the macro block centering of basic unit 8 (" B8_0 " among Fig. 4 a and " B8_2 ") are merged into 8 of single 8 x.But if any among 8 of corresponding 8 x do not segmented, then they are merged into 4 of two 8 x, and if have any to be segmented among 8 of corresponding 8 x, then they are merged into four 44 of x (" 401 " among Fig. 4 a).
2) 16 of 8 x of basic unit dwindle into 8 of 8 x, and 8 of 16 x dwindle into 4 of two 8 x that adjoin, and 16 of 16 x dwindle into 8 of 16 x.
If respective macroblock is centering to a rare macro block with the internal schema coding, then EL encoder 20 was at first carried out following process before the merger process.
If only have one to be internal schema in these two macro blocks, then macro block shown in Fig. 4 b, be copied into interior macro block such as the movable information macro-block partition mode, reference key and the motion vector, perhaps in macro block shown in Fig. 4 c, be considered to have 16 macro blocks of 16 x of 0 motion vector and 0 reference key.Perhaps, shown in Fig. 4 d, the reference key of interior macro block by will between the reference key of macro block copy in macro block be provided with, and 0 motion vector is distributed to interior macro block.Then, carry out above mentioned merger process, execution reference key as described below then and motion vector derivation program.
EL encoder 20 carry out following process with from respective macroblock to 410 reference key derivation current macro to 412 reference key.
If each piece corresponding to 8 centerings of 8 x of basic unit of 8 of current 8 x has been subdivided into a similar number part, then the reference key of 8 centerings of this 8 x one (jacking block or sole piece) is confirmed as the reference key of 8 of current 8 x.Otherwise that the reference key that 8 centerings of this 8 x have been subdivided into less number part is confirmed as the reference key of 8 of current 8 x.
In another embodiment of the present invention, for corresponding to 8 of 8 x of basic unit of 8 of current 8 x to a less reference key that is confirmed as 8 of current 8 x in the reference key that is provided with.This definite method in the example of Fig. 4 e can be expressed as follows:
Reference key=min of current B8_0 (reference key of the B8_0 of basic top frame MB, the reference key of the B8_2 of basic top frame MB)
Reference key=min of current B8_1 (reference key of the B8_1 of basic top frame MB, the reference key of the B8_3 of basic top frame MB)
Reference key=min of current B8_2 (reference key of the B8_0 of substrate frame MB, the reference key of the B8_2 of substrate frame MB), and
Reference key=min of current B8_3 (reference key of the B8_1 of substrate frame MB, the reference key of the B8_3 of substrate frame MB).
Above reference key derivation program applicable to top and bottom field macro block both.The reference key of 8 of each definite in this way 8 x be multiply by 2, and the reference key after will multiplying each other is defined as its final reference index.The reason of making this multiplication is that the number of picture is the twice of the number in the frame sequence when decoding, is divided into even and strange because belong to the field macro block of picture.Depend on decoding algorithm, the final reference index of field, end macro block can add 1 and determines by its reference key being multiply by 2 reference keys after will multiplying each other then.
It below is the derive program of the right motion vector of the macro block of virtual basic unit of EL encoder 20.
Motion vector is what to determine on the basis of 4 of 4 x, so 8 quilts of corresponding 4 x of basic unit identify out, shown in Fig. 4 f.If these 8 of corresponding 4 x have been segmented, then the motion vector of 4 of its top or the ends 4 x is confirmed as the motion vector of 4 of current 4 x.Otherwise, the motion vector of 8 of 4 x of correspondence is defined as the motion vector of 4 of current 4 x.Determined motion vector is used as the final motion vector of 4 of current 4 x after its vertical component is divided by 2.The reason of making this division is to be included in two iconic elements in the frame macro block corresponding to the iconic element of a field macro block thereby make the size of field picture reduce half in vertical direction.
In case the field macro block of virtual basic unit determines in this way to 412 movable information, the target field macro block that this movable information just is used to enhancement layer is to 413 inter-layer motion prediction.Equally, in the following description, in case the right movable information of the macro block of virtual basic unit or macro block is determined, this movable information just is used to respective macroblock or the right inter-layer motion prediction of respective macroblock when anterior layer.In the following description, suppose to be used to also be employed when respective macroblock or right this process of inter-layer motion prediction of respective macroblock of anterior layer even without the macro block of mentioning virtual basic unit or the right movable information of macro block.
The field macro block that Fig. 5 schematically shows the virtual basic unit that will be used to inter-layer prediction according to another embodiment of the invention to 500 movable information how from deriving corresponding to the right movable information of the right basic frame macro block of current macro.In the present embodiment, as shown in the figure, the field macro block that the reference key that the top of the top macro block that the frame macro block of basic unit is right or the ends 8 x are 8 is used as virtual basic unit is to the reference key of 8 of top 8 x of each macro block in 500, and the reference key of 8 of the top of the end macro block of basic unit or the ends 8 x is used as the reference key of this macro block to 8 of the ends 8 x of each macro block in 500.On the other hand, as shown in the figure, the motion vector that 4 x of the top of the top macro block that the frame macro block of basic unit is right 4 is common to the field macro block of virtual basic unit to 4 of each macro block 4 x topmost in 500, the motion vector that the 3rd 4 x of the top macro block that the frame macro block of basic unit is right is 4 is common to this macro block to 4 of second 4 x of each macro block in 500, the motion vector that right end macro block 4 x topmost of the frame macro block of basic unit 4 is common to this macro block to 4 of the 3rd 4 x of each macro block in 500, and the motion vector of 4 of the 3rd 4 x of the right end macro block of the frame macro block of basic unit is common to this macro block to 4 of the 4th 4 x of each macro block in 500.
Shown in Fig. 5 a, for being used for the motion vector of 4 of 4 x among field macro block that inter-layer prediction constructs uses basic unit to 4 501 of top 4 x among 8 of 8 x among 8 of 500 8 x and 4 502 of the ends 4 x 8 511 of different 8 x and 512.These motion vectors may be to use the motion vector of different reference pictures.That is, 8 511 of different 8 x may have different reference keys with 512.Correspondingly, in this case, for the macro block of constructing virtual basic unit to 500, EL encoder 20 will be made the motion vector of 4 502 of second 4 x of virtual basic unit for the motion vector that pushes up 4 503 of 4 501 corresponding 4 x that select of 4 x is shared, shown in Fig. 5 b (521).
In the embodiment that describes with reference to figure 4a to 4f, for the movable information of constructing virtual basic unit with the right movable information of prediction current macro, EL encoder 20 is based on the right movable information of the respective macroblock of basic unit sequentially derive partition mode, reference key and motion vector.Yet, in with reference to figure 5a and the described embodiment of 5b, EL encoder 20 is finally determined the right partition mode of macro block of virtual basic unit then at first based on the derive right reference key and the motion vector of macro block of virtual basic unit of the right movable information of the respective macroblock of basic unit based on the value of being derived.When partition mode is determined, 4 x, 4 module units with the identical motion vector of deriving and reference key are combined, if and the block mode after the combination is the partition mode that allows, then partition mode is arranged to the pattern after this combination, otherwise the pattern before partition mode is arranged to make up.
In the above-described embodiment, if the respective macroblock of basic unit all is an internal schema to two macro blocks in 410, then current macro is carried out prediction in the base to 413.In this case, do not carry out motion prediction.Certainly, the macro block of constructing virtual basic unit is not right under the situation of texture prediction.If the respective macroblock of basic unit is an internal schema to having only a macro block in 410, the movable information of macro block is copied to interior macro block between then inciting somebody to action shown in Fig. 4 b, shown in Fig. 4 c, the motion vector and the reference key of interior macro block are arranged to 0, perhaps shown in Fig. 4 d by will between the reference key of macro block copy in macro block be provided with in macro block reference key and the motion vector of interior macro block is arranged to 0.Then, the right movable information of the macro block of virtual basic unit is derived as described above.
Be as mentioned above inter-layer motion prediction constructing virtual basic unit macro block to after, EL encoder 20 use the right movable information of macro block that is constructed predict and encode when the front court macro block to 413 movable information.
Now inter-layer texture prediction will be described.Fig. 4 g is illustrated in texture Forecasting Methodology between exemplary layer under the situation of " frame MB-〉in the MBAFF frame field MB ".The respective frame macro block that EL encoder 20 identifies basic unit is to 410 block mode.If the respective frame macro block is to two macro blocks in 410 or all be internal schema or all be inter mode, then EL encoder 20 becomes interim field macro block to 421 to 410 conversions (conversion) respective macroblock of basic unit, so that or carry out when the front court macro block and predict (when two frame macro blocks 410 all are internal schema) or carry out its residual prediction (when two frame macro blocks 410 all are inter mode) in the mode that describes below in to 413 base.When two macro blocks of corresponding macro block in to 410 all were internal schema, this interim field macro block comprises 421 foregoingly finished the deblocked data of (that is, be used to deblock filtering) of decoding back under the situation of internal schema.In following description to various embodiment, for the interim macro block of deriving from the macro block of the basic unit that is used for texture prediction to so same.
Yet, do not carry out inter-layer texture prediction when in these two macro blocks, only having one to be inter mode.The macro block that is used for the basic unit of inter-layer texture prediction is the raw image data (or through decoded image data) that has un-encoded under the situation of internal schema to 410 at macro block, and is to have encoded residual error data (or through decoding residual error data) under the situation of inter mode at macro block.Following to the description of texture prediction in for the macro block of basic unit to so same.
Fig. 4 h illustrates and is used for the frame macro block converting the right method of field macro block that will be used for inter-layer texture prediction to.As shown in the figure, the idol row of sequentially selecting a pair of frame macro block A and B is with a structure top macro block A ', and sequentially select this to the strange row of frame macro block A and B with structure field, end macro block B '.When filling a field macro block with row, it is at first filled with idol (or strange) row (A_ occasionally A_ is strange) of jacking block A, fills with strange (or idol) row (B occasionally B_ is strange) of sole piece B then.
II. frame MB-〉the situation of field MB in the picture
In this case, when the macro block in the anterior layer is the field macro block that is encoded in the picture, and the macro block that will be used for when the basic unit of the inter-layer prediction of the macro block of anterior layer is the framing macro block that is encoded.Among the included vision signal composition of macro block centering in the basic unit and idol in anterior layer or strange in the macro block of coordination included vision signal composition identical.At first, inter-layer motion prediction is described below.
EL encoder 20 uses by the macro block with basic unit and merger is become the partition mode of the macro-block partition mode of single macro block (by being compressed to half size in vertical direction) acquisition as the occasionally strange macro block of virtual basic unit.Fig. 6 a illustrates the detailed example of this process.As shown in the figure, at first the respective macroblock with basic unit becomes single macro block 611 (S61) to 610 merger, and will be applied to be used for the macro block (S62) of virtual basic unit of the inter-layer motion prediction of current macro 613 by the partition mode that this merger obtains.Identical among merger rule and the previous situation I.Identical among processing method when corresponding macro block has at least a macro block to encode with internal schema in to 610 and the previous situation I.
The program of reference key and motion vector of being used for deriving is also carried out with the mode same way as of describing with top situation I formerly.In situation I, identical derivation program is applied to the top and bottom macro block, because the strange macro block of even summation is to being carried in the frame.Yet this situation II and situation I difference are the derivation program only is applied to a field macro block, shown in Fig. 6 b and 6c because want encoding and decoding only exist in the picture of front court one corresponding to base layer macro block to 610 macro block.
In above embodiment, for the movable information of the macro block of predicting virtual basic unit, EL encoder 20 is based on the right movable information of the respective macroblock of basic unit sequentially derive partition mode, reference key and the motion vector of this macro block.
In another embodiment of the present invention, EL encoder 20 is at first based on the derive reference key and the motion vector of macro block of virtual basic unit of the right movable information of the respective macroblock of basic unit, then, finally determine the block mode of the macro block of virtual basic unit based on the value of being derived.Fig. 7 a and 7b schematically show the field reference key of macro block of virtual basic unit and the derivation of motion vector.The class of operation that is used for deriving in this case is similar to the operation with reference to the situation I of figure 5a and 5b description, and difference is to push up or the movable information of end macro block is to utilize the right movable information of macro block of basic unit to derive.
When partition mode is finally determined, 4x4 module unit with the identical motion vector of deriving and reference key is combined, if and the block mode after the combination is the partition mode that allows, then partition mode is arranged to the pattern after this combination, otherwise the pattern before partition mode is arranged to make up.
In the above-described embodiment, if two macro blocks of the respective macroblock centering of basic unit all are internal schemas, then do not carry out motion prediction, the also right movable information of the macro block of constructing virtual basic unit not, if and only have one to be internal schema in these two macro blocks, then carry out motion prediction in this case as described previously.
Now inter-layer texture prediction will be described.Fig. 6 d is illustrated in texture Forecasting Methodology between exemplary layer under the situation of " frame MB-〉in the picture field MB ".The respective macroblock that EL encoder 20 identifies basic unit is to 610 block mode.If two macro blocks of this macro block centering or all be internal schema or all be inter mode, then EL encoder 20 by single to the interim macro block 621 of frame macro block 610 structures.If current macro 613 belongs to an idol picture, then EL encoder 20 is constructed an interim macro block 621 by respective macroblock to 610 idol row.If current macro 613 belongs to strange picture, then EL encoder 20 is constructed an interim macro block 621 by respective macroblock to 610 strange row.Building method is similar to the method for constructing single macro block A ' or B ' among Fig. 4 h.
In case an interim macro block 621 is configured out, EL encoder 20 is just carried out when predicting (when two macro blocks of corresponding macro block in to 610 all are internal schema) in the base of front court macro block 613 based on the texture information in the field macro block 621, or carries out its residual prediction (when two macro blocks of corresponding macro block in to 610 all are inter mode).
If corresponding macro block is an inter mode to having only a macro block in 610, then EL encoder 20 is not carried out inter-layer texture prediction.
MB-in the III.MBAFF frame〉situation of frame MB
In this case, the macro block in anterior layer is the framing macro block that is encoded, and the macro block that will be used for when the basic unit of the inter-layer prediction of the frame macro block of anterior layer is the field macro block that is encoded in the MBAFF frame.Field in the basic unit in the macro block included vision signal composition with work as anterior layer in the macro block of a pair of coordination in included vision signal composition be identical.At first, inter-layer motion prediction is described below.
EL encoder 20 uses macro-block partition mode by right top of expansion base layer macro block or end macro block (expanding to twice in vertical direction) acquisition as the right partition mode of the macro block in the virtual basic unit.Fig. 8 a illustrates the specific example of this process.Although be that top macro block is selected in following description and accompanying drawing, when field, end macro block is selected below described same being suitable for.
Shown in Fig. 8 a, the respective macroblock of basic unit is expanded to twice constructing two macro blocks 811 (S81) to 810 field, top macro block, and will be applied to the macro block of virtual basic unit to 812 (S82) by the partition mode that expansion obtains.
Yet, when the respective fields macro block is extended to twice in vertical direction, may be created on unallowed partition mode (or pattern) in the macro-block partition mode.In order to prevent this situation, EL encoder 20 by following rule according to determining partition mode through the partition mode of expansion.
1) 4 x, 4,8 x 4 of basic unit and 8 of 16 x are confirmed as after expansion by it being amplified in vertical direction 16 of 4 x, 8,8 x 8 that twice obtains and 16 x.
2) 4 x, 8,8 x 8 of basic unit and 16 of 16 x are confirmed as two of the top and bottoms of identical size separately after expansion.Shown in Fig. 8 a, 8 B8_0 of 8 x of basic unit are confirmed as two 88 of x (801).8 B8_0 of 8 x are that what adjoin may not be 8 x, 16 divided block through extension blocks on its left side or the right side in the reason that is not configured to 16 of 8 x after the expansion, and do not have which kind of macro-block partition mode supported in this case.
If respective macroblock has a macro block to encode with internal schema in 810, then EL encoder 20 be not select internal schema but select the top or field, the end macro block of inter mode, and it is carried out above expansion process to determine that macro block in the virtual basic unit is to 812 partition mode.
If corresponding macro block all is an internal schema to two macro blocks in 810, then 20 of EL encoders are carried out inter-layer texture prediction, do not determine and reference key described below and motion vector derivation and do not carry out the partition mode that is undertaken by above expansion process.
The right reference key of macro block for the virtual basic unit that derives from the reference key of respective fields macro block, EL encoder 20 is defined as each reference key among 8 of two 8 x of this top and bottom with the reference key of 8 B8_0 of corresponding 8 x of basic unit, shown in Fig. 8 b, and with the reference key of 8 of determined each 8 x divided by 2 to obtain its final reference key.The reason of making this division is in order to be applied to frame sequence, frame numbers need to be reduced half, being provided with because the number of reference pictures of a macro block is based on the picture that is divided into strange of even summation.
When the frame macro block of the virtual basic unit that derives during to 812 motion vector, the macro block that EL encoder 20 is defined as the motion vector of 4 of corresponding 4 x of basic unit virtual basic unit is to the motion vector of 8 of 4 x in 812, shown in Fig. 8 c, and with determined motion vector after its vertical component multiply by 2 as final motion vector.The reason of making this multiplication is iconic element included in the field macro block iconic element corresponding to two frame macro blocks, thereby makes the size of two field picture be increased to twice in vertical direction.
In the above-described embodiment, for the right movable information of the macro block of predicting virtual basic unit, EL encoder 20 is based on the movable information of the respective fields macro block of basic unit sequentially derive partition mode, reference key and the motion vector of this macro block.
In another embodiment of the present invention, when will be used for the right movable information of the macro block of virtual basic unit of the right inter-layer prediction of current macro when deriving, EL encoder 20 at first obtains the right reference key and the motion vector of macro block of virtual basic unit based on the movable information of the respective fields macro block of basic unit, finally determine the block mode of each macro block of the macro block centering of virtual basic unit then based on the value that is obtained, shown in Fig. 9 a.When partition mode is finally determined, 4 x, 4 module units with the identical motion vector of deriving and reference key are combined, if and the block mode after the combination is the partition mode that allows, then partition mode is arranged to the pattern after this combination, otherwise the pattern before partition mode is arranged to make up.
It below is the more detailed description of the embodiment of Fig. 9 a.As shown in the figure, the inter mode field macro block of basic unit is selected, and the right reference key and the motion vector of frame macro block of the virtual basic unit that uses the motion vector of selected macro block and reference key to derive will to be used for the right motion prediction of current macro.If these two macro blocks all are inter modes, then one of arbitrariness selected (901 or 902) in the top and bottom macro block, and use the motion vector and the reference index information of selected macro block.As shown in the figure, for the reference key of deriving, the analog value that top 8 x of selected macro block 8 is copied into the reference key of 8 of top and bottom 8 x of the top macro block of virtual basic unit, and the analog value of 8 of the ends 8 x of selected macro block is copied into the reference key of 8 of top and bottom 8 x of the end macro block of virtual basic unit.As shown in the figure, for the motion vector of deriving, the analog value that each of selected macro block 4 x 4 is by the motion vector of 4 of corresponding a pair of 4 x that vertically adjoin of the shared macro block centering of doing virtual basic unit.In another embodiment of the present invention, the right movable information of the respective macroblock of basic unit can be mixed and the right motion vector and the reference key of frame macro block of the virtual basic unit that is used to derive, and these are different with the embodiment shown in Fig. 9 a.Fig. 9 b illustrates and is used to derive the program of motion vector and reference key according to this embodiment.The reference key of the sub-piece of the macro block centering of virtual basic unit (8x8 piece and 4 of 4 x) and the related detailed description of copy of motion vector are here omitted because its can be from the illustration of the description of above-mentioned movable information derivation program and Fig. 9 b intuitivism apprehension.
Yet, because the movable information of two macro blocks of the field macro block centering of basic unit all is used in the embodiment of Fig. 9 b, if so the field macro block centering of basic unit has a macro block is internal schema, then utilize movable information as the movable information derivation internal schema macro block of another macro block of inter mode macro block.Particularly, can shown in Fig. 4 b, be copied to after the internal schema macro block constructs the motion vector and reference key of internal schema macro block by corresponding information with the inter mode macro block, or after shown in Fig. 4 c, the internal schema macro block being considered as having the inter mode macro block of 0 motion vector and 0 reference key, or be set to after 0 motion vector and reference index information that the macro block of the virtual basic unit that derives is right shown in Fig. 4 d, being copied to reference key and its motion vector that the internal schema macro block is provided with the internal schema macro block shown in Fig. 9 b by reference key with the inter mode macro block.In case derive right motion vector of the macro block of virtual basic unit and reference index information, just determine the block mode that macro block is right based on the information of being derived as discussed previouslyly.
On the other hand, if two macro blocks of the respective fields macro block centering of basic unit all are internal schemas, then do not carry out motion prediction.
Now inter-layer texture prediction will be described.Fig. 8 d be illustrated in " in the MBAFF frame the field MB-frame MB " situation under exemplary layer between the texture Forecasting Methodology.The respective fields macro block that EL encoder 20 identifies basic unit is to 810 block mode.If corresponding frame macro block is to two macro blocks in 810 or all be internal schema or all be inter mode, then EL encoder 20 converts the respective fields macro block of basic unit to interim frame macro block to 821 to 810, so that or carry out the present frame macro block and carry out its residual prediction (when these two frame macro blocks 810 all are inter mode) to prediction (when these two frame macro blocks 810 all are internal schema) in 813 the base or in the mode that describes below.When two macro blocks of respective macroblock in to 810 all were internal schema, macro block comprised decoded data to 810, and as discussed previously deblocking filter was applied to the frame macro block to 821.Fig. 8 e illustrates and is used for the field macro block the right method of conversion framing macro block.As shown in the figure, the row of a pair of macro block A and B begin from the top of each macro block sequentially by alternate selection (A-〉B-〉A-B-A-...), begin from the top then to arrange to construct a pair of frame macro block A ' and B ' by selected sequence ground.Because be the right row of marshalling yard (MY) macro block again in this way, thus top frame macro block A ' be by this to the row structure of the first half of field macro block A and B, and end frame macro block B ' is by the row structure of the latter half.
On the other hand, if the respective fields macro block of basic unit is an inter mode to having only a macro block in 810, then 813 block mode is selected a piece from the macro block of basic unit to 810, and selected block is used for inter-layer texture prediction according to current frame macro block.Perhaps, before determining that current frame macro block is to 813 block mode, can use each method described below earlier and come inter-layer prediction, can determine that then macro block is to 813 block mode.
Fig. 8 f and 8g illustrate and wherein select the example of a piece with inter-layer prediction.Is under the situation with inter mode coding (perhaps carrying out model prediction therebetween) at current frame macro block to 813, shown in Fig. 8 f, to 810, select inter mode piece 810a from the field macro block of basic unit, and selected is risen sampling in vertical direction to create two respective macroblock 831.Then these two macro blocks 831 are used for current frame macro block to 813 residual prediction.Is not under the situation with inter mode coding (perhaps carrying out its intra mode prediction) at current frame macro block to 813, shown in Fig. 8 g, to 810, select internal schema piece 810b from the field macro block of basic unit, and selected is risen sampling in vertical direction to create two respective macroblock 841.After deblocking filter being applied to these two macro blocks 841, these two macro blocks 841 are used for the present frame macro block to prediction in 813 the base.
Fig. 8 f is selected with one of them piece shown in the 8g and rise sampling and also can be suitable for when each layer has different picture rates to create the right method of macro block that will be used for inter-layer texture prediction.When the picture rate of enhancement layer is higher than the picture rate of basic unit, corresponding picture on some picture in the picture sequence of enhancement layer may have no time in basic unit.In basic unit not free in the enhancement layer picture of corresponding picture the included right inter-layer texture prediction of frame macro block can utilize in the basic unit on the time macro block in the field macro block of coordination on a pair of space in the picture the preceding to carry out.
Fig. 8 h is that the picture rate of enhancement layer is the example of this method under the situation of twice of basic unit's picture rate.
As shown in the figure, the picture rate of enhancement layer is the twice of the picture rate of basic unit.Therefore, having one in per two pictures of enhancement layer---such as the picture of picture order count (POC) for " n2 "---does not have the identical picture of picture order count (POC) in basic unit.Here, the identical consistency of POC on instruction time.
When in the basic unit not free when consistent picture (for example, when current POC is n2), previous picture (promptly, POC is than the picture of current POC low 1) in a pair of space on the field of coordination in the macro block included field, end macro block 802 risen sampling to create interim macro block by vertical to 852 (S82), use this interim macro block to carry out current macro to 815 inter-layer texture prediction then to 852.When in the basic unit free when consistent picture (for example, when current POC is n1), go up that included field, top macro block 801 is risen sampling to create interim macro block to 851 (S82) by vertical in the field macro block of coordination on a pair of space in the consistent picture this time, use this interim macro block to carry out current macro to 814 inter-layer texture prediction then 851.When comprise in by liter interim macro block that sampling is created 851 or 852 from the macro block of internal schema macro block decoding to the time, to this macro block to after using deblocking filter with this macro block to being used for inter-layer texture prediction.
In another embodiment of the present invention, when going up consistent picture if having time in the basic unit (when the current POC in the example of Fig. 8 h is n1), the frame macro block to not being to use the method shown in Fig. 8 h but can be according to the embodiment shown in Fig. 8 d by the field macro block to creating, can use it for inter-layer texture prediction then.In addition, during consistent picture (when the current POC in the example of Fig. 8 h is n2), inter-layer texture prediction can be carried out as Fig. 8 h ground, perhaps can be to the execution of the macro block in current picture inter-layer texture prediction current picture is not free in basic unit.
Correspondingly, embodiments of the invention assignment flag ' field_base_flag (a disjunction mark will) ' is to carry out or carry out according to the method shown in Fig. 8 h according to the method shown in Fig. 8 d with the indication inter-layer texture prediction, and this sign is incorporated in the coded message.For example, texture prediction be according to as the method for Fig. 8 d when carrying out this sign be set to ' 0 ', and when texture prediction be that this sign was set to ' 1 ' when basis was as the method execution of Fig. 8 h.This sign is defined within will be in the macroblock layer in the section head in the parameter sets in the sequential parameter in the sequence parameter set in the enhancement layer that decoder transmits, the expansion of liftable level, parameter sets, the expansion of liftable level, section head, the expansion of liftable level, macroblock layer or the expansion of liftable level.
IV. the field MB-in the picture〉situation of frame MB
In this case, the macro block in anterior layer (EL) is the framing macro block that is encoded, and the macro block that will be used for when the basic unit (BL) of the inter-layer prediction of the frame macro block of anterior layer is the field macro block that is encoded in the picture.Included vision signal composition is identical with the vision signal composition that is included in the macro block of a pair of coordination in the anterior layer in the field macro block in the basic unit.At first, inter-layer motion prediction is described below.
EL encoder 20 uses the partition mode of the partition mode of macro block (the expanding to twice in vertical direction) acquisition among by expansion basic unit occasionally strange as the macro block in the virtual basic unit.Figure 10 a illustrates the specific example of this process.The difference of top in the program shown in Figure 10 a and the MBAFF frame wherein or the program of the selecteed situation III of field, end macro block is to use very naturally the field macro block 1010 of coordination on the space in occasionally strange, and its similar part with the program of situation III is that the field macro block 1010 of coordination is expanded and the partition mode of two macro blocks by the expansion acquisition is applied to the macro block of virtual basic unit to 1012.When respective fields macro block 1010 is extended to twice in vertical direction, may be created on unallowed partition mode (or pattern) in the macro-block partition mode.In order to prevent this situation, EL encoder 20 by with the rule 1 of in situation III, advising) and 2) identical rule is according to determining partition mode through the partition mode of expansion.
If corresponding macro block is to encode by internal schema, then 20 of EL encoders are carried out inter-layer texture prediction, determine and reference key described below and motion vector derivation and do not carry out the partition mode that is undertaken by above expansion process.That is, EL encoder 20 is not carried out inter-layer motion prediction.
Reference key and motion vector derivation program be similar with described in front the situation III also.Yet this situation IV is different from situation III in the following areas.In situation III, because corresponding base layer macro block is carried at the strange macro block centering of even summation in the frame, so one of top and bottom macro block is selected and be applied to the derivation program.In this situation IV, because only there is a macro block in the basic unit corresponding to the current macro of wanting encoding and decoding, so the macro block of virtual basic unit is derived from the movable information of respective fields macro block to 1012 movable information, and the macro block option program shown in Figure 10 b and 10c not, and the movable information of deriving is used to current macro to 1013 inter-layer motion prediction.
Figure 11 schematically shows the right reference key of the macro block of virtual basic unit according to another embodiment of the invention and the derivation of motion vector.In this case, the right movable information of the macro block of virtual basic unit is to derive from the movable information of an occasionally strange macro block of basic unit, and this is with above different with reference to the described situation of figure 9a.The derivation operation identical with the situation of Fig. 9 a is applicable to this situation.Yet the mixing in the situation shown in Fig. 9 b also uses the process of the right movable information of macro block inapplicable in this situation IV, because there is not the top and bottom macro block pairing in the respective fields in basic unit.
In the embodiment that describes with reference to figure 10a to 10c, for the right movable information of the macro block of predicting virtual basic unit, EL encoder 20 is based on the movable information of the respective fields macro block of basic unit sequentially derive partition mode, reference key and motion vector.Yet, in another embodiment of Figure 11, EL encoder 20 is finally determined the right partition mode of macro block of virtual basic unit then at first based on the derive right reference key and the motion vector of macro block of virtual basic unit of the right movable information of the respective macroblock of basic unit based on the value of being derived.When partition mode is determined, 4 x, 4 module units with the identical motion vector of deriving and reference key are combined, if and the block mode after the combination is the partition mode that allows, then partition mode is arranged to the pattern after this combination, otherwise the pattern before partition mode is arranged to make up.
When carrying out texture prediction in the above-described embodiment,, then current macro is carried out prediction encoding and decoding in the base if the respective fields macro block of basic unit is an internal schema.If the respective fields macro block is an inter mode, and if current macro with inter mode coding, then carry out the inter-layer residue prediction encoding and decoding.Here, certainly, the field macro block that uses in prediction is used for texture prediction after it is risen sampling in vertical direction.
In another embodiment of the present invention, to create virtual macro block right with the structure macro block by being included in a field macro block in the strange or idol, then from the macro block that the is constructed movable information right to the macro block of the virtual basic unit that derives.Figure 12 a and Figure 12 b illustrate the example of this embodiment.
In this embodiment, the reference key and the motion vector of the corresponding idol of basic unit (or strange) macro block be copied (1201 and 1202) construct macro block to 1211 to create virtual strange (or idol) macro block, and the macro block that is constructed to the mixed macro block with the virtual basic unit that derives of 1211 movable information to 1212 movable information (1203 and 1204).Mixing and using in the exemplary method of movable information, shown in Figure 12 a and 12b, the macro block that the reference key that top 8 x of respective top macro block 8 is applied to virtual basic unit is to 8 of top 8 x of 1212 top macro block, the reference key that the ends 8 x is 8 is applied to 8 of top 8 x of end macro block, the macro block that the reference key that top 8 x of macro block of the corresponding end 8 is applied to virtual basic unit is to 8 of the ends 8 x of 1212 top macro block, and the reference key of 8 of the ends 8 x is applied to 8 of the x in the ends 8 (1203) of end macro block.Use motion vector (1204) according to reference key.Here omitted the description of this process, because it can be understood intuitively from Figure 12 a and 12b.
In the embodiment shown in Figure 12 a and the 12b, the macro block of virtual basic unit is to use with identical as mentioned above method 1212 partition mode to be determined based on the reference key of deriving and motion vector.
Now inter-layer texture prediction will be described.Figure 10 b illustrates at texture Forecasting Methodology between the exemplary layer of the situation of " the field MB-in the picture〉frame MB ".EL encoder 20 at first rises the respective fields macro block 1010 of sampling basic unit to create two interim macro blocks 1021.If respective fields macro block 1010 is internal schemas, then EL encoder 20 these two interim macro blocks 1021 that deblocking filter is applied to be created are carried out the present frame macro blocks to prediction in 1013 the base based on these two interim macro blocks 1021 then.If respective fields macro block 1010 is inter modes, then EL encoder 20 is carried out the present frame macro blocks to 1013 residual prediction based on these two that are created interim macro blocks 1021.
V. MB-〉situation of a MB
This situation is subdivided into following four kinds of situations, because a macro block is divided into field macro block that comprises in the picture on the scene and the field macro block that is included in the MBAFF frame.
I) basic unit and enhancement layer are the situations of MBAFF frame
This situation is shown in Figure 13 a.As shown in the figure, the movable information that the respective macroblock of basic unit is right (partition mode, reference key and motion vector) is the macro block movable information right to the macro block that is used as virtual basic unit that directly is copied to virtual basic unit by the movable information that respective macroblock is right.Here, movable information is to be copied having between the macro block of same parity.Particularly, the movable information of idol macro block is copied into an idol macro block, and the movable information of a strange macro block is copied into strange macro block, to be configured to the macro block when the virtual level of the motion prediction of the macro block of anterior layer.
When carrying out texture prediction, use the method for the inter-layer texture prediction between the known frame macro block.
Ii) basic unit comprises a picture and enhancement layer comprises the situation of MBAFF frame
This situation is shown in Figure 13 b.As shown in the figure, the movable information of the respective fields macro block of basic unit (partition mode, reference key and motion vector) is the movable information of each macro block that is used as the macro block centering of virtual basic unit by each macro block that the movable information with the respective fields macro block directly is copied to the macro block centering of virtual basic unit.Here, same parity copy rule is inapplicable because the movable information of single macro block be used to top and bottom field macro block both.
When carrying out texture prediction, between enhancement layer with identical (occasionally strange) attribute and base layer macro block, predict (when the corresponding blocks of basic unit is internal schema) in the as fired basis or use residual prediction (when the relevant block of basic unit is inter mode).
Iii) basic unit comprises the MBAFF frame and enhancement layer comprises the situation of a picture
This situation is shown in Figure 13 c.As shown in the figure, from corresponding to the field macro block of selecting to have same parity when the base layer macro block centering of front court macro block, and the field macro block that directly is copied to virtual basic unit by the movable information with institute's selected scenes macro block is with the movable information (partition mode, reference key and motion vector) of the institute's selected scenes macro block movable information as the field macro block of virtual basic unit.
When carrying out texture prediction, between enhancement layer with identical (occasionally strange) attribute and base layer macro block, predict (when the corresponding blocks of basic unit is internal schema) in the as fired basis or use residual prediction (when the relevant block of basic unit is inter mode).
Iv) basic unit and enhancement layer are the situations of a picture
This situation is shown in Figure 13 d.As shown in the figure, the field macro block that directly is copied to virtual basic unit by the movable information with the respective fields macro block of basic unit is with the movable information (partition mode, reference key and motion vector) of the respective fields macro block of the basic unit movable information as the field macro block of virtual basic unit.Equally in this case, movable information is to be copied having between the macro block of same parity.
When carrying out texture prediction, use the method for the inter-layer texture prediction between the known frame macro block.
The description of above inter-layer prediction is to provide at the situation that basic unit and enhancement layer have an equal resolution.Following description is put up with when the resolution of enhancement layer is higher than basic unit's resolution the type of macro block in picture type how (, when SpatialScalabilityType () greater than 0 time) to identify each layer (progressive frame, MBAFF frame, still interlaced field) and/or the picture and is provided according to the type application inter-layer prediction method that identifies.Inter-layer motion prediction is at first described.
M_A). basic unit's (progressive frame)-enhancement layer (MBAFF frame)
Figure 14 a illustrates the processing method at this situation.As shown in the figure, at first, the movable information of all macro blocks of the respective frame in the basic unit is copied to create virtual frames.Carry out then and rise sampling.Rise in the sampling at this, utilize the texture information of basic unit's picture to carry out interpolation with the interpolation rate that the resolution of working as anterior layer equates with the resolution (or being picture size) that allows this picture.In addition, the movable information of each macro block of the picture that is exaggerated by interpolation is based on that the movable information of each macro block of this virtual frames constructs.Can be with a kind of this structure that is used in the multiple known method.The picture of the provisional basic unit that constructs in this way has the identical resolution of picture with current (enhancing) layer.Correspondingly, can use above-mentioned inter-layer motion prediction in this case.
(Figure 14 a), basic unit and the macro block in the picture in the anterior layer are the field macro blocks in frame macro block and the MBAFF frame, comprise the MBAFF frame because basic unit comprises frame when anterior layer in this case.Correspondingly, the method for using above-mentioned situation I is carried out inter-layer motion prediction.Yet not only a macro block is right as mentioned above, and the frame macro block is to also being included in the same MBAFF frame.Correspondingly, corresponding to the macro block in the picture of provisional basic unit right when the right type of anterior layer macro block has been identified to frame macro block (mb) type rather than macro block (mb) type, use the method (frame-frame Forecasting Methodology) of the motion prediction of the known simple copy that comprises movable information between the frame macro block.
M_B). basic unit's (progressive frame)-enhancement layer (interlaced field)
Figure 14 b illustrates the processing method at this situation.As shown in the figure, at first, the movable information of all macro blocks of the respective frame in the basic unit is copied to create virtual frames.Carry out then and rise sampling.Rise in the sampling at this, utilize the texture information of basic unit's picture, carry out interpolation with the interpolation rate that the resolution that allows picture equates with the resolution of working as anterior layer.In addition, the movable information of each macro block of the picture that is exaggerated by interpolation is based on that the movable information of each macro block of the virtual frames of being created constructs.
Use the method for above-mentioned situation II and carry out inter-layer motion prediction because each macro block of the picture of Gou Zao provisional basic unit all is frame macro blocks in this way, and when each macro block of anterior layer all are field macro blocks in the picture.
M_C). basic unit's (MBAFF frame)-enhancement layer (progressive frame)
Figure 14 c illustrates the processing method at this situation.As shown in the figure, at first, the corresponding MBAFF frame transform of basic unit is become progressive frame.The method of above-mentioned situation III is applicable to field macro block with the MBAFF frame to being transformed into progressive frame, and known frame-frame Forecasting Methodology is applicable to the right conversion of frame macro block of MBAFF frame.Certainly, when the method with situation III is applied in this situation, be to utilize the movable information that the data that dope and the difference of the data of the actual layer of wanting encoding and decoding is carried out each macro block of data creation virtual frames that the inter-layer prediction of the operation of encoding and decoding obtains and this frame by not carrying out.
In case the acquisition virtual frames is just carried out this virtual frames and is risen sampling.Rise in the sampling at this, carry out interpolation with the interpolation rate that the resolution that allows basic unit equates with the resolution of working as anterior layer.In addition, utilize the movable information of each macro block of movable information structure through amplifying picture of a kind of each macro block based on virtual frames in the multiple known method.Here, carry out known frame macro block-macro block inter-layer motion prediction method because each macro block of the picture of the provisional basic unit that constructs in this way all is frame macro blocks, and when each macro block of anterior layer all are frame macro blocks.
M_D). basic unit's (interlaced field)-enhancement layer (progressive frame)
Figure 14 d illustrates a kind of processing method at this situation.In this case, the type of picture is identical with the type of the macro block of this picture.As shown in the figure, at first, the respective fields of basic unit is transformed into progressive frame.The frame that conversion goes out has and compares when the identical vertical/horizontal (in length and breadth) of the picture of anterior layer.The method that rises sampling process and above-mentioned situation IV is applicable to interlaced field is transformed into progressive frame.Certainly, when the method with situation IV is applied in this situation, be to utilize by not carrying out the data that dope and the difference of the data of the actual layer of wanting encoding and decoding are carried out the movable information that data that the inter-layer prediction of the operation of encoding and decoding obtains are created each macro block of the data texturing of virtual frames and this frame.
In case the acquisition virtual frames is just carried out this virtual frames and is risen sampling.Rise in the sampling at this, carry out interpolation equals to work as anterior layer with the resolution that allows virtual frames resolution.In addition, use the movable information of each macro block of the picture that the movable information structure interpolation of a kind of each macro block based on virtual frames in the multiple known method goes out.Here be to carry out known frame macro block-macro block inter-layer motion prediction method, because each macro block of the picture of Gou Zao provisional basic unit all is frame macro blocks in this way, and each macro block of working as anterior layer is the frame macro block.
Figure 14 e illustrate according to another embodiment of the invention at above situation M_D) processing method.As shown in the figure, this embodiment is transformed into progressive frame with strange or even respective fields.For interlaced field is transformed into progressive frame, shown in Figure 14 d, use the method that rises sampling and above-mentioned situation IV.In case acquisition virtual frames, motion prediction between---its in the multiple known method a kind of---picture of work as anterior layer of just virtual frames being used the method for the motion prediction between the picture with same aspect ratio and the provisional layer is worked as the prediction encoding and decoding of movable information of each macro block of the picture line by line of anterior layer with execution.
The difference of the method for the method shown in Figure 14 e and Figure 14 d is not generate interim prediction signal.
Figure 14 f illustrate according to another embodiment of the invention at above situation M_D) processing method.As shown in the figure, this embodiment copies the movable information of all macro blocks of respective fields of basic unit to create virtual screen.Carry out then and rise sampling.Rise in the sampling at this, use the texture information of the picture of basic unit, and different interpolation rates is used for vertical and horizontal interpolation so that the picture through amplifying has and the identical size (or being resolution) of picture of working as anterior layer.In addition, a kind of (for example, the expanding special liftable level (ESS)) in the multiple known Forecasting Methodology can be applied to virtual screen with structure through amplifying the various syntactic informations and the movable information of picture.The motion vector that constructs in this process is expanded according to magnification ratio.In case provisional basic unit is configured out through rising sampled picture, just this picture is used for carrying out inter-layer motion prediction, with the movable information of encoding and decoding when each macro block of the picture of anterior layer when each macro block of the picture of anterior layer.Here, use known frame macro block=macro block inter-layer motion prediction method.
Figure 14 g illustrate according to another embodiment of the invention at above situation M_D) processing method.As shown in the figure, this embodiment at first copies the movable information of all macro blocks of respective fields of basic unit to create virtual screen.Afterwards, the texture information of the picture of use basic unit is to carry out interpolation for the different ratio of vertical and horizontal interpolation.The texture information of creating by this operation is used to inter-layer texture prediction.In addition, the movable information in the virtual screen is used to carry out the inter-layer motion prediction of each macro block in the picture of anterior layer.Here, a kind of (for example, the special liftable level of expansion (ESS) of definition in associating liftable level video model (JSVM)) used in the multiple known method carried out the motion prediction encoding and decoding when the picture of anterior layer.
The difference of the method for the method shown in Figure 14 g and Figure 14 f is not generate interim prediction signal.
M_E). basic unit's (MBAFF frame)-enhancement layer (MBAFF frame)
Figure 14 h illustrates the processing method at this situation.As shown in the figure, at first, the corresponding MBAFF frame transform of basic unit is become progressive frame.For the MBAFF frame transform is become progressive frame, the method for above-mentioned situation III is applicable to the right conversion of field macro block of MBAFF frame, and frame-frame Forecasting Methodology is applicable to the right conversion of frame macro block of MBAFF frame.Certainly, when the method with situation III is applied in this situation, be to utilize the data that obtain by the inter-layer prediction that need not carry out data that encoding and decoding dope and the operation of the difference of the data of the actual layer of wanting encoding and decoding to create the movable information of each macro block of virtual frames and this frame.
In case the acquisition virtual frames is just carried out this virtual frames and is risen sampling.Rise in the sampling at this, carry out interpolation with the interpolation rate that the resolution that allows basic unit equates with the resolution of working as anterior layer.In addition, utilize the movable information of each macro block of the picture of movable information structure through amplifying of a kind of each macro block based on virtual frames in the multiple known method.Use the method for above-mentioned situation I and carry out inter-layer motion prediction because each macro block of the picture of Gou Zao provisional basic unit all is frame macro blocks in this way, and when each macro block of anterior layer all are field macro blocks in the MBAFF frame.Yet not only a macro block is right as mentioned above, and the frame macro block is to also being included in the same MBAFF frame.Correspondingly, corresponding to the macro block in the picture of provisional basic unit right when the anterior layer macro block when being frame macro block rather than macro block, use the method (frame-frame Forecasting Methodology) of the motion prediction of the known copy that comprises movable information between the frame macro block.
M_F). basic unit's (MBFF frame)-enhancement layer (interlaced field)
Figure 14 i illustrates the processing method of this situation.As shown in the figure, at first, the corresponding MBAFF frame transform of basic unit is become progressive frame.For the MBAFF frame transform is become progressive frame, the method for above-mentioned situation III is applicable to the right conversion of field macro block of MBAFF frame, and frame-frame Forecasting Methodology is applicable to the right conversion of frame macro block of MBAFF frame.Certainly, equally, when the method with situation III is applied in this situation, be to utilize the data that obtain by the inter-layer prediction that need not carry out data that encoding and decoding dope and the operation of the difference of the data of the actual layer of wanting encoding and decoding to create the movable information of each macro block of virtual frames and this frame.
In case obtain virtual frames, just to allow resolution to equal this virtual frames to be carried out interpolation when the interpolation rate of the resolution of anterior layer.In addition, use the movable information of each macro block of the picture of movable information structure through amplifying of a kind of each macro block based on virtual frames in the multiple known method.Use the method for above-mentioned situation II and carry out inter-layer motion prediction because each macro block of the picture of Gou Zao provisional basic unit all is frame macro blocks in this way, and when each macro block of anterior layer all are field macro blocks in occasionally strange.
M_G). basic unit's (interlaced field)-enhancement layer (MBAFF frame)
Figure 14 j illustrates the processing method at this situation.As shown in the figure, at first, the interlaced field of basic unit is transformed into progressive frame.Use the method that rises sampling and above-mentioned situation IV interlaced field is transformed into progressive frame.Certainly, equally, when the method with situation IV is applied in this situation, be to utilize by not carrying out the data that dope and the difference of the data of the actual layer of wanting encoding and decoding are carried out the movable information that data that the inter-layer prediction of the operation of encoding and decoding obtains are created each macro block of virtual frames and this frame.
In case the acquisition virtual frames is just carried out to rise to this virtual frames and is sampled to allow resolution to equal to work as the resolution of anterior layer.In addition, utilize the movable information of each macro block of the picture of a kind of structure in the multiple known method through amplifying.Use the method for above-mentioned situation I and carry out inter-layer motion prediction because each macro block of the picture of Gou Zao provisional basic unit all is frame macro blocks in this way, and when each macro block of anterior layer all are field macro blocks in the MBAFF frame.Yet not only a macro block is right as mentioned above, and the frame macro block is to also being included in the same MBAFF frame.Therefore, corresponding to the macro block in the picture of provisional basic unit right when the anterior layer macro block when comprising frame macro block rather than macro block, use the method (frame-frame Forecasting Methodology) of the known motion prediction between the frame macro block rather than the Forecasting Methodology of above-mentioned situation I.
M_H). basic unit's (interlaced field)-enhancement layer (interlaced field)
Figure 14 k illustrates the processing method at this situation.As shown in the figure, at first, the movable information of all macro blocks of the respective fields in the basic unit is copied to create virtual field, this virtual field is carried out rising sampling then.This rises sampling and carries out with the sample rate that rises that the resolution that allows basic unit equates with the resolution of working as anterior layer.In addition, use the movable information of each macro block of the picture of movable information structure through amplifying of a kind of each macro block based on the virtual frames of being created in the multiple known method.The situation method of using among the above-mentioned situation V is iv) carried out inter-layer motion prediction, because each macro block of the picture of Gou Zao provisional basic unit all is the field macro blocks in the picture in this way, and when each macro block of anterior layer also all are field macro blocks in the picture.
Although be to use the texture information of the picture of the texture information of the virtual field of provisional layer or frame rather than basic unit to carry out rising sampling in the description of the embodiment of Figure 14 a to 14k, the texture information of basic unit's picture also can be used for rising sampling.In addition, if not necessity, when will be used for the movable information of picture of provisional layer of the inter-layer motion prediction carried out in following stages when deriving, can omit the interpolation process that utilizes texture information in above-mentioned rising in the sampling process.
On the other hand, although the description of texture prediction is to provide at the situation that basic unit and enhancement layer have a same spatial resolution, these two layers may have different spatial resolutions as mentioned above.Be higher than in the resolution of enhancement layer under the situation of resolution of basic unit, at first, execution makes the resolution of the picture of basic unit equal the operation of resolution of the picture of enhancement layer, the basic unit's picture that has the resolution identical with establishment with the resolution of enhancement layer, and based on each macro block in this picture select with above-mentioned situation I-V in the corresponding texture prediction method of each situation to carry out the prediction encoding and decoding.Describe the program that the resolution that makes basic unit's picture equals the resolution of enhancement layer picture now in detail.
When considering to be used for inter-layer prediction two-layer, the combined number that is used for the picture format (line by line and interlaced format) of encoding and decoding between two-layer is 4, because two kinds of vision signal scan methods are arranged, a kind of is to line by line scan and another kind is interlacing scan.Therefore, will increase the resolution of basic unit's picture with texture forecast method between execution level at each description in these four kinds of situations respectively.
T_A). enhancement layer be line by line and basic unit is the situation of interlacing
Figure 15 a illustrates the embodiment that basic unit's picture is used for the method for inter-layer texture prediction at this situation.As shown in the figure, be included in strange of the even summation of different time output on the time corresponding to basic unit's picture 1501 of picture 1500 of current (enhancing) layer.Therefore, at first, EL encoder 20 is divided into even summation strange (S151) with the picture of basic unit.The internal schema macro block of basic unit's picture 1501 has the raw image data that is not encoded (or decoded view data) that is used for intra mode prediction, and macro block has the encoded residual error data (or the residual error data through decoding) that is used for residual prediction between its pattern.When describing texture prediction hereinafter, like this equally for base layer macro block or picture.
After respective picture 1501 was divided into field component, EL encoder 20 was carried out the interpolation of isolated 1501a and 1501b on vertical and/or horizontal direction, to create strange picture 1502a of even summation and the 1502b (S152) through amplifying.This interpolation is used a kind of in the multiple known method, such as 6 tap filtering and binary system linear filtering.Be used for increasing the vertical and level of resolution (that is size) of picture than the vertical and level ratio of the size that equals enhancement layer picture 1500 with the size of basic unit picture 1501 by interpolation.Vertical and level ratio can be equal to each other.For example, if the resolution between enhancement layer and the basic unit is 2, then strange 1501a of isolated even summation and 1501b are carried out interpolation, to create a pixel on vertical and the horizontal direction between each pixel in each again.
In case interpolation is finished, then make up strange 1502a of even summation through amplifying and 1502b with structure picture 1503 (S153).In this combination, alternately select strange 1502a of even summation through amplifying and 1502b row (1502a-〉1502b-〉1502a-1502b-..) then with it by the picture 1503 of order layout of selecting to construct combination.Here, the block mode of each macro block in the picture of determining to make up 1503.For example, the block mode of the macro block of the picture 1503 of combination is confirmed as equating with the block mode of macro block in the basic unit's picture 1501 that comprises the zone with identical image composition.This definite method can be applicable in any situation of the picture through amplifying described below.Because Gou Zao combined picture 1503 has the spatial resolution identical with the current picture 1500 of enhancement layer in this way, so the texture prediction (for example, between frame-frame macro block texture prediction) of carrying out the macro block in the current picture line by line 1500 based on the respective macroblock of combined picture 1503 (S154).
Figure 15 b illustrates the method for using basic unit's picture in inter-layer texture prediction according to another embodiment of the invention.As shown in the figure, this embodiment is absent from the scene and separates basic unit's picture on the basis of attribute (parity), but on vertical and/or horizontal direction, directly carry out the interpolation (S155) of basic unit's picture of strange of the even summation that is included in different time output, with the resolution (that is size) of structure resolution and enhancement layer picture identical through the amplification picture.Structure is used to carry out the inter-layer texture prediction (S156) of the current picture line by line of enhancement layer through amplifying picture in this way.
Figure 15 a separates the program that the picture with strange of even summation comes it is carried out interpolation on illustrating by the basis of attribute on the scene on the picture rank.Yet EL encoder 20 can be reached and identical result shown in Figure 15 a by the program shown in the execution graph 15a on the macro block rank.More specifically, the vision signal of the strange field component of even summation that when the basic unit with strange of even summation is when encode by MBAFF, it is right vertically to adjoin macro block in the picture 1501---itself and be subjected to macro block in the enhancement layer picture of texture prediction encoding and decoding at present to coordination---can comprise as among Figure 16 a or the 16b.Figure 16 a illustrates the frame MB that interweaves in the strange field component of even summation each macro block in a pair of macro block A and B wherein to pattern, and each macro block that Figure 16 b illustrates among wherein a pair of macro block A and the B comprises that the field MB of the video line with homologous field attribute is to pattern.
Under the situation of Figure 16 a, for the method shown in the application drawing 15a, select this idol row to construct an idol piece A ' to each macro block among macro block A and the B, and select its strange row to construct strange piece B ', thereby with the macro block that all is intertwined with the strange field component of even summation in each macro block to being divided into two piece A ' and the B ' that has the strange field component of even summation respectively.To among isolated two macro block A ' and the B ' in this way each carry out interpolation with structure through amplifying piece.Utilization in amplifying piece with the current enhancement layer picture that will be subjected to the texture prediction encoding and decoding in data in the corresponding zone of macro block of intra_BL (in the basic unit) or residual_prediction (residual prediction) pattern carry out texture prediction.Although it is not shown among Figure 16 a, but but on the attribute partly on the scene basis combination among all piece structural map 15a that amplifies individually through amplifying strange picture 1502a of even summation and 1502b, so can construct by the operation above every pair of macro block is repeated among Figure 15 a through amplifying strange picture 1502a of even summation and 1502b.
Cutting apart macro block based on the field attribute under with the situation of constructing each macro block as Figure 16 b, above-mentioned separable programming is to copying the process of each macro block with the macro block of constructing two separation simply from this macro block.Follow-up program is similar to the described program with reference to figure 16a.
T_B). enhancement layer be interlacing and basic unit is a situation line by line
Figure 17 a illustrates the embodiment that basic unit's picture is used for the method for inter-layer texture prediction at this situation.As shown in the figure, at first, EL encoder 20 is for constructing two pictures (S171) when anterior layer picture 1700.In the exemplary method of two pictures of application construction, select the idol row of respective picture 1701 to construct a picture 1701a, and select its strange row to construct another picture 1701b.EL encoder 20 is carried out the interpolation of two picture 1701a so constructing and 1701b then to create two picture 1702a and 1702b (S172) through amplifying on vertical and/or horizontal direction.This interpolation is used a kind of in the multiple known method, such as situation T_A) in 6 tap filtering and binary system linear filtering.The ratio that is used for increasing resolution also with situation T_A) describe those are identical.
In case interpolation is finished, just make up these two through amplifying field 1702a and 1702b with structure picture 1703 (S173).In this combination, alternately select these two through amplifying field 1702a and the row of 1702b (1702a-〉1702b-〉1702a-1702b-...) then with it by the picture 1703 of the order layout of selecting with tectonic association.Because the combined picture that constructs in this way 1703 has the spatial resolution identical with the current picture 1700 of enhancement layer, so the texture prediction (for example, between frame-frame macro block texture prediction or with reference to the texture prediction of figure 4g description) of carrying out the macro block in the current interlaced picture 1700 based on the respective macroblock of picture 1703 of combination (S174).
Figure 17 b illustrates the method for using basic unit's picture in inter-layer texture prediction according to another embodiment of the invention.As shown in the figure, this embodiment is not divided into basic unit's picture two pictures, but directly carries out the interpolation (S175) of basic unit's picture on vertical and/or horizontal direction, with structure resolution identical with enhancement layer screen resolution (that is size) through the amplification picture.Structure is used to carry out the inter-layer texture prediction (S176) of the current interlaced picture of enhancement layer through amplifying picture in this way.
Although the description of Figure 17 a also provides on the picture rank, EL encoder 20 can be as above situation T_A) described on the macro block rank, carry out the picture separation process.When single picture 1701 is considered as vertically adjoining macro block to the time, the method for Figure 17 b is similar to separation shown in Figure 17 a and interpolator.Here omitted the detailed description of this program, because it can be understood intuitively from Figure 17 a.
T_C). enhancement layer and the both of basic unit are the situations of interlacing
Figure 18 illustrates the embodiment that basic unit's picture is used for the method for inter-layer texture prediction at this situation.In this case, as shown in the figure, EL encoder 20 with situation T_A) in identical mode will go up the time corresponding to basic unit's picture 1801 and be divided into even summation strange (S181) when anterior layer picture 1800.EL encoder 20 is carried out the interpolation of isolated 1801a and 1801b then to create strange picture 1802a of even summation and the 1802b (S182) through amplifying on vertical and/or horizontal direction.EL encoder 20 makes up the strange 1802a of even summation through amplifying and 1802b then with structure picture 1803 (S182).The inter-layer texture prediction (for example, between frame-frame macro block texture prediction or with reference to the texture prediction of figure 4g description) that EL encoder 20 is carried out macro block in the current interlaced picture 1800 (the frame macro block of MBAFF coding to) based on the respective macroblock of combined picture 1803 then (S184).
Although two layers have identical picture format, but separate picture 1801 (S181) of basic unit on the basis of EL encoder 20 attributes on the scene and amplify isolated (S182) individually and make up picture (S183) then through amplifying, this is because if make up very picture 1801 direct interpolation of quilt when the characteristic that its vision signal with strange of even summation alters a great deal of field of even summation, then the interlaced picture 1800 of strange of the even summation that interweaves with having of enhancement layer is compared, picture through amplifying may have the image (for example, having the image that stretches the border) of distortion.Correspondingly, even two layers are interlacing all, according to the present invention, EL encoder 20 also separates the back on the basis of basic unit's picture attribute on the scene and uses it to obtain two fields, and individual amplifies this two fields, makes up the field through amplifying then.
Certainly, can not be when the picture of two layers all is interlacing, always to be to use method shown in Figure 180, but the video properties that replaces according to picture selectively use this method.
Figure 18 is illustrating according to the present invention the program of separating on the attribute on the scene basis and amplifying the picture with strange of even summation on the picture rank.Yet, as above T_A) described in, EL encoder 20 can be reached result same as shown in Figure 18 by carry out program shown in Figure 180 on the macro block rank, it comprise with reference to figure 16a and 16b describe based on macro block separate and interpolation process (particularly be with the frame macro block to the piece that is divided into the strange row of even summation and amplify isolated individually) and combination and inter-layer texture prediction process (particularly be the row of alternately selecting piece constructing a pair of piece, and utilize the piece that is constructed) to carrying out when the right texture prediction of the frame macro block of anterior layer through amplifying through amplifying through amplifying.
T_D). enhancement layer and the both of basic unit are situations line by line
In this case, basic unit's picture is amplified to the size identical with the enhancement layer picture, and will picture be used to have the inter-layer texture prediction of the current enhancement layer picture of same frame form through amplifying.
Although below described the embodiment of the texture prediction when having identical temporal resolution in basic unit with enhancement layer, two layers may have different temporal resolutions, that is, and and different picture rates.If even all layer picture also is different picture scan type when all layers have identical temporal resolution, then these pictures may comprise the vision signal with different output times, even they are pictures (that is the picture that corresponds to each other on the time) of identical POC.Inter-layer texture prediction method in this case will be described now.In the following description, suppose that two layers have identical spatial resolution at first.If two-layer have different spatial resolutions, then as described above at liter each picture of sampling basic unit so that spatial resolution equal to use method described below again after the resolution of enhancement layer.
A) enhancement layer comprises that progressive frame, basic unit comprise that the temporal resolution of MBAFF frame and enhancement layer reaches the high situation of twice
Figure 19 a illustrates the inter-layer texture prediction method at this situation.As shown in the figure, each MBAFF frame of basic unit comprises strange of the even summation with different output times, so EL encoder 20 is divided into even summation strange (S191) with each MABFF frame.EL encoder 20 is divided into idol field and strange field respectively with the even field component (for example, idol is gone) and the strange field component (for example, strange going) of each MBAFF frame.After in this way the MBAFF frame being divided into two fields, EL encoder 20 in vertical direction interpolation each so that it has the high resolution (S192) that reaches twice.This interpolation is used a kind of in the multiple known method, such as 6 tap filtering, binary system linear filtering and the capable zero padding of sample.In case interpolation is finished, each frame of enhancement layer in basic unit just free consistent picture, so the macro block of every frame of 20 pairs of enhancement layers of EL encoder is carried out known inter-layer texture prediction (for example, predicting between frame-frame macro block) (S193).
Also above program can be applied to inter-layer motion prediction.Here, when the MBAFF frame was divided into two fields, the movable information of each macro block of the field macro block centering in the EL encoder 20 copy MBAFF frames was as the movable information with macro block of homologous field attribute (parity), to use it for inter-layer motion prediction.Even in basic unit, not free during consistent picture (under the situation of t1, t3...), use this method also can go up consistent picture to carry out inter-layer motion prediction according to the said method time of creating out.
When one of two-layer resolution is twice high of resolution of another layer as in the example of Figure 19 a and even be that N doubly all can directly use above-mentioned method during (three times or more than) high at it.For example, when resolution is three times high, one of these two isolated can be copied in addition with structure and use three fields, and when resolution is four times high, in these two isolated each can be copied again to construct and to use four fields.Obviously, under the situation that any time differences in resolution is arranged, those skilled in the art need not any creative thinking just can come inter-layer prediction by using principle of the present invention simply.Therefore, being used for of not describing in this specification falls within the scope of the present invention naturally in any method that prediction between the layer of different time resolution is being arranged.Other situations described below are like this equally.
If base layer encoder has been become frame self-adaptive field and frame (PAFF) rather than MBAFF frame, then two-layer may have identical temporal resolution as among Figure 19 b.Therefore, in this case, frame is being divided into two process and directly frame is being carried out interpolation and construct and have and behind the picture of the identical temporal resolution of anterior layer, carry out inter-layer texture prediction again by need not to carry out.
B) enhancement layer comprises that MBAFF frame, basic unit comprise that the temporal resolution of progressive frame and enhancement layer is half a situation of basic unit
Figure 20 illustrates the inter-layer texture prediction method at this situation.As shown in the figure, each MBAFF frame of enhancement layer comprises strange of the even summation with different output times, so EL encoder 20 is divided into even summation strange (S201) with each MABFF frame.EL encoder 20 is divided into idol field and strange field respectively with the even field component (for example, idol is gone) and the strange field component (for example, strange going) of each MBAFF frame.EL encoder 20 is carried out the sub sampling of each frame of basic unit in vertical direction with the picture (S202) that reduces by half of structure resolution.This sub sampling can use a kind of in capable sub sampling or various other known down-sampled method, in the example of Figure 20, EL encoder 20 selects to have the idol row of picture (picture t0, t2, t4...) of even picture index obtaining half-sized picture, and the strange row of picture (picture t1, t3...) of selecting to have strange picture index is to obtain half-sized picture.Also can carry out frame in reverse order and separate (S201) and sub sampling (S202).
In case finish this two process S201 and S202, from the frame of enhancement layer isolated 2001 picture consistent with field 2001 and that have the spatial resolution identical with field 2001 was just arranged basic unit on time, EL encoder 20 is carried out known inter-layer texture prediction (for example, predicting between frame-frame macro block) (S203) to the macro block in each thus.
Also above program can be applied to inter-layer motion prediction.Here, when by sub sampling when each frame of basic unit obtains the picture (S202) that size reduces, EL encoder 20 can be according to the method that is fit to (for example, adopt the method for the movable information of the piece do not divided fully) from the movable information of each macro block of vertically adjoining macro block centering, obtain the movable information of respective macroblock, the movable information that is obtained can be used for inter-layer motion prediction then.
In this case, the picture of enhancement layer is encoded so that transmit by PAFF, because inter-layer prediction is to carrying out from isolated each picture 2001 of MBAFF frame.
C) enhancement layer comprises that MBAFF frame, basic unit comprise progressive frame and two-layer situation with identical temporal resolution
Figure 21 illustrates the inter-layer texture prediction method at this situation.As shown in the figure, each MBAFF frame of enhancement layer comprises strange of the even summation with different output times, so EL encoder 20 is divided into even summation strange (S211) with each MABFF frame.EL encoder 20 is divided into idol field and strange field respectively with the even field component (for example, idol is gone) and the strange field component (for example, strange going) of each MBAFF frame.EL encoder 20 is carried out the sub sampling of each frame of basic unit in vertical direction with the picture (S212) that reduces by half of structure resolution.This sub sampling can use a kind of in capable sub sampling or various other known down-sampled method.Also can carry out frame in reverse order and separate (S211) and sub sampling (S212).
EL encoder 20 also can be constructed (for example, an idol picture) by the MBAFF frame, rather than the MBAFF frame is divided into two fields.This is to have identical temporal resolution because of two-layer, so only has one (rather than both all) to have the respective frame that can be used for inter-layer prediction from a frame in isolated two field pictures in basic unit.
In case finish this two process S211 and S212, EL encoder 20 is just carried out known inter-layer texture prediction (for example, predicting between frame-frame macro block) (S213) based on corresponding picture through sub sampling in the basic unit to only idol (very) field in the isolated field from the frame of enhancement layer.
Equally in this case, can by with situation b) described identical mode carries out inter-layer motion prediction to isolated of the enhancement layer of carrying out inter-layer texture prediction for it.
Provide although above description is the inter-layer prediction operation of just being carried out by the EL encoder 20 of Fig. 2 a or 2b, all descriptions of inter-layer prediction operation can jointly be applicable to from basic unit and receive through the information of decoding and the EL decoder of decoding enhancement layer stream.In the Code And Decode program, above-mentioned inter-layer prediction operation (comprise and be used for separating, amplifies and the operation of the vision signal of combined picture or macro block) is carried out in an identical manner, but inter-layer prediction operation is afterwards carried out in a different manner.The example of this difference is: after carrying out motion and texture prediction, difference information between the information that encoder encodes dopes or information that dopes and the actual information, so that send it to decoder, and decoder by will by the information of carrying out with identical interlayer motion and texture prediction in the execution of encoder place obtains directly apply to current macro or by other use actual reception to the macro block coding/decoding information obtain actual motion information and texture information.Above details of describing from the coding angle of the present invention and principle are directly applied for the decoder of the two-layer data flow that decoding receives.
Yet, when the EL encoder transmitted the enhancement layer with MBAFF frame in the PAFF mode after as described in reference Figure 20 and 21 enhancement layer being separated into a sequence and inter-layer prediction, decoder was not carried out the above-mentioned program that the MBAFF frame is divided into a picture to the current layer that receives.
In addition, to go out to identify EL encoder 20 be shown in Fig. 8 d or carried out the sign ' field_base_flag ' of the inter-layer texture prediction between the macro block shown in Fig. 8 h to decoder from receiving signal decoding then.Based on the value of statistical indicant that decodes, decoder determines that it still is to carry out that the prediction between the macro block is carried out shown in Fig. 8 h shown in Fig. 8 d, and determine to obtain texture prediction information according to this.If do not receive sign ' field_base_flag ', then EL decoder hypothesis has received the sign with " 0 " value.That is, the texture prediction between the EL decoder hypothesis macro block is to carry out according to the method shown in Fig. 8 d, and it is right to rebuild current macro or macro block to obtain the right information of forecasting of current macro.
Have at least among the above-mentioned limited embodiment of the present invention one in addition can be when using the video signal source of different-format (or pattern) inter-layer prediction.Therefore, when a plurality of layers of encoding and decoding, can improve the digital coding rate, and be not limited to the picture type of vision signal, such as interlace signal, progressive signal, MBAFF frame picture and a picture.In addition, when being the interlaced video signal source, the image of the picture that uses in the prediction can be configured to more be similar to the original image that is used to predict encoding and decoding, thereby improves the digital coding rate when one of two-layer.
Although described the present invention with reference to preferred embodiment, it should be apparent to those skilled in the art that and to carry out various improvement, modification, replacement in the present invention and increase and can not depart from the scope of the present invention and spirit.Therefore, the present invention is intended to contain improvement of the present invention, modification, replacement and increases, as long as they drop in the scope of claims and equivalents thereof.

Claims (16)

1. one kind is used at coding or has the method for inter-layer prediction between the layer of different spatial resolutions during decoded video signal, said method comprising the steps of:
Identify lower floor and upper strata picture type and/or be included in the type of the piece in the described picture;
Result according to described sign uses the virtual screen that has the aspect ratio identical with the described picture on described upper strata with structure from single the right method of macroblock prediction frame macro block to the picture of described lower floor;
Rise the described virtual screen of sampling; And
Utilize described virtual screen inter-layer motion prediction to be applied to the picture on described upper strata through rising sampling.
2. the method for claim 1 is characterized in that, the described sampling step that rises comprises that liter described virtual screen of sampling is so that it has the resolution identical with the picture on described upper strata.
3. the method for claim 1 is characterized in that, the aspect ratio of the aspect ratio of the picture of described lower floor and the picture on described upper strata is different.
4. one kind is used at coding or has the method for inter-layer prediction between the layer of different spatial resolutions during decoded video signal, said method comprising the steps of:
The type of the picture on sign lower floor and upper strata and/or be included in the type of the piece in the described picture;
Result according to described sign uses the virtual screen that has the aspect ratio identical with the picture on described upper strata with structure from single the right method of macroblock prediction frame macro block to the picture of described lower floor; And
Utilize the described virtual screen that constructs that the picture on described upper strata is used inter-layer motion prediction.
5. method as claimed in claim 4 is characterized in that, described step to the upper layer application inter-layer motion prediction is included in and rises the virtual screen that sampling is constructed in the inter-layer prediction.
6. method as claimed in claim 5 is characterized in that, described liter sampling vertically with on the horizontal direction is being carried out with the identical sample rate that rises.
7. method as claimed in claim 4 is characterized in that, the aspect ratio of the aspect ratio of the picture of described lower floor and the picture on described upper strata is different.
8. one kind is used for said method comprising the steps of in the coding or the method for inter-layer prediction during decoded video signal:
Identify the type of the picture on lower floor and upper strata;
If the type of the picture of described lower floor is and the type of the picture on described upper strata is line by line, then copy the piece of the movable information of the piece in the picture of described lower floor with the constructing virtual picture;
Rise the described virtual screen of sampling; And
Described through rising sampling virtual screen and the picture on described upper strata between application of frame macro block-macroblock motion prediction.
9. method as claimed in claim 8 is characterized in that, the described sampling step that rises vertically with on the horizontal direction is being carried out with the different sample rates that rises.
10. method as claimed in claim 8 is characterized in that, the described sampling step that rises comprises the described virtual screen of amplification so that it has the resolution identical with the picture on described upper strata, and expands the motion vector of described virtual screen.
11. method as claimed in claim 8 is characterized in that, the aspect ratio of the aspect ratio of the picture of described lower floor and the picture on described upper strata is different.
12. a method that is used for inter-layer prediction when encoding/decoding video signal said method comprising the steps of:
Identify the type of picture on lower floor and upper strata and/or the type of the macro block in the described picture;
If the type of the macro block in the type of the picture of described lower floor or the picture of described lower floor is and the picture on the type of the picture on described upper strata or described upper strata in the type of macro block be line by line, then copy the piece of the movable information of the piece in the picture of described lower floor with the constructing virtual picture; And
The movable information that the described copy of the piece of described virtual screen is come is used for the inter-layer motion prediction to the macro block of the picture on described upper strata.
13. method as claimed in claim 12 is characterized in that, the step of the piece of described constructing virtual picture comprises that structure has the virtual screen of the copy information of the texture information of picture of described lower floor and movable information.
14. method as claimed in claim 12 is characterized in that, the step of the movable information that described use copy comes is included in and rises the described virtual screen of sampling in the inter-layer prediction.
15. method as claimed in claim 14 is characterized in that, described liter sampling vertically with on the horizontal direction is being carried out with the different sample rates that rises.
16. method as claimed in claim 12 is characterized in that, the aspect ratio of the aspect ratio of the picture of described lower floor and the picture on described upper strata is different.
CN200780005672XA 2006-01-09 2007-01-09 Inter-layer prediction method for video signal Active CN101385352B (en)

Priority Applications (1)

Application Number Priority Date Filing Date Title
CN201210585882.3A CN103096078B (en) 2006-01-09 2007-01-09 For inter-layer prediction method and the device of vision signal

Applications Claiming Priority (31)

Application Number Priority Date Filing Date Title
US75700906P 2006-01-09 2006-01-09
US60/757,009 2006-01-09
US75823506P 2006-01-12 2006-01-12
US60/758,235 2006-01-12
US77693506P 2006-02-28 2006-02-28
US60/776,935 2006-02-28
US78339506P 2006-03-20 2006-03-20
US60/783,395 2006-03-20
US60/786,741 2006-03-29
US60/787,496 2006-03-31
US60/816,340 2006-06-26
US60/830,600 2006-07-14
KR1020060111897A KR20070074453A (en) 2006-01-09 2006-11-13 Method for encoding and decoding video signal
KR1020060111893 2006-11-13
KR10-2006-0111895 2006-11-13
KR10-2006-0111897 2006-11-13
KR1020060111895A KR20070074452A (en) 2006-01-09 2006-11-13 Inter-layer prediction method for video signal
KR1020060111893A KR20070075257A (en) 2006-01-12 2006-11-13 Inter-layer motion prediction method for video signal
KR1020060111897 2006-11-13
KR1020060111894 2006-11-13
KR10-2006-0111893 2006-11-13
KR10-2006-0111894 2006-11-13
KR1020060111894A KR20070074451A (en) 2006-01-09 2006-11-13 Method for using video signals of a baselayer for interlayer prediction
KR1020060111895 2006-11-13
KR1020070001587 2007-01-05
KR10-2007-0001582 2007-01-05
KR1020070001587A KR20070075293A (en) 2006-01-12 2007-01-05 Inter-layer motion prediction method for video signal
KR1020070001582A KR20070095180A (en) 2006-03-20 2007-01-05 Inter-layer prediction method for video signal based on picture types
KR10-2007-0001587 2007-01-05
KR1020070001582 2007-01-05
PCT/KR2007/000147 WO2007081139A1 (en) 2006-01-09 2007-01-09 Inter-layer prediction method for video signal

Related Child Applications (1)

Application Number Title Priority Date Filing Date
CN201210585882.3A Division CN103096078B (en) 2006-01-09 2007-01-09 For inter-layer prediction method and the device of vision signal

Publications (2)

Publication Number Publication Date
CN101385352A true CN101385352A (en) 2009-03-11
CN101385352B CN101385352B (en) 2013-02-13

Family

ID=38508610

Family Applications (9)

Application Number Title Priority Date Filing Date
CN2007800057826A Active CN101385355B (en) 2006-01-09 2007-01-09 Inter-layer prediction method for video signal
CN2007800053238A Active CN101385348B (en) 2006-01-09 2007-01-09 Inter-layer prediction method for video signal
CN2007800057544A Active CN101385353B (en) 2006-01-09 2007-01-09 Inter-layer prediction method for video signal
CN2007800056166A Active CN101385350B (en) 2006-01-09 2007-01-09 Inter-layer prediction method for video signal
CN2007800056467A Active CN101385351B (en) 2006-01-09 2007-01-09 Inter-layer prediction method for video signal
CN200780005672XA Active CN101385352B (en) 2006-01-09 2007-01-09 Inter-layer prediction method for video signal
CN2007800056109A Active CN101385349B (en) 2006-01-09 2007-01-09 Inter-layer prediction method for video signal
CNA2007800057012A Pending CN101416522A (en) 2006-01-09 2007-01-09 Method for encoding and decoding video signal
CN2007800057614A Active CN101385354B (en) 2006-01-09 2007-01-09 Inter-layer prediction method for video signal

Family Applications Before (5)

Application Number Title Priority Date Filing Date
CN2007800057826A Active CN101385355B (en) 2006-01-09 2007-01-09 Inter-layer prediction method for video signal
CN2007800053238A Active CN101385348B (en) 2006-01-09 2007-01-09 Inter-layer prediction method for video signal
CN2007800057544A Active CN101385353B (en) 2006-01-09 2007-01-09 Inter-layer prediction method for video signal
CN2007800056166A Active CN101385350B (en) 2006-01-09 2007-01-09 Inter-layer prediction method for video signal
CN2007800056467A Active CN101385351B (en) 2006-01-09 2007-01-09 Inter-layer prediction method for video signal

Family Applications After (3)

Application Number Title Priority Date Filing Date
CN2007800056109A Active CN101385349B (en) 2006-01-09 2007-01-09 Inter-layer prediction method for video signal
CNA2007800057012A Pending CN101416522A (en) 2006-01-09 2007-01-09 Method for encoding and decoding video signal
CN2007800057614A Active CN101385354B (en) 2006-01-09 2007-01-09 Inter-layer prediction method for video signal

Country Status (2)

Country Link
KR (3) KR20070074452A (en)
CN (9) CN101385355B (en)

Families Citing this family (16)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US8792554B2 (en) 2006-01-09 2014-07-29 Lg Electronics Inc. Inter-layer prediction method for video signal
CN105933715B (en) * 2010-04-13 2019-04-12 Ge视频压缩有限责任公司 Across planar prediction
BR112012026400B1 (en) 2010-04-13 2021-08-10 Ge Video Compression, Ll INTER-PLANE PREDICTION
KR101549644B1 (en) 2010-04-13 2015-09-03 지이 비디오 컴프레션, 엘엘씨 Sample region merging
CN106231324B (en) 2010-04-13 2019-11-05 Ge视频压缩有限责任公司 Decoder, coding/decoding method, encoder and coding method
DK2559005T3 (en) 2010-04-13 2015-11-23 Ge Video Compression Llc Inheritance at sample array multitræsunderinddeling
WO2012097749A1 (en) * 2011-01-19 2012-07-26 Mediatek Inc. Method and apparatus for parsing error robustness of temporal motion vector prediction
EP3174295B1 (en) 2012-04-13 2018-12-12 GE Video Compression, LLC Low delay picture coding
PL2868103T3 (en) 2012-06-29 2017-06-30 Ge Video Compression, Llc Video data stream concept
WO2014038906A1 (en) * 2012-09-09 2014-03-13 엘지전자 주식회사 Image decoding method and apparatus using same
EP3748969B1 (en) * 2012-09-27 2024-01-03 Dolby Laboratories Licensing Corporation Inter-layer reference picture processing for coding standard scalability
US10375405B2 (en) * 2012-10-05 2019-08-06 Qualcomm Incorporated Motion field upsampling for scalable coding based on high efficiency video coding
WO2014163418A1 (en) * 2013-04-04 2014-10-09 한국전자통신연구원 Image encoding/decoding method and device
KR20140121315A (en) 2013-04-04 2014-10-15 한국전자통신연구원 Method and apparatus for image encoding and decoding based on multi-layer using reference picture list
US9756335B2 (en) * 2013-07-02 2017-09-05 Qualcomm Incorporated Optimizations on inter-layer prediction signalling for multi-layer video coding
KR20150009424A (en) 2013-07-15 2015-01-26 한국전자통신연구원 Method and apparatus for image encoding and decoding using inter-layer prediction based on temporal sub-layer information

Family Cites Families (15)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
GB9022326D0 (en) * 1990-10-15 1990-11-28 British Telecomm Signal coding
AU5632394A (en) * 1993-03-05 1994-09-08 Sony Corporation Apparatus and method for reproducing a prediction-encoded video signal
EP0731614B1 (en) * 1995-03-10 2002-02-06 Kabushiki Kaisha Toshiba Video coding/decoding apparatus
US6173013B1 (en) * 1996-11-08 2001-01-09 Sony Corporation Method and apparatus for encoding enhancement and base layer image signals using a predicted image signal
JP2001160971A (en) * 1996-11-12 2001-06-12 Matsushita Electric Ind Co Ltd Digital image coding method and digital image encoding device, digital image decoding method and digital image decoder, and data storage medium
JPH10271509A (en) * 1997-03-26 1998-10-09 Sharp Corp Image encoder and image decoder
US6332002B1 (en) * 1997-11-01 2001-12-18 Lg Electronics Inc. Motion prediction apparatus and method
KR100281462B1 (en) * 1998-03-30 2001-02-01 전주범 Method for encoding motion vector of binary shape signals in interlaced shape coding technique
US20020118742A1 (en) * 2001-02-26 2002-08-29 Philips Electronics North America Corporation. Prediction structures for enhancement layer in fine granular scalability video coding
ES2745058T3 (en) * 2002-04-19 2020-02-27 Panasonic Ip Corp America Motion vector calculation method
KR20040070490A (en) * 2003-02-03 2004-08-11 삼성전자주식회사 Method and apparatus for encoding/decoding video signal in interlaced video
CN1204753C (en) * 2003-05-19 2005-06-01 北京工业大学 Interframe predicting method based on adjacent pixel prediction
JP2005123732A (en) * 2003-10-14 2005-05-12 Matsushita Electric Ind Co Ltd Apparatus and method for deblocking filter processing
EP1585061A1 (en) * 2004-04-07 2005-10-12 Deutsche Thomson-Brandt Gmbh Block adaptive predictive coding
CN1268136C (en) * 2004-07-02 2006-08-02 上海广电(集团)有限公司中央研究院 Frame field adaptive coding method based on image slice structure

Also Published As

Publication number Publication date
CN101385349A (en) 2009-03-11
CN101385350A (en) 2009-03-11
CN101385348A (en) 2009-03-11
CN101385349B (en) 2013-07-10
CN101385353B (en) 2010-12-22
CN101416522A (en) 2009-04-22
KR20070074451A (en) 2007-07-12
CN101385354A (en) 2009-03-11
KR20070074453A (en) 2007-07-12
CN101385351B (en) 2011-10-05
CN101385350B (en) 2010-12-22
CN101385355A (en) 2009-03-11
CN101385354B (en) 2010-12-08
CN101385353A (en) 2009-03-11
KR20070074452A (en) 2007-07-12
CN101385348B (en) 2011-01-12
CN101385352B (en) 2013-02-13
CN101385355B (en) 2011-08-31
CN101385351A (en) 2009-03-11

Similar Documents

Publication Publication Date Title
CN101385354B (en) Inter-layer prediction method for video signal
KR100917829B1 (en) Inter-layer prediction method for video signal
CN103096078B (en) For inter-layer prediction method and the device of vision signal
RU2384970C1 (en) Interlayer forcasting method for video signal
MX2008008825A (en) Inter-layer prediction method for video signal

Legal Events

Date Code Title Description
C06 Publication
PB01 Publication
C10 Entry into substantive examination
SE01 Entry into force of request for substantive examination
C14 Grant of patent or utility model
GR01 Patent grant