WO2008091117A1 - A method and an apparatus for processing a video signal - Google Patents

A method and an apparatus for processing a video signal Download PDF

Info

Publication number
WO2008091117A1
WO2008091117A1 PCT/KR2008/000444 KR2008000444W WO2008091117A1 WO 2008091117 A1 WO2008091117 A1 WO 2008091117A1 KR 2008000444 W KR2008000444 W KR 2008000444W WO 2008091117 A1 WO2008091117 A1 WO 2008091117A1
Authority
WO
WIPO (PCT)
Prior art keywords
block
view
motion
motion information
corresponding block
Prior art date
Application number
PCT/KR2008/000444
Other languages
French (fr)
Inventor
Jeong Hyu Yang
Original Assignee
Lg Electronics Inc.
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Application filed by Lg Electronics Inc. filed Critical Lg Electronics Inc.
Priority to US12/449,062 priority Critical patent/US8923399B2/en
Priority to JP2009547172A priority patent/JP5249242B2/en
Priority to KR1020097016427A priority patent/KR20090113281A/en
Priority to EP08704944A priority patent/EP2123042A4/en
Priority to CN200880008776A priority patent/CN101647279A/en
Publication of WO2008091117A1 publication Critical patent/WO2008091117A1/en

Links

Classifications

    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04NPICTORIAL COMMUNICATION, e.g. TELEVISION
    • H04N19/00Methods or arrangements for coding, decoding, compressing or decompressing digital video signals
    • H04N19/50Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using predictive coding
    • H04N19/503Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using predictive coding involving temporal prediction
    • H04N19/51Motion estimation or motion compensation
    • H04N19/513Processing of motion vectors
    • H04N19/517Processing of motion vectors by encoding
    • H04N19/52Processing of motion vectors by encoding by predictive encoding
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04NPICTORIAL COMMUNICATION, e.g. TELEVISION
    • H04N13/00Stereoscopic video systems; Multi-view video systems; Details thereof
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04NPICTORIAL COMMUNICATION, e.g. TELEVISION
    • H04N19/00Methods or arrangements for coding, decoding, compressing or decompressing digital video signals
    • H04N19/10Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using adaptive coding
    • H04N19/102Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using adaptive coding characterised by the element, parameter or selection affected or controlled by the adaptive coding
    • H04N19/103Selection of coding mode or of prediction mode
    • H04N19/105Selection of the reference unit for prediction within a chosen coding or prediction mode, e.g. adaptive choice of position and number of pixels used for prediction
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04NPICTORIAL COMMUNICATION, e.g. TELEVISION
    • H04N19/00Methods or arrangements for coding, decoding, compressing or decompressing digital video signals
    • H04N19/10Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using adaptive coding
    • H04N19/102Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using adaptive coding characterised by the element, parameter or selection affected or controlled by the adaptive coding
    • H04N19/103Selection of coding mode or of prediction mode
    • H04N19/107Selection of coding mode or of prediction mode between spatial and temporal predictive coding, e.g. picture refresh
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04NPICTORIAL COMMUNICATION, e.g. TELEVISION
    • H04N19/00Methods or arrangements for coding, decoding, compressing or decompressing digital video signals
    • H04N19/10Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using adaptive coding
    • H04N19/134Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using adaptive coding characterised by the element, parameter or criterion affecting or controlling the adaptive coding
    • H04N19/136Incoming video signal characteristics or properties
    • H04N19/137Motion inside a coding unit, e.g. average field, frame or block difference
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04NPICTORIAL COMMUNICATION, e.g. TELEVISION
    • H04N19/00Methods or arrangements for coding, decoding, compressing or decompressing digital video signals
    • H04N19/10Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using adaptive coding
    • H04N19/169Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using adaptive coding characterised by the coding unit, i.e. the structural portion or semantic portion of the video signal being the object or the subject of the adaptive coding
    • H04N19/17Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using adaptive coding characterised by the coding unit, i.e. the structural portion or semantic portion of the video signal being the object or the subject of the adaptive coding the unit being an image region, e.g. an object
    • H04N19/176Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using adaptive coding characterised by the coding unit, i.e. the structural portion or semantic portion of the video signal being the object or the subject of the adaptive coding the unit being an image region, e.g. an object the region being a block, e.g. a macroblock
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04NPICTORIAL COMMUNICATION, e.g. TELEVISION
    • H04N19/00Methods or arrangements for coding, decoding, compressing or decompressing digital video signals
    • H04N19/50Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using predictive coding
    • H04N19/503Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using predictive coding involving temporal prediction
    • H04N19/51Motion estimation or motion compensation
    • H04N19/527Global motion vector estimation
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04NPICTORIAL COMMUNICATION, e.g. TELEVISION
    • H04N19/00Methods or arrangements for coding, decoding, compressing or decompressing digital video signals
    • H04N19/50Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using predictive coding
    • H04N19/597Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using predictive coding specially adapted for multi-view video sequence encoding
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04NPICTORIAL COMMUNICATION, e.g. TELEVISION
    • H04N19/00Methods or arrangements for coding, decoding, compressing or decompressing digital video signals
    • H04N19/60Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using transform coding
    • H04N19/61Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using transform coding in combination with predictive coding

Definitions

  • the present invention relates to video signal processing, and more particularly, to a method for processing a video signal and apparatus thereof.
  • the present invention is suitable for a wide scope of applications, it is particularly suitable for decoding video signals.
  • compression coding means a series of signal processing techniques for transferring digitalized information via a communication circuit or storing digitalized information in a format suitable for a storage medium.
  • Targets of compression coding include audio, video, character, etc.
  • video compression a technique of performing compression coding on video is called video compression.
  • Video sequence is generally characterized in having spatial redundancy and temporal redundancy.
  • inter-view pictures since inter-view pictures mostly have the difference caused by a camera position only, they tend to have very high relevance and redundancy. In case that the redundancy of the interview pictures is insufficiently eliminated, a compression ratio or a reconstruction rate is lowered.
  • the present invention is directed to a method for processing a video signal and apparatus thereof that substantially obviate one or more of the problems due to limitations and disadvantages of the related art.
  • An object of the present invention is to provide a method for processing a video signal and apparatus thereof, by which the video signal from which redundancy of interview pictures is eliminated can be decoded.
  • Another object of the present invention is to provide a method for processing a video signal and apparatus thereof, by which motion compensation can be carried out in a manner of obtaining motion information on a current picture based on relevance between inter-view pictures.
  • a further object of the present invention is to provide a method for processing a video signal and apparatus thereof, by which a reconstruction rate of a current picture can be raised using motion information in a reference view having high similarity to motion information on the current picture.
  • the present invention provides the following effects or advantages.
  • the present invention omits motion information having high redundancy in encoding a video signal, thereby raising a compression ratio.
  • the present invention raises a reconstruction rate of the video signal using motion information of a reference view having considerable similarity to motion information on the current picture .
  • the present invention uses motion information existing in a picture of another view.
  • FIG. 1 is a schematic block diagram of a video signal decoding apparatus according to the present invention
  • FIG. 2 is a block diagram of a video signal processing apparatus according to an embodiment of the present invention.
  • FIG. 3 is a flowchart of a video signal processing method according to an embodiment of the present invention.
  • FIG. 4 is a flowchart of an example for sub-steps of a motion skip determining step SlOO;
  • FIG. 5 is a conceptional diagram to explain view dependency and reference list
  • FIG. 6 is a diagram for one example of a reference list for a current view
  • FIG. 7 is a flowchart of one example for sub-steps of a step S300 of determining a reference view and a corresponding block
  • FIG. 8 is a diagram for another example of a reference list for a current view
  • FIG. 9 is a flowchart of another example for sub- steps of a step S300 of determining a reference view and a corresponding block;
  • FIG. 10 is a diagram for a further example of a reference list for a current view
  • FIG. 11 is a flowchart of a further example for sub- steps of a step S300 of determining a reference view and a corresponding block;
  • FIG. 12 is a diagram to explain a case that a mapping level is a macroblock level or 8x8 block level.
  • FIG. 13 is a diagram to explain a case that a mapping level is an 8x8 block level or a sub-pixel level.
  • a method of processing a video signal includes obtaining a block type of a first corresponding block within a first reference view included in a reference list, if the block type of the first corresponding block is an intra block, obtaining a block type and motion information of a second corresponding block within a second reference view included in the reference list, and if the block type of the first corresponding block is not the intra block, decoding a current block using the motion information of the second corresponding block.
  • the reference list includes a first reference list and a second reference list.
  • the first reference list includes a forward reference view and the second reference list includes a backward reference view.
  • the reference list is obtained based on view dependency on a current view.
  • the first reference view corresponds to a reference view having a lowest index indicating an order of being included in a bitstream of the video signal among reference views included in the reference list.
  • the first reference view corresponds to a reference view closest to a current view among reference views included in the reference list.
  • the first corresponding block is obtained based on a global motion vector for the first reference view.
  • the first corresponding block belongs to a reference view within a first reference list of the reference list and the second corresponding block belongs to a reference view within a second reference list of the reference list.
  • the second corresponding block is obtained based on a global motion vector for the second reference view.
  • the motion information includes a motion vector and a reference picture index. More preferably, the decoding further includes obtaining a reference picture in an intra-view direction.
  • the video signal is received on a broadcast signal.
  • the video signal is received via a digital medium.
  • a computer-readable recording medium comprising a program for executing a method including obtaining a block type of a first corresponding block within a first reference view included in a reference list, if the block type of the first corresponding block is an intra block, obtaining a block type and motion information of a second corresponding block within a second reference view included in the reference list, and if the block type of the first corresponding block is not the intra block, decoding a current block using the motion information of the second corresponding block.
  • an apparatus for processing a video signal includes a motion information obtaining unit obtaining a block type of a first corresponding block within a first reference view included in a reference list, the motion information obtaining unit, if the block type of the first corresponding block is an intra block, the motion information obtaining unit obtaining a block type and motion information of a second corresponding block within a second reference view included in the reference list and a decoding unit, if the block type of the first corresponding block is not the intra block, the decoding unit decoding a current block using the motion information of the second corresponding block.
  • a method of processing a video signal includes reordering at least one of a first reference list and a second reference list, deciding a motion vector using either a first motion information extracted from a bitstream or a second motion information of a corresponding block on a reference view, deciding a motion vector using at least one motion information of inter-motion information extracted from the bitstream and intra-motion information extracted from the corresponding block on the reference view, obtaining a reference block using the motion vector, and decoding a current block of a current picture using the reference block.
  • the motion vector uses the inter-motion formation only.
  • the reference block is associated with the inter-motion information and the reference block exists on a same time of the current block in a view different from that of the current block.
  • an apparatus for processing a video signal includes a motion information obtaining unit reordering at least one of a first reference list and a second reference list, the motion information obtaining unit deciding a motion vector using either a first motion information extracted from a bitstream or a second motion information of a corresponding block on a reference view, the motion information obtaining unit deciding a motion vector using at least one motion information of inter-motion information extracted from the bitstream and intra-motion information extracted from the corresponding block on the reference view and a decoding unit obtaining a reference block using the motion vector, the decoding unit decoding a current block of a current picture using the reference block.
  • motion information in the present invention should be construed as a concept that includes interview motion information as well as temporal motion information.
  • a motion vector should be construed as a concept that includes interview disparity offset as well as temporal motion offset.
  • coding should conceptionally include both encoding and decoding.
  • intra-view prediction means prediction between pictures existing within a same view and includes temporal prediction for example.
  • Inter-view prediction means prediction between pictures existing in different views .
  • FIG. 1 is a schematic block diagram of an apparatus for decoding a video signal according to the present invention.
  • a decoding apparatus includes a NAL parser unit 100, an entropy decoding unit 200, an inverse quantization/inverse transform unit 300, an intra prediction unit 400, a deblocking filter unit 500, a decoded picture buffer unit 600 and an inter prediction unit 700.
  • the inter prediction unit 700 includes an illumination (IC) offset predictor 710, an illumination compensation unit 720 and a motion compensation unit 730.
  • the NAL parser unit 100 performs parsing by a unit of NAL (network abstraction layer) to decode a received video sequence. Normally, before a slice header and slice data are decoded, at least one sequence parameter set and picture parameter set are transferred to a decoder.
  • a NAL header unit or an extension unit of a NAL header can contain various kinds of attribute informations .
  • MVC is a scheme additional to AVC scheme. And, it may be more advantageous to add various kinds of attribute informations for a case of an MVC bitstream only rather tan to add them unconditionally. For instance, it is able to add flag information for identifying whether it is an MVC bitstream in the NAL header unit or the extension unit of the NAL header.
  • the attribute information can include temporal level information, view level information, inter-view picture group identification information, view identification information and the like.
  • Parsed bitstream is entropy-decoded by the entropy decoding unit 20, and a coefficient, motion vector and the like of each macroblock is extracted.
  • the inverse dequnatization/transform unit 300 obtains a coefficient value transformed by multiplying a received quantized value by a predetermined constant and then reconstructs a pixel value by inverting the coefficient value.
  • the intra prediction unit 400 performs intra-frame prediction from a decoded sample within a current picture using the reconstructed pixel value.
  • the deblocking filter unit 500 is applied to each coded macroblock to decrease block distortion. In this case, a filter smoothens edges of a block to enhance an image quality of a decoded picture. And, a selection of this filtering process depends on boundary strength and a gradient of an image sample around a boundary. Filtered pictures are outputted or and stored in the decoded picture buffer unit 600 to be used as reference pictures.
  • the decoded picture buffer unit 600 plays a role in storing or opening previously coded pictures to perform inter- frame prediction. In doing so, to store/open the pictures in/from the decoded picture buffer unit 600, v frame_num" of each picture and POC (picture order count) are used. In MVC, since there exist pictures in a view different from that of a current picture among the previously coded pictures, view information for identifying a view of picture is usable as well as the v frame_num" and the POC to utilize theses pictures as reference pictures. The reference pictures managed in the above manner are usable for the inter prediction unit 700 as well. The inter prediction unit 700 carries out inter-frame prediction using the reference pictures stored in the decoded picture buffer unit 600.
  • the inter prediction unit 700 includes the illumination offset predictor 710, the illumination compensation unit 720 and the motion compensation unit 730.
  • each view sequence is a sequence captured by a different camera. So, an illumination difference is generated due to internal and external causes of the cameras.
  • the illumination compensation unit 720 carries out illumination compensation. In carrying out the illumination compensation, it is able to use flag information indicating whether to perform illumination compensation on a predetermined layer of a video signal. For instance, it is able to perform illumination compensation using flag information indicating whether to perform illumination compensation on a corresponding slice or macroblock. In performing illumination compensation using the flag information, the illumination compensation is applicable to various macroblock types (e.g., inter 16 x 16 mode, B-skip mode, direct mode) .
  • the motion compensation unit 730 compensates for a motion of a current block using motion informations transferred from the entropy decoding unit 200.
  • the motion compensation unit 730 extracts motion vectors of blocks neighbor to a current block from a video signal and then obtains a motion vector prediction value of the current block.
  • the motion compensation unit 730 compensates for a motion of the current block using the obtained motion vector prediction value and an offset vector extracted from the video signal.
  • Such motion compensation can be performed using a single reference picture or a plurality of pictures.
  • motion compensation can be carried out using information on a reference picture list for interview prediction which is stored in the decoded picture buffer unit 600. And, motion compensation can be also carried out using view information for identifying a view of a corresponding picture.
  • the motion skip mode is a scheme for deriving motion information on a current block from motion information on a block having been fully decoded. For instance, a reference picture index, a motion vector and the like of a current block can be derived from a corresponding block existing in a reference view having motion information similar to that of the current block. In this scheme, since the motion information on the current block is not transferred, it is able to save the number of bits necessary to encode motion information.
  • FIG. 2 is a block diagram of a video signal processing apparatus according to an embodiment of the present invention
  • FIG. 3 is a flowchart of a video signal processing method according to an embodiment of the present invention.
  • a video signal processing apparatus 1000 according to an embodiment of the present invention includes an information extracting unit 1050, a motion skip determining unit 1100, a reference list obtaining unit 1200, a offset obtaining unit 1250, a corresponding block deciding unit 1300, and a decoding unit 1400.
  • a motion information obtaining unit is able to include the reference list obtaining unit 1200, the offset obtaining unit 1250 and the corresponding block deciding unit 1300.
  • the embodiment of the present invention is schematically explained with reference to FIG. 2 and FIG. 3.
  • the information extracting unit 1050 extracts motion skip allowing flag information and motion skip applying flag information, and the motion skip determining unit 110 then decides whether a current block corresponds to a motion skip mode based on the extracted information [SIlO] . If the current block does not correspond to the motion skip mode, the current block is coded using the motion information extracted by the decoding unit 1400 through a step of extracting motion information on the current block from a bitstream. Meanwhile, the information extracting unit 1050 can be included in the entropy decoding unit 200 described with reference to FIG. 1, which does not put limitation on various implementations of the present invention. Besides, the step SIlO will be explained in detail later with reference to ⁇ l. Motion skip determining step'.
  • the reference list obtaining unit 1200 obtains a first reference list and a second reference list based on view dependency [S200] . Details of the view dependency and the reference list will be explained later with reference to ⁇ 2.1 Reference list obtaining step'.
  • the offset obtaining unit 1250 obtains offset (e.g., global motion vector, local motion vector, zero vector) necessary to search for a corresponding block [S250] . In this case, various embodiments for an offset type and an offset deriving method will be explained in detail later with reference to '2.2 Offset obtaining step'.
  • the corresponding block deciding unit 1300 decides a reference view and a corresponding block based on the reference list obtained in the step S200 and the offset obtained in the step S250 [S300] .
  • Reference view and corresponding block deciding step' a method of deciding a corresponding block in accordance with whether a corresponding block corresponds to a macroblock level or 8 x 8 block level can be diversified, which will be explained later.
  • the decoding unit 1400 decodes a current block using the motion information on the corresponding block which is decided in the step S300 [S400] . And, residual information can be used as well as the motion information on the corresponding block, of which details will be explained later with reference to M. Step of decoding a current block using motion information on corresponding block' . Meanwhile, the motion skip determining unit, the reference list obtaining unit 1200, the offset obtaining unit 1250, the corresponding block deciding unit 1300 and the decoding unit 1400 can be included in the motion compensation unit 730 described with reference to FIG. 1, which does not restrict various implementations of the present invention.
  • the information extracting unit 1050 extracts motion skip allowing flag information and motion skip applying flag information, and the motion skip determining unit 1100 decides whether a current block corresponds to a motion skip mode based on the extracted information.
  • Sub-steps of the step slOO are shown in FIG. 4.
  • motion skip allowing flag information is firstly extracted from a slice header [SIlO] .
  • the motion skip allowing flag information is the information indicating whether at least one block corresponding to a motion skip mode exists in blocks belonging to a slice.
  • the motion skip allowing flag information is the information indicating whether a motion skip mode is allowed for blocks within a slice.
  • a motion skip allowing flag can be included in a slice header, which does not put limitation on various implementations of the present invention.
  • An example of syntax, in which a motion skip allowing flag is included in a slice header is shown in the following table.
  • Table 1 Referring to a row indicates by (A) marked on a right side thereof, there exists a syntax element x motion_skip_enable_flag' .
  • This is a motion skip allowing flag and may include a 1-bit flag having a value set to 0 or 1. And, a meaning of each flag value can be defined as the following table.
  • motion skip is not applied to a current slice, it is unnecessary to transfer a motion skip allowing flag. So, as shown in the above table, it is able to add a condition for applying a motion skip. In particular, only if a current slice is a non-anchor picture and it has a reference relation with another view, it is sufficient to transfer a motion skip allowing flag.
  • a motion skip mode is allowed for a current slice [S120] . If the motion skip mode is not slowed [ ⁇ no' in the step S120] (e.g., if a motion skip allowing flag motion_skip_enable_flag is 0) , it means that any block corresponding to motion skip mode does not exist in slice at all. Hence, it is decided that a whole corresponding slice does not correspond to the motion skip mode. Motion information on a current block is then extracted at a block level [S150] .
  • motion skip applying flag information is extracted at a block level [S130] .
  • the motion skip applying flag information is the information indicating whether the current block corresponds to the motion skip mode.
  • each macroblock just needs to check a motion skip allowing flag only. So, in case that a motion skip allowing flag, as shown in the above table, in a current slice is 1, it is sufficient to transfer a motion skip applying flag for each macroblock.
  • a motion skip applying flag may include a 1-bit flag having a value set to 0 or 1 as well. Meaning according to each flag value can be defined as the following table.
  • the deciding step S140 if the current block does not correspond to the motion skip mode ['no' in the step S140] (e.g., motion skip applying flag motion_skip_flag is 0) , it is determined that it is not the motion skip mode. Motion information on the current block is then extracted from a bitstream [S150] [cf. rows (CCl) and (Pl) in Table 3] .
  • step SlOO is terminated and the steps S200 to S400 are then executed.
  • the view dependency means relation between a current view and a different view and may relate to whether a different view is necessary to decode a current view.
  • the view dependency can be estimated from information contained in a sequence extension field (SPS MVC extension) .
  • the step S200 can be executed based on view dependency (view_dependency) of a non-anchor picture.
  • an anchor picture is a picture for executing inter-view prediction only, refers to a picture within a different view but does not refer to a picture within the same view.
  • a non-anchor picture is a picture not corresponding to the anchor picture and includes a picture capable of referring to both a picture in a same view and a picture in a different view.
  • the view dependency of the non-anchor picture is the relation between non-anchor pictures of a current view and non-anchor pictures of a different view. Since an anchor picture is characterized in executing inter-view prediction only, a motion skip mode is not applied to the anchor picture.
  • the motion skip mode is applicable to a non-anchor picture only.
  • the step S200 can be executed based on the view dependency of the non-anchor picture.
  • view information having view dependency between a current view and a non-anchor picture can be divided into view information on a first reference list and view information on a second reference list.
  • FIG. 5 is a conceptional diagram to explain view dependency and reference list.
  • non-anchor pictures P C i, -/Pe n of a current view V c there exist non-anchor pictures P C i, -/Pe n of a current view V c .
  • a right dotted-line arrow indicates view dependency of a non- anchor picture between the current view V c and a different view Vc-i, V c- 2, Vc + i or V c+2 .
  • a first reference view list RLl conceptionally includes the view V c -i and the view Vc-2
  • a second reference view list RL2 conceptionally includes the view V c+ i and the view V c+ 2 •
  • An order of storage in a reference view list is not always defined as order that gets closer to a current view spatially or may be specified by an encoder.
  • a backward view can be contained in the first view list and a forward view can be contained in the second reference view list.
  • Reference picture list reordering (RPLR) operation enables the order to be set different per a slice in a basic list configuration .
  • a first reference block included in a reference view on a same time reference of a current block may be a reference block associated with inter-motion information and a second reference block included in a current view on a same view reference of a current block may e a reference block associated with intra-motion information.
  • an offset value between a current block and a corresponding block of a reference picture is needed.
  • the offset value can be derived in various ways a) zero vector, b) global motion vector, c) local motion vector, etc.
  • a corresponding block can be searched for in various ways using the offset value as well.
  • Global motion vector or global disparity vector is a value that represents a difference of overall motion or disparity between a current picture and a reference picture and may include an offset corresponding to a slice, a field, a frame, a sequence or the like as well as a picture. If a motion skip mode is to skip motion information in a time direction, a global motion vector becomes a difference of motion or disparity in a view direction. On the contrary, if a motion skip mode is to skip motion information in a view direction, a global motion vector becomes a difference of motion or disparity in a time direction.
  • the global motion vector can be transferred for each picture or for each anchor picture. And, it is also able to transfer one global motion vector for a whole sequence. In case that a global motion vector is transferred for each anchor, since a motion skip mode is applicable to a non-anchor picture only, it is able to transfer a global motion vector for a view having view dependency of a non-anchor picture only.
  • a global motion vector is transferred for each anchor picture, it is able to derive a global motion vector of a non-anchor picture using a global motion vector of a neighbor anchor picture.
  • This can be accomplished in various ways.
  • a current picture is a non-anchor picture, it is able to bring one of global motion vectors of a most nearest neighbor anchor picture as it is.
  • it is able to find a global motion vector by performing interpolation on global motion vectors of two neighbor anchor pictures by considering POC (picture order count) between the current picture and a neighbor anchor picture.
  • POC picture order count
  • Resolution of a global motion vector can have resolution of macroblock, 8x8 block, 4x4 block, pixel, subpixel or the like.
  • a reference picture is decided by 8x8 block unit and a motion vector is decided by 4x4 block unit. So, in case that a global motion vector is found by 8 x 8 block level unit, it is able to specify a more detailed corresponding block than that found by macroblock level unit.
  • Local motion vector normally has resolution of l/4pel not as an offset corresponding a whole picture but as an offset corresponding to a specific macroblock, a specific 8x8 block or a specific 4x4 block within a picture.
  • neighbor vectors adjacent to a current block have motion vectors (or disparity vectors) in a view direction
  • it is able to derive a location motion vector of the current block using the motion vectors of the neighbor blocks.
  • the reference picture is identical in accordance with a priority by giving the priority to a specific block (e.g., first priority given to a left upper block, second priority to an upper block, etc.), it is able to use the motion vector of the neighbor block having the same reference picture.
  • a specific block e.g., first priority given to a left upper block, second priority to an upper block, etc.
  • the zero or view-directional motion vector (global motion vector or local motion vector) derived in Clause 2.2.1 may differ from a real motion vector of a current block in a view direction. If the difference is transferred as the following table, a prediction value is set to a value of a derived motion vector and it is then able to obtain a motion vector by adding the difference value to the prediction value.
  • a prediction value is set to a value of a derived motion vector and it is then able to obtain a motion vector by adding the difference value to the prediction value.
  • the difference value can be a value of macroblock (16x16) level, 8x8 block level, 4x4 block level, pixel level or subpixel unit.
  • a corresponding block may belong to a picture on a same time zone. If a corresponding block belongs to a picture on a different time instance, it is able to further use POC (picture order count) information.
  • POC picture order count
  • the information on the reference view can be transferred via a bitstream by an encoder, which is an explicit scheme.
  • the information on the reference view can be arbitrarily decided by a decoder, which is an implicit scheme.
  • the explicit and implicit schemes will be explained in the following description.
  • mapping a current block and a corresponding block to each other there can exist various embodiments for varying such a level as a macroblock level, an 8x8 block level and the like. This will be explained in the following description as well.
  • first and second reference lists i.e., a view identifier of a reference view.
  • first and second reference lists will be set as a reference view.
  • it is able to decide a final reference view and a final corresponding block by checking whether motion formation exists in a corresponding block of each of the reference views.
  • various embodiments can exist for whether to firstly check a reference view belonging to which one of the first and second reference lists.
  • various embodiments can exist for, if motion information does not exist in the corresponding reference view, how to continue performing a check in prescribed order.
  • priorities between reference views belonging to a reference list first of all, it is able to check reference views included in the first or second reference list in order of a lower index indicating a reference view.
  • the index indicating the reference view may include a serial number of each reference view set in coding a bitstream by an encoder.
  • 'i' may be the index indicating the reference view.
  • FIG. 6 is a diagram for one example of a reference list for a current view
  • FIG. 7 is a flowchart of one example for sub-steps of a step S300 of deciding a reference view and a corresponding block. Referring to FIG. 6, it can be observed that there exist a first reference list RLl in a direction LO and a second reference list RL2 in a direction Ll with reference to a current view V 0 and a current block MB C .
  • the first corresponding block CBl is not an intra block, i.e., if motion information exists [ ⁇ no' in the step S320]
  • the first corresponding block is finally decided as a corresponding block and then motion information is obtained from the first corresponding block [S332] .
  • a block type of the corresponding block CBl is an intra-frame prediction block
  • Vc and the second reference view RV2 is decided as a second corresponding block CB2 [S334] .
  • third and fourth corresponding blocks CB3 and CB4 are sequentially checked.
  • a presence or non- presence of motion information is checked in a manner of alternating the reference views of the first and second reference lists RLl and RL2.
  • a selection reference for candidates i.e., the first reference view, the second reference view, etc.
  • a candidate selection reference can be the base view or an order that is closest to the base view, which does not restrict various implementations of the present invention.
  • FIG. 8 is a diagram for another example of a reference list for a current view
  • FIG. 9 is a flowchart of another example for sub-steps of a step S300 of determining a reference view and a corresponding block.
  • FIG. 8 and FIG. 9 show a first reference list RLl only.
  • IT is then checked whether a type of each of the corresponding blocks is an intra block or not.
  • a reference view and a corresponding block capable of using motion information are finally decided.
  • FIG. 10 is a diagram for a further example of a reference list for a current view
  • FIG. 11 is a flowchart of a further example for sub-steps of a step S300 of determining a reference view and a corresponding block.
  • FIG. 10 and FIG. 11 show a second reference list RL2 only. Similar to the case shown in FIG. 8 and FIG. 9, first reference view RVl and first corresponding block CBl, second reference view RV2 and second corresponding block CB2, third reference view RV3 and third corresponding block CB3 and the like are decided in order of a lower index in the second reference list RL2. It is then checked whether a type of each of the corresponding blocks is an intra block in corresponding order.
  • a reference view in a direction LO is not tested and block types of the corresponding blocks are checked in order from a view having a lower index among the reference views of the second reference list RL2.
  • a local motion vector corresponding to a macroblock (16x16 block) , 8x8 block, 4x4 block, pixel, subpixel or the like as well as a global motion vector.
  • a unit of a global or local motion vector can be a unit of a macroblock (16x16 block) , 8x8 block, 4x4 block, pixel, subpixel or the like.
  • FIG. 12 is a diagram to explain a case that a mapping level is a macroblock level or 8x8 block level
  • FIG. 13 is a diagram to explain a case that a mapping level is an 8x8 block level or a sub-pixel level. A method of deciding a corresponding block in each case is explained as follows.
  • a left upper end point of a block LB having a global or local motion vector applied thereto is marked. It is able to decide a macroblock CB 0 closest to the block LB having the global or local motion vector applied thereto as a corresponding block.
  • all motion information on a corresponding block such as block type information, a reference picture index, a motion vector and the like can be used as motion information on a current block as it is.
  • a corresponding block is such a skip block as P skip (P_skip) and B skip (B_skip) , it can be reset to an inter 8x8 mode.
  • 8x8 block CB 0 closest to a block LB having a global or local motion vector applied thereto and 8x8 blocks CB 1 , CB 2 and CB 3 neighbor to the block are shown.
  • 8x8 blocks as corresponding blocks, it is able to use motion informations of the blocks for decoding of a current block. If each of the four blocks has motion information as an inter block, a type of the current block is decided as "inter 8x8'. If the current block is a macroblock (16x16), it is divided into four 8x8 blocks (MBc 0 , MBci, MBc 2 , MBc 3 ) (not shown in the drawing) .
  • Motion information on a corresponding block corresponding to each of the 8x8 blocks is then used. For instance, in case that the current block is MBc 2 , motion information on a corresponding block CB 2 is used. On the other hand, in case that there exists an intra block in four corresponding blocks, it is able to use motion information corresponding to a different inter block.
  • a reference picture index is derived using index information on a reference picture of the 8x8 block including the 4x4 block CB 00 and a motion vector is derived by taking an intermediate or average value of motion vectors of the 4x4 block CB 00 and reference blocks CBi 0 , CB 2 O and CB 30 .
  • FIG. 13 4x4 block closest to a block LB having a global or local motion vector applied thereto and it s neighbor blocks CB 00 , CB 01 , CB 02 and CB 03 are shown.
  • a reference picture index and a motion vector of the closest 4x4 block are usable as they are or can be derived from reference picture indexes and motion vectors of four 4x4 blocks.
  • a method of an intermediate or average is usable. And, it is able to set a type of macroblock to v inter__8x8 mode'.
  • each of the per-level mapping methods is independently usable or any combinations thereof are possible.
  • motion information on a current block is derived using motion information on the corresponding block in the step S400.
  • various corresponding blocks of 16x16 level, 8x8 level and the like as mentioned in the foregoing description of ⁇ 3.2 Corresponding block
  • Each Mapping Level' there can exist various embodiments. Meanwhile, there can exist a method of deriving residual information and the like as well as motion information on a corresponding block. To raise accuracy of motion information, there can exist a method of transferring a difference between a motion vector value derived by a motion skip scheme and a real value. Theses methods are explained as follows.
  • a residual of a corresponding block is usable as a residual of a current block as well as motion information on the corresponding block.
  • a prediction value of a current block is obtained using a reference picture index and motion vector obtained in accordance with a process for a motion skip mode .
  • the current block is then reconstructed by adding the residual of the corresponding block to the obtained prediction value.
  • the motion vector obtained in accordance with the process for the motion skip mode may have a difference from a real motion vector of the current block. If the difference value is transferred via a bitstream, it is able to obtain a more accurate motion vector by adding the difference value to the motion vector derived in accordance with the motion skip mode in the course of decoding. By performing motion compensation suing the obtained motion vector, it is able to raise a reconstruction rate of the current block.
  • a residual of the current block may almost 0. In this case, it s able to reduce bits rate required for residual coding by not transferring a residual value at all.
  • flag information e.g., residual_coding_flag
  • residual decoding is not performed as well as residual data is not extracted. If the flag information is 1, residual data is extracted and the extracted residual data is then added to a prediction value of a current block.
  • the encoding/decoding method of the present invention can be implemented in a program recorded medium as computer-readable codes.
  • the computer-readable media include all kinds of recording devices in which data readable by a computer system are stored.
  • the computer- readable media include ROM, RAM, CD-ROM, magnetic tapes, floppy discs, optical data storage devices, and the like for example and also include carrier-wave type implementations (e.g., transmission via Internet).
  • carrier-wave type implementations e.g., transmission via Internet.
  • a bitstream produced by the encoding method is stored in a computer-readable recording medium or can be transmitted via wire/wireless communication network.
  • the present invention is applicable to encoding/decoding a video signal.

Abstract

An apparatus for processing a video signal and method thereof are disclosed, by which the video signal from which redundancy of inter-view pictures is eliminated can be decoded. The present invention includes obtaining a block type of a first corresponding block within a first reference view included in a reference list, if the block type of the first corresponding block is an intra block, obtaining a block type and motion information of a second corresponding block within a second reference view included in the reference list, and if the block type of the first corresponding block is not the intra block, decoding a current block using the motion information of the second corresponding block.

Description

A METHOD AND AN APPARATUS FOR PROCESSING A VIDEO SIGNAL
TECHNICAL FIELD
The present invention relates to video signal processing, and more particularly, to a method for processing a video signal and apparatus thereof. Although the present invention is suitable for a wide scope of applications, it is particularly suitable for decoding video signals.
BACKGROUND ART
Generally, compression coding means a series of signal processing techniques for transferring digitalized information via a communication circuit or storing digitalized information in a format suitable for a storage medium. Targets of compression coding include audio, video, character, etc. In particular, a technique of performing compression coding on video is called video compression. Video sequence is generally characterized in having spatial redundancy and temporal redundancy.
DISCLOSURE OF THE INVENTION TECHNICAL PROBLEM
However, if the spatial redundancy and the temporal redundancy are not sufficiently eliminated, a compression rate in coding a video signal is lowered. If the spatial redundancy and the temporal redundancy are excessively eliminated, it is unable to generate information required for decoding a video signal to degrade a reconstruction ratio.
Specifically, in a multi-view video signal, since inter-view pictures mostly have the difference caused by a camera position only, they tend to have very high relevance and redundancy. In case that the redundancy of the interview pictures is insufficiently eliminated, a compression ratio or a reconstruction rate is lowered.
TECHNICAL SOLUTION Accordingly, the present invention is directed to a method for processing a video signal and apparatus thereof that substantially obviate one or more of the problems due to limitations and disadvantages of the related art.
An object of the present invention is to provide a method for processing a video signal and apparatus thereof, by which the video signal from which redundancy of interview pictures is eliminated can be decoded.
Another object of the present invention is to provide a method for processing a video signal and apparatus thereof, by which motion compensation can be carried out in a manner of obtaining motion information on a current picture based on relevance between inter-view pictures.
A further object of the present invention is to provide a method for processing a video signal and apparatus thereof, by which a reconstruction rate of a current picture can be raised using motion information in a reference view having high similarity to motion information on the current picture.
ADVANTAGEOUS EFFECTS
Accordingly, the present invention provides the following effects or advantages.
First of all, the present invention omits motion information having high redundancy in encoding a video signal, thereby raising a compression ratio.
Secondly, in decoding a video signal, even if motion information on a current picture is not transferred, the present invention raises a reconstruction rate of the video signal using motion information of a reference view having considerable similarity to motion information on the current picture .
Thirdly, in decoding a video signal, in case that motion information fails to exist in a picture of a neighbor view, the present invention uses motion information existing in a picture of another view.
DESCRIPTION OF DRAWINGS The accompanying drawings, which are included to provide a further understanding of the invention and are incorporated in and constitute a part of this specification, illustrate embodiments of the invention and together with the description serve to explain the principles of the invention.
In the drawings :
FIG. 1 is a schematic block diagram of a video signal decoding apparatus according to the present invention;
FIG. 2 is a block diagram of a video signal processing apparatus according to an embodiment of the present invention;
FIG. 3 is a flowchart of a video signal processing method according to an embodiment of the present invention;
FIG. 4 is a flowchart of an example for sub-steps of a motion skip determining step SlOO;
FIG. 5 is a conceptional diagram to explain view dependency and reference list;
FIG. 6 is a diagram for one example of a reference list for a current view; FIG. 7 is a flowchart of one example for sub-steps of a step S300 of determining a reference view and a corresponding block;
FIG. 8 is a diagram for another example of a reference list for a current view;
FIG. 9 is a flowchart of another example for sub- steps of a step S300 of determining a reference view and a corresponding block;
FIG. 10 is a diagram for a further example of a reference list for a current view;
FIG. 11 is a flowchart of a further example for sub- steps of a step S300 of determining a reference view and a corresponding block;
FIG. 12 is a diagram to explain a case that a mapping level is a macroblock level or 8x8 block level; and
FIG. 13 is a diagram to explain a case that a mapping level is an 8x8 block level or a sub-pixel level.
BEST MODE Additional features and advantages of the invention will be set forth in the description which follows, and in part will be apparent from the description, or may be learned by practice of the invention. The objectives and other advantages of the invention will be realized and attained by the structure particularly pointed out in the written description and claims thereof as well as the appended drawings .
To achieve these and other advantages and in accordance with the purpose of the present invention, as embodied and broadly described, a method of processing a video signal according to the present invention includes obtaining a block type of a first corresponding block within a first reference view included in a reference list, if the block type of the first corresponding block is an intra block, obtaining a block type and motion information of a second corresponding block within a second reference view included in the reference list, and if the block type of the first corresponding block is not the intra block, decoding a current block using the motion information of the second corresponding block.
Preferably, the reference list includes a first reference list and a second reference list.
More preferably, the first reference list includes a forward reference view and the second reference list includes a backward reference view.
Preferably, the reference list is obtained based on view dependency on a current view.
Preferably, the first reference view corresponds to a reference view having a lowest index indicating an order of being included in a bitstream of the video signal among reference views included in the reference list.
Preferably, the first reference view corresponds to a reference view closest to a current view among reference views included in the reference list.
Preferably, the first corresponding block is obtained based on a global motion vector for the first reference view. Preferably, the first corresponding block belongs to a reference view within a first reference list of the reference list and the second corresponding block belongs to a reference view within a second reference list of the reference list. Preferably, the second corresponding block is obtained based on a global motion vector for the second reference view.
Preferably, the motion information includes a motion vector and a reference picture index. More preferably, the decoding further includes obtaining a reference picture in an intra-view direction.
Preferably, the video signal is received on a broadcast signal.
Preferably, the video signal is received via a digital medium.
To further achieve these and other advantages and in accordance with the purpose of the present invention, a computer-readable recording medium comprising a program for executing a method including obtaining a block type of a first corresponding block within a first reference view included in a reference list, if the block type of the first corresponding block is an intra block, obtaining a block type and motion information of a second corresponding block within a second reference view included in the reference list, and if the block type of the first corresponding block is not the intra block, decoding a current block using the motion information of the second corresponding block. To further achieve these and other advantages and in accordance with the purpose of the present invention, an apparatus for processing a video signal includes a motion information obtaining unit obtaining a block type of a first corresponding block within a first reference view included in a reference list, the motion information obtaining unit, if the block type of the first corresponding block is an intra block, the motion information obtaining unit obtaining a block type and motion information of a second corresponding block within a second reference view included in the reference list and a decoding unit, if the block type of the first corresponding block is not the intra block, the decoding unit decoding a current block using the motion information of the second corresponding block.
To further achieve these and other advantages and in accordance with the purpose of the present invention, a method of processing a video signal includes reordering at least one of a first reference list and a second reference list, deciding a motion vector using either a first motion information extracted from a bitstream or a second motion information of a corresponding block on a reference view, deciding a motion vector using at least one motion information of inter-motion information extracted from the bitstream and intra-motion information extracted from the corresponding block on the reference view, obtaining a reference block using the motion vector, and decoding a current block of a current picture using the reference block. Preferably, if the current picture is an anchor picture, the motion vector uses the inter-motion formation only.
More preferably, the reference block is associated with the inter-motion information and the reference block exists on a same time of the current block in a view different from that of the current block.
To further achieve these and other advantages and in accordance with the purpose of the present invention, an apparatus for processing a video signal includes a motion information obtaining unit reordering at least one of a first reference list and a second reference list, the motion information obtaining unit deciding a motion vector using either a first motion information extracted from a bitstream or a second motion information of a corresponding block on a reference view, the motion information obtaining unit deciding a motion vector using at least one motion information of inter-motion information extracted from the bitstream and intra-motion information extracted from the corresponding block on the reference view and a decoding unit obtaining a reference block using the motion vector, the decoding unit decoding a current block of a current picture using the reference block.
It is to be understood that both the foregoing general description and the following detailed description are exemplary and explanatory and are intended to provide further explanation of the invention as claimed. MODE FOR INVENTION
Reference will now be made in detail to the preferred embodiments of the present invention, examples of which are illustrated in the accompanying drawings.
First of all, motion information in the present invention should be construed as a concept that includes interview motion information as well as temporal motion information. Furthermore, a motion vector should be construed as a concept that includes interview disparity offset as well as temporal motion offset.
In the present invention, it is understood that coding should conceptionally include both encoding and decoding.
Meanwhile, intra-view prediction means prediction between pictures existing within a same view and includes temporal prediction for example. Inter-view prediction means prediction between pictures existing in different views .
FIG. 1 is a schematic block diagram of an apparatus for decoding a video signal according to the present invention. Referring to FIG. 1, a decoding apparatus includes a NAL parser unit 100, an entropy decoding unit 200, an inverse quantization/inverse transform unit 300, an intra prediction unit 400, a deblocking filter unit 500, a decoded picture buffer unit 600 and an inter prediction unit 700. And, the inter prediction unit 700 includes an illumination (IC) offset predictor 710, an illumination compensation unit 720 and a motion compensation unit 730.
The NAL parser unit 100 performs parsing by a unit of NAL (network abstraction layer) to decode a received video sequence. Normally, before a slice header and slice data are decoded, at least one sequence parameter set and picture parameter set are transferred to a decoder. In this case, a NAL header unit or an extension unit of a NAL header can contain various kinds of attribute informations . MVC is a scheme additional to AVC scheme. And, it may be more advantageous to add various kinds of attribute informations for a case of an MVC bitstream only rather tan to add them unconditionally. For instance, it is able to add flag information for identifying whether it is an MVC bitstream in the NAL header unit or the extension unit of the NAL header. Only if a bitstream inputted in accordance with the flag information is a multi-view sequence coded bitstream, it is able to add attribute informations on a multi-view sequence. For instance, the attribute information can include temporal level information, view level information, inter-view picture group identification information, view identification information and the like.
Parsed bitstream is entropy-decoded by the entropy decoding unit 20, and a coefficient, motion vector and the like of each macroblock is extracted. The inverse dequnatization/transform unit 300 obtains a coefficient value transformed by multiplying a received quantized value by a predetermined constant and then reconstructs a pixel value by inverting the coefficient value. The intra prediction unit 400 performs intra-frame prediction from a decoded sample within a current picture using the reconstructed pixel value. Meanwhile, the deblocking filter unit 500 is applied to each coded macroblock to decrease block distortion. In this case, a filter smoothens edges of a block to enhance an image quality of a decoded picture. And, a selection of this filtering process depends on boundary strength and a gradient of an image sample around a boundary. Filtered pictures are outputted or and stored in the decoded picture buffer unit 600 to be used as reference pictures.
The decoded picture buffer unit 600 plays a role in storing or opening previously coded pictures to perform inter- frame prediction. In doing so, to store/open the pictures in/from the decoded picture buffer unit 600, vframe_num" of each picture and POC (picture order count) are used. In MVC, since there exist pictures in a view different from that of a current picture among the previously coded pictures, view information for identifying a view of picture is usable as well as the v frame_num" and the POC to utilize theses pictures as reference pictures. The reference pictures managed in the above manner are usable for the inter prediction unit 700 as well. The inter prediction unit 700 carries out inter-frame prediction using the reference pictures stored in the decoded picture buffer unit 600. It is able to divide inter-coded macroblock into macroblock partitions. And, each of the macroblock partitions can be predicted from one or two reference pictures. The inter prediction unit 700 includes the illumination offset predictor 710, the illumination compensation unit 720 and the motion compensation unit 730.
In case that an inputted bitstream corresponds to a multi-view sequence, each view sequence is a sequence captured by a different camera. So, an illumination difference is generated due to internal and external causes of the cameras. To prevent the illumination difference, the illumination compensation unit 720 carries out illumination compensation. In carrying out the illumination compensation, it is able to use flag information indicating whether to perform illumination compensation on a predetermined layer of a video signal. For instance, it is able to perform illumination compensation using flag information indicating whether to perform illumination compensation on a corresponding slice or macroblock. In performing illumination compensation using the flag information, the illumination compensation is applicable to various macroblock types (e.g., inter 16 x 16 mode, B-skip mode, direct mode) .
The motion compensation unit 730 compensates for a motion of a current block using motion informations transferred from the entropy decoding unit 200. The motion compensation unit 730 extracts motion vectors of blocks neighbor to a current block from a video signal and then obtains a motion vector prediction value of the current block. The motion compensation unit 730 compensates for a motion of the current block using the obtained motion vector prediction value and an offset vector extracted from the video signal. Such motion compensation can be performed using a single reference picture or a plurality of pictures. In multi-view video coding, in case that a current picture refers to pictures in a view different from that of the current picture, motion compensation can be carried out using information on a reference picture list for interview prediction which is stored in the decoded picture buffer unit 600. And, motion compensation can be also carried out using view information for identifying a view of a corresponding picture.
Meanwhile, since motion information is needed to carry out the motion compensation in the above manner, motion information corresponding to a specific block may not be transferred. For this, there exists a scheme called a motion skip mode. In particular, the motion skip mode is a scheme for deriving motion information on a current block from motion information on a block having been fully decoded. For instance, a reference picture index, a motion vector and the like of a current block can be derived from a corresponding block existing in a reference view having motion information similar to that of the current block. In this scheme, since the motion information on the current block is not transferred, it is able to save the number of bits necessary to encode motion information.
Current picture is then reconstructed in a manner that the pictures inter-predicted or intra-predicted through the above process are selected in accordance with a prediction mode. Various embodiments of the motion skip mode are explained as follows.
FIG. 2 is a block diagram of a video signal processing apparatus according to an embodiment of the present invention, and FIG. 3 is a flowchart of a video signal processing method according to an embodiment of the present invention. Referring to FIG. 2, a video signal processing apparatus 1000 according to an embodiment of the present invention includes an information extracting unit 1050, a motion skip determining unit 1100, a reference list obtaining unit 1200, a offset obtaining unit 1250, a corresponding block deciding unit 1300, and a decoding unit 1400. And, a motion information obtaining unit is able to include the reference list obtaining unit 1200, the offset obtaining unit 1250 and the corresponding block deciding unit 1300. In the following description, the embodiment of the present invention is schematically explained with reference to FIG. 2 and FIG. 3.
First of all, the information extracting unit 1050 extracts motion skip allowing flag information and motion skip applying flag information, and the motion skip determining unit 110 then decides whether a current block corresponds to a motion skip mode based on the extracted information [SIlO] . If the current block does not correspond to the motion skip mode, the current block is coded using the motion information extracted by the decoding unit 1400 through a step of extracting motion information on the current block from a bitstream. Meanwhile, the information extracting unit 1050 can be included in the entropy decoding unit 200 described with reference to FIG. 1, which does not put limitation on various implementations of the present invention. Besides, the step SIlO will be explained in detail later with reference to λl. Motion skip determining step'. The reference list obtaining unit 1200 obtains a first reference list and a second reference list based on view dependency [S200] . Details of the view dependency and the reference list will be explained later with reference to λ2.1 Reference list obtaining step'. The offset obtaining unit 1250 obtains offset (e.g., global motion vector, local motion vector, zero vector) necessary to search for a corresponding block [S250] . In this case, various embodiments for an offset type and an offset deriving method will be explained in detail later with reference to '2.2 Offset obtaining step'.
The corresponding block deciding unit 1300 decides a reference view and a corresponding block based on the reference list obtained in the step S200 and the offset obtained in the step S250 [S300] . In this process, various embodiments about which one of at lest one reference views having view dependency on a current view will be preferentially searched, which view will be searched in a second-best way if motion information does not exist in a corresponding block of the searched view, and the like will be explained later with reference to λ3. Reference view and corresponding block deciding step' . Meanwhile, a method of deciding a corresponding block in accordance with whether a corresponding block corresponds to a macroblock level or 8 x 8 block level can be diversified, which will be explained later.
The decoding unit 1400 decodes a current block using the motion information on the corresponding block which is decided in the step S300 [S400] . And, residual information can be used as well as the motion information on the corresponding block, of which details will be explained later with reference to M. Step of decoding a current block using motion information on corresponding block' . Meanwhile, the motion skip determining unit, the reference list obtaining unit 1200, the offset obtaining unit 1250, the corresponding block deciding unit 1300 and the decoding unit 1400 can be included in the motion compensation unit 730 described with reference to FIG. 1, which does not restrict various implementations of the present invention.
1. Motion Skip Determining Step
In the step SlOO, the information extracting unit 1050 extracts motion skip allowing flag information and motion skip applying flag information, and the motion skip determining unit 1100 decides whether a current block corresponds to a motion skip mode based on the extracted information. Sub-steps of the step slOO are shown in FIG. 4.
Referring to FIG. 4, motion skip allowing flag information is firstly extracted from a slice header [SIlO] . In this case, the motion skip allowing flag information is the information indicating whether at least one block corresponding to a motion skip mode exists in blocks belonging to a slice. In other words, the motion skip allowing flag information is the information indicating whether a motion skip mode is allowed for blocks within a slice. A motion skip allowing flag can be included in a slice header, which does not put limitation on various implementations of the present invention. An example of syntax, in which a motion skip allowing flag is included in a slice header, is shown in the following table.
[Table 1] Motion Skip Allowing Flag
Figure imgf000021_0001
In Table 1, Referring to a row indicates by (A) marked on a right side thereof, there exists a syntax element xmotion_skip_enable_flag' . This is a motion skip allowing flag and may include a 1-bit flag having a value set to 0 or 1. And, a meaning of each flag value can be defined as the following table.
[Table 2] Meaning of Motion Skip Allowing Flag
Figure imgf000022_0001
[Modification of Table 1] Motion Skip Allowing Flag
Figure imgf000022_0002
In case that motion skip is not applied to a current slice, it is unnecessary to transfer a motion skip allowing flag. So, as shown in the above table, it is able to add a condition for applying a motion skip. In particular, only if a current slice is a non-anchor picture and it has a reference relation with another view, it is sufficient to transfer a motion skip allowing flag.
Based on the motion skip allowing flag extracted in the step SIlO, it is decided whether a motion skip mode is allowed for a current slice [S120] . If the motion skip mode is not slowed [λno' in the step S120] (e.g., if a motion skip allowing flag motion_skip_enable_flag is 0) , it means that any block corresponding to motion skip mode does not exist in slice at all. Hence, it is decided that a whole corresponding slice does not correspond to the motion skip mode. Motion information on a current block is then extracted at a block level [S150] .
Otherwise, if the motion skip is allowed for the current slice in the step S120 [Hyes' in the step S120] , it is able to execute a step of deciding whether each block within the slice corresponds to the motion skip mode. First of all, motion skip applying flag information is extracted at a block level [S130] . In this case, the motion skip applying flag information is the information indicating whether the current block corresponds to the motion skip mode. An example of extracting a motion skip applying flag from a macroblock layer based on the motion skip allowing flag extracted in the step SIlO is shown in the following table .
[Table 3] Motion Skip Applying Flag
Figure imgf000023_0001
Figure imgf000024_0001
Referring to rows having (Cl) and (F) marked thereon in a right part of Table 3, if a motion skip allowing flag is not 0 [cf. if(... motion_skip_enable_flag) on the row (Cl)], it can be observed that a motion skip applying flag [cf. motion_skip_flag on the row (F)] is extracted. In particular, if the motion skip allowing flag is 0, a motion skip applying flag is not extracted from a macroblock layer. [Modification of Table 3] Modification Motion Skip Applying Flag
Figure imgf000024_0002
Like the above [Modification of table 1] , if a slice header applies a condition for applying a motion skip to a slice, each macroblock just needs to check a motion skip allowing flag only. So, in case that a motion skip allowing flag, as shown in the above table, in a current slice is 1, it is sufficient to transfer a motion skip applying flag for each macroblock.
Based on the motion skip applying flag extracted in the step S130, it is decided whether a current block corresponds to a motion skip mode [S140] . In this case, a motion skip applying flag may include a 1-bit flag having a value set to 0 or 1 as well. Meaning according to each flag value can be defined as the following table.
[Table 4] Meaning of Motion Skip Applying Flag
As a result of the deciding step S140, if the current block does not correspond to the motion skip mode ['no' in the step S140] (e.g., motion skip applying flag motion_skip_flag is 0) , it is determined that it is not the motion skip mode. Motion information on the current block is then extracted from a bitstream [S150] [cf. rows (CCl) and (Pl) in Table 3] .
Otherwise, if the current block corresponds to the motion skip mode [λyes' in the step S140] , the step SlOO is terminated and the steps S200 to S400 are then executed.
2.1 Reference List Obtaining Step
As mentioned in the foregoing description, the first and second reference lists are obtained in the step S200 based on the view dependency. The view dependency means relation between a current view and a different view and may relate to whether a different view is necessary to decode a current view. And, the view dependency can be estimated from information contained in a sequence extension field (SPS MVC extension) .
The step S200 can be executed based on view dependency (view_dependency) of a non-anchor picture. In this case, an anchor picture is a picture for executing inter-view prediction only, refers to a picture within a different view but does not refer to a picture within the same view. A non-anchor picture is a picture not corresponding to the anchor picture and includes a picture capable of referring to both a picture in a same view and a picture in a different view. The view dependency of the non-anchor picture is the relation between non-anchor pictures of a current view and non-anchor pictures of a different view. Since an anchor picture is characterized in executing inter-view prediction only, a motion skip mode is not applied to the anchor picture. And, the motion skip mode is applicable to a non-anchor picture only. Hence, the step S200 can be executed based on the view dependency of the non-anchor picture. Meanwhile, view information having view dependency between a current view and a non-anchor picture (non_anchor) can be divided into view information on a first reference list and view information on a second reference list. For instance, the information can be divided into view information (non_anchor_ref10 [j] [i] , j=current view, i=0, ..., num_ anchor_ref_10-l) on LO- direction reference list and view information (non_anchor__ref11 [j ] [i] , j=current view, i=0, ..., num_anchor_ref_ll-l) on Ll-direction reference list. And, it is able to know a view identifier (view_id) of a different view having view dependency on a current view using the view information.
FIG. 5 is a conceptional diagram to explain view dependency and reference list. Referring to FIG. 5, there exist non-anchor pictures PCi, -/Pen of a current view Vc. A right dotted-line arrow indicates view dependency of a non- anchor picture between the current view Vc and a different view Vc-i, Vc-2, Vc+i or Vc+2. It can be observed that views (non_anchor-ref_10 [Vc] [i] , i=0, 1) having view dependency of the non-anchor picture on the current view Vc in a forward direction are views Vc-i. and VC-2- And, it can be also observed that views (non_anchor-ref_ll [Vc] [i] , i=0, 1) having view dependency of the non-anchor picture on the current view Vc in a backward direction are views Vc+1 and Vc+2- In the case shown in FIG. 5, a first reference view list RLl conceptionally includes the view Vc-i and the view Vc-2 and a second reference view list RL2 conceptionally includes the view Vc+i and the view Vc+2 • This technique is just an embodiment of the case shown in FIG. 5 only. An order of storage in a reference view list is not always defined as order that gets closer to a current view spatially or may be specified by an encoder. A backward view can be contained in the first view list and a forward view can be contained in the second reference view list. Reference picture list reordering (RPLR) operation enables the order to be set different per a slice in a basic list configuration . A first reference block included in a reference view on a same time reference of a current block may be a reference block associated with inter-motion information and a second reference block included in a current view on a same view reference of a current block may e a reference block associated with intra-motion information.
Thus, after the first and second reference view lists have been obtained, a reference view is decided in the step S300 using the obtained first and second reference view lists. 2.2 Offset Obtaining Step
To search for a corresponding block in the step S300, an offset value between a current block and a corresponding block of a reference picture is needed. The offset value can be derived in various ways a) zero vector, b) global motion vector, c) local motion vector, etc. And, a corresponding block can be searched for in various ways using the offset value as well. Various embodiments thereof are explained as follows.
2.2.1 Offset Type and Driving Method Thereof
(1) Zero Vector
It is able to set a zero vector to an offset between a current block and a corresponding block. This corresponds to a case that the corresponding block is set to a co- located block.
(2) Global Motion Vector
Global motion vector or global disparity vector is a value that represents a difference of overall motion or disparity between a current picture and a reference picture and may include an offset corresponding to a slice, a field, a frame, a sequence or the like as well as a picture. If a motion skip mode is to skip motion information in a time direction, a global motion vector becomes a difference of motion or disparity in a view direction. On the contrary, if a motion skip mode is to skip motion information in a view direction, a global motion vector becomes a difference of motion or disparity in a time direction.
It is able to transfer a global motion vector in various ways. The global motion vector can be transferred for each picture or for each anchor picture. And, it is also able to transfer one global motion vector for a whole sequence. In case that a global motion vector is transferred for each anchor, since a motion skip mode is applicable to a non-anchor picture only, it is able to transfer a global motion vector for a view having view dependency of a non-anchor picture only. For instance, if a view identifier of a current view is Vc, if a view having view dependency between non-anchor pictures is Vc-i in a direction LO, and if a view having view dependency between non-anchor pictures is Vc+2 in a direction Ll7 a global motion vector global_disparity_10 [0] (where, VIeWlD[O]=Vo1) can be transferred in the direction LO and a global motion vector global_disparity_ll [1] (where, ViewID [1] =Vc+2) can be transferred in the direction Ll.
Thus, in case that a global motion vector is transferred for each anchor picture, it is able to derive a global motion vector of a non-anchor picture using a global motion vector of a neighbor anchor picture. This can be accomplished in various ways. In case that a current picture is a non-anchor picture, it is able to bring one of global motion vectors of a most nearest neighbor anchor picture as it is. Alternatively, it is able to find a global motion vector by performing interpolation on global motion vectors of two neighbor anchor pictures by considering POC (picture order count) between the current picture and a neighbor anchor picture. And, the present invention enables more implementations thereof.
Resolution of a global motion vector can have resolution of macroblock, 8x8 block, 4x4 block, pixel, subpixel or the like. A reference picture is decided by 8x8 block unit and a motion vector is decided by 4x4 block unit. So, in case that a global motion vector is found by 8 x 8 block level unit, it is able to specify a more detailed corresponding block than that found by macroblock level unit. (3) Local Motion vector
Local motion vector normally has resolution of l/4pel not as an offset corresponding a whole picture but as an offset corresponding to a specific macroblock, a specific 8x8 block or a specific 4x4 block within a picture. In case that neighbor vectors adjacent to a current block have motion vectors (or disparity vectors) in a view direction, it is able to derive a location motion vector of the current block using the motion vectors of the neighbor blocks. In particular, based on whether the neighbor blocks have a same reference picture of the current block, it is able to use motion vectors of the neighbor blocks having the same reference picture of the current picture. Alternatively, by checking whether the reference picture is identical in accordance with a priority by giving the priority to a specific block (e.g., first priority given to a left upper block, second priority to an upper block, etc.), it is able to use the motion vector of the neighbor block having the same reference picture.
2.2.2 Offset Refinement Method
The zero or view-directional motion vector (global motion vector or local motion vector) derived in Clause 2.2.1 may differ from a real motion vector of a current block in a view direction. If the difference is transferred as the following table, a prediction value is set to a value of a derived motion vector and it is then able to obtain a motion vector by adding the difference value to the prediction value. [Table 5] Transfer Difference (Offset) from Prediction of Motion Vector
Figure imgf000033_0001
It is understood that the difference value can be a value of macroblock (16x16) level, 8x8 block level, 4x4 block level, pixel level or subpixel unit.
Thus, it is able to find an offset between a current block and a corresponding block according to various embodiments .
3. Reference View and Corresponding block Deciding Step
Using the reference list and offset obtained in the above process, it is able to decide a reference view and a corresponding block existing in the reference view. In case that motion information exists in a corresponding block of a view, the view becomes a final reference view of a current block and the corresponding block becomes a final corresponding block. Meanwhile, a corresponding block may belong to a picture on a same time zone. If a corresponding block belongs to a picture on a different time instance, it is able to further use POC (picture order count) information.
The information on the reference view can be transferred via a bitstream by an encoder, which is an explicit scheme. The information on the reference view can be arbitrarily decided by a decoder, which is an implicit scheme. The explicit and implicit schemes will be explained in the following description.
Besides, in mapping a current block and a corresponding block to each other, there can exist various embodiments for varying such a level as a macroblock level, an 8x8 block level and the like. This will be explained in the following description as well.
3.1 Scheme for Deciding Reference View and Corresponding block
(1) Explicit Scheme
First of all, it is able to explicitly transfer information indicating which one of views contained in first and second reference lists will be set as a reference view, i.e., a view identifier of a reference view.
Moreover, in case that it is set to check whether a view nearest to a current view among views belonging to each reference list can be set as a reference view, it is unnecessary to explicitly transfer a view identifier of the reference view. Yet, since both of the first and second reference lists can exist in such a case, it is able to explicitly transfer a motion skip reference list flag, which indicates which one of the two will be firstly checked, as shown in the following table.
[Table 6] Reference View Information Transfer
Figure imgf000035_0001
[Modification of Table 6] Reference View Information Transfer
Figure imgf000035_0002
Like [Modification of Table 1] , in case that a motion skip allowing condition is applied to a slice header, it sufficient to confirm motion skip flag information only like the above table.
[Table 7] Meaning of Motion Skip Reference List Flag
Figure imgf000035_0003
Figure imgf000036_0001
(2) Implicit Scheme
First of all, information indicating which one of views contained in first and second reference lists will be set as a reference view may not be transferred. In this case, it is able to decide a final reference view and a final corresponding block by checking whether motion formation exists in a corresponding block of each of the reference views. And, various embodiments can exist for whether to firstly check a reference view belonging to which one of the first and second reference lists. Moreover, various embodiments can exist for, if motion information does not exist in the corresponding reference view, how to continue performing a check in prescribed order. In priorities between reference views belonging to a reference list, first of all, it is able to check reference views included in the first or second reference list in order of a lower index indicating a reference view. In this case, the index indicating the reference view may include a serial number of each reference view set in coding a bitstream by an encoder. For instance, in case of representing a reference view of a non-anchor picture in sequence extension information (SPS extension) as non_anchor_ref_10 [i] or non_anchor_ref_ll [i] , 'i' may be the index indicating the reference view. The encoder is able to give a lower index in order closer to a current view, which does not restrict various implementations of the present invention. In case that an index λi' starts from 0, a reference view of i=0 is checked, a reference view of i=l is checked, and a reference view of i=2 is then checked.
Secondly, it is able to check reference views in order closer to a current view among reference views included in the first or second reference list.
Thirdly, it is able to check reference views in order closer to a base view among reference views included in the first or second reference list.
In priorities between the first and second reference lists, it is able to set reference views belonging to the first reference list to be checked prior to reference views belonging to the second reference list. On the assumption of the above setting, explained in the following description are a case that reference views exist in both of the first and second reference lists and a case that a reference list exists in either the first or second reference list.
(2) -1 Case That Reference Views Exist in Both of Two Reference Lists FIG. 6 is a diagram for one example of a reference list for a current view, and FIG. 7 is a flowchart of one example for sub-steps of a step S300 of deciding a reference view and a corresponding block. Referring to FIG. 6, it can be observed that there exist a first reference list RLl in a direction LO and a second reference list RL2 in a direction Ll with reference to a current view V0 and a current block MBC.
Referring to FIG. 6 and FIG. 7, a view (Vc- x=non_anchor_ref_10 [0] ) having a lowest index indicating a reference view in the first reference list RLl is decided as a first reference view RVl and a block indicated by an offset (GDV_10 [0] ) between the current view Vc and the first reference view RVl is decided as a first corresponding block CBl [S310] . If the first corresponding block CBl is not an intra block, i.e., if motion information exists [λno' in the step S320] , the first corresponding block is finally decided as a corresponding block and then motion information is obtained from the first corresponding block [S332] . On the other hand, in case that a block type of the corresponding block CBl is an intra-frame prediction block
Pyes' in the step S320] , a view (Vc+i=non_anchor_ref_ll [0] ) having a lowest index in the second reference list RL2 is decided as a second reference view RV2 , and a block indicates by an offset (GDV_ll[0]) between the current view
Vc and the second reference view RV2 is decided as a second corresponding block CB2 [S334] . Subsequently, like the formerly described steps S320, S332 and S334, if motion information fails to exist in the second corresponding block CB2, a view (Vc-2=non_anchor_ref_10 [1] ) having a second lowest index in the first reference list is decided as a third reference view RV3 and a view (Vc+2 =non_anchor_ref_ll [1] ) having a second lowest index in he second reference list is decided as a fourth reference view RV4. Thus, third and fourth corresponding blocks CB3 and CB4 are sequentially checked. In particular, by considering an index indicating a reference view, a presence or non- presence of motion information is checked in a manner of alternating the reference views of the first and second reference lists RLl and RL2.
If a view (e.g., non_anchor_ref_10 [i] , non_anchor_ref_ll [i] , i=0) having a lowest index in view dependency information on a current view is a view closest to the current view Vc, a selection reference for candidates (i.e., the first reference view, the second reference view, etc.) of a reference view may follow an order that is closest to the current view Vc. Meanwhile, if a view having a lowest index is a view close to a base view, a candidate selection reference can be the base view or an order that is closest to the base view, which does not restrict various implementations of the present invention.
(2) -2 Case that Reference View Exists in Single Reference List
FIG. 8 is a diagram for another example of a reference list for a current view, and FIG. 9 is a flowchart of another example for sub-steps of a step S300 of determining a reference view and a corresponding block. FIG. 8 and FIG. 9 show a first reference list RLl only. In this case, a view (Vc-i=non_anchor_ref__l0 [0] ) having a lowest index in the first reference list RLl is decided as a first reference view RVl and a first corresponding block CBl, a view (Vc-2=non_anchor_ref_10 [1] ) having a second lowest index is decided as a second reference view RV2 and a second corresponding block CB2, a view having a third lowest index is decided as a third reference view (RV3=non_anchor_ref_10 [2] ) (not shown in the drawing) and a third corresponding block CB3 (not shown in the drawing) . IT is then checked whether a type of each of the corresponding blocks is an intra block or not. Thus, a reference view and a corresponding block capable of using motion information are finally decided.
FIG. 10 is a diagram for a further example of a reference list for a current view, and FIG. 11 is a flowchart of a further example for sub-steps of a step S300 of determining a reference view and a corresponding block. FIG. 10 and FIG. 11 show a second reference list RL2 only. Similar to the case shown in FIG. 8 and FIG. 9, first reference view RVl and first corresponding block CBl, second reference view RV2 and second corresponding block CB2, third reference view RV3 and third corresponding block CB3 and the like are decided in order of a lower index in the second reference list RL2. It is then checked whether a type of each of the corresponding blocks is an intra block in corresponding order. In other words, if a reference view does not exist in the first reference list RLl, a reference view in a direction LO is not tested and block types of the corresponding blocks are checked in order from a view having a lower index among the reference views of the second reference list RL2.
Through the above process, a reference view and a corresponding block to obtain motion information in the step S300 are decided.
3.2 Corresponding block According to Each Mapping Level
To search for a corresponding block in a reference view, it is able to use a local motion vector corresponding to a macroblock (16x16 block) , 8x8 block, 4x4 block, pixel, subpixel or the like as well as a global motion vector. Meanwhile, a unit of a global or local motion vector can be a unit of a macroblock (16x16 block) , 8x8 block, 4x4 block, pixel, subpixel or the like.
In mapping a current block and a corresponding block using the global or local motion vector, various levels can be provided. FIG. 12 is a diagram to explain a case that a mapping level is a macroblock level or 8x8 block level, and FIG. 13 is a diagram to explain a case that a mapping level is an 8x8 block level or a sub-pixel level. A method of deciding a corresponding block in each case is explained as follows.
(1) Macroblock Level
Referring to (A) of FIG. 12, a left upper end point of a block LB having a global or local motion vector applied thereto is marked. It is able to decide a macroblock CB0 closest to the block LB having the global or local motion vector applied thereto as a corresponding block. In particular, all motion information on a corresponding block such as block type information, a reference picture index, a motion vector and the like can be used as motion information on a current block as it is. Yet, in case that a corresponding block is such a skip block as P skip (P_skip) and B skip (B_skip) , it can be reset to an inter 8x8 mode. (2) 8x8 Block Level
Referring to (B) of FIG. 12, 8x8 block CB0 closest to a block LB having a global or local motion vector applied thereto and 8x8 blocks CB1, CB2 and CB3 neighbor to the block are shown. Using the four 8 x 8 blocks as corresponding blocks, it is able to use motion informations of the blocks for decoding of a current block. If each of the four blocks has motion information as an inter block, a type of the current block is decided as "inter 8x8'. If the current block is a macroblock (16x16), it is divided into four 8x8 blocks (MBc0, MBci, MBc2, MBc3) (not shown in the drawing) . Motion information on a corresponding block corresponding to each of the 8x8 blocks is then used. For instance, in case that the current block is MBc2, motion information on a corresponding block CB2 is used. On the other hand, in case that there exists an intra block in four corresponding blocks, it is able to use motion information corresponding to a different inter block.
(3) 4x4 Block Level Referring to (A) of FIG. 13, 4x4 block CB0 closest to a block LB having a global or local motion vector applied thereto, 8x8 block including the 4x4 block CB0 and 8x8 blocks CB1, CB2 and CB3 neighbor to the 8x8 block are shown. Various embodiments can exist as a method of using motion information on these blocks. And, it is able to use motion vector and reference picture index information on a block CB00 closest to the block LB as they are. Alternatively, a reference picture index is derived using index information on a reference picture of the 8x8 block including the 4x4 block CB00 and a motion vector is derived by taking an intermediate or average value of motion vectors of the 4x4 block CB00 and reference blocks CBi0, CB2O and CB30.
Alternatively, it is ale to it is able to map motion information on 4 x 4 blocks CB00, CBi0, CB20 and CB30 to each of the 8x8 blocks of the current block. Alternatively, a smallest value of reference picture indexes of three 4x4 blocks CB0O/ CB10 and CB20 is decided as a reference picture index of the current block and a motion vector of the current block is then derived using motion vectors of the three 4x4 blocks. If motion information on the three 4x4 blocks is unusable, it is able to motion information on the 4x4 block CB30 at a right lower end. And, it is able to set a type of macroblock to xinter_8x8 mode' . (4) Pixel/Subpixel Level
Referring to (B) of FIG. 13, 4x4 block closest to a block LB having a global or local motion vector applied thereto and it s neighbor blocks CB00, CB01, CB02 and CB03 are shown. A reference picture index and a motion vector of the closest 4x4 block are usable as they are or can be derived from reference picture indexes and motion vectors of four 4x4 blocks. In deriving a motion vector, a method of an intermediate or average is usable. And, it is able to set a type of macroblock to vinter__8x8 mode'. In the above-explained levels of the four kinds of cases, each of the per-level mapping methods is independently usable or any combinations thereof are possible. Meanwhile, in case that an initial mapping position is located at a special point such as a corner, an edge, a block center and the like, exceptional processing is necessary. And, such order as left side, right side, upper side, lower side, left upper side, right upper side, left lower side, right lower side and the like can be arbitrarily designated. Besides, processing needs to be performed on an intra block or an exception that a reference index fails to exist on a same view.
4. Step of Decoding Current Block Using Motion Information on Corresponding block
After the reference view and corresponding block have been decided to obtain motion information in the step S300, motion information on a current block is derived using motion information on the corresponding block in the step S400. In case that there exist various corresponding blocks of 16x16 level, 8x8 level and the like, as mentioned in the foregoing description of λ3.2 Corresponding block
According to Each Mapping Level' , there can exist various embodiments. Meanwhile, there can exist a method of deriving residual information and the like as well as motion information on a corresponding block. To raise accuracy of motion information, there can exist a method of transferring a difference between a motion vector value derived by a motion skip scheme and a real value. Theses methods are explained as follows.
4.1 Residual Deriving Method
First of all, a residual of a corresponding block is usable as a residual of a current block as well as motion information on the corresponding block. In particular, a prediction value of a current block is obtained using a reference picture index and motion vector obtained in accordance with a process for a motion skip mode . The current block is then reconstructed by adding the residual of the corresponding block to the obtained prediction value.
4.2 Motion Vector Refinement Method
The motion vector obtained in accordance with the process for the motion skip mode may have a difference from a real motion vector of the current block. If the difference value is transferred via a bitstream, it is able to obtain a more accurate motion vector by adding the difference value to the motion vector derived in accordance with the motion skip mode in the course of decoding. By performing motion compensation suing the obtained motion vector, it is able to raise a reconstruction rate of the current block.
4.3 Residual Skip Method
In case that similarity between motion information on a current block and motion information on a corresponding block is considerably high and if a reconstruction rate after completion of motion compensation of the current block is considerably high, a residual of the current block may almost 0. In this case, it s able to reduce bits rate required for residual coding by not transferring a residual value at all. In case that a residual value is not included in a bitstream, flag information (e.g., residual_coding_flag) indicating whether residual data is transferred can be included in the bitstream. In particular, if the flag information is 0, residual decoding is not performed as well as residual data is not extracted. If the flag information is 1, residual data is extracted and the extracted residual data is then added to a prediction value of a current block. Moreover, the encoding/decoding method of the present invention can be implemented in a program recorded medium as computer-readable codes. The computer-readable media include all kinds of recording devices in which data readable by a computer system are stored. The computer- readable media include ROM, RAM, CD-ROM, magnetic tapes, floppy discs, optical data storage devices, and the like for example and also include carrier-wave type implementations (e.g., transmission via Internet). And, a bitstream produced by the encoding method is stored in a computer-readable recording medium or can be transmitted via wire/wireless communication network.
INDUSTRIAL APPLICABILITY
Accordingly, the present invention is applicable to encoding/decoding a video signal.
While the present invention has been described and illustrated herein with reference to the preferred embodiments thereof, it will be apparent to those skilled in the art that various modifications and variations can be made therein without departing from the spirit and scope of the invention. Thus, it is intended that the present invention covers the modifications and variations of this invention that come within the scope of the appended claims and their equivalents.

Claims

WHAT IS CLAIMED IS:
1. A method of processing a video signal, comprising: obtaining a block type of a first corresponding block within a first reference view included in a reference list; if the block type of the first corresponding block is an intra block, obtaining a block type and motion information of a second corresponding block within a second reference view included in the reference list; and if the block type of the second corresponding block is not the intra block, decoding a current block using the motion information of the second corresponding block.
2. The method of claim 1, wherein the reference list includes a first reference list and a second reference list.
3. The method of claim 2, wherein the first reference list includes a forward reference view, and the second reference list includes a backward reference view.
4. The method of claim 1, wherein the reference list is obtained based on view dependency on a current view.
5. The method of claim 1, wherein the first reference view corresponds to a reference view having a lowest index indicating an order of being included in a bitstream of the video signal among reference views included in the reference list.
6. The method of claim 1, wherein the first reference view corresponds to a reference view closest to a current view among reference views included in the reference list.
7. The method of claim 1, wherein the first corresponding block is obtained based on a global motion vector for the first reference view.
8. The method of claim 1, wherein the first corresponding block belongs to a reference view within a first reference list of the reference list, and the second corresponding block belongs to a reference view within a second reference list of the reference list.
9. The method of claim 1, wherein the second corresponding block is obtained based on a global motion vector for the second reference view.
10. The method of claim 1, wherein the motion information includes a motion vector and a reference picture index.
11. The method of claim 10, the decoding further comprising obtaining a reference picture in an intra-view direction.
12. The method of claim 1, wherein the video signal is received from a broadcast signal.
13. The method of claim 1, wherein the video signal is received on a digital medium.
14. A computer-readable recording medium comprising a program for executing the method of claim 1.
15. An apparatus for processing a video signal, comprising: a motion information obtaining unit obtaining a block type of a first corresponding block within a first reference view included in a reference list, the motion information obtaining unit, if the block type of the first corresponding block is an intra block, the motion information obtaining unit obtaining a block type and motion information of a second corresponding block within a second reference view included in the reference list; and a decoding unit, if the block type of the second corresponding block is not the intra block, the decoding unit decoding a current block using the motion information of the second corresponding block.
16. A method of processing a video signal, comprising: reordering at least one of a first reference list and a second reference list; deciding a motion vector using either a first motion information extracted from a bitstream or a second motion information of a corresponding block on a reference view; deciding a motion vector using at least one motion information of inter-motion information extracted from the bitstream and intra-motion information extracted from the corresponding block on the reference view; obtaining a reference block using the motion vector,- and decoding a current block of a current picture using the reference block.
17. The method of claim 16, wherein if the current picture is an anchor picture, the motion vector uses the inter-motion formation only.
18. The method of claim 17, wherein the reference block is associated with the inter-motion information, and the reference block exists on a same time of the current block in a view different from that of the current block.
19. An apparatus for processing a video signal, comprising: a motion information obtaining unit reordering at least one of a first reference list and a second reference list, the motion information obtaining unit deciding a motion vector using either a first motion information extracted from a bitstream or a second motion information of a corresponding block on a reference view, the motion information obtaining unit deciding a motion vector using at least one motion information of inter-motion information extracted from the bitstream and intra-motion information extracted from the corresponding block on the reference view; and a decoding unit obtaining a reference block using the motion vector, the decoding unit decoding a current block of a current picture using the reference block.
PCT/KR2008/000444 2007-01-24 2008-01-24 A method and an apparatus for processing a video signal WO2008091117A1 (en)

Priority Applications (5)

Application Number Priority Date Filing Date Title
US12/449,062 US8923399B2 (en) 2007-01-24 2008-01-24 Method and an apparatus for processing a video signal
JP2009547172A JP5249242B2 (en) 2007-01-24 2008-01-24 Video signal processing method and apparatus
KR1020097016427A KR20090113281A (en) 2007-01-24 2008-01-24 A method and an apparatus for processing a video signal
EP08704944A EP2123042A4 (en) 2007-01-24 2008-01-24 A method and an apparatus for processing a video signal
CN200880008776A CN101647279A (en) 2007-01-24 2008-01-24 A method and an apparatus for processing a video signal

Applications Claiming Priority (10)

Application Number Priority Date Filing Date Title
US88635007P 2007-01-24 2007-01-24
US60/886,350 2007-01-24
US90958207P 2007-04-02 2007-04-02
US60/909,582 2007-04-02
US90796407P 2007-04-25 2007-04-25
US60/907,964 2007-04-25
US92469307P 2007-05-29 2007-05-29
US60/924,693 2007-05-29
US94797907P 2007-07-04 2007-07-04
US60/947,979 2007-07-04

Publications (1)

Publication Number Publication Date
WO2008091117A1 true WO2008091117A1 (en) 2008-07-31

Family

ID=39644660

Family Applications (1)

Application Number Title Priority Date Filing Date
PCT/KR2008/000444 WO2008091117A1 (en) 2007-01-24 2008-01-24 A method and an apparatus for processing a video signal

Country Status (7)

Country Link
US (1) US8923399B2 (en)
EP (1) EP2123042A4 (en)
JP (1) JP5249242B2 (en)
KR (1) KR20090113281A (en)
CN (1) CN101647279A (en)
TW (1) TWI355205B (en)
WO (1) WO2008091117A1 (en)

Cited By (8)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN101901493A (en) * 2010-06-21 2010-12-01 清华大学 Method and system for multi-view image combined reconstruction based on compression sampling
WO2013073282A1 (en) * 2011-11-17 2013-05-23 Mitsubishi Electric Corporation Method and system for processing multiview videos of scene
US9204163B2 (en) 2011-11-08 2015-12-01 Samsung Electronics Co., Ltd. Method and apparatus for motion vector determination in video encoding or decoding
EP2839639A4 (en) * 2012-04-19 2016-01-06 Intel Corp 3d video coding including depth based disparity vector calibration
US9307239B2 (en) 2011-03-14 2016-04-05 Mediatek Inc. Method and apparatus for derivation of motion vector candidate and motion vector prediction candidate
US9510000B2 (en) 2011-05-10 2016-11-29 Qualcomm Incorporated Offset type and coefficients signaling method for sample adaptive offset
US9762924B2 (en) 2011-07-02 2017-09-12 Samsung Electronics Co., Ltd. Method and apparatus for coding video, and method and apparatus for decoding video accompanied by inter prediction using collocated image
US9781414B2 (en) 2011-03-17 2017-10-03 Hfi Innovation Inc. Method and apparatus for derivation of spatial motion vector candidate and motion vector prediction candidate

Families Citing this family (31)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US8917775B2 (en) * 2007-05-02 2014-12-23 Samsung Electronics Co., Ltd. Method and apparatus for encoding and decoding multi-view video data
KR101560182B1 (en) * 2008-01-07 2015-10-15 삼성전자주식회사 Method and apparatus for multi-view video encoding and method and apparatus for multi-view video decoding
US20090219985A1 (en) * 2008-02-28 2009-09-03 Vasanth Swaminathan Systems and Methods for Processing Multiple Projections of Video Data in a Single Video File
KR101484281B1 (en) * 2010-07-09 2015-01-21 삼성전자주식회사 Method and apparatus for video encoding using block merging, method and apparatus for video decoding using block merging
SI2924995T1 (en) 2010-07-09 2018-10-30 Samsung Electronics Co., Ltd. Method for decoding video by using block merging
JP2012023652A (en) * 2010-07-16 2012-02-02 Sony Corp Image processing device and image processing method
CA2807780C (en) * 2010-09-24 2018-09-18 Panasonic Corporation Image coding method, image decoding method, image coding apparatus, and image decoding apparatus
KR101893559B1 (en) * 2010-12-14 2018-08-31 삼성전자주식회사 Apparatus and method for encoding and decoding multi-view video
CA2833032C (en) * 2011-06-15 2016-08-30 Mediatek Inc. Method and apparatus of texture image compression in 3d video coding
WO2013001749A1 (en) * 2011-06-29 2013-01-03 パナソニック株式会社 Image encoding method, image decoding method, image encoding device, image decoding device, and image encoding/decoding device
US9712819B2 (en) 2011-10-12 2017-07-18 Lg Electronics Inc. Image encoding method and image decoding method
US9357161B1 (en) * 2011-10-17 2016-05-31 Pixelworks, Inc. Motion vector interpolation for motion compensation
US9525861B2 (en) * 2012-03-14 2016-12-20 Qualcomm Incorporated Disparity vector prediction in video coding
US9445076B2 (en) * 2012-03-14 2016-09-13 Qualcomm Incorporated Disparity vector construction method for 3D-HEVC
WO2013144144A1 (en) * 2012-03-30 2013-10-03 Panasonic Corporation Syntax and semantics for adaptive loop filter and sample adaptive offset
US9549180B2 (en) 2012-04-20 2017-01-17 Qualcomm Incorporated Disparity vector generation for inter-view prediction for video coding
US9952840B2 (en) * 2012-05-15 2018-04-24 International Business Machines Corporation Conditional sequential selection of external reference resolutions
US9998726B2 (en) 2012-06-20 2018-06-12 Nokia Technologies Oy Apparatus, a method and a computer program for video coding and decoding
AU2013285749B2 (en) * 2012-07-02 2016-06-16 Samsung Electronics Co., Ltd. Method and apparatus for predicting motion vector for coding video or decoding video
KR101966920B1 (en) 2012-07-10 2019-04-08 삼성전자주식회사 Method and apparatus for estimating motion of image using disparity information of multi view image
US9357195B2 (en) * 2012-08-16 2016-05-31 Qualcomm Incorporated Inter-view predicted motion vector for 3D video
WO2014058210A1 (en) 2012-10-08 2014-04-17 삼성전자 주식회사 Method and apparatus for coding video stream according to inter-layer prediction of multi-view video, and method and apparatus for decoding video stream according to inter-layer prediction of multi-view video
WO2014091933A1 (en) * 2012-12-11 2014-06-19 ソニー株式会社 Encoding device and encoding method, and decoding device and decoding method
US9521389B2 (en) * 2013-03-06 2016-12-13 Qualcomm Incorporated Derived disparity vector in 3D video coding
US9667990B2 (en) 2013-05-31 2017-05-30 Qualcomm Incorporated Parallel derived disparity vector for 3D video coding with neighbor-based disparity vector derivation
WO2015006922A1 (en) * 2013-07-16 2015-01-22 Mediatek Singapore Pte. Ltd. Methods for residual prediction
WO2015006984A1 (en) * 2013-07-19 2015-01-22 Mediatek Singapore Pte. Ltd. Reference view selection for 3d video coding
KR20160072104A (en) * 2013-10-18 2016-06-22 엘지전자 주식회사 Method and apparatus for coding/decoding video comprising multi-view
WO2015100710A1 (en) * 2014-01-02 2015-07-09 Mediatek Singapore Pte. Ltd. Existence of inter-view reference picture and availability of 3dvc coding tools
DK3091741T3 (en) * 2014-01-02 2022-01-31 Intellectual Discovery Co Ltd METHOD FOR DECODING MULTI-VIEW VIDEO
CN110855992A (en) 2014-01-03 2020-02-28 庆熙大学校产学协力团 Method and apparatus for deriving motion information between time points of sub-prediction units

Citations (3)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US20040247159A1 (en) * 2003-06-07 2004-12-09 Niranjan Damera-Venkata Motion estimation for compression of calibrated multi-view image sequences
US20060132610A1 (en) * 2004-12-17 2006-06-22 Jun Xin Multiview video decomposition and encoding
US20060146143A1 (en) * 2004-12-17 2006-07-06 Jun Xin Method and system for managing reference pictures in multiview videos

Family Cites Families (9)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
JP4015934B2 (en) * 2002-04-18 2007-11-28 株式会社東芝 Video coding method and apparatus
CN1312927C (en) * 2002-07-15 2007-04-25 株式会社日立制作所 Moving picture encoding method and decoding method
JP4355914B2 (en) * 2003-10-01 2009-11-04 日本電気株式会社 Multi-view image transmission system and method, multi-view image compression device and method, multi-view image decompression device and method, and program
US8854486B2 (en) * 2004-12-17 2014-10-07 Mitsubishi Electric Research Laboratories, Inc. Method and system for processing multiview videos for view synthesis using skip and direct modes
KR101246915B1 (en) * 2005-04-18 2013-03-25 삼성전자주식회사 Method and apparatus for encoding or decoding moving picture
US8228994B2 (en) * 2005-05-20 2012-07-24 Microsoft Corporation Multi-view video coding based on temporal and view decomposition
KR101276720B1 (en) * 2005-09-29 2013-06-19 삼성전자주식회사 Method for predicting disparity vector using camera parameter, apparatus for encoding and decoding muti-view image using method thereof, and a recording medium having a program to implement thereof
WO2007081177A1 (en) * 2006-01-12 2007-07-19 Lg Electronics Inc. Processing multiview video
KR20080066522A (en) 2007-01-11 2008-07-16 삼성전자주식회사 Method and apparatus for encoding and decoding multi-view image

Patent Citations (3)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US20040247159A1 (en) * 2003-06-07 2004-12-09 Niranjan Damera-Venkata Motion estimation for compression of calibrated multi-view image sequences
US20060132610A1 (en) * 2004-12-17 2006-06-22 Jun Xin Multiview video decomposition and encoding
US20060146143A1 (en) * 2004-12-17 2006-07-06 Jun Xin Method and system for managing reference pictures in multiview videos

Cited By (16)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN101901493A (en) * 2010-06-21 2010-12-01 清华大学 Method and system for multi-view image combined reconstruction based on compression sampling
US9860552B2 (en) 2011-03-14 2018-01-02 Hfi Innovation Inc. Method and apparatus for derivation of motion vector candidate and motion vector prediction candidate
US9307239B2 (en) 2011-03-14 2016-04-05 Mediatek Inc. Method and apparatus for derivation of motion vector candidate and motion vector prediction candidate
US9781414B2 (en) 2011-03-17 2017-10-03 Hfi Innovation Inc. Method and apparatus for derivation of spatial motion vector candidate and motion vector prediction candidate
US9510000B2 (en) 2011-05-10 2016-11-29 Qualcomm Incorporated Offset type and coefficients signaling method for sample adaptive offset
US10034014B2 (en) 2011-07-02 2018-07-24 Samsung Electronics Co., Ltd. Method and apparatus for coding video, and method and apparatus for decoding video accompanied by inter prediction using collocated image
US9762924B2 (en) 2011-07-02 2017-09-12 Samsung Electronics Co., Ltd. Method and apparatus for coding video, and method and apparatus for decoding video accompanied by inter prediction using collocated image
US10397601B2 (en) 2011-07-02 2019-08-27 Samsung Electronics Co., Ltd. Method and apparatus for coding video, and method and apparatus for decoding video accompanied by inter prediction using collocated image
US9225995B2 (en) 2011-11-08 2015-12-29 Samsung Electronics Co., Ltd. Method and apparatus for motion vector determination in video encoding or decoding
US9451282B2 (en) 2011-11-08 2016-09-20 Samsung Electronics Co., Ltd. Method and apparatus for motion vector determination in video encoding or decoding
US9332273B2 (en) 2011-11-08 2016-05-03 Samsung Electronics Co., Ltd. Method and apparatus for motion vector determination in video encoding or decoding
US9204163B2 (en) 2011-11-08 2015-12-01 Samsung Electronics Co., Ltd. Method and apparatus for motion vector determination in video encoding or decoding
WO2013073282A1 (en) * 2011-11-17 2013-05-23 Mitsubishi Electric Corporation Method and system for processing multiview videos of scene
EP2839639A4 (en) * 2012-04-19 2016-01-06 Intel Corp 3d video coding including depth based disparity vector calibration
US9860514B2 (en) 2012-04-19 2018-01-02 Intel Corporation 3D video coding including depth based disparity vector calibration
US9729849B2 (en) 2012-04-19 2017-08-08 Intel Corporation 3D video coding including depth based disparity vector calibration

Also Published As

Publication number Publication date
JP5249242B2 (en) 2013-07-31
CN101647279A (en) 2010-02-10
EP2123042A4 (en) 2010-03-10
US20100091858A1 (en) 2010-04-15
TW200845767A (en) 2008-11-16
TWI355205B (en) 2011-12-21
US8923399B2 (en) 2014-12-30
KR20090113281A (en) 2009-10-29
EP2123042A1 (en) 2009-11-25
JP2010517409A (en) 2010-05-20

Similar Documents

Publication Publication Date Title
US8923399B2 (en) Method and an apparatus for processing a video signal
US11190795B2 (en) Method and an apparatus for processing a video signal
JP5021739B2 (en) Signal processing method and apparatus
US20100111183A1 (en) Method and an apparatus for decording/encording a video signal
EP1982518A1 (en) Processing multiview video
EP2207357A1 (en) Method and apparatus for video coding using large macroblocks
WO2020008328A1 (en) Shape dependent merge mode and amvp mode coding
WO2020139903A1 (en) Video coding with triangular shape prediction units
CN114501019A (en) Partition prediction
KR20080006494A (en) A method and apparatus for decoding a video signal
KR101366289B1 (en) A method and apparatus for decoding/encoding a video signal
KR20080060188A (en) A method and apparatus for decoding a video signal
KR20080029944A (en) A method and apparatus for processing a video signal
WO2009139569A2 (en) Method and apparatus for decoding video signal
KR20080029788A (en) A method and apparatus for decoding a video signal

Legal Events

Date Code Title Description
WWE Wipo information: entry into national phase

Ref document number: 200880008776.0

Country of ref document: CN

121 Ep: the epo has been informed by wipo that ep was designated in this application

Ref document number: 08704944

Country of ref document: EP

Kind code of ref document: A1

ENP Entry into the national phase

Ref document number: 2009547172

Country of ref document: JP

Kind code of ref document: A

NENP Non-entry into the national phase

Ref country code: DE

WWE Wipo information: entry into national phase

Ref document number: 1020097016427

Country of ref document: KR

WWE Wipo information: entry into national phase

Ref document number: 2008704944

Country of ref document: EP

WWE Wipo information: entry into national phase

Ref document number: 12449062

Country of ref document: US